viethoangtranduong
commited on
Commit
•
6371b57
1
Parent(s):
df53ac2
Update README.md
Browse files
README.md
CHANGED
@@ -44,14 +44,13 @@ to learn more about "Programmatically scale human preferences and alignment in G
|
|
44 |
|
45 |
|
46 |
#### Result:
|
47 |
-
|
48 |
-
-
|
|
|
|
|
|
|
49 |
The best model on the leaderboard is "gpt-4-turbo".
|
50 |
|
51 |
-
We acknowledge that Alpaca-Eval 2.0 is not the full reflection of LLMs' performances.
|
52 |
-
However, in this work, as we are aligning toward general "human preferences", this benchmark serves as a compatible, representative benchmark.
|
53 |
-
We expect more word on new alignment axes from the community and perform evaluation on other suitable benchmarks.
|
54 |
-
|
55 |
We recognize that the Alpaca-Eval 2.0 benchmark does not entirely capture the full range of capabilities and performances of LLMs.
|
56 |
However, in our current work, where the goal is to align with general "human preferences," Alpaca-Eval 2.0 serves as a suitable and representative benchmark.
|
57 |
Moving forward, we anticipate further contributions from the community regarding new alignment axes, and conduct evaluations using other appropriate benchmarks.
|
@@ -61,7 +60,6 @@ The model is a quick demonstration that the LLMs can be programmatically aligned
|
|
61 |
It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
|
62 |
make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
|
63 |
|
64 |
-
|
65 |
## Acknowledgments
|
66 |
- The Mistral AI Team for developing and releasing the advanced Mistral-7B-Instruct-v0.2 model.
|
67 |
- The author of the [Direct Preference Optimization paper](https://arxiv.org/abs/2305.18290) for the innovative approach
|
|
|
44 |
|
45 |
|
46 |
#### Result:
|
47 |
+
On [**Alpaca-Eval 2.0**](https://tatsu-lab.github.io/alpaca_eval/):
|
48 |
+
- The base model: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) scored **14.72**.
|
49 |
+
After applying the above methodology:
|
50 |
+
- This model scored **30.2** - ranked 3rd and the highest for an open-source base model at the time of publication.
|
51 |
+
- Utilizing the model with PairRM, which involved generating 16 responses and submitting the highest-scoring response by PairRM, we scored **34.86** - ranked 2nd.
|
52 |
The best model on the leaderboard is "gpt-4-turbo".
|
53 |
|
|
|
|
|
|
|
|
|
54 |
We recognize that the Alpaca-Eval 2.0 benchmark does not entirely capture the full range of capabilities and performances of LLMs.
|
55 |
However, in our current work, where the goal is to align with general "human preferences," Alpaca-Eval 2.0 serves as a suitable and representative benchmark.
|
56 |
Moving forward, we anticipate further contributions from the community regarding new alignment axes, and conduct evaluations using other appropriate benchmarks.
|
|
|
60 |
It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
|
61 |
make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
|
62 |
|
|
|
63 |
## Acknowledgments
|
64 |
- The Mistral AI Team for developing and releasing the advanced Mistral-7B-Instruct-v0.2 model.
|
65 |
- The author of the [Direct Preference Optimization paper](https://arxiv.org/abs/2305.18290) for the innovative approach
|