viethoangtranduong commited on
Commit
6371b57
1 Parent(s): df53ac2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -44,14 +44,13 @@ to learn more about "Programmatically scale human preferences and alignment in G
44
 
45
 
46
  #### Result:
47
- - This model scored **30.2** on [Alpaca-Eval 2.0](https://tatsu-lab.github.io/alpaca_eval/) - ranked 3rd and the highest for an open source base model at the time of publication.
48
- - Utilizing the model with PairRM, which involved generating 16 responses and submitting the highest-scoring one by PairRM, we scored **34.86** - ranked 2nd.
 
 
 
49
  The best model on the leaderboard is "gpt-4-turbo".
50
 
51
- We acknowledge that Alpaca-Eval 2.0 is not the full reflection of LLMs' performances.
52
- However, in this work, as we are aligning toward general "human preferences", this benchmark serves as a compatible, representative benchmark.
53
- We expect more word on new alignment axes from the community and perform evaluation on other suitable benchmarks.
54
-
55
  We recognize that the Alpaca-Eval 2.0 benchmark does not entirely capture the full range of capabilities and performances of LLMs.
56
  However, in our current work, where the goal is to align with general "human preferences," Alpaca-Eval 2.0 serves as a suitable and representative benchmark.
57
  Moving forward, we anticipate further contributions from the community regarding new alignment axes, and conduct evaluations using other appropriate benchmarks.
@@ -61,7 +60,6 @@ The model is a quick demonstration that the LLMs can be programmatically aligned
61
  It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
62
  make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
63
 
64
-
65
  ## Acknowledgments
66
  - The Mistral AI Team for developing and releasing the advanced Mistral-7B-Instruct-v0.2 model.
67
  - The author of the [Direct Preference Optimization paper](https://arxiv.org/abs/2305.18290) for the innovative approach
 
44
 
45
 
46
  #### Result:
47
+ On [**Alpaca-Eval 2.0**](https://tatsu-lab.github.io/alpaca_eval/):
48
+ - The base model: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) scored **14.72**.
49
+ After applying the above methodology:
50
+ - This model scored **30.2** - ranked 3rd and the highest for an open-source base model at the time of publication.
51
+ - Utilizing the model with PairRM, which involved generating 16 responses and submitting the highest-scoring response by PairRM, we scored **34.86** - ranked 2nd.
52
  The best model on the leaderboard is "gpt-4-turbo".
53
 
 
 
 
 
54
  We recognize that the Alpaca-Eval 2.0 benchmark does not entirely capture the full range of capabilities and performances of LLMs.
55
  However, in our current work, where the goal is to align with general "human preferences," Alpaca-Eval 2.0 serves as a suitable and representative benchmark.
56
  Moving forward, we anticipate further contributions from the community regarding new alignment axes, and conduct evaluations using other appropriate benchmarks.
 
60
  It does not have any moderation mechanisms. We're looking forward to engaging with the community on ways to
61
  make the model finely respect guardrails, allowing for deployment in environments requiring moderated outputs.
62
 
 
63
  ## Acknowledgments
64
  - The Mistral AI Team for developing and releasing the advanced Mistral-7B-Instruct-v0.2 model.
65
  - The author of the [Direct Preference Optimization paper](https://arxiv.org/abs/2305.18290) for the innovative approach