Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ This model card contains the information and the results of our released Bailong
|
|
22 |
- **Bailong-instruct 7B:** Bailong-instruct 7B is the fine-tuned version of Bailong 7B optimized for multi-turn dialogue use cases. Similar to secondary pretraining stage, we use QLoRA to fine-tune the model. To facilitate the development and communication within the research community in Traditional Chinese NLP, we decide to release this model on Hugging Face.
|
23 |
- **Bailong-bench:** Most existing language models claiming to support Traditional Chinese are adapted from continuously pre-trained open-source models, primarily trained on English data. In certain cases, models fine-tuned with instructions using this approach may respond to Traditional Chinese instructions in English and vice versa. This could pose a significant problem when deploying the model for real-world applications. Consequently, it is essential to have a benchmark dataset specifically designed to assess a model's proficiency in following both English and Traditional Chinese instructions. To address this issue, we propose Bailong-bench, a benchmark dataset crafted not only to evaluate the model's performance in various real-world application scenarios but also to assess its ability to maintain language consistency.
|
24 |
- **Technical report:** In our [technical report](https://arxiv.org/abs/2404.00862), we document the model training process and the details regarding the sources of training data.
|
25 |
-
- **Bailong-orpo 7B:**
|
26 |
algorithm, [ORPO](https://arxiv.org/abs/2403.07691), we further fine-tune Bailong-instruct 7B with 180k preference pair data to derive Bailong-orpo 7B. We also provide f16 GGUF version of Bailong-orpo 7B for efficient inference and storage purposes.
|
27 |
|
28 |
## Model information
|
|
|
22 |
- **Bailong-instruct 7B:** Bailong-instruct 7B is the fine-tuned version of Bailong 7B optimized for multi-turn dialogue use cases. Similar to secondary pretraining stage, we use QLoRA to fine-tune the model. To facilitate the development and communication within the research community in Traditional Chinese NLP, we decide to release this model on Hugging Face.
|
23 |
- **Bailong-bench:** Most existing language models claiming to support Traditional Chinese are adapted from continuously pre-trained open-source models, primarily trained on English data. In certain cases, models fine-tuned with instructions using this approach may respond to Traditional Chinese instructions in English and vice versa. This could pose a significant problem when deploying the model for real-world applications. Consequently, it is essential to have a benchmark dataset specifically designed to assess a model's proficiency in following both English and Traditional Chinese instructions. To address this issue, we propose Bailong-bench, a benchmark dataset crafted not only to evaluate the model's performance in various real-world application scenarios but also to assess its ability to maintain language consistency.
|
24 |
- **Technical report:** In our [technical report](https://arxiv.org/abs/2404.00862), we document the model training process and the details regarding the sources of training data.
|
25 |
+
- **Bailong-orpo 7B:** By leveraging monolithic odds ratio preference optimization
|
26 |
algorithm, [ORPO](https://arxiv.org/abs/2403.07691), we further fine-tune Bailong-instruct 7B with 180k preference pair data to derive Bailong-orpo 7B. We also provide f16 GGUF version of Bailong-orpo 7B for efficient inference and storage purposes.
|
27 |
|
28 |
## Model information
|