Update README.md
Browse files
README.md
CHANGED
@@ -21,16 +21,17 @@ base_model:
|
|
21 |
|
22 |
|
23 |
# **Introduction**
|
|
|
24 |
|
25 |
-
We
|
26 |
|
27 |
-
We developed the Depth Up-Scaling technique. Built on the Llama2 architecture, SOLAR-10.7B incorporates the innovative Upstage Depth Up-Scaling. We then integrated Mistral 7B weights into the upscaled layers, and finally, continued pre-training for the entire model.
|
28 |
|
29 |
-
|
30 |
Solar 10.7B is an ideal choice for fine-tuning. SOLAR-10.7B offers robustness and adaptability for your fine-tuning needs. Our simple instruction fine-tuning using the SOLAR-10.7B pre-trained model yields significant performance improvements.
|
31 |
|
32 |
For full details of this model please read our [paper](https://arxiv.org/abs/2312.15166).
|
33 |
|
|
|
34 |
# **Instruction Fine-Tuning Strategy**
|
35 |
|
36 |
We utilize state-of-the-art instruction fine-tuning methods including supervised fine-tuning (SFT) and direct preference optimization (DPO) [1].
|
|
|
21 |
|
22 |
|
23 |
# **Introduction**
|
24 |
+
We introduce SOLAR-10.7B, an advanced large language model (LLM) with 10.7 billion parameters, demonstrating superior performance in various natural language processing (NLP) tasks. It's compact, yet remarkably powerful, and demonstrates unparalleled state-of-the-art performance in models with parameters under 30B.
|
25 |
|
26 |
+
We present a methodology for scaling LLMs called depth up-scaling (DUS) , which encompasses architectural modifications and continued pretraining. In other words, we integrated Mistral 7B weights into the upscaled layers, and finally, continued pre-training for the entire model.
|
27 |
|
|
|
28 |
|
29 |
+
SOLAR-10.7B has remarkable performance. It outperforms models with up to 30B parameters, even surpassing the recent Mixtral 8X7B model. For detailed information, please refer to the experimental table.
|
30 |
Solar 10.7B is an ideal choice for fine-tuning. SOLAR-10.7B offers robustness and adaptability for your fine-tuning needs. Our simple instruction fine-tuning using the SOLAR-10.7B pre-trained model yields significant performance improvements.
|
31 |
|
32 |
For full details of this model please read our [paper](https://arxiv.org/abs/2312.15166).
|
33 |
|
34 |
+
|
35 |
# **Instruction Fine-Tuning Strategy**
|
36 |
|
37 |
We utilize state-of-the-art instruction fine-tuning methods including supervised fine-tuning (SFT) and direct preference optimization (DPO) [1].
|