loubnabnl HF staff commited on
Commit
399e30b
1 Parent(s): 52379bc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -6,7 +6,7 @@ language:
6
  ---
7
 
8
 
9
- # SmolLM
10
 
11
  <center>
12
  <img src="https://huggingface.co/datasets/HuggingFaceTB/images/resolve/main/banner_smol.png" alt="SmolLM" width="1100" height="600">
@@ -27,6 +27,8 @@ SmolLM is a series of state-of-the-art small language models available in three
27
  To build SmolLM-Instruct, we instruction tuned the models using publicly available permissive instruction datasets. We trained all three models for one epoch on the permissive subset of the WebInstructSub dataset, combined with StarCoder2-Self-OSS-Instruct. Following this, we performed DPO (Direct Preference Optimization) for one epoch: using HelpSteer for the 135M and 1.7B models, and argilla/dpo-mix-7k for the 360M model. We followed the training parameters from the Zephyr-Gemma recipe in the alignment handbook, but adjusted the SFT (Supervised Fine-Tuning) learning rate to 3e-4.
28
  [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)
29
 
 
 
30
  ### Generation
31
  ```bash
32
  pip install transformers
 
6
  ---
7
 
8
 
9
+ # SmolLM-Instruct
10
 
11
  <center>
12
  <img src="https://huggingface.co/datasets/HuggingFaceTB/images/resolve/main/banner_smol.png" alt="SmolLM" width="1100" height="600">
 
27
  To build SmolLM-Instruct, we instruction tuned the models using publicly available permissive instruction datasets. We trained all three models for one epoch on the permissive subset of the WebInstructSub dataset, combined with StarCoder2-Self-OSS-Instruct. Following this, we performed DPO (Direct Preference Optimization) for one epoch: using HelpSteer for the 135M and 1.7B models, and argilla/dpo-mix-7k for the 360M model. We followed the training parameters from the Zephyr-Gemma recipe in the alignment handbook, but adjusted the SFT (Supervised Fine-Tuning) learning rate to 3e-4.
28
  [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)
29
 
30
+ This is the SmolLM-135M-Instruct.
31
+
32
  ### Generation
33
  ```bash
34
  pip install transformers