Update README.md
Browse files
README.md
CHANGED
@@ -30,8 +30,8 @@ They show the performance gain that OLMo base models can achieve with existing f
|
|
30 |
We release two adapted model versions:
|
31 |
| Model | Training Method(s) | Datasets | Context Length |
|
32 |
|------|--------|---------|--|
|
33 |
-
| [OLMo 7B July 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) |
|
34 |
-
| [OLMo 7B July 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) |
|
35 |
|
36 |
These models are both trained on top of OLMo 7b July 2024:
|
37 |
| Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |
|
|
|
30 |
We release two adapted model versions:
|
31 |
| Model | Training Method(s) | Datasets | Context Length |
|
32 |
|------|--------|---------|--|
|
33 |
+
| [OLMo 7B July 2024 SFT](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-SFT-hf) | SFT | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) | 4096 |
|
34 |
+
| [OLMo 7B July 2024 Instruct](https://huggingface.co/allenai/OLMo-1.7-7B-Nitro-Instruct-hf) | SFT + DPO | [Tulu 2 SFT Mix](https://huggingface.co/datasets/allenai/tulu-v2-sft-mixture-olmo-4096) + [Ultrafeedback Cleaned](https://huggingface.co/datasets/allenai/ultrafeedback_binarized_cleaned) | 4096 |
|
35 |
|
36 |
These models are both trained on top of OLMo 7b July 2024:
|
37 |
| Size | Training Tokens | Layers | Hidden Size | Attention Heads | Context Length |
|