Update README.md
Browse files
README.md
CHANGED
@@ -30,7 +30,7 @@ This model is based on Llama-3-70b, and is governed by [META LLAMA 3 COMMUNITY L
|
|
30 |
|
31 |
The base model has 8k context, and the qLoRA fine-tuning was with 8k sequence length.
|
32 |
|
33 |
-
It took 2.5 days on
|
34 |
|
35 |
This model was trained FFT on all parameters, using ChatML prompt template format.
|
36 |
|
|
|
30 |
|
31 |
The base model has 8k context, and the qLoRA fine-tuning was with 8k sequence length.
|
32 |
|
33 |
+
It took 2.5 days on 8xH100 node provided by Crusoe Cloud
|
34 |
|
35 |
This model was trained FFT on all parameters, using ChatML prompt template format.
|
36 |
|