End of training
Browse files- README.md +2 -2
- adapter_model.safetensors +1 -1
- trainer_state.json +8 -8
README.md
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
---
|
2 |
-
base_model: meta-llama/Meta-Llama-3-8B
|
3 |
library_name: peft
|
4 |
license: llama3
|
5 |
tags:
|
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
16 |
|
17 |
# chess-sft-lora-adapter
|
18 |
|
19 |
-
This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the None dataset.
|
20 |
|
21 |
## Model description
|
22 |
|
|
|
1 |
---
|
2 |
+
base_model: meta-llama/Meta-Llama-3-8B-Instruct
|
3 |
library_name: peft
|
4 |
license: llama3
|
5 |
tags:
|
|
|
16 |
|
17 |
# chess-sft-lora-adapter
|
18 |
|
19 |
+
This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the None dataset.
|
20 |
|
21 |
## Model description
|
22 |
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5544997664
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8436da1d13f39a78aec058c18a651032c7af8f7fed9ee51385a8f4c98f719644
|
3 |
size 5544997664
|
trainer_state.json
CHANGED
@@ -10,19 +10,19 @@
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 2.9940119760479043,
|
13 |
-
"grad_norm":
|
14 |
"learning_rate": 9.98003992015968e-08,
|
15 |
-
"loss":
|
16 |
"step": 500
|
17 |
},
|
18 |
{
|
19 |
"epoch": 3.0,
|
20 |
"step": 501,
|
21 |
-
"total_flos": 2.
|
22 |
-
"train_loss":
|
23 |
-
"train_runtime":
|
24 |
-
"train_samples_per_second":
|
25 |
-
"train_steps_per_second": 0.
|
26 |
}
|
27 |
],
|
28 |
"logging_steps": 500,
|
@@ -42,7 +42,7 @@
|
|
42 |
"attributes": {}
|
43 |
}
|
44 |
},
|
45 |
-
"total_flos": 2.
|
46 |
"train_batch_size": 1,
|
47 |
"trial_name": null,
|
48 |
"trial_params": null
|
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 2.9940119760479043,
|
13 |
+
"grad_norm": 0.9438353180885315,
|
14 |
"learning_rate": 9.98003992015968e-08,
|
15 |
+
"loss": 1.0378,
|
16 |
"step": 500
|
17 |
},
|
18 |
{
|
19 |
"epoch": 3.0,
|
20 |
"step": 501,
|
21 |
+
"total_flos": 2.6036531965722624e+16,
|
22 |
+
"train_loss": 1.0371553612326434,
|
23 |
+
"train_runtime": 1237.191,
|
24 |
+
"train_samples_per_second": 1.617,
|
25 |
+
"train_steps_per_second": 0.405
|
26 |
}
|
27 |
],
|
28 |
"logging_steps": 500,
|
|
|
42 |
"attributes": {}
|
43 |
}
|
44 |
},
|
45 |
+
"total_flos": 2.6036531965722624e+16,
|
46 |
"train_batch_size": 1,
|
47 |
"trial_name": null,
|
48 |
"trial_params": null
|