mkopecki commited on
Commit
d0bd4d5
1 Parent(s): a014297

End of training

Browse files
Files changed (3) hide show
  1. README.md +2 -2
  2. adapter_model.safetensors +1 -1
  3. trainer_state.json +8 -8
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: meta-llama/Meta-Llama-3-8B
3
  library_name: peft
4
  license: llama3
5
  tags:
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # chess-sft-lora-adapter
18
 
19
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on the None dataset.
20
 
21
  ## Model description
22
 
 
1
  ---
2
+ base_model: meta-llama/Meta-Llama-3-8B-Instruct
3
  library_name: peft
4
  license: llama3
5
  tags:
 
16
 
17
  # chess-sft-lora-adapter
18
 
19
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the None dataset.
20
 
21
  ## Model description
22
 
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac9fe8124dab0a7b6cbefd18116896f74607adc3059aa9292828dd90fb3803cc
3
  size 5544997664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8436da1d13f39a78aec058c18a651032c7af8f7fed9ee51385a8f4c98f719644
3
  size 5544997664
trainer_state.json CHANGED
@@ -10,19 +10,19 @@
10
  "log_history": [
11
  {
12
  "epoch": 2.9940119760479043,
13
- "grad_norm": 2.0264816284179688,
14
  "learning_rate": 9.98003992015968e-08,
15
- "loss": 0.9075,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 3.0,
20
  "step": 501,
21
- "total_flos": 2.157617842277581e+16,
22
- "train_loss": 0.9071252518547271,
23
- "train_runtime": 914.2434,
24
- "train_samples_per_second": 2.189,
25
- "train_steps_per_second": 0.548
26
  }
27
  ],
28
  "logging_steps": 500,
@@ -42,7 +42,7 @@
42
  "attributes": {}
43
  }
44
  },
45
- "total_flos": 2.157617842277581e+16,
46
  "train_batch_size": 1,
47
  "trial_name": null,
48
  "trial_params": null
 
10
  "log_history": [
11
  {
12
  "epoch": 2.9940119760479043,
13
+ "grad_norm": 0.9438353180885315,
14
  "learning_rate": 9.98003992015968e-08,
15
+ "loss": 1.0378,
16
  "step": 500
17
  },
18
  {
19
  "epoch": 3.0,
20
  "step": 501,
21
+ "total_flos": 2.6036531965722624e+16,
22
+ "train_loss": 1.0371553612326434,
23
+ "train_runtime": 1237.191,
24
+ "train_samples_per_second": 1.617,
25
+ "train_steps_per_second": 0.405
26
  }
27
  ],
28
  "logging_steps": 500,
 
42
  "attributes": {}
43
  }
44
  },
45
+ "total_flos": 2.6036531965722624e+16,
46
  "train_batch_size": 1,
47
  "trial_name": null,
48
  "trial_params": null