End of training
Browse files- README.md +12 -0
- adapter_model.bin +1 -1
- all_results.json +4 -4
- runs/Jul25_12-44-09_a96f9d5e146d/events.out.tfevents.1690289053.a96f9d5e146d.548.2 +3 -0
- runs/Jul25_13-13-45_a96f9d5e146d/events.out.tfevents.1690290829.a96f9d5e146d.548.3 +3 -0
- train_results.json +4 -4
- trainer_state.json +24 -24
- training_args.bin +1 -1
README.md
CHANGED
@@ -4,6 +4,17 @@ library_name: peft
|
|
4 |
## Training procedure
|
5 |
|
6 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
The following `bitsandbytes` quantization config was used during training:
|
8 |
- load_in_8bit: False
|
9 |
- load_in_4bit: True
|
@@ -16,5 +27,6 @@ The following `bitsandbytes` quantization config was used during training:
|
|
16 |
- bnb_4bit_compute_dtype: bfloat16
|
17 |
### Framework versions
|
18 |
|
|
|
19 |
|
20 |
- PEFT 0.5.0.dev0
|
|
|
4 |
## Training procedure
|
5 |
|
6 |
|
7 |
+
The following `bitsandbytes` quantization config was used during training:
|
8 |
+
- load_in_8bit: False
|
9 |
+
- load_in_4bit: True
|
10 |
+
- llm_int8_threshold: 6.0
|
11 |
+
- llm_int8_skip_modules: None
|
12 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
13 |
+
- llm_int8_has_fp16_weight: False
|
14 |
+
- bnb_4bit_quant_type: nf4
|
15 |
+
- bnb_4bit_use_double_quant: True
|
16 |
+
- bnb_4bit_compute_dtype: bfloat16
|
17 |
+
|
18 |
The following `bitsandbytes` quantization config was used during training:
|
19 |
- load_in_8bit: False
|
20 |
- load_in_4bit: True
|
|
|
27 |
- bnb_4bit_compute_dtype: bfloat16
|
28 |
### Framework versions
|
29 |
|
30 |
+
- PEFT 0.5.0.dev0
|
31 |
|
32 |
- PEFT 0.5.0.dev0
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 160069389
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:72430327e78de59be454a580331181e8890a36181da2b084c9534c4fe570c7bc
|
3 |
size 160069389
|
all_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 0.0,
|
3 |
"total_flos": 232910960836608.0,
|
4 |
-
"train_loss": 1.
|
5 |
-
"train_runtime":
|
6 |
-
"train_samples_per_second": 0.
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 0.0,
|
3 |
"total_flos": 232910960836608.0,
|
4 |
+
"train_loss": 1.4252451022466024,
|
5 |
+
"train_runtime": 119.4569,
|
6 |
+
"train_samples_per_second": 0.502,
|
7 |
+
"train_steps_per_second": 0.126
|
8 |
}
|
runs/Jul25_12-44-09_a96f9d5e146d/events.out.tfevents.1690289053.a96f9d5e146d.548.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3f39c5ee5b0075bb15a7fb07d879f5f3dbb9e46af3b1551827bd1ac6e1fe9260
|
3 |
+
size 7043
|
runs/Jul25_13-13-45_a96f9d5e146d/events.out.tfevents.1690290829.a96f9d5e146d.548.3
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:16f93248b75582710c0fc31bc883e563d1484693b1d8335815c75721035c0d6e
|
3 |
+
size 4693
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"epoch": 0.0,
|
3 |
"total_flos": 232910960836608.0,
|
4 |
-
"train_loss": 1.
|
5 |
-
"train_runtime":
|
6 |
-
"train_samples_per_second": 0.
|
7 |
-
"train_steps_per_second": 0.
|
8 |
}
|
|
|
1 |
{
|
2 |
"epoch": 0.0,
|
3 |
"total_flos": 232910960836608.0,
|
4 |
+
"train_loss": 1.4252451022466024,
|
5 |
+
"train_runtime": 119.4569,
|
6 |
+
"train_samples_per_second": 0.502,
|
7 |
+
"train_steps_per_second": 0.126
|
8 |
}
|
trainer_state.json
CHANGED
@@ -22,89 +22,89 @@
|
|
22 |
{
|
23 |
"epoch": 0.0,
|
24 |
"learning_rate": 0.00018461538461538463,
|
25 |
-
"loss": 1.
|
26 |
"step": 3
|
27 |
},
|
28 |
{
|
29 |
"epoch": 0.0,
|
30 |
"learning_rate": 0.00016923076923076923,
|
31 |
-
"loss": 1.
|
32 |
"step": 4
|
33 |
},
|
34 |
{
|
35 |
"epoch": 0.0,
|
36 |
"learning_rate": 0.00015384615384615385,
|
37 |
-
"loss": 1.
|
38 |
"step": 5
|
39 |
},
|
40 |
{
|
41 |
"epoch": 0.0,
|
42 |
"learning_rate": 0.00013846153846153847,
|
43 |
-
"loss": 1.
|
44 |
"step": 6
|
45 |
},
|
46 |
{
|
47 |
"epoch": 0.0,
|
48 |
"learning_rate": 0.0001230769230769231,
|
49 |
-
"loss": 1.
|
50 |
"step": 7
|
51 |
},
|
52 |
{
|
53 |
"epoch": 0.0,
|
54 |
"learning_rate": 0.0001076923076923077,
|
55 |
-
"loss": 1.
|
56 |
"step": 8
|
57 |
},
|
58 |
{
|
59 |
"epoch": 0.0,
|
60 |
-
"learning_rate":
|
61 |
-
"loss": 1.
|
62 |
"step": 9
|
63 |
},
|
64 |
{
|
65 |
"epoch": 0.0,
|
66 |
-
"learning_rate":
|
67 |
-
"loss": 1.
|
68 |
"step": 10
|
69 |
},
|
70 |
{
|
71 |
"epoch": 0.0,
|
72 |
-
"learning_rate":
|
73 |
-
"loss": 1.
|
74 |
"step": 11
|
75 |
},
|
76 |
{
|
77 |
"epoch": 0.0,
|
78 |
-
"learning_rate":
|
79 |
-
"loss": 0.
|
80 |
"step": 12
|
81 |
},
|
82 |
{
|
83 |
"epoch": 0.0,
|
84 |
-
"learning_rate":
|
85 |
-
"loss": 1.
|
86 |
"step": 13
|
87 |
},
|
88 |
{
|
89 |
"epoch": 0.0,
|
90 |
-
"learning_rate":
|
91 |
-
"loss": 1.
|
92 |
"step": 14
|
93 |
},
|
94 |
{
|
95 |
"epoch": 0.0,
|
96 |
-
"learning_rate":
|
97 |
-
"loss": 1.
|
98 |
"step": 15
|
99 |
},
|
100 |
{
|
101 |
"epoch": 0.0,
|
102 |
"step": 15,
|
103 |
"total_flos": 232910960836608.0,
|
104 |
-
"train_loss": 1.
|
105 |
-
"train_runtime":
|
106 |
-
"train_samples_per_second": 0.
|
107 |
-
"train_steps_per_second": 0.
|
108 |
}
|
109 |
],
|
110 |
"max_steps": 15,
|
|
|
22 |
{
|
23 |
"epoch": 0.0,
|
24 |
"learning_rate": 0.00018461538461538463,
|
25 |
+
"loss": 1.638,
|
26 |
"step": 3
|
27 |
},
|
28 |
{
|
29 |
"epoch": 0.0,
|
30 |
"learning_rate": 0.00016923076923076923,
|
31 |
+
"loss": 1.5763,
|
32 |
"step": 4
|
33 |
},
|
34 |
{
|
35 |
"epoch": 0.0,
|
36 |
"learning_rate": 0.00015384615384615385,
|
37 |
+
"loss": 1.5843,
|
38 |
"step": 5
|
39 |
},
|
40 |
{
|
41 |
"epoch": 0.0,
|
42 |
"learning_rate": 0.00013846153846153847,
|
43 |
+
"loss": 1.2616,
|
44 |
"step": 6
|
45 |
},
|
46 |
{
|
47 |
"epoch": 0.0,
|
48 |
"learning_rate": 0.0001230769230769231,
|
49 |
+
"loss": 1.3111,
|
50 |
"step": 7
|
51 |
},
|
52 |
{
|
53 |
"epoch": 0.0,
|
54 |
"learning_rate": 0.0001076923076923077,
|
55 |
+
"loss": 1.3044,
|
56 |
"step": 8
|
57 |
},
|
58 |
{
|
59 |
"epoch": 0.0,
|
60 |
+
"learning_rate": 0.0001076923076923077,
|
61 |
+
"loss": 1.299,
|
62 |
"step": 9
|
63 |
},
|
64 |
{
|
65 |
"epoch": 0.0,
|
66 |
+
"learning_rate": 9.230769230769232e-05,
|
67 |
+
"loss": 1.2492,
|
68 |
"step": 10
|
69 |
},
|
70 |
{
|
71 |
"epoch": 0.0,
|
72 |
+
"learning_rate": 7.692307692307693e-05,
|
73 |
+
"loss": 1.1069,
|
74 |
"step": 11
|
75 |
},
|
76 |
{
|
77 |
"epoch": 0.0,
|
78 |
+
"learning_rate": 6.153846153846155e-05,
|
79 |
+
"loss": 0.7699,
|
80 |
"step": 12
|
81 |
},
|
82 |
{
|
83 |
"epoch": 0.0,
|
84 |
+
"learning_rate": 4.615384615384616e-05,
|
85 |
+
"loss": 1.1393,
|
86 |
"step": 13
|
87 |
},
|
88 |
{
|
89 |
"epoch": 0.0,
|
90 |
+
"learning_rate": 3.0769230769230774e-05,
|
91 |
+
"loss": 1.4198,
|
92 |
"step": 14
|
93 |
},
|
94 |
{
|
95 |
"epoch": 0.0,
|
96 |
+
"learning_rate": 1.5384615384615387e-05,
|
97 |
+
"loss": 1.2558,
|
98 |
"step": 15
|
99 |
},
|
100 |
{
|
101 |
"epoch": 0.0,
|
102 |
"step": 15,
|
103 |
"total_flos": 232910960836608.0,
|
104 |
+
"train_loss": 1.4252451022466024,
|
105 |
+
"train_runtime": 119.4569,
|
106 |
+
"train_samples_per_second": 0.502,
|
107 |
+
"train_steps_per_second": 0.126
|
108 |
}
|
109 |
],
|
110 |
"max_steps": 15,
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3963
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc5ffe380d413e602923828d5dc8c1040fadc1541eda8020df9c5efb964a8927
|
3 |
size 3963
|