AbedFixed commited on
Commit
8e87895
1 Parent(s): 47d4d66

End of training

Browse files
README.md CHANGED
@@ -4,6 +4,17 @@ library_name: peft
4
  ## Training procedure
5
 
6
 
 
 
 
 
 
 
 
 
 
 
 
7
  The following `bitsandbytes` quantization config was used during training:
8
  - load_in_8bit: False
9
  - load_in_4bit: True
@@ -16,5 +27,6 @@ The following `bitsandbytes` quantization config was used during training:
16
  - bnb_4bit_compute_dtype: bfloat16
17
  ### Framework versions
18
 
 
19
 
20
  - PEFT 0.5.0.dev0
 
4
  ## Training procedure
5
 
6
 
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: False
9
+ - load_in_4bit: True
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: nf4
15
+ - bnb_4bit_use_double_quant: True
16
+ - bnb_4bit_compute_dtype: bfloat16
17
+
18
  The following `bitsandbytes` quantization config was used during training:
19
  - load_in_8bit: False
20
  - load_in_4bit: True
 
27
  - bnb_4bit_compute_dtype: bfloat16
28
  ### Framework versions
29
 
30
+ - PEFT 0.5.0.dev0
31
 
32
  - PEFT 0.5.0.dev0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e67d7745d3aeca35ff3b1c6478fffa90ede3bf05c782597b3761b48c46666b7d
3
  size 160069389
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72430327e78de59be454a580331181e8890a36181da2b084c9534c4fe570c7bc
3
  size 160069389
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.0,
3
  "total_flos": 232910960836608.0,
4
- "train_loss": 1.4102739254633585,
5
- "train_runtime": 116.3917,
6
- "train_samples_per_second": 0.516,
7
- "train_steps_per_second": 0.129
8
  }
 
1
  {
2
  "epoch": 0.0,
3
  "total_flos": 232910960836608.0,
4
+ "train_loss": 1.4252451022466024,
5
+ "train_runtime": 119.4569,
6
+ "train_samples_per_second": 0.502,
7
+ "train_steps_per_second": 0.126
8
  }
runs/Jul25_12-44-09_a96f9d5e146d/events.out.tfevents.1690289053.a96f9d5e146d.548.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f39c5ee5b0075bb15a7fb07d879f5f3dbb9e46af3b1551827bd1ac6e1fe9260
3
+ size 7043
runs/Jul25_13-13-45_a96f9d5e146d/events.out.tfevents.1690290829.a96f9d5e146d.548.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16f93248b75582710c0fc31bc883e563d1484693b1d8335815c75721035c0d6e
3
+ size 4693
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.0,
3
  "total_flos": 232910960836608.0,
4
- "train_loss": 1.4102739254633585,
5
- "train_runtime": 116.3917,
6
- "train_samples_per_second": 0.516,
7
- "train_steps_per_second": 0.129
8
  }
 
1
  {
2
  "epoch": 0.0,
3
  "total_flos": 232910960836608.0,
4
+ "train_loss": 1.4252451022466024,
5
+ "train_runtime": 119.4569,
6
+ "train_samples_per_second": 0.502,
7
+ "train_steps_per_second": 0.126
8
  }
trainer_state.json CHANGED
@@ -22,89 +22,89 @@
22
  {
23
  "epoch": 0.0,
24
  "learning_rate": 0.00018461538461538463,
25
- "loss": 1.6373,
26
  "step": 3
27
  },
28
  {
29
  "epoch": 0.0,
30
  "learning_rate": 0.00016923076923076923,
31
- "loss": 1.5748,
32
  "step": 4
33
  },
34
  {
35
  "epoch": 0.0,
36
  "learning_rate": 0.00015384615384615385,
37
- "loss": 1.5775,
38
  "step": 5
39
  },
40
  {
41
  "epoch": 0.0,
42
  "learning_rate": 0.00013846153846153847,
43
- "loss": 1.2566,
44
  "step": 6
45
  },
46
  {
47
  "epoch": 0.0,
48
  "learning_rate": 0.0001230769230769231,
49
- "loss": 1.3078,
50
  "step": 7
51
  },
52
  {
53
  "epoch": 0.0,
54
  "learning_rate": 0.0001076923076923077,
55
- "loss": 1.3014,
56
  "step": 8
57
  },
58
  {
59
  "epoch": 0.0,
60
- "learning_rate": 9.230769230769232e-05,
61
- "loss": 1.2879,
62
  "step": 9
63
  },
64
  {
65
  "epoch": 0.0,
66
- "learning_rate": 7.692307692307693e-05,
67
- "loss": 1.2206,
68
  "step": 10
69
  },
70
  {
71
  "epoch": 0.0,
72
- "learning_rate": 6.153846153846155e-05,
73
- "loss": 1.083,
74
  "step": 11
75
  },
76
  {
77
  "epoch": 0.0,
78
- "learning_rate": 4.615384615384616e-05,
79
- "loss": 0.7552,
80
  "step": 12
81
  },
82
  {
83
  "epoch": 0.0,
84
- "learning_rate": 3.0769230769230774e-05,
85
- "loss": 1.1158,
86
  "step": 13
87
  },
88
  {
89
  "epoch": 0.0,
90
- "learning_rate": 1.5384615384615387e-05,
91
- "loss": 1.3777,
92
  "step": 14
93
  },
94
  {
95
  "epoch": 0.0,
96
- "learning_rate": 0.0,
97
- "loss": 1.1955,
98
  "step": 15
99
  },
100
  {
101
  "epoch": 0.0,
102
  "step": 15,
103
  "total_flos": 232910960836608.0,
104
- "train_loss": 1.4102739254633585,
105
- "train_runtime": 116.3917,
106
- "train_samples_per_second": 0.516,
107
- "train_steps_per_second": 0.129
108
  }
109
  ],
110
  "max_steps": 15,
 
22
  {
23
  "epoch": 0.0,
24
  "learning_rate": 0.00018461538461538463,
25
+ "loss": 1.638,
26
  "step": 3
27
  },
28
  {
29
  "epoch": 0.0,
30
  "learning_rate": 0.00016923076923076923,
31
+ "loss": 1.5763,
32
  "step": 4
33
  },
34
  {
35
  "epoch": 0.0,
36
  "learning_rate": 0.00015384615384615385,
37
+ "loss": 1.5843,
38
  "step": 5
39
  },
40
  {
41
  "epoch": 0.0,
42
  "learning_rate": 0.00013846153846153847,
43
+ "loss": 1.2616,
44
  "step": 6
45
  },
46
  {
47
  "epoch": 0.0,
48
  "learning_rate": 0.0001230769230769231,
49
+ "loss": 1.3111,
50
  "step": 7
51
  },
52
  {
53
  "epoch": 0.0,
54
  "learning_rate": 0.0001076923076923077,
55
+ "loss": 1.3044,
56
  "step": 8
57
  },
58
  {
59
  "epoch": 0.0,
60
+ "learning_rate": 0.0001076923076923077,
61
+ "loss": 1.299,
62
  "step": 9
63
  },
64
  {
65
  "epoch": 0.0,
66
+ "learning_rate": 9.230769230769232e-05,
67
+ "loss": 1.2492,
68
  "step": 10
69
  },
70
  {
71
  "epoch": 0.0,
72
+ "learning_rate": 7.692307692307693e-05,
73
+ "loss": 1.1069,
74
  "step": 11
75
  },
76
  {
77
  "epoch": 0.0,
78
+ "learning_rate": 6.153846153846155e-05,
79
+ "loss": 0.7699,
80
  "step": 12
81
  },
82
  {
83
  "epoch": 0.0,
84
+ "learning_rate": 4.615384615384616e-05,
85
+ "loss": 1.1393,
86
  "step": 13
87
  },
88
  {
89
  "epoch": 0.0,
90
+ "learning_rate": 3.0769230769230774e-05,
91
+ "loss": 1.4198,
92
  "step": 14
93
  },
94
  {
95
  "epoch": 0.0,
96
+ "learning_rate": 1.5384615384615387e-05,
97
+ "loss": 1.2558,
98
  "step": 15
99
  },
100
  {
101
  "epoch": 0.0,
102
  "step": 15,
103
  "total_flos": 232910960836608.0,
104
+ "train_loss": 1.4252451022466024,
105
+ "train_runtime": 119.4569,
106
+ "train_samples_per_second": 0.502,
107
+ "train_steps_per_second": 0.126
108
  }
109
  ],
110
  "max_steps": 15,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e7ca42ee048951bf6929c30c03619fb336d7b3f80bbfeacde887d223d12e670
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc5ffe380d413e602923828d5dc8c1040fadc1541eda8020df9c5efb964a8927
3
  size 3963