Mark-Arcee commited on
Commit
f1fe70d
1 Parent(s): 4701bc4

Model save

Browse files
README.md CHANGED
@@ -2,15 +2,12 @@
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
  - trl
10
  - sft
 
11
  - generated_from_trainer
12
  datasets:
13
- - arcee-ai/Zilo-Filtered-SQL-Instruct
14
  model-index:
15
  - name: zilo-instruct-v2-sft-filtered
16
  results: []
@@ -21,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # zilo-instruct-v2-sft-filtered
23
 
24
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the arcee-ai/Zilo-Filtered-SQL-Instruct dataset.
25
  It achieves the following results on the evaluation set:
26
- - Loss: 0.4110
27
 
28
  ## Model description
29
 
@@ -53,15 +50,17 @@ The following hyperparameters were used during training:
53
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
  - lr_scheduler_type: cosine
55
  - lr_scheduler_warmup_ratio: 0.1
56
- - num_epochs: 3
57
 
58
  ### Training results
59
 
60
  | Training Loss | Epoch | Step | Validation Loss |
61
  |:-------------:|:-----:|:----:|:---------------:|
62
- | 0.612 | 1.0 | 15 | 0.5663 |
63
- | 0.4089 | 2.0 | 30 | 0.4292 |
64
- | 0.3032 | 3.0 | 45 | 0.4110 |
 
 
65
 
66
 
67
  ### Framework versions
 
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
 
 
 
 
5
  - trl
6
  - sft
7
+ - alignment-handbook
8
  - generated_from_trainer
9
  datasets:
10
+ - generator
11
  model-index:
12
  - name: zilo-instruct-v2-sft-filtered
13
  results: []
 
18
 
19
  # zilo-instruct-v2-sft-filtered
20
 
21
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4709
24
 
25
  ## Model description
26
 
 
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: cosine
52
  - lr_scheduler_warmup_ratio: 0.1
53
+ - num_epochs: 5
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss |
58
  |:-------------:|:-----:|:----:|:---------------:|
59
+ | 0.6543 | 1.0 | 15 | 0.5526 |
60
+ | 0.4164 | 2.0 | 30 | 0.4357 |
61
+ | 0.2956 | 3.0 | 45 | 0.4067 |
62
+ | 0.1871 | 4.0 | 60 | 0.4253 |
63
+ | 0.1363 | 5.0 | 75 | 0.4709 |
64
 
65
 
66
  ### Framework versions
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 3.0,
3
  "eval_loss": 0.4109707474708557,
4
  "eval_runtime": 4.066,
5
  "eval_samples": 2958,
6
  "eval_samples_per_second": 48.942,
7
  "eval_steps_per_second": 0.984,
8
- "total_flos": 18844169011200.0,
9
- "train_loss": 0.5794603930579292,
10
- "train_runtime": 365.3228,
11
  "train_samples": 26621,
12
- "train_samples_per_second": 14.765,
13
- "train_steps_per_second": 0.123
14
  }
 
1
  {
2
+ "epoch": 5.0,
3
  "eval_loss": 0.4109707474708557,
4
  "eval_runtime": 4.066,
5
  "eval_samples": 2958,
6
  "eval_samples_per_second": 48.942,
7
  "eval_steps_per_second": 0.984,
8
+ "total_flos": 31406948352000.0,
9
+ "train_loss": 0.4169128179550171,
10
+ "train_runtime": 607.0886,
11
  "train_samples": 26621,
12
+ "train_samples_per_second": 14.808,
13
+ "train_steps_per_second": 0.124
14
  }
config.json CHANGED
@@ -21,6 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.41.1",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.41.1",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f405459d29d4f24ca925cc753af33355908ad63f30f4385a2e4a85f9d3f7479
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc6da614ec17ae16d1e4117583f2468dde398379987e0368a6a7f30127ec8b8f
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2447be147c0ea3ec8d8acffff86e58e7449d6ef2067e5fe061246f32d036885
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afde27c7acd99cf421df092d50c893e786d8ff5bdcc984b8a89595cb88ca8ada
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9009683f866f59b1167a0e3d05306e477ba5bbb876214564f0458cc1e2d653d4
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f1e80481537111e6e033c6d7390504f3d0769dde8d969cb2f7599a32f7b1c64
3
  size 4540516344
runs/May24_22-46-04_5c31577a2818/events.out.tfevents.1716590804.5c31577a2818.6596.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a11d7ec9e457fb408aab58496acf88e3a86df481d5a61e958ad7811656e9964
3
+ size 6267
runs/May24_22-50-13_5c31577a2818/events.out.tfevents.1716591025.5c31577a2818.7690.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83985bee703a1a9564efc903e6385fa47698b1e1a066a72aba84dbe0eb10bcc0
3
+ size 8182
runs/May24_22-50-13_5c31577a2818/events.out.tfevents.1716591543.5c31577a2818.7690.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eaef88f2128a6d15206f9f6ed81e30a94f464f7eac30ebd23dd4a0809832280
3
+ size 354
runs/May24_23-58-51_5c31577a2818/events.out.tfevents.1716595145.5c31577a2818.11341.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7cd90bece5d909e72588e90ae6f1b13df19a4deeb4221c70a4738e57711a766
3
+ size 9965
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 3.0,
3
- "total_flos": 18844169011200.0,
4
- "train_loss": 0.5794603930579292,
5
- "train_runtime": 365.3228,
6
  "train_samples": 26621,
7
- "train_samples_per_second": 14.765,
8
- "train_steps_per_second": 0.123
9
  }
 
1
  {
2
+ "epoch": 5.0,
3
+ "total_flos": 31406948352000.0,
4
+ "train_loss": 0.4169128179550171,
5
+ "train_runtime": 607.0886,
6
  "train_samples": 26621,
7
+ "train_samples_per_second": 14.808,
8
+ "train_steps_per_second": 0.124
9
  }
trainer_state.json CHANGED
@@ -1,121 +1,179 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.0,
5
  "eval_steps": 500,
6
- "global_step": 45,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.06666666666666667,
13
- "grad_norm": 98.21532030503785,
14
- "learning_rate": 4.000000000000001e-06,
15
  "loss": 1.8595,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.3333333333333333,
20
- "grad_norm": 30.800740243895696,
21
- "learning_rate": 2e-05,
22
- "loss": 1.3856,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.6666666666666666,
27
- "grad_norm": 15.616799282095581,
28
- "learning_rate": 1.9238795325112867e-05,
29
- "loss": 0.8256,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 1.0,
34
- "grad_norm": 4.893885987624207,
35
- "learning_rate": 1.7071067811865477e-05,
36
- "loss": 0.612,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 1.0,
41
- "eval_loss": 0.566252589225769,
42
- "eval_runtime": 4.2539,
43
- "eval_samples_per_second": 46.78,
44
  "eval_steps_per_second": 0.94,
45
  "step": 15
46
  },
47
  {
48
  "epoch": 1.3333333333333333,
49
- "grad_norm": 3.3286961705391014,
50
- "learning_rate": 1.3826834323650899e-05,
51
- "loss": 0.5009,
52
  "step": 20
53
  },
54
  {
55
  "epoch": 1.6666666666666665,
56
- "grad_norm": 1.3920639558462544,
57
- "learning_rate": 1e-05,
58
- "loss": 0.4407,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 2.0,
63
- "grad_norm": 1.3655223364170395,
64
- "learning_rate": 6.173165676349103e-06,
65
- "loss": 0.4089,
66
  "step": 30
67
  },
68
  {
69
  "epoch": 2.0,
70
- "eval_loss": 0.429166316986084,
71
- "eval_runtime": 4.1164,
72
- "eval_samples_per_second": 48.343,
73
- "eval_steps_per_second": 0.972,
74
  "step": 30
75
  },
76
  {
77
  "epoch": 2.3333333333333335,
78
- "grad_norm": 1.2577240969960297,
79
- "learning_rate": 2.9289321881345257e-06,
80
- "loss": 0.3299,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 2.6666666666666665,
85
- "grad_norm": 0.9884829876056015,
86
- "learning_rate": 7.612046748871327e-07,
87
- "loss": 0.3136,
88
  "step": 40
89
  },
90
  {
91
  "epoch": 3.0,
92
- "grad_norm": 1.4077416282701332,
93
- "learning_rate": 0.0,
94
- "loss": 0.3032,
95
  "step": 45
96
  },
97
  {
98
  "epoch": 3.0,
99
- "eval_loss": 0.4109707474708557,
100
- "eval_runtime": 4.1074,
101
- "eval_samples_per_second": 48.449,
102
- "eval_steps_per_second": 0.974,
103
  "step": 45
104
  },
105
  {
106
- "epoch": 3.0,
107
- "step": 45,
108
- "total_flos": 18844169011200.0,
109
- "train_loss": 0.5794603930579292,
110
- "train_runtime": 365.3228,
111
- "train_samples_per_second": 14.765,
112
- "train_steps_per_second": 0.123
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
113
  }
114
  ],
115
  "logging_steps": 5,
116
- "max_steps": 45,
117
  "num_input_tokens_seen": 0,
118
- "num_train_epochs": 3,
119
  "save_steps": 100,
120
  "stateful_callbacks": {
121
  "TrainerControl": {
@@ -129,7 +187,7 @@
129
  "attributes": {}
130
  }
131
  },
132
- "total_flos": 18844169011200.0,
133
  "train_batch_size": 16,
134
  "trial_name": null,
135
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 5.0,
5
  "eval_steps": 500,
6
+ "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.06666666666666667,
13
+ "grad_norm": 98.20704487515755,
14
+ "learning_rate": 2.5e-06,
15
  "loss": 1.8595,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.3333333333333333,
20
+ "grad_norm": 26.158158491928763,
21
+ "learning_rate": 1.25e-05,
22
+ "loss": 1.3331,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.6666666666666666,
27
+ "grad_norm": 52.110102576445215,
28
+ "learning_rate": 1.9956059820218982e-05,
29
+ "loss": 0.7898,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 1.0,
34
+ "grad_norm": 7.630410550452039,
35
+ "learning_rate": 1.9466156752904344e-05,
36
+ "loss": 0.6543,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 1.0,
41
+ "eval_loss": 0.552646815776825,
42
+ "eval_runtime": 4.2535,
43
+ "eval_samples_per_second": 46.785,
44
  "eval_steps_per_second": 0.94,
45
  "step": 15
46
  },
47
  {
48
  "epoch": 1.3333333333333333,
49
+ "grad_norm": 17.50115629098683,
50
+ "learning_rate": 1.8458320592590976e-05,
51
+ "loss": 0.5122,
52
  "step": 20
53
  },
54
  {
55
  "epoch": 1.6666666666666665,
56
+ "grad_norm": 3.153164672454271,
57
+ "learning_rate": 1.698769427778842e-05,
58
+ "loss": 0.4774,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 2.0,
63
+ "grad_norm": 1.7735518447317222,
64
+ "learning_rate": 1.513474193514842e-05,
65
+ "loss": 0.4164,
66
  "step": 30
67
  },
68
  {
69
  "epoch": 2.0,
70
+ "eval_loss": 0.4356851875782013,
71
+ "eval_runtime": 4.105,
72
+ "eval_samples_per_second": 48.477,
73
+ "eval_steps_per_second": 0.974,
74
  "step": 30
75
  },
76
  {
77
  "epoch": 2.3333333333333335,
78
+ "grad_norm": 1.295932258196787,
79
+ "learning_rate": 1.300084635000341e-05,
80
+ "loss": 0.3318,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 2.6666666666666665,
85
+ "grad_norm": 1.0485408353072447,
86
+ "learning_rate": 1.070276188945293e-05,
87
+ "loss": 0.3158,
88
  "step": 40
89
  },
90
  {
91
  "epoch": 3.0,
92
+ "grad_norm": 1.4216337627274278,
93
+ "learning_rate": 8.366226381814698e-06,
94
+ "loss": 0.2956,
95
  "step": 45
96
  },
97
  {
98
  "epoch": 3.0,
99
+ "eval_loss": 0.4067309498786926,
100
+ "eval_runtime": 4.1095,
101
+ "eval_samples_per_second": 48.424,
102
+ "eval_steps_per_second": 0.973,
103
  "step": 45
104
  },
105
  {
106
+ "epoch": 3.3333333333333335,
107
+ "grad_norm": 1.290531056985457,
108
+ "learning_rate": 6.119081473277502e-06,
109
+ "loss": 0.2149,
110
+ "step": 50
111
+ },
112
+ {
113
+ "epoch": 3.6666666666666665,
114
+ "grad_norm": 1.1012108900548532,
115
+ "learning_rate": 4.084277875864776e-06,
116
+ "loss": 0.1992,
117
+ "step": 55
118
+ },
119
+ {
120
+ "epoch": 4.0,
121
+ "grad_norm": 1.1454117876381236,
122
+ "learning_rate": 2.373148218896182e-06,
123
+ "loss": 0.1871,
124
+ "step": 60
125
+ },
126
+ {
127
+ "epoch": 4.0,
128
+ "eval_loss": 0.42530176043510437,
129
+ "eval_runtime": 4.1023,
130
+ "eval_samples_per_second": 48.509,
131
+ "eval_steps_per_second": 0.975,
132
+ "step": 60
133
+ },
134
+ {
135
+ "epoch": 4.333333333333333,
136
+ "grad_norm": 0.947060255649263,
137
+ "learning_rate": 1.0793155744261352e-06,
138
+ "loss": 0.145,
139
+ "step": 65
140
+ },
141
+ {
142
+ "epoch": 4.666666666666667,
143
+ "grad_norm": 1.2233089117487779,
144
+ "learning_rate": 2.735709467518699e-07,
145
+ "loss": 0.1396,
146
+ "step": 70
147
+ },
148
+ {
149
+ "epoch": 5.0,
150
+ "grad_norm": 0.8371390612015485,
151
+ "learning_rate": 0.0,
152
+ "loss": 0.1363,
153
+ "step": 75
154
+ },
155
+ {
156
+ "epoch": 5.0,
157
+ "eval_loss": 0.4708780348300934,
158
+ "eval_runtime": 4.0889,
159
+ "eval_samples_per_second": 48.669,
160
+ "eval_steps_per_second": 0.978,
161
+ "step": 75
162
+ },
163
+ {
164
+ "epoch": 5.0,
165
+ "step": 75,
166
+ "total_flos": 31406948352000.0,
167
+ "train_loss": 0.4169128179550171,
168
+ "train_runtime": 607.0886,
169
+ "train_samples_per_second": 14.808,
170
+ "train_steps_per_second": 0.124
171
  }
172
  ],
173
  "logging_steps": 5,
174
+ "max_steps": 75,
175
  "num_input_tokens_seen": 0,
176
+ "num_train_epochs": 5,
177
  "save_steps": 100,
178
  "stateful_callbacks": {
179
  "TrainerControl": {
 
187
  "attributes": {}
188
  }
189
  },
190
+ "total_flos": 31406948352000.0,
191
  "train_batch_size": 16,
192
  "trial_name": null,
193
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6c1175d266c101e9bb917f7cfdadd36783940e29fb5163f64c8a645a0b647ca
3
- size 6328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcbd9b7605e96941aae059351fb08ef1d90212437053df7ce4e1a3a85ba81b73
3
+ size 6392