Mark-Arcee commited on
Commit
99f4d33
1 Parent(s): 429d938

Model save

Browse files
README.md CHANGED
@@ -2,16 +2,12 @@
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
5
- - alignment-handbook
6
- - trl
7
- - sft
8
- - generated_from_trainer
9
  - trl
10
  - sft
11
  - alignment-handbook
12
  - generated_from_trainer
13
  datasets:
14
- - arcee-ai/Zilo-Filtered-SQL-Instruct
15
  model-index:
16
  - name: zilo-instruct-v2-sft-filtered
17
  results: []
@@ -22,9 +18,9 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  # zilo-instruct-v2-sft-filtered
24
 
25
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the arcee-ai/Zilo-Filtered-SQL-Instruct dataset.
26
  It achieves the following results on the evaluation set:
27
- - Loss: 0.4709
28
 
29
  ## Model description
30
 
@@ -54,17 +50,15 @@ The following hyperparameters were used during training:
54
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
  - lr_scheduler_type: cosine
56
  - lr_scheduler_warmup_ratio: 0.1
57
- - num_epochs: 5
58
 
59
  ### Training results
60
 
61
  | Training Loss | Epoch | Step | Validation Loss |
62
  |:-------------:|:-----:|:----:|:---------------:|
63
- | 0.6543 | 1.0 | 15 | 0.5526 |
64
- | 0.4164 | 2.0 | 30 | 0.4357 |
65
- | 0.2956 | 3.0 | 45 | 0.4067 |
66
- | 0.1871 | 4.0 | 60 | 0.4253 |
67
- | 0.1363 | 5.0 | 75 | 0.4709 |
68
 
69
 
70
  ### Framework versions
 
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
 
 
 
 
5
  - trl
6
  - sft
7
  - alignment-handbook
8
  - generated_from_trainer
9
  datasets:
10
+ - generator
11
  model-index:
12
  - name: zilo-instruct-v2-sft-filtered
13
  results: []
 
18
 
19
  # zilo-instruct-v2-sft-filtered
20
 
21
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4072
24
 
25
  ## Model description
26
 
 
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: cosine
52
  - lr_scheduler_warmup_ratio: 0.1
53
+ - num_epochs: 3
54
 
55
  ### Training results
56
 
57
  | Training Loss | Epoch | Step | Validation Loss |
58
  |:-------------:|:-----:|:----:|:---------------:|
59
+ | 0.6043 | 1.0 | 15 | 0.5420 |
60
+ | 0.3922 | 2.0 | 30 | 0.4157 |
61
+ | 0.2791 | 3.0 | 45 | 0.4072 |
 
 
62
 
63
 
64
  ### Framework versions
all_results.json CHANGED
@@ -1,14 +1,14 @@
1
  {
2
- "epoch": 5.0,
3
  "eval_loss": 0.4708780348300934,
4
  "eval_runtime": 4.0651,
5
  "eval_samples": 2958,
6
  "eval_samples_per_second": 48.953,
7
  "eval_steps_per_second": 0.984,
8
- "total_flos": 31406948352000.0,
9
- "train_loss": 0.4169128179550171,
10
- "train_runtime": 607.0886,
11
  "train_samples": 26621,
12
- "train_samples_per_second": 14.808,
13
  "train_steps_per_second": 0.124
14
  }
 
1
  {
2
+ "epoch": 3.0,
3
  "eval_loss": 0.4708780348300934,
4
  "eval_runtime": 4.0651,
5
  "eval_samples": 2958,
6
  "eval_samples_per_second": 48.953,
7
  "eval_steps_per_second": 0.984,
8
+ "total_flos": 18844169011200.0,
9
+ "train_loss": 0.5649417373869154,
10
+ "train_runtime": 363.236,
11
  "train_samples": 26621,
12
+ "train_samples_per_second": 14.85,
13
  "train_steps_per_second": 0.124
14
  }
config.json CHANGED
@@ -21,6 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.41.1",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.41.1",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc6da614ec17ae16d1e4117583f2468dde398379987e0368a6a7f30127ec8b8f
3
  size 4943162336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce24b378ae43fe02b92e39dd4d0181fc4821e81b76e1ab3b48feb485cee8ad9a
3
  size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afde27c7acd99cf421df092d50c893e786d8ff5bdcc984b8a89595cb88ca8ada
3
  size 4999819336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:767652637ec5730e9771e3f2929b36cf446fdb16382700a907dde817a6c9b649
3
  size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f1e80481537111e6e033c6d7390504f3d0769dde8d969cb2f7599a32f7b1c64
3
  size 4540516344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90277adea4cd1317fb6624f070f075f16342cb8b18b00043f18cb5db49b56010
3
  size 4540516344
runs/May25_01-05-52_5c31577a2818/events.out.tfevents.1716599166.5c31577a2818.16226.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d42f23ba895e8799c2c78e19a872f0b213a0e952d6e3cf8a73f151d63c735ce
3
+ size 8191
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 5.0,
3
- "total_flos": 31406948352000.0,
4
- "train_loss": 0.4169128179550171,
5
- "train_runtime": 607.0886,
6
  "train_samples": 26621,
7
- "train_samples_per_second": 14.808,
8
  "train_steps_per_second": 0.124
9
  }
 
1
  {
2
+ "epoch": 3.0,
3
+ "total_flos": 18844169011200.0,
4
+ "train_loss": 0.5649417373869154,
5
+ "train_runtime": 363.236,
6
  "train_samples": 26621,
7
+ "train_samples_per_second": 14.85,
8
  "train_steps_per_second": 0.124
9
  }
trainer_state.json CHANGED
@@ -1,179 +1,121 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 5.0,
5
  "eval_steps": 500,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.06666666666666667,
13
- "grad_norm": 98.20704487515755,
14
- "learning_rate": 2.5e-06,
15
  "loss": 1.8595,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.3333333333333333,
20
- "grad_norm": 26.158158491928763,
21
- "learning_rate": 1.25e-05,
22
- "loss": 1.3331,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.6666666666666666,
27
- "grad_norm": 52.110102576445215,
28
- "learning_rate": 1.9956059820218982e-05,
29
- "loss": 0.7898,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 1.0,
34
- "grad_norm": 7.630410550452039,
35
- "learning_rate": 1.9466156752904344e-05,
36
- "loss": 0.6543,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 1.0,
41
- "eval_loss": 0.552646815776825,
42
- "eval_runtime": 4.2535,
43
- "eval_samples_per_second": 46.785,
44
- "eval_steps_per_second": 0.94,
45
  "step": 15
46
  },
47
  {
48
  "epoch": 1.3333333333333333,
49
- "grad_norm": 17.50115629098683,
50
- "learning_rate": 1.8458320592590976e-05,
51
- "loss": 0.5122,
52
  "step": 20
53
  },
54
  {
55
  "epoch": 1.6666666666666665,
56
- "grad_norm": 3.153164672454271,
57
- "learning_rate": 1.698769427778842e-05,
58
- "loss": 0.4774,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 2.0,
63
- "grad_norm": 1.7735518447317222,
64
- "learning_rate": 1.513474193514842e-05,
65
- "loss": 0.4164,
66
  "step": 30
67
  },
68
  {
69
  "epoch": 2.0,
70
- "eval_loss": 0.4356851875782013,
71
- "eval_runtime": 4.105,
72
- "eval_samples_per_second": 48.477,
73
- "eval_steps_per_second": 0.974,
74
  "step": 30
75
  },
76
  {
77
  "epoch": 2.3333333333333335,
78
- "grad_norm": 1.295932258196787,
79
- "learning_rate": 1.300084635000341e-05,
80
- "loss": 0.3318,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 2.6666666666666665,
85
- "grad_norm": 1.0485408353072447,
86
- "learning_rate": 1.070276188945293e-05,
87
- "loss": 0.3158,
88
  "step": 40
89
  },
90
  {
91
  "epoch": 3.0,
92
- "grad_norm": 1.4216337627274278,
93
- "learning_rate": 8.366226381814698e-06,
94
- "loss": 0.2956,
95
  "step": 45
96
  },
97
  {
98
  "epoch": 3.0,
99
- "eval_loss": 0.4067309498786926,
100
- "eval_runtime": 4.1095,
101
- "eval_samples_per_second": 48.424,
102
- "eval_steps_per_second": 0.973,
103
  "step": 45
104
  },
105
  {
106
- "epoch": 3.3333333333333335,
107
- "grad_norm": 1.290531056985457,
108
- "learning_rate": 6.119081473277502e-06,
109
- "loss": 0.2149,
110
- "step": 50
111
- },
112
- {
113
- "epoch": 3.6666666666666665,
114
- "grad_norm": 1.1012108900548532,
115
- "learning_rate": 4.084277875864776e-06,
116
- "loss": 0.1992,
117
- "step": 55
118
- },
119
- {
120
- "epoch": 4.0,
121
- "grad_norm": 1.1454117876381236,
122
- "learning_rate": 2.373148218896182e-06,
123
- "loss": 0.1871,
124
- "step": 60
125
- },
126
- {
127
- "epoch": 4.0,
128
- "eval_loss": 0.42530176043510437,
129
- "eval_runtime": 4.1023,
130
- "eval_samples_per_second": 48.509,
131
- "eval_steps_per_second": 0.975,
132
- "step": 60
133
- },
134
- {
135
- "epoch": 4.333333333333333,
136
- "grad_norm": 0.947060255649263,
137
- "learning_rate": 1.0793155744261352e-06,
138
- "loss": 0.145,
139
- "step": 65
140
- },
141
- {
142
- "epoch": 4.666666666666667,
143
- "grad_norm": 1.2233089117487779,
144
- "learning_rate": 2.735709467518699e-07,
145
- "loss": 0.1396,
146
- "step": 70
147
- },
148
- {
149
- "epoch": 5.0,
150
- "grad_norm": 0.8371390612015485,
151
- "learning_rate": 0.0,
152
- "loss": 0.1363,
153
- "step": 75
154
- },
155
- {
156
- "epoch": 5.0,
157
- "eval_loss": 0.4708780348300934,
158
- "eval_runtime": 4.0889,
159
- "eval_samples_per_second": 48.669,
160
- "eval_steps_per_second": 0.978,
161
- "step": 75
162
- },
163
- {
164
- "epoch": 5.0,
165
- "step": 75,
166
- "total_flos": 31406948352000.0,
167
- "train_loss": 0.4169128179550171,
168
- "train_runtime": 607.0886,
169
- "train_samples_per_second": 14.808,
170
  "train_steps_per_second": 0.124
171
  }
172
  ],
173
  "logging_steps": 5,
174
- "max_steps": 75,
175
  "num_input_tokens_seen": 0,
176
- "num_train_epochs": 5,
177
  "save_steps": 100,
178
  "stateful_callbacks": {
179
  "TrainerControl": {
@@ -187,7 +129,7 @@
187
  "attributes": {}
188
  }
189
  },
190
- "total_flos": 31406948352000.0,
191
  "train_batch_size": 16,
192
  "trial_name": null,
193
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
  "eval_steps": 500,
6
+ "global_step": 45,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.06666666666666667,
13
+ "grad_norm": 98.21472515172992,
14
+ "learning_rate": 4.000000000000001e-06,
15
  "loss": 1.8595,
16
  "step": 1
17
  },
18
  {
19
  "epoch": 0.3333333333333333,
20
+ "grad_norm": 30.799283162433554,
21
+ "learning_rate": 2e-05,
22
+ "loss": 1.3856,
23
  "step": 5
24
  },
25
  {
26
  "epoch": 0.6666666666666666,
27
+ "grad_norm": 7.8774399080468624,
28
+ "learning_rate": 1.9238795325112867e-05,
29
+ "loss": 0.8251,
30
  "step": 10
31
  },
32
  {
33
  "epoch": 1.0,
34
+ "grad_norm": 4.085781086210361,
35
+ "learning_rate": 1.7071067811865477e-05,
36
+ "loss": 0.6043,
37
  "step": 15
38
  },
39
  {
40
  "epoch": 1.0,
41
+ "eval_loss": 0.5420283079147339,
42
+ "eval_runtime": 4.2361,
43
+ "eval_samples_per_second": 46.977,
44
+ "eval_steps_per_second": 0.944,
45
  "step": 15
46
  },
47
  {
48
  "epoch": 1.3333333333333333,
49
+ "grad_norm": 2.0713099775720534,
50
+ "learning_rate": 1.3826834323650899e-05,
51
+ "loss": 0.4794,
52
  "step": 20
53
  },
54
  {
55
  "epoch": 1.6666666666666665,
56
+ "grad_norm": 1.3136102966664787,
57
+ "learning_rate": 1e-05,
58
+ "loss": 0.4248,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 2.0,
63
+ "grad_norm": 1.1912342263078068,
64
+ "learning_rate": 6.173165676349103e-06,
65
+ "loss": 0.3922,
66
  "step": 30
67
  },
68
  {
69
  "epoch": 2.0,
70
+ "eval_loss": 0.4156629145145416,
71
+ "eval_runtime": 4.0917,
72
+ "eval_samples_per_second": 48.635,
73
+ "eval_steps_per_second": 0.978,
74
  "step": 30
75
  },
76
  {
77
  "epoch": 2.3333333333333335,
78
+ "grad_norm": 1.1730848362574522,
79
+ "learning_rate": 2.9289321881345257e-06,
80
+ "loss": 0.3075,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 2.6666666666666665,
85
+ "grad_norm": 0.9598398180975349,
86
+ "learning_rate": 7.612046748871327e-07,
87
+ "loss": 0.2917,
88
  "step": 40
89
  },
90
  {
91
  "epoch": 3.0,
92
+ "grad_norm": 1.3661284541735195,
93
+ "learning_rate": 0.0,
94
+ "loss": 0.2791,
95
  "step": 45
96
  },
97
  {
98
  "epoch": 3.0,
99
+ "eval_loss": 0.40715718269348145,
100
+ "eval_runtime": 4.0987,
101
+ "eval_samples_per_second": 48.552,
102
+ "eval_steps_per_second": 0.976,
103
  "step": 45
104
  },
105
  {
106
+ "epoch": 3.0,
107
+ "step": 45,
108
+ "total_flos": 18844169011200.0,
109
+ "train_loss": 0.5649417373869154,
110
+ "train_runtime": 363.236,
111
+ "train_samples_per_second": 14.85,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112
  "train_steps_per_second": 0.124
113
  }
114
  ],
115
  "logging_steps": 5,
116
+ "max_steps": 45,
117
  "num_input_tokens_seen": 0,
118
+ "num_train_epochs": 3,
119
  "save_steps": 100,
120
  "stateful_callbacks": {
121
  "TrainerControl": {
 
129
  "attributes": {}
130
  }
131
  },
132
+ "total_flos": 18844169011200.0,
133
  "train_batch_size": 16,
134
  "trial_name": null,
135
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcbd9b7605e96941aae059351fb08ef1d90212437053df7ce4e1a3a85ba81b73
3
  size 6392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16468e20b0c8c5a6af8b75ded9d347ce0856a32d5e19a9f872bac737dba7ae17
3
  size 6392