csikasote commited on
Commit
e23ad97
1 Parent(s): 6d2a040

End of training

Browse files
README.md CHANGED
@@ -3,6 +3,8 @@ library_name: transformers
3
  license: apache-2.0
4
  base_model: facebook/wav2vec2-xls-r-1b
5
  tags:
 
 
6
  - generated_from_trainer
7
  metrics:
8
  - wer
@@ -16,10 +18,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # xls-r-1b-bem-genbed-all
18
 
19
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-1b](https://huggingface.co/facebook/wav2vec2-xls-r-1b) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.2316
22
- - Wer: 0.7116
23
 
24
  ## Model description
25
 
 
3
  license: apache-2.0
4
  base_model: facebook/wav2vec2-xls-r-1b
5
  tags:
6
+ - automatic-speech-recognition
7
+ - genbed
8
  - generated_from_trainer
9
  metrics:
10
  - wer
 
18
 
19
  # xls-r-1b-bem-genbed-all
20
 
21
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-1b](https://huggingface.co/facebook/wav2vec2-xls-r-1b) on the GENBED - BEM dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.2172
24
+ - Wer: 0.7294
25
 
26
  ## Model description
27
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.9966953073364175,
3
+ "eval_loss": 0.21721933782100677,
4
+ "eval_runtime": 124.0521,
5
+ "eval_samples": 2018,
6
+ "eval_samples_per_second": 16.267,
7
+ "eval_steps_per_second": 4.071,
8
+ "eval_wer": 0.7294350842418236,
9
+ "total_flos": 1.7289369341985591e+19,
10
+ "train_loss": 0.4604275799302197,
11
+ "train_runtime": 6372.0716,
12
+ "train_samples": 6049,
13
+ "train_samples_per_second": 4.746,
14
+ "train_steps_per_second": 0.593
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.9966953073364175,
3
+ "eval_loss": 0.21721933782100677,
4
+ "eval_runtime": 124.0521,
5
+ "eval_samples": 2018,
6
+ "eval_samples_per_second": 16.267,
7
+ "eval_steps_per_second": 4.071,
8
+ "eval_wer": 0.7294350842418236
9
+ }
runs/Sep07_01-16-38_02da24ddff73/events.out.tfevents.1725678589.02da24ddff73.7708.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ac0bad8ad18e8fb7f7515f7df815a97d45894f346cbe3e07e7eeb56ef4f259e
3
+ size 406
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.9966953073364175,
3
+ "total_flos": 1.7289369341985591e+19,
4
+ "train_loss": 0.4604275799302197,
5
+ "train_runtime": 6372.0716,
6
+ "train_samples": 6049,
7
+ "train_samples_per_second": 4.746,
8
+ "train_steps_per_second": 0.593
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,330 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.21721933782100677,
3
+ "best_model_checkpoint": "./xls-r-1b-bem-genbed-all/checkpoint-3000",
4
+ "epoch": 4.9966953073364175,
5
+ "eval_steps": 200,
6
+ "global_step": 3780,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.26437541308658297,
13
+ "grad_norm": 3.2960169315338135,
14
+ "learning_rate": 1.97e-05,
15
+ "loss": 4.6827,
16
+ "step": 200
17
+ },
18
+ {
19
+ "epoch": 0.26437541308658297,
20
+ "eval_loss": 2.834711790084839,
21
+ "eval_runtime": 123.0061,
22
+ "eval_samples_per_second": 16.406,
23
+ "eval_steps_per_second": 4.105,
24
+ "eval_wer": 1.0,
25
+ "step": 200
26
+ },
27
+ {
28
+ "epoch": 0.5287508261731659,
29
+ "grad_norm": 5.020152568817139,
30
+ "learning_rate": 3.97e-05,
31
+ "loss": 1.0401,
32
+ "step": 400
33
+ },
34
+ {
35
+ "epoch": 0.5287508261731659,
36
+ "eval_loss": 0.5636318922042847,
37
+ "eval_runtime": 121.385,
38
+ "eval_samples_per_second": 16.625,
39
+ "eval_steps_per_second": 4.16,
40
+ "eval_wer": 0.9410307234886026,
41
+ "step": 400
42
+ },
43
+ {
44
+ "epoch": 0.7931262392597488,
45
+ "grad_norm": 4.118492126464844,
46
+ "learning_rate": 4.852134146341464e-05,
47
+ "loss": 0.4289,
48
+ "step": 600
49
+ },
50
+ {
51
+ "epoch": 0.7931262392597488,
52
+ "eval_loss": 0.40182533860206604,
53
+ "eval_runtime": 122.3487,
54
+ "eval_samples_per_second": 16.494,
55
+ "eval_steps_per_second": 4.128,
56
+ "eval_wer": 0.9028741328047571,
57
+ "step": 600
58
+ },
59
+ {
60
+ "epoch": 1.0575016523463319,
61
+ "grad_norm": 1.0666348934173584,
62
+ "learning_rate": 4.5472560975609756e-05,
63
+ "loss": 0.3449,
64
+ "step": 800
65
+ },
66
+ {
67
+ "epoch": 1.0575016523463319,
68
+ "eval_loss": 0.3604024052619934,
69
+ "eval_runtime": 123.3539,
70
+ "eval_samples_per_second": 16.359,
71
+ "eval_steps_per_second": 4.094,
72
+ "eval_wer": 0.8771060455896927,
73
+ "step": 800
74
+ },
75
+ {
76
+ "epoch": 1.3218770654329148,
77
+ "grad_norm": 3.203847885131836,
78
+ "learning_rate": 4.242378048780488e-05,
79
+ "loss": 0.2954,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 1.3218770654329148,
84
+ "eval_loss": 0.33885934948921204,
85
+ "eval_runtime": 124.083,
86
+ "eval_samples_per_second": 16.263,
87
+ "eval_steps_per_second": 4.07,
88
+ "eval_wer": 0.8741328047571854,
89
+ "step": 1000
90
+ },
91
+ {
92
+ "epoch": 1.5862524785194978,
93
+ "grad_norm": 2.469949722290039,
94
+ "learning_rate": 3.9375e-05,
95
+ "loss": 0.2719,
96
+ "step": 1200
97
+ },
98
+ {
99
+ "epoch": 1.5862524785194978,
100
+ "eval_loss": 0.29620760679244995,
101
+ "eval_runtime": 124.1198,
102
+ "eval_samples_per_second": 16.258,
103
+ "eval_steps_per_second": 4.069,
104
+ "eval_wer": 0.8439048562933598,
105
+ "step": 1200
106
+ },
107
+ {
108
+ "epoch": 1.8506278916060808,
109
+ "grad_norm": 1.500631332397461,
110
+ "learning_rate": 3.632621951219513e-05,
111
+ "loss": 0.2472,
112
+ "step": 1400
113
+ },
114
+ {
115
+ "epoch": 1.8506278916060808,
116
+ "eval_loss": 0.2701094448566437,
117
+ "eval_runtime": 123.0962,
118
+ "eval_samples_per_second": 16.394,
119
+ "eval_steps_per_second": 4.102,
120
+ "eval_wer": 0.8052527254707631,
121
+ "step": 1400
122
+ },
123
+ {
124
+ "epoch": 2.1150033046926637,
125
+ "grad_norm": 0.6989238262176514,
126
+ "learning_rate": 3.327743902439025e-05,
127
+ "loss": 0.2093,
128
+ "step": 1600
129
+ },
130
+ {
131
+ "epoch": 2.1150033046926637,
132
+ "eval_loss": 0.2598518133163452,
133
+ "eval_runtime": 123.5685,
134
+ "eval_samples_per_second": 16.331,
135
+ "eval_steps_per_second": 4.087,
136
+ "eval_wer": 0.8285431119920713,
137
+ "step": 1600
138
+ },
139
+ {
140
+ "epoch": 2.3793787177792467,
141
+ "grad_norm": 0.9642230868339539,
142
+ "learning_rate": 3.022865853658537e-05,
143
+ "loss": 0.1725,
144
+ "step": 1800
145
+ },
146
+ {
147
+ "epoch": 2.3793787177792467,
148
+ "eval_loss": 0.2533758580684662,
149
+ "eval_runtime": 124.2174,
150
+ "eval_samples_per_second": 16.246,
151
+ "eval_steps_per_second": 4.065,
152
+ "eval_wer": 0.8374628344895937,
153
+ "step": 1800
154
+ },
155
+ {
156
+ "epoch": 2.6437541308658297,
157
+ "grad_norm": 0.6002829670906067,
158
+ "learning_rate": 2.717987804878049e-05,
159
+ "loss": 0.1675,
160
+ "step": 2000
161
+ },
162
+ {
163
+ "epoch": 2.6437541308658297,
164
+ "eval_loss": 0.240593820810318,
165
+ "eval_runtime": 125.6029,
166
+ "eval_samples_per_second": 16.067,
167
+ "eval_steps_per_second": 4.021,
168
+ "eval_wer": 0.7690782953419227,
169
+ "step": 2000
170
+ },
171
+ {
172
+ "epoch": 2.9081295439524126,
173
+ "grad_norm": 1.0664541721343994,
174
+ "learning_rate": 2.413109756097561e-05,
175
+ "loss": 0.1632,
176
+ "step": 2200
177
+ },
178
+ {
179
+ "epoch": 2.9081295439524126,
180
+ "eval_loss": 0.2308683693408966,
181
+ "eval_runtime": 124.4,
182
+ "eval_samples_per_second": 16.222,
183
+ "eval_steps_per_second": 4.059,
184
+ "eval_wer": 0.7616451932606542,
185
+ "step": 2200
186
+ },
187
+ {
188
+ "epoch": 3.1725049570389956,
189
+ "grad_norm": 0.6247605681419373,
190
+ "learning_rate": 2.108231707317073e-05,
191
+ "loss": 0.1295,
192
+ "step": 2400
193
+ },
194
+ {
195
+ "epoch": 3.1725049570389956,
196
+ "eval_loss": 0.23871323466300964,
197
+ "eval_runtime": 124.4212,
198
+ "eval_samples_per_second": 16.219,
199
+ "eval_steps_per_second": 4.059,
200
+ "eval_wer": 0.7556987115956393,
201
+ "step": 2400
202
+ },
203
+ {
204
+ "epoch": 3.4368803701255786,
205
+ "grad_norm": 0.5814207792282104,
206
+ "learning_rate": 1.8033536585365853e-05,
207
+ "loss": 0.1082,
208
+ "step": 2600
209
+ },
210
+ {
211
+ "epoch": 3.4368803701255786,
212
+ "eval_loss": 0.2275388240814209,
213
+ "eval_runtime": 124.2026,
214
+ "eval_samples_per_second": 16.248,
215
+ "eval_steps_per_second": 4.066,
216
+ "eval_wer": 0.7329038652130823,
217
+ "step": 2600
218
+ },
219
+ {
220
+ "epoch": 3.7012557832121615,
221
+ "grad_norm": 0.5891350507736206,
222
+ "learning_rate": 1.4984756097560976e-05,
223
+ "loss": 0.1059,
224
+ "step": 2800
225
+ },
226
+ {
227
+ "epoch": 3.7012557832121615,
228
+ "eval_loss": 0.223988875746727,
229
+ "eval_runtime": 124.3149,
230
+ "eval_samples_per_second": 16.233,
231
+ "eval_steps_per_second": 4.062,
232
+ "eval_wer": 0.7329038652130823,
233
+ "step": 2800
234
+ },
235
+ {
236
+ "epoch": 3.965631196298744,
237
+ "grad_norm": 1.1988484859466553,
238
+ "learning_rate": 1.1935975609756097e-05,
239
+ "loss": 0.1049,
240
+ "step": 3000
241
+ },
242
+ {
243
+ "epoch": 3.965631196298744,
244
+ "eval_loss": 0.21721933782100677,
245
+ "eval_runtime": 124.3819,
246
+ "eval_samples_per_second": 16.224,
247
+ "eval_steps_per_second": 4.06,
248
+ "eval_wer": 0.7294350842418236,
249
+ "step": 3000
250
+ },
251
+ {
252
+ "epoch": 4.2300066093853275,
253
+ "grad_norm": 0.32344383001327515,
254
+ "learning_rate": 8.88719512195122e-06,
255
+ "loss": 0.0657,
256
+ "step": 3200
257
+ },
258
+ {
259
+ "epoch": 4.2300066093853275,
260
+ "eval_loss": 0.23203983902931213,
261
+ "eval_runtime": 125.8598,
262
+ "eval_samples_per_second": 16.034,
263
+ "eval_steps_per_second": 4.012,
264
+ "eval_wer": 0.722001982160555,
265
+ "step": 3200
266
+ },
267
+ {
268
+ "epoch": 4.49438202247191,
269
+ "grad_norm": 0.8019347190856934,
270
+ "learning_rate": 5.838414634146342e-06,
271
+ "loss": 0.059,
272
+ "step": 3400
273
+ },
274
+ {
275
+ "epoch": 4.49438202247191,
276
+ "eval_loss": 0.23412850499153137,
277
+ "eval_runtime": 124.439,
278
+ "eval_samples_per_second": 16.217,
279
+ "eval_steps_per_second": 4.058,
280
+ "eval_wer": 0.7215064420218038,
281
+ "step": 3400
282
+ },
283
+ {
284
+ "epoch": 4.758757435558493,
285
+ "grad_norm": 0.9007149934768677,
286
+ "learning_rate": 2.7896341463414635e-06,
287
+ "loss": 0.0582,
288
+ "step": 3600
289
+ },
290
+ {
291
+ "epoch": 4.758757435558493,
292
+ "eval_loss": 0.2315901815891266,
293
+ "eval_runtime": 124.1565,
294
+ "eval_samples_per_second": 16.254,
295
+ "eval_steps_per_second": 4.067,
296
+ "eval_wer": 0.711595639246779,
297
+ "step": 3600
298
+ },
299
+ {
300
+ "epoch": 4.9966953073364175,
301
+ "step": 3780,
302
+ "total_flos": 1.7289369341985591e+19,
303
+ "train_loss": 0.4604275799302197,
304
+ "train_runtime": 6372.0716,
305
+ "train_samples_per_second": 4.746,
306
+ "train_steps_per_second": 0.593
307
+ }
308
+ ],
309
+ "logging_steps": 200,
310
+ "max_steps": 3780,
311
+ "num_input_tokens_seen": 0,
312
+ "num_train_epochs": 5,
313
+ "save_steps": 200,
314
+ "stateful_callbacks": {
315
+ "TrainerControl": {
316
+ "args": {
317
+ "should_epoch_stop": false,
318
+ "should_evaluate": false,
319
+ "should_log": false,
320
+ "should_save": true,
321
+ "should_training_stop": true
322
+ },
323
+ "attributes": {}
324
+ }
325
+ },
326
+ "total_flos": 1.7289369341985591e+19,
327
+ "train_batch_size": 4,
328
+ "trial_name": null,
329
+ "trial_params": null
330
+ }