Training checkpoint
Browse files- README.md +7 -5
- trainer_state.json +312 -14
README.md
CHANGED
@@ -1,21 +1,23 @@
|
|
1 |
---
|
|
|
|
|
2 |
license: apache-2.0
|
3 |
-
base_model:
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
-
- common_voice_16_1
|
8 |
model-index:
|
9 |
-
- name:
|
10 |
results: []
|
11 |
---
|
12 |
|
13 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
14 |
should probably proofread and complete it, then remove this comment. -->
|
15 |
|
16 |
-
#
|
17 |
|
18 |
-
This model is a fine-tuned version of [
|
19 |
It achieves the following results on the evaluation set:
|
20 |
- Loss: 0.3867
|
21 |
- Cer: 21.2804
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- zh
|
4 |
license: apache-2.0
|
5 |
+
base_model: openai/whisper-small
|
6 |
tags:
|
7 |
- generated_from_trainer
|
8 |
datasets:
|
9 |
+
- mozilla-foundation/common_voice_16_1
|
10 |
model-index:
|
11 |
+
- name: Wisper-Small-zh_test
|
12 |
results: []
|
13 |
---
|
14 |
|
15 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
should probably proofread and complete it, then remove this comment. -->
|
17 |
|
18 |
+
# Wisper-Small-zh_test
|
19 |
|
20 |
+
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 16.1 dataset.
|
21 |
It achieves the following results on the evaluation set:
|
22 |
- Loss: 0.3867
|
23 |
- Cer: 21.2804
|
trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric":
|
3 |
-
"best_model_checkpoint": "/kaggle/working/whisper-small/checkpoint-
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -307,19 +307,317 @@
|
|
307 |
"step": 1000
|
308 |
},
|
309 |
{
|
310 |
-
"epoch": 1.
|
311 |
-
"
|
312 |
-
"
|
313 |
-
"
|
314 |
-
"
|
315 |
-
|
316 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
317 |
}
|
318 |
],
|
319 |
"logging_steps": 25,
|
320 |
-
"max_steps":
|
321 |
"num_input_tokens_seen": 0,
|
322 |
-
"num_train_epochs":
|
323 |
"save_steps": 500,
|
324 |
"stateful_callbacks": {
|
325 |
"TrainerControl": {
|
@@ -333,7 +631,7 @@
|
|
333 |
"attributes": {}
|
334 |
}
|
335 |
},
|
336 |
-
"total_flos":
|
337 |
"train_batch_size": 16,
|
338 |
"trial_name": null,
|
339 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 21.28044187798142,
|
3 |
+
"best_model_checkpoint": "/kaggle/working/whisper-small/checkpoint-2000",
|
4 |
+
"epoch": 2.6490066225165565,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 2000,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
307 |
"step": 1000
|
308 |
},
|
309 |
{
|
310 |
+
"epoch": 1.3576158940397351,
|
311 |
+
"grad_norm": 15.552218437194824,
|
312 |
+
"learning_rate": 6.540000000000001e-06,
|
313 |
+
"loss": 0.501,
|
314 |
+
"step": 1025
|
315 |
+
},
|
316 |
+
{
|
317 |
+
"epoch": 1.390728476821192,
|
318 |
+
"grad_norm": 18.500070571899414,
|
319 |
+
"learning_rate": 6.373333333333334e-06,
|
320 |
+
"loss": 0.5625,
|
321 |
+
"step": 1050
|
322 |
+
},
|
323 |
+
{
|
324 |
+
"epoch": 1.423841059602649,
|
325 |
+
"grad_norm": 13.384895324707031,
|
326 |
+
"learning_rate": 6.206666666666668e-06,
|
327 |
+
"loss": 0.4785,
|
328 |
+
"step": 1075
|
329 |
+
},
|
330 |
+
{
|
331 |
+
"epoch": 1.4569536423841059,
|
332 |
+
"grad_norm": 11.979981422424316,
|
333 |
+
"learning_rate": 6.040000000000001e-06,
|
334 |
+
"loss": 0.4728,
|
335 |
+
"step": 1100
|
336 |
+
},
|
337 |
+
{
|
338 |
+
"epoch": 1.490066225165563,
|
339 |
+
"grad_norm": 13.125775337219238,
|
340 |
+
"learning_rate": 5.873333333333334e-06,
|
341 |
+
"loss": 0.4484,
|
342 |
+
"step": 1125
|
343 |
+
},
|
344 |
+
{
|
345 |
+
"epoch": 1.5231788079470199,
|
346 |
+
"grad_norm": 10.14282512664795,
|
347 |
+
"learning_rate": 5.713333333333334e-06,
|
348 |
+
"loss": 0.4227,
|
349 |
+
"step": 1150
|
350 |
+
},
|
351 |
+
{
|
352 |
+
"epoch": 1.5562913907284768,
|
353 |
+
"grad_norm": 14.51291561126709,
|
354 |
+
"learning_rate": 5.546666666666667e-06,
|
355 |
+
"loss": 0.4419,
|
356 |
+
"step": 1175
|
357 |
+
},
|
358 |
+
{
|
359 |
+
"epoch": 1.589403973509934,
|
360 |
+
"grad_norm": 13.863635063171387,
|
361 |
+
"learning_rate": 5.380000000000001e-06,
|
362 |
+
"loss": 0.4429,
|
363 |
+
"step": 1200
|
364 |
+
},
|
365 |
+
{
|
366 |
+
"epoch": 1.6225165562913908,
|
367 |
+
"grad_norm": 11.522802352905273,
|
368 |
+
"learning_rate": 5.213333333333334e-06,
|
369 |
+
"loss": 0.3981,
|
370 |
+
"step": 1225
|
371 |
+
},
|
372 |
+
{
|
373 |
+
"epoch": 1.6556291390728477,
|
374 |
+
"grad_norm": 13.733115196228027,
|
375 |
+
"learning_rate": 5.046666666666668e-06,
|
376 |
+
"loss": 0.3974,
|
377 |
+
"step": 1250
|
378 |
+
},
|
379 |
+
{
|
380 |
+
"epoch": 1.6887417218543046,
|
381 |
+
"grad_norm": 11.775789260864258,
|
382 |
+
"learning_rate": 4.880000000000001e-06,
|
383 |
+
"loss": 0.3978,
|
384 |
+
"step": 1275
|
385 |
+
},
|
386 |
+
{
|
387 |
+
"epoch": 1.7218543046357615,
|
388 |
+
"grad_norm": 8.765869140625,
|
389 |
+
"learning_rate": 4.713333333333334e-06,
|
390 |
+
"loss": 0.3812,
|
391 |
+
"step": 1300
|
392 |
+
},
|
393 |
+
{
|
394 |
+
"epoch": 1.7549668874172184,
|
395 |
+
"grad_norm": 10.572060585021973,
|
396 |
+
"learning_rate": 4.546666666666667e-06,
|
397 |
+
"loss": 0.4257,
|
398 |
+
"step": 1325
|
399 |
+
},
|
400 |
+
{
|
401 |
+
"epoch": 1.7880794701986755,
|
402 |
+
"grad_norm": 11.202813148498535,
|
403 |
+
"learning_rate": 4.38e-06,
|
404 |
+
"loss": 0.3635,
|
405 |
+
"step": 1350
|
406 |
+
},
|
407 |
+
{
|
408 |
+
"epoch": 1.8211920529801324,
|
409 |
+
"grad_norm": 10.402491569519043,
|
410 |
+
"learning_rate": 4.213333333333333e-06,
|
411 |
+
"loss": 0.3636,
|
412 |
+
"step": 1375
|
413 |
+
},
|
414 |
+
{
|
415 |
+
"epoch": 1.8543046357615895,
|
416 |
+
"grad_norm": 22.270811080932617,
|
417 |
+
"learning_rate": 4.046666666666667e-06,
|
418 |
+
"loss": 0.3696,
|
419 |
+
"step": 1400
|
420 |
+
},
|
421 |
+
{
|
422 |
+
"epoch": 1.8874172185430464,
|
423 |
+
"grad_norm": 8.238192558288574,
|
424 |
+
"learning_rate": 3.88e-06,
|
425 |
+
"loss": 0.397,
|
426 |
+
"step": 1425
|
427 |
+
},
|
428 |
+
{
|
429 |
+
"epoch": 1.9205298013245033,
|
430 |
+
"grad_norm": 9.551776885986328,
|
431 |
+
"learning_rate": 3.713333333333334e-06,
|
432 |
+
"loss": 0.3747,
|
433 |
+
"step": 1450
|
434 |
+
},
|
435 |
+
{
|
436 |
+
"epoch": 1.9536423841059603,
|
437 |
+
"grad_norm": 13.149867057800293,
|
438 |
+
"learning_rate": 3.5466666666666673e-06,
|
439 |
+
"loss": 0.3771,
|
440 |
+
"step": 1475
|
441 |
+
},
|
442 |
+
{
|
443 |
+
"epoch": 1.9867549668874172,
|
444 |
+
"grad_norm": 13.081770896911621,
|
445 |
+
"learning_rate": 3.3800000000000007e-06,
|
446 |
+
"loss": 0.352,
|
447 |
+
"step": 1500
|
448 |
+
},
|
449 |
+
{
|
450 |
+
"epoch": 1.9867549668874172,
|
451 |
+
"eval_cer": 23.299020838563898,
|
452 |
+
"eval_loss": 0.4214184284210205,
|
453 |
+
"eval_runtime": 2258.7536,
|
454 |
+
"eval_samples_per_second": 2.203,
|
455 |
+
"eval_steps_per_second": 0.275,
|
456 |
+
"step": 1500
|
457 |
+
},
|
458 |
+
{
|
459 |
+
"epoch": 2.019867549668874,
|
460 |
+
"grad_norm": 7.450255393981934,
|
461 |
+
"learning_rate": 3.213333333333334e-06,
|
462 |
+
"loss": 0.2911,
|
463 |
+
"step": 1525
|
464 |
+
},
|
465 |
+
{
|
466 |
+
"epoch": 2.052980132450331,
|
467 |
+
"grad_norm": 8.604903221130371,
|
468 |
+
"learning_rate": 3.0466666666666666e-06,
|
469 |
+
"loss": 0.2069,
|
470 |
+
"step": 1550
|
471 |
+
},
|
472 |
+
{
|
473 |
+
"epoch": 2.0860927152317883,
|
474 |
+
"grad_norm": 5.367754936218262,
|
475 |
+
"learning_rate": 2.88e-06,
|
476 |
+
"loss": 0.2461,
|
477 |
+
"step": 1575
|
478 |
+
},
|
479 |
+
{
|
480 |
+
"epoch": 2.119205298013245,
|
481 |
+
"grad_norm": 11.53250789642334,
|
482 |
+
"learning_rate": 2.7133333333333333e-06,
|
483 |
+
"loss": 0.2071,
|
484 |
+
"step": 1600
|
485 |
+
},
|
486 |
+
{
|
487 |
+
"epoch": 2.152317880794702,
|
488 |
+
"grad_norm": 9.057580947875977,
|
489 |
+
"learning_rate": 2.5466666666666667e-06,
|
490 |
+
"loss": 0.2255,
|
491 |
+
"step": 1625
|
492 |
+
},
|
493 |
+
{
|
494 |
+
"epoch": 2.185430463576159,
|
495 |
+
"grad_norm": 8.253719329833984,
|
496 |
+
"learning_rate": 2.38e-06,
|
497 |
+
"loss": 0.1968,
|
498 |
+
"step": 1650
|
499 |
+
},
|
500 |
+
{
|
501 |
+
"epoch": 2.218543046357616,
|
502 |
+
"grad_norm": 10.867476463317871,
|
503 |
+
"learning_rate": 2.2133333333333335e-06,
|
504 |
+
"loss": 0.2177,
|
505 |
+
"step": 1675
|
506 |
+
},
|
507 |
+
{
|
508 |
+
"epoch": 2.251655629139073,
|
509 |
+
"grad_norm": 10.779939651489258,
|
510 |
+
"learning_rate": 2.046666666666667e-06,
|
511 |
+
"loss": 0.2191,
|
512 |
+
"step": 1700
|
513 |
+
},
|
514 |
+
{
|
515 |
+
"epoch": 2.2847682119205297,
|
516 |
+
"grad_norm": 8.389144897460938,
|
517 |
+
"learning_rate": 1.8800000000000002e-06,
|
518 |
+
"loss": 0.2137,
|
519 |
+
"step": 1725
|
520 |
+
},
|
521 |
+
{
|
522 |
+
"epoch": 2.3178807947019866,
|
523 |
+
"grad_norm": 11.38824462890625,
|
524 |
+
"learning_rate": 1.7133333333333336e-06,
|
525 |
+
"loss": 0.2326,
|
526 |
+
"step": 1750
|
527 |
+
},
|
528 |
+
{
|
529 |
+
"epoch": 2.3509933774834435,
|
530 |
+
"grad_norm": 10.286229133605957,
|
531 |
+
"learning_rate": 1.546666666666667e-06,
|
532 |
+
"loss": 0.2014,
|
533 |
+
"step": 1775
|
534 |
+
},
|
535 |
+
{
|
536 |
+
"epoch": 2.384105960264901,
|
537 |
+
"grad_norm": 9.5010986328125,
|
538 |
+
"learning_rate": 1.3800000000000001e-06,
|
539 |
+
"loss": 0.2121,
|
540 |
+
"step": 1800
|
541 |
+
},
|
542 |
+
{
|
543 |
+
"epoch": 2.4172185430463577,
|
544 |
+
"grad_norm": 10.394664764404297,
|
545 |
+
"learning_rate": 1.2133333333333335e-06,
|
546 |
+
"loss": 0.2166,
|
547 |
+
"step": 1825
|
548 |
+
},
|
549 |
+
{
|
550 |
+
"epoch": 2.4503311258278146,
|
551 |
+
"grad_norm": 13.732166290283203,
|
552 |
+
"learning_rate": 1.0466666666666669e-06,
|
553 |
+
"loss": 0.2243,
|
554 |
+
"step": 1850
|
555 |
+
},
|
556 |
+
{
|
557 |
+
"epoch": 2.4834437086092715,
|
558 |
+
"grad_norm": 7.431657791137695,
|
559 |
+
"learning_rate": 8.8e-07,
|
560 |
+
"loss": 0.2186,
|
561 |
+
"step": 1875
|
562 |
+
},
|
563 |
+
{
|
564 |
+
"epoch": 2.5165562913907285,
|
565 |
+
"grad_norm": 5.561306476593018,
|
566 |
+
"learning_rate": 7.133333333333334e-07,
|
567 |
+
"loss": 0.1992,
|
568 |
+
"step": 1900
|
569 |
+
},
|
570 |
+
{
|
571 |
+
"epoch": 2.5496688741721854,
|
572 |
+
"grad_norm": 8.894646644592285,
|
573 |
+
"learning_rate": 5.466666666666667e-07,
|
574 |
+
"loss": 0.2195,
|
575 |
+
"step": 1925
|
576 |
+
},
|
577 |
+
{
|
578 |
+
"epoch": 2.5827814569536423,
|
579 |
+
"grad_norm": 8.13320541381836,
|
580 |
+
"learning_rate": 3.8e-07,
|
581 |
+
"loss": 0.2235,
|
582 |
+
"step": 1950
|
583 |
+
},
|
584 |
+
{
|
585 |
+
"epoch": 2.6158940397350996,
|
586 |
+
"grad_norm": 8.711597442626953,
|
587 |
+
"learning_rate": 2.1333333333333334e-07,
|
588 |
+
"loss": 0.2033,
|
589 |
+
"step": 1975
|
590 |
+
},
|
591 |
+
{
|
592 |
+
"epoch": 2.6490066225165565,
|
593 |
+
"grad_norm": 9.557293891906738,
|
594 |
+
"learning_rate": 4.6666666666666674e-08,
|
595 |
+
"loss": 0.2243,
|
596 |
+
"step": 2000
|
597 |
+
},
|
598 |
+
{
|
599 |
+
"epoch": 2.6490066225165565,
|
600 |
+
"eval_cer": 21.28044187798142,
|
601 |
+
"eval_loss": 0.38667094707489014,
|
602 |
+
"eval_runtime": 2236.5916,
|
603 |
+
"eval_samples_per_second": 2.225,
|
604 |
+
"eval_steps_per_second": 0.278,
|
605 |
+
"step": 2000
|
606 |
+
},
|
607 |
+
{
|
608 |
+
"epoch": 2.6490066225165565,
|
609 |
+
"step": 2000,
|
610 |
+
"total_flos": 9.31455866216448e+18,
|
611 |
+
"train_loss": 0.15934584045410155,
|
612 |
+
"train_runtime": 12035.0029,
|
613 |
+
"train_samples_per_second": 2.659,
|
614 |
+
"train_steps_per_second": 0.166
|
615 |
}
|
616 |
],
|
617 |
"logging_steps": 25,
|
618 |
+
"max_steps": 2000,
|
619 |
"num_input_tokens_seen": 0,
|
620 |
+
"num_train_epochs": 3,
|
621 |
"save_steps": 500,
|
622 |
"stateful_callbacks": {
|
623 |
"TrainerControl": {
|
|
|
631 |
"attributes": {}
|
632 |
}
|
633 |
},
|
634 |
+
"total_flos": 9.31455866216448e+18,
|
635 |
"train_batch_size": 16,
|
636 |
"trial_name": null,
|
637 |
"trial_params": null
|