elsayedissa
commited on
Commit
•
0ae67d4
1
Parent(s):
7d368e5
Training in progress, step 19000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:181c2bd989506934f36609329ae3277eb855e8acf49b3c12e326fb79cb004eb3
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15b5954cb47199f441cbbb70307c72621abf9248aa385660bfdb147a516624ef
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9460131d11a73df83270a4796c5192126900329b0b2274773520724725f4b12d
|
3 |
size 14575
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6a09c97ace5e5a49360efd6cac2da516ffa83c9ee08cb64155c59b3a316ab7a
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec3490e7ac28e19432d614c125578d2d66d232f9a51b472efd2d122fe9c6b708
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -4488,11 +4488,260 @@
|
|
4488 |
"eval_steps_per_second": 0.062,
|
4489 |
"eval_wer": 0.08993702591943148,
|
4490 |
"step": 18000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4491 |
}
|
4492 |
],
|
4493 |
"max_steps": 25000,
|
4494 |
"num_train_epochs": 1,
|
4495 |
-
"total_flos": 6.
|
4496 |
"trial_name": null,
|
4497 |
"trial_params": null
|
4498 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.6595161234336492,
|
5 |
+
"global_step": 19000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
4488 |
"eval_steps_per_second": 0.062,
|
4489 |
"eval_wer": 0.08993702591943148,
|
4490 |
"step": 18000
|
4491 |
+
},
|
4492 |
+
{
|
4493 |
+
"epoch": 0.63,
|
4494 |
+
"learning_rate": 2.8514285714285715e-06,
|
4495 |
+
"loss": 0.1658,
|
4496 |
+
"step": 18025
|
4497 |
+
},
|
4498 |
+
{
|
4499 |
+
"epoch": 0.63,
|
4500 |
+
"learning_rate": 2.8412244897959185e-06,
|
4501 |
+
"loss": 0.1237,
|
4502 |
+
"step": 18050
|
4503 |
+
},
|
4504 |
+
{
|
4505 |
+
"epoch": 0.63,
|
4506 |
+
"learning_rate": 2.831020408163265e-06,
|
4507 |
+
"loss": 0.1672,
|
4508 |
+
"step": 18075
|
4509 |
+
},
|
4510 |
+
{
|
4511 |
+
"epoch": 0.63,
|
4512 |
+
"learning_rate": 2.8208163265306127e-06,
|
4513 |
+
"loss": 0.1319,
|
4514 |
+
"step": 18100
|
4515 |
+
},
|
4516 |
+
{
|
4517 |
+
"epoch": 0.63,
|
4518 |
+
"learning_rate": 2.8106122448979593e-06,
|
4519 |
+
"loss": 0.1543,
|
4520 |
+
"step": 18125
|
4521 |
+
},
|
4522 |
+
{
|
4523 |
+
"epoch": 0.63,
|
4524 |
+
"learning_rate": 2.8004081632653064e-06,
|
4525 |
+
"loss": 0.1167,
|
4526 |
+
"step": 18150
|
4527 |
+
},
|
4528 |
+
{
|
4529 |
+
"epoch": 0.63,
|
4530 |
+
"learning_rate": 2.790204081632653e-06,
|
4531 |
+
"loss": 0.1708,
|
4532 |
+
"step": 18175
|
4533 |
+
},
|
4534 |
+
{
|
4535 |
+
"epoch": 0.63,
|
4536 |
+
"learning_rate": 2.7800000000000005e-06,
|
4537 |
+
"loss": 0.1265,
|
4538 |
+
"step": 18200
|
4539 |
+
},
|
4540 |
+
{
|
4541 |
+
"epoch": 0.63,
|
4542 |
+
"learning_rate": 2.769795918367347e-06,
|
4543 |
+
"loss": 0.1588,
|
4544 |
+
"step": 18225
|
4545 |
+
},
|
4546 |
+
{
|
4547 |
+
"epoch": 0.63,
|
4548 |
+
"learning_rate": 2.759591836734694e-06,
|
4549 |
+
"loss": 0.133,
|
4550 |
+
"step": 18250
|
4551 |
+
},
|
4552 |
+
{
|
4553 |
+
"epoch": 0.63,
|
4554 |
+
"learning_rate": 2.749387755102041e-06,
|
4555 |
+
"loss": 0.1744,
|
4556 |
+
"step": 18275
|
4557 |
+
},
|
4558 |
+
{
|
4559 |
+
"epoch": 0.64,
|
4560 |
+
"learning_rate": 2.739183673469388e-06,
|
4561 |
+
"loss": 0.1305,
|
4562 |
+
"step": 18300
|
4563 |
+
},
|
4564 |
+
{
|
4565 |
+
"epoch": 0.64,
|
4566 |
+
"learning_rate": 2.728979591836735e-06,
|
4567 |
+
"loss": 0.1486,
|
4568 |
+
"step": 18325
|
4569 |
+
},
|
4570 |
+
{
|
4571 |
+
"epoch": 0.64,
|
4572 |
+
"learning_rate": 2.718775510204082e-06,
|
4573 |
+
"loss": 0.1379,
|
4574 |
+
"step": 18350
|
4575 |
+
},
|
4576 |
+
{
|
4577 |
+
"epoch": 0.64,
|
4578 |
+
"learning_rate": 2.7085714285714287e-06,
|
4579 |
+
"loss": 0.1669,
|
4580 |
+
"step": 18375
|
4581 |
+
},
|
4582 |
+
{
|
4583 |
+
"epoch": 0.64,
|
4584 |
+
"learning_rate": 2.6983673469387757e-06,
|
4585 |
+
"loss": 0.1447,
|
4586 |
+
"step": 18400
|
4587 |
+
},
|
4588 |
+
{
|
4589 |
+
"epoch": 0.64,
|
4590 |
+
"learning_rate": 2.6881632653061228e-06,
|
4591 |
+
"loss": 0.1751,
|
4592 |
+
"step": 18425
|
4593 |
+
},
|
4594 |
+
{
|
4595 |
+
"epoch": 0.64,
|
4596 |
+
"learning_rate": 2.67795918367347e-06,
|
4597 |
+
"loss": 0.1259,
|
4598 |
+
"step": 18450
|
4599 |
+
},
|
4600 |
+
{
|
4601 |
+
"epoch": 0.64,
|
4602 |
+
"learning_rate": 2.6677551020408165e-06,
|
4603 |
+
"loss": 0.1406,
|
4604 |
+
"step": 18475
|
4605 |
+
},
|
4606 |
+
{
|
4607 |
+
"epoch": 0.64,
|
4608 |
+
"learning_rate": 2.6575510204081635e-06,
|
4609 |
+
"loss": 0.1463,
|
4610 |
+
"step": 18500
|
4611 |
+
},
|
4612 |
+
{
|
4613 |
+
"epoch": 0.64,
|
4614 |
+
"learning_rate": 2.64734693877551e-06,
|
4615 |
+
"loss": 0.164,
|
4616 |
+
"step": 18525
|
4617 |
+
},
|
4618 |
+
{
|
4619 |
+
"epoch": 0.64,
|
4620 |
+
"learning_rate": 2.6371428571428577e-06,
|
4621 |
+
"loss": 0.1241,
|
4622 |
+
"step": 18550
|
4623 |
+
},
|
4624 |
+
{
|
4625 |
+
"epoch": 0.64,
|
4626 |
+
"learning_rate": 2.6269387755102043e-06,
|
4627 |
+
"loss": 0.1574,
|
4628 |
+
"step": 18575
|
4629 |
+
},
|
4630 |
+
{
|
4631 |
+
"epoch": 0.65,
|
4632 |
+
"learning_rate": 2.6167346938775514e-06,
|
4633 |
+
"loss": 0.1365,
|
4634 |
+
"step": 18600
|
4635 |
+
},
|
4636 |
+
{
|
4637 |
+
"epoch": 0.65,
|
4638 |
+
"learning_rate": 2.606530612244898e-06,
|
4639 |
+
"loss": 0.1641,
|
4640 |
+
"step": 18625
|
4641 |
+
},
|
4642 |
+
{
|
4643 |
+
"epoch": 0.65,
|
4644 |
+
"learning_rate": 2.5963265306122455e-06,
|
4645 |
+
"loss": 0.1235,
|
4646 |
+
"step": 18650
|
4647 |
+
},
|
4648 |
+
{
|
4649 |
+
"epoch": 0.65,
|
4650 |
+
"learning_rate": 2.586122448979592e-06,
|
4651 |
+
"loss": 0.1604,
|
4652 |
+
"step": 18675
|
4653 |
+
},
|
4654 |
+
{
|
4655 |
+
"epoch": 0.65,
|
4656 |
+
"learning_rate": 2.575918367346939e-06,
|
4657 |
+
"loss": 0.1296,
|
4658 |
+
"step": 18700
|
4659 |
+
},
|
4660 |
+
{
|
4661 |
+
"epoch": 0.65,
|
4662 |
+
"learning_rate": 2.565714285714286e-06,
|
4663 |
+
"loss": 0.1757,
|
4664 |
+
"step": 18725
|
4665 |
+
},
|
4666 |
+
{
|
4667 |
+
"epoch": 0.65,
|
4668 |
+
"learning_rate": 2.5555102040816325e-06,
|
4669 |
+
"loss": 0.1293,
|
4670 |
+
"step": 18750
|
4671 |
+
},
|
4672 |
+
{
|
4673 |
+
"epoch": 0.65,
|
4674 |
+
"learning_rate": 2.54530612244898e-06,
|
4675 |
+
"loss": 0.1602,
|
4676 |
+
"step": 18775
|
4677 |
+
},
|
4678 |
+
{
|
4679 |
+
"epoch": 0.65,
|
4680 |
+
"learning_rate": 2.535102040816327e-06,
|
4681 |
+
"loss": 0.1641,
|
4682 |
+
"step": 18800
|
4683 |
+
},
|
4684 |
+
{
|
4685 |
+
"epoch": 0.65,
|
4686 |
+
"learning_rate": 2.5248979591836737e-06,
|
4687 |
+
"loss": 0.1445,
|
4688 |
+
"step": 18825
|
4689 |
+
},
|
4690 |
+
{
|
4691 |
+
"epoch": 0.65,
|
4692 |
+
"learning_rate": 2.5146938775510203e-06,
|
4693 |
+
"loss": 0.1292,
|
4694 |
+
"step": 18850
|
4695 |
+
},
|
4696 |
+
{
|
4697 |
+
"epoch": 0.66,
|
4698 |
+
"learning_rate": 2.5044897959183678e-06,
|
4699 |
+
"loss": 0.1642,
|
4700 |
+
"step": 18875
|
4701 |
+
},
|
4702 |
+
{
|
4703 |
+
"epoch": 0.66,
|
4704 |
+
"learning_rate": 2.4942857142857144e-06,
|
4705 |
+
"loss": 0.13,
|
4706 |
+
"step": 18900
|
4707 |
+
},
|
4708 |
+
{
|
4709 |
+
"epoch": 0.66,
|
4710 |
+
"learning_rate": 2.4840816326530615e-06,
|
4711 |
+
"loss": 0.1527,
|
4712 |
+
"step": 18925
|
4713 |
+
},
|
4714 |
+
{
|
4715 |
+
"epoch": 0.66,
|
4716 |
+
"learning_rate": 2.473877551020408e-06,
|
4717 |
+
"loss": 0.1325,
|
4718 |
+
"step": 18950
|
4719 |
+
},
|
4720 |
+
{
|
4721 |
+
"epoch": 0.66,
|
4722 |
+
"learning_rate": 2.463673469387755e-06,
|
4723 |
+
"loss": 0.1569,
|
4724 |
+
"step": 18975
|
4725 |
+
},
|
4726 |
+
{
|
4727 |
+
"epoch": 0.66,
|
4728 |
+
"learning_rate": 2.4534693877551023e-06,
|
4729 |
+
"loss": 0.148,
|
4730 |
+
"step": 19000
|
4731 |
+
},
|
4732 |
+
{
|
4733 |
+
"epoch": 0.66,
|
4734 |
+
"eval_loss": 0.15705841779708862,
|
4735 |
+
"eval_runtime": 31110.7125,
|
4736 |
+
"eval_samples_per_second": 0.998,
|
4737 |
+
"eval_steps_per_second": 0.062,
|
4738 |
+
"eval_wer": 0.08945159383534276,
|
4739 |
+
"step": 19000
|
4740 |
}
|
4741 |
],
|
4742 |
"max_steps": 25000,
|
4743 |
"num_train_epochs": 1,
|
4744 |
+
"total_flos": 6.454439903232e+20,
|
4745 |
"trial_name": null,
|
4746 |
"trial_params": null
|
4747 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:15b5954cb47199f441cbbb70307c72621abf9248aa385660bfdb147a516624ef
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6547ad8d8b1ed3f418b525a96f2b38ff44683cdf8d77997c10dc189dea1b0ef9
|
3 |
+
size 130084
|