elsayedissa commited on
Commit
0ae67d4
1 Parent(s): 7d368e5

Training in progress, step 19000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8a90d99fa5e493d103776713ee2ee6b9fca9948380c5ad509565e448a0d3c2f
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:181c2bd989506934f36609329ae3277eb855e8acf49b3c12e326fb79cb004eb3
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b690154ee77cdbea8a14f2965218b2b3790caf8955482e539a12fa93fdc53f8
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15b5954cb47199f441cbbb70307c72621abf9248aa385660bfdb147a516624ef
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:537f8d4bc472f85d8aaced00700f04502f899659cbfa067468feee29a0f0233a
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9460131d11a73df83270a4796c5192126900329b0b2274773520724725f4b12d
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90be1018263c5a8b0b93a055ffb96e8625ade303fea979c01cbf658c8aed8c95
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6a09c97ace5e5a49360efd6cac2da516ffa83c9ee08cb64155c59b3a316ab7a
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9205cfac3dc2d508e85efbdf6cfda796cb8dde46c8f1bd9e062253fea417d2b3
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec3490e7ac28e19432d614c125578d2d66d232f9a51b472efd2d122fe9c6b708
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.6248047485160887,
5
- "global_step": 18000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -4488,11 +4488,260 @@
4488
  "eval_steps_per_second": 0.062,
4489
  "eval_wer": 0.08993702591943148,
4490
  "step": 18000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4491
  }
4492
  ],
4493
  "max_steps": 25000,
4494
  "num_train_epochs": 1,
4495
- "total_flos": 6.114732539904e+20,
4496
  "trial_name": null,
4497
  "trial_params": null
4498
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.6595161234336492,
5
+ "global_step": 19000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
4488
  "eval_steps_per_second": 0.062,
4489
  "eval_wer": 0.08993702591943148,
4490
  "step": 18000
4491
+ },
4492
+ {
4493
+ "epoch": 0.63,
4494
+ "learning_rate": 2.8514285714285715e-06,
4495
+ "loss": 0.1658,
4496
+ "step": 18025
4497
+ },
4498
+ {
4499
+ "epoch": 0.63,
4500
+ "learning_rate": 2.8412244897959185e-06,
4501
+ "loss": 0.1237,
4502
+ "step": 18050
4503
+ },
4504
+ {
4505
+ "epoch": 0.63,
4506
+ "learning_rate": 2.831020408163265e-06,
4507
+ "loss": 0.1672,
4508
+ "step": 18075
4509
+ },
4510
+ {
4511
+ "epoch": 0.63,
4512
+ "learning_rate": 2.8208163265306127e-06,
4513
+ "loss": 0.1319,
4514
+ "step": 18100
4515
+ },
4516
+ {
4517
+ "epoch": 0.63,
4518
+ "learning_rate": 2.8106122448979593e-06,
4519
+ "loss": 0.1543,
4520
+ "step": 18125
4521
+ },
4522
+ {
4523
+ "epoch": 0.63,
4524
+ "learning_rate": 2.8004081632653064e-06,
4525
+ "loss": 0.1167,
4526
+ "step": 18150
4527
+ },
4528
+ {
4529
+ "epoch": 0.63,
4530
+ "learning_rate": 2.790204081632653e-06,
4531
+ "loss": 0.1708,
4532
+ "step": 18175
4533
+ },
4534
+ {
4535
+ "epoch": 0.63,
4536
+ "learning_rate": 2.7800000000000005e-06,
4537
+ "loss": 0.1265,
4538
+ "step": 18200
4539
+ },
4540
+ {
4541
+ "epoch": 0.63,
4542
+ "learning_rate": 2.769795918367347e-06,
4543
+ "loss": 0.1588,
4544
+ "step": 18225
4545
+ },
4546
+ {
4547
+ "epoch": 0.63,
4548
+ "learning_rate": 2.759591836734694e-06,
4549
+ "loss": 0.133,
4550
+ "step": 18250
4551
+ },
4552
+ {
4553
+ "epoch": 0.63,
4554
+ "learning_rate": 2.749387755102041e-06,
4555
+ "loss": 0.1744,
4556
+ "step": 18275
4557
+ },
4558
+ {
4559
+ "epoch": 0.64,
4560
+ "learning_rate": 2.739183673469388e-06,
4561
+ "loss": 0.1305,
4562
+ "step": 18300
4563
+ },
4564
+ {
4565
+ "epoch": 0.64,
4566
+ "learning_rate": 2.728979591836735e-06,
4567
+ "loss": 0.1486,
4568
+ "step": 18325
4569
+ },
4570
+ {
4571
+ "epoch": 0.64,
4572
+ "learning_rate": 2.718775510204082e-06,
4573
+ "loss": 0.1379,
4574
+ "step": 18350
4575
+ },
4576
+ {
4577
+ "epoch": 0.64,
4578
+ "learning_rate": 2.7085714285714287e-06,
4579
+ "loss": 0.1669,
4580
+ "step": 18375
4581
+ },
4582
+ {
4583
+ "epoch": 0.64,
4584
+ "learning_rate": 2.6983673469387757e-06,
4585
+ "loss": 0.1447,
4586
+ "step": 18400
4587
+ },
4588
+ {
4589
+ "epoch": 0.64,
4590
+ "learning_rate": 2.6881632653061228e-06,
4591
+ "loss": 0.1751,
4592
+ "step": 18425
4593
+ },
4594
+ {
4595
+ "epoch": 0.64,
4596
+ "learning_rate": 2.67795918367347e-06,
4597
+ "loss": 0.1259,
4598
+ "step": 18450
4599
+ },
4600
+ {
4601
+ "epoch": 0.64,
4602
+ "learning_rate": 2.6677551020408165e-06,
4603
+ "loss": 0.1406,
4604
+ "step": 18475
4605
+ },
4606
+ {
4607
+ "epoch": 0.64,
4608
+ "learning_rate": 2.6575510204081635e-06,
4609
+ "loss": 0.1463,
4610
+ "step": 18500
4611
+ },
4612
+ {
4613
+ "epoch": 0.64,
4614
+ "learning_rate": 2.64734693877551e-06,
4615
+ "loss": 0.164,
4616
+ "step": 18525
4617
+ },
4618
+ {
4619
+ "epoch": 0.64,
4620
+ "learning_rate": 2.6371428571428577e-06,
4621
+ "loss": 0.1241,
4622
+ "step": 18550
4623
+ },
4624
+ {
4625
+ "epoch": 0.64,
4626
+ "learning_rate": 2.6269387755102043e-06,
4627
+ "loss": 0.1574,
4628
+ "step": 18575
4629
+ },
4630
+ {
4631
+ "epoch": 0.65,
4632
+ "learning_rate": 2.6167346938775514e-06,
4633
+ "loss": 0.1365,
4634
+ "step": 18600
4635
+ },
4636
+ {
4637
+ "epoch": 0.65,
4638
+ "learning_rate": 2.606530612244898e-06,
4639
+ "loss": 0.1641,
4640
+ "step": 18625
4641
+ },
4642
+ {
4643
+ "epoch": 0.65,
4644
+ "learning_rate": 2.5963265306122455e-06,
4645
+ "loss": 0.1235,
4646
+ "step": 18650
4647
+ },
4648
+ {
4649
+ "epoch": 0.65,
4650
+ "learning_rate": 2.586122448979592e-06,
4651
+ "loss": 0.1604,
4652
+ "step": 18675
4653
+ },
4654
+ {
4655
+ "epoch": 0.65,
4656
+ "learning_rate": 2.575918367346939e-06,
4657
+ "loss": 0.1296,
4658
+ "step": 18700
4659
+ },
4660
+ {
4661
+ "epoch": 0.65,
4662
+ "learning_rate": 2.565714285714286e-06,
4663
+ "loss": 0.1757,
4664
+ "step": 18725
4665
+ },
4666
+ {
4667
+ "epoch": 0.65,
4668
+ "learning_rate": 2.5555102040816325e-06,
4669
+ "loss": 0.1293,
4670
+ "step": 18750
4671
+ },
4672
+ {
4673
+ "epoch": 0.65,
4674
+ "learning_rate": 2.54530612244898e-06,
4675
+ "loss": 0.1602,
4676
+ "step": 18775
4677
+ },
4678
+ {
4679
+ "epoch": 0.65,
4680
+ "learning_rate": 2.535102040816327e-06,
4681
+ "loss": 0.1641,
4682
+ "step": 18800
4683
+ },
4684
+ {
4685
+ "epoch": 0.65,
4686
+ "learning_rate": 2.5248979591836737e-06,
4687
+ "loss": 0.1445,
4688
+ "step": 18825
4689
+ },
4690
+ {
4691
+ "epoch": 0.65,
4692
+ "learning_rate": 2.5146938775510203e-06,
4693
+ "loss": 0.1292,
4694
+ "step": 18850
4695
+ },
4696
+ {
4697
+ "epoch": 0.66,
4698
+ "learning_rate": 2.5044897959183678e-06,
4699
+ "loss": 0.1642,
4700
+ "step": 18875
4701
+ },
4702
+ {
4703
+ "epoch": 0.66,
4704
+ "learning_rate": 2.4942857142857144e-06,
4705
+ "loss": 0.13,
4706
+ "step": 18900
4707
+ },
4708
+ {
4709
+ "epoch": 0.66,
4710
+ "learning_rate": 2.4840816326530615e-06,
4711
+ "loss": 0.1527,
4712
+ "step": 18925
4713
+ },
4714
+ {
4715
+ "epoch": 0.66,
4716
+ "learning_rate": 2.473877551020408e-06,
4717
+ "loss": 0.1325,
4718
+ "step": 18950
4719
+ },
4720
+ {
4721
+ "epoch": 0.66,
4722
+ "learning_rate": 2.463673469387755e-06,
4723
+ "loss": 0.1569,
4724
+ "step": 18975
4725
+ },
4726
+ {
4727
+ "epoch": 0.66,
4728
+ "learning_rate": 2.4534693877551023e-06,
4729
+ "loss": 0.148,
4730
+ "step": 19000
4731
+ },
4732
+ {
4733
+ "epoch": 0.66,
4734
+ "eval_loss": 0.15705841779708862,
4735
+ "eval_runtime": 31110.7125,
4736
+ "eval_samples_per_second": 0.998,
4737
+ "eval_steps_per_second": 0.062,
4738
+ "eval_wer": 0.08945159383534276,
4739
+ "step": 19000
4740
  }
4741
  ],
4742
  "max_steps": 25000,
4743
  "num_train_epochs": 1,
4744
+ "total_flos": 6.454439903232e+20,
4745
  "trial_name": null,
4746
  "trial_params": null
4747
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b690154ee77cdbea8a14f2965218b2b3790caf8955482e539a12fa93fdc53f8
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15b5954cb47199f441cbbb70307c72621abf9248aa385660bfdb147a516624ef
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54f4f30667756976745aebfb4e84b0aa810e69c9ffaa1a0e78323a5e4c0623c1
3
- size 123360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6547ad8d8b1ed3f418b525a96f2b38ff44683cdf8d77997c10dc189dea1b0ef9
3
+ size 130084