elsayedissa commited on
Commit
af4dead
1 Parent(s): ffc11ff

Training in progress, step 24000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2daba987c3635c07dbbfcc0f0b00435568b055cd9e65badd42c48ee0e8cc6eb7
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd7eb848685ef3f06accf08b0991ae5fb7aca4770e2f824f2a4e024e639e7fb2
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7110b14f56f7f4c10e6b71fbe6f5251480a4ae2ea07b69ac339d59349d6f9560
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e71d25993f756e62aa28d38555706c7d09a30a1163fed42767f77816851100d
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ee749995a148c38ebbb92f25aca7282f391ce6f8f7323887d7232d39b5ea5d4
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9eaf0badf99eb2f5a4bcf7352b7e7e60bdb9fb3eb0aafd7a775251c7a5269cfd
3
+ size 14511
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13a7589714ca42744cc53ba86bd24fc9f77beb3a4a23c480fa17e3779ac95997
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf5ce6709b8cf97417772fb30cceb44ce3363ceb800bc698892d9018ba93c60f
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0261b0ef9ff8bcf7f335c8ff2f03f795976c908bbc906ae5bd82b9ccd45be639
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0b3ee000403ded4a43ceaf77d66b9d28960f5cc5d0851756133984c89e3a2c4
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.7983616231038911,
5
- "global_step": 23000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -5733,11 +5733,260 @@
5733
  "eval_steps_per_second": 0.062,
5734
  "eval_wer": 0.08671732332088382,
5735
  "step": 23000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5736
  }
5737
  ],
5738
  "max_steps": 25000,
5739
  "num_train_epochs": 1,
5740
- "total_flos": 7.813269356544e+20,
5741
  "trial_name": null,
5742
  "trial_params": null
5743
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.8330729980214516,
5
+ "global_step": 24000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
5733
  "eval_steps_per_second": 0.062,
5734
  "eval_wer": 0.08671732332088382,
5735
  "step": 23000
5736
+ },
5737
+ {
5738
+ "epoch": 0.8,
5739
+ "learning_rate": 8.110204081632655e-07,
5740
+ "loss": 0.1585,
5741
+ "step": 23025
5742
+ },
5743
+ {
5744
+ "epoch": 0.8,
5745
+ "learning_rate": 8.008163265306124e-07,
5746
+ "loss": 0.1183,
5747
+ "step": 23050
5748
+ },
5749
+ {
5750
+ "epoch": 0.8,
5751
+ "learning_rate": 7.906122448979592e-07,
5752
+ "loss": 0.1367,
5753
+ "step": 23075
5754
+ },
5755
+ {
5756
+ "epoch": 0.8,
5757
+ "learning_rate": 7.804081632653061e-07,
5758
+ "loss": 0.1175,
5759
+ "step": 23100
5760
+ },
5761
+ {
5762
+ "epoch": 0.8,
5763
+ "learning_rate": 7.702040816326531e-07,
5764
+ "loss": 0.1487,
5765
+ "step": 23125
5766
+ },
5767
+ {
5768
+ "epoch": 0.8,
5769
+ "learning_rate": 7.6e-07,
5770
+ "loss": 0.119,
5771
+ "step": 23150
5772
+ },
5773
+ {
5774
+ "epoch": 0.8,
5775
+ "learning_rate": 7.49795918367347e-07,
5776
+ "loss": 0.1402,
5777
+ "step": 23175
5778
+ },
5779
+ {
5780
+ "epoch": 0.81,
5781
+ "learning_rate": 7.395918367346939e-07,
5782
+ "loss": 0.1307,
5783
+ "step": 23200
5784
+ },
5785
+ {
5786
+ "epoch": 0.81,
5787
+ "learning_rate": 7.293877551020409e-07,
5788
+ "loss": 0.1547,
5789
+ "step": 23225
5790
+ },
5791
+ {
5792
+ "epoch": 0.81,
5793
+ "learning_rate": 7.191836734693879e-07,
5794
+ "loss": 0.1208,
5795
+ "step": 23250
5796
+ },
5797
+ {
5798
+ "epoch": 0.81,
5799
+ "learning_rate": 7.089795918367348e-07,
5800
+ "loss": 0.1647,
5801
+ "step": 23275
5802
+ },
5803
+ {
5804
+ "epoch": 0.81,
5805
+ "learning_rate": 6.987755102040817e-07,
5806
+ "loss": 0.1025,
5807
+ "step": 23300
5808
+ },
5809
+ {
5810
+ "epoch": 0.81,
5811
+ "learning_rate": 6.885714285714286e-07,
5812
+ "loss": 0.1778,
5813
+ "step": 23325
5814
+ },
5815
+ {
5816
+ "epoch": 0.81,
5817
+ "learning_rate": 6.783673469387756e-07,
5818
+ "loss": 0.1232,
5819
+ "step": 23350
5820
+ },
5821
+ {
5822
+ "epoch": 0.81,
5823
+ "learning_rate": 6.681632653061225e-07,
5824
+ "loss": 0.1301,
5825
+ "step": 23375
5826
+ },
5827
+ {
5828
+ "epoch": 0.81,
5829
+ "learning_rate": 6.579591836734695e-07,
5830
+ "loss": 0.1401,
5831
+ "step": 23400
5832
+ },
5833
+ {
5834
+ "epoch": 0.81,
5835
+ "learning_rate": 6.477551020408164e-07,
5836
+ "loss": 0.1588,
5837
+ "step": 23425
5838
+ },
5839
+ {
5840
+ "epoch": 0.81,
5841
+ "learning_rate": 6.375510204081634e-07,
5842
+ "loss": 0.1226,
5843
+ "step": 23450
5844
+ },
5845
+ {
5846
+ "epoch": 0.81,
5847
+ "learning_rate": 6.273469387755104e-07,
5848
+ "loss": 0.1594,
5849
+ "step": 23475
5850
+ },
5851
+ {
5852
+ "epoch": 0.82,
5853
+ "learning_rate": 6.171428571428572e-07,
5854
+ "loss": 0.1341,
5855
+ "step": 23500
5856
+ },
5857
+ {
5858
+ "epoch": 0.82,
5859
+ "learning_rate": 6.069387755102042e-07,
5860
+ "loss": 0.1606,
5861
+ "step": 23525
5862
+ },
5863
+ {
5864
+ "epoch": 0.82,
5865
+ "learning_rate": 5.96734693877551e-07,
5866
+ "loss": 0.1174,
5867
+ "step": 23550
5868
+ },
5869
+ {
5870
+ "epoch": 0.82,
5871
+ "learning_rate": 5.86530612244898e-07,
5872
+ "loss": 0.156,
5873
+ "step": 23575
5874
+ },
5875
+ {
5876
+ "epoch": 0.82,
5877
+ "learning_rate": 5.763265306122449e-07,
5878
+ "loss": 0.1323,
5879
+ "step": 23600
5880
+ },
5881
+ {
5882
+ "epoch": 0.82,
5883
+ "learning_rate": 5.661224489795919e-07,
5884
+ "loss": 0.1532,
5885
+ "step": 23625
5886
+ },
5887
+ {
5888
+ "epoch": 0.82,
5889
+ "learning_rate": 5.559183673469388e-07,
5890
+ "loss": 0.119,
5891
+ "step": 23650
5892
+ },
5893
+ {
5894
+ "epoch": 0.82,
5895
+ "learning_rate": 5.457142857142857e-07,
5896
+ "loss": 0.1476,
5897
+ "step": 23675
5898
+ },
5899
+ {
5900
+ "epoch": 0.82,
5901
+ "learning_rate": 5.355102040816326e-07,
5902
+ "loss": 0.1331,
5903
+ "step": 23700
5904
+ },
5905
+ {
5906
+ "epoch": 0.82,
5907
+ "learning_rate": 5.253061224489796e-07,
5908
+ "loss": 0.1498,
5909
+ "step": 23725
5910
+ },
5911
+ {
5912
+ "epoch": 0.82,
5913
+ "learning_rate": 5.151020408163266e-07,
5914
+ "loss": 0.1268,
5915
+ "step": 23750
5916
+ },
5917
+ {
5918
+ "epoch": 0.83,
5919
+ "learning_rate": 5.048979591836735e-07,
5920
+ "loss": 0.1327,
5921
+ "step": 23775
5922
+ },
5923
+ {
5924
+ "epoch": 0.83,
5925
+ "learning_rate": 4.946938775510205e-07,
5926
+ "loss": 0.1253,
5927
+ "step": 23800
5928
+ },
5929
+ {
5930
+ "epoch": 0.83,
5931
+ "learning_rate": 4.844897959183674e-07,
5932
+ "loss": 0.1756,
5933
+ "step": 23825
5934
+ },
5935
+ {
5936
+ "epoch": 0.83,
5937
+ "learning_rate": 4.7428571428571434e-07,
5938
+ "loss": 0.1487,
5939
+ "step": 23850
5940
+ },
5941
+ {
5942
+ "epoch": 0.83,
5943
+ "learning_rate": 4.6408163265306124e-07,
5944
+ "loss": 0.1502,
5945
+ "step": 23875
5946
+ },
5947
+ {
5948
+ "epoch": 0.83,
5949
+ "learning_rate": 4.538775510204082e-07,
5950
+ "loss": 0.1231,
5951
+ "step": 23900
5952
+ },
5953
+ {
5954
+ "epoch": 0.83,
5955
+ "learning_rate": 4.4367346938775515e-07,
5956
+ "loss": 0.1352,
5957
+ "step": 23925
5958
+ },
5959
+ {
5960
+ "epoch": 0.83,
5961
+ "learning_rate": 4.334693877551021e-07,
5962
+ "loss": 0.141,
5963
+ "step": 23950
5964
+ },
5965
+ {
5966
+ "epoch": 0.83,
5967
+ "learning_rate": 4.2326530612244896e-07,
5968
+ "loss": 0.1539,
5969
+ "step": 23975
5970
+ },
5971
+ {
5972
+ "epoch": 0.83,
5973
+ "learning_rate": 4.130612244897959e-07,
5974
+ "loss": 0.1337,
5975
+ "step": 24000
5976
+ },
5977
+ {
5978
+ "epoch": 0.83,
5979
+ "eval_loss": 0.1472465544939041,
5980
+ "eval_runtime": 31230.9152,
5981
+ "eval_samples_per_second": 0.994,
5982
+ "eval_steps_per_second": 0.062,
5983
+ "eval_wer": 0.08535018806365434,
5984
+ "step": 24000
5985
  }
5986
  ],
5987
  "max_steps": 25000,
5988
  "num_train_epochs": 1,
5989
+ "total_flos": 8.152976719872e+20,
5990
  "trial_name": null,
5991
  "trial_params": null
5992
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7110b14f56f7f4c10e6b71fbe6f5251480a4ae2ea07b69ac339d59349d6f9560
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e71d25993f756e62aa28d38555706c7d09a30a1163fed42767f77816851100d
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40bb4a9db04f1abcaa80cc21618c523855a5b501d308ce0a0b153c875673d03d
3
- size 156980
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3be1bf9d6afc1185ac6ba7539df5045fb2a516aaf0676ce7cb1d8b30a8903c18
3
+ size 163704