elsayedissa commited on
Commit
7fb866c
1 Parent(s): f8064d0

Training in progress, step 12000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b15938f7058276e6bdfc2fddcc7a120da5e6a0c7c9c19c71e16cb873fc6886b3
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:178b5b502a485d7c2fe861b88b73958c4bdd88fbe0bd804aaac864b4c3d58b52
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d40f3344c48b8c1de58d7e8467befdcdcde144b9ff8ed4465cd30973a433d738
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:069bc79c0b97b8bebabb6777e867a3617901ff0e79b431f2ba4a791a90815e7f
3
- size 14639
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e348aa5f9f6cf6306d6ba94fb59dc91d14cd792144b93c44e821af2cb774ce
3
+ size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b4b38bb567a4c3bd8aafad6c492b89fde494be62ffddc238da23107b19a9945
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3ab4d8f5a2206989f1574e42f0b2264f32c39696956f3c8247ab96f9f887022
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e9a74567933ed8c342d4d89869156924da4e6db37efcc7105e4c34e754e42a4
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e22e62dec3718a0ec90f6becc88cb24f88c5c482567956f03f4b33d800c0df20
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.3818251240931653,
5
- "global_step": 11000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -2745,11 +2745,260 @@
2745
  "eval_steps_per_second": 0.062,
2746
  "eval_wer": 0.09984380314573199,
2747
  "step": 11000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2748
  }
2749
  ],
2750
  "max_steps": 25000,
2751
  "num_train_epochs": 1,
2752
- "total_flos": 3.736780996608e+20,
2753
  "trial_name": null,
2754
  "trial_params": null
2755
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.4165364990107258,
5
+ "global_step": 12000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
2745
  "eval_steps_per_second": 0.062,
2746
  "eval_wer": 0.09984380314573199,
2747
  "step": 11000
2748
+ },
2749
+ {
2750
+ "epoch": 0.38,
2751
+ "learning_rate": 5.7073469387755105e-06,
2752
+ "loss": 0.1738,
2753
+ "step": 11025
2754
+ },
2755
+ {
2756
+ "epoch": 0.38,
2757
+ "learning_rate": 5.697142857142858e-06,
2758
+ "loss": 0.156,
2759
+ "step": 11050
2760
+ },
2761
+ {
2762
+ "epoch": 0.38,
2763
+ "learning_rate": 5.686938775510205e-06,
2764
+ "loss": 0.1786,
2765
+ "step": 11075
2766
+ },
2767
+ {
2768
+ "epoch": 0.39,
2769
+ "learning_rate": 5.676734693877551e-06,
2770
+ "loss": 0.1681,
2771
+ "step": 11100
2772
+ },
2773
+ {
2774
+ "epoch": 0.39,
2775
+ "learning_rate": 5.666530612244898e-06,
2776
+ "loss": 0.1678,
2777
+ "step": 11125
2778
+ },
2779
+ {
2780
+ "epoch": 0.39,
2781
+ "learning_rate": 5.656326530612245e-06,
2782
+ "loss": 0.1696,
2783
+ "step": 11150
2784
+ },
2785
+ {
2786
+ "epoch": 0.39,
2787
+ "learning_rate": 5.646122448979592e-06,
2788
+ "loss": 0.1693,
2789
+ "step": 11175
2790
+ },
2791
+ {
2792
+ "epoch": 0.39,
2793
+ "learning_rate": 5.6359183673469395e-06,
2794
+ "loss": 0.1342,
2795
+ "step": 11200
2796
+ },
2797
+ {
2798
+ "epoch": 0.39,
2799
+ "learning_rate": 5.625714285714286e-06,
2800
+ "loss": 0.2051,
2801
+ "step": 11225
2802
+ },
2803
+ {
2804
+ "epoch": 0.39,
2805
+ "learning_rate": 5.615510204081634e-06,
2806
+ "loss": 0.1534,
2807
+ "step": 11250
2808
+ },
2809
+ {
2810
+ "epoch": 0.39,
2811
+ "learning_rate": 5.60530612244898e-06,
2812
+ "loss": 0.1683,
2813
+ "step": 11275
2814
+ },
2815
+ {
2816
+ "epoch": 0.39,
2817
+ "learning_rate": 5.595102040816327e-06,
2818
+ "loss": 0.1524,
2819
+ "step": 11300
2820
+ },
2821
+ {
2822
+ "epoch": 0.39,
2823
+ "learning_rate": 5.584897959183674e-06,
2824
+ "loss": 0.184,
2825
+ "step": 11325
2826
+ },
2827
+ {
2828
+ "epoch": 0.39,
2829
+ "learning_rate": 5.574693877551021e-06,
2830
+ "loss": 0.1486,
2831
+ "step": 11350
2832
+ },
2833
+ {
2834
+ "epoch": 0.39,
2835
+ "learning_rate": 5.564489795918368e-06,
2836
+ "loss": 0.1685,
2837
+ "step": 11375
2838
+ },
2839
+ {
2840
+ "epoch": 0.4,
2841
+ "learning_rate": 5.554285714285714e-06,
2842
+ "loss": 0.1645,
2843
+ "step": 11400
2844
+ },
2845
+ {
2846
+ "epoch": 0.4,
2847
+ "learning_rate": 5.544081632653062e-06,
2848
+ "loss": 0.1743,
2849
+ "step": 11425
2850
+ },
2851
+ {
2852
+ "epoch": 0.4,
2853
+ "learning_rate": 5.533877551020409e-06,
2854
+ "loss": 0.1377,
2855
+ "step": 11450
2856
+ },
2857
+ {
2858
+ "epoch": 0.4,
2859
+ "learning_rate": 5.524489795918367e-06,
2860
+ "loss": 0.1776,
2861
+ "step": 11475
2862
+ },
2863
+ {
2864
+ "epoch": 0.4,
2865
+ "learning_rate": 5.514285714285714e-06,
2866
+ "loss": 0.1365,
2867
+ "step": 11500
2868
+ },
2869
+ {
2870
+ "epoch": 0.4,
2871
+ "learning_rate": 5.504081632653062e-06,
2872
+ "loss": 0.1777,
2873
+ "step": 11525
2874
+ },
2875
+ {
2876
+ "epoch": 0.4,
2877
+ "learning_rate": 5.493877551020409e-06,
2878
+ "loss": 0.149,
2879
+ "step": 11550
2880
+ },
2881
+ {
2882
+ "epoch": 0.4,
2883
+ "learning_rate": 5.4836734693877555e-06,
2884
+ "loss": 0.1855,
2885
+ "step": 11575
2886
+ },
2887
+ {
2888
+ "epoch": 0.4,
2889
+ "learning_rate": 5.473469387755103e-06,
2890
+ "loss": 0.1543,
2891
+ "step": 11600
2892
+ },
2893
+ {
2894
+ "epoch": 0.4,
2895
+ "learning_rate": 5.46326530612245e-06,
2896
+ "loss": 0.1817,
2897
+ "step": 11625
2898
+ },
2899
+ {
2900
+ "epoch": 0.4,
2901
+ "learning_rate": 5.453061224489796e-06,
2902
+ "loss": 0.1614,
2903
+ "step": 11650
2904
+ },
2905
+ {
2906
+ "epoch": 0.41,
2907
+ "learning_rate": 5.442857142857143e-06,
2908
+ "loss": 0.1544,
2909
+ "step": 11675
2910
+ },
2911
+ {
2912
+ "epoch": 0.41,
2913
+ "learning_rate": 5.4326530612244895e-06,
2914
+ "loss": 0.1592,
2915
+ "step": 11700
2916
+ },
2917
+ {
2918
+ "epoch": 0.41,
2919
+ "learning_rate": 5.422448979591837e-06,
2920
+ "loss": 0.1658,
2921
+ "step": 11725
2922
+ },
2923
+ {
2924
+ "epoch": 0.41,
2925
+ "learning_rate": 5.4122448979591845e-06,
2926
+ "loss": 0.1596,
2927
+ "step": 11750
2928
+ },
2929
+ {
2930
+ "epoch": 0.41,
2931
+ "learning_rate": 5.402040816326531e-06,
2932
+ "loss": 0.1707,
2933
+ "step": 11775
2934
+ },
2935
+ {
2936
+ "epoch": 0.41,
2937
+ "learning_rate": 5.391836734693879e-06,
2938
+ "loss": 0.1383,
2939
+ "step": 11800
2940
+ },
2941
+ {
2942
+ "epoch": 0.41,
2943
+ "learning_rate": 5.381632653061225e-06,
2944
+ "loss": 0.175,
2945
+ "step": 11825
2946
+ },
2947
+ {
2948
+ "epoch": 0.41,
2949
+ "learning_rate": 5.371428571428572e-06,
2950
+ "loss": 0.1658,
2951
+ "step": 11850
2952
+ },
2953
+ {
2954
+ "epoch": 0.41,
2955
+ "learning_rate": 5.3612244897959186e-06,
2956
+ "loss": 0.1864,
2957
+ "step": 11875
2958
+ },
2959
+ {
2960
+ "epoch": 0.41,
2961
+ "learning_rate": 5.351020408163265e-06,
2962
+ "loss": 0.1441,
2963
+ "step": 11900
2964
+ },
2965
+ {
2966
+ "epoch": 0.41,
2967
+ "learning_rate": 5.340816326530613e-06,
2968
+ "loss": 0.1559,
2969
+ "step": 11925
2970
+ },
2971
+ {
2972
+ "epoch": 0.41,
2973
+ "learning_rate": 5.330612244897959e-06,
2974
+ "loss": 0.1559,
2975
+ "step": 11950
2976
+ },
2977
+ {
2978
+ "epoch": 0.42,
2979
+ "learning_rate": 5.320408163265307e-06,
2980
+ "loss": 0.1786,
2981
+ "step": 11975
2982
+ },
2983
+ {
2984
+ "epoch": 0.42,
2985
+ "learning_rate": 5.310204081632654e-06,
2986
+ "loss": 0.169,
2987
+ "step": 12000
2988
+ },
2989
+ {
2990
+ "epoch": 0.42,
2991
+ "eval_loss": 0.17712195217609406,
2992
+ "eval_runtime": 30926.8901,
2993
+ "eval_samples_per_second": 1.004,
2994
+ "eval_steps_per_second": 0.063,
2995
+ "eval_wer": 0.09824881201229761,
2996
+ "step": 12000
2997
  }
2998
  ],
2999
  "max_steps": 25000,
3000
  "num_train_epochs": 1,
3001
+ "total_flos": 4.076488359936e+20,
3002
  "trial_name": null,
3003
  "trial_params": null
3004
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bdd501f6153777369e4dc961faa1bf60f1f7535ded5d73c87f2ff06b5aa8064
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d40f3344c48b8c1de58d7e8467befdcdcde144b9ff8ed4465cd30973a433d738
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c954f05ee34517b3eb4c4f0da7bbc163f490b1b0ac3836bb3e4f0baf08b45f5
3
- size 76967
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc505845f9a549bc125c53e3e46d54bfa0b789b17c0fc518106e8b0e000bbc35
3
+ size 83565