elsayedissa
commited on
Commit
•
7fb866c
1
Parent(s):
f8064d0
Training in progress, step 12000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +2 -2
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:178b5b502a485d7c2fe861b88b73958c4bdd88fbe0bd804aaac864b4c3d58b52
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d40f3344c48b8c1de58d7e8467befdcdcde144b9ff8ed4465cd30973a433d738
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35e348aa5f9f6cf6306d6ba94fb59dc91d14cd792144b93c44e821af2cb774ce
|
3 |
+
size 14575
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3ab4d8f5a2206989f1574e42f0b2264f32c39696956f3c8247ab96f9f887022
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e22e62dec3718a0ec90f6becc88cb24f88c5c482567956f03f4b33d800c0df20
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -2745,11 +2745,260 @@
|
|
2745 |
"eval_steps_per_second": 0.062,
|
2746 |
"eval_wer": 0.09984380314573199,
|
2747 |
"step": 11000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2748 |
}
|
2749 |
],
|
2750 |
"max_steps": 25000,
|
2751 |
"num_train_epochs": 1,
|
2752 |
-
"total_flos":
|
2753 |
"trial_name": null,
|
2754 |
"trial_params": null
|
2755 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.4165364990107258,
|
5 |
+
"global_step": 12000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
2745 |
"eval_steps_per_second": 0.062,
|
2746 |
"eval_wer": 0.09984380314573199,
|
2747 |
"step": 11000
|
2748 |
+
},
|
2749 |
+
{
|
2750 |
+
"epoch": 0.38,
|
2751 |
+
"learning_rate": 5.7073469387755105e-06,
|
2752 |
+
"loss": 0.1738,
|
2753 |
+
"step": 11025
|
2754 |
+
},
|
2755 |
+
{
|
2756 |
+
"epoch": 0.38,
|
2757 |
+
"learning_rate": 5.697142857142858e-06,
|
2758 |
+
"loss": 0.156,
|
2759 |
+
"step": 11050
|
2760 |
+
},
|
2761 |
+
{
|
2762 |
+
"epoch": 0.38,
|
2763 |
+
"learning_rate": 5.686938775510205e-06,
|
2764 |
+
"loss": 0.1786,
|
2765 |
+
"step": 11075
|
2766 |
+
},
|
2767 |
+
{
|
2768 |
+
"epoch": 0.39,
|
2769 |
+
"learning_rate": 5.676734693877551e-06,
|
2770 |
+
"loss": 0.1681,
|
2771 |
+
"step": 11100
|
2772 |
+
},
|
2773 |
+
{
|
2774 |
+
"epoch": 0.39,
|
2775 |
+
"learning_rate": 5.666530612244898e-06,
|
2776 |
+
"loss": 0.1678,
|
2777 |
+
"step": 11125
|
2778 |
+
},
|
2779 |
+
{
|
2780 |
+
"epoch": 0.39,
|
2781 |
+
"learning_rate": 5.656326530612245e-06,
|
2782 |
+
"loss": 0.1696,
|
2783 |
+
"step": 11150
|
2784 |
+
},
|
2785 |
+
{
|
2786 |
+
"epoch": 0.39,
|
2787 |
+
"learning_rate": 5.646122448979592e-06,
|
2788 |
+
"loss": 0.1693,
|
2789 |
+
"step": 11175
|
2790 |
+
},
|
2791 |
+
{
|
2792 |
+
"epoch": 0.39,
|
2793 |
+
"learning_rate": 5.6359183673469395e-06,
|
2794 |
+
"loss": 0.1342,
|
2795 |
+
"step": 11200
|
2796 |
+
},
|
2797 |
+
{
|
2798 |
+
"epoch": 0.39,
|
2799 |
+
"learning_rate": 5.625714285714286e-06,
|
2800 |
+
"loss": 0.2051,
|
2801 |
+
"step": 11225
|
2802 |
+
},
|
2803 |
+
{
|
2804 |
+
"epoch": 0.39,
|
2805 |
+
"learning_rate": 5.615510204081634e-06,
|
2806 |
+
"loss": 0.1534,
|
2807 |
+
"step": 11250
|
2808 |
+
},
|
2809 |
+
{
|
2810 |
+
"epoch": 0.39,
|
2811 |
+
"learning_rate": 5.60530612244898e-06,
|
2812 |
+
"loss": 0.1683,
|
2813 |
+
"step": 11275
|
2814 |
+
},
|
2815 |
+
{
|
2816 |
+
"epoch": 0.39,
|
2817 |
+
"learning_rate": 5.595102040816327e-06,
|
2818 |
+
"loss": 0.1524,
|
2819 |
+
"step": 11300
|
2820 |
+
},
|
2821 |
+
{
|
2822 |
+
"epoch": 0.39,
|
2823 |
+
"learning_rate": 5.584897959183674e-06,
|
2824 |
+
"loss": 0.184,
|
2825 |
+
"step": 11325
|
2826 |
+
},
|
2827 |
+
{
|
2828 |
+
"epoch": 0.39,
|
2829 |
+
"learning_rate": 5.574693877551021e-06,
|
2830 |
+
"loss": 0.1486,
|
2831 |
+
"step": 11350
|
2832 |
+
},
|
2833 |
+
{
|
2834 |
+
"epoch": 0.39,
|
2835 |
+
"learning_rate": 5.564489795918368e-06,
|
2836 |
+
"loss": 0.1685,
|
2837 |
+
"step": 11375
|
2838 |
+
},
|
2839 |
+
{
|
2840 |
+
"epoch": 0.4,
|
2841 |
+
"learning_rate": 5.554285714285714e-06,
|
2842 |
+
"loss": 0.1645,
|
2843 |
+
"step": 11400
|
2844 |
+
},
|
2845 |
+
{
|
2846 |
+
"epoch": 0.4,
|
2847 |
+
"learning_rate": 5.544081632653062e-06,
|
2848 |
+
"loss": 0.1743,
|
2849 |
+
"step": 11425
|
2850 |
+
},
|
2851 |
+
{
|
2852 |
+
"epoch": 0.4,
|
2853 |
+
"learning_rate": 5.533877551020409e-06,
|
2854 |
+
"loss": 0.1377,
|
2855 |
+
"step": 11450
|
2856 |
+
},
|
2857 |
+
{
|
2858 |
+
"epoch": 0.4,
|
2859 |
+
"learning_rate": 5.524489795918367e-06,
|
2860 |
+
"loss": 0.1776,
|
2861 |
+
"step": 11475
|
2862 |
+
},
|
2863 |
+
{
|
2864 |
+
"epoch": 0.4,
|
2865 |
+
"learning_rate": 5.514285714285714e-06,
|
2866 |
+
"loss": 0.1365,
|
2867 |
+
"step": 11500
|
2868 |
+
},
|
2869 |
+
{
|
2870 |
+
"epoch": 0.4,
|
2871 |
+
"learning_rate": 5.504081632653062e-06,
|
2872 |
+
"loss": 0.1777,
|
2873 |
+
"step": 11525
|
2874 |
+
},
|
2875 |
+
{
|
2876 |
+
"epoch": 0.4,
|
2877 |
+
"learning_rate": 5.493877551020409e-06,
|
2878 |
+
"loss": 0.149,
|
2879 |
+
"step": 11550
|
2880 |
+
},
|
2881 |
+
{
|
2882 |
+
"epoch": 0.4,
|
2883 |
+
"learning_rate": 5.4836734693877555e-06,
|
2884 |
+
"loss": 0.1855,
|
2885 |
+
"step": 11575
|
2886 |
+
},
|
2887 |
+
{
|
2888 |
+
"epoch": 0.4,
|
2889 |
+
"learning_rate": 5.473469387755103e-06,
|
2890 |
+
"loss": 0.1543,
|
2891 |
+
"step": 11600
|
2892 |
+
},
|
2893 |
+
{
|
2894 |
+
"epoch": 0.4,
|
2895 |
+
"learning_rate": 5.46326530612245e-06,
|
2896 |
+
"loss": 0.1817,
|
2897 |
+
"step": 11625
|
2898 |
+
},
|
2899 |
+
{
|
2900 |
+
"epoch": 0.4,
|
2901 |
+
"learning_rate": 5.453061224489796e-06,
|
2902 |
+
"loss": 0.1614,
|
2903 |
+
"step": 11650
|
2904 |
+
},
|
2905 |
+
{
|
2906 |
+
"epoch": 0.41,
|
2907 |
+
"learning_rate": 5.442857142857143e-06,
|
2908 |
+
"loss": 0.1544,
|
2909 |
+
"step": 11675
|
2910 |
+
},
|
2911 |
+
{
|
2912 |
+
"epoch": 0.41,
|
2913 |
+
"learning_rate": 5.4326530612244895e-06,
|
2914 |
+
"loss": 0.1592,
|
2915 |
+
"step": 11700
|
2916 |
+
},
|
2917 |
+
{
|
2918 |
+
"epoch": 0.41,
|
2919 |
+
"learning_rate": 5.422448979591837e-06,
|
2920 |
+
"loss": 0.1658,
|
2921 |
+
"step": 11725
|
2922 |
+
},
|
2923 |
+
{
|
2924 |
+
"epoch": 0.41,
|
2925 |
+
"learning_rate": 5.4122448979591845e-06,
|
2926 |
+
"loss": 0.1596,
|
2927 |
+
"step": 11750
|
2928 |
+
},
|
2929 |
+
{
|
2930 |
+
"epoch": 0.41,
|
2931 |
+
"learning_rate": 5.402040816326531e-06,
|
2932 |
+
"loss": 0.1707,
|
2933 |
+
"step": 11775
|
2934 |
+
},
|
2935 |
+
{
|
2936 |
+
"epoch": 0.41,
|
2937 |
+
"learning_rate": 5.391836734693879e-06,
|
2938 |
+
"loss": 0.1383,
|
2939 |
+
"step": 11800
|
2940 |
+
},
|
2941 |
+
{
|
2942 |
+
"epoch": 0.41,
|
2943 |
+
"learning_rate": 5.381632653061225e-06,
|
2944 |
+
"loss": 0.175,
|
2945 |
+
"step": 11825
|
2946 |
+
},
|
2947 |
+
{
|
2948 |
+
"epoch": 0.41,
|
2949 |
+
"learning_rate": 5.371428571428572e-06,
|
2950 |
+
"loss": 0.1658,
|
2951 |
+
"step": 11850
|
2952 |
+
},
|
2953 |
+
{
|
2954 |
+
"epoch": 0.41,
|
2955 |
+
"learning_rate": 5.3612244897959186e-06,
|
2956 |
+
"loss": 0.1864,
|
2957 |
+
"step": 11875
|
2958 |
+
},
|
2959 |
+
{
|
2960 |
+
"epoch": 0.41,
|
2961 |
+
"learning_rate": 5.351020408163265e-06,
|
2962 |
+
"loss": 0.1441,
|
2963 |
+
"step": 11900
|
2964 |
+
},
|
2965 |
+
{
|
2966 |
+
"epoch": 0.41,
|
2967 |
+
"learning_rate": 5.340816326530613e-06,
|
2968 |
+
"loss": 0.1559,
|
2969 |
+
"step": 11925
|
2970 |
+
},
|
2971 |
+
{
|
2972 |
+
"epoch": 0.41,
|
2973 |
+
"learning_rate": 5.330612244897959e-06,
|
2974 |
+
"loss": 0.1559,
|
2975 |
+
"step": 11950
|
2976 |
+
},
|
2977 |
+
{
|
2978 |
+
"epoch": 0.42,
|
2979 |
+
"learning_rate": 5.320408163265307e-06,
|
2980 |
+
"loss": 0.1786,
|
2981 |
+
"step": 11975
|
2982 |
+
},
|
2983 |
+
{
|
2984 |
+
"epoch": 0.42,
|
2985 |
+
"learning_rate": 5.310204081632654e-06,
|
2986 |
+
"loss": 0.169,
|
2987 |
+
"step": 12000
|
2988 |
+
},
|
2989 |
+
{
|
2990 |
+
"epoch": 0.42,
|
2991 |
+
"eval_loss": 0.17712195217609406,
|
2992 |
+
"eval_runtime": 30926.8901,
|
2993 |
+
"eval_samples_per_second": 1.004,
|
2994 |
+
"eval_steps_per_second": 0.063,
|
2995 |
+
"eval_wer": 0.09824881201229761,
|
2996 |
+
"step": 12000
|
2997 |
}
|
2998 |
],
|
2999 |
"max_steps": 25000,
|
3000 |
"num_train_epochs": 1,
|
3001 |
+
"total_flos": 4.076488359936e+20,
|
3002 |
"trial_name": null,
|
3003 |
"trial_params": null
|
3004 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d40f3344c48b8c1de58d7e8467befdcdcde144b9ff8ed4465cd30973a433d738
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc505845f9a549bc125c53e3e46d54bfa0b789b17c0fc518106e8b0e000bbc35
|
3 |
+
size 83565
|