elsayedissa commited on
Commit
f7a6472
1 Parent(s): 08cf7b9

Training in progress, step 9000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63d424da125909272adbcd6febb158fb464d384fcc7052b7279506ac6f5969ac
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f166b67a2e5e8da54bcd327bceca4b864da3d9470f1e27ae8c07237158f24f2
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00f04d4ca24866b389dc2df4457f672ba5912421795c884ce9a09b3caf5317e1
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd38db0ab531900a0edf77b4884776dbc30584f579ce891d6129463490ea14aa
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbfadcfc0728788eb8464114731524e59e0eeea3de78fb600f90df25108582a6
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14d50af3ad80506d9b5eb2378638e4d6bd5b02f77a4e8ed4e65b44aee3f7410c
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2033db9ad36b6f8a091d260383726ce676844e1562b9251b94e781e66324da99
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a05fc2aa63e1878582f11ce1a8a19360e9ade1f13595895a6cf908de48f3707
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72ca6c784ba72f382b1bcb047ab8308017dc36fd5ba0484d4a1d02017086a52b
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adf20911833533da369ecd232dea5963d4a78412ad9ec6b3923960496ee6c41e
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.27769099934048386,
5
- "global_step": 8000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -1998,11 +1998,260 @@
1998
  "eval_steps_per_second": 0.062,
1999
  "eval_wer": 0.10428864386126549,
2000
  "step": 8000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2001
  }
2002
  ],
2003
  "max_steps": 25000,
2004
  "num_train_epochs": 1,
2005
- "total_flos": 2.717658906624e+20,
2006
  "trial_name": null,
2007
  "trial_params": null
2008
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.31240237425804435,
5
+ "global_step": 9000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
1998
  "eval_steps_per_second": 0.062,
1999
  "eval_wer": 0.10428864386126549,
2000
  "step": 8000
2001
+ },
2002
+ {
2003
+ "epoch": 0.28,
2004
+ "learning_rate": 6.931836734693878e-06,
2005
+ "loss": 0.179,
2006
+ "step": 8025
2007
+ },
2008
+ {
2009
+ "epoch": 0.28,
2010
+ "learning_rate": 6.921632653061226e-06,
2011
+ "loss": 0.1647,
2012
+ "step": 8050
2013
+ },
2014
+ {
2015
+ "epoch": 0.28,
2016
+ "learning_rate": 6.911428571428572e-06,
2017
+ "loss": 0.1792,
2018
+ "step": 8075
2019
+ },
2020
+ {
2021
+ "epoch": 0.28,
2022
+ "learning_rate": 6.901224489795919e-06,
2023
+ "loss": 0.1815,
2024
+ "step": 8100
2025
+ },
2026
+ {
2027
+ "epoch": 0.28,
2028
+ "learning_rate": 6.891020408163266e-06,
2029
+ "loss": 0.1943,
2030
+ "step": 8125
2031
+ },
2032
+ {
2033
+ "epoch": 0.28,
2034
+ "learning_rate": 6.880816326530612e-06,
2035
+ "loss": 0.1473,
2036
+ "step": 8150
2037
+ },
2038
+ {
2039
+ "epoch": 0.28,
2040
+ "learning_rate": 6.87061224489796e-06,
2041
+ "loss": 0.196,
2042
+ "step": 8175
2043
+ },
2044
+ {
2045
+ "epoch": 0.28,
2046
+ "learning_rate": 6.8604081632653065e-06,
2047
+ "loss": 0.1471,
2048
+ "step": 8200
2049
+ },
2050
+ {
2051
+ "epoch": 0.29,
2052
+ "learning_rate": 6.850204081632653e-06,
2053
+ "loss": 0.1928,
2054
+ "step": 8225
2055
+ },
2056
+ {
2057
+ "epoch": 0.29,
2058
+ "learning_rate": 6.8400000000000014e-06,
2059
+ "loss": 0.1693,
2060
+ "step": 8250
2061
+ },
2062
+ {
2063
+ "epoch": 0.29,
2064
+ "learning_rate": 6.829795918367348e-06,
2065
+ "loss": 0.1941,
2066
+ "step": 8275
2067
+ },
2068
+ {
2069
+ "epoch": 0.29,
2070
+ "learning_rate": 6.819591836734695e-06,
2071
+ "loss": 0.1627,
2072
+ "step": 8300
2073
+ },
2074
+ {
2075
+ "epoch": 0.29,
2076
+ "learning_rate": 6.809387755102041e-06,
2077
+ "loss": 0.1591,
2078
+ "step": 8325
2079
+ },
2080
+ {
2081
+ "epoch": 0.29,
2082
+ "learning_rate": 6.799183673469388e-06,
2083
+ "loss": 0.1641,
2084
+ "step": 8350
2085
+ },
2086
+ {
2087
+ "epoch": 0.29,
2088
+ "learning_rate": 6.7889795918367355e-06,
2089
+ "loss": 0.1943,
2090
+ "step": 8375
2091
+ },
2092
+ {
2093
+ "epoch": 0.29,
2094
+ "learning_rate": 6.778775510204082e-06,
2095
+ "loss": 0.1474,
2096
+ "step": 8400
2097
+ },
2098
+ {
2099
+ "epoch": 0.29,
2100
+ "learning_rate": 6.768571428571429e-06,
2101
+ "loss": 0.2049,
2102
+ "step": 8425
2103
+ },
2104
+ {
2105
+ "epoch": 0.29,
2106
+ "learning_rate": 6.758367346938775e-06,
2107
+ "loss": 0.1701,
2108
+ "step": 8450
2109
+ },
2110
+ {
2111
+ "epoch": 0.29,
2112
+ "learning_rate": 6.748163265306124e-06,
2113
+ "loss": 0.1948,
2114
+ "step": 8475
2115
+ },
2116
+ {
2117
+ "epoch": 0.3,
2118
+ "learning_rate": 6.73795918367347e-06,
2119
+ "loss": 0.176,
2120
+ "step": 8500
2121
+ },
2122
+ {
2123
+ "epoch": 0.3,
2124
+ "learning_rate": 6.727755102040817e-06,
2125
+ "loss": 0.2003,
2126
+ "step": 8525
2127
+ },
2128
+ {
2129
+ "epoch": 0.3,
2130
+ "learning_rate": 6.717551020408164e-06,
2131
+ "loss": 0.1687,
2132
+ "step": 8550
2133
+ },
2134
+ {
2135
+ "epoch": 0.3,
2136
+ "learning_rate": 6.707346938775511e-06,
2137
+ "loss": 0.1838,
2138
+ "step": 8575
2139
+ },
2140
+ {
2141
+ "epoch": 0.3,
2142
+ "learning_rate": 6.697142857142858e-06,
2143
+ "loss": 0.1492,
2144
+ "step": 8600
2145
+ },
2146
+ {
2147
+ "epoch": 0.3,
2148
+ "learning_rate": 6.686938775510204e-06,
2149
+ "loss": 0.1886,
2150
+ "step": 8625
2151
+ },
2152
+ {
2153
+ "epoch": 0.3,
2154
+ "learning_rate": 6.676734693877551e-06,
2155
+ "loss": 0.1604,
2156
+ "step": 8650
2157
+ },
2158
+ {
2159
+ "epoch": 0.3,
2160
+ "learning_rate": 6.666530612244898e-06,
2161
+ "loss": 0.1935,
2162
+ "step": 8675
2163
+ },
2164
+ {
2165
+ "epoch": 0.3,
2166
+ "learning_rate": 6.656326530612246e-06,
2167
+ "loss": 0.1571,
2168
+ "step": 8700
2169
+ },
2170
+ {
2171
+ "epoch": 0.3,
2172
+ "learning_rate": 6.646122448979593e-06,
2173
+ "loss": 0.1697,
2174
+ "step": 8725
2175
+ },
2176
+ {
2177
+ "epoch": 0.3,
2178
+ "learning_rate": 6.635918367346939e-06,
2179
+ "loss": 0.1644,
2180
+ "step": 8750
2181
+ },
2182
+ {
2183
+ "epoch": 0.3,
2184
+ "learning_rate": 6.625714285714287e-06,
2185
+ "loss": 0.1977,
2186
+ "step": 8775
2187
+ },
2188
+ {
2189
+ "epoch": 0.31,
2190
+ "learning_rate": 6.615510204081633e-06,
2191
+ "loss": 0.1566,
2192
+ "step": 8800
2193
+ },
2194
+ {
2195
+ "epoch": 0.31,
2196
+ "learning_rate": 6.60530612244898e-06,
2197
+ "loss": 0.1676,
2198
+ "step": 8825
2199
+ },
2200
+ {
2201
+ "epoch": 0.31,
2202
+ "learning_rate": 6.595102040816327e-06,
2203
+ "loss": 0.151,
2204
+ "step": 8850
2205
+ },
2206
+ {
2207
+ "epoch": 0.31,
2208
+ "learning_rate": 6.584897959183673e-06,
2209
+ "loss": 0.1828,
2210
+ "step": 8875
2211
+ },
2212
+ {
2213
+ "epoch": 0.31,
2214
+ "learning_rate": 6.574693877551021e-06,
2215
+ "loss": 0.1781,
2216
+ "step": 8900
2217
+ },
2218
+ {
2219
+ "epoch": 0.31,
2220
+ "learning_rate": 6.564489795918368e-06,
2221
+ "loss": 0.1689,
2222
+ "step": 8925
2223
+ },
2224
+ {
2225
+ "epoch": 0.31,
2226
+ "learning_rate": 6.554285714285715e-06,
2227
+ "loss": 0.1698,
2228
+ "step": 8950
2229
+ },
2230
+ {
2231
+ "epoch": 0.31,
2232
+ "learning_rate": 6.5440816326530624e-06,
2233
+ "loss": 0.1893,
2234
+ "step": 8975
2235
+ },
2236
+ {
2237
+ "epoch": 0.31,
2238
+ "learning_rate": 6.533877551020409e-06,
2239
+ "loss": 0.166,
2240
+ "step": 9000
2241
+ },
2242
+ {
2243
+ "epoch": 0.31,
2244
+ "eval_loss": 0.18502512574195862,
2245
+ "eval_runtime": 31506.0133,
2246
+ "eval_samples_per_second": 0.985,
2247
+ "eval_steps_per_second": 0.062,
2248
+ "eval_wer": 0.10222473193911955,
2249
+ "step": 9000
2250
  }
2251
  ],
2252
  "max_steps": 25000,
2253
  "num_train_epochs": 1,
2254
+ "total_flos": 3.057366269952e+20,
2255
  "trial_name": null,
2256
  "trial_params": null
2257
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00f04d4ca24866b389dc2df4457f672ba5912421795c884ce9a09b3caf5317e1
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd38db0ab531900a0edf77b4884776dbc30584f579ce891d6129463490ea14aa
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:076eb2bada8a4a94fa9d189e543bfecad75db25a7e323324e16521b9cba64240
3
- size 57173
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d63ccbe1acfa4c9224bc8fd7be72373fbc520dadad5e653bda1403d45e1622f
3
+ size 63771