elsayedissa
commited on
Commit
•
e5e85ab
1
Parent(s):
f7a6472
Training in progress, step 10000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:666bd295da5110fc8ed72c097e043b17b68a4f46ec1f780894808ac7d810702a
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aad3fc937f2c6f56a2a6ddd2d655b320e20d5cf3c0672d854ed64624351b0488
|
3 |
size 14575
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0ff19fa2317539f8bfb616fcdefc3caa2efd60d151a696a55bb38626512e9ba9
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1196c99a263cfa6a64b1d73b220e535cc282c309df0a28be120a2a37266e02ba
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -2247,11 +2247,260 @@
|
|
2247 |
"eval_steps_per_second": 0.062,
|
2248 |
"eval_wer": 0.10222473193911955,
|
2249 |
"step": 9000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2250 |
}
|
2251 |
],
|
2252 |
"max_steps": 25000,
|
2253 |
"num_train_epochs": 1,
|
2254 |
-
"total_flos": 3.
|
2255 |
"trial_name": null,
|
2256 |
"trial_params": null
|
2257 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.34711374917560484,
|
5 |
+
"global_step": 10000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
2247 |
"eval_steps_per_second": 0.062,
|
2248 |
"eval_wer": 0.10222473193911955,
|
2249 |
"step": 9000
|
2250 |
+
},
|
2251 |
+
{
|
2252 |
+
"epoch": 0.31,
|
2253 |
+
"learning_rate": 6.523673469387756e-06,
|
2254 |
+
"loss": 0.1991,
|
2255 |
+
"step": 9025
|
2256 |
+
},
|
2257 |
+
{
|
2258 |
+
"epoch": 0.31,
|
2259 |
+
"learning_rate": 6.513469387755102e-06,
|
2260 |
+
"loss": 0.1732,
|
2261 |
+
"step": 9050
|
2262 |
+
},
|
2263 |
+
{
|
2264 |
+
"epoch": 0.32,
|
2265 |
+
"learning_rate": 6.503265306122449e-06,
|
2266 |
+
"loss": 0.1864,
|
2267 |
+
"step": 9075
|
2268 |
+
},
|
2269 |
+
{
|
2270 |
+
"epoch": 0.32,
|
2271 |
+
"learning_rate": 6.4930612244897965e-06,
|
2272 |
+
"loss": 0.1535,
|
2273 |
+
"step": 9100
|
2274 |
+
},
|
2275 |
+
{
|
2276 |
+
"epoch": 0.32,
|
2277 |
+
"learning_rate": 6.482857142857143e-06,
|
2278 |
+
"loss": 0.1749,
|
2279 |
+
"step": 9125
|
2280 |
+
},
|
2281 |
+
{
|
2282 |
+
"epoch": 0.32,
|
2283 |
+
"learning_rate": 6.47265306122449e-06,
|
2284 |
+
"loss": 0.1707,
|
2285 |
+
"step": 9150
|
2286 |
+
},
|
2287 |
+
{
|
2288 |
+
"epoch": 0.32,
|
2289 |
+
"learning_rate": 6.462448979591838e-06,
|
2290 |
+
"loss": 0.1855,
|
2291 |
+
"step": 9175
|
2292 |
+
},
|
2293 |
+
{
|
2294 |
+
"epoch": 0.32,
|
2295 |
+
"learning_rate": 6.452244897959185e-06,
|
2296 |
+
"loss": 0.1667,
|
2297 |
+
"step": 9200
|
2298 |
+
},
|
2299 |
+
{
|
2300 |
+
"epoch": 0.32,
|
2301 |
+
"learning_rate": 6.442040816326531e-06,
|
2302 |
+
"loss": 0.1997,
|
2303 |
+
"step": 9225
|
2304 |
+
},
|
2305 |
+
{
|
2306 |
+
"epoch": 0.32,
|
2307 |
+
"learning_rate": 6.431836734693878e-06,
|
2308 |
+
"loss": 0.1476,
|
2309 |
+
"step": 9250
|
2310 |
+
},
|
2311 |
+
{
|
2312 |
+
"epoch": 0.32,
|
2313 |
+
"learning_rate": 6.421632653061225e-06,
|
2314 |
+
"loss": 0.1955,
|
2315 |
+
"step": 9275
|
2316 |
+
},
|
2317 |
+
{
|
2318 |
+
"epoch": 0.32,
|
2319 |
+
"learning_rate": 6.411428571428572e-06,
|
2320 |
+
"loss": 0.1701,
|
2321 |
+
"step": 9300
|
2322 |
+
},
|
2323 |
+
{
|
2324 |
+
"epoch": 0.32,
|
2325 |
+
"learning_rate": 6.401224489795919e-06,
|
2326 |
+
"loss": 0.1738,
|
2327 |
+
"step": 9325
|
2328 |
+
},
|
2329 |
+
{
|
2330 |
+
"epoch": 0.32,
|
2331 |
+
"learning_rate": 6.391020408163265e-06,
|
2332 |
+
"loss": 0.1659,
|
2333 |
+
"step": 9350
|
2334 |
+
},
|
2335 |
+
{
|
2336 |
+
"epoch": 0.33,
|
2337 |
+
"learning_rate": 6.380816326530612e-06,
|
2338 |
+
"loss": 0.1665,
|
2339 |
+
"step": 9375
|
2340 |
+
},
|
2341 |
+
{
|
2342 |
+
"epoch": 0.33,
|
2343 |
+
"learning_rate": 6.37061224489796e-06,
|
2344 |
+
"loss": 0.1668,
|
2345 |
+
"step": 9400
|
2346 |
+
},
|
2347 |
+
{
|
2348 |
+
"epoch": 0.33,
|
2349 |
+
"learning_rate": 6.360408163265307e-06,
|
2350 |
+
"loss": 0.2077,
|
2351 |
+
"step": 9425
|
2352 |
+
},
|
2353 |
+
{
|
2354 |
+
"epoch": 0.33,
|
2355 |
+
"learning_rate": 6.350204081632654e-06,
|
2356 |
+
"loss": 0.1615,
|
2357 |
+
"step": 9450
|
2358 |
+
},
|
2359 |
+
{
|
2360 |
+
"epoch": 0.33,
|
2361 |
+
"learning_rate": 6.34e-06,
|
2362 |
+
"loss": 0.1833,
|
2363 |
+
"step": 9475
|
2364 |
+
},
|
2365 |
+
{
|
2366 |
+
"epoch": 0.33,
|
2367 |
+
"learning_rate": 6.329795918367348e-06,
|
2368 |
+
"loss": 0.1696,
|
2369 |
+
"step": 9500
|
2370 |
+
},
|
2371 |
+
{
|
2372 |
+
"epoch": 0.33,
|
2373 |
+
"learning_rate": 6.319591836734694e-06,
|
2374 |
+
"loss": 0.183,
|
2375 |
+
"step": 9525
|
2376 |
+
},
|
2377 |
+
{
|
2378 |
+
"epoch": 0.33,
|
2379 |
+
"learning_rate": 6.309387755102041e-06,
|
2380 |
+
"loss": 0.1892,
|
2381 |
+
"step": 9550
|
2382 |
+
},
|
2383 |
+
{
|
2384 |
+
"epoch": 0.33,
|
2385 |
+
"learning_rate": 6.299183673469388e-06,
|
2386 |
+
"loss": 0.168,
|
2387 |
+
"step": 9575
|
2388 |
+
},
|
2389 |
+
{
|
2390 |
+
"epoch": 0.33,
|
2391 |
+
"learning_rate": 6.288979591836734e-06,
|
2392 |
+
"loss": 0.1645,
|
2393 |
+
"step": 9600
|
2394 |
+
},
|
2395 |
+
{
|
2396 |
+
"epoch": 0.33,
|
2397 |
+
"learning_rate": 6.278775510204083e-06,
|
2398 |
+
"loss": 0.181,
|
2399 |
+
"step": 9625
|
2400 |
+
},
|
2401 |
+
{
|
2402 |
+
"epoch": 0.33,
|
2403 |
+
"learning_rate": 6.268571428571429e-06,
|
2404 |
+
"loss": 0.1508,
|
2405 |
+
"step": 9650
|
2406 |
+
},
|
2407 |
+
{
|
2408 |
+
"epoch": 0.34,
|
2409 |
+
"learning_rate": 6.258367346938776e-06,
|
2410 |
+
"loss": 0.1776,
|
2411 |
+
"step": 9675
|
2412 |
+
},
|
2413 |
+
{
|
2414 |
+
"epoch": 0.34,
|
2415 |
+
"learning_rate": 6.248163265306123e-06,
|
2416 |
+
"loss": 0.161,
|
2417 |
+
"step": 9700
|
2418 |
+
},
|
2419 |
+
{
|
2420 |
+
"epoch": 0.34,
|
2421 |
+
"learning_rate": 6.23795918367347e-06,
|
2422 |
+
"loss": 0.1873,
|
2423 |
+
"step": 9725
|
2424 |
+
},
|
2425 |
+
{
|
2426 |
+
"epoch": 0.34,
|
2427 |
+
"learning_rate": 6.227755102040817e-06,
|
2428 |
+
"loss": 0.166,
|
2429 |
+
"step": 9750
|
2430 |
+
},
|
2431 |
+
{
|
2432 |
+
"epoch": 0.34,
|
2433 |
+
"learning_rate": 6.217551020408163e-06,
|
2434 |
+
"loss": 0.1739,
|
2435 |
+
"step": 9775
|
2436 |
+
},
|
2437 |
+
{
|
2438 |
+
"epoch": 0.34,
|
2439 |
+
"learning_rate": 6.20734693877551e-06,
|
2440 |
+
"loss": 0.1496,
|
2441 |
+
"step": 9800
|
2442 |
+
},
|
2443 |
+
{
|
2444 |
+
"epoch": 0.34,
|
2445 |
+
"learning_rate": 6.1971428571428575e-06,
|
2446 |
+
"loss": 0.1924,
|
2447 |
+
"step": 9825
|
2448 |
+
},
|
2449 |
+
{
|
2450 |
+
"epoch": 0.34,
|
2451 |
+
"learning_rate": 6.186938775510205e-06,
|
2452 |
+
"loss": 0.1453,
|
2453 |
+
"step": 9850
|
2454 |
+
},
|
2455 |
+
{
|
2456 |
+
"epoch": 0.34,
|
2457 |
+
"learning_rate": 6.176734693877552e-06,
|
2458 |
+
"loss": 0.196,
|
2459 |
+
"step": 9875
|
2460 |
+
},
|
2461 |
+
{
|
2462 |
+
"epoch": 0.34,
|
2463 |
+
"learning_rate": 6.166530612244899e-06,
|
2464 |
+
"loss": 0.1782,
|
2465 |
+
"step": 9900
|
2466 |
+
},
|
2467 |
+
{
|
2468 |
+
"epoch": 0.34,
|
2469 |
+
"learning_rate": 6.156326530612246e-06,
|
2470 |
+
"loss": 0.1855,
|
2471 |
+
"step": 9925
|
2472 |
+
},
|
2473 |
+
{
|
2474 |
+
"epoch": 0.35,
|
2475 |
+
"learning_rate": 6.146122448979592e-06,
|
2476 |
+
"loss": 0.1628,
|
2477 |
+
"step": 9950
|
2478 |
+
},
|
2479 |
+
{
|
2480 |
+
"epoch": 0.35,
|
2481 |
+
"learning_rate": 6.135918367346939e-06,
|
2482 |
+
"loss": 0.1733,
|
2483 |
+
"step": 9975
|
2484 |
+
},
|
2485 |
+
{
|
2486 |
+
"epoch": 0.35,
|
2487 |
+
"learning_rate": 6.125714285714286e-06,
|
2488 |
+
"loss": 0.1277,
|
2489 |
+
"step": 10000
|
2490 |
+
},
|
2491 |
+
{
|
2492 |
+
"epoch": 0.35,
|
2493 |
+
"eval_loss": 0.1820017397403717,
|
2494 |
+
"eval_runtime": 31277.5076,
|
2495 |
+
"eval_samples_per_second": 0.992,
|
2496 |
+
"eval_steps_per_second": 0.062,
|
2497 |
+
"eval_wer": 0.10322861869805133,
|
2498 |
+
"step": 10000
|
2499 |
}
|
2500 |
],
|
2501 |
"max_steps": 25000,
|
2502 |
"num_train_epochs": 1,
|
2503 |
+
"total_flos": 3.39707363328e+20,
|
2504 |
"trial_name": null,
|
2505 |
"trial_params": null
|
2506 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3649b56169acd2abda9b0ee07a52cc5ff3d90c7e405842df1439d1a6f490683e
|
3 |
+
size 70369
|