elsayedissa commited on
Commit
e5e85ab
1 Parent(s): f7a6472

Training in progress, step 10000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f166b67a2e5e8da54bcd327bceca4b864da3d9470f1e27ae8c07237158f24f2
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:666bd295da5110fc8ed72c097e043b17b68a4f46ec1f780894808ac7d810702a
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd38db0ab531900a0edf77b4884776dbc30584f579ce891d6129463490ea14aa
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14d50af3ad80506d9b5eb2378638e4d6bd5b02f77a4e8ed4e65b44aee3f7410c
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aad3fc937f2c6f56a2a6ddd2d655b320e20d5cf3c0672d854ed64624351b0488
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a05fc2aa63e1878582f11ce1a8a19360e9ade1f13595895a6cf908de48f3707
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ff19fa2317539f8bfb616fcdefc3caa2efd60d151a696a55bb38626512e9ba9
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:adf20911833533da369ecd232dea5963d4a78412ad9ec6b3923960496ee6c41e
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1196c99a263cfa6a64b1d73b220e535cc282c309df0a28be120a2a37266e02ba
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.31240237425804435,
5
- "global_step": 9000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -2247,11 +2247,260 @@
2247
  "eval_steps_per_second": 0.062,
2248
  "eval_wer": 0.10222473193911955,
2249
  "step": 9000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2250
  }
2251
  ],
2252
  "max_steps": 25000,
2253
  "num_train_epochs": 1,
2254
- "total_flos": 3.057366269952e+20,
2255
  "trial_name": null,
2256
  "trial_params": null
2257
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.34711374917560484,
5
+ "global_step": 10000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
2247
  "eval_steps_per_second": 0.062,
2248
  "eval_wer": 0.10222473193911955,
2249
  "step": 9000
2250
+ },
2251
+ {
2252
+ "epoch": 0.31,
2253
+ "learning_rate": 6.523673469387756e-06,
2254
+ "loss": 0.1991,
2255
+ "step": 9025
2256
+ },
2257
+ {
2258
+ "epoch": 0.31,
2259
+ "learning_rate": 6.513469387755102e-06,
2260
+ "loss": 0.1732,
2261
+ "step": 9050
2262
+ },
2263
+ {
2264
+ "epoch": 0.32,
2265
+ "learning_rate": 6.503265306122449e-06,
2266
+ "loss": 0.1864,
2267
+ "step": 9075
2268
+ },
2269
+ {
2270
+ "epoch": 0.32,
2271
+ "learning_rate": 6.4930612244897965e-06,
2272
+ "loss": 0.1535,
2273
+ "step": 9100
2274
+ },
2275
+ {
2276
+ "epoch": 0.32,
2277
+ "learning_rate": 6.482857142857143e-06,
2278
+ "loss": 0.1749,
2279
+ "step": 9125
2280
+ },
2281
+ {
2282
+ "epoch": 0.32,
2283
+ "learning_rate": 6.47265306122449e-06,
2284
+ "loss": 0.1707,
2285
+ "step": 9150
2286
+ },
2287
+ {
2288
+ "epoch": 0.32,
2289
+ "learning_rate": 6.462448979591838e-06,
2290
+ "loss": 0.1855,
2291
+ "step": 9175
2292
+ },
2293
+ {
2294
+ "epoch": 0.32,
2295
+ "learning_rate": 6.452244897959185e-06,
2296
+ "loss": 0.1667,
2297
+ "step": 9200
2298
+ },
2299
+ {
2300
+ "epoch": 0.32,
2301
+ "learning_rate": 6.442040816326531e-06,
2302
+ "loss": 0.1997,
2303
+ "step": 9225
2304
+ },
2305
+ {
2306
+ "epoch": 0.32,
2307
+ "learning_rate": 6.431836734693878e-06,
2308
+ "loss": 0.1476,
2309
+ "step": 9250
2310
+ },
2311
+ {
2312
+ "epoch": 0.32,
2313
+ "learning_rate": 6.421632653061225e-06,
2314
+ "loss": 0.1955,
2315
+ "step": 9275
2316
+ },
2317
+ {
2318
+ "epoch": 0.32,
2319
+ "learning_rate": 6.411428571428572e-06,
2320
+ "loss": 0.1701,
2321
+ "step": 9300
2322
+ },
2323
+ {
2324
+ "epoch": 0.32,
2325
+ "learning_rate": 6.401224489795919e-06,
2326
+ "loss": 0.1738,
2327
+ "step": 9325
2328
+ },
2329
+ {
2330
+ "epoch": 0.32,
2331
+ "learning_rate": 6.391020408163265e-06,
2332
+ "loss": 0.1659,
2333
+ "step": 9350
2334
+ },
2335
+ {
2336
+ "epoch": 0.33,
2337
+ "learning_rate": 6.380816326530612e-06,
2338
+ "loss": 0.1665,
2339
+ "step": 9375
2340
+ },
2341
+ {
2342
+ "epoch": 0.33,
2343
+ "learning_rate": 6.37061224489796e-06,
2344
+ "loss": 0.1668,
2345
+ "step": 9400
2346
+ },
2347
+ {
2348
+ "epoch": 0.33,
2349
+ "learning_rate": 6.360408163265307e-06,
2350
+ "loss": 0.2077,
2351
+ "step": 9425
2352
+ },
2353
+ {
2354
+ "epoch": 0.33,
2355
+ "learning_rate": 6.350204081632654e-06,
2356
+ "loss": 0.1615,
2357
+ "step": 9450
2358
+ },
2359
+ {
2360
+ "epoch": 0.33,
2361
+ "learning_rate": 6.34e-06,
2362
+ "loss": 0.1833,
2363
+ "step": 9475
2364
+ },
2365
+ {
2366
+ "epoch": 0.33,
2367
+ "learning_rate": 6.329795918367348e-06,
2368
+ "loss": 0.1696,
2369
+ "step": 9500
2370
+ },
2371
+ {
2372
+ "epoch": 0.33,
2373
+ "learning_rate": 6.319591836734694e-06,
2374
+ "loss": 0.183,
2375
+ "step": 9525
2376
+ },
2377
+ {
2378
+ "epoch": 0.33,
2379
+ "learning_rate": 6.309387755102041e-06,
2380
+ "loss": 0.1892,
2381
+ "step": 9550
2382
+ },
2383
+ {
2384
+ "epoch": 0.33,
2385
+ "learning_rate": 6.299183673469388e-06,
2386
+ "loss": 0.168,
2387
+ "step": 9575
2388
+ },
2389
+ {
2390
+ "epoch": 0.33,
2391
+ "learning_rate": 6.288979591836734e-06,
2392
+ "loss": 0.1645,
2393
+ "step": 9600
2394
+ },
2395
+ {
2396
+ "epoch": 0.33,
2397
+ "learning_rate": 6.278775510204083e-06,
2398
+ "loss": 0.181,
2399
+ "step": 9625
2400
+ },
2401
+ {
2402
+ "epoch": 0.33,
2403
+ "learning_rate": 6.268571428571429e-06,
2404
+ "loss": 0.1508,
2405
+ "step": 9650
2406
+ },
2407
+ {
2408
+ "epoch": 0.34,
2409
+ "learning_rate": 6.258367346938776e-06,
2410
+ "loss": 0.1776,
2411
+ "step": 9675
2412
+ },
2413
+ {
2414
+ "epoch": 0.34,
2415
+ "learning_rate": 6.248163265306123e-06,
2416
+ "loss": 0.161,
2417
+ "step": 9700
2418
+ },
2419
+ {
2420
+ "epoch": 0.34,
2421
+ "learning_rate": 6.23795918367347e-06,
2422
+ "loss": 0.1873,
2423
+ "step": 9725
2424
+ },
2425
+ {
2426
+ "epoch": 0.34,
2427
+ "learning_rate": 6.227755102040817e-06,
2428
+ "loss": 0.166,
2429
+ "step": 9750
2430
+ },
2431
+ {
2432
+ "epoch": 0.34,
2433
+ "learning_rate": 6.217551020408163e-06,
2434
+ "loss": 0.1739,
2435
+ "step": 9775
2436
+ },
2437
+ {
2438
+ "epoch": 0.34,
2439
+ "learning_rate": 6.20734693877551e-06,
2440
+ "loss": 0.1496,
2441
+ "step": 9800
2442
+ },
2443
+ {
2444
+ "epoch": 0.34,
2445
+ "learning_rate": 6.1971428571428575e-06,
2446
+ "loss": 0.1924,
2447
+ "step": 9825
2448
+ },
2449
+ {
2450
+ "epoch": 0.34,
2451
+ "learning_rate": 6.186938775510205e-06,
2452
+ "loss": 0.1453,
2453
+ "step": 9850
2454
+ },
2455
+ {
2456
+ "epoch": 0.34,
2457
+ "learning_rate": 6.176734693877552e-06,
2458
+ "loss": 0.196,
2459
+ "step": 9875
2460
+ },
2461
+ {
2462
+ "epoch": 0.34,
2463
+ "learning_rate": 6.166530612244899e-06,
2464
+ "loss": 0.1782,
2465
+ "step": 9900
2466
+ },
2467
+ {
2468
+ "epoch": 0.34,
2469
+ "learning_rate": 6.156326530612246e-06,
2470
+ "loss": 0.1855,
2471
+ "step": 9925
2472
+ },
2473
+ {
2474
+ "epoch": 0.35,
2475
+ "learning_rate": 6.146122448979592e-06,
2476
+ "loss": 0.1628,
2477
+ "step": 9950
2478
+ },
2479
+ {
2480
+ "epoch": 0.35,
2481
+ "learning_rate": 6.135918367346939e-06,
2482
+ "loss": 0.1733,
2483
+ "step": 9975
2484
+ },
2485
+ {
2486
+ "epoch": 0.35,
2487
+ "learning_rate": 6.125714285714286e-06,
2488
+ "loss": 0.1277,
2489
+ "step": 10000
2490
+ },
2491
+ {
2492
+ "epoch": 0.35,
2493
+ "eval_loss": 0.1820017397403717,
2494
+ "eval_runtime": 31277.5076,
2495
+ "eval_samples_per_second": 0.992,
2496
+ "eval_steps_per_second": 0.062,
2497
+ "eval_wer": 0.10322861869805133,
2498
+ "step": 10000
2499
  }
2500
  ],
2501
  "max_steps": 25000,
2502
  "num_train_epochs": 1,
2503
+ "total_flos": 3.39707363328e+20,
2504
  "trial_name": null,
2505
  "trial_params": null
2506
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd38db0ab531900a0edf77b4884776dbc30584f579ce891d6129463490ea14aa
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:492b321ad12ae8744ac23f006fa941f8bd7881d2b5e9aecb269dbdaf98b5ee88
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d63ccbe1acfa4c9224bc8fd7be72373fbc520dadad5e653bda1403d45e1622f
3
- size 63771
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3649b56169acd2abda9b0ee07a52cc5ff3d90c7e405842df1439d1a6f490683e
3
+ size 70369