elsayedissa commited on
Commit
34aa10e
1 Parent(s): 2685bd3

Training in progress, step 22000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c1e93fa65fde9ba67832b5fde4ad330506d95a23edfec7be98020d37f17559b
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53ad543f0df9ec43766eb0c4be86887e32e3e4bd5f26ee276eea2ab28ae1ddef
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f07b5f25c323d9a36c85d57f748183f9d3d14494dbfa6b195ed15224d9deafeb
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bffe463d62be009d062d805c732b0374b40de43d96b166556a53b6a5998e2271
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f36f42cb84b389111430939ff75a7277c877ebb6534c5db685d643483c1e7ae
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4828b52c0813e99f8e1771ed2daad93ca33f628aeb2f5d64700f0bf5092a00e
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75b2d5c8b12724e300963a719034bfe44737705d2de06deafe54683c78564862
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96e64a1f7ca480b449676f7014db4dccc62eddeb2f60d979889930ac4af623a1
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7394c13a986a89cb46fa0877cb8091fc4bf0e4f28ce1608b81abde4cc024dfc
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0810f9c5189858aa63fe3a8926a77123b48758bcece57c05805766ff2c826811
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.7289388732687702,
5
- "global_step": 21000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -5235,11 +5235,260 @@
5235
  "eval_steps_per_second": 0.062,
5236
  "eval_wer": 0.08848403192624074,
5237
  "step": 21000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5238
  }
5239
  ],
5240
  "max_steps": 25000,
5241
  "num_train_epochs": 1,
5242
- "total_flos": 7.133854629888e+20,
5243
  "trial_name": null,
5244
  "trial_params": null
5245
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.7636502481863306,
5
+ "global_step": 22000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
5235
  "eval_steps_per_second": 0.062,
5236
  "eval_wer": 0.08848403192624074,
5237
  "step": 21000
5238
+ },
5239
+ {
5240
+ "epoch": 0.73,
5241
+ "learning_rate": 1.6269387755102041e-06,
5242
+ "loss": 0.1562,
5243
+ "step": 21025
5244
+ },
5245
+ {
5246
+ "epoch": 0.73,
5247
+ "learning_rate": 1.616734693877551e-06,
5248
+ "loss": 0.1307,
5249
+ "step": 21050
5250
+ },
5251
+ {
5252
+ "epoch": 0.73,
5253
+ "learning_rate": 1.606530612244898e-06,
5254
+ "loss": 0.1696,
5255
+ "step": 21075
5256
+ },
5257
+ {
5258
+ "epoch": 0.73,
5259
+ "learning_rate": 1.596326530612245e-06,
5260
+ "loss": 0.1247,
5261
+ "step": 21100
5262
+ },
5263
+ {
5264
+ "epoch": 0.73,
5265
+ "learning_rate": 1.586122448979592e-06,
5266
+ "loss": 0.1712,
5267
+ "step": 21125
5268
+ },
5269
+ {
5270
+ "epoch": 0.73,
5271
+ "learning_rate": 1.5759183673469388e-06,
5272
+ "loss": 0.1345,
5273
+ "step": 21150
5274
+ },
5275
+ {
5276
+ "epoch": 0.74,
5277
+ "learning_rate": 1.5657142857142859e-06,
5278
+ "loss": 0.1496,
5279
+ "step": 21175
5280
+ },
5281
+ {
5282
+ "epoch": 0.74,
5283
+ "learning_rate": 1.5555102040816327e-06,
5284
+ "loss": 0.1436,
5285
+ "step": 21200
5286
+ },
5287
+ {
5288
+ "epoch": 0.74,
5289
+ "learning_rate": 1.5453061224489796e-06,
5290
+ "loss": 0.1344,
5291
+ "step": 21225
5292
+ },
5293
+ {
5294
+ "epoch": 0.74,
5295
+ "learning_rate": 1.5351020408163266e-06,
5296
+ "loss": 0.1455,
5297
+ "step": 21250
5298
+ },
5299
+ {
5300
+ "epoch": 0.74,
5301
+ "learning_rate": 1.5248979591836735e-06,
5302
+ "loss": 0.1667,
5303
+ "step": 21275
5304
+ },
5305
+ {
5306
+ "epoch": 0.74,
5307
+ "learning_rate": 1.5146938775510206e-06,
5308
+ "loss": 0.1242,
5309
+ "step": 21300
5310
+ },
5311
+ {
5312
+ "epoch": 0.74,
5313
+ "learning_rate": 1.5044897959183674e-06,
5314
+ "loss": 0.1368,
5315
+ "step": 21325
5316
+ },
5317
+ {
5318
+ "epoch": 0.74,
5319
+ "learning_rate": 1.4942857142857145e-06,
5320
+ "loss": 0.128,
5321
+ "step": 21350
5322
+ },
5323
+ {
5324
+ "epoch": 0.74,
5325
+ "learning_rate": 1.4840816326530613e-06,
5326
+ "loss": 0.1627,
5327
+ "step": 21375
5328
+ },
5329
+ {
5330
+ "epoch": 0.74,
5331
+ "learning_rate": 1.4738775510204084e-06,
5332
+ "loss": 0.1227,
5333
+ "step": 21400
5334
+ },
5335
+ {
5336
+ "epoch": 0.74,
5337
+ "learning_rate": 1.4636734693877552e-06,
5338
+ "loss": 0.1562,
5339
+ "step": 21425
5340
+ },
5341
+ {
5342
+ "epoch": 0.74,
5343
+ "learning_rate": 1.453469387755102e-06,
5344
+ "loss": 0.1284,
5345
+ "step": 21450
5346
+ },
5347
+ {
5348
+ "epoch": 0.75,
5349
+ "learning_rate": 1.4432653061224491e-06,
5350
+ "loss": 0.1524,
5351
+ "step": 21475
5352
+ },
5353
+ {
5354
+ "epoch": 0.75,
5355
+ "learning_rate": 1.433469387755102e-06,
5356
+ "loss": 0.1128,
5357
+ "step": 21500
5358
+ },
5359
+ {
5360
+ "epoch": 0.75,
5361
+ "learning_rate": 1.4232653061224491e-06,
5362
+ "loss": 0.1426,
5363
+ "step": 21525
5364
+ },
5365
+ {
5366
+ "epoch": 0.75,
5367
+ "learning_rate": 1.413061224489796e-06,
5368
+ "loss": 0.1194,
5369
+ "step": 21550
5370
+ },
5371
+ {
5372
+ "epoch": 0.75,
5373
+ "learning_rate": 1.402857142857143e-06,
5374
+ "loss": 0.1384,
5375
+ "step": 21575
5376
+ },
5377
+ {
5378
+ "epoch": 0.75,
5379
+ "learning_rate": 1.3926530612244899e-06,
5380
+ "loss": 0.1236,
5381
+ "step": 21600
5382
+ },
5383
+ {
5384
+ "epoch": 0.75,
5385
+ "learning_rate": 1.382448979591837e-06,
5386
+ "loss": 0.1448,
5387
+ "step": 21625
5388
+ },
5389
+ {
5390
+ "epoch": 0.75,
5391
+ "learning_rate": 1.3722448979591838e-06,
5392
+ "loss": 0.132,
5393
+ "step": 21650
5394
+ },
5395
+ {
5396
+ "epoch": 0.75,
5397
+ "learning_rate": 1.3620408163265309e-06,
5398
+ "loss": 0.1383,
5399
+ "step": 21675
5400
+ },
5401
+ {
5402
+ "epoch": 0.75,
5403
+ "learning_rate": 1.3518367346938777e-06,
5404
+ "loss": 0.1041,
5405
+ "step": 21700
5406
+ },
5407
+ {
5408
+ "epoch": 0.75,
5409
+ "learning_rate": 1.3416326530612246e-06,
5410
+ "loss": 0.1358,
5411
+ "step": 21725
5412
+ },
5413
+ {
5414
+ "epoch": 0.75,
5415
+ "learning_rate": 1.3314285714285716e-06,
5416
+ "loss": 0.1381,
5417
+ "step": 21750
5418
+ },
5419
+ {
5420
+ "epoch": 0.76,
5421
+ "learning_rate": 1.3212244897959185e-06,
5422
+ "loss": 0.1536,
5423
+ "step": 21775
5424
+ },
5425
+ {
5426
+ "epoch": 0.76,
5427
+ "learning_rate": 1.3110204081632655e-06,
5428
+ "loss": 0.1466,
5429
+ "step": 21800
5430
+ },
5431
+ {
5432
+ "epoch": 0.76,
5433
+ "learning_rate": 1.3008163265306124e-06,
5434
+ "loss": 0.148,
5435
+ "step": 21825
5436
+ },
5437
+ {
5438
+ "epoch": 0.76,
5439
+ "learning_rate": 1.2906122448979595e-06,
5440
+ "loss": 0.1201,
5441
+ "step": 21850
5442
+ },
5443
+ {
5444
+ "epoch": 0.76,
5445
+ "learning_rate": 1.2804081632653063e-06,
5446
+ "loss": 0.1492,
5447
+ "step": 21875
5448
+ },
5449
+ {
5450
+ "epoch": 0.76,
5451
+ "learning_rate": 1.2702040816326534e-06,
5452
+ "loss": 0.1376,
5453
+ "step": 21900
5454
+ },
5455
+ {
5456
+ "epoch": 0.76,
5457
+ "learning_rate": 1.26e-06,
5458
+ "loss": 0.147,
5459
+ "step": 21925
5460
+ },
5461
+ {
5462
+ "epoch": 0.76,
5463
+ "learning_rate": 1.249795918367347e-06,
5464
+ "loss": 0.145,
5465
+ "step": 21950
5466
+ },
5467
+ {
5468
+ "epoch": 0.76,
5469
+ "learning_rate": 1.239591836734694e-06,
5470
+ "loss": 0.1316,
5471
+ "step": 21975
5472
+ },
5473
+ {
5474
+ "epoch": 0.76,
5475
+ "learning_rate": 1.229387755102041e-06,
5476
+ "loss": 0.1304,
5477
+ "step": 22000
5478
+ },
5479
+ {
5480
+ "epoch": 0.76,
5481
+ "eval_loss": 0.15034320950508118,
5482
+ "eval_runtime": 31138.0589,
5483
+ "eval_samples_per_second": 0.997,
5484
+ "eval_steps_per_second": 0.062,
5485
+ "eval_wer": 0.08610310313285319,
5486
+ "step": 22000
5487
  }
5488
  ],
5489
  "max_steps": 25000,
5490
  "num_train_epochs": 1,
5491
+ "total_flos": 7.473561993216e+20,
5492
  "trial_name": null,
5493
  "trial_params": null
5494
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f07b5f25c323d9a36c85d57f748183f9d3d14494dbfa6b195ed15224d9deafeb
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bffe463d62be009d062d805c732b0374b40de43d96b166556a53b6a5998e2271
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a05d2b060ee61e97fa7da66f905940a11d0a5b14104c16f577979a5e29d6fa2d
3
- size 143532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1c65d9ff9a4749b75ba77cb308089cdd19fbec529f495b56e4e7b0c5010e235
3
+ size 150256