elsayedissa commited on
Commit
7d368e5
1 Parent(s): e30e002

Training in progress, step 18000

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0afdff77222472257de0ba34c36739c26555750dfdb27e7b6f2ffed753b60d4e
3
  size 12347192855
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8a90d99fa5e493d103776713ee2ee6b9fca9948380c5ad509565e448a0d3c2f
3
  size 12347192855
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a46b6b71b47c9b3cdde4a1d39cea580e81a6bbf969d5738793f41c7f51939ad5
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b690154ee77cdbea8a14f2965218b2b3790caf8955482e539a12fa93fdc53f8
3
  size 6173655480
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:009dbf954364aba5d9d07623c2bc3be59611e798b46569f55549a02c3eabd95e
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:537f8d4bc472f85d8aaced00700f04502f899659cbfa067468feee29a0f0233a
3
  size 14575
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b9006cc2f9bf0c15a1ea19a9fdf3bcc02dfdb969110e593a43ce6b82aac6cf3
3
  size 557
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90be1018263c5a8b0b93a055ffb96e8625ade303fea979c01cbf658c8aed8c95
3
  size 557
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3675e25dffe1ac318cb230ca76cdb84aec69a7bf2cb79f6d7b5f6265ec4df7c8
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9205cfac3dc2d508e85efbdf6cfda796cb8dde46c8f1bd9e062253fea417d2b3
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
- "epoch": 0.5900933735985282,
5
- "global_step": 17000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -4239,11 +4239,260 @@
4239
  "eval_steps_per_second": 0.062,
4240
  "eval_wer": 0.09140322894892396,
4241
  "step": 17000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4242
  }
4243
  ],
4244
  "max_steps": 25000,
4245
  "num_train_epochs": 1,
4246
- "total_flos": 5.775025176576e+20,
4247
  "trial_name": null,
4248
  "trial_params": null
4249
  }
 
1
  {
2
  "best_metric": 0.11536111854119403,
3
  "best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
4
+ "epoch": 0.6248047485160887,
5
+ "global_step": 18000,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
4239
  "eval_steps_per_second": 0.062,
4240
  "eval_wer": 0.09140322894892396,
4241
  "step": 17000
4242
+ },
4243
+ {
4244
+ "epoch": 0.59,
4245
+ "learning_rate": 3.259591836734694e-06,
4246
+ "loss": 0.1475,
4247
+ "step": 17025
4248
+ },
4249
+ {
4250
+ "epoch": 0.59,
4251
+ "learning_rate": 3.2493877551020407e-06,
4252
+ "loss": 0.1458,
4253
+ "step": 17050
4254
+ },
4255
+ {
4256
+ "epoch": 0.59,
4257
+ "learning_rate": 3.239183673469388e-06,
4258
+ "loss": 0.155,
4259
+ "step": 17075
4260
+ },
4261
+ {
4262
+ "epoch": 0.59,
4263
+ "learning_rate": 3.2289795918367352e-06,
4264
+ "loss": 0.1446,
4265
+ "step": 17100
4266
+ },
4267
+ {
4268
+ "epoch": 0.59,
4269
+ "learning_rate": 3.218775510204082e-06,
4270
+ "loss": 0.1645,
4271
+ "step": 17125
4272
+ },
4273
+ {
4274
+ "epoch": 0.6,
4275
+ "learning_rate": 3.2085714285714285e-06,
4276
+ "loss": 0.1445,
4277
+ "step": 17150
4278
+ },
4279
+ {
4280
+ "epoch": 0.6,
4281
+ "learning_rate": 3.198367346938776e-06,
4282
+ "loss": 0.1561,
4283
+ "step": 17175
4284
+ },
4285
+ {
4286
+ "epoch": 0.6,
4287
+ "learning_rate": 3.1881632653061227e-06,
4288
+ "loss": 0.1518,
4289
+ "step": 17200
4290
+ },
4291
+ {
4292
+ "epoch": 0.6,
4293
+ "learning_rate": 3.1779591836734697e-06,
4294
+ "loss": 0.1521,
4295
+ "step": 17225
4296
+ },
4297
+ {
4298
+ "epoch": 0.6,
4299
+ "learning_rate": 3.1677551020408164e-06,
4300
+ "loss": 0.1426,
4301
+ "step": 17250
4302
+ },
4303
+ {
4304
+ "epoch": 0.6,
4305
+ "learning_rate": 3.1575510204081634e-06,
4306
+ "loss": 0.145,
4307
+ "step": 17275
4308
+ },
4309
+ {
4310
+ "epoch": 0.6,
4311
+ "learning_rate": 3.1473469387755105e-06,
4312
+ "loss": 0.1287,
4313
+ "step": 17300
4314
+ },
4315
+ {
4316
+ "epoch": 0.6,
4317
+ "learning_rate": 3.1371428571428575e-06,
4318
+ "loss": 0.1598,
4319
+ "step": 17325
4320
+ },
4321
+ {
4322
+ "epoch": 0.6,
4323
+ "learning_rate": 3.126938775510204e-06,
4324
+ "loss": 0.1581,
4325
+ "step": 17350
4326
+ },
4327
+ {
4328
+ "epoch": 0.6,
4329
+ "learning_rate": 3.1167346938775512e-06,
4330
+ "loss": 0.1566,
4331
+ "step": 17375
4332
+ },
4333
+ {
4334
+ "epoch": 0.6,
4335
+ "learning_rate": 3.106530612244898e-06,
4336
+ "loss": 0.1291,
4337
+ "step": 17400
4338
+ },
4339
+ {
4340
+ "epoch": 0.6,
4341
+ "learning_rate": 3.0963265306122454e-06,
4342
+ "loss": 0.1532,
4343
+ "step": 17425
4344
+ },
4345
+ {
4346
+ "epoch": 0.61,
4347
+ "learning_rate": 3.086122448979592e-06,
4348
+ "loss": 0.146,
4349
+ "step": 17450
4350
+ },
4351
+ {
4352
+ "epoch": 0.61,
4353
+ "learning_rate": 3.075918367346939e-06,
4354
+ "loss": 0.157,
4355
+ "step": 17475
4356
+ },
4357
+ {
4358
+ "epoch": 0.61,
4359
+ "learning_rate": 3.0657142857142857e-06,
4360
+ "loss": 0.1252,
4361
+ "step": 17500
4362
+ },
4363
+ {
4364
+ "epoch": 0.61,
4365
+ "learning_rate": 3.055510204081633e-06,
4366
+ "loss": 0.1701,
4367
+ "step": 17525
4368
+ },
4369
+ {
4370
+ "epoch": 0.61,
4371
+ "learning_rate": 3.04530612244898e-06,
4372
+ "loss": 0.1354,
4373
+ "step": 17550
4374
+ },
4375
+ {
4376
+ "epoch": 0.61,
4377
+ "learning_rate": 3.035102040816327e-06,
4378
+ "loss": 0.1687,
4379
+ "step": 17575
4380
+ },
4381
+ {
4382
+ "epoch": 0.61,
4383
+ "learning_rate": 3.0248979591836735e-06,
4384
+ "loss": 0.1178,
4385
+ "step": 17600
4386
+ },
4387
+ {
4388
+ "epoch": 0.61,
4389
+ "learning_rate": 3.0146938775510206e-06,
4390
+ "loss": 0.1564,
4391
+ "step": 17625
4392
+ },
4393
+ {
4394
+ "epoch": 0.61,
4395
+ "learning_rate": 3.0044897959183677e-06,
4396
+ "loss": 0.1446,
4397
+ "step": 17650
4398
+ },
4399
+ {
4400
+ "epoch": 0.61,
4401
+ "learning_rate": 2.9942857142857147e-06,
4402
+ "loss": 0.1463,
4403
+ "step": 17675
4404
+ },
4405
+ {
4406
+ "epoch": 0.61,
4407
+ "learning_rate": 2.9840816326530614e-06,
4408
+ "loss": 0.1206,
4409
+ "step": 17700
4410
+ },
4411
+ {
4412
+ "epoch": 0.62,
4413
+ "learning_rate": 2.973877551020408e-06,
4414
+ "loss": 0.1473,
4415
+ "step": 17725
4416
+ },
4417
+ {
4418
+ "epoch": 0.62,
4419
+ "learning_rate": 2.9636734693877555e-06,
4420
+ "loss": 0.1327,
4421
+ "step": 17750
4422
+ },
4423
+ {
4424
+ "epoch": 0.62,
4425
+ "learning_rate": 2.9534693877551025e-06,
4426
+ "loss": 0.1654,
4427
+ "step": 17775
4428
+ },
4429
+ {
4430
+ "epoch": 0.62,
4431
+ "learning_rate": 2.943265306122449e-06,
4432
+ "loss": 0.1237,
4433
+ "step": 17800
4434
+ },
4435
+ {
4436
+ "epoch": 0.62,
4437
+ "learning_rate": 2.933061224489796e-06,
4438
+ "loss": 0.1615,
4439
+ "step": 17825
4440
+ },
4441
+ {
4442
+ "epoch": 0.62,
4443
+ "learning_rate": 2.922857142857143e-06,
4444
+ "loss": 0.142,
4445
+ "step": 17850
4446
+ },
4447
+ {
4448
+ "epoch": 0.62,
4449
+ "learning_rate": 2.9126530612244904e-06,
4450
+ "loss": 0.1608,
4451
+ "step": 17875
4452
+ },
4453
+ {
4454
+ "epoch": 0.62,
4455
+ "learning_rate": 2.902448979591837e-06,
4456
+ "loss": 0.1296,
4457
+ "step": 17900
4458
+ },
4459
+ {
4460
+ "epoch": 0.62,
4461
+ "learning_rate": 2.8922448979591836e-06,
4462
+ "loss": 0.1569,
4463
+ "step": 17925
4464
+ },
4465
+ {
4466
+ "epoch": 0.62,
4467
+ "learning_rate": 2.8820408163265307e-06,
4468
+ "loss": 0.1331,
4469
+ "step": 17950
4470
+ },
4471
+ {
4472
+ "epoch": 0.62,
4473
+ "learning_rate": 2.871836734693878e-06,
4474
+ "loss": 0.1872,
4475
+ "step": 17975
4476
+ },
4477
+ {
4478
+ "epoch": 0.62,
4479
+ "learning_rate": 2.861632653061225e-06,
4480
+ "loss": 0.1441,
4481
+ "step": 18000
4482
+ },
4483
+ {
4484
+ "epoch": 0.62,
4485
+ "eval_loss": 0.15962253510951996,
4486
+ "eval_runtime": 31151.8818,
4487
+ "eval_samples_per_second": 0.996,
4488
+ "eval_steps_per_second": 0.062,
4489
+ "eval_wer": 0.08993702591943148,
4490
+ "step": 18000
4491
  }
4492
  ],
4493
  "max_steps": 25000,
4494
  "num_train_epochs": 1,
4495
+ "total_flos": 6.114732539904e+20,
4496
  "trial_name": null,
4497
  "trial_params": null
4498
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a46b6b71b47c9b3cdde4a1d39cea580e81a6bbf969d5738793f41c7f51939ad5
3
  size 6173655480
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b690154ee77cdbea8a14f2965218b2b3790caf8955482e539a12fa93fdc53f8
3
  size 6173655480
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e78bc9cee84c391c8ae997316cec7200e593d160d602a2fd280af39d860d9f6a
3
- size 116636
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54f4f30667756976745aebfb4e84b0aa810e69c9ffaa1a0e78323a5e4c0623c1
3
+ size 123360