elsayedissa
commited on
Commit
•
34aa10e
1
Parent(s):
2685bd3
Training in progress, step 22000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +1 -1
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53ad543f0df9ec43766eb0c4be86887e32e3e4bd5f26ee276eea2ab28ae1ddef
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bffe463d62be009d062d805c732b0374b40de43d96b166556a53b6a5998e2271
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14575
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c4828b52c0813e99f8e1771ed2daad93ca33f628aeb2f5d64700f0bf5092a00e
|
3 |
size 14575
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:96e64a1f7ca480b449676f7014db4dccc62eddeb2f60d979889930ac4af623a1
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0810f9c5189858aa63fe3a8926a77123b48758bcece57c05805766ff2c826811
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -5235,11 +5235,260 @@
|
|
5235 |
"eval_steps_per_second": 0.062,
|
5236 |
"eval_wer": 0.08848403192624074,
|
5237 |
"step": 21000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5238 |
}
|
5239 |
],
|
5240 |
"max_steps": 25000,
|
5241 |
"num_train_epochs": 1,
|
5242 |
-
"total_flos": 7.
|
5243 |
"trial_name": null,
|
5244 |
"trial_params": null
|
5245 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.7636502481863306,
|
5 |
+
"global_step": 22000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
5235 |
"eval_steps_per_second": 0.062,
|
5236 |
"eval_wer": 0.08848403192624074,
|
5237 |
"step": 21000
|
5238 |
+
},
|
5239 |
+
{
|
5240 |
+
"epoch": 0.73,
|
5241 |
+
"learning_rate": 1.6269387755102041e-06,
|
5242 |
+
"loss": 0.1562,
|
5243 |
+
"step": 21025
|
5244 |
+
},
|
5245 |
+
{
|
5246 |
+
"epoch": 0.73,
|
5247 |
+
"learning_rate": 1.616734693877551e-06,
|
5248 |
+
"loss": 0.1307,
|
5249 |
+
"step": 21050
|
5250 |
+
},
|
5251 |
+
{
|
5252 |
+
"epoch": 0.73,
|
5253 |
+
"learning_rate": 1.606530612244898e-06,
|
5254 |
+
"loss": 0.1696,
|
5255 |
+
"step": 21075
|
5256 |
+
},
|
5257 |
+
{
|
5258 |
+
"epoch": 0.73,
|
5259 |
+
"learning_rate": 1.596326530612245e-06,
|
5260 |
+
"loss": 0.1247,
|
5261 |
+
"step": 21100
|
5262 |
+
},
|
5263 |
+
{
|
5264 |
+
"epoch": 0.73,
|
5265 |
+
"learning_rate": 1.586122448979592e-06,
|
5266 |
+
"loss": 0.1712,
|
5267 |
+
"step": 21125
|
5268 |
+
},
|
5269 |
+
{
|
5270 |
+
"epoch": 0.73,
|
5271 |
+
"learning_rate": 1.5759183673469388e-06,
|
5272 |
+
"loss": 0.1345,
|
5273 |
+
"step": 21150
|
5274 |
+
},
|
5275 |
+
{
|
5276 |
+
"epoch": 0.74,
|
5277 |
+
"learning_rate": 1.5657142857142859e-06,
|
5278 |
+
"loss": 0.1496,
|
5279 |
+
"step": 21175
|
5280 |
+
},
|
5281 |
+
{
|
5282 |
+
"epoch": 0.74,
|
5283 |
+
"learning_rate": 1.5555102040816327e-06,
|
5284 |
+
"loss": 0.1436,
|
5285 |
+
"step": 21200
|
5286 |
+
},
|
5287 |
+
{
|
5288 |
+
"epoch": 0.74,
|
5289 |
+
"learning_rate": 1.5453061224489796e-06,
|
5290 |
+
"loss": 0.1344,
|
5291 |
+
"step": 21225
|
5292 |
+
},
|
5293 |
+
{
|
5294 |
+
"epoch": 0.74,
|
5295 |
+
"learning_rate": 1.5351020408163266e-06,
|
5296 |
+
"loss": 0.1455,
|
5297 |
+
"step": 21250
|
5298 |
+
},
|
5299 |
+
{
|
5300 |
+
"epoch": 0.74,
|
5301 |
+
"learning_rate": 1.5248979591836735e-06,
|
5302 |
+
"loss": 0.1667,
|
5303 |
+
"step": 21275
|
5304 |
+
},
|
5305 |
+
{
|
5306 |
+
"epoch": 0.74,
|
5307 |
+
"learning_rate": 1.5146938775510206e-06,
|
5308 |
+
"loss": 0.1242,
|
5309 |
+
"step": 21300
|
5310 |
+
},
|
5311 |
+
{
|
5312 |
+
"epoch": 0.74,
|
5313 |
+
"learning_rate": 1.5044897959183674e-06,
|
5314 |
+
"loss": 0.1368,
|
5315 |
+
"step": 21325
|
5316 |
+
},
|
5317 |
+
{
|
5318 |
+
"epoch": 0.74,
|
5319 |
+
"learning_rate": 1.4942857142857145e-06,
|
5320 |
+
"loss": 0.128,
|
5321 |
+
"step": 21350
|
5322 |
+
},
|
5323 |
+
{
|
5324 |
+
"epoch": 0.74,
|
5325 |
+
"learning_rate": 1.4840816326530613e-06,
|
5326 |
+
"loss": 0.1627,
|
5327 |
+
"step": 21375
|
5328 |
+
},
|
5329 |
+
{
|
5330 |
+
"epoch": 0.74,
|
5331 |
+
"learning_rate": 1.4738775510204084e-06,
|
5332 |
+
"loss": 0.1227,
|
5333 |
+
"step": 21400
|
5334 |
+
},
|
5335 |
+
{
|
5336 |
+
"epoch": 0.74,
|
5337 |
+
"learning_rate": 1.4636734693877552e-06,
|
5338 |
+
"loss": 0.1562,
|
5339 |
+
"step": 21425
|
5340 |
+
},
|
5341 |
+
{
|
5342 |
+
"epoch": 0.74,
|
5343 |
+
"learning_rate": 1.453469387755102e-06,
|
5344 |
+
"loss": 0.1284,
|
5345 |
+
"step": 21450
|
5346 |
+
},
|
5347 |
+
{
|
5348 |
+
"epoch": 0.75,
|
5349 |
+
"learning_rate": 1.4432653061224491e-06,
|
5350 |
+
"loss": 0.1524,
|
5351 |
+
"step": 21475
|
5352 |
+
},
|
5353 |
+
{
|
5354 |
+
"epoch": 0.75,
|
5355 |
+
"learning_rate": 1.433469387755102e-06,
|
5356 |
+
"loss": 0.1128,
|
5357 |
+
"step": 21500
|
5358 |
+
},
|
5359 |
+
{
|
5360 |
+
"epoch": 0.75,
|
5361 |
+
"learning_rate": 1.4232653061224491e-06,
|
5362 |
+
"loss": 0.1426,
|
5363 |
+
"step": 21525
|
5364 |
+
},
|
5365 |
+
{
|
5366 |
+
"epoch": 0.75,
|
5367 |
+
"learning_rate": 1.413061224489796e-06,
|
5368 |
+
"loss": 0.1194,
|
5369 |
+
"step": 21550
|
5370 |
+
},
|
5371 |
+
{
|
5372 |
+
"epoch": 0.75,
|
5373 |
+
"learning_rate": 1.402857142857143e-06,
|
5374 |
+
"loss": 0.1384,
|
5375 |
+
"step": 21575
|
5376 |
+
},
|
5377 |
+
{
|
5378 |
+
"epoch": 0.75,
|
5379 |
+
"learning_rate": 1.3926530612244899e-06,
|
5380 |
+
"loss": 0.1236,
|
5381 |
+
"step": 21600
|
5382 |
+
},
|
5383 |
+
{
|
5384 |
+
"epoch": 0.75,
|
5385 |
+
"learning_rate": 1.382448979591837e-06,
|
5386 |
+
"loss": 0.1448,
|
5387 |
+
"step": 21625
|
5388 |
+
},
|
5389 |
+
{
|
5390 |
+
"epoch": 0.75,
|
5391 |
+
"learning_rate": 1.3722448979591838e-06,
|
5392 |
+
"loss": 0.132,
|
5393 |
+
"step": 21650
|
5394 |
+
},
|
5395 |
+
{
|
5396 |
+
"epoch": 0.75,
|
5397 |
+
"learning_rate": 1.3620408163265309e-06,
|
5398 |
+
"loss": 0.1383,
|
5399 |
+
"step": 21675
|
5400 |
+
},
|
5401 |
+
{
|
5402 |
+
"epoch": 0.75,
|
5403 |
+
"learning_rate": 1.3518367346938777e-06,
|
5404 |
+
"loss": 0.1041,
|
5405 |
+
"step": 21700
|
5406 |
+
},
|
5407 |
+
{
|
5408 |
+
"epoch": 0.75,
|
5409 |
+
"learning_rate": 1.3416326530612246e-06,
|
5410 |
+
"loss": 0.1358,
|
5411 |
+
"step": 21725
|
5412 |
+
},
|
5413 |
+
{
|
5414 |
+
"epoch": 0.75,
|
5415 |
+
"learning_rate": 1.3314285714285716e-06,
|
5416 |
+
"loss": 0.1381,
|
5417 |
+
"step": 21750
|
5418 |
+
},
|
5419 |
+
{
|
5420 |
+
"epoch": 0.76,
|
5421 |
+
"learning_rate": 1.3212244897959185e-06,
|
5422 |
+
"loss": 0.1536,
|
5423 |
+
"step": 21775
|
5424 |
+
},
|
5425 |
+
{
|
5426 |
+
"epoch": 0.76,
|
5427 |
+
"learning_rate": 1.3110204081632655e-06,
|
5428 |
+
"loss": 0.1466,
|
5429 |
+
"step": 21800
|
5430 |
+
},
|
5431 |
+
{
|
5432 |
+
"epoch": 0.76,
|
5433 |
+
"learning_rate": 1.3008163265306124e-06,
|
5434 |
+
"loss": 0.148,
|
5435 |
+
"step": 21825
|
5436 |
+
},
|
5437 |
+
{
|
5438 |
+
"epoch": 0.76,
|
5439 |
+
"learning_rate": 1.2906122448979595e-06,
|
5440 |
+
"loss": 0.1201,
|
5441 |
+
"step": 21850
|
5442 |
+
},
|
5443 |
+
{
|
5444 |
+
"epoch": 0.76,
|
5445 |
+
"learning_rate": 1.2804081632653063e-06,
|
5446 |
+
"loss": 0.1492,
|
5447 |
+
"step": 21875
|
5448 |
+
},
|
5449 |
+
{
|
5450 |
+
"epoch": 0.76,
|
5451 |
+
"learning_rate": 1.2702040816326534e-06,
|
5452 |
+
"loss": 0.1376,
|
5453 |
+
"step": 21900
|
5454 |
+
},
|
5455 |
+
{
|
5456 |
+
"epoch": 0.76,
|
5457 |
+
"learning_rate": 1.26e-06,
|
5458 |
+
"loss": 0.147,
|
5459 |
+
"step": 21925
|
5460 |
+
},
|
5461 |
+
{
|
5462 |
+
"epoch": 0.76,
|
5463 |
+
"learning_rate": 1.249795918367347e-06,
|
5464 |
+
"loss": 0.145,
|
5465 |
+
"step": 21950
|
5466 |
+
},
|
5467 |
+
{
|
5468 |
+
"epoch": 0.76,
|
5469 |
+
"learning_rate": 1.239591836734694e-06,
|
5470 |
+
"loss": 0.1316,
|
5471 |
+
"step": 21975
|
5472 |
+
},
|
5473 |
+
{
|
5474 |
+
"epoch": 0.76,
|
5475 |
+
"learning_rate": 1.229387755102041e-06,
|
5476 |
+
"loss": 0.1304,
|
5477 |
+
"step": 22000
|
5478 |
+
},
|
5479 |
+
{
|
5480 |
+
"epoch": 0.76,
|
5481 |
+
"eval_loss": 0.15034320950508118,
|
5482 |
+
"eval_runtime": 31138.0589,
|
5483 |
+
"eval_samples_per_second": 0.997,
|
5484 |
+
"eval_steps_per_second": 0.062,
|
5485 |
+
"eval_wer": 0.08610310313285319,
|
5486 |
+
"step": 22000
|
5487 |
}
|
5488 |
],
|
5489 |
"max_steps": 25000,
|
5490 |
"num_train_epochs": 1,
|
5491 |
+
"total_flos": 7.473561993216e+20,
|
5492 |
"trial_name": null,
|
5493 |
"trial_params": null
|
5494 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bffe463d62be009d062d805c732b0374b40de43d96b166556a53b6a5998e2271
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a1c65d9ff9a4749b75ba77cb308089cdd19fbec529f495b56e4e7b0c5010e235
|
3 |
+
size 150256
|