elsayedissa
commited on
Commit
•
39d68ac
1
Parent(s):
fefba33
Training in progress, step 6000
Browse files- last-checkpoint/optimizer.pt +1 -1
- last-checkpoint/pytorch_model.bin +1 -1
- last-checkpoint/rng_state.pth +2 -2
- last-checkpoint/scaler.pt +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +252 -3
- pytorch_model.bin +1 -1
- runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0 +2 -2
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 12347192855
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f0460400a279e7766ba1a5c0ca78380731e0774cf4d291edfbc84654ee4a6aca
|
3 |
size 12347192855
|
last-checkpoint/pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e0ff690a614ad4feb1584be1b22e3316419b9b45d45121399a24f7b7785e65c
|
3 |
size 6173655480
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3788ae40196f57221f45ac4d182ab0f36e1ff61517502c4d51687db9b9b71e1f
|
3 |
+
size 14511
|
last-checkpoint/scaler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 557
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da3defc4121d1071efa04d9a3c24a084fc57cd835b7f2d21ed3bc283ad291901
|
3 |
size 557
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 627
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c16fd3410fdd3660983e7e486d0b41995f7675ed13e0a940e3532e9d1a7db51
|
3 |
size 627
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
-
"epoch": 0.
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -1251,11 +1251,260 @@
|
|
1251 |
"eval_steps_per_second": 0.063,
|
1252 |
"eval_wer": 0.10732672221066432,
|
1253 |
"step": 5000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1254 |
}
|
1255 |
],
|
1256 |
"max_steps": 25000,
|
1257 |
"num_train_epochs": 1,
|
1258 |
-
"total_flos":
|
1259 |
"trial_name": null,
|
1260 |
"trial_params": null
|
1261 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 0.11536111854119403,
|
3 |
"best_model_checkpoint": "/storage/elsayedissa/whisper-large-v2-spanish/checkpoint-1000",
|
4 |
+
"epoch": 0.2082682495053629,
|
5 |
+
"global_step": 6000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
1251 |
"eval_steps_per_second": 0.063,
|
1252 |
"eval_wer": 0.10732672221066432,
|
1253 |
"step": 5000
|
1254 |
+
},
|
1255 |
+
{
|
1256 |
+
"epoch": 0.17,
|
1257 |
+
"learning_rate": 8.155918367346939e-06,
|
1258 |
+
"loss": 0.2001,
|
1259 |
+
"step": 5025
|
1260 |
+
},
|
1261 |
+
{
|
1262 |
+
"epoch": 0.18,
|
1263 |
+
"learning_rate": 8.145714285714287e-06,
|
1264 |
+
"loss": 0.1835,
|
1265 |
+
"step": 5050
|
1266 |
+
},
|
1267 |
+
{
|
1268 |
+
"epoch": 0.18,
|
1269 |
+
"learning_rate": 8.135510204081634e-06,
|
1270 |
+
"loss": 0.2016,
|
1271 |
+
"step": 5075
|
1272 |
+
},
|
1273 |
+
{
|
1274 |
+
"epoch": 0.18,
|
1275 |
+
"learning_rate": 8.12530612244898e-06,
|
1276 |
+
"loss": 0.1907,
|
1277 |
+
"step": 5100
|
1278 |
+
},
|
1279 |
+
{
|
1280 |
+
"epoch": 0.18,
|
1281 |
+
"learning_rate": 8.115102040816327e-06,
|
1282 |
+
"loss": 0.1955,
|
1283 |
+
"step": 5125
|
1284 |
+
},
|
1285 |
+
{
|
1286 |
+
"epoch": 0.18,
|
1287 |
+
"learning_rate": 8.104897959183675e-06,
|
1288 |
+
"loss": 0.1904,
|
1289 |
+
"step": 5150
|
1290 |
+
},
|
1291 |
+
{
|
1292 |
+
"epoch": 0.18,
|
1293 |
+
"learning_rate": 8.09469387755102e-06,
|
1294 |
+
"loss": 0.1986,
|
1295 |
+
"step": 5175
|
1296 |
+
},
|
1297 |
+
{
|
1298 |
+
"epoch": 0.18,
|
1299 |
+
"learning_rate": 8.084489795918368e-06,
|
1300 |
+
"loss": 0.1499,
|
1301 |
+
"step": 5200
|
1302 |
+
},
|
1303 |
+
{
|
1304 |
+
"epoch": 0.18,
|
1305 |
+
"learning_rate": 8.074285714285714e-06,
|
1306 |
+
"loss": 0.1963,
|
1307 |
+
"step": 5225
|
1308 |
+
},
|
1309 |
+
{
|
1310 |
+
"epoch": 0.18,
|
1311 |
+
"learning_rate": 8.064081632653063e-06,
|
1312 |
+
"loss": 0.1821,
|
1313 |
+
"step": 5250
|
1314 |
+
},
|
1315 |
+
{
|
1316 |
+
"epoch": 0.18,
|
1317 |
+
"learning_rate": 8.053877551020409e-06,
|
1318 |
+
"loss": 0.1766,
|
1319 |
+
"step": 5275
|
1320 |
+
},
|
1321 |
+
{
|
1322 |
+
"epoch": 0.18,
|
1323 |
+
"learning_rate": 8.043673469387756e-06,
|
1324 |
+
"loss": 0.1827,
|
1325 |
+
"step": 5300
|
1326 |
+
},
|
1327 |
+
{
|
1328 |
+
"epoch": 0.18,
|
1329 |
+
"learning_rate": 8.033469387755102e-06,
|
1330 |
+
"loss": 0.1911,
|
1331 |
+
"step": 5325
|
1332 |
+
},
|
1333 |
+
{
|
1334 |
+
"epoch": 0.19,
|
1335 |
+
"learning_rate": 8.02326530612245e-06,
|
1336 |
+
"loss": 0.1809,
|
1337 |
+
"step": 5350
|
1338 |
+
},
|
1339 |
+
{
|
1340 |
+
"epoch": 0.19,
|
1341 |
+
"learning_rate": 8.013061224489797e-06,
|
1342 |
+
"loss": 0.2068,
|
1343 |
+
"step": 5375
|
1344 |
+
},
|
1345 |
+
{
|
1346 |
+
"epoch": 0.19,
|
1347 |
+
"learning_rate": 8.002857142857143e-06,
|
1348 |
+
"loss": 0.1841,
|
1349 |
+
"step": 5400
|
1350 |
+
},
|
1351 |
+
{
|
1352 |
+
"epoch": 0.19,
|
1353 |
+
"learning_rate": 7.99265306122449e-06,
|
1354 |
+
"loss": 0.1834,
|
1355 |
+
"step": 5425
|
1356 |
+
},
|
1357 |
+
{
|
1358 |
+
"epoch": 0.19,
|
1359 |
+
"learning_rate": 7.982448979591836e-06,
|
1360 |
+
"loss": 0.1848,
|
1361 |
+
"step": 5450
|
1362 |
+
},
|
1363 |
+
{
|
1364 |
+
"epoch": 0.19,
|
1365 |
+
"learning_rate": 7.972244897959184e-06,
|
1366 |
+
"loss": 0.1938,
|
1367 |
+
"step": 5475
|
1368 |
+
},
|
1369 |
+
{
|
1370 |
+
"epoch": 0.19,
|
1371 |
+
"learning_rate": 7.962040816326531e-06,
|
1372 |
+
"loss": 0.1747,
|
1373 |
+
"step": 5500
|
1374 |
+
},
|
1375 |
+
{
|
1376 |
+
"epoch": 0.19,
|
1377 |
+
"learning_rate": 7.951836734693879e-06,
|
1378 |
+
"loss": 0.1929,
|
1379 |
+
"step": 5525
|
1380 |
+
},
|
1381 |
+
{
|
1382 |
+
"epoch": 0.19,
|
1383 |
+
"learning_rate": 7.941632653061226e-06,
|
1384 |
+
"loss": 0.1785,
|
1385 |
+
"step": 5550
|
1386 |
+
},
|
1387 |
+
{
|
1388 |
+
"epoch": 0.19,
|
1389 |
+
"learning_rate": 7.931428571428572e-06,
|
1390 |
+
"loss": 0.1894,
|
1391 |
+
"step": 5575
|
1392 |
+
},
|
1393 |
+
{
|
1394 |
+
"epoch": 0.19,
|
1395 |
+
"learning_rate": 7.92122448979592e-06,
|
1396 |
+
"loss": 0.172,
|
1397 |
+
"step": 5600
|
1398 |
+
},
|
1399 |
+
{
|
1400 |
+
"epoch": 0.2,
|
1401 |
+
"learning_rate": 7.911020408163265e-06,
|
1402 |
+
"loss": 0.2007,
|
1403 |
+
"step": 5625
|
1404 |
+
},
|
1405 |
+
{
|
1406 |
+
"epoch": 0.2,
|
1407 |
+
"learning_rate": 7.900816326530613e-06,
|
1408 |
+
"loss": 0.1493,
|
1409 |
+
"step": 5650
|
1410 |
+
},
|
1411 |
+
{
|
1412 |
+
"epoch": 0.2,
|
1413 |
+
"learning_rate": 7.89061224489796e-06,
|
1414 |
+
"loss": 0.2092,
|
1415 |
+
"step": 5675
|
1416 |
+
},
|
1417 |
+
{
|
1418 |
+
"epoch": 0.2,
|
1419 |
+
"learning_rate": 7.880408163265306e-06,
|
1420 |
+
"loss": 0.1795,
|
1421 |
+
"step": 5700
|
1422 |
+
},
|
1423 |
+
{
|
1424 |
+
"epoch": 0.2,
|
1425 |
+
"learning_rate": 7.870204081632653e-06,
|
1426 |
+
"loss": 0.1958,
|
1427 |
+
"step": 5725
|
1428 |
+
},
|
1429 |
+
{
|
1430 |
+
"epoch": 0.2,
|
1431 |
+
"learning_rate": 7.860000000000001e-06,
|
1432 |
+
"loss": 0.165,
|
1433 |
+
"step": 5750
|
1434 |
+
},
|
1435 |
+
{
|
1436 |
+
"epoch": 0.2,
|
1437 |
+
"learning_rate": 7.849795918367348e-06,
|
1438 |
+
"loss": 0.1826,
|
1439 |
+
"step": 5775
|
1440 |
+
},
|
1441 |
+
{
|
1442 |
+
"epoch": 0.2,
|
1443 |
+
"learning_rate": 7.839591836734694e-06,
|
1444 |
+
"loss": 0.162,
|
1445 |
+
"step": 5800
|
1446 |
+
},
|
1447 |
+
{
|
1448 |
+
"epoch": 0.2,
|
1449 |
+
"learning_rate": 7.829387755102042e-06,
|
1450 |
+
"loss": 0.1729,
|
1451 |
+
"step": 5825
|
1452 |
+
},
|
1453 |
+
{
|
1454 |
+
"epoch": 0.2,
|
1455 |
+
"learning_rate": 7.819183673469388e-06,
|
1456 |
+
"loss": 0.1912,
|
1457 |
+
"step": 5850
|
1458 |
+
},
|
1459 |
+
{
|
1460 |
+
"epoch": 0.2,
|
1461 |
+
"learning_rate": 7.808979591836735e-06,
|
1462 |
+
"loss": 0.181,
|
1463 |
+
"step": 5875
|
1464 |
+
},
|
1465 |
+
{
|
1466 |
+
"epoch": 0.2,
|
1467 |
+
"learning_rate": 7.798775510204083e-06,
|
1468 |
+
"loss": 0.1547,
|
1469 |
+
"step": 5900
|
1470 |
+
},
|
1471 |
+
{
|
1472 |
+
"epoch": 0.21,
|
1473 |
+
"learning_rate": 7.788571428571428e-06,
|
1474 |
+
"loss": 0.2268,
|
1475 |
+
"step": 5925
|
1476 |
+
},
|
1477 |
+
{
|
1478 |
+
"epoch": 0.21,
|
1479 |
+
"learning_rate": 7.778367346938776e-06,
|
1480 |
+
"loss": 0.1674,
|
1481 |
+
"step": 5950
|
1482 |
+
},
|
1483 |
+
{
|
1484 |
+
"epoch": 0.21,
|
1485 |
+
"learning_rate": 7.768163265306123e-06,
|
1486 |
+
"loss": 0.1934,
|
1487 |
+
"step": 5975
|
1488 |
+
},
|
1489 |
+
{
|
1490 |
+
"epoch": 0.21,
|
1491 |
+
"learning_rate": 7.75795918367347e-06,
|
1492 |
+
"loss": 0.1535,
|
1493 |
+
"step": 6000
|
1494 |
+
},
|
1495 |
+
{
|
1496 |
+
"epoch": 0.21,
|
1497 |
+
"eval_loss": 0.1971120834350586,
|
1498 |
+
"eval_runtime": 31128.1173,
|
1499 |
+
"eval_samples_per_second": 0.997,
|
1500 |
+
"eval_steps_per_second": 0.062,
|
1501 |
+
"eval_wer": 0.10863441680453598,
|
1502 |
+
"step": 6000
|
1503 |
}
|
1504 |
],
|
1505 |
"max_steps": 25000,
|
1506 |
"num_train_epochs": 1,
|
1507 |
+
"total_flos": 2.038244179968e+20,
|
1508 |
"trial_name": null,
|
1509 |
"trial_params": null
|
1510 |
}
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 6173655480
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5e0ff690a614ad4feb1584be1b22e3316419b9b45d45121399a24f7b7785e65c
|
3 |
size 6173655480
|
runs/Dec26_19-57-19_gpu07.cyverse.org/events.out.tfevents.1672109970.gpu07.cyverse.org.126369.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c471ec8b7d00e44f926ec8e72e6951f01dad13c670f4fcd66f750928299dbb9
|
3 |
+
size 43977
|