|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 9.727626459143968, |
|
"eval_steps": 1000, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.048638132295719845, |
|
"grad_norm": 7.16464900970459, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.0297, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.09727626459143969, |
|
"grad_norm": 4.811949729919434, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.786, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.14591439688715954, |
|
"grad_norm": 4.57177734375, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.5046, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.19455252918287938, |
|
"grad_norm": 4.517965793609619, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.4671, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.24319066147859922, |
|
"grad_norm": 4.065127849578857, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.4381, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2918287937743191, |
|
"grad_norm": 3.5869154930114746, |
|
"learning_rate": 3e-06, |
|
"loss": 0.4062, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.3404669260700389, |
|
"grad_norm": 4.262431621551514, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.3951, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.38910505836575876, |
|
"grad_norm": 4.2176713943481445, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.3862, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4377431906614786, |
|
"grad_norm": 3.8358681201934814, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.364, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.48638132295719844, |
|
"grad_norm": 3.8311147689819336, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3425, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.5350194552529183, |
|
"grad_norm": 3.7137532234191895, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 0.339, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5836575875486382, |
|
"grad_norm": 4.2145280838012695, |
|
"learning_rate": 6e-06, |
|
"loss": 0.3393, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.632295719844358, |
|
"grad_norm": 3.540754556655884, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 0.3154, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.6809338521400778, |
|
"grad_norm": 3.7712759971618652, |
|
"learning_rate": 7e-06, |
|
"loss": 0.319, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.7295719844357976, |
|
"grad_norm": 3.6147518157958984, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.3127, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.7782101167315175, |
|
"grad_norm": 3.6948869228363037, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.2965, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8268482490272373, |
|
"grad_norm": 3.350135087966919, |
|
"learning_rate": 8.5e-06, |
|
"loss": 0.3001, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.8754863813229572, |
|
"grad_norm": 3.7692646980285645, |
|
"learning_rate": 9e-06, |
|
"loss": 0.281, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.9241245136186771, |
|
"grad_norm": 3.2747561931610107, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.2707, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.9727626459143969, |
|
"grad_norm": 3.260094165802002, |
|
"learning_rate": 1e-05, |
|
"loss": 0.2708, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.0214007782101167, |
|
"grad_norm": 2.917370080947876, |
|
"learning_rate": 9.944444444444445e-06, |
|
"loss": 0.2415, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.0700389105058365, |
|
"grad_norm": 2.564760208129883, |
|
"learning_rate": 9.88888888888889e-06, |
|
"loss": 0.1714, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.1186770428015564, |
|
"grad_norm": 2.844514846801758, |
|
"learning_rate": 9.833333333333333e-06, |
|
"loss": 0.1842, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.1673151750972763, |
|
"grad_norm": 3.1123480796813965, |
|
"learning_rate": 9.777777777777779e-06, |
|
"loss": 0.171, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.2159533073929962, |
|
"grad_norm": 3.253563642501831, |
|
"learning_rate": 9.722222222222223e-06, |
|
"loss": 0.1773, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.264591439688716, |
|
"grad_norm": 2.3145253658294678, |
|
"learning_rate": 9.666666666666667e-06, |
|
"loss": 0.1779, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.3132295719844358, |
|
"grad_norm": 2.493607521057129, |
|
"learning_rate": 9.611111111111112e-06, |
|
"loss": 0.1717, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.3618677042801557, |
|
"grad_norm": 2.081571578979492, |
|
"learning_rate": 9.555555555555556e-06, |
|
"loss": 0.1747, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.4105058365758754, |
|
"grad_norm": 2.9359965324401855, |
|
"learning_rate": 9.5e-06, |
|
"loss": 0.1591, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.4591439688715953, |
|
"grad_norm": 2.9549202919006348, |
|
"learning_rate": 9.444444444444445e-06, |
|
"loss": 0.1556, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.5077821011673151, |
|
"grad_norm": 3.0995261669158936, |
|
"learning_rate": 9.38888888888889e-06, |
|
"loss": 0.1553, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.556420233463035, |
|
"grad_norm": 2.3519606590270996, |
|
"learning_rate": 9.333333333333334e-06, |
|
"loss": 0.1502, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.605058365758755, |
|
"grad_norm": 2.3519246578216553, |
|
"learning_rate": 9.277777777777778e-06, |
|
"loss": 0.1398, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.6536964980544746, |
|
"grad_norm": 2.292020797729492, |
|
"learning_rate": 9.222222222222224e-06, |
|
"loss": 0.1518, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.7023346303501945, |
|
"grad_norm": 2.7264564037323, |
|
"learning_rate": 9.166666666666666e-06, |
|
"loss": 0.1393, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.7509727626459144, |
|
"grad_norm": 2.202409267425537, |
|
"learning_rate": 9.111111111111112e-06, |
|
"loss": 0.1432, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.7996108949416343, |
|
"grad_norm": 2.6967380046844482, |
|
"learning_rate": 9.055555555555556e-06, |
|
"loss": 0.1492, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.8482490272373542, |
|
"grad_norm": 2.8609344959259033, |
|
"learning_rate": 9e-06, |
|
"loss": 0.1351, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.8968871595330739, |
|
"grad_norm": 1.9660629034042358, |
|
"learning_rate": 8.944444444444446e-06, |
|
"loss": 0.1378, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 1.9455252918287937, |
|
"grad_norm": 2.9408161640167236, |
|
"learning_rate": 8.888888888888888e-06, |
|
"loss": 0.1429, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.9455252918287937, |
|
"eval_loss": 0.2754097580909729, |
|
"eval_runtime": 2154.7003, |
|
"eval_samples_per_second": 2.497, |
|
"eval_steps_per_second": 0.156, |
|
"eval_wer": 0.22082388961787744, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.9941634241245136, |
|
"grad_norm": 2.337907075881958, |
|
"learning_rate": 8.833333333333334e-06, |
|
"loss": 0.137, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.0428015564202333, |
|
"grad_norm": 1.9045900106430054, |
|
"learning_rate": 8.777777777777778e-06, |
|
"loss": 0.0728, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.0914396887159534, |
|
"grad_norm": 2.0690503120422363, |
|
"learning_rate": 8.722222222222224e-06, |
|
"loss": 0.0615, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.140077821011673, |
|
"grad_norm": 2.0722174644470215, |
|
"learning_rate": 8.666666666666668e-06, |
|
"loss": 0.0582, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.188715953307393, |
|
"grad_norm": 1.7977361679077148, |
|
"learning_rate": 8.611111111111112e-06, |
|
"loss": 0.0589, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 2.237354085603113, |
|
"grad_norm": 1.5375025272369385, |
|
"learning_rate": 8.555555555555556e-06, |
|
"loss": 0.0657, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.2859922178988326, |
|
"grad_norm": 1.9439505338668823, |
|
"learning_rate": 8.5e-06, |
|
"loss": 0.0648, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 2.3346303501945527, |
|
"grad_norm": 1.7909560203552246, |
|
"learning_rate": 8.444444444444446e-06, |
|
"loss": 0.0594, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.3832684824902723, |
|
"grad_norm": 1.7498406171798706, |
|
"learning_rate": 8.38888888888889e-06, |
|
"loss": 0.0627, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 2.4319066147859925, |
|
"grad_norm": 1.557906985282898, |
|
"learning_rate": 8.333333333333334e-06, |
|
"loss": 0.0585, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 2.480544747081712, |
|
"grad_norm": 1.3372323513031006, |
|
"learning_rate": 8.277777777777778e-06, |
|
"loss": 0.0614, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 2.529182879377432, |
|
"grad_norm": 1.8402326107025146, |
|
"learning_rate": 8.222222222222222e-06, |
|
"loss": 0.0601, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.5778210116731515, |
|
"grad_norm": 2.003889560699463, |
|
"learning_rate": 8.166666666666668e-06, |
|
"loss": 0.0575, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 2.6264591439688716, |
|
"grad_norm": 1.5208065509796143, |
|
"learning_rate": 8.111111111111112e-06, |
|
"loss": 0.0563, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.6750972762645917, |
|
"grad_norm": 2.1009633541107178, |
|
"learning_rate": 8.055555555555557e-06, |
|
"loss": 0.0633, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 2.7237354085603114, |
|
"grad_norm": 1.367943286895752, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.0572, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.772373540856031, |
|
"grad_norm": 1.914768934249878, |
|
"learning_rate": 7.944444444444445e-06, |
|
"loss": 0.0541, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 2.8210116731517507, |
|
"grad_norm": 2.1714184284210205, |
|
"learning_rate": 7.88888888888889e-06, |
|
"loss": 0.059, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.869649805447471, |
|
"grad_norm": 1.951029658317566, |
|
"learning_rate": 7.833333333333333e-06, |
|
"loss": 0.0571, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 2.9182879377431905, |
|
"grad_norm": 1.6378835439682007, |
|
"learning_rate": 7.77777777777778e-06, |
|
"loss": 0.0563, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.9669260700389106, |
|
"grad_norm": 1.8460659980773926, |
|
"learning_rate": 7.722222222222223e-06, |
|
"loss": 0.0556, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 3.0155642023346303, |
|
"grad_norm": 1.0332473516464233, |
|
"learning_rate": 7.666666666666667e-06, |
|
"loss": 0.045, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 3.06420233463035, |
|
"grad_norm": 1.5045939683914185, |
|
"learning_rate": 7.611111111111111e-06, |
|
"loss": 0.0227, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 3.11284046692607, |
|
"grad_norm": 2.1660637855529785, |
|
"learning_rate": 7.555555555555556e-06, |
|
"loss": 0.0275, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 3.1614785992217898, |
|
"grad_norm": 1.159175992012024, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.024, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 3.21011673151751, |
|
"grad_norm": 0.9240212440490723, |
|
"learning_rate": 7.444444444444445e-06, |
|
"loss": 0.0271, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 3.2587548638132295, |
|
"grad_norm": 0.9973633885383606, |
|
"learning_rate": 7.38888888888889e-06, |
|
"loss": 0.0266, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 3.307392996108949, |
|
"grad_norm": 1.5561943054199219, |
|
"learning_rate": 7.333333333333333e-06, |
|
"loss": 0.0267, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 3.3560311284046693, |
|
"grad_norm": 1.2981663942337036, |
|
"learning_rate": 7.277777777777778e-06, |
|
"loss": 0.0284, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 3.404669260700389, |
|
"grad_norm": 1.1085954904556274, |
|
"learning_rate": 7.222222222222223e-06, |
|
"loss": 0.0239, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 3.453307392996109, |
|
"grad_norm": 1.219834566116333, |
|
"learning_rate": 7.166666666666667e-06, |
|
"loss": 0.0244, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 3.501945525291829, |
|
"grad_norm": 1.1478172540664673, |
|
"learning_rate": 7.111111111111112e-06, |
|
"loss": 0.0273, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 3.5505836575875485, |
|
"grad_norm": 1.5546625852584839, |
|
"learning_rate": 7.055555555555557e-06, |
|
"loss": 0.0242, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 3.5992217898832686, |
|
"grad_norm": 1.5303540229797363, |
|
"learning_rate": 7e-06, |
|
"loss": 0.0248, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 3.6478599221789882, |
|
"grad_norm": 1.2220957279205322, |
|
"learning_rate": 6.944444444444445e-06, |
|
"loss": 0.0268, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 3.6964980544747084, |
|
"grad_norm": 1.076720952987671, |
|
"learning_rate": 6.88888888888889e-06, |
|
"loss": 0.0246, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 3.745136186770428, |
|
"grad_norm": 0.6398268342018127, |
|
"learning_rate": 6.833333333333334e-06, |
|
"loss": 0.0251, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 3.7937743190661477, |
|
"grad_norm": 1.0957626104354858, |
|
"learning_rate": 6.777777777777779e-06, |
|
"loss": 0.024, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 3.842412451361868, |
|
"grad_norm": 1.3786637783050537, |
|
"learning_rate": 6.7222222222222235e-06, |
|
"loss": 0.0221, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 3.8910505836575875, |
|
"grad_norm": 1.0472965240478516, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.0232, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.8910505836575875, |
|
"eval_loss": 0.2916410565376282, |
|
"eval_runtime": 2132.1371, |
|
"eval_samples_per_second": 2.524, |
|
"eval_steps_per_second": 0.158, |
|
"eval_wer": 0.19905364030378941, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.9396887159533076, |
|
"grad_norm": 1.0177409648895264, |
|
"learning_rate": 6.6111111111111115e-06, |
|
"loss": 0.0238, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 3.9883268482490273, |
|
"grad_norm": 1.0486767292022705, |
|
"learning_rate": 6.555555555555556e-06, |
|
"loss": 0.0245, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 4.036964980544747, |
|
"grad_norm": 0.7433611750602722, |
|
"learning_rate": 6.5000000000000004e-06, |
|
"loss": 0.0138, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 4.085603112840467, |
|
"grad_norm": 0.4771675765514374, |
|
"learning_rate": 6.444444444444445e-06, |
|
"loss": 0.0109, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 4.134241245136187, |
|
"grad_norm": 0.6232401132583618, |
|
"learning_rate": 6.3888888888888885e-06, |
|
"loss": 0.0109, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 4.182879377431907, |
|
"grad_norm": 0.6728478670120239, |
|
"learning_rate": 6.333333333333333e-06, |
|
"loss": 0.0103, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 4.2315175097276265, |
|
"grad_norm": 1.6818970441818237, |
|
"learning_rate": 6.277777777777778e-06, |
|
"loss": 0.0114, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 4.280155642023346, |
|
"grad_norm": 0.36247119307518005, |
|
"learning_rate": 6.222222222222223e-06, |
|
"loss": 0.011, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 4.328793774319066, |
|
"grad_norm": 1.3438676595687866, |
|
"learning_rate": 6.166666666666667e-06, |
|
"loss": 0.0126, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 4.377431906614786, |
|
"grad_norm": 0.8739299774169922, |
|
"learning_rate": 6.111111111111112e-06, |
|
"loss": 0.0104, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 4.426070038910506, |
|
"grad_norm": 0.6817359328269958, |
|
"learning_rate": 6.055555555555555e-06, |
|
"loss": 0.0118, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 4.474708171206226, |
|
"grad_norm": 0.5845357775688171, |
|
"learning_rate": 6e-06, |
|
"loss": 0.0125, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 4.523346303501945, |
|
"grad_norm": 1.067460298538208, |
|
"learning_rate": 5.944444444444445e-06, |
|
"loss": 0.0099, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 4.571984435797665, |
|
"grad_norm": 0.7725332379341125, |
|
"learning_rate": 5.88888888888889e-06, |
|
"loss": 0.0123, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 4.620622568093385, |
|
"grad_norm": 0.8168444633483887, |
|
"learning_rate": 5.833333333333334e-06, |
|
"loss": 0.01, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 4.669260700389105, |
|
"grad_norm": 0.7538411617279053, |
|
"learning_rate": 5.777777777777778e-06, |
|
"loss": 0.0098, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 4.717898832684825, |
|
"grad_norm": 0.3141120970249176, |
|
"learning_rate": 5.722222222222222e-06, |
|
"loss": 0.0107, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 4.766536964980545, |
|
"grad_norm": 0.887184202671051, |
|
"learning_rate": 5.666666666666667e-06, |
|
"loss": 0.0095, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 4.815175097276264, |
|
"grad_norm": 0.5846179127693176, |
|
"learning_rate": 5.611111111111112e-06, |
|
"loss": 0.0099, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 4.863813229571985, |
|
"grad_norm": 0.3938518464565277, |
|
"learning_rate": 5.555555555555557e-06, |
|
"loss": 0.0091, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 4.912451361867705, |
|
"grad_norm": 0.9777388572692871, |
|
"learning_rate": 5.500000000000001e-06, |
|
"loss": 0.0093, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 4.961089494163424, |
|
"grad_norm": 0.7862864136695862, |
|
"learning_rate": 5.444444444444445e-06, |
|
"loss": 0.0111, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 5.009727626459144, |
|
"grad_norm": 0.7504919171333313, |
|
"learning_rate": 5.388888888888889e-06, |
|
"loss": 0.0094, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 5.058365758754864, |
|
"grad_norm": 0.2351156324148178, |
|
"learning_rate": 5.333333333333334e-06, |
|
"loss": 0.0049, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 5.107003891050583, |
|
"grad_norm": 0.13562794029712677, |
|
"learning_rate": 5.2777777777777785e-06, |
|
"loss": 0.0043, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 5.155642023346304, |
|
"grad_norm": 0.33899495005607605, |
|
"learning_rate": 5.2222222222222226e-06, |
|
"loss": 0.0048, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 5.2042801556420235, |
|
"grad_norm": 0.9408302307128906, |
|
"learning_rate": 5.1666666666666675e-06, |
|
"loss": 0.0051, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 5.252918287937743, |
|
"grad_norm": 0.35573306679725647, |
|
"learning_rate": 5.1111111111111115e-06, |
|
"loss": 0.0053, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 5.301556420233463, |
|
"grad_norm": 0.7676318883895874, |
|
"learning_rate": 5.0555555555555555e-06, |
|
"loss": 0.0042, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 5.3501945525291825, |
|
"grad_norm": 0.6776612401008606, |
|
"learning_rate": 5e-06, |
|
"loss": 0.004, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 5.398832684824903, |
|
"grad_norm": 1.0338044166564941, |
|
"learning_rate": 4.944444444444445e-06, |
|
"loss": 0.0047, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 5.447470817120623, |
|
"grad_norm": 0.2744717001914978, |
|
"learning_rate": 4.888888888888889e-06, |
|
"loss": 0.0068, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 5.496108949416342, |
|
"grad_norm": 0.7075155973434448, |
|
"learning_rate": 4.833333333333333e-06, |
|
"loss": 0.0061, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 5.544747081712062, |
|
"grad_norm": 0.1820390671491623, |
|
"learning_rate": 4.777777777777778e-06, |
|
"loss": 0.0052, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 5.593385214007782, |
|
"grad_norm": 0.33624935150146484, |
|
"learning_rate": 4.722222222222222e-06, |
|
"loss": 0.0038, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 5.642023346303502, |
|
"grad_norm": 0.1461501270532608, |
|
"learning_rate": 4.666666666666667e-06, |
|
"loss": 0.0046, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 5.690661478599222, |
|
"grad_norm": 0.33367106318473816, |
|
"learning_rate": 4.611111111111112e-06, |
|
"loss": 0.0042, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 5.739299610894942, |
|
"grad_norm": 0.32830071449279785, |
|
"learning_rate": 4.555555555555556e-06, |
|
"loss": 0.0043, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 5.787937743190661, |
|
"grad_norm": 1.2332854270935059, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.0039, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 5.836575875486381, |
|
"grad_norm": 0.6161640286445618, |
|
"learning_rate": 4.444444444444444e-06, |
|
"loss": 0.0046, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 5.836575875486381, |
|
"eval_loss": 0.3218872845172882, |
|
"eval_runtime": 2114.549, |
|
"eval_samples_per_second": 2.545, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 0.18784047079406735, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 5.885214007782102, |
|
"grad_norm": 2.608947992324829, |
|
"learning_rate": 4.388888888888889e-06, |
|
"loss": 0.0061, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 5.933852140077821, |
|
"grad_norm": 0.25549063086509705, |
|
"learning_rate": 4.333333333333334e-06, |
|
"loss": 0.0044, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 5.982490272373541, |
|
"grad_norm": 1.4720717668533325, |
|
"learning_rate": 4.277777777777778e-06, |
|
"loss": 0.0051, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 6.031128404669261, |
|
"grad_norm": 0.7722542881965637, |
|
"learning_rate": 4.222222222222223e-06, |
|
"loss": 0.0037, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 6.07976653696498, |
|
"grad_norm": 0.35967299342155457, |
|
"learning_rate": 4.166666666666667e-06, |
|
"loss": 0.0029, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 6.1284046692607, |
|
"grad_norm": 0.7472084164619446, |
|
"learning_rate": 4.111111111111111e-06, |
|
"loss": 0.0029, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 6.1770428015564205, |
|
"grad_norm": 0.08338561654090881, |
|
"learning_rate": 4.055555555555556e-06, |
|
"loss": 0.0023, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 6.22568093385214, |
|
"grad_norm": 0.49297797679901123, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.0022, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 6.27431906614786, |
|
"grad_norm": 0.08119112253189087, |
|
"learning_rate": 3.944444444444445e-06, |
|
"loss": 0.0019, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 6.3229571984435795, |
|
"grad_norm": 0.13106492161750793, |
|
"learning_rate": 3.88888888888889e-06, |
|
"loss": 0.0027, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 6.3715953307393, |
|
"grad_norm": 0.18573686480522156, |
|
"learning_rate": 3.833333333333334e-06, |
|
"loss": 0.003, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 6.42023346303502, |
|
"grad_norm": 0.09228133410215378, |
|
"learning_rate": 3.777777777777778e-06, |
|
"loss": 0.004, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 6.468871595330739, |
|
"grad_norm": 0.8169698715209961, |
|
"learning_rate": 3.7222222222222225e-06, |
|
"loss": 0.0029, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 6.517509727626459, |
|
"grad_norm": 0.23130229115486145, |
|
"learning_rate": 3.6666666666666666e-06, |
|
"loss": 0.0026, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 6.566147859922179, |
|
"grad_norm": 0.10798731446266174, |
|
"learning_rate": 3.6111111111111115e-06, |
|
"loss": 0.0036, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 6.614785992217898, |
|
"grad_norm": 0.09145894646644592, |
|
"learning_rate": 3.555555555555556e-06, |
|
"loss": 0.0024, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 6.663424124513619, |
|
"grad_norm": 0.3099443018436432, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.0018, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 6.712062256809339, |
|
"grad_norm": 0.13762035965919495, |
|
"learning_rate": 3.444444444444445e-06, |
|
"loss": 0.0021, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 6.760700389105058, |
|
"grad_norm": 0.11253529787063599, |
|
"learning_rate": 3.3888888888888893e-06, |
|
"loss": 0.0012, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 6.809338521400778, |
|
"grad_norm": 0.054800793528556824, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 0.0019, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 6.857976653696498, |
|
"grad_norm": 0.06803842633962631, |
|
"learning_rate": 3.277777777777778e-06, |
|
"loss": 0.0024, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 6.906614785992218, |
|
"grad_norm": 0.08114957064390182, |
|
"learning_rate": 3.2222222222222227e-06, |
|
"loss": 0.0019, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 6.955252918287938, |
|
"grad_norm": 0.06155654788017273, |
|
"learning_rate": 3.1666666666666667e-06, |
|
"loss": 0.0017, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 7.003891050583658, |
|
"grad_norm": 0.09375619143247604, |
|
"learning_rate": 3.1111111111111116e-06, |
|
"loss": 0.0015, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 7.052529182879377, |
|
"grad_norm": 0.04578198865056038, |
|
"learning_rate": 3.055555555555556e-06, |
|
"loss": 0.0011, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 7.101167315175097, |
|
"grad_norm": 0.036920215934515, |
|
"learning_rate": 3e-06, |
|
"loss": 0.0012, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 7.1498054474708175, |
|
"grad_norm": 0.07471055537462234, |
|
"learning_rate": 2.944444444444445e-06, |
|
"loss": 0.0009, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 7.198443579766537, |
|
"grad_norm": 0.05370425805449486, |
|
"learning_rate": 2.888888888888889e-06, |
|
"loss": 0.0009, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 7.247081712062257, |
|
"grad_norm": 0.04795600846409798, |
|
"learning_rate": 2.8333333333333335e-06, |
|
"loss": 0.001, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 7.2957198443579765, |
|
"grad_norm": 0.030901705846190453, |
|
"learning_rate": 2.7777777777777783e-06, |
|
"loss": 0.0007, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 7.344357976653696, |
|
"grad_norm": 0.05389130115509033, |
|
"learning_rate": 2.7222222222222224e-06, |
|
"loss": 0.0012, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 7.392996108949417, |
|
"grad_norm": 0.05646834522485733, |
|
"learning_rate": 2.666666666666667e-06, |
|
"loss": 0.0009, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 7.441634241245136, |
|
"grad_norm": 0.04085018113255501, |
|
"learning_rate": 2.6111111111111113e-06, |
|
"loss": 0.001, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 7.490272373540856, |
|
"grad_norm": 0.024638062343001366, |
|
"learning_rate": 2.5555555555555557e-06, |
|
"loss": 0.0009, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 7.538910505836576, |
|
"grad_norm": 0.03571788966655731, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.0011, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 7.587548638132295, |
|
"grad_norm": 0.05909154936671257, |
|
"learning_rate": 2.4444444444444447e-06, |
|
"loss": 0.0007, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 7.636186770428015, |
|
"grad_norm": 0.028353404253721237, |
|
"learning_rate": 2.388888888888889e-06, |
|
"loss": 0.001, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 7.684824902723736, |
|
"grad_norm": 0.16029401123523712, |
|
"learning_rate": 2.3333333333333336e-06, |
|
"loss": 0.0009, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 7.733463035019455, |
|
"grad_norm": 0.04438905417919159, |
|
"learning_rate": 2.277777777777778e-06, |
|
"loss": 0.0008, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 7.782101167315175, |
|
"grad_norm": 0.035433579236269, |
|
"learning_rate": 2.222222222222222e-06, |
|
"loss": 0.0009, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 7.782101167315175, |
|
"eval_loss": 0.3454054296016693, |
|
"eval_runtime": 2148.9325, |
|
"eval_samples_per_second": 2.504, |
|
"eval_steps_per_second": 0.157, |
|
"eval_wer": 0.1831881983379061, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 7.830739299610895, |
|
"grad_norm": 0.028996312990784645, |
|
"learning_rate": 2.166666666666667e-06, |
|
"loss": 0.0008, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 7.879377431906615, |
|
"grad_norm": 0.038502488285303116, |
|
"learning_rate": 2.1111111111111114e-06, |
|
"loss": 0.0014, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 7.928015564202335, |
|
"grad_norm": 0.06861083209514618, |
|
"learning_rate": 2.0555555555555555e-06, |
|
"loss": 0.0009, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 7.976653696498055, |
|
"grad_norm": 0.07425787299871445, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.0011, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 8.025291828793774, |
|
"grad_norm": 0.025021173059940338, |
|
"learning_rate": 1.944444444444445e-06, |
|
"loss": 0.0006, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 8.073929961089494, |
|
"grad_norm": 0.031610701233148575, |
|
"learning_rate": 1.888888888888889e-06, |
|
"loss": 0.0006, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 8.122568093385214, |
|
"grad_norm": 0.02210886962711811, |
|
"learning_rate": 1.8333333333333333e-06, |
|
"loss": 0.0008, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 8.171206225680933, |
|
"grad_norm": 0.03183664008975029, |
|
"learning_rate": 1.777777777777778e-06, |
|
"loss": 0.0007, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 8.219844357976653, |
|
"grad_norm": 0.024653365835547447, |
|
"learning_rate": 1.7222222222222224e-06, |
|
"loss": 0.0005, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 8.268482490272374, |
|
"grad_norm": 0.02229795977473259, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 0.0005, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 8.317120622568094, |
|
"grad_norm": 0.025678085163235664, |
|
"learning_rate": 1.6111111111111113e-06, |
|
"loss": 0.0005, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 8.365758754863814, |
|
"grad_norm": 0.025166384875774384, |
|
"learning_rate": 1.5555555555555558e-06, |
|
"loss": 0.0005, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 8.414396887159533, |
|
"grad_norm": 0.017260603606700897, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.0005, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 8.463035019455253, |
|
"grad_norm": 0.016240952536463737, |
|
"learning_rate": 1.4444444444444445e-06, |
|
"loss": 0.0005, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 8.511673151750973, |
|
"grad_norm": 0.021515797823667526, |
|
"learning_rate": 1.3888888888888892e-06, |
|
"loss": 0.0006, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 8.560311284046692, |
|
"grad_norm": 0.023731861263513565, |
|
"learning_rate": 1.3333333333333334e-06, |
|
"loss": 0.0006, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 8.608949416342412, |
|
"grad_norm": 0.027586709707975388, |
|
"learning_rate": 1.2777777777777779e-06, |
|
"loss": 0.0006, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 8.657587548638132, |
|
"grad_norm": 0.04710078611969948, |
|
"learning_rate": 1.2222222222222223e-06, |
|
"loss": 0.0005, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 8.706225680933851, |
|
"grad_norm": 0.030895188450813293, |
|
"learning_rate": 1.1666666666666668e-06, |
|
"loss": 0.0005, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 8.754863813229573, |
|
"grad_norm": 0.04257024824619293, |
|
"learning_rate": 1.111111111111111e-06, |
|
"loss": 0.0006, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 8.803501945525293, |
|
"grad_norm": 0.022679895162582397, |
|
"learning_rate": 1.0555555555555557e-06, |
|
"loss": 0.0006, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 8.852140077821012, |
|
"grad_norm": 0.028052836656570435, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.0006, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 8.900778210116732, |
|
"grad_norm": 0.022373031824827194, |
|
"learning_rate": 9.444444444444445e-07, |
|
"loss": 0.0005, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 8.949416342412452, |
|
"grad_norm": 0.02304161712527275, |
|
"learning_rate": 8.88888888888889e-07, |
|
"loss": 0.0005, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 8.998054474708171, |
|
"grad_norm": 0.020944086834788322, |
|
"learning_rate": 8.333333333333333e-07, |
|
"loss": 0.0005, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 9.04669260700389, |
|
"grad_norm": 0.018772481009364128, |
|
"learning_rate": 7.777777777777779e-07, |
|
"loss": 0.0005, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 9.09533073929961, |
|
"grad_norm": 0.019685110077261925, |
|
"learning_rate": 7.222222222222222e-07, |
|
"loss": 0.0005, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 9.14396887159533, |
|
"grad_norm": 0.017540233209729195, |
|
"learning_rate": 6.666666666666667e-07, |
|
"loss": 0.0004, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 9.19260700389105, |
|
"grad_norm": 0.022134965285658836, |
|
"learning_rate": 6.111111111111112e-07, |
|
"loss": 0.0004, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 9.24124513618677, |
|
"grad_norm": 0.01725279912352562, |
|
"learning_rate": 5.555555555555555e-07, |
|
"loss": 0.0005, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 9.289883268482491, |
|
"grad_norm": 0.01975845918059349, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 0.0004, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 9.33852140077821, |
|
"grad_norm": 0.021399999037384987, |
|
"learning_rate": 4.444444444444445e-07, |
|
"loss": 0.0005, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 9.38715953307393, |
|
"grad_norm": 0.020681940019130707, |
|
"learning_rate": 3.8888888888888895e-07, |
|
"loss": 0.0004, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 9.43579766536965, |
|
"grad_norm": 0.020026126876473427, |
|
"learning_rate": 3.3333333333333335e-07, |
|
"loss": 0.0005, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 9.48443579766537, |
|
"grad_norm": 0.030826667323708534, |
|
"learning_rate": 2.7777777777777776e-07, |
|
"loss": 0.0004, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 9.53307392996109, |
|
"grad_norm": 0.017975399270653725, |
|
"learning_rate": 2.2222222222222224e-07, |
|
"loss": 0.0004, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 9.581712062256809, |
|
"grad_norm": 0.014251296408474445, |
|
"learning_rate": 1.6666666666666668e-07, |
|
"loss": 0.0005, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 9.630350194552529, |
|
"grad_norm": 0.018128497526049614, |
|
"learning_rate": 1.1111111111111112e-07, |
|
"loss": 0.0005, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 9.678988326848248, |
|
"grad_norm": 0.018901441246271133, |
|
"learning_rate": 5.555555555555556e-08, |
|
"loss": 0.0005, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 9.727626459143968, |
|
"grad_norm": 0.01760680228471756, |
|
"learning_rate": 0.0, |
|
"loss": 0.0004, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 9.727626459143968, |
|
"eval_loss": 0.36376532912254333, |
|
"eval_runtime": 2157.1617, |
|
"eval_samples_per_second": 2.494, |
|
"eval_steps_per_second": 0.156, |
|
"eval_wer": 0.18173684838363355, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 9.727626459143968, |
|
"step": 5000, |
|
"total_flos": 5.4332453168676864e+20, |
|
"train_loss": 0.06844911024216563, |
|
"train_runtime": 58020.7767, |
|
"train_samples_per_second": 2.758, |
|
"train_steps_per_second": 0.086 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.4332453168676864e+20, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|