|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 1682, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0005945303210463733, |
|
"grad_norm": 23.17915866800298, |
|
"learning_rate": 5.91715976331361e-08, |
|
"loss": 1.3012, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.002972651605231867, |
|
"grad_norm": 23.37449539925455, |
|
"learning_rate": 2.958579881656805e-07, |
|
"loss": 1.3439, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.005945303210463734, |
|
"grad_norm": 10.391844977150628, |
|
"learning_rate": 5.91715976331361e-07, |
|
"loss": 1.2902, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0089179548156956, |
|
"grad_norm": 12.298658554924508, |
|
"learning_rate": 8.875739644970415e-07, |
|
"loss": 1.1528, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.011890606420927468, |
|
"grad_norm": 5.836227197737131, |
|
"learning_rate": 1.183431952662722e-06, |
|
"loss": 1.0462, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.014863258026159334, |
|
"grad_norm": 3.59897554922516, |
|
"learning_rate": 1.4792899408284026e-06, |
|
"loss": 0.9416, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0178359096313912, |
|
"grad_norm": 3.2741630752466273, |
|
"learning_rate": 1.775147928994083e-06, |
|
"loss": 0.9146, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.020808561236623068, |
|
"grad_norm": 2.905060653147668, |
|
"learning_rate": 2.0710059171597635e-06, |
|
"loss": 0.8965, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.023781212841854936, |
|
"grad_norm": 2.9848388765396066, |
|
"learning_rate": 2.366863905325444e-06, |
|
"loss": 0.8532, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0267538644470868, |
|
"grad_norm": 2.928858994497014, |
|
"learning_rate": 2.6627218934911246e-06, |
|
"loss": 0.8607, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.029726516052318668, |
|
"grad_norm": 2.9699519304565105, |
|
"learning_rate": 2.958579881656805e-06, |
|
"loss": 0.8293, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.032699167657550536, |
|
"grad_norm": 2.9924496822372846, |
|
"learning_rate": 3.2544378698224853e-06, |
|
"loss": 0.8376, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.0356718192627824, |
|
"grad_norm": 3.0189144878653127, |
|
"learning_rate": 3.550295857988166e-06, |
|
"loss": 0.8223, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.03864447086801427, |
|
"grad_norm": 3.022152891252942, |
|
"learning_rate": 3.846153846153847e-06, |
|
"loss": 0.808, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.041617122473246136, |
|
"grad_norm": 3.0781659112469577, |
|
"learning_rate": 4.142011834319527e-06, |
|
"loss": 0.7936, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.044589774078478, |
|
"grad_norm": 3.084476562677424, |
|
"learning_rate": 4.437869822485207e-06, |
|
"loss": 0.7752, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.04756242568370987, |
|
"grad_norm": 3.2098986890021237, |
|
"learning_rate": 4.733727810650888e-06, |
|
"loss": 0.7763, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.050535077288941736, |
|
"grad_norm": 3.19384567951823, |
|
"learning_rate": 5.029585798816569e-06, |
|
"loss": 0.7746, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0535077288941736, |
|
"grad_norm": 2.964927416048633, |
|
"learning_rate": 5.325443786982249e-06, |
|
"loss": 0.7682, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.05648038049940547, |
|
"grad_norm": 3.1019748122364277, |
|
"learning_rate": 5.621301775147929e-06, |
|
"loss": 0.7518, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.059453032104637336, |
|
"grad_norm": 3.2201430089223386, |
|
"learning_rate": 5.91715976331361e-06, |
|
"loss": 0.7436, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0624256837098692, |
|
"grad_norm": 2.927109670294165, |
|
"learning_rate": 6.21301775147929e-06, |
|
"loss": 0.7426, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.06539833531510107, |
|
"grad_norm": 2.8885844881352742, |
|
"learning_rate": 6.5088757396449705e-06, |
|
"loss": 0.74, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.06837098692033294, |
|
"grad_norm": 3.172326235032539, |
|
"learning_rate": 6.8047337278106515e-06, |
|
"loss": 0.7647, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.0713436385255648, |
|
"grad_norm": 2.8788604196009047, |
|
"learning_rate": 7.100591715976332e-06, |
|
"loss": 0.7341, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.07431629013079667, |
|
"grad_norm": 2.7082199945019956, |
|
"learning_rate": 7.396449704142013e-06, |
|
"loss": 0.7347, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.07728894173602854, |
|
"grad_norm": 2.877616745493528, |
|
"learning_rate": 7.692307692307694e-06, |
|
"loss": 0.7266, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.0802615933412604, |
|
"grad_norm": 2.873705604654048, |
|
"learning_rate": 7.988165680473373e-06, |
|
"loss": 0.7461, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.08323424494649227, |
|
"grad_norm": 3.188866051330491, |
|
"learning_rate": 8.284023668639054e-06, |
|
"loss": 0.7267, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.08620689655172414, |
|
"grad_norm": 2.7709923287552787, |
|
"learning_rate": 8.579881656804735e-06, |
|
"loss": 0.7183, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.089179548156956, |
|
"grad_norm": 2.8980920523843032, |
|
"learning_rate": 8.875739644970414e-06, |
|
"loss": 0.7204, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.09215219976218787, |
|
"grad_norm": 2.8127651120484893, |
|
"learning_rate": 9.171597633136095e-06, |
|
"loss": 0.7263, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.09512485136741974, |
|
"grad_norm": 3.099526975800466, |
|
"learning_rate": 9.467455621301776e-06, |
|
"loss": 0.7266, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.0980975029726516, |
|
"grad_norm": 2.847472407806089, |
|
"learning_rate": 9.763313609467457e-06, |
|
"loss": 0.7148, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.10107015457788347, |
|
"grad_norm": 3.1542218699949243, |
|
"learning_rate": 9.999989221415223e-06, |
|
"loss": 0.73, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.10404280618311534, |
|
"grad_norm": 2.7886695417326197, |
|
"learning_rate": 9.999611975827441e-06, |
|
"loss": 0.7205, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.1070154577883472, |
|
"grad_norm": 2.863907914649196, |
|
"learning_rate": 9.998695847471015e-06, |
|
"loss": 0.7291, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.10998810939357907, |
|
"grad_norm": 2.682699306725325, |
|
"learning_rate": 9.997240935090764e-06, |
|
"loss": 0.7179, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.11296076099881094, |
|
"grad_norm": 2.6714047020353893, |
|
"learning_rate": 9.9952473955043e-06, |
|
"loss": 0.7095, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.1159334126040428, |
|
"grad_norm": 2.919385127037625, |
|
"learning_rate": 9.992715443585123e-06, |
|
"loss": 0.7183, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.11890606420927467, |
|
"grad_norm": 2.758146983625847, |
|
"learning_rate": 9.989645352239467e-06, |
|
"loss": 0.7097, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.12187871581450654, |
|
"grad_norm": 2.478949177408348, |
|
"learning_rate": 9.986037452376875e-06, |
|
"loss": 0.7053, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.1248513674197384, |
|
"grad_norm": 2.7139657440292004, |
|
"learning_rate": 9.98189213287454e-06, |
|
"loss": 0.7141, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1278240190249703, |
|
"grad_norm": 2.525450968321056, |
|
"learning_rate": 9.977209840535388e-06, |
|
"loss": 0.712, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.13079667063020214, |
|
"grad_norm": 2.5660210355660085, |
|
"learning_rate": 9.971991080039912e-06, |
|
"loss": 0.7029, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.133769322235434, |
|
"grad_norm": 2.719791305848057, |
|
"learning_rate": 9.966236413891787e-06, |
|
"loss": 0.6874, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.1367419738406659, |
|
"grad_norm": 2.7013784687029174, |
|
"learning_rate": 9.959946462357237e-06, |
|
"loss": 0.7053, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.13971462544589774, |
|
"grad_norm": 2.4163450862939273, |
|
"learning_rate": 9.953121903398172e-06, |
|
"loss": 0.69, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.1426872770511296, |
|
"grad_norm": 2.6651187768588884, |
|
"learning_rate": 9.945763472599125e-06, |
|
"loss": 0.6974, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.1456599286563615, |
|
"grad_norm": 2.61783925406702, |
|
"learning_rate": 9.937871963087958e-06, |
|
"loss": 0.6964, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.14863258026159334, |
|
"grad_norm": 2.6420632329467826, |
|
"learning_rate": 9.929448225450375e-06, |
|
"loss": 0.6814, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.1516052318668252, |
|
"grad_norm": 2.585841442008173, |
|
"learning_rate": 9.920493167638256e-06, |
|
"loss": 0.6904, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.1545778834720571, |
|
"grad_norm": 2.772149995259441, |
|
"learning_rate": 9.911007754871773e-06, |
|
"loss": 0.6998, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.15755053507728894, |
|
"grad_norm": 2.633310745015164, |
|
"learning_rate": 9.900993009535365e-06, |
|
"loss": 0.6936, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.1605231866825208, |
|
"grad_norm": 2.5624767016075016, |
|
"learning_rate": 9.890450011067544e-06, |
|
"loss": 0.6901, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.1634958382877527, |
|
"grad_norm": 2.4863789713476803, |
|
"learning_rate": 9.879379895844537e-06, |
|
"loss": 0.7005, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.16646848989298454, |
|
"grad_norm": 2.6081863153211535, |
|
"learning_rate": 9.867783857057808e-06, |
|
"loss": 0.6739, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.1694411414982164, |
|
"grad_norm": 2.5731120093742854, |
|
"learning_rate": 9.855663144585456e-06, |
|
"loss": 0.6788, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.1724137931034483, |
|
"grad_norm": 2.4603786025187375, |
|
"learning_rate": 9.843019064857481e-06, |
|
"loss": 0.682, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.17538644470868014, |
|
"grad_norm": 2.5203036804775083, |
|
"learning_rate": 9.829852980714986e-06, |
|
"loss": 0.6592, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.178359096313912, |
|
"grad_norm": 2.4982207442211846, |
|
"learning_rate": 9.816166311263276e-06, |
|
"loss": 0.6678, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.1813317479191439, |
|
"grad_norm": 2.7749883913033395, |
|
"learning_rate": 9.801960531718898e-06, |
|
"loss": 0.6906, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.18430439952437574, |
|
"grad_norm": 2.604444101827745, |
|
"learning_rate": 9.787237173250641e-06, |
|
"loss": 0.6591, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.1872770511296076, |
|
"grad_norm": 2.4391472456168812, |
|
"learning_rate": 9.771997822814496e-06, |
|
"loss": 0.6697, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.1902497027348395, |
|
"grad_norm": 2.4616817852361366, |
|
"learning_rate": 9.756244122982608e-06, |
|
"loss": 0.6734, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.19322235434007135, |
|
"grad_norm": 2.5774348327106984, |
|
"learning_rate": 9.739977771766225e-06, |
|
"loss": 0.6588, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.1961950059453032, |
|
"grad_norm": 2.5087234779825116, |
|
"learning_rate": 9.723200522432683e-06, |
|
"loss": 0.6664, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.1991676575505351, |
|
"grad_norm": 2.3989927240738607, |
|
"learning_rate": 9.705914183316435e-06, |
|
"loss": 0.6803, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.20214030915576695, |
|
"grad_norm": 2.401172984855569, |
|
"learning_rate": 9.68812061762413e-06, |
|
"loss": 0.6636, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.2051129607609988, |
|
"grad_norm": 2.4421036295734426, |
|
"learning_rate": 9.669821743233794e-06, |
|
"loss": 0.6389, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.2080856123662307, |
|
"grad_norm": 2.577840676624431, |
|
"learning_rate": 9.65101953248811e-06, |
|
"loss": 0.6586, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.21105826397146255, |
|
"grad_norm": 2.3940226323264917, |
|
"learning_rate": 9.631716011981823e-06, |
|
"loss": 0.6479, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.2140309155766944, |
|
"grad_norm": 2.5250690081911786, |
|
"learning_rate": 9.611913262343321e-06, |
|
"loss": 0.6506, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.2170035671819263, |
|
"grad_norm": 2.4264301363563545, |
|
"learning_rate": 9.591613418010355e-06, |
|
"loss": 0.6357, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.21997621878715815, |
|
"grad_norm": 2.751369457099525, |
|
"learning_rate": 9.57081866699999e-06, |
|
"loss": 0.6551, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.22294887039239, |
|
"grad_norm": 2.7219604193356672, |
|
"learning_rate": 9.549531250672764e-06, |
|
"loss": 0.6463, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.2259215219976219, |
|
"grad_norm": 2.45624731672107, |
|
"learning_rate": 9.52775346349111e-06, |
|
"loss": 0.646, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.22889417360285375, |
|
"grad_norm": 2.544975764712231, |
|
"learning_rate": 9.505487652772038e-06, |
|
"loss": 0.6354, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.2318668252080856, |
|
"grad_norm": 2.374830073694884, |
|
"learning_rate": 9.482736218434144e-06, |
|
"loss": 0.6502, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.2348394768133175, |
|
"grad_norm": 2.422663231385471, |
|
"learning_rate": 9.459501612738915e-06, |
|
"loss": 0.6502, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.23781212841854935, |
|
"grad_norm": 2.490293745714411, |
|
"learning_rate": 9.435786340026434e-06, |
|
"loss": 0.6303, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.2407847800237812, |
|
"grad_norm": 2.6066004181742124, |
|
"learning_rate": 9.41159295644543e-06, |
|
"loss": 0.6486, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.2437574316290131, |
|
"grad_norm": 2.5218752196441434, |
|
"learning_rate": 9.38692406967778e-06, |
|
"loss": 0.6209, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.24673008323424495, |
|
"grad_norm": 2.389380886868179, |
|
"learning_rate": 9.36178233865743e-06, |
|
"loss": 0.6168, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.2497027348394768, |
|
"grad_norm": 2.5150661226784794, |
|
"learning_rate": 9.336170473283803e-06, |
|
"loss": 0.6171, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.25267538644470866, |
|
"grad_norm": 2.435724631506405, |
|
"learning_rate": 9.310091234129718e-06, |
|
"loss": 0.6299, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.2556480380499406, |
|
"grad_norm": 2.394787101037759, |
|
"learning_rate": 9.283547432143842e-06, |
|
"loss": 0.6051, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.25862068965517243, |
|
"grad_norm": 2.3112769318054602, |
|
"learning_rate": 9.256541928347701e-06, |
|
"loss": 0.6332, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.2615933412604043, |
|
"grad_norm": 2.395884463823803, |
|
"learning_rate": 9.229077633527317e-06, |
|
"loss": 0.6363, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.26456599286563615, |
|
"grad_norm": 2.5353977000342245, |
|
"learning_rate": 9.201157507919466e-06, |
|
"loss": 0.6096, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.267538644470868, |
|
"grad_norm": 2.4062321379781957, |
|
"learning_rate": 9.17278456089261e-06, |
|
"loss": 0.6214, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.27051129607609986, |
|
"grad_norm": 2.3853947296491254, |
|
"learning_rate": 9.143961850622523e-06, |
|
"loss": 0.6357, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.2734839476813318, |
|
"grad_norm": 2.3411923936500587, |
|
"learning_rate": 9.114692483762682e-06, |
|
"loss": 0.6111, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.27645659928656363, |
|
"grad_norm": 2.805658532417448, |
|
"learning_rate": 9.084979615109402e-06, |
|
"loss": 0.6338, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.2794292508917955, |
|
"grad_norm": 2.519691039204996, |
|
"learning_rate": 9.054826447261807e-06, |
|
"loss": 0.616, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.28240190249702735, |
|
"grad_norm": 2.5502443998981295, |
|
"learning_rate": 9.02423623027663e-06, |
|
"loss": 0.6215, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.2853745541022592, |
|
"grad_norm": 2.3581542332305374, |
|
"learning_rate": 8.993212261317911e-06, |
|
"loss": 0.6143, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.28834720570749106, |
|
"grad_norm": 2.2185371096529267, |
|
"learning_rate": 8.961757884301614e-06, |
|
"loss": 0.6103, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.291319857312723, |
|
"grad_norm": 2.388161926408038, |
|
"learning_rate": 8.929876489535196e-06, |
|
"loss": 0.6165, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.29429250891795483, |
|
"grad_norm": 2.3962973629889466, |
|
"learning_rate": 8.89757151335219e-06, |
|
"loss": 0.6124, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.2972651605231867, |
|
"grad_norm": 2.3456101381645746, |
|
"learning_rate": 8.864846437741817e-06, |
|
"loss": 0.6168, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.30023781212841855, |
|
"grad_norm": 2.857739205060792, |
|
"learning_rate": 8.831704789973677e-06, |
|
"loss": 0.601, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.3032104637336504, |
|
"grad_norm": 2.404200391558541, |
|
"learning_rate": 8.79815014221757e-06, |
|
"loss": 0.6159, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.30618311533888226, |
|
"grad_norm": 2.5066882094814185, |
|
"learning_rate": 8.764186111158468e-06, |
|
"loss": 0.6107, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.3091557669441142, |
|
"grad_norm": 2.439281663409053, |
|
"learning_rate": 8.729816357606683e-06, |
|
"loss": 0.6082, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.31212841854934603, |
|
"grad_norm": 2.434841960288868, |
|
"learning_rate": 8.695044586103297e-06, |
|
"loss": 0.6085, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.3151010701545779, |
|
"grad_norm": 2.3045925871070394, |
|
"learning_rate": 8.659874544520862e-06, |
|
"loss": 0.5752, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.31807372175980975, |
|
"grad_norm": 2.515840298273436, |
|
"learning_rate": 8.624310023659437e-06, |
|
"loss": 0.5954, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.3210463733650416, |
|
"grad_norm": 2.514641603087097, |
|
"learning_rate": 8.588354856838002e-06, |
|
"loss": 0.5923, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.32401902497027346, |
|
"grad_norm": 2.353964195966498, |
|
"learning_rate": 8.552012919481278e-06, |
|
"loss": 0.5785, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.3269916765755054, |
|
"grad_norm": 2.3620191543194657, |
|
"learning_rate": 8.515288128702015e-06, |
|
"loss": 0.5942, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.32996432818073723, |
|
"grad_norm": 2.5008196020127964, |
|
"learning_rate": 8.478184442878787e-06, |
|
"loss": 0.6116, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.3329369797859691, |
|
"grad_norm": 2.3614970118631833, |
|
"learning_rate": 8.440705861229344e-06, |
|
"loss": 0.5838, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.33590963139120095, |
|
"grad_norm": 2.3019970993481813, |
|
"learning_rate": 8.40285642337955e-06, |
|
"loss": 0.5979, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.3388822829964328, |
|
"grad_norm": 2.437131942480355, |
|
"learning_rate": 8.36464020892797e-06, |
|
"loss": 0.5583, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.34185493460166466, |
|
"grad_norm": 2.326966178552118, |
|
"learning_rate": 8.326061337006161e-06, |
|
"loss": 0.5877, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.3448275862068966, |
|
"grad_norm": 2.5231196817030455, |
|
"learning_rate": 8.287123965834682e-06, |
|
"loss": 0.5774, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.34780023781212843, |
|
"grad_norm": 2.445974147547591, |
|
"learning_rate": 8.247832292274909e-06, |
|
"loss": 0.5957, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.3507728894173603, |
|
"grad_norm": 2.30995967302817, |
|
"learning_rate": 8.208190551376668e-06, |
|
"loss": 0.5913, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.35374554102259215, |
|
"grad_norm": 2.304615444057916, |
|
"learning_rate": 8.16820301592176e-06, |
|
"loss": 0.5951, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.356718192627824, |
|
"grad_norm": 2.2337712440097466, |
|
"learning_rate": 8.127873995963439e-06, |
|
"loss": 0.5717, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.35969084423305586, |
|
"grad_norm": 2.352192550598562, |
|
"learning_rate": 8.087207838361819e-06, |
|
"loss": 0.5559, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.3626634958382878, |
|
"grad_norm": 2.4383288854885086, |
|
"learning_rate": 8.046208926315381e-06, |
|
"loss": 0.5621, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.36563614744351963, |
|
"grad_norm": 2.4669126509592982, |
|
"learning_rate": 8.004881678888518e-06, |
|
"loss": 0.5822, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.3686087990487515, |
|
"grad_norm": 2.350359506686674, |
|
"learning_rate": 7.963230550535212e-06, |
|
"loss": 0.5726, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.37158145065398335, |
|
"grad_norm": 2.3573597155480965, |
|
"learning_rate": 7.921260030618935e-06, |
|
"loss": 0.5615, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.3745541022592152, |
|
"grad_norm": 2.768754472930866, |
|
"learning_rate": 7.878974642928752e-06, |
|
"loss": 0.566, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.37752675386444706, |
|
"grad_norm": 2.460878358484002, |
|
"learning_rate": 7.836378945191718e-06, |
|
"loss": 0.5534, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.380499405469679, |
|
"grad_norm": 2.280914113206368, |
|
"learning_rate": 7.793477528581642e-06, |
|
"loss": 0.5655, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.38347205707491083, |
|
"grad_norm": 2.481778359904293, |
|
"learning_rate": 7.750275017224208e-06, |
|
"loss": 0.5676, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.3864447086801427, |
|
"grad_norm": 2.616616432524906, |
|
"learning_rate": 7.706776067698574e-06, |
|
"loss": 0.564, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.38941736028537455, |
|
"grad_norm": 2.371851297853059, |
|
"learning_rate": 7.662985368535465e-06, |
|
"loss": 0.5633, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.3923900118906064, |
|
"grad_norm": 2.3788259328454413, |
|
"learning_rate": 7.618907639711816e-06, |
|
"loss": 0.5478, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.39536266349583826, |
|
"grad_norm": 2.329685605293642, |
|
"learning_rate": 7.574547632142029e-06, |
|
"loss": 0.5496, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.3983353151010702, |
|
"grad_norm": 2.3986113985448867, |
|
"learning_rate": 7.529910127165897e-06, |
|
"loss": 0.5488, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.40130796670630203, |
|
"grad_norm": 2.314354656341197, |
|
"learning_rate": 7.48499993603325e-06, |
|
"loss": 0.5564, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.4042806183115339, |
|
"grad_norm": 2.301009486833614, |
|
"learning_rate": 7.4398218993853765e-06, |
|
"loss": 0.5579, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.40725326991676575, |
|
"grad_norm": 2.529267389965807, |
|
"learning_rate": 7.394380886733263e-06, |
|
"loss": 0.5418, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.4102259215219976, |
|
"grad_norm": 2.273496393101965, |
|
"learning_rate": 7.3486817959327524e-06, |
|
"loss": 0.5486, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.41319857312722946, |
|
"grad_norm": 2.362055285648788, |
|
"learning_rate": 7.302729552656613e-06, |
|
"loss": 0.5473, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.4161712247324614, |
|
"grad_norm": 2.568398004847708, |
|
"learning_rate": 7.256529109863637e-06, |
|
"loss": 0.5635, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.41914387633769323, |
|
"grad_norm": 2.334007582897662, |
|
"learning_rate": 7.210085447264777e-06, |
|
"loss": 0.5569, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.4221165279429251, |
|
"grad_norm": 2.27944109416587, |
|
"learning_rate": 7.163403570786416e-06, |
|
"loss": 0.5314, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.42508917954815695, |
|
"grad_norm": 2.439475832033061, |
|
"learning_rate": 7.116488512030798e-06, |
|
"loss": 0.5511, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.4280618311533888, |
|
"grad_norm": 2.3531415117791354, |
|
"learning_rate": 7.069345327733701e-06, |
|
"loss": 0.5457, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.43103448275862066, |
|
"grad_norm": 2.3295198969589035, |
|
"learning_rate": 7.021979099219396e-06, |
|
"loss": 0.5239, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.4340071343638526, |
|
"grad_norm": 2.5063503569942385, |
|
"learning_rate": 6.974394931852957e-06, |
|
"loss": 0.5384, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.43697978596908443, |
|
"grad_norm": 2.355478191724581, |
|
"learning_rate": 6.926597954489979e-06, |
|
"loss": 0.5098, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.4399524375743163, |
|
"grad_norm": 2.4091145093686372, |
|
"learning_rate": 6.878593318923763e-06, |
|
"loss": 0.5146, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.44292508917954815, |
|
"grad_norm": 2.3684497417326913, |
|
"learning_rate": 6.830386199330036e-06, |
|
"loss": 0.5146, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.44589774078478, |
|
"grad_norm": 2.4513900163292113, |
|
"learning_rate": 6.781981791709247e-06, |
|
"loss": 0.5152, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.44887039239001186, |
|
"grad_norm": 2.453629962173857, |
|
"learning_rate": 6.73338531332652e-06, |
|
"loss": 0.536, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.4518430439952438, |
|
"grad_norm": 2.474092048043491, |
|
"learning_rate": 6.6846020021493105e-06, |
|
"loss": 0.5236, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.45481569560047563, |
|
"grad_norm": 2.4532902284126523, |
|
"learning_rate": 6.63563711628283e-06, |
|
"loss": 0.518, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.4577883472057075, |
|
"grad_norm": 2.4470168591075163, |
|
"learning_rate": 6.586495933403306e-06, |
|
"loss": 0.5192, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.46076099881093935, |
|
"grad_norm": 2.483107107350786, |
|
"learning_rate": 6.53718375018912e-06, |
|
"loss": 0.5217, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.4637336504161712, |
|
"grad_norm": 2.433139401600069, |
|
"learning_rate": 6.4877058817499174e-06, |
|
"loss": 0.4948, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.4667063020214031, |
|
"grad_norm": 2.3018625992746684, |
|
"learning_rate": 6.4380676610537075e-06, |
|
"loss": 0.5156, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.469678953626635, |
|
"grad_norm": 2.425100167363479, |
|
"learning_rate": 6.388274438352053e-06, |
|
"loss": 0.5186, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.47265160523186683, |
|
"grad_norm": 2.4243934526405493, |
|
"learning_rate": 6.3383315806034014e-06, |
|
"loss": 0.5123, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.4756242568370987, |
|
"grad_norm": 2.3005885180264545, |
|
"learning_rate": 6.288244470894596e-06, |
|
"loss": 0.5147, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.47859690844233055, |
|
"grad_norm": 2.461720715630345, |
|
"learning_rate": 6.238018507860667e-06, |
|
"loss": 0.5192, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.4815695600475624, |
|
"grad_norm": 2.391513806976481, |
|
"learning_rate": 6.187659105102941e-06, |
|
"loss": 0.5114, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.4845422116527943, |
|
"grad_norm": 2.2043491836548363, |
|
"learning_rate": 6.1371716906055336e-06, |
|
"loss": 0.5277, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.4875148632580262, |
|
"grad_norm": 2.4124196938652074, |
|
"learning_rate": 6.086561706150292e-06, |
|
"loss": 0.5271, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.49048751486325803, |
|
"grad_norm": 2.365747909189894, |
|
"learning_rate": 6.035834606730258e-06, |
|
"loss": 0.4949, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.4934601664684899, |
|
"grad_norm": 2.4935417512792273, |
|
"learning_rate": 5.984995859961707e-06, |
|
"loss": 0.4807, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.49643281807372175, |
|
"grad_norm": 2.23752471394695, |
|
"learning_rate": 5.934050945494803e-06, |
|
"loss": 0.5053, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.4994054696789536, |
|
"grad_norm": 2.3920061376574187, |
|
"learning_rate": 5.883005354422995e-06, |
|
"loss": 0.5087, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.5023781212841855, |
|
"grad_norm": 2.35431470302546, |
|
"learning_rate": 5.83186458869115e-06, |
|
"loss": 0.51, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.5053507728894173, |
|
"grad_norm": 2.3257622845834653, |
|
"learning_rate": 5.7806341605025264e-06, |
|
"loss": 0.5092, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.5083234244946492, |
|
"grad_norm": 2.236833604753037, |
|
"learning_rate": 5.7293195917246426e-06, |
|
"loss": 0.5014, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.5112960760998811, |
|
"grad_norm": 2.408857775213423, |
|
"learning_rate": 5.67792641329411e-06, |
|
"loss": 0.4932, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.514268727705113, |
|
"grad_norm": 2.4984651682967782, |
|
"learning_rate": 5.6264601646204645e-06, |
|
"loss": 0.5, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.5172413793103449, |
|
"grad_norm": 2.6589983137124964, |
|
"learning_rate": 5.574926392989118e-06, |
|
"loss": 0.4953, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.5202140309155767, |
|
"grad_norm": 2.2920034526589474, |
|
"learning_rate": 5.523330652963443e-06, |
|
"loss": 0.4918, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.5231866825208086, |
|
"grad_norm": 2.4555058214366414, |
|
"learning_rate": 5.471678505786063e-06, |
|
"loss": 0.5018, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.5261593341260404, |
|
"grad_norm": 2.5114137922653446, |
|
"learning_rate": 5.419975518779443e-06, |
|
"loss": 0.4941, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.5291319857312723, |
|
"grad_norm": 2.271061287454104, |
|
"learning_rate": 5.36822726474582e-06, |
|
"loss": 0.4931, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.5321046373365041, |
|
"grad_norm": 2.381608236846904, |
|
"learning_rate": 5.3164393213665275e-06, |
|
"loss": 0.4785, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.535077288941736, |
|
"grad_norm": 2.3066959393356106, |
|
"learning_rate": 5.2646172706008154e-06, |
|
"loss": 0.49, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.5380499405469679, |
|
"grad_norm": 2.357712136319602, |
|
"learning_rate": 5.212766698084196e-06, |
|
"loss": 0.5027, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.5410225921521997, |
|
"grad_norm": 2.5267610099653406, |
|
"learning_rate": 5.160893192526395e-06, |
|
"loss": 0.4777, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.5439952437574316, |
|
"grad_norm": 2.359927663823654, |
|
"learning_rate": 5.109002345108971e-06, |
|
"loss": 0.4937, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.5469678953626635, |
|
"grad_norm": 2.41091168049318, |
|
"learning_rate": 5.0570997488826775e-06, |
|
"loss": 0.4757, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.5499405469678954, |
|
"grad_norm": 2.2779085550054305, |
|
"learning_rate": 5.005190998164614e-06, |
|
"loss": 0.4951, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.5529131985731273, |
|
"grad_norm": 2.2695911460870417, |
|
"learning_rate": 4.953281687935236e-06, |
|
"loss": 0.4712, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.5558858501783591, |
|
"grad_norm": 2.3975975966184198, |
|
"learning_rate": 4.901377413235312e-06, |
|
"loss": 0.4808, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.558858501783591, |
|
"grad_norm": 2.3802468272374253, |
|
"learning_rate": 4.849483768562849e-06, |
|
"loss": 0.4611, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.5618311533888228, |
|
"grad_norm": 2.324549924901845, |
|
"learning_rate": 4.797606347270104e-06, |
|
"loss": 0.4606, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.5648038049940547, |
|
"grad_norm": 2.8248050417928727, |
|
"learning_rate": 4.7457507409606954e-06, |
|
"loss": 0.4586, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.5677764565992865, |
|
"grad_norm": 2.3408008419334867, |
|
"learning_rate": 4.693922538886914e-06, |
|
"loss": 0.4673, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.5707491082045184, |
|
"grad_norm": 2.455876677604174, |
|
"learning_rate": 4.642127327347292e-06, |
|
"loss": 0.4803, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.5737217598097503, |
|
"grad_norm": 2.4097659884184717, |
|
"learning_rate": 4.590370689084474e-06, |
|
"loss": 0.4634, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.5766944114149821, |
|
"grad_norm": 2.5668401837693033, |
|
"learning_rate": 4.53865820268349e-06, |
|
"loss": 0.4598, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.579667063020214, |
|
"grad_norm": 2.2667882653983566, |
|
"learning_rate": 4.486995441970468e-06, |
|
"loss": 0.4606, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.582639714625446, |
|
"grad_norm": 2.3952069585748728, |
|
"learning_rate": 4.435387975411849e-06, |
|
"loss": 0.4554, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.5856123662306778, |
|
"grad_norm": 2.280919958503882, |
|
"learning_rate": 4.383841365514208e-06, |
|
"loss": 0.4765, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.5885850178359097, |
|
"grad_norm": 2.342556124656158, |
|
"learning_rate": 4.332361168224678e-06, |
|
"loss": 0.4613, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.5915576694411415, |
|
"grad_norm": 2.2952593312149405, |
|
"learning_rate": 4.280952932332128e-06, |
|
"loss": 0.446, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.5945303210463734, |
|
"grad_norm": 2.3321748153973183, |
|
"learning_rate": 4.229622198869063e-06, |
|
"loss": 0.452, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.5975029726516052, |
|
"grad_norm": 2.4823844105009423, |
|
"learning_rate": 4.178374500514397e-06, |
|
"loss": 0.449, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.6004756242568371, |
|
"grad_norm": 2.2950369831862134, |
|
"learning_rate": 4.127215360997117e-06, |
|
"loss": 0.4549, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.603448275862069, |
|
"grad_norm": 2.5166575033757477, |
|
"learning_rate": 4.076150294500897e-06, |
|
"loss": 0.4519, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.6064209274673008, |
|
"grad_norm": 2.3807622926087664, |
|
"learning_rate": 4.0251848050697615e-06, |
|
"loss": 0.4559, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.6093935790725327, |
|
"grad_norm": 2.351343124866575, |
|
"learning_rate": 3.974324386014828e-06, |
|
"loss": 0.4562, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.6123662306777645, |
|
"grad_norm": 2.2523827712367988, |
|
"learning_rate": 3.92357451932221e-06, |
|
"loss": 0.4453, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.6153388822829964, |
|
"grad_norm": 2.30496980280335, |
|
"learning_rate": 3.8729406750621486e-06, |
|
"loss": 0.4571, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.6183115338882283, |
|
"grad_norm": 2.27622501534253, |
|
"learning_rate": 3.82242831079941e-06, |
|
"loss": 0.4337, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.6212841854934602, |
|
"grad_norm": 2.2971541253075114, |
|
"learning_rate": 3.772042871005057e-06, |
|
"loss": 0.4417, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.6242568370986921, |
|
"grad_norm": 2.295568197082708, |
|
"learning_rate": 3.721789786469602e-06, |
|
"loss": 0.4482, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.6272294887039239, |
|
"grad_norm": 2.418242134420845, |
|
"learning_rate": 3.6716744737176567e-06, |
|
"loss": 0.4452, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.6302021403091558, |
|
"grad_norm": 2.258447637407748, |
|
"learning_rate": 3.6217023344241198e-06, |
|
"loss": 0.458, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.6331747919143876, |
|
"grad_norm": 2.3383623329057732, |
|
"learning_rate": 3.57187875483194e-06, |
|
"loss": 0.4484, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.6361474435196195, |
|
"grad_norm": 2.2709297436937885, |
|
"learning_rate": 3.5222091051715803e-06, |
|
"loss": 0.4385, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.6391200951248514, |
|
"grad_norm": 2.3001869154507357, |
|
"learning_rate": 3.472698739082172e-06, |
|
"loss": 0.4409, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.6420927467300832, |
|
"grad_norm": 2.2279803463182395, |
|
"learning_rate": 3.423352993034482e-06, |
|
"loss": 0.428, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.6450653983353151, |
|
"grad_norm": 2.1572728551701723, |
|
"learning_rate": 3.374177185755721e-06, |
|
"loss": 0.4195, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.6480380499405469, |
|
"grad_norm": 2.30995305892372, |
|
"learning_rate": 3.3251766176562605e-06, |
|
"loss": 0.4387, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.6510107015457788, |
|
"grad_norm": 2.363018968058027, |
|
"learning_rate": 3.2763565702583322e-06, |
|
"loss": 0.428, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.6539833531510107, |
|
"grad_norm": 2.3825197647163137, |
|
"learning_rate": 3.227722305626761e-06, |
|
"loss": 0.4299, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.6569560047562426, |
|
"grad_norm": 2.16179977498919, |
|
"learning_rate": 3.1792790658017874e-06, |
|
"loss": 0.4327, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.6599286563614745, |
|
"grad_norm": 2.175753238496728, |
|
"learning_rate": 3.1310320722340636e-06, |
|
"loss": 0.4204, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.6629013079667063, |
|
"grad_norm": 2.222441260772541, |
|
"learning_rate": 3.082986525221851e-06, |
|
"loss": 0.4316, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.6658739595719382, |
|
"grad_norm": 2.2178029194036415, |
|
"learning_rate": 3.035147603350511e-06, |
|
"loss": 0.4231, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.66884661117717, |
|
"grad_norm": 2.3519506089063786, |
|
"learning_rate": 2.9875204629343333e-06, |
|
"loss": 0.4231, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.6718192627824019, |
|
"grad_norm": 2.248473095466383, |
|
"learning_rate": 2.940110237460758e-06, |
|
"loss": 0.4166, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.6747919143876338, |
|
"grad_norm": 2.3278826893891793, |
|
"learning_rate": 2.8929220370370705e-06, |
|
"loss": 0.4184, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.6777645659928656, |
|
"grad_norm": 2.2527103466010794, |
|
"learning_rate": 2.845960947839598e-06, |
|
"loss": 0.4253, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.6807372175980975, |
|
"grad_norm": 2.3636767593489836, |
|
"learning_rate": 2.7992320315655043e-06, |
|
"loss": 0.4294, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.6837098692033293, |
|
"grad_norm": 2.2384821444252543, |
|
"learning_rate": 2.752740324887219e-06, |
|
"loss": 0.4164, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.6866825208085613, |
|
"grad_norm": 2.3602120268768783, |
|
"learning_rate": 2.706490838909547e-06, |
|
"loss": 0.4244, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.6896551724137931, |
|
"grad_norm": 2.474112669222775, |
|
"learning_rate": 2.6604885586295636e-06, |
|
"loss": 0.4137, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.692627824019025, |
|
"grad_norm": 2.2552175118807605, |
|
"learning_rate": 2.6147384423992893e-06, |
|
"loss": 0.428, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.6956004756242569, |
|
"grad_norm": 2.4189395286363258, |
|
"learning_rate": 2.5692454213912655e-06, |
|
"loss": 0.4244, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.6985731272294887, |
|
"grad_norm": 2.240213088622091, |
|
"learning_rate": 2.5240143990670513e-06, |
|
"loss": 0.4117, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.7015457788347206, |
|
"grad_norm": 2.2726272120032283, |
|
"learning_rate": 2.479050250648692e-06, |
|
"loss": 0.4296, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.7045184304399524, |
|
"grad_norm": 2.1850341107989975, |
|
"learning_rate": 2.4343578225932573e-06, |
|
"loss": 0.4132, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.7074910820451843, |
|
"grad_norm": 2.272873233207882, |
|
"learning_rate": 2.389941932070455e-06, |
|
"loss": 0.3981, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.7104637336504162, |
|
"grad_norm": 2.252150495601297, |
|
"learning_rate": 2.345807366443419e-06, |
|
"loss": 0.4115, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.713436385255648, |
|
"grad_norm": 2.266551085513327, |
|
"learning_rate": 2.3019588827527017e-06, |
|
"loss": 0.4086, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.7164090368608799, |
|
"grad_norm": 2.332350830052582, |
|
"learning_rate": 2.2584012072035366e-06, |
|
"loss": 0.4089, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.7193816884661117, |
|
"grad_norm": 2.2344702253310715, |
|
"learning_rate": 2.21513903465643e-06, |
|
"loss": 0.3998, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.7223543400713437, |
|
"grad_norm": 2.3267615219996935, |
|
"learning_rate": 2.1721770281211162e-06, |
|
"loss": 0.4039, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.7253269916765755, |
|
"grad_norm": 2.2320045041894945, |
|
"learning_rate": 2.1295198182539627e-06, |
|
"loss": 0.4106, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.7282996432818074, |
|
"grad_norm": 2.3252249478957574, |
|
"learning_rate": 2.0871720028588516e-06, |
|
"loss": 0.4067, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.7312722948870393, |
|
"grad_norm": 2.2267058392084604, |
|
"learning_rate": 2.045138146391607e-06, |
|
"loss": 0.4086, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.7342449464922711, |
|
"grad_norm": 2.411234372389695, |
|
"learning_rate": 2.0034227794680195e-06, |
|
"loss": 0.3969, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.737217598097503, |
|
"grad_norm": 2.1311399653064518, |
|
"learning_rate": 1.962030398375506e-06, |
|
"loss": 0.4165, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.7401902497027348, |
|
"grad_norm": 2.3782226385738836, |
|
"learning_rate": 1.920965464588481e-06, |
|
"loss": 0.3987, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.7431629013079667, |
|
"grad_norm": 2.3739915931205076, |
|
"learning_rate": 1.8802324042874847e-06, |
|
"loss": 0.3995, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.7461355529131986, |
|
"grad_norm": 2.3894925514110965, |
|
"learning_rate": 1.8398356078820934e-06, |
|
"loss": 0.3906, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.7491082045184304, |
|
"grad_norm": 2.3093439961839657, |
|
"learning_rate": 1.7997794295377098e-06, |
|
"loss": 0.3917, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.7520808561236623, |
|
"grad_norm": 2.319382769614651, |
|
"learning_rate": 1.7600681867062469e-06, |
|
"loss": 0.3885, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.7550535077288941, |
|
"grad_norm": 2.2803304350831413, |
|
"learning_rate": 1.7207061596607694e-06, |
|
"loss": 0.3927, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.7580261593341261, |
|
"grad_norm": 2.3220946916005567, |
|
"learning_rate": 1.6816975910341526e-06, |
|
"loss": 0.3962, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.760998810939358, |
|
"grad_norm": 2.264317075137385, |
|
"learning_rate": 1.6430466853617788e-06, |
|
"loss": 0.3978, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.7639714625445898, |
|
"grad_norm": 2.1705446882611654, |
|
"learning_rate": 1.6047576086283613e-06, |
|
"loss": 0.3912, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.7669441141498217, |
|
"grad_norm": 2.291864149582287, |
|
"learning_rate": 1.566834487818909e-06, |
|
"loss": 0.3997, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.7699167657550535, |
|
"grad_norm": 2.2779523408752276, |
|
"learning_rate": 1.5292814104739012e-06, |
|
"loss": 0.4015, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.7728894173602854, |
|
"grad_norm": 2.2360566853344532, |
|
"learning_rate": 1.492102424248717e-06, |
|
"loss": 0.3936, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.7758620689655172, |
|
"grad_norm": 2.3193326093728395, |
|
"learning_rate": 1.455301536477348e-06, |
|
"loss": 0.4006, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.7788347205707491, |
|
"grad_norm": 2.241071015791803, |
|
"learning_rate": 1.4188827137404814e-06, |
|
"loss": 0.3953, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.781807372175981, |
|
"grad_norm": 2.3345320959262397, |
|
"learning_rate": 1.3828498814379526e-06, |
|
"loss": 0.3827, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.7847800237812128, |
|
"grad_norm": 2.255050340004342, |
|
"learning_rate": 1.3472069233656455e-06, |
|
"loss": 0.395, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.7877526753864447, |
|
"grad_norm": 2.182203091512302, |
|
"learning_rate": 1.3119576812968893e-06, |
|
"loss": 0.3889, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.7907253269916765, |
|
"grad_norm": 2.266179435014493, |
|
"learning_rate": 1.2771059545683612e-06, |
|
"loss": 0.3974, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.7936979785969085, |
|
"grad_norm": 2.3267812132219663, |
|
"learning_rate": 1.242655499670583e-06, |
|
"loss": 0.3827, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.7966706302021404, |
|
"grad_norm": 2.1058933825314736, |
|
"learning_rate": 1.2086100298430225e-06, |
|
"loss": 0.3833, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.7996432818073722, |
|
"grad_norm": 2.1268029483679514, |
|
"learning_rate": 1.1749732146738651e-06, |
|
"loss": 0.3866, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.8026159334126041, |
|
"grad_norm": 2.1738354617676854, |
|
"learning_rate": 1.1417486797044886e-06, |
|
"loss": 0.3791, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.8055885850178359, |
|
"grad_norm": 2.160878200122144, |
|
"learning_rate": 1.1089400060386801e-06, |
|
"loss": 0.3755, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.8085612366230678, |
|
"grad_norm": 2.129626397999257, |
|
"learning_rate": 1.0765507299566552e-06, |
|
"loss": 0.372, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.8115338882282996, |
|
"grad_norm": 2.3501282748597947, |
|
"learning_rate": 1.0445843425338902e-06, |
|
"loss": 0.3856, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.8145065398335315, |
|
"grad_norm": 2.2841395141791883, |
|
"learning_rate": 1.0130442892648434e-06, |
|
"loss": 0.383, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.8174791914387634, |
|
"grad_norm": 2.2929912740203333, |
|
"learning_rate": 9.819339696915848e-07, |
|
"loss": 0.3826, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.8204518430439952, |
|
"grad_norm": 2.0935559084604436, |
|
"learning_rate": 9.512567370373643e-07, |
|
"loss": 0.3846, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.8234244946492271, |
|
"grad_norm": 2.2050607660086574, |
|
"learning_rate": 9.210158978452033e-07, |
|
"loss": 0.3831, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.8263971462544589, |
|
"grad_norm": 2.365214622091771, |
|
"learning_rate": 8.912147116214842e-07, |
|
"loss": 0.3766, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.8293697978596909, |
|
"grad_norm": 2.2601705624692547, |
|
"learning_rate": 8.618563904846294e-07, |
|
"loss": 0.3804, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.8323424494649228, |
|
"grad_norm": 2.4159041452669117, |
|
"learning_rate": 8.329440988188886e-07, |
|
"loss": 0.3924, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.8353151010701546, |
|
"grad_norm": 2.1795173707942643, |
|
"learning_rate": 8.04480952933257e-07, |
|
"loss": 0.3754, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.8382877526753865, |
|
"grad_norm": 2.2351438825547896, |
|
"learning_rate": 7.764700207255904e-07, |
|
"loss": 0.3897, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.8412604042806183, |
|
"grad_norm": 2.20058726362207, |
|
"learning_rate": 7.489143213519301e-07, |
|
"loss": 0.3775, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.8442330558858502, |
|
"grad_norm": 2.3495313166309213, |
|
"learning_rate": 7.218168249010821e-07, |
|
"loss": 0.3795, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.847205707491082, |
|
"grad_norm": 2.068114427890453, |
|
"learning_rate": 6.951804520744915e-07, |
|
"loss": 0.3754, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.8501783590963139, |
|
"grad_norm": 2.3055094145170445, |
|
"learning_rate": 6.690080738714267e-07, |
|
"loss": 0.3871, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.8531510107015458, |
|
"grad_norm": 2.1607153035130953, |
|
"learning_rate": 6.433025112795383e-07, |
|
"loss": 0.3649, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.8561236623067776, |
|
"grad_norm": 2.3452133211542847, |
|
"learning_rate": 6.180665349707892e-07, |
|
"loss": 0.3789, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.8590963139120095, |
|
"grad_norm": 2.270134523374029, |
|
"learning_rate": 5.93302865002825e-07, |
|
"loss": 0.3774, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.8620689655172413, |
|
"grad_norm": 2.188053068739294, |
|
"learning_rate": 5.69014170525794e-07, |
|
"loss": 0.3769, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.8650416171224733, |
|
"grad_norm": 2.1512178822513772, |
|
"learning_rate": 5.452030694946448e-07, |
|
"loss": 0.3766, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.8680142687277052, |
|
"grad_norm": 2.5248216489482314, |
|
"learning_rate": 5.21872128386961e-07, |
|
"loss": 0.3685, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.870986920332937, |
|
"grad_norm": 2.1936261850407455, |
|
"learning_rate": 4.990238619263232e-07, |
|
"loss": 0.3755, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.8739595719381689, |
|
"grad_norm": 2.0674457126577956, |
|
"learning_rate": 4.7666073281126713e-07, |
|
"loss": 0.37, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.8769322235434007, |
|
"grad_norm": 2.4393074814620705, |
|
"learning_rate": 4.547851514498386e-07, |
|
"loss": 0.3743, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.8799048751486326, |
|
"grad_norm": 2.2218679471161984, |
|
"learning_rate": 4.333994756997878e-07, |
|
"loss": 0.3737, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.8828775267538644, |
|
"grad_norm": 2.326578195058314, |
|
"learning_rate": 4.125060106144313e-07, |
|
"loss": 0.3706, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.8858501783590963, |
|
"grad_norm": 2.159799023836089, |
|
"learning_rate": 3.921070081941969e-07, |
|
"loss": 0.3757, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.8888228299643282, |
|
"grad_norm": 2.0971056501716543, |
|
"learning_rate": 3.72204667143895e-07, |
|
"loss": 0.3662, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.89179548156956, |
|
"grad_norm": 2.2898930751852817, |
|
"learning_rate": 3.528011326357306e-07, |
|
"loss": 0.3803, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.8947681331747919, |
|
"grad_norm": 2.276170558283118, |
|
"learning_rate": 3.338984960780856e-07, |
|
"loss": 0.3696, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.8977407847800237, |
|
"grad_norm": 2.151976309053238, |
|
"learning_rate": 3.154987948901006e-07, |
|
"loss": 0.3856, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.9007134363852557, |
|
"grad_norm": 2.168451215561875, |
|
"learning_rate": 2.9760401228206394e-07, |
|
"loss": 0.3773, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.9036860879904876, |
|
"grad_norm": 2.2792459501817155, |
|
"learning_rate": 2.802160770416584e-07, |
|
"loss": 0.3707, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.9066587395957194, |
|
"grad_norm": 2.4152961249318436, |
|
"learning_rate": 2.633368633260658e-07, |
|
"loss": 0.3664, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.9096313912009513, |
|
"grad_norm": 2.2767769677463066, |
|
"learning_rate": 2.469681904599569e-07, |
|
"loss": 0.3786, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.9126040428061831, |
|
"grad_norm": 2.3341884568313795, |
|
"learning_rate": 2.3111182273940103e-07, |
|
"loss": 0.366, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.915576694411415, |
|
"grad_norm": 2.3090738993253064, |
|
"learning_rate": 2.1576946924169785e-07, |
|
"loss": 0.3682, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.9185493460166468, |
|
"grad_norm": 2.2708470444789968, |
|
"learning_rate": 2.009427836411665e-07, |
|
"loss": 0.367, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.9215219976218787, |
|
"grad_norm": 2.0757645629730543, |
|
"learning_rate": 1.8663336403090482e-07, |
|
"loss": 0.3664, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.9244946492271106, |
|
"grad_norm": 2.424017703704812, |
|
"learning_rate": 1.7284275275053663e-07, |
|
"loss": 0.3792, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.9274673008323424, |
|
"grad_norm": 2.127097574037306, |
|
"learning_rate": 1.59572436219973e-07, |
|
"loss": 0.3542, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.9304399524375743, |
|
"grad_norm": 2.3260445722240015, |
|
"learning_rate": 1.4682384477919808e-07, |
|
"loss": 0.3675, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.9334126040428062, |
|
"grad_norm": 2.39014963013141, |
|
"learning_rate": 1.345983525340988e-07, |
|
"loss": 0.3667, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.9363852556480381, |
|
"grad_norm": 2.2694068787756247, |
|
"learning_rate": 1.228972772083592e-07, |
|
"loss": 0.37, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.93935790725327, |
|
"grad_norm": 2.3620320947289604, |
|
"learning_rate": 1.1172188000142803e-07, |
|
"loss": 0.3727, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.9423305588585018, |
|
"grad_norm": 2.187390021455131, |
|
"learning_rate": 1.0107336545258084e-07, |
|
"loss": 0.3587, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.9453032104637337, |
|
"grad_norm": 2.2100838643930034, |
|
"learning_rate": 9.095288131108937e-08, |
|
"loss": 0.3673, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.9482758620689655, |
|
"grad_norm": 2.293416030314438, |
|
"learning_rate": 8.136151841250938e-08, |
|
"loss": 0.3783, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.9512485136741974, |
|
"grad_norm": 2.178738314835226, |
|
"learning_rate": 7.230031056110864e-08, |
|
"loss": 0.3731, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.9542211652794292, |
|
"grad_norm": 2.2670575062833165, |
|
"learning_rate": 6.377023441843599e-08, |
|
"loss": 0.373, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.9571938168846611, |
|
"grad_norm": 2.1567401585206514, |
|
"learning_rate": 5.577220939805117e-08, |
|
"loss": 0.3699, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.960166468489893, |
|
"grad_norm": 2.0789730650126, |
|
"learning_rate": 4.830709756642904e-08, |
|
"loss": 0.366, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.9631391200951248, |
|
"grad_norm": 2.157814350971754, |
|
"learning_rate": 4.1375703550038394e-08, |
|
"loss": 0.3633, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.9661117717003567, |
|
"grad_norm": 2.1100943054104344, |
|
"learning_rate": 3.4978774448617414e-08, |
|
"loss": 0.3519, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.9690844233055886, |
|
"grad_norm": 2.2083883883509743, |
|
"learning_rate": 2.9116999754646437e-08, |
|
"loss": 0.3613, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.9720570749108205, |
|
"grad_norm": 2.2428508865311527, |
|
"learning_rate": 2.3791011279033492e-08, |
|
"loss": 0.3644, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.9750297265160524, |
|
"grad_norm": 2.1651670922102872, |
|
"learning_rate": 1.9001383083011028e-08, |
|
"loss": 0.3662, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.9780023781212842, |
|
"grad_norm": 2.3186672122830716, |
|
"learning_rate": 1.4748631416262593e-08, |
|
"loss": 0.3606, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.9809750297265161, |
|
"grad_norm": 2.2616261488795617, |
|
"learning_rate": 1.103321466128071e-08, |
|
"loss": 0.3603, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.9839476813317479, |
|
"grad_norm": 2.277277885300596, |
|
"learning_rate": 7.855533283955829e-09, |
|
"loss": 0.3657, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.9869203329369798, |
|
"grad_norm": 2.1694357282951566, |
|
"learning_rate": 5.215929790416407e-09, |
|
"loss": 0.363, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.9898929845422116, |
|
"grad_norm": 2.2318921404664795, |
|
"learning_rate": 3.1146886901090024e-09, |
|
"loss": 0.3603, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.9928656361474435, |
|
"grad_norm": 2.209506235328767, |
|
"learning_rate": 1.5520364651344655e-09, |
|
"loss": 0.3624, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.9958382877526754, |
|
"grad_norm": 2.184944707544881, |
|
"learning_rate": 5.281415458346928e-10, |
|
"loss": 0.3688, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.9988109393579072, |
|
"grad_norm": 2.267638514087637, |
|
"learning_rate": 4.3114292641033286e-11, |
|
"loss": 0.3663, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_runtime": 3.3713, |
|
"eval_samples_per_second": 2.966, |
|
"eval_steps_per_second": 0.89, |
|
"step": 1682 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1682, |
|
"total_flos": 176088290426880.0, |
|
"train_loss": 0.0, |
|
"train_runtime": 0.008, |
|
"train_samples_per_second": 3378420.528, |
|
"train_steps_per_second": 211190.52 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 1682, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 176088290426880.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|