{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9996544063194273, "eval_steps": 500, "global_step": 2531, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003949642063687978, "grad_norm": 0.7681758999824524, "learning_rate": 3e-05, "loss": 3.3209, "step": 10 }, { "epoch": 0.007899284127375956, "grad_norm": 0.7526655793190002, "learning_rate": 2.9999953706169412e-05, "loss": 0.1659, "step": 20 }, { "epoch": 0.011848926191063935, "grad_norm": 0.573715090751648, "learning_rate": 2.999981482496339e-05, "loss": 0.1512, "step": 30 }, { "epoch": 0.015798568254751913, "grad_norm": 0.602210521697998, "learning_rate": 2.9999583357239188e-05, "loss": 0.1438, "step": 40 }, { "epoch": 0.01974821031843989, "grad_norm": 0.6056187748908997, "learning_rate": 2.9999259304425536e-05, "loss": 0.1456, "step": 50 }, { "epoch": 0.02369785238212787, "grad_norm": 0.6448855400085449, "learning_rate": 2.9998842668522657e-05, "loss": 0.1573, "step": 60 }, { "epoch": 0.027647494445815848, "grad_norm": 0.5019297003746033, "learning_rate": 2.9998333452102237e-05, "loss": 0.1432, "step": 70 }, { "epoch": 0.031597136509503826, "grad_norm": 0.5146121382713318, "learning_rate": 2.9997731658307427e-05, "loss": 0.1414, "step": 80 }, { "epoch": 0.03554677857319181, "grad_norm": 0.5668686628341675, "learning_rate": 2.99970372908528e-05, "loss": 0.1406, "step": 90 }, { "epoch": 0.03949642063687978, "grad_norm": 0.5978260040283203, "learning_rate": 2.9996250354024345e-05, "loss": 0.1389, "step": 100 }, { "epoch": 0.043446062700567764, "grad_norm": 0.7445759773254395, "learning_rate": 2.9995370852679447e-05, "loss": 0.1468, "step": 110 }, { "epoch": 0.04739570476425574, "grad_norm": 0.6418613791465759, "learning_rate": 2.9994398792246826e-05, "loss": 0.1486, "step": 120 }, { "epoch": 0.05134534682794372, "grad_norm": 0.4584648907184601, "learning_rate": 2.9993334178726546e-05, "loss": 0.1404, "step": 130 }, { "epoch": 0.055294988891631695, "grad_norm": 0.5104970932006836, "learning_rate": 2.9992177018689935e-05, "loss": 0.1398, "step": 140 }, { "epoch": 0.05924463095531968, "grad_norm": 0.6897855401039124, "learning_rate": 2.9990927319279584e-05, "loss": 0.1546, "step": 150 }, { "epoch": 0.06319427301900765, "grad_norm": 0.5376682281494141, "learning_rate": 2.998958508820927e-05, "loss": 0.15, "step": 160 }, { "epoch": 0.06714391508269563, "grad_norm": 0.5601758360862732, "learning_rate": 2.9988150333763933e-05, "loss": 0.1471, "step": 170 }, { "epoch": 0.07109355714638362, "grad_norm": 0.4657880961894989, "learning_rate": 2.998662306479961e-05, "loss": 0.1394, "step": 180 }, { "epoch": 0.07504319921007159, "grad_norm": 0.5285632014274597, "learning_rate": 2.9985003290743385e-05, "loss": 0.1452, "step": 190 }, { "epoch": 0.07899284127375956, "grad_norm": 0.4982571601867676, "learning_rate": 2.9983291021593326e-05, "loss": 0.1402, "step": 200 }, { "epoch": 0.08294248333744754, "grad_norm": 0.48557624220848083, "learning_rate": 2.998148626791844e-05, "loss": 0.142, "step": 210 }, { "epoch": 0.08689212540113553, "grad_norm": 0.5023711919784546, "learning_rate": 2.9979589040858586e-05, "loss": 0.1624, "step": 220 }, { "epoch": 0.0908417674648235, "grad_norm": 0.47005024552345276, "learning_rate": 2.9977599352124413e-05, "loss": 0.155, "step": 230 }, { "epoch": 0.09479140952851148, "grad_norm": 0.5279797315597534, "learning_rate": 2.9975517213997302e-05, "loss": 0.1532, "step": 240 }, { "epoch": 0.09874105159219945, "grad_norm": 0.43386149406433105, "learning_rate": 2.9973342639329272e-05, "loss": 0.1481, "step": 250 }, { "epoch": 0.10269069365588744, "grad_norm": 0.5564565062522888, "learning_rate": 2.997107564154291e-05, "loss": 0.1358, "step": 260 }, { "epoch": 0.10664033571957542, "grad_norm": 0.6061131358146667, "learning_rate": 2.996871623463128e-05, "loss": 0.1464, "step": 270 }, { "epoch": 0.11058997778326339, "grad_norm": 0.5967995524406433, "learning_rate": 2.996626443315785e-05, "loss": 0.1451, "step": 280 }, { "epoch": 0.11453961984695137, "grad_norm": 0.5291288495063782, "learning_rate": 2.9963720252256387e-05, "loss": 0.1436, "step": 290 }, { "epoch": 0.11848926191063935, "grad_norm": 0.5956757068634033, "learning_rate": 2.9961083707630877e-05, "loss": 0.1492, "step": 300 }, { "epoch": 0.12243890397432733, "grad_norm": 0.5079193711280823, "learning_rate": 2.9958354815555426e-05, "loss": 0.1388, "step": 310 }, { "epoch": 0.1263885460380153, "grad_norm": 0.44773226976394653, "learning_rate": 2.995553359287414e-05, "loss": 0.1311, "step": 320 }, { "epoch": 0.13033818810170328, "grad_norm": 0.5200162529945374, "learning_rate": 2.9952620057001055e-05, "loss": 0.1401, "step": 330 }, { "epoch": 0.13428783016539125, "grad_norm": 0.5840851068496704, "learning_rate": 2.994961422591999e-05, "loss": 0.1484, "step": 340 }, { "epoch": 0.13823747222907923, "grad_norm": 0.5336028933525085, "learning_rate": 2.9946516118184484e-05, "loss": 0.1298, "step": 350 }, { "epoch": 0.14218711429276723, "grad_norm": 0.7243465781211853, "learning_rate": 2.9943325752917633e-05, "loss": 0.1463, "step": 360 }, { "epoch": 0.1461367563564552, "grad_norm": 0.5500628352165222, "learning_rate": 2.9940043149812006e-05, "loss": 0.1465, "step": 370 }, { "epoch": 0.15008639842014318, "grad_norm": 0.5050541162490845, "learning_rate": 2.993666832912949e-05, "loss": 0.1434, "step": 380 }, { "epoch": 0.15403604048383115, "grad_norm": 0.5059782266616821, "learning_rate": 2.9933201311701222e-05, "loss": 0.1385, "step": 390 }, { "epoch": 0.15798568254751913, "grad_norm": 0.5439670085906982, "learning_rate": 2.9929642118927397e-05, "loss": 0.1421, "step": 400 }, { "epoch": 0.1619353246112071, "grad_norm": 0.5325440168380737, "learning_rate": 2.992599077277717e-05, "loss": 0.1482, "step": 410 }, { "epoch": 0.16588496667489508, "grad_norm": 0.5057934522628784, "learning_rate": 2.992224729578851e-05, "loss": 0.1415, "step": 420 }, { "epoch": 0.16983460873858305, "grad_norm": 0.5029751062393188, "learning_rate": 2.9918411711068074e-05, "loss": 0.1517, "step": 430 }, { "epoch": 0.17378425080227106, "grad_norm": 0.38729503750801086, "learning_rate": 2.9914484042291053e-05, "loss": 0.1367, "step": 440 }, { "epoch": 0.17773389286595903, "grad_norm": 0.4877079725265503, "learning_rate": 2.991046431370102e-05, "loss": 0.1446, "step": 450 }, { "epoch": 0.181683534929647, "grad_norm": 0.48335975408554077, "learning_rate": 2.9906352550109787e-05, "loss": 0.1372, "step": 460 }, { "epoch": 0.18563317699333498, "grad_norm": 0.48935794830322266, "learning_rate": 2.990214877689727e-05, "loss": 0.144, "step": 470 }, { "epoch": 0.18958281905702296, "grad_norm": 0.44793424010276794, "learning_rate": 2.9897853020011298e-05, "loss": 0.1298, "step": 480 }, { "epoch": 0.19353246112071093, "grad_norm": 0.6513413190841675, "learning_rate": 2.9893465305967483e-05, "loss": 0.1361, "step": 490 }, { "epoch": 0.1974821031843989, "grad_norm": 0.5480667948722839, "learning_rate": 2.9888985661849028e-05, "loss": 0.1497, "step": 500 }, { "epoch": 0.20143174524808688, "grad_norm": 0.4912254810333252, "learning_rate": 2.988441411530659e-05, "loss": 0.1461, "step": 510 }, { "epoch": 0.20538138731177488, "grad_norm": 0.4925342798233032, "learning_rate": 2.987975069455809e-05, "loss": 0.1418, "step": 520 }, { "epoch": 0.20933102937546286, "grad_norm": 0.46876174211502075, "learning_rate": 2.987499542838854e-05, "loss": 0.1409, "step": 530 }, { "epoch": 0.21328067143915083, "grad_norm": 0.5577364563941956, "learning_rate": 2.9870148346149865e-05, "loss": 0.1423, "step": 540 }, { "epoch": 0.2172303135028388, "grad_norm": 0.48950615525245667, "learning_rate": 2.9865209477760746e-05, "loss": 0.1367, "step": 550 }, { "epoch": 0.22117995556652678, "grad_norm": 0.4494156241416931, "learning_rate": 2.9860178853706397e-05, "loss": 0.1384, "step": 560 }, { "epoch": 0.22512959763021476, "grad_norm": 0.4439913034439087, "learning_rate": 2.9855056505038395e-05, "loss": 0.1447, "step": 570 }, { "epoch": 0.22907923969390273, "grad_norm": 0.5027551054954529, "learning_rate": 2.984984246337449e-05, "loss": 0.1526, "step": 580 }, { "epoch": 0.2330288817575907, "grad_norm": 0.4503665566444397, "learning_rate": 2.984453676089842e-05, "loss": 0.1333, "step": 590 }, { "epoch": 0.2369785238212787, "grad_norm": 0.5127356052398682, "learning_rate": 2.9839139430359684e-05, "loss": 0.1372, "step": 600 }, { "epoch": 0.24092816588496668, "grad_norm": 0.672027051448822, "learning_rate": 2.983365050507336e-05, "loss": 0.1359, "step": 610 }, { "epoch": 0.24487780794865466, "grad_norm": 0.5182546377182007, "learning_rate": 2.9828070018919902e-05, "loss": 0.1504, "step": 620 }, { "epoch": 0.24882745001234263, "grad_norm": 0.5020663738250732, "learning_rate": 2.9822398006344923e-05, "loss": 0.1416, "step": 630 }, { "epoch": 0.2527770920760306, "grad_norm": 0.41956228017807007, "learning_rate": 2.9816634502358976e-05, "loss": 0.1306, "step": 640 }, { "epoch": 0.2567267341397186, "grad_norm": 0.42414599657058716, "learning_rate": 2.9810779542537357e-05, "loss": 0.1412, "step": 650 }, { "epoch": 0.26067637620340656, "grad_norm": 0.5931263566017151, "learning_rate": 2.9804833163019866e-05, "loss": 0.1354, "step": 660 }, { "epoch": 0.26462601826709453, "grad_norm": 0.49056369066238403, "learning_rate": 2.9798795400510588e-05, "loss": 0.1313, "step": 670 }, { "epoch": 0.2685756603307825, "grad_norm": 0.5098786950111389, "learning_rate": 2.9792666292277687e-05, "loss": 0.1264, "step": 680 }, { "epoch": 0.2725253023944705, "grad_norm": 0.5509768128395081, "learning_rate": 2.9786445876153147e-05, "loss": 0.1418, "step": 690 }, { "epoch": 0.27647494445815846, "grad_norm": 0.6949456930160522, "learning_rate": 2.978013419053255e-05, "loss": 0.1399, "step": 700 }, { "epoch": 0.28042458652184643, "grad_norm": 0.5233505368232727, "learning_rate": 2.9773731274374847e-05, "loss": 0.1415, "step": 710 }, { "epoch": 0.28437422858553446, "grad_norm": 0.539608895778656, "learning_rate": 2.9767237167202105e-05, "loss": 0.1458, "step": 720 }, { "epoch": 0.28832387064922244, "grad_norm": 0.4299115240573883, "learning_rate": 2.976065190909927e-05, "loss": 0.1351, "step": 730 }, { "epoch": 0.2922735127129104, "grad_norm": 0.46829068660736084, "learning_rate": 2.975397554071392e-05, "loss": 0.1349, "step": 740 }, { "epoch": 0.2962231547765984, "grad_norm": 0.490376353263855, "learning_rate": 2.9747208103256007e-05, "loss": 0.1439, "step": 750 }, { "epoch": 0.30017279684028636, "grad_norm": 0.5873934030532837, "learning_rate": 2.9740349638497614e-05, "loss": 0.1395, "step": 760 }, { "epoch": 0.30412243890397433, "grad_norm": 0.5811406373977661, "learning_rate": 2.973340018877269e-05, "loss": 0.1342, "step": 770 }, { "epoch": 0.3080720809676623, "grad_norm": 0.5323910713195801, "learning_rate": 2.972635979697678e-05, "loss": 0.1471, "step": 780 }, { "epoch": 0.3120217230313503, "grad_norm": 0.5084981918334961, "learning_rate": 2.9719228506566792e-05, "loss": 0.1296, "step": 790 }, { "epoch": 0.31597136509503826, "grad_norm": 0.5692681670188904, "learning_rate": 2.9712006361560685e-05, "loss": 0.1341, "step": 800 }, { "epoch": 0.31992100715872623, "grad_norm": 0.525729775428772, "learning_rate": 2.9704693406537222e-05, "loss": 0.1454, "step": 810 }, { "epoch": 0.3238706492224142, "grad_norm": 0.4046003520488739, "learning_rate": 2.9697289686635703e-05, "loss": 0.1342, "step": 820 }, { "epoch": 0.3278202912861022, "grad_norm": 0.47330015897750854, "learning_rate": 2.968979524755567e-05, "loss": 0.1417, "step": 830 }, { "epoch": 0.33176993334979016, "grad_norm": 0.4547816812992096, "learning_rate": 2.968221013555662e-05, "loss": 0.1298, "step": 840 }, { "epoch": 0.33571957541347813, "grad_norm": 0.4318365156650543, "learning_rate": 2.9674534397457747e-05, "loss": 0.1454, "step": 850 }, { "epoch": 0.3396692174771661, "grad_norm": 0.4720039665699005, "learning_rate": 2.9666768080637622e-05, "loss": 0.1321, "step": 860 }, { "epoch": 0.3436188595408541, "grad_norm": 0.5289425849914551, "learning_rate": 2.965891123303392e-05, "loss": 0.1301, "step": 870 }, { "epoch": 0.3475685016045421, "grad_norm": 0.49078208208084106, "learning_rate": 2.9650963903143124e-05, "loss": 0.1452, "step": 880 }, { "epoch": 0.3515181436682301, "grad_norm": 0.7317320704460144, "learning_rate": 2.9642926140020203e-05, "loss": 0.1516, "step": 890 }, { "epoch": 0.35546778573191806, "grad_norm": 0.5247913599014282, "learning_rate": 2.9634797993278337e-05, "loss": 0.1408, "step": 900 }, { "epoch": 0.35941742779560604, "grad_norm": 0.5061410665512085, "learning_rate": 2.9626579513088606e-05, "loss": 0.1396, "step": 910 }, { "epoch": 0.363367069859294, "grad_norm": 0.5871759057044983, "learning_rate": 2.9618270750179665e-05, "loss": 0.152, "step": 920 }, { "epoch": 0.367316711922982, "grad_norm": 0.4584594666957855, "learning_rate": 2.9609871755837436e-05, "loss": 0.1274, "step": 930 }, { "epoch": 0.37126635398666996, "grad_norm": 0.405700147151947, "learning_rate": 2.9601382581904816e-05, "loss": 0.1284, "step": 940 }, { "epoch": 0.37521599605035794, "grad_norm": 0.4593953788280487, "learning_rate": 2.9592803280781306e-05, "loss": 0.1359, "step": 950 }, { "epoch": 0.3791656381140459, "grad_norm": 0.5641497373580933, "learning_rate": 2.9584133905422744e-05, "loss": 0.1454, "step": 960 }, { "epoch": 0.3831152801777339, "grad_norm": 0.3449844419956207, "learning_rate": 2.9575374509340935e-05, "loss": 0.1385, "step": 970 }, { "epoch": 0.38706492224142186, "grad_norm": 0.46647313237190247, "learning_rate": 2.956652514660336e-05, "loss": 0.1328, "step": 980 }, { "epoch": 0.39101456430510984, "grad_norm": 0.4442940950393677, "learning_rate": 2.9557585871832787e-05, "loss": 0.1379, "step": 990 }, { "epoch": 0.3949642063687978, "grad_norm": 0.5622376799583435, "learning_rate": 2.9548556740207e-05, "loss": 0.1525, "step": 1000 }, { "epoch": 0.3989138484324858, "grad_norm": 0.5095304250717163, "learning_rate": 2.9539437807458404e-05, "loss": 0.1317, "step": 1010 }, { "epoch": 0.40286349049617376, "grad_norm": 0.36938050389289856, "learning_rate": 2.9530229129873715e-05, "loss": 0.1361, "step": 1020 }, { "epoch": 0.4068131325598618, "grad_norm": 0.4959389865398407, "learning_rate": 2.9520930764293586e-05, "loss": 0.1475, "step": 1030 }, { "epoch": 0.41076277462354976, "grad_norm": 0.5631204843521118, "learning_rate": 2.9511542768112284e-05, "loss": 0.1391, "step": 1040 }, { "epoch": 0.41471241668723774, "grad_norm": 0.4227543771266937, "learning_rate": 2.9502065199277312e-05, "loss": 0.1402, "step": 1050 }, { "epoch": 0.4186620587509257, "grad_norm": 0.43038052320480347, "learning_rate": 2.9492498116289072e-05, "loss": 0.1239, "step": 1060 }, { "epoch": 0.4226117008146137, "grad_norm": 0.5115047097206116, "learning_rate": 2.9482841578200486e-05, "loss": 0.1417, "step": 1070 }, { "epoch": 0.42656134287830166, "grad_norm": 0.4372217059135437, "learning_rate": 2.9473095644616634e-05, "loss": 0.139, "step": 1080 }, { "epoch": 0.43051098494198964, "grad_norm": 0.792674720287323, "learning_rate": 2.94632603756944e-05, "loss": 0.1355, "step": 1090 }, { "epoch": 0.4344606270056776, "grad_norm": 0.46272650361061096, "learning_rate": 2.945333583214208e-05, "loss": 0.1513, "step": 1100 }, { "epoch": 0.4384102690693656, "grad_norm": 0.3727450668811798, "learning_rate": 2.9443322075219036e-05, "loss": 0.1317, "step": 1110 }, { "epoch": 0.44235991113305356, "grad_norm": 0.39475393295288086, "learning_rate": 2.9433219166735285e-05, "loss": 0.126, "step": 1120 }, { "epoch": 0.44630955319674154, "grad_norm": 0.5749325156211853, "learning_rate": 2.9423027169051134e-05, "loss": 0.1509, "step": 1130 }, { "epoch": 0.4502591952604295, "grad_norm": 0.44618451595306396, "learning_rate": 2.9412746145076804e-05, "loss": 0.1257, "step": 1140 }, { "epoch": 0.4542088373241175, "grad_norm": 0.46040260791778564, "learning_rate": 2.9402376158272026e-05, "loss": 0.1306, "step": 1150 }, { "epoch": 0.45815847938780546, "grad_norm": 0.6470154523849487, "learning_rate": 2.9391917272645654e-05, "loss": 0.147, "step": 1160 }, { "epoch": 0.46210812145149344, "grad_norm": 0.4042102098464966, "learning_rate": 2.9381369552755268e-05, "loss": 0.1358, "step": 1170 }, { "epoch": 0.4660577635151814, "grad_norm": 0.5040680766105652, "learning_rate": 2.937073306370679e-05, "loss": 0.1364, "step": 1180 }, { "epoch": 0.47000740557886944, "grad_norm": 0.44574257731437683, "learning_rate": 2.936000787115406e-05, "loss": 0.1468, "step": 1190 }, { "epoch": 0.4739570476425574, "grad_norm": 0.4155598282814026, "learning_rate": 2.9349194041298437e-05, "loss": 0.138, "step": 1200 }, { "epoch": 0.4779066897062454, "grad_norm": 0.43807128071784973, "learning_rate": 2.9338291640888413e-05, "loss": 0.1376, "step": 1210 }, { "epoch": 0.48185633176993337, "grad_norm": 0.6164836883544922, "learning_rate": 2.9327300737219164e-05, "loss": 0.1415, "step": 1220 }, { "epoch": 0.48580597383362134, "grad_norm": 0.4064141511917114, "learning_rate": 2.9316221398132163e-05, "loss": 0.1457, "step": 1230 }, { "epoch": 0.4897556158973093, "grad_norm": 0.47821277379989624, "learning_rate": 2.930505369201475e-05, "loss": 0.144, "step": 1240 }, { "epoch": 0.4937052579609973, "grad_norm": 0.4229309558868408, "learning_rate": 2.9293797687799717e-05, "loss": 0.1286, "step": 1250 }, { "epoch": 0.49765490002468526, "grad_norm": 0.42858126759529114, "learning_rate": 2.9282453454964856e-05, "loss": 0.1388, "step": 1260 }, { "epoch": 0.5016045420883732, "grad_norm": 0.47248193621635437, "learning_rate": 2.9271021063532586e-05, "loss": 0.1279, "step": 1270 }, { "epoch": 0.5055541841520612, "grad_norm": 0.5147600769996643, "learning_rate": 2.9259500584069444e-05, "loss": 0.1281, "step": 1280 }, { "epoch": 0.5095038262157492, "grad_norm": 0.4137686789035797, "learning_rate": 2.924789208768573e-05, "loss": 0.1441, "step": 1290 }, { "epoch": 0.5134534682794372, "grad_norm": 0.484967440366745, "learning_rate": 2.923619564603501e-05, "loss": 0.1328, "step": 1300 }, { "epoch": 0.5174031103431251, "grad_norm": 0.5038776397705078, "learning_rate": 2.922441133131369e-05, "loss": 0.1442, "step": 1310 }, { "epoch": 0.5213527524068131, "grad_norm": 0.4918186366558075, "learning_rate": 2.921253921626058e-05, "loss": 0.1285, "step": 1320 }, { "epoch": 0.5253023944705011, "grad_norm": 0.447346568107605, "learning_rate": 2.9200579374156447e-05, "loss": 0.1261, "step": 1330 }, { "epoch": 0.5292520365341891, "grad_norm": 0.4736550748348236, "learning_rate": 2.9188531878823532e-05, "loss": 0.133, "step": 1340 }, { "epoch": 0.533201678597877, "grad_norm": 0.586494505405426, "learning_rate": 2.9176396804625135e-05, "loss": 0.1409, "step": 1350 }, { "epoch": 0.537151320661565, "grad_norm": 0.49870672821998596, "learning_rate": 2.9164174226465134e-05, "loss": 0.1444, "step": 1360 }, { "epoch": 0.541100962725253, "grad_norm": 0.404547780752182, "learning_rate": 2.9151864219787522e-05, "loss": 0.1303, "step": 1370 }, { "epoch": 0.545050604788941, "grad_norm": 0.42132681608200073, "learning_rate": 2.913946686057595e-05, "loss": 0.1276, "step": 1380 }, { "epoch": 0.5490002468526289, "grad_norm": 0.4928096830844879, "learning_rate": 2.9126982225353243e-05, "loss": 0.1348, "step": 1390 }, { "epoch": 0.5529498889163169, "grad_norm": 0.44450655579566956, "learning_rate": 2.911441039118095e-05, "loss": 0.1417, "step": 1400 }, { "epoch": 0.5568995309800049, "grad_norm": 0.5710647702217102, "learning_rate": 2.910175143565886e-05, "loss": 0.1284, "step": 1410 }, { "epoch": 0.5608491730436929, "grad_norm": 0.3675592243671417, "learning_rate": 2.9089005436924506e-05, "loss": 0.1505, "step": 1420 }, { "epoch": 0.564798815107381, "grad_norm": 0.4794444441795349, "learning_rate": 2.90761724736527e-05, "loss": 0.1325, "step": 1430 }, { "epoch": 0.5687484571710689, "grad_norm": 0.5743889808654785, "learning_rate": 2.906325262505505e-05, "loss": 0.1358, "step": 1440 }, { "epoch": 0.5726980992347569, "grad_norm": 0.4955087900161743, "learning_rate": 2.9050245970879455e-05, "loss": 0.1387, "step": 1450 }, { "epoch": 0.5766477412984449, "grad_norm": 0.42035970091819763, "learning_rate": 2.9037152591409635e-05, "loss": 0.1369, "step": 1460 }, { "epoch": 0.5805973833621328, "grad_norm": 0.4199492335319519, "learning_rate": 2.9023972567464606e-05, "loss": 0.1461, "step": 1470 }, { "epoch": 0.5845470254258208, "grad_norm": 0.43724125623703003, "learning_rate": 2.9010705980398217e-05, "loss": 0.1219, "step": 1480 }, { "epoch": 0.5884966674895088, "grad_norm": 0.39386245608329773, "learning_rate": 2.8997352912098616e-05, "loss": 0.1255, "step": 1490 }, { "epoch": 0.5924463095531968, "grad_norm": 0.3640863597393036, "learning_rate": 2.8983913444987754e-05, "loss": 0.1273, "step": 1500 }, { "epoch": 0.5963959516168847, "grad_norm": 0.40772178769111633, "learning_rate": 2.8970387662020898e-05, "loss": 0.1326, "step": 1510 }, { "epoch": 0.6003455936805727, "grad_norm": 0.4535306990146637, "learning_rate": 2.895677564668608e-05, "loss": 0.1273, "step": 1520 }, { "epoch": 0.6042952357442607, "grad_norm": 0.5429089665412903, "learning_rate": 2.894307748300361e-05, "loss": 0.1245, "step": 1530 }, { "epoch": 0.6082448778079487, "grad_norm": 0.38951486349105835, "learning_rate": 2.8929293255525563e-05, "loss": 0.1437, "step": 1540 }, { "epoch": 0.6121945198716366, "grad_norm": 0.4131280183792114, "learning_rate": 2.8915423049335214e-05, "loss": 0.1249, "step": 1550 }, { "epoch": 0.6161441619353246, "grad_norm": 0.44423356652259827, "learning_rate": 2.890146695004657e-05, "loss": 0.1315, "step": 1560 }, { "epoch": 0.6200938039990126, "grad_norm": 0.4929848313331604, "learning_rate": 2.88874250438038e-05, "loss": 0.1399, "step": 1570 }, { "epoch": 0.6240434460627006, "grad_norm": 0.44524630904197693, "learning_rate": 2.8873297417280724e-05, "loss": 0.1304, "step": 1580 }, { "epoch": 0.6279930881263885, "grad_norm": 0.4765247404575348, "learning_rate": 2.885908415768027e-05, "loss": 0.1422, "step": 1590 }, { "epoch": 0.6319427301900765, "grad_norm": 0.44227954745292664, "learning_rate": 2.884478535273393e-05, "loss": 0.1573, "step": 1600 }, { "epoch": 0.6358923722537645, "grad_norm": 0.4304993152618408, "learning_rate": 2.8830401090701234e-05, "loss": 0.1365, "step": 1610 }, { "epoch": 0.6398420143174525, "grad_norm": 0.42231836915016174, "learning_rate": 2.8815931460369198e-05, "loss": 0.1328, "step": 1620 }, { "epoch": 0.6437916563811404, "grad_norm": 0.44187867641448975, "learning_rate": 2.880137655105176e-05, "loss": 0.1228, "step": 1630 }, { "epoch": 0.6477412984448284, "grad_norm": 0.433136910200119, "learning_rate": 2.8786736452589265e-05, "loss": 0.133, "step": 1640 }, { "epoch": 0.6516909405085164, "grad_norm": 0.4308445453643799, "learning_rate": 2.8772011255347875e-05, "loss": 0.127, "step": 1650 }, { "epoch": 0.6556405825722044, "grad_norm": 0.4352281391620636, "learning_rate": 2.8757201050219027e-05, "loss": 0.1276, "step": 1660 }, { "epoch": 0.6595902246358923, "grad_norm": 0.450520396232605, "learning_rate": 2.874230592861887e-05, "loss": 0.1233, "step": 1670 }, { "epoch": 0.6635398666995803, "grad_norm": 0.4648306369781494, "learning_rate": 2.8727325982487705e-05, "loss": 0.1243, "step": 1680 }, { "epoch": 0.6674895087632683, "grad_norm": 0.5166367888450623, "learning_rate": 2.871226130428941e-05, "loss": 0.1308, "step": 1690 }, { "epoch": 0.6714391508269563, "grad_norm": 0.6115042567253113, "learning_rate": 2.8697111987010868e-05, "loss": 0.1339, "step": 1700 }, { "epoch": 0.6753887928906442, "grad_norm": 0.3470801115036011, "learning_rate": 2.868187812416141e-05, "loss": 0.1305, "step": 1710 }, { "epoch": 0.6793384349543322, "grad_norm": 0.40242600440979004, "learning_rate": 2.8666559809772217e-05, "loss": 0.1325, "step": 1720 }, { "epoch": 0.6832880770180202, "grad_norm": 0.4116344749927521, "learning_rate": 2.8651157138395744e-05, "loss": 0.1385, "step": 1730 }, { "epoch": 0.6872377190817082, "grad_norm": 0.39455336332321167, "learning_rate": 2.863567020510515e-05, "loss": 0.1291, "step": 1740 }, { "epoch": 0.6911873611453963, "grad_norm": 0.49655675888061523, "learning_rate": 2.86200991054937e-05, "loss": 0.1363, "step": 1750 }, { "epoch": 0.6951370032090842, "grad_norm": 0.4002642035484314, "learning_rate": 2.8604443935674164e-05, "loss": 0.1421, "step": 1760 }, { "epoch": 0.6990866452727722, "grad_norm": 0.43481770157814026, "learning_rate": 2.8588704792278248e-05, "loss": 0.1254, "step": 1770 }, { "epoch": 0.7030362873364602, "grad_norm": 0.49691149592399597, "learning_rate": 2.8572881772455993e-05, "loss": 0.1393, "step": 1780 }, { "epoch": 0.7069859294001481, "grad_norm": 0.47778138518333435, "learning_rate": 2.8556974973875152e-05, "loss": 0.1387, "step": 1790 }, { "epoch": 0.7109355714638361, "grad_norm": 0.3887634873390198, "learning_rate": 2.854098449472061e-05, "loss": 0.1301, "step": 1800 }, { "epoch": 0.7148852135275241, "grad_norm": 0.3825758695602417, "learning_rate": 2.852491043369377e-05, "loss": 0.1292, "step": 1810 }, { "epoch": 0.7188348555912121, "grad_norm": 0.44277575612068176, "learning_rate": 2.8508752890011957e-05, "loss": 0.1263, "step": 1820 }, { "epoch": 0.7227844976549, "grad_norm": 0.54979008436203, "learning_rate": 2.849251196340777e-05, "loss": 0.1487, "step": 1830 }, { "epoch": 0.726734139718588, "grad_norm": 0.5191593170166016, "learning_rate": 2.847618775412851e-05, "loss": 0.1355, "step": 1840 }, { "epoch": 0.730683781782276, "grad_norm": 0.42348307371139526, "learning_rate": 2.8459780362935532e-05, "loss": 0.1356, "step": 1850 }, { "epoch": 0.734633423845964, "grad_norm": 0.41457122564315796, "learning_rate": 2.8443289891103634e-05, "loss": 0.1268, "step": 1860 }, { "epoch": 0.738583065909652, "grad_norm": 0.559184193611145, "learning_rate": 2.842671644042043e-05, "loss": 0.1273, "step": 1870 }, { "epoch": 0.7425327079733399, "grad_norm": 0.46100959181785583, "learning_rate": 2.8410060113185724e-05, "loss": 0.1357, "step": 1880 }, { "epoch": 0.7464823500370279, "grad_norm": 0.5634859204292297, "learning_rate": 2.8393321012210877e-05, "loss": 0.1271, "step": 1890 }, { "epoch": 0.7504319921007159, "grad_norm": 0.4173336327075958, "learning_rate": 2.8376499240818164e-05, "loss": 0.1302, "step": 1900 }, { "epoch": 0.7543816341644038, "grad_norm": 0.40243804454803467, "learning_rate": 2.8359594902840152e-05, "loss": 0.1333, "step": 1910 }, { "epoch": 0.7583312762280918, "grad_norm": 0.3762458562850952, "learning_rate": 2.8342608102619052e-05, "loss": 0.1271, "step": 1920 }, { "epoch": 0.7622809182917798, "grad_norm": 0.43715864419937134, "learning_rate": 2.832553894500607e-05, "loss": 0.1484, "step": 1930 }, { "epoch": 0.7662305603554678, "grad_norm": 0.3971126675605774, "learning_rate": 2.8308387535360763e-05, "loss": 0.1258, "step": 1940 }, { "epoch": 0.7701802024191557, "grad_norm": 0.40626007318496704, "learning_rate": 2.829115397955039e-05, "loss": 0.1336, "step": 1950 }, { "epoch": 0.7741298444828437, "grad_norm": 0.503835141658783, "learning_rate": 2.827383838394926e-05, "loss": 0.135, "step": 1960 }, { "epoch": 0.7780794865465317, "grad_norm": 0.5298701524734497, "learning_rate": 2.8256440855438074e-05, "loss": 0.1409, "step": 1970 }, { "epoch": 0.7820291286102197, "grad_norm": 0.5498703122138977, "learning_rate": 2.8238961501403266e-05, "loss": 0.1453, "step": 1980 }, { "epoch": 0.7859787706739076, "grad_norm": 0.4256785809993744, "learning_rate": 2.8221400429736332e-05, "loss": 0.1297, "step": 1990 }, { "epoch": 0.7899284127375956, "grad_norm": 0.38886457681655884, "learning_rate": 2.820375774883318e-05, "loss": 0.13, "step": 2000 }, { "epoch": 0.7938780548012836, "grad_norm": 0.5477973222732544, "learning_rate": 2.8186033567593445e-05, "loss": 0.1398, "step": 2010 }, { "epoch": 0.7978276968649716, "grad_norm": 0.4944402277469635, "learning_rate": 2.8168227995419828e-05, "loss": 0.1259, "step": 2020 }, { "epoch": 0.8017773389286595, "grad_norm": 0.4402163624763489, "learning_rate": 2.8150341142217407e-05, "loss": 0.1368, "step": 2030 }, { "epoch": 0.8057269809923475, "grad_norm": 0.4140058755874634, "learning_rate": 2.8132373118392985e-05, "loss": 0.1402, "step": 2040 }, { "epoch": 0.8096766230560355, "grad_norm": 0.5238107442855835, "learning_rate": 2.8114324034854378e-05, "loss": 0.1336, "step": 2050 }, { "epoch": 0.8136262651197236, "grad_norm": 0.45435237884521484, "learning_rate": 2.809619400300975e-05, "loss": 0.1421, "step": 2060 }, { "epoch": 0.8175759071834116, "grad_norm": 0.5276714563369751, "learning_rate": 2.8077983134766914e-05, "loss": 0.1234, "step": 2070 }, { "epoch": 0.8215255492470995, "grad_norm": 0.4083622395992279, "learning_rate": 2.8059691542532657e-05, "loss": 0.13, "step": 2080 }, { "epoch": 0.8254751913107875, "grad_norm": 0.3944040834903717, "learning_rate": 2.8041319339212017e-05, "loss": 0.1229, "step": 2090 }, { "epoch": 0.8294248333744755, "grad_norm": 0.5149401426315308, "learning_rate": 2.802286663820763e-05, "loss": 0.1349, "step": 2100 }, { "epoch": 0.8333744754381635, "grad_norm": 0.5086573362350464, "learning_rate": 2.800433355341898e-05, "loss": 0.1367, "step": 2110 }, { "epoch": 0.8373241175018514, "grad_norm": 0.47434648871421814, "learning_rate": 2.7985720199241736e-05, "loss": 0.1458, "step": 2120 }, { "epoch": 0.8412737595655394, "grad_norm": 0.5867214798927307, "learning_rate": 2.796702669056703e-05, "loss": 0.1319, "step": 2130 }, { "epoch": 0.8452234016292274, "grad_norm": 0.4446616768836975, "learning_rate": 2.794825314278074e-05, "loss": 0.1266, "step": 2140 }, { "epoch": 0.8491730436929154, "grad_norm": 0.44527551531791687, "learning_rate": 2.7929399671762794e-05, "loss": 0.1396, "step": 2150 }, { "epoch": 0.8531226857566033, "grad_norm": 0.4233611524105072, "learning_rate": 2.791046639388644e-05, "loss": 0.1265, "step": 2160 }, { "epoch": 0.8570723278202913, "grad_norm": 0.42697539925575256, "learning_rate": 2.7891453426017552e-05, "loss": 0.129, "step": 2170 }, { "epoch": 0.8610219698839793, "grad_norm": 0.5311276912689209, "learning_rate": 2.7872360885513862e-05, "loss": 0.1351, "step": 2180 }, { "epoch": 0.8649716119476673, "grad_norm": 0.45064228773117065, "learning_rate": 2.7853188890224292e-05, "loss": 0.1132, "step": 2190 }, { "epoch": 0.8689212540113552, "grad_norm": 0.39009493589401245, "learning_rate": 2.7833937558488185e-05, "loss": 0.1327, "step": 2200 }, { "epoch": 0.8728708960750432, "grad_norm": 0.39206671714782715, "learning_rate": 2.7814607009134595e-05, "loss": 0.1209, "step": 2210 }, { "epoch": 0.8768205381387312, "grad_norm": 0.35631102323532104, "learning_rate": 2.7795197361481545e-05, "loss": 0.1267, "step": 2220 }, { "epoch": 0.8807701802024192, "grad_norm": 0.4283501207828522, "learning_rate": 2.7775708735335293e-05, "loss": 0.135, "step": 2230 }, { "epoch": 0.8847198222661071, "grad_norm": 0.3623165190219879, "learning_rate": 2.7756141250989593e-05, "loss": 0.1277, "step": 2240 }, { "epoch": 0.8886694643297951, "grad_norm": 0.42114606499671936, "learning_rate": 2.773649502922495e-05, "loss": 0.1378, "step": 2250 }, { "epoch": 0.8926191063934831, "grad_norm": 0.4476473033428192, "learning_rate": 2.7716770191307887e-05, "loss": 0.1296, "step": 2260 }, { "epoch": 0.896568748457171, "grad_norm": 0.3927001655101776, "learning_rate": 2.7696966858990172e-05, "loss": 0.1348, "step": 2270 }, { "epoch": 0.900518390520859, "grad_norm": 0.4335472881793976, "learning_rate": 2.7677085154508085e-05, "loss": 0.1243, "step": 2280 }, { "epoch": 0.904468032584547, "grad_norm": 0.432326078414917, "learning_rate": 2.7657125200581666e-05, "loss": 0.1232, "step": 2290 }, { "epoch": 0.908417674648235, "grad_norm": 0.42572349309921265, "learning_rate": 2.7637087120413937e-05, "loss": 0.1197, "step": 2300 }, { "epoch": 0.912367316711923, "grad_norm": 0.5097776651382446, "learning_rate": 2.761697103769017e-05, "loss": 0.1106, "step": 2310 }, { "epoch": 0.9163169587756109, "grad_norm": 0.4214634895324707, "learning_rate": 2.7596777076577105e-05, "loss": 0.1306, "step": 2320 }, { "epoch": 0.9202666008392989, "grad_norm": 0.5993767380714417, "learning_rate": 2.7576505361722174e-05, "loss": 0.1308, "step": 2330 }, { "epoch": 0.9242162429029869, "grad_norm": 0.44176799058914185, "learning_rate": 2.755615601825276e-05, "loss": 0.1348, "step": 2340 }, { "epoch": 0.9281658849666748, "grad_norm": 0.4011238217353821, "learning_rate": 2.7535729171775406e-05, "loss": 0.1357, "step": 2350 }, { "epoch": 0.9321155270303628, "grad_norm": 0.35617443919181824, "learning_rate": 2.7515224948375038e-05, "loss": 0.1299, "step": 2360 }, { "epoch": 0.9360651690940508, "grad_norm": 0.3995439112186432, "learning_rate": 2.7494643474614197e-05, "loss": 0.1327, "step": 2370 }, { "epoch": 0.9400148111577389, "grad_norm": 0.35780492424964905, "learning_rate": 2.7473984877532247e-05, "loss": 0.1407, "step": 2380 }, { "epoch": 0.9439644532214269, "grad_norm": 0.46763497591018677, "learning_rate": 2.745324928464461e-05, "loss": 0.1316, "step": 2390 }, { "epoch": 0.9479140952851148, "grad_norm": 0.5247623324394226, "learning_rate": 2.743243682394195e-05, "loss": 0.1353, "step": 2400 }, { "epoch": 0.9518637373488028, "grad_norm": 0.5231720805168152, "learning_rate": 2.7411547623889397e-05, "loss": 0.127, "step": 2410 }, { "epoch": 0.9558133794124908, "grad_norm": 0.38919833302497864, "learning_rate": 2.7390581813425776e-05, "loss": 0.1197, "step": 2420 }, { "epoch": 0.9597630214761788, "grad_norm": 0.4457249045372009, "learning_rate": 2.736953952196277e-05, "loss": 0.1333, "step": 2430 }, { "epoch": 0.9637126635398667, "grad_norm": 0.5078391432762146, "learning_rate": 2.734842087938415e-05, "loss": 0.1318, "step": 2440 }, { "epoch": 0.9676623056035547, "grad_norm": 0.5152226090431213, "learning_rate": 2.7327226016044965e-05, "loss": 0.133, "step": 2450 }, { "epoch": 0.9716119476672427, "grad_norm": 0.4484505355358124, "learning_rate": 2.7305955062770738e-05, "loss": 0.1291, "step": 2460 }, { "epoch": 0.9755615897309307, "grad_norm": 0.38752976059913635, "learning_rate": 2.728460815085665e-05, "loss": 0.1274, "step": 2470 }, { "epoch": 0.9795112317946186, "grad_norm": 0.433149129152298, "learning_rate": 2.7263185412066756e-05, "loss": 0.1205, "step": 2480 }, { "epoch": 0.9834608738583066, "grad_norm": 0.4824409782886505, "learning_rate": 2.724168697863313e-05, "loss": 0.1369, "step": 2490 }, { "epoch": 0.9874105159219946, "grad_norm": 0.4385254383087158, "learning_rate": 2.722011298325509e-05, "loss": 0.1249, "step": 2500 }, { "epoch": 0.9913601579856826, "grad_norm": 0.44593289494514465, "learning_rate": 2.719846355909835e-05, "loss": 0.1336, "step": 2510 }, { "epoch": 0.9953098000493705, "grad_norm": 0.5583507418632507, "learning_rate": 2.7176738839794218e-05, "loss": 0.1402, "step": 2520 }, { "epoch": 0.9992594421130585, "grad_norm": 0.47735145688056946, "learning_rate": 2.7154938959438757e-05, "loss": 0.1241, "step": 2530 } ], "logging_steps": 10, "max_steps": 12655, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.5381326285910835e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }