{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 183, "global_step": 549, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0018214936247723133, "grad_norm": 8.483417195876436, "learning_rate": 4.705882352941176e-07, "loss": 0.792, "step": 1 }, { "epoch": 0.0036429872495446266, "grad_norm": 7.370293355302168, "learning_rate": 9.411764705882352e-07, "loss": 0.7618, "step": 2 }, { "epoch": 0.00546448087431694, "grad_norm": 8.828281470296401, "learning_rate": 1.411764705882353e-06, "loss": 0.798, "step": 3 }, { "epoch": 0.007285974499089253, "grad_norm": 7.996641905843589, "learning_rate": 1.8823529411764705e-06, "loss": 0.799, "step": 4 }, { "epoch": 0.009107468123861567, "grad_norm": 5.679308683510637, "learning_rate": 2.352941176470588e-06, "loss": 0.703, "step": 5 }, { "epoch": 0.01092896174863388, "grad_norm": 4.017882280804956, "learning_rate": 2.823529411764706e-06, "loss": 0.6024, "step": 6 }, { "epoch": 0.012750455373406194, "grad_norm": 2.6645091422211795, "learning_rate": 3.294117647058823e-06, "loss": 0.716, "step": 7 }, { "epoch": 0.014571948998178506, "grad_norm": 3.0108132390816054, "learning_rate": 3.764705882352941e-06, "loss": 0.6157, "step": 8 }, { "epoch": 0.01639344262295082, "grad_norm": 2.748696274813349, "learning_rate": 4.235294117647058e-06, "loss": 0.674, "step": 9 }, { "epoch": 0.018214936247723135, "grad_norm": 2.765963400363717, "learning_rate": 4.705882352941176e-06, "loss": 0.662, "step": 10 }, { "epoch": 0.020036429872495445, "grad_norm": 2.0283815116624813, "learning_rate": 5.176470588235294e-06, "loss": 0.5478, "step": 11 }, { "epoch": 0.02185792349726776, "grad_norm": 1.6884205496189686, "learning_rate": 5.647058823529412e-06, "loss": 0.6197, "step": 12 }, { "epoch": 0.023679417122040074, "grad_norm": 1.9209865174719332, "learning_rate": 6.1176470588235285e-06, "loss": 0.6172, "step": 13 }, { "epoch": 0.025500910746812388, "grad_norm": 1.6671170743222523, "learning_rate": 6.588235294117646e-06, "loss": 0.7003, "step": 14 }, { "epoch": 0.0273224043715847, "grad_norm": 1.7653937318251773, "learning_rate": 7.058823529411764e-06, "loss": 0.6144, "step": 15 }, { "epoch": 0.029143897996357013, "grad_norm": 1.6694163536385278, "learning_rate": 7.529411764705882e-06, "loss": 0.6644, "step": 16 }, { "epoch": 0.030965391621129327, "grad_norm": 1.24459662447303, "learning_rate": 8e-06, "loss": 0.6239, "step": 17 }, { "epoch": 0.03278688524590164, "grad_norm": 1.4386147481898333, "learning_rate": 7.999930256262932e-06, "loss": 0.598, "step": 18 }, { "epoch": 0.03460837887067395, "grad_norm": 1.6183710176455128, "learning_rate": 7.999721027483818e-06, "loss": 0.6282, "step": 19 }, { "epoch": 0.03642987249544627, "grad_norm": 2.2610452246136794, "learning_rate": 7.999372320958861e-06, "loss": 0.5843, "step": 20 }, { "epoch": 0.03825136612021858, "grad_norm": 1.7393146801183827, "learning_rate": 7.998884148848109e-06, "loss": 0.6479, "step": 21 }, { "epoch": 0.04007285974499089, "grad_norm": 1.3703584066908119, "learning_rate": 7.998256528175033e-06, "loss": 0.5559, "step": 22 }, { "epoch": 0.04189435336976321, "grad_norm": 1.3670186569963991, "learning_rate": 7.997489480825941e-06, "loss": 0.5666, "step": 23 }, { "epoch": 0.04371584699453552, "grad_norm": 1.383027766675318, "learning_rate": 7.996583033549204e-06, "loss": 0.6172, "step": 24 }, { "epoch": 0.04553734061930783, "grad_norm": 1.4278949896780369, "learning_rate": 7.995537217954335e-06, "loss": 0.6508, "step": 25 }, { "epoch": 0.04735883424408015, "grad_norm": 1.3763761366268141, "learning_rate": 7.994352070510876e-06, "loss": 0.5794, "step": 26 }, { "epoch": 0.04918032786885246, "grad_norm": 1.269210429185682, "learning_rate": 7.993027632547137e-06, "loss": 0.6494, "step": 27 }, { "epoch": 0.051001821493624776, "grad_norm": 1.293399215498077, "learning_rate": 7.991563950248739e-06, "loss": 0.6337, "step": 28 }, { "epoch": 0.052823315118397086, "grad_norm": 1.3443794138088765, "learning_rate": 7.989961074657023e-06, "loss": 0.6124, "step": 29 }, { "epoch": 0.0546448087431694, "grad_norm": 1.4797305271067247, "learning_rate": 7.988219061667252e-06, "loss": 0.64, "step": 30 }, { "epoch": 0.056466302367941715, "grad_norm": 2.20965791616683, "learning_rate": 7.986337972026678e-06, "loss": 0.5096, "step": 31 }, { "epoch": 0.058287795992714025, "grad_norm": 1.3472988619381234, "learning_rate": 7.98431787133241e-06, "loss": 0.5845, "step": 32 }, { "epoch": 0.060109289617486336, "grad_norm": 1.922191039308769, "learning_rate": 7.982158830029133e-06, "loss": 0.5812, "step": 33 }, { "epoch": 0.061930783242258654, "grad_norm": 1.5178941257266974, "learning_rate": 7.979860923406654e-06, "loss": 0.6065, "step": 34 }, { "epoch": 0.06375227686703097, "grad_norm": 1.2582259195864762, "learning_rate": 7.977424231597266e-06, "loss": 0.6219, "step": 35 }, { "epoch": 0.06557377049180328, "grad_norm": 1.440820716830347, "learning_rate": 7.97484883957297e-06, "loss": 0.6243, "step": 36 }, { "epoch": 0.06739526411657559, "grad_norm": 1.2398972840100644, "learning_rate": 7.972134837142497e-06, "loss": 0.5014, "step": 37 }, { "epoch": 0.0692167577413479, "grad_norm": 1.2685725818359594, "learning_rate": 7.969282318948179e-06, "loss": 0.6028, "step": 38 }, { "epoch": 0.07103825136612021, "grad_norm": 1.3336293286163448, "learning_rate": 7.966291384462662e-06, "loss": 0.5084, "step": 39 }, { "epoch": 0.07285974499089254, "grad_norm": 1.3736224319100918, "learning_rate": 7.963162137985416e-06, "loss": 0.6208, "step": 40 }, { "epoch": 0.07468123861566485, "grad_norm": 1.1372782370001664, "learning_rate": 7.959894688639114e-06, "loss": 0.5566, "step": 41 }, { "epoch": 0.07650273224043716, "grad_norm": 1.1370294392240514, "learning_rate": 7.956489150365818e-06, "loss": 0.5316, "step": 42 }, { "epoch": 0.07832422586520947, "grad_norm": 1.3638397949259238, "learning_rate": 7.952945641923014e-06, "loss": 0.5665, "step": 43 }, { "epoch": 0.08014571948998178, "grad_norm": 1.3721010139488579, "learning_rate": 7.949264286879461e-06, "loss": 0.6686, "step": 44 }, { "epoch": 0.08196721311475409, "grad_norm": 1.3079030861579901, "learning_rate": 7.94544521361089e-06, "loss": 0.5678, "step": 45 }, { "epoch": 0.08378870673952642, "grad_norm": 1.3169311772009933, "learning_rate": 7.941488555295519e-06, "loss": 0.5883, "step": 46 }, { "epoch": 0.08561020036429873, "grad_norm": 1.187489505664262, "learning_rate": 7.937394449909417e-06, "loss": 0.6398, "step": 47 }, { "epoch": 0.08743169398907104, "grad_norm": 1.3538084566677648, "learning_rate": 7.933163040221691e-06, "loss": 0.6001, "step": 48 }, { "epoch": 0.08925318761384335, "grad_norm": 1.282030778796261, "learning_rate": 7.928794473789502e-06, "loss": 0.6299, "step": 49 }, { "epoch": 0.09107468123861566, "grad_norm": 1.291046733294755, "learning_rate": 7.924288902952924e-06, "loss": 0.6643, "step": 50 }, { "epoch": 0.09289617486338798, "grad_norm": 1.2600029350057804, "learning_rate": 7.91964648482963e-06, "loss": 0.6266, "step": 51 }, { "epoch": 0.0947176684881603, "grad_norm": 1.383888278177588, "learning_rate": 7.914867381309417e-06, "loss": 0.5565, "step": 52 }, { "epoch": 0.0965391621129326, "grad_norm": 1.2920207521690645, "learning_rate": 7.909951759048553e-06, "loss": 0.5539, "step": 53 }, { "epoch": 0.09836065573770492, "grad_norm": 1.4806098875225748, "learning_rate": 7.904899789463974e-06, "loss": 0.4836, "step": 54 }, { "epoch": 0.10018214936247723, "grad_norm": 1.3236781330536374, "learning_rate": 7.899711648727295e-06, "loss": 0.6716, "step": 55 }, { "epoch": 0.10200364298724955, "grad_norm": 1.307628398165257, "learning_rate": 7.894387517758679e-06, "loss": 0.5806, "step": 56 }, { "epoch": 0.10382513661202186, "grad_norm": 1.2494530103588137, "learning_rate": 7.888927582220521e-06, "loss": 0.5255, "step": 57 }, { "epoch": 0.10564663023679417, "grad_norm": 1.1220353680400945, "learning_rate": 7.883332032510978e-06, "loss": 0.5951, "step": 58 }, { "epoch": 0.10746812386156648, "grad_norm": 1.2679713293816075, "learning_rate": 7.877601063757322e-06, "loss": 0.6249, "step": 59 }, { "epoch": 0.1092896174863388, "grad_norm": 1.367413675973217, "learning_rate": 7.871734875809141e-06, "loss": 0.5839, "step": 60 }, { "epoch": 0.1111111111111111, "grad_norm": 1.3310878119591731, "learning_rate": 7.86573367323137e-06, "loss": 0.5441, "step": 61 }, { "epoch": 0.11293260473588343, "grad_norm": 1.2672834778790825, "learning_rate": 7.859597665297158e-06, "loss": 0.5799, "step": 62 }, { "epoch": 0.11475409836065574, "grad_norm": 1.4949905078759134, "learning_rate": 7.853327065980567e-06, "loss": 0.5811, "step": 63 }, { "epoch": 0.11657559198542805, "grad_norm": 1.2772501302963368, "learning_rate": 7.84692209394911e-06, "loss": 0.5454, "step": 64 }, { "epoch": 0.11839708561020036, "grad_norm": 1.4486798634496518, "learning_rate": 7.84038297255613e-06, "loss": 0.6189, "step": 65 }, { "epoch": 0.12021857923497267, "grad_norm": 1.247869533369392, "learning_rate": 7.83370992983301e-06, "loss": 0.6109, "step": 66 }, { "epoch": 0.122040072859745, "grad_norm": 1.4278988044592194, "learning_rate": 7.826903198481218e-06, "loss": 0.6362, "step": 67 }, { "epoch": 0.12386156648451731, "grad_norm": 1.4421005783850482, "learning_rate": 7.819963015864195e-06, "loss": 0.5953, "step": 68 }, { "epoch": 0.12568306010928962, "grad_norm": 1.2853062012314034, "learning_rate": 7.812889623999077e-06, "loss": 0.6449, "step": 69 }, { "epoch": 0.12750455373406194, "grad_norm": 1.1806756552505506, "learning_rate": 7.805683269548253e-06, "loss": 0.4961, "step": 70 }, { "epoch": 0.12932604735883424, "grad_norm": 1.1874724065893263, "learning_rate": 7.798344203810772e-06, "loss": 0.5731, "step": 71 }, { "epoch": 0.13114754098360656, "grad_norm": 1.4302204052792575, "learning_rate": 7.790872682713567e-06, "loss": 0.5694, "step": 72 }, { "epoch": 0.13296903460837886, "grad_norm": 1.621022341496324, "learning_rate": 7.783268966802538e-06, "loss": 0.5683, "step": 73 }, { "epoch": 0.13479052823315119, "grad_norm": 1.266495179779122, "learning_rate": 7.77553332123347e-06, "loss": 0.6351, "step": 74 }, { "epoch": 0.1366120218579235, "grad_norm": 1.4124120835397862, "learning_rate": 7.767666015762775e-06, "loss": 0.5941, "step": 75 }, { "epoch": 0.1384335154826958, "grad_norm": 1.3413868354070095, "learning_rate": 7.7596673247381e-06, "loss": 0.5407, "step": 76 }, { "epoch": 0.14025500910746813, "grad_norm": 1.305749508434761, "learning_rate": 7.751537527088742e-06, "loss": 0.5331, "step": 77 }, { "epoch": 0.14207650273224043, "grad_norm": 1.2388240813097555, "learning_rate": 7.743276906315936e-06, "loss": 0.6214, "step": 78 }, { "epoch": 0.14389799635701275, "grad_norm": 1.4526353435545132, "learning_rate": 7.734885750482967e-06, "loss": 0.5391, "step": 79 }, { "epoch": 0.14571948998178508, "grad_norm": 1.3313440090499233, "learning_rate": 7.726364352205116e-06, "loss": 0.6254, "step": 80 }, { "epoch": 0.14754098360655737, "grad_norm": 1.336317757950119, "learning_rate": 7.717713008639463e-06, "loss": 0.4752, "step": 81 }, { "epoch": 0.1493624772313297, "grad_norm": 1.2089723824473042, "learning_rate": 7.708932021474524e-06, "loss": 0.7128, "step": 82 }, { "epoch": 0.151183970856102, "grad_norm": 1.4923241883131872, "learning_rate": 7.70002169691973e-06, "loss": 0.6695, "step": 83 }, { "epoch": 0.15300546448087432, "grad_norm": 1.1956673865353409, "learning_rate": 7.690982345694746e-06, "loss": 0.5046, "step": 84 }, { "epoch": 0.15482695810564662, "grad_norm": 1.3110716423599806, "learning_rate": 7.68181428301864e-06, "loss": 0.5885, "step": 85 }, { "epoch": 0.15664845173041894, "grad_norm": 1.340432170090615, "learning_rate": 7.67251782859889e-06, "loss": 0.5455, "step": 86 }, { "epoch": 0.15846994535519127, "grad_norm": 1.2506027715723123, "learning_rate": 7.663093306620228e-06, "loss": 0.5864, "step": 87 }, { "epoch": 0.16029143897996356, "grad_norm": 1.3974083197789515, "learning_rate": 7.653541045733351e-06, "loss": 0.5428, "step": 88 }, { "epoch": 0.1621129326047359, "grad_norm": 1.5848220165792029, "learning_rate": 7.643861379043442e-06, "loss": 0.577, "step": 89 }, { "epoch": 0.16393442622950818, "grad_norm": 1.319727068673188, "learning_rate": 7.634054644098566e-06, "loss": 0.5787, "step": 90 }, { "epoch": 0.1657559198542805, "grad_norm": 1.3075529460521724, "learning_rate": 7.624121182877892e-06, "loss": 0.6063, "step": 91 }, { "epoch": 0.16757741347905283, "grad_norm": 1.3869090444503254, "learning_rate": 7.614061341779777e-06, "loss": 0.5533, "step": 92 }, { "epoch": 0.16939890710382513, "grad_norm": 1.3038369655118094, "learning_rate": 7.6038754716096755e-06, "loss": 0.6403, "step": 93 }, { "epoch": 0.17122040072859745, "grad_norm": 1.201815586378659, "learning_rate": 7.593563927567915e-06, "loss": 0.5341, "step": 94 }, { "epoch": 0.17304189435336975, "grad_norm": 1.2948505897209415, "learning_rate": 7.583127069237302e-06, "loss": 0.5988, "step": 95 }, { "epoch": 0.17486338797814208, "grad_norm": 1.2777544112935704, "learning_rate": 7.5725652605705876e-06, "loss": 0.5467, "step": 96 }, { "epoch": 0.1766848816029144, "grad_norm": 1.347734488800424, "learning_rate": 7.561878869877778e-06, "loss": 0.6734, "step": 97 }, { "epoch": 0.1785063752276867, "grad_norm": 1.3934793951098927, "learning_rate": 7.551068269813282e-06, "loss": 0.5408, "step": 98 }, { "epoch": 0.18032786885245902, "grad_norm": 1.2265308094279608, "learning_rate": 7.540133837362924e-06, "loss": 0.5798, "step": 99 }, { "epoch": 0.18214936247723132, "grad_norm": 1.2042353648647974, "learning_rate": 7.5290759538307944e-06, "loss": 0.5349, "step": 100 }, { "epoch": 0.18397085610200364, "grad_norm": 1.198872560339186, "learning_rate": 7.517895004825955e-06, "loss": 0.6078, "step": 101 }, { "epoch": 0.18579234972677597, "grad_norm": 1.5341596102925161, "learning_rate": 7.506591380248991e-06, "loss": 0.669, "step": 102 }, { "epoch": 0.18761384335154827, "grad_norm": 1.2181363687268225, "learning_rate": 7.495165474278411e-06, "loss": 0.5508, "step": 103 }, { "epoch": 0.1894353369763206, "grad_norm": 1.2061034316572241, "learning_rate": 7.483617685356906e-06, "loss": 0.6374, "step": 104 }, { "epoch": 0.1912568306010929, "grad_norm": 1.270570136570114, "learning_rate": 7.471948416177452e-06, "loss": 0.6582, "step": 105 }, { "epoch": 0.1930783242258652, "grad_norm": 1.2220914581087463, "learning_rate": 7.460158073669271e-06, "loss": 0.5195, "step": 106 }, { "epoch": 0.19489981785063754, "grad_norm": 1.1327340940905815, "learning_rate": 7.448247068983638e-06, "loss": 0.6108, "step": 107 }, { "epoch": 0.19672131147540983, "grad_norm": 1.2389921551395293, "learning_rate": 7.43621581747954e-06, "loss": 0.5197, "step": 108 }, { "epoch": 0.19854280510018216, "grad_norm": 1.3119245038499585, "learning_rate": 7.4240647387092e-06, "loss": 0.5787, "step": 109 }, { "epoch": 0.20036429872495445, "grad_norm": 1.2985603971431927, "learning_rate": 7.411794256403439e-06, "loss": 0.5408, "step": 110 }, { "epoch": 0.20218579234972678, "grad_norm": 1.3547972287329184, "learning_rate": 7.399404798456901e-06, "loss": 0.5645, "step": 111 }, { "epoch": 0.2040072859744991, "grad_norm": 1.247671888832723, "learning_rate": 7.3868967969131364e-06, "loss": 0.594, "step": 112 }, { "epoch": 0.2058287795992714, "grad_norm": 1.1653540957694166, "learning_rate": 7.374270687949531e-06, "loss": 0.6022, "step": 113 }, { "epoch": 0.20765027322404372, "grad_norm": 1.110075329220651, "learning_rate": 7.3615269118620945e-06, "loss": 0.4952, "step": 114 }, { "epoch": 0.20947176684881602, "grad_norm": 1.1773358008446368, "learning_rate": 7.348665913050114e-06, "loss": 0.5796, "step": 115 }, { "epoch": 0.21129326047358835, "grad_norm": 1.1762014415226016, "learning_rate": 7.3356881400006485e-06, "loss": 0.5019, "step": 116 }, { "epoch": 0.21311475409836064, "grad_norm": 1.2857996230820723, "learning_rate": 7.3225940452728915e-06, "loss": 0.6128, "step": 117 }, { "epoch": 0.21493624772313297, "grad_norm": 1.278319933723205, "learning_rate": 7.309384085482396e-06, "loss": 0.6233, "step": 118 }, { "epoch": 0.2167577413479053, "grad_norm": 1.3246399685797854, "learning_rate": 7.29605872128514e-06, "loss": 0.5597, "step": 119 }, { "epoch": 0.2185792349726776, "grad_norm": 1.3738785761031618, "learning_rate": 7.282618417361476e-06, "loss": 0.6105, "step": 120 }, { "epoch": 0.2204007285974499, "grad_norm": 1.510255253584312, "learning_rate": 7.269063642399912e-06, "loss": 0.5677, "step": 121 }, { "epoch": 0.2222222222222222, "grad_norm": 1.3705154463280955, "learning_rate": 7.25539486908078e-06, "loss": 0.5172, "step": 122 }, { "epoch": 0.22404371584699453, "grad_norm": 1.4229862954590966, "learning_rate": 7.241612574059745e-06, "loss": 0.5987, "step": 123 }, { "epoch": 0.22586520947176686, "grad_norm": 1.3813913047228632, "learning_rate": 7.227717237951189e-06, "loss": 0.6519, "step": 124 }, { "epoch": 0.22768670309653916, "grad_norm": 1.4086532700921839, "learning_rate": 7.213709345311444e-06, "loss": 0.5507, "step": 125 }, { "epoch": 0.22950819672131148, "grad_norm": 1.3222980277500922, "learning_rate": 7.1995893846219035e-06, "loss": 0.5821, "step": 126 }, { "epoch": 0.23132969034608378, "grad_norm": 1.2772860113277071, "learning_rate": 7.185357848271977e-06, "loss": 0.5552, "step": 127 }, { "epoch": 0.2331511839708561, "grad_norm": 1.3184892568778903, "learning_rate": 7.17101523254193e-06, "loss": 0.4548, "step": 128 }, { "epoch": 0.23497267759562843, "grad_norm": 1.417969805143716, "learning_rate": 7.156562037585575e-06, "loss": 0.5907, "step": 129 }, { "epoch": 0.23679417122040072, "grad_norm": 1.3194920301350814, "learning_rate": 7.1419987674128225e-06, "loss": 0.5377, "step": 130 }, { "epoch": 0.23861566484517305, "grad_norm": 1.1915857881381822, "learning_rate": 7.127325929872119e-06, "loss": 0.5803, "step": 131 }, { "epoch": 0.24043715846994534, "grad_norm": 1.294311599245113, "learning_rate": 7.1125440366327245e-06, "loss": 0.533, "step": 132 }, { "epoch": 0.24225865209471767, "grad_norm": 1.2869026066621274, "learning_rate": 7.0976536031668775e-06, "loss": 0.5978, "step": 133 }, { "epoch": 0.24408014571949, "grad_norm": 1.2882193076261699, "learning_rate": 7.082655148731815e-06, "loss": 0.5998, "step": 134 }, { "epoch": 0.2459016393442623, "grad_norm": 1.2783117350056112, "learning_rate": 7.067549196351669e-06, "loss": 0.5792, "step": 135 }, { "epoch": 0.24772313296903462, "grad_norm": 1.2465596890374582, "learning_rate": 7.052336272799226e-06, "loss": 0.6114, "step": 136 }, { "epoch": 0.2495446265938069, "grad_norm": 1.3115289926959033, "learning_rate": 7.037016908577555e-06, "loss": 0.5711, "step": 137 }, { "epoch": 0.25136612021857924, "grad_norm": 1.2945372654161171, "learning_rate": 7.02159163790151e-06, "loss": 0.6108, "step": 138 }, { "epoch": 0.25318761384335153, "grad_norm": 1.179854303445542, "learning_rate": 7.006060998679105e-06, "loss": 0.5712, "step": 139 }, { "epoch": 0.2550091074681239, "grad_norm": 1.3505552689452494, "learning_rate": 6.990425532492747e-06, "loss": 0.607, "step": 140 }, { "epoch": 0.2568306010928962, "grad_norm": 1.293079490327599, "learning_rate": 6.974685784580359e-06, "loss": 0.5907, "step": 141 }, { "epoch": 0.2586520947176685, "grad_norm": 1.2632127589893052, "learning_rate": 6.958842303816359e-06, "loss": 0.6264, "step": 142 }, { "epoch": 0.2604735883424408, "grad_norm": 1.3425160590690703, "learning_rate": 6.942895642692527e-06, "loss": 0.4246, "step": 143 }, { "epoch": 0.26229508196721313, "grad_norm": 1.5771791991344761, "learning_rate": 6.926846357298732e-06, "loss": 0.5806, "step": 144 }, { "epoch": 0.2641165755919854, "grad_norm": 1.4030090930033172, "learning_rate": 6.910695007303544e-06, "loss": 0.5502, "step": 145 }, { "epoch": 0.2659380692167577, "grad_norm": 1.3313664381580208, "learning_rate": 6.894442155934719e-06, "loss": 0.5853, "step": 146 }, { "epoch": 0.2677595628415301, "grad_norm": 1.2706667077309404, "learning_rate": 6.878088369959553e-06, "loss": 0.5632, "step": 147 }, { "epoch": 0.26958105646630237, "grad_norm": 1.2404396727083782, "learning_rate": 6.861634219665117e-06, "loss": 0.6277, "step": 148 }, { "epoch": 0.27140255009107467, "grad_norm": 1.266069671654807, "learning_rate": 6.845080278838381e-06, "loss": 0.5883, "step": 149 }, { "epoch": 0.273224043715847, "grad_norm": 1.4551654748561211, "learning_rate": 6.82842712474619e-06, "loss": 0.5585, "step": 150 }, { "epoch": 0.2750455373406193, "grad_norm": 1.4120447459169916, "learning_rate": 6.811675338115146e-06, "loss": 0.5859, "step": 151 }, { "epoch": 0.2768670309653916, "grad_norm": 1.4794301221707373, "learning_rate": 6.7948255031113505e-06, "loss": 0.5735, "step": 152 }, { "epoch": 0.2786885245901639, "grad_norm": 1.4445882004520383, "learning_rate": 6.777878207320034e-06, "loss": 0.5308, "step": 153 }, { "epoch": 0.28051001821493626, "grad_norm": 1.259161824393288, "learning_rate": 6.760834041725068e-06, "loss": 0.5683, "step": 154 }, { "epoch": 0.28233151183970856, "grad_norm": 1.5252085328619271, "learning_rate": 6.743693600688352e-06, "loss": 0.625, "step": 155 }, { "epoch": 0.28415300546448086, "grad_norm": 1.4215413394995613, "learning_rate": 6.726457481929095e-06, "loss": 0.5376, "step": 156 }, { "epoch": 0.2859744990892532, "grad_norm": 1.2834803437487838, "learning_rate": 6.7091262865029645e-06, "loss": 0.5586, "step": 157 }, { "epoch": 0.2877959927140255, "grad_norm": 1.320167275314629, "learning_rate": 6.691700618781126e-06, "loss": 0.5789, "step": 158 }, { "epoch": 0.2896174863387978, "grad_norm": 1.8005652724042749, "learning_rate": 6.674181086429177e-06, "loss": 0.5188, "step": 159 }, { "epoch": 0.29143897996357016, "grad_norm": 1.2690750482423212, "learning_rate": 6.656568300385944e-06, "loss": 0.5046, "step": 160 }, { "epoch": 0.29326047358834245, "grad_norm": 1.31082889556021, "learning_rate": 6.6388628748421895e-06, "loss": 0.578, "step": 161 }, { "epoch": 0.29508196721311475, "grad_norm": 1.3247928047009208, "learning_rate": 6.62106542721918e-06, "loss": 0.6034, "step": 162 }, { "epoch": 0.29690346083788705, "grad_norm": 1.2229208641682885, "learning_rate": 6.603176578147174e-06, "loss": 0.4996, "step": 163 }, { "epoch": 0.2987249544626594, "grad_norm": 1.3325184853391303, "learning_rate": 6.585196951443763e-06, "loss": 0.4985, "step": 164 }, { "epoch": 0.3005464480874317, "grad_norm": 1.326999208770471, "learning_rate": 6.5671271740921266e-06, "loss": 0.5379, "step": 165 }, { "epoch": 0.302367941712204, "grad_norm": 1.2382829663292223, "learning_rate": 6.548967876219163e-06, "loss": 0.5481, "step": 166 }, { "epoch": 0.30418943533697634, "grad_norm": 1.3992391592431397, "learning_rate": 6.530719691073521e-06, "loss": 0.5941, "step": 167 }, { "epoch": 0.30601092896174864, "grad_norm": 1.4237814669133988, "learning_rate": 6.5123832550035165e-06, "loss": 0.5879, "step": 168 }, { "epoch": 0.30783242258652094, "grad_norm": 1.2434267947044435, "learning_rate": 6.493959207434934e-06, "loss": 0.5608, "step": 169 }, { "epoch": 0.30965391621129323, "grad_norm": 1.2949876195852905, "learning_rate": 6.47544819084874e-06, "loss": 0.5477, "step": 170 }, { "epoch": 0.3114754098360656, "grad_norm": 1.2709700316435848, "learning_rate": 6.4568508507586715e-06, "loss": 0.5913, "step": 171 }, { "epoch": 0.3132969034608379, "grad_norm": 1.1917528631285046, "learning_rate": 6.438167835688725e-06, "loss": 0.5811, "step": 172 }, { "epoch": 0.3151183970856102, "grad_norm": 1.4420942251279008, "learning_rate": 6.41939979715055e-06, "loss": 0.5258, "step": 173 }, { "epoch": 0.31693989071038253, "grad_norm": 1.3898239712198435, "learning_rate": 6.400547389620716e-06, "loss": 0.4382, "step": 174 }, { "epoch": 0.31876138433515483, "grad_norm": 1.1944425406473265, "learning_rate": 6.3816112705178984e-06, "loss": 0.4981, "step": 175 }, { "epoch": 0.3205828779599271, "grad_norm": 1.2940463807027722, "learning_rate": 6.362592100179958e-06, "loss": 0.5679, "step": 176 }, { "epoch": 0.3224043715846995, "grad_norm": 1.1827687350225293, "learning_rate": 6.343490541840899e-06, "loss": 0.5778, "step": 177 }, { "epoch": 0.3242258652094718, "grad_norm": 1.2684254658308312, "learning_rate": 6.3243072616077535e-06, "loss": 0.5077, "step": 178 }, { "epoch": 0.32604735883424407, "grad_norm": 1.4375046609341398, "learning_rate": 6.3050429284373465e-06, "loss": 0.5428, "step": 179 }, { "epoch": 0.32786885245901637, "grad_norm": 1.2298035631627269, "learning_rate": 6.285698214112974e-06, "loss": 0.5333, "step": 180 }, { "epoch": 0.3296903460837887, "grad_norm": 1.3091190041481056, "learning_rate": 6.2662737932209695e-06, "loss": 0.4829, "step": 181 }, { "epoch": 0.331511839708561, "grad_norm": 1.399739792081344, "learning_rate": 6.246770343127185e-06, "loss": 0.5039, "step": 182 }, { "epoch": 0.3333333333333333, "grad_norm": 1.4099523447236924, "learning_rate": 6.227188543953368e-06, "loss": 0.4718, "step": 183 }, { "epoch": 0.3333333333333333, "eval_accuracy": 0.8100095999478772, "eval_accuracy_first_token": 0.9386401326699834, "eval_accuracy_first_token_all": 0.9740702326270558, "eval_accuracy_first_token_all_total": 6749, "eval_accuracy_first_token_calculate": 0.8181818181818182, "eval_accuracy_first_token_calculate_total": 44, "eval_accuracy_first_token_execute": 0.995049504950495, "eval_accuracy_first_token_execute_total": 202, "eval_accuracy_first_token_get": 0.9409190371991247, "eval_accuracy_first_token_get_total": 457, "eval_accuracy_first_token_python": 0.7979899497487437, "eval_accuracy_first_token_python_total": 995, "eval_loss": 0.5817924737930298, "eval_perplexity": 1.1146453885480792, "eval_runtime": 276.9015, "eval_samples_per_second": 2.506, "eval_steps_per_second": 0.628, "eval_total_number_first_token": 9648, "step": 183 }, { "epoch": 0.33515482695810567, "grad_norm": 1.3753718658598428, "learning_rate": 6.207529078553444e-06, "loss": 0.4554, "step": 184 }, { "epoch": 0.33697632058287796, "grad_norm": 1.23834147665285, "learning_rate": 6.1877926324897085e-06, "loss": 0.5276, "step": 185 }, { "epoch": 0.33879781420765026, "grad_norm": 1.4417460479810351, "learning_rate": 6.16797989400891e-06, "loss": 0.5891, "step": 186 }, { "epoch": 0.3406193078324226, "grad_norm": 1.212881090816042, "learning_rate": 6.148091554018264e-06, "loss": 0.51, "step": 187 }, { "epoch": 0.3424408014571949, "grad_norm": 1.3090197042208114, "learning_rate": 6.128128306061346e-06, "loss": 0.5551, "step": 188 }, { "epoch": 0.3442622950819672, "grad_norm": 1.4028504868912617, "learning_rate": 6.108090846293915e-06, "loss": 0.571, "step": 189 }, { "epoch": 0.3460837887067395, "grad_norm": 1.2383621150830924, "learning_rate": 6.087979873459634e-06, "loss": 0.4812, "step": 190 }, { "epoch": 0.34790528233151186, "grad_norm": 1.1655822819675992, "learning_rate": 6.0677960888657015e-06, "loss": 0.5027, "step": 191 }, { "epoch": 0.34972677595628415, "grad_norm": 1.2330586732923372, "learning_rate": 6.047540196358404e-06, "loss": 0.5472, "step": 192 }, { "epoch": 0.35154826958105645, "grad_norm": 1.2120668135216859, "learning_rate": 6.02721290229856e-06, "loss": 0.4723, "step": 193 }, { "epoch": 0.3533697632058288, "grad_norm": 1.2127776371529766, "learning_rate": 6.006814915536894e-06, "loss": 0.5994, "step": 194 }, { "epoch": 0.3551912568306011, "grad_norm": 1.3781231876403488, "learning_rate": 5.9863469473893225e-06, "loss": 0.6055, "step": 195 }, { "epoch": 0.3570127504553734, "grad_norm": 1.2377625062579942, "learning_rate": 5.965809711612137e-06, "loss": 0.6285, "step": 196 }, { "epoch": 0.3588342440801457, "grad_norm": 1.2436458612174006, "learning_rate": 5.945203924377125e-06, "loss": 0.5051, "step": 197 }, { "epoch": 0.36065573770491804, "grad_norm": 1.067560558211732, "learning_rate": 5.92453030424659e-06, "loss": 0.557, "step": 198 }, { "epoch": 0.36247723132969034, "grad_norm": 1.339153659611944, "learning_rate": 5.903789572148294e-06, "loss": 0.5624, "step": 199 }, { "epoch": 0.36429872495446264, "grad_norm": 1.3921715569996027, "learning_rate": 5.88298245135032e-06, "loss": 0.6122, "step": 200 }, { "epoch": 0.366120218579235, "grad_norm": 1.0724617492860091, "learning_rate": 5.862109667435853e-06, "loss": 0.406, "step": 201 }, { "epoch": 0.3679417122040073, "grad_norm": 1.3579568036237248, "learning_rate": 5.8411719482778645e-06, "loss": 0.5615, "step": 202 }, { "epoch": 0.3697632058287796, "grad_norm": 1.392362486259979, "learning_rate": 5.820170024013746e-06, "loss": 0.4783, "step": 203 }, { "epoch": 0.37158469945355194, "grad_norm": 1.4663688083889719, "learning_rate": 5.79910462701984e-06, "loss": 0.6462, "step": 204 }, { "epoch": 0.37340619307832423, "grad_norm": 1.5906378406347372, "learning_rate": 5.777976491885903e-06, "loss": 0.5989, "step": 205 }, { "epoch": 0.37522768670309653, "grad_norm": 1.2690398154345799, "learning_rate": 5.756786355389481e-06, "loss": 0.4675, "step": 206 }, { "epoch": 0.3770491803278688, "grad_norm": 1.340814834772367, "learning_rate": 5.735534956470232e-06, "loss": 0.4978, "step": 207 }, { "epoch": 0.3788706739526412, "grad_norm": 1.2798668544381595, "learning_rate": 5.714223036204144e-06, "loss": 0.5032, "step": 208 }, { "epoch": 0.3806921675774135, "grad_norm": 1.2535515686753846, "learning_rate": 5.6928513377777e-06, "loss": 0.53, "step": 209 }, { "epoch": 0.3825136612021858, "grad_norm": 1.2457137330041876, "learning_rate": 5.671420606461956e-06, "loss": 0.5851, "step": 210 }, { "epoch": 0.3843351548269581, "grad_norm": 1.4696601482702962, "learning_rate": 5.649931589586557e-06, "loss": 0.654, "step": 211 }, { "epoch": 0.3861566484517304, "grad_norm": 1.2987011689556953, "learning_rate": 5.628385036513676e-06, "loss": 0.5669, "step": 212 }, { "epoch": 0.3879781420765027, "grad_norm": 1.3159510108963395, "learning_rate": 5.606781698611878e-06, "loss": 0.5532, "step": 213 }, { "epoch": 0.38979963570127507, "grad_norm": 1.1640550154922134, "learning_rate": 5.585122329229923e-06, "loss": 0.4804, "step": 214 }, { "epoch": 0.39162112932604737, "grad_norm": 1.257926153405232, "learning_rate": 5.56340768367049e-06, "loss": 0.5356, "step": 215 }, { "epoch": 0.39344262295081966, "grad_norm": 1.2768971849963524, "learning_rate": 5.541638519163849e-06, "loss": 0.5141, "step": 216 }, { "epoch": 0.39526411657559196, "grad_norm": 1.1748448811474719, "learning_rate": 5.51981559484144e-06, "loss": 0.6517, "step": 217 }, { "epoch": 0.3970856102003643, "grad_norm": 2.0539459488388427, "learning_rate": 5.49793967170941e-06, "loss": 0.5137, "step": 218 }, { "epoch": 0.3989071038251366, "grad_norm": 1.3340041273272538, "learning_rate": 5.476011512622076e-06, "loss": 0.5224, "step": 219 }, { "epoch": 0.4007285974499089, "grad_norm": 1.2742409163320605, "learning_rate": 5.454031882255319e-06, "loss": 0.593, "step": 220 }, { "epoch": 0.40255009107468126, "grad_norm": 1.4312401458156725, "learning_rate": 5.43200154707992e-06, "loss": 0.6052, "step": 221 }, { "epoch": 0.40437158469945356, "grad_norm": 1.1725298771962427, "learning_rate": 5.4099212753348294e-06, "loss": 0.5926, "step": 222 }, { "epoch": 0.40619307832422585, "grad_norm": 1.2329994896181276, "learning_rate": 5.3877918370003806e-06, "loss": 0.4412, "step": 223 }, { "epoch": 0.4080145719489982, "grad_norm": 1.3077181540697282, "learning_rate": 5.365614003771439e-06, "loss": 0.4822, "step": 224 }, { "epoch": 0.4098360655737705, "grad_norm": 1.3216563432171085, "learning_rate": 5.343388549030491e-06, "loss": 0.5746, "step": 225 }, { "epoch": 0.4116575591985428, "grad_norm": 1.4069728629721256, "learning_rate": 5.321116247820669e-06, "loss": 0.4688, "step": 226 }, { "epoch": 0.4134790528233151, "grad_norm": 1.1054828696912837, "learning_rate": 5.298797876818734e-06, "loss": 0.5574, "step": 227 }, { "epoch": 0.41530054644808745, "grad_norm": 1.3971039327183563, "learning_rate": 5.276434214307986e-06, "loss": 0.5568, "step": 228 }, { "epoch": 0.41712204007285975, "grad_norm": 1.2052367676116202, "learning_rate": 5.2540260401511255e-06, "loss": 0.4715, "step": 229 }, { "epoch": 0.41894353369763204, "grad_norm": 1.2975437539065673, "learning_rate": 5.231574135763052e-06, "loss": 0.531, "step": 230 }, { "epoch": 0.4207650273224044, "grad_norm": 1.2072349064171715, "learning_rate": 5.209079284083626e-06, "loss": 0.4933, "step": 231 }, { "epoch": 0.4225865209471767, "grad_norm": 1.28034314366732, "learning_rate": 5.186542269550359e-06, "loss": 0.6263, "step": 232 }, { "epoch": 0.424408014571949, "grad_norm": 1.2774088433481947, "learning_rate": 5.163963878071058e-06, "loss": 0.5761, "step": 233 }, { "epoch": 0.4262295081967213, "grad_norm": 1.2827700082849656, "learning_rate": 5.141344896996421e-06, "loss": 0.5472, "step": 234 }, { "epoch": 0.42805100182149364, "grad_norm": 1.2399786652373073, "learning_rate": 5.1186861150925844e-06, "loss": 0.4721, "step": 235 }, { "epoch": 0.42987249544626593, "grad_norm": 1.4866480679484804, "learning_rate": 5.09598832251361e-06, "loss": 0.5454, "step": 236 }, { "epoch": 0.43169398907103823, "grad_norm": 1.2892625562378888, "learning_rate": 5.073252310773939e-06, "loss": 0.522, "step": 237 }, { "epoch": 0.4335154826958106, "grad_norm": 1.4228683878206347, "learning_rate": 5.050478872720782e-06, "loss": 0.5789, "step": 238 }, { "epoch": 0.4353369763205829, "grad_norm": 1.2557222729475248, "learning_rate": 5.027668802506476e-06, "loss": 0.542, "step": 239 }, { "epoch": 0.4371584699453552, "grad_norm": 1.3719564205247436, "learning_rate": 5.004822895560794e-06, "loss": 0.5978, "step": 240 }, { "epoch": 0.43897996357012753, "grad_norm": 1.2002491118216863, "learning_rate": 4.981941948563196e-06, "loss": 0.567, "step": 241 }, { "epoch": 0.4408014571948998, "grad_norm": 1.23705839412736, "learning_rate": 4.959026759415061e-06, "loss": 0.5622, "step": 242 }, { "epoch": 0.4426229508196721, "grad_norm": 1.2443680094563607, "learning_rate": 4.936078127211849e-06, "loss": 0.4652, "step": 243 }, { "epoch": 0.4444444444444444, "grad_norm": 1.2292169146784966, "learning_rate": 4.913096852215248e-06, "loss": 0.4945, "step": 244 }, { "epoch": 0.44626593806921677, "grad_norm": 1.2199454755400951, "learning_rate": 4.890083735825257e-06, "loss": 0.4975, "step": 245 }, { "epoch": 0.44808743169398907, "grad_norm": 1.1351286410957175, "learning_rate": 4.867039580552247e-06, "loss": 0.6239, "step": 246 }, { "epoch": 0.44990892531876137, "grad_norm": 2.489162682585333, "learning_rate": 4.843965189988969e-06, "loss": 0.4521, "step": 247 }, { "epoch": 0.4517304189435337, "grad_norm": 1.3760193870465582, "learning_rate": 4.820861368782537e-06, "loss": 0.5452, "step": 248 }, { "epoch": 0.453551912568306, "grad_norm": 1.9314160205055073, "learning_rate": 4.79772892260637e-06, "loss": 0.524, "step": 249 }, { "epoch": 0.4553734061930783, "grad_norm": 1.381935096821563, "learning_rate": 4.774568658132086e-06, "loss": 0.5054, "step": 250 }, { "epoch": 0.45719489981785066, "grad_norm": 1.4937455355331937, "learning_rate": 4.751381383001386e-06, "loss": 0.5263, "step": 251 }, { "epoch": 0.45901639344262296, "grad_norm": 1.2817536849435798, "learning_rate": 4.728167905797877e-06, "loss": 0.4674, "step": 252 }, { "epoch": 0.46083788706739526, "grad_norm": 1.2354902761906357, "learning_rate": 4.7049290360188875e-06, "loss": 0.4812, "step": 253 }, { "epoch": 0.46265938069216755, "grad_norm": 1.278394919844599, "learning_rate": 4.681665584047227e-06, "loss": 0.5462, "step": 254 }, { "epoch": 0.4644808743169399, "grad_norm": 1.2675137056539372, "learning_rate": 4.658378361122936e-06, "loss": 0.5193, "step": 255 }, { "epoch": 0.4663023679417122, "grad_norm": 1.1337542247765147, "learning_rate": 4.6350681793149884e-06, "loss": 0.4399, "step": 256 }, { "epoch": 0.4681238615664845, "grad_norm": 1.1926312684911093, "learning_rate": 4.611735851492984e-06, "loss": 0.392, "step": 257 }, { "epoch": 0.46994535519125685, "grad_norm": 1.5088135476737818, "learning_rate": 4.588382191298787e-06, "loss": 0.6285, "step": 258 }, { "epoch": 0.47176684881602915, "grad_norm": 1.2008807604194953, "learning_rate": 4.5650080131181675e-06, "loss": 0.6111, "step": 259 }, { "epoch": 0.47358834244080145, "grad_norm": 1.231409667649599, "learning_rate": 4.541614132052393e-06, "loss": 0.5028, "step": 260 }, { "epoch": 0.47540983606557374, "grad_norm": 1.0843389080843633, "learning_rate": 4.51820136388981e-06, "loss": 0.5245, "step": 261 }, { "epoch": 0.4772313296903461, "grad_norm": 1.3503009764858278, "learning_rate": 4.494770525077391e-06, "loss": 0.5069, "step": 262 }, { "epoch": 0.4790528233151184, "grad_norm": 1.2840200177200862, "learning_rate": 4.4713224326922655e-06, "loss": 0.5877, "step": 263 }, { "epoch": 0.4808743169398907, "grad_norm": 1.1859994564826624, "learning_rate": 4.447857904413231e-06, "loss": 0.5355, "step": 264 }, { "epoch": 0.48269581056466304, "grad_norm": 1.1755333908210361, "learning_rate": 4.424377758492232e-06, "loss": 0.5262, "step": 265 }, { "epoch": 0.48451730418943534, "grad_norm": 1.3928180479700631, "learning_rate": 4.40088281372583e-06, "loss": 0.5019, "step": 266 }, { "epoch": 0.48633879781420764, "grad_norm": 1.1971102520263333, "learning_rate": 4.377373889426649e-06, "loss": 0.5133, "step": 267 }, { "epoch": 0.48816029143898, "grad_norm": 1.3847287014410063, "learning_rate": 4.353851805394808e-06, "loss": 0.5929, "step": 268 }, { "epoch": 0.4899817850637523, "grad_norm": 1.3913642906155013, "learning_rate": 4.33031738188933e-06, "loss": 0.5482, "step": 269 }, { "epoch": 0.4918032786885246, "grad_norm": 1.2060593356882499, "learning_rate": 4.306771439599534e-06, "loss": 0.5505, "step": 270 }, { "epoch": 0.4936247723132969, "grad_norm": 1.1459886490704883, "learning_rate": 4.283214799616428e-06, "loss": 0.5472, "step": 271 }, { "epoch": 0.49544626593806923, "grad_norm": 1.388116383059367, "learning_rate": 4.259648283404062e-06, "loss": 0.5006, "step": 272 }, { "epoch": 0.4972677595628415, "grad_norm": 1.267084901831605, "learning_rate": 4.236072712770891e-06, "loss": 0.5578, "step": 273 }, { "epoch": 0.4990892531876138, "grad_norm": 1.2590176689522474, "learning_rate": 4.2124889098411175e-06, "loss": 0.4804, "step": 274 }, { "epoch": 0.5009107468123861, "grad_norm": 1.2991828184628833, "learning_rate": 4.1888976970260135e-06, "loss": 0.4865, "step": 275 }, { "epoch": 0.5027322404371585, "grad_norm": 1.3502024483330752, "learning_rate": 4.165299896995252e-06, "loss": 0.6084, "step": 276 }, { "epoch": 0.5045537340619308, "grad_norm": 1.24662926047003, "learning_rate": 4.141696332648216e-06, "loss": 0.5299, "step": 277 }, { "epoch": 0.5063752276867031, "grad_norm": 1.1169437339901658, "learning_rate": 4.118087827085294e-06, "loss": 0.4495, "step": 278 }, { "epoch": 0.5081967213114754, "grad_norm": 1.5004394388158784, "learning_rate": 4.094475203579191e-06, "loss": 0.5311, "step": 279 }, { "epoch": 0.5100182149362478, "grad_norm": 1.383813491474933, "learning_rate": 4.070859285546209e-06, "loss": 0.5176, "step": 280 }, { "epoch": 0.51183970856102, "grad_norm": 1.4019514520014746, "learning_rate": 4.047240896517539e-06, "loss": 0.5066, "step": 281 }, { "epoch": 0.5136612021857924, "grad_norm": 1.2338591970281994, "learning_rate": 4.023620860110533e-06, "loss": 0.5499, "step": 282 }, { "epoch": 0.5154826958105647, "grad_norm": 1.4230250812531824, "learning_rate": 4e-06, "loss": 0.5209, "step": 283 }, { "epoch": 0.517304189435337, "grad_norm": 1.3829946645846831, "learning_rate": 3.976379139889467e-06, "loss": 0.5384, "step": 284 }, { "epoch": 0.5191256830601093, "grad_norm": 1.3807362256783324, "learning_rate": 3.9527591034824616e-06, "loss": 0.4828, "step": 285 }, { "epoch": 0.5209471766848816, "grad_norm": 1.2837504835980802, "learning_rate": 3.929140714453791e-06, "loss": 0.5297, "step": 286 }, { "epoch": 0.5227686703096539, "grad_norm": 1.2883082935052266, "learning_rate": 3.9055247964208075e-06, "loss": 0.5485, "step": 287 }, { "epoch": 0.5245901639344263, "grad_norm": 1.2663249100766605, "learning_rate": 3.8819121729147055e-06, "loss": 0.4972, "step": 288 }, { "epoch": 0.5264116575591985, "grad_norm": 1.325952034342347, "learning_rate": 3.8583036673517845e-06, "loss": 0.529, "step": 289 }, { "epoch": 0.5282331511839709, "grad_norm": 1.4160215190484486, "learning_rate": 3.834700103004747e-06, "loss": 0.4542, "step": 290 }, { "epoch": 0.5300546448087432, "grad_norm": 1.2682053508757782, "learning_rate": 3.8111023029739866e-06, "loss": 0.4907, "step": 291 }, { "epoch": 0.5318761384335154, "grad_norm": 1.1893254878402797, "learning_rate": 3.787511090158884e-06, "loss": 0.5084, "step": 292 }, { "epoch": 0.5336976320582878, "grad_norm": 1.2969065452294093, "learning_rate": 3.763927287229109e-06, "loss": 0.4577, "step": 293 }, { "epoch": 0.5355191256830601, "grad_norm": 1.2651027758547007, "learning_rate": 3.740351716595939e-06, "loss": 0.6394, "step": 294 }, { "epoch": 0.5373406193078324, "grad_norm": 1.3693523974507902, "learning_rate": 3.7167852003835723e-06, "loss": 0.5101, "step": 295 }, { "epoch": 0.5391621129326047, "grad_norm": 1.2607338758277125, "learning_rate": 3.6932285604004656e-06, "loss": 0.5171, "step": 296 }, { "epoch": 0.5409836065573771, "grad_norm": 1.3133605575947254, "learning_rate": 3.669682618110671e-06, "loss": 0.5532, "step": 297 }, { "epoch": 0.5428051001821493, "grad_norm": 1.1776402028496125, "learning_rate": 3.646148194605191e-06, "loss": 0.488, "step": 298 }, { "epoch": 0.5446265938069217, "grad_norm": 1.3163668976765623, "learning_rate": 3.622626110573351e-06, "loss": 0.5082, "step": 299 }, { "epoch": 0.546448087431694, "grad_norm": 1.234697706107032, "learning_rate": 3.5991171862741713e-06, "loss": 0.6067, "step": 300 }, { "epoch": 0.5482695810564663, "grad_norm": 1.4127072359417963, "learning_rate": 3.575622241507768e-06, "loss": 0.5798, "step": 301 }, { "epoch": 0.5500910746812386, "grad_norm": 1.2117668842288387, "learning_rate": 3.5521420955867683e-06, "loss": 0.5147, "step": 302 }, { "epoch": 0.5519125683060109, "grad_norm": 1.2967372119013212, "learning_rate": 3.5286775673077332e-06, "loss": 0.5121, "step": 303 }, { "epoch": 0.5537340619307832, "grad_norm": 1.2224035940340945, "learning_rate": 3.505229474922609e-06, "loss": 0.5522, "step": 304 }, { "epoch": 0.5555555555555556, "grad_norm": 1.249619173678802, "learning_rate": 3.481798636110191e-06, "loss": 0.5408, "step": 305 }, { "epoch": 0.5573770491803278, "grad_norm": 1.3457532512289554, "learning_rate": 3.458385867947607e-06, "loss": 0.4521, "step": 306 }, { "epoch": 0.5591985428051002, "grad_norm": 1.360942386694437, "learning_rate": 3.434991986881833e-06, "loss": 0.493, "step": 307 }, { "epoch": 0.5610200364298725, "grad_norm": 1.3130428998562595, "learning_rate": 3.4116178087012136e-06, "loss": 0.5506, "step": 308 }, { "epoch": 0.5628415300546448, "grad_norm": 1.1190533809006658, "learning_rate": 3.388264148507016e-06, "loss": 0.5126, "step": 309 }, { "epoch": 0.5646630236794171, "grad_norm": 1.2987174512716828, "learning_rate": 3.3649318206850116e-06, "loss": 0.4693, "step": 310 }, { "epoch": 0.5664845173041895, "grad_norm": 1.199702517512821, "learning_rate": 3.3416216388770635e-06, "loss": 0.6163, "step": 311 }, { "epoch": 0.5683060109289617, "grad_norm": 1.4350182469129544, "learning_rate": 3.3183344159527736e-06, "loss": 0.4995, "step": 312 }, { "epoch": 0.5701275045537341, "grad_norm": 1.1958548262022524, "learning_rate": 3.2950709639811134e-06, "loss": 0.49, "step": 313 }, { "epoch": 0.5719489981785064, "grad_norm": 1.3945811458876904, "learning_rate": 3.271832094202123e-06, "loss": 0.5864, "step": 314 }, { "epoch": 0.5737704918032787, "grad_norm": 1.4302437878542158, "learning_rate": 3.2486186169986153e-06, "loss": 0.5064, "step": 315 }, { "epoch": 0.575591985428051, "grad_norm": 1.2460171889983833, "learning_rate": 3.2254313418679154e-06, "loss": 0.4993, "step": 316 }, { "epoch": 0.5774134790528234, "grad_norm": 1.3262935951428076, "learning_rate": 3.2022710773936304e-06, "loss": 0.4744, "step": 317 }, { "epoch": 0.5792349726775956, "grad_norm": 1.3389955195972856, "learning_rate": 3.1791386312174633e-06, "loss": 0.4928, "step": 318 }, { "epoch": 0.581056466302368, "grad_norm": 1.2940542358009448, "learning_rate": 3.1560348100110315e-06, "loss": 0.5806, "step": 319 }, { "epoch": 0.5828779599271403, "grad_norm": 1.3088940762820873, "learning_rate": 3.1329604194477535e-06, "loss": 0.4656, "step": 320 }, { "epoch": 0.5846994535519126, "grad_norm": 1.27093825835977, "learning_rate": 3.1099162641747427e-06, "loss": 0.4405, "step": 321 }, { "epoch": 0.5865209471766849, "grad_norm": 1.2054913034625938, "learning_rate": 3.0869031477847507e-06, "loss": 0.5713, "step": 322 }, { "epoch": 0.5883424408014571, "grad_norm": 1.2012296656668295, "learning_rate": 3.0639218727881508e-06, "loss": 0.5791, "step": 323 }, { "epoch": 0.5901639344262295, "grad_norm": 1.2982012651862458, "learning_rate": 3.04097324058494e-06, "loss": 0.5778, "step": 324 }, { "epoch": 0.5919854280510018, "grad_norm": 1.2933931880238856, "learning_rate": 3.0180580514368034e-06, "loss": 0.5219, "step": 325 }, { "epoch": 0.5938069216757741, "grad_norm": 1.4720885088406803, "learning_rate": 2.9951771044392066e-06, "loss": 0.5007, "step": 326 }, { "epoch": 0.5956284153005464, "grad_norm": 1.4987465964654607, "learning_rate": 2.972331197493523e-06, "loss": 0.4449, "step": 327 }, { "epoch": 0.5974499089253188, "grad_norm": 1.24446661500654, "learning_rate": 2.949521127279218e-06, "loss": 0.5865, "step": 328 }, { "epoch": 0.599271402550091, "grad_norm": 1.2063717392200657, "learning_rate": 2.926747689226062e-06, "loss": 0.5257, "step": 329 }, { "epoch": 0.6010928961748634, "grad_norm": 1.3076236400800156, "learning_rate": 2.9040116774863896e-06, "loss": 0.4262, "step": 330 }, { "epoch": 0.6029143897996357, "grad_norm": 1.486619018414579, "learning_rate": 2.881313884907416e-06, "loss": 0.5119, "step": 331 }, { "epoch": 0.604735883424408, "grad_norm": 1.4291961410481424, "learning_rate": 2.8586551030035797e-06, "loss": 0.5522, "step": 332 }, { "epoch": 0.6065573770491803, "grad_norm": 1.363647670787008, "learning_rate": 2.836036121928942e-06, "loss": 0.4966, "step": 333 }, { "epoch": 0.6083788706739527, "grad_norm": 1.3510346555153874, "learning_rate": 2.813457730449641e-06, "loss": 0.557, "step": 334 }, { "epoch": 0.6102003642987249, "grad_norm": 1.3056707121910645, "learning_rate": 2.790920715916372e-06, "loss": 0.4749, "step": 335 }, { "epoch": 0.6120218579234973, "grad_norm": 1.2674665608771611, "learning_rate": 2.7684258642369484e-06, "loss": 0.5118, "step": 336 }, { "epoch": 0.6138433515482696, "grad_norm": 1.1704712319064594, "learning_rate": 2.7459739598488762e-06, "loss": 0.5549, "step": 337 }, { "epoch": 0.6156648451730419, "grad_norm": 1.4094367691998289, "learning_rate": 2.723565785692013e-06, "loss": 0.5281, "step": 338 }, { "epoch": 0.6174863387978142, "grad_norm": 1.1692398204308077, "learning_rate": 2.701202123181266e-06, "loss": 0.5117, "step": 339 }, { "epoch": 0.6193078324225865, "grad_norm": 1.1555269052808808, "learning_rate": 2.6788837521793328e-06, "loss": 0.5099, "step": 340 }, { "epoch": 0.6211293260473588, "grad_norm": 1.2795356877186783, "learning_rate": 2.6566114509695096e-06, "loss": 0.5776, "step": 341 }, { "epoch": 0.6229508196721312, "grad_norm": 1.402365827706012, "learning_rate": 2.634385996228561e-06, "loss": 0.5233, "step": 342 }, { "epoch": 0.6247723132969034, "grad_norm": 1.1473627159160211, "learning_rate": 2.6122081629996195e-06, "loss": 0.6332, "step": 343 }, { "epoch": 0.6265938069216758, "grad_norm": 1.363153716598952, "learning_rate": 2.5900787246651715e-06, "loss": 0.5023, "step": 344 }, { "epoch": 0.6284153005464481, "grad_norm": 1.3207498781298728, "learning_rate": 2.567998452920081e-06, "loss": 0.5304, "step": 345 }, { "epoch": 0.6302367941712204, "grad_norm": 1.3224239362488786, "learning_rate": 2.5459681177446797e-06, "loss": 0.4926, "step": 346 }, { "epoch": 0.6320582877959927, "grad_norm": 1.1481892109560257, "learning_rate": 2.523988487377924e-06, "loss": 0.4879, "step": 347 }, { "epoch": 0.6338797814207651, "grad_norm": 1.3002026066911114, "learning_rate": 2.50206032829059e-06, "loss": 0.5203, "step": 348 }, { "epoch": 0.6357012750455373, "grad_norm": 1.278028126213566, "learning_rate": 2.4801844051585604e-06, "loss": 0.3582, "step": 349 }, { "epoch": 0.6375227686703097, "grad_norm": 1.2059444131681067, "learning_rate": 2.4583614808361508e-06, "loss": 0.6105, "step": 350 }, { "epoch": 0.639344262295082, "grad_norm": 1.216009601153044, "learning_rate": 2.4365923163295083e-06, "loss": 0.5276, "step": 351 }, { "epoch": 0.6411657559198543, "grad_norm": 1.3096172150636922, "learning_rate": 2.4148776707700775e-06, "loss": 0.5187, "step": 352 }, { "epoch": 0.6429872495446266, "grad_norm": 1.288189929013976, "learning_rate": 2.393218301388123e-06, "loss": 0.5577, "step": 353 }, { "epoch": 0.644808743169399, "grad_norm": 1.3155373684757863, "learning_rate": 2.3716149634863244e-06, "loss": 0.5057, "step": 354 }, { "epoch": 0.6466302367941712, "grad_norm": 1.3419874756397705, "learning_rate": 2.3500684104134433e-06, "loss": 0.4088, "step": 355 }, { "epoch": 0.6484517304189436, "grad_norm": 1.433794079225407, "learning_rate": 2.328579393538046e-06, "loss": 0.5522, "step": 356 }, { "epoch": 0.6502732240437158, "grad_norm": 1.4381038473896983, "learning_rate": 2.3071486622223e-06, "loss": 0.5745, "step": 357 }, { "epoch": 0.6520947176684881, "grad_norm": 1.2039623063737521, "learning_rate": 2.2857769637958554e-06, "loss": 0.4083, "step": 358 }, { "epoch": 0.6539162112932605, "grad_norm": 1.1858586411225656, "learning_rate": 2.2644650435297675e-06, "loss": 0.5221, "step": 359 }, { "epoch": 0.6557377049180327, "grad_norm": 1.2748116882040539, "learning_rate": 2.243213644610519e-06, "loss": 0.5085, "step": 360 }, { "epoch": 0.6575591985428051, "grad_norm": 1.1732780097848132, "learning_rate": 2.2220235081140985e-06, "loss": 0.4498, "step": 361 }, { "epoch": 0.6593806921675774, "grad_norm": 1.4343186803172785, "learning_rate": 2.2008953729801583e-06, "loss": 0.4823, "step": 362 }, { "epoch": 0.6612021857923497, "grad_norm": 1.496026602053126, "learning_rate": 2.1798299759862545e-06, "loss": 0.4367, "step": 363 }, { "epoch": 0.663023679417122, "grad_norm": 1.3025991152153253, "learning_rate": 2.158828051722137e-06, "loss": 0.571, "step": 364 }, { "epoch": 0.6648451730418944, "grad_norm": 1.1193584120775901, "learning_rate": 2.137890332564147e-06, "loss": 0.4189, "step": 365 }, { "epoch": 0.6666666666666666, "grad_norm": 1.2674961696906415, "learning_rate": 2.117017548649678e-06, "loss": 0.5501, "step": 366 }, { "epoch": 0.6666666666666666, "eval_accuracy": 0.813597109631812, "eval_accuracy_first_token": 0.9505597014925373, "eval_accuracy_first_token_all": 0.9743665728256038, "eval_accuracy_first_token_all_total": 6749, "eval_accuracy_first_token_calculate": 0.8409090909090909, "eval_accuracy_first_token_calculate_total": 44, "eval_accuracy_first_token_execute": 1.0, "eval_accuracy_first_token_execute_total": 202, "eval_accuracy_first_token_get": 0.912472647702407, "eval_accuracy_first_token_get_total": 457, "eval_accuracy_first_token_python": 0.871356783919598, "eval_accuracy_first_token_python_total": 995, "eval_loss": 0.5661724805831909, "eval_perplexity": 1.1117590100934498, "eval_runtime": 274.9553, "eval_samples_per_second": 2.524, "eval_steps_per_second": 0.633, "eval_total_number_first_token": 9648, "step": 366 }, { "epoch": 0.668488160291439, "grad_norm": 1.2382082369437504, "learning_rate": 2.0962104278517058e-06, "loss": 0.4206, "step": 367 }, { "epoch": 0.6703096539162113, "grad_norm": 1.3495235838757789, "learning_rate": 2.0754696957534105e-06, "loss": 0.5516, "step": 368 }, { "epoch": 0.6721311475409836, "grad_norm": 1.192568387623234, "learning_rate": 2.0547960756228746e-06, "loss": 0.4107, "step": 369 }, { "epoch": 0.6739526411657559, "grad_norm": 1.180512972457561, "learning_rate": 2.0341902883878626e-06, "loss": 0.486, "step": 370 }, { "epoch": 0.6757741347905283, "grad_norm": 1.4275500278475863, "learning_rate": 2.013653052610678e-06, "loss": 0.4959, "step": 371 }, { "epoch": 0.6775956284153005, "grad_norm": 1.2902691782551479, "learning_rate": 1.993185084463106e-06, "loss": 0.4299, "step": 372 }, { "epoch": 0.6794171220400729, "grad_norm": 1.1323245803277089, "learning_rate": 1.97278709770144e-06, "loss": 0.449, "step": 373 }, { "epoch": 0.6812386156648452, "grad_norm": 1.1729164825962006, "learning_rate": 1.952459803641597e-06, "loss": 0.5054, "step": 374 }, { "epoch": 0.6830601092896175, "grad_norm": 1.302619775593189, "learning_rate": 1.9322039111342977e-06, "loss": 0.5069, "step": 375 }, { "epoch": 0.6848816029143898, "grad_norm": 1.3225743591209795, "learning_rate": 1.912020126540366e-06, "loss": 0.4455, "step": 376 }, { "epoch": 0.6867030965391621, "grad_norm": 1.292024983046111, "learning_rate": 1.8919091537060847e-06, "loss": 0.5328, "step": 377 }, { "epoch": 0.6885245901639344, "grad_norm": 1.2278749426015196, "learning_rate": 1.8718716939386541e-06, "loss": 0.4648, "step": 378 }, { "epoch": 0.6903460837887068, "grad_norm": 1.2977606986012467, "learning_rate": 1.8519084459817362e-06, "loss": 0.4924, "step": 379 }, { "epoch": 0.692167577413479, "grad_norm": 1.227698057643303, "learning_rate": 1.83202010599109e-06, "loss": 0.4792, "step": 380 }, { "epoch": 0.6939890710382514, "grad_norm": 1.3176757819527871, "learning_rate": 1.8122073675102932e-06, "loss": 0.4661, "step": 381 }, { "epoch": 0.6958105646630237, "grad_norm": 1.60497524399408, "learning_rate": 1.792470921446557e-06, "loss": 0.4696, "step": 382 }, { "epoch": 0.697632058287796, "grad_norm": 1.3498126136904973, "learning_rate": 1.7728114560466324e-06, "loss": 0.5222, "step": 383 }, { "epoch": 0.6994535519125683, "grad_norm": 1.2558395717712512, "learning_rate": 1.753229656872815e-06, "loss": 0.4103, "step": 384 }, { "epoch": 0.7012750455373407, "grad_norm": 1.3697944306184828, "learning_rate": 1.7337262067790319e-06, "loss": 0.4544, "step": 385 }, { "epoch": 0.7030965391621129, "grad_norm": 1.368204116383326, "learning_rate": 1.7143017858870259e-06, "loss": 0.4975, "step": 386 }, { "epoch": 0.7049180327868853, "grad_norm": 1.2028196112608587, "learning_rate": 1.6949570715626532e-06, "loss": 0.5793, "step": 387 }, { "epoch": 0.7067395264116576, "grad_norm": 1.2193898504656977, "learning_rate": 1.675692738392247e-06, "loss": 0.4841, "step": 388 }, { "epoch": 0.7085610200364298, "grad_norm": 1.2121663225442056, "learning_rate": 1.6565094581591015e-06, "loss": 0.4891, "step": 389 }, { "epoch": 0.7103825136612022, "grad_norm": 1.2920888769630756, "learning_rate": 1.6374078998200424e-06, "loss": 0.5242, "step": 390 }, { "epoch": 0.7122040072859745, "grad_norm": 1.2474786830204843, "learning_rate": 1.6183887294820995e-06, "loss": 0.4941, "step": 391 }, { "epoch": 0.7140255009107468, "grad_norm": 1.2316800527163396, "learning_rate": 1.5994526103792852e-06, "loss": 0.4402, "step": 392 }, { "epoch": 0.7158469945355191, "grad_norm": 1.267209993054678, "learning_rate": 1.5806002028494509e-06, "loss": 0.4665, "step": 393 }, { "epoch": 0.7176684881602914, "grad_norm": 1.0939936334202054, "learning_rate": 1.5618321643112738e-06, "loss": 0.5385, "step": 394 }, { "epoch": 0.7194899817850637, "grad_norm": 1.2229947652634188, "learning_rate": 1.5431491492413286e-06, "loss": 0.563, "step": 395 }, { "epoch": 0.7213114754098361, "grad_norm": 1.3179598742616503, "learning_rate": 1.52455180915126e-06, "loss": 0.5043, "step": 396 }, { "epoch": 0.7231329690346083, "grad_norm": 1.177007091506137, "learning_rate": 1.506040792565066e-06, "loss": 0.5163, "step": 397 }, { "epoch": 0.7249544626593807, "grad_norm": 1.1922238886428833, "learning_rate": 1.487616744996484e-06, "loss": 0.4374, "step": 398 }, { "epoch": 0.726775956284153, "grad_norm": 1.434536851445206, "learning_rate": 1.4692803089264772e-06, "loss": 0.4976, "step": 399 }, { "epoch": 0.7285974499089253, "grad_norm": 1.1083154204777248, "learning_rate": 1.4510321237808377e-06, "loss": 0.4397, "step": 400 }, { "epoch": 0.7304189435336976, "grad_norm": 1.446108741355073, "learning_rate": 1.4328728259078746e-06, "loss": 0.4707, "step": 401 }, { "epoch": 0.73224043715847, "grad_norm": 1.3655307747592895, "learning_rate": 1.414803048556236e-06, "loss": 0.4509, "step": 402 }, { "epoch": 0.7340619307832422, "grad_norm": 1.2590237151009611, "learning_rate": 1.396823421852825e-06, "loss": 0.4137, "step": 403 }, { "epoch": 0.7358834244080146, "grad_norm": 1.3245324404535452, "learning_rate": 1.3789345727808207e-06, "loss": 0.5792, "step": 404 }, { "epoch": 0.7377049180327869, "grad_norm": 1.2386242097070082, "learning_rate": 1.3611371251578114e-06, "loss": 0.5098, "step": 405 }, { "epoch": 0.7395264116575592, "grad_norm": 1.1915406492227152, "learning_rate": 1.3434316996140553e-06, "loss": 0.5163, "step": 406 }, { "epoch": 0.7413479052823315, "grad_norm": 1.4026256649335254, "learning_rate": 1.3258189135708229e-06, "loss": 0.4404, "step": 407 }, { "epoch": 0.7431693989071039, "grad_norm": 1.2980107514185337, "learning_rate": 1.3082993812188735e-06, "loss": 0.54, "step": 408 }, { "epoch": 0.7449908925318761, "grad_norm": 1.2715578342804374, "learning_rate": 1.2908737134970364e-06, "loss": 0.4448, "step": 409 }, { "epoch": 0.7468123861566485, "grad_norm": 1.143848333980239, "learning_rate": 1.2735425180709039e-06, "loss": 0.4223, "step": 410 }, { "epoch": 0.7486338797814208, "grad_norm": 1.1255231193873407, "learning_rate": 1.2563063993116482e-06, "loss": 0.46, "step": 411 }, { "epoch": 0.7504553734061931, "grad_norm": 1.3419216896904982, "learning_rate": 1.239165958274933e-06, "loss": 0.5361, "step": 412 }, { "epoch": 0.7522768670309654, "grad_norm": 1.4023805964702172, "learning_rate": 1.2221217926799652e-06, "loss": 0.4397, "step": 413 }, { "epoch": 0.7540983606557377, "grad_norm": 1.330897728853757, "learning_rate": 1.2051744968886489e-06, "loss": 0.4957, "step": 414 }, { "epoch": 0.75591985428051, "grad_norm": 1.1838422813407694, "learning_rate": 1.1883246618848533e-06, "loss": 0.4462, "step": 415 }, { "epoch": 0.7577413479052824, "grad_norm": 1.1723617674919102, "learning_rate": 1.1715728752538101e-06, "loss": 0.451, "step": 416 }, { "epoch": 0.7595628415300546, "grad_norm": 1.206209172251872, "learning_rate": 1.1549197211616203e-06, "loss": 0.415, "step": 417 }, { "epoch": 0.761384335154827, "grad_norm": 1.3707028766377027, "learning_rate": 1.1383657803348835e-06, "loss": 0.4914, "step": 418 }, { "epoch": 0.7632058287795993, "grad_norm": 1.8225393046839982, "learning_rate": 1.1219116300404486e-06, "loss": 0.5696, "step": 419 }, { "epoch": 0.7650273224043715, "grad_norm": 1.2961370764387552, "learning_rate": 1.10555784406528e-06, "loss": 0.4906, "step": 420 }, { "epoch": 0.7668488160291439, "grad_norm": 1.273661302678146, "learning_rate": 1.089304992696455e-06, "loss": 0.4176, "step": 421 }, { "epoch": 0.7686703096539163, "grad_norm": 1.1358792116562773, "learning_rate": 1.0731536427012695e-06, "loss": 0.5669, "step": 422 }, { "epoch": 0.7704918032786885, "grad_norm": 1.1543809051499938, "learning_rate": 1.0571043573074736e-06, "loss": 0.4731, "step": 423 }, { "epoch": 0.7723132969034608, "grad_norm": 1.268995763936549, "learning_rate": 1.041157696183641e-06, "loss": 0.4944, "step": 424 }, { "epoch": 0.7741347905282332, "grad_norm": 1.3502204931718385, "learning_rate": 1.0253142154196415e-06, "loss": 0.5632, "step": 425 }, { "epoch": 0.7759562841530054, "grad_norm": 1.2861585316767994, "learning_rate": 1.0095744675072525e-06, "loss": 0.5321, "step": 426 }, { "epoch": 0.7777777777777778, "grad_norm": 1.2500169007804636, "learning_rate": 9.93939001320895e-07, "loss": 0.5018, "step": 427 }, { "epoch": 0.7795992714025501, "grad_norm": 1.1927946674927596, "learning_rate": 9.784083620984884e-07, "loss": 0.4495, "step": 428 }, { "epoch": 0.7814207650273224, "grad_norm": 1.211453529275201, "learning_rate": 9.62983091422446e-07, "loss": 0.459, "step": 429 }, { "epoch": 0.7832422586520947, "grad_norm": 1.191077248020614, "learning_rate": 9.476637272007746e-07, "loss": 0.5285, "step": 430 }, { "epoch": 0.785063752276867, "grad_norm": 1.5578769325659387, "learning_rate": 9.324508036483303e-07, "loss": 0.4826, "step": 431 }, { "epoch": 0.7868852459016393, "grad_norm": 1.143092830687595, "learning_rate": 9.173448512681848e-07, "loss": 0.4689, "step": 432 }, { "epoch": 0.7887067395264117, "grad_norm": 1.3178766788423166, "learning_rate": 9.023463968331238e-07, "loss": 0.423, "step": 433 }, { "epoch": 0.7905282331511839, "grad_norm": 1.270598374226235, "learning_rate": 8.874559633672754e-07, "loss": 0.5087, "step": 434 }, { "epoch": 0.7923497267759563, "grad_norm": 1.3159704576848927, "learning_rate": 8.726740701278808e-07, "loss": 0.4658, "step": 435 }, { "epoch": 0.7941712204007286, "grad_norm": 1.308828411323957, "learning_rate": 8.580012325871773e-07, "loss": 0.4874, "step": 436 }, { "epoch": 0.7959927140255009, "grad_norm": 1.1534207003253993, "learning_rate": 8.434379624144261e-07, "loss": 0.559, "step": 437 }, { "epoch": 0.7978142076502732, "grad_norm": 1.2149317958628532, "learning_rate": 8.289847674580702e-07, "loss": 0.4374, "step": 438 }, { "epoch": 0.7996357012750456, "grad_norm": 1.7952195261148596, "learning_rate": 8.146421517280226e-07, "loss": 0.4826, "step": 439 }, { "epoch": 0.8014571948998178, "grad_norm": 1.204321359741705, "learning_rate": 8.004106153780967e-07, "loss": 0.5445, "step": 440 }, { "epoch": 0.8032786885245902, "grad_norm": 1.3038648482659696, "learning_rate": 7.862906546885559e-07, "loss": 0.4759, "step": 441 }, { "epoch": 0.8051001821493625, "grad_norm": 1.3025645847276193, "learning_rate": 7.722827620488108e-07, "loss": 0.4493, "step": 442 }, { "epoch": 0.8069216757741348, "grad_norm": 1.3302596060388403, "learning_rate": 7.583874259402545e-07, "loss": 0.5191, "step": 443 }, { "epoch": 0.8087431693989071, "grad_norm": 1.2390558736887658, "learning_rate": 7.446051309192203e-07, "loss": 0.5035, "step": 444 }, { "epoch": 0.8105646630236795, "grad_norm": 1.2987933255329953, "learning_rate": 7.30936357600088e-07, "loss": 0.5051, "step": 445 }, { "epoch": 0.8123861566484517, "grad_norm": 1.3795445751301032, "learning_rate": 7.173815826385246e-07, "loss": 0.4551, "step": 446 }, { "epoch": 0.8142076502732241, "grad_norm": 1.2599528493900025, "learning_rate": 7.039412787148586e-07, "loss": 0.4923, "step": 447 }, { "epoch": 0.8160291438979964, "grad_norm": 1.1755103488000105, "learning_rate": 6.906159145176049e-07, "loss": 0.5205, "step": 448 }, { "epoch": 0.8178506375227687, "grad_norm": 1.301352850179719, "learning_rate": 6.774059547271087e-07, "loss": 0.516, "step": 449 }, { "epoch": 0.819672131147541, "grad_norm": 1.2962192786547895, "learning_rate": 6.643118599993518e-07, "loss": 0.4872, "step": 450 }, { "epoch": 0.8214936247723132, "grad_norm": 1.5585419654309247, "learning_rate": 6.513340869498858e-07, "loss": 0.5804, "step": 451 }, { "epoch": 0.8233151183970856, "grad_norm": 1.387119348514666, "learning_rate": 6.384730881379048e-07, "loss": 0.5619, "step": 452 }, { "epoch": 0.825136612021858, "grad_norm": 1.4119192925549904, "learning_rate": 6.257293120504692e-07, "loss": 0.631, "step": 453 }, { "epoch": 0.8269581056466302, "grad_norm": 1.3308860456915488, "learning_rate": 6.131032030868635e-07, "loss": 0.485, "step": 454 }, { "epoch": 0.8287795992714025, "grad_norm": 1.4307724793695609, "learning_rate": 6.005952015430993e-07, "loss": 0.5642, "step": 455 }, { "epoch": 0.8306010928961749, "grad_norm": 1.2106456273020985, "learning_rate": 5.882057435965619e-07, "loss": 0.4985, "step": 456 }, { "epoch": 0.8324225865209471, "grad_norm": 1.3115620293405426, "learning_rate": 5.759352612907999e-07, "loss": 0.5307, "step": 457 }, { "epoch": 0.8342440801457195, "grad_norm": 1.248365530305497, "learning_rate": 5.637841825204588e-07, "loss": 0.5194, "step": 458 }, { "epoch": 0.8360655737704918, "grad_norm": 1.2238266091876802, "learning_rate": 5.517529310163627e-07, "loss": 0.5482, "step": 459 }, { "epoch": 0.8378870673952641, "grad_norm": 1.4210179624401589, "learning_rate": 5.398419263307281e-07, "loss": 0.4293, "step": 460 }, { "epoch": 0.8397085610200364, "grad_norm": 1.2412519735070302, "learning_rate": 5.280515838225477e-07, "loss": 0.4852, "step": 461 }, { "epoch": 0.8415300546448088, "grad_norm": 1.2092346961035594, "learning_rate": 5.163823146430944e-07, "loss": 0.5908, "step": 462 }, { "epoch": 0.843351548269581, "grad_norm": 1.3880323085259179, "learning_rate": 5.048345257215892e-07, "loss": 0.5084, "step": 463 }, { "epoch": 0.8451730418943534, "grad_norm": 1.2676463291192996, "learning_rate": 4.934086197510088e-07, "loss": 0.5547, "step": 464 }, { "epoch": 0.8469945355191257, "grad_norm": 1.2609901690813994, "learning_rate": 4.821049951740441e-07, "loss": 0.5434, "step": 465 }, { "epoch": 0.848816029143898, "grad_norm": 1.3054902000162272, "learning_rate": 4.7092404616920547e-07, "loss": 0.4394, "step": 466 }, { "epoch": 0.8506375227686703, "grad_norm": 1.323905342489299, "learning_rate": 4.59866162637077e-07, "loss": 0.5067, "step": 467 }, { "epoch": 0.8524590163934426, "grad_norm": 1.4483470172100243, "learning_rate": 4.4893173018671816e-07, "loss": 0.4553, "step": 468 }, { "epoch": 0.8542805100182149, "grad_norm": 1.2403981869880174, "learning_rate": 4.3812113012222164e-07, "loss": 0.5209, "step": 469 }, { "epoch": 0.8561020036429873, "grad_norm": 1.8179855460416823, "learning_rate": 4.2743473942941177e-07, "loss": 0.4736, "step": 470 }, { "epoch": 0.8579234972677595, "grad_norm": 1.2339086882585282, "learning_rate": 4.168729307626977e-07, "loss": 0.5098, "step": 471 }, { "epoch": 0.8597449908925319, "grad_norm": 1.276189228460651, "learning_rate": 4.0643607243208455e-07, "loss": 0.4989, "step": 472 }, { "epoch": 0.8615664845173042, "grad_norm": 1.2203026643363581, "learning_rate": 3.9612452839032384e-07, "loss": 0.4462, "step": 473 }, { "epoch": 0.8633879781420765, "grad_norm": 1.242544747026145, "learning_rate": 3.859386582202231e-07, "loss": 0.5232, "step": 474 }, { "epoch": 0.8652094717668488, "grad_norm": 1.4374137888249117, "learning_rate": 3.758788171221079e-07, "loss": 0.5248, "step": 475 }, { "epoch": 0.8670309653916212, "grad_norm": 1.1588668244652476, "learning_rate": 3.659453559014345e-07, "loss": 0.4631, "step": 476 }, { "epoch": 0.8688524590163934, "grad_norm": 1.3937157924335801, "learning_rate": 3.561386209565582e-07, "loss": 0.4675, "step": 477 }, { "epoch": 0.8706739526411658, "grad_norm": 1.376102069963406, "learning_rate": 3.464589542666485e-07, "loss": 0.458, "step": 478 }, { "epoch": 0.8724954462659381, "grad_norm": 1.2706836490870992, "learning_rate": 3.3690669337976996e-07, "loss": 0.4851, "step": 479 }, { "epoch": 0.8743169398907104, "grad_norm": 1.221114304425684, "learning_rate": 3.2748217140111e-07, "loss": 0.5032, "step": 480 }, { "epoch": 0.8761384335154827, "grad_norm": 1.3883301701023747, "learning_rate": 3.1818571698135976e-07, "loss": 0.4749, "step": 481 }, { "epoch": 0.8779599271402551, "grad_norm": 1.1787371549756893, "learning_rate": 3.0901765430525337e-07, "loss": 0.5106, "step": 482 }, { "epoch": 0.8797814207650273, "grad_norm": 1.346207438421883, "learning_rate": 2.9997830308027003e-07, "loss": 0.4383, "step": 483 }, { "epoch": 0.8816029143897997, "grad_norm": 1.348993463677863, "learning_rate": 2.9106797852547483e-07, "loss": 0.4673, "step": 484 }, { "epoch": 0.8834244080145719, "grad_norm": 1.201847671893057, "learning_rate": 2.8228699136053726e-07, "loss": 0.5759, "step": 485 }, { "epoch": 0.8852459016393442, "grad_norm": 1.340461205562142, "learning_rate": 2.7363564779488446e-07, "loss": 0.4465, "step": 486 }, { "epoch": 0.8870673952641166, "grad_norm": 1.1752701027649086, "learning_rate": 2.6511424951703244e-07, "loss": 0.4118, "step": 487 }, { "epoch": 0.8888888888888888, "grad_norm": 1.2789959623625877, "learning_rate": 2.567230936840632e-07, "loss": 0.5079, "step": 488 }, { "epoch": 0.8907103825136612, "grad_norm": 1.2640082921345366, "learning_rate": 2.4846247291125897e-07, "loss": 0.4711, "step": 489 }, { "epoch": 0.8925318761384335, "grad_norm": 1.1282519699248084, "learning_rate": 2.4033267526190057e-07, "loss": 0.5091, "step": 490 }, { "epoch": 0.8943533697632058, "grad_norm": 1.2710706301896888, "learning_rate": 2.323339842372234e-07, "loss": 0.4149, "step": 491 }, { "epoch": 0.8961748633879781, "grad_norm": 1.4092310204502283, "learning_rate": 2.2446667876652968e-07, "loss": 0.5713, "step": 492 }, { "epoch": 0.8979963570127505, "grad_norm": 1.1624712007482816, "learning_rate": 2.1673103319746146e-07, "loss": 0.4281, "step": 493 }, { "epoch": 0.8998178506375227, "grad_norm": 1.2035233792963602, "learning_rate": 2.0912731728643362e-07, "loss": 0.4497, "step": 494 }, { "epoch": 0.9016393442622951, "grad_norm": 1.2087725899212483, "learning_rate": 2.0165579618922757e-07, "loss": 0.4554, "step": 495 }, { "epoch": 0.9034608378870674, "grad_norm": 1.3284543377876135, "learning_rate": 1.943167304517459e-07, "loss": 0.4693, "step": 496 }, { "epoch": 0.9052823315118397, "grad_norm": 1.4011820069689847, "learning_rate": 1.871103760009234e-07, "loss": 0.5348, "step": 497 }, { "epoch": 0.907103825136612, "grad_norm": 1.2193252171112197, "learning_rate": 1.8003698413580427e-07, "loss": 0.4572, "step": 498 }, { "epoch": 0.9089253187613844, "grad_norm": 1.1222006814434855, "learning_rate": 1.7309680151878126e-07, "loss": 0.5136, "step": 499 }, { "epoch": 0.9107468123861566, "grad_norm": 1.2891850954780644, "learning_rate": 1.6629007016698916e-07, "loss": 0.4979, "step": 500 }, { "epoch": 0.912568306010929, "grad_norm": 1.3322769925178861, "learning_rate": 1.5961702744386973e-07, "loss": 0.5423, "step": 501 }, { "epoch": 0.9143897996357013, "grad_norm": 1.160971888154378, "learning_rate": 1.5307790605089045e-07, "loss": 0.4617, "step": 502 }, { "epoch": 0.9162112932604736, "grad_norm": 1.364660127297357, "learning_rate": 1.4667293401943393e-07, "loss": 0.4879, "step": 503 }, { "epoch": 0.9180327868852459, "grad_norm": 1.220502331061608, "learning_rate": 1.404023347028418e-07, "loss": 0.4495, "step": 504 }, { "epoch": 0.9198542805100182, "grad_norm": 1.5127357404088118, "learning_rate": 1.342663267686297e-07, "loss": 0.4893, "step": 505 }, { "epoch": 0.9216757741347905, "grad_norm": 1.0948101040744869, "learning_rate": 1.2826512419085922e-07, "loss": 0.4689, "step": 506 }, { "epoch": 0.9234972677595629, "grad_norm": 1.2102894214186626, "learning_rate": 1.223989362426785e-07, "loss": 0.4728, "step": 507 }, { "epoch": 0.9253187613843351, "grad_norm": 1.2964990563502878, "learning_rate": 1.1666796748902142e-07, "loss": 0.4591, "step": 508 }, { "epoch": 0.9271402550091075, "grad_norm": 1.3179562572508972, "learning_rate": 1.1107241777947774e-07, "loss": 0.5085, "step": 509 }, { "epoch": 0.9289617486338798, "grad_norm": 1.124362582099017, "learning_rate": 1.0561248224132091e-07, "loss": 0.43, "step": 510 }, { "epoch": 0.930783242258652, "grad_norm": 1.263872009096953, "learning_rate": 1.0028835127270552e-07, "loss": 0.4993, "step": 511 }, { "epoch": 0.9326047358834244, "grad_norm": 1.2997351666975574, "learning_rate": 9.510021053602679e-08, "loss": 0.4404, "step": 512 }, { "epoch": 0.9344262295081968, "grad_norm": 1.329253107069971, "learning_rate": 9.004824095144581e-08, "loss": 0.5285, "step": 513 }, { "epoch": 0.936247723132969, "grad_norm": 1.0886350408667627, "learning_rate": 8.513261869058209e-08, "loss": 0.4606, "step": 514 }, { "epoch": 0.9380692167577414, "grad_norm": 1.2876516793995008, "learning_rate": 8.035351517036914e-08, "loss": 0.4777, "step": 515 }, { "epoch": 0.9398907103825137, "grad_norm": 1.2692698581081259, "learning_rate": 7.571109704707623e-08, "loss": 0.4554, "step": 516 }, { "epoch": 0.941712204007286, "grad_norm": 1.2791868173597516, "learning_rate": 7.120552621049825e-08, "loss": 0.4488, "step": 517 }, { "epoch": 0.9435336976320583, "grad_norm": 1.208557284284356, "learning_rate": 6.68369597783096e-08, "loss": 0.4034, "step": 518 }, { "epoch": 0.9453551912568307, "grad_norm": 1.2963091541864111, "learning_rate": 6.260555009058288e-08, "loss": 0.4838, "step": 519 }, { "epoch": 0.9471766848816029, "grad_norm": 1.3727003252206291, "learning_rate": 5.851144470448144e-08, "loss": 0.4996, "step": 520 }, { "epoch": 0.9489981785063752, "grad_norm": 1.3210931262200563, "learning_rate": 5.455478638911071e-08, "loss": 0.5332, "step": 521 }, { "epoch": 0.9508196721311475, "grad_norm": 1.2713274273103237, "learning_rate": 5.073571312053815e-08, "loss": 0.4783, "step": 522 }, { "epoch": 0.9526411657559198, "grad_norm": 1.2497269961588167, "learning_rate": 4.705435807698555e-08, "loss": 0.4485, "step": 523 }, { "epoch": 0.9544626593806922, "grad_norm": 1.2309482179042086, "learning_rate": 4.351084963418117e-08, "loss": 0.4331, "step": 524 }, { "epoch": 0.9562841530054644, "grad_norm": 1.1950945289140942, "learning_rate": 4.010531136088691e-08, "loss": 0.4604, "step": 525 }, { "epoch": 0.9581056466302368, "grad_norm": 1.2931423414382452, "learning_rate": 3.683786201458439e-08, "loss": 0.3767, "step": 526 }, { "epoch": 0.9599271402550091, "grad_norm": 1.389899139107159, "learning_rate": 3.370861553733784e-08, "loss": 0.4897, "step": 527 }, { "epoch": 0.9617486338797814, "grad_norm": 1.362732666201115, "learning_rate": 3.071768105181993e-08, "loss": 0.499, "step": 528 }, { "epoch": 0.9635701275045537, "grad_norm": 1.1771528381852838, "learning_rate": 2.786516285750373e-08, "loss": 0.5627, "step": 529 }, { "epoch": 0.9653916211293261, "grad_norm": 1.2404159607795244, "learning_rate": 2.5151160427029582e-08, "loss": 0.4175, "step": 530 }, { "epoch": 0.9672131147540983, "grad_norm": 1.3555377819209309, "learning_rate": 2.2575768402733232e-08, "loss": 0.4961, "step": 531 }, { "epoch": 0.9690346083788707, "grad_norm": 1.4576298033097894, "learning_rate": 2.013907659334624e-08, "loss": 0.5458, "step": 532 }, { "epoch": 0.970856102003643, "grad_norm": 1.4545178388397544, "learning_rate": 1.7841169970866042e-08, "loss": 0.557, "step": 533 }, { "epoch": 0.9726775956284153, "grad_norm": 1.2516530596811049, "learning_rate": 1.5682128667589e-08, "loss": 0.5176, "step": 534 }, { "epoch": 0.9744990892531876, "grad_norm": 1.395868262099181, "learning_rate": 1.3662027973320612e-08, "loss": 0.6025, "step": 535 }, { "epoch": 0.97632058287796, "grad_norm": 1.228580089232679, "learning_rate": 1.1780938332746515e-08, "loss": 0.4352, "step": 536 }, { "epoch": 0.9781420765027322, "grad_norm": 1.307882851191111, "learning_rate": 1.0038925342977122e-08, "loss": 0.4912, "step": 537 }, { "epoch": 0.9799635701275046, "grad_norm": 1.412138408392618, "learning_rate": 8.43604975126011e-09, "loss": 0.5536, "step": 538 }, { "epoch": 0.9817850637522769, "grad_norm": 1.3540327476191583, "learning_rate": 6.972367452863004e-09, "loss": 0.4909, "step": 539 }, { "epoch": 0.9836065573770492, "grad_norm": 1.2531752735063777, "learning_rate": 5.647929489122738e-09, "loss": 0.4375, "step": 540 }, { "epoch": 0.9854280510018215, "grad_norm": 1.2878472135871981, "learning_rate": 4.462782045664859e-09, "loss": 0.4857, "step": 541 }, { "epoch": 0.9872495446265938, "grad_norm": 1.2374394809536116, "learning_rate": 3.4169664507959216e-09, "loss": 0.4435, "step": 542 }, { "epoch": 0.9890710382513661, "grad_norm": 1.3708150206530674, "learning_rate": 2.5105191740597553e-09, "loss": 0.5403, "step": 543 }, { "epoch": 0.9908925318761385, "grad_norm": 1.1853808948747229, "learning_rate": 1.7434718249664803e-09, "loss": 0.4532, "step": 544 }, { "epoch": 0.9927140255009107, "grad_norm": 1.190207243267208, "learning_rate": 1.1158511518902791e-09, "loss": 0.4696, "step": 545 }, { "epoch": 0.994535519125683, "grad_norm": 1.2336375950670702, "learning_rate": 6.276790411372524e-10, "loss": 0.4729, "step": 546 }, { "epoch": 0.9963570127504554, "grad_norm": 1.1739246355744901, "learning_rate": 2.789725161806977e-10, "loss": 0.5413, "step": 547 }, { "epoch": 0.9981785063752276, "grad_norm": 1.2297320117715114, "learning_rate": 6.974373706869486e-11, "loss": 0.4542, "step": 548 }, { "epoch": 1.0, "grad_norm": 1.2871587171589334, "learning_rate": 0.0, "loss": 0.4884, "step": 549 }, { "epoch": 1.0, "eval_accuracy": 0.8146578886019198, "eval_accuracy_first_token": 0.9528399668325042, "eval_accuracy_first_token_all": 0.9712550007408505, "eval_accuracy_first_token_all_total": 6749, "eval_accuracy_first_token_calculate": 0.8636363636363636, "eval_accuracy_first_token_calculate_total": 44, "eval_accuracy_first_token_execute": 1.0, "eval_accuracy_first_token_execute_total": 202, "eval_accuracy_first_token_get": 0.9343544857768052, "eval_accuracy_first_token_get_total": 457, "eval_accuracy_first_token_python": 0.8753768844221106, "eval_accuracy_first_token_python_total": 995, "eval_loss": 0.5619608163833618, "eval_perplexity": 1.1109600199882654, "eval_runtime": 283.0681, "eval_samples_per_second": 2.452, "eval_steps_per_second": 0.615, "eval_total_number_first_token": 9648, "step": 549 }, { "epoch": 1.0, "step": 549, "total_flos": 229794172108800.0, "train_loss": 0.5332211447652355, "train_runtime": 12454.5665, "train_samples_per_second": 0.705, "train_steps_per_second": 0.044 } ], "logging_steps": 1.0, "max_steps": 549, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 229794172108800.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }