diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,6111 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 4325, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00023121387283236994, + "grad_norm": 0.6965160472814916, + "learning_rate": 4.6189376443418015e-07, + "loss": 1.086, + "step": 1 + }, + { + "epoch": 0.0011560693641618498, + "grad_norm": 0.6835787858040867, + "learning_rate": 2.309468822170901e-06, + "loss": 1.128, + "step": 5 + }, + { + "epoch": 0.0023121387283236996, + "grad_norm": 0.6184758767742636, + "learning_rate": 4.618937644341802e-06, + "loss": 1.1188, + "step": 10 + }, + { + "epoch": 0.003468208092485549, + "grad_norm": 0.6078620387437725, + "learning_rate": 6.928406466512702e-06, + "loss": 1.1243, + "step": 15 + }, + { + "epoch": 0.004624277456647399, + "grad_norm": 0.5222130979899404, + "learning_rate": 9.237875288683604e-06, + "loss": 1.1013, + "step": 20 + }, + { + "epoch": 0.005780346820809248, + "grad_norm": 0.4982345960164489, + "learning_rate": 1.1547344110854504e-05, + "loss": 1.0928, + "step": 25 + }, + { + "epoch": 0.006936416184971098, + "grad_norm": 0.40171891380185876, + "learning_rate": 1.3856812933025404e-05, + "loss": 1.0736, + "step": 30 + }, + { + "epoch": 0.008092485549132947, + "grad_norm": 0.41106693801534305, + "learning_rate": 1.6166281755196306e-05, + "loss": 1.0672, + "step": 35 + }, + { + "epoch": 0.009248554913294798, + "grad_norm": 0.4109497681037648, + "learning_rate": 1.8475750577367208e-05, + "loss": 1.0258, + "step": 40 + }, + { + "epoch": 0.010404624277456647, + "grad_norm": 0.379459396453402, + "learning_rate": 2.0785219399538107e-05, + "loss": 1.0102, + "step": 45 + }, + { + "epoch": 0.011560693641618497, + "grad_norm": 0.3721675211937845, + "learning_rate": 2.309468822170901e-05, + "loss": 1.0234, + "step": 50 + }, + { + "epoch": 0.012716763005780347, + "grad_norm": 0.373222147309766, + "learning_rate": 2.540415704387991e-05, + "loss": 0.992, + "step": 55 + }, + { + "epoch": 0.013872832369942197, + "grad_norm": 0.3702598509393352, + "learning_rate": 2.771362586605081e-05, + "loss": 1.0326, + "step": 60 + }, + { + "epoch": 0.015028901734104046, + "grad_norm": 0.35490433661120324, + "learning_rate": 3.0023094688221707e-05, + "loss": 1.0246, + "step": 65 + }, + { + "epoch": 0.016184971098265895, + "grad_norm": 0.34072939280119285, + "learning_rate": 3.233256351039261e-05, + "loss": 1.0115, + "step": 70 + }, + { + "epoch": 0.017341040462427744, + "grad_norm": 0.3831809542290894, + "learning_rate": 3.464203233256351e-05, + "loss": 1.0331, + "step": 75 + }, + { + "epoch": 0.018497109826589597, + "grad_norm": 0.3623596879709559, + "learning_rate": 3.6951501154734416e-05, + "loss": 1.0022, + "step": 80 + }, + { + "epoch": 0.019653179190751446, + "grad_norm": 0.3931577119185935, + "learning_rate": 3.9260969976905315e-05, + "loss": 0.9997, + "step": 85 + }, + { + "epoch": 0.020809248554913295, + "grad_norm": 0.4109841798218973, + "learning_rate": 4.1570438799076213e-05, + "loss": 0.9914, + "step": 90 + }, + { + "epoch": 0.021965317919075144, + "grad_norm": 0.40213543229926174, + "learning_rate": 4.387990762124711e-05, + "loss": 0.9643, + "step": 95 + }, + { + "epoch": 0.023121387283236993, + "grad_norm": 0.3657003831875091, + "learning_rate": 4.618937644341802e-05, + "loss": 0.9957, + "step": 100 + }, + { + "epoch": 0.024277456647398842, + "grad_norm": 0.38609844859658143, + "learning_rate": 4.8498845265588916e-05, + "loss": 1.0111, + "step": 105 + }, + { + "epoch": 0.025433526011560695, + "grad_norm": 0.3583592530014769, + "learning_rate": 5.080831408775982e-05, + "loss": 0.9963, + "step": 110 + }, + { + "epoch": 0.026589595375722544, + "grad_norm": 0.3554977804638662, + "learning_rate": 5.311778290993071e-05, + "loss": 0.9817, + "step": 115 + }, + { + "epoch": 0.027745664739884393, + "grad_norm": 0.3840657570276599, + "learning_rate": 5.542725173210162e-05, + "loss": 1.0232, + "step": 120 + }, + { + "epoch": 0.028901734104046242, + "grad_norm": 0.3774759292695808, + "learning_rate": 5.7736720554272516e-05, + "loss": 1.0139, + "step": 125 + }, + { + "epoch": 0.03005780346820809, + "grad_norm": 0.360754142577289, + "learning_rate": 6.0046189376443415e-05, + "loss": 0.9982, + "step": 130 + }, + { + "epoch": 0.03121387283236994, + "grad_norm": 0.3811963374124325, + "learning_rate": 6.235565819861431e-05, + "loss": 1.0088, + "step": 135 + }, + { + "epoch": 0.03236994219653179, + "grad_norm": 0.3667939507833288, + "learning_rate": 6.466512702078523e-05, + "loss": 0.9887, + "step": 140 + }, + { + "epoch": 0.03352601156069364, + "grad_norm": 0.3988058302280759, + "learning_rate": 6.697459584295612e-05, + "loss": 0.9513, + "step": 145 + }, + { + "epoch": 0.03468208092485549, + "grad_norm": 0.3603023427371918, + "learning_rate": 6.928406466512702e-05, + "loss": 1.0239, + "step": 150 + }, + { + "epoch": 0.035838150289017344, + "grad_norm": 0.321539455275297, + "learning_rate": 7.159353348729792e-05, + "loss": 0.9821, + "step": 155 + }, + { + "epoch": 0.03699421965317919, + "grad_norm": 0.3548185437448054, + "learning_rate": 7.390300230946883e-05, + "loss": 1.0028, + "step": 160 + }, + { + "epoch": 0.03815028901734104, + "grad_norm": 0.33174112797447736, + "learning_rate": 7.621247113163973e-05, + "loss": 0.9852, + "step": 165 + }, + { + "epoch": 0.03930635838150289, + "grad_norm": 0.3428952862407286, + "learning_rate": 7.852193995381063e-05, + "loss": 1.0035, + "step": 170 + }, + { + "epoch": 0.04046242774566474, + "grad_norm": 0.3164096194909282, + "learning_rate": 8.083140877598153e-05, + "loss": 0.9594, + "step": 175 + }, + { + "epoch": 0.04161849710982659, + "grad_norm": 0.3210772711947753, + "learning_rate": 8.314087759815243e-05, + "loss": 0.9751, + "step": 180 + }, + { + "epoch": 0.04277456647398844, + "grad_norm": 0.31913209612903376, + "learning_rate": 8.545034642032334e-05, + "loss": 0.9816, + "step": 185 + }, + { + "epoch": 0.04393063583815029, + "grad_norm": 0.32476583892299626, + "learning_rate": 8.775981524249422e-05, + "loss": 1.0382, + "step": 190 + }, + { + "epoch": 0.04508670520231214, + "grad_norm": 0.33804033601960015, + "learning_rate": 9.006928406466512e-05, + "loss": 1.0278, + "step": 195 + }, + { + "epoch": 0.046242774566473986, + "grad_norm": 0.3034185595284857, + "learning_rate": 9.237875288683603e-05, + "loss": 0.9209, + "step": 200 + }, + { + "epoch": 0.047398843930635835, + "grad_norm": 0.30966667195251285, + "learning_rate": 9.468822170900693e-05, + "loss": 0.9929, + "step": 205 + }, + { + "epoch": 0.048554913294797684, + "grad_norm": 0.3110271670167101, + "learning_rate": 9.699769053117783e-05, + "loss": 0.999, + "step": 210 + }, + { + "epoch": 0.04971098265895954, + "grad_norm": 0.3168339881753396, + "learning_rate": 9.930715935334873e-05, + "loss": 0.9801, + "step": 215 + }, + { + "epoch": 0.05086705202312139, + "grad_norm": 0.2980534767723837, + "learning_rate": 0.00010161662817551964, + "loss": 0.964, + "step": 220 + }, + { + "epoch": 0.05202312138728324, + "grad_norm": 0.29439975733255125, + "learning_rate": 0.00010392609699769054, + "loss": 1.0141, + "step": 225 + }, + { + "epoch": 0.05317919075144509, + "grad_norm": 0.2939300503984728, + "learning_rate": 0.00010623556581986143, + "loss": 1.029, + "step": 230 + }, + { + "epoch": 0.05433526011560694, + "grad_norm": 0.2921065506824694, + "learning_rate": 0.00010854503464203234, + "loss": 1.0008, + "step": 235 + }, + { + "epoch": 0.055491329479768786, + "grad_norm": 0.29281624873588324, + "learning_rate": 0.00011085450346420324, + "loss": 1.0286, + "step": 240 + }, + { + "epoch": 0.056647398843930635, + "grad_norm": 0.30008365055838865, + "learning_rate": 0.00011316397228637415, + "loss": 1.0009, + "step": 245 + }, + { + "epoch": 0.057803468208092484, + "grad_norm": 0.2821304520615669, + "learning_rate": 0.00011547344110854503, + "loss": 1.0172, + "step": 250 + }, + { + "epoch": 0.058959537572254334, + "grad_norm": 0.28925572801814625, + "learning_rate": 0.00011778290993071594, + "loss": 1.0106, + "step": 255 + }, + { + "epoch": 0.06011560693641618, + "grad_norm": 0.2979979142081887, + "learning_rate": 0.00012009237875288683, + "loss": 0.9851, + "step": 260 + }, + { + "epoch": 0.06127167630057803, + "grad_norm": 0.27608874909832487, + "learning_rate": 0.00012240184757505776, + "loss": 0.9386, + "step": 265 + }, + { + "epoch": 0.06242774566473988, + "grad_norm": 0.270653606190121, + "learning_rate": 0.00012471131639722863, + "loss": 0.9823, + "step": 270 + }, + { + "epoch": 0.06358381502890173, + "grad_norm": 0.2889956899438749, + "learning_rate": 0.00012702078521939955, + "loss": 1.0096, + "step": 275 + }, + { + "epoch": 0.06473988439306358, + "grad_norm": 0.30488051626437424, + "learning_rate": 0.00012933025404157045, + "loss": 0.9819, + "step": 280 + }, + { + "epoch": 0.06589595375722543, + "grad_norm": 0.2884447508894878, + "learning_rate": 0.00013163972286374135, + "loss": 0.9468, + "step": 285 + }, + { + "epoch": 0.06705202312138728, + "grad_norm": 0.2755861252646099, + "learning_rate": 0.00013394919168591225, + "loss": 0.9828, + "step": 290 + }, + { + "epoch": 0.06820809248554913, + "grad_norm": 0.3025742735794824, + "learning_rate": 0.00013625866050808315, + "loss": 1.0178, + "step": 295 + }, + { + "epoch": 0.06936416184971098, + "grad_norm": 0.28423295151616335, + "learning_rate": 0.00013856812933025404, + "loss": 0.9945, + "step": 300 + }, + { + "epoch": 0.07052023121387284, + "grad_norm": 0.27585086154702476, + "learning_rate": 0.00014087759815242494, + "loss": 0.9695, + "step": 305 + }, + { + "epoch": 0.07167630057803469, + "grad_norm": 0.3038442212974576, + "learning_rate": 0.00014318706697459584, + "loss": 0.9808, + "step": 310 + }, + { + "epoch": 0.07283236994219654, + "grad_norm": 0.27277935404650916, + "learning_rate": 0.00014549653579676674, + "loss": 0.9846, + "step": 315 + }, + { + "epoch": 0.07398843930635839, + "grad_norm": 0.29591421870168255, + "learning_rate": 0.00014780600461893767, + "loss": 0.9954, + "step": 320 + }, + { + "epoch": 0.07514450867052024, + "grad_norm": 0.2823015918956852, + "learning_rate": 0.00015011547344110854, + "loss": 0.9674, + "step": 325 + }, + { + "epoch": 0.07630057803468208, + "grad_norm": 0.27344790533462154, + "learning_rate": 0.00015242494226327946, + "loss": 1.0236, + "step": 330 + }, + { + "epoch": 0.07745664739884393, + "grad_norm": 0.2949347092864385, + "learning_rate": 0.00015473441108545036, + "loss": 0.9813, + "step": 335 + }, + { + "epoch": 0.07861271676300578, + "grad_norm": 0.272235850063355, + "learning_rate": 0.00015704387990762126, + "loss": 0.987, + "step": 340 + }, + { + "epoch": 0.07976878612716763, + "grad_norm": 0.28565982534338485, + "learning_rate": 0.00015935334872979216, + "loss": 0.9716, + "step": 345 + }, + { + "epoch": 0.08092485549132948, + "grad_norm": 0.28274111136822716, + "learning_rate": 0.00016166281755196306, + "loss": 0.9779, + "step": 350 + }, + { + "epoch": 0.08208092485549133, + "grad_norm": 0.2878620812403313, + "learning_rate": 0.00016397228637413396, + "loss": 1.0209, + "step": 355 + }, + { + "epoch": 0.08323699421965318, + "grad_norm": 0.2865629307216328, + "learning_rate": 0.00016628175519630485, + "loss": 0.9921, + "step": 360 + }, + { + "epoch": 0.08439306358381503, + "grad_norm": 0.2826816337246457, + "learning_rate": 0.00016859122401847575, + "loss": 0.9348, + "step": 365 + }, + { + "epoch": 0.08554913294797688, + "grad_norm": 0.2881907360895622, + "learning_rate": 0.00017090069284064668, + "loss": 1.0231, + "step": 370 + }, + { + "epoch": 0.08670520231213873, + "grad_norm": 0.27638308021239116, + "learning_rate": 0.00017321016166281755, + "loss": 1.0079, + "step": 375 + }, + { + "epoch": 0.08786127167630058, + "grad_norm": 0.28798387568151884, + "learning_rate": 0.00017551963048498845, + "loss": 0.9165, + "step": 380 + }, + { + "epoch": 0.08901734104046242, + "grad_norm": 0.2767437621962458, + "learning_rate": 0.00017782909930715937, + "loss": 0.9489, + "step": 385 + }, + { + "epoch": 0.09017341040462427, + "grad_norm": 0.2877748564441354, + "learning_rate": 0.00018013856812933024, + "loss": 1.0022, + "step": 390 + }, + { + "epoch": 0.09132947976878612, + "grad_norm": 0.28925322685576144, + "learning_rate": 0.00018244803695150117, + "loss": 1.004, + "step": 395 + }, + { + "epoch": 0.09248554913294797, + "grad_norm": 0.31070512225405156, + "learning_rate": 0.00018475750577367207, + "loss": 0.9753, + "step": 400 + }, + { + "epoch": 0.09364161849710982, + "grad_norm": 0.31735644814371816, + "learning_rate": 0.00018706697459584297, + "loss": 0.9781, + "step": 405 + }, + { + "epoch": 0.09479768786127167, + "grad_norm": 0.2974104968051762, + "learning_rate": 0.00018937644341801387, + "loss": 0.9659, + "step": 410 + }, + { + "epoch": 0.09595375722543352, + "grad_norm": 0.2701025540904289, + "learning_rate": 0.00019168591224018476, + "loss": 0.9294, + "step": 415 + }, + { + "epoch": 0.09710982658959537, + "grad_norm": 0.27428411358071536, + "learning_rate": 0.00019399538106235566, + "loss": 0.9513, + "step": 420 + }, + { + "epoch": 0.09826589595375723, + "grad_norm": 0.2745240121214777, + "learning_rate": 0.0001963048498845266, + "loss": 1.0152, + "step": 425 + }, + { + "epoch": 0.09942196531791908, + "grad_norm": 0.274405426645142, + "learning_rate": 0.00019861431870669746, + "loss": 0.9863, + "step": 430 + }, + { + "epoch": 0.10057803468208093, + "grad_norm": 0.9048729498402529, + "learning_rate": 0.00019999986968812804, + "loss": 0.9992, + "step": 435 + }, + { + "epoch": 0.10173410404624278, + "grad_norm": 0.29062903824052616, + "learning_rate": 0.00019999840368346898, + "loss": 1.0509, + "step": 440 + }, + { + "epoch": 0.10289017341040463, + "grad_norm": 0.28317998532870003, + "learning_rate": 0.0001999953088082702, + "loss": 0.9774, + "step": 445 + }, + { + "epoch": 0.10404624277456648, + "grad_norm": 0.3000365098515979, + "learning_rate": 0.000199990585112944, + "loss": 0.9822, + "step": 450 + }, + { + "epoch": 0.10520231213872833, + "grad_norm": 0.28508661433222776, + "learning_rate": 0.00019998423267443454, + "loss": 0.9991, + "step": 455 + }, + { + "epoch": 0.10635838150289018, + "grad_norm": 0.300390972643917, + "learning_rate": 0.00019997625159621642, + "loss": 0.9411, + "step": 460 + }, + { + "epoch": 0.10751445086705202, + "grad_norm": 0.3065580188577536, + "learning_rate": 0.0001999666420082932, + "loss": 1.0153, + "step": 465 + }, + { + "epoch": 0.10867052023121387, + "grad_norm": 0.30176770344761106, + "learning_rate": 0.00019995540406719507, + "loss": 0.9451, + "step": 470 + }, + { + "epoch": 0.10982658959537572, + "grad_norm": 0.28363462288539226, + "learning_rate": 0.0001999425379559765, + "loss": 1.0229, + "step": 475 + }, + { + "epoch": 0.11098265895953757, + "grad_norm": 0.2979648349669768, + "learning_rate": 0.00019992804388421312, + "loss": 0.9615, + "step": 480 + }, + { + "epoch": 0.11213872832369942, + "grad_norm": 0.3533699255705701, + "learning_rate": 0.00019991192208799837, + "loss": 0.9945, + "step": 485 + }, + { + "epoch": 0.11329479768786127, + "grad_norm": 0.29374690183301444, + "learning_rate": 0.0001998941728299396, + "loss": 0.9481, + "step": 490 + }, + { + "epoch": 0.11445086705202312, + "grad_norm": 0.27638233259638606, + "learning_rate": 0.0001998747963991539, + "loss": 0.976, + "step": 495 + }, + { + "epoch": 0.11560693641618497, + "grad_norm": 0.29591583372167063, + "learning_rate": 0.00019985379311126327, + "loss": 0.9776, + "step": 500 + }, + { + "epoch": 0.11676300578034682, + "grad_norm": 0.28596466519406494, + "learning_rate": 0.00019983116330838955, + "loss": 1.0003, + "step": 505 + }, + { + "epoch": 0.11791907514450867, + "grad_norm": 0.27570121782043344, + "learning_rate": 0.00019980690735914877, + "loss": 0.9797, + "step": 510 + }, + { + "epoch": 0.11907514450867052, + "grad_norm": 0.30038501431153675, + "learning_rate": 0.0001997810256586453, + "loss": 0.9865, + "step": 515 + }, + { + "epoch": 0.12023121387283237, + "grad_norm": 0.31132305182282943, + "learning_rate": 0.00019975351862846523, + "loss": 1.0071, + "step": 520 + }, + { + "epoch": 0.12138728323699421, + "grad_norm": 0.28621069496048757, + "learning_rate": 0.00019972438671666967, + "loss": 0.9877, + "step": 525 + }, + { + "epoch": 0.12254335260115606, + "grad_norm": 0.3030051142967184, + "learning_rate": 0.00019969363039778728, + "loss": 0.9894, + "step": 530 + }, + { + "epoch": 0.12369942196531791, + "grad_norm": 0.2866405334142299, + "learning_rate": 0.0001996612501728067, + "loss": 0.9893, + "step": 535 + }, + { + "epoch": 0.12485549132947976, + "grad_norm": 0.2867249309375898, + "learning_rate": 0.00019962724656916826, + "loss": 0.9765, + "step": 540 + }, + { + "epoch": 0.1260115606936416, + "grad_norm": 0.31798324367678865, + "learning_rate": 0.00019959162014075553, + "loss": 0.9465, + "step": 545 + }, + { + "epoch": 0.12716763005780346, + "grad_norm": 0.30253196958089823, + "learning_rate": 0.0001995543714678861, + "loss": 0.9635, + "step": 550 + }, + { + "epoch": 0.1283236994219653, + "grad_norm": 0.28954658171708875, + "learning_rate": 0.00019951550115730244, + "loss": 1.0041, + "step": 555 + }, + { + "epoch": 0.12947976878612716, + "grad_norm": 0.28359238422516453, + "learning_rate": 0.00019947500984216157, + "loss": 0.9837, + "step": 560 + }, + { + "epoch": 0.130635838150289, + "grad_norm": 0.29624734780777734, + "learning_rate": 0.00019943289818202519, + "loss": 0.9375, + "step": 565 + }, + { + "epoch": 0.13179190751445086, + "grad_norm": 0.291198302886137, + "learning_rate": 0.0001993891668628486, + "loss": 0.9665, + "step": 570 + }, + { + "epoch": 0.1329479768786127, + "grad_norm": 0.31056892991094237, + "learning_rate": 0.00019934381659696989, + "loss": 0.9414, + "step": 575 + }, + { + "epoch": 0.13410404624277455, + "grad_norm": 0.3136680287460065, + "learning_rate": 0.0001992968481230978, + "loss": 1.0442, + "step": 580 + }, + { + "epoch": 0.1352601156069364, + "grad_norm": 0.30128304774644027, + "learning_rate": 0.0001992482622063003, + "loss": 0.9916, + "step": 585 + }, + { + "epoch": 0.13641618497109825, + "grad_norm": 0.29097714851626455, + "learning_rate": 0.00019919805963799166, + "loss": 0.9947, + "step": 590 + }, + { + "epoch": 0.1375722543352601, + "grad_norm": 0.280417582339227, + "learning_rate": 0.0001991462412359198, + "loss": 0.9825, + "step": 595 + }, + { + "epoch": 0.13872832369942195, + "grad_norm": 0.28842456996684646, + "learning_rate": 0.00019909280784415287, + "loss": 1.0237, + "step": 600 + }, + { + "epoch": 0.13988439306358383, + "grad_norm": 0.2874190385868597, + "learning_rate": 0.00019903776033306555, + "loss": 0.9611, + "step": 605 + }, + { + "epoch": 0.14104046242774568, + "grad_norm": 0.2854599682323198, + "learning_rate": 0.00019898109959932478, + "loss": 0.9879, + "step": 610 + }, + { + "epoch": 0.14219653179190753, + "grad_norm": 0.2986976522432421, + "learning_rate": 0.0001989228265658754, + "loss": 0.9911, + "step": 615 + }, + { + "epoch": 0.14335260115606938, + "grad_norm": 0.329057703574734, + "learning_rate": 0.00019886294218192477, + "loss": 0.9714, + "step": 620 + }, + { + "epoch": 0.14450867052023122, + "grad_norm": 0.29538740567126964, + "learning_rate": 0.00019880144742292753, + "loss": 0.955, + "step": 625 + }, + { + "epoch": 0.14566473988439307, + "grad_norm": 0.32615973261215037, + "learning_rate": 0.00019873834329056975, + "loss": 0.9789, + "step": 630 + }, + { + "epoch": 0.14682080924855492, + "grad_norm": 0.26819671057976713, + "learning_rate": 0.00019867363081275242, + "loss": 0.9471, + "step": 635 + }, + { + "epoch": 0.14797687861271677, + "grad_norm": 0.28488339516835476, + "learning_rate": 0.00019860731104357485, + "loss": 0.9779, + "step": 640 + }, + { + "epoch": 0.14913294797687862, + "grad_norm": 0.27998012828783303, + "learning_rate": 0.00019853938506331749, + "loss": 0.951, + "step": 645 + }, + { + "epoch": 0.15028901734104047, + "grad_norm": 0.29388395642325527, + "learning_rate": 0.00019846985397842427, + "loss": 0.9554, + "step": 650 + }, + { + "epoch": 0.15144508670520232, + "grad_norm": 0.2896808457234832, + "learning_rate": 0.0001983987189214846, + "loss": 0.9651, + "step": 655 + }, + { + "epoch": 0.15260115606936417, + "grad_norm": 0.30032384976542736, + "learning_rate": 0.000198325981051215, + "loss": 0.943, + "step": 660 + }, + { + "epoch": 0.15375722543352602, + "grad_norm": 0.2977312487124492, + "learning_rate": 0.00019825164155244012, + "loss": 0.9887, + "step": 665 + }, + { + "epoch": 0.15491329479768787, + "grad_norm": 0.341662364901083, + "learning_rate": 0.00019817570163607347, + "loss": 1.0059, + "step": 670 + }, + { + "epoch": 0.15606936416184972, + "grad_norm": 0.2996891760902693, + "learning_rate": 0.00019809816253909773, + "loss": 0.9673, + "step": 675 + }, + { + "epoch": 0.15722543352601157, + "grad_norm": 0.2882952073089595, + "learning_rate": 0.00019801902552454454, + "loss": 0.9558, + "step": 680 + }, + { + "epoch": 0.15838150289017341, + "grad_norm": 0.3203389676438109, + "learning_rate": 0.00019793829188147406, + "loss": 1.0122, + "step": 685 + }, + { + "epoch": 0.15953757225433526, + "grad_norm": 0.30270844487268483, + "learning_rate": 0.00019785596292495376, + "loss": 0.9822, + "step": 690 + }, + { + "epoch": 0.1606936416184971, + "grad_norm": 0.27999850496563145, + "learning_rate": 0.00019777203999603717, + "loss": 0.9841, + "step": 695 + }, + { + "epoch": 0.16184971098265896, + "grad_norm": 0.29383524324706056, + "learning_rate": 0.000197686524461742, + "loss": 1.0269, + "step": 700 + }, + { + "epoch": 0.1630057803468208, + "grad_norm": 0.28066838492867907, + "learning_rate": 0.0001975994177150278, + "loss": 0.9927, + "step": 705 + }, + { + "epoch": 0.16416184971098266, + "grad_norm": 0.31122106212206363, + "learning_rate": 0.0001975107211747734, + "loss": 0.9632, + "step": 710 + }, + { + "epoch": 0.1653179190751445, + "grad_norm": 0.30120413493928255, + "learning_rate": 0.00019742043628575364, + "loss": 0.9739, + "step": 715 + }, + { + "epoch": 0.16647398843930636, + "grad_norm": 0.30660646504263267, + "learning_rate": 0.00019732856451861594, + "loss": 0.976, + "step": 720 + }, + { + "epoch": 0.1676300578034682, + "grad_norm": 0.2856850255835453, + "learning_rate": 0.0001972351073698564, + "loss": 0.9848, + "step": 725 + }, + { + "epoch": 0.16878612716763006, + "grad_norm": 0.3147493299962656, + "learning_rate": 0.0001971400663617952, + "loss": 0.9921, + "step": 730 + }, + { + "epoch": 0.1699421965317919, + "grad_norm": 0.29251745238857035, + "learning_rate": 0.0001970434430425521, + "loss": 0.9398, + "step": 735 + }, + { + "epoch": 0.17109826589595376, + "grad_norm": 0.2938440274420457, + "learning_rate": 0.000196945238986021, + "loss": 1.0153, + "step": 740 + }, + { + "epoch": 0.1722543352601156, + "grad_norm": 0.2898363733115959, + "learning_rate": 0.00019684545579184433, + "loss": 0.9576, + "step": 745 + }, + { + "epoch": 0.17341040462427745, + "grad_norm": 0.3088888037796155, + "learning_rate": 0.00019674409508538718, + "loss": 0.9696, + "step": 750 + }, + { + "epoch": 0.1745664739884393, + "grad_norm": 0.2942346284280998, + "learning_rate": 0.0001966411585177105, + "loss": 1.0203, + "step": 755 + }, + { + "epoch": 0.17572254335260115, + "grad_norm": 0.2852125727236794, + "learning_rate": 0.00019653664776554455, + "loss": 0.9556, + "step": 760 + }, + { + "epoch": 0.176878612716763, + "grad_norm": 0.3016262447783914, + "learning_rate": 0.0001964305645312613, + "loss": 0.9896, + "step": 765 + }, + { + "epoch": 0.17803468208092485, + "grad_norm": 0.30255168675186533, + "learning_rate": 0.00019632291054284693, + "loss": 0.9839, + "step": 770 + }, + { + "epoch": 0.1791907514450867, + "grad_norm": 0.2806238558760804, + "learning_rate": 0.0001962136875538735, + "loss": 0.9748, + "step": 775 + }, + { + "epoch": 0.18034682080924855, + "grad_norm": 0.2945672906624874, + "learning_rate": 0.00019610289734347053, + "loss": 0.9479, + "step": 780 + }, + { + "epoch": 0.1815028901734104, + "grad_norm": 0.3012247410303452, + "learning_rate": 0.00019599054171629595, + "loss": 1.0132, + "step": 785 + }, + { + "epoch": 0.18265895953757225, + "grad_norm": 0.2944499117709193, + "learning_rate": 0.0001958766225025066, + "loss": 0.9336, + "step": 790 + }, + { + "epoch": 0.1838150289017341, + "grad_norm": 0.2989322109974369, + "learning_rate": 0.0001957611415577287, + "loss": 0.9857, + "step": 795 + }, + { + "epoch": 0.18497109826589594, + "grad_norm": 0.2859334477998094, + "learning_rate": 0.0001956441007630273, + "loss": 0.9831, + "step": 800 + }, + { + "epoch": 0.1861271676300578, + "grad_norm": 0.27928427673168016, + "learning_rate": 0.0001955255020248759, + "loss": 0.9582, + "step": 805 + }, + { + "epoch": 0.18728323699421964, + "grad_norm": 0.3044716011707441, + "learning_rate": 0.00019540534727512522, + "loss": 1.0061, + "step": 810 + }, + { + "epoch": 0.1884393063583815, + "grad_norm": 0.30123009109430604, + "learning_rate": 0.00019528363847097185, + "loss": 1.0015, + "step": 815 + }, + { + "epoch": 0.18959537572254334, + "grad_norm": 0.2983431588431708, + "learning_rate": 0.00019516037759492627, + "loss": 0.9917, + "step": 820 + }, + { + "epoch": 0.1907514450867052, + "grad_norm": 0.3045474200889231, + "learning_rate": 0.00019503556665478067, + "loss": 0.9924, + "step": 825 + }, + { + "epoch": 0.19190751445086704, + "grad_norm": 0.3023598788495744, + "learning_rate": 0.00019490920768357607, + "loss": 0.9824, + "step": 830 + }, + { + "epoch": 0.1930635838150289, + "grad_norm": 0.288004788186244, + "learning_rate": 0.00019478130273956943, + "loss": 0.9756, + "step": 835 + }, + { + "epoch": 0.19421965317919074, + "grad_norm": 0.28774556008482255, + "learning_rate": 0.00019465185390619996, + "loss": 0.9292, + "step": 840 + }, + { + "epoch": 0.19537572254335261, + "grad_norm": 0.29766392431820693, + "learning_rate": 0.00019452086329205522, + "loss": 1.008, + "step": 845 + }, + { + "epoch": 0.19653179190751446, + "grad_norm": 0.2968523784854959, + "learning_rate": 0.00019438833303083678, + "loss": 0.9469, + "step": 850 + }, + { + "epoch": 0.1976878612716763, + "grad_norm": 0.32050710908212127, + "learning_rate": 0.00019425426528132546, + "loss": 0.9584, + "step": 855 + }, + { + "epoch": 0.19884393063583816, + "grad_norm": 0.30039833468190924, + "learning_rate": 0.00019411866222734627, + "loss": 0.9345, + "step": 860 + }, + { + "epoch": 0.2, + "grad_norm": 0.3003418210003542, + "learning_rate": 0.00019398152607773264, + "loss": 1.0149, + "step": 865 + }, + { + "epoch": 0.20115606936416186, + "grad_norm": 0.28740303119481114, + "learning_rate": 0.00019384285906629055, + "loss": 0.9453, + "step": 870 + }, + { + "epoch": 0.2023121387283237, + "grad_norm": 0.3022020472256075, + "learning_rate": 0.00019370266345176214, + "loss": 0.9964, + "step": 875 + }, + { + "epoch": 0.20346820809248556, + "grad_norm": 0.2887054624910223, + "learning_rate": 0.00019356094151778895, + "loss": 0.9857, + "step": 880 + }, + { + "epoch": 0.2046242774566474, + "grad_norm": 0.38190567119224494, + "learning_rate": 0.00019341769557287467, + "loss": 0.9554, + "step": 885 + }, + { + "epoch": 0.20578034682080926, + "grad_norm": 0.29810678010523667, + "learning_rate": 0.00019327292795034753, + "loss": 0.9138, + "step": 890 + }, + { + "epoch": 0.2069364161849711, + "grad_norm": 0.30694177853911697, + "learning_rate": 0.00019312664100832233, + "loss": 0.9504, + "step": 895 + }, + { + "epoch": 0.20809248554913296, + "grad_norm": 0.31819590085942273, + "learning_rate": 0.00019297883712966204, + "loss": 0.9969, + "step": 900 + }, + { + "epoch": 0.2092485549132948, + "grad_norm": 0.2915158769948663, + "learning_rate": 0.00019282951872193885, + "loss": 1.0207, + "step": 905 + }, + { + "epoch": 0.21040462427745665, + "grad_norm": 0.3020525391391881, + "learning_rate": 0.0001926786882173952, + "loss": 0.9424, + "step": 910 + }, + { + "epoch": 0.2115606936416185, + "grad_norm": 0.3095938351879466, + "learning_rate": 0.000192526348072904, + "loss": 0.9762, + "step": 915 + }, + { + "epoch": 0.21271676300578035, + "grad_norm": 0.32067080864793646, + "learning_rate": 0.0001923725007699285, + "loss": 0.9423, + "step": 920 + }, + { + "epoch": 0.2138728323699422, + "grad_norm": 0.3144763229931657, + "learning_rate": 0.00019221714881448217, + "loss": 0.9824, + "step": 925 + }, + { + "epoch": 0.21502890173410405, + "grad_norm": 0.3179959968229521, + "learning_rate": 0.0001920602947370876, + "loss": 0.9829, + "step": 930 + }, + { + "epoch": 0.2161849710982659, + "grad_norm": 0.30113560950220297, + "learning_rate": 0.00019190194109273544, + "loss": 0.9463, + "step": 935 + }, + { + "epoch": 0.21734104046242775, + "grad_norm": 0.2896899865821755, + "learning_rate": 0.00019174209046084276, + "loss": 0.9786, + "step": 940 + }, + { + "epoch": 0.2184971098265896, + "grad_norm": 0.2969332481357098, + "learning_rate": 0.00019158074544521094, + "loss": 1.0183, + "step": 945 + }, + { + "epoch": 0.21965317919075145, + "grad_norm": 0.29984068010125464, + "learning_rate": 0.0001914179086739834, + "loss": 0.9688, + "step": 950 + }, + { + "epoch": 0.2208092485549133, + "grad_norm": 0.3179298483108417, + "learning_rate": 0.0001912535827996026, + "loss": 1.0213, + "step": 955 + }, + { + "epoch": 0.22196531791907514, + "grad_norm": 0.30071667875773894, + "learning_rate": 0.0001910877704987671, + "loss": 0.9947, + "step": 960 + }, + { + "epoch": 0.223121387283237, + "grad_norm": 0.3162017726279279, + "learning_rate": 0.00019092047447238773, + "loss": 0.9765, + "step": 965 + }, + { + "epoch": 0.22427745664739884, + "grad_norm": 0.31999805840625895, + "learning_rate": 0.0001907516974455436, + "loss": 0.9956, + "step": 970 + }, + { + "epoch": 0.2254335260115607, + "grad_norm": 0.2931503664377131, + "learning_rate": 0.00019058144216743797, + "loss": 1.002, + "step": 975 + }, + { + "epoch": 0.22658959537572254, + "grad_norm": 0.30298932952701096, + "learning_rate": 0.0001904097114113531, + "loss": 1.0159, + "step": 980 + }, + { + "epoch": 0.2277456647398844, + "grad_norm": 0.29804869616385626, + "learning_rate": 0.0001902365079746054, + "loss": 0.9771, + "step": 985 + }, + { + "epoch": 0.22890173410404624, + "grad_norm": 0.30761841093550274, + "learning_rate": 0.00019006183467849957, + "loss": 1.0157, + "step": 990 + }, + { + "epoch": 0.2300578034682081, + "grad_norm": 0.3080044180069082, + "learning_rate": 0.000189885694368283, + "loss": 0.9463, + "step": 995 + }, + { + "epoch": 0.23121387283236994, + "grad_norm": 0.29852967122969754, + "learning_rate": 0.00018970808991309904, + "loss": 1.0021, + "step": 1000 + }, + { + "epoch": 0.2323699421965318, + "grad_norm": 0.2987990831868652, + "learning_rate": 0.00018952902420594058, + "loss": 0.9492, + "step": 1005 + }, + { + "epoch": 0.23352601156069364, + "grad_norm": 0.30446826555699585, + "learning_rate": 0.0001893485001636026, + "loss": 1.0019, + "step": 1010 + }, + { + "epoch": 0.23468208092485549, + "grad_norm": 0.30998357590060016, + "learning_rate": 0.00018916652072663515, + "loss": 0.9478, + "step": 1015 + }, + { + "epoch": 0.23583815028901733, + "grad_norm": 0.3109448921090665, + "learning_rate": 0.0001889830888592949, + "loss": 0.9797, + "step": 1020 + }, + { + "epoch": 0.23699421965317918, + "grad_norm": 0.30658444282489306, + "learning_rate": 0.00018879820754949718, + "loss": 0.9976, + "step": 1025 + }, + { + "epoch": 0.23815028901734103, + "grad_norm": 0.32337481964353393, + "learning_rate": 0.0001886118798087673, + "loss": 0.9622, + "step": 1030 + }, + { + "epoch": 0.23930635838150288, + "grad_norm": 0.30934913326885294, + "learning_rate": 0.00018842410867219136, + "loss": 1.0095, + "step": 1035 + }, + { + "epoch": 0.24046242774566473, + "grad_norm": 0.32554022935815935, + "learning_rate": 0.0001882348971983669, + "loss": 1.0082, + "step": 1040 + }, + { + "epoch": 0.24161849710982658, + "grad_norm": 0.28387403686918444, + "learning_rate": 0.0001880442484693531, + "loss": 0.9433, + "step": 1045 + }, + { + "epoch": 0.24277456647398843, + "grad_norm": 0.2889713456813008, + "learning_rate": 0.0001878521655906205, + "loss": 0.994, + "step": 1050 + }, + { + "epoch": 0.24393063583815028, + "grad_norm": 0.3038933468668014, + "learning_rate": 0.00018765865169100048, + "loss": 0.966, + "step": 1055 + }, + { + "epoch": 0.24508670520231213, + "grad_norm": 0.294597892145829, + "learning_rate": 0.00018746370992263423, + "loss": 0.9501, + "step": 1060 + }, + { + "epoch": 0.24624277456647398, + "grad_norm": 0.30895343503048994, + "learning_rate": 0.00018726734346092148, + "loss": 0.9663, + "step": 1065 + }, + { + "epoch": 0.24739884393063583, + "grad_norm": 0.32867090123543136, + "learning_rate": 0.00018706955550446878, + "loss": 1.0, + "step": 1070 + }, + { + "epoch": 0.24855491329479767, + "grad_norm": 0.2860732876796574, + "learning_rate": 0.00018687034927503728, + "loss": 0.9282, + "step": 1075 + }, + { + "epoch": 0.24971098265895952, + "grad_norm": 0.3622792683930793, + "learning_rate": 0.00018666972801749035, + "loss": 0.9534, + "step": 1080 + }, + { + "epoch": 0.2508670520231214, + "grad_norm": 0.3215073797028641, + "learning_rate": 0.00018646769499974076, + "loss": 1.0177, + "step": 1085 + }, + { + "epoch": 0.2520231213872832, + "grad_norm": 0.30391881014243827, + "learning_rate": 0.00018626425351269733, + "loss": 1.0213, + "step": 1090 + }, + { + "epoch": 0.25317919075144507, + "grad_norm": 0.2959182253635083, + "learning_rate": 0.00018605940687021133, + "loss": 0.9265, + "step": 1095 + }, + { + "epoch": 0.2543352601156069, + "grad_norm": 0.3054358776400607, + "learning_rate": 0.00018585315840902275, + "loss": 0.9566, + "step": 1100 + }, + { + "epoch": 0.25549132947976877, + "grad_norm": 0.3038159760611022, + "learning_rate": 0.00018564551148870563, + "loss": 0.9728, + "step": 1105 + }, + { + "epoch": 0.2566473988439306, + "grad_norm": 0.295408221908172, + "learning_rate": 0.0001854364694916134, + "loss": 0.9769, + "step": 1110 + }, + { + "epoch": 0.25780346820809247, + "grad_norm": 0.3040191717007843, + "learning_rate": 0.00018522603582282396, + "loss": 0.9745, + "step": 1115 + }, + { + "epoch": 0.2589595375722543, + "grad_norm": 0.3276795968564327, + "learning_rate": 0.0001850142139100841, + "loss": 0.9843, + "step": 1120 + }, + { + "epoch": 0.26011560693641617, + "grad_norm": 0.30521432519770036, + "learning_rate": 0.0001848010072037536, + "loss": 0.9958, + "step": 1125 + }, + { + "epoch": 0.261271676300578, + "grad_norm": 0.3074723851239691, + "learning_rate": 0.0001845864191767491, + "loss": 1.0061, + "step": 1130 + }, + { + "epoch": 0.26242774566473986, + "grad_norm": 0.38542353780888683, + "learning_rate": 0.0001843704533244876, + "loss": 0.9527, + "step": 1135 + }, + { + "epoch": 0.2635838150289017, + "grad_norm": 0.31655887417225065, + "learning_rate": 0.00018415311316482934, + "loss": 0.9677, + "step": 1140 + }, + { + "epoch": 0.26473988439306356, + "grad_norm": 0.296622948449929, + "learning_rate": 0.00018393440223802077, + "loss": 1.0277, + "step": 1145 + }, + { + "epoch": 0.2658959537572254, + "grad_norm": 0.30339302168936083, + "learning_rate": 0.0001837143241066365, + "loss": 0.9996, + "step": 1150 + }, + { + "epoch": 0.26705202312138726, + "grad_norm": 0.321536892508796, + "learning_rate": 0.00018349288235552168, + "loss": 0.9599, + "step": 1155 + }, + { + "epoch": 0.2682080924855491, + "grad_norm": 0.3220388229874098, + "learning_rate": 0.0001832700805917333, + "loss": 0.9957, + "step": 1160 + }, + { + "epoch": 0.26936416184971096, + "grad_norm": 0.2951235948754925, + "learning_rate": 0.00018304592244448156, + "loss": 1.0261, + "step": 1165 + }, + { + "epoch": 0.2705202312138728, + "grad_norm": 0.3044130581874429, + "learning_rate": 0.0001828204115650708, + "loss": 0.9871, + "step": 1170 + }, + { + "epoch": 0.27167630057803466, + "grad_norm": 0.3009718363326464, + "learning_rate": 0.00018259355162684, + "loss": 0.9963, + "step": 1175 + }, + { + "epoch": 0.2728323699421965, + "grad_norm": 0.28905525049692726, + "learning_rate": 0.00018236534632510277, + "loss": 0.9848, + "step": 1180 + }, + { + "epoch": 0.27398843930635836, + "grad_norm": 0.35621478263055945, + "learning_rate": 0.00018213579937708735, + "loss": 0.9529, + "step": 1185 + }, + { + "epoch": 0.2751445086705202, + "grad_norm": 0.3185821009153887, + "learning_rate": 0.00018190491452187613, + "loss": 1.0045, + "step": 1190 + }, + { + "epoch": 0.27630057803468205, + "grad_norm": 0.3174337839488537, + "learning_rate": 0.00018167269552034446, + "loss": 0.9549, + "step": 1195 + }, + { + "epoch": 0.2774566473988439, + "grad_norm": 0.2929697771476232, + "learning_rate": 0.00018143914615509967, + "loss": 0.9463, + "step": 1200 + }, + { + "epoch": 0.2786127167630058, + "grad_norm": 0.3218921944038643, + "learning_rate": 0.00018120427023041925, + "loss": 1.0432, + "step": 1205 + }, + { + "epoch": 0.27976878612716766, + "grad_norm": 0.29522034854514784, + "learning_rate": 0.00018096807157218909, + "loss": 0.9354, + "step": 1210 + }, + { + "epoch": 0.2809248554913295, + "grad_norm": 0.31103148068930697, + "learning_rate": 0.0001807305540278409, + "loss": 0.9275, + "step": 1215 + }, + { + "epoch": 0.28208092485549136, + "grad_norm": 0.3268759634134707, + "learning_rate": 0.00018049172146628975, + "loss": 1.0122, + "step": 1220 + }, + { + "epoch": 0.2832369942196532, + "grad_norm": 0.2992667041917287, + "learning_rate": 0.00018025157777787102, + "loss": 0.9422, + "step": 1225 + }, + { + "epoch": 0.28439306358381505, + "grad_norm": 0.31617502406730474, + "learning_rate": 0.00018001012687427688, + "loss": 0.9909, + "step": 1230 + }, + { + "epoch": 0.2855491329479769, + "grad_norm": 0.31251457847228237, + "learning_rate": 0.0001797673726884928, + "loss": 0.9885, + "step": 1235 + }, + { + "epoch": 0.28670520231213875, + "grad_norm": 0.32021979333618866, + "learning_rate": 0.00017952331917473336, + "loss": 0.9396, + "step": 1240 + }, + { + "epoch": 0.2878612716763006, + "grad_norm": 0.34841403249467495, + "learning_rate": 0.00017927797030837768, + "loss": 0.9188, + "step": 1245 + }, + { + "epoch": 0.28901734104046245, + "grad_norm": 0.40965133837398776, + "learning_rate": 0.0001790313300859051, + "loss": 0.9582, + "step": 1250 + }, + { + "epoch": 0.2901734104046243, + "grad_norm": 0.3128265434514317, + "learning_rate": 0.00017878340252482956, + "loss": 0.9891, + "step": 1255 + }, + { + "epoch": 0.29132947976878615, + "grad_norm": 0.3762731976913158, + "learning_rate": 0.00017853419166363458, + "loss": 0.973, + "step": 1260 + }, + { + "epoch": 0.292485549132948, + "grad_norm": 0.3150565577316399, + "learning_rate": 0.00017828370156170727, + "loss": 0.9777, + "step": 1265 + }, + { + "epoch": 0.29364161849710985, + "grad_norm": 0.3124944800404496, + "learning_rate": 0.00017803193629927223, + "loss": 0.984, + "step": 1270 + }, + { + "epoch": 0.2947976878612717, + "grad_norm": 0.3456192268358289, + "learning_rate": 0.0001777788999773251, + "loss": 0.9881, + "step": 1275 + }, + { + "epoch": 0.29595375722543354, + "grad_norm": 0.309939339205246, + "learning_rate": 0.0001775245967175658, + "loss": 0.9483, + "step": 1280 + }, + { + "epoch": 0.2971098265895954, + "grad_norm": 0.29097211271370754, + "learning_rate": 0.00017726903066233134, + "loss": 0.9438, + "step": 1285 + }, + { + "epoch": 0.29826589595375724, + "grad_norm": 0.29586840025056343, + "learning_rate": 0.00017701220597452833, + "loss": 0.9754, + "step": 1290 + }, + { + "epoch": 0.2994219653179191, + "grad_norm": 0.3017342795593241, + "learning_rate": 0.0001767541268375652, + "loss": 0.9805, + "step": 1295 + }, + { + "epoch": 0.30057803468208094, + "grad_norm": 0.32061970031794484, + "learning_rate": 0.00017649479745528417, + "loss": 0.9818, + "step": 1300 + }, + { + "epoch": 0.3017341040462428, + "grad_norm": 0.3011478634925178, + "learning_rate": 0.00017623422205189252, + "loss": 0.9815, + "step": 1305 + }, + { + "epoch": 0.30289017341040464, + "grad_norm": 0.3086640164396148, + "learning_rate": 0.000175972404871894, + "loss": 0.9756, + "step": 1310 + }, + { + "epoch": 0.3040462427745665, + "grad_norm": 0.31272520758866784, + "learning_rate": 0.0001757093501800196, + "loss": 0.9993, + "step": 1315 + }, + { + "epoch": 0.30520231213872834, + "grad_norm": 0.3574223169536535, + "learning_rate": 0.0001754450622611581, + "loss": 0.9468, + "step": 1320 + }, + { + "epoch": 0.3063583815028902, + "grad_norm": 0.3147796469526059, + "learning_rate": 0.0001751795454202863, + "loss": 0.9848, + "step": 1325 + }, + { + "epoch": 0.30751445086705204, + "grad_norm": 0.3118480322453381, + "learning_rate": 0.0001749128039823988, + "loss": 1.0249, + "step": 1330 + }, + { + "epoch": 0.3086705202312139, + "grad_norm": 0.3023964221368843, + "learning_rate": 0.00017464484229243768, + "loss": 0.9232, + "step": 1335 + }, + { + "epoch": 0.30982658959537573, + "grad_norm": 0.3310385612654826, + "learning_rate": 0.0001743756647152216, + "loss": 0.9975, + "step": 1340 + }, + { + "epoch": 0.3109826589595376, + "grad_norm": 0.3318179078521207, + "learning_rate": 0.00017410527563537488, + "loss": 0.9776, + "step": 1345 + }, + { + "epoch": 0.31213872832369943, + "grad_norm": 0.31582505608794464, + "learning_rate": 0.00017383367945725584, + "loss": 0.9191, + "step": 1350 + }, + { + "epoch": 0.3132947976878613, + "grad_norm": 0.30919652895574773, + "learning_rate": 0.00017356088060488525, + "loss": 0.9813, + "step": 1355 + }, + { + "epoch": 0.31445086705202313, + "grad_norm": 0.30484218478884034, + "learning_rate": 0.00017328688352187416, + "loss": 0.9791, + "step": 1360 + }, + { + "epoch": 0.315606936416185, + "grad_norm": 0.3016535900655947, + "learning_rate": 0.00017301169267135163, + "loss": 0.9918, + "step": 1365 + }, + { + "epoch": 0.31676300578034683, + "grad_norm": 0.2828981498735541, + "learning_rate": 0.00017273531253589187, + "loss": 0.9266, + "step": 1370 + }, + { + "epoch": 0.3179190751445087, + "grad_norm": 0.3077602888091621, + "learning_rate": 0.00017245774761744134, + "loss": 0.9674, + "step": 1375 + }, + { + "epoch": 0.3190751445086705, + "grad_norm": 0.3099506136784924, + "learning_rate": 0.00017217900243724543, + "loss": 0.9836, + "step": 1380 + }, + { + "epoch": 0.3202312138728324, + "grad_norm": 0.31771107668630955, + "learning_rate": 0.00017189908153577473, + "loss": 0.9387, + "step": 1385 + }, + { + "epoch": 0.3213872832369942, + "grad_norm": 0.3121605825107148, + "learning_rate": 0.0001716179894726511, + "loss": 1.0108, + "step": 1390 + }, + { + "epoch": 0.3225433526011561, + "grad_norm": 0.32262006217730343, + "learning_rate": 0.0001713357308265735, + "loss": 1.0374, + "step": 1395 + }, + { + "epoch": 0.3236994219653179, + "grad_norm": 0.3209655049858561, + "learning_rate": 0.0001710523101952432, + "loss": 0.9936, + "step": 1400 + }, + { + "epoch": 0.3248554913294798, + "grad_norm": 0.32077514063057966, + "learning_rate": 0.00017076773219528905, + "loss": 0.9704, + "step": 1405 + }, + { + "epoch": 0.3260115606936416, + "grad_norm": 0.30932663652684755, + "learning_rate": 0.0001704820014621923, + "loss": 1.003, + "step": 1410 + }, + { + "epoch": 0.32716763005780347, + "grad_norm": 0.32273320904553404, + "learning_rate": 0.00017019512265021097, + "loss": 1.0388, + "step": 1415 + }, + { + "epoch": 0.3283236994219653, + "grad_norm": 0.31422034130330717, + "learning_rate": 0.00016990710043230406, + "loss": 0.9556, + "step": 1420 + }, + { + "epoch": 0.32947976878612717, + "grad_norm": 0.31168769800585655, + "learning_rate": 0.00016961793950005558, + "loss": 0.9746, + "step": 1425 + }, + { + "epoch": 0.330635838150289, + "grad_norm": 0.3052533915264079, + "learning_rate": 0.00016932764456359793, + "loss": 0.9542, + "step": 1430 + }, + { + "epoch": 0.33179190751445087, + "grad_norm": 0.2905803213236271, + "learning_rate": 0.0001690362203515353, + "loss": 0.97, + "step": 1435 + }, + { + "epoch": 0.3329479768786127, + "grad_norm": 0.3350090558904066, + "learning_rate": 0.00016874367161086662, + "loss": 1.0443, + "step": 1440 + }, + { + "epoch": 0.33410404624277457, + "grad_norm": 0.3060737171676248, + "learning_rate": 0.00016845000310690815, + "loss": 1.043, + "step": 1445 + }, + { + "epoch": 0.3352601156069364, + "grad_norm": 0.3129317388012455, + "learning_rate": 0.00016815521962321604, + "loss": 0.9288, + "step": 1450 + }, + { + "epoch": 0.33641618497109826, + "grad_norm": 0.3197265364198292, + "learning_rate": 0.00016785932596150827, + "loss": 0.984, + "step": 1455 + }, + { + "epoch": 0.3375722543352601, + "grad_norm": 0.29604487460555096, + "learning_rate": 0.0001675623269415864, + "loss": 1.0016, + "step": 1460 + }, + { + "epoch": 0.33872832369942196, + "grad_norm": 0.30262162872771553, + "learning_rate": 0.00016726422740125728, + "loss": 0.9856, + "step": 1465 + }, + { + "epoch": 0.3398843930635838, + "grad_norm": 0.3087826779620012, + "learning_rate": 0.000166965032196254, + "loss": 0.9957, + "step": 1470 + }, + { + "epoch": 0.34104046242774566, + "grad_norm": 0.3024417730810271, + "learning_rate": 0.00016666474620015686, + "loss": 0.939, + "step": 1475 + }, + { + "epoch": 0.3421965317919075, + "grad_norm": 0.3138788978371571, + "learning_rate": 0.0001663633743043141, + "loss": 0.9363, + "step": 1480 + }, + { + "epoch": 0.34335260115606936, + "grad_norm": 0.30022727252981385, + "learning_rate": 0.0001660609214177621, + "loss": 1.0356, + "step": 1485 + }, + { + "epoch": 0.3445086705202312, + "grad_norm": 0.3554930404416277, + "learning_rate": 0.00016575739246714547, + "loss": 0.9741, + "step": 1490 + }, + { + "epoch": 0.34566473988439306, + "grad_norm": 0.29207690202141207, + "learning_rate": 0.00016545279239663682, + "loss": 0.9914, + "step": 1495 + }, + { + "epoch": 0.3468208092485549, + "grad_norm": 0.3029574756031942, + "learning_rate": 0.00016514712616785612, + "loss": 0.9421, + "step": 1500 + }, + { + "epoch": 0.34797687861271676, + "grad_norm": 0.3003131478682294, + "learning_rate": 0.00016484039875979005, + "loss": 0.9536, + "step": 1505 + }, + { + "epoch": 0.3491329479768786, + "grad_norm": 0.31513567646148855, + "learning_rate": 0.00016453261516871068, + "loss": 0.9426, + "step": 1510 + }, + { + "epoch": 0.35028901734104045, + "grad_norm": 0.2996325149723125, + "learning_rate": 0.00016422378040809437, + "loss": 1.0104, + "step": 1515 + }, + { + "epoch": 0.3514450867052023, + "grad_norm": 0.31000327701469227, + "learning_rate": 0.00016391389950853977, + "loss": 0.9899, + "step": 1520 + }, + { + "epoch": 0.35260115606936415, + "grad_norm": 0.3077861834938761, + "learning_rate": 0.0001636029775176862, + "loss": 0.9865, + "step": 1525 + }, + { + "epoch": 0.353757225433526, + "grad_norm": 0.3065445301834393, + "learning_rate": 0.00016329101950013122, + "loss": 0.9833, + "step": 1530 + }, + { + "epoch": 0.35491329479768785, + "grad_norm": 0.3317987511497053, + "learning_rate": 0.00016297803053734816, + "loss": 0.9549, + "step": 1535 + }, + { + "epoch": 0.3560693641618497, + "grad_norm": 0.3196645968421778, + "learning_rate": 0.0001626640157276034, + "loss": 0.9675, + "step": 1540 + }, + { + "epoch": 0.35722543352601155, + "grad_norm": 0.3134584834564055, + "learning_rate": 0.00016234898018587337, + "loss": 0.9516, + "step": 1545 + }, + { + "epoch": 0.3583815028901734, + "grad_norm": 0.3117834490142485, + "learning_rate": 0.00016203292904376105, + "loss": 0.9846, + "step": 1550 + }, + { + "epoch": 0.35953757225433525, + "grad_norm": 0.3217830266213364, + "learning_rate": 0.00016171586744941264, + "loss": 0.967, + "step": 1555 + }, + { + "epoch": 0.3606936416184971, + "grad_norm": 0.30934961569422764, + "learning_rate": 0.00016139780056743342, + "loss": 0.9649, + "step": 1560 + }, + { + "epoch": 0.36184971098265895, + "grad_norm": 0.3292849028870402, + "learning_rate": 0.00016107873357880384, + "loss": 1.0175, + "step": 1565 + }, + { + "epoch": 0.3630057803468208, + "grad_norm": 0.31062236776235347, + "learning_rate": 0.00016075867168079507, + "loss": 0.9696, + "step": 1570 + }, + { + "epoch": 0.36416184971098264, + "grad_norm": 0.318894494073892, + "learning_rate": 0.00016043762008688433, + "loss": 0.9286, + "step": 1575 + }, + { + "epoch": 0.3653179190751445, + "grad_norm": 0.3029890310755476, + "learning_rate": 0.00016011558402666983, + "loss": 0.9594, + "step": 1580 + }, + { + "epoch": 0.36647398843930634, + "grad_norm": 0.3081821304700694, + "learning_rate": 0.00015979256874578594, + "loss": 0.987, + "step": 1585 + }, + { + "epoch": 0.3676300578034682, + "grad_norm": 0.29222596605397133, + "learning_rate": 0.00015946857950581734, + "loss": 0.9919, + "step": 1590 + }, + { + "epoch": 0.36878612716763004, + "grad_norm": 0.2955787813149893, + "learning_rate": 0.0001591436215842135, + "loss": 0.9653, + "step": 1595 + }, + { + "epoch": 0.3699421965317919, + "grad_norm": 0.2991599664341822, + "learning_rate": 0.0001588177002742029, + "loss": 0.9874, + "step": 1600 + }, + { + "epoch": 0.37109826589595374, + "grad_norm": 0.44600654437638615, + "learning_rate": 0.00015849082088470638, + "loss": 0.9504, + "step": 1605 + }, + { + "epoch": 0.3722543352601156, + "grad_norm": 0.33315550683583905, + "learning_rate": 0.00015816298874025102, + "loss": 1.0328, + "step": 1610 + }, + { + "epoch": 0.37341040462427744, + "grad_norm": 0.292061245143086, + "learning_rate": 0.00015783420918088337, + "loss": 0.9762, + "step": 1615 + }, + { + "epoch": 0.3745664739884393, + "grad_norm": 0.31976882089395187, + "learning_rate": 0.0001575044875620822, + "loss": 1.026, + "step": 1620 + }, + { + "epoch": 0.37572254335260113, + "grad_norm": 0.3238783922087859, + "learning_rate": 0.0001571738292546716, + "loss": 0.9496, + "step": 1625 + }, + { + "epoch": 0.376878612716763, + "grad_norm": 0.2959033885199569, + "learning_rate": 0.00015684223964473337, + "loss": 0.9656, + "step": 1630 + }, + { + "epoch": 0.37803468208092483, + "grad_norm": 0.3043686833702477, + "learning_rate": 0.0001565097241335191, + "loss": 0.954, + "step": 1635 + }, + { + "epoch": 0.3791907514450867, + "grad_norm": 0.3015192978319062, + "learning_rate": 0.00015617628813736247, + "loss": 0.9908, + "step": 1640 + }, + { + "epoch": 0.38034682080924853, + "grad_norm": 0.3082482576635595, + "learning_rate": 0.00015584193708759094, + "loss": 0.9477, + "step": 1645 + }, + { + "epoch": 0.3815028901734104, + "grad_norm": 0.30690197747994147, + "learning_rate": 0.00015550667643043716, + "loss": 0.9547, + "step": 1650 + }, + { + "epoch": 0.38265895953757223, + "grad_norm": 0.32095813086552044, + "learning_rate": 0.0001551705116269504, + "loss": 0.9946, + "step": 1655 + }, + { + "epoch": 0.3838150289017341, + "grad_norm": 0.3102088056503803, + "learning_rate": 0.0001548334481529075, + "loss": 0.9755, + "step": 1660 + }, + { + "epoch": 0.38497109826589593, + "grad_norm": 0.31740424344750273, + "learning_rate": 0.00015449549149872376, + "loss": 0.986, + "step": 1665 + }, + { + "epoch": 0.3861271676300578, + "grad_norm": 0.29760230294640583, + "learning_rate": 0.00015415664716936345, + "loss": 0.9736, + "step": 1670 + }, + { + "epoch": 0.3872832369942196, + "grad_norm": 0.33696332230509884, + "learning_rate": 0.00015381692068425004, + "loss": 0.9833, + "step": 1675 + }, + { + "epoch": 0.3884393063583815, + "grad_norm": 0.2971442840267895, + "learning_rate": 0.0001534763175771766, + "loss": 0.9787, + "step": 1680 + }, + { + "epoch": 0.3895953757225434, + "grad_norm": 0.2985513113352265, + "learning_rate": 0.00015313484339621534, + "loss": 0.9586, + "step": 1685 + }, + { + "epoch": 0.39075144508670523, + "grad_norm": 0.3070566835370781, + "learning_rate": 0.00015279250370362735, + "loss": 0.9878, + "step": 1690 + }, + { + "epoch": 0.3919075144508671, + "grad_norm": 0.3277524257534511, + "learning_rate": 0.00015244930407577205, + "loss": 1.0016, + "step": 1695 + }, + { + "epoch": 0.3930635838150289, + "grad_norm": 0.30050526813256595, + "learning_rate": 0.00015210525010301638, + "loss": 0.9553, + "step": 1700 + }, + { + "epoch": 0.3942196531791908, + "grad_norm": 0.30810810818757023, + "learning_rate": 0.0001517603473896435, + "loss": 0.9559, + "step": 1705 + }, + { + "epoch": 0.3953757225433526, + "grad_norm": 0.299950087627466, + "learning_rate": 0.00015141460155376182, + "loss": 0.9609, + "step": 1710 + }, + { + "epoch": 0.3965317919075145, + "grad_norm": 0.2925418247045739, + "learning_rate": 0.00015106801822721338, + "loss": 0.9763, + "step": 1715 + }, + { + "epoch": 0.3976878612716763, + "grad_norm": 0.30196182567060115, + "learning_rate": 0.00015072060305548187, + "loss": 0.9959, + "step": 1720 + }, + { + "epoch": 0.3988439306358382, + "grad_norm": 0.30989108180452857, + "learning_rate": 0.0001503723616976011, + "loss": 1.0003, + "step": 1725 + }, + { + "epoch": 0.4, + "grad_norm": 0.3039991146331331, + "learning_rate": 0.00015002329982606255, + "loss": 1.0345, + "step": 1730 + }, + { + "epoch": 0.40115606936416187, + "grad_norm": 0.3211973783178471, + "learning_rate": 0.00014967342312672283, + "loss": 0.9384, + "step": 1735 + }, + { + "epoch": 0.4023121387283237, + "grad_norm": 0.29388068969488124, + "learning_rate": 0.00014932273729871152, + "loss": 0.9051, + "step": 1740 + }, + { + "epoch": 0.40346820809248557, + "grad_norm": 0.3265810444042218, + "learning_rate": 0.0001489712480543379, + "loss": 0.9835, + "step": 1745 + }, + { + "epoch": 0.4046242774566474, + "grad_norm": 0.303317950576793, + "learning_rate": 0.0001486189611189981, + "loss": 0.9446, + "step": 1750 + }, + { + "epoch": 0.40578034682080927, + "grad_norm": 0.2967360524243329, + "learning_rate": 0.00014826588223108185, + "loss": 0.9908, + "step": 1755 + }, + { + "epoch": 0.4069364161849711, + "grad_norm": 0.30908569033672595, + "learning_rate": 0.00014791201714187897, + "loss": 0.9118, + "step": 1760 + }, + { + "epoch": 0.40809248554913297, + "grad_norm": 0.2921540136185523, + "learning_rate": 0.0001475573716154856, + "loss": 1.0177, + "step": 1765 + }, + { + "epoch": 0.4092485549132948, + "grad_norm": 0.30891513255558445, + "learning_rate": 0.00014720195142871054, + "loss": 0.9528, + "step": 1770 + }, + { + "epoch": 0.41040462427745666, + "grad_norm": 0.3116110416347837, + "learning_rate": 0.00014684576237098082, + "loss": 1.0153, + "step": 1775 + }, + { + "epoch": 0.4115606936416185, + "grad_norm": 0.29853304778547163, + "learning_rate": 0.00014648881024424774, + "loss": 0.9607, + "step": 1780 + }, + { + "epoch": 0.41271676300578036, + "grad_norm": 0.29621019258375375, + "learning_rate": 0.00014613110086289218, + "loss": 1.0178, + "step": 1785 + }, + { + "epoch": 0.4138728323699422, + "grad_norm": 0.32202847876721696, + "learning_rate": 0.00014577264005362985, + "loss": 0.9274, + "step": 1790 + }, + { + "epoch": 0.41502890173410406, + "grad_norm": 0.30619120168198916, + "learning_rate": 0.00014541343365541645, + "loss": 0.9435, + "step": 1795 + }, + { + "epoch": 0.4161849710982659, + "grad_norm": 0.3001554889949122, + "learning_rate": 0.00014505348751935263, + "loss": 0.9738, + "step": 1800 + }, + { + "epoch": 0.41734104046242776, + "grad_norm": 0.3065569332725715, + "learning_rate": 0.00014469280750858854, + "loss": 0.9627, + "step": 1805 + }, + { + "epoch": 0.4184971098265896, + "grad_norm": 0.3346803326675102, + "learning_rate": 0.00014433139949822837, + "loss": 1.0008, + "step": 1810 + }, + { + "epoch": 0.41965317919075146, + "grad_norm": 0.30285946590074797, + "learning_rate": 0.00014396926937523477, + "loss": 0.9681, + "step": 1815 + }, + { + "epoch": 0.4208092485549133, + "grad_norm": 0.323926429665197, + "learning_rate": 0.0001436064230383327, + "loss": 0.9883, + "step": 1820 + }, + { + "epoch": 0.42196531791907516, + "grad_norm": 0.30822208654391275, + "learning_rate": 0.00014324286639791367, + "loss": 0.9471, + "step": 1825 + }, + { + "epoch": 0.423121387283237, + "grad_norm": 0.3043728994137006, + "learning_rate": 0.00014287860537593917, + "loss": 0.9837, + "step": 1830 + }, + { + "epoch": 0.42427745664739885, + "grad_norm": 0.3042147218697011, + "learning_rate": 0.00014251364590584444, + "loss": 0.9576, + "step": 1835 + }, + { + "epoch": 0.4254335260115607, + "grad_norm": 0.32773524321463, + "learning_rate": 0.00014214799393244166, + "loss": 0.9356, + "step": 1840 + }, + { + "epoch": 0.42658959537572255, + "grad_norm": 0.30439292824288355, + "learning_rate": 0.00014178165541182312, + "loss": 0.9421, + "step": 1845 + }, + { + "epoch": 0.4277456647398844, + "grad_norm": 0.321382456218326, + "learning_rate": 0.00014141463631126442, + "loss": 0.9515, + "step": 1850 + }, + { + "epoch": 0.42890173410404625, + "grad_norm": 0.3130786437336031, + "learning_rate": 0.0001410469426091269, + "loss": 0.9715, + "step": 1855 + }, + { + "epoch": 0.4300578034682081, + "grad_norm": 0.3135399317680074, + "learning_rate": 0.00014067858029476063, + "loss": 0.9474, + "step": 1860 + }, + { + "epoch": 0.43121387283236995, + "grad_norm": 0.31368812931362966, + "learning_rate": 0.00014030955536840656, + "loss": 1.0225, + "step": 1865 + }, + { + "epoch": 0.4323699421965318, + "grad_norm": 0.3332336708705887, + "learning_rate": 0.00013993987384109898, + "loss": 1.0098, + "step": 1870 + }, + { + "epoch": 0.43352601156069365, + "grad_norm": 0.2950594206550405, + "learning_rate": 0.00013956954173456747, + "loss": 0.9846, + "step": 1875 + }, + { + "epoch": 0.4346820809248555, + "grad_norm": 0.3001574947847011, + "learning_rate": 0.000139198565081139, + "loss": 0.9853, + "step": 1880 + }, + { + "epoch": 0.43583815028901735, + "grad_norm": 0.29311121484460284, + "learning_rate": 0.00013882694992363936, + "loss": 1.0175, + "step": 1885 + }, + { + "epoch": 0.4369942196531792, + "grad_norm": 0.3286518580874968, + "learning_rate": 0.00013845470231529502, + "loss": 0.9845, + "step": 1890 + }, + { + "epoch": 0.43815028901734104, + "grad_norm": 0.31669776510548286, + "learning_rate": 0.00013808182831963442, + "loss": 1.0096, + "step": 1895 + }, + { + "epoch": 0.4393063583815029, + "grad_norm": 0.3230500170692119, + "learning_rate": 0.00013770833401038912, + "loss": 0.9652, + "step": 1900 + }, + { + "epoch": 0.44046242774566474, + "grad_norm": 0.314774115101565, + "learning_rate": 0.0001373342254713951, + "loss": 0.9884, + "step": 1905 + }, + { + "epoch": 0.4416184971098266, + "grad_norm": 0.309673258526753, + "learning_rate": 0.00013695950879649338, + "loss": 0.9617, + "step": 1910 + }, + { + "epoch": 0.44277456647398844, + "grad_norm": 0.31688401964004287, + "learning_rate": 0.00013658419008943088, + "loss": 1.0007, + "step": 1915 + }, + { + "epoch": 0.4439306358381503, + "grad_norm": 0.3115831394799577, + "learning_rate": 0.00013620827546376112, + "loss": 0.9837, + "step": 1920 + }, + { + "epoch": 0.44508670520231214, + "grad_norm": 0.3134574472279371, + "learning_rate": 0.00013583177104274435, + "loss": 0.9748, + "step": 1925 + }, + { + "epoch": 0.446242774566474, + "grad_norm": 0.3388852510725773, + "learning_rate": 0.00013545468295924812, + "loss": 0.9825, + "step": 1930 + }, + { + "epoch": 0.44739884393063584, + "grad_norm": 0.33068272114069625, + "learning_rate": 0.00013507701735564716, + "loss": 0.9552, + "step": 1935 + }, + { + "epoch": 0.4485549132947977, + "grad_norm": 0.3070498417777377, + "learning_rate": 0.00013469878038372348, + "loss": 0.9842, + "step": 1940 + }, + { + "epoch": 0.44971098265895953, + "grad_norm": 0.32088136083821, + "learning_rate": 0.00013431997820456592, + "loss": 0.9635, + "step": 1945 + }, + { + "epoch": 0.4508670520231214, + "grad_norm": 0.317988833081837, + "learning_rate": 0.00013394061698847022, + "loss": 0.9922, + "step": 1950 + }, + { + "epoch": 0.45202312138728323, + "grad_norm": 0.29873515381181037, + "learning_rate": 0.000133560702914838, + "loss": 0.9808, + "step": 1955 + }, + { + "epoch": 0.4531791907514451, + "grad_norm": 0.334307242594275, + "learning_rate": 0.00013318024217207652, + "loss": 0.9285, + "step": 1960 + }, + { + "epoch": 0.45433526011560693, + "grad_norm": 0.3161167520009514, + "learning_rate": 0.00013279924095749768, + "loss": 0.9721, + "step": 1965 + }, + { + "epoch": 0.4554913294797688, + "grad_norm": 0.3219859959358273, + "learning_rate": 0.00013241770547721703, + "loss": 1.008, + "step": 1970 + }, + { + "epoch": 0.45664739884393063, + "grad_norm": 0.27935961874420406, + "learning_rate": 0.00013203564194605284, + "loss": 0.9502, + "step": 1975 + }, + { + "epoch": 0.4578034682080925, + "grad_norm": 0.29913347538052254, + "learning_rate": 0.0001316530565874248, + "loss": 0.9791, + "step": 1980 + }, + { + "epoch": 0.45895953757225433, + "grad_norm": 0.30294269390330414, + "learning_rate": 0.00013126995563325254, + "loss": 0.9763, + "step": 1985 + }, + { + "epoch": 0.4601156069364162, + "grad_norm": 0.32096224239126736, + "learning_rate": 0.00013088634532385424, + "loss": 0.9238, + "step": 1990 + }, + { + "epoch": 0.461271676300578, + "grad_norm": 0.2960586714061201, + "learning_rate": 0.000130502231907845, + "loss": 0.9533, + "step": 1995 + }, + { + "epoch": 0.4624277456647399, + "grad_norm": 0.29803737045431256, + "learning_rate": 0.000130117621642035, + "loss": 0.9526, + "step": 2000 + }, + { + "epoch": 0.4635838150289017, + "grad_norm": 0.31720967984437226, + "learning_rate": 0.00012973252079132749, + "loss": 0.9566, + "step": 2005 + }, + { + "epoch": 0.4647398843930636, + "grad_norm": 0.31204560106706253, + "learning_rate": 0.00012934693562861692, + "loss": 0.9821, + "step": 2010 + }, + { + "epoch": 0.4658959537572254, + "grad_norm": 0.3452586478382497, + "learning_rate": 0.00012896087243468673, + "loss": 0.9866, + "step": 2015 + }, + { + "epoch": 0.46705202312138727, + "grad_norm": 0.30419391343270963, + "learning_rate": 0.00012857433749810691, + "loss": 0.9465, + "step": 2020 + }, + { + "epoch": 0.4682080924855491, + "grad_norm": 0.302216494494177, + "learning_rate": 0.00012818733711513164, + "loss": 0.9928, + "step": 2025 + }, + { + "epoch": 0.46936416184971097, + "grad_norm": 0.29660145267520094, + "learning_rate": 0.00012779987758959683, + "loss": 0.9714, + "step": 2030 + }, + { + "epoch": 0.4705202312138728, + "grad_norm": 0.3375993332751583, + "learning_rate": 0.00012741196523281728, + "loss": 1.004, + "step": 2035 + }, + { + "epoch": 0.47167630057803467, + "grad_norm": 0.31685124172490736, + "learning_rate": 0.0001270236063634839, + "loss": 0.9686, + "step": 2040 + }, + { + "epoch": 0.4728323699421965, + "grad_norm": 0.30517277761996336, + "learning_rate": 0.00012663480730756095, + "loss": 0.97, + "step": 2045 + }, + { + "epoch": 0.47398843930635837, + "grad_norm": 0.3075134986191579, + "learning_rate": 0.00012624557439818275, + "loss": 0.9535, + "step": 2050 + }, + { + "epoch": 0.4751445086705202, + "grad_norm": 0.2914116111037525, + "learning_rate": 0.00012585591397555078, + "loss": 0.9549, + "step": 2055 + }, + { + "epoch": 0.47630057803468207, + "grad_norm": 0.3065733883077486, + "learning_rate": 0.00012546583238683015, + "loss": 0.9694, + "step": 2060 + }, + { + "epoch": 0.4774566473988439, + "grad_norm": 0.30076466916700556, + "learning_rate": 0.00012507533598604632, + "loss": 0.9802, + "step": 2065 + }, + { + "epoch": 0.47861271676300576, + "grad_norm": 0.29670240314259055, + "learning_rate": 0.00012468443113398175, + "loss": 0.9366, + "step": 2070 + }, + { + "epoch": 0.4797687861271676, + "grad_norm": 0.31183074363125884, + "learning_rate": 0.00012429312419807198, + "loss": 0.966, + "step": 2075 + }, + { + "epoch": 0.48092485549132946, + "grad_norm": 0.31278790481596425, + "learning_rate": 0.00012390142155230217, + "loss": 0.9893, + "step": 2080 + }, + { + "epoch": 0.4820809248554913, + "grad_norm": 0.28207826631174193, + "learning_rate": 0.0001235093295771032, + "loss": 0.9472, + "step": 2085 + }, + { + "epoch": 0.48323699421965316, + "grad_norm": 0.32635523820738965, + "learning_rate": 0.00012311685465924774, + "loss": 0.9089, + "step": 2090 + }, + { + "epoch": 0.484393063583815, + "grad_norm": 0.2977916871662523, + "learning_rate": 0.00012272400319174607, + "loss": 0.9834, + "step": 2095 + }, + { + "epoch": 0.48554913294797686, + "grad_norm": 0.31990291106992935, + "learning_rate": 0.00012233078157374217, + "loss": 0.9312, + "step": 2100 + }, + { + "epoch": 0.4867052023121387, + "grad_norm": 0.33073575866363214, + "learning_rate": 0.00012193719621040942, + "loss": 0.9795, + "step": 2105 + }, + { + "epoch": 0.48786127167630056, + "grad_norm": 0.2915838062263623, + "learning_rate": 0.00012154325351284618, + "loss": 0.9789, + "step": 2110 + }, + { + "epoch": 0.4890173410404624, + "grad_norm": 0.33822747094942934, + "learning_rate": 0.00012114895989797144, + "loss": 0.9304, + "step": 2115 + }, + { + "epoch": 0.49017341040462425, + "grad_norm": 0.32370602693562334, + "learning_rate": 0.00012075432178842021, + "loss": 0.9428, + "step": 2120 + }, + { + "epoch": 0.4913294797687861, + "grad_norm": 0.32375527811459415, + "learning_rate": 0.00012035934561243905, + "loss": 0.9718, + "step": 2125 + }, + { + "epoch": 0.49248554913294795, + "grad_norm": 0.2943351586407433, + "learning_rate": 0.00011996403780378123, + "loss": 0.9712, + "step": 2130 + }, + { + "epoch": 0.4936416184971098, + "grad_norm": 0.3223304229208655, + "learning_rate": 0.00011956840480160194, + "loss": 1.0046, + "step": 2135 + }, + { + "epoch": 0.49479768786127165, + "grad_norm": 0.3032804365412004, + "learning_rate": 0.00011917245305035354, + "loss": 0.9596, + "step": 2140 + }, + { + "epoch": 0.4959537572254335, + "grad_norm": 0.3086608217360584, + "learning_rate": 0.00011877618899968037, + "loss": 0.9473, + "step": 2145 + }, + { + "epoch": 0.49710982658959535, + "grad_norm": 0.3131187277547376, + "learning_rate": 0.00011837961910431383, + "loss": 1.0065, + "step": 2150 + }, + { + "epoch": 0.4982658959537572, + "grad_norm": 0.2975605931413944, + "learning_rate": 0.00011798274982396726, + "loss": 0.9481, + "step": 2155 + }, + { + "epoch": 0.49942196531791905, + "grad_norm": 0.29498275298057963, + "learning_rate": 0.00011758558762323067, + "loss": 0.9884, + "step": 2160 + }, + { + "epoch": 0.500578034682081, + "grad_norm": 0.2906355713880922, + "learning_rate": 0.00011718813897146535, + "loss": 0.9643, + "step": 2165 + }, + { + "epoch": 0.5017341040462427, + "grad_norm": 0.2879177716708955, + "learning_rate": 0.00011679041034269869, + "loss": 0.9496, + "step": 2170 + }, + { + "epoch": 0.5028901734104047, + "grad_norm": 0.3107002014222183, + "learning_rate": 0.00011639240821551858, + "loss": 0.9489, + "step": 2175 + }, + { + "epoch": 0.5040462427745664, + "grad_norm": 0.30854297451303886, + "learning_rate": 0.00011599413907296785, + "loss": 0.9887, + "step": 2180 + }, + { + "epoch": 0.5052023121387283, + "grad_norm": 0.3250596343211611, + "learning_rate": 0.00011559560940243888, + "loss": 0.9421, + "step": 2185 + }, + { + "epoch": 0.5063583815028901, + "grad_norm": 0.303124134082483, + "learning_rate": 0.00011519682569556758, + "loss": 0.967, + "step": 2190 + }, + { + "epoch": 0.507514450867052, + "grad_norm": 0.29292499319855175, + "learning_rate": 0.00011479779444812808, + "loss": 0.9679, + "step": 2195 + }, + { + "epoch": 0.5086705202312138, + "grad_norm": 0.30291810874235703, + "learning_rate": 0.00011439852215992647, + "loss": 0.997, + "step": 2200 + }, + { + "epoch": 0.5098265895953757, + "grad_norm": 0.3234308605182878, + "learning_rate": 0.0001139990153346953, + "loss": 0.9876, + "step": 2205 + }, + { + "epoch": 0.5109826589595375, + "grad_norm": 0.3137214941805028, + "learning_rate": 0.00011359928047998744, + "loss": 1.0407, + "step": 2210 + }, + { + "epoch": 0.5121387283236994, + "grad_norm": 0.32171251618436913, + "learning_rate": 0.0001131993241070701, + "loss": 0.9783, + "step": 2215 + }, + { + "epoch": 0.5132947976878612, + "grad_norm": 0.2971713793214721, + "learning_rate": 0.00011279915273081876, + "loss": 0.9678, + "step": 2220 + }, + { + "epoch": 0.5144508670520231, + "grad_norm": 0.30876666041737444, + "learning_rate": 0.00011239877286961122, + "loss": 0.9717, + "step": 2225 + }, + { + "epoch": 0.5156069364161849, + "grad_norm": 0.31611543232380335, + "learning_rate": 0.00011199819104522114, + "loss": 0.9611, + "step": 2230 + }, + { + "epoch": 0.5167630057803468, + "grad_norm": 0.31365800007794736, + "learning_rate": 0.000111597413782712, + "loss": 0.986, + "step": 2235 + }, + { + "epoch": 0.5179190751445086, + "grad_norm": 0.3387876838248837, + "learning_rate": 0.00011119644761033078, + "loss": 0.9865, + "step": 2240 + }, + { + "epoch": 0.5190751445086705, + "grad_norm": 0.3090392049931908, + "learning_rate": 0.00011079529905940163, + "loss": 0.9264, + "step": 2245 + }, + { + "epoch": 0.5202312138728323, + "grad_norm": 0.30547601371038785, + "learning_rate": 0.0001103939746642194, + "loss": 0.9293, + "step": 2250 + }, + { + "epoch": 0.5213872832369942, + "grad_norm": 0.30920860300711217, + "learning_rate": 0.00010999248096194326, + "loss": 0.9759, + "step": 2255 + }, + { + "epoch": 0.522543352601156, + "grad_norm": 0.30207153503695156, + "learning_rate": 0.00010959082449249026, + "loss": 0.9557, + "step": 2260 + }, + { + "epoch": 0.5236994219653179, + "grad_norm": 0.29504681849985004, + "learning_rate": 0.00010918901179842877, + "loss": 0.9686, + "step": 2265 + }, + { + "epoch": 0.5248554913294797, + "grad_norm": 0.29286119267320176, + "learning_rate": 0.00010878704942487183, + "loss": 1.0042, + "step": 2270 + }, + { + "epoch": 0.5260115606936416, + "grad_norm": 0.3062243965654378, + "learning_rate": 0.00010838494391937064, + "loss": 0.9784, + "step": 2275 + }, + { + "epoch": 0.5271676300578034, + "grad_norm": 0.3079900592590067, + "learning_rate": 0.00010798270183180794, + "loss": 0.9503, + "step": 2280 + }, + { + "epoch": 0.5283236994219653, + "grad_norm": 0.3154259872984876, + "learning_rate": 0.0001075803297142911, + "loss": 0.9509, + "step": 2285 + }, + { + "epoch": 0.5294797687861271, + "grad_norm": 0.29229323286742326, + "learning_rate": 0.00010717783412104568, + "loss": 0.9557, + "step": 2290 + }, + { + "epoch": 0.530635838150289, + "grad_norm": 0.31601404726519217, + "learning_rate": 0.00010677522160830848, + "loss": 0.9042, + "step": 2295 + }, + { + "epoch": 0.5317919075144508, + "grad_norm": 0.28505318471127816, + "learning_rate": 0.00010637249873422077, + "loss": 0.9692, + "step": 2300 + }, + { + "epoch": 0.5329479768786127, + "grad_norm": 0.32485305181812835, + "learning_rate": 0.00010596967205872154, + "loss": 1.0065, + "step": 2305 + }, + { + "epoch": 0.5341040462427745, + "grad_norm": 0.3078039739362063, + "learning_rate": 0.00010556674814344059, + "loss": 0.9284, + "step": 2310 + }, + { + "epoch": 0.5352601156069364, + "grad_norm": 0.3208549342399588, + "learning_rate": 0.00010516373355159159, + "loss": 0.9477, + "step": 2315 + }, + { + "epoch": 0.5364161849710982, + "grad_norm": 0.30748525424656054, + "learning_rate": 0.00010476063484786535, + "loss": 0.9629, + "step": 2320 + }, + { + "epoch": 0.5375722543352601, + "grad_norm": 0.35275688331568944, + "learning_rate": 0.0001043574585983227, + "loss": 1.0113, + "step": 2325 + }, + { + "epoch": 0.5387283236994219, + "grad_norm": 0.3253346439794717, + "learning_rate": 0.00010395421137028761, + "loss": 1.0346, + "step": 2330 + }, + { + "epoch": 0.5398843930635838, + "grad_norm": 0.3095697055293057, + "learning_rate": 0.00010355089973224026, + "loss": 0.9546, + "step": 2335 + }, + { + "epoch": 0.5410404624277456, + "grad_norm": 0.307599147895419, + "learning_rate": 0.00010314753025370991, + "loss": 0.9836, + "step": 2340 + }, + { + "epoch": 0.5421965317919075, + "grad_norm": 0.31630009518917934, + "learning_rate": 0.00010274410950516815, + "loss": 1.0071, + "step": 2345 + }, + { + "epoch": 0.5433526011560693, + "grad_norm": 0.2834312244927633, + "learning_rate": 0.00010234064405792154, + "loss": 0.9489, + "step": 2350 + }, + { + "epoch": 0.5445086705202312, + "grad_norm": 0.2970129068571585, + "learning_rate": 0.0001019371404840048, + "loss": 0.9351, + "step": 2355 + }, + { + "epoch": 0.545664739884393, + "grad_norm": 0.3025660920447834, + "learning_rate": 0.0001015336053560737, + "loss": 0.99, + "step": 2360 + }, + { + "epoch": 0.5468208092485549, + "grad_norm": 0.30633618369872967, + "learning_rate": 0.00010113004524729799, + "loss": 0.954, + "step": 2365 + }, + { + "epoch": 0.5479768786127167, + "grad_norm": 0.28629344395996326, + "learning_rate": 0.00010072646673125432, + "loss": 0.9783, + "step": 2370 + }, + { + "epoch": 0.5491329479768786, + "grad_norm": 0.3097159968894246, + "learning_rate": 0.00010032287638181919, + "loss": 1.0303, + "step": 2375 + }, + { + "epoch": 0.5502890173410404, + "grad_norm": 0.30322928513172714, + "learning_rate": 9.991928077306183e-05, + "loss": 1.0, + "step": 2380 + }, + { + "epoch": 0.5514450867052023, + "grad_norm": 0.4644501347952281, + "learning_rate": 9.951568647913718e-05, + "loss": 0.9294, + "step": 2385 + }, + { + "epoch": 0.5526011560693641, + "grad_norm": 0.3124185608435605, + "learning_rate": 9.911210007417869e-05, + "loss": 0.9847, + "step": 2390 + }, + { + "epoch": 0.553757225433526, + "grad_norm": 0.32991063517383223, + "learning_rate": 9.870852813219143e-05, + "loss": 0.9755, + "step": 2395 + }, + { + "epoch": 0.5549132947976878, + "grad_norm": 0.31527757887354024, + "learning_rate": 9.830497722694478e-05, + "loss": 0.9819, + "step": 2400 + }, + { + "epoch": 0.5560693641618497, + "grad_norm": 0.3094199650313883, + "learning_rate": 9.790145393186541e-05, + "loss": 0.9409, + "step": 2405 + }, + { + "epoch": 0.5572254335260116, + "grad_norm": 0.3055144703983288, + "learning_rate": 9.749796481993042e-05, + "loss": 0.9674, + "step": 2410 + }, + { + "epoch": 0.5583815028901734, + "grad_norm": 0.31607857296755776, + "learning_rate": 9.709451646355996e-05, + "loss": 1.0174, + "step": 2415 + }, + { + "epoch": 0.5595375722543353, + "grad_norm": 0.2975421743209585, + "learning_rate": 9.669111543451033e-05, + "loss": 0.9683, + "step": 2420 + }, + { + "epoch": 0.5606936416184971, + "grad_norm": 0.30435556233851097, + "learning_rate": 9.628776830376698e-05, + "loss": 0.9539, + "step": 2425 + }, + { + "epoch": 0.561849710982659, + "grad_norm": 0.31164699307502913, + "learning_rate": 9.588448164143739e-05, + "loss": 0.9716, + "step": 2430 + }, + { + "epoch": 0.5630057803468208, + "grad_norm": 0.3029837051614153, + "learning_rate": 9.548126201664398e-05, + "loss": 0.9488, + "step": 2435 + }, + { + "epoch": 0.5641618497109827, + "grad_norm": 0.30375216989550236, + "learning_rate": 9.507811599741735e-05, + "loss": 0.9241, + "step": 2440 + }, + { + "epoch": 0.5653179190751445, + "grad_norm": 0.3044100811963364, + "learning_rate": 9.467505015058901e-05, + "loss": 1.0212, + "step": 2445 + }, + { + "epoch": 0.5664739884393064, + "grad_norm": 0.3079385379059537, + "learning_rate": 9.427207104168467e-05, + "loss": 0.9341, + "step": 2450 + }, + { + "epoch": 0.5676300578034682, + "grad_norm": 0.3012241533223519, + "learning_rate": 9.386918523481709e-05, + "loss": 0.9533, + "step": 2455 + }, + { + "epoch": 0.5687861271676301, + "grad_norm": 0.32392283471513106, + "learning_rate": 9.346639929257916e-05, + "loss": 0.9888, + "step": 2460 + }, + { + "epoch": 0.5699421965317919, + "grad_norm": 0.28835183242553575, + "learning_rate": 9.306371977593726e-05, + "loss": 0.9847, + "step": 2465 + }, + { + "epoch": 0.5710982658959538, + "grad_norm": 0.29592049768177997, + "learning_rate": 9.26611532441241e-05, + "loss": 0.9075, + "step": 2470 + }, + { + "epoch": 0.5722543352601156, + "grad_norm": 0.30510640464990846, + "learning_rate": 9.225870625453192e-05, + "loss": 0.9276, + "step": 2475 + }, + { + "epoch": 0.5734104046242775, + "grad_norm": 0.32217649840985024, + "learning_rate": 9.18563853626059e-05, + "loss": 0.9936, + "step": 2480 + }, + { + "epoch": 0.5745664739884393, + "grad_norm": 0.2945778811908368, + "learning_rate": 9.145419712173713e-05, + "loss": 0.9775, + "step": 2485 + }, + { + "epoch": 0.5757225433526012, + "grad_norm": 0.2937161972776823, + "learning_rate": 9.105214808315588e-05, + "loss": 0.9293, + "step": 2490 + }, + { + "epoch": 0.576878612716763, + "grad_norm": 0.34400282833354867, + "learning_rate": 9.065024479582513e-05, + "loss": 1.0045, + "step": 2495 + }, + { + "epoch": 0.5780346820809249, + "grad_norm": 0.30233509314710183, + "learning_rate": 9.024849380633359e-05, + "loss": 0.9786, + "step": 2500 + }, + { + "epoch": 0.5791907514450867, + "grad_norm": 0.29454389871868664, + "learning_rate": 8.984690165878921e-05, + "loss": 0.9584, + "step": 2505 + }, + { + "epoch": 0.5803468208092486, + "grad_norm": 0.30071228762049557, + "learning_rate": 8.944547489471265e-05, + "loss": 0.955, + "step": 2510 + }, + { + "epoch": 0.5815028901734104, + "grad_norm": 0.3169803125653554, + "learning_rate": 8.904422005293052e-05, + "loss": 1.0198, + "step": 2515 + }, + { + "epoch": 0.5826589595375723, + "grad_norm": 0.31831872317303483, + "learning_rate": 8.864314366946913e-05, + "loss": 0.9781, + "step": 2520 + }, + { + "epoch": 0.5838150289017341, + "grad_norm": 0.30646509617401063, + "learning_rate": 8.824225227744782e-05, + "loss": 0.9556, + "step": 2525 + }, + { + "epoch": 0.584971098265896, + "grad_norm": 0.28426966152836436, + "learning_rate": 8.784155240697254e-05, + "loss": 0.9811, + "step": 2530 + }, + { + "epoch": 0.5861271676300578, + "grad_norm": 0.29076631427987554, + "learning_rate": 8.74410505850297e-05, + "loss": 0.9653, + "step": 2535 + }, + { + "epoch": 0.5872832369942197, + "grad_norm": 0.2993968457362386, + "learning_rate": 8.704075333537963e-05, + "loss": 0.9267, + "step": 2540 + }, + { + "epoch": 0.5884393063583815, + "grad_norm": 0.28060562517633875, + "learning_rate": 8.66406671784503e-05, + "loss": 0.9767, + "step": 2545 + }, + { + "epoch": 0.5895953757225434, + "grad_norm": 0.29394163379907895, + "learning_rate": 8.624079863123135e-05, + "loss": 0.9692, + "step": 2550 + }, + { + "epoch": 0.5907514450867052, + "grad_norm": 0.29752047867631554, + "learning_rate": 8.584115420716777e-05, + "loss": 1.0218, + "step": 2555 + }, + { + "epoch": 0.5919075144508671, + "grad_norm": 0.2960489971126818, + "learning_rate": 8.544174041605363e-05, + "loss": 0.9386, + "step": 2560 + }, + { + "epoch": 0.5930635838150289, + "grad_norm": 0.2991003815095411, + "learning_rate": 8.504256376392647e-05, + "loss": 0.951, + "step": 2565 + }, + { + "epoch": 0.5942196531791908, + "grad_norm": 0.2875329275503883, + "learning_rate": 8.464363075296095e-05, + "loss": 0.9595, + "step": 2570 + }, + { + "epoch": 0.5953757225433526, + "grad_norm": 0.29629340379738117, + "learning_rate": 8.424494788136303e-05, + "loss": 0.946, + "step": 2575 + }, + { + "epoch": 0.5965317919075145, + "grad_norm": 0.3039710490176798, + "learning_rate": 8.384652164326432e-05, + "loss": 0.9297, + "step": 2580 + }, + { + "epoch": 0.5976878612716763, + "grad_norm": 0.2867197601664578, + "learning_rate": 8.344835852861595e-05, + "loss": 0.9655, + "step": 2585 + }, + { + "epoch": 0.5988439306358382, + "grad_norm": 0.3021141932586307, + "learning_rate": 8.305046502308319e-05, + "loss": 0.9388, + "step": 2590 + }, + { + "epoch": 0.6, + "grad_norm": 0.32608595836134247, + "learning_rate": 8.265284760793957e-05, + "loss": 0.948, + "step": 2595 + }, + { + "epoch": 0.6011560693641619, + "grad_norm": 0.3077259465134535, + "learning_rate": 8.225551275996138e-05, + "loss": 1.0123, + "step": 2600 + }, + { + "epoch": 0.6023121387283237, + "grad_norm": 0.28311149702035393, + "learning_rate": 8.185846695132227e-05, + "loss": 0.9456, + "step": 2605 + }, + { + "epoch": 0.6034682080924856, + "grad_norm": 0.31276946079278556, + "learning_rate": 8.146171664948769e-05, + "loss": 0.9755, + "step": 2610 + }, + { + "epoch": 0.6046242774566474, + "grad_norm": 0.3264513747288462, + "learning_rate": 8.10652683171095e-05, + "loss": 0.9619, + "step": 2615 + }, + { + "epoch": 0.6057803468208093, + "grad_norm": 0.30285428517053464, + "learning_rate": 8.066912841192099e-05, + "loss": 0.9344, + "step": 2620 + }, + { + "epoch": 0.6069364161849711, + "grad_norm": 0.3644501014383341, + "learning_rate": 8.027330338663132e-05, + "loss": 0.9794, + "step": 2625 + }, + { + "epoch": 0.608092485549133, + "grad_norm": 0.28904372054279964, + "learning_rate": 7.987779968882061e-05, + "loss": 0.941, + "step": 2630 + }, + { + "epoch": 0.6092485549132948, + "grad_norm": 0.2895705859655337, + "learning_rate": 7.9482623760835e-05, + "loss": 0.951, + "step": 2635 + }, + { + "epoch": 0.6104046242774567, + "grad_norm": 0.29749813183970186, + "learning_rate": 7.908778203968146e-05, + "loss": 0.9244, + "step": 2640 + }, + { + "epoch": 0.6115606936416185, + "grad_norm": 0.31971185227728377, + "learning_rate": 7.869328095692312e-05, + "loss": 0.9645, + "step": 2645 + }, + { + "epoch": 0.6127167630057804, + "grad_norm": 0.2873929264676909, + "learning_rate": 7.829912693857454e-05, + "loss": 0.9739, + "step": 2650 + }, + { + "epoch": 0.6138728323699422, + "grad_norm": 0.2962385037449908, + "learning_rate": 7.79053264049968e-05, + "loss": 1.0025, + "step": 2655 + }, + { + "epoch": 0.6150289017341041, + "grad_norm": 0.2889472954071748, + "learning_rate": 7.751188577079327e-05, + "loss": 0.9764, + "step": 2660 + }, + { + "epoch": 0.6161849710982659, + "grad_norm": 0.2914115227852295, + "learning_rate": 7.711881144470481e-05, + "loss": 0.9575, + "step": 2665 + }, + { + "epoch": 0.6173410404624278, + "grad_norm": 0.31913799827615985, + "learning_rate": 7.672610982950546e-05, + "loss": 0.9684, + "step": 2670 + }, + { + "epoch": 0.6184971098265896, + "grad_norm": 0.3005365331976947, + "learning_rate": 7.633378732189833e-05, + "loss": 0.941, + "step": 2675 + }, + { + "epoch": 0.6196531791907515, + "grad_norm": 0.29855395413135466, + "learning_rate": 7.594185031241115e-05, + "loss": 0.9504, + "step": 2680 + }, + { + "epoch": 0.6208092485549133, + "grad_norm": 0.30079166608986607, + "learning_rate": 7.555030518529227e-05, + "loss": 0.9489, + "step": 2685 + }, + { + "epoch": 0.6219653179190752, + "grad_norm": 0.2835216761875895, + "learning_rate": 7.515915831840682e-05, + "loss": 1.0283, + "step": 2690 + }, + { + "epoch": 0.623121387283237, + "grad_norm": 0.296763030995749, + "learning_rate": 7.476841608313253e-05, + "loss": 0.9495, + "step": 2695 + }, + { + "epoch": 0.6242774566473989, + "grad_norm": 0.2965441757959662, + "learning_rate": 7.437808484425614e-05, + "loss": 0.9207, + "step": 2700 + }, + { + "epoch": 0.6254335260115607, + "grad_norm": 0.28563226791333124, + "learning_rate": 7.398817095986978e-05, + "loss": 0.9529, + "step": 2705 + }, + { + "epoch": 0.6265895953757226, + "grad_norm": 0.2958876716229884, + "learning_rate": 7.359868078126714e-05, + "loss": 0.9415, + "step": 2710 + }, + { + "epoch": 0.6277456647398844, + "grad_norm": 0.29947263183760775, + "learning_rate": 7.320962065284032e-05, + "loss": 0.9153, + "step": 2715 + }, + { + "epoch": 0.6289017341040463, + "grad_norm": 0.2921294166805471, + "learning_rate": 7.282099691197632e-05, + "loss": 1.0061, + "step": 2720 + }, + { + "epoch": 0.630057803468208, + "grad_norm": 0.3065806154787742, + "learning_rate": 7.243281588895374e-05, + "loss": 0.9713, + "step": 2725 + }, + { + "epoch": 0.63121387283237, + "grad_norm": 0.29670093707685785, + "learning_rate": 7.204508390683991e-05, + "loss": 0.9152, + "step": 2730 + }, + { + "epoch": 0.6323699421965318, + "grad_norm": 0.28334644508416545, + "learning_rate": 7.165780728138769e-05, + "loss": 0.9216, + "step": 2735 + }, + { + "epoch": 0.6335260115606937, + "grad_norm": 0.2983458223496864, + "learning_rate": 7.127099232093252e-05, + "loss": 0.9684, + "step": 2740 + }, + { + "epoch": 0.6346820809248555, + "grad_norm": 0.297707639389454, + "learning_rate": 7.08846453262901e-05, + "loss": 0.9677, + "step": 2745 + }, + { + "epoch": 0.6358381502890174, + "grad_norm": 0.3030445642630547, + "learning_rate": 7.049877259065312e-05, + "loss": 0.991, + "step": 2750 + }, + { + "epoch": 0.6369942196531792, + "grad_norm": 0.2967929219140181, + "learning_rate": 7.011338039948925e-05, + "loss": 0.9331, + "step": 2755 + }, + { + "epoch": 0.638150289017341, + "grad_norm": 0.30678542010472665, + "learning_rate": 6.972847503043864e-05, + "loss": 0.9952, + "step": 2760 + }, + { + "epoch": 0.6393063583815028, + "grad_norm": 0.2869522581890146, + "learning_rate": 6.934406275321147e-05, + "loss": 0.9817, + "step": 2765 + }, + { + "epoch": 0.6404624277456648, + "grad_norm": 0.30227078806254565, + "learning_rate": 6.896014982948602e-05, + "loss": 0.9713, + "step": 2770 + }, + { + "epoch": 0.6416184971098265, + "grad_norm": 0.29642633769940174, + "learning_rate": 6.857674251280671e-05, + "loss": 0.9495, + "step": 2775 + }, + { + "epoch": 0.6427745664739885, + "grad_norm": 0.3324961966488245, + "learning_rate": 6.819384704848199e-05, + "loss": 1.0679, + "step": 2780 + }, + { + "epoch": 0.6439306358381502, + "grad_norm": 0.2888690447316259, + "learning_rate": 6.781146967348284e-05, + "loss": 0.941, + "step": 2785 + }, + { + "epoch": 0.6450867052023121, + "grad_norm": 0.29579888471073373, + "learning_rate": 6.742961661634115e-05, + "loss": 1.0323, + "step": 2790 + }, + { + "epoch": 0.6462427745664739, + "grad_norm": 0.2989300466555907, + "learning_rate": 6.704829409704809e-05, + "loss": 0.9723, + "step": 2795 + }, + { + "epoch": 0.6473988439306358, + "grad_norm": 0.27150342274150757, + "learning_rate": 6.666750832695306e-05, + "loss": 0.9397, + "step": 2800 + }, + { + "epoch": 0.6485549132947976, + "grad_norm": 0.29998837345134693, + "learning_rate": 6.628726550866227e-05, + "loss": 0.9835, + "step": 2805 + }, + { + "epoch": 0.6497109826589595, + "grad_norm": 0.28399979161419353, + "learning_rate": 6.59075718359378e-05, + "loss": 0.9509, + "step": 2810 + }, + { + "epoch": 0.6508670520231213, + "grad_norm": 0.30370562141401386, + "learning_rate": 6.552843349359688e-05, + "loss": 0.967, + "step": 2815 + }, + { + "epoch": 0.6520231213872832, + "grad_norm": 0.3363515271195273, + "learning_rate": 6.514985665741073e-05, + "loss": 1.0054, + "step": 2820 + }, + { + "epoch": 0.653179190751445, + "grad_norm": 0.2941850764424752, + "learning_rate": 6.477184749400438e-05, + "loss": 0.9958, + "step": 2825 + }, + { + "epoch": 0.6543352601156069, + "grad_norm": 0.2845955247940358, + "learning_rate": 6.439441216075605e-05, + "loss": 0.9199, + "step": 2830 + }, + { + "epoch": 0.6554913294797687, + "grad_norm": 0.30264607648319775, + "learning_rate": 6.401755680569683e-05, + "loss": 0.9974, + "step": 2835 + }, + { + "epoch": 0.6566473988439306, + "grad_norm": 0.3049902419448789, + "learning_rate": 6.36412875674105e-05, + "loss": 0.9412, + "step": 2840 + }, + { + "epoch": 0.6578034682080924, + "grad_norm": 0.28427694016822674, + "learning_rate": 6.326561057493376e-05, + "loss": 0.9618, + "step": 2845 + }, + { + "epoch": 0.6589595375722543, + "grad_norm": 0.2923557343391648, + "learning_rate": 6.28905319476561e-05, + "loss": 0.9738, + "step": 2850 + }, + { + "epoch": 0.6601156069364161, + "grad_norm": 0.3011922931000182, + "learning_rate": 6.251605779522032e-05, + "loss": 0.9547, + "step": 2855 + }, + { + "epoch": 0.661271676300578, + "grad_norm": 0.2778141274186397, + "learning_rate": 6.214219421742295e-05, + "loss": 1.0062, + "step": 2860 + }, + { + "epoch": 0.6624277456647398, + "grad_norm": 0.28425833137988876, + "learning_rate": 6.176894730411483e-05, + "loss": 0.9529, + "step": 2865 + }, + { + "epoch": 0.6635838150289017, + "grad_norm": 0.2914968347637112, + "learning_rate": 6.139632313510212e-05, + "loss": 0.944, + "step": 2870 + }, + { + "epoch": 0.6647398843930635, + "grad_norm": 0.2910244571300189, + "learning_rate": 6.1024327780046944e-05, + "loss": 1.0063, + "step": 2875 + }, + { + "epoch": 0.6658959537572254, + "grad_norm": 0.28730856037506486, + "learning_rate": 6.065296729836879e-05, + "loss": 0.978, + "step": 2880 + }, + { + "epoch": 0.6670520231213873, + "grad_norm": 0.26692983480937005, + "learning_rate": 6.028224773914575e-05, + "loss": 0.953, + "step": 2885 + }, + { + "epoch": 0.6682080924855491, + "grad_norm": 0.28871757324400055, + "learning_rate": 5.991217514101586e-05, + "loss": 0.9275, + "step": 2890 + }, + { + "epoch": 0.669364161849711, + "grad_norm": 0.29493246673699086, + "learning_rate": 5.9542755532078856e-05, + "loss": 1.0001, + "step": 2895 + }, + { + "epoch": 0.6705202312138728, + "grad_norm": 0.3100505532341507, + "learning_rate": 5.917399492979805e-05, + "loss": 0.9716, + "step": 2900 + }, + { + "epoch": 0.6716763005780347, + "grad_norm": 0.296746629997667, + "learning_rate": 5.880589934090206e-05, + "loss": 0.987, + "step": 2905 + }, + { + "epoch": 0.6728323699421965, + "grad_norm": 0.3175494758298602, + "learning_rate": 5.843847476128722e-05, + "loss": 0.9643, + "step": 2910 + }, + { + "epoch": 0.6739884393063584, + "grad_norm": 0.310484738297623, + "learning_rate": 5.807172717591984e-05, + "loss": 0.9322, + "step": 2915 + }, + { + "epoch": 0.6751445086705202, + "grad_norm": 0.29571778911697455, + "learning_rate": 5.770566255873866e-05, + "loss": 0.929, + "step": 2920 + }, + { + "epoch": 0.6763005780346821, + "grad_norm": 0.2876079644684902, + "learning_rate": 5.734028687255751e-05, + "loss": 0.9644, + "step": 2925 + }, + { + "epoch": 0.6774566473988439, + "grad_norm": 0.3002944071515893, + "learning_rate": 5.697560606896839e-05, + "loss": 0.9987, + "step": 2930 + }, + { + "epoch": 0.6786127167630058, + "grad_norm": 0.2931767870621979, + "learning_rate": 5.6611626088244194e-05, + "loss": 0.9474, + "step": 2935 + }, + { + "epoch": 0.6797687861271676, + "grad_norm": 0.28720352206351696, + "learning_rate": 5.6248352859242314e-05, + "loss": 1.0355, + "step": 2940 + }, + { + "epoch": 0.6809248554913295, + "grad_norm": 0.3191718472968375, + "learning_rate": 5.588579229930784e-05, + "loss": 0.9699, + "step": 2945 + }, + { + "epoch": 0.6820809248554913, + "grad_norm": 0.30266420054049115, + "learning_rate": 5.552395031417712e-05, + "loss": 0.9358, + "step": 2950 + }, + { + "epoch": 0.6832369942196532, + "grad_norm": 0.30130721873082394, + "learning_rate": 5.516283279788183e-05, + "loss": 0.9496, + "step": 2955 + }, + { + "epoch": 0.684393063583815, + "grad_norm": 0.2929120042219513, + "learning_rate": 5.4802445632652634e-05, + "loss": 0.9248, + "step": 2960 + }, + { + "epoch": 0.6855491329479769, + "grad_norm": 0.31437058476192303, + "learning_rate": 5.444279468882358e-05, + "loss": 0.9745, + "step": 2965 + }, + { + "epoch": 0.6867052023121387, + "grad_norm": 0.2878711246279315, + "learning_rate": 5.408388582473651e-05, + "loss": 1.0008, + "step": 2970 + }, + { + "epoch": 0.6878612716763006, + "grad_norm": 0.2808303288729568, + "learning_rate": 5.3725724886645526e-05, + "loss": 0.9381, + "step": 2975 + }, + { + "epoch": 0.6890173410404624, + "grad_norm": 0.3006592351672636, + "learning_rate": 5.3368317708621674e-05, + "loss": 1.0076, + "step": 2980 + }, + { + "epoch": 0.6901734104046243, + "grad_norm": 0.30870133430305197, + "learning_rate": 5.3011670112458224e-05, + "loss": 0.9616, + "step": 2985 + }, + { + "epoch": 0.6913294797687861, + "grad_norm": 0.29489554268370943, + "learning_rate": 5.2655787907575436e-05, + "loss": 0.908, + "step": 2990 + }, + { + "epoch": 0.692485549132948, + "grad_norm": 0.30595244192322063, + "learning_rate": 5.230067689092629e-05, + "loss": 0.9902, + "step": 2995 + }, + { + "epoch": 0.6936416184971098, + "grad_norm": 0.3007538157511167, + "learning_rate": 5.19463428469019e-05, + "loss": 1.0259, + "step": 3000 + }, + { + "epoch": 0.6947976878612717, + "grad_norm": 0.28815906202526853, + "learning_rate": 5.159279154723715e-05, + "loss": 0.9497, + "step": 3005 + }, + { + "epoch": 0.6959537572254335, + "grad_norm": 0.2752889596438775, + "learning_rate": 5.124002875091704e-05, + "loss": 0.9586, + "step": 3010 + }, + { + "epoch": 0.6971098265895954, + "grad_norm": 0.2850467130558078, + "learning_rate": 5.088806020408252e-05, + "loss": 0.9049, + "step": 3015 + }, + { + "epoch": 0.6982658959537572, + "grad_norm": 0.2750258968098293, + "learning_rate": 5.053689163993703e-05, + "loss": 0.937, + "step": 3020 + }, + { + "epoch": 0.6994219653179191, + "grad_norm": 0.29519726643012756, + "learning_rate": 5.018652877865322e-05, + "loss": 0.9325, + "step": 3025 + }, + { + "epoch": 0.7005780346820809, + "grad_norm": 0.31869143160605945, + "learning_rate": 4.983697732727964e-05, + "loss": 0.9484, + "step": 3030 + }, + { + "epoch": 0.7017341040462428, + "grad_norm": 0.30604777371620057, + "learning_rate": 4.948824297964774e-05, + "loss": 0.9497, + "step": 3035 + }, + { + "epoch": 0.7028901734104046, + "grad_norm": 0.2781484111225009, + "learning_rate": 4.914033141627931e-05, + "loss": 0.9732, + "step": 3040 + }, + { + "epoch": 0.7040462427745665, + "grad_norm": 0.287979860552131, + "learning_rate": 4.87932483042937e-05, + "loss": 0.9513, + "step": 3045 + }, + { + "epoch": 0.7052023121387283, + "grad_norm": 0.28553452094386605, + "learning_rate": 4.8446999297315764e-05, + "loss": 0.9528, + "step": 3050 + }, + { + "epoch": 0.7063583815028902, + "grad_norm": 0.26876115159064556, + "learning_rate": 4.810159003538365e-05, + "loss": 0.9513, + "step": 3055 + }, + { + "epoch": 0.707514450867052, + "grad_norm": 0.29148448420091455, + "learning_rate": 4.775702614485678e-05, + "loss": 0.9732, + "step": 3060 + }, + { + "epoch": 0.7086705202312139, + "grad_norm": 0.2858967419780267, + "learning_rate": 4.7413313238324556e-05, + "loss": 0.9874, + "step": 3065 + }, + { + "epoch": 0.7098265895953757, + "grad_norm": 0.2853440571230289, + "learning_rate": 4.707045691451456e-05, + "loss": 0.9365, + "step": 3070 + }, + { + "epoch": 0.7109826589595376, + "grad_norm": 0.2756362587234919, + "learning_rate": 4.6728462758201574e-05, + "loss": 0.918, + "step": 3075 + }, + { + "epoch": 0.7121387283236994, + "grad_norm": 0.2933985662205089, + "learning_rate": 4.638733634011663e-05, + "loss": 0.9652, + "step": 3080 + }, + { + "epoch": 0.7132947976878613, + "grad_norm": 0.3088785959772121, + "learning_rate": 4.604708321685618e-05, + "loss": 0.9468, + "step": 3085 + }, + { + "epoch": 0.7144508670520231, + "grad_norm": 0.2897414636494522, + "learning_rate": 4.5707708930791514e-05, + "loss": 0.9136, + "step": 3090 + }, + { + "epoch": 0.715606936416185, + "grad_norm": 0.2977453761954692, + "learning_rate": 4.536921900997872e-05, + "loss": 0.9684, + "step": 3095 + }, + { + "epoch": 0.7167630057803468, + "grad_norm": 0.31568624217549895, + "learning_rate": 4.5031618968068325e-05, + "loss": 0.9804, + "step": 3100 + }, + { + "epoch": 0.7179190751445087, + "grad_norm": 0.2937121832237494, + "learning_rate": 4.4694914304215796e-05, + "loss": 0.9923, + "step": 3105 + }, + { + "epoch": 0.7190751445086705, + "grad_norm": 0.31223322912252777, + "learning_rate": 4.4359110502991773e-05, + "loss": 0.9493, + "step": 3110 + }, + { + "epoch": 0.7202312138728324, + "grad_norm": 0.2948338805474285, + "learning_rate": 4.402421303429274e-05, + "loss": 1.0058, + "step": 3115 + }, + { + "epoch": 0.7213872832369942, + "grad_norm": 0.2998471688331991, + "learning_rate": 4.3690227353251944e-05, + "loss": 0.9793, + "step": 3120 + }, + { + "epoch": 0.7225433526011561, + "grad_norm": 0.30105760378227486, + "learning_rate": 4.335715890015067e-05, + "loss": 0.9636, + "step": 3125 + }, + { + "epoch": 0.7236994219653179, + "grad_norm": 0.298969154483284, + "learning_rate": 4.302501310032937e-05, + "loss": 0.9849, + "step": 3130 + }, + { + "epoch": 0.7248554913294798, + "grad_norm": 0.30686646202523543, + "learning_rate": 4.26937953640995e-05, + "loss": 0.943, + "step": 3135 + }, + { + "epoch": 0.7260115606936416, + "grad_norm": 0.2906008606262198, + "learning_rate": 4.236351108665537e-05, + "loss": 0.9584, + "step": 3140 + }, + { + "epoch": 0.7271676300578035, + "grad_norm": 0.2965620261809724, + "learning_rate": 4.203416564798608e-05, + "loss": 0.9376, + "step": 3145 + }, + { + "epoch": 0.7283236994219653, + "grad_norm": 0.29569547507554944, + "learning_rate": 4.170576441278815e-05, + "loss": 1.0158, + "step": 3150 + }, + { + "epoch": 0.7294797687861272, + "grad_norm": 0.2873361360138618, + "learning_rate": 4.137831273037793e-05, + "loss": 0.956, + "step": 3155 + }, + { + "epoch": 0.730635838150289, + "grad_norm": 0.28863435820628536, + "learning_rate": 4.1051815934604465e-05, + "loss": 0.9456, + "step": 3160 + }, + { + "epoch": 0.7317919075144509, + "grad_norm": 0.30820662139709337, + "learning_rate": 4.072627934376292e-05, + "loss": 0.9627, + "step": 3165 + }, + { + "epoch": 0.7329479768786127, + "grad_norm": 0.2733290689658499, + "learning_rate": 4.0401708260507495e-05, + "loss": 0.9896, + "step": 3170 + }, + { + "epoch": 0.7341040462427746, + "grad_norm": 0.29448552437974784, + "learning_rate": 4.00781079717653e-05, + "loss": 0.9817, + "step": 3175 + }, + { + "epoch": 0.7352601156069364, + "grad_norm": 0.31792876220744604, + "learning_rate": 3.975548374865034e-05, + "loss": 0.98, + "step": 3180 + }, + { + "epoch": 0.7364161849710983, + "grad_norm": 0.2867123384035597, + "learning_rate": 3.943384084637732e-05, + "loss": 0.9845, + "step": 3185 + }, + { + "epoch": 0.7375722543352601, + "grad_norm": 0.29809867628254705, + "learning_rate": 3.9113184504176426e-05, + "loss": 0.9589, + "step": 3190 + }, + { + "epoch": 0.738728323699422, + "grad_norm": 0.2962633412794635, + "learning_rate": 3.879351994520774e-05, + "loss": 0.9644, + "step": 3195 + }, + { + "epoch": 0.7398843930635838, + "grad_norm": 0.3173965169375924, + "learning_rate": 3.847485237647614e-05, + "loss": 0.9243, + "step": 3200 + }, + { + "epoch": 0.7410404624277457, + "grad_norm": 0.2802160296625643, + "learning_rate": 3.815718698874672e-05, + "loss": 0.9627, + "step": 3205 + }, + { + "epoch": 0.7421965317919075, + "grad_norm": 0.30606251064274426, + "learning_rate": 3.7840528956459956e-05, + "loss": 1.0133, + "step": 3210 + }, + { + "epoch": 0.7433526011560694, + "grad_norm": 0.29567079239489624, + "learning_rate": 3.752488343764751e-05, + "loss": 1.0064, + "step": 3215 + }, + { + "epoch": 0.7445086705202312, + "grad_norm": 0.2908365758289181, + "learning_rate": 3.721025557384845e-05, + "loss": 0.939, + "step": 3220 + }, + { + "epoch": 0.7456647398843931, + "grad_norm": 0.2821772422176111, + "learning_rate": 3.689665049002513e-05, + "loss": 0.9176, + "step": 3225 + }, + { + "epoch": 0.7468208092485549, + "grad_norm": 0.28323988627298013, + "learning_rate": 3.658407329447986e-05, + "loss": 0.9504, + "step": 3230 + }, + { + "epoch": 0.7479768786127168, + "grad_norm": 0.3052577077686943, + "learning_rate": 3.627252907877184e-05, + "loss": 0.963, + "step": 3235 + }, + { + "epoch": 0.7491329479768786, + "grad_norm": 0.29287895420892424, + "learning_rate": 3.5962022917633976e-05, + "loss": 0.9758, + "step": 3240 + }, + { + "epoch": 0.7502890173410405, + "grad_norm": 0.26219424592081725, + "learning_rate": 3.56525598688904e-05, + "loss": 0.9501, + "step": 3245 + }, + { + "epoch": 0.7514450867052023, + "grad_norm": 0.2999575076011054, + "learning_rate": 3.534414497337406e-05, + "loss": 0.9979, + "step": 3250 + }, + { + "epoch": 0.7526011560693642, + "grad_norm": 0.30401739845864295, + "learning_rate": 3.503678325484448e-05, + "loss": 0.96, + "step": 3255 + }, + { + "epoch": 0.753757225433526, + "grad_norm": 0.2861448274685923, + "learning_rate": 3.473047971990605e-05, + "loss": 0.9956, + "step": 3260 + }, + { + "epoch": 0.7549132947976879, + "grad_norm": 0.2749463140020902, + "learning_rate": 3.442523935792651e-05, + "loss": 1.0207, + "step": 3265 + }, + { + "epoch": 0.7560693641618497, + "grad_norm": 0.2987193247584773, + "learning_rate": 3.4121067140955455e-05, + "loss": 0.9469, + "step": 3270 + }, + { + "epoch": 0.7572254335260116, + "grad_norm": 0.29389883736035816, + "learning_rate": 3.3817968023643766e-05, + "loss": 0.9934, + "step": 3275 + }, + { + "epoch": 0.7583815028901734, + "grad_norm": 0.28653966462371155, + "learning_rate": 3.351594694316239e-05, + "loss": 0.9634, + "step": 3280 + }, + { + "epoch": 0.7595375722543353, + "grad_norm": 0.29816187415165, + "learning_rate": 3.321500881912225e-05, + "loss": 0.9398, + "step": 3285 + }, + { + "epoch": 0.7606936416184971, + "grad_norm": 0.289319657394509, + "learning_rate": 3.29151585534941e-05, + "loss": 0.8928, + "step": 3290 + }, + { + "epoch": 0.761849710982659, + "grad_norm": 0.29521868021236264, + "learning_rate": 3.261640103052849e-05, + "loss": 0.9653, + "step": 3295 + }, + { + "epoch": 0.7630057803468208, + "grad_norm": 0.29934248191757035, + "learning_rate": 3.23187411166764e-05, + "loss": 0.97, + "step": 3300 + }, + { + "epoch": 0.7641618497109827, + "grad_norm": 0.32270679766275234, + "learning_rate": 3.2022183660509916e-05, + "loss": 0.9495, + "step": 3305 + }, + { + "epoch": 0.7653179190751445, + "grad_norm": 0.30814076544629976, + "learning_rate": 3.172673349264316e-05, + "loss": 0.9897, + "step": 3310 + }, + { + "epoch": 0.7664739884393064, + "grad_norm": 0.2731527495101352, + "learning_rate": 3.143239542565365e-05, + "loss": 0.9922, + "step": 3315 + }, + { + "epoch": 0.7676300578034682, + "grad_norm": 0.29822125577085, + "learning_rate": 3.113917425400406e-05, + "loss": 0.974, + "step": 3320 + }, + { + "epoch": 0.7687861271676301, + "grad_norm": 0.27831179956286994, + "learning_rate": 3.084707475396385e-05, + "loss": 0.9309, + "step": 3325 + }, + { + "epoch": 0.7699421965317919, + "grad_norm": 0.30061057019675924, + "learning_rate": 3.05561016835317e-05, + "loss": 0.9889, + "step": 3330 + }, + { + "epoch": 0.7710982658959538, + "grad_norm": 0.2839150865870422, + "learning_rate": 3.026625978235793e-05, + "loss": 0.9279, + "step": 3335 + }, + { + "epoch": 0.7722543352601156, + "grad_norm": 0.28191298234209533, + "learning_rate": 2.9977553771667178e-05, + "loss": 0.9043, + "step": 3340 + }, + { + "epoch": 0.7734104046242775, + "grad_norm": 0.2839643289974997, + "learning_rate": 2.968998835418174e-05, + "loss": 0.958, + "step": 3345 + }, + { + "epoch": 0.7745664739884393, + "grad_norm": 0.2727626163667981, + "learning_rate": 2.9403568214044687e-05, + "loss": 1.0045, + "step": 3350 + }, + { + "epoch": 0.7757225433526012, + "grad_norm": 0.2829586617421094, + "learning_rate": 2.9118298016743815e-05, + "loss": 0.9477, + "step": 3355 + }, + { + "epoch": 0.776878612716763, + "grad_norm": 0.2863725325553883, + "learning_rate": 2.8834182409035527e-05, + "loss": 1.0095, + "step": 3360 + }, + { + "epoch": 0.7780346820809249, + "grad_norm": 0.2981971073978521, + "learning_rate": 2.8551226018869105e-05, + "loss": 1.0128, + "step": 3365 + }, + { + "epoch": 0.7791907514450868, + "grad_norm": 0.2783020472744863, + "learning_rate": 2.8269433455311378e-05, + "loss": 0.9581, + "step": 3370 + }, + { + "epoch": 0.7803468208092486, + "grad_norm": 0.29194953070328256, + "learning_rate": 2.798880930847173e-05, + "loss": 0.9402, + "step": 3375 + }, + { + "epoch": 0.7815028901734105, + "grad_norm": 0.31785104915649315, + "learning_rate": 2.7709358149427113e-05, + "loss": 0.9912, + "step": 3380 + }, + { + "epoch": 0.7826589595375723, + "grad_norm": 0.27416031521109446, + "learning_rate": 2.7431084530147834e-05, + "loss": 0.9413, + "step": 3385 + }, + { + "epoch": 0.7838150289017342, + "grad_norm": 0.3023857012875686, + "learning_rate": 2.7153992983423283e-05, + "loss": 1.0101, + "step": 3390 + }, + { + "epoch": 0.784971098265896, + "grad_norm": 0.27883119977909965, + "learning_rate": 2.687808802278805e-05, + "loss": 0.9699, + "step": 3395 + }, + { + "epoch": 0.7861271676300579, + "grad_norm": 0.28656925239378356, + "learning_rate": 2.6603374142448467e-05, + "loss": 0.9459, + "step": 3400 + }, + { + "epoch": 0.7872832369942196, + "grad_norm": 0.2805856931300408, + "learning_rate": 2.632985581720947e-05, + "loss": 0.9347, + "step": 3405 + }, + { + "epoch": 0.7884393063583816, + "grad_norm": 0.28925155839879374, + "learning_rate": 2.6057537502401598e-05, + "loss": 0.9549, + "step": 3410 + }, + { + "epoch": 0.7895953757225433, + "grad_norm": 0.30326180151228377, + "learning_rate": 2.5786423633808487e-05, + "loss": 0.9954, + "step": 3415 + }, + { + "epoch": 0.7907514450867053, + "grad_norm": 0.310861152386562, + "learning_rate": 2.5516518627594542e-05, + "loss": 0.981, + "step": 3420 + }, + { + "epoch": 0.791907514450867, + "grad_norm": 0.29730609595809665, + "learning_rate": 2.524782688023305e-05, + "loss": 0.9933, + "step": 3425 + }, + { + "epoch": 0.793063583815029, + "grad_norm": 0.29011083826669604, + "learning_rate": 2.4980352768434643e-05, + "loss": 1.0396, + "step": 3430 + }, + { + "epoch": 0.7942196531791907, + "grad_norm": 0.2941491077374515, + "learning_rate": 2.4714100649075833e-05, + "loss": 0.9676, + "step": 3435 + }, + { + "epoch": 0.7953757225433526, + "grad_norm": 0.28795285220192574, + "learning_rate": 2.4449074859128197e-05, + "loss": 0.9545, + "step": 3440 + }, + { + "epoch": 0.7965317919075144, + "grad_norm": 0.2973713924271482, + "learning_rate": 2.4185279715587704e-05, + "loss": 0.925, + "step": 3445 + }, + { + "epoch": 0.7976878612716763, + "grad_norm": 0.28132269243331093, + "learning_rate": 2.39227195154043e-05, + "loss": 0.974, + "step": 3450 + }, + { + "epoch": 0.7988439306358381, + "grad_norm": 0.3130593570967486, + "learning_rate": 2.366139853541197e-05, + "loss": 0.9429, + "step": 3455 + }, + { + "epoch": 0.8, + "grad_norm": 0.28725553924621783, + "learning_rate": 2.340132103225916e-05, + "loss": 0.9609, + "step": 3460 + }, + { + "epoch": 0.8011560693641618, + "grad_norm": 0.2927313493050735, + "learning_rate": 2.3142491242339338e-05, + "loss": 0.9801, + "step": 3465 + }, + { + "epoch": 0.8023121387283237, + "grad_norm": 0.2959751791171696, + "learning_rate": 2.288491338172196e-05, + "loss": 0.9983, + "step": 3470 + }, + { + "epoch": 0.8034682080924855, + "grad_norm": 0.29009100429119816, + "learning_rate": 2.262859164608393e-05, + "loss": 0.9435, + "step": 3475 + }, + { + "epoch": 0.8046242774566474, + "grad_norm": 0.28065545174572615, + "learning_rate": 2.2373530210641103e-05, + "loss": 0.9664, + "step": 3480 + }, + { + "epoch": 0.8057803468208092, + "grad_norm": 0.3079082233043266, + "learning_rate": 2.2119733230080408e-05, + "loss": 0.9339, + "step": 3485 + }, + { + "epoch": 0.8069364161849711, + "grad_norm": 0.2994856214599522, + "learning_rate": 2.186720483849206e-05, + "loss": 0.9642, + "step": 3490 + }, + { + "epoch": 0.8080924855491329, + "grad_norm": 0.3150825598363239, + "learning_rate": 2.1615949149302305e-05, + "loss": 0.9391, + "step": 3495 + }, + { + "epoch": 0.8092485549132948, + "grad_norm": 0.28663989647188765, + "learning_rate": 2.1365970255206402e-05, + "loss": 0.9533, + "step": 3500 + }, + { + "epoch": 0.8104046242774566, + "grad_norm": 0.3064963810102998, + "learning_rate": 2.1117272228101902e-05, + "loss": 0.9269, + "step": 3505 + }, + { + "epoch": 0.8115606936416185, + "grad_norm": 0.2888662994868131, + "learning_rate": 2.0869859119022328e-05, + "loss": 0.9684, + "step": 3510 + }, + { + "epoch": 0.8127167630057803, + "grad_norm": 0.29449459192953004, + "learning_rate": 2.0623734958071296e-05, + "loss": 0.9588, + "step": 3515 + }, + { + "epoch": 0.8138728323699422, + "grad_norm": 0.28505307744008485, + "learning_rate": 2.037890375435677e-05, + "loss": 0.969, + "step": 3520 + }, + { + "epoch": 0.815028901734104, + "grad_norm": 0.264553046638561, + "learning_rate": 2.0135369495925714e-05, + "loss": 0.8865, + "step": 3525 + }, + { + "epoch": 0.8161849710982659, + "grad_norm": 0.2854615552898905, + "learning_rate": 1.9893136149699287e-05, + "loss": 0.9606, + "step": 3530 + }, + { + "epoch": 0.8173410404624277, + "grad_norm": 0.30132043472669484, + "learning_rate": 1.9652207661408073e-05, + "loss": 0.9783, + "step": 3535 + }, + { + "epoch": 0.8184971098265896, + "grad_norm": 0.28158707938088035, + "learning_rate": 1.941258795552785e-05, + "loss": 0.9702, + "step": 3540 + }, + { + "epoch": 0.8196531791907514, + "grad_norm": 0.3117678798282903, + "learning_rate": 1.917428093521576e-05, + "loss": 0.9381, + "step": 3545 + }, + { + "epoch": 0.8208092485549133, + "grad_norm": 0.2721446216347946, + "learning_rate": 1.8937290482246606e-05, + "loss": 0.9294, + "step": 3550 + }, + { + "epoch": 0.8219653179190751, + "grad_norm": 0.274947743569076, + "learning_rate": 1.870162045694971e-05, + "loss": 0.9874, + "step": 3555 + }, + { + "epoch": 0.823121387283237, + "grad_norm": 0.29164661845373113, + "learning_rate": 1.8467274698145942e-05, + "loss": 0.9641, + "step": 3560 + }, + { + "epoch": 0.8242774566473988, + "grad_norm": 0.2857254550756751, + "learning_rate": 1.8234257023085234e-05, + "loss": 0.9629, + "step": 3565 + }, + { + "epoch": 0.8254335260115607, + "grad_norm": 0.28344620289201494, + "learning_rate": 1.8002571227384467e-05, + "loss": 0.9213, + "step": 3570 + }, + { + "epoch": 0.8265895953757225, + "grad_norm": 0.2940307676845195, + "learning_rate": 1.777222108496558e-05, + "loss": 0.9567, + "step": 3575 + }, + { + "epoch": 0.8277456647398844, + "grad_norm": 0.29951372831586726, + "learning_rate": 1.7543210347994022e-05, + "loss": 0.9813, + "step": 3580 + }, + { + "epoch": 0.8289017341040462, + "grad_norm": 0.2878040658280256, + "learning_rate": 1.7315542746817825e-05, + "loss": 0.9152, + "step": 3585 + }, + { + "epoch": 0.8300578034682081, + "grad_norm": 0.30016211006512195, + "learning_rate": 1.7089221989906633e-05, + "loss": 0.9656, + "step": 3590 + }, + { + "epoch": 0.8312138728323699, + "grad_norm": 0.2744245197404456, + "learning_rate": 1.6864251763791428e-05, + "loss": 1.0234, + "step": 3595 + }, + { + "epoch": 0.8323699421965318, + "grad_norm": 0.28064881967639976, + "learning_rate": 1.664063573300446e-05, + "loss": 1.0006, + "step": 3600 + }, + { + "epoch": 0.8335260115606936, + "grad_norm": 0.28500326120194347, + "learning_rate": 1.6418377540019536e-05, + "loss": 0.9442, + "step": 3605 + }, + { + "epoch": 0.8346820809248555, + "grad_norm": 0.27741047130235014, + "learning_rate": 1.6197480805192634e-05, + "loss": 0.9173, + "step": 3610 + }, + { + "epoch": 0.8358381502890173, + "grad_norm": 0.2952544695479881, + "learning_rate": 1.5977949126703084e-05, + "loss": 0.9546, + "step": 3615 + }, + { + "epoch": 0.8369942196531792, + "grad_norm": 0.2753923526465453, + "learning_rate": 1.5759786080494743e-05, + "loss": 0.9922, + "step": 3620 + }, + { + "epoch": 0.838150289017341, + "grad_norm": 0.2713440737201456, + "learning_rate": 1.554299522021796e-05, + "loss": 0.9215, + "step": 3625 + }, + { + "epoch": 0.8393063583815029, + "grad_norm": 0.2924976737824655, + "learning_rate": 1.5327580077171587e-05, + "loss": 0.9553, + "step": 3630 + }, + { + "epoch": 0.8404624277456647, + "grad_norm": 0.28640397728398614, + "learning_rate": 1.5113544160245397e-05, + "loss": 0.9395, + "step": 3635 + }, + { + "epoch": 0.8416184971098266, + "grad_norm": 0.2859545519050961, + "learning_rate": 1.4900890955863067e-05, + "loss": 0.9711, + "step": 3640 + }, + { + "epoch": 0.8427745664739884, + "grad_norm": 0.3089713704921713, + "learning_rate": 1.4689623927925289e-05, + "loss": 1.0092, + "step": 3645 + }, + { + "epoch": 0.8439306358381503, + "grad_norm": 0.2837926097401027, + "learning_rate": 1.4479746517753335e-05, + "loss": 1.0052, + "step": 3650 + }, + { + "epoch": 0.8450867052023121, + "grad_norm": 0.30261005244224143, + "learning_rate": 1.4271262144033116e-05, + "loss": 0.9991, + "step": 3655 + }, + { + "epoch": 0.846242774566474, + "grad_norm": 0.2674381336270983, + "learning_rate": 1.4064174202759407e-05, + "loss": 0.9552, + "step": 3660 + }, + { + "epoch": 0.8473988439306358, + "grad_norm": 0.2767756118448426, + "learning_rate": 1.3858486067180465e-05, + "loss": 0.9573, + "step": 3665 + }, + { + "epoch": 0.8485549132947977, + "grad_norm": 0.2836915452259556, + "learning_rate": 1.3654201087743279e-05, + "loss": 0.9466, + "step": 3670 + }, + { + "epoch": 0.8497109826589595, + "grad_norm": 0.30898236268087276, + "learning_rate": 1.3451322592038774e-05, + "loss": 0.9379, + "step": 3675 + }, + { + "epoch": 0.8508670520231214, + "grad_norm": 0.30463841544406706, + "learning_rate": 1.3249853884747753e-05, + "loss": 0.9506, + "step": 3680 + }, + { + "epoch": 0.8520231213872832, + "grad_norm": 0.29760105531487135, + "learning_rate": 1.3049798247587064e-05, + "loss": 0.9603, + "step": 3685 + }, + { + "epoch": 0.8531791907514451, + "grad_norm": 0.2910112980743013, + "learning_rate": 1.2851158939256002e-05, + "loss": 0.9903, + "step": 3690 + }, + { + "epoch": 0.8543352601156069, + "grad_norm": 0.28051123703308095, + "learning_rate": 1.2653939195383446e-05, + "loss": 0.9552, + "step": 3695 + }, + { + "epoch": 0.8554913294797688, + "grad_norm": 0.32807409572083474, + "learning_rate": 1.2458142228474967e-05, + "loss": 0.9391, + "step": 3700 + }, + { + "epoch": 0.8566473988439306, + "grad_norm": 0.2857511643393787, + "learning_rate": 1.2263771227860555e-05, + "loss": 0.9249, + "step": 3705 + }, + { + "epoch": 0.8578034682080925, + "grad_norm": 0.3130878799401169, + "learning_rate": 1.2070829359642743e-05, + "loss": 0.9801, + "step": 3710 + }, + { + "epoch": 0.8589595375722543, + "grad_norm": 0.2857834049135047, + "learning_rate": 1.1879319766644969e-05, + "loss": 0.9881, + "step": 3715 + }, + { + "epoch": 0.8601156069364162, + "grad_norm": 0.28295250626185514, + "learning_rate": 1.168924556836034e-05, + "loss": 0.9212, + "step": 3720 + }, + { + "epoch": 0.861271676300578, + "grad_norm": 0.270720741833164, + "learning_rate": 1.1500609860900934e-05, + "loss": 0.9208, + "step": 3725 + }, + { + "epoch": 0.8624277456647399, + "grad_norm": 0.3310546829847928, + "learning_rate": 1.131341571694724e-05, + "loss": 0.9134, + "step": 3730 + }, + { + "epoch": 0.8635838150289017, + "grad_norm": 0.2906866125100536, + "learning_rate": 1.1127666185698183e-05, + "loss": 0.9292, + "step": 3735 + }, + { + "epoch": 0.8647398843930636, + "grad_norm": 0.2857402985603114, + "learning_rate": 1.0943364292821478e-05, + "loss": 0.9782, + "step": 3740 + }, + { + "epoch": 0.8658959537572254, + "grad_norm": 0.30621748628878964, + "learning_rate": 1.0760513040404275e-05, + "loss": 0.9538, + "step": 3745 + }, + { + "epoch": 0.8670520231213873, + "grad_norm": 0.3076325870742748, + "learning_rate": 1.0579115406904327e-05, + "loss": 0.9814, + "step": 3750 + }, + { + "epoch": 0.8682080924855491, + "grad_norm": 0.2875679417313269, + "learning_rate": 1.0399174347101404e-05, + "loss": 0.9682, + "step": 3755 + }, + { + "epoch": 0.869364161849711, + "grad_norm": 0.2827349143068363, + "learning_rate": 1.0220692792049169e-05, + "loss": 0.9715, + "step": 3760 + }, + { + "epoch": 0.8705202312138728, + "grad_norm": 0.2898950516764476, + "learning_rate": 1.0043673649027518e-05, + "loss": 0.9809, + "step": 3765 + }, + { + "epoch": 0.8716763005780347, + "grad_norm": 0.29127323235770597, + "learning_rate": 9.86811980149519e-06, + "loss": 0.9738, + "step": 3770 + }, + { + "epoch": 0.8728323699421965, + "grad_norm": 0.289489961872677, + "learning_rate": 9.694034109042694e-06, + "loss": 1.0206, + "step": 3775 + }, + { + "epoch": 0.8739884393063584, + "grad_norm": 0.2820602963126248, + "learning_rate": 9.521419407345878e-06, + "loss": 0.8967, + "step": 3780 + }, + { + "epoch": 0.8751445086705202, + "grad_norm": 0.31061259437004146, + "learning_rate": 9.350278508119636e-06, + "loss": 0.9786, + "step": 3785 + }, + { + "epoch": 0.8763005780346821, + "grad_norm": 0.2815496774647879, + "learning_rate": 9.180614199072146e-06, + "loss": 0.9485, + "step": 3790 + }, + { + "epoch": 0.8774566473988439, + "grad_norm": 0.3518473115040322, + "learning_rate": 9.012429243859487e-06, + "loss": 0.9903, + "step": 3795 + }, + { + "epoch": 0.8786127167630058, + "grad_norm": 0.2916712847778055, + "learning_rate": 8.845726382040597e-06, + "loss": 0.9397, + "step": 3800 + }, + { + "epoch": 0.8797687861271676, + "grad_norm": 0.29392008724784624, + "learning_rate": 8.680508329032589e-06, + "loss": 0.9629, + "step": 3805 + }, + { + "epoch": 0.8809248554913295, + "grad_norm": 0.28601993585254176, + "learning_rate": 8.516777776066643e-06, + "loss": 0.982, + "step": 3810 + }, + { + "epoch": 0.8820809248554913, + "grad_norm": 0.28346541614728327, + "learning_rate": 8.354537390144057e-06, + "loss": 0.9446, + "step": 3815 + }, + { + "epoch": 0.8832369942196532, + "grad_norm": 0.289284064978837, + "learning_rate": 8.19378981399287e-06, + "loss": 0.9647, + "step": 3820 + }, + { + "epoch": 0.884393063583815, + "grad_norm": 0.2867377684057541, + "learning_rate": 8.034537666024822e-06, + "loss": 0.9455, + "step": 3825 + }, + { + "epoch": 0.8855491329479769, + "grad_norm": 0.2756548624705791, + "learning_rate": 7.876783540292599e-06, + "loss": 1.0273, + "step": 3830 + }, + { + "epoch": 0.8867052023121387, + "grad_norm": 0.2936449029949148, + "learning_rate": 7.720530006447736e-06, + "loss": 0.9742, + "step": 3835 + }, + { + "epoch": 0.8878612716763006, + "grad_norm": 0.297967509970897, + "learning_rate": 7.565779609698631e-06, + "loss": 0.9329, + "step": 3840 + }, + { + "epoch": 0.8890173410404625, + "grad_norm": 0.2942735618729235, + "learning_rate": 7.412534870769116e-06, + "loss": 0.9581, + "step": 3845 + }, + { + "epoch": 0.8901734104046243, + "grad_norm": 0.28806737344976613, + "learning_rate": 7.260798285857484e-06, + "loss": 1.0075, + "step": 3850 + }, + { + "epoch": 0.8913294797687862, + "grad_norm": 0.2831951777045236, + "learning_rate": 7.110572326595711e-06, + "loss": 0.9971, + "step": 3855 + }, + { + "epoch": 0.892485549132948, + "grad_norm": 0.28520928967193043, + "learning_rate": 6.961859440009233e-06, + "loss": 1.0001, + "step": 3860 + }, + { + "epoch": 0.8936416184971099, + "grad_norm": 0.27689057228854386, + "learning_rate": 6.8146620484771495e-06, + "loss": 0.9889, + "step": 3865 + }, + { + "epoch": 0.8947976878612717, + "grad_norm": 0.285699685858991, + "learning_rate": 6.668982549692649e-06, + "loss": 0.9478, + "step": 3870 + }, + { + "epoch": 0.8959537572254336, + "grad_norm": 0.2754348125812895, + "learning_rate": 6.524823316624063e-06, + "loss": 0.9343, + "step": 3875 + }, + { + "epoch": 0.8971098265895954, + "grad_norm": 0.29989809168873266, + "learning_rate": 6.382186697476167e-06, + "loss": 0.9789, + "step": 3880 + }, + { + "epoch": 0.8982658959537573, + "grad_norm": 0.2828138414197748, + "learning_rate": 6.2410750156518985e-06, + "loss": 1.0062, + "step": 3885 + }, + { + "epoch": 0.8994219653179191, + "grad_norm": 0.2889085637993367, + "learning_rate": 6.101490569714574e-06, + "loss": 0.9696, + "step": 3890 + }, + { + "epoch": 0.900578034682081, + "grad_norm": 0.29960325302848234, + "learning_rate": 5.963435633350412e-06, + "loss": 0.9162, + "step": 3895 + }, + { + "epoch": 0.9017341040462428, + "grad_norm": 0.3005589289503204, + "learning_rate": 5.826912455331468e-06, + "loss": 0.9903, + "step": 3900 + }, + { + "epoch": 0.9028901734104047, + "grad_norm": 0.2928261089403607, + "learning_rate": 5.691923259479093e-06, + "loss": 0.9736, + "step": 3905 + }, + { + "epoch": 0.9040462427745665, + "grad_norm": 0.28429998893881614, + "learning_rate": 5.558470244627634e-06, + "loss": 0.9711, + "step": 3910 + }, + { + "epoch": 0.9052023121387284, + "grad_norm": 0.2784468816364227, + "learning_rate": 5.4265555845886215e-06, + "loss": 0.9811, + "step": 3915 + }, + { + "epoch": 0.9063583815028902, + "grad_norm": 0.2872050965542848, + "learning_rate": 5.29618142811541e-06, + "loss": 0.954, + "step": 3920 + }, + { + "epoch": 0.9075144508670521, + "grad_norm": 0.2948554706499198, + "learning_rate": 5.16734989886809e-06, + "loss": 0.9811, + "step": 3925 + }, + { + "epoch": 0.9086705202312139, + "grad_norm": 0.28399320997388067, + "learning_rate": 5.040063095379011e-06, + "loss": 0.9175, + "step": 3930 + }, + { + "epoch": 0.9098265895953758, + "grad_norm": 0.2855808882463355, + "learning_rate": 4.914323091018535e-06, + "loss": 0.9901, + "step": 3935 + }, + { + "epoch": 0.9109826589595376, + "grad_norm": 0.3021329131025513, + "learning_rate": 4.790131933961206e-06, + "loss": 0.9613, + "step": 3940 + }, + { + "epoch": 0.9121387283236995, + "grad_norm": 0.2862589716791116, + "learning_rate": 4.6674916471524995e-06, + "loss": 0.9961, + "step": 3945 + }, + { + "epoch": 0.9132947976878613, + "grad_norm": 0.2998472712073513, + "learning_rate": 4.546404228275824e-06, + "loss": 0.9845, + "step": 3950 + }, + { + "epoch": 0.9144508670520232, + "grad_norm": 0.28603652316011496, + "learning_rate": 4.426871649719932e-06, + "loss": 0.9765, + "step": 3955 + }, + { + "epoch": 0.915606936416185, + "grad_norm": 0.2872761079408887, + "learning_rate": 4.3088958585468686e-06, + "loss": 0.9997, + "step": 3960 + }, + { + "epoch": 0.9167630057803469, + "grad_norm": 0.2873199733628149, + "learning_rate": 4.192478776460229e-06, + "loss": 1.0064, + "step": 3965 + }, + { + "epoch": 0.9179190751445087, + "grad_norm": 0.27515435582137365, + "learning_rate": 4.077622299773831e-06, + "loss": 0.9189, + "step": 3970 + }, + { + "epoch": 0.9190751445086706, + "grad_norm": 0.27999010046883294, + "learning_rate": 3.96432829938086e-06, + "loss": 0.9885, + "step": 3975 + }, + { + "epoch": 0.9202312138728324, + "grad_norm": 0.2635891865916151, + "learning_rate": 3.8525986207233465e-06, + "loss": 0.9352, + "step": 3980 + }, + { + "epoch": 0.9213872832369943, + "grad_norm": 0.28633688125843376, + "learning_rate": 3.742435083762186e-06, + "loss": 0.9599, + "step": 3985 + }, + { + "epoch": 0.922543352601156, + "grad_norm": 0.2826637970887437, + "learning_rate": 3.633839482947421e-06, + "loss": 0.9948, + "step": 3990 + }, + { + "epoch": 0.923699421965318, + "grad_norm": 0.30153590546133957, + "learning_rate": 3.526813587189026e-06, + "loss": 0.9804, + "step": 3995 + }, + { + "epoch": 0.9248554913294798, + "grad_norm": 0.2858269823618572, + "learning_rate": 3.4213591398281175e-06, + "loss": 0.974, + "step": 4000 + }, + { + "epoch": 0.9260115606936417, + "grad_norm": 0.2820688133929364, + "learning_rate": 3.3174778586085643e-06, + "loss": 0.9566, + "step": 4005 + }, + { + "epoch": 0.9271676300578034, + "grad_norm": 0.3000844480451751, + "learning_rate": 3.2151714356489225e-06, + "loss": 1.0118, + "step": 4010 + }, + { + "epoch": 0.9283236994219654, + "grad_norm": 0.2790666100533931, + "learning_rate": 3.114441537415014e-06, + "loss": 0.9533, + "step": 4015 + }, + { + "epoch": 0.9294797687861271, + "grad_norm": 0.29749566810305794, + "learning_rate": 3.0152898046926557e-06, + "loss": 0.9569, + "step": 4020 + }, + { + "epoch": 0.930635838150289, + "grad_norm": 0.28219907290129675, + "learning_rate": 2.917717852560997e-06, + "loss": 0.9512, + "step": 4025 + }, + { + "epoch": 0.9317919075144508, + "grad_norm": 0.29888043306041123, + "learning_rate": 2.8217272703661923e-06, + "loss": 1.0231, + "step": 4030 + }, + { + "epoch": 0.9329479768786128, + "grad_norm": 0.26919643142805166, + "learning_rate": 2.727319621695501e-06, + "loss": 0.9873, + "step": 4035 + }, + { + "epoch": 0.9341040462427745, + "grad_norm": 0.29380224096591, + "learning_rate": 2.6344964443518526e-06, + "loss": 0.9385, + "step": 4040 + }, + { + "epoch": 0.9352601156069364, + "grad_norm": 0.28299962979190113, + "learning_rate": 2.5432592503288e-06, + "loss": 0.9292, + "step": 4045 + }, + { + "epoch": 0.9364161849710982, + "grad_norm": 0.29962326178158855, + "learning_rate": 2.453609525785816e-06, + "loss": 0.9773, + "step": 4050 + }, + { + "epoch": 0.9375722543352601, + "grad_norm": 0.32360192327986553, + "learning_rate": 2.3655487310241585e-06, + "loss": 0.9529, + "step": 4055 + }, + { + "epoch": 0.9387283236994219, + "grad_norm": 0.30410924441522585, + "learning_rate": 2.279078300463089e-06, + "loss": 1.0106, + "step": 4060 + }, + { + "epoch": 0.9398843930635838, + "grad_norm": 0.30431377413637145, + "learning_rate": 2.1941996426164344e-06, + "loss": 0.9735, + "step": 4065 + }, + { + "epoch": 0.9410404624277456, + "grad_norm": 0.2903468628517127, + "learning_rate": 2.1109141400697418e-06, + "loss": 0.972, + "step": 4070 + }, + { + "epoch": 0.9421965317919075, + "grad_norm": 0.2954528106164249, + "learning_rate": 2.029223149457682e-06, + "loss": 0.9662, + "step": 4075 + }, + { + "epoch": 0.9433526011560693, + "grad_norm": 0.27439734246785746, + "learning_rate": 1.949128001441969e-06, + "loss": 0.9509, + "step": 4080 + }, + { + "epoch": 0.9445086705202312, + "grad_norm": 0.27172305484407105, + "learning_rate": 1.8706300006896882e-06, + "loss": 0.8628, + "step": 4085 + }, + { + "epoch": 0.945664739884393, + "grad_norm": 0.27216092212065524, + "learning_rate": 1.7937304258520692e-06, + "loss": 0.9785, + "step": 4090 + }, + { + "epoch": 0.9468208092485549, + "grad_norm": 0.2918352349773759, + "learning_rate": 1.718430529543613e-06, + "loss": 0.9571, + "step": 4095 + }, + { + "epoch": 0.9479768786127167, + "grad_norm": 0.272999376680257, + "learning_rate": 1.6447315383217643e-06, + "loss": 0.918, + "step": 4100 + }, + { + "epoch": 0.9491329479768786, + "grad_norm": 0.27901699608664005, + "learning_rate": 1.5726346526668156e-06, + "loss": 0.9528, + "step": 4105 + }, + { + "epoch": 0.9502890173410404, + "grad_norm": 0.28289047639247433, + "learning_rate": 1.5021410469624465e-06, + "loss": 0.9067, + "step": 4110 + }, + { + "epoch": 0.9514450867052023, + "grad_norm": 0.2825322282067992, + "learning_rate": 1.4332518694765707e-06, + "loss": 0.8735, + "step": 4115 + }, + { + "epoch": 0.9526011560693641, + "grad_norm": 0.28837647678016226, + "learning_rate": 1.3659682423425968e-06, + "loss": 0.9432, + "step": 4120 + }, + { + "epoch": 0.953757225433526, + "grad_norm": 0.2876129817030531, + "learning_rate": 1.30029126154122e-06, + "loss": 0.9268, + "step": 4125 + }, + { + "epoch": 0.9549132947976878, + "grad_norm": 0.2792371887058164, + "learning_rate": 1.236221996882514e-06, + "loss": 0.9726, + "step": 4130 + }, + { + "epoch": 0.9560693641618497, + "grad_norm": 0.29099718253951457, + "learning_rate": 1.1737614919885008e-06, + "loss": 0.9209, + "step": 4135 + }, + { + "epoch": 0.9572254335260115, + "grad_norm": 0.2995617326772327, + "learning_rate": 1.1129107642761872e-06, + "loss": 0.9187, + "step": 4140 + }, + { + "epoch": 0.9583815028901734, + "grad_norm": 0.2762400721384854, + "learning_rate": 1.053670804940987e-06, + "loss": 0.9142, + "step": 4145 + }, + { + "epoch": 0.9595375722543352, + "grad_norm": 0.2931863784593866, + "learning_rate": 9.960425789405258e-07, + "loss": 0.9084, + "step": 4150 + }, + { + "epoch": 0.9606936416184971, + "grad_norm": 0.2880567882354411, + "learning_rate": 9.400270249789955e-07, + "loss": 0.9868, + "step": 4155 + }, + { + "epoch": 0.9618497109826589, + "grad_norm": 0.2871636561882668, + "learning_rate": 8.856250554918344e-07, + "loss": 0.9152, + "step": 4160 + }, + { + "epoch": 0.9630057803468208, + "grad_norm": 0.28517050924588744, + "learning_rate": 8.328375566308166e-07, + "loss": 0.9757, + "step": 4165 + }, + { + "epoch": 0.9641618497109826, + "grad_norm": 0.28587044682235146, + "learning_rate": 7.816653882496971e-07, + "loss": 0.9537, + "step": 4170 + }, + { + "epoch": 0.9653179190751445, + "grad_norm": 0.29706875055091925, + "learning_rate": 7.321093838901449e-07, + "loss": 0.9247, + "step": 4175 + }, + { + "epoch": 0.9664739884393063, + "grad_norm": 0.2861635930847044, + "learning_rate": 6.841703507682206e-07, + "loss": 0.8905, + "step": 4180 + }, + { + "epoch": 0.9676300578034682, + "grad_norm": 0.2813627882340819, + "learning_rate": 6.378490697611761e-07, + "loss": 0.9749, + "step": 4185 + }, + { + "epoch": 0.96878612716763, + "grad_norm": 0.2865848575770194, + "learning_rate": 5.93146295394742e-07, + "loss": 0.9443, + "step": 4190 + }, + { + "epoch": 0.9699421965317919, + "grad_norm": 0.27882077648739195, + "learning_rate": 5.500627558308713e-07, + "loss": 0.9398, + "step": 4195 + }, + { + "epoch": 0.9710982658959537, + "grad_norm": 0.3065313177380281, + "learning_rate": 5.085991528558487e-07, + "loss": 0.9236, + "step": 4200 + }, + { + "epoch": 0.9722543352601156, + "grad_norm": 0.2871839420940439, + "learning_rate": 4.687561618688663e-07, + "loss": 0.9676, + "step": 4205 + }, + { + "epoch": 0.9734104046242774, + "grad_norm": 0.2637381920338272, + "learning_rate": 4.3053443187103247e-07, + "loss": 0.8815, + "step": 4210 + }, + { + "epoch": 0.9745664739884393, + "grad_norm": 0.285755856830912, + "learning_rate": 3.9393458545479157e-07, + "loss": 0.9926, + "step": 4215 + }, + { + "epoch": 0.9757225433526011, + "grad_norm": 0.30359812477643444, + "learning_rate": 3.589572187937651e-07, + "loss": 0.9604, + "step": 4220 + }, + { + "epoch": 0.976878612716763, + "grad_norm": 0.29778116804881444, + "learning_rate": 3.2560290163307083e-07, + "loss": 0.9788, + "step": 4225 + }, + { + "epoch": 0.9780346820809248, + "grad_norm": 0.2794224102916479, + "learning_rate": 2.93872177280019e-07, + "loss": 0.9427, + "step": 4230 + }, + { + "epoch": 0.9791907514450867, + "grad_norm": 0.3008883017130643, + "learning_rate": 2.637655625952973e-07, + "loss": 0.9475, + "step": 4235 + }, + { + "epoch": 0.9803468208092485, + "grad_norm": 0.28456190197348324, + "learning_rate": 2.3528354798451058e-07, + "loss": 0.9461, + "step": 4240 + }, + { + "epoch": 0.9815028901734104, + "grad_norm": 0.2829229588306782, + "learning_rate": 2.0842659739019887e-07, + "loss": 0.9524, + "step": 4245 + }, + { + "epoch": 0.9826589595375722, + "grad_norm": 0.283105235302632, + "learning_rate": 1.8319514828430973e-07, + "loss": 0.9474, + "step": 4250 + }, + { + "epoch": 0.9838150289017341, + "grad_norm": 0.27690208166745106, + "learning_rate": 1.5958961166104847e-07, + "loss": 0.9351, + "step": 4255 + }, + { + "epoch": 0.9849710982658959, + "grad_norm": 0.25712842285219684, + "learning_rate": 1.3761037203017245e-07, + "loss": 0.914, + "step": 4260 + }, + { + "epoch": 0.9861271676300578, + "grad_norm": 0.2735837511043363, + "learning_rate": 1.1725778741076276e-07, + "loss": 0.8924, + "step": 4265 + }, + { + "epoch": 0.9872832369942196, + "grad_norm": 0.2820464817023967, + "learning_rate": 9.853218932536212e-08, + "loss": 0.9359, + "step": 4270 + }, + { + "epoch": 0.9884393063583815, + "grad_norm": 0.27653871874823904, + "learning_rate": 8.143388279460151e-08, + "loss": 0.9704, + "step": 4275 + }, + { + "epoch": 0.9895953757225433, + "grad_norm": 0.2836276601553615, + "learning_rate": 6.596314633219303e-08, + "loss": 0.9966, + "step": 4280 + }, + { + "epoch": 0.9907514450867052, + "grad_norm": 0.27880904921583405, + "learning_rate": 5.2120231940433474e-08, + "loss": 0.9227, + "step": 4285 + }, + { + "epoch": 0.991907514450867, + "grad_norm": 0.2657505634085022, + "learning_rate": 3.990536510608544e-08, + "loss": 0.9657, + "step": 4290 + }, + { + "epoch": 0.9930635838150289, + "grad_norm": 0.2908473414800192, + "learning_rate": 2.9318744796669183e-08, + "loss": 0.9206, + "step": 4295 + }, + { + "epoch": 0.9942196531791907, + "grad_norm": 0.2957217704373017, + "learning_rate": 2.036054345729843e-08, + "loss": 0.9528, + "step": 4300 + }, + { + "epoch": 0.9953757225433526, + "grad_norm": 0.2910406474246415, + "learning_rate": 1.3030907007793857e-08, + "loss": 1.0005, + "step": 4305 + }, + { + "epoch": 0.9965317919075144, + "grad_norm": 0.2796254503042765, + "learning_rate": 7.329954840362696e-09, + "loss": 1.0259, + "step": 4310 + }, + { + "epoch": 0.9976878612716763, + "grad_norm": 0.2844197423108749, + "learning_rate": 3.257779817600337e-09, + "loss": 0.9363, + "step": 4315 + }, + { + "epoch": 0.9988439306358381, + "grad_norm": 0.3388694141182201, + "learning_rate": 8.144482710248369e-10, + "loss": 1.0117, + "step": 4320 + }, + { + "epoch": 1.0, + "grad_norm": 0.29406054557823574, + "learning_rate": 0.0, + "loss": 0.9703, + "step": 4325 + }, + { + "epoch": 1.0, + "eval_runtime": 3.3948, + "eval_samples_per_second": 2.946, + "eval_steps_per_second": 0.884, + "step": 4325 + }, + { + "epoch": 1.0, + "step": 4325, + "total_flos": 1.71674607550464e+16, + "train_loss": 0.9719910388461427, + "train_runtime": 20047.9922, + "train_samples_per_second": 3.452, + "train_steps_per_second": 0.216 + } + ], + "logging_steps": 5, + "max_steps": 4325, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 100, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.71674607550464e+16, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}