{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2230, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004484304932735426, "grad_norm": 4.697020980518476, "learning_rate": 4.4843049327354265e-08, "loss": 0.9912, "step": 1 }, { "epoch": 0.004484304932735426, "grad_norm": 5.023039914107906, "learning_rate": 4.484304932735426e-07, "loss": 1.0335, "step": 10 }, { "epoch": 0.008968609865470852, "grad_norm": 4.618075407201601, "learning_rate": 8.968609865470852e-07, "loss": 1.0355, "step": 20 }, { "epoch": 0.013452914798206279, "grad_norm": 2.181160774619855, "learning_rate": 1.345291479820628e-06, "loss": 0.9824, "step": 30 }, { "epoch": 0.017937219730941704, "grad_norm": 2.008569343268214, "learning_rate": 1.7937219730941704e-06, "loss": 0.9383, "step": 40 }, { "epoch": 0.02242152466367713, "grad_norm": 1.7133805307500916, "learning_rate": 2.242152466367713e-06, "loss": 0.87, "step": 50 }, { "epoch": 0.026905829596412557, "grad_norm": 1.6402285214131669, "learning_rate": 2.690582959641256e-06, "loss": 0.8608, "step": 60 }, { "epoch": 0.03139013452914798, "grad_norm": 1.548076900788033, "learning_rate": 3.1390134529147986e-06, "loss": 0.8675, "step": 70 }, { "epoch": 0.03587443946188341, "grad_norm": 1.5686573872779637, "learning_rate": 3.587443946188341e-06, "loss": 0.8581, "step": 80 }, { "epoch": 0.04035874439461883, "grad_norm": 1.8398749741247653, "learning_rate": 4.0358744394618836e-06, "loss": 0.8272, "step": 90 }, { "epoch": 0.04484304932735426, "grad_norm": 1.431522929426351, "learning_rate": 4.484304932735426e-06, "loss": 0.837, "step": 100 }, { "epoch": 0.04932735426008968, "grad_norm": 1.5183209831163855, "learning_rate": 4.932735426008969e-06, "loss": 0.7993, "step": 110 }, { "epoch": 0.053811659192825115, "grad_norm": 1.440741479973038, "learning_rate": 5.381165919282512e-06, "loss": 0.7768, "step": 120 }, { "epoch": 0.05829596412556054, "grad_norm": 1.6957957755628803, "learning_rate": 5.8295964125560544e-06, "loss": 0.793, "step": 130 }, { "epoch": 0.06278026905829596, "grad_norm": 1.5847929433417824, "learning_rate": 6.278026905829597e-06, "loss": 0.8145, "step": 140 }, { "epoch": 0.06726457399103139, "grad_norm": 1.721638040499256, "learning_rate": 6.72645739910314e-06, "loss": 0.7677, "step": 150 }, { "epoch": 0.07174887892376682, "grad_norm": 1.8900581111883776, "learning_rate": 7.174887892376682e-06, "loss": 0.784, "step": 160 }, { "epoch": 0.07623318385650224, "grad_norm": 1.554784427401762, "learning_rate": 7.6233183856502244e-06, "loss": 0.7595, "step": 170 }, { "epoch": 0.08071748878923767, "grad_norm": 1.8326262271816938, "learning_rate": 8.071748878923767e-06, "loss": 0.7722, "step": 180 }, { "epoch": 0.08520179372197309, "grad_norm": 1.4707798772479566, "learning_rate": 8.52017937219731e-06, "loss": 0.764, "step": 190 }, { "epoch": 0.08968609865470852, "grad_norm": 1.449476026227771, "learning_rate": 8.968609865470853e-06, "loss": 0.773, "step": 200 }, { "epoch": 0.09417040358744394, "grad_norm": 1.5945882538155545, "learning_rate": 9.417040358744395e-06, "loss": 0.7549, "step": 210 }, { "epoch": 0.09865470852017937, "grad_norm": 1.4765341188830319, "learning_rate": 9.865470852017938e-06, "loss": 0.784, "step": 220 }, { "epoch": 0.1031390134529148, "grad_norm": 1.452955640004767, "learning_rate": 9.999699851108367e-06, "loss": 0.7909, "step": 230 }, { "epoch": 0.10762331838565023, "grad_norm": 1.5997501860541272, "learning_rate": 9.9982298208374e-06, "loss": 0.7651, "step": 240 }, { "epoch": 0.11210762331838565, "grad_norm": 1.554511883278329, "learning_rate": 9.995535139530904e-06, "loss": 0.7621, "step": 250 }, { "epoch": 0.11659192825112108, "grad_norm": 1.6319742757477633, "learning_rate": 9.991616467431486e-06, "loss": 0.7906, "step": 260 }, { "epoch": 0.1210762331838565, "grad_norm": 1.6863475759473823, "learning_rate": 9.986474764680236e-06, "loss": 0.7684, "step": 270 }, { "epoch": 0.12556053811659193, "grad_norm": 1.5563949560805244, "learning_rate": 9.98011129108149e-06, "loss": 0.793, "step": 280 }, { "epoch": 0.13004484304932734, "grad_norm": 1.6203440473254576, "learning_rate": 9.972527605794151e-06, "loss": 0.771, "step": 290 }, { "epoch": 0.13452914798206278, "grad_norm": 1.5211967329248808, "learning_rate": 9.963725566949674e-06, "loss": 0.7865, "step": 300 }, { "epoch": 0.13901345291479822, "grad_norm": 1.3996609263865165, "learning_rate": 9.953707331196787e-06, "loss": 0.7397, "step": 310 }, { "epoch": 0.14349775784753363, "grad_norm": 1.7467903395183983, "learning_rate": 9.94247535317308e-06, "loss": 0.802, "step": 320 }, { "epoch": 0.14798206278026907, "grad_norm": 1.4561862746225176, "learning_rate": 9.930032384903566e-06, "loss": 0.78, "step": 330 }, { "epoch": 0.15246636771300448, "grad_norm": 1.261592243705757, "learning_rate": 9.916381475126406e-06, "loss": 0.7678, "step": 340 }, { "epoch": 0.15695067264573992, "grad_norm": 1.3845441861346746, "learning_rate": 9.901525968545907e-06, "loss": 0.7462, "step": 350 }, { "epoch": 0.16143497757847533, "grad_norm": 1.330783975393604, "learning_rate": 9.885469505013006e-06, "loss": 0.7516, "step": 360 }, { "epoch": 0.16591928251121077, "grad_norm": 1.5411411223309597, "learning_rate": 9.868216018633456e-06, "loss": 0.7617, "step": 370 }, { "epoch": 0.17040358744394618, "grad_norm": 1.4997661320107978, "learning_rate": 9.8497697368039e-06, "loss": 0.7433, "step": 380 }, { "epoch": 0.17488789237668162, "grad_norm": 1.4613487182945122, "learning_rate": 9.830135179176086e-06, "loss": 0.7977, "step": 390 }, { "epoch": 0.17937219730941703, "grad_norm": 1.3823002584421413, "learning_rate": 9.809317156549476e-06, "loss": 0.7668, "step": 400 }, { "epoch": 0.18385650224215247, "grad_norm": 1.3405610913825137, "learning_rate": 9.787320769692517e-06, "loss": 0.755, "step": 410 }, { "epoch": 0.18834080717488788, "grad_norm": 1.321174566440371, "learning_rate": 9.76415140809287e-06, "loss": 0.7712, "step": 420 }, { "epoch": 0.19282511210762332, "grad_norm": 1.5855307049280556, "learning_rate": 9.739814748636892e-06, "loss": 0.7876, "step": 430 }, { "epoch": 0.19730941704035873, "grad_norm": 1.3277684374580685, "learning_rate": 9.7143167542187e-06, "loss": 0.7497, "step": 440 }, { "epoch": 0.20179372197309417, "grad_norm": 1.3316052843966186, "learning_rate": 9.687663672279167e-06, "loss": 0.7742, "step": 450 }, { "epoch": 0.2062780269058296, "grad_norm": 1.3422686862026139, "learning_rate": 9.659862033275188e-06, "loss": 0.7443, "step": 460 }, { "epoch": 0.21076233183856502, "grad_norm": 1.3794382112115433, "learning_rate": 9.630918649079606e-06, "loss": 0.7423, "step": 470 }, { "epoch": 0.21524663677130046, "grad_norm": 1.3473781761704757, "learning_rate": 9.600840611312198e-06, "loss": 0.756, "step": 480 }, { "epoch": 0.21973094170403587, "grad_norm": 1.477159074283593, "learning_rate": 9.569635289602098e-06, "loss": 0.758, "step": 490 }, { "epoch": 0.2242152466367713, "grad_norm": 1.4406496372056654, "learning_rate": 9.537310329782109e-06, "loss": 0.7373, "step": 500 }, { "epoch": 0.22869955156950672, "grad_norm": 1.3236307975388621, "learning_rate": 9.503873652015358e-06, "loss": 0.7485, "step": 510 }, { "epoch": 0.23318385650224216, "grad_norm": 1.2168213830447414, "learning_rate": 9.469333448854713e-06, "loss": 0.7518, "step": 520 }, { "epoch": 0.23766816143497757, "grad_norm": 1.4695219464522695, "learning_rate": 9.433698183235468e-06, "loss": 0.7389, "step": 530 }, { "epoch": 0.242152466367713, "grad_norm": 1.5460371795366352, "learning_rate": 9.39697658640179e-06, "loss": 0.7606, "step": 540 }, { "epoch": 0.24663677130044842, "grad_norm": 1.36384817307445, "learning_rate": 9.359177655767398e-06, "loss": 0.7573, "step": 550 }, { "epoch": 0.25112107623318386, "grad_norm": 1.2974873879986306, "learning_rate": 9.320310652711062e-06, "loss": 0.7447, "step": 560 }, { "epoch": 0.2556053811659193, "grad_norm": 1.3924395179060305, "learning_rate": 9.2803851003074e-06, "loss": 0.7346, "step": 570 }, { "epoch": 0.2600896860986547, "grad_norm": 1.3646463661968233, "learning_rate": 9.239410780993565e-06, "loss": 0.7637, "step": 580 }, { "epoch": 0.2645739910313901, "grad_norm": 1.5500883315093192, "learning_rate": 9.197397734172381e-06, "loss": 0.7352, "step": 590 }, { "epoch": 0.26905829596412556, "grad_norm": 1.2384578247121611, "learning_rate": 9.154356253752519e-06, "loss": 0.7467, "step": 600 }, { "epoch": 0.273542600896861, "grad_norm": 1.4632407956897133, "learning_rate": 9.110296885626315e-06, "loss": 0.7592, "step": 610 }, { "epoch": 0.27802690582959644, "grad_norm": 1.4743516068749583, "learning_rate": 9.065230425085849e-06, "loss": 0.7471, "step": 620 }, { "epoch": 0.2825112107623318, "grad_norm": 1.609459180353317, "learning_rate": 9.01916791417792e-06, "loss": 0.7411, "step": 630 }, { "epoch": 0.28699551569506726, "grad_norm": 1.4163468213726333, "learning_rate": 8.97212063899854e-06, "loss": 0.7583, "step": 640 }, { "epoch": 0.2914798206278027, "grad_norm": 1.2789206960042645, "learning_rate": 8.924100126927672e-06, "loss": 0.7637, "step": 650 }, { "epoch": 0.29596412556053814, "grad_norm": 1.4348891847742615, "learning_rate": 8.87511814380481e-06, "loss": 0.7376, "step": 660 }, { "epoch": 0.3004484304932735, "grad_norm": 1.4055551714674843, "learning_rate": 8.825186691046156e-06, "loss": 0.7544, "step": 670 }, { "epoch": 0.30493273542600896, "grad_norm": 1.3076750983715024, "learning_rate": 8.774318002704072e-06, "loss": 0.7388, "step": 680 }, { "epoch": 0.3094170403587444, "grad_norm": 1.2832634236951583, "learning_rate": 8.722524542469517e-06, "loss": 0.7386, "step": 690 }, { "epoch": 0.31390134529147984, "grad_norm": 1.3612825976793976, "learning_rate": 8.669819000618248e-06, "loss": 0.768, "step": 700 }, { "epoch": 0.3183856502242152, "grad_norm": 1.3343477240590562, "learning_rate": 8.616214290901474e-06, "loss": 0.7244, "step": 710 }, { "epoch": 0.32286995515695066, "grad_norm": 1.3061898085537815, "learning_rate": 8.56172354738178e-06, "loss": 0.7368, "step": 720 }, { "epoch": 0.3273542600896861, "grad_norm": 1.423008056010291, "learning_rate": 8.506360121215046e-06, "loss": 0.7297, "step": 730 }, { "epoch": 0.33183856502242154, "grad_norm": 1.258418518673196, "learning_rate": 8.4501375773792e-06, "loss": 0.7322, "step": 740 }, { "epoch": 0.336322869955157, "grad_norm": 1.4805131801943718, "learning_rate": 8.39306969135056e-06, "loss": 0.7284, "step": 750 }, { "epoch": 0.34080717488789236, "grad_norm": 1.3260411374117855, "learning_rate": 8.335170445728609e-06, "loss": 0.7618, "step": 760 }, { "epoch": 0.3452914798206278, "grad_norm": 1.3351013471897553, "learning_rate": 8.276454026810026e-06, "loss": 0.7454, "step": 770 }, { "epoch": 0.34977578475336324, "grad_norm": 1.267536863507402, "learning_rate": 8.216934821112803e-06, "loss": 0.742, "step": 780 }, { "epoch": 0.3542600896860987, "grad_norm": 1.4791659570865663, "learning_rate": 8.156627411851295e-06, "loss": 0.7483, "step": 790 }, { "epoch": 0.35874439461883406, "grad_norm": 1.4042375363677306, "learning_rate": 8.095546575363098e-06, "loss": 0.7134, "step": 800 }, { "epoch": 0.3632286995515695, "grad_norm": 1.2552675050466975, "learning_rate": 8.033707277488585e-06, "loss": 0.7186, "step": 810 }, { "epoch": 0.36771300448430494, "grad_norm": 1.281776208560821, "learning_rate": 7.97112466990403e-06, "loss": 0.7367, "step": 820 }, { "epoch": 0.3721973094170404, "grad_norm": 1.417675486477273, "learning_rate": 7.907814086409183e-06, "loss": 0.7399, "step": 830 }, { "epoch": 0.37668161434977576, "grad_norm": 1.4911782746859528, "learning_rate": 7.843791039170232e-06, "loss": 0.738, "step": 840 }, { "epoch": 0.3811659192825112, "grad_norm": 1.3628167592658191, "learning_rate": 7.779071214919068e-06, "loss": 0.7404, "step": 850 }, { "epoch": 0.38565022421524664, "grad_norm": 1.256534318600846, "learning_rate": 7.713670471109749e-06, "loss": 0.7364, "step": 860 }, { "epoch": 0.3901345291479821, "grad_norm": 1.3061985247991272, "learning_rate": 7.647604832033178e-06, "loss": 0.7535, "step": 870 }, { "epoch": 0.39461883408071746, "grad_norm": 1.3626743194809932, "learning_rate": 7.580890484890864e-06, "loss": 0.7212, "step": 880 }, { "epoch": 0.3991031390134529, "grad_norm": 1.283405821881289, "learning_rate": 7.513543775828791e-06, "loss": 0.7336, "step": 890 }, { "epoch": 0.40358744394618834, "grad_norm": 1.4217858183255085, "learning_rate": 7.445581205932335e-06, "loss": 0.7349, "step": 900 }, { "epoch": 0.4080717488789238, "grad_norm": 1.3304359323377464, "learning_rate": 7.377019427183213e-06, "loss": 0.7265, "step": 910 }, { "epoch": 0.4125560538116592, "grad_norm": 1.3324230635898773, "learning_rate": 7.30787523837947e-06, "loss": 0.7451, "step": 920 }, { "epoch": 0.4170403587443946, "grad_norm": 1.2976336030674929, "learning_rate": 7.238165581019488e-06, "loss": 0.7415, "step": 930 }, { "epoch": 0.42152466367713004, "grad_norm": 1.3082251443805497, "learning_rate": 7.167907535151027e-06, "loss": 0.7405, "step": 940 }, { "epoch": 0.4260089686098655, "grad_norm": 1.368310990039969, "learning_rate": 7.097118315186335e-06, "loss": 0.7141, "step": 950 }, { "epoch": 0.4304932735426009, "grad_norm": 1.3507885090012453, "learning_rate": 7.025815265684315e-06, "loss": 0.744, "step": 960 }, { "epoch": 0.4349775784753363, "grad_norm": 1.3870418574466943, "learning_rate": 6.9540158571008105e-06, "loss": 0.7344, "step": 970 }, { "epoch": 0.43946188340807174, "grad_norm": 1.2702369270052376, "learning_rate": 6.881737681508065e-06, "loss": 0.7131, "step": 980 }, { "epoch": 0.4439461883408072, "grad_norm": 1.3694518474478212, "learning_rate": 6.808998448284347e-06, "loss": 0.7516, "step": 990 }, { "epoch": 0.4484304932735426, "grad_norm": 1.3967080566658139, "learning_rate": 6.735815979774865e-06, "loss": 0.7208, "step": 1000 }, { "epoch": 0.452914798206278, "grad_norm": 1.4517646369323314, "learning_rate": 6.662208206924986e-06, "loss": 0.7455, "step": 1010 }, { "epoch": 0.45739910313901344, "grad_norm": 1.2752229454885209, "learning_rate": 6.588193164886847e-06, "loss": 0.7555, "step": 1020 }, { "epoch": 0.4618834080717489, "grad_norm": 1.3528261238688069, "learning_rate": 6.513788988600441e-06, "loss": 0.7428, "step": 1030 }, { "epoch": 0.4663677130044843, "grad_norm": 1.340831615562883, "learning_rate": 6.439013908350249e-06, "loss": 0.7446, "step": 1040 }, { "epoch": 0.47085201793721976, "grad_norm": 1.3233981064583105, "learning_rate": 6.363886245298514e-06, "loss": 0.6945, "step": 1050 }, { "epoch": 0.47533632286995514, "grad_norm": 1.29683857481815, "learning_rate": 6.288424406996237e-06, "loss": 0.7085, "step": 1060 }, { "epoch": 0.4798206278026906, "grad_norm": 1.3009424922765027, "learning_rate": 6.2126468828730225e-06, "loss": 0.7294, "step": 1070 }, { "epoch": 0.484304932735426, "grad_norm": 1.244204104767217, "learning_rate": 6.136572239706854e-06, "loss": 0.7091, "step": 1080 }, { "epoch": 0.48878923766816146, "grad_norm": 1.4080424781447183, "learning_rate": 6.060219117074914e-06, "loss": 0.724, "step": 1090 }, { "epoch": 0.49327354260089684, "grad_norm": 1.4328593518662633, "learning_rate": 5.983606222786577e-06, "loss": 0.7106, "step": 1100 }, { "epoch": 0.4977578475336323, "grad_norm": 1.3515007322722326, "learning_rate": 5.9067523282996775e-06, "loss": 0.7111, "step": 1110 }, { "epoch": 0.5022421524663677, "grad_norm": 1.3343622572750706, "learning_rate": 5.829676264121184e-06, "loss": 0.7323, "step": 1120 }, { "epoch": 0.5067264573991032, "grad_norm": 1.3346769678858248, "learning_rate": 5.752396915193403e-06, "loss": 0.744, "step": 1130 }, { "epoch": 0.5112107623318386, "grad_norm": 1.4071259742434898, "learning_rate": 5.6749332162668525e-06, "loss": 0.7181, "step": 1140 }, { "epoch": 0.515695067264574, "grad_norm": 1.314030857289351, "learning_rate": 5.5973041472609265e-06, "loss": 0.7278, "step": 1150 }, { "epoch": 0.5201793721973094, "grad_norm": 1.2893634754111014, "learning_rate": 5.519528728613491e-06, "loss": 0.722, "step": 1160 }, { "epoch": 0.5246636771300448, "grad_norm": 1.2894378069172643, "learning_rate": 5.4416260166205525e-06, "loss": 0.7282, "step": 1170 }, { "epoch": 0.5291479820627802, "grad_norm": 1.2691657277439665, "learning_rate": 5.363615098767149e-06, "loss": 0.7439, "step": 1180 }, { "epoch": 0.5336322869955157, "grad_norm": 1.3478124675299437, "learning_rate": 5.285515089050587e-06, "loss": 0.7164, "step": 1190 }, { "epoch": 0.5381165919282511, "grad_norm": 1.3052899854504807, "learning_rate": 5.207345123297187e-06, "loss": 0.7171, "step": 1200 }, { "epoch": 0.5426008968609866, "grad_norm": 1.27131464716386, "learning_rate": 5.129124354473688e-06, "loss": 0.7235, "step": 1210 }, { "epoch": 0.547085201793722, "grad_norm": 1.257329353514232, "learning_rate": 5.050871947994443e-06, "loss": 0.6999, "step": 1220 }, { "epoch": 0.5515695067264574, "grad_norm": 1.3863466652299603, "learning_rate": 4.972607077025563e-06, "loss": 0.7251, "step": 1230 }, { "epoch": 0.5560538116591929, "grad_norm": 1.394706696869187, "learning_rate": 4.894348917787174e-06, "loss": 0.6963, "step": 1240 }, { "epoch": 0.5605381165919282, "grad_norm": 1.288516832053435, "learning_rate": 4.816116644854912e-06, "loss": 0.7207, "step": 1250 }, { "epoch": 0.5650224215246636, "grad_norm": 1.291568839795731, "learning_rate": 4.73792942646183e-06, "loss": 0.7168, "step": 1260 }, { "epoch": 0.5695067264573991, "grad_norm": 1.3692501176044114, "learning_rate": 4.659806419801855e-06, "loss": 0.7311, "step": 1270 }, { "epoch": 0.5739910313901345, "grad_norm": 1.2214111714932894, "learning_rate": 4.581766766335953e-06, "loss": 0.7175, "step": 1280 }, { "epoch": 0.57847533632287, "grad_norm": 1.2261021201524966, "learning_rate": 4.503829587102138e-06, "loss": 0.722, "step": 1290 }, { "epoch": 0.5829596412556054, "grad_norm": 1.279552435182188, "learning_rate": 4.426013978030508e-06, "loss": 0.7407, "step": 1300 }, { "epoch": 0.5874439461883408, "grad_norm": 1.2639914975822624, "learning_rate": 4.348339005264406e-06, "loss": 0.7174, "step": 1310 }, { "epoch": 0.5919282511210763, "grad_norm": 1.2931987303236723, "learning_rate": 4.270823700488896e-06, "loss": 0.7236, "step": 1320 }, { "epoch": 0.5964125560538116, "grad_norm": 1.3083460389811294, "learning_rate": 4.19348705626768e-06, "loss": 0.7247, "step": 1330 }, { "epoch": 0.600896860986547, "grad_norm": 1.359946232054244, "learning_rate": 4.116348021389595e-06, "loss": 0.7289, "step": 1340 }, { "epoch": 0.6053811659192825, "grad_norm": 1.3038409489179155, "learning_rate": 4.039425496225834e-06, "loss": 0.723, "step": 1350 }, { "epoch": 0.6098654708520179, "grad_norm": 1.337031671252276, "learning_rate": 3.962738328099047e-06, "loss": 0.718, "step": 1360 }, { "epoch": 0.6143497757847534, "grad_norm": 1.3006506595251124, "learning_rate": 3.88630530666542e-06, "loss": 0.7372, "step": 1370 }, { "epoch": 0.6188340807174888, "grad_norm": 1.3038120667886732, "learning_rate": 3.8101451593108816e-06, "loss": 0.732, "step": 1380 }, { "epoch": 0.6233183856502242, "grad_norm": 1.2544712968929104, "learning_rate": 3.7342765465625953e-06, "loss": 0.7347, "step": 1390 }, { "epoch": 0.6278026905829597, "grad_norm": 1.3352755014667614, "learning_rate": 3.658718057516803e-06, "loss": 0.7332, "step": 1400 }, { "epoch": 0.6322869955156951, "grad_norm": 1.3389617347187606, "learning_rate": 3.5834882052841744e-06, "loss": 0.7154, "step": 1410 }, { "epoch": 0.6367713004484304, "grad_norm": 1.2654799213890686, "learning_rate": 3.508605422453799e-06, "loss": 0.7002, "step": 1420 }, { "epoch": 0.6412556053811659, "grad_norm": 1.311398422880929, "learning_rate": 3.4340880565768707e-06, "loss": 0.7098, "step": 1430 }, { "epoch": 0.6457399103139013, "grad_norm": 1.3577211660369808, "learning_rate": 3.359954365671241e-06, "loss": 0.7024, "step": 1440 }, { "epoch": 0.6502242152466368, "grad_norm": 1.145375472595952, "learning_rate": 3.2862225137478897e-06, "loss": 0.7097, "step": 1450 }, { "epoch": 0.6547085201793722, "grad_norm": 1.3358950336855993, "learning_rate": 3.2129105663604275e-06, "loss": 0.7148, "step": 1460 }, { "epoch": 0.6591928251121076, "grad_norm": 1.2495395838860759, "learning_rate": 3.1400364861787434e-06, "loss": 0.7483, "step": 1470 }, { "epoch": 0.6636771300448431, "grad_norm": 1.2742823576583961, "learning_rate": 3.0676181285878343e-06, "loss": 0.7063, "step": 1480 }, { "epoch": 0.6681614349775785, "grad_norm": 1.2596513250823083, "learning_rate": 2.9956732373129378e-06, "loss": 0.7201, "step": 1490 }, { "epoch": 0.672645739910314, "grad_norm": 1.4183600229677984, "learning_rate": 2.9242194400720157e-06, "loss": 0.7202, "step": 1500 }, { "epoch": 0.6771300448430493, "grad_norm": 1.224500332009707, "learning_rate": 2.8532742442566735e-06, "loss": 0.7228, "step": 1510 }, { "epoch": 0.6816143497757847, "grad_norm": 1.20210393613667, "learning_rate": 2.782855032642535e-06, "loss": 0.7386, "step": 1520 }, { "epoch": 0.6860986547085202, "grad_norm": 1.2835056973370584, "learning_rate": 2.712979059130187e-06, "loss": 0.7207, "step": 1530 }, { "epoch": 0.6905829596412556, "grad_norm": 1.180714987729606, "learning_rate": 2.643663444517671e-06, "loss": 0.6981, "step": 1540 }, { "epoch": 0.695067264573991, "grad_norm": 1.2871858226590431, "learning_rate": 2.5749251723055933e-06, "loss": 0.6853, "step": 1550 }, { "epoch": 0.6995515695067265, "grad_norm": 1.3219720717807693, "learning_rate": 2.5067810845358926e-06, "loss": 0.7192, "step": 1560 }, { "epoch": 0.7040358744394619, "grad_norm": 1.391893981214182, "learning_rate": 2.439247877665244e-06, "loss": 0.7103, "step": 1570 }, { "epoch": 0.7085201793721974, "grad_norm": 1.2636799158865641, "learning_rate": 2.3723420984741417e-06, "loss": 0.684, "step": 1580 }, { "epoch": 0.7130044843049327, "grad_norm": 1.3557464635552046, "learning_rate": 2.3060801400126693e-06, "loss": 0.7207, "step": 1590 }, { "epoch": 0.7174887892376681, "grad_norm": 1.3569256088684083, "learning_rate": 2.240478237583915e-06, "loss": 0.7077, "step": 1600 }, { "epoch": 0.7219730941704036, "grad_norm": 1.37192661939199, "learning_rate": 2.1755524647660514e-06, "loss": 0.693, "step": 1610 }, { "epoch": 0.726457399103139, "grad_norm": 1.248391921620642, "learning_rate": 2.1113187294740294e-06, "loss": 0.6911, "step": 1620 }, { "epoch": 0.7309417040358744, "grad_norm": 1.357856212054366, "learning_rate": 2.047792770061881e-06, "loss": 0.6838, "step": 1630 }, { "epoch": 0.7354260089686099, "grad_norm": 1.4147463764446673, "learning_rate": 1.9849901514665458e-06, "loss": 0.7122, "step": 1640 }, { "epoch": 0.7399103139013453, "grad_norm": 1.3587967144630926, "learning_rate": 1.922926261394206e-06, "loss": 0.6927, "step": 1650 }, { "epoch": 0.7443946188340808, "grad_norm": 1.326112035320906, "learning_rate": 1.8616163065500231e-06, "loss": 0.6931, "step": 1660 }, { "epoch": 0.7488789237668162, "grad_norm": 1.2337199420172695, "learning_rate": 1.8010753089122572e-06, "loss": 0.6934, "step": 1670 }, { "epoch": 0.7533632286995515, "grad_norm": 1.2176765529306792, "learning_rate": 1.7413181020516146e-06, "loss": 0.7164, "step": 1680 }, { "epoch": 0.757847533632287, "grad_norm": 1.2724919849788396, "learning_rate": 1.6823593274967703e-06, "loss": 0.7267, "step": 1690 }, { "epoch": 0.7623318385650224, "grad_norm": 1.3274428154634692, "learning_rate": 1.6242134311469538e-06, "loss": 0.6824, "step": 1700 }, { "epoch": 0.7668161434977578, "grad_norm": 1.3661211722387332, "learning_rate": 1.5668946597324558e-06, "loss": 0.7182, "step": 1710 }, { "epoch": 0.7713004484304933, "grad_norm": 1.2959174525555186, "learning_rate": 1.51041705732393e-06, "loss": 0.7118, "step": 1720 }, { "epoch": 0.7757847533632287, "grad_norm": 1.2773849053623296, "learning_rate": 1.4547944618913706e-06, "loss": 0.6929, "step": 1730 }, { "epoch": 0.7802690582959642, "grad_norm": 1.297150269984566, "learning_rate": 1.4000405019135676e-06, "loss": 0.6883, "step": 1740 }, { "epoch": 0.7847533632286996, "grad_norm": 1.400539444832659, "learning_rate": 1.3461685930388958e-06, "loss": 0.6911, "step": 1750 }, { "epoch": 0.7892376681614349, "grad_norm": 1.2337657683341194, "learning_rate": 1.2931919347982607e-06, "loss": 0.6921, "step": 1760 }, { "epoch": 0.7937219730941704, "grad_norm": 1.2072415924732212, "learning_rate": 1.2411235073709883e-06, "loss": 0.7102, "step": 1770 }, { "epoch": 0.7982062780269058, "grad_norm": 1.381367292756694, "learning_rate": 1.1899760684044515e-06, "loss": 0.6838, "step": 1780 }, { "epoch": 0.8026905829596412, "grad_norm": 1.3102643665842388, "learning_rate": 1.1397621498882471e-06, "loss": 0.6945, "step": 1790 }, { "epoch": 0.8071748878923767, "grad_norm": 1.3214717866253802, "learning_rate": 1.0904940550836285e-06, "loss": 0.7016, "step": 1800 }, { "epoch": 0.8116591928251121, "grad_norm": 1.1240219719921938, "learning_rate": 1.0421838555090119e-06, "loss": 0.7018, "step": 1810 }, { "epoch": 0.8161434977578476, "grad_norm": 1.2844428659053804, "learning_rate": 9.948433879822428e-07, "loss": 0.7361, "step": 1820 }, { "epoch": 0.820627802690583, "grad_norm": 1.249492840397432, "learning_rate": 9.484842517203735e-07, "loss": 0.707, "step": 1830 }, { "epoch": 0.8251121076233184, "grad_norm": 1.2793612843154911, "learning_rate": 9.031178054976636e-07, "loss": 0.7226, "step": 1840 }, { "epoch": 0.8295964125560538, "grad_norm": 1.3447642426455195, "learning_rate": 8.587551648624859e-07, "loss": 0.6906, "step": 1850 }, { "epoch": 0.8340807174887892, "grad_norm": 1.282168137173779, "learning_rate": 8.154071994138241e-07, "loss": 0.698, "step": 1860 }, { "epoch": 0.8385650224215246, "grad_norm": 1.4629920435128856, "learning_rate": 7.730845301380441e-07, "loss": 0.7212, "step": 1870 }, { "epoch": 0.8430493273542601, "grad_norm": 1.3143459341111923, "learning_rate": 7.317975268065685e-07, "loss": 0.6942, "step": 1880 }, { "epoch": 0.8475336322869955, "grad_norm": 1.3274198546425116, "learning_rate": 6.915563054351037e-07, "loss": 0.6944, "step": 1890 }, { "epoch": 0.852017937219731, "grad_norm": 1.2958298618538142, "learning_rate": 6.523707258050516e-07, "loss": 0.6692, "step": 1900 }, { "epoch": 0.8565022421524664, "grad_norm": 1.3403229426415875, "learning_rate": 6.14250389047692e-07, "loss": 0.7034, "step": 1910 }, { "epoch": 0.8609865470852018, "grad_norm": 1.258601487476543, "learning_rate": 5.772046352917399e-07, "loss": 0.7144, "step": 1920 }, { "epoch": 0.8654708520179372, "grad_norm": 1.1222145256140716, "learning_rate": 5.412425413748623e-07, "loss": 0.6988, "step": 1930 }, { "epoch": 0.8699551569506726, "grad_norm": 1.2674996138499859, "learning_rate": 5.063729186196948e-07, "loss": 0.7089, "step": 1940 }, { "epoch": 0.874439461883408, "grad_norm": 1.3633828734617575, "learning_rate": 4.7260431067491617e-07, "loss": 0.733, "step": 1950 }, { "epoch": 0.8789237668161435, "grad_norm": 1.2652212276863493, "learning_rate": 4.399449914219167e-07, "loss": 0.7209, "step": 1960 }, { "epoch": 0.8834080717488789, "grad_norm": 1.2975811004623845, "learning_rate": 4.084029629475478e-07, "loss": 0.7252, "step": 1970 }, { "epoch": 0.8878923766816144, "grad_norm": 1.2744740809644324, "learning_rate": 3.7798595358348457e-07, "loss": 0.7083, "step": 1980 }, { "epoch": 0.8923766816143498, "grad_norm": 1.2924446335095698, "learning_rate": 3.487014160126467e-07, "loss": 0.7077, "step": 1990 }, { "epoch": 0.8968609865470852, "grad_norm": 1.283938673421291, "learning_rate": 3.2055652544316695e-07, "loss": 0.7038, "step": 2000 }, { "epoch": 0.9013452914798207, "grad_norm": 1.3428158271865258, "learning_rate": 2.9355817785034325e-07, "loss": 0.7177, "step": 2010 }, { "epoch": 0.905829596412556, "grad_norm": 1.3237588390932393, "learning_rate": 2.6771298828700885e-07, "loss": 0.7079, "step": 2020 }, { "epoch": 0.9103139013452914, "grad_norm": 1.379188659318735, "learning_rate": 2.4302728926273224e-07, "loss": 0.7159, "step": 2030 }, { "epoch": 0.9147982062780269, "grad_norm": 1.2853687562313414, "learning_rate": 2.195071291922435e-07, "loss": 0.6842, "step": 2040 }, { "epoch": 0.9192825112107623, "grad_norm": 1.3212853096546313, "learning_rate": 1.9715827091347005e-07, "loss": 0.6994, "step": 2050 }, { "epoch": 0.9237668161434978, "grad_norm": 1.4231610496002076, "learning_rate": 1.7598619027554553e-07, "loss": 0.7032, "step": 2060 }, { "epoch": 0.9282511210762332, "grad_norm": 1.288660209861224, "learning_rate": 1.5599607479713396e-07, "loss": 0.6856, "step": 2070 }, { "epoch": 0.9327354260089686, "grad_norm": 1.291080758227027, "learning_rate": 1.3719282239539722e-07, "loss": 0.7183, "step": 2080 }, { "epoch": 0.9372197309417041, "grad_norm": 1.3527938026636528, "learning_rate": 1.1958104018592376e-07, "loss": 0.7022, "step": 2090 }, { "epoch": 0.9417040358744395, "grad_norm": 1.306930563286188, "learning_rate": 1.0316504335390775e-07, "loss": 0.7202, "step": 2100 }, { "epoch": 0.9461883408071748, "grad_norm": 1.34728936572921, "learning_rate": 8.79488540968565e-08, "loss": 0.7128, "step": 2110 }, { "epoch": 0.9506726457399103, "grad_norm": 1.5138997567977701, "learning_rate": 7.39362006390798e-08, "loss": 0.6841, "step": 2120 }, { "epoch": 0.9551569506726457, "grad_norm": 1.2518587224976194, "learning_rate": 6.113051631821631e-08, "loss": 0.71, "step": 2130 }, { "epoch": 0.9596412556053812, "grad_norm": 1.4419766059722212, "learning_rate": 4.9534938744004723e-08, "loss": 0.6944, "step": 2140 }, { "epoch": 0.9641255605381166, "grad_norm": 1.3591089992663685, "learning_rate": 3.915230902951761e-08, "loss": 0.717, "step": 2150 }, { "epoch": 0.968609865470852, "grad_norm": 1.281084457332566, "learning_rate": 2.9985171095041066e-08, "loss": 0.703, "step": 2160 }, { "epoch": 0.9730941704035875, "grad_norm": 1.1922556783094496, "learning_rate": 2.203577104476773e-08, "loss": 0.7085, "step": 2170 }, { "epoch": 0.9775784753363229, "grad_norm": 1.3917918614257676, "learning_rate": 1.5306056616468666e-08, "loss": 0.709, "step": 2180 }, { "epoch": 0.9820627802690582, "grad_norm": 1.178733370314234, "learning_rate": 9.797676704259574e-09, "loss": 0.7009, "step": 2190 }, { "epoch": 0.9865470852017937, "grad_norm": 1.2352452900540438, "learning_rate": 5.511980954596152e-09, "loss": 0.7085, "step": 2200 }, { "epoch": 0.9910313901345291, "grad_norm": 1.519249312059187, "learning_rate": 2.4500194355880913e-09, "loss": 0.6865, "step": 2210 }, { "epoch": 0.9955156950672646, "grad_norm": 1.2975150315542752, "learning_rate": 6.125423797137541e-10, "loss": 0.6935, "step": 2220 }, { "epoch": 1.0, "grad_norm": 1.2570272992562246, "learning_rate": 0.0, "loss": 0.6946, "step": 2230 }, { "epoch": 1.0, "eval_loss": 0.7042345404624939, "eval_runtime": 381.2635, "eval_samples_per_second": 131.345, "eval_steps_per_second": 2.054, "step": 2230 }, { "epoch": 1.0, "step": 2230, "total_flos": 244955314192384.0, "train_loss": 0.7376568270371099, "train_runtime": 8720.6202, "train_samples_per_second": 32.731, "train_steps_per_second": 0.256 } ], "logging_steps": 10, "max_steps": 2230, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 244955314192384.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }