{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.5159621490167481, "eval_steps": 500, "global_step": 1500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003439747660111654, "grad_norm": 4.688905239105225, "learning_rate": 1.7199862401100792e-07, "loss": 13.1365, "step": 10 }, { "epoch": 0.006879495320223308, "grad_norm": 4.80711030960083, "learning_rate": 3.4399724802201585e-07, "loss": 13.154, "step": 20 }, { "epoch": 0.010319242980334962, "grad_norm": 4.933587074279785, "learning_rate": 5.159958720330238e-07, "loss": 13.1537, "step": 30 }, { "epoch": 0.013758990640446616, "grad_norm": 4.939642906188965, "learning_rate": 6.879944960440317e-07, "loss": 13.1529, "step": 40 }, { "epoch": 0.01719873830055827, "grad_norm": 5.070457935333252, "learning_rate": 8.599931200550396e-07, "loss": 13.1224, "step": 50 }, { "epoch": 0.020638485960669924, "grad_norm": 5.011481285095215, "learning_rate": 1.0319917440660475e-06, "loss": 13.0346, "step": 60 }, { "epoch": 0.02407823362078158, "grad_norm": 5.395399570465088, "learning_rate": 1.2039903680770555e-06, "loss": 12.9762, "step": 70 }, { "epoch": 0.027517981280893233, "grad_norm": 5.811434745788574, "learning_rate": 1.3759889920880634e-06, "loss": 12.853, "step": 80 }, { "epoch": 0.03095772894100489, "grad_norm": 6.088072776794434, "learning_rate": 1.5479876160990713e-06, "loss": 12.6852, "step": 90 }, { "epoch": 0.03439747660111654, "grad_norm": 6.066171646118164, "learning_rate": 1.7199862401100792e-06, "loss": 12.4486, "step": 100 }, { "epoch": 0.037837224261228194, "grad_norm": 6.659308910369873, "learning_rate": 1.891984864121087e-06, "loss": 12.1745, "step": 110 }, { "epoch": 0.04127697192133985, "grad_norm": 7.230446815490723, "learning_rate": 2.063983488132095e-06, "loss": 11.8162, "step": 120 }, { "epoch": 0.04471671958145151, "grad_norm": 7.676384449005127, "learning_rate": 2.2359821121431026e-06, "loss": 11.3668, "step": 130 }, { "epoch": 0.04815646724156316, "grad_norm": 8.068939208984375, "learning_rate": 2.407980736154111e-06, "loss": 10.9014, "step": 140 }, { "epoch": 0.05159621490167481, "grad_norm": 8.794692039489746, "learning_rate": 2.5799793601651184e-06, "loss": 10.327, "step": 150 }, { "epoch": 0.055035962561786465, "grad_norm": 9.102154731750488, "learning_rate": 2.7519779841761268e-06, "loss": 9.6304, "step": 160 }, { "epoch": 0.05847571022189812, "grad_norm": 8.208027839660645, "learning_rate": 2.9239766081871343e-06, "loss": 8.8888, "step": 170 }, { "epoch": 0.06191545788200978, "grad_norm": 6.106681823730469, "learning_rate": 3.0959752321981426e-06, "loss": 8.1024, "step": 180 }, { "epoch": 0.06535520554212143, "grad_norm": 3.8156211376190186, "learning_rate": 3.2679738562091506e-06, "loss": 7.5017, "step": 190 }, { "epoch": 0.06879495320223308, "grad_norm": 3.4960334300994873, "learning_rate": 3.4399724802201585e-06, "loss": 7.0312, "step": 200 }, { "epoch": 0.07223470086234474, "grad_norm": 3.1875991821289062, "learning_rate": 3.611971104231166e-06, "loss": 6.7019, "step": 210 }, { "epoch": 0.07567444852245639, "grad_norm": 3.0836124420166016, "learning_rate": 3.783969728242174e-06, "loss": 6.3059, "step": 220 }, { "epoch": 0.07911419618256804, "grad_norm": 2.8337793350219727, "learning_rate": 3.955968352253183e-06, "loss": 5.9064, "step": 230 }, { "epoch": 0.0825539438426797, "grad_norm": 2.593134641647339, "learning_rate": 4.12796697626419e-06, "loss": 5.5292, "step": 240 }, { "epoch": 0.08599369150279136, "grad_norm": 2.23457670211792, "learning_rate": 4.299965600275198e-06, "loss": 5.1508, "step": 250 }, { "epoch": 0.08943343916290301, "grad_norm": 2.220684051513672, "learning_rate": 4.471964224286205e-06, "loss": 4.9064, "step": 260 }, { "epoch": 0.09287318682301467, "grad_norm": 2.066509962081909, "learning_rate": 4.643962848297214e-06, "loss": 4.5587, "step": 270 }, { "epoch": 0.09631293448312632, "grad_norm": 1.9317891597747803, "learning_rate": 4.815961472308222e-06, "loss": 4.2968, "step": 280 }, { "epoch": 0.09975268214323797, "grad_norm": 1.852779746055603, "learning_rate": 4.987960096319229e-06, "loss": 3.9759, "step": 290 }, { "epoch": 0.10319242980334963, "grad_norm": 1.7298812866210938, "learning_rate": 5.159958720330237e-06, "loss": 3.6811, "step": 300 }, { "epoch": 0.10663217746346128, "grad_norm": 1.5647941827774048, "learning_rate": 5.331957344341246e-06, "loss": 3.4349, "step": 310 }, { "epoch": 0.11007192512357293, "grad_norm": 1.423771619796753, "learning_rate": 5.5039559683522536e-06, "loss": 3.1899, "step": 320 }, { "epoch": 0.11351167278368458, "grad_norm": 1.2160062789916992, "learning_rate": 5.675954592363261e-06, "loss": 3.0158, "step": 330 }, { "epoch": 0.11695142044379624, "grad_norm": 1.162079095840454, "learning_rate": 5.8479532163742686e-06, "loss": 2.8798, "step": 340 }, { "epoch": 0.1203911681039079, "grad_norm": 0.9763687252998352, "learning_rate": 6.019951840385278e-06, "loss": 2.7573, "step": 350 }, { "epoch": 0.12383091576401956, "grad_norm": 0.9514108300209045, "learning_rate": 6.191950464396285e-06, "loss": 2.6355, "step": 360 }, { "epoch": 0.1272706634241312, "grad_norm": 0.8850705623626709, "learning_rate": 6.363949088407294e-06, "loss": 2.5371, "step": 370 }, { "epoch": 0.13071041108424286, "grad_norm": 0.8824434280395508, "learning_rate": 6.535947712418301e-06, "loss": 2.4404, "step": 380 }, { "epoch": 0.13415015874435451, "grad_norm": 0.9136425256729126, "learning_rate": 6.7079463364293095e-06, "loss": 2.3809, "step": 390 }, { "epoch": 0.13758990640446617, "grad_norm": 1.0447123050689697, "learning_rate": 6.879944960440317e-06, "loss": 2.3013, "step": 400 }, { "epoch": 0.14102965406457782, "grad_norm": 1.0016720294952393, "learning_rate": 7.0519435844513245e-06, "loss": 2.2526, "step": 410 }, { "epoch": 0.14446940172468947, "grad_norm": 0.9885135293006897, "learning_rate": 7.223942208462332e-06, "loss": 2.2099, "step": 420 }, { "epoch": 0.14790914938480113, "grad_norm": 0.9515292048454285, "learning_rate": 7.39594083247334e-06, "loss": 2.1602, "step": 430 }, { "epoch": 0.15134889704491278, "grad_norm": 0.8287838101387024, "learning_rate": 7.567939456484348e-06, "loss": 2.1251, "step": 440 }, { "epoch": 0.15478864470502443, "grad_norm": 1.160489559173584, "learning_rate": 7.739938080495356e-06, "loss": 2.0836, "step": 450 }, { "epoch": 0.15822839236513608, "grad_norm": 1.1019588708877563, "learning_rate": 7.911936704506365e-06, "loss": 2.0474, "step": 460 }, { "epoch": 0.16166814002524774, "grad_norm": 0.9389374256134033, "learning_rate": 8.083935328517373e-06, "loss": 2.013, "step": 470 }, { "epoch": 0.1651078876853594, "grad_norm": 0.9914370775222778, "learning_rate": 8.25593395252838e-06, "loss": 1.9829, "step": 480 }, { "epoch": 0.16854763534547107, "grad_norm": 0.8827924132347107, "learning_rate": 8.427932576539388e-06, "loss": 1.9574, "step": 490 }, { "epoch": 0.17198738300558272, "grad_norm": 1.0819416046142578, "learning_rate": 8.599931200550395e-06, "loss": 1.9333, "step": 500 }, { "epoch": 0.17542713066569438, "grad_norm": 1.006886601448059, "learning_rate": 8.771929824561403e-06, "loss": 1.8845, "step": 510 }, { "epoch": 0.17886687832580603, "grad_norm": 0.8748846054077148, "learning_rate": 8.94392844857241e-06, "loss": 1.8757, "step": 520 }, { "epoch": 0.18230662598591768, "grad_norm": 0.9551517963409424, "learning_rate": 9.11592707258342e-06, "loss": 1.8478, "step": 530 }, { "epoch": 0.18574637364602933, "grad_norm": 1.4100078344345093, "learning_rate": 9.287925696594429e-06, "loss": 1.8186, "step": 540 }, { "epoch": 0.189186121306141, "grad_norm": 1.5030393600463867, "learning_rate": 9.459924320605436e-06, "loss": 1.7939, "step": 550 }, { "epoch": 0.19262586896625264, "grad_norm": 1.0008896589279175, "learning_rate": 9.631922944616444e-06, "loss": 1.7711, "step": 560 }, { "epoch": 0.1960656166263643, "grad_norm": 1.5640122890472412, "learning_rate": 9.803921568627451e-06, "loss": 1.7555, "step": 570 }, { "epoch": 0.19950536428647594, "grad_norm": 1.041014313697815, "learning_rate": 9.975920192638459e-06, "loss": 1.7305, "step": 580 }, { "epoch": 0.2029451119465876, "grad_norm": 1.2513008117675781, "learning_rate": 1.0147918816649466e-05, "loss": 1.7083, "step": 590 }, { "epoch": 0.20638485960669925, "grad_norm": 1.2905975580215454, "learning_rate": 1.0319917440660474e-05, "loss": 1.7061, "step": 600 }, { "epoch": 0.2098246072668109, "grad_norm": 1.0839684009552002, "learning_rate": 1.0491916064671483e-05, "loss": 1.6722, "step": 610 }, { "epoch": 0.21326435492692256, "grad_norm": 1.2286721467971802, "learning_rate": 1.0663914688682492e-05, "loss": 1.6444, "step": 620 }, { "epoch": 0.2167041025870342, "grad_norm": 1.3250325918197632, "learning_rate": 1.08359133126935e-05, "loss": 1.6229, "step": 630 }, { "epoch": 0.22014385024714586, "grad_norm": 1.176580548286438, "learning_rate": 1.1007911936704507e-05, "loss": 1.6143, "step": 640 }, { "epoch": 0.22358359790725751, "grad_norm": 1.4954670667648315, "learning_rate": 1.1179910560715515e-05, "loss": 1.597, "step": 650 }, { "epoch": 0.22702334556736917, "grad_norm": 1.0406630039215088, "learning_rate": 1.1351909184726522e-05, "loss": 1.5877, "step": 660 }, { "epoch": 0.23046309322748082, "grad_norm": 1.5156970024108887, "learning_rate": 1.152390780873753e-05, "loss": 1.5601, "step": 670 }, { "epoch": 0.23390284088759247, "grad_norm": 1.3882994651794434, "learning_rate": 1.1695906432748537e-05, "loss": 1.5456, "step": 680 }, { "epoch": 0.23734258854770415, "grad_norm": 1.5137856006622314, "learning_rate": 1.1867905056759546e-05, "loss": 1.5101, "step": 690 }, { "epoch": 0.2407823362078158, "grad_norm": 1.2775577306747437, "learning_rate": 1.2039903680770556e-05, "loss": 1.5042, "step": 700 }, { "epoch": 0.24422208386792746, "grad_norm": 1.3952692747116089, "learning_rate": 1.2211902304781563e-05, "loss": 1.4864, "step": 710 }, { "epoch": 0.2476618315280391, "grad_norm": 1.5909626483917236, "learning_rate": 1.238390092879257e-05, "loss": 1.4703, "step": 720 }, { "epoch": 0.25110157918815074, "grad_norm": 1.4731673002243042, "learning_rate": 1.255589955280358e-05, "loss": 1.4434, "step": 730 }, { "epoch": 0.2545413268482624, "grad_norm": 1.370367169380188, "learning_rate": 1.2727898176814587e-05, "loss": 1.4522, "step": 740 }, { "epoch": 0.25798107450837404, "grad_norm": 1.7441356182098389, "learning_rate": 1.2899896800825595e-05, "loss": 1.4124, "step": 750 }, { "epoch": 0.2614208221684857, "grad_norm": 1.6607742309570312, "learning_rate": 1.3071895424836602e-05, "loss": 1.409, "step": 760 }, { "epoch": 0.26486056982859735, "grad_norm": 1.423419713973999, "learning_rate": 1.324389404884761e-05, "loss": 1.3912, "step": 770 }, { "epoch": 0.26830031748870903, "grad_norm": 1.6139755249023438, "learning_rate": 1.3415892672858619e-05, "loss": 1.3753, "step": 780 }, { "epoch": 0.2717400651488207, "grad_norm": 1.9674487113952637, "learning_rate": 1.3587891296869626e-05, "loss": 1.3718, "step": 790 }, { "epoch": 0.27517981280893233, "grad_norm": 2.2793004512786865, "learning_rate": 1.3759889920880634e-05, "loss": 1.3426, "step": 800 }, { "epoch": 0.278619560469044, "grad_norm": 1.6088732481002808, "learning_rate": 1.3931888544891641e-05, "loss": 1.3336, "step": 810 }, { "epoch": 0.28205930812915564, "grad_norm": 2.0331807136535645, "learning_rate": 1.4103887168902649e-05, "loss": 1.3224, "step": 820 }, { "epoch": 0.2854990557892673, "grad_norm": 1.7430800199508667, "learning_rate": 1.4275885792913656e-05, "loss": 1.3212, "step": 830 }, { "epoch": 0.28893880344937894, "grad_norm": 1.4452030658721924, "learning_rate": 1.4447884416924664e-05, "loss": 1.3141, "step": 840 }, { "epoch": 0.2923785511094906, "grad_norm": 1.6565144062042236, "learning_rate": 1.4619883040935673e-05, "loss": 1.2885, "step": 850 }, { "epoch": 0.29581829876960225, "grad_norm": 1.3976999521255493, "learning_rate": 1.479188166494668e-05, "loss": 1.2906, "step": 860 }, { "epoch": 0.29925804642971393, "grad_norm": 2.229381561279297, "learning_rate": 1.4963880288957688e-05, "loss": 1.267, "step": 870 }, { "epoch": 0.30269779408982556, "grad_norm": 1.6807219982147217, "learning_rate": 1.5135878912968696e-05, "loss": 1.2696, "step": 880 }, { "epoch": 0.30613754174993724, "grad_norm": 1.2693103551864624, "learning_rate": 1.5307877536979705e-05, "loss": 1.2572, "step": 890 }, { "epoch": 0.30957728941004886, "grad_norm": 1.9241502285003662, "learning_rate": 1.5479876160990712e-05, "loss": 1.2568, "step": 900 }, { "epoch": 0.31301703707016054, "grad_norm": 1.660744547843933, "learning_rate": 1.5651874785001723e-05, "loss": 1.2521, "step": 910 }, { "epoch": 0.31645678473027217, "grad_norm": 1.706275463104248, "learning_rate": 1.582387340901273e-05, "loss": 1.2357, "step": 920 }, { "epoch": 0.31989653239038385, "grad_norm": 1.8540180921554565, "learning_rate": 1.5995872033023738e-05, "loss": 1.2167, "step": 930 }, { "epoch": 0.3233362800504955, "grad_norm": 1.8119511604309082, "learning_rate": 1.6167870657034746e-05, "loss": 1.2544, "step": 940 }, { "epoch": 0.32677602771060715, "grad_norm": 1.9702472686767578, "learning_rate": 1.6339869281045753e-05, "loss": 1.2094, "step": 950 }, { "epoch": 0.3302157753707188, "grad_norm": 2.0417630672454834, "learning_rate": 1.651186790505676e-05, "loss": 1.1975, "step": 960 }, { "epoch": 0.33365552303083046, "grad_norm": 1.4115933179855347, "learning_rate": 1.6683866529067768e-05, "loss": 1.1897, "step": 970 }, { "epoch": 0.33709527069094214, "grad_norm": 2.0552473068237305, "learning_rate": 1.6855865153078776e-05, "loss": 1.2015, "step": 980 }, { "epoch": 0.34053501835105376, "grad_norm": 2.383172035217285, "learning_rate": 1.7027863777089783e-05, "loss": 1.1849, "step": 990 }, { "epoch": 0.34397476601116544, "grad_norm": 2.141005516052246, "learning_rate": 1.719986240110079e-05, "loss": 1.1892, "step": 1000 }, { "epoch": 0.34741451367127707, "grad_norm": 2.242393732070923, "learning_rate": 1.7371861025111798e-05, "loss": 1.1812, "step": 1010 }, { "epoch": 0.35085426133138875, "grad_norm": 2.2587263584136963, "learning_rate": 1.7543859649122806e-05, "loss": 1.1722, "step": 1020 }, { "epoch": 0.3542940089915004, "grad_norm": 2.2541704177856445, "learning_rate": 1.7715858273133813e-05, "loss": 1.1677, "step": 1030 }, { "epoch": 0.35773375665161206, "grad_norm": 1.5090018510818481, "learning_rate": 1.788785689714482e-05, "loss": 1.1496, "step": 1040 }, { "epoch": 0.3611735043117237, "grad_norm": 2.1420912742614746, "learning_rate": 1.805985552115583e-05, "loss": 1.152, "step": 1050 }, { "epoch": 0.36461325197183536, "grad_norm": 1.6955997943878174, "learning_rate": 1.823185414516684e-05, "loss": 1.1514, "step": 1060 }, { "epoch": 0.368052999631947, "grad_norm": 2.603424549102783, "learning_rate": 1.840385276917785e-05, "loss": 1.1543, "step": 1070 }, { "epoch": 0.37149274729205867, "grad_norm": 1.6074223518371582, "learning_rate": 1.8575851393188857e-05, "loss": 1.1632, "step": 1080 }, { "epoch": 0.3749324949521703, "grad_norm": 2.813610553741455, "learning_rate": 1.8747850017199865e-05, "loss": 1.1383, "step": 1090 }, { "epoch": 0.378372242612282, "grad_norm": 2.169351816177368, "learning_rate": 1.8919848641210872e-05, "loss": 1.1316, "step": 1100 }, { "epoch": 0.3818119902723936, "grad_norm": 2.653705596923828, "learning_rate": 1.909184726522188e-05, "loss": 1.1433, "step": 1110 }, { "epoch": 0.3852517379325053, "grad_norm": 3.17232084274292, "learning_rate": 1.9263845889232888e-05, "loss": 1.1385, "step": 1120 }, { "epoch": 0.3886914855926169, "grad_norm": 2.2379255294799805, "learning_rate": 1.9435844513243895e-05, "loss": 1.1205, "step": 1130 }, { "epoch": 0.3921312332527286, "grad_norm": 2.5751757621765137, "learning_rate": 1.9607843137254903e-05, "loss": 1.1125, "step": 1140 }, { "epoch": 0.3955709809128402, "grad_norm": 1.6813164949417114, "learning_rate": 1.977984176126591e-05, "loss": 1.1147, "step": 1150 }, { "epoch": 0.3990107285729519, "grad_norm": 2.8093101978302, "learning_rate": 1.9951840385276918e-05, "loss": 1.1057, "step": 1160 }, { "epoch": 0.40245047623306357, "grad_norm": 2.059581995010376, "learning_rate": 2.0123839009287925e-05, "loss": 1.1037, "step": 1170 }, { "epoch": 0.4058902238931752, "grad_norm": 1.9385766983032227, "learning_rate": 2.0295837633298933e-05, "loss": 1.0987, "step": 1180 }, { "epoch": 0.4093299715532869, "grad_norm": 1.7843406200408936, "learning_rate": 2.046783625730994e-05, "loss": 1.0917, "step": 1190 }, { "epoch": 0.4127697192133985, "grad_norm": 1.7654945850372314, "learning_rate": 2.0639834881320948e-05, "loss": 1.1, "step": 1200 }, { "epoch": 0.4162094668735102, "grad_norm": 2.3094961643218994, "learning_rate": 2.081183350533196e-05, "loss": 1.0891, "step": 1210 }, { "epoch": 0.4196492145336218, "grad_norm": 2.654541492462158, "learning_rate": 2.0983832129342966e-05, "loss": 1.0909, "step": 1220 }, { "epoch": 0.4230889621937335, "grad_norm": 2.7142221927642822, "learning_rate": 2.1155830753353977e-05, "loss": 1.1068, "step": 1230 }, { "epoch": 0.4265287098538451, "grad_norm": 1.7891792058944702, "learning_rate": 2.1327829377364984e-05, "loss": 1.076, "step": 1240 }, { "epoch": 0.4299684575139568, "grad_norm": 1.8982404470443726, "learning_rate": 2.1499828001375992e-05, "loss": 1.0703, "step": 1250 }, { "epoch": 0.4334082051740684, "grad_norm": 2.5560953617095947, "learning_rate": 2.1671826625387e-05, "loss": 1.0696, "step": 1260 }, { "epoch": 0.4368479528341801, "grad_norm": 2.6159865856170654, "learning_rate": 2.1843825249398007e-05, "loss": 1.069, "step": 1270 }, { "epoch": 0.4402877004942917, "grad_norm": 2.1036102771759033, "learning_rate": 2.2015823873409014e-05, "loss": 1.054, "step": 1280 }, { "epoch": 0.4437274481544034, "grad_norm": 1.6638193130493164, "learning_rate": 2.2187822497420022e-05, "loss": 1.0554, "step": 1290 }, { "epoch": 0.44716719581451503, "grad_norm": 1.7467694282531738, "learning_rate": 2.235982112143103e-05, "loss": 1.0356, "step": 1300 }, { "epoch": 0.4506069434746267, "grad_norm": 1.9570752382278442, "learning_rate": 2.2531819745442037e-05, "loss": 1.0501, "step": 1310 }, { "epoch": 0.45404669113473833, "grad_norm": 2.062814950942993, "learning_rate": 2.2703818369453044e-05, "loss": 1.041, "step": 1320 }, { "epoch": 0.45748643879485, "grad_norm": 3.205482006072998, "learning_rate": 2.2875816993464052e-05, "loss": 1.0428, "step": 1330 }, { "epoch": 0.46092618645496164, "grad_norm": 3.798652172088623, "learning_rate": 2.304781561747506e-05, "loss": 1.055, "step": 1340 }, { "epoch": 0.4643659341150733, "grad_norm": 2.6361844539642334, "learning_rate": 2.3219814241486067e-05, "loss": 1.051, "step": 1350 }, { "epoch": 0.46780568177518494, "grad_norm": 3.9871771335601807, "learning_rate": 2.3391812865497074e-05, "loss": 1.0576, "step": 1360 }, { "epoch": 0.4712454294352966, "grad_norm": 2.9001245498657227, "learning_rate": 2.3563811489508085e-05, "loss": 1.0521, "step": 1370 }, { "epoch": 0.4746851770954083, "grad_norm": 3.8362972736358643, "learning_rate": 2.3735810113519093e-05, "loss": 1.0392, "step": 1380 }, { "epoch": 0.47812492475551993, "grad_norm": 3.222538709640503, "learning_rate": 2.39078087375301e-05, "loss": 1.0449, "step": 1390 }, { "epoch": 0.4815646724156316, "grad_norm": 4.081332683563232, "learning_rate": 2.407980736154111e-05, "loss": 1.0402, "step": 1400 }, { "epoch": 0.48500442007574324, "grad_norm": 3.8474042415618896, "learning_rate": 2.425180598555212e-05, "loss": 1.0199, "step": 1410 }, { "epoch": 0.4884441677358549, "grad_norm": 3.6466848850250244, "learning_rate": 2.4423804609563126e-05, "loss": 1.0352, "step": 1420 }, { "epoch": 0.49188391539596654, "grad_norm": 3.44462251663208, "learning_rate": 2.4595803233574134e-05, "loss": 1.0476, "step": 1430 }, { "epoch": 0.4953236630560782, "grad_norm": 2.647538661956787, "learning_rate": 2.476780185758514e-05, "loss": 1.0094, "step": 1440 }, { "epoch": 0.49876341071618985, "grad_norm": 2.3128774166107178, "learning_rate": 2.493980048159615e-05, "loss": 1.0291, "step": 1450 }, { "epoch": 0.5022031583763015, "grad_norm": 3.0225725173950195, "learning_rate": 2.511179910560716e-05, "loss": 1.0321, "step": 1460 }, { "epoch": 0.5056429060364132, "grad_norm": 2.3091073036193848, "learning_rate": 2.5283797729618164e-05, "loss": 1.0078, "step": 1470 }, { "epoch": 0.5090826536965248, "grad_norm": 2.2101924419403076, "learning_rate": 2.5455796353629174e-05, "loss": 1.0316, "step": 1480 }, { "epoch": 0.5125224013566365, "grad_norm": 2.288010597229004, "learning_rate": 2.562779497764018e-05, "loss": 1.0158, "step": 1490 }, { "epoch": 0.5159621490167481, "grad_norm": 2.550391674041748, "learning_rate": 2.579979360165119e-05, "loss": 1.0131, "step": 1500 } ], "logging_steps": 10, "max_steps": 14535, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4.63721472e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }