{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.991416309012875, "eval_steps": 500, "global_step": 5820, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0017167381974248926, "grad_norm": 6.089517593383789, "learning_rate": 3.436426116838488e-07, "loss": 2.0708, "step": 1 }, { "epoch": 0.008583690987124463, "grad_norm": 6.525425434112549, "learning_rate": 1.7182130584192443e-06, "loss": 2.0164, "step": 5 }, { "epoch": 0.017167381974248927, "grad_norm": 5.859015464782715, "learning_rate": 3.4364261168384886e-06, "loss": 2.0159, "step": 10 }, { "epoch": 0.02575107296137339, "grad_norm": 4.282891750335693, "learning_rate": 5.154639175257732e-06, "loss": 1.9357, "step": 15 }, { "epoch": 0.034334763948497854, "grad_norm": 2.575316905975342, "learning_rate": 6.872852233676977e-06, "loss": 1.8673, "step": 20 }, { "epoch": 0.04291845493562232, "grad_norm": 1.4356714487075806, "learning_rate": 8.591065292096221e-06, "loss": 1.7557, "step": 25 }, { "epoch": 0.05150214592274678, "grad_norm": 0.756867527961731, "learning_rate": 1.0309278350515464e-05, "loss": 1.6729, "step": 30 }, { "epoch": 0.060085836909871244, "grad_norm": 0.5709408521652222, "learning_rate": 1.2027491408934708e-05, "loss": 1.6029, "step": 35 }, { "epoch": 0.06866952789699571, "grad_norm": 0.6336522698402405, "learning_rate": 1.3745704467353954e-05, "loss": 1.5371, "step": 40 }, { "epoch": 0.07725321888412018, "grad_norm": 0.48585963249206543, "learning_rate": 1.5463917525773197e-05, "loss": 1.4468, "step": 45 }, { "epoch": 0.08583690987124463, "grad_norm": 0.45667070150375366, "learning_rate": 1.7182130584192442e-05, "loss": 1.3687, "step": 50 }, { "epoch": 0.0944206008583691, "grad_norm": 0.4194663465023041, "learning_rate": 1.8900343642611683e-05, "loss": 1.3212, "step": 55 }, { "epoch": 0.10300429184549356, "grad_norm": 0.3849167227745056, "learning_rate": 2.0618556701030927e-05, "loss": 1.2499, "step": 60 }, { "epoch": 0.11158798283261803, "grad_norm": 0.32962682843208313, "learning_rate": 2.2336769759450175e-05, "loss": 1.2345, "step": 65 }, { "epoch": 0.12017167381974249, "grad_norm": 0.32320886850357056, "learning_rate": 2.4054982817869417e-05, "loss": 1.1963, "step": 70 }, { "epoch": 0.12875536480686695, "grad_norm": 0.3570818603038788, "learning_rate": 2.5773195876288658e-05, "loss": 1.1658, "step": 75 }, { "epoch": 0.13733905579399142, "grad_norm": 0.3045337498188019, "learning_rate": 2.749140893470791e-05, "loss": 1.1628, "step": 80 }, { "epoch": 0.1459227467811159, "grad_norm": 0.2675187289714813, "learning_rate": 2.920962199312715e-05, "loss": 1.153, "step": 85 }, { "epoch": 0.15450643776824036, "grad_norm": 0.2941209077835083, "learning_rate": 3.0927835051546395e-05, "loss": 1.1233, "step": 90 }, { "epoch": 0.1630901287553648, "grad_norm": 0.30070310831069946, "learning_rate": 3.2646048109965636e-05, "loss": 1.1199, "step": 95 }, { "epoch": 0.17167381974248927, "grad_norm": 0.2994473874568939, "learning_rate": 3.4364261168384884e-05, "loss": 1.1063, "step": 100 }, { "epoch": 0.18025751072961374, "grad_norm": 0.301921010017395, "learning_rate": 3.6082474226804125e-05, "loss": 1.0991, "step": 105 }, { "epoch": 0.1888412017167382, "grad_norm": 0.31418925523757935, "learning_rate": 3.7800687285223366e-05, "loss": 1.0939, "step": 110 }, { "epoch": 0.19742489270386265, "grad_norm": 0.31536900997161865, "learning_rate": 3.9518900343642614e-05, "loss": 1.0961, "step": 115 }, { "epoch": 0.20600858369098712, "grad_norm": 0.2873052954673767, "learning_rate": 4.1237113402061855e-05, "loss": 1.0807, "step": 120 }, { "epoch": 0.2145922746781116, "grad_norm": 0.34555503726005554, "learning_rate": 4.2955326460481096e-05, "loss": 1.0645, "step": 125 }, { "epoch": 0.22317596566523606, "grad_norm": 0.313761442899704, "learning_rate": 4.467353951890035e-05, "loss": 1.0513, "step": 130 }, { "epoch": 0.2317596566523605, "grad_norm": 0.32121869921684265, "learning_rate": 4.639175257731959e-05, "loss": 1.0735, "step": 135 }, { "epoch": 0.24034334763948498, "grad_norm": 0.3088555335998535, "learning_rate": 4.810996563573883e-05, "loss": 1.0552, "step": 140 }, { "epoch": 0.24892703862660945, "grad_norm": 0.35492023825645447, "learning_rate": 4.982817869415808e-05, "loss": 1.0546, "step": 145 }, { "epoch": 0.2575107296137339, "grad_norm": 0.30434370040893555, "learning_rate": 5.1546391752577315e-05, "loss": 1.0422, "step": 150 }, { "epoch": 0.26609442060085836, "grad_norm": 0.30399689078330994, "learning_rate": 5.326460481099656e-05, "loss": 1.0465, "step": 155 }, { "epoch": 0.27467811158798283, "grad_norm": 0.30653682351112366, "learning_rate": 5.498281786941582e-05, "loss": 1.0404, "step": 160 }, { "epoch": 0.2832618025751073, "grad_norm": 0.3112322986125946, "learning_rate": 5.670103092783505e-05, "loss": 1.0369, "step": 165 }, { "epoch": 0.2918454935622318, "grad_norm": 0.3165334165096283, "learning_rate": 5.84192439862543e-05, "loss": 1.0433, "step": 170 }, { "epoch": 0.30042918454935624, "grad_norm": 0.34382325410842896, "learning_rate": 6.013745704467354e-05, "loss": 1.042, "step": 175 }, { "epoch": 0.3090128755364807, "grad_norm": 0.3302488625049591, "learning_rate": 6.185567010309279e-05, "loss": 1.0166, "step": 180 }, { "epoch": 0.31759656652360513, "grad_norm": 0.3078051209449768, "learning_rate": 6.357388316151203e-05, "loss": 1.0138, "step": 185 }, { "epoch": 0.3261802575107296, "grad_norm": 0.30805814266204834, "learning_rate": 6.529209621993127e-05, "loss": 0.9951, "step": 190 }, { "epoch": 0.33476394849785407, "grad_norm": 0.4036194980144501, "learning_rate": 6.701030927835051e-05, "loss": 1.013, "step": 195 }, { "epoch": 0.34334763948497854, "grad_norm": 0.4557146430015564, "learning_rate": 6.872852233676977e-05, "loss": 0.9959, "step": 200 }, { "epoch": 0.351931330472103, "grad_norm": 0.36137068271636963, "learning_rate": 7.044673539518901e-05, "loss": 1.0068, "step": 205 }, { "epoch": 0.3605150214592275, "grad_norm": 0.3014100193977356, "learning_rate": 7.216494845360825e-05, "loss": 0.9923, "step": 210 }, { "epoch": 0.36909871244635195, "grad_norm": 0.290464848279953, "learning_rate": 7.38831615120275e-05, "loss": 0.9969, "step": 215 }, { "epoch": 0.3776824034334764, "grad_norm": 0.31709638237953186, "learning_rate": 7.560137457044673e-05, "loss": 1.0221, "step": 220 }, { "epoch": 0.38626609442060084, "grad_norm": 0.2746679484844208, "learning_rate": 7.731958762886599e-05, "loss": 0.9829, "step": 225 }, { "epoch": 0.3948497854077253, "grad_norm": 0.28260111808776855, "learning_rate": 7.903780068728523e-05, "loss": 0.9868, "step": 230 }, { "epoch": 0.4034334763948498, "grad_norm": 0.3063802421092987, "learning_rate": 8.075601374570447e-05, "loss": 0.9789, "step": 235 }, { "epoch": 0.41201716738197425, "grad_norm": 0.29451537132263184, "learning_rate": 8.247422680412371e-05, "loss": 1.0036, "step": 240 }, { "epoch": 0.4206008583690987, "grad_norm": 0.29433488845825195, "learning_rate": 8.419243986254296e-05, "loss": 0.9936, "step": 245 }, { "epoch": 0.4291845493562232, "grad_norm": 0.2725401520729065, "learning_rate": 8.591065292096219e-05, "loss": 0.9786, "step": 250 }, { "epoch": 0.43776824034334766, "grad_norm": 0.2759503722190857, "learning_rate": 8.762886597938145e-05, "loss": 0.9656, "step": 255 }, { "epoch": 0.44635193133047213, "grad_norm": 0.2762455642223358, "learning_rate": 8.93470790378007e-05, "loss": 0.9837, "step": 260 }, { "epoch": 0.45493562231759654, "grad_norm": 0.2803398668766022, "learning_rate": 9.106529209621993e-05, "loss": 0.9757, "step": 265 }, { "epoch": 0.463519313304721, "grad_norm": 0.29107633233070374, "learning_rate": 9.278350515463918e-05, "loss": 0.9692, "step": 270 }, { "epoch": 0.4721030042918455, "grad_norm": 0.32915282249450684, "learning_rate": 9.450171821305843e-05, "loss": 0.9911, "step": 275 }, { "epoch": 0.48068669527896996, "grad_norm": 0.2902511954307556, "learning_rate": 9.621993127147767e-05, "loss": 0.9635, "step": 280 }, { "epoch": 0.4892703862660944, "grad_norm": 0.28177133202552795, "learning_rate": 9.793814432989691e-05, "loss": 0.9704, "step": 285 }, { "epoch": 0.4978540772532189, "grad_norm": 0.26041337847709656, "learning_rate": 9.965635738831616e-05, "loss": 0.9668, "step": 290 }, { "epoch": 0.5064377682403434, "grad_norm": 0.27237018942832947, "learning_rate": 0.00010137457044673539, "loss": 0.9751, "step": 295 }, { "epoch": 0.5150214592274678, "grad_norm": 0.2928680181503296, "learning_rate": 0.00010309278350515463, "loss": 0.9828, "step": 300 }, { "epoch": 0.5236051502145923, "grad_norm": 0.3540053367614746, "learning_rate": 0.0001048109965635739, "loss": 0.9725, "step": 305 }, { "epoch": 0.5321888412017167, "grad_norm": 0.34892937541007996, "learning_rate": 0.00010652920962199313, "loss": 0.9688, "step": 310 }, { "epoch": 0.5407725321888412, "grad_norm": 0.2727091908454895, "learning_rate": 0.00010824742268041237, "loss": 0.9797, "step": 315 }, { "epoch": 0.5493562231759657, "grad_norm": 0.2613857090473175, "learning_rate": 0.00010996563573883164, "loss": 0.9751, "step": 320 }, { "epoch": 0.5579399141630901, "grad_norm": 0.24695193767547607, "learning_rate": 0.00011168384879725086, "loss": 0.9661, "step": 325 }, { "epoch": 0.5665236051502146, "grad_norm": 0.25507402420043945, "learning_rate": 0.0001134020618556701, "loss": 0.9703, "step": 330 }, { "epoch": 0.575107296137339, "grad_norm": 0.2496771663427353, "learning_rate": 0.00011512027491408935, "loss": 0.9766, "step": 335 }, { "epoch": 0.5836909871244635, "grad_norm": 0.2503701448440552, "learning_rate": 0.0001168384879725086, "loss": 0.9506, "step": 340 }, { "epoch": 0.592274678111588, "grad_norm": 0.26991888880729675, "learning_rate": 0.00011855670103092784, "loss": 0.9587, "step": 345 }, { "epoch": 0.6008583690987125, "grad_norm": 0.24292829632759094, "learning_rate": 0.00012027491408934708, "loss": 0.951, "step": 350 }, { "epoch": 0.6094420600858369, "grad_norm": 0.26761141419410706, "learning_rate": 0.00012199312714776634, "loss": 0.9525, "step": 355 }, { "epoch": 0.6180257510729614, "grad_norm": 0.24770408868789673, "learning_rate": 0.00012371134020618558, "loss": 0.9606, "step": 360 }, { "epoch": 0.6266094420600858, "grad_norm": 0.24938061833381653, "learning_rate": 0.00012542955326460482, "loss": 0.9577, "step": 365 }, { "epoch": 0.6351931330472103, "grad_norm": 0.22758124768733978, "learning_rate": 0.00012714776632302406, "loss": 0.9693, "step": 370 }, { "epoch": 0.6437768240343348, "grad_norm": 0.24254348874092102, "learning_rate": 0.0001288659793814433, "loss": 0.9492, "step": 375 }, { "epoch": 0.6523605150214592, "grad_norm": 0.37186160683631897, "learning_rate": 0.00013058419243986254, "loss": 0.9722, "step": 380 }, { "epoch": 0.6609442060085837, "grad_norm": 0.23567767441272736, "learning_rate": 0.00013230240549828178, "loss": 0.9567, "step": 385 }, { "epoch": 0.6695278969957081, "grad_norm": 0.2288963794708252, "learning_rate": 0.00013402061855670103, "loss": 0.9564, "step": 390 }, { "epoch": 0.6781115879828327, "grad_norm": 0.24152550101280212, "learning_rate": 0.0001357388316151203, "loss": 0.9532, "step": 395 }, { "epoch": 0.6866952789699571, "grad_norm": 0.261593759059906, "learning_rate": 0.00013745704467353953, "loss": 0.9471, "step": 400 }, { "epoch": 0.6952789699570815, "grad_norm": 0.27105116844177246, "learning_rate": 0.00013917525773195878, "loss": 0.9639, "step": 405 }, { "epoch": 0.703862660944206, "grad_norm": 0.361182302236557, "learning_rate": 0.00014089347079037802, "loss": 0.9516, "step": 410 }, { "epoch": 0.7124463519313304, "grad_norm": 0.2614869475364685, "learning_rate": 0.00014261168384879726, "loss": 0.9615, "step": 415 }, { "epoch": 0.721030042918455, "grad_norm": 0.2598520517349243, "learning_rate": 0.0001443298969072165, "loss": 0.9587, "step": 420 }, { "epoch": 0.7296137339055794, "grad_norm": 0.26025694608688354, "learning_rate": 0.00014604810996563574, "loss": 0.9483, "step": 425 }, { "epoch": 0.7381974248927039, "grad_norm": 0.27435532212257385, "learning_rate": 0.000147766323024055, "loss": 0.9555, "step": 430 }, { "epoch": 0.7467811158798283, "grad_norm": 0.2179042547941208, "learning_rate": 0.00014948453608247422, "loss": 0.9719, "step": 435 }, { "epoch": 0.7553648068669528, "grad_norm": 0.25120726227760315, "learning_rate": 0.00015120274914089346, "loss": 0.9601, "step": 440 }, { "epoch": 0.7639484978540773, "grad_norm": 0.21970221400260925, "learning_rate": 0.0001529209621993127, "loss": 0.9545, "step": 445 }, { "epoch": 0.7725321888412017, "grad_norm": 0.2047254890203476, "learning_rate": 0.00015463917525773197, "loss": 0.9448, "step": 450 }, { "epoch": 0.7811158798283262, "grad_norm": 0.21874375641345978, "learning_rate": 0.00015635738831615121, "loss": 0.9426, "step": 455 }, { "epoch": 0.7896995708154506, "grad_norm": 0.21705736219882965, "learning_rate": 0.00015807560137457046, "loss": 0.941, "step": 460 }, { "epoch": 0.7982832618025751, "grad_norm": 0.20229893922805786, "learning_rate": 0.0001597938144329897, "loss": 0.9495, "step": 465 }, { "epoch": 0.8068669527896996, "grad_norm": 0.21079690754413605, "learning_rate": 0.00016151202749140894, "loss": 0.9488, "step": 470 }, { "epoch": 0.8154506437768241, "grad_norm": 0.21350346505641937, "learning_rate": 0.00016323024054982818, "loss": 0.941, "step": 475 }, { "epoch": 0.8240343347639485, "grad_norm": 0.22696025669574738, "learning_rate": 0.00016494845360824742, "loss": 0.9468, "step": 480 }, { "epoch": 0.8326180257510729, "grad_norm": 0.2032315880060196, "learning_rate": 0.0001666666666666667, "loss": 0.9649, "step": 485 }, { "epoch": 0.8412017167381974, "grad_norm": 0.24577978253364563, "learning_rate": 0.00016838487972508593, "loss": 0.9457, "step": 490 }, { "epoch": 0.8497854077253219, "grad_norm": 0.2154797613620758, "learning_rate": 0.00017010309278350517, "loss": 0.9322, "step": 495 }, { "epoch": 0.8583690987124464, "grad_norm": 0.20580855011940002, "learning_rate": 0.00017182130584192438, "loss": 0.9417, "step": 500 }, { "epoch": 0.8669527896995708, "grad_norm": 0.2095131129026413, "learning_rate": 0.00017353951890034365, "loss": 0.9447, "step": 505 }, { "epoch": 0.8755364806866953, "grad_norm": 0.20167525112628937, "learning_rate": 0.0001752577319587629, "loss": 0.9504, "step": 510 }, { "epoch": 0.8841201716738197, "grad_norm": 0.21665619313716888, "learning_rate": 0.00017697594501718214, "loss": 0.9386, "step": 515 }, { "epoch": 0.8927038626609443, "grad_norm": 0.2125951647758484, "learning_rate": 0.0001786941580756014, "loss": 0.9373, "step": 520 }, { "epoch": 0.9012875536480687, "grad_norm": 0.19751432538032532, "learning_rate": 0.00018041237113402062, "loss": 0.935, "step": 525 }, { "epoch": 0.9098712446351931, "grad_norm": 0.20792262256145477, "learning_rate": 0.00018213058419243986, "loss": 0.9624, "step": 530 }, { "epoch": 0.9184549356223176, "grad_norm": 0.2099096179008484, "learning_rate": 0.0001838487972508591, "loss": 0.9392, "step": 535 }, { "epoch": 0.927038626609442, "grad_norm": 0.21597731113433838, "learning_rate": 0.00018556701030927837, "loss": 0.942, "step": 540 }, { "epoch": 0.9356223175965666, "grad_norm": 0.2026844620704651, "learning_rate": 0.0001872852233676976, "loss": 0.9579, "step": 545 }, { "epoch": 0.944206008583691, "grad_norm": 0.20321713387966156, "learning_rate": 0.00018900343642611685, "loss": 0.9519, "step": 550 }, { "epoch": 0.9527896995708155, "grad_norm": 0.20575563609600067, "learning_rate": 0.0001907216494845361, "loss": 0.9363, "step": 555 }, { "epoch": 0.9613733905579399, "grad_norm": 0.21118000149726868, "learning_rate": 0.00019243986254295533, "loss": 0.9493, "step": 560 }, { "epoch": 0.9699570815450643, "grad_norm": 0.2216077297925949, "learning_rate": 0.00019415807560137457, "loss": 0.9499, "step": 565 }, { "epoch": 0.9785407725321889, "grad_norm": 0.20766399800777435, "learning_rate": 0.00019587628865979381, "loss": 0.9359, "step": 570 }, { "epoch": 0.9871244635193133, "grad_norm": 0.30319133400917053, "learning_rate": 0.00019759450171821308, "loss": 0.9483, "step": 575 }, { "epoch": 0.9957081545064378, "grad_norm": 0.21939022839069366, "learning_rate": 0.00019931271477663232, "loss": 0.9422, "step": 580 }, { "epoch": 0.9991416309012876, "eval_loss": 1.976241946220398, "eval_runtime": 0.3946, "eval_samples_per_second": 15.206, "eval_steps_per_second": 2.534, "step": 582 }, { "epoch": 1.0042918454935623, "grad_norm": 0.21883882582187653, "learning_rate": 0.00019999983812448848, "loss": 0.915, "step": 585 }, { "epoch": 1.0128755364806867, "grad_norm": 0.21552623808383942, "learning_rate": 0.0001999988488871492, "loss": 0.9101, "step": 590 }, { "epoch": 1.0214592274678111, "grad_norm": 2.067782402038574, "learning_rate": 0.00019999696035219593, "loss": 0.9213, "step": 595 }, { "epoch": 1.0300429184549356, "grad_norm": 0.22093655169010162, "learning_rate": 0.00019999417253661235, "loss": 0.9244, "step": 600 }, { "epoch": 1.0386266094420602, "grad_norm": 0.2211354523897171, "learning_rate": 0.00019999048546546954, "loss": 0.8949, "step": 605 }, { "epoch": 1.0472103004291846, "grad_norm": 0.21562980115413666, "learning_rate": 0.00019998589917192568, "loss": 0.9248, "step": 610 }, { "epoch": 1.055793991416309, "grad_norm": 0.22060342133045197, "learning_rate": 0.00019998041369722556, "loss": 0.907, "step": 615 }, { "epoch": 1.0643776824034334, "grad_norm": 0.20447732508182526, "learning_rate": 0.00019997402909070059, "loss": 0.9155, "step": 620 }, { "epoch": 1.0729613733905579, "grad_norm": 0.21326489746570587, "learning_rate": 0.000199966745409768, "loss": 0.9148, "step": 625 }, { "epoch": 1.0815450643776825, "grad_norm": 0.21152488887310028, "learning_rate": 0.0001999585627199305, "loss": 0.8954, "step": 630 }, { "epoch": 1.090128755364807, "grad_norm": 0.21141602098941803, "learning_rate": 0.0001999494810947757, "loss": 0.9091, "step": 635 }, { "epoch": 1.0987124463519313, "grad_norm": 0.21263065934181213, "learning_rate": 0.00019993950061597535, "loss": 0.9065, "step": 640 }, { "epoch": 1.1072961373390557, "grad_norm": 0.21794655919075012, "learning_rate": 0.00019992862137328474, "loss": 0.9029, "step": 645 }, { "epoch": 1.1158798283261802, "grad_norm": 0.21535712480545044, "learning_rate": 0.00019991684346454172, "loss": 0.9139, "step": 650 }, { "epoch": 1.1244635193133048, "grad_norm": 0.1971653550863266, "learning_rate": 0.00019990416699566598, "loss": 0.8918, "step": 655 }, { "epoch": 1.1330472103004292, "grad_norm": 0.2062826007604599, "learning_rate": 0.000199890592080658, "loss": 0.9188, "step": 660 }, { "epoch": 1.1416309012875536, "grad_norm": 0.2263791412115097, "learning_rate": 0.0001998761188415981, "loss": 0.904, "step": 665 }, { "epoch": 1.150214592274678, "grad_norm": 0.19670893251895905, "learning_rate": 0.00019986074740864526, "loss": 0.9165, "step": 670 }, { "epoch": 1.1587982832618025, "grad_norm": 0.21204271912574768, "learning_rate": 0.000199844477920036, "loss": 0.8874, "step": 675 }, { "epoch": 1.167381974248927, "grad_norm": 0.19298429787158966, "learning_rate": 0.00019982731052208309, "loss": 0.9102, "step": 680 }, { "epoch": 1.1759656652360515, "grad_norm": 0.21324272453784943, "learning_rate": 0.00019980924536917437, "loss": 0.9119, "step": 685 }, { "epoch": 1.184549356223176, "grad_norm": 0.192140594124794, "learning_rate": 0.00019979028262377118, "loss": 0.8957, "step": 690 }, { "epoch": 1.1931330472103003, "grad_norm": 0.20233942568302155, "learning_rate": 0.00019977042245640698, "loss": 0.8969, "step": 695 }, { "epoch": 1.201716738197425, "grad_norm": 0.21660216152668, "learning_rate": 0.00019974966504568583, "loss": 0.9064, "step": 700 }, { "epoch": 1.2103004291845494, "grad_norm": 0.2243824154138565, "learning_rate": 0.0001997280105782808, "loss": 0.9166, "step": 705 }, { "epoch": 1.2188841201716738, "grad_norm": 0.20581458508968353, "learning_rate": 0.00019970545924893226, "loss": 0.9064, "step": 710 }, { "epoch": 1.2274678111587982, "grad_norm": 0.19686444103717804, "learning_rate": 0.00019968201126044604, "loss": 0.9126, "step": 715 }, { "epoch": 1.2360515021459229, "grad_norm": 0.21029411256313324, "learning_rate": 0.00019965766682369186, "loss": 0.892, "step": 720 }, { "epoch": 1.2446351931330473, "grad_norm": 0.21291205286979675, "learning_rate": 0.0001996324261576011, "loss": 0.8936, "step": 725 }, { "epoch": 1.2532188841201717, "grad_norm": 0.2174995094537735, "learning_rate": 0.00019960628948916518, "loss": 0.9157, "step": 730 }, { "epoch": 1.261802575107296, "grad_norm": 0.2011156976222992, "learning_rate": 0.0001995792570534331, "loss": 0.9161, "step": 735 }, { "epoch": 1.2703862660944205, "grad_norm": 0.20497067272663116, "learning_rate": 0.00019955132909350984, "loss": 0.8999, "step": 740 }, { "epoch": 1.2789699570815452, "grad_norm": 0.19909746944904327, "learning_rate": 0.0001995225058605537, "loss": 0.9142, "step": 745 }, { "epoch": 1.2875536480686696, "grad_norm": 0.22116069495677948, "learning_rate": 0.0001994927876137743, "loss": 0.8992, "step": 750 }, { "epoch": 1.296137339055794, "grad_norm": 0.22861087322235107, "learning_rate": 0.00019946217462043025, "loss": 0.898, "step": 755 }, { "epoch": 1.3047210300429184, "grad_norm": 0.20132282376289368, "learning_rate": 0.0001994306671558266, "loss": 0.9147, "step": 760 }, { "epoch": 1.3133047210300428, "grad_norm": 0.21803739666938782, "learning_rate": 0.00019939826550331252, "loss": 0.9, "step": 765 }, { "epoch": 1.3218884120171674, "grad_norm": 0.19455976784229279, "learning_rate": 0.0001993649699542786, "loss": 0.9126, "step": 770 }, { "epoch": 1.3304721030042919, "grad_norm": 0.18571655452251434, "learning_rate": 0.0001993307808081544, "loss": 0.9006, "step": 775 }, { "epoch": 1.3390557939914163, "grad_norm": 0.20103998482227325, "learning_rate": 0.00019929569837240564, "loss": 0.8881, "step": 780 }, { "epoch": 1.3476394849785407, "grad_norm": 0.19315999746322632, "learning_rate": 0.00019925972296253145, "loss": 0.901, "step": 785 }, { "epoch": 1.356223175965665, "grad_norm": 0.2066372036933899, "learning_rate": 0.00019922285490206156, "loss": 0.888, "step": 790 }, { "epoch": 1.3648068669527897, "grad_norm": 0.20879539847373962, "learning_rate": 0.00019918509452255338, "loss": 0.901, "step": 795 }, { "epoch": 1.3733905579399142, "grad_norm": 0.20333191752433777, "learning_rate": 0.000199146442163589, "loss": 0.9099, "step": 800 }, { "epoch": 1.3819742489270386, "grad_norm": 0.1949775516986847, "learning_rate": 0.00019910689817277216, "loss": 0.904, "step": 805 }, { "epoch": 1.3905579399141632, "grad_norm": 0.20540495216846466, "learning_rate": 0.00019906646290572514, "loss": 0.8965, "step": 810 }, { "epoch": 1.3991416309012876, "grad_norm": 0.19921506941318512, "learning_rate": 0.00019902513672608553, "loss": 0.8991, "step": 815 }, { "epoch": 1.407725321888412, "grad_norm": 0.21238817274570465, "learning_rate": 0.0001989829200055029, "loss": 0.9026, "step": 820 }, { "epoch": 1.4163090128755365, "grad_norm": 0.2081788033246994, "learning_rate": 0.00019893981312363562, "loss": 0.9052, "step": 825 }, { "epoch": 1.4248927038626609, "grad_norm": 0.20578624308109283, "learning_rate": 0.00019889581646814728, "loss": 0.9038, "step": 830 }, { "epoch": 1.4334763948497855, "grad_norm": 0.2119644731283188, "learning_rate": 0.00019885093043470336, "loss": 0.8936, "step": 835 }, { "epoch": 1.44206008583691, "grad_norm": 0.19631995260715485, "learning_rate": 0.0001988051554269675, "loss": 0.9059, "step": 840 }, { "epoch": 1.4506437768240343, "grad_norm": 0.22262215614318848, "learning_rate": 0.00019875849185659798, "loss": 0.9172, "step": 845 }, { "epoch": 1.4592274678111588, "grad_norm": 0.19081105291843414, "learning_rate": 0.00019871094014324404, "loss": 0.9025, "step": 850 }, { "epoch": 1.4678111587982832, "grad_norm": 0.18824172019958496, "learning_rate": 0.000198662500714542, "loss": 0.9141, "step": 855 }, { "epoch": 1.4763948497854078, "grad_norm": 0.20280902087688446, "learning_rate": 0.0001986131740061115, "loss": 0.8889, "step": 860 }, { "epoch": 1.4849785407725322, "grad_norm": 0.19314704835414886, "learning_rate": 0.00019856296046155157, "loss": 0.8919, "step": 865 }, { "epoch": 1.4935622317596566, "grad_norm": 0.1936980038881302, "learning_rate": 0.00019851186053243666, "loss": 0.9015, "step": 870 }, { "epoch": 1.5021459227467813, "grad_norm": 0.21349290013313293, "learning_rate": 0.00019845987467831242, "loss": 0.9068, "step": 875 }, { "epoch": 1.5107296137339055, "grad_norm": 0.1915241926908493, "learning_rate": 0.00019840700336669183, "loss": 0.9148, "step": 880 }, { "epoch": 1.51931330472103, "grad_norm": 0.1982114166021347, "learning_rate": 0.00019835324707305076, "loss": 0.9043, "step": 885 }, { "epoch": 1.5278969957081545, "grad_norm": 0.18504977226257324, "learning_rate": 0.0001982986062808239, "loss": 0.8926, "step": 890 }, { "epoch": 1.536480686695279, "grad_norm": 0.22229517996311188, "learning_rate": 0.0001982430814814002, "loss": 0.8849, "step": 895 }, { "epoch": 1.5450643776824036, "grad_norm": 0.21088634431362152, "learning_rate": 0.00019818667317411865, "loss": 0.9075, "step": 900 }, { "epoch": 1.5536480686695278, "grad_norm": 0.20124419033527374, "learning_rate": 0.0001981293818662636, "loss": 0.8914, "step": 905 }, { "epoch": 1.5622317596566524, "grad_norm": 0.19154104590415955, "learning_rate": 0.0001980712080730604, "loss": 0.8816, "step": 910 }, { "epoch": 1.5708154506437768, "grad_norm": 0.1901169866323471, "learning_rate": 0.00019801215231767056, "loss": 0.9051, "step": 915 }, { "epoch": 1.5793991416309012, "grad_norm": 0.19122549891471863, "learning_rate": 0.00019795221513118722, "loss": 0.8965, "step": 920 }, { "epoch": 1.5879828326180259, "grad_norm": 0.192024827003479, "learning_rate": 0.00019789139705263026, "loss": 0.8958, "step": 925 }, { "epoch": 1.59656652360515, "grad_norm": 0.19915080070495605, "learning_rate": 0.0001978296986289415, "loss": 0.8924, "step": 930 }, { "epoch": 1.6051502145922747, "grad_norm": 0.19441018998622894, "learning_rate": 0.0001977671204149798, "loss": 0.8997, "step": 935 }, { "epoch": 1.613733905579399, "grad_norm": 0.20060202479362488, "learning_rate": 0.000197703662973516, "loss": 0.8808, "step": 940 }, { "epoch": 1.6223175965665235, "grad_norm": 0.20653241872787476, "learning_rate": 0.00019763932687522794, "loss": 0.9032, "step": 945 }, { "epoch": 1.6309012875536482, "grad_norm": 0.1972658932209015, "learning_rate": 0.00019757411269869527, "loss": 0.9093, "step": 950 }, { "epoch": 1.6394849785407726, "grad_norm": 0.18751849234104156, "learning_rate": 0.0001975080210303943, "loss": 0.8842, "step": 955 }, { "epoch": 1.648068669527897, "grad_norm": 0.1898711621761322, "learning_rate": 0.00019744105246469263, "loss": 0.8914, "step": 960 }, { "epoch": 1.6566523605150214, "grad_norm": 0.1935146301984787, "learning_rate": 0.0001973732076038439, "loss": 0.881, "step": 965 }, { "epoch": 1.6652360515021458, "grad_norm": 0.19017855823040009, "learning_rate": 0.00019730448705798239, "loss": 0.9005, "step": 970 }, { "epoch": 1.6738197424892705, "grad_norm": 0.1929733008146286, "learning_rate": 0.00019723489144511742, "loss": 0.8898, "step": 975 }, { "epoch": 1.6824034334763949, "grad_norm": 0.1935940533876419, "learning_rate": 0.0001971644213911279, "loss": 0.8889, "step": 980 }, { "epoch": 1.6909871244635193, "grad_norm": 0.19845978915691376, "learning_rate": 0.0001970930775297566, "loss": 0.8891, "step": 985 }, { "epoch": 1.699570815450644, "grad_norm": 0.18910686671733856, "learning_rate": 0.00019702086050260456, "loss": 0.8909, "step": 990 }, { "epoch": 1.7081545064377681, "grad_norm": 0.20300810039043427, "learning_rate": 0.00019694777095912534, "loss": 0.9012, "step": 995 }, { "epoch": 1.7167381974248928, "grad_norm": 0.21887531876564026, "learning_rate": 0.0001968738095566189, "loss": 0.9116, "step": 1000 }, { "epoch": 1.7253218884120172, "grad_norm": 0.21735869348049164, "learning_rate": 0.00019679897696022608, "loss": 0.8873, "step": 1005 }, { "epoch": 1.7339055793991416, "grad_norm": 0.1856721192598343, "learning_rate": 0.0001967232738429224, "loss": 0.8889, "step": 1010 }, { "epoch": 1.7424892703862662, "grad_norm": 0.2046109437942505, "learning_rate": 0.000196646700885512, "loss": 0.89, "step": 1015 }, { "epoch": 1.7510729613733904, "grad_norm": 0.19593974947929382, "learning_rate": 0.0001965692587766216, "loss": 0.9065, "step": 1020 }, { "epoch": 1.759656652360515, "grad_norm": 0.18540222942829132, "learning_rate": 0.00019649094821269425, "loss": 0.8877, "step": 1025 }, { "epoch": 1.7682403433476395, "grad_norm": 0.19899272918701172, "learning_rate": 0.00019641176989798305, "loss": 0.8965, "step": 1030 }, { "epoch": 1.7768240343347639, "grad_norm": 0.18957588076591492, "learning_rate": 0.00019633172454454497, "loss": 0.8876, "step": 1035 }, { "epoch": 1.7854077253218885, "grad_norm": 0.20278845727443695, "learning_rate": 0.0001962508128722342, "loss": 0.8982, "step": 1040 }, { "epoch": 1.7939914163090127, "grad_norm": 0.1874280869960785, "learning_rate": 0.00019616903560869584, "loss": 0.9049, "step": 1045 }, { "epoch": 1.8025751072961373, "grad_norm": 0.19025950133800507, "learning_rate": 0.0001960863934893594, "loss": 0.8901, "step": 1050 }, { "epoch": 1.8111587982832618, "grad_norm": 0.20806211233139038, "learning_rate": 0.00019600288725743194, "loss": 0.9015, "step": 1055 }, { "epoch": 1.8197424892703862, "grad_norm": 0.2008458375930786, "learning_rate": 0.00019591851766389176, "loss": 0.9029, "step": 1060 }, { "epoch": 1.8283261802575108, "grad_norm": 0.1986788511276245, "learning_rate": 0.00019583328546748127, "loss": 0.8942, "step": 1065 }, { "epoch": 1.8369098712446352, "grad_norm": 0.18976636230945587, "learning_rate": 0.00019574719143470044, "loss": 0.8962, "step": 1070 }, { "epoch": 1.8454935622317596, "grad_norm": 0.19317425787448883, "learning_rate": 0.00019566023633979976, "loss": 0.8918, "step": 1075 }, { "epoch": 1.8540772532188843, "grad_norm": 0.1889304369688034, "learning_rate": 0.00019557242096477327, "loss": 0.8934, "step": 1080 }, { "epoch": 1.8626609442060085, "grad_norm": 0.18771173059940338, "learning_rate": 0.00019548374609935172, "loss": 0.8782, "step": 1085 }, { "epoch": 1.871244635193133, "grad_norm": 0.18727517127990723, "learning_rate": 0.00019539421254099519, "loss": 0.9014, "step": 1090 }, { "epoch": 1.8798283261802575, "grad_norm": 0.19307033717632294, "learning_rate": 0.0001953038210948861, "loss": 0.896, "step": 1095 }, { "epoch": 1.888412017167382, "grad_norm": 0.1863000988960266, "learning_rate": 0.00019521257257392192, "loss": 0.8855, "step": 1100 }, { "epoch": 1.8969957081545066, "grad_norm": 0.1884726732969284, "learning_rate": 0.0001951204677987079, "loss": 0.8902, "step": 1105 }, { "epoch": 1.9055793991416308, "grad_norm": 0.20304642617702484, "learning_rate": 0.00019502750759754962, "loss": 0.8892, "step": 1110 }, { "epoch": 1.9141630901287554, "grad_norm": 0.1887015998363495, "learning_rate": 0.00019493369280644554, "loss": 0.8946, "step": 1115 }, { "epoch": 1.9227467811158798, "grad_norm": 0.18979288637638092, "learning_rate": 0.00019483902426907954, "loss": 0.8825, "step": 1120 }, { "epoch": 1.9313304721030042, "grad_norm": 0.18896907567977905, "learning_rate": 0.00019474350283681338, "loss": 0.887, "step": 1125 }, { "epoch": 1.9399141630901289, "grad_norm": 0.17926710844039917, "learning_rate": 0.00019464712936867885, "loss": 0.8832, "step": 1130 }, { "epoch": 1.948497854077253, "grad_norm": 0.19314360618591309, "learning_rate": 0.00019454990473137028, "loss": 0.89, "step": 1135 }, { "epoch": 1.9570815450643777, "grad_norm": 0.19700467586517334, "learning_rate": 0.00019445182979923654, "loss": 0.8844, "step": 1140 }, { "epoch": 1.9656652360515021, "grad_norm": 0.20681554079055786, "learning_rate": 0.00019435290545427328, "loss": 0.896, "step": 1145 }, { "epoch": 1.9742489270386265, "grad_norm": 0.1876552402973175, "learning_rate": 0.0001942531325861151, "loss": 0.886, "step": 1150 }, { "epoch": 1.9828326180257512, "grad_norm": 0.18256564438343048, "learning_rate": 0.0001941525120920273, "loss": 0.9008, "step": 1155 }, { "epoch": 1.9914163090128756, "grad_norm": 0.19153741002082825, "learning_rate": 0.00019405104487689798, "loss": 0.8804, "step": 1160 }, { "epoch": 2.0, "grad_norm": 0.19448307156562805, "learning_rate": 0.0001939487318532299, "loss": 0.8939, "step": 1165 }, { "epoch": 2.0, "eval_loss": 2.0231504440307617, "eval_runtime": 0.3938, "eval_samples_per_second": 15.238, "eval_steps_per_second": 2.54, "step": 1165 }, { "epoch": 2.0085836909871246, "grad_norm": 0.2327311784029007, "learning_rate": 0.00019384557394113228, "loss": 0.8256, "step": 1170 }, { "epoch": 2.017167381974249, "grad_norm": 0.23671980202198029, "learning_rate": 0.00019374157206831236, "loss": 0.8223, "step": 1175 }, { "epoch": 2.0257510729613735, "grad_norm": 0.23530033230781555, "learning_rate": 0.00019363672717006734, "loss": 0.8109, "step": 1180 }, { "epoch": 2.0343347639484977, "grad_norm": 0.21592716872692108, "learning_rate": 0.00019353104018927567, "loss": 0.8071, "step": 1185 }, { "epoch": 2.0429184549356223, "grad_norm": 0.22698843479156494, "learning_rate": 0.0001934245120763889, "loss": 0.8047, "step": 1190 }, { "epoch": 2.051502145922747, "grad_norm": 0.20773455500602722, "learning_rate": 0.0001933171437894227, "loss": 0.8081, "step": 1195 }, { "epoch": 2.060085836909871, "grad_norm": 0.21392963826656342, "learning_rate": 0.00019320893629394873, "loss": 0.8267, "step": 1200 }, { "epoch": 2.0686695278969958, "grad_norm": 0.1993769258260727, "learning_rate": 0.00019309989056308556, "loss": 0.8122, "step": 1205 }, { "epoch": 2.0772532188841204, "grad_norm": 0.21093840897083282, "learning_rate": 0.00019299000757749016, "loss": 0.8135, "step": 1210 }, { "epoch": 2.0858369098712446, "grad_norm": 0.21615874767303467, "learning_rate": 0.00019287928832534897, "loss": 0.8098, "step": 1215 }, { "epoch": 2.0944206008583692, "grad_norm": 0.21068502962589264, "learning_rate": 0.00019276773380236904, "loss": 0.813, "step": 1220 }, { "epoch": 2.1030042918454934, "grad_norm": 0.21265622973442078, "learning_rate": 0.00019265534501176906, "loss": 0.8034, "step": 1225 }, { "epoch": 2.111587982832618, "grad_norm": 0.22396954894065857, "learning_rate": 0.00019254212296427044, "loss": 0.8221, "step": 1230 }, { "epoch": 2.1201716738197427, "grad_norm": 0.2072274535894394, "learning_rate": 0.00019242806867808798, "loss": 0.8235, "step": 1235 }, { "epoch": 2.128755364806867, "grad_norm": 0.20798753201961517, "learning_rate": 0.00019231318317892106, "loss": 0.8137, "step": 1240 }, { "epoch": 2.1373390557939915, "grad_norm": 0.2030133157968521, "learning_rate": 0.00019219746749994405, "loss": 0.8154, "step": 1245 }, { "epoch": 2.1459227467811157, "grad_norm": 0.22628700733184814, "learning_rate": 0.0001920809226817973, "loss": 0.8261, "step": 1250 }, { "epoch": 2.1545064377682404, "grad_norm": 0.21634644269943237, "learning_rate": 0.00019196354977257766, "loss": 0.8221, "step": 1255 }, { "epoch": 2.163090128755365, "grad_norm": 0.2259581983089447, "learning_rate": 0.00019184534982782904, "loss": 0.8287, "step": 1260 }, { "epoch": 2.171673819742489, "grad_norm": 0.23607933521270752, "learning_rate": 0.00019172632391053294, "loss": 0.8218, "step": 1265 }, { "epoch": 2.180257510729614, "grad_norm": 0.20960725843906403, "learning_rate": 0.0001916064730910989, "loss": 0.8233, "step": 1270 }, { "epoch": 2.188841201716738, "grad_norm": 0.19818070530891418, "learning_rate": 0.00019148579844735497, "loss": 0.8253, "step": 1275 }, { "epoch": 2.1974248927038627, "grad_norm": 0.2142871767282486, "learning_rate": 0.00019136430106453777, "loss": 0.8289, "step": 1280 }, { "epoch": 2.2060085836909873, "grad_norm": 0.21934735774993896, "learning_rate": 0.0001912419820352829, "loss": 0.8191, "step": 1285 }, { "epoch": 2.2145922746781115, "grad_norm": 0.21653762459754944, "learning_rate": 0.00019111884245961522, "loss": 0.8194, "step": 1290 }, { "epoch": 2.223175965665236, "grad_norm": 0.21233248710632324, "learning_rate": 0.00019099488344493873, "loss": 0.8247, "step": 1295 }, { "epoch": 2.2317596566523603, "grad_norm": 0.23292584717273712, "learning_rate": 0.00019087010610602668, "loss": 0.8197, "step": 1300 }, { "epoch": 2.240343347639485, "grad_norm": 0.20501044392585754, "learning_rate": 0.00019074451156501164, "loss": 0.8152, "step": 1305 }, { "epoch": 2.2489270386266096, "grad_norm": 0.23035867512226105, "learning_rate": 0.00019061810095137533, "loss": 0.8168, "step": 1310 }, { "epoch": 2.257510729613734, "grad_norm": 0.21323524415493011, "learning_rate": 0.00019049087540193847, "loss": 0.8118, "step": 1315 }, { "epoch": 2.2660944206008584, "grad_norm": 0.20477545261383057, "learning_rate": 0.00019036283606085053, "loss": 0.8164, "step": 1320 }, { "epoch": 2.274678111587983, "grad_norm": 0.21431773900985718, "learning_rate": 0.00019023398407957956, "loss": 0.8258, "step": 1325 }, { "epoch": 2.2832618025751072, "grad_norm": 0.21260547637939453, "learning_rate": 0.00019010432061690165, "loss": 0.8166, "step": 1330 }, { "epoch": 2.291845493562232, "grad_norm": 0.20846493542194366, "learning_rate": 0.00018997384683889067, "loss": 0.8066, "step": 1335 }, { "epoch": 2.300429184549356, "grad_norm": 0.20466403663158417, "learning_rate": 0.00018984256391890765, "loss": 0.8251, "step": 1340 }, { "epoch": 2.3090128755364807, "grad_norm": 0.20920304954051971, "learning_rate": 0.0001897104730375904, "loss": 0.8164, "step": 1345 }, { "epoch": 2.317596566523605, "grad_norm": 0.22407568991184235, "learning_rate": 0.00018957757538284273, "loss": 0.8156, "step": 1350 }, { "epoch": 2.3261802575107295, "grad_norm": 0.22706876695156097, "learning_rate": 0.00018944387214982382, "loss": 0.8348, "step": 1355 }, { "epoch": 2.334763948497854, "grad_norm": 0.20964165031909943, "learning_rate": 0.00018930936454093753, "loss": 0.8258, "step": 1360 }, { "epoch": 2.3433476394849784, "grad_norm": 0.23025156557559967, "learning_rate": 0.00018917405376582145, "loss": 0.8205, "step": 1365 }, { "epoch": 2.351931330472103, "grad_norm": 0.23171818256378174, "learning_rate": 0.0001890379410413362, "loss": 0.8224, "step": 1370 }, { "epoch": 2.3605150214592276, "grad_norm": 0.2035280019044876, "learning_rate": 0.0001889010275915543, "loss": 0.8363, "step": 1375 }, { "epoch": 2.369098712446352, "grad_norm": 0.2576422691345215, "learning_rate": 0.00018876331464774945, "loss": 0.8216, "step": 1380 }, { "epoch": 2.3776824034334765, "grad_norm": 0.21184222400188446, "learning_rate": 0.00018862480344838495, "loss": 0.8161, "step": 1385 }, { "epoch": 2.3862660944206007, "grad_norm": 0.22491346299648285, "learning_rate": 0.00018848549523910313, "loss": 0.8261, "step": 1390 }, { "epoch": 2.3948497854077253, "grad_norm": 0.21227188408374786, "learning_rate": 0.0001883453912727138, "loss": 0.8377, "step": 1395 }, { "epoch": 2.40343347639485, "grad_norm": 0.21044416725635529, "learning_rate": 0.0001882044928091831, "loss": 0.819, "step": 1400 }, { "epoch": 2.412017167381974, "grad_norm": 0.20745404064655304, "learning_rate": 0.00018806280111562215, "loss": 0.8265, "step": 1405 }, { "epoch": 2.4206008583690988, "grad_norm": 0.2179802805185318, "learning_rate": 0.00018792031746627563, "loss": 0.8382, "step": 1410 }, { "epoch": 2.429184549356223, "grad_norm": 0.20480507612228394, "learning_rate": 0.00018777704314251032, "loss": 0.8312, "step": 1415 }, { "epoch": 2.4377682403433476, "grad_norm": 0.23306381702423096, "learning_rate": 0.00018763297943280368, "loss": 0.8161, "step": 1420 }, { "epoch": 2.4463519313304722, "grad_norm": 0.21607355773448944, "learning_rate": 0.00018748812763273208, "loss": 0.8197, "step": 1425 }, { "epoch": 2.4549356223175964, "grad_norm": 0.21942569315433502, "learning_rate": 0.0001873424890449593, "loss": 0.8291, "step": 1430 }, { "epoch": 2.463519313304721, "grad_norm": 0.2144131362438202, "learning_rate": 0.00018719606497922476, "loss": 0.8203, "step": 1435 }, { "epoch": 2.4721030042918457, "grad_norm": 0.21602974832057953, "learning_rate": 0.0001870488567523318, "loss": 0.8154, "step": 1440 }, { "epoch": 2.48068669527897, "grad_norm": 0.2094966620206833, "learning_rate": 0.0001869008656881357, "loss": 0.8197, "step": 1445 }, { "epoch": 2.4892703862660945, "grad_norm": 0.21330519020557404, "learning_rate": 0.00018675209311753185, "loss": 0.8325, "step": 1450 }, { "epoch": 2.4978540772532187, "grad_norm": 0.22934697568416595, "learning_rate": 0.00018660254037844388, "loss": 0.8238, "step": 1455 }, { "epoch": 2.5064377682403434, "grad_norm": 0.23202557861804962, "learning_rate": 0.00018645220881581144, "loss": 0.8277, "step": 1460 }, { "epoch": 2.5150214592274676, "grad_norm": 0.22423741221427917, "learning_rate": 0.0001863010997815783, "loss": 0.8205, "step": 1465 }, { "epoch": 2.523605150214592, "grad_norm": 0.2139664888381958, "learning_rate": 0.00018614921463468002, "loss": 0.833, "step": 1470 }, { "epoch": 2.532188841201717, "grad_norm": 0.22042877972126007, "learning_rate": 0.00018599655474103182, "loss": 0.8281, "step": 1475 }, { "epoch": 2.540772532188841, "grad_norm": 0.21639470756053925, "learning_rate": 0.0001858431214735163, "loss": 0.8353, "step": 1480 }, { "epoch": 2.5493562231759657, "grad_norm": 0.21406595408916473, "learning_rate": 0.00018568891621197103, "loss": 0.8351, "step": 1485 }, { "epoch": 2.5579399141630903, "grad_norm": 0.20837725698947906, "learning_rate": 0.00018553394034317622, "loss": 0.8251, "step": 1490 }, { "epoch": 2.5665236051502145, "grad_norm": 0.21612149477005005, "learning_rate": 0.0001853781952608422, "loss": 0.846, "step": 1495 }, { "epoch": 2.575107296137339, "grad_norm": 0.21887291967868805, "learning_rate": 0.00018522168236559695, "loss": 0.8388, "step": 1500 }, { "epoch": 2.5836909871244638, "grad_norm": 0.20973001420497894, "learning_rate": 0.00018506440306497335, "loss": 0.839, "step": 1505 }, { "epoch": 2.592274678111588, "grad_norm": 0.21462783217430115, "learning_rate": 0.00018490635877339666, "loss": 0.8276, "step": 1510 }, { "epoch": 2.6008583690987126, "grad_norm": 0.210985004901886, "learning_rate": 0.00018474755091217186, "loss": 0.8221, "step": 1515 }, { "epoch": 2.609442060085837, "grad_norm": 0.20986580848693848, "learning_rate": 0.00018458798090947065, "loss": 0.8234, "step": 1520 }, { "epoch": 2.6180257510729614, "grad_norm": 0.22892533242702484, "learning_rate": 0.00018442765020031877, "loss": 0.8242, "step": 1525 }, { "epoch": 2.6266094420600856, "grad_norm": 0.2284938395023346, "learning_rate": 0.0001842665602265831, "loss": 0.8161, "step": 1530 }, { "epoch": 2.6351931330472103, "grad_norm": 0.22317782044410706, "learning_rate": 0.00018410471243695856, "loss": 0.8284, "step": 1535 }, { "epoch": 2.643776824034335, "grad_norm": 0.21049915254116058, "learning_rate": 0.00018394210828695523, "loss": 0.8183, "step": 1540 }, { "epoch": 2.652360515021459, "grad_norm": 0.21207213401794434, "learning_rate": 0.0001837787492388852, "loss": 0.8287, "step": 1545 }, { "epoch": 2.6609442060085837, "grad_norm": 0.2118200659751892, "learning_rate": 0.0001836146367618494, "loss": 0.8204, "step": 1550 }, { "epoch": 2.6695278969957084, "grad_norm": 0.22095955908298492, "learning_rate": 0.00018344977233172437, "loss": 0.8335, "step": 1555 }, { "epoch": 2.6781115879828326, "grad_norm": 0.21252469718456268, "learning_rate": 0.00018328415743114912, "loss": 0.8191, "step": 1560 }, { "epoch": 2.686695278969957, "grad_norm": 0.20323017239570618, "learning_rate": 0.0001831177935495116, "loss": 0.8231, "step": 1565 }, { "epoch": 2.6952789699570814, "grad_norm": 0.21805858612060547, "learning_rate": 0.00018295068218293547, "loss": 0.8341, "step": 1570 }, { "epoch": 2.703862660944206, "grad_norm": 0.21013419330120087, "learning_rate": 0.00018278282483426658, "loss": 0.839, "step": 1575 }, { "epoch": 2.71244635193133, "grad_norm": 0.21768461167812347, "learning_rate": 0.0001826142230130594, "loss": 0.8356, "step": 1580 }, { "epoch": 2.721030042918455, "grad_norm": 0.21069899201393127, "learning_rate": 0.00018244487823556357, "loss": 0.8188, "step": 1585 }, { "epoch": 2.7296137339055795, "grad_norm": 0.21788835525512695, "learning_rate": 0.00018227479202471015, "loss": 0.8408, "step": 1590 }, { "epoch": 2.7381974248927037, "grad_norm": 0.21380050480365753, "learning_rate": 0.00018210396591009795, "loss": 0.8358, "step": 1595 }, { "epoch": 2.7467811158798283, "grad_norm": 0.21521276235580444, "learning_rate": 0.00018193240142797988, "loss": 0.8328, "step": 1600 }, { "epoch": 2.755364806866953, "grad_norm": 0.20885252952575684, "learning_rate": 0.000181760100121249, "loss": 0.8238, "step": 1605 }, { "epoch": 2.763948497854077, "grad_norm": 0.21117731928825378, "learning_rate": 0.00018158706353942463, "loss": 0.8301, "step": 1610 }, { "epoch": 2.772532188841202, "grad_norm": 0.22012095153331757, "learning_rate": 0.0001814132932386386, "loss": 0.8357, "step": 1615 }, { "epoch": 2.7811158798283264, "grad_norm": 0.22017072141170502, "learning_rate": 0.00018123879078162097, "loss": 0.8323, "step": 1620 }, { "epoch": 2.7896995708154506, "grad_norm": 0.2259422242641449, "learning_rate": 0.00018106355773768638, "loss": 0.848, "step": 1625 }, { "epoch": 2.7982832618025753, "grad_norm": 0.21191255748271942, "learning_rate": 0.0001808875956827194, "loss": 0.823, "step": 1630 }, { "epoch": 2.8068669527896994, "grad_norm": 0.21371833980083466, "learning_rate": 0.00018071090619916093, "loss": 0.8194, "step": 1635 }, { "epoch": 2.815450643776824, "grad_norm": 0.22189456224441528, "learning_rate": 0.00018053349087599353, "loss": 0.8329, "step": 1640 }, { "epoch": 2.8240343347639483, "grad_norm": 0.20956319570541382, "learning_rate": 0.00018035535130872732, "loss": 0.8293, "step": 1645 }, { "epoch": 2.832618025751073, "grad_norm": 0.21734033524990082, "learning_rate": 0.0001801764890993856, "loss": 0.8334, "step": 1650 }, { "epoch": 2.8412017167381975, "grad_norm": 0.2138412892818451, "learning_rate": 0.00017999690585649052, "loss": 0.8354, "step": 1655 }, { "epoch": 2.8497854077253217, "grad_norm": 0.21562372148036957, "learning_rate": 0.00017981660319504845, "loss": 0.8384, "step": 1660 }, { "epoch": 2.8583690987124464, "grad_norm": 0.21281686425209045, "learning_rate": 0.0001796355827365356, "loss": 0.8312, "step": 1665 }, { "epoch": 2.866952789699571, "grad_norm": 0.21461673080921173, "learning_rate": 0.00017945384610888341, "loss": 0.8344, "step": 1670 }, { "epoch": 2.875536480686695, "grad_norm": 0.20743022859096527, "learning_rate": 0.00017927139494646377, "loss": 0.8215, "step": 1675 }, { "epoch": 2.88412017167382, "grad_norm": 0.21129368245601654, "learning_rate": 0.00017908823089007457, "loss": 0.8274, "step": 1680 }, { "epoch": 2.8927038626609445, "grad_norm": 0.2333795428276062, "learning_rate": 0.00017890435558692475, "loss": 0.8307, "step": 1685 }, { "epoch": 2.9012875536480687, "grad_norm": 0.21824228763580322, "learning_rate": 0.0001787197706906196, "loss": 0.8498, "step": 1690 }, { "epoch": 2.909871244635193, "grad_norm": 0.21459732949733734, "learning_rate": 0.0001785344778611457, "loss": 0.8265, "step": 1695 }, { "epoch": 2.9184549356223175, "grad_norm": 0.20637935400009155, "learning_rate": 0.00017834847876485629, "loss": 0.8309, "step": 1700 }, { "epoch": 2.927038626609442, "grad_norm": 0.2137777954339981, "learning_rate": 0.0001781617750744561, "loss": 0.8345, "step": 1705 }, { "epoch": 2.9356223175965663, "grad_norm": 0.23476457595825195, "learning_rate": 0.00017797436846898619, "loss": 0.8335, "step": 1710 }, { "epoch": 2.944206008583691, "grad_norm": 0.20995980501174927, "learning_rate": 0.00017778626063380917, "loss": 0.8209, "step": 1715 }, { "epoch": 2.9527896995708156, "grad_norm": 0.2296920269727707, "learning_rate": 0.00017759745326059379, "loss": 0.8426, "step": 1720 }, { "epoch": 2.96137339055794, "grad_norm": 0.20545101165771484, "learning_rate": 0.00017740794804729969, "loss": 0.8324, "step": 1725 }, { "epoch": 2.9699570815450644, "grad_norm": 0.21105705201625824, "learning_rate": 0.00017721774669816252, "loss": 0.8212, "step": 1730 }, { "epoch": 2.978540772532189, "grad_norm": 0.21741057932376862, "learning_rate": 0.000177026850923678, "loss": 0.8333, "step": 1735 }, { "epoch": 2.9871244635193133, "grad_norm": 0.22390629351139069, "learning_rate": 0.00017683526244058716, "loss": 0.8364, "step": 1740 }, { "epoch": 2.995708154506438, "grad_norm": 0.21623565256595612, "learning_rate": 0.00017664298297186042, "loss": 0.8255, "step": 1745 }, { "epoch": 2.9991416309012875, "eval_loss": 2.1085665225982666, "eval_runtime": 0.3945, "eval_samples_per_second": 15.208, "eval_steps_per_second": 2.535, "step": 1747 }, { "epoch": 3.004291845493562, "grad_norm": 0.21203316748142242, "learning_rate": 0.00017645001424668237, "loss": 0.7739, "step": 1750 }, { "epoch": 3.0128755364806867, "grad_norm": 0.23032112419605255, "learning_rate": 0.00017625635800043617, "loss": 0.741, "step": 1755 }, { "epoch": 3.0214592274678114, "grad_norm": 0.24847178161144257, "learning_rate": 0.00017606201597468782, "loss": 0.7348, "step": 1760 }, { "epoch": 3.0300429184549356, "grad_norm": 0.24480335414409637, "learning_rate": 0.00017586698991717064, "loss": 0.7212, "step": 1765 }, { "epoch": 3.03862660944206, "grad_norm": 0.24489726126194, "learning_rate": 0.00017567128158176953, "loss": 0.7312, "step": 1770 }, { "epoch": 3.0472103004291844, "grad_norm": 0.24028155207633972, "learning_rate": 0.00017547489272850511, "loss": 0.7271, "step": 1775 }, { "epoch": 3.055793991416309, "grad_norm": 0.24730311334133148, "learning_rate": 0.00017527782512351804, "loss": 0.7344, "step": 1780 }, { "epoch": 3.0643776824034337, "grad_norm": 0.23651528358459473, "learning_rate": 0.00017508008053905295, "loss": 0.7297, "step": 1785 }, { "epoch": 3.072961373390558, "grad_norm": 0.23505684733390808, "learning_rate": 0.0001748816607534426, "loss": 0.7214, "step": 1790 }, { "epoch": 3.0815450643776825, "grad_norm": 0.2424248605966568, "learning_rate": 0.00017468256755109199, "loss": 0.721, "step": 1795 }, { "epoch": 3.0901287553648067, "grad_norm": 0.243468776345253, "learning_rate": 0.00017448280272246212, "loss": 0.7203, "step": 1800 }, { "epoch": 3.0987124463519313, "grad_norm": 0.2453926056623459, "learning_rate": 0.000174282368064054, "loss": 0.7326, "step": 1805 }, { "epoch": 3.107296137339056, "grad_norm": 0.24548988044261932, "learning_rate": 0.00017408126537839252, "loss": 0.7345, "step": 1810 }, { "epoch": 3.11587982832618, "grad_norm": 0.2359829694032669, "learning_rate": 0.00017387949647401012, "loss": 0.748, "step": 1815 }, { "epoch": 3.124463519313305, "grad_norm": 0.25862741470336914, "learning_rate": 0.00017367706316543063, "loss": 0.7338, "step": 1820 }, { "epoch": 3.133047210300429, "grad_norm": 0.25267720222473145, "learning_rate": 0.00017347396727315296, "loss": 0.7403, "step": 1825 }, { "epoch": 3.1416309012875536, "grad_norm": 0.2458384782075882, "learning_rate": 0.00017327021062363458, "loss": 0.7432, "step": 1830 }, { "epoch": 3.1502145922746783, "grad_norm": 0.2578388750553131, "learning_rate": 0.0001730657950492753, "loss": 0.7447, "step": 1835 }, { "epoch": 3.1587982832618025, "grad_norm": 0.23755429685115814, "learning_rate": 0.00017286072238840067, "loss": 0.7389, "step": 1840 }, { "epoch": 3.167381974248927, "grad_norm": 0.24692735075950623, "learning_rate": 0.0001726549944852455, "loss": 0.7584, "step": 1845 }, { "epoch": 3.1759656652360517, "grad_norm": 0.24396221339702606, "learning_rate": 0.00017244861318993713, "loss": 0.7386, "step": 1850 }, { "epoch": 3.184549356223176, "grad_norm": 0.25548049807548523, "learning_rate": 0.00017224158035847905, "loss": 0.738, "step": 1855 }, { "epoch": 3.1931330472103006, "grad_norm": 0.2472919523715973, "learning_rate": 0.000172033897852734, "loss": 0.7519, "step": 1860 }, { "epoch": 3.2017167381974247, "grad_norm": 0.245948925614357, "learning_rate": 0.0001718255675404073, "loss": 0.7461, "step": 1865 }, { "epoch": 3.2103004291845494, "grad_norm": 0.2513918876647949, "learning_rate": 0.00017161659129503003, "loss": 0.7458, "step": 1870 }, { "epoch": 3.218884120171674, "grad_norm": 0.24049414694309235, "learning_rate": 0.0001714069709959422, "loss": 0.7344, "step": 1875 }, { "epoch": 3.227467811158798, "grad_norm": 0.25180676579475403, "learning_rate": 0.00017119670852827588, "loss": 0.7378, "step": 1880 }, { "epoch": 3.236051502145923, "grad_norm": 0.2704819440841675, "learning_rate": 0.0001709858057829382, "loss": 0.7491, "step": 1885 }, { "epoch": 3.244635193133047, "grad_norm": 0.2382296621799469, "learning_rate": 0.00017077426465659433, "loss": 0.7433, "step": 1890 }, { "epoch": 3.2532188841201717, "grad_norm": 0.25334346294403076, "learning_rate": 0.00017056208705165045, "loss": 0.7505, "step": 1895 }, { "epoch": 3.2618025751072963, "grad_norm": 0.2550380527973175, "learning_rate": 0.0001703492748762367, "loss": 0.7531, "step": 1900 }, { "epoch": 3.2703862660944205, "grad_norm": 0.257135808467865, "learning_rate": 0.00017013583004418993, "loss": 0.7453, "step": 1905 }, { "epoch": 3.278969957081545, "grad_norm": 0.27000248432159424, "learning_rate": 0.0001699217544750365, "loss": 0.7512, "step": 1910 }, { "epoch": 3.2875536480686693, "grad_norm": 0.25268518924713135, "learning_rate": 0.00016970705009397504, "loss": 0.7397, "step": 1915 }, { "epoch": 3.296137339055794, "grad_norm": 0.26630303263664246, "learning_rate": 0.00016949171883185918, "loss": 0.7436, "step": 1920 }, { "epoch": 3.3047210300429186, "grad_norm": 0.24609267711639404, "learning_rate": 0.0001692757626251801, "loss": 0.7402, "step": 1925 }, { "epoch": 3.313304721030043, "grad_norm": 0.26151612401008606, "learning_rate": 0.00016905918341604922, "loss": 0.7424, "step": 1930 }, { "epoch": 3.3218884120171674, "grad_norm": 0.2529394030570984, "learning_rate": 0.00016884198315218055, "loss": 0.7566, "step": 1935 }, { "epoch": 3.3304721030042916, "grad_norm": 0.2545251250267029, "learning_rate": 0.0001686241637868734, "loss": 0.762, "step": 1940 }, { "epoch": 3.3390557939914163, "grad_norm": 0.249998539686203, "learning_rate": 0.00016840572727899462, "loss": 0.7399, "step": 1945 }, { "epoch": 3.347639484978541, "grad_norm": 0.24969734251499176, "learning_rate": 0.0001681866755929612, "loss": 0.748, "step": 1950 }, { "epoch": 3.356223175965665, "grad_norm": 0.262955904006958, "learning_rate": 0.00016796701069872238, "loss": 0.754, "step": 1955 }, { "epoch": 3.3648068669527897, "grad_norm": 0.2457767277956009, "learning_rate": 0.00016774673457174206, "loss": 0.7443, "step": 1960 }, { "epoch": 3.3733905579399144, "grad_norm": 0.2644675374031067, "learning_rate": 0.00016752584919298093, "loss": 0.7519, "step": 1965 }, { "epoch": 3.3819742489270386, "grad_norm": 0.2620808482170105, "learning_rate": 0.0001673043565488789, "loss": 0.7556, "step": 1970 }, { "epoch": 3.390557939914163, "grad_norm": 0.2510511577129364, "learning_rate": 0.00016708225863133693, "loss": 0.7556, "step": 1975 }, { "epoch": 3.3991416309012874, "grad_norm": 0.2542615830898285, "learning_rate": 0.0001668595574376992, "loss": 0.7502, "step": 1980 }, { "epoch": 3.407725321888412, "grad_norm": 0.25436341762542725, "learning_rate": 0.0001666362549707354, "loss": 0.7505, "step": 1985 }, { "epoch": 3.4163090128755362, "grad_norm": 0.24043235182762146, "learning_rate": 0.00016641235323862236, "loss": 0.7433, "step": 1990 }, { "epoch": 3.424892703862661, "grad_norm": 0.25933003425598145, "learning_rate": 0.00016618785425492617, "loss": 0.7595, "step": 1995 }, { "epoch": 3.4334763948497855, "grad_norm": 0.24922600388526917, "learning_rate": 0.00016596276003858412, "loss": 0.7489, "step": 2000 }, { "epoch": 3.4420600858369097, "grad_norm": 0.23722489178180695, "learning_rate": 0.0001657370726138864, "loss": 0.7447, "step": 2005 }, { "epoch": 3.4506437768240343, "grad_norm": 0.273787260055542, "learning_rate": 0.000165510794010458, "loss": 0.75, "step": 2010 }, { "epoch": 3.459227467811159, "grad_norm": 0.23869618773460388, "learning_rate": 0.0001652839262632404, "loss": 0.7463, "step": 2015 }, { "epoch": 3.467811158798283, "grad_norm": 0.264568030834198, "learning_rate": 0.0001650564714124734, "loss": 0.7566, "step": 2020 }, { "epoch": 3.476394849785408, "grad_norm": 0.2636789083480835, "learning_rate": 0.0001648284315036765, "loss": 0.749, "step": 2025 }, { "epoch": 3.484978540772532, "grad_norm": 0.24988381564617157, "learning_rate": 0.0001645998085876308, "loss": 0.7523, "step": 2030 }, { "epoch": 3.4935622317596566, "grad_norm": 0.26084083318710327, "learning_rate": 0.00016437060472036046, "loss": 0.7541, "step": 2035 }, { "epoch": 3.5021459227467813, "grad_norm": 0.2548128664493561, "learning_rate": 0.000164140821963114, "loss": 0.7593, "step": 2040 }, { "epoch": 3.5107296137339055, "grad_norm": 0.2459845244884491, "learning_rate": 0.00016391046238234616, "loss": 0.7485, "step": 2045 }, { "epoch": 3.51931330472103, "grad_norm": 0.26049911975860596, "learning_rate": 0.00016367952804969895, "loss": 0.7492, "step": 2050 }, { "epoch": 3.5278969957081543, "grad_norm": 0.2775178551673889, "learning_rate": 0.00016344802104198324, "loss": 0.7534, "step": 2055 }, { "epoch": 3.536480686695279, "grad_norm": 0.26411354541778564, "learning_rate": 0.00016321594344115997, "loss": 0.7597, "step": 2060 }, { "epoch": 3.5450643776824036, "grad_norm": 0.26565033197402954, "learning_rate": 0.00016298329733432153, "loss": 0.7659, "step": 2065 }, { "epoch": 3.5536480686695278, "grad_norm": 0.2576376795768738, "learning_rate": 0.00016275008481367287, "loss": 0.7632, "step": 2070 }, { "epoch": 3.5622317596566524, "grad_norm": 0.25273096561431885, "learning_rate": 0.00016251630797651276, "loss": 0.7604, "step": 2075 }, { "epoch": 3.570815450643777, "grad_norm": 0.25641995668411255, "learning_rate": 0.0001622819689252149, "loss": 0.7575, "step": 2080 }, { "epoch": 3.5793991416309012, "grad_norm": 0.24130167067050934, "learning_rate": 0.0001620470697672091, "loss": 0.7512, "step": 2085 }, { "epoch": 3.587982832618026, "grad_norm": 0.2561969757080078, "learning_rate": 0.00016181161261496216, "loss": 0.7555, "step": 2090 }, { "epoch": 3.59656652360515, "grad_norm": 0.2512848675251007, "learning_rate": 0.000161575599585959, "loss": 0.7556, "step": 2095 }, { "epoch": 3.6051502145922747, "grad_norm": 0.2550983130931854, "learning_rate": 0.00016133903280268362, "loss": 0.7673, "step": 2100 }, { "epoch": 3.613733905579399, "grad_norm": 0.2565702795982361, "learning_rate": 0.00016110191439259997, "loss": 0.7662, "step": 2105 }, { "epoch": 3.6223175965665235, "grad_norm": 0.24696961045265198, "learning_rate": 0.00016086424648813273, "loss": 0.742, "step": 2110 }, { "epoch": 3.630901287553648, "grad_norm": 0.2504982054233551, "learning_rate": 0.00016062603122664833, "loss": 0.7514, "step": 2115 }, { "epoch": 3.6394849785407724, "grad_norm": 0.2710771858692169, "learning_rate": 0.00016038727075043562, "loss": 0.7681, "step": 2120 }, { "epoch": 3.648068669527897, "grad_norm": 0.26008063554763794, "learning_rate": 0.0001601479672066865, "loss": 0.7655, "step": 2125 }, { "epoch": 3.6566523605150216, "grad_norm": 0.25275853276252747, "learning_rate": 0.00015990812274747692, "loss": 0.761, "step": 2130 }, { "epoch": 3.665236051502146, "grad_norm": 0.26592886447906494, "learning_rate": 0.00015966773952974715, "loss": 0.7529, "step": 2135 }, { "epoch": 3.6738197424892705, "grad_norm": 0.24650226533412933, "learning_rate": 0.0001594268197152826, "loss": 0.7538, "step": 2140 }, { "epoch": 3.682403433476395, "grad_norm": 0.25551459193229675, "learning_rate": 0.00015918536547069435, "loss": 0.7719, "step": 2145 }, { "epoch": 3.6909871244635193, "grad_norm": 0.24978633224964142, "learning_rate": 0.0001589433789673997, "loss": 0.7545, "step": 2150 }, { "epoch": 3.699570815450644, "grad_norm": 0.2471482902765274, "learning_rate": 0.0001587008623816025, "loss": 0.749, "step": 2155 }, { "epoch": 3.708154506437768, "grad_norm": 0.24815160036087036, "learning_rate": 0.00015845781789427377, "loss": 0.7506, "step": 2160 }, { "epoch": 3.7167381974248928, "grad_norm": 0.26376664638519287, "learning_rate": 0.00015821424769113193, "loss": 0.7569, "step": 2165 }, { "epoch": 3.725321888412017, "grad_norm": 0.25356897711753845, "learning_rate": 0.0001579701539626232, "loss": 0.7707, "step": 2170 }, { "epoch": 3.7339055793991416, "grad_norm": 0.25035008788108826, "learning_rate": 0.00015772553890390197, "loss": 0.76, "step": 2175 }, { "epoch": 3.742489270386266, "grad_norm": 0.2481870949268341, "learning_rate": 0.0001574804047148109, "loss": 0.7521, "step": 2180 }, { "epoch": 3.7510729613733904, "grad_norm": 0.25502651929855347, "learning_rate": 0.00015723475359986127, "loss": 0.7713, "step": 2185 }, { "epoch": 3.759656652360515, "grad_norm": 0.25871410965919495, "learning_rate": 0.0001569885877682132, "loss": 0.7668, "step": 2190 }, { "epoch": 3.7682403433476397, "grad_norm": 0.25364378094673157, "learning_rate": 0.00015674190943365556, "loss": 0.754, "step": 2195 }, { "epoch": 3.776824034334764, "grad_norm": 0.2515285909175873, "learning_rate": 0.0001564947208145863, "loss": 0.7689, "step": 2200 }, { "epoch": 3.7854077253218885, "grad_norm": 0.24017582833766937, "learning_rate": 0.00015624702413399231, "loss": 0.7718, "step": 2205 }, { "epoch": 3.7939914163090127, "grad_norm": 0.26583361625671387, "learning_rate": 0.00015599882161942966, "loss": 0.7668, "step": 2210 }, { "epoch": 3.8025751072961373, "grad_norm": 0.2555334270000458, "learning_rate": 0.00015575011550300323, "loss": 0.7507, "step": 2215 }, { "epoch": 3.8111587982832615, "grad_norm": 0.2571168839931488, "learning_rate": 0.000155500908021347, "loss": 0.752, "step": 2220 }, { "epoch": 3.819742489270386, "grad_norm": 0.26001662015914917, "learning_rate": 0.0001552512014156037, "loss": 0.7633, "step": 2225 }, { "epoch": 3.828326180257511, "grad_norm": 0.25432994961738586, "learning_rate": 0.00015500099793140475, "loss": 0.758, "step": 2230 }, { "epoch": 3.836909871244635, "grad_norm": 0.24695011973381042, "learning_rate": 0.00015475029981884992, "loss": 0.7674, "step": 2235 }, { "epoch": 3.8454935622317596, "grad_norm": 0.24964170157909393, "learning_rate": 0.00015449910933248743, "loss": 0.7611, "step": 2240 }, { "epoch": 3.8540772532188843, "grad_norm": 0.2555537223815918, "learning_rate": 0.00015424742873129324, "loss": 0.7657, "step": 2245 }, { "epoch": 3.8626609442060085, "grad_norm": 0.27087053656578064, "learning_rate": 0.00015399526027865107, "loss": 0.7645, "step": 2250 }, { "epoch": 3.871244635193133, "grad_norm": 0.2602386176586151, "learning_rate": 0.00015374260624233195, "loss": 0.7604, "step": 2255 }, { "epoch": 3.8798283261802577, "grad_norm": 0.24609869718551636, "learning_rate": 0.00015348946889447368, "loss": 0.7596, "step": 2260 }, { "epoch": 3.888412017167382, "grad_norm": 0.2546113133430481, "learning_rate": 0.0001532358505115607, "loss": 0.7594, "step": 2265 }, { "epoch": 3.8969957081545066, "grad_norm": 0.25102177262306213, "learning_rate": 0.0001529817533744032, "loss": 0.7606, "step": 2270 }, { "epoch": 3.9055793991416308, "grad_norm": 0.2544861435890198, "learning_rate": 0.00015272717976811708, "loss": 0.7535, "step": 2275 }, { "epoch": 3.9141630901287554, "grad_norm": 0.2670022249221802, "learning_rate": 0.000152472131982103, "loss": 0.7609, "step": 2280 }, { "epoch": 3.9227467811158796, "grad_norm": 0.2539633810520172, "learning_rate": 0.00015221661231002605, "loss": 0.7556, "step": 2285 }, { "epoch": 3.9313304721030042, "grad_norm": 0.2580619156360626, "learning_rate": 0.00015196062304979497, "loss": 0.7717, "step": 2290 }, { "epoch": 3.939914163090129, "grad_norm": 0.24921616911888123, "learning_rate": 0.00015170416650354157, "loss": 0.7642, "step": 2295 }, { "epoch": 3.948497854077253, "grad_norm": 0.2630159556865692, "learning_rate": 0.00015144724497760003, "loss": 0.7522, "step": 2300 }, { "epoch": 3.9570815450643777, "grad_norm": 0.2687895894050598, "learning_rate": 0.00015118986078248612, "loss": 0.7623, "step": 2305 }, { "epoch": 3.9656652360515023, "grad_norm": 0.27267009019851685, "learning_rate": 0.00015093201623287631, "loss": 0.7755, "step": 2310 }, { "epoch": 3.9742489270386265, "grad_norm": 0.2584875524044037, "learning_rate": 0.00015067371364758727, "loss": 0.7582, "step": 2315 }, { "epoch": 3.982832618025751, "grad_norm": 0.25810128450393677, "learning_rate": 0.00015041495534955467, "loss": 0.7738, "step": 2320 }, { "epoch": 3.991416309012876, "grad_norm": 0.42070698738098145, "learning_rate": 0.00015015574366581257, "loss": 0.7609, "step": 2325 }, { "epoch": 4.0, "grad_norm": 0.2550402581691742, "learning_rate": 0.0001498960809274722, "loss": 0.7584, "step": 2330 }, { "epoch": 4.0, "eval_loss": 2.2541472911834717, "eval_runtime": 0.3945, "eval_samples_per_second": 15.209, "eval_steps_per_second": 2.535, "step": 2330 }, { "epoch": 4.008583690987124, "grad_norm": 0.34891819953918457, "learning_rate": 0.00014963596946970128, "loss": 0.6641, "step": 2335 }, { "epoch": 4.017167381974249, "grad_norm": 0.26585909724235535, "learning_rate": 0.0001493754116317029, "loss": 0.657, "step": 2340 }, { "epoch": 4.0257510729613735, "grad_norm": 0.3037340044975281, "learning_rate": 0.00014911440975669447, "loss": 0.6571, "step": 2345 }, { "epoch": 4.034334763948498, "grad_norm": 0.3071967363357544, "learning_rate": 0.00014885296619188658, "loss": 0.6607, "step": 2350 }, { "epoch": 4.042918454935623, "grad_norm": 0.28058749437332153, "learning_rate": 0.00014859108328846204, "loss": 0.6558, "step": 2355 }, { "epoch": 4.051502145922747, "grad_norm": 0.27246275544166565, "learning_rate": 0.00014832876340155476, "loss": 0.6464, "step": 2360 }, { "epoch": 4.060085836909871, "grad_norm": 0.30583012104034424, "learning_rate": 0.00014806600889022824, "loss": 0.6602, "step": 2365 }, { "epoch": 4.068669527896995, "grad_norm": 0.2768241763114929, "learning_rate": 0.0001478028221174548, "loss": 0.6601, "step": 2370 }, { "epoch": 4.07725321888412, "grad_norm": 0.2900557219982147, "learning_rate": 0.00014753920545009408, "loss": 0.6612, "step": 2375 }, { "epoch": 4.085836909871245, "grad_norm": 0.28369641304016113, "learning_rate": 0.00014727516125887175, "loss": 0.6497, "step": 2380 }, { "epoch": 4.094420600858369, "grad_norm": 0.2893315553665161, "learning_rate": 0.0001470106919183582, "loss": 0.6615, "step": 2385 }, { "epoch": 4.103004291845494, "grad_norm": 0.2940422296524048, "learning_rate": 0.00014674579980694736, "loss": 0.6583, "step": 2390 }, { "epoch": 4.111587982832618, "grad_norm": 0.28732019662857056, "learning_rate": 0.00014648048730683507, "loss": 0.6717, "step": 2395 }, { "epoch": 4.120171673819742, "grad_norm": 0.31053388118743896, "learning_rate": 0.0001462147568039977, "loss": 0.654, "step": 2400 }, { "epoch": 4.128755364806867, "grad_norm": 0.29770082235336304, "learning_rate": 0.00014594861068817095, "loss": 0.6641, "step": 2405 }, { "epoch": 4.1373390557939915, "grad_norm": 0.28309738636016846, "learning_rate": 0.00014568205135282795, "loss": 0.6715, "step": 2410 }, { "epoch": 4.145922746781116, "grad_norm": 0.29497766494750977, "learning_rate": 0.00014541508119515808, "loss": 0.6622, "step": 2415 }, { "epoch": 4.154506437768241, "grad_norm": 0.3100745975971222, "learning_rate": 0.00014514770261604522, "loss": 0.6714, "step": 2420 }, { "epoch": 4.163090128755365, "grad_norm": 0.29295915365219116, "learning_rate": 0.00014487991802004623, "loss": 0.6617, "step": 2425 }, { "epoch": 4.171673819742489, "grad_norm": 0.2942890524864197, "learning_rate": 0.0001446117298153693, "loss": 0.6627, "step": 2430 }, { "epoch": 4.180257510729613, "grad_norm": 0.2925000786781311, "learning_rate": 0.0001443431404138524, "loss": 0.6747, "step": 2435 }, { "epoch": 4.1888412017167385, "grad_norm": 0.31883692741394043, "learning_rate": 0.00014407415223094132, "loss": 0.6645, "step": 2440 }, { "epoch": 4.197424892703863, "grad_norm": 0.29616445302963257, "learning_rate": 0.00014380476768566824, "loss": 0.6653, "step": 2445 }, { "epoch": 4.206008583690987, "grad_norm": 0.2958747148513794, "learning_rate": 0.00014353498920062987, "loss": 0.6663, "step": 2450 }, { "epoch": 4.214592274678112, "grad_norm": 0.2944903075695038, "learning_rate": 0.00014326481920196556, "loss": 0.6584, "step": 2455 }, { "epoch": 4.223175965665236, "grad_norm": 0.30497610569000244, "learning_rate": 0.00014299426011933568, "loss": 0.6734, "step": 2460 }, { "epoch": 4.23175965665236, "grad_norm": 0.2795341908931732, "learning_rate": 0.0001427233143858996, "loss": 0.6664, "step": 2465 }, { "epoch": 4.240343347639485, "grad_norm": 0.2952185273170471, "learning_rate": 0.00014245198443829383, "loss": 0.675, "step": 2470 }, { "epoch": 4.24892703862661, "grad_norm": 0.29675596952438354, "learning_rate": 0.0001421802727166103, "loss": 0.6726, "step": 2475 }, { "epoch": 4.257510729613734, "grad_norm": 0.2960766553878784, "learning_rate": 0.0001419081816643741, "loss": 0.6657, "step": 2480 }, { "epoch": 4.266094420600858, "grad_norm": 0.2973078191280365, "learning_rate": 0.00014163571372852177, "loss": 0.6781, "step": 2485 }, { "epoch": 4.274678111587983, "grad_norm": 0.293087363243103, "learning_rate": 0.00014136287135937915, "loss": 0.6715, "step": 2490 }, { "epoch": 4.283261802575107, "grad_norm": 0.30738070607185364, "learning_rate": 0.00014108965701063942, "loss": 0.6692, "step": 2495 }, { "epoch": 4.291845493562231, "grad_norm": 0.29339906573295593, "learning_rate": 0.0001408160731393409, "loss": 0.6632, "step": 2500 }, { "epoch": 4.3004291845493565, "grad_norm": 0.3105657696723938, "learning_rate": 0.00014054212220584525, "loss": 0.6768, "step": 2505 }, { "epoch": 4.309012875536481, "grad_norm": 0.29471009969711304, "learning_rate": 0.00014026780667381498, "loss": 0.6703, "step": 2510 }, { "epoch": 4.317596566523605, "grad_norm": 0.30494722723960876, "learning_rate": 0.0001399931290101915, "loss": 0.6725, "step": 2515 }, { "epoch": 4.32618025751073, "grad_norm": 0.2980051040649414, "learning_rate": 0.00013971809168517298, "loss": 0.6652, "step": 2520 }, { "epoch": 4.334763948497854, "grad_norm": 0.2986336350440979, "learning_rate": 0.00013944269717219198, "loss": 0.6579, "step": 2525 }, { "epoch": 4.343347639484978, "grad_norm": 0.2928684949874878, "learning_rate": 0.00013916694794789325, "loss": 0.6797, "step": 2530 }, { "epoch": 4.3519313304721035, "grad_norm": 0.30945298075675964, "learning_rate": 0.00013889084649211156, "loss": 0.6781, "step": 2535 }, { "epoch": 4.360515021459228, "grad_norm": 0.29915961623191833, "learning_rate": 0.0001386143952878493, "loss": 0.6802, "step": 2540 }, { "epoch": 4.369098712446352, "grad_norm": 0.309627503156662, "learning_rate": 0.0001383375968212542, "loss": 0.6728, "step": 2545 }, { "epoch": 4.377682403433476, "grad_norm": 0.292521595954895, "learning_rate": 0.00013806045358159683, "loss": 0.6739, "step": 2550 }, { "epoch": 4.386266094420601, "grad_norm": 0.30214038491249084, "learning_rate": 0.00013778296806124852, "loss": 0.6821, "step": 2555 }, { "epoch": 4.394849785407725, "grad_norm": 0.30407920479774475, "learning_rate": 0.0001375051427556586, "loss": 0.6723, "step": 2560 }, { "epoch": 4.4034334763948495, "grad_norm": 0.3059447109699249, "learning_rate": 0.00013722698016333218, "loss": 0.6787, "step": 2565 }, { "epoch": 4.412017167381975, "grad_norm": 0.2976439893245697, "learning_rate": 0.00013694848278580763, "loss": 0.6675, "step": 2570 }, { "epoch": 4.420600858369099, "grad_norm": 0.30373451113700867, "learning_rate": 0.00013666965312763405, "loss": 0.6743, "step": 2575 }, { "epoch": 4.429184549356223, "grad_norm": 0.2918217182159424, "learning_rate": 0.00013639049369634876, "loss": 0.6753, "step": 2580 }, { "epoch": 4.437768240343348, "grad_norm": 0.2900952696800232, "learning_rate": 0.00013611100700245478, "loss": 0.6732, "step": 2585 }, { "epoch": 4.446351931330472, "grad_norm": 0.307317852973938, "learning_rate": 0.00013583119555939815, "loss": 0.6825, "step": 2590 }, { "epoch": 4.454935622317596, "grad_norm": 0.298483282327652, "learning_rate": 0.00013555106188354557, "loss": 0.6744, "step": 2595 }, { "epoch": 4.463519313304721, "grad_norm": 0.3024675250053406, "learning_rate": 0.00013527060849416144, "loss": 0.6786, "step": 2600 }, { "epoch": 4.472103004291846, "grad_norm": 0.30732661485671997, "learning_rate": 0.00013498983791338545, "loss": 0.6745, "step": 2605 }, { "epoch": 4.48068669527897, "grad_norm": 0.29272985458374023, "learning_rate": 0.00013470875266620978, "loss": 0.6782, "step": 2610 }, { "epoch": 4.489270386266094, "grad_norm": 0.3057718873023987, "learning_rate": 0.00013442735528045643, "loss": 0.6844, "step": 2615 }, { "epoch": 4.497854077253219, "grad_norm": 0.31071603298187256, "learning_rate": 0.00013414564828675456, "loss": 0.6827, "step": 2620 }, { "epoch": 4.506437768240343, "grad_norm": 0.32284069061279297, "learning_rate": 0.00013386363421851756, "loss": 0.6809, "step": 2625 }, { "epoch": 4.515021459227468, "grad_norm": 0.3142790198326111, "learning_rate": 0.00013358131561192046, "loss": 0.6859, "step": 2630 }, { "epoch": 4.523605150214593, "grad_norm": 0.30643147230148315, "learning_rate": 0.00013329869500587694, "loss": 0.6742, "step": 2635 }, { "epoch": 4.532188841201717, "grad_norm": 0.2984931468963623, "learning_rate": 0.00013301577494201664, "loss": 0.677, "step": 2640 }, { "epoch": 4.540772532188841, "grad_norm": 0.30939051508903503, "learning_rate": 0.00013273255796466223, "loss": 0.6742, "step": 2645 }, { "epoch": 4.549356223175966, "grad_norm": 0.3024001717567444, "learning_rate": 0.00013244904662080653, "loss": 0.6718, "step": 2650 }, { "epoch": 4.55793991416309, "grad_norm": 0.294514536857605, "learning_rate": 0.0001321652434600897, "loss": 0.6787, "step": 2655 }, { "epoch": 4.5665236051502145, "grad_norm": 0.2984825372695923, "learning_rate": 0.00013188115103477606, "loss": 0.6905, "step": 2660 }, { "epoch": 4.575107296137339, "grad_norm": 0.298879474401474, "learning_rate": 0.00013159677189973152, "loss": 0.6882, "step": 2665 }, { "epoch": 4.583690987124464, "grad_norm": 0.30825379490852356, "learning_rate": 0.00013131210861240026, "loss": 0.6752, "step": 2670 }, { "epoch": 4.592274678111588, "grad_norm": 0.3206503093242645, "learning_rate": 0.00013102716373278192, "loss": 0.6827, "step": 2675 }, { "epoch": 4.600858369098712, "grad_norm": 0.3049548268318176, "learning_rate": 0.00013074193982340847, "loss": 0.6871, "step": 2680 }, { "epoch": 4.609442060085837, "grad_norm": 0.29572340846061707, "learning_rate": 0.00013045643944932126, "loss": 0.6796, "step": 2685 }, { "epoch": 4.618025751072961, "grad_norm": 0.2995782792568207, "learning_rate": 0.00013017066517804793, "loss": 0.6827, "step": 2690 }, { "epoch": 4.626609442060086, "grad_norm": 0.31238484382629395, "learning_rate": 0.00012988461957957922, "loss": 0.6755, "step": 2695 }, { "epoch": 4.63519313304721, "grad_norm": 0.29801589250564575, "learning_rate": 0.00012959830522634596, "loss": 0.6825, "step": 2700 }, { "epoch": 4.643776824034335, "grad_norm": 0.318678081035614, "learning_rate": 0.000129311724693196, "loss": 0.6898, "step": 2705 }, { "epoch": 4.652360515021459, "grad_norm": 0.3065158426761627, "learning_rate": 0.00012902488055737093, "loss": 0.6765, "step": 2710 }, { "epoch": 4.660944206008583, "grad_norm": 0.30059394240379333, "learning_rate": 0.00012873777539848283, "loss": 0.6938, "step": 2715 }, { "epoch": 4.669527896995708, "grad_norm": 0.30598220229148865, "learning_rate": 0.00012845041179849128, "loss": 0.686, "step": 2720 }, { "epoch": 4.6781115879828326, "grad_norm": 0.30764105916023254, "learning_rate": 0.00012816279234168008, "loss": 0.6886, "step": 2725 }, { "epoch": 4.686695278969957, "grad_norm": 0.302541583776474, "learning_rate": 0.0001278749196146339, "loss": 0.6848, "step": 2730 }, { "epoch": 4.695278969957082, "grad_norm": 0.2911517322063446, "learning_rate": 0.00012758679620621503, "loss": 0.6859, "step": 2735 }, { "epoch": 4.703862660944206, "grad_norm": 0.2968668043613434, "learning_rate": 0.00012729842470754032, "loss": 0.68, "step": 2740 }, { "epoch": 4.71244635193133, "grad_norm": 0.30458763241767883, "learning_rate": 0.00012700980771195762, "loss": 0.6785, "step": 2745 }, { "epoch": 4.721030042918455, "grad_norm": 0.3035382032394409, "learning_rate": 0.00012672094781502252, "loss": 0.6896, "step": 2750 }, { "epoch": 4.7296137339055795, "grad_norm": 0.2893989682197571, "learning_rate": 0.00012643184761447512, "loss": 0.6786, "step": 2755 }, { "epoch": 4.738197424892704, "grad_norm": 0.3313075006008148, "learning_rate": 0.00012614250971021657, "loss": 0.6859, "step": 2760 }, { "epoch": 4.746781115879829, "grad_norm": 0.3178950250148773, "learning_rate": 0.00012585293670428564, "loss": 0.6855, "step": 2765 }, { "epoch": 4.755364806866953, "grad_norm": 0.28977376222610474, "learning_rate": 0.00012556313120083546, "loss": 0.6761, "step": 2770 }, { "epoch": 4.763948497854077, "grad_norm": 0.30826762318611145, "learning_rate": 0.00012527309580611005, "loss": 0.677, "step": 2775 }, { "epoch": 4.772532188841201, "grad_norm": 0.3121355473995209, "learning_rate": 0.0001249828331284207, "loss": 0.6854, "step": 2780 }, { "epoch": 4.781115879828326, "grad_norm": 0.30394992232322693, "learning_rate": 0.00012469234577812296, "loss": 0.6927, "step": 2785 }, { "epoch": 4.789699570815451, "grad_norm": 0.28112301230430603, "learning_rate": 0.0001244016363675926, "loss": 0.6917, "step": 2790 }, { "epoch": 4.798283261802575, "grad_norm": 0.2926720678806305, "learning_rate": 0.00012411070751120254, "loss": 0.6703, "step": 2795 }, { "epoch": 4.8068669527897, "grad_norm": 0.296567440032959, "learning_rate": 0.00012381956182529918, "loss": 0.6831, "step": 2800 }, { "epoch": 4.815450643776824, "grad_norm": 0.29968711733818054, "learning_rate": 0.00012352820192817877, "loss": 0.68, "step": 2805 }, { "epoch": 4.824034334763948, "grad_norm": 0.31971994042396545, "learning_rate": 0.0001232366304400642, "loss": 0.6936, "step": 2810 }, { "epoch": 4.8326180257510725, "grad_norm": 0.29575350880622864, "learning_rate": 0.000122944849983081, "loss": 0.689, "step": 2815 }, { "epoch": 4.8412017167381975, "grad_norm": 0.33436283469200134, "learning_rate": 0.00012265286318123415, "loss": 0.6905, "step": 2820 }, { "epoch": 4.849785407725322, "grad_norm": 0.27712202072143555, "learning_rate": 0.00012236067266038414, "loss": 0.6856, "step": 2825 }, { "epoch": 4.858369098712446, "grad_norm": 0.32512983679771423, "learning_rate": 0.00012206828104822363, "loss": 0.6906, "step": 2830 }, { "epoch": 4.866952789699571, "grad_norm": 0.3128452003002167, "learning_rate": 0.00012177569097425368, "loss": 0.6814, "step": 2835 }, { "epoch": 4.875536480686695, "grad_norm": 0.28740525245666504, "learning_rate": 0.00012148290506976012, "loss": 0.683, "step": 2840 }, { "epoch": 4.884120171673819, "grad_norm": 0.3094848394393921, "learning_rate": 0.00012118992596778995, "loss": 0.6751, "step": 2845 }, { "epoch": 4.8927038626609445, "grad_norm": 0.28997689485549927, "learning_rate": 0.00012089675630312754, "loss": 0.6918, "step": 2850 }, { "epoch": 4.901287553648069, "grad_norm": 0.29608073830604553, "learning_rate": 0.00012060339871227101, "loss": 0.687, "step": 2855 }, { "epoch": 4.909871244635193, "grad_norm": 0.2982884347438812, "learning_rate": 0.00012030985583340861, "loss": 0.6886, "step": 2860 }, { "epoch": 4.918454935622318, "grad_norm": 0.31985583901405334, "learning_rate": 0.00012001613030639478, "loss": 0.6787, "step": 2865 }, { "epoch": 4.927038626609442, "grad_norm": 0.31187719106674194, "learning_rate": 0.00011972222477272663, "loss": 0.6944, "step": 2870 }, { "epoch": 4.935622317596566, "grad_norm": 0.28702715039253235, "learning_rate": 0.00011942814187552005, "loss": 0.6849, "step": 2875 }, { "epoch": 4.944206008583691, "grad_norm": 0.28225070238113403, "learning_rate": 0.00011913388425948584, "loss": 0.6791, "step": 2880 }, { "epoch": 4.952789699570816, "grad_norm": 0.3197003901004791, "learning_rate": 0.00011883945457090633, "loss": 0.6905, "step": 2885 }, { "epoch": 4.96137339055794, "grad_norm": 0.30299967527389526, "learning_rate": 0.00011854485545761108, "loss": 0.6794, "step": 2890 }, { "epoch": 4.969957081545064, "grad_norm": 0.3089299499988556, "learning_rate": 0.00011825008956895338, "loss": 0.6815, "step": 2895 }, { "epoch": 4.978540772532189, "grad_norm": 0.31284767389297485, "learning_rate": 0.0001179551595557864, "loss": 0.6878, "step": 2900 }, { "epoch": 4.987124463519313, "grad_norm": 0.2950330078601837, "learning_rate": 0.00011766006807043921, "loss": 0.6765, "step": 2905 }, { "epoch": 4.9957081545064375, "grad_norm": 0.31097424030303955, "learning_rate": 0.00011736481776669306, "loss": 0.6928, "step": 2910 }, { "epoch": 4.9991416309012875, "eval_loss": 2.442364454269409, "eval_runtime": 0.3946, "eval_samples_per_second": 15.205, "eval_steps_per_second": 2.534, "step": 2912 }, { "epoch": 5.0042918454935625, "grad_norm": 0.24456113576889038, "learning_rate": 0.00011706941129975741, "loss": 0.6479, "step": 2915 }, { "epoch": 5.012875536480687, "grad_norm": 0.3765704333782196, "learning_rate": 0.00011677385132624621, "loss": 0.5981, "step": 2920 }, { "epoch": 5.021459227467811, "grad_norm": 0.30039721727371216, "learning_rate": 0.00011647814050415378, "loss": 0.5899, "step": 2925 }, { "epoch": 5.030042918454936, "grad_norm": 0.3503376245498657, "learning_rate": 0.00011618228149283117, "loss": 0.586, "step": 2930 }, { "epoch": 5.03862660944206, "grad_norm": 0.315405935049057, "learning_rate": 0.000115886276952962, "loss": 0.5842, "step": 2935 }, { "epoch": 5.047210300429184, "grad_norm": 0.3637129068374634, "learning_rate": 0.00011559012954653865, "loss": 0.5894, "step": 2940 }, { "epoch": 5.055793991416309, "grad_norm": 0.3133433759212494, "learning_rate": 0.00011529384193683838, "loss": 0.5889, "step": 2945 }, { "epoch": 5.064377682403434, "grad_norm": 0.3363387882709503, "learning_rate": 0.00011499741678839928, "loss": 0.5992, "step": 2950 }, { "epoch": 5.072961373390558, "grad_norm": 0.33112606406211853, "learning_rate": 0.00011470085676699627, "loss": 0.584, "step": 2955 }, { "epoch": 5.081545064377682, "grad_norm": 0.33626094460487366, "learning_rate": 0.00011440416453961728, "loss": 0.5907, "step": 2960 }, { "epoch": 5.090128755364807, "grad_norm": 0.32099804282188416, "learning_rate": 0.00011410734277443915, "loss": 0.5875, "step": 2965 }, { "epoch": 5.098712446351931, "grad_norm": 0.33036282658576965, "learning_rate": 0.00011381039414080365, "loss": 0.5883, "step": 2970 }, { "epoch": 5.1072961373390555, "grad_norm": 0.33885952830314636, "learning_rate": 0.00011351332130919348, "loss": 0.5857, "step": 2975 }, { "epoch": 5.115879828326181, "grad_norm": 0.31977617740631104, "learning_rate": 0.00011321612695120832, "loss": 0.5855, "step": 2980 }, { "epoch": 5.124463519313305, "grad_norm": 0.3337278664112091, "learning_rate": 0.00011291881373954065, "loss": 0.5969, "step": 2985 }, { "epoch": 5.133047210300429, "grad_norm": 0.33656007051467896, "learning_rate": 0.00011262138434795191, "loss": 0.5811, "step": 2990 }, { "epoch": 5.141630901287554, "grad_norm": 0.34739845991134644, "learning_rate": 0.00011232384145124831, "loss": 0.5932, "step": 2995 }, { "epoch": 5.150214592274678, "grad_norm": 0.33286526799201965, "learning_rate": 0.0001120261877252568, "loss": 0.5983, "step": 3000 }, { "epoch": 5.1587982832618025, "grad_norm": 0.3316696584224701, "learning_rate": 0.00011172842584680107, "loss": 0.5976, "step": 3005 }, { "epoch": 5.167381974248927, "grad_norm": 0.3135989308357239, "learning_rate": 0.00011143055849367738, "loss": 0.5978, "step": 3010 }, { "epoch": 5.175965665236052, "grad_norm": 0.3531875014305115, "learning_rate": 0.00011113258834463063, "loss": 0.5965, "step": 3015 }, { "epoch": 5.184549356223176, "grad_norm": 0.3395566940307617, "learning_rate": 0.00011083451807933008, "loss": 0.5933, "step": 3020 }, { "epoch": 5.1931330472103, "grad_norm": 0.3231445252895355, "learning_rate": 0.00011053635037834532, "loss": 0.6071, "step": 3025 }, { "epoch": 5.201716738197425, "grad_norm": 0.33721092343330383, "learning_rate": 0.00011023808792312227, "loss": 0.6049, "step": 3030 }, { "epoch": 5.210300429184549, "grad_norm": 0.32114890217781067, "learning_rate": 0.00010993973339595896, "loss": 0.6038, "step": 3035 }, { "epoch": 5.218884120171674, "grad_norm": 0.3454916477203369, "learning_rate": 0.00010964128947998142, "loss": 0.5998, "step": 3040 }, { "epoch": 5.227467811158799, "grad_norm": 0.31621354818344116, "learning_rate": 0.00010934275885911956, "loss": 0.5977, "step": 3045 }, { "epoch": 5.236051502145923, "grad_norm": 0.3424486517906189, "learning_rate": 0.00010904414421808303, "loss": 0.5923, "step": 3050 }, { "epoch": 5.244635193133047, "grad_norm": 0.3184787333011627, "learning_rate": 0.00010874544824233705, "loss": 0.5986, "step": 3055 }, { "epoch": 5.253218884120171, "grad_norm": 0.3448573052883148, "learning_rate": 0.00010844667361807842, "loss": 0.5931, "step": 3060 }, { "epoch": 5.261802575107296, "grad_norm": 0.314815878868103, "learning_rate": 0.00010814782303221105, "loss": 0.6008, "step": 3065 }, { "epoch": 5.2703862660944205, "grad_norm": 0.3566027581691742, "learning_rate": 0.00010784889917232206, "loss": 0.585, "step": 3070 }, { "epoch": 5.278969957081545, "grad_norm": 0.31585681438446045, "learning_rate": 0.0001075499047266576, "loss": 0.5927, "step": 3075 }, { "epoch": 5.28755364806867, "grad_norm": 0.33874404430389404, "learning_rate": 0.00010725084238409848, "loss": 0.6047, "step": 3080 }, { "epoch": 5.296137339055794, "grad_norm": 0.3274739682674408, "learning_rate": 0.00010695171483413619, "loss": 0.5919, "step": 3085 }, { "epoch": 5.304721030042918, "grad_norm": 0.3326583802700043, "learning_rate": 0.00010665252476684864, "loss": 0.5982, "step": 3090 }, { "epoch": 5.313304721030043, "grad_norm": 0.34962788224220276, "learning_rate": 0.00010635327487287591, "loss": 0.5999, "step": 3095 }, { "epoch": 5.3218884120171674, "grad_norm": 0.33242544531822205, "learning_rate": 0.00010605396784339612, "loss": 0.6037, "step": 3100 }, { "epoch": 5.330472103004292, "grad_norm": 0.38023149967193604, "learning_rate": 0.00010575460637010128, "loss": 0.6068, "step": 3105 }, { "epoch": 5.339055793991417, "grad_norm": 0.3179317116737366, "learning_rate": 0.00010545519314517291, "loss": 0.5993, "step": 3110 }, { "epoch": 5.347639484978541, "grad_norm": 0.343841016292572, "learning_rate": 0.00010515573086125805, "loss": 0.6029, "step": 3115 }, { "epoch": 5.356223175965665, "grad_norm": 0.3552517890930176, "learning_rate": 0.00010485622221144484, "loss": 0.599, "step": 3120 }, { "epoch": 5.364806866952789, "grad_norm": 0.3441089391708374, "learning_rate": 0.00010455666988923842, "loss": 0.598, "step": 3125 }, { "epoch": 5.373390557939914, "grad_norm": 0.35372617840766907, "learning_rate": 0.00010425707658853672, "loss": 0.597, "step": 3130 }, { "epoch": 5.381974248927039, "grad_norm": 0.33165809512138367, "learning_rate": 0.00010395744500360612, "loss": 0.6077, "step": 3135 }, { "epoch": 5.390557939914163, "grad_norm": 0.3485681414604187, "learning_rate": 0.00010365777782905735, "loss": 0.5956, "step": 3140 }, { "epoch": 5.399141630901288, "grad_norm": 0.3294559717178345, "learning_rate": 0.00010335807775982116, "loss": 0.6076, "step": 3145 }, { "epoch": 5.407725321888412, "grad_norm": 0.3176850378513336, "learning_rate": 0.00010305834749112421, "loss": 0.594, "step": 3150 }, { "epoch": 5.416309012875536, "grad_norm": 0.3580548167228699, "learning_rate": 0.00010275858971846463, "loss": 0.6029, "step": 3155 }, { "epoch": 5.424892703862661, "grad_norm": 0.32710379362106323, "learning_rate": 0.00010245880713758793, "loss": 0.6063, "step": 3160 }, { "epoch": 5.4334763948497855, "grad_norm": 0.37140390276908875, "learning_rate": 0.00010215900244446279, "loss": 0.6108, "step": 3165 }, { "epoch": 5.44206008583691, "grad_norm": 0.3271103501319885, "learning_rate": 0.00010185917833525669, "loss": 0.6086, "step": 3170 }, { "epoch": 5.450643776824034, "grad_norm": 0.3177226781845093, "learning_rate": 0.00010155933750631172, "loss": 0.5976, "step": 3175 }, { "epoch": 5.459227467811159, "grad_norm": 0.33881354331970215, "learning_rate": 0.00010125948265412033, "loss": 0.5975, "step": 3180 }, { "epoch": 5.467811158798283, "grad_norm": 0.35869210958480835, "learning_rate": 0.0001009596164753011, "loss": 0.6032, "step": 3185 }, { "epoch": 5.476394849785407, "grad_norm": 0.34474968910217285, "learning_rate": 0.00010065974166657448, "loss": 0.6014, "step": 3190 }, { "epoch": 5.484978540772532, "grad_norm": 0.3329688012599945, "learning_rate": 0.00010035986092473847, "loss": 0.6042, "step": 3195 }, { "epoch": 5.493562231759657, "grad_norm": 0.34499508142471313, "learning_rate": 0.00010005997694664451, "loss": 0.5998, "step": 3200 }, { "epoch": 5.502145922746781, "grad_norm": 0.3572762608528137, "learning_rate": 9.976009242917307e-05, "loss": 0.6031, "step": 3205 }, { "epoch": 5.510729613733906, "grad_norm": 0.33951064944267273, "learning_rate": 9.946021006920959e-05, "loss": 0.6046, "step": 3210 }, { "epoch": 5.51931330472103, "grad_norm": 0.3542894423007965, "learning_rate": 9.91603325636199e-05, "loss": 0.6063, "step": 3215 }, { "epoch": 5.527896995708154, "grad_norm": 0.3496350646018982, "learning_rate": 9.886046260922634e-05, "loss": 0.6059, "step": 3220 }, { "epoch": 5.536480686695279, "grad_norm": 0.36511626839637756, "learning_rate": 9.856060290278337e-05, "loss": 0.6011, "step": 3225 }, { "epoch": 5.545064377682404, "grad_norm": 0.3340092897415161, "learning_rate": 9.826075614095311e-05, "loss": 0.6098, "step": 3230 }, { "epoch": 5.553648068669528, "grad_norm": 0.35916590690612793, "learning_rate": 9.796092502028145e-05, "loss": 0.6047, "step": 3235 }, { "epoch": 5.562231759656653, "grad_norm": 0.3391835689544678, "learning_rate": 9.766111223717352e-05, "loss": 0.61, "step": 3240 }, { "epoch": 5.570815450643777, "grad_norm": 0.3280404508113861, "learning_rate": 9.736132048786954e-05, "loss": 0.6088, "step": 3245 }, { "epoch": 5.579399141630901, "grad_norm": 0.31755104660987854, "learning_rate": 9.706155246842062e-05, "loss": 0.6051, "step": 3250 }, { "epoch": 5.587982832618025, "grad_norm": 0.33683615922927856, "learning_rate": 9.676181087466444e-05, "loss": 0.6071, "step": 3255 }, { "epoch": 5.5965665236051505, "grad_norm": 0.34338319301605225, "learning_rate": 9.646209840220098e-05, "loss": 0.6083, "step": 3260 }, { "epoch": 5.605150214592275, "grad_norm": 0.35656723380088806, "learning_rate": 9.616241774636845e-05, "loss": 0.6032, "step": 3265 }, { "epoch": 5.613733905579399, "grad_norm": 0.33608657121658325, "learning_rate": 9.586277160221884e-05, "loss": 0.6176, "step": 3270 }, { "epoch": 5.622317596566524, "grad_norm": 0.35035207867622375, "learning_rate": 9.556316266449377e-05, "loss": 0.6037, "step": 3275 }, { "epoch": 5.630901287553648, "grad_norm": 0.3345491290092468, "learning_rate": 9.526359362760032e-05, "loss": 0.6022, "step": 3280 }, { "epoch": 5.639484978540772, "grad_norm": 0.3335653841495514, "learning_rate": 9.496406718558665e-05, "loss": 0.6091, "step": 3285 }, { "epoch": 5.6480686695278965, "grad_norm": 0.3456047773361206, "learning_rate": 9.466458603211796e-05, "loss": 0.615, "step": 3290 }, { "epoch": 5.656652360515022, "grad_norm": 0.3243827223777771, "learning_rate": 9.436515286045214e-05, "loss": 0.6058, "step": 3295 }, { "epoch": 5.665236051502146, "grad_norm": 0.329574853181839, "learning_rate": 9.406577036341548e-05, "loss": 0.6054, "step": 3300 }, { "epoch": 5.67381974248927, "grad_norm": 0.3418329954147339, "learning_rate": 9.376644123337867e-05, "loss": 0.6098, "step": 3305 }, { "epoch": 5.682403433476395, "grad_norm": 0.36660462617874146, "learning_rate": 9.346716816223245e-05, "loss": 0.6187, "step": 3310 }, { "epoch": 5.690987124463519, "grad_norm": 0.3241427540779114, "learning_rate": 9.316795384136333e-05, "loss": 0.6121, "step": 3315 }, { "epoch": 5.6995708154506435, "grad_norm": 0.36070680618286133, "learning_rate": 9.286880096162956e-05, "loss": 0.6095, "step": 3320 }, { "epoch": 5.708154506437769, "grad_norm": 0.33510082960128784, "learning_rate": 9.256971221333685e-05, "loss": 0.6126, "step": 3325 }, { "epoch": 5.716738197424893, "grad_norm": 0.3640751838684082, "learning_rate": 9.227069028621406e-05, "loss": 0.6199, "step": 3330 }, { "epoch": 5.725321888412017, "grad_norm": 0.34373095631599426, "learning_rate": 9.197173786938926e-05, "loss": 0.6091, "step": 3335 }, { "epoch": 5.733905579399142, "grad_norm": 0.33689820766448975, "learning_rate": 9.167285765136533e-05, "loss": 0.6032, "step": 3340 }, { "epoch": 5.742489270386266, "grad_norm": 0.33981624245643616, "learning_rate": 9.137405231999593e-05, "loss": 0.6069, "step": 3345 }, { "epoch": 5.75107296137339, "grad_norm": 0.34156525135040283, "learning_rate": 9.107532456246114e-05, "loss": 0.6006, "step": 3350 }, { "epoch": 5.7596566523605155, "grad_norm": 0.34921032190322876, "learning_rate": 9.077667706524348e-05, "loss": 0.6165, "step": 3355 }, { "epoch": 5.76824034334764, "grad_norm": 0.3511927127838135, "learning_rate": 9.047811251410376e-05, "loss": 0.6149, "step": 3360 }, { "epoch": 5.776824034334764, "grad_norm": 0.3529740571975708, "learning_rate": 9.01796335940567e-05, "loss": 0.6088, "step": 3365 }, { "epoch": 5.785407725321888, "grad_norm": 0.33018723130226135, "learning_rate": 8.9881242989347e-05, "loss": 0.6089, "step": 3370 }, { "epoch": 5.793991416309013, "grad_norm": 0.33075249195098877, "learning_rate": 8.95829433834252e-05, "loss": 0.6107, "step": 3375 }, { "epoch": 5.802575107296137, "grad_norm": 0.35009533166885376, "learning_rate": 8.928473745892339e-05, "loss": 0.6092, "step": 3380 }, { "epoch": 5.8111587982832615, "grad_norm": 0.337410569190979, "learning_rate": 8.898662789763115e-05, "loss": 0.6049, "step": 3385 }, { "epoch": 5.819742489270387, "grad_norm": 0.332481324672699, "learning_rate": 8.868861738047158e-05, "loss": 0.625, "step": 3390 }, { "epoch": 5.828326180257511, "grad_norm": 0.34015268087387085, "learning_rate": 8.839070858747697e-05, "loss": 0.615, "step": 3395 }, { "epoch": 5.836909871244635, "grad_norm": 0.35073375701904297, "learning_rate": 8.809290419776488e-05, "loss": 0.6038, "step": 3400 }, { "epoch": 5.845493562231759, "grad_norm": 0.3357756733894348, "learning_rate": 8.779520688951383e-05, "loss": 0.6077, "step": 3405 }, { "epoch": 5.854077253218884, "grad_norm": 0.33854493498802185, "learning_rate": 8.749761933993945e-05, "loss": 0.6058, "step": 3410 }, { "epoch": 5.8626609442060085, "grad_norm": 0.3343227505683899, "learning_rate": 8.720014422527034e-05, "loss": 0.6113, "step": 3415 }, { "epoch": 5.871244635193133, "grad_norm": 0.35862088203430176, "learning_rate": 8.690278422072384e-05, "loss": 0.606, "step": 3420 }, { "epoch": 5.879828326180258, "grad_norm": 0.37902429699897766, "learning_rate": 8.660554200048215e-05, "loss": 0.6108, "step": 3425 }, { "epoch": 5.888412017167382, "grad_norm": 0.3685941994190216, "learning_rate": 8.630842023766831e-05, "loss": 0.6138, "step": 3430 }, { "epoch": 5.896995708154506, "grad_norm": 0.3437183201313019, "learning_rate": 8.601142160432193e-05, "loss": 0.6156, "step": 3435 }, { "epoch": 5.905579399141631, "grad_norm": 0.34756121039390564, "learning_rate": 8.571454877137539e-05, "loss": 0.6076, "step": 3440 }, { "epoch": 5.914163090128755, "grad_norm": 0.33176571130752563, "learning_rate": 8.541780440862977e-05, "loss": 0.6065, "step": 3445 }, { "epoch": 5.92274678111588, "grad_norm": 0.3526177704334259, "learning_rate": 8.512119118473067e-05, "loss": 0.6125, "step": 3450 }, { "epoch": 5.931330472103005, "grad_norm": 0.3659914433956146, "learning_rate": 8.482471176714454e-05, "loss": 0.6095, "step": 3455 }, { "epoch": 5.939914163090129, "grad_norm": 0.34757936000823975, "learning_rate": 8.45283688221344e-05, "loss": 0.6134, "step": 3460 }, { "epoch": 5.948497854077253, "grad_norm": 0.34589987993240356, "learning_rate": 8.423216501473585e-05, "loss": 0.6074, "step": 3465 }, { "epoch": 5.957081545064378, "grad_norm": 0.34556832909584045, "learning_rate": 8.393610300873345e-05, "loss": 0.609, "step": 3470 }, { "epoch": 5.965665236051502, "grad_norm": 0.3431447148323059, "learning_rate": 8.364018546663634e-05, "loss": 0.6157, "step": 3475 }, { "epoch": 5.9742489270386265, "grad_norm": 0.35311102867126465, "learning_rate": 8.334441504965455e-05, "loss": 0.6143, "step": 3480 }, { "epoch": 5.982832618025751, "grad_norm": 0.3422008156776428, "learning_rate": 8.304879441767504e-05, "loss": 0.6046, "step": 3485 }, { "epoch": 5.991416309012876, "grad_norm": 0.3697339594364166, "learning_rate": 8.275332622923769e-05, "loss": 0.6141, "step": 3490 }, { "epoch": 6.0, "grad_norm": 0.332302987575531, "learning_rate": 8.245801314151139e-05, "loss": 0.6102, "step": 3495 }, { "epoch": 6.0, "eval_loss": 2.7089035511016846, "eval_runtime": 0.3939, "eval_samples_per_second": 15.232, "eval_steps_per_second": 2.539, "step": 3495 }, { "epoch": 6.008583690987124, "grad_norm": 0.3818419575691223, "learning_rate": 8.216285781027036e-05, "loss": 0.5334, "step": 3500 }, { "epoch": 6.017167381974249, "grad_norm": 0.35173356533050537, "learning_rate": 8.186786288986992e-05, "loss": 0.5409, "step": 3505 }, { "epoch": 6.0257510729613735, "grad_norm": 0.348001092672348, "learning_rate": 8.157303103322296e-05, "loss": 0.5294, "step": 3510 }, { "epoch": 6.034334763948498, "grad_norm": 0.38012921810150146, "learning_rate": 8.127836489177584e-05, "loss": 0.5289, "step": 3515 }, { "epoch": 6.042918454935623, "grad_norm": 0.3508910536766052, "learning_rate": 8.098386711548458e-05, "loss": 0.5375, "step": 3520 }, { "epoch": 6.051502145922747, "grad_norm": 0.3329296410083771, "learning_rate": 8.068954035279121e-05, "loss": 0.5298, "step": 3525 }, { "epoch": 6.060085836909871, "grad_norm": 0.3630905747413635, "learning_rate": 8.039538725059976e-05, "loss": 0.5349, "step": 3530 }, { "epoch": 6.068669527896995, "grad_norm": 0.32074230909347534, "learning_rate": 8.010141045425244e-05, "loss": 0.5376, "step": 3535 }, { "epoch": 6.07725321888412, "grad_norm": 0.3476736545562744, "learning_rate": 7.980761260750607e-05, "loss": 0.5279, "step": 3540 }, { "epoch": 6.085836909871245, "grad_norm": 0.340426504611969, "learning_rate": 7.951399635250806e-05, "loss": 0.5323, "step": 3545 }, { "epoch": 6.094420600858369, "grad_norm": 0.3367501497268677, "learning_rate": 7.922056432977275e-05, "loss": 0.5486, "step": 3550 }, { "epoch": 6.103004291845494, "grad_norm": 0.3827115595340729, "learning_rate": 7.892731917815774e-05, "loss": 0.528, "step": 3555 }, { "epoch": 6.111587982832618, "grad_norm": 0.33026212453842163, "learning_rate": 7.863426353484002e-05, "loss": 0.5303, "step": 3560 }, { "epoch": 6.120171673819742, "grad_norm": 0.3674776256084442, "learning_rate": 7.834140003529238e-05, "loss": 0.5355, "step": 3565 }, { "epoch": 6.128755364806867, "grad_norm": 0.3238033056259155, "learning_rate": 7.804873131325954e-05, "loss": 0.5382, "step": 3570 }, { "epoch": 6.1373390557939915, "grad_norm": 0.3594464063644409, "learning_rate": 7.775626000073463e-05, "loss": 0.5319, "step": 3575 }, { "epoch": 6.145922746781116, "grad_norm": 0.3666881322860718, "learning_rate": 7.74639887279355e-05, "loss": 0.5387, "step": 3580 }, { "epoch": 6.154506437768241, "grad_norm": 0.34055057168006897, "learning_rate": 7.7171920123281e-05, "loss": 0.5296, "step": 3585 }, { "epoch": 6.163090128755365, "grad_norm": 0.3717866837978363, "learning_rate": 7.688005681336729e-05, "loss": 0.5371, "step": 3590 }, { "epoch": 6.171673819742489, "grad_norm": 0.3607926368713379, "learning_rate": 7.658840142294444e-05, "loss": 0.5446, "step": 3595 }, { "epoch": 6.180257510729613, "grad_norm": 0.3458268940448761, "learning_rate": 7.629695657489257e-05, "loss": 0.5393, "step": 3600 }, { "epoch": 6.1888412017167385, "grad_norm": 0.3347238600254059, "learning_rate": 7.600572489019842e-05, "loss": 0.5254, "step": 3605 }, { "epoch": 6.197424892703863, "grad_norm": 0.3612549901008606, "learning_rate": 7.571470898793173e-05, "loss": 0.5275, "step": 3610 }, { "epoch": 6.206008583690987, "grad_norm": 0.3335767686367035, "learning_rate": 7.542391148522168e-05, "loss": 0.5304, "step": 3615 }, { "epoch": 6.214592274678112, "grad_norm": 0.3704369068145752, "learning_rate": 7.513333499723343e-05, "loss": 0.5404, "step": 3620 }, { "epoch": 6.223175965665236, "grad_norm": 0.3659282922744751, "learning_rate": 7.484298213714442e-05, "loss": 0.5357, "step": 3625 }, { "epoch": 6.23175965665236, "grad_norm": 0.34985628724098206, "learning_rate": 7.455285551612105e-05, "loss": 0.5411, "step": 3630 }, { "epoch": 6.240343347639485, "grad_norm": 0.3511213958263397, "learning_rate": 7.426295774329524e-05, "loss": 0.5361, "step": 3635 }, { "epoch": 6.24892703862661, "grad_norm": 0.36643365025520325, "learning_rate": 7.397329142574063e-05, "loss": 0.5256, "step": 3640 }, { "epoch": 6.257510729613734, "grad_norm": 0.3581530451774597, "learning_rate": 7.36838591684495e-05, "loss": 0.5364, "step": 3645 }, { "epoch": 6.266094420600858, "grad_norm": 0.3588225245475769, "learning_rate": 7.339466357430928e-05, "loss": 0.5433, "step": 3650 }, { "epoch": 6.274678111587983, "grad_norm": 0.3423435389995575, "learning_rate": 7.310570724407892e-05, "loss": 0.5481, "step": 3655 }, { "epoch": 6.283261802575107, "grad_norm": 0.37222719192504883, "learning_rate": 7.281699277636572e-05, "loss": 0.5418, "step": 3660 }, { "epoch": 6.291845493562231, "grad_norm": 0.37203842401504517, "learning_rate": 7.252852276760193e-05, "loss": 0.5401, "step": 3665 }, { "epoch": 6.3004291845493565, "grad_norm": 0.3703368306159973, "learning_rate": 7.224029981202122e-05, "loss": 0.5381, "step": 3670 }, { "epoch": 6.309012875536481, "grad_norm": 0.36878761649131775, "learning_rate": 7.195232650163575e-05, "loss": 0.5257, "step": 3675 }, { "epoch": 6.317596566523605, "grad_norm": 0.3529357314109802, "learning_rate": 7.16646054262123e-05, "loss": 0.542, "step": 3680 }, { "epoch": 6.32618025751073, "grad_norm": 0.3673364520072937, "learning_rate": 7.137713917324945e-05, "loss": 0.5351, "step": 3685 }, { "epoch": 6.334763948497854, "grad_norm": 0.3675953149795532, "learning_rate": 7.108993032795418e-05, "loss": 0.5455, "step": 3690 }, { "epoch": 6.343347639484978, "grad_norm": 0.3677637279033661, "learning_rate": 7.080298147321844e-05, "loss": 0.5352, "step": 3695 }, { "epoch": 6.3519313304721035, "grad_norm": 0.36320698261260986, "learning_rate": 7.051629518959614e-05, "loss": 0.5445, "step": 3700 }, { "epoch": 6.360515021459228, "grad_norm": 0.36490681767463684, "learning_rate": 7.022987405527997e-05, "loss": 0.5379, "step": 3705 }, { "epoch": 6.369098712446352, "grad_norm": 0.37636715173721313, "learning_rate": 6.994372064607788e-05, "loss": 0.5416, "step": 3710 }, { "epoch": 6.377682403433476, "grad_norm": 0.3601493239402771, "learning_rate": 6.96578375353903e-05, "loss": 0.544, "step": 3715 }, { "epoch": 6.386266094420601, "grad_norm": 0.3615502715110779, "learning_rate": 6.93722272941869e-05, "loss": 0.5368, "step": 3720 }, { "epoch": 6.394849785407725, "grad_norm": 0.3495427668094635, "learning_rate": 6.908689249098321e-05, "loss": 0.5414, "step": 3725 }, { "epoch": 6.4034334763948495, "grad_norm": 0.36156705021858215, "learning_rate": 6.880183569181795e-05, "loss": 0.5379, "step": 3730 }, { "epoch": 6.412017167381975, "grad_norm": 0.38569867610931396, "learning_rate": 6.85170594602296e-05, "loss": 0.5445, "step": 3735 }, { "epoch": 6.420600858369099, "grad_norm": 0.37031883001327515, "learning_rate": 6.823256635723343e-05, "loss": 0.55, "step": 3740 }, { "epoch": 6.429184549356223, "grad_norm": 0.33991584181785583, "learning_rate": 6.794835894129865e-05, "loss": 0.5427, "step": 3745 }, { "epoch": 6.437768240343348, "grad_norm": 0.35758277773857117, "learning_rate": 6.766443976832517e-05, "loss": 0.5393, "step": 3750 }, { "epoch": 6.446351931330472, "grad_norm": 0.37792807817459106, "learning_rate": 6.738081139162072e-05, "loss": 0.5466, "step": 3755 }, { "epoch": 6.454935622317596, "grad_norm": 0.3743073046207428, "learning_rate": 6.709747636187789e-05, "loss": 0.539, "step": 3760 }, { "epoch": 6.463519313304721, "grad_norm": 0.397246390581131, "learning_rate": 6.68144372271512e-05, "loss": 0.5417, "step": 3765 }, { "epoch": 6.472103004291846, "grad_norm": 0.3632808029651642, "learning_rate": 6.653169653283406e-05, "loss": 0.5403, "step": 3770 }, { "epoch": 6.48068669527897, "grad_norm": 0.3583681583404541, "learning_rate": 6.624925682163614e-05, "loss": 0.5462, "step": 3775 }, { "epoch": 6.489270386266094, "grad_norm": 0.3529057502746582, "learning_rate": 6.59671206335602e-05, "loss": 0.5425, "step": 3780 }, { "epoch": 6.497854077253219, "grad_norm": 0.3624202311038971, "learning_rate": 6.568529050587953e-05, "loss": 0.5462, "step": 3785 }, { "epoch": 6.506437768240343, "grad_norm": 0.3762538433074951, "learning_rate": 6.540376897311489e-05, "loss": 0.538, "step": 3790 }, { "epoch": 6.515021459227468, "grad_norm": 0.37663355469703674, "learning_rate": 6.512255856701177e-05, "loss": 0.5432, "step": 3795 }, { "epoch": 6.523605150214593, "grad_norm": 0.37055134773254395, "learning_rate": 6.484166181651785e-05, "loss": 0.5405, "step": 3800 }, { "epoch": 6.532188841201717, "grad_norm": 0.3634713590145111, "learning_rate": 6.456108124775999e-05, "loss": 0.5442, "step": 3805 }, { "epoch": 6.540772532188841, "grad_norm": 0.3575882017612457, "learning_rate": 6.428081938402149e-05, "loss": 0.5395, "step": 3810 }, { "epoch": 6.549356223175966, "grad_norm": 0.3856394290924072, "learning_rate": 6.400087874571973e-05, "loss": 0.5417, "step": 3815 }, { "epoch": 6.55793991416309, "grad_norm": 0.3542211651802063, "learning_rate": 6.372126185038313e-05, "loss": 0.5337, "step": 3820 }, { "epoch": 6.5665236051502145, "grad_norm": 0.3997708559036255, "learning_rate": 6.344197121262868e-05, "loss": 0.5431, "step": 3825 }, { "epoch": 6.575107296137339, "grad_norm": 0.3704608082771301, "learning_rate": 6.316300934413935e-05, "loss": 0.5356, "step": 3830 }, { "epoch": 6.583690987124464, "grad_norm": 0.3824236989021301, "learning_rate": 6.288437875364141e-05, "loss": 0.5406, "step": 3835 }, { "epoch": 6.592274678111588, "grad_norm": 0.3561914563179016, "learning_rate": 6.260608194688206e-05, "loss": 0.5405, "step": 3840 }, { "epoch": 6.600858369098712, "grad_norm": 0.3756065368652344, "learning_rate": 6.232812142660658e-05, "loss": 0.5365, "step": 3845 }, { "epoch": 6.609442060085837, "grad_norm": 0.3645598292350769, "learning_rate": 6.205049969253605e-05, "loss": 0.5358, "step": 3850 }, { "epoch": 6.618025751072961, "grad_norm": 0.37949660420417786, "learning_rate": 6.17732192413449e-05, "loss": 0.5452, "step": 3855 }, { "epoch": 6.626609442060086, "grad_norm": 0.36608970165252686, "learning_rate": 6.149628256663827e-05, "loss": 0.545, "step": 3860 }, { "epoch": 6.63519313304721, "grad_norm": 0.3485977351665497, "learning_rate": 6.121969215892972e-05, "loss": 0.5445, "step": 3865 }, { "epoch": 6.643776824034335, "grad_norm": 0.38575315475463867, "learning_rate": 6.0943450505618917e-05, "loss": 0.5448, "step": 3870 }, { "epoch": 6.652360515021459, "grad_norm": 0.3650740087032318, "learning_rate": 6.066756009096896e-05, "loss": 0.5489, "step": 3875 }, { "epoch": 6.660944206008583, "grad_norm": 0.3552764058113098, "learning_rate": 6.039202339608432e-05, "loss": 0.5461, "step": 3880 }, { "epoch": 6.669527896995708, "grad_norm": 0.3777913749217987, "learning_rate": 6.01168428988885e-05, "loss": 0.5437, "step": 3885 }, { "epoch": 6.6781115879828326, "grad_norm": 0.37254467606544495, "learning_rate": 5.9842021074101605e-05, "loss": 0.5351, "step": 3890 }, { "epoch": 6.686695278969957, "grad_norm": 0.36322537064552307, "learning_rate": 5.956756039321825e-05, "loss": 0.5503, "step": 3895 }, { "epoch": 6.695278969957082, "grad_norm": 0.3798597753047943, "learning_rate": 5.929346332448511e-05, "loss": 0.5336, "step": 3900 }, { "epoch": 6.703862660944206, "grad_norm": 0.3622066080570221, "learning_rate": 5.901973233287901e-05, "loss": 0.5472, "step": 3905 }, { "epoch": 6.71244635193133, "grad_norm": 0.37123680114746094, "learning_rate": 5.874636988008457e-05, "loss": 0.5376, "step": 3910 }, { "epoch": 6.721030042918455, "grad_norm": 0.3789604902267456, "learning_rate": 5.847337842447209e-05, "loss": 0.5472, "step": 3915 }, { "epoch": 6.7296137339055795, "grad_norm": 0.3710649311542511, "learning_rate": 5.820076042107545e-05, "loss": 0.5459, "step": 3920 }, { "epoch": 6.738197424892704, "grad_norm": 0.41028717160224915, "learning_rate": 5.792851832157014e-05, "loss": 0.5415, "step": 3925 }, { "epoch": 6.746781115879829, "grad_norm": 0.3710199296474457, "learning_rate": 5.765665457425102e-05, "loss": 0.5376, "step": 3930 }, { "epoch": 6.755364806866953, "grad_norm": 0.37828171253204346, "learning_rate": 5.7385171624010346e-05, "loss": 0.5474, "step": 3935 }, { "epoch": 6.763948497854077, "grad_norm": 0.35286852717399597, "learning_rate": 5.711407191231602e-05, "loss": 0.5435, "step": 3940 }, { "epoch": 6.772532188841201, "grad_norm": 0.39667871594429016, "learning_rate": 5.684335787718932e-05, "loss": 0.5471, "step": 3945 }, { "epoch": 6.781115879828326, "grad_norm": 0.3569738268852234, "learning_rate": 5.657303195318311e-05, "loss": 0.5362, "step": 3950 }, { "epoch": 6.789699570815451, "grad_norm": 0.3528185784816742, "learning_rate": 5.630309657135997e-05, "loss": 0.5383, "step": 3955 }, { "epoch": 6.798283261802575, "grad_norm": 0.3892223834991455, "learning_rate": 5.6033554159270294e-05, "loss": 0.5446, "step": 3960 }, { "epoch": 6.8068669527897, "grad_norm": 0.3695877492427826, "learning_rate": 5.576440714093046e-05, "loss": 0.5488, "step": 3965 }, { "epoch": 6.815450643776824, "grad_norm": 0.3762911856174469, "learning_rate": 5.549565793680105e-05, "loss": 0.5398, "step": 3970 }, { "epoch": 6.824034334763948, "grad_norm": 0.36472398042678833, "learning_rate": 5.522730896376506e-05, "loss": 0.5457, "step": 3975 }, { "epoch": 6.8326180257510725, "grad_norm": 0.37586814165115356, "learning_rate": 5.495936263510617e-05, "loss": 0.5396, "step": 3980 }, { "epoch": 6.8412017167381975, "grad_norm": 0.37699511647224426, "learning_rate": 5.4691821360487086e-05, "loss": 0.5394, "step": 3985 }, { "epoch": 6.849785407725322, "grad_norm": 0.3839593529701233, "learning_rate": 5.4424687545927776e-05, "loss": 0.5429, "step": 3990 }, { "epoch": 6.858369098712446, "grad_norm": 0.3746870756149292, "learning_rate": 5.415796359378393e-05, "loss": 0.538, "step": 3995 }, { "epoch": 6.866952789699571, "grad_norm": 0.36862820386886597, "learning_rate": 5.389165190272527e-05, "loss": 0.5433, "step": 4000 }, { "epoch": 6.875536480686695, "grad_norm": 0.3722948729991913, "learning_rate": 5.362575486771414e-05, "loss": 0.5533, "step": 4005 }, { "epoch": 6.884120171673819, "grad_norm": 0.3654981255531311, "learning_rate": 5.3360274879983654e-05, "loss": 0.5345, "step": 4010 }, { "epoch": 6.8927038626609445, "grad_norm": 0.35018405318260193, "learning_rate": 5.3095214327016474e-05, "loss": 0.5484, "step": 4015 }, { "epoch": 6.901287553648069, "grad_norm": 0.37176111340522766, "learning_rate": 5.283057559252341e-05, "loss": 0.5422, "step": 4020 }, { "epoch": 6.909871244635193, "grad_norm": 0.37949976325035095, "learning_rate": 5.256636105642154e-05, "loss": 0.5501, "step": 4025 }, { "epoch": 6.918454935622318, "grad_norm": 0.35853028297424316, "learning_rate": 5.2302573094813266e-05, "loss": 0.5427, "step": 4030 }, { "epoch": 6.927038626609442, "grad_norm": 0.370491623878479, "learning_rate": 5.2039214079964836e-05, "loss": 0.5426, "step": 4035 }, { "epoch": 6.935622317596566, "grad_norm": 0.3726717233657837, "learning_rate": 5.177628638028472e-05, "loss": 0.5447, "step": 4040 }, { "epoch": 6.944206008583691, "grad_norm": 0.37951403856277466, "learning_rate": 5.1513792360302696e-05, "loss": 0.5369, "step": 4045 }, { "epoch": 6.952789699570816, "grad_norm": 0.3583022654056549, "learning_rate": 5.12517343806485e-05, "loss": 0.5586, "step": 4050 }, { "epoch": 6.96137339055794, "grad_norm": 0.37766262888908386, "learning_rate": 5.099011479803025e-05, "loss": 0.537, "step": 4055 }, { "epoch": 6.969957081545064, "grad_norm": 0.36432990431785583, "learning_rate": 5.0728935965213834e-05, "loss": 0.5462, "step": 4060 }, { "epoch": 6.978540772532189, "grad_norm": 0.37999647855758667, "learning_rate": 5.0468200231001286e-05, "loss": 0.5474, "step": 4065 }, { "epoch": 6.987124463519313, "grad_norm": 0.3890798091888428, "learning_rate": 5.020790994020972e-05, "loss": 0.5395, "step": 4070 }, { "epoch": 6.9957081545064375, "grad_norm": 0.3609655797481537, "learning_rate": 4.994806743365057e-05, "loss": 0.5466, "step": 4075 }, { "epoch": 6.9991416309012875, "eval_loss": 3.0554237365722656, "eval_runtime": 0.3942, "eval_samples_per_second": 15.219, "eval_steps_per_second": 2.537, "step": 4077 }, { "epoch": 7.0042918454935625, "grad_norm": 0.2814909517765045, "learning_rate": 4.96886750481082e-05, "loss": 0.5172, "step": 4080 }, { "epoch": 7.012875536480687, "grad_norm": 0.39267781376838684, "learning_rate": 4.942973511631889e-05, "loss": 0.4929, "step": 4085 }, { "epoch": 7.021459227467811, "grad_norm": 0.35972246527671814, "learning_rate": 4.9171249966950175e-05, "loss": 0.4891, "step": 4090 }, { "epoch": 7.030042918454936, "grad_norm": 0.31733205914497375, "learning_rate": 4.8913221924579554e-05, "loss": 0.4899, "step": 4095 }, { "epoch": 7.03862660944206, "grad_norm": 0.37763702869415283, "learning_rate": 4.8655653309673776e-05, "loss": 0.4899, "step": 4100 }, { "epoch": 7.047210300429184, "grad_norm": 0.39104217290878296, "learning_rate": 4.839854643856792e-05, "loss": 0.4843, "step": 4105 }, { "epoch": 7.055793991416309, "grad_norm": 0.3271881937980652, "learning_rate": 4.814190362344454e-05, "loss": 0.4903, "step": 4110 }, { "epoch": 7.064377682403434, "grad_norm": 0.35317346453666687, "learning_rate": 4.788572717231293e-05, "loss": 0.4916, "step": 4115 }, { "epoch": 7.072961373390558, "grad_norm": 0.37199559807777405, "learning_rate": 4.763001938898832e-05, "loss": 0.4865, "step": 4120 }, { "epoch": 7.081545064377682, "grad_norm": 0.36147797107696533, "learning_rate": 4.7374782573071176e-05, "loss": 0.4884, "step": 4125 }, { "epoch": 7.090128755364807, "grad_norm": 0.3491626977920532, "learning_rate": 4.712001901992652e-05, "loss": 0.4926, "step": 4130 }, { "epoch": 7.098712446351931, "grad_norm": 0.36010846495628357, "learning_rate": 4.686573102066326e-05, "loss": 0.4942, "step": 4135 }, { "epoch": 7.1072961373390555, "grad_norm": 0.34614065289497375, "learning_rate": 4.661192086211366e-05, "loss": 0.4888, "step": 4140 }, { "epoch": 7.115879828326181, "grad_norm": 0.37029707431793213, "learning_rate": 4.6358590826812664e-05, "loss": 0.493, "step": 4145 }, { "epoch": 7.124463519313305, "grad_norm": 0.35328662395477295, "learning_rate": 4.610574319297748e-05, "loss": 0.4949, "step": 4150 }, { "epoch": 7.133047210300429, "grad_norm": 0.33476021885871887, "learning_rate": 4.585338023448702e-05, "loss": 0.488, "step": 4155 }, { "epoch": 7.141630901287554, "grad_norm": 0.36379748582839966, "learning_rate": 4.560150422086147e-05, "loss": 0.4927, "step": 4160 }, { "epoch": 7.150214592274678, "grad_norm": 0.3896268308162689, "learning_rate": 4.535011741724184e-05, "loss": 0.4917, "step": 4165 }, { "epoch": 7.1587982832618025, "grad_norm": 0.3854301869869232, "learning_rate": 4.5099222084369805e-05, "loss": 0.4853, "step": 4170 }, { "epoch": 7.167381974248927, "grad_norm": 0.36009612679481506, "learning_rate": 4.4848820478566966e-05, "loss": 0.4905, "step": 4175 }, { "epoch": 7.175965665236052, "grad_norm": 0.3766346573829651, "learning_rate": 4.45989148517149e-05, "loss": 0.4936, "step": 4180 }, { "epoch": 7.184549356223176, "grad_norm": 0.3666467070579529, "learning_rate": 4.4349507451234894e-05, "loss": 0.487, "step": 4185 }, { "epoch": 7.1931330472103, "grad_norm": 0.3508441150188446, "learning_rate": 4.410060052006758e-05, "loss": 0.4916, "step": 4190 }, { "epoch": 7.201716738197425, "grad_norm": 0.3494192957878113, "learning_rate": 4.3852196296652706e-05, "loss": 0.4901, "step": 4195 }, { "epoch": 7.210300429184549, "grad_norm": 0.35362881422042847, "learning_rate": 4.360429701490934e-05, "loss": 0.4933, "step": 4200 }, { "epoch": 7.218884120171674, "grad_norm": 0.35061484575271606, "learning_rate": 4.335690490421548e-05, "loss": 0.4883, "step": 4205 }, { "epoch": 7.227467811158799, "grad_norm": 0.3576537072658539, "learning_rate": 4.311002218938798e-05, "loss": 0.4896, "step": 4210 }, { "epoch": 7.236051502145923, "grad_norm": 0.35517919063568115, "learning_rate": 4.286365109066285e-05, "loss": 0.4873, "step": 4215 }, { "epoch": 7.244635193133047, "grad_norm": 0.3709685206413269, "learning_rate": 4.261779382367499e-05, "loss": 0.495, "step": 4220 }, { "epoch": 7.253218884120171, "grad_norm": 0.39842909574508667, "learning_rate": 4.237245259943837e-05, "loss": 0.4957, "step": 4225 }, { "epoch": 7.261802575107296, "grad_norm": 0.3722572922706604, "learning_rate": 4.212762962432619e-05, "loss": 0.4978, "step": 4230 }, { "epoch": 7.2703862660944205, "grad_norm": 0.3434411287307739, "learning_rate": 4.188332710005094e-05, "loss": 0.4925, "step": 4235 }, { "epoch": 7.278969957081545, "grad_norm": 0.3870338797569275, "learning_rate": 4.1639547223644706e-05, "loss": 0.4802, "step": 4240 }, { "epoch": 7.28755364806867, "grad_norm": 0.3743104040622711, "learning_rate": 4.139629218743931e-05, "loss": 0.4847, "step": 4245 }, { "epoch": 7.296137339055794, "grad_norm": 0.3608282506465912, "learning_rate": 4.11535641790467e-05, "loss": 0.486, "step": 4250 }, { "epoch": 7.304721030042918, "grad_norm": 0.3679661452770233, "learning_rate": 4.091136538133916e-05, "loss": 0.4942, "step": 4255 }, { "epoch": 7.313304721030043, "grad_norm": 0.3837164044380188, "learning_rate": 4.06696979724298e-05, "loss": 0.4881, "step": 4260 }, { "epoch": 7.3218884120171674, "grad_norm": 0.37015727162361145, "learning_rate": 4.042856412565287e-05, "loss": 0.4875, "step": 4265 }, { "epoch": 7.330472103004292, "grad_norm": 0.3824974596500397, "learning_rate": 4.0187966009544255e-05, "loss": 0.4895, "step": 4270 }, { "epoch": 7.339055793991417, "grad_norm": 0.356283038854599, "learning_rate": 3.994790578782198e-05, "loss": 0.4961, "step": 4275 }, { "epoch": 7.347639484978541, "grad_norm": 0.3656464219093323, "learning_rate": 3.970838561936675e-05, "loss": 0.5015, "step": 4280 }, { "epoch": 7.356223175965665, "grad_norm": 0.3856269419193268, "learning_rate": 3.9469407658202514e-05, "loss": 0.4941, "step": 4285 }, { "epoch": 7.364806866952789, "grad_norm": 0.4023449122905731, "learning_rate": 3.9230974053477086e-05, "loss": 0.4943, "step": 4290 }, { "epoch": 7.373390557939914, "grad_norm": 0.34391605854034424, "learning_rate": 3.899308694944298e-05, "loss": 0.5006, "step": 4295 }, { "epoch": 7.381974248927039, "grad_norm": 0.3965080976486206, "learning_rate": 3.875574848543774e-05, "loss": 0.4925, "step": 4300 }, { "epoch": 7.390557939914163, "grad_norm": 0.37056249380111694, "learning_rate": 3.85189607958651e-05, "loss": 0.5052, "step": 4305 }, { "epoch": 7.399141630901288, "grad_norm": 0.3915135860443115, "learning_rate": 3.8282726010175715e-05, "loss": 0.4885, "step": 4310 }, { "epoch": 7.407725321888412, "grad_norm": 0.3784487247467041, "learning_rate": 3.804704625284774e-05, "loss": 0.4902, "step": 4315 }, { "epoch": 7.416309012875536, "grad_norm": 0.36971473693847656, "learning_rate": 3.7811923643367974e-05, "loss": 0.4971, "step": 4320 }, { "epoch": 7.424892703862661, "grad_norm": 0.36764466762542725, "learning_rate": 3.757736029621292e-05, "loss": 0.4873, "step": 4325 }, { "epoch": 7.4334763948497855, "grad_norm": 0.3773200809955597, "learning_rate": 3.734335832082927e-05, "loss": 0.5019, "step": 4330 }, { "epoch": 7.44206008583691, "grad_norm": 0.34619271755218506, "learning_rate": 3.710991982161555e-05, "loss": 0.4919, "step": 4335 }, { "epoch": 7.450643776824034, "grad_norm": 0.33658042550086975, "learning_rate": 3.687704689790277e-05, "loss": 0.4883, "step": 4340 }, { "epoch": 7.459227467811159, "grad_norm": 0.36298757791519165, "learning_rate": 3.66447416439356e-05, "loss": 0.5003, "step": 4345 }, { "epoch": 7.467811158798283, "grad_norm": 0.35422852635383606, "learning_rate": 3.641300614885378e-05, "loss": 0.4923, "step": 4350 }, { "epoch": 7.476394849785407, "grad_norm": 0.3848954737186432, "learning_rate": 3.618184249667308e-05, "loss": 0.4977, "step": 4355 }, { "epoch": 7.484978540772532, "grad_norm": 0.3532540798187256, "learning_rate": 3.595125276626653e-05, "loss": 0.5032, "step": 4360 }, { "epoch": 7.493562231759657, "grad_norm": 0.35986649990081787, "learning_rate": 3.5721239031346066e-05, "loss": 0.4964, "step": 4365 }, { "epoch": 7.502145922746781, "grad_norm": 0.3666352927684784, "learning_rate": 3.549180336044352e-05, "loss": 0.4992, "step": 4370 }, { "epoch": 7.510729613733906, "grad_norm": 0.3380297124385834, "learning_rate": 3.526294781689206e-05, "loss": 0.4817, "step": 4375 }, { "epoch": 7.51931330472103, "grad_norm": 0.36128494143486023, "learning_rate": 3.503467445880789e-05, "loss": 0.483, "step": 4380 }, { "epoch": 7.527896995708154, "grad_norm": 0.3538447916507721, "learning_rate": 3.480698533907152e-05, "loss": 0.4921, "step": 4385 }, { "epoch": 7.536480686695279, "grad_norm": 0.36427024006843567, "learning_rate": 3.457988250530931e-05, "loss": 0.4993, "step": 4390 }, { "epoch": 7.545064377682404, "grad_norm": 0.37009111046791077, "learning_rate": 3.435336799987514e-05, "loss": 0.4961, "step": 4395 }, { "epoch": 7.553648068669528, "grad_norm": 0.36786022782325745, "learning_rate": 3.412744385983201e-05, "loss": 0.4894, "step": 4400 }, { "epoch": 7.562231759656653, "grad_norm": 0.3731597363948822, "learning_rate": 3.390211211693369e-05, "loss": 0.4916, "step": 4405 }, { "epoch": 7.570815450643777, "grad_norm": 0.35336822271347046, "learning_rate": 3.367737479760652e-05, "loss": 0.4891, "step": 4410 }, { "epoch": 7.579399141630901, "grad_norm": 0.35434237122535706, "learning_rate": 3.3453233922931094e-05, "loss": 0.4937, "step": 4415 }, { "epoch": 7.587982832618025, "grad_norm": 0.3567320704460144, "learning_rate": 3.322969150862416e-05, "loss": 0.4979, "step": 4420 }, { "epoch": 7.5965665236051505, "grad_norm": 0.3649292588233948, "learning_rate": 3.300674956502047e-05, "loss": 0.4925, "step": 4425 }, { "epoch": 7.605150214592275, "grad_norm": 0.39201802015304565, "learning_rate": 3.2784410097054666e-05, "loss": 0.4866, "step": 4430 }, { "epoch": 7.613733905579399, "grad_norm": 0.3701328933238983, "learning_rate": 3.25626751042433e-05, "loss": 0.4876, "step": 4435 }, { "epoch": 7.622317596566524, "grad_norm": 0.3631632328033447, "learning_rate": 3.2341546580666796e-05, "loss": 0.4944, "step": 4440 }, { "epoch": 7.630901287553648, "grad_norm": 0.3858960270881653, "learning_rate": 3.212102651495167e-05, "loss": 0.4971, "step": 4445 }, { "epoch": 7.639484978540772, "grad_norm": 0.37257277965545654, "learning_rate": 3.1901116890252345e-05, "loss": 0.4971, "step": 4450 }, { "epoch": 7.6480686695278965, "grad_norm": 0.37199750542640686, "learning_rate": 3.1681819684233605e-05, "loss": 0.4989, "step": 4455 }, { "epoch": 7.656652360515022, "grad_norm": 0.3896372318267822, "learning_rate": 3.146313686905279e-05, "loss": 0.4939, "step": 4460 }, { "epoch": 7.665236051502146, "grad_norm": 0.3580029010772705, "learning_rate": 3.124507041134177e-05, "loss": 0.4945, "step": 4465 }, { "epoch": 7.67381974248927, "grad_norm": 0.36499252915382385, "learning_rate": 3.102762227218957e-05, "loss": 0.4912, "step": 4470 }, { "epoch": 7.682403433476395, "grad_norm": 0.3598448634147644, "learning_rate": 3.081079440712473e-05, "loss": 0.4994, "step": 4475 }, { "epoch": 7.690987124463519, "grad_norm": 0.3923290967941284, "learning_rate": 3.059458876609742e-05, "loss": 0.4894, "step": 4480 }, { "epoch": 7.6995708154506435, "grad_norm": 0.3650890588760376, "learning_rate": 3.0379007293462192e-05, "loss": 0.4905, "step": 4485 }, { "epoch": 7.708154506437769, "grad_norm": 0.38070616126060486, "learning_rate": 3.0164051927960492e-05, "loss": 0.4996, "step": 4490 }, { "epoch": 7.716738197424893, "grad_norm": 0.3461267054080963, "learning_rate": 2.994972460270291e-05, "loss": 0.4939, "step": 4495 }, { "epoch": 7.725321888412017, "grad_norm": 0.36452245712280273, "learning_rate": 2.9736027245152275e-05, "loss": 0.5021, "step": 4500 }, { "epoch": 7.733905579399142, "grad_norm": 0.4071807861328125, "learning_rate": 2.9522961777105897e-05, "loss": 0.5019, "step": 4505 }, { "epoch": 7.742489270386266, "grad_norm": 0.36440128087997437, "learning_rate": 2.9310530114678502e-05, "loss": 0.5024, "step": 4510 }, { "epoch": 7.75107296137339, "grad_norm": 0.3590448796749115, "learning_rate": 2.9098734168284968e-05, "loss": 0.4874, "step": 4515 }, { "epoch": 7.7596566523605155, "grad_norm": 0.3638148903846741, "learning_rate": 2.8887575842623093e-05, "loss": 0.483, "step": 4520 }, { "epoch": 7.76824034334764, "grad_norm": 0.36555618047714233, "learning_rate": 2.867705703665654e-05, "loss": 0.4917, "step": 4525 }, { "epoch": 7.776824034334764, "grad_norm": 0.3763795793056488, "learning_rate": 2.8467179643597697e-05, "loss": 0.4886, "step": 4530 }, { "epoch": 7.785407725321888, "grad_norm": 0.3643328845500946, "learning_rate": 2.8257945550890665e-05, "loss": 0.4981, "step": 4535 }, { "epoch": 7.793991416309013, "grad_norm": 0.3772119879722595, "learning_rate": 2.8049356640194314e-05, "loss": 0.4868, "step": 4540 }, { "epoch": 7.802575107296137, "grad_norm": 0.3641767203807831, "learning_rate": 2.784141478736534e-05, "loss": 0.4928, "step": 4545 }, { "epoch": 7.8111587982832615, "grad_norm": 0.3673217296600342, "learning_rate": 2.7634121862441386e-05, "loss": 0.4922, "step": 4550 }, { "epoch": 7.819742489270387, "grad_norm": 0.3594400882720947, "learning_rate": 2.742747972962424e-05, "loss": 0.5024, "step": 4555 }, { "epoch": 7.828326180257511, "grad_norm": 0.36666861176490784, "learning_rate": 2.722149024726307e-05, "loss": 0.5001, "step": 4560 }, { "epoch": 7.836909871244635, "grad_norm": 0.3865159749984741, "learning_rate": 2.7016155267837684e-05, "loss": 0.4909, "step": 4565 }, { "epoch": 7.845493562231759, "grad_norm": 0.3859226107597351, "learning_rate": 2.6811476637941922e-05, "loss": 0.4917, "step": 4570 }, { "epoch": 7.854077253218884, "grad_norm": 0.37502434849739075, "learning_rate": 2.660745619826701e-05, "loss": 0.4934, "step": 4575 }, { "epoch": 7.8626609442060085, "grad_norm": 0.3713277280330658, "learning_rate": 2.6404095783585002e-05, "loss": 0.5048, "step": 4580 }, { "epoch": 7.871244635193133, "grad_norm": 0.39273905754089355, "learning_rate": 2.6201397222732316e-05, "loss": 0.4937, "step": 4585 }, { "epoch": 7.879828326180258, "grad_norm": 0.377205491065979, "learning_rate": 2.599936233859326e-05, "loss": 0.4989, "step": 4590 }, { "epoch": 7.888412017167382, "grad_norm": 0.3574148714542389, "learning_rate": 2.5797992948083592e-05, "loss": 0.492, "step": 4595 }, { "epoch": 7.896995708154506, "grad_norm": 0.3615160286426544, "learning_rate": 2.5597290862134405e-05, "loss": 0.4859, "step": 4600 }, { "epoch": 7.905579399141631, "grad_norm": 0.37071695923805237, "learning_rate": 2.5397257885675397e-05, "loss": 0.4884, "step": 4605 }, { "epoch": 7.914163090128755, "grad_norm": 0.36150577664375305, "learning_rate": 2.5197895817619153e-05, "loss": 0.4903, "step": 4610 }, { "epoch": 7.92274678111588, "grad_norm": 0.3787161409854889, "learning_rate": 2.499920645084465e-05, "loss": 0.498, "step": 4615 }, { "epoch": 7.931330472103005, "grad_norm": 0.36254122853279114, "learning_rate": 2.480119157218108e-05, "loss": 0.4968, "step": 4620 }, { "epoch": 7.939914163090129, "grad_norm": 0.3832210302352905, "learning_rate": 2.4603852962392125e-05, "loss": 0.4936, "step": 4625 }, { "epoch": 7.948497854077253, "grad_norm": 0.39253130555152893, "learning_rate": 2.4407192396159627e-05, "loss": 0.4941, "step": 4630 }, { "epoch": 7.957081545064378, "grad_norm": 0.3705868422985077, "learning_rate": 2.4211211642067623e-05, "loss": 0.4864, "step": 4635 }, { "epoch": 7.965665236051502, "grad_norm": 0.38986867666244507, "learning_rate": 2.401591246258673e-05, "loss": 0.4971, "step": 4640 }, { "epoch": 7.9742489270386265, "grad_norm": 0.3880539536476135, "learning_rate": 2.3821296614058054e-05, "loss": 0.4966, "step": 4645 }, { "epoch": 7.982832618025751, "grad_norm": 0.3790036141872406, "learning_rate": 2.3627365846677306e-05, "loss": 0.5004, "step": 4650 }, { "epoch": 7.991416309012876, "grad_norm": 0.3554070293903351, "learning_rate": 2.3434121904479434e-05, "loss": 0.4865, "step": 4655 }, { "epoch": 8.0, "grad_norm": 0.3582840859889984, "learning_rate": 2.3241566525322554e-05, "loss": 0.5038, "step": 4660 }, { "epoch": 8.0, "eval_loss": 3.4052770137786865, "eval_runtime": 0.394, "eval_samples_per_second": 15.23, "eval_steps_per_second": 2.538, "step": 4660 }, { "epoch": 8.008583690987125, "grad_norm": 0.29518163204193115, "learning_rate": 2.304970144087255e-05, "loss": 0.4553, "step": 4665 }, { "epoch": 8.017167381974248, "grad_norm": 0.3456011414527893, "learning_rate": 2.2858528376587407e-05, "loss": 0.4638, "step": 4670 }, { "epoch": 8.025751072961373, "grad_norm": 0.3549324572086334, "learning_rate": 2.2668049051701713e-05, "loss": 0.461, "step": 4675 }, { "epoch": 8.034334763948499, "grad_norm": 0.31662818789482117, "learning_rate": 2.247826517921121e-05, "loss": 0.4662, "step": 4680 }, { "epoch": 8.042918454935622, "grad_norm": 0.3052162230014801, "learning_rate": 2.2289178465857397e-05, "loss": 0.4645, "step": 4685 }, { "epoch": 8.051502145922747, "grad_norm": 0.34132641553878784, "learning_rate": 2.2100790612112133e-05, "loss": 0.461, "step": 4690 }, { "epoch": 8.060085836909872, "grad_norm": 0.3659987449645996, "learning_rate": 2.19131033121624e-05, "loss": 0.4602, "step": 4695 }, { "epoch": 8.068669527896995, "grad_norm": 0.3580094575881958, "learning_rate": 2.1726118253895034e-05, "loss": 0.4593, "step": 4700 }, { "epoch": 8.07725321888412, "grad_norm": 0.32578280568122864, "learning_rate": 2.1539837118881567e-05, "loss": 0.4593, "step": 4705 }, { "epoch": 8.085836909871245, "grad_norm": 0.3422725200653076, "learning_rate": 2.135426158236309e-05, "loss": 0.4624, "step": 4710 }, { "epoch": 8.094420600858369, "grad_norm": 0.33877745270729065, "learning_rate": 2.116939331323514e-05, "loss": 0.465, "step": 4715 }, { "epoch": 8.103004291845494, "grad_norm": 0.3325134813785553, "learning_rate": 2.098523397403288e-05, "loss": 0.4585, "step": 4720 }, { "epoch": 8.111587982832617, "grad_norm": 0.3371487259864807, "learning_rate": 2.080178522091585e-05, "loss": 0.4631, "step": 4725 }, { "epoch": 8.120171673819742, "grad_norm": 0.3465471565723419, "learning_rate": 2.0619048703653266e-05, "loss": 0.469, "step": 4730 }, { "epoch": 8.128755364806867, "grad_norm": 0.3354833424091339, "learning_rate": 2.04370260656093e-05, "loss": 0.4656, "step": 4735 }, { "epoch": 8.13733905579399, "grad_norm": 0.335443913936615, "learning_rate": 2.025571894372794e-05, "loss": 0.4591, "step": 4740 }, { "epoch": 8.145922746781116, "grad_norm": 0.35047757625579834, "learning_rate": 2.0075128968518573e-05, "loss": 0.4656, "step": 4745 }, { "epoch": 8.15450643776824, "grad_norm": 0.373524934053421, "learning_rate": 1.989525776404132e-05, "loss": 0.4612, "step": 4750 }, { "epoch": 8.163090128755364, "grad_norm": 0.3468015491962433, "learning_rate": 1.9716106947892164e-05, "loss": 0.4594, "step": 4755 }, { "epoch": 8.17167381974249, "grad_norm": 0.3522886335849762, "learning_rate": 1.9537678131188674e-05, "loss": 0.4635, "step": 4760 }, { "epoch": 8.180257510729614, "grad_norm": 0.350538045167923, "learning_rate": 1.9359972918555492e-05, "loss": 0.4615, "step": 4765 }, { "epoch": 8.188841201716738, "grad_norm": 0.31984084844589233, "learning_rate": 1.9182992908109644e-05, "loss": 0.4623, "step": 4770 }, { "epoch": 8.197424892703863, "grad_norm": 0.34820571541786194, "learning_rate": 1.900673969144653e-05, "loss": 0.465, "step": 4775 }, { "epoch": 8.206008583690988, "grad_norm": 0.3526110053062439, "learning_rate": 1.883121485362538e-05, "loss": 0.4608, "step": 4780 }, { "epoch": 8.214592274678111, "grad_norm": 0.3859311044216156, "learning_rate": 1.865641997315496e-05, "loss": 0.4705, "step": 4785 }, { "epoch": 8.223175965665236, "grad_norm": 0.3409660756587982, "learning_rate": 1.8482356621979645e-05, "loss": 0.4647, "step": 4790 }, { "epoch": 8.231759656652361, "grad_norm": 0.34335795044898987, "learning_rate": 1.8309026365464998e-05, "loss": 0.4647, "step": 4795 }, { "epoch": 8.240343347639485, "grad_norm": 0.33586952090263367, "learning_rate": 1.813643076238375e-05, "loss": 0.4626, "step": 4800 }, { "epoch": 8.24892703862661, "grad_norm": 0.343476265668869, "learning_rate": 1.7964571364902005e-05, "loss": 0.4604, "step": 4805 }, { "epoch": 8.257510729613735, "grad_norm": 0.3527016341686249, "learning_rate": 1.779344971856497e-05, "loss": 0.4645, "step": 4810 }, { "epoch": 8.266094420600858, "grad_norm": 0.34603551030158997, "learning_rate": 1.7623067362283243e-05, "loss": 0.4641, "step": 4815 }, { "epoch": 8.274678111587983, "grad_norm": 0.3567690849304199, "learning_rate": 1.7453425828318936e-05, "loss": 0.4622, "step": 4820 }, { "epoch": 8.283261802575108, "grad_norm": 0.3398036062717438, "learning_rate": 1.728452664227187e-05, "loss": 0.457, "step": 4825 }, { "epoch": 8.291845493562231, "grad_norm": 0.359521746635437, "learning_rate": 1.7116371323065883e-05, "loss": 0.4616, "step": 4830 }, { "epoch": 8.300429184549357, "grad_norm": 0.3150378465652466, "learning_rate": 1.694896138293516e-05, "loss": 0.4578, "step": 4835 }, { "epoch": 8.309012875536482, "grad_norm": 0.3591357469558716, "learning_rate": 1.6782298327410616e-05, "loss": 0.4604, "step": 4840 }, { "epoch": 8.317596566523605, "grad_norm": 0.33606967329978943, "learning_rate": 1.66163836553064e-05, "loss": 0.4521, "step": 4845 }, { "epoch": 8.32618025751073, "grad_norm": 0.3668070435523987, "learning_rate": 1.6451218858706374e-05, "loss": 0.4737, "step": 4850 }, { "epoch": 8.334763948497853, "grad_norm": 0.36258599162101746, "learning_rate": 1.628680542295069e-05, "loss": 0.4691, "step": 4855 }, { "epoch": 8.343347639484978, "grad_norm": 0.3564538061618805, "learning_rate": 1.6123144826622504e-05, "loss": 0.4634, "step": 4860 }, { "epoch": 8.351931330472103, "grad_norm": 0.36181172728538513, "learning_rate": 1.5960238541534578e-05, "loss": 0.4555, "step": 4865 }, { "epoch": 8.360515021459227, "grad_norm": 0.36802351474761963, "learning_rate": 1.579808803271612e-05, "loss": 0.4605, "step": 4870 }, { "epoch": 8.369098712446352, "grad_norm": 0.37159237265586853, "learning_rate": 1.563669475839956e-05, "loss": 0.46, "step": 4875 }, { "epoch": 8.377682403433477, "grad_norm": 0.36820727586746216, "learning_rate": 1.5476060170007457e-05, "loss": 0.467, "step": 4880 }, { "epoch": 8.3862660944206, "grad_norm": 0.3330000340938568, "learning_rate": 1.531618571213953e-05, "loss": 0.469, "step": 4885 }, { "epoch": 8.394849785407725, "grad_norm": 0.38085103034973145, "learning_rate": 1.5157072822559437e-05, "loss": 0.4644, "step": 4890 }, { "epoch": 8.40343347639485, "grad_norm": 0.35326817631721497, "learning_rate": 1.4998722932182074e-05, "loss": 0.4659, "step": 4895 }, { "epoch": 8.412017167381974, "grad_norm": 0.3420933187007904, "learning_rate": 1.4841137465060672e-05, "loss": 0.4673, "step": 4900 }, { "epoch": 8.420600858369099, "grad_norm": 0.3507622480392456, "learning_rate": 1.4684317838373884e-05, "loss": 0.4721, "step": 4905 }, { "epoch": 8.429184549356224, "grad_norm": 0.35186630487442017, "learning_rate": 1.4528265462413038e-05, "loss": 0.4667, "step": 4910 }, { "epoch": 8.437768240343347, "grad_norm": 0.3655546009540558, "learning_rate": 1.4372981740569646e-05, "loss": 0.4675, "step": 4915 }, { "epoch": 8.446351931330472, "grad_norm": 0.3504914343357086, "learning_rate": 1.4218468069322578e-05, "loss": 0.4657, "step": 4920 }, { "epoch": 8.454935622317597, "grad_norm": 0.3535081446170807, "learning_rate": 1.4064725838225568e-05, "loss": 0.4672, "step": 4925 }, { "epoch": 8.46351931330472, "grad_norm": 0.38395631313323975, "learning_rate": 1.3911756429894763e-05, "loss": 0.4684, "step": 4930 }, { "epoch": 8.472103004291846, "grad_norm": 0.3384489417076111, "learning_rate": 1.3759561219996242e-05, "loss": 0.4515, "step": 4935 }, { "epoch": 8.48068669527897, "grad_norm": 0.3759305477142334, "learning_rate": 1.3608141577233636e-05, "loss": 0.4604, "step": 4940 }, { "epoch": 8.489270386266094, "grad_norm": 0.3741336464881897, "learning_rate": 1.345749886333586e-05, "loss": 0.4683, "step": 4945 }, { "epoch": 8.49785407725322, "grad_norm": 0.3483313322067261, "learning_rate": 1.3307634433044846e-05, "loss": 0.4639, "step": 4950 }, { "epoch": 8.506437768240342, "grad_norm": 0.36218151450157166, "learning_rate": 1.3158549634103357e-05, "loss": 0.466, "step": 4955 }, { "epoch": 8.515021459227468, "grad_norm": 0.363930344581604, "learning_rate": 1.3010245807242849e-05, "loss": 0.4617, "step": 4960 }, { "epoch": 8.523605150214593, "grad_norm": 0.35775625705718994, "learning_rate": 1.2862724286171467e-05, "loss": 0.4717, "step": 4965 }, { "epoch": 8.532188841201716, "grad_norm": 0.3388819098472595, "learning_rate": 1.2715986397561997e-05, "loss": 0.467, "step": 4970 }, { "epoch": 8.540772532188841, "grad_norm": 0.3473096787929535, "learning_rate": 1.2570033461039954e-05, "loss": 0.4569, "step": 4975 }, { "epoch": 8.549356223175966, "grad_norm": 0.36242905259132385, "learning_rate": 1.2424866789171729e-05, "loss": 0.4631, "step": 4980 }, { "epoch": 8.55793991416309, "grad_norm": 0.33919695019721985, "learning_rate": 1.2280487687452768e-05, "loss": 0.4658, "step": 4985 }, { "epoch": 8.566523605150214, "grad_norm": 0.36114802956581116, "learning_rate": 1.2136897454295837e-05, "loss": 0.4615, "step": 4990 }, { "epoch": 8.57510729613734, "grad_norm": 0.3717144727706909, "learning_rate": 1.199409738101933e-05, "loss": 0.4604, "step": 4995 }, { "epoch": 8.583690987124463, "grad_norm": 0.3811343014240265, "learning_rate": 1.1852088751835689e-05, "loss": 0.4623, "step": 5000 }, { "epoch": 8.592274678111588, "grad_norm": 0.35531142354011536, "learning_rate": 1.1710872843839804e-05, "loss": 0.4609, "step": 5005 }, { "epoch": 8.600858369098713, "grad_norm": 0.3563953936100006, "learning_rate": 1.1570450926997655e-05, "loss": 0.4699, "step": 5010 }, { "epoch": 8.609442060085836, "grad_norm": 0.3635469377040863, "learning_rate": 1.1430824264134654e-05, "loss": 0.4632, "step": 5015 }, { "epoch": 8.618025751072961, "grad_norm": 0.3603283762931824, "learning_rate": 1.1291994110924509e-05, "loss": 0.4671, "step": 5020 }, { "epoch": 8.626609442060087, "grad_norm": 0.35889148712158203, "learning_rate": 1.1153961715877914e-05, "loss": 0.4586, "step": 5025 }, { "epoch": 8.63519313304721, "grad_norm": 0.38485071063041687, "learning_rate": 1.1016728320331093e-05, "loss": 0.4698, "step": 5030 }, { "epoch": 8.643776824034335, "grad_norm": 0.3366287052631378, "learning_rate": 1.0880295158434983e-05, "loss": 0.4598, "step": 5035 }, { "epoch": 8.65236051502146, "grad_norm": 0.3784838914871216, "learning_rate": 1.0744663457143878e-05, "loss": 0.4637, "step": 5040 }, { "epoch": 8.660944206008583, "grad_norm": 0.35765987634658813, "learning_rate": 1.0609834436204403e-05, "loss": 0.462, "step": 5045 }, { "epoch": 8.669527896995708, "grad_norm": 0.37458154559135437, "learning_rate": 1.0475809308144747e-05, "loss": 0.4613, "step": 5050 }, { "epoch": 8.678111587982833, "grad_norm": 0.374141126871109, "learning_rate": 1.0342589278263559e-05, "loss": 0.4614, "step": 5055 }, { "epoch": 8.686695278969957, "grad_norm": 0.34101325273513794, "learning_rate": 1.0210175544619116e-05, "loss": 0.4627, "step": 5060 }, { "epoch": 8.695278969957082, "grad_norm": 0.345047265291214, "learning_rate": 1.0078569298018758e-05, "loss": 0.4708, "step": 5065 }, { "epoch": 8.703862660944207, "grad_norm": 0.3726472854614258, "learning_rate": 9.947771722007915e-06, "loss": 0.464, "step": 5070 }, { "epoch": 8.71244635193133, "grad_norm": 0.3675495386123657, "learning_rate": 9.817783992859564e-06, "loss": 0.4633, "step": 5075 }, { "epoch": 8.721030042918455, "grad_norm": 0.32659244537353516, "learning_rate": 9.688607279563766e-06, "loss": 0.4685, "step": 5080 }, { "epoch": 8.729613733905579, "grad_norm": 0.3733295798301697, "learning_rate": 9.560242743816972e-06, "loss": 0.4532, "step": 5085 }, { "epoch": 8.738197424892704, "grad_norm": 0.35878074169158936, "learning_rate": 9.432691540011674e-06, "loss": 0.4678, "step": 5090 }, { "epoch": 8.746781115879829, "grad_norm": 0.3598923087120056, "learning_rate": 9.305954815226014e-06, "loss": 0.4715, "step": 5095 }, { "epoch": 8.755364806866952, "grad_norm": 0.34524357318878174, "learning_rate": 9.180033709213454e-06, "loss": 0.463, "step": 5100 }, { "epoch": 8.763948497854077, "grad_norm": 0.34148141741752625, "learning_rate": 9.054929354392527e-06, "loss": 0.4693, "step": 5105 }, { "epoch": 8.772532188841202, "grad_norm": 0.35487231612205505, "learning_rate": 8.93064287583667e-06, "loss": 0.4625, "step": 5110 }, { "epoch": 8.781115879828326, "grad_norm": 0.36163830757141113, "learning_rate": 8.807175391264067e-06, "loss": 0.4619, "step": 5115 }, { "epoch": 8.78969957081545, "grad_norm": 0.34637895226478577, "learning_rate": 8.684528011027659e-06, "loss": 0.4612, "step": 5120 }, { "epoch": 8.798283261802576, "grad_norm": 0.3432014584541321, "learning_rate": 8.562701838105115e-06, "loss": 0.4666, "step": 5125 }, { "epoch": 8.806866952789699, "grad_norm": 0.34569093585014343, "learning_rate": 8.441697968088891e-06, "loss": 0.4659, "step": 5130 }, { "epoch": 8.815450643776824, "grad_norm": 0.3551480770111084, "learning_rate": 8.321517489176433e-06, "loss": 0.4619, "step": 5135 }, { "epoch": 8.82403433476395, "grad_norm": 0.35777968168258667, "learning_rate": 8.202161482160353e-06, "loss": 0.4583, "step": 5140 }, { "epoch": 8.832618025751072, "grad_norm": 0.3783648908138275, "learning_rate": 8.083631020418791e-06, "loss": 0.4596, "step": 5145 }, { "epoch": 8.841201716738198, "grad_norm": 0.33539873361587524, "learning_rate": 7.965927169905551e-06, "loss": 0.4711, "step": 5150 }, { "epoch": 8.849785407725323, "grad_norm": 0.36662939190864563, "learning_rate": 7.84905098914076e-06, "loss": 0.4665, "step": 5155 }, { "epoch": 8.858369098712446, "grad_norm": 0.34115639328956604, "learning_rate": 7.733003529201278e-06, "loss": 0.4581, "step": 5160 }, { "epoch": 8.866952789699571, "grad_norm": 0.3474951386451721, "learning_rate": 7.617785833711077e-06, "loss": 0.4662, "step": 5165 }, { "epoch": 8.875536480686696, "grad_norm": 0.34105169773101807, "learning_rate": 7.503398938832107e-06, "loss": 0.4575, "step": 5170 }, { "epoch": 8.88412017167382, "grad_norm": 0.381610631942749, "learning_rate": 7.389843873254843e-06, "loss": 0.4616, "step": 5175 }, { "epoch": 8.892703862660944, "grad_norm": 0.3617483079433441, "learning_rate": 7.277121658189001e-06, "loss": 0.4629, "step": 5180 }, { "epoch": 8.901287553648068, "grad_norm": 0.3416938781738281, "learning_rate": 7.165233307354446e-06, "loss": 0.465, "step": 5185 }, { "epoch": 8.909871244635193, "grad_norm": 0.35436323285102844, "learning_rate": 7.054179826972074e-06, "loss": 0.4628, "step": 5190 }, { "epoch": 8.918454935622318, "grad_norm": 0.35174670815467834, "learning_rate": 6.943962215754618e-06, "loss": 0.4704, "step": 5195 }, { "epoch": 8.927038626609441, "grad_norm": 0.3784787356853485, "learning_rate": 6.834581464897871e-06, "loss": 0.4683, "step": 5200 }, { "epoch": 8.935622317596566, "grad_norm": 0.34359362721443176, "learning_rate": 6.726038558071656e-06, "loss": 0.4634, "step": 5205 }, { "epoch": 8.944206008583691, "grad_norm": 0.35282644629478455, "learning_rate": 6.618334471410925e-06, "loss": 0.4608, "step": 5210 }, { "epoch": 8.952789699570815, "grad_norm": 0.3536522388458252, "learning_rate": 6.511470173507161e-06, "loss": 0.4631, "step": 5215 }, { "epoch": 8.96137339055794, "grad_norm": 0.34291592240333557, "learning_rate": 6.405446625399481e-06, "loss": 0.4628, "step": 5220 }, { "epoch": 8.969957081545065, "grad_norm": 0.33180317282676697, "learning_rate": 6.300264780566112e-06, "loss": 0.4615, "step": 5225 }, { "epoch": 8.978540772532188, "grad_norm": 0.3489115536212921, "learning_rate": 6.195925584915752e-06, "loss": 0.4596, "step": 5230 }, { "epoch": 8.987124463519313, "grad_norm": 0.34033530950546265, "learning_rate": 6.0924299767791126e-06, "loss": 0.47, "step": 5235 }, { "epoch": 8.995708154506438, "grad_norm": 0.37230873107910156, "learning_rate": 5.989778886900432e-06, "loss": 0.4624, "step": 5240 }, { "epoch": 8.999141630901288, "eval_loss": 3.695244073867798, "eval_runtime": 0.3944, "eval_samples_per_second": 15.212, "eval_steps_per_second": 2.535, "step": 5242 }, { "epoch": 9.004291845493562, "grad_norm": 0.28615859150886536, "learning_rate": 5.887973238429145e-06, "loss": 0.4573, "step": 5245 }, { "epoch": 9.012875536480687, "grad_norm": 0.3141264319419861, "learning_rate": 5.787013946911546e-06, "loss": 0.4503, "step": 5250 }, { "epoch": 9.021459227467812, "grad_norm": 0.32362473011016846, "learning_rate": 5.686901920282606e-06, "loss": 0.4558, "step": 5255 }, { "epoch": 9.030042918454935, "grad_norm": 0.32775941491127014, "learning_rate": 5.587638058857736e-06, "loss": 0.445, "step": 5260 }, { "epoch": 9.03862660944206, "grad_norm": 0.33696043491363525, "learning_rate": 5.48922325532476e-06, "loss": 0.4521, "step": 5265 }, { "epoch": 9.047210300429185, "grad_norm": 0.3470819294452667, "learning_rate": 5.391658394735855e-06, "loss": 0.4513, "step": 5270 }, { "epoch": 9.055793991416309, "grad_norm": 0.3222349286079407, "learning_rate": 5.2949443544995644e-06, "loss": 0.4488, "step": 5275 }, { "epoch": 9.064377682403434, "grad_norm": 0.33785441517829895, "learning_rate": 5.199082004372957e-06, "loss": 0.4493, "step": 5280 }, { "epoch": 9.072961373390559, "grad_norm": 0.3577852249145508, "learning_rate": 5.104072206453802e-06, "loss": 0.4615, "step": 5285 }, { "epoch": 9.081545064377682, "grad_norm": 0.32605546712875366, "learning_rate": 5.009915815172772e-06, "loss": 0.4482, "step": 5290 }, { "epoch": 9.090128755364807, "grad_norm": 0.320216566324234, "learning_rate": 4.916613677285786e-06, "loss": 0.4518, "step": 5295 }, { "epoch": 9.098712446351932, "grad_norm": 0.323912650346756, "learning_rate": 4.8241666318664115e-06, "loss": 0.4442, "step": 5300 }, { "epoch": 9.107296137339056, "grad_norm": 0.342655748128891, "learning_rate": 4.732575510298276e-06, "loss": 0.4437, "step": 5305 }, { "epoch": 9.11587982832618, "grad_norm": 0.34046629071235657, "learning_rate": 4.641841136267666e-06, "loss": 0.4497, "step": 5310 }, { "epoch": 9.124463519313304, "grad_norm": 0.3281947374343872, "learning_rate": 4.551964325756031e-06, "loss": 0.4569, "step": 5315 }, { "epoch": 9.133047210300429, "grad_norm": 0.3604039251804352, "learning_rate": 4.462945887032632e-06, "loss": 0.451, "step": 5320 }, { "epoch": 9.141630901287554, "grad_norm": 0.3501492738723755, "learning_rate": 4.374786620647442e-06, "loss": 0.448, "step": 5325 }, { "epoch": 9.150214592274677, "grad_norm": 0.3506092429161072, "learning_rate": 4.287487319423756e-06, "loss": 0.4459, "step": 5330 }, { "epoch": 9.158798283261802, "grad_norm": 0.3382214307785034, "learning_rate": 4.20104876845111e-06, "loss": 0.452, "step": 5335 }, { "epoch": 9.167381974248928, "grad_norm": 0.3224546015262604, "learning_rate": 4.115471745078314e-06, "loss": 0.4535, "step": 5340 }, { "epoch": 9.17596566523605, "grad_norm": 0.3321012854576111, "learning_rate": 4.03075701890635e-06, "loss": 0.4477, "step": 5345 }, { "epoch": 9.184549356223176, "grad_norm": 0.32435712218284607, "learning_rate": 3.946905351781472e-06, "loss": 0.4494, "step": 5350 }, { "epoch": 9.193133047210301, "grad_norm": 0.33920931816101074, "learning_rate": 3.863917497788438e-06, "loss": 0.456, "step": 5355 }, { "epoch": 9.201716738197424, "grad_norm": 0.33260124921798706, "learning_rate": 3.7817942032436048e-06, "loss": 0.4471, "step": 5360 }, { "epoch": 9.21030042918455, "grad_norm": 0.3275390863418579, "learning_rate": 3.700536206688321e-06, "loss": 0.4493, "step": 5365 }, { "epoch": 9.218884120171674, "grad_norm": 0.35647067427635193, "learning_rate": 3.620144238882206e-06, "loss": 0.4491, "step": 5370 }, { "epoch": 9.227467811158798, "grad_norm": 0.3307458162307739, "learning_rate": 3.5406190227966427e-06, "loss": 0.4504, "step": 5375 }, { "epoch": 9.236051502145923, "grad_norm": 0.35020336508750916, "learning_rate": 3.4619612736082273e-06, "loss": 0.4577, "step": 5380 }, { "epoch": 9.244635193133048, "grad_norm": 0.33766666054725647, "learning_rate": 3.3841716986923624e-06, "loss": 0.4531, "step": 5385 }, { "epoch": 9.253218884120171, "grad_norm": 0.33843091130256653, "learning_rate": 3.3072509976169065e-06, "loss": 0.4564, "step": 5390 }, { "epoch": 9.261802575107296, "grad_norm": 0.3248330056667328, "learning_rate": 3.2311998621358363e-06, "loss": 0.4526, "step": 5395 }, { "epoch": 9.270386266094421, "grad_norm": 0.3351515829563141, "learning_rate": 3.1560189761830728e-06, "loss": 0.4544, "step": 5400 }, { "epoch": 9.278969957081545, "grad_norm": 0.3289077877998352, "learning_rate": 3.0817090158663185e-06, "loss": 0.4449, "step": 5405 }, { "epoch": 9.28755364806867, "grad_norm": 0.32089975476264954, "learning_rate": 3.008270649460965e-06, "loss": 0.4496, "step": 5410 }, { "epoch": 9.296137339055793, "grad_norm": 0.2968757748603821, "learning_rate": 2.9357045374040825e-06, "loss": 0.4458, "step": 5415 }, { "epoch": 9.304721030042918, "grad_norm": 0.34240734577178955, "learning_rate": 2.8640113322885185e-06, "loss": 0.4469, "step": 5420 }, { "epoch": 9.313304721030043, "grad_norm": 0.33385157585144043, "learning_rate": 2.7931916788569545e-06, "loss": 0.4527, "step": 5425 }, { "epoch": 9.321888412017167, "grad_norm": 0.34486281871795654, "learning_rate": 2.723246213996178e-06, "loss": 0.4542, "step": 5430 }, { "epoch": 9.330472103004292, "grad_norm": 0.3246801495552063, "learning_rate": 2.654175566731365e-06, "loss": 0.4574, "step": 5435 }, { "epoch": 9.339055793991417, "grad_norm": 0.33539149165153503, "learning_rate": 2.5859803582202968e-06, "loss": 0.4457, "step": 5440 }, { "epoch": 9.34763948497854, "grad_norm": 0.33203625679016113, "learning_rate": 2.518661201747918e-06, "loss": 0.4567, "step": 5445 }, { "epoch": 9.356223175965665, "grad_norm": 0.32282063364982605, "learning_rate": 2.452218702720821e-06, "loss": 0.4427, "step": 5450 }, { "epoch": 9.36480686695279, "grad_norm": 0.333141028881073, "learning_rate": 2.3866534586616364e-06, "loss": 0.4548, "step": 5455 }, { "epoch": 9.373390557939913, "grad_norm": 0.3323938250541687, "learning_rate": 2.3219660592038285e-06, "loss": 0.4558, "step": 5460 }, { "epoch": 9.381974248927039, "grad_norm": 0.33186817169189453, "learning_rate": 2.258157086086388e-06, "loss": 0.4499, "step": 5465 }, { "epoch": 9.390557939914164, "grad_norm": 0.33666694164276123, "learning_rate": 2.1952271131484236e-06, "loss": 0.4533, "step": 5470 }, { "epoch": 9.399141630901287, "grad_norm": 0.3561409115791321, "learning_rate": 2.133176706324236e-06, "loss": 0.4574, "step": 5475 }, { "epoch": 9.407725321888412, "grad_norm": 0.3282804489135742, "learning_rate": 2.0720064236380842e-06, "loss": 0.4511, "step": 5480 }, { "epoch": 9.416309012875537, "grad_norm": 0.3417915403842926, "learning_rate": 2.0117168151991606e-06, "loss": 0.4517, "step": 5485 }, { "epoch": 9.42489270386266, "grad_norm": 0.35541415214538574, "learning_rate": 1.9523084231967358e-06, "loss": 0.4498, "step": 5490 }, { "epoch": 9.433476394849786, "grad_norm": 0.33606576919555664, "learning_rate": 1.893781781895232e-06, "loss": 0.4466, "step": 5495 }, { "epoch": 9.44206008583691, "grad_norm": 0.333290159702301, "learning_rate": 1.8361374176293467e-06, "loss": 0.4514, "step": 5500 }, { "epoch": 9.450643776824034, "grad_norm": 0.3518344461917877, "learning_rate": 1.7793758487994694e-06, "loss": 0.4566, "step": 5505 }, { "epoch": 9.459227467811159, "grad_norm": 0.44788244366645813, "learning_rate": 1.7234975858669178e-06, "loss": 0.4564, "step": 5510 }, { "epoch": 9.467811158798284, "grad_norm": 0.34256601333618164, "learning_rate": 1.6685031313493416e-06, "loss": 0.4493, "step": 5515 }, { "epoch": 9.476394849785407, "grad_norm": 0.33245575428009033, "learning_rate": 1.6143929798162704e-06, "loss": 0.4479, "step": 5520 }, { "epoch": 9.484978540772532, "grad_norm": 0.34188759326934814, "learning_rate": 1.5611676178845958e-06, "loss": 0.4459, "step": 5525 }, { "epoch": 9.493562231759658, "grad_norm": 0.32970142364501953, "learning_rate": 1.5088275242142402e-06, "loss": 0.45, "step": 5530 }, { "epoch": 9.50214592274678, "grad_norm": 0.34352561831474304, "learning_rate": 1.4573731695038395e-06, "loss": 0.452, "step": 5535 }, { "epoch": 9.510729613733906, "grad_norm": 0.35988888144493103, "learning_rate": 1.4068050164864898e-06, "loss": 0.4497, "step": 5540 }, { "epoch": 9.51931330472103, "grad_norm": 0.32545995712280273, "learning_rate": 1.3571235199256405e-06, "loss": 0.4515, "step": 5545 }, { "epoch": 9.527896995708154, "grad_norm": 0.3102465569972992, "learning_rate": 1.30832912661093e-06, "loss": 0.4405, "step": 5550 }, { "epoch": 9.53648068669528, "grad_norm": 0.32020366191864014, "learning_rate": 1.2604222753542339e-06, "loss": 0.4479, "step": 5555 }, { "epoch": 9.545064377682403, "grad_norm": 0.357705295085907, "learning_rate": 1.2134033969856907e-06, "loss": 0.4435, "step": 5560 }, { "epoch": 9.553648068669528, "grad_norm": 0.3494960367679596, "learning_rate": 1.1672729143497929e-06, "loss": 0.4502, "step": 5565 }, { "epoch": 9.562231759656653, "grad_norm": 0.33992525935173035, "learning_rate": 1.1220312423016687e-06, "loss": 0.4597, "step": 5570 }, { "epoch": 9.570815450643776, "grad_norm": 0.3502410352230072, "learning_rate": 1.0776787877032736e-06, "loss": 0.4532, "step": 5575 }, { "epoch": 9.579399141630901, "grad_norm": 0.3116472065448761, "learning_rate": 1.034215949419748e-06, "loss": 0.4447, "step": 5580 }, { "epoch": 9.587982832618026, "grad_norm": 0.31818586587905884, "learning_rate": 9.916431183158881e-07, "loss": 0.449, "step": 5585 }, { "epoch": 9.59656652360515, "grad_norm": 0.31284070014953613, "learning_rate": 9.499606772525371e-07, "loss": 0.4426, "step": 5590 }, { "epoch": 9.605150214592275, "grad_norm": 0.35043418407440186, "learning_rate": 9.091690010831988e-07, "loss": 0.4521, "step": 5595 }, { "epoch": 9.6137339055794, "grad_norm": 0.352905809879303, "learning_rate": 8.692684566506959e-07, "loss": 0.4451, "step": 5600 }, { "epoch": 9.622317596566523, "grad_norm": 0.31967219710350037, "learning_rate": 8.30259402783784e-07, "loss": 0.4576, "step": 5605 }, { "epoch": 9.630901287553648, "grad_norm": 0.3698691129684448, "learning_rate": 7.921421902939874e-07, "loss": 0.4494, "step": 5610 }, { "epoch": 9.639484978540773, "grad_norm": 0.32335957884788513, "learning_rate": 7.54917161972446e-07, "loss": 0.4464, "step": 5615 }, { "epoch": 9.648068669527897, "grad_norm": 0.3521655201911926, "learning_rate": 7.185846525867956e-07, "loss": 0.4571, "step": 5620 }, { "epoch": 9.656652360515022, "grad_norm": 0.3388623893260956, "learning_rate": 6.831449888781926e-07, "loss": 0.453, "step": 5625 }, { "epoch": 9.665236051502147, "grad_norm": 0.3277793228626251, "learning_rate": 6.485984895583608e-07, "loss": 0.4486, "step": 5630 }, { "epoch": 9.67381974248927, "grad_norm": 0.32895511388778687, "learning_rate": 6.149454653067044e-07, "loss": 0.4509, "step": 5635 }, { "epoch": 9.682403433476395, "grad_norm": 0.34782108664512634, "learning_rate": 5.821862187675775e-07, "loss": 0.4537, "step": 5640 }, { "epoch": 9.690987124463518, "grad_norm": 0.33518868684768677, "learning_rate": 5.503210445474638e-07, "loss": 0.4543, "step": 5645 }, { "epoch": 9.699570815450643, "grad_norm": 0.33902445435523987, "learning_rate": 5.193502292124341e-07, "loss": 0.4487, "step": 5650 }, { "epoch": 9.708154506437769, "grad_norm": 0.3404318690299988, "learning_rate": 4.892740512854932e-07, "loss": 0.4597, "step": 5655 }, { "epoch": 9.716738197424892, "grad_norm": 0.33036890625953674, "learning_rate": 4.600927812441036e-07, "loss": 0.4472, "step": 5660 }, { "epoch": 9.725321888412017, "grad_norm": 0.34414026141166687, "learning_rate": 4.318066815177435e-07, "loss": 0.4452, "step": 5665 }, { "epoch": 9.733905579399142, "grad_norm": 0.33726632595062256, "learning_rate": 4.044160064855751e-07, "loss": 0.45, "step": 5670 }, { "epoch": 9.742489270386265, "grad_norm": 0.31507617235183716, "learning_rate": 3.779210024741131e-07, "loss": 0.4436, "step": 5675 }, { "epoch": 9.75107296137339, "grad_norm": 0.35751578211784363, "learning_rate": 3.523219077550488e-07, "loss": 0.4514, "step": 5680 }, { "epoch": 9.759656652360515, "grad_norm": 0.3282091021537781, "learning_rate": 3.2761895254306287e-07, "loss": 0.4472, "step": 5685 }, { "epoch": 9.768240343347639, "grad_norm": 0.3441978693008423, "learning_rate": 3.038123589938047e-07, "loss": 0.4516, "step": 5690 }, { "epoch": 9.776824034334764, "grad_norm": 0.33709728717803955, "learning_rate": 2.8090234120188295e-07, "loss": 0.4508, "step": 5695 }, { "epoch": 9.785407725321889, "grad_norm": 0.32524409890174866, "learning_rate": 2.588891051988895e-07, "loss": 0.445, "step": 5700 }, { "epoch": 9.793991416309012, "grad_norm": 0.3647370934486389, "learning_rate": 2.3777284895162288e-07, "loss": 0.444, "step": 5705 }, { "epoch": 9.802575107296137, "grad_norm": 0.3221174478530884, "learning_rate": 2.1755376236025637e-07, "loss": 0.4478, "step": 5710 }, { "epoch": 9.811158798283262, "grad_norm": 0.35065704584121704, "learning_rate": 1.9823202725665068e-07, "loss": 0.4538, "step": 5715 }, { "epoch": 9.819742489270386, "grad_norm": 0.3380087912082672, "learning_rate": 1.7980781740268848e-07, "loss": 0.4477, "step": 5720 }, { "epoch": 9.82832618025751, "grad_norm": 0.3177869915962219, "learning_rate": 1.622812984887867e-07, "loss": 0.4496, "step": 5725 }, { "epoch": 9.836909871244636, "grad_norm": 0.3140113353729248, "learning_rate": 1.4565262813230894e-07, "loss": 0.4485, "step": 5730 }, { "epoch": 9.84549356223176, "grad_norm": 0.32537147402763367, "learning_rate": 1.2992195587619993e-07, "loss": 0.4433, "step": 5735 }, { "epoch": 9.854077253218884, "grad_norm": 0.3178805410861969, "learning_rate": 1.1508942318767535e-07, "loss": 0.4465, "step": 5740 }, { "epoch": 9.86266094420601, "grad_norm": 0.33153483271598816, "learning_rate": 1.0115516345686749e-07, "loss": 0.4567, "step": 5745 }, { "epoch": 9.871244635193133, "grad_norm": 0.31883201003074646, "learning_rate": 8.811930199568163e-08, "loss": 0.449, "step": 5750 }, { "epoch": 9.879828326180258, "grad_norm": 0.3389532268047333, "learning_rate": 7.598195603666369e-08, "loss": 0.4533, "step": 5755 }, { "epoch": 9.888412017167383, "grad_norm": 0.3354385197162628, "learning_rate": 6.474323473194543e-08, "loss": 0.4558, "step": 5760 }, { "epoch": 9.896995708154506, "grad_norm": 0.3773539364337921, "learning_rate": 5.4403239152212013e-08, "loss": 0.4524, "step": 5765 }, { "epoch": 9.905579399141631, "grad_norm": 0.30916520953178406, "learning_rate": 4.4962062285902607e-08, "loss": 0.4474, "step": 5770 }, { "epoch": 9.914163090128756, "grad_norm": 0.3748643100261688, "learning_rate": 3.6419789038244504e-08, "loss": 0.4449, "step": 5775 }, { "epoch": 9.92274678111588, "grad_norm": 0.33963683247566223, "learning_rate": 2.877649623059808e-08, "loss": 0.4491, "step": 5780 }, { "epoch": 9.931330472103005, "grad_norm": 0.3399483859539032, "learning_rate": 2.2032252599690773e-08, "loss": 0.4523, "step": 5785 }, { "epoch": 9.939914163090128, "grad_norm": 0.3306158781051636, "learning_rate": 1.6187118797061917e-08, "loss": 0.4551, "step": 5790 }, { "epoch": 9.948497854077253, "grad_norm": 0.3417421579360962, "learning_rate": 1.1241147388452167e-08, "loss": 0.4527, "step": 5795 }, { "epoch": 9.957081545064378, "grad_norm": 0.34274721145629883, "learning_rate": 7.194382853370485e-09, "loss": 0.4442, "step": 5800 }, { "epoch": 9.965665236051501, "grad_norm": 0.36596229672431946, "learning_rate": 4.046861584705575e-09, "loss": 0.4545, "step": 5805 }, { "epoch": 9.974248927038627, "grad_norm": 0.3520822823047638, "learning_rate": 1.798611888370605e-09, "loss": 0.439, "step": 5810 }, { "epoch": 9.982832618025752, "grad_norm": 0.3437098562717438, "learning_rate": 4.4965398303675745e-10, "loss": 0.4463, "step": 5815 }, { "epoch": 9.991416309012875, "grad_norm": 0.3176514208316803, "learning_rate": 0.0, "loss": 0.454, "step": 5820 }, { "epoch": 9.991416309012875, "eval_loss": 3.8423588275909424, "eval_runtime": 0.4223, "eval_samples_per_second": 14.208, "eval_steps_per_second": 2.368, "step": 5820 }, { "epoch": 9.991416309012875, "step": 5820, "total_flos": 8.683561975386472e+18, "train_loss": 0.6765515476977293, "train_runtime": 24420.5896, "train_samples_per_second": 7.63, "train_steps_per_second": 0.238 } ], "logging_steps": 5, "max_steps": 5820, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.683561975386472e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }