{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 3848, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005197505197505198, "grad_norm": 37.05835422762717, "learning_rate": 5.194805194805195e-08, "loss": 2.0466, "step": 1 }, { "epoch": 0.002598752598752599, "grad_norm": 37.36458385503662, "learning_rate": 2.597402597402598e-07, "loss": 2.0439, "step": 5 }, { "epoch": 0.005197505197505198, "grad_norm": 39.070563651666575, "learning_rate": 5.194805194805196e-07, "loss": 2.0133, "step": 10 }, { "epoch": 0.007796257796257797, "grad_norm": 11.361027583243034, "learning_rate": 7.792207792207792e-07, "loss": 1.8666, "step": 15 }, { "epoch": 0.010395010395010396, "grad_norm": 5.863826986672296, "learning_rate": 1.0389610389610392e-06, "loss": 1.7677, "step": 20 }, { "epoch": 0.012993762993762994, "grad_norm": 4.522921155466077, "learning_rate": 1.2987012987012986e-06, "loss": 1.6467, "step": 25 }, { "epoch": 0.015592515592515593, "grad_norm": 3.1296619021160326, "learning_rate": 1.5584415584415584e-06, "loss": 1.4777, "step": 30 }, { "epoch": 0.018191268191268192, "grad_norm": 2.12702049381069, "learning_rate": 1.8181818181818183e-06, "loss": 1.3423, "step": 35 }, { "epoch": 0.02079002079002079, "grad_norm": 2.381326820034051, "learning_rate": 2.0779220779220784e-06, "loss": 1.2129, "step": 40 }, { "epoch": 0.02338877338877339, "grad_norm": 1.745648282948505, "learning_rate": 2.337662337662338e-06, "loss": 1.117, "step": 45 }, { "epoch": 0.02598752598752599, "grad_norm": 1.2151402583346058, "learning_rate": 2.597402597402597e-06, "loss": 1.0561, "step": 50 }, { "epoch": 0.028586278586278588, "grad_norm": 1.1469155851046635, "learning_rate": 2.8571428571428573e-06, "loss": 0.9923, "step": 55 }, { "epoch": 0.031185031185031187, "grad_norm": 1.1501290412070948, "learning_rate": 3.116883116883117e-06, "loss": 0.9596, "step": 60 }, { "epoch": 0.033783783783783786, "grad_norm": 1.2437833266166898, "learning_rate": 3.376623376623377e-06, "loss": 0.9281, "step": 65 }, { "epoch": 0.036382536382536385, "grad_norm": 1.1628173200126817, "learning_rate": 3.6363636363636366e-06, "loss": 0.9026, "step": 70 }, { "epoch": 0.03898128898128898, "grad_norm": 1.1632550654052534, "learning_rate": 3.896103896103897e-06, "loss": 0.9004, "step": 75 }, { "epoch": 0.04158004158004158, "grad_norm": 1.2551305602329859, "learning_rate": 4.155844155844157e-06, "loss": 0.8755, "step": 80 }, { "epoch": 0.04417879417879418, "grad_norm": 1.1474998623142625, "learning_rate": 4.415584415584416e-06, "loss": 0.8516, "step": 85 }, { "epoch": 0.04677754677754678, "grad_norm": 1.1143075622189529, "learning_rate": 4.675324675324676e-06, "loss": 0.8517, "step": 90 }, { "epoch": 0.04937629937629938, "grad_norm": 1.1206710563724116, "learning_rate": 4.935064935064935e-06, "loss": 0.8458, "step": 95 }, { "epoch": 0.05197505197505198, "grad_norm": 1.0838734153124963, "learning_rate": 5.194805194805194e-06, "loss": 0.8363, "step": 100 }, { "epoch": 0.05457380457380458, "grad_norm": 1.5339125140531156, "learning_rate": 5.4545454545454545e-06, "loss": 0.8195, "step": 105 }, { "epoch": 0.057172557172557176, "grad_norm": 1.1439071531812668, "learning_rate": 5.7142857142857145e-06, "loss": 0.8394, "step": 110 }, { "epoch": 0.059771309771309775, "grad_norm": 1.4061392934636887, "learning_rate": 5.9740259740259746e-06, "loss": 0.8239, "step": 115 }, { "epoch": 0.062370062370062374, "grad_norm": 1.1592154562156236, "learning_rate": 6.233766233766234e-06, "loss": 0.8039, "step": 120 }, { "epoch": 0.06496881496881497, "grad_norm": 1.0248228760326474, "learning_rate": 6.493506493506494e-06, "loss": 0.8022, "step": 125 }, { "epoch": 0.06756756756756757, "grad_norm": 1.1650242675138267, "learning_rate": 6.753246753246754e-06, "loss": 0.8032, "step": 130 }, { "epoch": 0.07016632016632017, "grad_norm": 1.1631059955651202, "learning_rate": 7.012987012987014e-06, "loss": 0.8076, "step": 135 }, { "epoch": 0.07276507276507277, "grad_norm": 0.9623641373070422, "learning_rate": 7.272727272727273e-06, "loss": 0.8039, "step": 140 }, { "epoch": 0.07536382536382537, "grad_norm": 1.1042092302715827, "learning_rate": 7.532467532467533e-06, "loss": 0.8041, "step": 145 }, { "epoch": 0.07796257796257797, "grad_norm": 1.034610728564867, "learning_rate": 7.792207792207793e-06, "loss": 0.7979, "step": 150 }, { "epoch": 0.08056133056133057, "grad_norm": 1.0441346842973676, "learning_rate": 8.051948051948052e-06, "loss": 0.7889, "step": 155 }, { "epoch": 0.08316008316008316, "grad_norm": 1.1314811400194547, "learning_rate": 8.311688311688313e-06, "loss": 0.8095, "step": 160 }, { "epoch": 0.08575883575883576, "grad_norm": 1.083993546178677, "learning_rate": 8.571428571428571e-06, "loss": 0.7838, "step": 165 }, { "epoch": 0.08835758835758836, "grad_norm": 0.9844284926961596, "learning_rate": 8.831168831168832e-06, "loss": 0.8007, "step": 170 }, { "epoch": 0.09095634095634096, "grad_norm": 0.9748756413462465, "learning_rate": 9.090909090909091e-06, "loss": 0.7839, "step": 175 }, { "epoch": 0.09355509355509356, "grad_norm": 1.054078525965627, "learning_rate": 9.350649350649352e-06, "loss": 0.7819, "step": 180 }, { "epoch": 0.09615384615384616, "grad_norm": 1.1412040734751607, "learning_rate": 9.610389610389611e-06, "loss": 0.7874, "step": 185 }, { "epoch": 0.09875259875259876, "grad_norm": 1.163328897434868, "learning_rate": 9.87012987012987e-06, "loss": 0.7762, "step": 190 }, { "epoch": 0.10135135135135136, "grad_norm": 1.0536579529578631, "learning_rate": 1.012987012987013e-05, "loss": 0.7665, "step": 195 }, { "epoch": 0.10395010395010396, "grad_norm": 1.048914553080904, "learning_rate": 1.0389610389610389e-05, "loss": 0.7849, "step": 200 }, { "epoch": 0.10654885654885655, "grad_norm": 0.9748667183768033, "learning_rate": 1.064935064935065e-05, "loss": 0.7839, "step": 205 }, { "epoch": 0.10914760914760915, "grad_norm": 0.9705690959742787, "learning_rate": 1.0909090909090909e-05, "loss": 0.7827, "step": 210 }, { "epoch": 0.11174636174636175, "grad_norm": 0.9527544873685447, "learning_rate": 1.116883116883117e-05, "loss": 0.7823, "step": 215 }, { "epoch": 0.11434511434511435, "grad_norm": 1.1194233425550948, "learning_rate": 1.1428571428571429e-05, "loss": 0.7795, "step": 220 }, { "epoch": 0.11694386694386695, "grad_norm": 2.120471539383005, "learning_rate": 1.1688311688311688e-05, "loss": 0.7778, "step": 225 }, { "epoch": 0.11954261954261955, "grad_norm": 0.9808803038153814, "learning_rate": 1.1948051948051949e-05, "loss": 0.7634, "step": 230 }, { "epoch": 0.12214137214137215, "grad_norm": 0.94060918910957, "learning_rate": 1.2207792207792208e-05, "loss": 0.7578, "step": 235 }, { "epoch": 0.12474012474012475, "grad_norm": 0.9878811168887828, "learning_rate": 1.2467532467532468e-05, "loss": 0.7837, "step": 240 }, { "epoch": 0.12733887733887733, "grad_norm": 1.0526434374561247, "learning_rate": 1.2727272727272728e-05, "loss": 0.777, "step": 245 }, { "epoch": 0.12993762993762994, "grad_norm": 1.067255251886244, "learning_rate": 1.2987012987012988e-05, "loss": 0.7606, "step": 250 }, { "epoch": 0.13253638253638253, "grad_norm": 1.019011419752788, "learning_rate": 1.3246753246753249e-05, "loss": 0.7783, "step": 255 }, { "epoch": 0.13513513513513514, "grad_norm": 0.9625368636583931, "learning_rate": 1.3506493506493508e-05, "loss": 0.7643, "step": 260 }, { "epoch": 0.13773388773388773, "grad_norm": 0.8964922122751663, "learning_rate": 1.3766233766233767e-05, "loss": 0.769, "step": 265 }, { "epoch": 0.14033264033264034, "grad_norm": 1.0861405810538973, "learning_rate": 1.4025974025974028e-05, "loss": 0.772, "step": 270 }, { "epoch": 0.14293139293139293, "grad_norm": 0.7725004042430159, "learning_rate": 1.4285714285714287e-05, "loss": 0.7684, "step": 275 }, { "epoch": 0.14553014553014554, "grad_norm": 0.9641597887627689, "learning_rate": 1.4545454545454546e-05, "loss": 0.7665, "step": 280 }, { "epoch": 0.14812889812889812, "grad_norm": 0.8799231599980127, "learning_rate": 1.4805194805194807e-05, "loss": 0.7553, "step": 285 }, { "epoch": 0.15072765072765074, "grad_norm": 0.865183845501612, "learning_rate": 1.5064935064935066e-05, "loss": 0.776, "step": 290 }, { "epoch": 0.15332640332640332, "grad_norm": 0.9280023423503597, "learning_rate": 1.5324675324675326e-05, "loss": 0.7655, "step": 295 }, { "epoch": 0.15592515592515593, "grad_norm": 0.9412086108813551, "learning_rate": 1.5584415584415587e-05, "loss": 0.7541, "step": 300 }, { "epoch": 0.15852390852390852, "grad_norm": 1.049389569363405, "learning_rate": 1.5844155844155847e-05, "loss": 0.7609, "step": 305 }, { "epoch": 0.16112266112266113, "grad_norm": 0.8632998959240495, "learning_rate": 1.6103896103896105e-05, "loss": 0.7572, "step": 310 }, { "epoch": 0.16372141372141372, "grad_norm": 0.926940763705136, "learning_rate": 1.6363636363636366e-05, "loss": 0.763, "step": 315 }, { "epoch": 0.16632016632016633, "grad_norm": 0.8169022540417774, "learning_rate": 1.6623376623376627e-05, "loss": 0.7564, "step": 320 }, { "epoch": 0.16891891891891891, "grad_norm": 0.8913503941645097, "learning_rate": 1.6883116883116884e-05, "loss": 0.768, "step": 325 }, { "epoch": 0.17151767151767153, "grad_norm": 0.9677009736871893, "learning_rate": 1.7142857142857142e-05, "loss": 0.7577, "step": 330 }, { "epoch": 0.1741164241164241, "grad_norm": 0.891596010196128, "learning_rate": 1.7402597402597403e-05, "loss": 0.7482, "step": 335 }, { "epoch": 0.17671517671517672, "grad_norm": 0.922259390367936, "learning_rate": 1.7662337662337664e-05, "loss": 0.7545, "step": 340 }, { "epoch": 0.1793139293139293, "grad_norm": 0.8012139878255935, "learning_rate": 1.792207792207792e-05, "loss": 0.748, "step": 345 }, { "epoch": 0.18191268191268192, "grad_norm": 0.8885070497602618, "learning_rate": 1.8181818181818182e-05, "loss": 0.7506, "step": 350 }, { "epoch": 0.1845114345114345, "grad_norm": 0.8139071250420767, "learning_rate": 1.8441558441558443e-05, "loss": 0.7657, "step": 355 }, { "epoch": 0.18711018711018712, "grad_norm": 0.8597830390928304, "learning_rate": 1.8701298701298704e-05, "loss": 0.7482, "step": 360 }, { "epoch": 0.1897089397089397, "grad_norm": 0.8215436811633837, "learning_rate": 1.896103896103896e-05, "loss": 0.7563, "step": 365 }, { "epoch": 0.19230769230769232, "grad_norm": 0.7888662364197664, "learning_rate": 1.9220779220779222e-05, "loss": 0.7638, "step": 370 }, { "epoch": 0.1949064449064449, "grad_norm": 0.7758295937643592, "learning_rate": 1.9480519480519483e-05, "loss": 0.7578, "step": 375 }, { "epoch": 0.19750519750519752, "grad_norm": 0.7321462182850453, "learning_rate": 1.974025974025974e-05, "loss": 0.7478, "step": 380 }, { "epoch": 0.2001039501039501, "grad_norm": 0.8972812591629451, "learning_rate": 2e-05, "loss": 0.7364, "step": 385 }, { "epoch": 0.20270270270270271, "grad_norm": 0.8504197871382875, "learning_rate": 1.9999897126378044e-05, "loss": 0.7531, "step": 390 }, { "epoch": 0.2053014553014553, "grad_norm": 1.0008077674950657, "learning_rate": 1.9999588507628768e-05, "loss": 0.7555, "step": 395 }, { "epoch": 0.2079002079002079, "grad_norm": 0.8219960180575271, "learning_rate": 1.999907415010192e-05, "loss": 0.7412, "step": 400 }, { "epoch": 0.2104989604989605, "grad_norm": 0.7975912262836815, "learning_rate": 1.9998354064380263e-05, "loss": 0.7342, "step": 405 }, { "epoch": 0.2130977130977131, "grad_norm": 0.7386294167625675, "learning_rate": 1.9997428265279365e-05, "loss": 0.7414, "step": 410 }, { "epoch": 0.2156964656964657, "grad_norm": 0.7996558972435127, "learning_rate": 1.999629677184728e-05, "loss": 0.7684, "step": 415 }, { "epoch": 0.2182952182952183, "grad_norm": 0.8532223218359839, "learning_rate": 1.999495960736418e-05, "loss": 0.7318, "step": 420 }, { "epoch": 0.2208939708939709, "grad_norm": 0.8099184350785439, "learning_rate": 1.999341679934186e-05, "loss": 0.7483, "step": 425 }, { "epoch": 0.2234927234927235, "grad_norm": 0.8032456938404925, "learning_rate": 1.999166837952316e-05, "loss": 0.7464, "step": 430 }, { "epoch": 0.2260914760914761, "grad_norm": 0.7613353132794639, "learning_rate": 1.998971438388134e-05, "loss": 0.7241, "step": 435 }, { "epoch": 0.2286902286902287, "grad_norm": 0.739838242733476, "learning_rate": 1.9987554852619325e-05, "loss": 0.7523, "step": 440 }, { "epoch": 0.2312889812889813, "grad_norm": 0.7456120762278995, "learning_rate": 1.998518983016887e-05, "loss": 0.7461, "step": 445 }, { "epoch": 0.2338877338877339, "grad_norm": 0.8064413740728865, "learning_rate": 1.9982619365189662e-05, "loss": 0.7524, "step": 450 }, { "epoch": 0.23648648648648649, "grad_norm": 0.8264553841158012, "learning_rate": 1.9979843510568312e-05, "loss": 0.7569, "step": 455 }, { "epoch": 0.2390852390852391, "grad_norm": 0.7700326026252647, "learning_rate": 1.9976862323417262e-05, "loss": 0.7421, "step": 460 }, { "epoch": 0.24168399168399168, "grad_norm": 0.8078133521219729, "learning_rate": 1.9973675865073614e-05, "loss": 0.7459, "step": 465 }, { "epoch": 0.2442827442827443, "grad_norm": 0.8834562269681555, "learning_rate": 1.9970284201097874e-05, "loss": 0.7419, "step": 470 }, { "epoch": 0.24688149688149688, "grad_norm": 0.8201197358624366, "learning_rate": 1.996668740127259e-05, "loss": 0.7264, "step": 475 }, { "epoch": 0.2494802494802495, "grad_norm": 0.7832673791812143, "learning_rate": 1.996288553960093e-05, "loss": 0.7326, "step": 480 }, { "epoch": 0.2520790020790021, "grad_norm": 0.7491102594223664, "learning_rate": 1.9958878694305147e-05, "loss": 0.7423, "step": 485 }, { "epoch": 0.25467775467775466, "grad_norm": 0.7711246686689462, "learning_rate": 1.9954666947824983e-05, "loss": 0.7601, "step": 490 }, { "epoch": 0.25727650727650725, "grad_norm": 0.8431355399537503, "learning_rate": 1.9950250386815953e-05, "loss": 0.7508, "step": 495 }, { "epoch": 0.2598752598752599, "grad_norm": 0.7491422333746851, "learning_rate": 1.9945629102147593e-05, "loss": 0.7365, "step": 500 }, { "epoch": 0.2624740124740125, "grad_norm": 0.6891526342744494, "learning_rate": 1.9940803188901556e-05, "loss": 0.7315, "step": 505 }, { "epoch": 0.26507276507276506, "grad_norm": 0.7000282425722503, "learning_rate": 1.9935772746369678e-05, "loss": 0.7524, "step": 510 }, { "epoch": 0.26767151767151764, "grad_norm": 0.6734603177161373, "learning_rate": 1.9930537878051927e-05, "loss": 0.746, "step": 515 }, { "epoch": 0.2702702702702703, "grad_norm": 0.6632065052188117, "learning_rate": 1.9925098691654275e-05, "loss": 0.7225, "step": 520 }, { "epoch": 0.27286902286902287, "grad_norm": 0.6985881279460795, "learning_rate": 1.9919455299086485e-05, "loss": 0.7451, "step": 525 }, { "epoch": 0.27546777546777546, "grad_norm": 0.6540248071379781, "learning_rate": 1.99136078164598e-05, "loss": 0.7322, "step": 530 }, { "epoch": 0.27806652806652804, "grad_norm": 0.661603660243552, "learning_rate": 1.9907556364084568e-05, "loss": 0.7544, "step": 535 }, { "epoch": 0.2806652806652807, "grad_norm": 0.7446420340679871, "learning_rate": 1.990130106646775e-05, "loss": 0.7476, "step": 540 }, { "epoch": 0.28326403326403327, "grad_norm": 0.8425291153171641, "learning_rate": 1.9894842052310373e-05, "loss": 0.733, "step": 545 }, { "epoch": 0.28586278586278585, "grad_norm": 0.6481748553031429, "learning_rate": 1.9888179454504875e-05, "loss": 0.7439, "step": 550 }, { "epoch": 0.28846153846153844, "grad_norm": 0.679678646955314, "learning_rate": 1.9881313410132365e-05, "loss": 0.7418, "step": 555 }, { "epoch": 0.2910602910602911, "grad_norm": 0.6504991149281961, "learning_rate": 1.9874244060459816e-05, "loss": 0.7338, "step": 560 }, { "epoch": 0.29365904365904366, "grad_norm": 0.7240524101878886, "learning_rate": 1.986697155093715e-05, "loss": 0.7384, "step": 565 }, { "epoch": 0.29625779625779625, "grad_norm": 0.677804845061605, "learning_rate": 1.9859496031194242e-05, "loss": 0.7368, "step": 570 }, { "epoch": 0.29885654885654883, "grad_norm": 0.6836994514744595, "learning_rate": 1.9851817655037854e-05, "loss": 0.7361, "step": 575 }, { "epoch": 0.30145530145530147, "grad_norm": 0.6837922941788586, "learning_rate": 1.9843936580448457e-05, "loss": 0.7243, "step": 580 }, { "epoch": 0.30405405405405406, "grad_norm": 0.7000337048592958, "learning_rate": 1.983585296957699e-05, "loss": 0.7329, "step": 585 }, { "epoch": 0.30665280665280664, "grad_norm": 0.7178435855481708, "learning_rate": 1.9827566988741525e-05, "loss": 0.719, "step": 590 }, { "epoch": 0.3092515592515592, "grad_norm": 0.7066252470023224, "learning_rate": 1.9819078808423825e-05, "loss": 0.7277, "step": 595 }, { "epoch": 0.31185031185031187, "grad_norm": 0.7131086872189417, "learning_rate": 1.981038860326586e-05, "loss": 0.7408, "step": 600 }, { "epoch": 0.31444906444906445, "grad_norm": 0.6862600699294851, "learning_rate": 1.980149655206621e-05, "loss": 0.7338, "step": 605 }, { "epoch": 0.31704781704781704, "grad_norm": 0.7265535072671346, "learning_rate": 1.9792402837776377e-05, "loss": 0.7209, "step": 610 }, { "epoch": 0.3196465696465696, "grad_norm": 0.7076896535093625, "learning_rate": 1.978310764749703e-05, "loss": 0.7442, "step": 615 }, { "epoch": 0.32224532224532226, "grad_norm": 0.6473108719176737, "learning_rate": 1.9773611172474143e-05, "loss": 0.7314, "step": 620 }, { "epoch": 0.32484407484407485, "grad_norm": 0.6958993345177985, "learning_rate": 1.976391360809507e-05, "loss": 0.7297, "step": 625 }, { "epoch": 0.32744282744282743, "grad_norm": 0.6980561880943195, "learning_rate": 1.9754015153884533e-05, "loss": 0.7507, "step": 630 }, { "epoch": 0.33004158004158, "grad_norm": 0.725526089648866, "learning_rate": 1.974391601350049e-05, "loss": 0.7316, "step": 635 }, { "epoch": 0.33264033264033266, "grad_norm": 0.6555475191719049, "learning_rate": 1.9733616394729975e-05, "loss": 0.7415, "step": 640 }, { "epoch": 0.33523908523908524, "grad_norm": 0.7126270164163553, "learning_rate": 1.9723116509484807e-05, "loss": 0.7084, "step": 645 }, { "epoch": 0.33783783783783783, "grad_norm": 0.6175823835566073, "learning_rate": 1.971241657379723e-05, "loss": 0.7437, "step": 650 }, { "epoch": 0.3404365904365904, "grad_norm": 0.641705701909066, "learning_rate": 1.9701516807815472e-05, "loss": 0.7227, "step": 655 }, { "epoch": 0.34303534303534305, "grad_norm": 0.6271473392148214, "learning_rate": 1.9690417435799217e-05, "loss": 0.7131, "step": 660 }, { "epoch": 0.34563409563409564, "grad_norm": 0.7368709959215286, "learning_rate": 1.967911868611498e-05, "loss": 0.7366, "step": 665 }, { "epoch": 0.3482328482328482, "grad_norm": 0.645812200849196, "learning_rate": 1.9667620791231422e-05, "loss": 0.7174, "step": 670 }, { "epoch": 0.3508316008316008, "grad_norm": 0.6331883051927242, "learning_rate": 1.965592398771456e-05, "loss": 0.7255, "step": 675 }, { "epoch": 0.35343035343035345, "grad_norm": 0.7047104196046573, "learning_rate": 1.9644028516222912e-05, "loss": 0.7349, "step": 680 }, { "epoch": 0.35602910602910603, "grad_norm": 0.6573670034252512, "learning_rate": 1.9631934621502514e-05, "loss": 0.7294, "step": 685 }, { "epoch": 0.3586278586278586, "grad_norm": 0.64012763867233, "learning_rate": 1.9619642552381924e-05, "loss": 0.7253, "step": 690 }, { "epoch": 0.3612266112266112, "grad_norm": 0.7085641762966977, "learning_rate": 1.9607152561767077e-05, "loss": 0.7302, "step": 695 }, { "epoch": 0.36382536382536385, "grad_norm": 0.6270235430166686, "learning_rate": 1.9594464906636083e-05, "loss": 0.7215, "step": 700 }, { "epoch": 0.36642411642411643, "grad_norm": 0.6347937985689194, "learning_rate": 1.958157984803395e-05, "loss": 0.7198, "step": 705 }, { "epoch": 0.369022869022869, "grad_norm": 0.6404396997260027, "learning_rate": 1.956849765106721e-05, "loss": 0.7355, "step": 710 }, { "epoch": 0.3716216216216216, "grad_norm": 0.5736442010381354, "learning_rate": 1.9555218584898457e-05, "loss": 0.7181, "step": 715 }, { "epoch": 0.37422037422037424, "grad_norm": 0.6408880853261398, "learning_rate": 1.954174292274082e-05, "loss": 0.713, "step": 720 }, { "epoch": 0.3768191268191268, "grad_norm": 0.6668059771868257, "learning_rate": 1.9528070941852334e-05, "loss": 0.7477, "step": 725 }, { "epoch": 0.3794178794178794, "grad_norm": 0.6868239822957647, "learning_rate": 1.9514202923530233e-05, "loss": 0.7209, "step": 730 }, { "epoch": 0.382016632016632, "grad_norm": 0.6673930879897976, "learning_rate": 1.9500139153105183e-05, "loss": 0.7256, "step": 735 }, { "epoch": 0.38461538461538464, "grad_norm": 0.6481855409470721, "learning_rate": 1.948587991993537e-05, "loss": 0.6989, "step": 740 }, { "epoch": 0.3872141372141372, "grad_norm": 0.7003926465709457, "learning_rate": 1.94714255174006e-05, "loss": 0.7079, "step": 745 }, { "epoch": 0.3898128898128898, "grad_norm": 0.7900259673292569, "learning_rate": 1.945677624289621e-05, "loss": 0.7238, "step": 750 }, { "epoch": 0.3924116424116424, "grad_norm": 0.6964569074530326, "learning_rate": 1.9441932397826993e-05, "loss": 0.7097, "step": 755 }, { "epoch": 0.39501039501039503, "grad_norm": 0.6487993082304472, "learning_rate": 1.9426894287600966e-05, "loss": 0.7079, "step": 760 }, { "epoch": 0.3976091476091476, "grad_norm": 0.688310939038027, "learning_rate": 1.9411662221623103e-05, "loss": 0.7154, "step": 765 }, { "epoch": 0.4002079002079002, "grad_norm": 0.67435617407303, "learning_rate": 1.939623651328897e-05, "loss": 0.7094, "step": 770 }, { "epoch": 0.4028066528066528, "grad_norm": 0.6183714426118461, "learning_rate": 1.9380617479978255e-05, "loss": 0.7341, "step": 775 }, { "epoch": 0.40540540540540543, "grad_norm": 0.6617362465155076, "learning_rate": 1.9364805443048266e-05, "loss": 0.7173, "step": 780 }, { "epoch": 0.408004158004158, "grad_norm": 0.5678238642277171, "learning_rate": 1.9348800727827307e-05, "loss": 0.7168, "step": 785 }, { "epoch": 0.4106029106029106, "grad_norm": 0.5910729067701429, "learning_rate": 1.9332603663607983e-05, "loss": 0.712, "step": 790 }, { "epoch": 0.4132016632016632, "grad_norm": 0.6197819946059432, "learning_rate": 1.9316214583640427e-05, "loss": 0.7247, "step": 795 }, { "epoch": 0.4158004158004158, "grad_norm": 0.6305700579142718, "learning_rate": 1.929963382512544e-05, "loss": 0.7139, "step": 800 }, { "epoch": 0.4183991683991684, "grad_norm": 0.6439900936581443, "learning_rate": 1.9282861729207555e-05, "loss": 0.7106, "step": 805 }, { "epoch": 0.420997920997921, "grad_norm": 0.6990297176777683, "learning_rate": 1.926589864096803e-05, "loss": 0.7234, "step": 810 }, { "epoch": 0.4235966735966736, "grad_norm": 0.6967169647297736, "learning_rate": 1.9248744909417728e-05, "loss": 0.7178, "step": 815 }, { "epoch": 0.4261954261954262, "grad_norm": 0.6433170596765804, "learning_rate": 1.923140088748995e-05, "loss": 0.7159, "step": 820 }, { "epoch": 0.4287941787941788, "grad_norm": 0.5983245551968179, "learning_rate": 1.9213866932033164e-05, "loss": 0.7057, "step": 825 }, { "epoch": 0.4313929313929314, "grad_norm": 0.6087581373525092, "learning_rate": 1.9196143403803667e-05, "loss": 0.7154, "step": 830 }, { "epoch": 0.433991683991684, "grad_norm": 0.636489262543138, "learning_rate": 1.9178230667458175e-05, "loss": 0.7057, "step": 835 }, { "epoch": 0.4365904365904366, "grad_norm": 0.6439601891575886, "learning_rate": 1.91601290915463e-05, "loss": 0.7279, "step": 840 }, { "epoch": 0.4391891891891892, "grad_norm": 0.6225933290388692, "learning_rate": 1.9141839048502974e-05, "loss": 0.7187, "step": 845 }, { "epoch": 0.4417879417879418, "grad_norm": 0.6035545819895592, "learning_rate": 1.9123360914640794e-05, "loss": 0.716, "step": 850 }, { "epoch": 0.44438669438669437, "grad_norm": 0.6607567088647912, "learning_rate": 1.9104695070142273e-05, "loss": 0.7309, "step": 855 }, { "epoch": 0.446985446985447, "grad_norm": 0.6708322724799578, "learning_rate": 1.9085841899052014e-05, "loss": 0.7042, "step": 860 }, { "epoch": 0.4495841995841996, "grad_norm": 0.5887928003237825, "learning_rate": 1.9066801789268815e-05, "loss": 0.7227, "step": 865 }, { "epoch": 0.4521829521829522, "grad_norm": 0.6055286727657766, "learning_rate": 1.9047575132537694e-05, "loss": 0.7092, "step": 870 }, { "epoch": 0.45478170478170477, "grad_norm": 0.6202428781057725, "learning_rate": 1.902816232444181e-05, "loss": 0.7104, "step": 875 }, { "epoch": 0.4573804573804574, "grad_norm": 0.6274570988856961, "learning_rate": 1.9008563764394334e-05, "loss": 0.7229, "step": 880 }, { "epoch": 0.45997920997921, "grad_norm": 0.6605818513974558, "learning_rate": 1.8988779855630246e-05, "loss": 0.7286, "step": 885 }, { "epoch": 0.4625779625779626, "grad_norm": 0.6410281151386927, "learning_rate": 1.896881100519801e-05, "loss": 0.725, "step": 890 }, { "epoch": 0.46517671517671516, "grad_norm": 0.5697116895525344, "learning_rate": 1.8948657623951224e-05, "loss": 0.7012, "step": 895 }, { "epoch": 0.4677754677754678, "grad_norm": 0.6221008774646193, "learning_rate": 1.8928320126540154e-05, "loss": 0.7088, "step": 900 }, { "epoch": 0.4703742203742204, "grad_norm": 0.6783327380563849, "learning_rate": 1.89077989314032e-05, "loss": 0.7012, "step": 905 }, { "epoch": 0.47297297297297297, "grad_norm": 0.6364455758589938, "learning_rate": 1.8887094460758298e-05, "loss": 0.7173, "step": 910 }, { "epoch": 0.47557172557172556, "grad_norm": 0.6423883619688115, "learning_rate": 1.8866207140594223e-05, "loss": 0.7221, "step": 915 }, { "epoch": 0.4781704781704782, "grad_norm": 0.6078098531540564, "learning_rate": 1.8845137400661832e-05, "loss": 0.7112, "step": 920 }, { "epoch": 0.4807692307692308, "grad_norm": 0.6855240964444005, "learning_rate": 1.882388567446522e-05, "loss": 0.701, "step": 925 }, { "epoch": 0.48336798336798337, "grad_norm": 0.6214299025960558, "learning_rate": 1.880245239925279e-05, "loss": 0.7152, "step": 930 }, { "epoch": 0.48596673596673595, "grad_norm": 0.7011949443886023, "learning_rate": 1.878083801600828e-05, "loss": 0.7099, "step": 935 }, { "epoch": 0.4885654885654886, "grad_norm": 0.6305195514031546, "learning_rate": 1.8759042969441666e-05, "loss": 0.7119, "step": 940 }, { "epoch": 0.4911642411642412, "grad_norm": 0.6284331380737156, "learning_rate": 1.8737067707980018e-05, "loss": 0.7283, "step": 945 }, { "epoch": 0.49376299376299376, "grad_norm": 0.6834839648993508, "learning_rate": 1.8714912683758292e-05, "loss": 0.7224, "step": 950 }, { "epoch": 0.49636174636174635, "grad_norm": 0.5860761907825943, "learning_rate": 1.869257835261e-05, "loss": 0.7006, "step": 955 }, { "epoch": 0.498960498960499, "grad_norm": 0.6589606191667001, "learning_rate": 1.8670065174057854e-05, "loss": 0.7179, "step": 960 }, { "epoch": 0.5015592515592515, "grad_norm": 0.6369606574521753, "learning_rate": 1.8647373611304295e-05, "loss": 0.715, "step": 965 }, { "epoch": 0.5041580041580042, "grad_norm": 0.604142259445137, "learning_rate": 1.862450413122197e-05, "loss": 0.7014, "step": 970 }, { "epoch": 0.5067567567567568, "grad_norm": 0.6290172817388325, "learning_rate": 1.8601457204344134e-05, "loss": 0.71, "step": 975 }, { "epoch": 0.5093555093555093, "grad_norm": 0.5861860996181917, "learning_rate": 1.8578233304854952e-05, "loss": 0.7041, "step": 980 }, { "epoch": 0.511954261954262, "grad_norm": 0.6582907547385, "learning_rate": 1.855483291057976e-05, "loss": 0.7259, "step": 985 }, { "epoch": 0.5145530145530145, "grad_norm": 0.6056438043018836, "learning_rate": 1.8531256502975218e-05, "loss": 0.7141, "step": 990 }, { "epoch": 0.5171517671517671, "grad_norm": 0.6299326122139869, "learning_rate": 1.850750456711941e-05, "loss": 0.7037, "step": 995 }, { "epoch": 0.5197505197505198, "grad_norm": 0.5922842182974687, "learning_rate": 1.8483577591701876e-05, "loss": 0.7227, "step": 1000 }, { "epoch": 0.5223492723492723, "grad_norm": 0.6335533442723702, "learning_rate": 1.8459476069013537e-05, "loss": 0.7331, "step": 1005 }, { "epoch": 0.524948024948025, "grad_norm": 0.603529261601813, "learning_rate": 1.8435200494936583e-05, "loss": 0.7204, "step": 1010 }, { "epoch": 0.5275467775467776, "grad_norm": 0.5784152517101885, "learning_rate": 1.841075136893426e-05, "loss": 0.7045, "step": 1015 }, { "epoch": 0.5301455301455301, "grad_norm": 0.5775827914904406, "learning_rate": 1.8386129194040597e-05, "loss": 0.71, "step": 1020 }, { "epoch": 0.5327442827442828, "grad_norm": 0.6037013614687209, "learning_rate": 1.8361334476850052e-05, "loss": 0.6973, "step": 1025 }, { "epoch": 0.5353430353430353, "grad_norm": 0.5969176345454209, "learning_rate": 1.8336367727507104e-05, "loss": 0.7102, "step": 1030 }, { "epoch": 0.5379417879417879, "grad_norm": 0.5880052937664013, "learning_rate": 1.8311229459695735e-05, "loss": 0.702, "step": 1035 }, { "epoch": 0.5405405405405406, "grad_norm": 0.5665171424206621, "learning_rate": 1.828592019062888e-05, "loss": 0.7059, "step": 1040 }, { "epoch": 0.5431392931392931, "grad_norm": 0.5776240714899302, "learning_rate": 1.826044044103777e-05, "loss": 0.7124, "step": 1045 }, { "epoch": 0.5457380457380457, "grad_norm": 0.5961454904432945, "learning_rate": 1.8234790735161233e-05, "loss": 0.7003, "step": 1050 }, { "epoch": 0.5483367983367984, "grad_norm": 0.61619821889063, "learning_rate": 1.82089716007349e-05, "loss": 0.7112, "step": 1055 }, { "epoch": 0.5509355509355509, "grad_norm": 0.5663452621669731, "learning_rate": 1.8182983568980347e-05, "loss": 0.7089, "step": 1060 }, { "epoch": 0.5535343035343036, "grad_norm": 0.5804355767396245, "learning_rate": 1.8156827174594157e-05, "loss": 0.6992, "step": 1065 }, { "epoch": 0.5561330561330561, "grad_norm": 0.6266224913727885, "learning_rate": 1.8130502955736945e-05, "loss": 0.7054, "step": 1070 }, { "epoch": 0.5587318087318087, "grad_norm": 0.5683126721009722, "learning_rate": 1.810401145402225e-05, "loss": 0.7132, "step": 1075 }, { "epoch": 0.5613305613305614, "grad_norm": 0.6083470685034659, "learning_rate": 1.8077353214505428e-05, "loss": 0.7058, "step": 1080 }, { "epoch": 0.5639293139293139, "grad_norm": 0.6109321123987571, "learning_rate": 1.8050528785672403e-05, "loss": 0.7096, "step": 1085 }, { "epoch": 0.5665280665280665, "grad_norm": 0.5553746362917324, "learning_rate": 1.8023538719428405e-05, "loss": 0.7009, "step": 1090 }, { "epoch": 0.5691268191268192, "grad_norm": 0.6062718001611305, "learning_rate": 1.7996383571086614e-05, "loss": 0.7098, "step": 1095 }, { "epoch": 0.5717255717255717, "grad_norm": 0.5458923136243866, "learning_rate": 1.796906389935672e-05, "loss": 0.6999, "step": 1100 }, { "epoch": 0.5743243243243243, "grad_norm": 0.5655005675331972, "learning_rate": 1.7941580266333433e-05, "loss": 0.7072, "step": 1105 }, { "epoch": 0.5769230769230769, "grad_norm": 0.5924848136830811, "learning_rate": 1.7913933237484936e-05, "loss": 0.7096, "step": 1110 }, { "epoch": 0.5795218295218295, "grad_norm": 0.5639064277581654, "learning_rate": 1.7886123381641226e-05, "loss": 0.7238, "step": 1115 }, { "epoch": 0.5821205821205822, "grad_norm": 0.6027713277537942, "learning_rate": 1.7858151270982422e-05, "loss": 0.7053, "step": 1120 }, { "epoch": 0.5847193347193347, "grad_norm": 0.5726133330263367, "learning_rate": 1.7830017481026992e-05, "loss": 0.6965, "step": 1125 }, { "epoch": 0.5873180873180873, "grad_norm": 0.5747640646170223, "learning_rate": 1.7801722590619905e-05, "loss": 0.6996, "step": 1130 }, { "epoch": 0.58991683991684, "grad_norm": 0.5720989361581698, "learning_rate": 1.777326718192074e-05, "loss": 0.7132, "step": 1135 }, { "epoch": 0.5925155925155925, "grad_norm": 0.6153266265302803, "learning_rate": 1.7744651840391686e-05, "loss": 0.7077, "step": 1140 }, { "epoch": 0.5951143451143451, "grad_norm": 0.6101914919623531, "learning_rate": 1.7715877154785504e-05, "loss": 0.7017, "step": 1145 }, { "epoch": 0.5977130977130977, "grad_norm": 0.5892227087436086, "learning_rate": 1.768694371713343e-05, "loss": 0.6901, "step": 1150 }, { "epoch": 0.6003118503118503, "grad_norm": 0.5792852971539604, "learning_rate": 1.7657852122732958e-05, "loss": 0.6915, "step": 1155 }, { "epoch": 0.6029106029106029, "grad_norm": 0.5812633503316377, "learning_rate": 1.7628602970135635e-05, "loss": 0.6983, "step": 1160 }, { "epoch": 0.6055093555093555, "grad_norm": 0.5776629062328037, "learning_rate": 1.7599196861134706e-05, "loss": 0.7033, "step": 1165 }, { "epoch": 0.6081081081081081, "grad_norm": 0.6291058371783688, "learning_rate": 1.7569634400752763e-05, "loss": 0.6976, "step": 1170 }, { "epoch": 0.6107068607068608, "grad_norm": 0.5613902945792599, "learning_rate": 1.753991619722928e-05, "loss": 0.7093, "step": 1175 }, { "epoch": 0.6133056133056133, "grad_norm": 0.6383056669880055, "learning_rate": 1.7510042862008102e-05, "loss": 0.7049, "step": 1180 }, { "epoch": 0.6159043659043659, "grad_norm": 0.6039268449286719, "learning_rate": 1.7480015009724873e-05, "loss": 0.702, "step": 1185 }, { "epoch": 0.6185031185031185, "grad_norm": 1.092030131308809, "learning_rate": 1.7449833258194376e-05, "loss": 0.6967, "step": 1190 }, { "epoch": 0.6211018711018711, "grad_norm": 0.6403075410117294, "learning_rate": 1.7419498228397826e-05, "loss": 0.707, "step": 1195 }, { "epoch": 0.6237006237006237, "grad_norm": 0.6802172468521516, "learning_rate": 1.7389010544470105e-05, "loss": 0.7178, "step": 1200 }, { "epoch": 0.6262993762993763, "grad_norm": 0.5789178063330651, "learning_rate": 1.7358370833686907e-05, "loss": 0.6981, "step": 1205 }, { "epoch": 0.6288981288981289, "grad_norm": 0.5616632715895711, "learning_rate": 1.732757972645183e-05, "loss": 0.6886, "step": 1210 }, { "epoch": 0.6314968814968815, "grad_norm": 0.5702377179814124, "learning_rate": 1.7296637856283432e-05, "loss": 0.6876, "step": 1215 }, { "epoch": 0.6340956340956341, "grad_norm": 0.5951629365447559, "learning_rate": 1.7265545859802154e-05, "loss": 0.6976, "step": 1220 }, { "epoch": 0.6366943866943867, "grad_norm": 0.5872853132760645, "learning_rate": 1.7234304376717252e-05, "loss": 0.6992, "step": 1225 }, { "epoch": 0.6392931392931392, "grad_norm": 0.570568895020055, "learning_rate": 1.720291404981364e-05, "loss": 0.6927, "step": 1230 }, { "epoch": 0.6418918918918919, "grad_norm": 0.5451578107308829, "learning_rate": 1.7171375524938638e-05, "loss": 0.7083, "step": 1235 }, { "epoch": 0.6444906444906445, "grad_norm": 0.5444907291949577, "learning_rate": 1.71396894509887e-05, "loss": 0.7107, "step": 1240 }, { "epoch": 0.6470893970893971, "grad_norm": 0.5605468917786897, "learning_rate": 1.710785647989607e-05, "loss": 0.7209, "step": 1245 }, { "epoch": 0.6496881496881497, "grad_norm": 0.578999844192668, "learning_rate": 1.7075877266615347e-05, "loss": 0.7045, "step": 1250 }, { "epoch": 0.6522869022869023, "grad_norm": 0.5894696454082586, "learning_rate": 1.704375246911004e-05, "loss": 0.7134, "step": 1255 }, { "epoch": 0.6548856548856549, "grad_norm": 0.6210688729366013, "learning_rate": 1.7011482748338998e-05, "loss": 0.6995, "step": 1260 }, { "epoch": 0.6574844074844075, "grad_norm": 0.5323403441786221, "learning_rate": 1.697906876824283e-05, "loss": 0.6931, "step": 1265 }, { "epoch": 0.66008316008316, "grad_norm": 0.5639934756275331, "learning_rate": 1.6946511195730254e-05, "loss": 0.7006, "step": 1270 }, { "epoch": 0.6626819126819127, "grad_norm": 0.5421699251312687, "learning_rate": 1.6913810700664337e-05, "loss": 0.6995, "step": 1275 }, { "epoch": 0.6652806652806653, "grad_norm": 0.5498793219946022, "learning_rate": 1.6880967955848762e-05, "loss": 0.6978, "step": 1280 }, { "epoch": 0.6678794178794178, "grad_norm": 0.5867341882206146, "learning_rate": 1.6847983637013946e-05, "loss": 0.7097, "step": 1285 }, { "epoch": 0.6704781704781705, "grad_norm": 0.5399055609353709, "learning_rate": 1.6814858422803162e-05, "loss": 0.7007, "step": 1290 }, { "epoch": 0.6730769230769231, "grad_norm": 0.5807292558166693, "learning_rate": 1.6781592994758563e-05, "loss": 0.7129, "step": 1295 }, { "epoch": 0.6756756756756757, "grad_norm": 0.5656285736891911, "learning_rate": 1.674818803730716e-05, "loss": 0.697, "step": 1300 }, { "epoch": 0.6782744282744283, "grad_norm": 0.6062022250444103, "learning_rate": 1.671464423774675e-05, "loss": 0.6933, "step": 1305 }, { "epoch": 0.6808731808731808, "grad_norm": 0.5432055936520285, "learning_rate": 1.668096228623176e-05, "loss": 0.7042, "step": 1310 }, { "epoch": 0.6834719334719335, "grad_norm": 0.5696528007526865, "learning_rate": 1.664714287575906e-05, "loss": 0.6921, "step": 1315 }, { "epoch": 0.6860706860706861, "grad_norm": 0.5723069912815502, "learning_rate": 1.66131867021537e-05, "loss": 0.7016, "step": 1320 }, { "epoch": 0.6886694386694386, "grad_norm": 0.5709607134990504, "learning_rate": 1.6579094464054595e-05, "loss": 0.7113, "step": 1325 }, { "epoch": 0.6912681912681913, "grad_norm": 0.6010864530745209, "learning_rate": 1.6544866862900146e-05, "loss": 0.695, "step": 1330 }, { "epoch": 0.6938669438669439, "grad_norm": 0.5800151181524591, "learning_rate": 1.6510504602913815e-05, "loss": 0.6952, "step": 1335 }, { "epoch": 0.6964656964656964, "grad_norm": 0.6157566120129612, "learning_rate": 1.6476008391089628e-05, "loss": 0.7105, "step": 1340 }, { "epoch": 0.6990644490644491, "grad_norm": 0.5752568805255817, "learning_rate": 1.644137893717764e-05, "loss": 0.7161, "step": 1345 }, { "epoch": 0.7016632016632016, "grad_norm": 0.5767294919201806, "learning_rate": 1.6406616953669317e-05, "loss": 0.6926, "step": 1350 }, { "epoch": 0.7042619542619543, "grad_norm": 0.6028116817361912, "learning_rate": 1.637172315578289e-05, "loss": 0.6878, "step": 1355 }, { "epoch": 0.7068607068607069, "grad_norm": 0.5528228994127236, "learning_rate": 1.6336698261448632e-05, "loss": 0.702, "step": 1360 }, { "epoch": 0.7094594594594594, "grad_norm": 0.5213097969559868, "learning_rate": 1.6301542991294094e-05, "loss": 0.6947, "step": 1365 }, { "epoch": 0.7120582120582121, "grad_norm": 0.5370559355171456, "learning_rate": 1.626625806862927e-05, "loss": 0.6938, "step": 1370 }, { "epoch": 0.7146569646569647, "grad_norm": 0.5319720586513772, "learning_rate": 1.623084421943172e-05, "loss": 0.691, "step": 1375 }, { "epoch": 0.7172557172557172, "grad_norm": 0.5688111345080934, "learning_rate": 1.6195302172331624e-05, "loss": 0.688, "step": 1380 }, { "epoch": 0.7198544698544699, "grad_norm": 0.5580488505367257, "learning_rate": 1.6159632658596815e-05, "loss": 0.6832, "step": 1385 }, { "epoch": 0.7224532224532224, "grad_norm": 0.5186643724448894, "learning_rate": 1.6123836412117702e-05, "loss": 0.7004, "step": 1390 }, { "epoch": 0.725051975051975, "grad_norm": 0.5624601265524769, "learning_rate": 1.608791416939219e-05, "loss": 0.7, "step": 1395 }, { "epoch": 0.7276507276507277, "grad_norm": 0.5541391195602022, "learning_rate": 1.605186666951052e-05, "loss": 0.6734, "step": 1400 }, { "epoch": 0.7302494802494802, "grad_norm": 0.5332911572620831, "learning_rate": 1.6015694654140076e-05, "loss": 0.6896, "step": 1405 }, { "epoch": 0.7328482328482329, "grad_norm": 0.5548630655118624, "learning_rate": 1.59793988675101e-05, "loss": 0.7001, "step": 1410 }, { "epoch": 0.7354469854469855, "grad_norm": 0.5047178273406415, "learning_rate": 1.5942980056396394e-05, "loss": 0.6959, "step": 1415 }, { "epoch": 0.738045738045738, "grad_norm": 0.5934685974493243, "learning_rate": 1.5906438970105965e-05, "loss": 0.6904, "step": 1420 }, { "epoch": 0.7406444906444907, "grad_norm": 0.5527382042918192, "learning_rate": 1.5869776360461596e-05, "loss": 0.6905, "step": 1425 }, { "epoch": 0.7432432432432432, "grad_norm": 0.5528065876901935, "learning_rate": 1.5832992981786362e-05, "loss": 0.6906, "step": 1430 }, { "epoch": 0.7458419958419958, "grad_norm": 0.5265702785034122, "learning_rate": 1.5796089590888154e-05, "loss": 0.6857, "step": 1435 }, { "epoch": 0.7484407484407485, "grad_norm": 0.5438966896525291, "learning_rate": 1.575906694704407e-05, "loss": 0.6926, "step": 1440 }, { "epoch": 0.751039501039501, "grad_norm": 0.5388783204520624, "learning_rate": 1.5721925811984795e-05, "loss": 0.6973, "step": 1445 }, { "epoch": 0.7536382536382537, "grad_norm": 0.5389485246395991, "learning_rate": 1.568466694987895e-05, "loss": 0.6962, "step": 1450 }, { "epoch": 0.7562370062370062, "grad_norm": 0.5935682313499335, "learning_rate": 1.5647291127317353e-05, "loss": 0.6725, "step": 1455 }, { "epoch": 0.7588357588357588, "grad_norm": 0.5822513257760918, "learning_rate": 1.560979911329725e-05, "loss": 0.6926, "step": 1460 }, { "epoch": 0.7614345114345115, "grad_norm": 0.591628155657702, "learning_rate": 1.5572191679206503e-05, "loss": 0.6724, "step": 1465 }, { "epoch": 0.764033264033264, "grad_norm": 0.5308640431230437, "learning_rate": 1.5534469598807703e-05, "loss": 0.693, "step": 1470 }, { "epoch": 0.7666320166320166, "grad_norm": 0.5305514116153979, "learning_rate": 1.5496633648222246e-05, "loss": 0.6781, "step": 1475 }, { "epoch": 0.7692307692307693, "grad_norm": 0.5426925004836652, "learning_rate": 1.54586846059144e-05, "loss": 0.7164, "step": 1480 }, { "epoch": 0.7718295218295218, "grad_norm": 0.523224567107794, "learning_rate": 1.542062325267525e-05, "loss": 0.6934, "step": 1485 }, { "epoch": 0.7744282744282744, "grad_norm": 0.5414111517541614, "learning_rate": 1.5382450371606643e-05, "loss": 0.6829, "step": 1490 }, { "epoch": 0.777027027027027, "grad_norm": 0.5960821029242948, "learning_rate": 1.534416674810509e-05, "loss": 0.6924, "step": 1495 }, { "epoch": 0.7796257796257796, "grad_norm": 0.5781538636667527, "learning_rate": 1.5305773169845598e-05, "loss": 0.695, "step": 1500 }, { "epoch": 0.7822245322245323, "grad_norm": 0.5764921412336972, "learning_rate": 1.5267270426765445e-05, "loss": 0.6887, "step": 1505 }, { "epoch": 0.7848232848232848, "grad_norm": 0.65716343258798, "learning_rate": 1.5228659311047969e-05, "loss": 0.6934, "step": 1510 }, { "epoch": 0.7874220374220374, "grad_norm": 0.5421938070789174, "learning_rate": 1.5189940617106231e-05, "loss": 0.6791, "step": 1515 }, { "epoch": 0.7900207900207901, "grad_norm": 0.5269524597532019, "learning_rate": 1.5151115141566686e-05, "loss": 0.6901, "step": 1520 }, { "epoch": 0.7926195426195426, "grad_norm": 0.5653499412081372, "learning_rate": 1.5112183683252797e-05, "loss": 0.6915, "step": 1525 }, { "epoch": 0.7952182952182952, "grad_norm": 0.5439207742834801, "learning_rate": 1.5073147043168589e-05, "loss": 0.6775, "step": 1530 }, { "epoch": 0.7978170478170478, "grad_norm": 0.5489788736531123, "learning_rate": 1.5034006024482164e-05, "loss": 0.7036, "step": 1535 }, { "epoch": 0.8004158004158004, "grad_norm": 0.5550872564267669, "learning_rate": 1.4994761432509201e-05, "loss": 0.6959, "step": 1540 }, { "epoch": 0.803014553014553, "grad_norm": 0.5490474276501166, "learning_rate": 1.4955414074696369e-05, "loss": 0.6906, "step": 1545 }, { "epoch": 0.8056133056133056, "grad_norm": 0.5125906404793841, "learning_rate": 1.4915964760604698e-05, "loss": 0.6828, "step": 1550 }, { "epoch": 0.8082120582120582, "grad_norm": 0.5203971242963676, "learning_rate": 1.4876414301892963e-05, "loss": 0.6825, "step": 1555 }, { "epoch": 0.8108108108108109, "grad_norm": 0.5658527313991132, "learning_rate": 1.4836763512300946e-05, "loss": 0.6928, "step": 1560 }, { "epoch": 0.8134095634095634, "grad_norm": 0.535836579476803, "learning_rate": 1.4797013207632718e-05, "loss": 0.7018, "step": 1565 }, { "epoch": 0.816008316008316, "grad_norm": 0.5645034857163748, "learning_rate": 1.475716420573984e-05, "loss": 0.685, "step": 1570 }, { "epoch": 0.8186070686070686, "grad_norm": 0.5217450999852206, "learning_rate": 1.4717217326504542e-05, "loss": 0.6849, "step": 1575 }, { "epoch": 0.8212058212058212, "grad_norm": 0.5480983222260942, "learning_rate": 1.467717339182286e-05, "loss": 0.7002, "step": 1580 }, { "epoch": 0.8238045738045738, "grad_norm": 0.542270358042675, "learning_rate": 1.4637033225587707e-05, "loss": 0.6694, "step": 1585 }, { "epoch": 0.8264033264033264, "grad_norm": 0.5543644512440328, "learning_rate": 1.4596797653671947e-05, "loss": 0.6921, "step": 1590 }, { "epoch": 0.829002079002079, "grad_norm": 0.5568668245488615, "learning_rate": 1.4556467503911376e-05, "loss": 0.6852, "step": 1595 }, { "epoch": 0.8316008316008316, "grad_norm": 0.5656786288010156, "learning_rate": 1.4516043606087712e-05, "loss": 0.6988, "step": 1600 }, { "epoch": 0.8341995841995842, "grad_norm": 0.5127153017778545, "learning_rate": 1.4475526791911516e-05, "loss": 0.6993, "step": 1605 }, { "epoch": 0.8367983367983368, "grad_norm": 0.5775499987771998, "learning_rate": 1.4434917895005071e-05, "loss": 0.695, "step": 1610 }, { "epoch": 0.8393970893970893, "grad_norm": 0.5330354393061656, "learning_rate": 1.4394217750885233e-05, "loss": 0.701, "step": 1615 }, { "epoch": 0.841995841995842, "grad_norm": 0.5419503755472291, "learning_rate": 1.4353427196946257e-05, "loss": 0.6913, "step": 1620 }, { "epoch": 0.8445945945945946, "grad_norm": 0.5311784231584057, "learning_rate": 1.4312547072442544e-05, "loss": 0.674, "step": 1625 }, { "epoch": 0.8471933471933472, "grad_norm": 0.5563211583759982, "learning_rate": 1.4271578218471392e-05, "loss": 0.6907, "step": 1630 }, { "epoch": 0.8497920997920998, "grad_norm": 0.543414093186935, "learning_rate": 1.4230521477955677e-05, "loss": 0.6933, "step": 1635 }, { "epoch": 0.8523908523908524, "grad_norm": 0.5240436124612864, "learning_rate": 1.418937769562652e-05, "loss": 0.686, "step": 1640 }, { "epoch": 0.854989604989605, "grad_norm": 0.5430088145666387, "learning_rate": 1.4148147718005907e-05, "loss": 0.6896, "step": 1645 }, { "epoch": 0.8575883575883576, "grad_norm": 0.5458623547814142, "learning_rate": 1.4106832393389258e-05, "loss": 0.6908, "step": 1650 }, { "epoch": 0.8601871101871101, "grad_norm": 0.5634088926920675, "learning_rate": 1.4065432571827992e-05, "loss": 0.7065, "step": 1655 }, { "epoch": 0.8627858627858628, "grad_norm": 0.5501420194013241, "learning_rate": 1.4023949105112034e-05, "loss": 0.6988, "step": 1660 }, { "epoch": 0.8653846153846154, "grad_norm": 0.545143784026079, "learning_rate": 1.3982382846752265e-05, "loss": 0.7001, "step": 1665 }, { "epoch": 0.867983367983368, "grad_norm": 0.5514351435035997, "learning_rate": 1.3940734651963008e-05, "loss": 0.6864, "step": 1670 }, { "epoch": 0.8705821205821206, "grad_norm": 0.578025130173629, "learning_rate": 1.3899005377644389e-05, "loss": 0.6951, "step": 1675 }, { "epoch": 0.8731808731808732, "grad_norm": 0.5486337422568206, "learning_rate": 1.3857195882364717e-05, "loss": 0.6875, "step": 1680 }, { "epoch": 0.8757796257796258, "grad_norm": 0.5987234793423925, "learning_rate": 1.3815307026342847e-05, "loss": 0.6895, "step": 1685 }, { "epoch": 0.8783783783783784, "grad_norm": 0.5723737998517776, "learning_rate": 1.3773339671430441e-05, "loss": 0.6934, "step": 1690 }, { "epoch": 0.8809771309771309, "grad_norm": 0.5157754043967531, "learning_rate": 1.3731294681094257e-05, "loss": 0.7008, "step": 1695 }, { "epoch": 0.8835758835758836, "grad_norm": 0.5484329818025037, "learning_rate": 1.3689172920398385e-05, "loss": 0.6588, "step": 1700 }, { "epoch": 0.8861746361746362, "grad_norm": 0.5313559372585769, "learning_rate": 1.3646975255986448e-05, "loss": 0.682, "step": 1705 }, { "epoch": 0.8887733887733887, "grad_norm": 0.5576768736741219, "learning_rate": 1.3604702556063751e-05, "loss": 0.6964, "step": 1710 }, { "epoch": 0.8913721413721414, "grad_norm": 0.5719657285080421, "learning_rate": 1.3562355690379455e-05, "loss": 0.7023, "step": 1715 }, { "epoch": 0.893970893970894, "grad_norm": 0.5445374932912297, "learning_rate": 1.3519935530208642e-05, "loss": 0.6969, "step": 1720 }, { "epoch": 0.8965696465696466, "grad_norm": 0.5332385586498977, "learning_rate": 1.347744294833442e-05, "loss": 0.7009, "step": 1725 }, { "epoch": 0.8991683991683992, "grad_norm": 0.5518972270290637, "learning_rate": 1.3434878819029949e-05, "loss": 0.6783, "step": 1730 }, { "epoch": 0.9017671517671517, "grad_norm": 0.5111471501880547, "learning_rate": 1.3392244018040459e-05, "loss": 0.6733, "step": 1735 }, { "epoch": 0.9043659043659044, "grad_norm": 0.5204670594020995, "learning_rate": 1.3349539422565227e-05, "loss": 0.6768, "step": 1740 }, { "epoch": 0.906964656964657, "grad_norm": 0.5185707128359598, "learning_rate": 1.3306765911239536e-05, "loss": 0.6836, "step": 1745 }, { "epoch": 0.9095634095634095, "grad_norm": 0.5371823929518809, "learning_rate": 1.3263924364116598e-05, "loss": 0.6757, "step": 1750 }, { "epoch": 0.9121621621621622, "grad_norm": 0.5512113239584037, "learning_rate": 1.3221015662649435e-05, "loss": 0.6874, "step": 1755 }, { "epoch": 0.9147609147609148, "grad_norm": 0.5413143469260978, "learning_rate": 1.3178040689672746e-05, "loss": 0.6903, "step": 1760 }, { "epoch": 0.9173596673596673, "grad_norm": 0.5448713291982036, "learning_rate": 1.3135000329384768e-05, "loss": 0.6749, "step": 1765 }, { "epoch": 0.91995841995842, "grad_norm": 0.5694338929447293, "learning_rate": 1.3091895467329043e-05, "loss": 0.6992, "step": 1770 }, { "epoch": 0.9225571725571725, "grad_norm": 0.5680751056661771, "learning_rate": 1.304872699037623e-05, "loss": 0.6778, "step": 1775 }, { "epoch": 0.9251559251559252, "grad_norm": 0.5827527502015905, "learning_rate": 1.3005495786705847e-05, "loss": 0.6801, "step": 1780 }, { "epoch": 0.9277546777546778, "grad_norm": 0.5414753069080744, "learning_rate": 1.2962202745787993e-05, "loss": 0.6828, "step": 1785 }, { "epoch": 0.9303534303534303, "grad_norm": 0.5954708948041362, "learning_rate": 1.2918848758365047e-05, "loss": 0.6804, "step": 1790 }, { "epoch": 0.932952182952183, "grad_norm": 0.4986624974413991, "learning_rate": 1.2875434716433364e-05, "loss": 0.6807, "step": 1795 }, { "epoch": 0.9355509355509356, "grad_norm": 0.512872956001043, "learning_rate": 1.2831961513224883e-05, "loss": 0.6758, "step": 1800 }, { "epoch": 0.9381496881496881, "grad_norm": 0.5424188290664714, "learning_rate": 1.278843004318878e-05, "loss": 0.6684, "step": 1805 }, { "epoch": 0.9407484407484408, "grad_norm": 0.5279446908433959, "learning_rate": 1.274484120197305e-05, "loss": 0.6732, "step": 1810 }, { "epoch": 0.9433471933471933, "grad_norm": 0.5019944493922829, "learning_rate": 1.2701195886406094e-05, "loss": 0.6736, "step": 1815 }, { "epoch": 0.9459459459459459, "grad_norm": 0.5441846652026447, "learning_rate": 1.2657494994478248e-05, "loss": 0.6803, "step": 1820 }, { "epoch": 0.9485446985446986, "grad_norm": 0.5027297117206215, "learning_rate": 1.2613739425323318e-05, "loss": 0.6916, "step": 1825 }, { "epoch": 0.9511434511434511, "grad_norm": 0.5208646951194689, "learning_rate": 1.2569930079200083e-05, "loss": 0.6674, "step": 1830 }, { "epoch": 0.9537422037422038, "grad_norm": 0.5229959541344918, "learning_rate": 1.2526067857473763e-05, "loss": 0.6808, "step": 1835 }, { "epoch": 0.9563409563409564, "grad_norm": 0.5436749548564572, "learning_rate": 1.2482153662597478e-05, "loss": 0.682, "step": 1840 }, { "epoch": 0.9589397089397089, "grad_norm": 0.525934850158874, "learning_rate": 1.2438188398093689e-05, "loss": 0.6684, "step": 1845 }, { "epoch": 0.9615384615384616, "grad_norm": 0.5241344769977555, "learning_rate": 1.2394172968535596e-05, "loss": 0.6845, "step": 1850 }, { "epoch": 0.9641372141372141, "grad_norm": 0.5334714104019, "learning_rate": 1.235010827952853e-05, "loss": 0.6752, "step": 1855 }, { "epoch": 0.9667359667359667, "grad_norm": 0.4981678215882384, "learning_rate": 1.230599523769132e-05, "loss": 0.6824, "step": 1860 }, { "epoch": 0.9693347193347194, "grad_norm": 0.529971709936305, "learning_rate": 1.2261834750637648e-05, "loss": 0.695, "step": 1865 }, { "epoch": 0.9719334719334719, "grad_norm": 0.5288630054972472, "learning_rate": 1.2217627726957362e-05, "loss": 0.6833, "step": 1870 }, { "epoch": 0.9745322245322245, "grad_norm": 0.5109511701940944, "learning_rate": 1.217337507619779e-05, "loss": 0.6688, "step": 1875 }, { "epoch": 0.9771309771309772, "grad_norm": 0.5251862301897194, "learning_rate": 1.212907770884503e-05, "loss": 0.6752, "step": 1880 }, { "epoch": 0.9797297297297297, "grad_norm": 0.5081841213925302, "learning_rate": 1.2084736536305199e-05, "loss": 0.6802, "step": 1885 }, { "epoch": 0.9823284823284824, "grad_norm": 0.5199130862705122, "learning_rate": 1.2040352470885705e-05, "loss": 0.6585, "step": 1890 }, { "epoch": 0.9849272349272349, "grad_norm": 0.5127627640317641, "learning_rate": 1.1995926425776463e-05, "loss": 0.6753, "step": 1895 }, { "epoch": 0.9875259875259875, "grad_norm": 0.5057020713500769, "learning_rate": 1.1951459315031103e-05, "loss": 0.6823, "step": 1900 }, { "epoch": 0.9901247401247402, "grad_norm": 0.5368825129232229, "learning_rate": 1.1906952053548173e-05, "loss": 0.6775, "step": 1905 }, { "epoch": 0.9927234927234927, "grad_norm": 0.48766377489485674, "learning_rate": 1.1862405557052316e-05, "loss": 0.6731, "step": 1910 }, { "epoch": 0.9953222453222453, "grad_norm": 0.5077162603005374, "learning_rate": 1.1817820742075418e-05, "loss": 0.659, "step": 1915 }, { "epoch": 0.997920997920998, "grad_norm": 0.5273871660362291, "learning_rate": 1.1773198525937758e-05, "loss": 0.679, "step": 1920 }, { "epoch": 1.0, "eval_loss": 0.7630624771118164, "eval_runtime": 104.7865, "eval_samples_per_second": 78.35, "eval_steps_per_second": 1.231, "step": 1924 }, { "epoch": 1.0005197505197505, "grad_norm": 0.7167016181446053, "learning_rate": 1.1728539826729135e-05, "loss": 0.6587, "step": 1925 }, { "epoch": 1.003118503118503, "grad_norm": 0.6041178102787718, "learning_rate": 1.168384556328998e-05, "loss": 0.6121, "step": 1930 }, { "epoch": 1.0057172557172558, "grad_norm": 0.5921694491638905, "learning_rate": 1.163911665519244e-05, "loss": 0.6028, "step": 1935 }, { "epoch": 1.0083160083160083, "grad_norm": 0.6013653429933059, "learning_rate": 1.1594354022721475e-05, "loss": 0.6156, "step": 1940 }, { "epoch": 1.0109147609147608, "grad_norm": 0.6105316954631931, "learning_rate": 1.1549558586855909e-05, "loss": 0.5945, "step": 1945 }, { "epoch": 1.0135135135135136, "grad_norm": 0.535707927809498, "learning_rate": 1.150473126924949e-05, "loss": 0.5922, "step": 1950 }, { "epoch": 1.0161122661122661, "grad_norm": 0.5556825438783977, "learning_rate": 1.1459872992211923e-05, "loss": 0.5953, "step": 1955 }, { "epoch": 1.0187110187110187, "grad_norm": 0.5691489620123985, "learning_rate": 1.1414984678689895e-05, "loss": 0.599, "step": 1960 }, { "epoch": 1.0213097713097714, "grad_norm": 0.5936266445205919, "learning_rate": 1.1370067252248085e-05, "loss": 0.5949, "step": 1965 }, { "epoch": 1.023908523908524, "grad_norm": 0.6066096499835257, "learning_rate": 1.1325121637050161e-05, "loss": 0.5971, "step": 1970 }, { "epoch": 1.0265072765072765, "grad_norm": 0.5681394506425539, "learning_rate": 1.128014875783977e-05, "loss": 0.5864, "step": 1975 }, { "epoch": 1.0291060291060292, "grad_norm": 0.5628786880019524, "learning_rate": 1.1235149539921509e-05, "loss": 0.592, "step": 1980 }, { "epoch": 1.0317047817047817, "grad_norm": 0.5467263882406993, "learning_rate": 1.1190124909141877e-05, "loss": 0.6075, "step": 1985 }, { "epoch": 1.0343035343035343, "grad_norm": 0.5440093360987802, "learning_rate": 1.1145075791870253e-05, "loss": 0.5988, "step": 1990 }, { "epoch": 1.0369022869022868, "grad_norm": 0.5752544838637329, "learning_rate": 1.1100003114979803e-05, "loss": 0.615, "step": 1995 }, { "epoch": 1.0395010395010396, "grad_norm": 0.5310516110561704, "learning_rate": 1.1054907805828427e-05, "loss": 0.6005, "step": 2000 }, { "epoch": 1.042099792099792, "grad_norm": 0.5432466126208855, "learning_rate": 1.1009790792239692e-05, "loss": 0.5866, "step": 2005 }, { "epoch": 1.0446985446985446, "grad_norm": 0.5644600605835878, "learning_rate": 1.0964653002483714e-05, "loss": 0.5936, "step": 2010 }, { "epoch": 1.0472972972972974, "grad_norm": 0.5272489276362843, "learning_rate": 1.0919495365258077e-05, "loss": 0.597, "step": 2015 }, { "epoch": 1.04989604989605, "grad_norm": 0.5617695811131541, "learning_rate": 1.0874318809668717e-05, "loss": 0.6065, "step": 2020 }, { "epoch": 1.0524948024948024, "grad_norm": 0.5192198208489455, "learning_rate": 1.0829124265210822e-05, "loss": 0.6075, "step": 2025 }, { "epoch": 1.0550935550935552, "grad_norm": 0.5440321607158946, "learning_rate": 1.0783912661749682e-05, "loss": 0.5993, "step": 2030 }, { "epoch": 1.0576923076923077, "grad_norm": 0.5380219431230713, "learning_rate": 1.0738684929501577e-05, "loss": 0.6047, "step": 2035 }, { "epoch": 1.0602910602910602, "grad_norm": 0.5309774372945627, "learning_rate": 1.069344199901464e-05, "loss": 0.6057, "step": 2040 }, { "epoch": 1.062889812889813, "grad_norm": 0.5317011424510519, "learning_rate": 1.064818480114969e-05, "loss": 0.6087, "step": 2045 }, { "epoch": 1.0654885654885655, "grad_norm": 0.5636006054084751, "learning_rate": 1.0602914267061102e-05, "loss": 0.6016, "step": 2050 }, { "epoch": 1.068087318087318, "grad_norm": 0.5385118881739976, "learning_rate": 1.0557631328177635e-05, "loss": 0.6009, "step": 2055 }, { "epoch": 1.0706860706860706, "grad_norm": 0.5387163558386009, "learning_rate": 1.0512336916183282e-05, "loss": 0.5919, "step": 2060 }, { "epoch": 1.0732848232848233, "grad_norm": 0.5550215843050625, "learning_rate": 1.046703196299808e-05, "loss": 0.6032, "step": 2065 }, { "epoch": 1.0758835758835759, "grad_norm": 0.5317572302029103, "learning_rate": 1.042171740075896e-05, "loss": 0.6061, "step": 2070 }, { "epoch": 1.0784823284823284, "grad_norm": 0.5953262222453396, "learning_rate": 1.037639416180055e-05, "loss": 0.5968, "step": 2075 }, { "epoch": 1.0810810810810811, "grad_norm": 0.5468172409024149, "learning_rate": 1.0331063178635991e-05, "loss": 0.5942, "step": 2080 }, { "epoch": 1.0836798336798337, "grad_norm": 0.5337399744353215, "learning_rate": 1.028572538393778e-05, "loss": 0.5784, "step": 2085 }, { "epoch": 1.0862785862785862, "grad_norm": 0.5918184869466311, "learning_rate": 1.0240381710518544e-05, "loss": 0.5825, "step": 2090 }, { "epoch": 1.088877338877339, "grad_norm": 0.5415874642045884, "learning_rate": 1.0195033091311866e-05, "loss": 0.6081, "step": 2095 }, { "epoch": 1.0914760914760915, "grad_norm": 0.5414796664701005, "learning_rate": 1.0149680459353085e-05, "loss": 0.5916, "step": 2100 }, { "epoch": 1.094074844074844, "grad_norm": 0.5523534552531586, "learning_rate": 1.0104324747760103e-05, "loss": 0.6108, "step": 2105 }, { "epoch": 1.0966735966735968, "grad_norm": 0.691647511565078, "learning_rate": 1.0058966889714192e-05, "loss": 0.6, "step": 2110 }, { "epoch": 1.0992723492723493, "grad_norm": 0.5912743887434098, "learning_rate": 1.0013607818440775e-05, "loss": 0.6006, "step": 2115 }, { "epoch": 1.1018711018711018, "grad_norm": 0.5460338682601912, "learning_rate": 9.968248467190246e-06, "loss": 0.6017, "step": 2120 }, { "epoch": 1.1044698544698546, "grad_norm": 0.5664759203624757, "learning_rate": 9.922889769218754e-06, "loss": 0.6093, "step": 2125 }, { "epoch": 1.107068607068607, "grad_norm": 0.5666035631954701, "learning_rate": 9.877532657769006e-06, "loss": 0.5971, "step": 2130 }, { "epoch": 1.1096673596673596, "grad_norm": 0.5200286477603056, "learning_rate": 9.832178066051074e-06, "loss": 0.5934, "step": 2135 }, { "epoch": 1.1122661122661124, "grad_norm": 0.6074145785434165, "learning_rate": 9.78682692722318e-06, "loss": 0.5886, "step": 2140 }, { "epoch": 1.114864864864865, "grad_norm": 0.5541608986138071, "learning_rate": 9.741480174372505e-06, "loss": 0.6028, "step": 2145 }, { "epoch": 1.1174636174636174, "grad_norm": 0.5324738113711789, "learning_rate": 9.696138740495992e-06, "loss": 0.5963, "step": 2150 }, { "epoch": 1.12006237006237, "grad_norm": 0.5848651297296711, "learning_rate": 9.650803558481147e-06, "loss": 0.5807, "step": 2155 }, { "epoch": 1.1226611226611227, "grad_norm": 0.5602080935436595, "learning_rate": 9.605475561086842e-06, "loss": 0.5888, "step": 2160 }, { "epoch": 1.1252598752598753, "grad_norm": 0.5653703789523455, "learning_rate": 9.560155680924137e-06, "loss": 0.5916, "step": 2165 }, { "epoch": 1.1278586278586278, "grad_norm": 0.5356042555591374, "learning_rate": 9.514844850437067e-06, "loss": 0.5877, "step": 2170 }, { "epoch": 1.1304573804573805, "grad_norm": 0.5682089324184239, "learning_rate": 9.469544001883491e-06, "loss": 0.5867, "step": 2175 }, { "epoch": 1.133056133056133, "grad_norm": 0.5334569975639222, "learning_rate": 9.424254067315875e-06, "loss": 0.6082, "step": 2180 }, { "epoch": 1.1356548856548856, "grad_norm": 0.6011244426639291, "learning_rate": 9.378975978562147e-06, "loss": 0.5934, "step": 2185 }, { "epoch": 1.1382536382536383, "grad_norm": 0.578376500402116, "learning_rate": 9.3337106672065e-06, "loss": 0.5836, "step": 2190 }, { "epoch": 1.1408523908523909, "grad_norm": 0.5588463035978978, "learning_rate": 9.288459064570239e-06, "loss": 0.6047, "step": 2195 }, { "epoch": 1.1434511434511434, "grad_norm": 0.5657302559519196, "learning_rate": 9.243222101692617e-06, "loss": 0.5858, "step": 2200 }, { "epoch": 1.1460498960498962, "grad_norm": 0.583919168737973, "learning_rate": 9.19800070931168e-06, "loss": 0.607, "step": 2205 }, { "epoch": 1.1486486486486487, "grad_norm": 0.5817136037512669, "learning_rate": 9.15279581784511e-06, "loss": 0.589, "step": 2210 }, { "epoch": 1.1512474012474012, "grad_norm": 0.538036225434835, "learning_rate": 9.107608357371092e-06, "loss": 0.6052, "step": 2215 }, { "epoch": 1.1538461538461537, "grad_norm": 0.5265174041614111, "learning_rate": 9.062439257609165e-06, "loss": 0.5953, "step": 2220 }, { "epoch": 1.1564449064449065, "grad_norm": 0.548975305771916, "learning_rate": 9.01728944790112e-06, "loss": 0.5931, "step": 2225 }, { "epoch": 1.159043659043659, "grad_norm": 0.5481727349763981, "learning_rate": 8.972159857191838e-06, "loss": 0.5951, "step": 2230 }, { "epoch": 1.1616424116424116, "grad_norm": 0.5808749569289294, "learning_rate": 8.927051414010213e-06, "loss": 0.5925, "step": 2235 }, { "epoch": 1.1642411642411643, "grad_norm": 0.5597977821004193, "learning_rate": 8.881965046450033e-06, "loss": 0.5983, "step": 2240 }, { "epoch": 1.1668399168399168, "grad_norm": 0.5732606987628547, "learning_rate": 8.83690168215088e-06, "loss": 0.6003, "step": 2245 }, { "epoch": 1.1694386694386694, "grad_norm": 0.5317615005846271, "learning_rate": 8.791862248279059e-06, "loss": 0.6011, "step": 2250 }, { "epoch": 1.1720374220374221, "grad_norm": 0.5175378118538798, "learning_rate": 8.746847671508506e-06, "loss": 0.5906, "step": 2255 }, { "epoch": 1.1746361746361746, "grad_norm": 0.5827570884628767, "learning_rate": 8.701858878001731e-06, "loss": 0.5947, "step": 2260 }, { "epoch": 1.1772349272349272, "grad_norm": 0.5832582553430286, "learning_rate": 8.656896793390763e-06, "loss": 0.6021, "step": 2265 }, { "epoch": 1.17983367983368, "grad_norm": 0.5252465600060544, "learning_rate": 8.6119623427581e-06, "loss": 0.5946, "step": 2270 }, { "epoch": 1.1824324324324325, "grad_norm": 0.5535883003187351, "learning_rate": 8.567056450617678e-06, "loss": 0.5997, "step": 2275 }, { "epoch": 1.185031185031185, "grad_norm": 0.5265834205804355, "learning_rate": 8.522180040895855e-06, "loss": 0.5919, "step": 2280 }, { "epoch": 1.1876299376299375, "grad_norm": 0.5588080207098279, "learning_rate": 8.477334036912392e-06, "loss": 0.599, "step": 2285 }, { "epoch": 1.1902286902286903, "grad_norm": 0.5254040822916115, "learning_rate": 8.432519361361458e-06, "loss": 0.5947, "step": 2290 }, { "epoch": 1.1928274428274428, "grad_norm": 0.5455269093593619, "learning_rate": 8.387736936292654e-06, "loss": 0.5939, "step": 2295 }, { "epoch": 1.1954261954261955, "grad_norm": 0.5639163666347963, "learning_rate": 8.342987683092033e-06, "loss": 0.5949, "step": 2300 }, { "epoch": 1.198024948024948, "grad_norm": 0.6084945001801156, "learning_rate": 8.298272522463144e-06, "loss": 0.596, "step": 2305 }, { "epoch": 1.2006237006237006, "grad_norm": 0.5836727328380208, "learning_rate": 8.253592374408095e-06, "loss": 0.5963, "step": 2310 }, { "epoch": 1.2032224532224531, "grad_norm": 0.5454765026633895, "learning_rate": 8.20894815820862e-06, "loss": 0.6003, "step": 2315 }, { "epoch": 1.2058212058212059, "grad_norm": 0.5490153680134165, "learning_rate": 8.164340792407159e-06, "loss": 0.594, "step": 2320 }, { "epoch": 1.2084199584199584, "grad_norm": 0.5442982828513857, "learning_rate": 8.119771194787966e-06, "loss": 0.5939, "step": 2325 }, { "epoch": 1.211018711018711, "grad_norm": 0.5224447370874735, "learning_rate": 8.07524028235824e-06, "loss": 0.5895, "step": 2330 }, { "epoch": 1.2136174636174637, "grad_norm": 0.5648657256694566, "learning_rate": 8.030748971329224e-06, "loss": 0.6101, "step": 2335 }, { "epoch": 1.2162162162162162, "grad_norm": 0.5616352744995109, "learning_rate": 7.986298177097377e-06, "loss": 0.6042, "step": 2340 }, { "epoch": 1.2188149688149688, "grad_norm": 0.5771472821711124, "learning_rate": 7.941888814225545e-06, "loss": 0.5987, "step": 2345 }, { "epoch": 1.2214137214137215, "grad_norm": 0.5512099431582866, "learning_rate": 7.897521796424129e-06, "loss": 0.6056, "step": 2350 }, { "epoch": 1.224012474012474, "grad_norm": 0.5098551231289585, "learning_rate": 7.853198036532288e-06, "loss": 0.6031, "step": 2355 }, { "epoch": 1.2266112266112266, "grad_norm": 0.5717033338675165, "learning_rate": 7.80891844649917e-06, "loss": 0.592, "step": 2360 }, { "epoch": 1.2292099792099793, "grad_norm": 0.5550621676748785, "learning_rate": 7.764683937365134e-06, "loss": 0.5877, "step": 2365 }, { "epoch": 1.2318087318087318, "grad_norm": 0.5823877655186512, "learning_rate": 7.720495419243014e-06, "loss": 0.6226, "step": 2370 }, { "epoch": 1.2344074844074844, "grad_norm": 0.547937317341142, "learning_rate": 7.676353801299388e-06, "loss": 0.5936, "step": 2375 }, { "epoch": 1.237006237006237, "grad_norm": 0.5341568281577489, "learning_rate": 7.63225999173589e-06, "loss": 0.5899, "step": 2380 }, { "epoch": 1.2396049896049897, "grad_norm": 0.5432593899885211, "learning_rate": 7.5882148977704914e-06, "loss": 0.602, "step": 2385 }, { "epoch": 1.2422037422037422, "grad_norm": 0.5421541547325941, "learning_rate": 7.544219425618862e-06, "loss": 0.5858, "step": 2390 }, { "epoch": 1.2448024948024947, "grad_norm": 0.5370495755166194, "learning_rate": 7.500274480475716e-06, "loss": 0.6071, "step": 2395 }, { "epoch": 1.2474012474012475, "grad_norm": 0.5449432404989222, "learning_rate": 7.45638096649619e-06, "loss": 0.5972, "step": 2400 }, { "epoch": 1.25, "grad_norm": 0.5660658221047755, "learning_rate": 7.412539786777234e-06, "loss": 0.5943, "step": 2405 }, { "epoch": 1.2525987525987525, "grad_norm": 0.5247570244603685, "learning_rate": 7.3687518433390414e-06, "loss": 0.5924, "step": 2410 }, { "epoch": 1.255197505197505, "grad_norm": 0.5207188548015057, "learning_rate": 7.325018037106474e-06, "loss": 0.5917, "step": 2415 }, { "epoch": 1.2577962577962578, "grad_norm": 0.550416355369781, "learning_rate": 7.281339267890548e-06, "loss": 0.5857, "step": 2420 }, { "epoch": 1.2603950103950103, "grad_norm": 0.5155652437752087, "learning_rate": 7.237716434369899e-06, "loss": 0.5946, "step": 2425 }, { "epoch": 1.262993762993763, "grad_norm": 0.5446737196259831, "learning_rate": 7.1941504340723e-06, "loss": 0.602, "step": 2430 }, { "epoch": 1.2655925155925156, "grad_norm": 0.5241476632554306, "learning_rate": 7.150642163356207e-06, "loss": 0.5848, "step": 2435 }, { "epoch": 1.2681912681912682, "grad_norm": 0.5368234251185683, "learning_rate": 7.107192517392296e-06, "loss": 0.593, "step": 2440 }, { "epoch": 1.2707900207900207, "grad_norm": 0.5609013153184366, "learning_rate": 7.063802390145058e-06, "loss": 0.5926, "step": 2445 }, { "epoch": 1.2733887733887734, "grad_norm": 0.5601427654877619, "learning_rate": 7.020472674354399e-06, "loss": 0.5955, "step": 2450 }, { "epoch": 1.275987525987526, "grad_norm": 0.5573446244898519, "learning_rate": 6.9772042615172855e-06, "loss": 0.596, "step": 2455 }, { "epoch": 1.2785862785862787, "grad_norm": 0.5409848408904747, "learning_rate": 6.933998041869381e-06, "loss": 0.5935, "step": 2460 }, { "epoch": 1.2811850311850312, "grad_norm": 0.535048843399521, "learning_rate": 6.890854904366749e-06, "loss": 0.5887, "step": 2465 }, { "epoch": 1.2837837837837838, "grad_norm": 0.5178312289683921, "learning_rate": 6.8477757366675526e-06, "loss": 0.5915, "step": 2470 }, { "epoch": 1.2863825363825363, "grad_norm": 0.5746689510719801, "learning_rate": 6.804761425113796e-06, "loss": 0.5952, "step": 2475 }, { "epoch": 1.288981288981289, "grad_norm": 0.5494461776895628, "learning_rate": 6.761812854713079e-06, "loss": 0.5905, "step": 2480 }, { "epoch": 1.2915800415800416, "grad_norm": 0.5437436470618007, "learning_rate": 6.718930909120404e-06, "loss": 0.5881, "step": 2485 }, { "epoch": 1.2941787941787941, "grad_norm": 0.5670019639541221, "learning_rate": 6.676116470619988e-06, "loss": 0.5945, "step": 2490 }, { "epoch": 1.2967775467775469, "grad_norm": 0.5754824812127866, "learning_rate": 6.633370420107097e-06, "loss": 0.5975, "step": 2495 }, { "epoch": 1.2993762993762994, "grad_norm": 0.548088254326578, "learning_rate": 6.590693637069938e-06, "loss": 0.5966, "step": 2500 }, { "epoch": 1.301975051975052, "grad_norm": 0.5243431037903644, "learning_rate": 6.548086999571566e-06, "loss": 0.5938, "step": 2505 }, { "epoch": 1.3045738045738045, "grad_norm": 0.565611456496591, "learning_rate": 6.505551384231801e-06, "loss": 0.592, "step": 2510 }, { "epoch": 1.3071725571725572, "grad_norm": 0.561890682944344, "learning_rate": 6.463087666209203e-06, "loss": 0.5818, "step": 2515 }, { "epoch": 1.3097713097713097, "grad_norm": 0.5872998314400075, "learning_rate": 6.420696719183072e-06, "loss": 0.5849, "step": 2520 }, { "epoch": 1.3123700623700625, "grad_norm": 0.5165776292315633, "learning_rate": 6.378379415335456e-06, "loss": 0.5913, "step": 2525 }, { "epoch": 1.314968814968815, "grad_norm": 0.5271295585343699, "learning_rate": 6.336136625333218e-06, "loss": 0.6001, "step": 2530 }, { "epoch": 1.3175675675675675, "grad_norm": 0.5541093558682746, "learning_rate": 6.293969218310122e-06, "loss": 0.5986, "step": 2535 }, { "epoch": 1.32016632016632, "grad_norm": 0.533798380054359, "learning_rate": 6.251878061848948e-06, "loss": 0.5759, "step": 2540 }, { "epoch": 1.3227650727650728, "grad_norm": 0.5573470744869463, "learning_rate": 6.209864021963638e-06, "loss": 0.5879, "step": 2545 }, { "epoch": 1.3253638253638254, "grad_norm": 0.5493774653074921, "learning_rate": 6.16792796308148e-06, "loss": 0.6017, "step": 2550 }, { "epoch": 1.3279625779625779, "grad_norm": 0.582071520963931, "learning_rate": 6.126070748025332e-06, "loss": 0.6025, "step": 2555 }, { "epoch": 1.3305613305613306, "grad_norm": 0.5144065189940694, "learning_rate": 6.084293237995855e-06, "loss": 0.5827, "step": 2560 }, { "epoch": 1.3331600831600832, "grad_norm": 0.5139551250609696, "learning_rate": 6.042596292553803e-06, "loss": 0.585, "step": 2565 }, { "epoch": 1.3357588357588357, "grad_norm": 0.512609402226577, "learning_rate": 6.0009807696023385e-06, "loss": 0.5865, "step": 2570 }, { "epoch": 1.3383575883575882, "grad_norm": 0.5648181415304806, "learning_rate": 5.959447525369375e-06, "loss": 0.585, "step": 2575 }, { "epoch": 1.340956340956341, "grad_norm": 0.4961717842990021, "learning_rate": 5.917997414389966e-06, "loss": 0.6065, "step": 2580 }, { "epoch": 1.3435550935550935, "grad_norm": 0.5369467703030896, "learning_rate": 5.87663128948872e-06, "loss": 0.586, "step": 2585 }, { "epoch": 1.3461538461538463, "grad_norm": 0.5805263679243358, "learning_rate": 5.835350001762258e-06, "loss": 0.5793, "step": 2590 }, { "epoch": 1.3487525987525988, "grad_norm": 0.5472679842232231, "learning_rate": 5.794154400561698e-06, "loss": 0.5936, "step": 2595 }, { "epoch": 1.3513513513513513, "grad_norm": 0.543074701262151, "learning_rate": 5.753045333475181e-06, "loss": 0.6068, "step": 2600 }, { "epoch": 1.3539501039501038, "grad_norm": 0.501759332855818, "learning_rate": 5.712023646310429e-06, "loss": 0.5923, "step": 2605 }, { "epoch": 1.3565488565488566, "grad_norm": 0.5627676287799811, "learning_rate": 5.671090183077355e-06, "loss": 0.5908, "step": 2610 }, { "epoch": 1.3591476091476091, "grad_norm": 0.5396141346304916, "learning_rate": 5.630245785970676e-06, "loss": 0.5835, "step": 2615 }, { "epoch": 1.3617463617463619, "grad_norm": 0.5237498968336539, "learning_rate": 5.589491295352607e-06, "loss": 0.6023, "step": 2620 }, { "epoch": 1.3643451143451144, "grad_norm": 0.5714671269121867, "learning_rate": 5.548827549735565e-06, "loss": 0.5875, "step": 2625 }, { "epoch": 1.366943866943867, "grad_norm": 0.5420021477246512, "learning_rate": 5.508255385764908e-06, "loss": 0.5962, "step": 2630 }, { "epoch": 1.3695426195426195, "grad_norm": 0.5213053662389652, "learning_rate": 5.46777563820172e-06, "loss": 0.5775, "step": 2635 }, { "epoch": 1.3721413721413722, "grad_norm": 0.5940088174557574, "learning_rate": 5.427389139905661e-06, "loss": 0.5925, "step": 2640 }, { "epoch": 1.3747401247401247, "grad_norm": 0.5222712044920537, "learning_rate": 5.387096721817799e-06, "loss": 0.5939, "step": 2645 }, { "epoch": 1.3773388773388773, "grad_norm": 0.533759193353919, "learning_rate": 5.346899212943529e-06, "loss": 0.5891, "step": 2650 }, { "epoch": 1.37993762993763, "grad_norm": 0.5684539829097972, "learning_rate": 5.306797440335516e-06, "loss": 0.598, "step": 2655 }, { "epoch": 1.3825363825363826, "grad_norm": 0.577296227438705, "learning_rate": 5.266792229076683e-06, "loss": 0.5908, "step": 2660 }, { "epoch": 1.385135135135135, "grad_norm": 0.5367239875824128, "learning_rate": 5.2268844022632255e-06, "loss": 0.5933, "step": 2665 }, { "epoch": 1.3877338877338876, "grad_norm": 0.5512610722369685, "learning_rate": 5.187074780987682e-06, "loss": 0.5952, "step": 2670 }, { "epoch": 1.3903326403326404, "grad_norm": 0.5045101270772827, "learning_rate": 5.147364184322035e-06, "loss": 0.5793, "step": 2675 }, { "epoch": 1.392931392931393, "grad_norm": 0.5505398932307536, "learning_rate": 5.1077534293008635e-06, "loss": 0.5973, "step": 2680 }, { "epoch": 1.3955301455301456, "grad_norm": 0.5618823440106911, "learning_rate": 5.068243330904543e-06, "loss": 0.5954, "step": 2685 }, { "epoch": 1.3981288981288982, "grad_norm": 0.547767493340866, "learning_rate": 5.028834702042451e-06, "loss": 0.5965, "step": 2690 }, { "epoch": 1.4007276507276507, "grad_norm": 0.5311004214157612, "learning_rate": 4.9895283535362744e-06, "loss": 0.5761, "step": 2695 }, { "epoch": 1.4033264033264032, "grad_norm": 0.5522752500932042, "learning_rate": 4.9503250941032944e-06, "loss": 0.5858, "step": 2700 }, { "epoch": 1.405925155925156, "grad_norm": 0.5236156885167432, "learning_rate": 4.911225730339769e-06, "loss": 0.5949, "step": 2705 }, { "epoch": 1.4085239085239085, "grad_norm": 0.5167097373944585, "learning_rate": 4.872231066704334e-06, "loss": 0.5961, "step": 2710 }, { "epoch": 1.411122661122661, "grad_norm": 0.563080292650056, "learning_rate": 4.83334190550144e-06, "loss": 0.5749, "step": 2715 }, { "epoch": 1.4137214137214138, "grad_norm": 0.5327245199013042, "learning_rate": 4.794559046864869e-06, "loss": 0.5862, "step": 2720 }, { "epoch": 1.4163201663201663, "grad_norm": 0.5289775705397598, "learning_rate": 4.7558832887412445e-06, "loss": 0.5877, "step": 2725 }, { "epoch": 1.4189189189189189, "grad_norm": 0.5162142489701083, "learning_rate": 4.717315426873631e-06, "loss": 0.601, "step": 2730 }, { "epoch": 1.4215176715176714, "grad_norm": 0.5405562294471293, "learning_rate": 4.678856254785158e-06, "loss": 0.6065, "step": 2735 }, { "epoch": 1.4241164241164241, "grad_norm": 0.534465583683509, "learning_rate": 4.6405065637626855e-06, "loss": 0.5871, "step": 2740 }, { "epoch": 1.4267151767151767, "grad_norm": 0.5179677159166826, "learning_rate": 4.602267142840545e-06, "loss": 0.5873, "step": 2745 }, { "epoch": 1.4293139293139294, "grad_norm": 0.5480531904822096, "learning_rate": 4.564138778784276e-06, "loss": 0.5895, "step": 2750 }, { "epoch": 1.431912681912682, "grad_norm": 0.5288015351179746, "learning_rate": 4.526122256074468e-06, "loss": 0.588, "step": 2755 }, { "epoch": 1.4345114345114345, "grad_norm": 0.5232295069807825, "learning_rate": 4.488218356890594e-06, "loss": 0.5802, "step": 2760 }, { "epoch": 1.437110187110187, "grad_norm": 0.5460332455155361, "learning_rate": 4.450427861094933e-06, "loss": 0.5809, "step": 2765 }, { "epoch": 1.4397089397089398, "grad_norm": 0.5451685990728384, "learning_rate": 4.412751546216521e-06, "loss": 0.5879, "step": 2770 }, { "epoch": 1.4423076923076923, "grad_norm": 0.5068743803171689, "learning_rate": 4.375190187435151e-06, "loss": 0.5819, "step": 2775 }, { "epoch": 1.444906444906445, "grad_norm": 0.602224317860505, "learning_rate": 4.3377445575654256e-06, "loss": 0.5808, "step": 2780 }, { "epoch": 1.4475051975051976, "grad_norm": 0.5584038619724856, "learning_rate": 4.300415427040867e-06, "loss": 0.586, "step": 2785 }, { "epoch": 1.45010395010395, "grad_norm": 0.5402693196977059, "learning_rate": 4.263203563898038e-06, "loss": 0.5828, "step": 2790 }, { "epoch": 1.4527027027027026, "grad_norm": 0.5393224079014185, "learning_rate": 4.226109733760777e-06, "loss": 0.5933, "step": 2795 }, { "epoch": 1.4553014553014554, "grad_norm": 0.519264378128423, "learning_rate": 4.1891346998244096e-06, "loss": 0.5902, "step": 2800 }, { "epoch": 1.457900207900208, "grad_norm": 0.5061001580587425, "learning_rate": 4.15227922284007e-06, "loss": 0.5965, "step": 2805 }, { "epoch": 1.4604989604989604, "grad_norm": 0.5250324465767509, "learning_rate": 4.11554406109904e-06, "loss": 0.5805, "step": 2810 }, { "epoch": 1.4630977130977132, "grad_norm": 0.5215691344217085, "learning_rate": 4.07892997041714e-06, "loss": 0.5847, "step": 2815 }, { "epoch": 1.4656964656964657, "grad_norm": 0.5100852244380407, "learning_rate": 4.042437704119207e-06, "loss": 0.5863, "step": 2820 }, { "epoch": 1.4682952182952183, "grad_norm": 0.5683610331681553, "learning_rate": 4.006068013023556e-06, "loss": 0.5916, "step": 2825 }, { "epoch": 1.4708939708939708, "grad_norm": 0.5388381909277385, "learning_rate": 3.969821645426559e-06, "loss": 0.5879, "step": 2830 }, { "epoch": 1.4734927234927235, "grad_norm": 0.5470869871655806, "learning_rate": 3.933699347087239e-06, "loss": 0.579, "step": 2835 }, { "epoch": 1.476091476091476, "grad_norm": 0.5410808556361425, "learning_rate": 3.897701861211924e-06, "loss": 0.5882, "step": 2840 }, { "epoch": 1.4786902286902288, "grad_norm": 0.5445488241193585, "learning_rate": 3.861829928438975e-06, "loss": 0.5745, "step": 2845 }, { "epoch": 1.4812889812889813, "grad_norm": 0.5601094797596367, "learning_rate": 3.8260842868235105e-06, "loss": 0.5845, "step": 2850 }, { "epoch": 1.4838877338877339, "grad_norm": 0.5108174266653619, "learning_rate": 3.7904656718222653e-06, "loss": 0.5817, "step": 2855 }, { "epoch": 1.4864864864864864, "grad_norm": 0.5380258071504634, "learning_rate": 3.7549748162784216e-06, "loss": 0.5734, "step": 2860 }, { "epoch": 1.4890852390852392, "grad_norm": 0.5056247324728291, "learning_rate": 3.7196124504065512e-06, "loss": 0.5837, "step": 2865 }, { "epoch": 1.4916839916839917, "grad_norm": 0.5164676599197356, "learning_rate": 3.684379301777585e-06, "loss": 0.5737, "step": 2870 }, { "epoch": 1.4942827442827442, "grad_norm": 0.5588163356512155, "learning_rate": 3.649276095303843e-06, "loss": 0.5896, "step": 2875 }, { "epoch": 1.496881496881497, "grad_norm": 0.4785867733453047, "learning_rate": 3.61430355322413e-06, "loss": 0.5829, "step": 2880 }, { "epoch": 1.4994802494802495, "grad_norm": 0.5236230514577889, "learning_rate": 3.579462395088856e-06, "loss": 0.5943, "step": 2885 }, { "epoch": 1.502079002079002, "grad_norm": 0.5243126400259197, "learning_rate": 3.544753337745249e-06, "loss": 0.583, "step": 2890 }, { "epoch": 1.5046777546777546, "grad_norm": 0.5400667626881714, "learning_rate": 3.5101770953225932e-06, "loss": 0.5868, "step": 2895 }, { "epoch": 1.5072765072765073, "grad_norm": 0.513023426174068, "learning_rate": 3.475734379217555e-06, "loss": 0.5833, "step": 2900 }, { "epoch": 1.5098752598752598, "grad_norm": 0.5157665498993713, "learning_rate": 3.4414258980795202e-06, "loss": 0.589, "step": 2905 }, { "epoch": 1.5124740124740126, "grad_norm": 0.5357953624143235, "learning_rate": 3.4072523577960304e-06, "loss": 0.5964, "step": 2910 }, { "epoch": 1.5150727650727651, "grad_norm": 0.5527070418499459, "learning_rate": 3.3732144614782657e-06, "loss": 0.5837, "step": 2915 }, { "epoch": 1.5176715176715176, "grad_norm": 0.5458675642969252, "learning_rate": 3.339312909446557e-06, "loss": 0.5796, "step": 2920 }, { "epoch": 1.5202702702702702, "grad_norm": 0.5152283307254787, "learning_rate": 3.305548399215994e-06, "loss": 0.5788, "step": 2925 }, { "epoch": 1.5228690228690227, "grad_norm": 0.530799104763713, "learning_rate": 3.2719216254820697e-06, "loss": 0.5838, "step": 2930 }, { "epoch": 1.5254677754677755, "grad_norm": 0.5403421099817592, "learning_rate": 3.2384332801063846e-06, "loss": 0.5656, "step": 2935 }, { "epoch": 1.5280665280665282, "grad_norm": 0.558670507000079, "learning_rate": 3.205084052102414e-06, "loss": 0.588, "step": 2940 }, { "epoch": 1.5306652806652807, "grad_norm": 0.5087922370395622, "learning_rate": 3.1718746276213343e-06, "loss": 0.5852, "step": 2945 }, { "epoch": 1.5332640332640333, "grad_norm": 0.5482047516259143, "learning_rate": 3.1388056899379036e-06, "loss": 0.5946, "step": 2950 }, { "epoch": 1.5358627858627858, "grad_norm": 0.5239979050184048, "learning_rate": 3.105877919436401e-06, "loss": 0.5961, "step": 2955 }, { "epoch": 1.5384615384615383, "grad_norm": 0.6264561692888647, "learning_rate": 3.0730919935966262e-06, "loss": 0.5896, "step": 2960 }, { "epoch": 1.541060291060291, "grad_norm": 0.5079701430841874, "learning_rate": 3.0404485869799684e-06, "loss": 0.5804, "step": 2965 }, { "epoch": 1.5436590436590436, "grad_norm": 0.5227318565543926, "learning_rate": 3.007948371215521e-06, "loss": 0.5881, "step": 2970 }, { "epoch": 1.5462577962577964, "grad_norm": 0.4870518164479936, "learning_rate": 2.975592014986265e-06, "loss": 0.5952, "step": 2975 }, { "epoch": 1.5488565488565489, "grad_norm": 0.5650630749989504, "learning_rate": 2.9433801840153166e-06, "loss": 0.5836, "step": 2980 }, { "epoch": 1.5514553014553014, "grad_norm": 0.5358798465667378, "learning_rate": 2.9113135410522173e-06, "loss": 0.5962, "step": 2985 }, { "epoch": 1.554054054054054, "grad_norm": 0.5158817153128504, "learning_rate": 2.8793927458593096e-06, "loss": 0.57, "step": 2990 }, { "epoch": 1.5566528066528067, "grad_norm": 0.5442534042057305, "learning_rate": 2.8476184551981547e-06, "loss": 0.5821, "step": 2995 }, { "epoch": 1.5592515592515592, "grad_norm": 0.5778592324539092, "learning_rate": 2.8159913228160242e-06, "loss": 0.5862, "step": 3000 }, { "epoch": 1.561850311850312, "grad_norm": 0.545713870580316, "learning_rate": 2.7845119994324567e-06, "loss": 0.5837, "step": 3005 }, { "epoch": 1.5644490644490645, "grad_norm": 0.5027605038010509, "learning_rate": 2.753181132725846e-06, "loss": 0.5908, "step": 3010 }, { "epoch": 1.567047817047817, "grad_norm": 0.5296784509548438, "learning_rate": 2.7219993673201483e-06, "loss": 0.5845, "step": 3015 }, { "epoch": 1.5696465696465696, "grad_norm": 0.5434115009675748, "learning_rate": 2.69096734477159e-06, "loss": 0.5761, "step": 3020 }, { "epoch": 1.572245322245322, "grad_norm": 0.5490780089111318, "learning_rate": 2.6600857035554805e-06, "loss": 0.5759, "step": 3025 }, { "epoch": 1.5748440748440748, "grad_norm": 0.5434069633378591, "learning_rate": 2.6293550790530776e-06, "loss": 0.5856, "step": 3030 }, { "epoch": 1.5774428274428276, "grad_norm": 0.5295051711364253, "learning_rate": 2.5987761035385074e-06, "loss": 0.5776, "step": 3035 }, { "epoch": 1.5800415800415801, "grad_norm": 0.5294382951858779, "learning_rate": 2.5683494061657698e-06, "loss": 0.6003, "step": 3040 }, { "epoch": 1.5826403326403327, "grad_norm": 0.5487264444738784, "learning_rate": 2.538075612955775e-06, "loss": 0.5905, "step": 3045 }, { "epoch": 1.5852390852390852, "grad_norm": 0.5151850538678816, "learning_rate": 2.5079553467834694e-06, "loss": 0.5945, "step": 3050 }, { "epoch": 1.5878378378378377, "grad_norm": 0.5241891607063889, "learning_rate": 2.4779892273650363e-06, "loss": 0.5809, "step": 3055 }, { "epoch": 1.5904365904365905, "grad_norm": 0.5490608540434708, "learning_rate": 2.44817787124512e-06, "loss": 0.5864, "step": 3060 }, { "epoch": 1.593035343035343, "grad_norm": 0.5085552474004695, "learning_rate": 2.4185218917841557e-06, "loss": 0.5839, "step": 3065 }, { "epoch": 1.5956340956340958, "grad_norm": 0.5237874429368263, "learning_rate": 2.389021899145745e-06, "loss": 0.583, "step": 3070 }, { "epoch": 1.5982328482328483, "grad_norm": 0.5462699535427128, "learning_rate": 2.3596785002841126e-06, "loss": 0.582, "step": 3075 }, { "epoch": 1.6008316008316008, "grad_norm": 0.5305897465130409, "learning_rate": 2.330492298931599e-06, "loss": 0.588, "step": 3080 }, { "epoch": 1.6034303534303533, "grad_norm": 0.5283228272742915, "learning_rate": 2.3014638955862534e-06, "loss": 0.5895, "step": 3085 }, { "epoch": 1.6060291060291059, "grad_norm": 0.5375248398279345, "learning_rate": 2.2725938874994733e-06, "loss": 0.5712, "step": 3090 }, { "epoch": 1.6086278586278586, "grad_norm": 0.5387607107517319, "learning_rate": 2.2438828686637193e-06, "loss": 0.5717, "step": 3095 }, { "epoch": 1.6112266112266114, "grad_norm": 0.5333070714415671, "learning_rate": 2.2153314298002883e-06, "loss": 0.5869, "step": 3100 }, { "epoch": 1.613825363825364, "grad_norm": 0.5130040335910431, "learning_rate": 2.1869401583471674e-06, "loss": 0.6017, "step": 3105 }, { "epoch": 1.6164241164241164, "grad_norm": 0.5340427602820074, "learning_rate": 2.1587096384469465e-06, "loss": 0.5792, "step": 3110 }, { "epoch": 1.619022869022869, "grad_norm": 0.5241334616500023, "learning_rate": 2.130640450934787e-06, "loss": 0.5766, "step": 3115 }, { "epoch": 1.6216216216216215, "grad_norm": 0.5307440437206696, "learning_rate": 2.1027331733264877e-06, "loss": 0.5841, "step": 3120 }, { "epoch": 1.6242203742203742, "grad_norm": 0.5111869577979938, "learning_rate": 2.074988379806593e-06, "loss": 0.5772, "step": 3125 }, { "epoch": 1.6268191268191268, "grad_norm": 0.5232894395546558, "learning_rate": 2.047406641216583e-06, "loss": 0.5798, "step": 3130 }, { "epoch": 1.6294178794178795, "grad_norm": 0.5427662431229768, "learning_rate": 2.019988525043125e-06, "loss": 0.5852, "step": 3135 }, { "epoch": 1.632016632016632, "grad_norm": 0.50753581880278, "learning_rate": 1.992734595406408e-06, "loss": 0.5756, "step": 3140 }, { "epoch": 1.6346153846153846, "grad_norm": 0.5026065449648726, "learning_rate": 1.965645413048519e-06, "loss": 0.5756, "step": 3145 }, { "epoch": 1.637214137214137, "grad_norm": 0.5135278822828021, "learning_rate": 1.9387215353219213e-06, "loss": 0.6112, "step": 3150 }, { "epoch": 1.6398128898128899, "grad_norm": 0.5576890383464939, "learning_rate": 1.9119635161779738e-06, "loss": 0.5933, "step": 3155 }, { "epoch": 1.6424116424116424, "grad_norm": 0.5005852274685088, "learning_rate": 1.885371906155552e-06, "loss": 0.575, "step": 3160 }, { "epoch": 1.6450103950103951, "grad_norm": 0.552519914509362, "learning_rate": 1.8589472523697017e-06, "loss": 0.5832, "step": 3165 }, { "epoch": 1.6476091476091477, "grad_norm": 0.5313038853576302, "learning_rate": 1.832690098500386e-06, "loss": 0.5799, "step": 3170 }, { "epoch": 1.6502079002079002, "grad_norm": 0.5307430737754275, "learning_rate": 1.8066009847813171e-06, "loss": 0.5933, "step": 3175 }, { "epoch": 1.6528066528066527, "grad_norm": 0.5092540679964042, "learning_rate": 1.7806804479888151e-06, "loss": 0.5857, "step": 3180 }, { "epoch": 1.6554054054054053, "grad_norm": 0.527226302557591, "learning_rate": 1.7549290214307824e-06, "loss": 0.5826, "step": 3185 }, { "epoch": 1.658004158004158, "grad_norm": 0.5117048033630656, "learning_rate": 1.729347234935721e-06, "loss": 0.5866, "step": 3190 }, { "epoch": 1.6606029106029108, "grad_norm": 0.5338263684846878, "learning_rate": 1.7039356148418363e-06, "loss": 0.5815, "step": 3195 }, { "epoch": 1.6632016632016633, "grad_norm": 0.514319228309042, "learning_rate": 1.6786946839862127e-06, "loss": 0.5704, "step": 3200 }, { "epoch": 1.6658004158004158, "grad_norm": 0.5093476576764645, "learning_rate": 1.6536249616940414e-06, "loss": 0.5861, "step": 3205 }, { "epoch": 1.6683991683991684, "grad_norm": 0.507273641369432, "learning_rate": 1.6287269637679538e-06, "loss": 0.5827, "step": 3210 }, { "epoch": 1.6709979209979209, "grad_norm": 0.5145115612127696, "learning_rate": 1.604001202477391e-06, "loss": 0.5768, "step": 3215 }, { "epoch": 1.6735966735966736, "grad_norm": 0.5149567719955702, "learning_rate": 1.5794481865480805e-06, "loss": 0.581, "step": 3220 }, { "epoch": 1.6761954261954262, "grad_norm": 0.5471640458477693, "learning_rate": 1.5550684211515544e-06, "loss": 0.6, "step": 3225 }, { "epoch": 1.678794178794179, "grad_norm": 0.5394179199712156, "learning_rate": 1.5308624078947666e-06, "loss": 0.5801, "step": 3230 }, { "epoch": 1.6813929313929314, "grad_norm": 0.506276316904012, "learning_rate": 1.5068306448097714e-06, "loss": 0.5689, "step": 3235 }, { "epoch": 1.683991683991684, "grad_norm": 0.5038655785707594, "learning_rate": 1.4829736263434702e-06, "loss": 0.5907, "step": 3240 }, { "epoch": 1.6865904365904365, "grad_norm": 0.5272673427698651, "learning_rate": 1.4592918433474424e-06, "loss": 0.586, "step": 3245 }, { "epoch": 1.689189189189189, "grad_norm": 0.5298471417100465, "learning_rate": 1.4357857830678445e-06, "loss": 0.5937, "step": 3250 }, { "epoch": 1.6917879417879418, "grad_norm": 0.5181589625133596, "learning_rate": 1.4124559291353878e-06, "loss": 0.5679, "step": 3255 }, { "epoch": 1.6943866943866945, "grad_norm": 0.5012265675078679, "learning_rate": 1.3893027615553922e-06, "loss": 0.5657, "step": 3260 }, { "epoch": 1.696985446985447, "grad_norm": 0.5568040197261597, "learning_rate": 1.366326756697892e-06, "loss": 0.5961, "step": 3265 }, { "epoch": 1.6995841995841996, "grad_norm": 0.5718957039548288, "learning_rate": 1.3435283872878623e-06, "loss": 0.5787, "step": 3270 }, { "epoch": 1.7021829521829521, "grad_norm": 0.5432446627166708, "learning_rate": 1.3209081223954678e-06, "loss": 0.5808, "step": 3275 }, { "epoch": 1.7047817047817047, "grad_norm": 0.5392794940489283, "learning_rate": 1.2984664274264214e-06, "loss": 0.571, "step": 3280 }, { "epoch": 1.7073804573804574, "grad_norm": 0.5619472295000788, "learning_rate": 1.2762037641124147e-06, "loss": 0.5752, "step": 3285 }, { "epoch": 1.70997920997921, "grad_norm": 0.5144706465312519, "learning_rate": 1.2541205905016097e-06, "loss": 0.5831, "step": 3290 }, { "epoch": 1.7125779625779627, "grad_norm": 0.5359125394441557, "learning_rate": 1.2322173609492138e-06, "loss": 0.571, "step": 3295 }, { "epoch": 1.7151767151767152, "grad_norm": 0.5219188144811212, "learning_rate": 1.2104945261081425e-06, "loss": 0.5866, "step": 3300 }, { "epoch": 1.7177754677754677, "grad_norm": 0.48682253952056287, "learning_rate": 1.1889525329197337e-06, "loss": 0.5659, "step": 3305 }, { "epoch": 1.7203742203742203, "grad_norm": 0.552850063663083, "learning_rate": 1.1675918246045592e-06, "loss": 0.5892, "step": 3310 }, { "epoch": 1.722972972972973, "grad_norm": 0.5129345549943813, "learning_rate": 1.1464128406533082e-06, "loss": 0.5726, "step": 3315 }, { "epoch": 1.7255717255717256, "grad_norm": 0.5215339840888276, "learning_rate": 1.1254160168177363e-06, "loss": 0.5755, "step": 3320 }, { "epoch": 1.7281704781704783, "grad_norm": 0.5346007889564445, "learning_rate": 1.1046017851017077e-06, "loss": 0.5934, "step": 3325 }, { "epoch": 1.7307692307692308, "grad_norm": 0.5241493511931715, "learning_rate": 1.083970573752301e-06, "loss": 0.5687, "step": 3330 }, { "epoch": 1.7333679833679834, "grad_norm": 0.5144495715767252, "learning_rate": 1.0635228072510084e-06, "loss": 0.5717, "step": 3335 }, { "epoch": 1.735966735966736, "grad_norm": 0.555636699696863, "learning_rate": 1.043258906304987e-06, "loss": 0.5717, "step": 3340 }, { "epoch": 1.7385654885654884, "grad_norm": 0.5398880461538901, "learning_rate": 1.023179287838414e-06, "loss": 0.5904, "step": 3345 }, { "epoch": 1.7411642411642412, "grad_norm": 0.5443352220533247, "learning_rate": 1.0032843649839052e-06, "loss": 0.5633, "step": 3350 }, { "epoch": 1.743762993762994, "grad_norm": 0.5176425518248007, "learning_rate": 9.835745470740143e-07, "loss": 0.5659, "step": 3355 }, { "epoch": 1.7463617463617465, "grad_norm": 0.5093779378294409, "learning_rate": 9.640502396328133e-07, "loss": 0.5727, "step": 3360 }, { "epoch": 1.748960498960499, "grad_norm": 0.5447675578056951, "learning_rate": 9.447118443675496e-07, "loss": 0.57, "step": 3365 }, { "epoch": 1.7515592515592515, "grad_norm": 0.5213010548532966, "learning_rate": 9.255597591603727e-07, "loss": 0.5714, "step": 3370 }, { "epoch": 1.754158004158004, "grad_norm": 0.5398854243774627, "learning_rate": 9.065943780601583e-07, "loss": 0.5811, "step": 3375 }, { "epoch": 1.7567567567567568, "grad_norm": 0.5056718137516323, "learning_rate": 8.87816091274396e-07, "loss": 0.5674, "step": 3380 }, { "epoch": 1.7593555093555093, "grad_norm": 0.4990869376715757, "learning_rate": 8.692252851611594e-07, "loss": 0.5813, "step": 3385 }, { "epoch": 1.761954261954262, "grad_norm": 0.53552516213898, "learning_rate": 8.508223422211593e-07, "loss": 0.5814, "step": 3390 }, { "epoch": 1.7645530145530146, "grad_norm": 0.5060595322478018, "learning_rate": 8.326076410898798e-07, "loss": 0.5653, "step": 3395 }, { "epoch": 1.7671517671517671, "grad_norm": 0.5333772606550752, "learning_rate": 8.145815565297721e-07, "loss": 0.575, "step": 3400 }, { "epoch": 1.7697505197505197, "grad_norm": 0.5193976811229921, "learning_rate": 7.967444594225604e-07, "loss": 0.5821, "step": 3405 }, { "epoch": 1.7723492723492722, "grad_norm": 0.4986756303541811, "learning_rate": 7.790967167615982e-07, "loss": 0.5878, "step": 3410 }, { "epoch": 1.774948024948025, "grad_norm": 0.5443836974981098, "learning_rate": 7.616386916443308e-07, "loss": 0.5788, "step": 3415 }, { "epoch": 1.7775467775467777, "grad_norm": 0.5018572354435884, "learning_rate": 7.443707432648117e-07, "loss": 0.5757, "step": 3420 }, { "epoch": 1.7801455301455302, "grad_norm": 0.5221278089522104, "learning_rate": 7.272932269063182e-07, "loss": 0.5804, "step": 3425 }, { "epoch": 1.7827442827442828, "grad_norm": 0.5309066311243344, "learning_rate": 7.10406493934046e-07, "loss": 0.5725, "step": 3430 }, { "epoch": 1.7853430353430353, "grad_norm": 0.523703068520167, "learning_rate": 6.937108917878699e-07, "loss": 0.5851, "step": 3435 }, { "epoch": 1.7879417879417878, "grad_norm": 0.4990677586423772, "learning_rate": 6.772067639752023e-07, "loss": 0.5738, "step": 3440 }, { "epoch": 1.7905405405405406, "grad_norm": 0.5118619723842658, "learning_rate": 6.608944500639247e-07, "loss": 0.5759, "step": 3445 }, { "epoch": 1.793139293139293, "grad_norm": 0.5335785839494404, "learning_rate": 6.447742856754014e-07, "loss": 0.5765, "step": 3450 }, { "epoch": 1.7957380457380459, "grad_norm": 0.5153869448178949, "learning_rate": 6.288466024775674e-07, "loss": 0.5696, "step": 3455 }, { "epoch": 1.7983367983367984, "grad_norm": 0.48901624634329477, "learning_rate": 6.131117281781196e-07, "loss": 0.5674, "step": 3460 }, { "epoch": 1.800935550935551, "grad_norm": 0.5102458462355101, "learning_rate": 5.975699865177575e-07, "loss": 0.5567, "step": 3465 }, { "epoch": 1.8035343035343034, "grad_norm": 0.5723265636044084, "learning_rate": 5.822216972635331e-07, "loss": 0.5926, "step": 3470 }, { "epoch": 1.806133056133056, "grad_norm": 0.5094713754422882, "learning_rate": 5.670671762022684e-07, "loss": 0.5655, "step": 3475 }, { "epoch": 1.8087318087318087, "grad_norm": 0.5039283824986392, "learning_rate": 5.52106735134057e-07, "loss": 0.5804, "step": 3480 }, { "epoch": 1.8113305613305615, "grad_norm": 0.5144912329273745, "learning_rate": 5.3734068186585e-07, "loss": 0.5836, "step": 3485 }, { "epoch": 1.813929313929314, "grad_norm": 0.5153264587268782, "learning_rate": 5.227693202051232e-07, "loss": 0.5847, "step": 3490 }, { "epoch": 1.8165280665280665, "grad_norm": 0.5415875159805758, "learning_rate": 5.083929499536289e-07, "loss": 0.577, "step": 3495 }, { "epoch": 1.819126819126819, "grad_norm": 0.5049450382278391, "learning_rate": 4.942118669012208e-07, "loss": 0.5822, "step": 3500 }, { "epoch": 1.8217255717255716, "grad_norm": 0.5051461545256577, "learning_rate": 4.802263628197756e-07, "loss": 0.5644, "step": 3505 }, { "epoch": 1.8243243243243243, "grad_norm": 0.49660831452275783, "learning_rate": 4.664367254571855e-07, "loss": 0.5709, "step": 3510 }, { "epoch": 1.8269230769230769, "grad_norm": 0.49492888592400675, "learning_rate": 4.528432385314352e-07, "loss": 0.5813, "step": 3515 }, { "epoch": 1.8295218295218296, "grad_norm": 0.4865514561195398, "learning_rate": 4.394461817247753e-07, "loss": 0.5686, "step": 3520 }, { "epoch": 1.8321205821205822, "grad_norm": 0.5225610714631679, "learning_rate": 4.262458306779571e-07, "loss": 0.5806, "step": 3525 }, { "epoch": 1.8347193347193347, "grad_norm": 0.5340600886285485, "learning_rate": 4.13242456984565e-07, "loss": 0.566, "step": 3530 }, { "epoch": 1.8373180873180872, "grad_norm": 0.5237864452094557, "learning_rate": 4.0043632818542845e-07, "loss": 0.576, "step": 3535 }, { "epoch": 1.83991683991684, "grad_norm": 0.5512703745424207, "learning_rate": 3.878277077631176e-07, "loss": 0.5786, "step": 3540 }, { "epoch": 1.8425155925155925, "grad_norm": 0.47905215589036726, "learning_rate": 3.754168551365234e-07, "loss": 0.5772, "step": 3545 }, { "epoch": 1.8451143451143452, "grad_norm": 0.549871032924406, "learning_rate": 3.6320402565551585e-07, "loss": 0.5779, "step": 3550 }, { "epoch": 1.8477130977130978, "grad_norm": 0.5305355260984184, "learning_rate": 3.5118947059569974e-07, "loss": 0.5814, "step": 3555 }, { "epoch": 1.8503118503118503, "grad_norm": 0.5361560006783601, "learning_rate": 3.393734371532309e-07, "loss": 0.5668, "step": 3560 }, { "epoch": 1.8529106029106028, "grad_norm": 0.5205186123699476, "learning_rate": 3.277561684397412e-07, "loss": 0.5921, "step": 3565 }, { "epoch": 1.8555093555093554, "grad_norm": 0.523345976738322, "learning_rate": 3.163379034773328e-07, "loss": 0.5889, "step": 3570 }, { "epoch": 1.8581081081081081, "grad_norm": 0.5080494795310655, "learning_rate": 3.0511887719366193e-07, "loss": 0.5707, "step": 3575 }, { "epoch": 1.8607068607068609, "grad_norm": 0.5483892874723801, "learning_rate": 2.9409932041710056e-07, "loss": 0.5867, "step": 3580 }, { "epoch": 1.8633056133056134, "grad_norm": 0.5028157875697813, "learning_rate": 2.8327945987199234e-07, "loss": 0.5735, "step": 3585 }, { "epoch": 1.865904365904366, "grad_norm": 0.542730797020021, "learning_rate": 2.726595181739866e-07, "loss": 0.5607, "step": 3590 }, { "epoch": 1.8685031185031185, "grad_norm": 0.5229545949985743, "learning_rate": 2.6223971382545623e-07, "loss": 0.5805, "step": 3595 }, { "epoch": 1.871101871101871, "grad_norm": 0.528588646817702, "learning_rate": 2.5202026121100453e-07, "loss": 0.5809, "step": 3600 }, { "epoch": 1.8737006237006237, "grad_norm": 0.5053433572695779, "learning_rate": 2.420013705930524e-07, "loss": 0.5695, "step": 3605 }, { "epoch": 1.8762993762993763, "grad_norm": 0.6061825606893263, "learning_rate": 2.321832481075137e-07, "loss": 0.5821, "step": 3610 }, { "epoch": 1.878898128898129, "grad_norm": 0.49099322033756715, "learning_rate": 2.225660957595499e-07, "loss": 0.5721, "step": 3615 }, { "epoch": 1.8814968814968815, "grad_norm": 0.5193378188817532, "learning_rate": 2.131501114194212e-07, "loss": 0.568, "step": 3620 }, { "epoch": 1.884095634095634, "grad_norm": 0.5088241393434693, "learning_rate": 2.0393548881841308e-07, "loss": 0.5937, "step": 3625 }, { "epoch": 1.8866943866943866, "grad_norm": 0.5249257025900219, "learning_rate": 1.9492241754484275e-07, "loss": 0.5696, "step": 3630 }, { "epoch": 1.8892931392931391, "grad_norm": 0.5063117403822439, "learning_rate": 1.861110830401691e-07, "loss": 0.5776, "step": 3635 }, { "epoch": 1.8918918918918919, "grad_norm": 0.5247823235703282, "learning_rate": 1.7750166659517009e-07, "loss": 0.5805, "step": 3640 }, { "epoch": 1.8944906444906446, "grad_norm": 0.530597215782001, "learning_rate": 1.6909434534621683e-07, "loss": 0.5833, "step": 3645 }, { "epoch": 1.8970893970893972, "grad_norm": 0.535514568090153, "learning_rate": 1.6088929227162542e-07, "loss": 0.5722, "step": 3650 }, { "epoch": 1.8996881496881497, "grad_norm": 0.5247060245361456, "learning_rate": 1.5288667618810426e-07, "loss": 0.5761, "step": 3655 }, { "epoch": 1.9022869022869022, "grad_norm": 0.5199277346646484, "learning_rate": 1.4508666174727237e-07, "loss": 0.5625, "step": 3660 }, { "epoch": 1.9048856548856548, "grad_norm": 0.5016083095240303, "learning_rate": 1.374894094322765e-07, "loss": 0.5754, "step": 3665 }, { "epoch": 1.9074844074844075, "grad_norm": 0.7501294618373303, "learning_rate": 1.300950755544894e-07, "loss": 0.5914, "step": 3670 }, { "epoch": 1.91008316008316, "grad_norm": 0.5599461348657118, "learning_rate": 1.2290381225029345e-07, "loss": 0.5661, "step": 3675 }, { "epoch": 1.9126819126819128, "grad_norm": 0.5324363544661017, "learning_rate": 1.1591576747794875e-07, "loss": 0.5734, "step": 3680 }, { "epoch": 1.9152806652806653, "grad_norm": 0.4949940507745294, "learning_rate": 1.0913108501454994e-07, "loss": 0.5691, "step": 3685 }, { "epoch": 1.9178794178794178, "grad_norm": 0.5201949950592929, "learning_rate": 1.0254990445306978e-07, "loss": 0.5702, "step": 3690 }, { "epoch": 1.9204781704781704, "grad_norm": 0.5435052866885572, "learning_rate": 9.617236119948359e-08, "loss": 0.5737, "step": 3695 }, { "epoch": 1.9230769230769231, "grad_norm": 0.5284636555155561, "learning_rate": 8.999858646998705e-08, "loss": 0.5691, "step": 3700 }, { "epoch": 1.9256756756756757, "grad_norm": 0.48133491617434554, "learning_rate": 8.402870728829283e-08, "loss": 0.5679, "step": 3705 }, { "epoch": 1.9282744282744284, "grad_norm": 0.525724615350644, "learning_rate": 7.82628464830193e-08, "loss": 0.5753, "step": 3710 }, { "epoch": 1.930873180873181, "grad_norm": 0.5316420453721764, "learning_rate": 7.270112268516482e-08, "loss": 0.5769, "step": 3715 }, { "epoch": 1.9334719334719335, "grad_norm": 0.5256504385548955, "learning_rate": 6.734365032566192e-08, "loss": 0.5818, "step": 3720 }, { "epoch": 1.936070686070686, "grad_norm": 0.5309069869526528, "learning_rate": 6.219053963303023e-08, "loss": 0.5734, "step": 3725 }, { "epoch": 1.9386694386694385, "grad_norm": 0.5354454111890438, "learning_rate": 5.724189663110058e-08, "loss": 0.5809, "step": 3730 }, { "epoch": 1.9412681912681913, "grad_norm": 0.5053818186885027, "learning_rate": 5.249782313683893e-08, "loss": 0.5815, "step": 3735 }, { "epoch": 1.943866943866944, "grad_norm": 0.5087952292369419, "learning_rate": 4.7958416758249196e-08, "loss": 0.572, "step": 3740 }, { "epoch": 1.9464656964656966, "grad_norm": 0.5399877020308321, "learning_rate": 4.3623770892368136e-08, "loss": 0.5745, "step": 3745 }, { "epoch": 1.949064449064449, "grad_norm": 0.4970333597526285, "learning_rate": 3.949397472333805e-08, "loss": 0.5677, "step": 3750 }, { "epoch": 1.9516632016632016, "grad_norm": 0.51125198825636, "learning_rate": 3.556911322057821e-08, "loss": 0.5807, "step": 3755 }, { "epoch": 1.9542619542619541, "grad_norm": 0.5453706817478942, "learning_rate": 3.184926713703185e-08, "loss": 0.5696, "step": 3760 }, { "epoch": 1.956860706860707, "grad_norm": 0.5259685279196025, "learning_rate": 2.8334513007507447e-08, "loss": 0.5874, "step": 3765 }, { "epoch": 1.9594594594594594, "grad_norm": 0.5299460024442244, "learning_rate": 2.5024923147101144e-08, "loss": 0.5802, "step": 3770 }, { "epoch": 1.9620582120582122, "grad_norm": 0.5085531168100673, "learning_rate": 2.1920565649713454e-08, "loss": 0.5633, "step": 3775 }, { "epoch": 1.9646569646569647, "grad_norm": 0.5066062776158768, "learning_rate": 1.902150438664374e-08, "loss": 0.5815, "step": 3780 }, { "epoch": 1.9672557172557172, "grad_norm": 0.5304292915843402, "learning_rate": 1.6327799005277923e-08, "loss": 0.5818, "step": 3785 }, { "epoch": 1.9698544698544698, "grad_norm": 0.5429547592576485, "learning_rate": 1.3839504927863901e-08, "loss": 0.5866, "step": 3790 }, { "epoch": 1.9724532224532223, "grad_norm": 0.5025697826189373, "learning_rate": 1.15566733503647e-08, "loss": 0.5805, "step": 3795 }, { "epoch": 1.975051975051975, "grad_norm": 0.5279650041637034, "learning_rate": 9.479351241410418e-09, "loss": 0.5664, "step": 3800 }, { "epoch": 1.9776507276507278, "grad_norm": 0.5231820617917791, "learning_rate": 7.607581341332326e-09, "loss": 0.5712, "step": 3805 }, { "epoch": 1.9802494802494803, "grad_norm": 0.5278123420036012, "learning_rate": 5.941402161279142e-09, "loss": 0.5737, "step": 3810 }, { "epoch": 1.9828482328482329, "grad_norm": 0.5309246932396601, "learning_rate": 4.48084798242987e-09, "loss": 0.5817, "step": 3815 }, { "epoch": 1.9854469854469854, "grad_norm": 0.5575208848453578, "learning_rate": 3.2259488552832676e-09, "loss": 0.5787, "step": 3820 }, { "epoch": 1.988045738045738, "grad_norm": 0.512653822002805, "learning_rate": 2.1767305990416653e-09, "loss": 0.5848, "step": 3825 }, { "epoch": 1.9906444906444907, "grad_norm": 0.5291714453451785, "learning_rate": 1.3332148010836154e-09, "loss": 0.5831, "step": 3830 }, { "epoch": 1.9932432432432432, "grad_norm": 0.5268775795002444, "learning_rate": 6.954188165131382e-10, "loss": 0.5895, "step": 3835 }, { "epoch": 1.995841995841996, "grad_norm": 0.5255058213556102, "learning_rate": 2.6335576780667227e-10, "loss": 0.5759, "step": 3840 }, { "epoch": 1.9984407484407485, "grad_norm": 0.5086937679965747, "learning_rate": 3.7034544543290694e-11, "loss": 0.5727, "step": 3845 }, { "epoch": 2.0, "eval_loss": 0.7488190531730652, "eval_runtime": 104.7896, "eval_samples_per_second": 78.347, "eval_steps_per_second": 1.231, "step": 3848 }, { "epoch": 2.0, "step": 3848, "total_flos": 805692915056640.0, "train_loss": 0.6651840950991657, "train_runtime": 11913.1492, "train_samples_per_second": 20.671, "train_steps_per_second": 0.323 } ], "logging_steps": 5, "max_steps": 3848, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 805692915056640.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }