{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3094198378041173, "eval_steps": 248, "global_step": 248, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012476606363069245, "grad_norm": 0.3369293212890625, "learning_rate": 4.0000000000000003e-07, "loss": 1.4366, "step": 1 }, { "epoch": 0.002495321272613849, "grad_norm": 0.36413154006004333, "learning_rate": 8.000000000000001e-07, "loss": 1.4996, "step": 2 }, { "epoch": 0.0037429819089207735, "grad_norm": 0.3281562626361847, "learning_rate": 1.2000000000000002e-06, "loss": 1.5808, "step": 3 }, { "epoch": 0.004990642545227698, "grad_norm": 0.20739880204200745, "learning_rate": 1.6000000000000001e-06, "loss": 1.5504, "step": 4 }, { "epoch": 0.006238303181534623, "grad_norm": 0.3129860460758209, "learning_rate": 2.0000000000000003e-06, "loss": 1.6271, "step": 5 }, { "epoch": 0.007485963817841547, "grad_norm": 0.2682516574859619, "learning_rate": 2.4000000000000003e-06, "loss": 1.5808, "step": 6 }, { "epoch": 0.008733624454148471, "grad_norm": 0.2960676848888397, "learning_rate": 2.8000000000000003e-06, "loss": 1.5214, "step": 7 }, { "epoch": 0.009981285090455396, "grad_norm": 0.1701919436454773, "learning_rate": 3.2000000000000003e-06, "loss": 1.5882, "step": 8 }, { "epoch": 0.011228945726762321, "grad_norm": 0.2361917495727539, "learning_rate": 3.6000000000000003e-06, "loss": 1.5905, "step": 9 }, { "epoch": 0.012476606363069246, "grad_norm": 0.30789047479629517, "learning_rate": 4.000000000000001e-06, "loss": 1.4745, "step": 10 }, { "epoch": 0.01372426699937617, "grad_norm": 0.3127429187297821, "learning_rate": 4.4e-06, "loss": 1.5997, "step": 11 }, { "epoch": 0.014971927635683094, "grad_norm": 0.4279906451702118, "learning_rate": 4.800000000000001e-06, "loss": 1.4534, "step": 12 }, { "epoch": 0.016219588271990017, "grad_norm": 0.1982978880405426, "learning_rate": 5.2e-06, "loss": 1.5678, "step": 13 }, { "epoch": 0.017467248908296942, "grad_norm": 0.35204875469207764, "learning_rate": 5.600000000000001e-06, "loss": 1.5609, "step": 14 }, { "epoch": 0.018714909544603867, "grad_norm": 0.3509620130062103, "learning_rate": 6e-06, "loss": 1.6769, "step": 15 }, { "epoch": 0.019962570180910792, "grad_norm": 0.3329513370990753, "learning_rate": 6.4000000000000006e-06, "loss": 1.4596, "step": 16 }, { "epoch": 0.021210230817217717, "grad_norm": 0.2940622568130493, "learning_rate": 6.800000000000001e-06, "loss": 1.4965, "step": 17 }, { "epoch": 0.022457891453524642, "grad_norm": 0.2941367030143738, "learning_rate": 7.2000000000000005e-06, "loss": 1.498, "step": 18 }, { "epoch": 0.023705552089831567, "grad_norm": 0.41950124502182007, "learning_rate": 7.600000000000001e-06, "loss": 1.5426, "step": 19 }, { "epoch": 0.024953212726138492, "grad_norm": 0.22718903422355652, "learning_rate": 8.000000000000001e-06, "loss": 1.5473, "step": 20 }, { "epoch": 0.026200873362445413, "grad_norm": 0.3167245388031006, "learning_rate": 8.400000000000001e-06, "loss": 1.5591, "step": 21 }, { "epoch": 0.02744853399875234, "grad_norm": 0.24035727977752686, "learning_rate": 8.8e-06, "loss": 1.5008, "step": 22 }, { "epoch": 0.028696194635059263, "grad_norm": 0.32235047221183777, "learning_rate": 9.200000000000002e-06, "loss": 1.6414, "step": 23 }, { "epoch": 0.02994385527136619, "grad_norm": 0.27163419127464294, "learning_rate": 9.600000000000001e-06, "loss": 1.534, "step": 24 }, { "epoch": 0.031191515907673113, "grad_norm": 0.19682233035564423, "learning_rate": 1e-05, "loss": 1.839, "step": 25 }, { "epoch": 0.032439176543980035, "grad_norm": 0.19141852855682373, "learning_rate": 1.04e-05, "loss": 1.4647, "step": 26 }, { "epoch": 0.03368683718028696, "grad_norm": 0.19754448533058167, "learning_rate": 1.0800000000000002e-05, "loss": 1.5247, "step": 27 }, { "epoch": 0.034934497816593885, "grad_norm": 0.21717581152915955, "learning_rate": 1.1200000000000001e-05, "loss": 1.4311, "step": 28 }, { "epoch": 0.03618215845290081, "grad_norm": 0.18977436423301697, "learning_rate": 1.16e-05, "loss": 1.4796, "step": 29 }, { "epoch": 0.037429819089207735, "grad_norm": 0.20318299531936646, "learning_rate": 1.2e-05, "loss": 1.4515, "step": 30 }, { "epoch": 0.03867747972551466, "grad_norm": 0.18727326393127441, "learning_rate": 1.2400000000000002e-05, "loss": 1.5346, "step": 31 }, { "epoch": 0.039925140361821584, "grad_norm": 0.2555878460407257, "learning_rate": 1.2800000000000001e-05, "loss": 1.4717, "step": 32 }, { "epoch": 0.041172800998128506, "grad_norm": 0.2864621579647064, "learning_rate": 1.3200000000000002e-05, "loss": 1.2702, "step": 33 }, { "epoch": 0.042420461634435434, "grad_norm": 0.14290520548820496, "learning_rate": 1.3600000000000002e-05, "loss": 1.4323, "step": 34 }, { "epoch": 0.043668122270742356, "grad_norm": 0.14048892259597778, "learning_rate": 1.4e-05, "loss": 1.4812, "step": 35 }, { "epoch": 0.044915782907049284, "grad_norm": 0.12439600378274918, "learning_rate": 1.4400000000000001e-05, "loss": 1.3209, "step": 36 }, { "epoch": 0.046163443543356206, "grad_norm": 0.11550629138946533, "learning_rate": 1.48e-05, "loss": 1.2587, "step": 37 }, { "epoch": 0.047411104179663134, "grad_norm": 0.1708095371723175, "learning_rate": 1.5200000000000002e-05, "loss": 1.4372, "step": 38 }, { "epoch": 0.048658764815970056, "grad_norm": 0.12109667807817459, "learning_rate": 1.5600000000000003e-05, "loss": 1.3698, "step": 39 }, { "epoch": 0.049906425452276984, "grad_norm": 0.3102639615535736, "learning_rate": 1.6000000000000003e-05, "loss": 1.2817, "step": 40 }, { "epoch": 0.051154086088583905, "grad_norm": 0.08972104638814926, "learning_rate": 1.64e-05, "loss": 1.2466, "step": 41 }, { "epoch": 0.05240174672489083, "grad_norm": 0.09888940304517746, "learning_rate": 1.6800000000000002e-05, "loss": 1.4295, "step": 42 }, { "epoch": 0.053649407361197755, "grad_norm": 0.09405261278152466, "learning_rate": 1.72e-05, "loss": 1.3466, "step": 43 }, { "epoch": 0.05489706799750468, "grad_norm": 0.08513078093528748, "learning_rate": 1.76e-05, "loss": 1.4428, "step": 44 }, { "epoch": 0.056144728633811605, "grad_norm": 0.09294164925813675, "learning_rate": 1.8e-05, "loss": 1.0991, "step": 45 }, { "epoch": 0.05739238927011853, "grad_norm": 0.09842169284820557, "learning_rate": 1.8400000000000003e-05, "loss": 1.1395, "step": 46 }, { "epoch": 0.058640049906425455, "grad_norm": 0.09997382760047913, "learning_rate": 1.88e-05, "loss": 1.2846, "step": 47 }, { "epoch": 0.05988771054273238, "grad_norm": 0.08888943493366241, "learning_rate": 1.9200000000000003e-05, "loss": 1.2464, "step": 48 }, { "epoch": 0.0611353711790393, "grad_norm": 0.08796142041683197, "learning_rate": 1.9600000000000002e-05, "loss": 1.2413, "step": 49 }, { "epoch": 0.06238303181534623, "grad_norm": 0.10694792866706848, "learning_rate": 2e-05, "loss": 1.1867, "step": 50 }, { "epoch": 0.06363069245165315, "grad_norm": 0.13788695633411407, "learning_rate": 1.999999108696051e-05, "loss": 1.1745, "step": 51 }, { "epoch": 0.06487835308796007, "grad_norm": 0.10375945270061493, "learning_rate": 1.999996434785792e-05, "loss": 1.1885, "step": 52 }, { "epoch": 0.066126013724267, "grad_norm": 0.10151001065969467, "learning_rate": 1.9999919782739897e-05, "loss": 1.1994, "step": 53 }, { "epoch": 0.06737367436057393, "grad_norm": 0.10222937166690826, "learning_rate": 1.999985739168588e-05, "loss": 1.1777, "step": 54 }, { "epoch": 0.06862133499688085, "grad_norm": 0.10297010838985443, "learning_rate": 1.9999777174807097e-05, "loss": 1.0911, "step": 55 }, { "epoch": 0.06986899563318777, "grad_norm": 0.1051320731639862, "learning_rate": 1.9999679132246536e-05, "loss": 1.1694, "step": 56 }, { "epoch": 0.07111665626949469, "grad_norm": 0.13086318969726562, "learning_rate": 1.999956326417897e-05, "loss": 1.2241, "step": 57 }, { "epoch": 0.07236431690580163, "grad_norm": 0.11484101414680481, "learning_rate": 1.9999429570810946e-05, "loss": 1.0468, "step": 58 }, { "epoch": 0.07361197754210855, "grad_norm": 0.11556559056043625, "learning_rate": 1.999927805238079e-05, "loss": 1.0311, "step": 59 }, { "epoch": 0.07485963817841547, "grad_norm": 0.11713112890720367, "learning_rate": 1.9999108709158594e-05, "loss": 1.0457, "step": 60 }, { "epoch": 0.07610729881472239, "grad_norm": 0.1211632490158081, "learning_rate": 1.9998921541446235e-05, "loss": 1.0551, "step": 61 }, { "epoch": 0.07735495945102933, "grad_norm": 0.12310254573822021, "learning_rate": 1.999871654957736e-05, "loss": 1.082, "step": 62 }, { "epoch": 0.07860262008733625, "grad_norm": 0.1967160552740097, "learning_rate": 1.9998493733917385e-05, "loss": 1.1005, "step": 63 }, { "epoch": 0.07985028072364317, "grad_norm": 0.12039731442928314, "learning_rate": 1.999825309486351e-05, "loss": 1.1583, "step": 64 }, { "epoch": 0.08109794135995009, "grad_norm": 0.1448286473751068, "learning_rate": 1.999799463284469e-05, "loss": 0.9149, "step": 65 }, { "epoch": 0.08234560199625701, "grad_norm": 0.13981443643569946, "learning_rate": 1.9997718348321672e-05, "loss": 0.9094, "step": 66 }, { "epoch": 0.08359326263256395, "grad_norm": 0.16553014516830444, "learning_rate": 1.9997424241786956e-05, "loss": 0.9401, "step": 67 }, { "epoch": 0.08484092326887087, "grad_norm": 0.16061550378799438, "learning_rate": 1.999711231376482e-05, "loss": 0.6806, "step": 68 }, { "epoch": 0.08608858390517779, "grad_norm": 0.157967671751976, "learning_rate": 1.999678256481131e-05, "loss": 0.7469, "step": 69 }, { "epoch": 0.08733624454148471, "grad_norm": 0.2233240157365799, "learning_rate": 1.9996434995514242e-05, "loss": 0.7202, "step": 70 }, { "epoch": 0.08858390517779165, "grad_norm": 0.18542423844337463, "learning_rate": 1.999606960649319e-05, "loss": 0.9828, "step": 71 }, { "epoch": 0.08983156581409857, "grad_norm": 0.18575641512870789, "learning_rate": 1.9995686398399505e-05, "loss": 0.6079, "step": 72 }, { "epoch": 0.09107922645040549, "grad_norm": 0.15296655893325806, "learning_rate": 1.9995285371916294e-05, "loss": 0.5708, "step": 73 }, { "epoch": 0.09232688708671241, "grad_norm": 0.21740297973155975, "learning_rate": 1.9994866527758427e-05, "loss": 0.8158, "step": 74 }, { "epoch": 0.09357454772301933, "grad_norm": 0.13831280171871185, "learning_rate": 1.9994429866672543e-05, "loss": 0.7365, "step": 75 }, { "epoch": 0.09482220835932627, "grad_norm": 0.12471487373113632, "learning_rate": 1.999397538943704e-05, "loss": 0.7776, "step": 76 }, { "epoch": 0.09606986899563319, "grad_norm": 0.11238758265972137, "learning_rate": 1.9993503096862066e-05, "loss": 0.591, "step": 77 }, { "epoch": 0.09731752963194011, "grad_norm": 0.12608778476715088, "learning_rate": 1.999301298978954e-05, "loss": 0.6285, "step": 78 }, { "epoch": 0.09856519026824703, "grad_norm": 0.12795987725257874, "learning_rate": 1.9992505069093124e-05, "loss": 0.4695, "step": 79 }, { "epoch": 0.09981285090455397, "grad_norm": 0.14311431348323822, "learning_rate": 1.9991979335678242e-05, "loss": 0.7291, "step": 80 }, { "epoch": 0.10106051154086089, "grad_norm": 0.12212368845939636, "learning_rate": 1.999143579048208e-05, "loss": 0.7634, "step": 81 }, { "epoch": 0.10230817217716781, "grad_norm": 0.0974789410829544, "learning_rate": 1.9990874434473554e-05, "loss": 0.4982, "step": 82 }, { "epoch": 0.10355583281347473, "grad_norm": 0.14408168196678162, "learning_rate": 1.999029526865335e-05, "loss": 0.5957, "step": 83 }, { "epoch": 0.10480349344978165, "grad_norm": 0.09892342239618301, "learning_rate": 1.9989698294053892e-05, "loss": 0.5647, "step": 84 }, { "epoch": 0.10605115408608859, "grad_norm": 0.10291016846895218, "learning_rate": 1.9989083511739348e-05, "loss": 0.6392, "step": 85 }, { "epoch": 0.10729881472239551, "grad_norm": 0.15446709096431732, "learning_rate": 1.9988450922805635e-05, "loss": 0.5549, "step": 86 }, { "epoch": 0.10854647535870243, "grad_norm": 0.08281419426202774, "learning_rate": 1.998780052838041e-05, "loss": 0.5886, "step": 87 }, { "epoch": 0.10979413599500935, "grad_norm": 0.09740785509347916, "learning_rate": 1.998713232962308e-05, "loss": 0.6006, "step": 88 }, { "epoch": 0.11104179663131628, "grad_norm": 0.09569819271564484, "learning_rate": 1.998644632772477e-05, "loss": 0.5268, "step": 89 }, { "epoch": 0.11228945726762321, "grad_norm": 0.08716901391744614, "learning_rate": 1.998574252390836e-05, "loss": 0.5426, "step": 90 }, { "epoch": 0.11353711790393013, "grad_norm": 0.08949075639247894, "learning_rate": 1.998502091942845e-05, "loss": 0.5637, "step": 91 }, { "epoch": 0.11478477854023705, "grad_norm": 0.10287146270275116, "learning_rate": 1.9984281515571383e-05, "loss": 0.4135, "step": 92 }, { "epoch": 0.11603243917654397, "grad_norm": 0.09598784893751144, "learning_rate": 1.9983524313655225e-05, "loss": 0.5652, "step": 93 }, { "epoch": 0.11728009981285091, "grad_norm": 0.07580853998661041, "learning_rate": 1.998274931502977e-05, "loss": 0.5787, "step": 94 }, { "epoch": 0.11852776044915783, "grad_norm": 0.0887116864323616, "learning_rate": 1.998195652107653e-05, "loss": 0.5494, "step": 95 }, { "epoch": 0.11977542108546475, "grad_norm": 0.09782399237155914, "learning_rate": 1.998114593320876e-05, "loss": 0.4127, "step": 96 }, { "epoch": 0.12102308172177167, "grad_norm": 0.09533873945474625, "learning_rate": 1.9980317552871407e-05, "loss": 0.5192, "step": 97 }, { "epoch": 0.1222707423580786, "grad_norm": 0.08982633054256439, "learning_rate": 1.9979471381541158e-05, "loss": 0.4699, "step": 98 }, { "epoch": 0.12351840299438553, "grad_norm": 0.0871262177824974, "learning_rate": 1.9978607420726396e-05, "loss": 0.5772, "step": 99 }, { "epoch": 0.12476606363069245, "grad_norm": 0.08864148706197739, "learning_rate": 1.997772567196723e-05, "loss": 0.6495, "step": 100 }, { "epoch": 0.1260137242669994, "grad_norm": 0.5612615942955017, "learning_rate": 1.997682613683547e-05, "loss": 0.3912, "step": 101 }, { "epoch": 0.1272613849033063, "grad_norm": 0.18454571068286896, "learning_rate": 1.997590881693464e-05, "loss": 0.6133, "step": 102 }, { "epoch": 0.12850904553961323, "grad_norm": 0.11318469792604446, "learning_rate": 1.997497371389995e-05, "loss": 0.4185, "step": 103 }, { "epoch": 0.12975670617592014, "grad_norm": 0.09561355412006378, "learning_rate": 1.9974020829398335e-05, "loss": 0.5241, "step": 104 }, { "epoch": 0.13100436681222707, "grad_norm": 0.1466185599565506, "learning_rate": 1.9973050165128406e-05, "loss": 0.559, "step": 105 }, { "epoch": 0.132252027448534, "grad_norm": 0.0754123404622078, "learning_rate": 1.9972061722820483e-05, "loss": 0.4629, "step": 106 }, { "epoch": 0.13349968808484092, "grad_norm": 0.10411768406629562, "learning_rate": 1.997105550423656e-05, "loss": 0.3849, "step": 107 }, { "epoch": 0.13474734872114785, "grad_norm": 0.09193742275238037, "learning_rate": 1.997003151117034e-05, "loss": 0.3919, "step": 108 }, { "epoch": 0.13599500935745476, "grad_norm": 0.16121919453144073, "learning_rate": 1.9968989745447203e-05, "loss": 0.6253, "step": 109 }, { "epoch": 0.1372426699937617, "grad_norm": 0.09475496411323547, "learning_rate": 1.99679302089242e-05, "loss": 0.5236, "step": 110 }, { "epoch": 0.13849033063006863, "grad_norm": 0.09533608704805374, "learning_rate": 1.9966852903490077e-05, "loss": 0.4389, "step": 111 }, { "epoch": 0.13973799126637554, "grad_norm": 0.08322104066610336, "learning_rate": 1.996575783106524e-05, "loss": 0.5455, "step": 112 }, { "epoch": 0.14098565190268247, "grad_norm": 0.1089191734790802, "learning_rate": 1.9964644993601782e-05, "loss": 0.4437, "step": 113 }, { "epoch": 0.14223331253898938, "grad_norm": 0.117027148604393, "learning_rate": 1.9963514393083452e-05, "loss": 0.4787, "step": 114 }, { "epoch": 0.14348097317529632, "grad_norm": 0.08248787373304367, "learning_rate": 1.9962366031525663e-05, "loss": 0.4845, "step": 115 }, { "epoch": 0.14472863381160325, "grad_norm": 0.08650378882884979, "learning_rate": 1.9961199910975497e-05, "loss": 0.3742, "step": 116 }, { "epoch": 0.14597629444791016, "grad_norm": 0.07683917135000229, "learning_rate": 1.9960016033511692e-05, "loss": 0.4035, "step": 117 }, { "epoch": 0.1472239550842171, "grad_norm": 0.08383751660585403, "learning_rate": 1.9958814401244634e-05, "loss": 0.5135, "step": 118 }, { "epoch": 0.14847161572052403, "grad_norm": 0.12660759687423706, "learning_rate": 1.995759501631637e-05, "loss": 0.6356, "step": 119 }, { "epoch": 0.14971927635683094, "grad_norm": 0.09009061008691788, "learning_rate": 1.995635788090057e-05, "loss": 0.4618, "step": 120 }, { "epoch": 0.15096693699313787, "grad_norm": 0.10818025469779968, "learning_rate": 1.9955102997202575e-05, "loss": 0.5522, "step": 121 }, { "epoch": 0.15221459762944478, "grad_norm": 0.07673201709985733, "learning_rate": 1.9953830367459343e-05, "loss": 0.3801, "step": 122 }, { "epoch": 0.15346225826575172, "grad_norm": 0.09158243238925934, "learning_rate": 1.9952539993939478e-05, "loss": 0.413, "step": 123 }, { "epoch": 0.15470991890205865, "grad_norm": 0.11646990478038788, "learning_rate": 1.9951231878943207e-05, "loss": 0.4732, "step": 124 }, { "epoch": 0.15595757953836556, "grad_norm": 0.0766897052526474, "learning_rate": 1.994990602480239e-05, "loss": 0.4355, "step": 125 }, { "epoch": 0.1572052401746725, "grad_norm": 0.10383328795433044, "learning_rate": 1.9948562433880498e-05, "loss": 0.433, "step": 126 }, { "epoch": 0.1584529008109794, "grad_norm": 0.07712090015411377, "learning_rate": 1.9947201108572636e-05, "loss": 0.4848, "step": 127 }, { "epoch": 0.15970056144728634, "grad_norm": 0.14091262221336365, "learning_rate": 1.994582205130551e-05, "loss": 0.439, "step": 128 }, { "epoch": 0.16094822208359327, "grad_norm": 0.10989029705524445, "learning_rate": 1.9944425264537432e-05, "loss": 0.4429, "step": 129 }, { "epoch": 0.16219588271990018, "grad_norm": 0.1106753721833229, "learning_rate": 1.9943010750758335e-05, "loss": 0.4662, "step": 130 }, { "epoch": 0.16344354335620712, "grad_norm": 0.09498463571071625, "learning_rate": 1.994157851248973e-05, "loss": 0.4164, "step": 131 }, { "epoch": 0.16469120399251402, "grad_norm": 0.0848456621170044, "learning_rate": 1.9940128552284756e-05, "loss": 0.4634, "step": 132 }, { "epoch": 0.16593886462882096, "grad_norm": 0.06310152262449265, "learning_rate": 1.99386608727281e-05, "loss": 0.4834, "step": 133 }, { "epoch": 0.1671865252651279, "grad_norm": 0.10514630377292633, "learning_rate": 1.9937175476436078e-05, "loss": 0.4392, "step": 134 }, { "epoch": 0.1684341859014348, "grad_norm": 0.47956305742263794, "learning_rate": 1.9935672366056557e-05, "loss": 0.4975, "step": 135 }, { "epoch": 0.16968184653774174, "grad_norm": 0.12076468020677567, "learning_rate": 1.9934151544269e-05, "loss": 0.425, "step": 136 }, { "epoch": 0.17092950717404864, "grad_norm": 0.07710213214159012, "learning_rate": 1.9932613013784436e-05, "loss": 0.4426, "step": 137 }, { "epoch": 0.17217716781035558, "grad_norm": 0.1181124597787857, "learning_rate": 1.993105677734546e-05, "loss": 0.4272, "step": 138 }, { "epoch": 0.17342482844666252, "grad_norm": 0.08521714061498642, "learning_rate": 1.9929482837726227e-05, "loss": 0.4041, "step": 139 }, { "epoch": 0.17467248908296942, "grad_norm": 0.0910830870270729, "learning_rate": 1.992789119773246e-05, "loss": 0.4372, "step": 140 }, { "epoch": 0.17592014971927636, "grad_norm": 0.09005699306726456, "learning_rate": 1.9926281860201427e-05, "loss": 0.3363, "step": 141 }, { "epoch": 0.1771678103555833, "grad_norm": 0.1397104263305664, "learning_rate": 1.9924654828001947e-05, "loss": 0.5186, "step": 142 }, { "epoch": 0.1784154709918902, "grad_norm": 0.0795428678393364, "learning_rate": 1.9923010104034382e-05, "loss": 0.4258, "step": 143 }, { "epoch": 0.17966313162819714, "grad_norm": 0.11801055818796158, "learning_rate": 1.9921347691230622e-05, "loss": 0.4102, "step": 144 }, { "epoch": 0.18091079226450404, "grad_norm": 0.09254780411720276, "learning_rate": 1.9919667592554107e-05, "loss": 0.4247, "step": 145 }, { "epoch": 0.18215845290081098, "grad_norm": 0.09013550728559494, "learning_rate": 1.991796981099979e-05, "loss": 0.4397, "step": 146 }, { "epoch": 0.18340611353711792, "grad_norm": 0.10347352921962738, "learning_rate": 1.991625434959415e-05, "loss": 0.3864, "step": 147 }, { "epoch": 0.18465377417342482, "grad_norm": 0.08151372522115707, "learning_rate": 1.991452121139518e-05, "loss": 0.3866, "step": 148 }, { "epoch": 0.18590143480973176, "grad_norm": 0.07901383191347122, "learning_rate": 1.9912770399492385e-05, "loss": 0.4413, "step": 149 }, { "epoch": 0.18714909544603867, "grad_norm": 0.10809416323900223, "learning_rate": 1.9911001917006782e-05, "loss": 0.4892, "step": 150 }, { "epoch": 0.1883967560823456, "grad_norm": 0.0716921016573906, "learning_rate": 1.990921576709088e-05, "loss": 0.377, "step": 151 }, { "epoch": 0.18964441671865254, "grad_norm": 0.07814662158489227, "learning_rate": 1.9907411952928676e-05, "loss": 0.3738, "step": 152 }, { "epoch": 0.19089207735495944, "grad_norm": 0.0741223469376564, "learning_rate": 1.9905590477735677e-05, "loss": 0.4715, "step": 153 }, { "epoch": 0.19213973799126638, "grad_norm": 0.11747987568378448, "learning_rate": 1.990375134475885e-05, "loss": 0.3903, "step": 154 }, { "epoch": 0.1933873986275733, "grad_norm": 0.08143888413906097, "learning_rate": 1.9901894557276647e-05, "loss": 0.4747, "step": 155 }, { "epoch": 0.19463505926388022, "grad_norm": 0.08514941483736038, "learning_rate": 1.9900020118598996e-05, "loss": 0.4365, "step": 156 }, { "epoch": 0.19588271990018716, "grad_norm": 0.07622472196817398, "learning_rate": 1.9898128032067287e-05, "loss": 0.4311, "step": 157 }, { "epoch": 0.19713038053649407, "grad_norm": 0.09147831797599792, "learning_rate": 1.9896218301054365e-05, "loss": 0.4376, "step": 158 }, { "epoch": 0.198378041172801, "grad_norm": 0.09080328047275543, "learning_rate": 1.9894290928964536e-05, "loss": 0.4817, "step": 159 }, { "epoch": 0.19962570180910794, "grad_norm": 0.0705234482884407, "learning_rate": 1.9892345919233543e-05, "loss": 0.3602, "step": 160 }, { "epoch": 0.20087336244541484, "grad_norm": 0.3075416088104248, "learning_rate": 1.9890383275328577e-05, "loss": 0.3686, "step": 161 }, { "epoch": 0.20212102308172178, "grad_norm": 0.0864262729883194, "learning_rate": 1.9888403000748268e-05, "loss": 0.5076, "step": 162 }, { "epoch": 0.2033686837180287, "grad_norm": 0.07593845576047897, "learning_rate": 1.988640509902266e-05, "loss": 0.3319, "step": 163 }, { "epoch": 0.20461634435433562, "grad_norm": 0.0860196128487587, "learning_rate": 1.988438957371324e-05, "loss": 0.4247, "step": 164 }, { "epoch": 0.20586400499064256, "grad_norm": 0.08619983494281769, "learning_rate": 1.9882356428412887e-05, "loss": 0.4515, "step": 165 }, { "epoch": 0.20711166562694946, "grad_norm": 0.08582403510808945, "learning_rate": 1.988030566674591e-05, "loss": 0.4861, "step": 166 }, { "epoch": 0.2083593262632564, "grad_norm": 0.09421785920858383, "learning_rate": 1.9878237292368014e-05, "loss": 0.3536, "step": 167 }, { "epoch": 0.2096069868995633, "grad_norm": 0.08474161475896835, "learning_rate": 1.9876151308966293e-05, "loss": 0.4128, "step": 168 }, { "epoch": 0.21085464753587024, "grad_norm": 0.07902742177248001, "learning_rate": 1.987404772025924e-05, "loss": 0.3909, "step": 169 }, { "epoch": 0.21210230817217718, "grad_norm": 0.12060698121786118, "learning_rate": 1.987192652999673e-05, "loss": 0.357, "step": 170 }, { "epoch": 0.2133499688084841, "grad_norm": 0.06981324404478073, "learning_rate": 1.9869787741960013e-05, "loss": 0.4047, "step": 171 }, { "epoch": 0.21459762944479102, "grad_norm": 0.11325976252555847, "learning_rate": 1.9867631359961712e-05, "loss": 0.4094, "step": 172 }, { "epoch": 0.21584529008109793, "grad_norm": 0.07985086739063263, "learning_rate": 1.9865457387845806e-05, "loss": 0.4056, "step": 173 }, { "epoch": 0.21709295071740486, "grad_norm": 0.08571935445070267, "learning_rate": 1.9863265829487637e-05, "loss": 0.3637, "step": 174 }, { "epoch": 0.2183406113537118, "grad_norm": 0.11255648732185364, "learning_rate": 1.9861056688793896e-05, "loss": 0.4475, "step": 175 }, { "epoch": 0.2195882719900187, "grad_norm": 0.08018800616264343, "learning_rate": 1.985882996970261e-05, "loss": 0.4584, "step": 176 }, { "epoch": 0.22083593262632564, "grad_norm": 0.07600877434015274, "learning_rate": 1.985658567618315e-05, "loss": 0.3794, "step": 177 }, { "epoch": 0.22208359326263255, "grad_norm": 0.07584904879331589, "learning_rate": 1.985432381223621e-05, "loss": 0.3718, "step": 178 }, { "epoch": 0.22333125389893949, "grad_norm": 0.07508600503206253, "learning_rate": 1.9852044381893806e-05, "loss": 0.4392, "step": 179 }, { "epoch": 0.22457891453524642, "grad_norm": 0.09708955883979797, "learning_rate": 1.9849747389219272e-05, "loss": 0.3792, "step": 180 }, { "epoch": 0.22582657517155333, "grad_norm": 0.11540783196687698, "learning_rate": 1.984743283830724e-05, "loss": 0.4075, "step": 181 }, { "epoch": 0.22707423580786026, "grad_norm": 0.08834560215473175, "learning_rate": 1.984510073328365e-05, "loss": 0.3469, "step": 182 }, { "epoch": 0.2283218964441672, "grad_norm": 0.09154046326875687, "learning_rate": 1.9842751078305735e-05, "loss": 0.411, "step": 183 }, { "epoch": 0.2295695570804741, "grad_norm": 0.07195434719324112, "learning_rate": 1.9840383877561998e-05, "loss": 0.385, "step": 184 }, { "epoch": 0.23081721771678104, "grad_norm": 0.10304310917854309, "learning_rate": 1.9837999135272235e-05, "loss": 0.3727, "step": 185 }, { "epoch": 0.23206487835308795, "grad_norm": 0.0923156812787056, "learning_rate": 1.983559685568751e-05, "loss": 0.4777, "step": 186 }, { "epoch": 0.23331253898939489, "grad_norm": 0.07710360735654831, "learning_rate": 1.983317704309014e-05, "loss": 0.3419, "step": 187 }, { "epoch": 0.23456019962570182, "grad_norm": 0.072215236723423, "learning_rate": 1.9830739701793705e-05, "loss": 0.342, "step": 188 }, { "epoch": 0.23580786026200873, "grad_norm": 0.0766647532582283, "learning_rate": 1.982828483614303e-05, "loss": 0.3672, "step": 189 }, { "epoch": 0.23705552089831566, "grad_norm": 0.07266882061958313, "learning_rate": 1.9825812450514174e-05, "loss": 0.4228, "step": 190 }, { "epoch": 0.23830318153462257, "grad_norm": 0.06986986100673676, "learning_rate": 1.9823322549314433e-05, "loss": 0.3741, "step": 191 }, { "epoch": 0.2395508421709295, "grad_norm": 0.08523140847682953, "learning_rate": 1.9820815136982327e-05, "loss": 0.4154, "step": 192 }, { "epoch": 0.24079850280723644, "grad_norm": 0.07692936062812805, "learning_rate": 1.9818290217987587e-05, "loss": 0.4984, "step": 193 }, { "epoch": 0.24204616344354335, "grad_norm": 0.11332568526268005, "learning_rate": 1.9815747796831154e-05, "loss": 0.377, "step": 194 }, { "epoch": 0.24329382407985028, "grad_norm": 0.06331691890954971, "learning_rate": 1.9813187878045164e-05, "loss": 0.3134, "step": 195 }, { "epoch": 0.2445414847161572, "grad_norm": 0.09051994234323502, "learning_rate": 1.9810610466192955e-05, "loss": 0.3818, "step": 196 }, { "epoch": 0.24578914535246413, "grad_norm": 0.09058686345815659, "learning_rate": 1.980801556586904e-05, "loss": 0.35, "step": 197 }, { "epoch": 0.24703680598877106, "grad_norm": 0.07872316241264343, "learning_rate": 1.9805403181699104e-05, "loss": 0.4229, "step": 198 }, { "epoch": 0.24828446662507797, "grad_norm": 0.07484429329633713, "learning_rate": 1.980277331834001e-05, "loss": 0.4025, "step": 199 }, { "epoch": 0.2495321272613849, "grad_norm": 0.08303457498550415, "learning_rate": 1.9800125980479768e-05, "loss": 0.4001, "step": 200 }, { "epoch": 0.25077978789769184, "grad_norm": 0.06768237799406052, "learning_rate": 1.9797461172837545e-05, "loss": 0.2898, "step": 201 }, { "epoch": 0.2520274485339988, "grad_norm": 0.10813165456056595, "learning_rate": 1.9794778900163653e-05, "loss": 0.3832, "step": 202 }, { "epoch": 0.25327510917030566, "grad_norm": 0.07230770587921143, "learning_rate": 1.9792079167239526e-05, "loss": 0.4103, "step": 203 }, { "epoch": 0.2545227698066126, "grad_norm": 0.07701678574085236, "learning_rate": 1.978936197887773e-05, "loss": 0.3438, "step": 204 }, { "epoch": 0.2557704304429195, "grad_norm": 0.07222580909729004, "learning_rate": 1.978662733992195e-05, "loss": 0.3693, "step": 205 }, { "epoch": 0.25701809107922646, "grad_norm": 0.07187530398368835, "learning_rate": 1.9783875255246972e-05, "loss": 0.3753, "step": 206 }, { "epoch": 0.2582657517155334, "grad_norm": 0.06922349333763123, "learning_rate": 1.9781105729758688e-05, "loss": 0.4076, "step": 207 }, { "epoch": 0.2595134123518403, "grad_norm": 0.07130351662635803, "learning_rate": 1.9778318768394072e-05, "loss": 0.411, "step": 208 }, { "epoch": 0.2607610729881472, "grad_norm": 0.07194481790065765, "learning_rate": 1.9775514376121184e-05, "loss": 0.3342, "step": 209 }, { "epoch": 0.26200873362445415, "grad_norm": 0.13812008500099182, "learning_rate": 1.977269255793915e-05, "loss": 0.3947, "step": 210 }, { "epoch": 0.2632563942607611, "grad_norm": 0.08995859324932098, "learning_rate": 1.9769853318878177e-05, "loss": 0.4331, "step": 211 }, { "epoch": 0.264504054897068, "grad_norm": 0.07644990086555481, "learning_rate": 1.976699666399951e-05, "loss": 0.3967, "step": 212 }, { "epoch": 0.2657517155333749, "grad_norm": 0.08435816317796707, "learning_rate": 1.9764122598395444e-05, "loss": 0.3821, "step": 213 }, { "epoch": 0.26699937616968183, "grad_norm": 0.10862424224615097, "learning_rate": 1.976123112718931e-05, "loss": 0.3533, "step": 214 }, { "epoch": 0.26824703680598877, "grad_norm": 0.5484840869903564, "learning_rate": 1.9758322255535468e-05, "loss": 0.4116, "step": 215 }, { "epoch": 0.2694946974422957, "grad_norm": 0.09223023056983948, "learning_rate": 1.9755395988619295e-05, "loss": 0.3947, "step": 216 }, { "epoch": 0.27074235807860264, "grad_norm": 0.12186647951602936, "learning_rate": 1.975245233165718e-05, "loss": 0.3597, "step": 217 }, { "epoch": 0.2719900187149095, "grad_norm": 0.08124908804893494, "learning_rate": 1.974949128989651e-05, "loss": 0.4222, "step": 218 }, { "epoch": 0.27323767935121646, "grad_norm": 0.05763809010386467, "learning_rate": 1.9746512868615656e-05, "loss": 0.3295, "step": 219 }, { "epoch": 0.2744853399875234, "grad_norm": 0.07339370250701904, "learning_rate": 1.974351707312398e-05, "loss": 0.4178, "step": 220 }, { "epoch": 0.2757330006238303, "grad_norm": 0.10172763466835022, "learning_rate": 1.974050390876181e-05, "loss": 0.4779, "step": 221 }, { "epoch": 0.27698066126013726, "grad_norm": 0.06345459818840027, "learning_rate": 1.973747338090043e-05, "loss": 0.3174, "step": 222 }, { "epoch": 0.27822832189644414, "grad_norm": 0.10512516647577286, "learning_rate": 1.9734425494942097e-05, "loss": 0.3712, "step": 223 }, { "epoch": 0.2794759825327511, "grad_norm": 0.07345092296600342, "learning_rate": 1.9731360256319982e-05, "loss": 0.3379, "step": 224 }, { "epoch": 0.280723643169058, "grad_norm": 0.08379875868558884, "learning_rate": 1.9728277670498212e-05, "loss": 0.4001, "step": 225 }, { "epoch": 0.28197130380536495, "grad_norm": 0.0922863557934761, "learning_rate": 1.972517774297183e-05, "loss": 0.4129, "step": 226 }, { "epoch": 0.2832189644416719, "grad_norm": 0.07143756747245789, "learning_rate": 1.9722060479266782e-05, "loss": 0.4302, "step": 227 }, { "epoch": 0.28446662507797876, "grad_norm": 0.07472452521324158, "learning_rate": 1.9718925884939935e-05, "loss": 0.3454, "step": 228 }, { "epoch": 0.2857142857142857, "grad_norm": 0.12237786501646042, "learning_rate": 1.9715773965579038e-05, "loss": 0.4297, "step": 229 }, { "epoch": 0.28696194635059263, "grad_norm": 0.10508041828870773, "learning_rate": 1.9712604726802732e-05, "loss": 0.4559, "step": 230 }, { "epoch": 0.28820960698689957, "grad_norm": 0.07032033056020737, "learning_rate": 1.9709418174260523e-05, "loss": 0.3811, "step": 231 }, { "epoch": 0.2894572676232065, "grad_norm": 0.1084074005484581, "learning_rate": 1.9706214313632786e-05, "loss": 0.3989, "step": 232 }, { "epoch": 0.29070492825951344, "grad_norm": 0.09418239444494247, "learning_rate": 1.9702993150630744e-05, "loss": 0.4491, "step": 233 }, { "epoch": 0.2919525888958203, "grad_norm": 0.07199984788894653, "learning_rate": 1.9699754690996476e-05, "loss": 0.408, "step": 234 }, { "epoch": 0.29320024953212726, "grad_norm": 0.10438424348831177, "learning_rate": 1.969649894050288e-05, "loss": 0.4234, "step": 235 }, { "epoch": 0.2944479101684342, "grad_norm": 0.07043947279453278, "learning_rate": 1.9693225904953682e-05, "loss": 0.3366, "step": 236 }, { "epoch": 0.2956955708047411, "grad_norm": 0.08545084297657013, "learning_rate": 1.968993559018342e-05, "loss": 0.506, "step": 237 }, { "epoch": 0.29694323144104806, "grad_norm": 0.0881681963801384, "learning_rate": 1.9686628002057445e-05, "loss": 0.4111, "step": 238 }, { "epoch": 0.29819089207735494, "grad_norm": 0.08068764954805374, "learning_rate": 1.9683303146471878e-05, "loss": 0.3981, "step": 239 }, { "epoch": 0.2994385527136619, "grad_norm": 0.17502640187740326, "learning_rate": 1.9679961029353637e-05, "loss": 0.4002, "step": 240 }, { "epoch": 0.3006862133499688, "grad_norm": 0.09356331080198288, "learning_rate": 1.967660165666041e-05, "loss": 0.4083, "step": 241 }, { "epoch": 0.30193387398627575, "grad_norm": 0.0942590981721878, "learning_rate": 1.9673225034380635e-05, "loss": 0.3643, "step": 242 }, { "epoch": 0.3031815346225827, "grad_norm": 0.09707184135913849, "learning_rate": 1.9669831168533513e-05, "loss": 0.4216, "step": 243 }, { "epoch": 0.30442919525888956, "grad_norm": 0.1203550398349762, "learning_rate": 1.9666420065168967e-05, "loss": 0.3555, "step": 244 }, { "epoch": 0.3056768558951965, "grad_norm": 0.09338900446891785, "learning_rate": 1.9662991730367664e-05, "loss": 0.3969, "step": 245 }, { "epoch": 0.30692451653150343, "grad_norm": 0.08367209136486053, "learning_rate": 1.9659546170240977e-05, "loss": 0.4305, "step": 246 }, { "epoch": 0.30817217716781037, "grad_norm": 0.08842061460018158, "learning_rate": 1.965608339093099e-05, "loss": 0.3676, "step": 247 }, { "epoch": 0.3094198378041173, "grad_norm": 0.07321275770664215, "learning_rate": 1.965260339861048e-05, "loss": 0.4718, "step": 248 }, { "epoch": 0.3094198378041173, "eval_loss": 0.37761175632476807, "eval_runtime": 2823.1422, "eval_samples_per_second": 0.253, "eval_steps_per_second": 0.126, "step": 248 } ], "logging_steps": 1, "max_steps": 2403, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 248, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.031565294349386e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }