{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.6188396756082346, "eval_steps": 248, "global_step": 496, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0012476606363069245, "grad_norm": 0.3369293212890625, "learning_rate": 4.0000000000000003e-07, "loss": 1.4366, "step": 1 }, { "epoch": 0.002495321272613849, "grad_norm": 0.36413154006004333, "learning_rate": 8.000000000000001e-07, "loss": 1.4996, "step": 2 }, { "epoch": 0.0037429819089207735, "grad_norm": 0.3281562626361847, "learning_rate": 1.2000000000000002e-06, "loss": 1.5808, "step": 3 }, { "epoch": 0.004990642545227698, "grad_norm": 0.20739880204200745, "learning_rate": 1.6000000000000001e-06, "loss": 1.5504, "step": 4 }, { "epoch": 0.006238303181534623, "grad_norm": 0.3129860460758209, "learning_rate": 2.0000000000000003e-06, "loss": 1.6271, "step": 5 }, { "epoch": 0.007485963817841547, "grad_norm": 0.2682516574859619, "learning_rate": 2.4000000000000003e-06, "loss": 1.5808, "step": 6 }, { "epoch": 0.008733624454148471, "grad_norm": 0.2960676848888397, "learning_rate": 2.8000000000000003e-06, "loss": 1.5214, "step": 7 }, { "epoch": 0.009981285090455396, "grad_norm": 0.1701919436454773, "learning_rate": 3.2000000000000003e-06, "loss": 1.5882, "step": 8 }, { "epoch": 0.011228945726762321, "grad_norm": 0.2361917495727539, "learning_rate": 3.6000000000000003e-06, "loss": 1.5905, "step": 9 }, { "epoch": 0.012476606363069246, "grad_norm": 0.30789047479629517, "learning_rate": 4.000000000000001e-06, "loss": 1.4745, "step": 10 }, { "epoch": 0.01372426699937617, "grad_norm": 0.3127429187297821, "learning_rate": 4.4e-06, "loss": 1.5997, "step": 11 }, { "epoch": 0.014971927635683094, "grad_norm": 0.4279906451702118, "learning_rate": 4.800000000000001e-06, "loss": 1.4534, "step": 12 }, { "epoch": 0.016219588271990017, "grad_norm": 0.1982978880405426, "learning_rate": 5.2e-06, "loss": 1.5678, "step": 13 }, { "epoch": 0.017467248908296942, "grad_norm": 0.35204875469207764, "learning_rate": 5.600000000000001e-06, "loss": 1.5609, "step": 14 }, { "epoch": 0.018714909544603867, "grad_norm": 0.3509620130062103, "learning_rate": 6e-06, "loss": 1.6769, "step": 15 }, { "epoch": 0.019962570180910792, "grad_norm": 0.3329513370990753, "learning_rate": 6.4000000000000006e-06, "loss": 1.4596, "step": 16 }, { "epoch": 0.021210230817217717, "grad_norm": 0.2940622568130493, "learning_rate": 6.800000000000001e-06, "loss": 1.4965, "step": 17 }, { "epoch": 0.022457891453524642, "grad_norm": 0.2941367030143738, "learning_rate": 7.2000000000000005e-06, "loss": 1.498, "step": 18 }, { "epoch": 0.023705552089831567, "grad_norm": 0.41950124502182007, "learning_rate": 7.600000000000001e-06, "loss": 1.5426, "step": 19 }, { "epoch": 0.024953212726138492, "grad_norm": 0.22718903422355652, "learning_rate": 8.000000000000001e-06, "loss": 1.5473, "step": 20 }, { "epoch": 0.026200873362445413, "grad_norm": 0.3167245388031006, "learning_rate": 8.400000000000001e-06, "loss": 1.5591, "step": 21 }, { "epoch": 0.02744853399875234, "grad_norm": 0.24035727977752686, "learning_rate": 8.8e-06, "loss": 1.5008, "step": 22 }, { "epoch": 0.028696194635059263, "grad_norm": 0.32235047221183777, "learning_rate": 9.200000000000002e-06, "loss": 1.6414, "step": 23 }, { "epoch": 0.02994385527136619, "grad_norm": 0.27163419127464294, "learning_rate": 9.600000000000001e-06, "loss": 1.534, "step": 24 }, { "epoch": 0.031191515907673113, "grad_norm": 0.19682233035564423, "learning_rate": 1e-05, "loss": 1.839, "step": 25 }, { "epoch": 0.032439176543980035, "grad_norm": 0.19141852855682373, "learning_rate": 1.04e-05, "loss": 1.4647, "step": 26 }, { "epoch": 0.03368683718028696, "grad_norm": 0.19754448533058167, "learning_rate": 1.0800000000000002e-05, "loss": 1.5247, "step": 27 }, { "epoch": 0.034934497816593885, "grad_norm": 0.21717581152915955, "learning_rate": 1.1200000000000001e-05, "loss": 1.4311, "step": 28 }, { "epoch": 0.03618215845290081, "grad_norm": 0.18977436423301697, "learning_rate": 1.16e-05, "loss": 1.4796, "step": 29 }, { "epoch": 0.037429819089207735, "grad_norm": 0.20318299531936646, "learning_rate": 1.2e-05, "loss": 1.4515, "step": 30 }, { "epoch": 0.03867747972551466, "grad_norm": 0.18727326393127441, "learning_rate": 1.2400000000000002e-05, "loss": 1.5346, "step": 31 }, { "epoch": 0.039925140361821584, "grad_norm": 0.2555878460407257, "learning_rate": 1.2800000000000001e-05, "loss": 1.4717, "step": 32 }, { "epoch": 0.041172800998128506, "grad_norm": 0.2864621579647064, "learning_rate": 1.3200000000000002e-05, "loss": 1.2702, "step": 33 }, { "epoch": 0.042420461634435434, "grad_norm": 0.14290520548820496, "learning_rate": 1.3600000000000002e-05, "loss": 1.4323, "step": 34 }, { "epoch": 0.043668122270742356, "grad_norm": 0.14048892259597778, "learning_rate": 1.4e-05, "loss": 1.4812, "step": 35 }, { "epoch": 0.044915782907049284, "grad_norm": 0.12439600378274918, "learning_rate": 1.4400000000000001e-05, "loss": 1.3209, "step": 36 }, { "epoch": 0.046163443543356206, "grad_norm": 0.11550629138946533, "learning_rate": 1.48e-05, "loss": 1.2587, "step": 37 }, { "epoch": 0.047411104179663134, "grad_norm": 0.1708095371723175, "learning_rate": 1.5200000000000002e-05, "loss": 1.4372, "step": 38 }, { "epoch": 0.048658764815970056, "grad_norm": 0.12109667807817459, "learning_rate": 1.5600000000000003e-05, "loss": 1.3698, "step": 39 }, { "epoch": 0.049906425452276984, "grad_norm": 0.3102639615535736, "learning_rate": 1.6000000000000003e-05, "loss": 1.2817, "step": 40 }, { "epoch": 0.051154086088583905, "grad_norm": 0.08972104638814926, "learning_rate": 1.64e-05, "loss": 1.2466, "step": 41 }, { "epoch": 0.05240174672489083, "grad_norm": 0.09888940304517746, "learning_rate": 1.6800000000000002e-05, "loss": 1.4295, "step": 42 }, { "epoch": 0.053649407361197755, "grad_norm": 0.09405261278152466, "learning_rate": 1.72e-05, "loss": 1.3466, "step": 43 }, { "epoch": 0.05489706799750468, "grad_norm": 0.08513078093528748, "learning_rate": 1.76e-05, "loss": 1.4428, "step": 44 }, { "epoch": 0.056144728633811605, "grad_norm": 0.09294164925813675, "learning_rate": 1.8e-05, "loss": 1.0991, "step": 45 }, { "epoch": 0.05739238927011853, "grad_norm": 0.09842169284820557, "learning_rate": 1.8400000000000003e-05, "loss": 1.1395, "step": 46 }, { "epoch": 0.058640049906425455, "grad_norm": 0.09997382760047913, "learning_rate": 1.88e-05, "loss": 1.2846, "step": 47 }, { "epoch": 0.05988771054273238, "grad_norm": 0.08888943493366241, "learning_rate": 1.9200000000000003e-05, "loss": 1.2464, "step": 48 }, { "epoch": 0.0611353711790393, "grad_norm": 0.08796142041683197, "learning_rate": 1.9600000000000002e-05, "loss": 1.2413, "step": 49 }, { "epoch": 0.06238303181534623, "grad_norm": 0.10694792866706848, "learning_rate": 2e-05, "loss": 1.1867, "step": 50 }, { "epoch": 0.06363069245165315, "grad_norm": 0.13788695633411407, "learning_rate": 1.999999108696051e-05, "loss": 1.1745, "step": 51 }, { "epoch": 0.06487835308796007, "grad_norm": 0.10375945270061493, "learning_rate": 1.999996434785792e-05, "loss": 1.1885, "step": 52 }, { "epoch": 0.066126013724267, "grad_norm": 0.10151001065969467, "learning_rate": 1.9999919782739897e-05, "loss": 1.1994, "step": 53 }, { "epoch": 0.06737367436057393, "grad_norm": 0.10222937166690826, "learning_rate": 1.999985739168588e-05, "loss": 1.1777, "step": 54 }, { "epoch": 0.06862133499688085, "grad_norm": 0.10297010838985443, "learning_rate": 1.9999777174807097e-05, "loss": 1.0911, "step": 55 }, { "epoch": 0.06986899563318777, "grad_norm": 0.1051320731639862, "learning_rate": 1.9999679132246536e-05, "loss": 1.1694, "step": 56 }, { "epoch": 0.07111665626949469, "grad_norm": 0.13086318969726562, "learning_rate": 1.999956326417897e-05, "loss": 1.2241, "step": 57 }, { "epoch": 0.07236431690580163, "grad_norm": 0.11484101414680481, "learning_rate": 1.9999429570810946e-05, "loss": 1.0468, "step": 58 }, { "epoch": 0.07361197754210855, "grad_norm": 0.11556559056043625, "learning_rate": 1.999927805238079e-05, "loss": 1.0311, "step": 59 }, { "epoch": 0.07485963817841547, "grad_norm": 0.11713112890720367, "learning_rate": 1.9999108709158594e-05, "loss": 1.0457, "step": 60 }, { "epoch": 0.07610729881472239, "grad_norm": 0.1211632490158081, "learning_rate": 1.9998921541446235e-05, "loss": 1.0551, "step": 61 }, { "epoch": 0.07735495945102933, "grad_norm": 0.12310254573822021, "learning_rate": 1.999871654957736e-05, "loss": 1.082, "step": 62 }, { "epoch": 0.07860262008733625, "grad_norm": 0.1967160552740097, "learning_rate": 1.9998493733917385e-05, "loss": 1.1005, "step": 63 }, { "epoch": 0.07985028072364317, "grad_norm": 0.12039731442928314, "learning_rate": 1.999825309486351e-05, "loss": 1.1583, "step": 64 }, { "epoch": 0.08109794135995009, "grad_norm": 0.1448286473751068, "learning_rate": 1.999799463284469e-05, "loss": 0.9149, "step": 65 }, { "epoch": 0.08234560199625701, "grad_norm": 0.13981443643569946, "learning_rate": 1.9997718348321672e-05, "loss": 0.9094, "step": 66 }, { "epoch": 0.08359326263256395, "grad_norm": 0.16553014516830444, "learning_rate": 1.9997424241786956e-05, "loss": 0.9401, "step": 67 }, { "epoch": 0.08484092326887087, "grad_norm": 0.16061550378799438, "learning_rate": 1.999711231376482e-05, "loss": 0.6806, "step": 68 }, { "epoch": 0.08608858390517779, "grad_norm": 0.157967671751976, "learning_rate": 1.999678256481131e-05, "loss": 0.7469, "step": 69 }, { "epoch": 0.08733624454148471, "grad_norm": 0.2233240157365799, "learning_rate": 1.9996434995514242e-05, "loss": 0.7202, "step": 70 }, { "epoch": 0.08858390517779165, "grad_norm": 0.18542423844337463, "learning_rate": 1.999606960649319e-05, "loss": 0.9828, "step": 71 }, { "epoch": 0.08983156581409857, "grad_norm": 0.18575641512870789, "learning_rate": 1.9995686398399505e-05, "loss": 0.6079, "step": 72 }, { "epoch": 0.09107922645040549, "grad_norm": 0.15296655893325806, "learning_rate": 1.9995285371916294e-05, "loss": 0.5708, "step": 73 }, { "epoch": 0.09232688708671241, "grad_norm": 0.21740297973155975, "learning_rate": 1.9994866527758427e-05, "loss": 0.8158, "step": 74 }, { "epoch": 0.09357454772301933, "grad_norm": 0.13831280171871185, "learning_rate": 1.9994429866672543e-05, "loss": 0.7365, "step": 75 }, { "epoch": 0.09482220835932627, "grad_norm": 0.12471487373113632, "learning_rate": 1.999397538943704e-05, "loss": 0.7776, "step": 76 }, { "epoch": 0.09606986899563319, "grad_norm": 0.11238758265972137, "learning_rate": 1.9993503096862066e-05, "loss": 0.591, "step": 77 }, { "epoch": 0.09731752963194011, "grad_norm": 0.12608778476715088, "learning_rate": 1.999301298978954e-05, "loss": 0.6285, "step": 78 }, { "epoch": 0.09856519026824703, "grad_norm": 0.12795987725257874, "learning_rate": 1.9992505069093124e-05, "loss": 0.4695, "step": 79 }, { "epoch": 0.09981285090455397, "grad_norm": 0.14311431348323822, "learning_rate": 1.9991979335678242e-05, "loss": 0.7291, "step": 80 }, { "epoch": 0.10106051154086089, "grad_norm": 0.12212368845939636, "learning_rate": 1.999143579048208e-05, "loss": 0.7634, "step": 81 }, { "epoch": 0.10230817217716781, "grad_norm": 0.0974789410829544, "learning_rate": 1.9990874434473554e-05, "loss": 0.4982, "step": 82 }, { "epoch": 0.10355583281347473, "grad_norm": 0.14408168196678162, "learning_rate": 1.999029526865335e-05, "loss": 0.5957, "step": 83 }, { "epoch": 0.10480349344978165, "grad_norm": 0.09892342239618301, "learning_rate": 1.9989698294053892e-05, "loss": 0.5647, "step": 84 }, { "epoch": 0.10605115408608859, "grad_norm": 0.10291016846895218, "learning_rate": 1.9989083511739348e-05, "loss": 0.6392, "step": 85 }, { "epoch": 0.10729881472239551, "grad_norm": 0.15446709096431732, "learning_rate": 1.9988450922805635e-05, "loss": 0.5549, "step": 86 }, { "epoch": 0.10854647535870243, "grad_norm": 0.08281419426202774, "learning_rate": 1.998780052838041e-05, "loss": 0.5886, "step": 87 }, { "epoch": 0.10979413599500935, "grad_norm": 0.09740785509347916, "learning_rate": 1.998713232962308e-05, "loss": 0.6006, "step": 88 }, { "epoch": 0.11104179663131628, "grad_norm": 0.09569819271564484, "learning_rate": 1.998644632772477e-05, "loss": 0.5268, "step": 89 }, { "epoch": 0.11228945726762321, "grad_norm": 0.08716901391744614, "learning_rate": 1.998574252390836e-05, "loss": 0.5426, "step": 90 }, { "epoch": 0.11353711790393013, "grad_norm": 0.08949075639247894, "learning_rate": 1.998502091942845e-05, "loss": 0.5637, "step": 91 }, { "epoch": 0.11478477854023705, "grad_norm": 0.10287146270275116, "learning_rate": 1.9984281515571383e-05, "loss": 0.4135, "step": 92 }, { "epoch": 0.11603243917654397, "grad_norm": 0.09598784893751144, "learning_rate": 1.9983524313655225e-05, "loss": 0.5652, "step": 93 }, { "epoch": 0.11728009981285091, "grad_norm": 0.07580853998661041, "learning_rate": 1.998274931502977e-05, "loss": 0.5787, "step": 94 }, { "epoch": 0.11852776044915783, "grad_norm": 0.0887116864323616, "learning_rate": 1.998195652107653e-05, "loss": 0.5494, "step": 95 }, { "epoch": 0.11977542108546475, "grad_norm": 0.09782399237155914, "learning_rate": 1.998114593320876e-05, "loss": 0.4127, "step": 96 }, { "epoch": 0.12102308172177167, "grad_norm": 0.09533873945474625, "learning_rate": 1.9980317552871407e-05, "loss": 0.5192, "step": 97 }, { "epoch": 0.1222707423580786, "grad_norm": 0.08982633054256439, "learning_rate": 1.9979471381541158e-05, "loss": 0.4699, "step": 98 }, { "epoch": 0.12351840299438553, "grad_norm": 0.0871262177824974, "learning_rate": 1.9978607420726396e-05, "loss": 0.5772, "step": 99 }, { "epoch": 0.12476606363069245, "grad_norm": 0.08864148706197739, "learning_rate": 1.997772567196723e-05, "loss": 0.6495, "step": 100 }, { "epoch": 0.1260137242669994, "grad_norm": 0.5612615942955017, "learning_rate": 1.997682613683547e-05, "loss": 0.3912, "step": 101 }, { "epoch": 0.1272613849033063, "grad_norm": 0.18454571068286896, "learning_rate": 1.997590881693464e-05, "loss": 0.6133, "step": 102 }, { "epoch": 0.12850904553961323, "grad_norm": 0.11318469792604446, "learning_rate": 1.997497371389995e-05, "loss": 0.4185, "step": 103 }, { "epoch": 0.12975670617592014, "grad_norm": 0.09561355412006378, "learning_rate": 1.9974020829398335e-05, "loss": 0.5241, "step": 104 }, { "epoch": 0.13100436681222707, "grad_norm": 0.1466185599565506, "learning_rate": 1.9973050165128406e-05, "loss": 0.559, "step": 105 }, { "epoch": 0.132252027448534, "grad_norm": 0.0754123404622078, "learning_rate": 1.9972061722820483e-05, "loss": 0.4629, "step": 106 }, { "epoch": 0.13349968808484092, "grad_norm": 0.10411768406629562, "learning_rate": 1.997105550423656e-05, "loss": 0.3849, "step": 107 }, { "epoch": 0.13474734872114785, "grad_norm": 0.09193742275238037, "learning_rate": 1.997003151117034e-05, "loss": 0.3919, "step": 108 }, { "epoch": 0.13599500935745476, "grad_norm": 0.16121919453144073, "learning_rate": 1.9968989745447203e-05, "loss": 0.6253, "step": 109 }, { "epoch": 0.1372426699937617, "grad_norm": 0.09475496411323547, "learning_rate": 1.99679302089242e-05, "loss": 0.5236, "step": 110 }, { "epoch": 0.13849033063006863, "grad_norm": 0.09533608704805374, "learning_rate": 1.9966852903490077e-05, "loss": 0.4389, "step": 111 }, { "epoch": 0.13973799126637554, "grad_norm": 0.08322104066610336, "learning_rate": 1.996575783106524e-05, "loss": 0.5455, "step": 112 }, { "epoch": 0.14098565190268247, "grad_norm": 0.1089191734790802, "learning_rate": 1.9964644993601782e-05, "loss": 0.4437, "step": 113 }, { "epoch": 0.14223331253898938, "grad_norm": 0.117027148604393, "learning_rate": 1.9963514393083452e-05, "loss": 0.4787, "step": 114 }, { "epoch": 0.14348097317529632, "grad_norm": 0.08248787373304367, "learning_rate": 1.9962366031525663e-05, "loss": 0.4845, "step": 115 }, { "epoch": 0.14472863381160325, "grad_norm": 0.08650378882884979, "learning_rate": 1.9961199910975497e-05, "loss": 0.3742, "step": 116 }, { "epoch": 0.14597629444791016, "grad_norm": 0.07683917135000229, "learning_rate": 1.9960016033511692e-05, "loss": 0.4035, "step": 117 }, { "epoch": 0.1472239550842171, "grad_norm": 0.08383751660585403, "learning_rate": 1.9958814401244634e-05, "loss": 0.5135, "step": 118 }, { "epoch": 0.14847161572052403, "grad_norm": 0.12660759687423706, "learning_rate": 1.995759501631637e-05, "loss": 0.6356, "step": 119 }, { "epoch": 0.14971927635683094, "grad_norm": 0.09009061008691788, "learning_rate": 1.995635788090057e-05, "loss": 0.4618, "step": 120 }, { "epoch": 0.15096693699313787, "grad_norm": 0.10818025469779968, "learning_rate": 1.9955102997202575e-05, "loss": 0.5522, "step": 121 }, { "epoch": 0.15221459762944478, "grad_norm": 0.07673201709985733, "learning_rate": 1.9953830367459343e-05, "loss": 0.3801, "step": 122 }, { "epoch": 0.15346225826575172, "grad_norm": 0.09158243238925934, "learning_rate": 1.9952539993939478e-05, "loss": 0.413, "step": 123 }, { "epoch": 0.15470991890205865, "grad_norm": 0.11646990478038788, "learning_rate": 1.9951231878943207e-05, "loss": 0.4732, "step": 124 }, { "epoch": 0.15595757953836556, "grad_norm": 0.0766897052526474, "learning_rate": 1.994990602480239e-05, "loss": 0.4355, "step": 125 }, { "epoch": 0.1572052401746725, "grad_norm": 0.10383328795433044, "learning_rate": 1.9948562433880498e-05, "loss": 0.433, "step": 126 }, { "epoch": 0.1584529008109794, "grad_norm": 0.07712090015411377, "learning_rate": 1.9947201108572636e-05, "loss": 0.4848, "step": 127 }, { "epoch": 0.15970056144728634, "grad_norm": 0.14091262221336365, "learning_rate": 1.994582205130551e-05, "loss": 0.439, "step": 128 }, { "epoch": 0.16094822208359327, "grad_norm": 0.10989029705524445, "learning_rate": 1.9944425264537432e-05, "loss": 0.4429, "step": 129 }, { "epoch": 0.16219588271990018, "grad_norm": 0.1106753721833229, "learning_rate": 1.9943010750758335e-05, "loss": 0.4662, "step": 130 }, { "epoch": 0.16344354335620712, "grad_norm": 0.09498463571071625, "learning_rate": 1.994157851248973e-05, "loss": 0.4164, "step": 131 }, { "epoch": 0.16469120399251402, "grad_norm": 0.0848456621170044, "learning_rate": 1.9940128552284756e-05, "loss": 0.4634, "step": 132 }, { "epoch": 0.16593886462882096, "grad_norm": 0.06310152262449265, "learning_rate": 1.99386608727281e-05, "loss": 0.4834, "step": 133 }, { "epoch": 0.1671865252651279, "grad_norm": 0.10514630377292633, "learning_rate": 1.9937175476436078e-05, "loss": 0.4392, "step": 134 }, { "epoch": 0.1684341859014348, "grad_norm": 0.47956305742263794, "learning_rate": 1.9935672366056557e-05, "loss": 0.4975, "step": 135 }, { "epoch": 0.16968184653774174, "grad_norm": 0.12076468020677567, "learning_rate": 1.9934151544269e-05, "loss": 0.425, "step": 136 }, { "epoch": 0.17092950717404864, "grad_norm": 0.07710213214159012, "learning_rate": 1.9932613013784436e-05, "loss": 0.4426, "step": 137 }, { "epoch": 0.17217716781035558, "grad_norm": 0.1181124597787857, "learning_rate": 1.993105677734546e-05, "loss": 0.4272, "step": 138 }, { "epoch": 0.17342482844666252, "grad_norm": 0.08521714061498642, "learning_rate": 1.9929482837726227e-05, "loss": 0.4041, "step": 139 }, { "epoch": 0.17467248908296942, "grad_norm": 0.0910830870270729, "learning_rate": 1.992789119773246e-05, "loss": 0.4372, "step": 140 }, { "epoch": 0.17592014971927636, "grad_norm": 0.09005699306726456, "learning_rate": 1.9926281860201427e-05, "loss": 0.3363, "step": 141 }, { "epoch": 0.1771678103555833, "grad_norm": 0.1397104263305664, "learning_rate": 1.9924654828001947e-05, "loss": 0.5186, "step": 142 }, { "epoch": 0.1784154709918902, "grad_norm": 0.0795428678393364, "learning_rate": 1.9923010104034382e-05, "loss": 0.4258, "step": 143 }, { "epoch": 0.17966313162819714, "grad_norm": 0.11801055818796158, "learning_rate": 1.9921347691230622e-05, "loss": 0.4102, "step": 144 }, { "epoch": 0.18091079226450404, "grad_norm": 0.09254780411720276, "learning_rate": 1.9919667592554107e-05, "loss": 0.4247, "step": 145 }, { "epoch": 0.18215845290081098, "grad_norm": 0.09013550728559494, "learning_rate": 1.991796981099979e-05, "loss": 0.4397, "step": 146 }, { "epoch": 0.18340611353711792, "grad_norm": 0.10347352921962738, "learning_rate": 1.991625434959415e-05, "loss": 0.3864, "step": 147 }, { "epoch": 0.18465377417342482, "grad_norm": 0.08151372522115707, "learning_rate": 1.991452121139518e-05, "loss": 0.3866, "step": 148 }, { "epoch": 0.18590143480973176, "grad_norm": 0.07901383191347122, "learning_rate": 1.9912770399492385e-05, "loss": 0.4413, "step": 149 }, { "epoch": 0.18714909544603867, "grad_norm": 0.10809416323900223, "learning_rate": 1.9911001917006782e-05, "loss": 0.4892, "step": 150 }, { "epoch": 0.1883967560823456, "grad_norm": 0.0716921016573906, "learning_rate": 1.990921576709088e-05, "loss": 0.377, "step": 151 }, { "epoch": 0.18964441671865254, "grad_norm": 0.07814662158489227, "learning_rate": 1.9907411952928676e-05, "loss": 0.3738, "step": 152 }, { "epoch": 0.19089207735495944, "grad_norm": 0.0741223469376564, "learning_rate": 1.9905590477735677e-05, "loss": 0.4715, "step": 153 }, { "epoch": 0.19213973799126638, "grad_norm": 0.11747987568378448, "learning_rate": 1.990375134475885e-05, "loss": 0.3903, "step": 154 }, { "epoch": 0.1933873986275733, "grad_norm": 0.08143888413906097, "learning_rate": 1.9901894557276647e-05, "loss": 0.4747, "step": 155 }, { "epoch": 0.19463505926388022, "grad_norm": 0.08514941483736038, "learning_rate": 1.9900020118598996e-05, "loss": 0.4365, "step": 156 }, { "epoch": 0.19588271990018716, "grad_norm": 0.07622472196817398, "learning_rate": 1.9898128032067287e-05, "loss": 0.4311, "step": 157 }, { "epoch": 0.19713038053649407, "grad_norm": 0.09147831797599792, "learning_rate": 1.9896218301054365e-05, "loss": 0.4376, "step": 158 }, { "epoch": 0.198378041172801, "grad_norm": 0.09080328047275543, "learning_rate": 1.9894290928964536e-05, "loss": 0.4817, "step": 159 }, { "epoch": 0.19962570180910794, "grad_norm": 0.0705234482884407, "learning_rate": 1.9892345919233543e-05, "loss": 0.3602, "step": 160 }, { "epoch": 0.20087336244541484, "grad_norm": 0.3075416088104248, "learning_rate": 1.9890383275328577e-05, "loss": 0.3686, "step": 161 }, { "epoch": 0.20212102308172178, "grad_norm": 0.0864262729883194, "learning_rate": 1.9888403000748268e-05, "loss": 0.5076, "step": 162 }, { "epoch": 0.2033686837180287, "grad_norm": 0.07593845576047897, "learning_rate": 1.988640509902266e-05, "loss": 0.3319, "step": 163 }, { "epoch": 0.20461634435433562, "grad_norm": 0.0860196128487587, "learning_rate": 1.988438957371324e-05, "loss": 0.4247, "step": 164 }, { "epoch": 0.20586400499064256, "grad_norm": 0.08619983494281769, "learning_rate": 1.9882356428412887e-05, "loss": 0.4515, "step": 165 }, { "epoch": 0.20711166562694946, "grad_norm": 0.08582403510808945, "learning_rate": 1.988030566674591e-05, "loss": 0.4861, "step": 166 }, { "epoch": 0.2083593262632564, "grad_norm": 0.09421785920858383, "learning_rate": 1.9878237292368014e-05, "loss": 0.3536, "step": 167 }, { "epoch": 0.2096069868995633, "grad_norm": 0.08474161475896835, "learning_rate": 1.9876151308966293e-05, "loss": 0.4128, "step": 168 }, { "epoch": 0.21085464753587024, "grad_norm": 0.07902742177248001, "learning_rate": 1.987404772025924e-05, "loss": 0.3909, "step": 169 }, { "epoch": 0.21210230817217718, "grad_norm": 0.12060698121786118, "learning_rate": 1.987192652999673e-05, "loss": 0.357, "step": 170 }, { "epoch": 0.2133499688084841, "grad_norm": 0.06981324404478073, "learning_rate": 1.9869787741960013e-05, "loss": 0.4047, "step": 171 }, { "epoch": 0.21459762944479102, "grad_norm": 0.11325976252555847, "learning_rate": 1.9867631359961712e-05, "loss": 0.4094, "step": 172 }, { "epoch": 0.21584529008109793, "grad_norm": 0.07985086739063263, "learning_rate": 1.9865457387845806e-05, "loss": 0.4056, "step": 173 }, { "epoch": 0.21709295071740486, "grad_norm": 0.08571935445070267, "learning_rate": 1.9863265829487637e-05, "loss": 0.3637, "step": 174 }, { "epoch": 0.2183406113537118, "grad_norm": 0.11255648732185364, "learning_rate": 1.9861056688793896e-05, "loss": 0.4475, "step": 175 }, { "epoch": 0.2195882719900187, "grad_norm": 0.08018800616264343, "learning_rate": 1.985882996970261e-05, "loss": 0.4584, "step": 176 }, { "epoch": 0.22083593262632564, "grad_norm": 0.07600877434015274, "learning_rate": 1.985658567618315e-05, "loss": 0.3794, "step": 177 }, { "epoch": 0.22208359326263255, "grad_norm": 0.07584904879331589, "learning_rate": 1.985432381223621e-05, "loss": 0.3718, "step": 178 }, { "epoch": 0.22333125389893949, "grad_norm": 0.07508600503206253, "learning_rate": 1.9852044381893806e-05, "loss": 0.4392, "step": 179 }, { "epoch": 0.22457891453524642, "grad_norm": 0.09708955883979797, "learning_rate": 1.9849747389219272e-05, "loss": 0.3792, "step": 180 }, { "epoch": 0.22582657517155333, "grad_norm": 0.11540783196687698, "learning_rate": 1.984743283830724e-05, "loss": 0.4075, "step": 181 }, { "epoch": 0.22707423580786026, "grad_norm": 0.08834560215473175, "learning_rate": 1.984510073328365e-05, "loss": 0.3469, "step": 182 }, { "epoch": 0.2283218964441672, "grad_norm": 0.09154046326875687, "learning_rate": 1.9842751078305735e-05, "loss": 0.411, "step": 183 }, { "epoch": 0.2295695570804741, "grad_norm": 0.07195434719324112, "learning_rate": 1.9840383877561998e-05, "loss": 0.385, "step": 184 }, { "epoch": 0.23081721771678104, "grad_norm": 0.10304310917854309, "learning_rate": 1.9837999135272235e-05, "loss": 0.3727, "step": 185 }, { "epoch": 0.23206487835308795, "grad_norm": 0.0923156812787056, "learning_rate": 1.983559685568751e-05, "loss": 0.4777, "step": 186 }, { "epoch": 0.23331253898939489, "grad_norm": 0.07710360735654831, "learning_rate": 1.983317704309014e-05, "loss": 0.3419, "step": 187 }, { "epoch": 0.23456019962570182, "grad_norm": 0.072215236723423, "learning_rate": 1.9830739701793705e-05, "loss": 0.342, "step": 188 }, { "epoch": 0.23580786026200873, "grad_norm": 0.0766647532582283, "learning_rate": 1.982828483614303e-05, "loss": 0.3672, "step": 189 }, { "epoch": 0.23705552089831566, "grad_norm": 0.07266882061958313, "learning_rate": 1.9825812450514174e-05, "loss": 0.4228, "step": 190 }, { "epoch": 0.23830318153462257, "grad_norm": 0.06986986100673676, "learning_rate": 1.9823322549314433e-05, "loss": 0.3741, "step": 191 }, { "epoch": 0.2395508421709295, "grad_norm": 0.08523140847682953, "learning_rate": 1.9820815136982327e-05, "loss": 0.4154, "step": 192 }, { "epoch": 0.24079850280723644, "grad_norm": 0.07692936062812805, "learning_rate": 1.9818290217987587e-05, "loss": 0.4984, "step": 193 }, { "epoch": 0.24204616344354335, "grad_norm": 0.11332568526268005, "learning_rate": 1.9815747796831154e-05, "loss": 0.377, "step": 194 }, { "epoch": 0.24329382407985028, "grad_norm": 0.06331691890954971, "learning_rate": 1.9813187878045164e-05, "loss": 0.3134, "step": 195 }, { "epoch": 0.2445414847161572, "grad_norm": 0.09051994234323502, "learning_rate": 1.9810610466192955e-05, "loss": 0.3818, "step": 196 }, { "epoch": 0.24578914535246413, "grad_norm": 0.09058686345815659, "learning_rate": 1.980801556586904e-05, "loss": 0.35, "step": 197 }, { "epoch": 0.24703680598877106, "grad_norm": 0.07872316241264343, "learning_rate": 1.9805403181699104e-05, "loss": 0.4229, "step": 198 }, { "epoch": 0.24828446662507797, "grad_norm": 0.07484429329633713, "learning_rate": 1.980277331834001e-05, "loss": 0.4025, "step": 199 }, { "epoch": 0.2495321272613849, "grad_norm": 0.08303457498550415, "learning_rate": 1.9800125980479768e-05, "loss": 0.4001, "step": 200 }, { "epoch": 0.25077978789769184, "grad_norm": 0.06768237799406052, "learning_rate": 1.9797461172837545e-05, "loss": 0.2898, "step": 201 }, { "epoch": 0.2520274485339988, "grad_norm": 0.10813165456056595, "learning_rate": 1.9794778900163653e-05, "loss": 0.3832, "step": 202 }, { "epoch": 0.25327510917030566, "grad_norm": 0.07230770587921143, "learning_rate": 1.9792079167239526e-05, "loss": 0.4103, "step": 203 }, { "epoch": 0.2545227698066126, "grad_norm": 0.07701678574085236, "learning_rate": 1.978936197887773e-05, "loss": 0.3438, "step": 204 }, { "epoch": 0.2557704304429195, "grad_norm": 0.07222580909729004, "learning_rate": 1.978662733992195e-05, "loss": 0.3693, "step": 205 }, { "epoch": 0.25701809107922646, "grad_norm": 0.07187530398368835, "learning_rate": 1.9783875255246972e-05, "loss": 0.3753, "step": 206 }, { "epoch": 0.2582657517155334, "grad_norm": 0.06922349333763123, "learning_rate": 1.9781105729758688e-05, "loss": 0.4076, "step": 207 }, { "epoch": 0.2595134123518403, "grad_norm": 0.07130351662635803, "learning_rate": 1.9778318768394072e-05, "loss": 0.411, "step": 208 }, { "epoch": 0.2607610729881472, "grad_norm": 0.07194481790065765, "learning_rate": 1.9775514376121184e-05, "loss": 0.3342, "step": 209 }, { "epoch": 0.26200873362445415, "grad_norm": 0.13812008500099182, "learning_rate": 1.977269255793915e-05, "loss": 0.3947, "step": 210 }, { "epoch": 0.2632563942607611, "grad_norm": 0.08995859324932098, "learning_rate": 1.9769853318878177e-05, "loss": 0.4331, "step": 211 }, { "epoch": 0.264504054897068, "grad_norm": 0.07644990086555481, "learning_rate": 1.976699666399951e-05, "loss": 0.3967, "step": 212 }, { "epoch": 0.2657517155333749, "grad_norm": 0.08435816317796707, "learning_rate": 1.9764122598395444e-05, "loss": 0.3821, "step": 213 }, { "epoch": 0.26699937616968183, "grad_norm": 0.10862424224615097, "learning_rate": 1.976123112718931e-05, "loss": 0.3533, "step": 214 }, { "epoch": 0.26824703680598877, "grad_norm": 0.5484840869903564, "learning_rate": 1.9758322255535468e-05, "loss": 0.4116, "step": 215 }, { "epoch": 0.2694946974422957, "grad_norm": 0.09223023056983948, "learning_rate": 1.9755395988619295e-05, "loss": 0.3947, "step": 216 }, { "epoch": 0.27074235807860264, "grad_norm": 0.12186647951602936, "learning_rate": 1.975245233165718e-05, "loss": 0.3597, "step": 217 }, { "epoch": 0.2719900187149095, "grad_norm": 0.08124908804893494, "learning_rate": 1.974949128989651e-05, "loss": 0.4222, "step": 218 }, { "epoch": 0.27323767935121646, "grad_norm": 0.05763809010386467, "learning_rate": 1.9746512868615656e-05, "loss": 0.3295, "step": 219 }, { "epoch": 0.2744853399875234, "grad_norm": 0.07339370250701904, "learning_rate": 1.974351707312398e-05, "loss": 0.4178, "step": 220 }, { "epoch": 0.2757330006238303, "grad_norm": 0.10172763466835022, "learning_rate": 1.974050390876181e-05, "loss": 0.4779, "step": 221 }, { "epoch": 0.27698066126013726, "grad_norm": 0.06345459818840027, "learning_rate": 1.973747338090043e-05, "loss": 0.3174, "step": 222 }, { "epoch": 0.27822832189644414, "grad_norm": 0.10512516647577286, "learning_rate": 1.9734425494942097e-05, "loss": 0.3712, "step": 223 }, { "epoch": 0.2794759825327511, "grad_norm": 0.07345092296600342, "learning_rate": 1.9731360256319982e-05, "loss": 0.3379, "step": 224 }, { "epoch": 0.280723643169058, "grad_norm": 0.08379875868558884, "learning_rate": 1.9728277670498212e-05, "loss": 0.4001, "step": 225 }, { "epoch": 0.28197130380536495, "grad_norm": 0.0922863557934761, "learning_rate": 1.972517774297183e-05, "loss": 0.4129, "step": 226 }, { "epoch": 0.2832189644416719, "grad_norm": 0.07143756747245789, "learning_rate": 1.9722060479266782e-05, "loss": 0.4302, "step": 227 }, { "epoch": 0.28446662507797876, "grad_norm": 0.07472452521324158, "learning_rate": 1.9718925884939935e-05, "loss": 0.3454, "step": 228 }, { "epoch": 0.2857142857142857, "grad_norm": 0.12237786501646042, "learning_rate": 1.9715773965579038e-05, "loss": 0.4297, "step": 229 }, { "epoch": 0.28696194635059263, "grad_norm": 0.10508041828870773, "learning_rate": 1.9712604726802732e-05, "loss": 0.4559, "step": 230 }, { "epoch": 0.28820960698689957, "grad_norm": 0.07032033056020737, "learning_rate": 1.9709418174260523e-05, "loss": 0.3811, "step": 231 }, { "epoch": 0.2894572676232065, "grad_norm": 0.1084074005484581, "learning_rate": 1.9706214313632786e-05, "loss": 0.3989, "step": 232 }, { "epoch": 0.29070492825951344, "grad_norm": 0.09418239444494247, "learning_rate": 1.9702993150630744e-05, "loss": 0.4491, "step": 233 }, { "epoch": 0.2919525888958203, "grad_norm": 0.07199984788894653, "learning_rate": 1.9699754690996476e-05, "loss": 0.408, "step": 234 }, { "epoch": 0.29320024953212726, "grad_norm": 0.10438424348831177, "learning_rate": 1.969649894050288e-05, "loss": 0.4234, "step": 235 }, { "epoch": 0.2944479101684342, "grad_norm": 0.07043947279453278, "learning_rate": 1.9693225904953682e-05, "loss": 0.3366, "step": 236 }, { "epoch": 0.2956955708047411, "grad_norm": 0.08545084297657013, "learning_rate": 1.968993559018342e-05, "loss": 0.506, "step": 237 }, { "epoch": 0.29694323144104806, "grad_norm": 0.0881681963801384, "learning_rate": 1.9686628002057445e-05, "loss": 0.4111, "step": 238 }, { "epoch": 0.29819089207735494, "grad_norm": 0.08068764954805374, "learning_rate": 1.9683303146471878e-05, "loss": 0.3981, "step": 239 }, { "epoch": 0.2994385527136619, "grad_norm": 0.17502640187740326, "learning_rate": 1.9679961029353637e-05, "loss": 0.4002, "step": 240 }, { "epoch": 0.3006862133499688, "grad_norm": 0.09356331080198288, "learning_rate": 1.967660165666041e-05, "loss": 0.4083, "step": 241 }, { "epoch": 0.30193387398627575, "grad_norm": 0.0942590981721878, "learning_rate": 1.9673225034380635e-05, "loss": 0.3643, "step": 242 }, { "epoch": 0.3031815346225827, "grad_norm": 0.09707184135913849, "learning_rate": 1.9669831168533513e-05, "loss": 0.4216, "step": 243 }, { "epoch": 0.30442919525888956, "grad_norm": 0.1203550398349762, "learning_rate": 1.9666420065168967e-05, "loss": 0.3555, "step": 244 }, { "epoch": 0.3056768558951965, "grad_norm": 0.09338900446891785, "learning_rate": 1.9662991730367664e-05, "loss": 0.3969, "step": 245 }, { "epoch": 0.30692451653150343, "grad_norm": 0.08367209136486053, "learning_rate": 1.9659546170240977e-05, "loss": 0.4305, "step": 246 }, { "epoch": 0.30817217716781037, "grad_norm": 0.08842061460018158, "learning_rate": 1.965608339093099e-05, "loss": 0.3676, "step": 247 }, { "epoch": 0.3094198378041173, "grad_norm": 0.07321275770664215, "learning_rate": 1.965260339861048e-05, "loss": 0.4718, "step": 248 }, { "epoch": 0.3094198378041173, "eval_loss": 0.37761175632476807, "eval_runtime": 2823.1422, "eval_samples_per_second": 0.253, "eval_steps_per_second": 0.126, "step": 248 }, { "epoch": 0.3106674984404242, "grad_norm": 0.12405838817358017, "learning_rate": 1.964910619948291e-05, "loss": 0.4069, "step": 249 }, { "epoch": 0.3119151590767311, "grad_norm": 0.08042417466640472, "learning_rate": 1.9645591799782416e-05, "loss": 0.4346, "step": 250 }, { "epoch": 0.31316281971303805, "grad_norm": 0.07625039666891098, "learning_rate": 1.9642060205773793e-05, "loss": 0.4323, "step": 251 }, { "epoch": 0.314410480349345, "grad_norm": 0.07860994338989258, "learning_rate": 1.9638511423752486e-05, "loss": 0.3794, "step": 252 }, { "epoch": 0.3156581409856519, "grad_norm": 0.10293662548065186, "learning_rate": 1.9634945460044587e-05, "loss": 0.3621, "step": 253 }, { "epoch": 0.3169058016219588, "grad_norm": 0.07868197560310364, "learning_rate": 1.9631362321006807e-05, "loss": 0.4549, "step": 254 }, { "epoch": 0.31815346225826574, "grad_norm": 0.08555343747138977, "learning_rate": 1.9627762013026474e-05, "loss": 0.3852, "step": 255 }, { "epoch": 0.3194011228945727, "grad_norm": 0.08398033678531647, "learning_rate": 1.9624144542521538e-05, "loss": 0.3722, "step": 256 }, { "epoch": 0.3206487835308796, "grad_norm": 0.07331738620996475, "learning_rate": 1.962050991594052e-05, "loss": 0.3561, "step": 257 }, { "epoch": 0.32189644416718655, "grad_norm": 0.07319743931293488, "learning_rate": 1.9616858139762534e-05, "loss": 0.3152, "step": 258 }, { "epoch": 0.3231441048034934, "grad_norm": 0.11726346611976624, "learning_rate": 1.9613189220497273e-05, "loss": 0.4616, "step": 259 }, { "epoch": 0.32439176543980036, "grad_norm": 0.10728958249092102, "learning_rate": 1.9609503164684974e-05, "loss": 0.3708, "step": 260 }, { "epoch": 0.3256394260761073, "grad_norm": 0.08347611129283905, "learning_rate": 1.9605799978896433e-05, "loss": 0.406, "step": 261 }, { "epoch": 0.32688708671241423, "grad_norm": 0.08457101881504059, "learning_rate": 1.9602079669732975e-05, "loss": 0.3541, "step": 262 }, { "epoch": 0.32813474734872117, "grad_norm": 0.10014831274747849, "learning_rate": 1.9598342243826457e-05, "loss": 0.4757, "step": 263 }, { "epoch": 0.32938240798502805, "grad_norm": 0.09075610339641571, "learning_rate": 1.9594587707839242e-05, "loss": 0.4567, "step": 264 }, { "epoch": 0.330630068621335, "grad_norm": 0.06062496826052666, "learning_rate": 1.9590816068464192e-05, "loss": 0.2811, "step": 265 }, { "epoch": 0.3318777292576419, "grad_norm": 0.10323761403560638, "learning_rate": 1.9587027332424668e-05, "loss": 0.4135, "step": 266 }, { "epoch": 0.33312538989394885, "grad_norm": 0.0712403953075409, "learning_rate": 1.9583221506474494e-05, "loss": 0.3841, "step": 267 }, { "epoch": 0.3343730505302558, "grad_norm": 0.12237131595611572, "learning_rate": 1.957939859739797e-05, "loss": 0.3045, "step": 268 }, { "epoch": 0.33562071116656267, "grad_norm": 0.09809068590402603, "learning_rate": 1.957555861200984e-05, "loss": 0.3938, "step": 269 }, { "epoch": 0.3368683718028696, "grad_norm": 0.15238316357135773, "learning_rate": 1.9571701557155295e-05, "loss": 0.3669, "step": 270 }, { "epoch": 0.33811603243917654, "grad_norm": 0.08934526145458221, "learning_rate": 1.9567827439709954e-05, "loss": 0.3647, "step": 271 }, { "epoch": 0.3393636930754835, "grad_norm": 0.0997021347284317, "learning_rate": 1.9563936266579845e-05, "loss": 0.3665, "step": 272 }, { "epoch": 0.3406113537117904, "grad_norm": 0.09582076221704483, "learning_rate": 1.9560028044701404e-05, "loss": 0.3686, "step": 273 }, { "epoch": 0.3418590143480973, "grad_norm": 0.1995895355939865, "learning_rate": 1.955610278104146e-05, "loss": 0.3884, "step": 274 }, { "epoch": 0.3431066749844042, "grad_norm": 0.13371475040912628, "learning_rate": 1.955216048259722e-05, "loss": 0.4411, "step": 275 }, { "epoch": 0.34435433562071116, "grad_norm": 0.0868840217590332, "learning_rate": 1.9548201156396255e-05, "loss": 0.4014, "step": 276 }, { "epoch": 0.3456019962570181, "grad_norm": 0.06779041141271591, "learning_rate": 1.954422480949649e-05, "loss": 0.2838, "step": 277 }, { "epoch": 0.34684965689332503, "grad_norm": 0.08861416578292847, "learning_rate": 1.9540231448986193e-05, "loss": 0.3724, "step": 278 }, { "epoch": 0.34809731752963197, "grad_norm": 0.12394684553146362, "learning_rate": 1.953622108198396e-05, "loss": 0.388, "step": 279 }, { "epoch": 0.34934497816593885, "grad_norm": 0.06695356220006943, "learning_rate": 1.9532193715638703e-05, "loss": 0.2598, "step": 280 }, { "epoch": 0.3505926388022458, "grad_norm": 0.07081321626901627, "learning_rate": 1.9528149357129637e-05, "loss": 0.3504, "step": 281 }, { "epoch": 0.3518402994385527, "grad_norm": 0.07354036718606949, "learning_rate": 1.9524088013666272e-05, "loss": 0.3052, "step": 282 }, { "epoch": 0.35308796007485965, "grad_norm": 0.06471217423677444, "learning_rate": 1.9520009692488382e-05, "loss": 0.3449, "step": 283 }, { "epoch": 0.3543356207111666, "grad_norm": 0.07577944546937943, "learning_rate": 1.9515914400866022e-05, "loss": 0.3782, "step": 284 }, { "epoch": 0.35558328134747347, "grad_norm": 0.06845004856586456, "learning_rate": 1.9511802146099486e-05, "loss": 0.282, "step": 285 }, { "epoch": 0.3568309419837804, "grad_norm": 0.09513509273529053, "learning_rate": 1.9507672935519318e-05, "loss": 0.3563, "step": 286 }, { "epoch": 0.35807860262008734, "grad_norm": 0.06815797835588455, "learning_rate": 1.9503526776486275e-05, "loss": 0.298, "step": 287 }, { "epoch": 0.3593262632563943, "grad_norm": 0.0746559351682663, "learning_rate": 1.9499363676391337e-05, "loss": 0.3185, "step": 288 }, { "epoch": 0.3605739238927012, "grad_norm": 0.10398710519075394, "learning_rate": 1.9495183642655674e-05, "loss": 0.4161, "step": 289 }, { "epoch": 0.3618215845290081, "grad_norm": 0.09173806756734848, "learning_rate": 1.949098668273065e-05, "loss": 0.4049, "step": 290 }, { "epoch": 0.363069245165315, "grad_norm": 0.08424055576324463, "learning_rate": 1.9486772804097807e-05, "loss": 0.2972, "step": 291 }, { "epoch": 0.36431690580162196, "grad_norm": 0.09673412144184113, "learning_rate": 1.948254201426883e-05, "loss": 0.2899, "step": 292 }, { "epoch": 0.3655645664379289, "grad_norm": 0.07556866854429245, "learning_rate": 1.9478294320785553e-05, "loss": 0.4131, "step": 293 }, { "epoch": 0.36681222707423583, "grad_norm": 0.13352152705192566, "learning_rate": 1.9474029731219958e-05, "loss": 0.3592, "step": 294 }, { "epoch": 0.3680598877105427, "grad_norm": 0.08789486438035965, "learning_rate": 1.9469748253174132e-05, "loss": 0.4613, "step": 295 }, { "epoch": 0.36930754834684965, "grad_norm": 0.2030758261680603, "learning_rate": 1.946544989428027e-05, "loss": 0.3532, "step": 296 }, { "epoch": 0.3705552089831566, "grad_norm": 0.058304548263549805, "learning_rate": 1.9461134662200667e-05, "loss": 0.2787, "step": 297 }, { "epoch": 0.3718028696194635, "grad_norm": 2.858107089996338, "learning_rate": 1.9456802564627682e-05, "loss": 0.4308, "step": 298 }, { "epoch": 0.37305053025577045, "grad_norm": 0.07007814943790436, "learning_rate": 1.9452453609283754e-05, "loss": 0.3738, "step": 299 }, { "epoch": 0.37429819089207733, "grad_norm": 0.07296311110258102, "learning_rate": 1.9448087803921358e-05, "loss": 0.3494, "step": 300 }, { "epoch": 0.37554585152838427, "grad_norm": 0.0859873816370964, "learning_rate": 1.9443705156323016e-05, "loss": 0.3019, "step": 301 }, { "epoch": 0.3767935121646912, "grad_norm": 0.06829937547445297, "learning_rate": 1.9439305674301265e-05, "loss": 0.3248, "step": 302 }, { "epoch": 0.37804117280099814, "grad_norm": 0.13362844288349152, "learning_rate": 1.9434889365698668e-05, "loss": 0.3045, "step": 303 }, { "epoch": 0.3792888334373051, "grad_norm": 0.07973142713308334, "learning_rate": 1.9430456238387763e-05, "loss": 0.3811, "step": 304 }, { "epoch": 0.38053649407361195, "grad_norm": 0.08348105847835541, "learning_rate": 1.9426006300271082e-05, "loss": 0.2735, "step": 305 }, { "epoch": 0.3817841547099189, "grad_norm": 0.0983782708644867, "learning_rate": 1.9421539559281114e-05, "loss": 0.4023, "step": 306 }, { "epoch": 0.3830318153462258, "grad_norm": 0.10556434094905853, "learning_rate": 1.9417056023380312e-05, "loss": 0.3741, "step": 307 }, { "epoch": 0.38427947598253276, "grad_norm": 0.10459857434034348, "learning_rate": 1.941255570056106e-05, "loss": 0.4195, "step": 308 }, { "epoch": 0.3855271366188397, "grad_norm": 0.09314591437578201, "learning_rate": 1.9408038598845672e-05, "loss": 0.4166, "step": 309 }, { "epoch": 0.3867747972551466, "grad_norm": 0.09022191166877747, "learning_rate": 1.9403504726286367e-05, "loss": 0.3708, "step": 310 }, { "epoch": 0.3880224578914535, "grad_norm": 0.07517941296100616, "learning_rate": 1.939895409096526e-05, "loss": 0.4089, "step": 311 }, { "epoch": 0.38927011852776044, "grad_norm": 0.08791400492191315, "learning_rate": 1.9394386700994357e-05, "loss": 0.3952, "step": 312 }, { "epoch": 0.3905177791640674, "grad_norm": 0.08254127204418182, "learning_rate": 1.9389802564515513e-05, "loss": 0.4023, "step": 313 }, { "epoch": 0.3917654398003743, "grad_norm": 0.09163601696491241, "learning_rate": 1.938520168970045e-05, "loss": 0.4002, "step": 314 }, { "epoch": 0.3930131004366812, "grad_norm": 0.10828031599521637, "learning_rate": 1.938058408475073e-05, "loss": 0.3885, "step": 315 }, { "epoch": 0.39426076107298813, "grad_norm": 0.10482138395309448, "learning_rate": 1.9375949757897725e-05, "loss": 0.3525, "step": 316 }, { "epoch": 0.39550842170929507, "grad_norm": 0.07083115726709366, "learning_rate": 1.9371298717402625e-05, "loss": 0.3372, "step": 317 }, { "epoch": 0.396756082345602, "grad_norm": 0.08137016743421555, "learning_rate": 1.9366630971556407e-05, "loss": 0.4064, "step": 318 }, { "epoch": 0.39800374298190894, "grad_norm": 0.07335782796144485, "learning_rate": 1.9361946528679837e-05, "loss": 0.3163, "step": 319 }, { "epoch": 0.39925140361821587, "grad_norm": 0.08302006125450134, "learning_rate": 1.9357245397123443e-05, "loss": 0.3502, "step": 320 }, { "epoch": 0.40049906425452275, "grad_norm": 0.07668264955282211, "learning_rate": 1.9352527585267493e-05, "loss": 0.3226, "step": 321 }, { "epoch": 0.4017467248908297, "grad_norm": 0.10439937561750412, "learning_rate": 1.9347793101521993e-05, "loss": 0.3212, "step": 322 }, { "epoch": 0.4029943855271366, "grad_norm": 0.10303515940904617, "learning_rate": 1.934304195432668e-05, "loss": 0.4068, "step": 323 }, { "epoch": 0.40424204616344356, "grad_norm": 0.09272205829620361, "learning_rate": 1.9338274152150983e-05, "loss": 0.3365, "step": 324 }, { "epoch": 0.4054897067997505, "grad_norm": 0.06783345341682434, "learning_rate": 1.933348970349402e-05, "loss": 0.3289, "step": 325 }, { "epoch": 0.4067373674360574, "grad_norm": 0.07962651550769806, "learning_rate": 1.932868861688459e-05, "loss": 0.384, "step": 326 }, { "epoch": 0.4079850280723643, "grad_norm": 0.08090052008628845, "learning_rate": 1.932387090088115e-05, "loss": 0.3748, "step": 327 }, { "epoch": 0.40923268870867124, "grad_norm": 0.06662657111883163, "learning_rate": 1.9319036564071797e-05, "loss": 0.3277, "step": 328 }, { "epoch": 0.4104803493449782, "grad_norm": 0.08814183622598648, "learning_rate": 1.9314185615074253e-05, "loss": 0.4001, "step": 329 }, { "epoch": 0.4117280099812851, "grad_norm": 0.06779827177524567, "learning_rate": 1.9309318062535864e-05, "loss": 0.2929, "step": 330 }, { "epoch": 0.412975670617592, "grad_norm": 0.08183370530605316, "learning_rate": 1.9304433915133567e-05, "loss": 0.3979, "step": 331 }, { "epoch": 0.41422333125389893, "grad_norm": 0.0669468492269516, "learning_rate": 1.9299533181573883e-05, "loss": 0.3279, "step": 332 }, { "epoch": 0.41547099189020587, "grad_norm": 0.08377107977867126, "learning_rate": 1.9294615870592892e-05, "loss": 0.3775, "step": 333 }, { "epoch": 0.4167186525265128, "grad_norm": 0.08826244622468948, "learning_rate": 1.928968199095624e-05, "loss": 0.4081, "step": 334 }, { "epoch": 0.41796631316281974, "grad_norm": 0.07477085292339325, "learning_rate": 1.928473155145909e-05, "loss": 0.3794, "step": 335 }, { "epoch": 0.4192139737991266, "grad_norm": 0.08051132410764694, "learning_rate": 1.9279764560926142e-05, "loss": 0.4369, "step": 336 }, { "epoch": 0.42046163443543355, "grad_norm": 0.08119165897369385, "learning_rate": 1.927478102821159e-05, "loss": 0.4217, "step": 337 }, { "epoch": 0.4217092950717405, "grad_norm": 0.07305929064750671, "learning_rate": 1.926978096219912e-05, "loss": 0.3603, "step": 338 }, { "epoch": 0.4229569557080474, "grad_norm": 0.08188523352146149, "learning_rate": 1.926476437180189e-05, "loss": 0.3972, "step": 339 }, { "epoch": 0.42420461634435436, "grad_norm": 0.08422026038169861, "learning_rate": 1.925973126596251e-05, "loss": 0.3614, "step": 340 }, { "epoch": 0.42545227698066124, "grad_norm": 0.07257169485092163, "learning_rate": 1.9254681653653037e-05, "loss": 0.3365, "step": 341 }, { "epoch": 0.4266999376169682, "grad_norm": 0.0878138542175293, "learning_rate": 1.9249615543874947e-05, "loss": 0.4388, "step": 342 }, { "epoch": 0.4279475982532751, "grad_norm": 0.08215674012899399, "learning_rate": 1.9244532945659132e-05, "loss": 0.4097, "step": 343 }, { "epoch": 0.42919525888958204, "grad_norm": 0.07723066210746765, "learning_rate": 1.9239433868065867e-05, "loss": 0.3869, "step": 344 }, { "epoch": 0.430442919525889, "grad_norm": 0.10846715420484543, "learning_rate": 1.923431832018481e-05, "loss": 0.3279, "step": 345 }, { "epoch": 0.43169058016219586, "grad_norm": 0.08006098121404648, "learning_rate": 1.922918631113498e-05, "loss": 0.3437, "step": 346 }, { "epoch": 0.4329382407985028, "grad_norm": 0.07850154489278793, "learning_rate": 1.922403785006473e-05, "loss": 0.32, "step": 347 }, { "epoch": 0.43418590143480973, "grad_norm": 0.07446196675300598, "learning_rate": 1.9218872946151753e-05, "loss": 0.3478, "step": 348 }, { "epoch": 0.43543356207111666, "grad_norm": 0.07685361057519913, "learning_rate": 1.9213691608603046e-05, "loss": 0.397, "step": 349 }, { "epoch": 0.4366812227074236, "grad_norm": 0.08373807370662689, "learning_rate": 1.9208493846654905e-05, "loss": 0.3756, "step": 350 }, { "epoch": 0.4379288833437305, "grad_norm": 0.08705881237983704, "learning_rate": 1.9203279669572893e-05, "loss": 0.3636, "step": 351 }, { "epoch": 0.4391765439800374, "grad_norm": 0.07912624627351761, "learning_rate": 1.9198049086651852e-05, "loss": 0.3214, "step": 352 }, { "epoch": 0.44042420461634435, "grad_norm": 0.07962213456630707, "learning_rate": 1.9192802107215857e-05, "loss": 0.3488, "step": 353 }, { "epoch": 0.4416718652526513, "grad_norm": 0.06611789017915726, "learning_rate": 1.9187538740618215e-05, "loss": 0.3276, "step": 354 }, { "epoch": 0.4429195258889582, "grad_norm": 0.0737161785364151, "learning_rate": 1.9182258996241446e-05, "loss": 0.344, "step": 355 }, { "epoch": 0.4441671865252651, "grad_norm": 0.0750458687543869, "learning_rate": 1.917696288349726e-05, "loss": 0.3952, "step": 356 }, { "epoch": 0.44541484716157204, "grad_norm": 0.09110225737094879, "learning_rate": 1.9171650411826556e-05, "loss": 0.3767, "step": 357 }, { "epoch": 0.44666250779787897, "grad_norm": 0.06446138769388199, "learning_rate": 1.9166321590699385e-05, "loss": 0.3164, "step": 358 }, { "epoch": 0.4479101684341859, "grad_norm": 0.07691174745559692, "learning_rate": 1.916097642961494e-05, "loss": 0.4026, "step": 359 }, { "epoch": 0.44915782907049284, "grad_norm": 0.07001764327287674, "learning_rate": 1.9155614938101557e-05, "loss": 0.3439, "step": 360 }, { "epoch": 0.4504054897067998, "grad_norm": 0.09792986512184143, "learning_rate": 1.915023712571667e-05, "loss": 0.3888, "step": 361 }, { "epoch": 0.45165315034310666, "grad_norm": 0.07331310957670212, "learning_rate": 1.9144843002046805e-05, "loss": 0.3222, "step": 362 }, { "epoch": 0.4529008109794136, "grad_norm": 0.06881300359964371, "learning_rate": 1.9139432576707575e-05, "loss": 0.324, "step": 363 }, { "epoch": 0.45414847161572053, "grad_norm": 0.07379250973463058, "learning_rate": 1.9134005859343647e-05, "loss": 0.3464, "step": 364 }, { "epoch": 0.45539613225202746, "grad_norm": 0.09389679878950119, "learning_rate": 1.9128562859628725e-05, "loss": 0.3314, "step": 365 }, { "epoch": 0.4566437928883344, "grad_norm": 0.10510922968387604, "learning_rate": 1.9123103587265545e-05, "loss": 0.3702, "step": 366 }, { "epoch": 0.4578914535246413, "grad_norm": 0.09789140522480011, "learning_rate": 1.911762805198585e-05, "loss": 0.3061, "step": 367 }, { "epoch": 0.4591391141609482, "grad_norm": 0.07343242317438126, "learning_rate": 1.9112136263550376e-05, "loss": 0.3674, "step": 368 }, { "epoch": 0.46038677479725515, "grad_norm": 0.07687333226203918, "learning_rate": 1.9106628231748825e-05, "loss": 0.3227, "step": 369 }, { "epoch": 0.4616344354335621, "grad_norm": 0.09030032902956009, "learning_rate": 1.9101103966399856e-05, "loss": 0.3499, "step": 370 }, { "epoch": 0.462882096069869, "grad_norm": 0.09658922255039215, "learning_rate": 1.9095563477351075e-05, "loss": 0.3838, "step": 371 }, { "epoch": 0.4641297567061759, "grad_norm": 0.42790326476097107, "learning_rate": 1.909000677447899e-05, "loss": 0.3827, "step": 372 }, { "epoch": 0.46537741734248284, "grad_norm": 0.07787511497735977, "learning_rate": 1.9084433867689035e-05, "loss": 0.3981, "step": 373 }, { "epoch": 0.46662507797878977, "grad_norm": 0.07825618982315063, "learning_rate": 1.9078844766915514e-05, "loss": 0.3584, "step": 374 }, { "epoch": 0.4678727386150967, "grad_norm": 0.09160970151424408, "learning_rate": 1.9073239482121598e-05, "loss": 0.3451, "step": 375 }, { "epoch": 0.46912039925140364, "grad_norm": 0.06959247589111328, "learning_rate": 1.9067618023299318e-05, "loss": 0.3168, "step": 376 }, { "epoch": 0.4703680598877105, "grad_norm": 0.09179724752902985, "learning_rate": 1.9061980400469524e-05, "loss": 0.3784, "step": 377 }, { "epoch": 0.47161572052401746, "grad_norm": 0.10104795545339584, "learning_rate": 1.9056326623681896e-05, "loss": 0.3234, "step": 378 }, { "epoch": 0.4728633811603244, "grad_norm": 0.07413583993911743, "learning_rate": 1.9050656703014893e-05, "loss": 0.3372, "step": 379 }, { "epoch": 0.4741110417966313, "grad_norm": 0.07793621718883514, "learning_rate": 1.9044970648575763e-05, "loss": 0.3489, "step": 380 }, { "epoch": 0.47535870243293826, "grad_norm": 0.08598163723945618, "learning_rate": 1.9039268470500515e-05, "loss": 0.4078, "step": 381 }, { "epoch": 0.47660636306924514, "grad_norm": 0.0972554013133049, "learning_rate": 1.903355017895389e-05, "loss": 0.3983, "step": 382 }, { "epoch": 0.4778540237055521, "grad_norm": 0.0939214900135994, "learning_rate": 1.9027815784129365e-05, "loss": 0.3407, "step": 383 }, { "epoch": 0.479101684341859, "grad_norm": 0.16490474343299866, "learning_rate": 1.9022065296249116e-05, "loss": 0.3577, "step": 384 }, { "epoch": 0.48034934497816595, "grad_norm": 0.07326449453830719, "learning_rate": 1.9016298725564012e-05, "loss": 0.3071, "step": 385 }, { "epoch": 0.4815970056144729, "grad_norm": 0.07553815096616745, "learning_rate": 1.901051608235358e-05, "loss": 0.3218, "step": 386 }, { "epoch": 0.48284466625077976, "grad_norm": 0.08704483509063721, "learning_rate": 1.9004717376926007e-05, "loss": 0.3309, "step": 387 }, { "epoch": 0.4840923268870867, "grad_norm": 0.07856780290603638, "learning_rate": 1.8998902619618117e-05, "loss": 0.3173, "step": 388 }, { "epoch": 0.48533998752339363, "grad_norm": 0.19682346284389496, "learning_rate": 1.8993071820795336e-05, "loss": 0.3777, "step": 389 }, { "epoch": 0.48658764815970057, "grad_norm": 0.21116366982460022, "learning_rate": 1.8987224990851697e-05, "loss": 0.3842, "step": 390 }, { "epoch": 0.4878353087960075, "grad_norm": 0.09021289646625519, "learning_rate": 1.8981362140209798e-05, "loss": 0.3869, "step": 391 }, { "epoch": 0.4890829694323144, "grad_norm": 0.10774579644203186, "learning_rate": 1.897548327932081e-05, "loss": 0.3502, "step": 392 }, { "epoch": 0.4903306300686213, "grad_norm": 0.09023426473140717, "learning_rate": 1.8969588418664434e-05, "loss": 0.4253, "step": 393 }, { "epoch": 0.49157829070492826, "grad_norm": 0.07410671561956406, "learning_rate": 1.89636775687489e-05, "loss": 0.3141, "step": 394 }, { "epoch": 0.4928259513412352, "grad_norm": 0.07652512937784195, "learning_rate": 1.8957750740110924e-05, "loss": 0.3277, "step": 395 }, { "epoch": 0.4940736119775421, "grad_norm": 0.0774303525686264, "learning_rate": 1.895180794331573e-05, "loss": 0.3339, "step": 396 }, { "epoch": 0.495321272613849, "grad_norm": 0.08927808701992035, "learning_rate": 1.8945849188956986e-05, "loss": 0.362, "step": 397 }, { "epoch": 0.49656893325015594, "grad_norm": 0.08084219694137573, "learning_rate": 1.893987448765682e-05, "loss": 0.3304, "step": 398 }, { "epoch": 0.4978165938864629, "grad_norm": 0.08156754076480865, "learning_rate": 1.8933883850065778e-05, "loss": 0.3759, "step": 399 }, { "epoch": 0.4990642545227698, "grad_norm": 0.14657171070575714, "learning_rate": 1.892787728686282e-05, "loss": 0.3648, "step": 400 }, { "epoch": 0.5003119151590767, "grad_norm": 0.07986622303724289, "learning_rate": 1.8921854808755295e-05, "loss": 0.3212, "step": 401 }, { "epoch": 0.5015595757953837, "grad_norm": 0.08847914636135101, "learning_rate": 1.8915816426478914e-05, "loss": 0.4114, "step": 402 }, { "epoch": 0.5028072364316906, "grad_norm": 0.08348351716995239, "learning_rate": 1.8909762150797752e-05, "loss": 0.4262, "step": 403 }, { "epoch": 0.5040548970679976, "grad_norm": 0.07425021380186081, "learning_rate": 1.8903691992504204e-05, "loss": 0.3396, "step": 404 }, { "epoch": 0.5053025577043044, "grad_norm": 0.09293273836374283, "learning_rate": 1.889760596241898e-05, "loss": 0.4121, "step": 405 }, { "epoch": 0.5065502183406113, "grad_norm": 0.08483749628067017, "learning_rate": 1.8891504071391092e-05, "loss": 0.3955, "step": 406 }, { "epoch": 0.5077978789769183, "grad_norm": 0.09369031339883804, "learning_rate": 1.8885386330297817e-05, "loss": 0.3644, "step": 407 }, { "epoch": 0.5090455396132252, "grad_norm": 0.06889427453279495, "learning_rate": 1.8879252750044686e-05, "loss": 0.2993, "step": 408 }, { "epoch": 0.5102932002495322, "grad_norm": 0.07942546904087067, "learning_rate": 1.887310334156547e-05, "loss": 0.3975, "step": 409 }, { "epoch": 0.511540860885839, "grad_norm": 0.0964096263051033, "learning_rate": 1.8866938115822148e-05, "loss": 0.4178, "step": 410 }, { "epoch": 0.5127885215221459, "grad_norm": 0.19631971418857574, "learning_rate": 1.8860757083804908e-05, "loss": 0.2797, "step": 411 }, { "epoch": 0.5140361821584529, "grad_norm": 0.08998764306306839, "learning_rate": 1.8854560256532098e-05, "loss": 0.3698, "step": 412 }, { "epoch": 0.5152838427947598, "grad_norm": 0.07443209737539291, "learning_rate": 1.8848347645050242e-05, "loss": 0.2805, "step": 413 }, { "epoch": 0.5165315034310668, "grad_norm": 0.11266658455133438, "learning_rate": 1.8842119260433984e-05, "loss": 0.4253, "step": 414 }, { "epoch": 0.5177791640673737, "grad_norm": 0.07977344840765, "learning_rate": 1.8835875113786088e-05, "loss": 0.4107, "step": 415 }, { "epoch": 0.5190268247036806, "grad_norm": 0.07444226741790771, "learning_rate": 1.8829615216237426e-05, "loss": 0.2756, "step": 416 }, { "epoch": 0.5202744853399875, "grad_norm": 0.08746916055679321, "learning_rate": 1.8823339578946935e-05, "loss": 0.3661, "step": 417 }, { "epoch": 0.5215221459762944, "grad_norm": 0.07776310294866562, "learning_rate": 1.881704821310162e-05, "loss": 0.356, "step": 418 }, { "epoch": 0.5227698066126014, "grad_norm": 0.07319823652505875, "learning_rate": 1.8810741129916516e-05, "loss": 0.2972, "step": 419 }, { "epoch": 0.5240174672489083, "grad_norm": 0.1166319027543068, "learning_rate": 1.880441834063468e-05, "loss": 0.3385, "step": 420 }, { "epoch": 0.5252651278852152, "grad_norm": 0.07309556007385254, "learning_rate": 1.8798079856527175e-05, "loss": 0.3663, "step": 421 }, { "epoch": 0.5265127885215222, "grad_norm": 0.08942176401615143, "learning_rate": 1.8791725688893018e-05, "loss": 0.366, "step": 422 }, { "epoch": 0.527760449157829, "grad_norm": 0.09301096200942993, "learning_rate": 1.8785355849059204e-05, "loss": 0.3379, "step": 423 }, { "epoch": 0.529008109794136, "grad_norm": 0.11961039900779724, "learning_rate": 1.877897034838067e-05, "loss": 0.3297, "step": 424 }, { "epoch": 0.5302557704304429, "grad_norm": 0.0748453140258789, "learning_rate": 1.8772569198240244e-05, "loss": 0.3437, "step": 425 }, { "epoch": 0.5315034310667498, "grad_norm": 0.120552197098732, "learning_rate": 1.8766152410048676e-05, "loss": 0.3903, "step": 426 }, { "epoch": 0.5327510917030568, "grad_norm": 0.09532199800014496, "learning_rate": 1.875971999524458e-05, "loss": 0.2942, "step": 427 }, { "epoch": 0.5339987523393637, "grad_norm": 0.07045169919729233, "learning_rate": 1.8753271965294437e-05, "loss": 0.2739, "step": 428 }, { "epoch": 0.5352464129756707, "grad_norm": 0.0874415785074234, "learning_rate": 1.8746808331692542e-05, "loss": 0.4072, "step": 429 }, { "epoch": 0.5364940736119775, "grad_norm": 0.09460163861513138, "learning_rate": 1.874032910596103e-05, "loss": 0.3936, "step": 430 }, { "epoch": 0.5377417342482844, "grad_norm": 0.1052878126502037, "learning_rate": 1.8733834299649818e-05, "loss": 0.3533, "step": 431 }, { "epoch": 0.5389893948845914, "grad_norm": 0.09336161613464355, "learning_rate": 1.87273239243366e-05, "loss": 0.3473, "step": 432 }, { "epoch": 0.5402370555208983, "grad_norm": 0.08659809827804565, "learning_rate": 1.872079799162682e-05, "loss": 0.4026, "step": 433 }, { "epoch": 0.5414847161572053, "grad_norm": 0.0875949040055275, "learning_rate": 1.871425651315365e-05, "loss": 0.4091, "step": 434 }, { "epoch": 0.5427323767935122, "grad_norm": 0.08493121713399887, "learning_rate": 1.8707699500577997e-05, "loss": 0.3439, "step": 435 }, { "epoch": 0.543980037429819, "grad_norm": 0.12357327342033386, "learning_rate": 1.8701126965588433e-05, "loss": 0.4433, "step": 436 }, { "epoch": 0.545227698066126, "grad_norm": 0.0841773971915245, "learning_rate": 1.8694538919901216e-05, "loss": 0.3916, "step": 437 }, { "epoch": 0.5464753587024329, "grad_norm": 0.08911257982254028, "learning_rate": 1.8687935375260244e-05, "loss": 0.3349, "step": 438 }, { "epoch": 0.5477230193387399, "grad_norm": 0.1051001101732254, "learning_rate": 1.8681316343437048e-05, "loss": 0.3307, "step": 439 }, { "epoch": 0.5489706799750468, "grad_norm": 0.08096891641616821, "learning_rate": 1.867468183623077e-05, "loss": 0.3684, "step": 440 }, { "epoch": 0.5502183406113537, "grad_norm": 0.06814990937709808, "learning_rate": 1.866803186546813e-05, "loss": 0.2707, "step": 441 }, { "epoch": 0.5514660012476607, "grad_norm": 0.08140693604946136, "learning_rate": 1.8661366443003426e-05, "loss": 0.3693, "step": 442 }, { "epoch": 0.5527136618839675, "grad_norm": 0.08824186027050018, "learning_rate": 1.8654685580718482e-05, "loss": 0.3401, "step": 443 }, { "epoch": 0.5539613225202745, "grad_norm": 0.07940424978733063, "learning_rate": 1.8647989290522667e-05, "loss": 0.3297, "step": 444 }, { "epoch": 0.5552089831565814, "grad_norm": 0.08890607953071594, "learning_rate": 1.8641277584352832e-05, "loss": 0.3961, "step": 445 }, { "epoch": 0.5564566437928883, "grad_norm": 0.09835802763700485, "learning_rate": 1.8634550474173322e-05, "loss": 0.3718, "step": 446 }, { "epoch": 0.5577043044291953, "grad_norm": 0.07689094543457031, "learning_rate": 1.862780797197594e-05, "loss": 0.3069, "step": 447 }, { "epoch": 0.5589519650655022, "grad_norm": 0.07634837925434113, "learning_rate": 1.8621050089779912e-05, "loss": 0.3772, "step": 448 }, { "epoch": 0.5601996257018091, "grad_norm": 0.08531224727630615, "learning_rate": 1.8614276839631902e-05, "loss": 0.3344, "step": 449 }, { "epoch": 0.561447286338116, "grad_norm": 0.08031272888183594, "learning_rate": 1.860748823360596e-05, "loss": 0.419, "step": 450 }, { "epoch": 0.5626949469744229, "grad_norm": 0.09414199739694595, "learning_rate": 1.86006842838035e-05, "loss": 0.4198, "step": 451 }, { "epoch": 0.5639426076107299, "grad_norm": 0.08871858566999435, "learning_rate": 1.8593865002353307e-05, "loss": 0.348, "step": 452 }, { "epoch": 0.5651902682470368, "grad_norm": 0.07083774358034134, "learning_rate": 1.858703040141148e-05, "loss": 0.3334, "step": 453 }, { "epoch": 0.5664379288833438, "grad_norm": 0.10692565143108368, "learning_rate": 1.8580180493161434e-05, "loss": 0.4008, "step": 454 }, { "epoch": 0.5676855895196506, "grad_norm": 0.07992343604564667, "learning_rate": 1.8573315289813866e-05, "loss": 0.3971, "step": 455 }, { "epoch": 0.5689332501559575, "grad_norm": 0.08050104975700378, "learning_rate": 1.856643480360675e-05, "loss": 0.3488, "step": 456 }, { "epoch": 0.5701809107922645, "grad_norm": 0.07934119552373886, "learning_rate": 1.855953904680529e-05, "loss": 0.2948, "step": 457 }, { "epoch": 0.5714285714285714, "grad_norm": 0.09823425859212875, "learning_rate": 1.855262803170191e-05, "loss": 0.3186, "step": 458 }, { "epoch": 0.5726762320648784, "grad_norm": 0.23104049265384674, "learning_rate": 1.8545701770616254e-05, "loss": 0.3841, "step": 459 }, { "epoch": 0.5739238927011853, "grad_norm": 0.07883207499980927, "learning_rate": 1.8538760275895118e-05, "loss": 0.4124, "step": 460 }, { "epoch": 0.5751715533374921, "grad_norm": 0.07763822376728058, "learning_rate": 1.853180355991247e-05, "loss": 0.2878, "step": 461 }, { "epoch": 0.5764192139737991, "grad_norm": 0.09326838701963425, "learning_rate": 1.852483163506941e-05, "loss": 0.322, "step": 462 }, { "epoch": 0.577666874610106, "grad_norm": 0.08444075286388397, "learning_rate": 1.8517844513794144e-05, "loss": 0.2674, "step": 463 }, { "epoch": 0.578914535246413, "grad_norm": 0.07540516555309296, "learning_rate": 1.8510842208541962e-05, "loss": 0.284, "step": 464 }, { "epoch": 0.5801621958827199, "grad_norm": 0.08852772414684296, "learning_rate": 1.850382473179524e-05, "loss": 0.4359, "step": 465 }, { "epoch": 0.5814098565190269, "grad_norm": 0.07915528863668442, "learning_rate": 1.849679209606338e-05, "loss": 0.3875, "step": 466 }, { "epoch": 0.5826575171553338, "grad_norm": 0.08570799231529236, "learning_rate": 1.8489744313882818e-05, "loss": 0.3781, "step": 467 }, { "epoch": 0.5839051777916406, "grad_norm": 0.0977482944726944, "learning_rate": 1.8482681397816986e-05, "loss": 0.3402, "step": 468 }, { "epoch": 0.5851528384279476, "grad_norm": 0.07194789499044418, "learning_rate": 1.847560336045629e-05, "loss": 0.3404, "step": 469 }, { "epoch": 0.5864004990642545, "grad_norm": 0.08206473290920258, "learning_rate": 1.84685102144181e-05, "loss": 0.2855, "step": 470 }, { "epoch": 0.5876481597005615, "grad_norm": 0.1329026073217392, "learning_rate": 1.8461401972346713e-05, "loss": 0.4132, "step": 471 }, { "epoch": 0.5888958203368684, "grad_norm": 0.11516719311475754, "learning_rate": 1.8454278646913336e-05, "loss": 0.3963, "step": 472 }, { "epoch": 0.5901434809731753, "grad_norm": 0.08752531558275223, "learning_rate": 1.8447140250816065e-05, "loss": 0.3965, "step": 473 }, { "epoch": 0.5913911416094823, "grad_norm": 0.06656806170940399, "learning_rate": 1.8439986796779866e-05, "loss": 0.2943, "step": 474 }, { "epoch": 0.5926388022457891, "grad_norm": 0.07554040104150772, "learning_rate": 1.8432818297556535e-05, "loss": 0.3709, "step": 475 }, { "epoch": 0.5938864628820961, "grad_norm": 0.07638383656740189, "learning_rate": 1.84256347659247e-05, "loss": 0.3587, "step": 476 }, { "epoch": 0.595134123518403, "grad_norm": 0.09883903712034225, "learning_rate": 1.841843621468978e-05, "loss": 0.2965, "step": 477 }, { "epoch": 0.5963817841547099, "grad_norm": 0.09742535650730133, "learning_rate": 1.8411222656683967e-05, "loss": 0.3927, "step": 478 }, { "epoch": 0.5976294447910169, "grad_norm": 0.08859851956367493, "learning_rate": 1.8403994104766214e-05, "loss": 0.3119, "step": 479 }, { "epoch": 0.5988771054273238, "grad_norm": 0.08222465217113495, "learning_rate": 1.8396750571822188e-05, "loss": 0.3199, "step": 480 }, { "epoch": 0.6001247660636307, "grad_norm": 0.08288297802209854, "learning_rate": 1.838949207076427e-05, "loss": 0.3581, "step": 481 }, { "epoch": 0.6013724266999376, "grad_norm": 0.09967238456010818, "learning_rate": 1.8382218614531518e-05, "loss": 0.4083, "step": 482 }, { "epoch": 0.6026200873362445, "grad_norm": 0.08456137031316757, "learning_rate": 1.8374930216089663e-05, "loss": 0.3454, "step": 483 }, { "epoch": 0.6038677479725515, "grad_norm": 0.061698053032159805, "learning_rate": 1.836762688843105e-05, "loss": 0.2286, "step": 484 }, { "epoch": 0.6051154086088584, "grad_norm": 0.09728894382715225, "learning_rate": 1.8360308644574654e-05, "loss": 0.362, "step": 485 }, { "epoch": 0.6063630692451654, "grad_norm": 0.07975105941295624, "learning_rate": 1.8352975497566032e-05, "loss": 0.3507, "step": 486 }, { "epoch": 0.6076107298814722, "grad_norm": 0.41635558009147644, "learning_rate": 1.8345627460477312e-05, "loss": 0.3044, "step": 487 }, { "epoch": 0.6088583905177791, "grad_norm": 0.07460454106330872, "learning_rate": 1.833826454640716e-05, "loss": 0.2928, "step": 488 }, { "epoch": 0.6101060511540861, "grad_norm": 0.08211305737495422, "learning_rate": 1.833088676848077e-05, "loss": 0.3005, "step": 489 }, { "epoch": 0.611353711790393, "grad_norm": 0.22251826524734497, "learning_rate": 1.8323494139849824e-05, "loss": 0.361, "step": 490 }, { "epoch": 0.6126013724267, "grad_norm": 0.07368269562721252, "learning_rate": 1.8316086673692477e-05, "loss": 0.2952, "step": 491 }, { "epoch": 0.6138490330630069, "grad_norm": 0.08986833691596985, "learning_rate": 1.8308664383213343e-05, "loss": 0.3222, "step": 492 }, { "epoch": 0.6150966936993137, "grad_norm": 0.07270447164773941, "learning_rate": 1.8301227281643453e-05, "loss": 0.3357, "step": 493 }, { "epoch": 0.6163443543356207, "grad_norm": 0.1448320597410202, "learning_rate": 1.829377538224024e-05, "loss": 0.4107, "step": 494 }, { "epoch": 0.6175920149719276, "grad_norm": 0.08396106213331223, "learning_rate": 1.8286308698287524e-05, "loss": 0.3513, "step": 495 }, { "epoch": 0.6188396756082346, "grad_norm": 0.08713540434837341, "learning_rate": 1.827882724309547e-05, "loss": 0.3977, "step": 496 }, { "epoch": 0.6188396756082346, "eval_loss": 0.34908154606819153, "eval_runtime": 2820.4987, "eval_samples_per_second": 0.253, "eval_steps_per_second": 0.127, "step": 496 } ], "logging_steps": 1, "max_steps": 2403, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 248, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.2079825954766143e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }