{ "best_metric": null, "best_model_checkpoint": null, "epoch": 21.0, "global_step": 80997, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 5e-09, "loss": 10.5291, "step": 1 }, { "epoch": 0.13, "learning_rate": 2.5e-06, "loss": 9.5195, "step": 500 }, { "epoch": 0.26, "learning_rate": 5e-06, "loss": 7.8918, "step": 1000 }, { "epoch": 0.39, "learning_rate": 7.5e-06, "loss": 6.8556, "step": 1500 }, { "epoch": 0.52, "learning_rate": 1e-05, "loss": 6.5569, "step": 2000 }, { "epoch": 0.65, "learning_rate": 1.25e-05, "loss": 6.3942, "step": 2500 }, { "epoch": 0.78, "learning_rate": 1.5e-05, "loss": 6.286, "step": 3000 }, { "epoch": 0.91, "learning_rate": 1.75e-05, "loss": 6.1963, "step": 3500 }, { "epoch": 1.04, "learning_rate": 2e-05, "loss": 6.1271, "step": 4000 }, { "epoch": 1.17, "learning_rate": 2.25e-05, "loss": 6.0662, "step": 4500 }, { "epoch": 1.3, "learning_rate": 2.5e-05, "loss": 6.0132, "step": 5000 }, { "epoch": 1.43, "learning_rate": 2.7500000000000004e-05, "loss": 5.9636, "step": 5500 }, { "epoch": 1.56, "learning_rate": 3e-05, "loss": 5.9214, "step": 6000 }, { "epoch": 1.69, "learning_rate": 3.2500000000000004e-05, "loss": 5.893, "step": 6500 }, { "epoch": 1.81, "learning_rate": 3.5e-05, "loss": 5.8604, "step": 7000 }, { "epoch": 1.94, "learning_rate": 3.7500000000000003e-05, "loss": 5.8285, "step": 7500 }, { "epoch": 2.07, "learning_rate": 4e-05, "loss": 5.8056, "step": 8000 }, { "epoch": 2.2, "learning_rate": 4.2495e-05, "loss": 5.7834, "step": 8500 }, { "epoch": 2.33, "learning_rate": 4.4995000000000005e-05, "loss": 5.7641, "step": 9000 }, { "epoch": 2.46, "learning_rate": 4.7495e-05, "loss": 5.7448, "step": 9500 }, { "epoch": 2.59, "learning_rate": 4.9995000000000005e-05, "loss": 5.7285, "step": 10000 }, { "epoch": 2.72, "learning_rate": 4.998375521920668e-05, "loss": 5.7139, "step": 10500 }, { "epoch": 2.85, "learning_rate": 4.996744519832985e-05, "loss": 5.6956, "step": 11000 }, { "epoch": 2.98, "learning_rate": 4.995113517745303e-05, "loss": 5.6856, "step": 11500 }, { "epoch": 3.11, "learning_rate": 4.9934825156576205e-05, "loss": 5.6724, "step": 12000 }, { "epoch": 3.24, "learning_rate": 4.991854775574113e-05, "loss": 5.6621, "step": 12500 }, { "epoch": 3.37, "learning_rate": 4.9902237734864304e-05, "loss": 5.6517, "step": 13000 }, { "epoch": 3.5, "learning_rate": 4.988592771398747e-05, "loss": 5.643, "step": 13500 }, { "epoch": 3.63, "learning_rate": 4.986961769311065e-05, "loss": 5.6321, "step": 14000 }, { "epoch": 3.76, "learning_rate": 4.9853307672233826e-05, "loss": 5.6268, "step": 14500 }, { "epoch": 3.89, "learning_rate": 4.983703027139875e-05, "loss": 5.6193, "step": 15000 }, { "epoch": 4.02, "learning_rate": 4.9820752870563676e-05, "loss": 5.6084, "step": 15500 }, { "epoch": 4.15, "learning_rate": 4.980444284968685e-05, "loss": 5.6034, "step": 16000 }, { "epoch": 4.28, "learning_rate": 4.978813282881002e-05, "loss": 5.597, "step": 16500 }, { "epoch": 4.41, "learning_rate": 4.977182280793319e-05, "loss": 5.5927, "step": 17000 }, { "epoch": 4.54, "learning_rate": 4.975551278705637e-05, "loss": 5.5841, "step": 17500 }, { "epoch": 4.67, "learning_rate": 4.9739202766179545e-05, "loss": 5.5778, "step": 18000 }, { "epoch": 4.8, "learning_rate": 4.972289274530272e-05, "loss": 5.5748, "step": 18500 }, { "epoch": 4.93, "learning_rate": 4.970658272442589e-05, "loss": 5.5692, "step": 19000 }, { "epoch": 5.06, "learning_rate": 4.969030532359082e-05, "loss": 5.5608, "step": 19500 }, { "epoch": 5.19, "learning_rate": 4.967399530271399e-05, "loss": 5.5594, "step": 20000 }, { "epoch": 5.32, "learning_rate": 4.965771790187892e-05, "loss": 5.5563, "step": 20500 }, { "epoch": 5.44, "learning_rate": 4.964140788100209e-05, "loss": 5.5528, "step": 21000 }, { "epoch": 5.57, "learning_rate": 4.962509786012526e-05, "loss": 5.5461, "step": 21500 }, { "epoch": 5.7, "learning_rate": 4.960878783924844e-05, "loss": 5.5413, "step": 22000 }, { "epoch": 5.83, "learning_rate": 4.959251043841336e-05, "loss": 5.5347, "step": 22500 }, { "epoch": 5.96, "learning_rate": 4.957623303757829e-05, "loss": 5.535, "step": 23000 }, { "epoch": 6.09, "learning_rate": 4.9559923016701466e-05, "loss": 5.5296, "step": 23500 }, { "epoch": 6.22, "learning_rate": 4.954361299582464e-05, "loss": 5.5252, "step": 24000 }, { "epoch": 6.35, "learning_rate": 4.952730297494781e-05, "loss": 5.5228, "step": 24500 }, { "epoch": 6.48, "learning_rate": 4.951099295407098e-05, "loss": 5.5196, "step": 25000 }, { "epoch": 6.61, "learning_rate": 4.949468293319416e-05, "loss": 5.5208, "step": 25500 }, { "epoch": 6.74, "learning_rate": 4.947837291231733e-05, "loss": 5.5133, "step": 26000 }, { "epoch": 6.87, "learning_rate": 4.9462062891440504e-05, "loss": 5.5112, "step": 26500 }, { "epoch": 7.0, "learning_rate": 4.9445785490605426e-05, "loss": 5.5082, "step": 27000 }, { "epoch": 7.13, "learning_rate": 4.94294754697286e-05, "loss": 5.4993, "step": 27500 }, { "epoch": 7.26, "learning_rate": 4.941316544885178e-05, "loss": 5.5017, "step": 28000 }, { "epoch": 7.39, "learning_rate": 4.9396855427974955e-05, "loss": 5.4975, "step": 28500 }, { "epoch": 7.52, "learning_rate": 4.9380545407098125e-05, "loss": 5.4939, "step": 29000 }, { "epoch": 7.65, "learning_rate": 4.9364268006263046e-05, "loss": 5.4887, "step": 29500 }, { "epoch": 7.78, "learning_rate": 4.934795798538622e-05, "loss": 5.4916, "step": 30000 }, { "epoch": 7.91, "learning_rate": 4.93316479645094e-05, "loss": 5.4877, "step": 30500 }, { "epoch": 8.04, "learning_rate": 4.9315337943632576e-05, "loss": 5.4827, "step": 31000 }, { "epoch": 8.17, "learning_rate": 4.929902792275574e-05, "loss": 5.4813, "step": 31500 }, { "epoch": 8.3, "learning_rate": 4.9282750521920674e-05, "loss": 5.4788, "step": 32000 }, { "epoch": 8.43, "learning_rate": 4.926644050104384e-05, "loss": 5.4751, "step": 32500 }, { "epoch": 8.56, "learning_rate": 4.925013048016702e-05, "loss": 5.4706, "step": 33000 }, { "epoch": 8.69, "learning_rate": 4.923382045929019e-05, "loss": 5.4727, "step": 33500 }, { "epoch": 8.82, "learning_rate": 4.921754305845512e-05, "loss": 5.4667, "step": 34000 }, { "epoch": 8.94, "learning_rate": 4.9201233037578294e-05, "loss": 5.3861, "step": 34500 }, { "epoch": 9.07, "learning_rate": 4.9184923016701464e-05, "loss": 5.1493, "step": 35000 }, { "epoch": 9.2, "learning_rate": 4.9168612995824633e-05, "loss": 4.972, "step": 35500 }, { "epoch": 9.33, "learning_rate": 4.915230297494781e-05, "loss": 4.7871, "step": 36000 }, { "epoch": 9.46, "learning_rate": 4.9135992954070986e-05, "loss": 4.5924, "step": 36500 }, { "epoch": 9.59, "learning_rate": 4.9119715553235915e-05, "loss": 4.4234, "step": 37000 }, { "epoch": 9.72, "learning_rate": 4.9103405532359084e-05, "loss": 4.2503, "step": 37500 }, { "epoch": 9.85, "learning_rate": 4.9087095511482254e-05, "loss": 4.0878, "step": 38000 }, { "epoch": 9.98, "learning_rate": 4.907078549060543e-05, "loss": 3.8941, "step": 38500 }, { "epoch": 10.11, "learning_rate": 4.905447546972861e-05, "loss": 3.4534, "step": 39000 }, { "epoch": 10.24, "learning_rate": 4.9038165448851776e-05, "loss": 2.8029, "step": 39500 }, { "epoch": 10.37, "learning_rate": 4.902185542797495e-05, "loss": 2.3458, "step": 40000 }, { "epoch": 10.5, "learning_rate": 4.900554540709812e-05, "loss": 2.1352, "step": 40500 }, { "epoch": 10.63, "learning_rate": 4.89893006263048e-05, "loss": 2.0089, "step": 41000 }, { "epoch": 10.76, "learning_rate": 4.897299060542798e-05, "loss": 1.9194, "step": 41500 }, { "epoch": 10.89, "learning_rate": 4.895668058455115e-05, "loss": 1.8505, "step": 42000 }, { "epoch": 11.02, "learning_rate": 4.8940370563674325e-05, "loss": 1.7901, "step": 42500 }, { "epoch": 11.15, "learning_rate": 4.8924060542797495e-05, "loss": 1.7405, "step": 43000 }, { "epoch": 11.28, "learning_rate": 4.890775052192067e-05, "loss": 1.6962, "step": 43500 }, { "epoch": 11.41, "learning_rate": 4.889144050104385e-05, "loss": 1.6585, "step": 44000 }, { "epoch": 11.54, "learning_rate": 4.887516310020877e-05, "loss": 1.6184, "step": 44500 }, { "epoch": 11.67, "learning_rate": 4.8858853079331946e-05, "loss": 1.5853, "step": 45000 }, { "epoch": 11.8, "learning_rate": 4.8842543058455116e-05, "loss": 1.5588, "step": 45500 }, { "epoch": 11.93, "learning_rate": 4.882623303757829e-05, "loss": 1.5296, "step": 46000 }, { "epoch": 12.06, "learning_rate": 4.880992301670146e-05, "loss": 1.504, "step": 46500 }, { "epoch": 12.19, "learning_rate": 4.879364561586639e-05, "loss": 1.4818, "step": 47000 }, { "epoch": 12.32, "learning_rate": 4.8777335594989566e-05, "loss": 1.4608, "step": 47500 }, { "epoch": 12.44, "learning_rate": 4.876102557411274e-05, "loss": 1.4441, "step": 48000 }, { "epoch": 12.57, "learning_rate": 4.8744715553235906e-05, "loss": 1.4253, "step": 48500 }, { "epoch": 12.7, "learning_rate": 4.872840553235908e-05, "loss": 1.4099, "step": 49000 }, { "epoch": 12.83, "learning_rate": 4.871209551148226e-05, "loss": 1.3957, "step": 49500 }, { "epoch": 12.96, "learning_rate": 4.869578549060543e-05, "loss": 1.3802, "step": 50000 }, { "epoch": 13.09, "learning_rate": 4.8679475469728604e-05, "loss": 1.3644, "step": 50500 }, { "epoch": 13.22, "learning_rate": 4.8663198068893526e-05, "loss": 1.3536, "step": 51000 }, { "epoch": 13.35, "learning_rate": 4.86468880480167e-05, "loss": 1.3423, "step": 51500 }, { "epoch": 13.48, "learning_rate": 4.863061064718163e-05, "loss": 1.328, "step": 52000 }, { "epoch": 13.61, "learning_rate": 4.86143006263048e-05, "loss": 1.3191, "step": 52500 }, { "epoch": 13.74, "learning_rate": 4.859799060542798e-05, "loss": 1.3084, "step": 53000 }, { "epoch": 13.87, "learning_rate": 4.858168058455115e-05, "loss": 1.2963, "step": 53500 }, { "epoch": 14.0, "learning_rate": 4.856540318371608e-05, "loss": 1.2866, "step": 54000 }, { "epoch": 14.13, "learning_rate": 4.854909316283925e-05, "loss": 1.2748, "step": 54500 }, { "epoch": 14.26, "learning_rate": 4.853278314196242e-05, "loss": 1.2662, "step": 55000 }, { "epoch": 14.39, "learning_rate": 4.85164731210856e-05, "loss": 1.2603, "step": 55500 }, { "epoch": 14.52, "learning_rate": 4.8500163100208774e-05, "loss": 1.2479, "step": 56000 }, { "epoch": 14.65, "learning_rate": 4.8483885699373696e-05, "loss": 1.2429, "step": 56500 }, { "epoch": 14.78, "learning_rate": 4.8467575678496865e-05, "loss": 1.2346, "step": 57000 }, { "epoch": 14.91, "learning_rate": 4.845126565762004e-05, "loss": 1.227, "step": 57500 }, { "epoch": 15.04, "learning_rate": 4.843495563674322e-05, "loss": 1.2201, "step": 58000 }, { "epoch": 15.17, "learning_rate": 4.8418645615866394e-05, "loss": 1.2107, "step": 58500 }, { "epoch": 15.3, "learning_rate": 4.8402368215031316e-05, "loss": 1.2042, "step": 59000 }, { "epoch": 15.43, "learning_rate": 4.838605819415449e-05, "loss": 1.1966, "step": 59500 }, { "epoch": 15.56, "learning_rate": 4.836974817327766e-05, "loss": 1.1893, "step": 60000 }, { "epoch": 15.69, "learning_rate": 4.835343815240084e-05, "loss": 1.1849, "step": 60500 }, { "epoch": 15.82, "learning_rate": 4.833716075156576e-05, "loss": 1.1777, "step": 61000 }, { "epoch": 15.95, "learning_rate": 4.832085073068894e-05, "loss": 1.1731, "step": 61500 }, { "epoch": 16.07, "learning_rate": 4.830454070981211e-05, "loss": 1.1659, "step": 62000 }, { "epoch": 16.2, "learning_rate": 4.828823068893528e-05, "loss": 1.159, "step": 62500 }, { "epoch": 16.33, "learning_rate": 4.827192066805846e-05, "loss": 1.1537, "step": 63000 }, { "epoch": 16.46, "learning_rate": 4.825564326722338e-05, "loss": 1.1486, "step": 63500 }, { "epoch": 16.59, "learning_rate": 4.823936586638831e-05, "loss": 1.1415, "step": 64000 }, { "epoch": 16.72, "learning_rate": 4.8223055845511486e-05, "loss": 1.138, "step": 64500 }, { "epoch": 16.85, "learning_rate": 4.8206745824634655e-05, "loss": 1.1307, "step": 65000 }, { "epoch": 16.98, "learning_rate": 4.819043580375783e-05, "loss": 1.127, "step": 65500 }, { "epoch": 17.11, "learning_rate": 4.817415840292276e-05, "loss": 1.1194, "step": 66000 }, { "epoch": 17.24, "learning_rate": 4.815784838204594e-05, "loss": 1.1152, "step": 66500 }, { "epoch": 17.37, "learning_rate": 4.8141538361169106e-05, "loss": 1.1109, "step": 67000 }, { "epoch": 17.5, "learning_rate": 4.8125228340292276e-05, "loss": 1.1096, "step": 67500 }, { "epoch": 17.63, "learning_rate": 4.8108950939457204e-05, "loss": 1.1036, "step": 68000 }, { "epoch": 17.76, "learning_rate": 4.809264091858038e-05, "loss": 1.0995, "step": 68500 }, { "epoch": 17.89, "learning_rate": 4.807633089770355e-05, "loss": 1.0941, "step": 69000 }, { "epoch": 18.02, "learning_rate": 4.806002087682672e-05, "loss": 1.09, "step": 69500 }, { "epoch": 18.15, "learning_rate": 4.8043710855949896e-05, "loss": 1.0841, "step": 70000 }, { "epoch": 18.28, "learning_rate": 4.802740083507307e-05, "loss": 1.0802, "step": 70500 }, { "epoch": 18.41, "learning_rate": 4.801109081419625e-05, "loss": 1.0769, "step": 71000 }, { "epoch": 18.54, "learning_rate": 4.799481341336117e-05, "loss": 1.0729, "step": 71500 }, { "epoch": 18.67, "learning_rate": 4.79785360125261e-05, "loss": 1.0696, "step": 72000 }, { "epoch": 18.8, "learning_rate": 4.7962225991649276e-05, "loss": 1.0645, "step": 72500 }, { "epoch": 18.93, "learning_rate": 4.7945915970772445e-05, "loss": 1.06, "step": 73000 }, { "epoch": 19.06, "learning_rate": 4.7929605949895615e-05, "loss": 1.0546, "step": 73500 }, { "epoch": 19.19, "learning_rate": 4.791329592901879e-05, "loss": 1.0523, "step": 74000 }, { "epoch": 19.32, "learning_rate": 4.789698590814197e-05, "loss": 1.0483, "step": 74500 }, { "epoch": 19.45, "learning_rate": 4.788067588726514e-05, "loss": 1.0445, "step": 75000 }, { "epoch": 19.57, "learning_rate": 4.7864365866388314e-05, "loss": 1.0402, "step": 75500 }, { "epoch": 19.7, "learning_rate": 4.7848088465553235e-05, "loss": 1.0354, "step": 76000 }, { "epoch": 19.83, "learning_rate": 4.783177844467641e-05, "loss": 1.0329, "step": 76500 }, { "epoch": 19.96, "learning_rate": 4.781546842379959e-05, "loss": 1.0308, "step": 77000 }, { "epoch": 20.09, "learning_rate": 4.779919102296451e-05, "loss": 1.0272, "step": 77500 }, { "epoch": 20.22, "learning_rate": 4.7782881002087686e-05, "loss": 1.0218, "step": 78000 }, { "epoch": 20.35, "learning_rate": 4.7766570981210856e-05, "loss": 1.0197, "step": 78500 }, { "epoch": 20.48, "learning_rate": 4.775026096033403e-05, "loss": 1.0147, "step": 79000 }, { "epoch": 20.61, "learning_rate": 4.773395093945721e-05, "loss": 1.0115, "step": 79500 }, { "epoch": 20.74, "learning_rate": 4.771764091858038e-05, "loss": 1.0095, "step": 80000 }, { "epoch": 20.87, "learning_rate": 4.770136351774531e-05, "loss": 1.0087, "step": 80500 } ], "max_steps": 1542800, "num_train_epochs": 400, "total_flos": 2.1827467373904396e+19, "trial_name": null, "trial_params": null }