|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.995708154506438, |
|
"eval_steps": 500, |
|
"global_step": 349, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 5.055780862946804, |
|
"learning_rate": 9.997135147120633e-05, |
|
"loss": 2.3426, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 3.9099604179547867, |
|
"learning_rate": 9.98854387143534e-05, |
|
"loss": 1.9269, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 2.3917278619890237, |
|
"learning_rate": 9.974236018040474e-05, |
|
"loss": 1.6971, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 2.0300075071526407, |
|
"learning_rate": 9.954227982894034e-05, |
|
"loss": 1.5024, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 2.5503247326265703, |
|
"learning_rate": 9.928542694026862e-05, |
|
"loss": 1.4291, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 2.8860783306647813, |
|
"learning_rate": 9.897209585268458e-05, |
|
"loss": 1.3888, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.7598771943831562, |
|
"learning_rate": 9.86026456251757e-05, |
|
"loss": 1.3614, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.8472495254954506, |
|
"learning_rate": 9.817749962596115e-05, |
|
"loss": 1.3218, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.8685305808519006, |
|
"learning_rate": 9.769714504733694e-05, |
|
"loss": 1.3144, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.6934593149050314, |
|
"learning_rate": 9.716213234738215e-05, |
|
"loss": 1.3105, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.6536690261785199, |
|
"learning_rate": 9.657307461916635e-05, |
|
"loss": 1.287, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.8152365885852073, |
|
"learning_rate": 9.59306468881811e-05, |
|
"loss": 1.2877, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 1.6831619071424775, |
|
"learning_rate": 9.52355853388003e-05, |
|
"loss": 1.2504, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.8966578603527602, |
|
"learning_rate": 9.448868647065642e-05, |
|
"loss": 1.2782, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.9475911961026766, |
|
"learning_rate": 9.369080618589864e-05, |
|
"loss": 1.2517, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.7144643185356307, |
|
"learning_rate": 9.284285880837946e-05, |
|
"loss": 1.2442, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 1.7507182276034248, |
|
"learning_rate": 9.194581603589328e-05, |
|
"loss": 1.2174, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.782170598732355, |
|
"learning_rate": 9.100070582666795e-05, |
|
"loss": 1.2231, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.884782447980567, |
|
"learning_rate": 9.000861122138517e-05, |
|
"loss": 1.1901, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.9269085187568622, |
|
"learning_rate": 8.897066910207958e-05, |
|
"loss": 1.1959, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 2.0913092281668977, |
|
"learning_rate": 8.788806888933881e-05, |
|
"loss": 1.1953, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.6532109062340223, |
|
"learning_rate": 8.676205117929752e-05, |
|
"loss": 1.2135, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.9499528694077117, |
|
"learning_rate": 8.559390632198723e-05, |
|
"loss": 1.2141, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.2135138511657715, |
|
"eval_runtime": 39.1621, |
|
"eval_samples_per_second": 7.66, |
|
"eval_steps_per_second": 0.97, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1.6359783575613835, |
|
"learning_rate": 8.438497294267117e-05, |
|
"loss": 1.1555, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 1.9376629758093418, |
|
"learning_rate": 8.313663640785839e-05, |
|
"loss": 1.123, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 1.8078455578062511, |
|
"learning_rate": 8.185032723775539e-05, |
|
"loss": 1.1545, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 1.78298002573264, |
|
"learning_rate": 8.052751946697403e-05, |
|
"loss": 1.0917, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1.8433442214623892, |
|
"learning_rate": 7.916972895537471e-05, |
|
"loss": 1.1507, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.7754964498370087, |
|
"learning_rate": 7.777851165098012e-05, |
|
"loss": 1.1461, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 1.9524999874543483, |
|
"learning_rate": 7.635546180695038e-05, |
|
"loss": 1.0643, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 1.7260817926145178, |
|
"learning_rate": 7.490221015466279e-05, |
|
"loss": 1.1075, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 2.253848011637847, |
|
"learning_rate": 7.342042203498951e-05, |
|
"loss": 1.1714, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 1.9774736597698745, |
|
"learning_rate": 7.191179548991507e-05, |
|
"loss": 1.1076, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 2.0714596334335673, |
|
"learning_rate": 7.037805931668005e-05, |
|
"loss": 1.1452, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.7678144581225992, |
|
"learning_rate": 6.882097108668132e-05, |
|
"loss": 1.1356, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 1.709740560594397, |
|
"learning_rate": 6.724231513139852e-05, |
|
"loss": 1.1126, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 1.7902700297237863, |
|
"learning_rate": 6.564390049765528e-05, |
|
"loss": 1.0937, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 1.8180035012469915, |
|
"learning_rate": 6.402755887455792e-05, |
|
"loss": 1.1122, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"grad_norm": 1.8387833131421265, |
|
"learning_rate": 6.239514249448767e-05, |
|
"loss": 1.1156, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 1.7206104055323197, |
|
"learning_rate": 6.0748522010551215e-05, |
|
"loss": 1.0624, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"grad_norm": 1.96126387756175, |
|
"learning_rate": 5.908958435292241e-05, |
|
"loss": 1.0896, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 1.7505678845455834, |
|
"learning_rate": 5.742023056653131e-05, |
|
"loss": 1.1093, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 1.88819291262154, |
|
"learning_rate": 5.574237363257858e-05, |
|
"loss": 1.1057, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 1.981696800102056, |
|
"learning_rate": 5.4057936276371565e-05, |
|
"loss": 1.0809, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"grad_norm": 2.0184048752186734, |
|
"learning_rate": 5.236884876399429e-05, |
|
"loss": 1.1245, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 1.7913247269002772, |
|
"learning_rate": 5.0677046690336096e-05, |
|
"loss": 1.1107, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.147458553314209, |
|
"eval_runtime": 38.5013, |
|
"eval_samples_per_second": 7.792, |
|
"eval_steps_per_second": 0.987, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"grad_norm": 1.573903357064543, |
|
"learning_rate": 4.898446876101379e-05, |
|
"loss": 1.0289, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"grad_norm": 1.886593037989885, |
|
"learning_rate": 4.729305457072913e-05, |
|
"loss": 1.0287, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 2.3850699585618855, |
|
"learning_rate": 4.560474238060739e-05, |
|
"loss": 1.0079, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 1.8897004088761564, |
|
"learning_rate": 4.392146689706425e-05, |
|
"loss": 1.0176, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 1.6637180057615024, |
|
"learning_rate": 4.224515705474603e-05, |
|
"loss": 1.0393, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 1.9082972238615958, |
|
"learning_rate": 4.057773380608411e-05, |
|
"loss": 0.9883, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 1.6248508379655304, |
|
"learning_rate": 3.892110791999649e-05, |
|
"loss": 1.0335, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"grad_norm": 2.057742891690489, |
|
"learning_rate": 3.7277177792259114e-05, |
|
"loss": 1.0055, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 2.1082750754536725, |
|
"learning_rate": 3.5647827270055945e-05, |
|
"loss": 1.0363, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 1.954319411917034, |
|
"learning_rate": 3.403492349320101e-05, |
|
"loss": 1.0049, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 1.835816144902083, |
|
"learning_rate": 3.244031475450599e-05, |
|
"loss": 1.0382, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 1.905542574366058, |
|
"learning_rate": 3.086582838174551e-05, |
|
"loss": 1.0012, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 1.9640737297497406, |
|
"learning_rate": 2.9313268643646986e-05, |
|
"loss": 1.0183, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 1.8759801566914525, |
|
"learning_rate": 2.7784414682304832e-05, |
|
"loss": 1.0513, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 1.7782596177203436, |
|
"learning_rate": 2.628101847438835e-05, |
|
"loss": 1.0053, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 1.8683159360283232, |
|
"learning_rate": 2.4804802823479613e-05, |
|
"loss": 1.0159, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 2.234906406234719, |
|
"learning_rate": 2.3357459385841823e-05, |
|
"loss": 1.0041, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 1.8995547320155095, |
|
"learning_rate": 2.194064673188089e-05, |
|
"loss": 1.0403, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 1.8065108976199886, |
|
"learning_rate": 2.055598844552129e-05, |
|
"loss": 1.0004, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 1.7757725659621284, |
|
"learning_rate": 1.920507126367448e-05, |
|
"loss": 1.0075, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 1.7478001464098316, |
|
"learning_rate": 1.7889443257931737e-05, |
|
"loss": 1.0163, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 1.7453331826305865, |
|
"learning_rate": 1.6610612060565234e-05, |
|
"loss": 1.0246, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 2.015231393689711, |
|
"learning_rate": 1.5370043136870148e-05, |
|
"loss": 1.0339, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.123692274093628, |
|
"eval_runtime": 38.4705, |
|
"eval_samples_per_second": 7.798, |
|
"eval_steps_per_second": 0.988, |
|
"step": 349 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 464, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"total_flos": 1663515060338688.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|