|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.201680672268908, |
|
"eval_steps": 100, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.10504201680672269, |
|
"grad_norm": 8.704133033752441, |
|
"learning_rate": 4.166666666666667e-05, |
|
"loss": 3.5481, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.21008403361344538, |
|
"grad_norm": 10.129899978637695, |
|
"learning_rate": 4.9947570655942796e-05, |
|
"loss": 3.0062, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.31512605042016806, |
|
"grad_norm": 6.111865997314453, |
|
"learning_rate": 4.9734953280908904e-05, |
|
"loss": 2.9095, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.42016806722689076, |
|
"grad_norm": 4.366275787353516, |
|
"learning_rate": 4.936026311617316e-05, |
|
"loss": 2.9062, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.42016806722689076, |
|
"eval_loss": 2.872298240661621, |
|
"eval_runtime": 35.9441, |
|
"eval_samples_per_second": 13.02, |
|
"eval_steps_per_second": 1.641, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.5252100840336135, |
|
"grad_norm": 25.31424903869629, |
|
"learning_rate": 4.882595527372152e-05, |
|
"loss": 2.874, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.6302521008403361, |
|
"grad_norm": 53.229515075683594, |
|
"learning_rate": 4.813553074106761e-05, |
|
"loss": 2.887, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7352941176470589, |
|
"grad_norm": 7.568065166473389, |
|
"learning_rate": 4.7293513441455364e-05, |
|
"loss": 2.8903, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.8403361344537815, |
|
"grad_norm": 6.394404888153076, |
|
"learning_rate": 4.630542059139924e-05, |
|
"loss": 2.7455, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.8403361344537815, |
|
"eval_loss": 2.743553638458252, |
|
"eval_runtime": 35.9282, |
|
"eval_samples_per_second": 13.026, |
|
"eval_steps_per_second": 1.642, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9453781512605042, |
|
"grad_norm": 3.6301393508911133, |
|
"learning_rate": 4.517772654979023e-05, |
|
"loss": 2.7609, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.050420168067227, |
|
"grad_norm": 21.900226593017578, |
|
"learning_rate": 4.391782039544238e-05, |
|
"loss": 2.7411, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.1554621848739495, |
|
"grad_norm": 9.00839900970459, |
|
"learning_rate": 4.253395751104748e-05, |
|
"loss": 2.3212, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.2605042016806722, |
|
"grad_norm": 3.9047327041625977, |
|
"learning_rate": 4.10352054907785e-05, |
|
"loss": 2.0506, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.2605042016806722, |
|
"eval_loss": 2.8556525707244873, |
|
"eval_runtime": 35.8984, |
|
"eval_samples_per_second": 13.037, |
|
"eval_steps_per_second": 1.644, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.365546218487395, |
|
"grad_norm": 3.455016613006592, |
|
"learning_rate": 3.943138472597549e-05, |
|
"loss": 1.826, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.4705882352941178, |
|
"grad_norm": 10.288714408874512, |
|
"learning_rate": 3.773300405821908e-05, |
|
"loss": 1.8779, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.5756302521008403, |
|
"grad_norm": 2.957385540008545, |
|
"learning_rate": 3.595119192141706e-05, |
|
"loss": 1.85, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.680672268907563, |
|
"grad_norm": 3.40120267868042, |
|
"learning_rate": 3.409762342408719e-05, |
|
"loss": 1.835, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.680672268907563, |
|
"eval_loss": 2.7827320098876953, |
|
"eval_runtime": 36.0584, |
|
"eval_samples_per_second": 12.979, |
|
"eval_steps_per_second": 1.636, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.7857142857142856, |
|
"grad_norm": 3.446842670440674, |
|
"learning_rate": 3.218444384962071e-05, |
|
"loss": 1.8477, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.8907563025210083, |
|
"grad_norm": 3.1697998046875, |
|
"learning_rate": 3.0224189075781884e-05, |
|
"loss": 1.7934, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.995798319327731, |
|
"grad_norm": 2.8750455379486084, |
|
"learning_rate": 2.8229703434885163e-05, |
|
"loss": 1.7526, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.100840336134454, |
|
"grad_norm": 2.977962017059326, |
|
"learning_rate": 2.621405555286121e-05, |
|
"loss": 0.8246, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.100840336134454, |
|
"eval_loss": 3.001242160797119, |
|
"eval_runtime": 35.8998, |
|
"eval_samples_per_second": 13.036, |
|
"eval_steps_per_second": 1.643, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.2058823529411766, |
|
"grad_norm": 2.8460147380828857, |
|
"learning_rate": 2.419045271866611e-05, |
|
"loss": 0.7854, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.310924369747899, |
|
"grad_norm": 3.5654945373535156, |
|
"learning_rate": 2.2172154345117894e-05, |
|
"loss": 0.8065, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.4159663865546217, |
|
"grad_norm": 2.3066728115081787, |
|
"learning_rate": 2.0172385088197803e-05, |
|
"loss": 0.7753, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.5210084033613445, |
|
"grad_norm": 2.5332283973693848, |
|
"learning_rate": 1.820424819409143e-05, |
|
"loss": 0.824, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.5210084033613445, |
|
"eval_loss": 2.8449320793151855, |
|
"eval_runtime": 35.8956, |
|
"eval_samples_per_second": 13.038, |
|
"eval_steps_per_second": 1.644, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.6260504201680672, |
|
"grad_norm": 3.0348715782165527, |
|
"learning_rate": 1.6280639641752942e-05, |
|
"loss": 0.7481, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.73109243697479, |
|
"grad_norm": 2.7579290866851807, |
|
"learning_rate": 1.4414163643562755e-05, |
|
"loss": 0.7543, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.8361344537815127, |
|
"grad_norm": 2.145768404006958, |
|
"learning_rate": 1.2617050057750322e-05, |
|
"loss": 0.753, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.9411764705882355, |
|
"grad_norm": 2.6759989261627197, |
|
"learning_rate": 1.0901074253727336e-05, |
|
"loss": 0.7418, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.9411764705882355, |
|
"eval_loss": 2.796262741088867, |
|
"eval_runtime": 35.8962, |
|
"eval_samples_per_second": 13.038, |
|
"eval_steps_per_second": 1.644, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.046218487394958, |
|
"grad_norm": 1.9286493062973022, |
|
"learning_rate": 9.277479955403887e-06, |
|
"loss": 0.5768, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 3.1512605042016806, |
|
"grad_norm": 1.2430107593536377, |
|
"learning_rate": 7.756905568047393e-06, |
|
"loss": 0.2213, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.2563025210084033, |
|
"grad_norm": 1.5327359437942505, |
|
"learning_rate": 6.349314471418849e-06, |
|
"loss": 0.2563, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 3.361344537815126, |
|
"grad_norm": 1.448890209197998, |
|
"learning_rate": 5.063929735931985e-06, |
|
"loss": 0.2227, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.361344537815126, |
|
"eval_loss": 3.311384439468384, |
|
"eval_runtime": 35.9006, |
|
"eval_samples_per_second": 13.036, |
|
"eval_steps_per_second": 1.643, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 3.466386554621849, |
|
"grad_norm": 1.3534725904464722, |
|
"learning_rate": 3.90917368959989e-06, |
|
"loss": 0.2464, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 3.571428571428571, |
|
"grad_norm": 1.3151746988296509, |
|
"learning_rate": 2.892612731749414e-06, |
|
"loss": 0.2439, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.6764705882352944, |
|
"grad_norm": 1.4781558513641357, |
|
"learning_rate": 2.020907755104698e-06, |
|
"loss": 0.2355, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 3.7815126050420167, |
|
"grad_norm": 1.4184162616729736, |
|
"learning_rate": 1.2997705010932393e-06, |
|
"loss": 0.2198, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.7815126050420167, |
|
"eval_loss": 3.312718152999878, |
|
"eval_runtime": 35.8966, |
|
"eval_samples_per_second": 13.037, |
|
"eval_steps_per_second": 1.644, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.8865546218487395, |
|
"grad_norm": 1.545154333114624, |
|
"learning_rate": 7.339261343510206e-07, |
|
"loss": 0.222, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 3.991596638655462, |
|
"grad_norm": 1.5649031400680542, |
|
"learning_rate": 3.270822816527325e-07, |
|
"loss": 0.2069, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.0966386554621845, |
|
"grad_norm": 0.9737259149551392, |
|
"learning_rate": 8.190473813576572e-08, |
|
"loss": 0.1163, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 4.201680672268908, |
|
"grad_norm": 0.9735974669456482, |
|
"learning_rate": 0.0, |
|
"loss": 0.1478, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.201680672268908, |
|
"eval_loss": 3.372861385345459, |
|
"eval_runtime": 35.8981, |
|
"eval_samples_per_second": 13.037, |
|
"eval_steps_per_second": 1.644, |
|
"step": 1000 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 1000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 250, |
|
"total_flos": 9.22205177249792e+16, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|