|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 6411, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 6479046246400.0, |
|
"learning_rate": 3.3255451713395644e-05, |
|
"loss": 31699221108.3091, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 8700652158976.0, |
|
"learning_rate": 4.8162593170393484e-05, |
|
"loss": 14378762930.6604, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 16.28143310546875, |
|
"learning_rate": 4.446177847113885e-05, |
|
"loss": 3635725784.4309, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 6.706409454345703, |
|
"learning_rate": 4.076096377188421e-05, |
|
"loss": 1017393142.4075, |
|
"step": 1708 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 9.295268058776855, |
|
"learning_rate": 3.7060149072629574e-05, |
|
"loss": 1230655574.3326, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 7.9164958000183105, |
|
"learning_rate": 3.335933437337493e-05, |
|
"loss": 387735187.4848, |
|
"step": 2562 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 18.9678897857666, |
|
"learning_rate": 2.96585196741203e-05, |
|
"loss": 610998305.5738, |
|
"step": 2989 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 22.521636962890625, |
|
"learning_rate": 2.595770497486566e-05, |
|
"loss": 286763645.9016, |
|
"step": 3416 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 6.2225751876831055, |
|
"learning_rate": 2.2256890275611026e-05, |
|
"loss": 146726173.377, |
|
"step": 3843 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 4.1397271156311035, |
|
"learning_rate": 1.8556075576356388e-05, |
|
"loss": 153057112.1311, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 20.3540096282959, |
|
"learning_rate": 1.4855260877101752e-05, |
|
"loss": 167597749.0585, |
|
"step": 4697 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 2.9089903831481934, |
|
"learning_rate": 1.1154446177847114e-05, |
|
"loss": 31491796.2342, |
|
"step": 5124 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 11.952803611755371, |
|
"learning_rate": 7.453631478592478e-06, |
|
"loss": 228096436.459, |
|
"step": 5551 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 8.380680084228516, |
|
"learning_rate": 3.7528167793378402e-06, |
|
"loss": 170669039.2131, |
|
"step": 5978 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 11.055083274841309, |
|
"learning_rate": 5.2002080083203335e-08, |
|
"loss": 77525903.2881, |
|
"step": 6405 |
|
} |
|
], |
|
"logging_steps": 427, |
|
"max_steps": 6411, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 1.089435891661996e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|