|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 5.0, |
|
"eval_steps": 500, |
|
"global_step": 450, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 8.953761100769043, |
|
"learning_rate": 1.888888888888889e-05, |
|
"loss": 1.3623, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 10.632478713989258, |
|
"learning_rate": 3.777777777777778e-05, |
|
"loss": 1.0225, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 16.316967010498047, |
|
"learning_rate": 4.925925925925926e-05, |
|
"loss": 0.6166, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 21.347354888916016, |
|
"learning_rate": 4.7160493827160495e-05, |
|
"loss": 0.5825, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 8.19096565246582, |
|
"learning_rate": 4.506172839506173e-05, |
|
"loss": 0.4259, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 12.711194038391113, |
|
"learning_rate": 4.296296296296296e-05, |
|
"loss": 0.2901, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"grad_norm": 18.733543395996094, |
|
"learning_rate": 4.0864197530864204e-05, |
|
"loss": 0.2684, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 6.416273593902588, |
|
"learning_rate": 3.876543209876544e-05, |
|
"loss": 0.2797, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 4.896417617797852, |
|
"learning_rate": 3.6666666666666666e-05, |
|
"loss": 0.3023, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 17.5325984954834, |
|
"learning_rate": 3.45679012345679e-05, |
|
"loss": 0.2397, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 9.514878273010254, |
|
"learning_rate": 3.2469135802469134e-05, |
|
"loss": 0.2499, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 6.783230304718018, |
|
"learning_rate": 3.037037037037037e-05, |
|
"loss": 0.2137, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 9.813042640686035, |
|
"learning_rate": 2.8271604938271606e-05, |
|
"loss": 0.1946, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 4.386660099029541, |
|
"learning_rate": 2.617283950617284e-05, |
|
"loss": 0.2326, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"grad_norm": 13.430244445800781, |
|
"learning_rate": 2.4074074074074074e-05, |
|
"loss": 0.1931, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"grad_norm": 10.026385307312012, |
|
"learning_rate": 2.1975308641975308e-05, |
|
"loss": 0.1584, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"grad_norm": 5.96959924697876, |
|
"learning_rate": 1.9876543209876546e-05, |
|
"loss": 0.1612, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"grad_norm": 8.849898338317871, |
|
"learning_rate": 1.777777777777778e-05, |
|
"loss": 0.1843, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"grad_norm": 4.787600517272949, |
|
"learning_rate": 1.5679012345679014e-05, |
|
"loss": 0.1395, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"grad_norm": 5.98640251159668, |
|
"learning_rate": 1.3580246913580247e-05, |
|
"loss": 0.1351, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"grad_norm": 8.119280815124512, |
|
"learning_rate": 1.1481481481481482e-05, |
|
"loss": 0.1626, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"grad_norm": 7.947086334228516, |
|
"learning_rate": 9.382716049382717e-06, |
|
"loss": 0.1658, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"grad_norm": 2.4981260299682617, |
|
"learning_rate": 7.283950617283951e-06, |
|
"loss": 0.1147, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"grad_norm": 5.954956531524658, |
|
"learning_rate": 5.185185185185185e-06, |
|
"loss": 0.1295, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"grad_norm": 9.320396423339844, |
|
"learning_rate": 3.0864197530864196e-06, |
|
"loss": 0.1309, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"grad_norm": 9.413307189941406, |
|
"learning_rate": 9.876543209876544e-07, |
|
"loss": 0.095, |
|
"step": 442 |
|
} |
|
], |
|
"logging_steps": 17, |
|
"max_steps": 450, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"total_flos": 1.1242852922068992e+18, |
|
"train_batch_size": 32, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|