|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 6, |
|
"global_step": 26, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.038461538461538464, |
|
"grad_norm": 8.71969985961914, |
|
"learning_rate": 1e-05, |
|
"loss": 1.7925, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.038461538461538464, |
|
"eval_loss": 2.0411531925201416, |
|
"eval_runtime": 1.6566, |
|
"eval_samples_per_second": 13.28, |
|
"eval_steps_per_second": 1.207, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.07692307692307693, |
|
"grad_norm": 4.587398052215576, |
|
"learning_rate": 2e-05, |
|
"loss": 2.1894, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.11538461538461539, |
|
"grad_norm": 4.494429111480713, |
|
"learning_rate": 3e-05, |
|
"loss": 1.9438, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.15384615384615385, |
|
"grad_norm": 7.78571891784668, |
|
"learning_rate": 4e-05, |
|
"loss": 1.9463, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.19230769230769232, |
|
"grad_norm": 6.792930603027344, |
|
"learning_rate": 5e-05, |
|
"loss": 1.8652, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.23076923076923078, |
|
"grad_norm": 10.765946388244629, |
|
"learning_rate": 6e-05, |
|
"loss": 1.6872, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.23076923076923078, |
|
"eval_loss": 1.6089171171188354, |
|
"eval_runtime": 1.667, |
|
"eval_samples_per_second": 13.198, |
|
"eval_steps_per_second": 1.2, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.2692307692307692, |
|
"grad_norm": 3.0110535621643066, |
|
"learning_rate": 7e-05, |
|
"loss": 1.6877, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.3076923076923077, |
|
"grad_norm": 2.871809959411621, |
|
"learning_rate": 8e-05, |
|
"loss": 1.4855, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.34615384615384615, |
|
"grad_norm": 2.223036050796509, |
|
"learning_rate": 9e-05, |
|
"loss": 1.2108, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 2.2891507148742676, |
|
"learning_rate": 0.0001, |
|
"loss": 1.0259, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.4230769230769231, |
|
"grad_norm": 2.058546543121338, |
|
"learning_rate": 9.997207818651274e-05, |
|
"loss": 0.8116, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.46153846153846156, |
|
"grad_norm": 2.8107495307922363, |
|
"learning_rate": 9.988834393115767e-05, |
|
"loss": 0.6967, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.46153846153846156, |
|
"eval_loss": 0.6327503323554993, |
|
"eval_runtime": 1.6315, |
|
"eval_samples_per_second": 13.485, |
|
"eval_steps_per_second": 1.226, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 2.0721328258514404, |
|
"learning_rate": 9.974889075442521e-05, |
|
"loss": 0.5983, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.5384615384615384, |
|
"grad_norm": 1.9303380250930786, |
|
"learning_rate": 9.9553874407739e-05, |
|
"loss": 0.4963, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.5769230769230769, |
|
"grad_norm": 1.40817391872406, |
|
"learning_rate": 9.930351269950143e-05, |
|
"loss": 0.3881, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.6153846153846154, |
|
"grad_norm": 1.4169831275939941, |
|
"learning_rate": 9.899808525182935e-05, |
|
"loss": 0.3139, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.6538461538461539, |
|
"grad_norm": 0.9847315549850464, |
|
"learning_rate": 9.863793318825186e-05, |
|
"loss": 0.2776, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.6923076923076923, |
|
"grad_norm": 1.2575572729110718, |
|
"learning_rate": 9.822345875271883e-05, |
|
"loss": 0.3327, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.6923076923076923, |
|
"eval_loss": 0.2710803747177124, |
|
"eval_runtime": 1.6739, |
|
"eval_samples_per_second": 13.143, |
|
"eval_steps_per_second": 1.195, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.7307692307692307, |
|
"grad_norm": 0.7373743057250977, |
|
"learning_rate": 9.775512486034563e-05, |
|
"loss": 0.1967, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 0.8841055035591125, |
|
"learning_rate": 9.723345458039594e-05, |
|
"loss": 0.244, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.8076923076923077, |
|
"grad_norm": 0.8621382713317871, |
|
"learning_rate": 9.665903055208014e-05, |
|
"loss": 0.1927, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.8461538461538461, |
|
"grad_norm": 0.9871663451194763, |
|
"learning_rate": 9.603249433382144e-05, |
|
"loss": 0.2435, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.8846153846153846, |
|
"grad_norm": 0.6780880689620972, |
|
"learning_rate": 9.535454568671704e-05, |
|
"loss": 0.1498, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.9230769230769231, |
|
"grad_norm": 0.7006078958511353, |
|
"learning_rate": 9.462594179299406e-05, |
|
"loss": 0.1784, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.9230769230769231, |
|
"eval_loss": 0.17325076460838318, |
|
"eval_runtime": 1.6307, |
|
"eval_samples_per_second": 13.491, |
|
"eval_steps_per_second": 1.226, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.9615384615384616, |
|
"grad_norm": 0.69418865442276, |
|
"learning_rate": 9.384749641033359e-05, |
|
"loss": 0.2146, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.5629783272743225, |
|
"learning_rate": 9.302007896300698e-05, |
|
"loss": 0.1474, |
|
"step": 26 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 104, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.3435852608765952e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|