dima806's picture
Upload folder using huggingface_hub
466bae2 verified
raw
history blame
4.28 kB
{
"best_metric": 0.832238495349884,
"best_model_checkpoint": "cat_breed_image_detection/checkpoint-7623",
"epoch": 3.0,
"eval_steps": 500,
"global_step": 7623,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.19677292404565133,
"grad_norm": 10.799985885620117,
"learning_rate": 6.584048593688102e-07,
"loss": 0.7525,
"step": 500
},
{
"epoch": 0.39354584809130266,
"grad_norm": 13.88831615447998,
"learning_rate": 6.12188036445266e-07,
"loss": 0.7606,
"step": 1000
},
{
"epoch": 0.5903187721369539,
"grad_norm": 10.507097244262695,
"learning_rate": 5.659712135217218e-07,
"loss": 0.7568,
"step": 1500
},
{
"epoch": 0.7870916961826053,
"grad_norm": 14.891112327575684,
"learning_rate": 5.197543905981778e-07,
"loss": 0.7615,
"step": 2000
},
{
"epoch": 0.9838646202282566,
"grad_norm": 14.09949016571045,
"learning_rate": 4.7353756767463353e-07,
"loss": 0.7659,
"step": 2500
},
{
"epoch": 1.0,
"eval_accuracy": 0.7616899173309714,
"eval_loss": 0.8394082188606262,
"eval_model_preparation_time": 0.0058,
"eval_runtime": 1186.6978,
"eval_samples_per_second": 91.332,
"eval_steps_per_second": 2.854,
"step": 2541
},
{
"epoch": 1.1806375442739079,
"grad_norm": 12.55162525177002,
"learning_rate": 4.273207447510894e-07,
"loss": 0.7569,
"step": 3000
},
{
"epoch": 1.3774104683195592,
"grad_norm": 8.38792896270752,
"learning_rate": 3.8110392182754523e-07,
"loss": 0.7385,
"step": 3500
},
{
"epoch": 1.5741833923652107,
"grad_norm": 14.774518013000488,
"learning_rate": 3.34887098904001e-07,
"loss": 0.7586,
"step": 4000
},
{
"epoch": 1.770956316410862,
"grad_norm": 12.208711624145508,
"learning_rate": 2.886702759804569e-07,
"loss": 0.7445,
"step": 4500
},
{
"epoch": 1.9677292404565132,
"grad_norm": 10.817965507507324,
"learning_rate": 2.424534530569127e-07,
"loss": 0.7536,
"step": 5000
},
{
"epoch": 2.0,
"eval_accuracy": 0.7624833923826395,
"eval_loss": 0.8353193402290344,
"eval_model_preparation_time": 0.0058,
"eval_runtime": 1195.2016,
"eval_samples_per_second": 90.683,
"eval_steps_per_second": 2.834,
"step": 5082
},
{
"epoch": 2.1645021645021645,
"grad_norm": 8.807418823242188,
"learning_rate": 1.9623663013336853e-07,
"loss": 0.75,
"step": 5500
},
{
"epoch": 2.3612750885478158,
"grad_norm": 11.618365287780762,
"learning_rate": 1.5001980720982439e-07,
"loss": 0.7474,
"step": 6000
},
{
"epoch": 2.558048012593467,
"grad_norm": 8.761614799499512,
"learning_rate": 1.0380298428628021e-07,
"loss": 0.7416,
"step": 6500
},
{
"epoch": 2.7548209366391183,
"grad_norm": 10.143717765808105,
"learning_rate": 5.7586161362736035e-08,
"loss": 0.7421,
"step": 7000
},
{
"epoch": 2.9515938606847696,
"grad_norm": 11.86281967163086,
"learning_rate": 1.1369338439191866e-08,
"loss": 0.7394,
"step": 7500
},
{
"epoch": 3.0,
"eval_accuracy": 0.7636090197815175,
"eval_loss": 0.832238495349884,
"eval_model_preparation_time": 0.0058,
"eval_runtime": 1186.9177,
"eval_samples_per_second": 91.316,
"eval_steps_per_second": 2.854,
"step": 7623
}
],
"logging_steps": 500,
"max_steps": 7623,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.781059536914691e+19,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}