StevenLimcorn's picture
Upload 9 files
81a6bc8
raw
history blame
4.5 kB
{
"best_metric": 0.7913689415224279,
"best_model_checkpoint": "/content/result/semeval-unsup-promcse-bert-base-uncased-semeval2016-laptops",
"epoch": 30.0,
"global_step": 2370,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.58,
"eval_avg_sts": 0.7662035438889292,
"eval_sickr_spearman": 0.7410381462554306,
"eval_stsb_spearman": 0.7913689415224279,
"step": 125
},
{
"epoch": 3.16,
"eval_avg_sts": 0.7622844935042077,
"eval_sickr_spearman": 0.7436556491594021,
"eval_stsb_spearman": 0.7809133378490134,
"step": 250
},
{
"epoch": 4.75,
"eval_avg_sts": 0.7643160286763296,
"eval_sickr_spearman": 0.7401467850749421,
"eval_stsb_spearman": 0.7884852722777171,
"step": 375
},
{
"epoch": 6.33,
"learning_rate": 0.023670886075949367,
"loss": 0.0008,
"step": 500
},
{
"epoch": 6.33,
"eval_avg_sts": 0.736147879900914,
"eval_sickr_spearman": 0.7162090447402155,
"eval_stsb_spearman": 0.7560867150616126,
"step": 500
},
{
"epoch": 7.91,
"eval_avg_sts": 0.7436231350118074,
"eval_sickr_spearman": 0.7186257776401757,
"eval_stsb_spearman": 0.7686204923834391,
"step": 625
},
{
"epoch": 9.49,
"eval_avg_sts": 0.7393926207347856,
"eval_sickr_spearman": 0.7195540747376701,
"eval_stsb_spearman": 0.7592311667319013,
"step": 750
},
{
"epoch": 11.08,
"eval_avg_sts": 0.7452291779176539,
"eval_sickr_spearman": 0.723164908850483,
"eval_stsb_spearman": 0.7672934469848247,
"step": 875
},
{
"epoch": 12.66,
"learning_rate": 0.017341772151898735,
"loss": 0.0012,
"step": 1000
},
{
"epoch": 12.66,
"eval_avg_sts": 0.7565861527485951,
"eval_sickr_spearman": 0.7287989090180722,
"eval_stsb_spearman": 0.7843733964791179,
"step": 1000
},
{
"epoch": 14.24,
"eval_avg_sts": 0.7438608566065805,
"eval_sickr_spearman": 0.7231068392488443,
"eval_stsb_spearman": 0.7646148739643166,
"step": 1125
},
{
"epoch": 15.82,
"eval_avg_sts": 0.7467030170884223,
"eval_sickr_spearman": 0.7322541703933272,
"eval_stsb_spearman": 0.7611518637835176,
"step": 1250
},
{
"epoch": 17.41,
"eval_avg_sts": 0.7515738063184808,
"eval_sickr_spearman": 0.7319930252948081,
"eval_stsb_spearman": 0.7711545873421535,
"step": 1375
},
{
"epoch": 18.99,
"learning_rate": 0.011012658227848101,
"loss": 0.0008,
"step": 1500
},
{
"epoch": 18.99,
"eval_avg_sts": 0.7547593596864199,
"eval_sickr_spearman": 0.7316824081618079,
"eval_stsb_spearman": 0.777836311211032,
"step": 1500
},
{
"epoch": 20.57,
"eval_avg_sts": 0.7546503164567119,
"eval_sickr_spearman": 0.7284250829555794,
"eval_stsb_spearman": 0.7808755499578444,
"step": 1625
},
{
"epoch": 22.15,
"eval_avg_sts": 0.7412961666365114,
"eval_sickr_spearman": 0.7141825165083387,
"eval_stsb_spearman": 0.7684098167646841,
"step": 1750
},
{
"epoch": 23.73,
"eval_avg_sts": 0.7444813377835273,
"eval_sickr_spearman": 0.7180804325144473,
"eval_stsb_spearman": 0.7708822430526072,
"step": 1875
},
{
"epoch": 25.32,
"learning_rate": 0.0046835443037974685,
"loss": 0.0005,
"step": 2000
},
{
"epoch": 25.32,
"eval_avg_sts": 0.7457798773906168,
"eval_sickr_spearman": 0.7201111394521492,
"eval_stsb_spearman": 0.7714486153290844,
"step": 2000
},
{
"epoch": 26.9,
"eval_avg_sts": 0.7454552356177395,
"eval_sickr_spearman": 0.7205938039905009,
"eval_stsb_spearman": 0.7703166672449782,
"step": 2125
},
{
"epoch": 28.48,
"eval_avg_sts": 0.7438784407146768,
"eval_sickr_spearman": 0.7192442261022928,
"eval_stsb_spearman": 0.7685126553270608,
"step": 2250
},
{
"epoch": 30.0,
"step": 2370,
"train_runtime": 1196.069,
"train_samples_per_second": 1.981
}
],
"max_steps": 2370,
"num_train_epochs": 30,
"total_flos": 3153372403064832,
"trial_name": null,
"trial_params": null
}