|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.400250941028858, |
|
"eval_steps": 479, |
|
"global_step": 15304, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0075282308657465494, |
|
"grad_norm": 87.12931823730469, |
|
"learning_rate": 3.147218736930155e-07, |
|
"loss": 12.3074, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.015056461731493099, |
|
"grad_norm": 64.15751647949219, |
|
"learning_rate": 6.660393140945211e-07, |
|
"loss": 15.7221, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.02258469259723965, |
|
"grad_norm": 81.39651489257812, |
|
"learning_rate": 1.0173567544960265e-06, |
|
"loss": 10.8027, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.030112923462986198, |
|
"grad_norm": 22.048904418945312, |
|
"learning_rate": 1.3686741948975323e-06, |
|
"loss": 8.9559, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.037641154328732745, |
|
"grad_norm": 11.542724609375, |
|
"learning_rate": 1.7126725219573398e-06, |
|
"loss": 8.8511, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.0451693851944793, |
|
"grad_norm": 10.625059127807617, |
|
"learning_rate": 2.063989962358846e-06, |
|
"loss": 9.3478, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.05269761606022585, |
|
"grad_norm": 14.21434211730957, |
|
"learning_rate": 2.415307402760351e-06, |
|
"loss": 8.8892, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.060225846925972396, |
|
"grad_norm": 13.216053009033203, |
|
"learning_rate": 2.7666248431618565e-06, |
|
"loss": 8.3008, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.06775407779171895, |
|
"grad_norm": 22.503334045410156, |
|
"learning_rate": 3.117942283563362e-06, |
|
"loss": 7.3455, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_nli-pairs_loss": 6.591032028198242, |
|
"eval_nli-pairs_runtime": 4.3469, |
|
"eval_nli-pairs_samples_per_second": 46.01, |
|
"eval_nli-pairs_steps_per_second": 1.15, |
|
"eval_sts-test_pearson_cosine": 0.39488461174644296, |
|
"eval_sts-test_pearson_dot": 0.15593446481859455, |
|
"eval_sts-test_pearson_euclidean": 0.39975070029693277, |
|
"eval_sts-test_pearson_manhattan": 0.4314268556737928, |
|
"eval_sts-test_pearson_max": 0.4314268556737928, |
|
"eval_sts-test_spearman_cosine": 0.3997824055251076, |
|
"eval_sts-test_spearman_dot": 0.14324216739430146, |
|
"eval_sts-test_spearman_euclidean": 0.40262274612650517, |
|
"eval_sts-test_spearman_manhattan": 0.42925492969387746, |
|
"eval_sts-test_spearman_max": 0.42925492969387746, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_vitaminc-pairs_loss": 6.247874736785889, |
|
"eval_vitaminc-pairs_runtime": 1.4382, |
|
"eval_vitaminc-pairs_samples_per_second": 115.426, |
|
"eval_vitaminc-pairs_steps_per_second": 2.781, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_sts-label_loss": 3.371708869934082, |
|
"eval_sts-label_runtime": 0.3925, |
|
"eval_sts-label_samples_per_second": 509.603, |
|
"eval_sts-label_steps_per_second": 12.74, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_qnli-contrastive_loss": 3.5311310291290283, |
|
"eval_qnli-contrastive_runtime": 0.2814, |
|
"eval_qnli-contrastive_samples_per_second": 710.798, |
|
"eval_qnli-contrastive_steps_per_second": 17.77, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_scitail-pairs-qa_loss": 5.4017333984375, |
|
"eval_scitail-pairs-qa_runtime": 1.1087, |
|
"eval_scitail-pairs-qa_samples_per_second": 180.387, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.51, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_scitail-pairs-pos_loss": 4.29502534866333, |
|
"eval_scitail-pairs-pos_runtime": 2.4269, |
|
"eval_scitail-pairs-pos_samples_per_second": 82.41, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.06, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_xsum-pairs_loss": 3.540722370147705, |
|
"eval_xsum-pairs_runtime": 1.0447, |
|
"eval_xsum-pairs_samples_per_second": 191.444, |
|
"eval_xsum-pairs_steps_per_second": 4.786, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_compression-pairs_loss": 2.8060033321380615, |
|
"eval_compression-pairs_runtime": 0.2399, |
|
"eval_compression-pairs_samples_per_second": 833.621, |
|
"eval_compression-pairs_steps_per_second": 20.841, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_sciq_pairs_loss": 10.621454238891602, |
|
"eval_sciq_pairs_runtime": 9.0638, |
|
"eval_sciq_pairs_samples_per_second": 22.066, |
|
"eval_sciq_pairs_steps_per_second": 0.552, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_qasc_pairs_loss": 7.7197771072387695, |
|
"eval_qasc_pairs_runtime": 1.2078, |
|
"eval_qasc_pairs_samples_per_second": 165.595, |
|
"eval_qasc_pairs_steps_per_second": 4.14, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_openbookqa_pairs_loss": 7.620975494384766, |
|
"eval_openbookqa_pairs_runtime": 1.053, |
|
"eval_openbookqa_pairs_samples_per_second": 189.941, |
|
"eval_openbookqa_pairs_steps_per_second": 4.749, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_msmarco_pairs_loss": 8.353594779968262, |
|
"eval_msmarco_pairs_runtime": 2.5338, |
|
"eval_msmarco_pairs_samples_per_second": 78.932, |
|
"eval_msmarco_pairs_steps_per_second": 1.973, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_nq_pairs_loss": 7.995354652404785, |
|
"eval_nq_pairs_runtime": 5.7107, |
|
"eval_nq_pairs_samples_per_second": 35.022, |
|
"eval_nq_pairs_steps_per_second": 0.876, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_trivia_pairs_loss": 8.177907943725586, |
|
"eval_trivia_pairs_runtime": 9.1824, |
|
"eval_trivia_pairs_samples_per_second": 21.781, |
|
"eval_trivia_pairs_steps_per_second": 0.545, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_quora_pairs_loss": 1.1983369588851929, |
|
"eval_quora_pairs_runtime": 0.659, |
|
"eval_quora_pairs_samples_per_second": 303.48, |
|
"eval_quora_pairs_steps_per_second": 7.587, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_gooaq_pairs_loss": 7.4573974609375, |
|
"eval_gooaq_pairs_runtime": 1.5917, |
|
"eval_gooaq_pairs_samples_per_second": 125.65, |
|
"eval_gooaq_pairs_steps_per_second": 3.141, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07512547051442911, |
|
"eval_mrpc_pairs_loss": 2.1152825355529785, |
|
"eval_mrpc_pairs_runtime": 0.241, |
|
"eval_mrpc_pairs_samples_per_second": 829.751, |
|
"eval_mrpc_pairs_steps_per_second": 20.744, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.07528230865746549, |
|
"grad_norm": 17.753456115722656, |
|
"learning_rate": 3.469259723964868e-06, |
|
"loss": 8.0369, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.08281053952321205, |
|
"grad_norm": 12.012594223022461, |
|
"learning_rate": 3.820577164366374e-06, |
|
"loss": 6.2732, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.0903387703889586, |
|
"grad_norm": 34.92698287963867, |
|
"learning_rate": 4.1718946047678796e-06, |
|
"loss": 7.8529, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.09786700125470514, |
|
"grad_norm": 23.564632415771484, |
|
"learning_rate": 4.523212045169385e-06, |
|
"loss": 5.8643, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.1053952321204517, |
|
"grad_norm": 22.126293182373047, |
|
"learning_rate": 4.874529485570891e-06, |
|
"loss": 6.3179, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.11292346298619825, |
|
"grad_norm": 25.067686080932617, |
|
"learning_rate": 5.225846925972396e-06, |
|
"loss": 6.1175, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.12045169385194479, |
|
"grad_norm": 29.170730590820312, |
|
"learning_rate": 5.577164366373902e-06, |
|
"loss": 5.2392, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.12797992471769135, |
|
"grad_norm": 29.377540588378906, |
|
"learning_rate": 5.928481806775407e-06, |
|
"loss": 5.8324, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.1355081555834379, |
|
"grad_norm": 33.512088775634766, |
|
"learning_rate": 6.279799247176913e-06, |
|
"loss": 5.1523, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.14303638644918445, |
|
"grad_norm": 32.54931640625, |
|
"learning_rate": 6.6311166875784185e-06, |
|
"loss": 6.0303, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_nli-pairs_loss": 4.317643165588379, |
|
"eval_nli-pairs_runtime": 4.0158, |
|
"eval_nli-pairs_samples_per_second": 49.803, |
|
"eval_nli-pairs_steps_per_second": 1.245, |
|
"eval_sts-test_pearson_cosine": 0.6751726661173544, |
|
"eval_sts-test_pearson_dot": 0.5308173325280101, |
|
"eval_sts-test_pearson_euclidean": 0.6660519042507951, |
|
"eval_sts-test_pearson_manhattan": 0.6723411683739887, |
|
"eval_sts-test_pearson_max": 0.6751726661173544, |
|
"eval_sts-test_spearman_cosine": 0.6458966208807124, |
|
"eval_sts-test_spearman_dot": 0.5040208096497271, |
|
"eval_sts-test_spearman_euclidean": 0.6447205374312966, |
|
"eval_sts-test_spearman_manhattan": 0.6498560301461127, |
|
"eval_sts-test_spearman_max": 0.6498560301461127, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_vitaminc-pairs_loss": 6.306981563568115, |
|
"eval_vitaminc-pairs_runtime": 1.4737, |
|
"eval_vitaminc-pairs_samples_per_second": 112.643, |
|
"eval_vitaminc-pairs_steps_per_second": 2.714, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_sts-label_loss": 3.7677345275878906, |
|
"eval_sts-label_runtime": 0.4136, |
|
"eval_sts-label_samples_per_second": 483.59, |
|
"eval_sts-label_steps_per_second": 12.09, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_qnli-contrastive_loss": 2.8633975982666016, |
|
"eval_qnli-contrastive_runtime": 0.2816, |
|
"eval_qnli-contrastive_samples_per_second": 710.123, |
|
"eval_qnli-contrastive_steps_per_second": 17.753, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_scitail-pairs-qa_loss": 1.4301409721374512, |
|
"eval_scitail-pairs-qa_runtime": 1.0526, |
|
"eval_scitail-pairs-qa_samples_per_second": 190.003, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.75, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_scitail-pairs-pos_loss": 2.468087673187256, |
|
"eval_scitail-pairs-pos_runtime": 2.3275, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.928, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.148, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_xsum-pairs_loss": 2.153658151626587, |
|
"eval_xsum-pairs_runtime": 1.0409, |
|
"eval_xsum-pairs_samples_per_second": 192.151, |
|
"eval_xsum-pairs_steps_per_second": 4.804, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_compression-pairs_loss": 1.6288033723831177, |
|
"eval_compression-pairs_runtime": 0.2383, |
|
"eval_compression-pairs_samples_per_second": 839.253, |
|
"eval_compression-pairs_steps_per_second": 20.981, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_sciq_pairs_loss": 9.78779411315918, |
|
"eval_sciq_pairs_runtime": 9.0233, |
|
"eval_sciq_pairs_samples_per_second": 22.165, |
|
"eval_sciq_pairs_steps_per_second": 0.554, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_qasc_pairs_loss": 3.7814972400665283, |
|
"eval_qasc_pairs_runtime": 1.2108, |
|
"eval_qasc_pairs_samples_per_second": 165.179, |
|
"eval_qasc_pairs_steps_per_second": 4.129, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_openbookqa_pairs_loss": 5.005772590637207, |
|
"eval_openbookqa_pairs_runtime": 1.0415, |
|
"eval_openbookqa_pairs_samples_per_second": 192.037, |
|
"eval_openbookqa_pairs_steps_per_second": 4.801, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_msmarco_pairs_loss": 4.574879169464111, |
|
"eval_msmarco_pairs_runtime": 2.527, |
|
"eval_msmarco_pairs_samples_per_second": 79.146, |
|
"eval_msmarco_pairs_steps_per_second": 1.979, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_nq_pairs_loss": 5.281248569488525, |
|
"eval_nq_pairs_runtime": 5.6503, |
|
"eval_nq_pairs_samples_per_second": 35.397, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_trivia_pairs_loss": 4.913428783416748, |
|
"eval_trivia_pairs_runtime": 9.0564, |
|
"eval_trivia_pairs_samples_per_second": 22.084, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_quora_pairs_loss": 0.9212128520011902, |
|
"eval_quora_pairs_runtime": 0.6293, |
|
"eval_quora_pairs_samples_per_second": 317.839, |
|
"eval_quora_pairs_steps_per_second": 7.946, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_gooaq_pairs_loss": 3.932173490524292, |
|
"eval_gooaq_pairs_runtime": 1.6066, |
|
"eval_gooaq_pairs_samples_per_second": 124.483, |
|
"eval_gooaq_pairs_steps_per_second": 3.112, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15025094102885822, |
|
"eval_mrpc_pairs_loss": 1.0853501558303833, |
|
"eval_mrpc_pairs_runtime": 0.2402, |
|
"eval_mrpc_pairs_samples_per_second": 832.56, |
|
"eval_mrpc_pairs_steps_per_second": 20.814, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 0.15056461731493098, |
|
"grad_norm": 28.37123680114746, |
|
"learning_rate": 6.982434127979924e-06, |
|
"loss": 5.7748, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.15809284818067754, |
|
"grad_norm": 26.708221435546875, |
|
"learning_rate": 7.33375156838143e-06, |
|
"loss": 4.8728, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.1656210790464241, |
|
"grad_norm": 26.786447525024414, |
|
"learning_rate": 7.685069008782934e-06, |
|
"loss": 4.7375, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.17314930991217065, |
|
"grad_norm": 26.215879440307617, |
|
"learning_rate": 8.03638644918444e-06, |
|
"loss": 4.6766, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.1806775407779172, |
|
"grad_norm": 35.618831634521484, |
|
"learning_rate": 8.387703889585947e-06, |
|
"loss": 4.3209, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.18820577164366373, |
|
"grad_norm": 37.166072845458984, |
|
"learning_rate": 8.739021329987453e-06, |
|
"loss": 3.7761, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.19573400250941028, |
|
"grad_norm": 35.78367233276367, |
|
"learning_rate": 9.090338770388957e-06, |
|
"loss": 4.2161, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.20326223337515684, |
|
"grad_norm": 36.299678802490234, |
|
"learning_rate": 9.441656210790464e-06, |
|
"loss": 4.9089, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 0.2107904642409034, |
|
"grad_norm": 18.610933303833008, |
|
"learning_rate": 9.792973651191968e-06, |
|
"loss": 4.3406, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 0.21831869510664995, |
|
"grad_norm": 9.592538833618164, |
|
"learning_rate": 1.0144291091593475e-05, |
|
"loss": 3.5664, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_nli-pairs_loss": 3.2245519161224365, |
|
"eval_nli-pairs_runtime": 4.0436, |
|
"eval_nli-pairs_samples_per_second": 49.461, |
|
"eval_nli-pairs_steps_per_second": 1.237, |
|
"eval_sts-test_pearson_cosine": 0.7002978854888552, |
|
"eval_sts-test_pearson_dot": 0.5685392445320393, |
|
"eval_sts-test_pearson_euclidean": 0.6963744527231541, |
|
"eval_sts-test_pearson_manhattan": 0.7050517306003169, |
|
"eval_sts-test_pearson_max": 0.7050517306003169, |
|
"eval_sts-test_spearman_cosine": 0.6718756239728468, |
|
"eval_sts-test_spearman_dot": 0.5416448961602434, |
|
"eval_sts-test_spearman_euclidean": 0.6742379556154348, |
|
"eval_sts-test_spearman_manhattan": 0.6824201536078427, |
|
"eval_sts-test_spearman_max": 0.6824201536078427, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_vitaminc-pairs_loss": 6.715206623077393, |
|
"eval_vitaminc-pairs_runtime": 1.4251, |
|
"eval_vitaminc-pairs_samples_per_second": 116.481, |
|
"eval_vitaminc-pairs_steps_per_second": 2.807, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_sts-label_loss": 4.016364097595215, |
|
"eval_sts-label_runtime": 0.4049, |
|
"eval_sts-label_samples_per_second": 493.95, |
|
"eval_sts-label_steps_per_second": 12.349, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_qnli-contrastive_loss": 1.999517560005188, |
|
"eval_qnli-contrastive_runtime": 0.2804, |
|
"eval_qnli-contrastive_samples_per_second": 713.282, |
|
"eval_qnli-contrastive_steps_per_second": 17.832, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_scitail-pairs-qa_loss": 1.0403239727020264, |
|
"eval_scitail-pairs-qa_runtime": 1.0483, |
|
"eval_scitail-pairs-qa_samples_per_second": 190.793, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.77, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_scitail-pairs-pos_loss": 1.9232473373413086, |
|
"eval_scitail-pairs-pos_runtime": 2.3447, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.298, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.132, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_xsum-pairs_loss": 1.6821197271347046, |
|
"eval_xsum-pairs_runtime": 1.0422, |
|
"eval_xsum-pairs_samples_per_second": 191.901, |
|
"eval_xsum-pairs_steps_per_second": 4.798, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_compression-pairs_loss": 1.1713249683380127, |
|
"eval_compression-pairs_runtime": 0.2392, |
|
"eval_compression-pairs_samples_per_second": 836.05, |
|
"eval_compression-pairs_steps_per_second": 20.901, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_sciq_pairs_loss": 9.443825721740723, |
|
"eval_sciq_pairs_runtime": 8.9916, |
|
"eval_sciq_pairs_samples_per_second": 22.243, |
|
"eval_sciq_pairs_steps_per_second": 0.556, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_qasc_pairs_loss": 2.9044029712677, |
|
"eval_qasc_pairs_runtime": 1.2182, |
|
"eval_qasc_pairs_samples_per_second": 164.182, |
|
"eval_qasc_pairs_steps_per_second": 4.105, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_openbookqa_pairs_loss": 4.360418796539307, |
|
"eval_openbookqa_pairs_runtime": 1.0522, |
|
"eval_openbookqa_pairs_samples_per_second": 190.077, |
|
"eval_openbookqa_pairs_steps_per_second": 4.752, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_msmarco_pairs_loss": 3.516049861907959, |
|
"eval_msmarco_pairs_runtime": 2.5595, |
|
"eval_msmarco_pairs_samples_per_second": 78.139, |
|
"eval_msmarco_pairs_steps_per_second": 1.953, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_nq_pairs_loss": 4.016308784484863, |
|
"eval_nq_pairs_runtime": 5.6561, |
|
"eval_nq_pairs_samples_per_second": 35.36, |
|
"eval_nq_pairs_steps_per_second": 0.884, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_trivia_pairs_loss": 3.781872272491455, |
|
"eval_trivia_pairs_runtime": 9.0801, |
|
"eval_trivia_pairs_samples_per_second": 22.026, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_quora_pairs_loss": 0.8747495412826538, |
|
"eval_quora_pairs_runtime": 0.6229, |
|
"eval_quora_pairs_samples_per_second": 321.076, |
|
"eval_quora_pairs_steps_per_second": 8.027, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_gooaq_pairs_loss": 3.0769765377044678, |
|
"eval_gooaq_pairs_runtime": 1.552, |
|
"eval_gooaq_pairs_samples_per_second": 128.863, |
|
"eval_gooaq_pairs_steps_per_second": 3.222, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.22537641154328733, |
|
"eval_mrpc_pairs_loss": 0.7370794415473938, |
|
"eval_mrpc_pairs_runtime": 0.2401, |
|
"eval_mrpc_pairs_samples_per_second": 832.867, |
|
"eval_mrpc_pairs_steps_per_second": 20.822, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 0.2258469259723965, |
|
"grad_norm": 29.73522186279297, |
|
"learning_rate": 1.049560853199498e-05, |
|
"loss": 4.7194, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.23337515683814303, |
|
"grad_norm": 30.467117309570312, |
|
"learning_rate": 1.0846925972396486e-05, |
|
"loss": 3.6345, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 0.24090338770388958, |
|
"grad_norm": 24.454021453857422, |
|
"learning_rate": 1.1198243412797992e-05, |
|
"loss": 3.5947, |
|
"step": 1536 |
|
}, |
|
{ |
|
"epoch": 0.24843161856963614, |
|
"grad_norm": 20.165475845336914, |
|
"learning_rate": 1.1549560853199497e-05, |
|
"loss": 4.0526, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 0.2559598494353827, |
|
"grad_norm": 34.79319381713867, |
|
"learning_rate": 1.1900878293601003e-05, |
|
"loss": 3.7962, |
|
"step": 1632 |
|
}, |
|
{ |
|
"epoch": 0.26348808030112925, |
|
"grad_norm": 127.97925567626953, |
|
"learning_rate": 1.2252195734002508e-05, |
|
"loss": 4.1927, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.2710163111668758, |
|
"grad_norm": 27.80243682861328, |
|
"learning_rate": 1.2603513174404014e-05, |
|
"loss": 3.6351, |
|
"step": 1728 |
|
}, |
|
{ |
|
"epoch": 0.27854454203262236, |
|
"grad_norm": 31.81105613708496, |
|
"learning_rate": 1.295483061480552e-05, |
|
"loss": 3.4256, |
|
"step": 1776 |
|
}, |
|
{ |
|
"epoch": 0.2860727728983689, |
|
"grad_norm": 32.932865142822266, |
|
"learning_rate": 1.3306148055207025e-05, |
|
"loss": 3.3175, |
|
"step": 1824 |
|
}, |
|
{ |
|
"epoch": 0.2936010037641154, |
|
"grad_norm": 31.197385787963867, |
|
"learning_rate": 1.365746549560853e-05, |
|
"loss": 3.4984, |
|
"step": 1872 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_nli-pairs_loss": 2.852742910385132, |
|
"eval_nli-pairs_runtime": 4.1529, |
|
"eval_nli-pairs_samples_per_second": 48.16, |
|
"eval_nli-pairs_steps_per_second": 1.204, |
|
"eval_sts-test_pearson_cosine": 0.7132313507241694, |
|
"eval_sts-test_pearson_dot": 0.559846529627866, |
|
"eval_sts-test_pearson_euclidean": 0.7145939583366395, |
|
"eval_sts-test_pearson_manhattan": 0.724552982808093, |
|
"eval_sts-test_pearson_max": 0.724552982808093, |
|
"eval_sts-test_spearman_cosine": 0.6912239915389706, |
|
"eval_sts-test_spearman_dot": 0.5394217029355446, |
|
"eval_sts-test_spearman_euclidean": 0.6946616748545426, |
|
"eval_sts-test_spearman_manhattan": 0.70491424059339, |
|
"eval_sts-test_spearman_max": 0.70491424059339, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_vitaminc-pairs_loss": 6.633151054382324, |
|
"eval_vitaminc-pairs_runtime": 1.4454, |
|
"eval_vitaminc-pairs_samples_per_second": 114.849, |
|
"eval_vitaminc-pairs_steps_per_second": 2.767, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_sts-label_loss": 3.8717281818389893, |
|
"eval_sts-label_runtime": 0.4016, |
|
"eval_sts-label_samples_per_second": 498.049, |
|
"eval_sts-label_steps_per_second": 12.451, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_qnli-contrastive_loss": 1.4170150756835938, |
|
"eval_qnli-contrastive_runtime": 0.2814, |
|
"eval_qnli-contrastive_samples_per_second": 710.85, |
|
"eval_qnli-contrastive_steps_per_second": 17.771, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_scitail-pairs-qa_loss": 0.6900365948677063, |
|
"eval_scitail-pairs-qa_runtime": 1.0611, |
|
"eval_scitail-pairs-qa_samples_per_second": 188.492, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.712, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_scitail-pairs-pos_loss": 1.352358102798462, |
|
"eval_scitail-pairs-pos_runtime": 2.3596, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.761, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.119, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_xsum-pairs_loss": 1.376610517501831, |
|
"eval_xsum-pairs_runtime": 1.0393, |
|
"eval_xsum-pairs_samples_per_second": 192.443, |
|
"eval_xsum-pairs_steps_per_second": 4.811, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_compression-pairs_loss": 0.871735692024231, |
|
"eval_compression-pairs_runtime": 0.2351, |
|
"eval_compression-pairs_samples_per_second": 850.579, |
|
"eval_compression-pairs_steps_per_second": 21.264, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_sciq_pairs_loss": 9.191713333129883, |
|
"eval_sciq_pairs_runtime": 9.1572, |
|
"eval_sciq_pairs_samples_per_second": 21.841, |
|
"eval_sciq_pairs_steps_per_second": 0.546, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_qasc_pairs_loss": 2.369694709777832, |
|
"eval_qasc_pairs_runtime": 1.2239, |
|
"eval_qasc_pairs_samples_per_second": 163.415, |
|
"eval_qasc_pairs_steps_per_second": 4.085, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_openbookqa_pairs_loss": 3.9601967334747314, |
|
"eval_openbookqa_pairs_runtime": 1.0681, |
|
"eval_openbookqa_pairs_samples_per_second": 187.247, |
|
"eval_openbookqa_pairs_steps_per_second": 4.681, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_msmarco_pairs_loss": 3.0808801651000977, |
|
"eval_msmarco_pairs_runtime": 2.5507, |
|
"eval_msmarco_pairs_samples_per_second": 78.409, |
|
"eval_msmarco_pairs_steps_per_second": 1.96, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_nq_pairs_loss": 3.4922549724578857, |
|
"eval_nq_pairs_runtime": 5.7154, |
|
"eval_nq_pairs_samples_per_second": 34.993, |
|
"eval_nq_pairs_steps_per_second": 0.875, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_trivia_pairs_loss": 3.4910638332366943, |
|
"eval_trivia_pairs_runtime": 9.1195, |
|
"eval_trivia_pairs_samples_per_second": 21.931, |
|
"eval_trivia_pairs_steps_per_second": 0.548, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_quora_pairs_loss": 0.833874523639679, |
|
"eval_quora_pairs_runtime": 0.6419, |
|
"eval_quora_pairs_samples_per_second": 311.554, |
|
"eval_quora_pairs_steps_per_second": 7.789, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_gooaq_pairs_loss": 2.622526168823242, |
|
"eval_gooaq_pairs_runtime": 1.5751, |
|
"eval_gooaq_pairs_samples_per_second": 126.977, |
|
"eval_gooaq_pairs_steps_per_second": 3.174, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30050188205771644, |
|
"eval_mrpc_pairs_loss": 0.4888114929199219, |
|
"eval_mrpc_pairs_runtime": 0.2398, |
|
"eval_mrpc_pairs_samples_per_second": 833.994, |
|
"eval_mrpc_pairs_steps_per_second": 20.85, |
|
"step": 1916 |
|
}, |
|
{ |
|
"epoch": 0.30112923462986196, |
|
"grad_norm": 6.916851997375488, |
|
"learning_rate": 1.4008782936010036e-05, |
|
"loss": 2.7233, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.3086574654956085, |
|
"grad_norm": 17.678085327148438, |
|
"learning_rate": 1.4360100376411543e-05, |
|
"loss": 3.6816, |
|
"step": 1968 |
|
}, |
|
{ |
|
"epoch": 0.3161856963613551, |
|
"grad_norm": 92.62138366699219, |
|
"learning_rate": 1.4711417816813047e-05, |
|
"loss": 3.3232, |
|
"step": 2016 |
|
}, |
|
{ |
|
"epoch": 0.3237139272271016, |
|
"grad_norm": 27.2542781829834, |
|
"learning_rate": 1.5062735257214554e-05, |
|
"loss": 3.3469, |
|
"step": 2064 |
|
}, |
|
{ |
|
"epoch": 0.3312421580928482, |
|
"grad_norm": 6.005978584289551, |
|
"learning_rate": 1.5414052697616058e-05, |
|
"loss": 3.7509, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 0.33877038895859474, |
|
"grad_norm": 16.488624572753906, |
|
"learning_rate": 1.5765370138017566e-05, |
|
"loss": 3.1811, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.3462986198243413, |
|
"grad_norm": 11.462204933166504, |
|
"learning_rate": 1.611668757841907e-05, |
|
"loss": 3.3341, |
|
"step": 2208 |
|
}, |
|
{ |
|
"epoch": 0.35382685069008785, |
|
"grad_norm": 33.485206604003906, |
|
"learning_rate": 1.6468005018820577e-05, |
|
"loss": 2.764, |
|
"step": 2256 |
|
}, |
|
{ |
|
"epoch": 0.3613550815558344, |
|
"grad_norm": 25.066240310668945, |
|
"learning_rate": 1.681932245922208e-05, |
|
"loss": 3.6488, |
|
"step": 2304 |
|
}, |
|
{ |
|
"epoch": 0.36888331242158096, |
|
"grad_norm": 28.305265426635742, |
|
"learning_rate": 1.7170639899623588e-05, |
|
"loss": 2.721, |
|
"step": 2352 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_nli-pairs_loss": 2.527458667755127, |
|
"eval_nli-pairs_runtime": 4.1153, |
|
"eval_nli-pairs_samples_per_second": 48.599, |
|
"eval_nli-pairs_steps_per_second": 1.215, |
|
"eval_sts-test_pearson_cosine": 0.7258900302408404, |
|
"eval_sts-test_pearson_dot": 0.5655223839113195, |
|
"eval_sts-test_pearson_euclidean": 0.7228747263710285, |
|
"eval_sts-test_pearson_manhattan": 0.732591374373909, |
|
"eval_sts-test_pearson_max": 0.732591374373909, |
|
"eval_sts-test_spearman_cosine": 0.707910346125958, |
|
"eval_sts-test_spearman_dot": 0.5482635095738919, |
|
"eval_sts-test_spearman_euclidean": 0.7064759533156177, |
|
"eval_sts-test_spearman_manhattan": 0.7166423493246757, |
|
"eval_sts-test_spearman_max": 0.7166423493246757, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_vitaminc-pairs_loss": 6.437549114227295, |
|
"eval_vitaminc-pairs_runtime": 1.4278, |
|
"eval_vitaminc-pairs_samples_per_second": 116.261, |
|
"eval_vitaminc-pairs_steps_per_second": 2.801, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_sts-label_loss": 4.1980671882629395, |
|
"eval_sts-label_runtime": 0.3956, |
|
"eval_sts-label_samples_per_second": 505.555, |
|
"eval_sts-label_steps_per_second": 12.639, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_qnli-contrastive_loss": 1.0682133436203003, |
|
"eval_qnli-contrastive_runtime": 0.2789, |
|
"eval_qnli-contrastive_samples_per_second": 717.152, |
|
"eval_qnli-contrastive_steps_per_second": 17.929, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_scitail-pairs-qa_loss": 0.5046552419662476, |
|
"eval_scitail-pairs-qa_runtime": 1.0451, |
|
"eval_scitail-pairs-qa_samples_per_second": 191.365, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.784, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_scitail-pairs-pos_loss": 1.1998459100723267, |
|
"eval_scitail-pairs-pos_runtime": 2.3442, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.316, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.133, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_xsum-pairs_loss": 1.1817097663879395, |
|
"eval_xsum-pairs_runtime": 1.0372, |
|
"eval_xsum-pairs_samples_per_second": 192.835, |
|
"eval_xsum-pairs_steps_per_second": 4.821, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_compression-pairs_loss": 0.6974765062332153, |
|
"eval_compression-pairs_runtime": 0.2369, |
|
"eval_compression-pairs_samples_per_second": 844.401, |
|
"eval_compression-pairs_steps_per_second": 21.11, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_sciq_pairs_loss": 8.970888137817383, |
|
"eval_sciq_pairs_runtime": 9.0441, |
|
"eval_sciq_pairs_samples_per_second": 22.114, |
|
"eval_sciq_pairs_steps_per_second": 0.553, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_qasc_pairs_loss": 1.9235339164733887, |
|
"eval_qasc_pairs_runtime": 1.2061, |
|
"eval_qasc_pairs_samples_per_second": 165.828, |
|
"eval_qasc_pairs_steps_per_second": 4.146, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_openbookqa_pairs_loss": 3.6225194931030273, |
|
"eval_openbookqa_pairs_runtime": 1.0455, |
|
"eval_openbookqa_pairs_samples_per_second": 191.296, |
|
"eval_openbookqa_pairs_steps_per_second": 4.782, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_msmarco_pairs_loss": 2.664341926574707, |
|
"eval_msmarco_pairs_runtime": 2.5305, |
|
"eval_msmarco_pairs_samples_per_second": 79.036, |
|
"eval_msmarco_pairs_steps_per_second": 1.976, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_nq_pairs_loss": 3.055206298828125, |
|
"eval_nq_pairs_runtime": 5.6527, |
|
"eval_nq_pairs_samples_per_second": 35.381, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_trivia_pairs_loss": 2.9497525691986084, |
|
"eval_trivia_pairs_runtime": 9.0334, |
|
"eval_trivia_pairs_samples_per_second": 22.14, |
|
"eval_trivia_pairs_steps_per_second": 0.554, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_quora_pairs_loss": 0.7771684527397156, |
|
"eval_quora_pairs_runtime": 0.626, |
|
"eval_quora_pairs_samples_per_second": 319.495, |
|
"eval_quora_pairs_steps_per_second": 7.987, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_gooaq_pairs_loss": 2.266879081726074, |
|
"eval_gooaq_pairs_runtime": 1.5425, |
|
"eval_gooaq_pairs_samples_per_second": 129.664, |
|
"eval_gooaq_pairs_steps_per_second": 3.242, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.3756273525721455, |
|
"eval_mrpc_pairs_loss": 0.36913084983825684, |
|
"eval_mrpc_pairs_runtime": 0.2383, |
|
"eval_mrpc_pairs_samples_per_second": 839.153, |
|
"eval_mrpc_pairs_steps_per_second": 20.979, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.37641154328732745, |
|
"grad_norm": 36.60768127441406, |
|
"learning_rate": 1.752195734002509e-05, |
|
"loss": 3.3609, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.383939774153074, |
|
"grad_norm": 24.15782928466797, |
|
"learning_rate": 1.7873274780426595e-05, |
|
"loss": 2.6252, |
|
"step": 2448 |
|
}, |
|
{ |
|
"epoch": 0.39146800501882056, |
|
"grad_norm": 7.509932041168213, |
|
"learning_rate": 1.8224592220828106e-05, |
|
"loss": 3.5142, |
|
"step": 2496 |
|
}, |
|
{ |
|
"epoch": 0.3989962358845671, |
|
"grad_norm": 29.380950927734375, |
|
"learning_rate": 1.857590966122961e-05, |
|
"loss": 4.0597, |
|
"step": 2544 |
|
}, |
|
{ |
|
"epoch": 0.4065244667503137, |
|
"grad_norm": 28.593975067138672, |
|
"learning_rate": 1.8927227101631114e-05, |
|
"loss": 2.8512, |
|
"step": 2592 |
|
}, |
|
{ |
|
"epoch": 0.41405269761606023, |
|
"grad_norm": 21.228628158569336, |
|
"learning_rate": 1.927854454203262e-05, |
|
"loss": 2.717, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.4215809284818068, |
|
"grad_norm": 43.00386047363281, |
|
"learning_rate": 1.962986198243413e-05, |
|
"loss": 3.4717, |
|
"step": 2688 |
|
}, |
|
{ |
|
"epoch": 0.42910915934755334, |
|
"grad_norm": 25.004785537719727, |
|
"learning_rate": 1.9981179422835632e-05, |
|
"loss": 3.1105, |
|
"step": 2736 |
|
}, |
|
{ |
|
"epoch": 0.4366373902132999, |
|
"grad_norm": 7.555154323577881, |
|
"learning_rate": 2.0332496863237136e-05, |
|
"loss": 2.7798, |
|
"step": 2784 |
|
}, |
|
{ |
|
"epoch": 0.44416562107904645, |
|
"grad_norm": 30.839733123779297, |
|
"learning_rate": 2.0683814303638643e-05, |
|
"loss": 3.3606, |
|
"step": 2832 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_nli-pairs_loss": 2.284590721130371, |
|
"eval_nli-pairs_runtime": 4.0714, |
|
"eval_nli-pairs_samples_per_second": 49.123, |
|
"eval_nli-pairs_steps_per_second": 1.228, |
|
"eval_sts-test_pearson_cosine": 0.7382507781851606, |
|
"eval_sts-test_pearson_dot": 0.5710221319397019, |
|
"eval_sts-test_pearson_euclidean": 0.7307583601561211, |
|
"eval_sts-test_pearson_manhattan": 0.7394202696141936, |
|
"eval_sts-test_pearson_max": 0.7394202696141936, |
|
"eval_sts-test_spearman_cosine": 0.7211579109789371, |
|
"eval_sts-test_spearman_dot": 0.5515579746967598, |
|
"eval_sts-test_spearman_euclidean": 0.7142073811971875, |
|
"eval_sts-test_spearman_manhattan": 0.7240537218564107, |
|
"eval_sts-test_spearman_max": 0.7240537218564107, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_vitaminc-pairs_loss": 6.35264253616333, |
|
"eval_vitaminc-pairs_runtime": 1.4349, |
|
"eval_vitaminc-pairs_samples_per_second": 115.687, |
|
"eval_vitaminc-pairs_steps_per_second": 2.788, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_sts-label_loss": 4.186042308807373, |
|
"eval_sts-label_runtime": 0.3983, |
|
"eval_sts-label_samples_per_second": 502.129, |
|
"eval_sts-label_steps_per_second": 12.553, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_qnli-contrastive_loss": 0.781445324420929, |
|
"eval_qnli-contrastive_runtime": 0.2765, |
|
"eval_qnli-contrastive_samples_per_second": 723.448, |
|
"eval_qnli-contrastive_steps_per_second": 18.086, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_scitail-pairs-qa_loss": 0.4217279851436615, |
|
"eval_scitail-pairs-qa_runtime": 1.0438, |
|
"eval_scitail-pairs-qa_samples_per_second": 191.612, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.79, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_scitail-pairs-pos_loss": 1.051362156867981, |
|
"eval_scitail-pairs-pos_runtime": 2.3425, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.379, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.134, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_xsum-pairs_loss": 1.0554753541946411, |
|
"eval_xsum-pairs_runtime": 1.044, |
|
"eval_xsum-pairs_samples_per_second": 191.573, |
|
"eval_xsum-pairs_steps_per_second": 4.789, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_compression-pairs_loss": 0.6035106778144836, |
|
"eval_compression-pairs_runtime": 0.241, |
|
"eval_compression-pairs_samples_per_second": 830.038, |
|
"eval_compression-pairs_steps_per_second": 20.751, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_sciq_pairs_loss": 8.811105728149414, |
|
"eval_sciq_pairs_runtime": 9.0357, |
|
"eval_sciq_pairs_samples_per_second": 22.134, |
|
"eval_sciq_pairs_steps_per_second": 0.553, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_qasc_pairs_loss": 1.615903377532959, |
|
"eval_qasc_pairs_runtime": 1.214, |
|
"eval_qasc_pairs_samples_per_second": 164.746, |
|
"eval_qasc_pairs_steps_per_second": 4.119, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_openbookqa_pairs_loss": 3.4049320220947266, |
|
"eval_openbookqa_pairs_runtime": 1.0554, |
|
"eval_openbookqa_pairs_samples_per_second": 189.509, |
|
"eval_openbookqa_pairs_steps_per_second": 4.738, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_msmarco_pairs_loss": 2.3909060955047607, |
|
"eval_msmarco_pairs_runtime": 2.5301, |
|
"eval_msmarco_pairs_samples_per_second": 79.048, |
|
"eval_msmarco_pairs_steps_per_second": 1.976, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_nq_pairs_loss": 2.794445753097534, |
|
"eval_nq_pairs_runtime": 5.6752, |
|
"eval_nq_pairs_samples_per_second": 35.241, |
|
"eval_nq_pairs_steps_per_second": 0.881, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_trivia_pairs_loss": 2.753361701965332, |
|
"eval_trivia_pairs_runtime": 9.0766, |
|
"eval_trivia_pairs_samples_per_second": 22.035, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_quora_pairs_loss": 0.205492302775383, |
|
"eval_quora_pairs_runtime": 0.6182, |
|
"eval_quora_pairs_samples_per_second": 323.536, |
|
"eval_quora_pairs_steps_per_second": 8.088, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_gooaq_pairs_loss": 2.038878917694092, |
|
"eval_gooaq_pairs_runtime": 1.5488, |
|
"eval_gooaq_pairs_samples_per_second": 129.134, |
|
"eval_gooaq_pairs_steps_per_second": 3.228, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.45075282308657466, |
|
"eval_mrpc_pairs_loss": 0.3230588436126709, |
|
"eval_mrpc_pairs_runtime": 0.2358, |
|
"eval_mrpc_pairs_samples_per_second": 848.229, |
|
"eval_mrpc_pairs_steps_per_second": 21.206, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 0.451693851944793, |
|
"grad_norm": 5.271574020385742, |
|
"learning_rate": 2.103513174404015e-05, |
|
"loss": 2.6918, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.4592220828105395, |
|
"grad_norm": 21.954103469848633, |
|
"learning_rate": 2.1386449184441654e-05, |
|
"loss": 2.8354, |
|
"step": 2928 |
|
}, |
|
{ |
|
"epoch": 0.46675031367628605, |
|
"grad_norm": 28.671293258666992, |
|
"learning_rate": 2.173776662484316e-05, |
|
"loss": 2.9499, |
|
"step": 2976 |
|
}, |
|
{ |
|
"epoch": 0.4742785445420326, |
|
"grad_norm": 26.562397003173828, |
|
"learning_rate": 2.2089084065244666e-05, |
|
"loss": 2.6211, |
|
"step": 3024 |
|
}, |
|
{ |
|
"epoch": 0.48180677540777916, |
|
"grad_norm": 48.511756896972656, |
|
"learning_rate": 2.2440401505646173e-05, |
|
"loss": 3.3356, |
|
"step": 3072 |
|
}, |
|
{ |
|
"epoch": 0.4893350062735257, |
|
"grad_norm": 46.71563720703125, |
|
"learning_rate": 2.2791718946047677e-05, |
|
"loss": 2.846, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.4968632371392723, |
|
"grad_norm": 24.524322509765625, |
|
"learning_rate": 2.3143036386449184e-05, |
|
"loss": 2.4866, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 0.5043914680050188, |
|
"grad_norm": 85.22843933105469, |
|
"learning_rate": 2.3494353826850688e-05, |
|
"loss": 2.6334, |
|
"step": 3216 |
|
}, |
|
{ |
|
"epoch": 0.5119196988707654, |
|
"grad_norm": 28.435443878173828, |
|
"learning_rate": 2.384567126725219e-05, |
|
"loss": 2.6118, |
|
"step": 3264 |
|
}, |
|
{ |
|
"epoch": 0.5194479297365119, |
|
"grad_norm": 21.590103149414062, |
|
"learning_rate": 2.41969887076537e-05, |
|
"loss": 2.5833, |
|
"step": 3312 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_nli-pairs_loss": 2.0752949714660645, |
|
"eval_nli-pairs_runtime": 4.0304, |
|
"eval_nli-pairs_samples_per_second": 49.623, |
|
"eval_nli-pairs_steps_per_second": 1.241, |
|
"eval_sts-test_pearson_cosine": 0.7401847199967786, |
|
"eval_sts-test_pearson_dot": 0.5441501995975192, |
|
"eval_sts-test_pearson_euclidean": 0.7344996320188322, |
|
"eval_sts-test_pearson_manhattan": 0.7394640598472787, |
|
"eval_sts-test_pearson_max": 0.7401847199967786, |
|
"eval_sts-test_spearman_cosine": 0.7300085598018916, |
|
"eval_sts-test_spearman_dot": 0.5241747185593542, |
|
"eval_sts-test_spearman_euclidean": 0.7194131601167465, |
|
"eval_sts-test_spearman_manhattan": 0.726961581928453, |
|
"eval_sts-test_spearman_max": 0.7300085598018916, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_vitaminc-pairs_loss": 6.441956996917725, |
|
"eval_vitaminc-pairs_runtime": 1.4416, |
|
"eval_vitaminc-pairs_samples_per_second": 115.149, |
|
"eval_vitaminc-pairs_steps_per_second": 2.775, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_sts-label_loss": 4.200085639953613, |
|
"eval_sts-label_runtime": 0.3949, |
|
"eval_sts-label_samples_per_second": 506.436, |
|
"eval_sts-label_steps_per_second": 12.661, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_qnli-contrastive_loss": 0.5195684432983398, |
|
"eval_qnli-contrastive_runtime": 0.2809, |
|
"eval_qnli-contrastive_samples_per_second": 712.107, |
|
"eval_qnli-contrastive_steps_per_second": 17.803, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_scitail-pairs-qa_loss": 0.35189124941825867, |
|
"eval_scitail-pairs-qa_runtime": 1.0578, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.064, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.727, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_scitail-pairs-pos_loss": 0.8873756527900696, |
|
"eval_scitail-pairs-pos_runtime": 2.4029, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.232, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.081, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_xsum-pairs_loss": 0.939339280128479, |
|
"eval_xsum-pairs_runtime": 1.041, |
|
"eval_xsum-pairs_samples_per_second": 192.121, |
|
"eval_xsum-pairs_steps_per_second": 4.803, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_compression-pairs_loss": 0.5007131695747375, |
|
"eval_compression-pairs_runtime": 0.2338, |
|
"eval_compression-pairs_samples_per_second": 855.479, |
|
"eval_compression-pairs_steps_per_second": 21.387, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_sciq_pairs_loss": 8.558987617492676, |
|
"eval_sciq_pairs_runtime": 9.0984, |
|
"eval_sciq_pairs_samples_per_second": 21.982, |
|
"eval_sciq_pairs_steps_per_second": 0.55, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_qasc_pairs_loss": 1.4318852424621582, |
|
"eval_qasc_pairs_runtime": 1.2286, |
|
"eval_qasc_pairs_samples_per_second": 162.79, |
|
"eval_qasc_pairs_steps_per_second": 4.07, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_openbookqa_pairs_loss": 3.1973114013671875, |
|
"eval_openbookqa_pairs_runtime": 1.0491, |
|
"eval_openbookqa_pairs_samples_per_second": 190.633, |
|
"eval_openbookqa_pairs_steps_per_second": 4.766, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_msmarco_pairs_loss": 2.2080254554748535, |
|
"eval_msmarco_pairs_runtime": 2.5223, |
|
"eval_msmarco_pairs_samples_per_second": 79.294, |
|
"eval_msmarco_pairs_steps_per_second": 1.982, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_nq_pairs_loss": 2.5810558795928955, |
|
"eval_nq_pairs_runtime": 5.6341, |
|
"eval_nq_pairs_samples_per_second": 35.498, |
|
"eval_nq_pairs_steps_per_second": 0.887, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_trivia_pairs_loss": 2.655771255493164, |
|
"eval_trivia_pairs_runtime": 9.0716, |
|
"eval_trivia_pairs_samples_per_second": 22.047, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_quora_pairs_loss": 0.5028819441795349, |
|
"eval_quora_pairs_runtime": 0.6144, |
|
"eval_quora_pairs_samples_per_second": 325.522, |
|
"eval_quora_pairs_steps_per_second": 8.138, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_gooaq_pairs_loss": 1.8867437839508057, |
|
"eval_gooaq_pairs_runtime": 1.5505, |
|
"eval_gooaq_pairs_samples_per_second": 128.994, |
|
"eval_gooaq_pairs_steps_per_second": 3.225, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5258782936010038, |
|
"eval_mrpc_pairs_loss": 0.2580638825893402, |
|
"eval_mrpc_pairs_runtime": 0.2364, |
|
"eval_mrpc_pairs_samples_per_second": 846.008, |
|
"eval_mrpc_pairs_steps_per_second": 21.15, |
|
"step": 3353 |
|
}, |
|
{ |
|
"epoch": 0.5269761606022585, |
|
"grad_norm": 2.6962711811065674, |
|
"learning_rate": 2.4548306148055206e-05, |
|
"loss": 2.3251, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.534504391468005, |
|
"grad_norm": 35.47948455810547, |
|
"learning_rate": 2.489962358845671e-05, |
|
"loss": 2.8494, |
|
"step": 3408 |
|
}, |
|
{ |
|
"epoch": 0.5420326223337516, |
|
"grad_norm": 8.13453483581543, |
|
"learning_rate": 2.5250941028858214e-05, |
|
"loss": 2.4009, |
|
"step": 3456 |
|
}, |
|
{ |
|
"epoch": 0.5495608531994981, |
|
"grad_norm": 20.041057586669922, |
|
"learning_rate": 2.560225846925972e-05, |
|
"loss": 2.5952, |
|
"step": 3504 |
|
}, |
|
{ |
|
"epoch": 0.5570890840652447, |
|
"grad_norm": 23.942073822021484, |
|
"learning_rate": 2.595357590966123e-05, |
|
"loss": 2.2798, |
|
"step": 3552 |
|
}, |
|
{ |
|
"epoch": 0.5646173149309912, |
|
"grad_norm": 17.675006866455078, |
|
"learning_rate": 2.6304893350062732e-05, |
|
"loss": 2.308, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.5721455457967378, |
|
"grad_norm": 24.20000457763672, |
|
"learning_rate": 2.6656210790464236e-05, |
|
"loss": 2.122, |
|
"step": 3648 |
|
}, |
|
{ |
|
"epoch": 0.5796737766624843, |
|
"grad_norm": 30.06256866455078, |
|
"learning_rate": 2.7007528230865747e-05, |
|
"loss": 2.7901, |
|
"step": 3696 |
|
}, |
|
{ |
|
"epoch": 0.5872020075282308, |
|
"grad_norm": 22.547115325927734, |
|
"learning_rate": 2.735884567126725e-05, |
|
"loss": 2.0671, |
|
"step": 3744 |
|
}, |
|
{ |
|
"epoch": 0.5947302383939774, |
|
"grad_norm": 34.11716079711914, |
|
"learning_rate": 2.7710163111668754e-05, |
|
"loss": 2.366, |
|
"step": 3792 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_nli-pairs_loss": 1.8900150060653687, |
|
"eval_nli-pairs_runtime": 4.0481, |
|
"eval_nli-pairs_samples_per_second": 49.406, |
|
"eval_nli-pairs_steps_per_second": 1.235, |
|
"eval_sts-test_pearson_cosine": 0.752143976340549, |
|
"eval_sts-test_pearson_dot": 0.5694102087200895, |
|
"eval_sts-test_pearson_euclidean": 0.7457585181878474, |
|
"eval_sts-test_pearson_manhattan": 0.7525316002813096, |
|
"eval_sts-test_pearson_max": 0.7525316002813096, |
|
"eval_sts-test_spearman_cosine": 0.7404216272264129, |
|
"eval_sts-test_spearman_dot": 0.5485789739808921, |
|
"eval_sts-test_spearman_euclidean": 0.728675089641457, |
|
"eval_sts-test_spearman_manhattan": 0.7367562035227414, |
|
"eval_sts-test_spearman_max": 0.7404216272264129, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_vitaminc-pairs_loss": 6.0831098556518555, |
|
"eval_vitaminc-pairs_runtime": 1.4528, |
|
"eval_vitaminc-pairs_samples_per_second": 114.264, |
|
"eval_vitaminc-pairs_steps_per_second": 2.753, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_sts-label_loss": 4.197264671325684, |
|
"eval_sts-label_runtime": 0.4176, |
|
"eval_sts-label_samples_per_second": 478.893, |
|
"eval_sts-label_steps_per_second": 11.972, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_qnli-contrastive_loss": 0.5115653872489929, |
|
"eval_qnli-contrastive_runtime": 0.3027, |
|
"eval_qnli-contrastive_samples_per_second": 660.784, |
|
"eval_qnli-contrastive_steps_per_second": 16.52, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_scitail-pairs-qa_loss": 0.29788386821746826, |
|
"eval_scitail-pairs-qa_runtime": 1.0654, |
|
"eval_scitail-pairs-qa_samples_per_second": 187.719, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.693, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_scitail-pairs-pos_loss": 0.8727617859840393, |
|
"eval_scitail-pairs-pos_runtime": 2.3677, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.471, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.112, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_xsum-pairs_loss": 0.8608022928237915, |
|
"eval_xsum-pairs_runtime": 1.0435, |
|
"eval_xsum-pairs_samples_per_second": 191.671, |
|
"eval_xsum-pairs_steps_per_second": 4.792, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_compression-pairs_loss": 0.4411359429359436, |
|
"eval_compression-pairs_runtime": 0.2345, |
|
"eval_compression-pairs_samples_per_second": 852.821, |
|
"eval_compression-pairs_steps_per_second": 21.321, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_sciq_pairs_loss": 8.294719696044922, |
|
"eval_sciq_pairs_runtime": 9.2141, |
|
"eval_sciq_pairs_samples_per_second": 21.706, |
|
"eval_sciq_pairs_steps_per_second": 0.543, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_qasc_pairs_loss": 1.1894803047180176, |
|
"eval_qasc_pairs_runtime": 1.2518, |
|
"eval_qasc_pairs_samples_per_second": 159.774, |
|
"eval_qasc_pairs_steps_per_second": 3.994, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_openbookqa_pairs_loss": 2.8579885959625244, |
|
"eval_openbookqa_pairs_runtime": 1.0874, |
|
"eval_openbookqa_pairs_samples_per_second": 183.92, |
|
"eval_openbookqa_pairs_steps_per_second": 4.598, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_msmarco_pairs_loss": 1.9733755588531494, |
|
"eval_msmarco_pairs_runtime": 2.5486, |
|
"eval_msmarco_pairs_samples_per_second": 78.476, |
|
"eval_msmarco_pairs_steps_per_second": 1.962, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_nq_pairs_loss": 2.206907033920288, |
|
"eval_nq_pairs_runtime": 5.7528, |
|
"eval_nq_pairs_samples_per_second": 34.766, |
|
"eval_nq_pairs_steps_per_second": 0.869, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_trivia_pairs_loss": 2.332620620727539, |
|
"eval_trivia_pairs_runtime": 9.1703, |
|
"eval_trivia_pairs_samples_per_second": 21.809, |
|
"eval_trivia_pairs_steps_per_second": 0.545, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_quora_pairs_loss": 0.48870089650154114, |
|
"eval_quora_pairs_runtime": 0.6491, |
|
"eval_quora_pairs_samples_per_second": 308.142, |
|
"eval_quora_pairs_steps_per_second": 7.704, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_gooaq_pairs_loss": 1.598087191581726, |
|
"eval_gooaq_pairs_runtime": 1.5759, |
|
"eval_gooaq_pairs_samples_per_second": 126.912, |
|
"eval_gooaq_pairs_steps_per_second": 3.173, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6010037641154329, |
|
"eval_mrpc_pairs_loss": 0.2343733161687851, |
|
"eval_mrpc_pairs_runtime": 0.2484, |
|
"eval_mrpc_pairs_samples_per_second": 805.097, |
|
"eval_mrpc_pairs_steps_per_second": 20.127, |
|
"step": 3832 |
|
}, |
|
{ |
|
"epoch": 0.6022584692597239, |
|
"grad_norm": 1.486786127090454, |
|
"learning_rate": 2.806148055207026e-05, |
|
"loss": 1.9614, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.6097867001254705, |
|
"grad_norm": 23.297300338745117, |
|
"learning_rate": 2.841279799247177e-05, |
|
"loss": 2.3589, |
|
"step": 3888 |
|
}, |
|
{ |
|
"epoch": 0.617314930991217, |
|
"grad_norm": 16.00516700744629, |
|
"learning_rate": 2.8764115432873273e-05, |
|
"loss": 2.1475, |
|
"step": 3936 |
|
}, |
|
{ |
|
"epoch": 0.6248431618569636, |
|
"grad_norm": 24.357616424560547, |
|
"learning_rate": 2.9115432873274777e-05, |
|
"loss": 2.1312, |
|
"step": 3984 |
|
}, |
|
{ |
|
"epoch": 0.6323713927227101, |
|
"grad_norm": 28.798917770385742, |
|
"learning_rate": 2.946675031367628e-05, |
|
"loss": 2.5716, |
|
"step": 4032 |
|
}, |
|
{ |
|
"epoch": 0.6398996235884568, |
|
"grad_norm": 18.239490509033203, |
|
"learning_rate": 2.981806775407779e-05, |
|
"loss": 2.2249, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.6474278544542033, |
|
"grad_norm": 19.50409507751465, |
|
"learning_rate": 3.0169385194479295e-05, |
|
"loss": 2.6331, |
|
"step": 4128 |
|
}, |
|
{ |
|
"epoch": 0.6549560853199499, |
|
"grad_norm": 12.110575675964355, |
|
"learning_rate": 3.05207026348808e-05, |
|
"loss": 2.7637, |
|
"step": 4176 |
|
}, |
|
{ |
|
"epoch": 0.6624843161856964, |
|
"grad_norm": 6.904999256134033, |
|
"learning_rate": 3.087202007528231e-05, |
|
"loss": 1.8973, |
|
"step": 4224 |
|
}, |
|
{ |
|
"epoch": 0.6700125470514429, |
|
"grad_norm": 9.007365226745605, |
|
"learning_rate": 3.1223337515683813e-05, |
|
"loss": 2.3181, |
|
"step": 4272 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_nli-pairs_loss": 1.7111084461212158, |
|
"eval_nli-pairs_runtime": 4.0305, |
|
"eval_nli-pairs_samples_per_second": 49.622, |
|
"eval_nli-pairs_steps_per_second": 1.241, |
|
"eval_sts-test_pearson_cosine": 0.7375865838793885, |
|
"eval_sts-test_pearson_dot": 0.5355907015359193, |
|
"eval_sts-test_pearson_euclidean": 0.7266850031847317, |
|
"eval_sts-test_pearson_manhattan": 0.7357621558005936, |
|
"eval_sts-test_pearson_max": 0.7375865838793885, |
|
"eval_sts-test_spearman_cosine": 0.7273524041973777, |
|
"eval_sts-test_spearman_dot": 0.5084902224306463, |
|
"eval_sts-test_spearman_euclidean": 0.7071419579928555, |
|
"eval_sts-test_spearman_manhattan": 0.7177664681655631, |
|
"eval_sts-test_spearman_max": 0.7273524041973777, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_vitaminc-pairs_loss": 6.216845989227295, |
|
"eval_vitaminc-pairs_runtime": 1.4703, |
|
"eval_vitaminc-pairs_samples_per_second": 112.902, |
|
"eval_vitaminc-pairs_steps_per_second": 2.721, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_sts-label_loss": 4.384557723999023, |
|
"eval_sts-label_runtime": 0.3912, |
|
"eval_sts-label_samples_per_second": 511.22, |
|
"eval_sts-label_steps_per_second": 12.78, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_qnli-contrastive_loss": 0.40437957644462585, |
|
"eval_qnli-contrastive_runtime": 0.281, |
|
"eval_qnli-contrastive_samples_per_second": 711.864, |
|
"eval_qnli-contrastive_steps_per_second": 17.797, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_scitail-pairs-qa_loss": 0.2210184931755066, |
|
"eval_scitail-pairs-qa_runtime": 1.0575, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.117, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.728, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_scitail-pairs-pos_loss": 0.9065079689025879, |
|
"eval_scitail-pairs-pos_runtime": 2.3488, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.151, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.129, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_xsum-pairs_loss": 0.8169436454772949, |
|
"eval_xsum-pairs_runtime": 1.0409, |
|
"eval_xsum-pairs_samples_per_second": 192.145, |
|
"eval_xsum-pairs_steps_per_second": 4.804, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_compression-pairs_loss": 0.391815721988678, |
|
"eval_compression-pairs_runtime": 0.2361, |
|
"eval_compression-pairs_samples_per_second": 847.066, |
|
"eval_compression-pairs_steps_per_second": 21.177, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_sciq_pairs_loss": 0.6230970025062561, |
|
"eval_sciq_pairs_runtime": 9.0874, |
|
"eval_sciq_pairs_samples_per_second": 22.008, |
|
"eval_sciq_pairs_steps_per_second": 0.55, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_qasc_pairs_loss": 1.1559942960739136, |
|
"eval_qasc_pairs_runtime": 1.2507, |
|
"eval_qasc_pairs_samples_per_second": 159.907, |
|
"eval_qasc_pairs_steps_per_second": 3.998, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_openbookqa_pairs_loss": 2.8303356170654297, |
|
"eval_openbookqa_pairs_runtime": 1.0524, |
|
"eval_openbookqa_pairs_samples_per_second": 190.041, |
|
"eval_openbookqa_pairs_steps_per_second": 4.751, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_msmarco_pairs_loss": 1.9672399759292603, |
|
"eval_msmarco_pairs_runtime": 2.518, |
|
"eval_msmarco_pairs_samples_per_second": 79.428, |
|
"eval_msmarco_pairs_steps_per_second": 1.986, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_nq_pairs_loss": 2.169950008392334, |
|
"eval_nq_pairs_runtime": 5.6541, |
|
"eval_nq_pairs_samples_per_second": 35.372, |
|
"eval_nq_pairs_steps_per_second": 0.884, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_trivia_pairs_loss": 2.198312520980835, |
|
"eval_trivia_pairs_runtime": 9.0535, |
|
"eval_trivia_pairs_samples_per_second": 22.091, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_quora_pairs_loss": 0.3780948519706726, |
|
"eval_quora_pairs_runtime": 0.6375, |
|
"eval_quora_pairs_samples_per_second": 313.737, |
|
"eval_quora_pairs_steps_per_second": 7.843, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_gooaq_pairs_loss": 1.5646275281906128, |
|
"eval_gooaq_pairs_runtime": 1.5531, |
|
"eval_gooaq_pairs_samples_per_second": 128.772, |
|
"eval_gooaq_pairs_steps_per_second": 3.219, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.676129234629862, |
|
"eval_mrpc_pairs_loss": 0.18545588850975037, |
|
"eval_mrpc_pairs_runtime": 0.2382, |
|
"eval_mrpc_pairs_samples_per_second": 839.569, |
|
"eval_mrpc_pairs_steps_per_second": 20.989, |
|
"step": 4311 |
|
}, |
|
{ |
|
"epoch": 0.6775407779171895, |
|
"grad_norm": 15.639892578125, |
|
"learning_rate": 3.157465495608532e-05, |
|
"loss": 2.2329, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.685069008782936, |
|
"grad_norm": 7.185269832611084, |
|
"learning_rate": 3.192597239648682e-05, |
|
"loss": 2.7864, |
|
"step": 4368 |
|
}, |
|
{ |
|
"epoch": 0.6925972396486826, |
|
"grad_norm": 10.165898323059082, |
|
"learning_rate": 3.227728983688833e-05, |
|
"loss": 2.5277, |
|
"step": 4416 |
|
}, |
|
{ |
|
"epoch": 0.7001254705144291, |
|
"grad_norm": 20.132612228393555, |
|
"learning_rate": 3.2628607277289836e-05, |
|
"loss": 2.526, |
|
"step": 4464 |
|
}, |
|
{ |
|
"epoch": 0.7076537013801757, |
|
"grad_norm": 6.147126197814941, |
|
"learning_rate": 3.297992471769134e-05, |
|
"loss": 1.5993, |
|
"step": 4512 |
|
}, |
|
{ |
|
"epoch": 0.7151819322459222, |
|
"grad_norm": 9.857342720031738, |
|
"learning_rate": 3.3331242158092843e-05, |
|
"loss": 2.0452, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.7227101631116688, |
|
"grad_norm": 22.434364318847656, |
|
"learning_rate": 3.3682559598494354e-05, |
|
"loss": 2.0458, |
|
"step": 4608 |
|
}, |
|
{ |
|
"epoch": 0.7302383939774153, |
|
"grad_norm": 22.420066833496094, |
|
"learning_rate": 3.403387703889586e-05, |
|
"loss": 1.854, |
|
"step": 4656 |
|
}, |
|
{ |
|
"epoch": 0.7377666248431619, |
|
"grad_norm": 24.213205337524414, |
|
"learning_rate": 3.438519447929736e-05, |
|
"loss": 2.31, |
|
"step": 4704 |
|
}, |
|
{ |
|
"epoch": 0.7452948557089084, |
|
"grad_norm": 2.1184492111206055, |
|
"learning_rate": 3.473651191969887e-05, |
|
"loss": 1.7484, |
|
"step": 4752 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_nli-pairs_loss": 1.5204579830169678, |
|
"eval_nli-pairs_runtime": 4.3227, |
|
"eval_nli-pairs_samples_per_second": 46.267, |
|
"eval_nli-pairs_steps_per_second": 1.157, |
|
"eval_sts-test_pearson_cosine": 0.753550468294361, |
|
"eval_sts-test_pearson_dot": 0.576164453162354, |
|
"eval_sts-test_pearson_euclidean": 0.7433413992355353, |
|
"eval_sts-test_pearson_manhattan": 0.7483177470711824, |
|
"eval_sts-test_pearson_max": 0.753550468294361, |
|
"eval_sts-test_spearman_cosine": 0.7510075785449373, |
|
"eval_sts-test_spearman_dot": 0.5438417987754244, |
|
"eval_sts-test_spearman_euclidean": 0.7271758422639625, |
|
"eval_sts-test_spearman_manhattan": 0.7334567781451864, |
|
"eval_sts-test_spearman_max": 0.7510075785449373, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_vitaminc-pairs_loss": 5.992164134979248, |
|
"eval_vitaminc-pairs_runtime": 1.4716, |
|
"eval_vitaminc-pairs_samples_per_second": 112.802, |
|
"eval_vitaminc-pairs_steps_per_second": 2.718, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_sts-label_loss": 4.175446033477783, |
|
"eval_sts-label_runtime": 0.3991, |
|
"eval_sts-label_samples_per_second": 501.105, |
|
"eval_sts-label_steps_per_second": 12.528, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_qnli-contrastive_loss": 0.4543713629245758, |
|
"eval_qnli-contrastive_runtime": 0.2765, |
|
"eval_qnli-contrastive_samples_per_second": 723.231, |
|
"eval_qnli-contrastive_steps_per_second": 18.081, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_scitail-pairs-qa_loss": 0.20784999430179596, |
|
"eval_scitail-pairs-qa_runtime": 1.0567, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.26, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.732, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_scitail-pairs-pos_loss": 0.8679056167602539, |
|
"eval_scitail-pairs-pos_runtime": 2.3456, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.268, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.132, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_xsum-pairs_loss": 0.6944636702537537, |
|
"eval_xsum-pairs_runtime": 1.0379, |
|
"eval_xsum-pairs_samples_per_second": 192.694, |
|
"eval_xsum-pairs_steps_per_second": 4.817, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_compression-pairs_loss": 0.34194332361221313, |
|
"eval_compression-pairs_runtime": 0.2354, |
|
"eval_compression-pairs_samples_per_second": 849.798, |
|
"eval_compression-pairs_steps_per_second": 21.245, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_sciq_pairs_loss": 7.352969646453857, |
|
"eval_sciq_pairs_runtime": 9.0502, |
|
"eval_sciq_pairs_samples_per_second": 22.099, |
|
"eval_sciq_pairs_steps_per_second": 0.552, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_qasc_pairs_loss": 0.9425787329673767, |
|
"eval_qasc_pairs_runtime": 1.2081, |
|
"eval_qasc_pairs_samples_per_second": 165.543, |
|
"eval_qasc_pairs_steps_per_second": 4.139, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_openbookqa_pairs_loss": 2.5082011222839355, |
|
"eval_openbookqa_pairs_runtime": 1.0492, |
|
"eval_openbookqa_pairs_samples_per_second": 190.618, |
|
"eval_openbookqa_pairs_steps_per_second": 4.765, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_msmarco_pairs_loss": 1.696744441986084, |
|
"eval_msmarco_pairs_runtime": 2.5162, |
|
"eval_msmarco_pairs_samples_per_second": 79.484, |
|
"eval_msmarco_pairs_steps_per_second": 1.987, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_nq_pairs_loss": 1.8095602989196777, |
|
"eval_nq_pairs_runtime": 5.6348, |
|
"eval_nq_pairs_samples_per_second": 35.494, |
|
"eval_nq_pairs_steps_per_second": 0.887, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_trivia_pairs_loss": 2.0272486209869385, |
|
"eval_trivia_pairs_runtime": 9.0671, |
|
"eval_trivia_pairs_samples_per_second": 22.058, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_quora_pairs_loss": 0.15705542266368866, |
|
"eval_quora_pairs_runtime": 0.6326, |
|
"eval_quora_pairs_samples_per_second": 316.138, |
|
"eval_quora_pairs_steps_per_second": 7.903, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_gooaq_pairs_loss": 1.374332308769226, |
|
"eval_gooaq_pairs_runtime": 1.5484, |
|
"eval_gooaq_pairs_samples_per_second": 129.164, |
|
"eval_gooaq_pairs_steps_per_second": 3.229, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.751254705144291, |
|
"eval_mrpc_pairs_loss": 0.17204828560352325, |
|
"eval_mrpc_pairs_runtime": 0.2358, |
|
"eval_mrpc_pairs_samples_per_second": 848.332, |
|
"eval_mrpc_pairs_steps_per_second": 21.208, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.7528230865746549, |
|
"grad_norm": 1.4021190404891968, |
|
"learning_rate": 3.4999758305020584e-05, |
|
"loss": 1.7113, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.7603513174404015, |
|
"grad_norm": 19.776817321777344, |
|
"learning_rate": 3.499395795931671e-05, |
|
"loss": 2.3696, |
|
"step": 4848 |
|
}, |
|
{ |
|
"epoch": 0.767879548306148, |
|
"grad_norm": 28.693845748901367, |
|
"learning_rate": 3.49804263115427e-05, |
|
"loss": 2.2947, |
|
"step": 4896 |
|
}, |
|
{ |
|
"epoch": 0.7754077791718946, |
|
"grad_norm": 1.3631008863449097, |
|
"learning_rate": 3.495916934189221e-05, |
|
"loss": 1.8841, |
|
"step": 4944 |
|
}, |
|
{ |
|
"epoch": 0.7829360100376411, |
|
"grad_norm": 40.640262603759766, |
|
"learning_rate": 3.4930196444697477e-05, |
|
"loss": 2.084, |
|
"step": 4992 |
|
}, |
|
{ |
|
"epoch": 0.7904642409033877, |
|
"grad_norm": 20.45759391784668, |
|
"learning_rate": 3.489352042427762e-05, |
|
"loss": 1.9297, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.7979924717691342, |
|
"grad_norm": 24.276058197021484, |
|
"learning_rate": 3.484915748927982e-05, |
|
"loss": 2.0521, |
|
"step": 5088 |
|
}, |
|
{ |
|
"epoch": 0.8055207026348808, |
|
"grad_norm": 24.93791389465332, |
|
"learning_rate": 3.4797127245516105e-05, |
|
"loss": 1.7092, |
|
"step": 5136 |
|
}, |
|
{ |
|
"epoch": 0.8130489335006273, |
|
"grad_norm": 25.131153106689453, |
|
"learning_rate": 3.4737452687298694e-05, |
|
"loss": 1.7394, |
|
"step": 5184 |
|
}, |
|
{ |
|
"epoch": 0.820577164366374, |
|
"grad_norm": 3.779459238052368, |
|
"learning_rate": 3.467016018727788e-05, |
|
"loss": 2.567, |
|
"step": 5232 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_nli-pairs_loss": 1.4584167003631592, |
|
"eval_nli-pairs_runtime": 3.9955, |
|
"eval_nli-pairs_samples_per_second": 50.056, |
|
"eval_nli-pairs_steps_per_second": 1.251, |
|
"eval_sts-test_pearson_cosine": 0.7484577894142428, |
|
"eval_sts-test_pearson_dot": 0.5289676422936789, |
|
"eval_sts-test_pearson_euclidean": 0.743677607180833, |
|
"eval_sts-test_pearson_manhattan": 0.7474581577502462, |
|
"eval_sts-test_pearson_max": 0.7484577894142428, |
|
"eval_sts-test_spearman_cosine": 0.7507798204197761, |
|
"eval_sts-test_spearman_dot": 0.5016451185199292, |
|
"eval_sts-test_spearman_euclidean": 0.7307379850546868, |
|
"eval_sts-test_spearman_manhattan": 0.7367432097081014, |
|
"eval_sts-test_spearman_max": 0.7507798204197761, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_vitaminc-pairs_loss": 5.56383752822876, |
|
"eval_vitaminc-pairs_runtime": 1.4339, |
|
"eval_vitaminc-pairs_samples_per_second": 115.772, |
|
"eval_vitaminc-pairs_steps_per_second": 2.79, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_sts-label_loss": 4.355674743652344, |
|
"eval_sts-label_runtime": 0.401, |
|
"eval_sts-label_samples_per_second": 498.723, |
|
"eval_sts-label_steps_per_second": 12.468, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_qnli-contrastive_loss": 0.2912294566631317, |
|
"eval_qnli-contrastive_runtime": 0.2801, |
|
"eval_qnli-contrastive_samples_per_second": 714.044, |
|
"eval_qnli-contrastive_steps_per_second": 17.851, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_scitail-pairs-qa_loss": 0.19145721197128296, |
|
"eval_scitail-pairs-qa_runtime": 1.0728, |
|
"eval_scitail-pairs-qa_samples_per_second": 186.429, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.661, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_scitail-pairs-pos_loss": 0.7433645725250244, |
|
"eval_scitail-pairs-pos_runtime": 2.3785, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.086, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.102, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_xsum-pairs_loss": 0.6145637631416321, |
|
"eval_xsum-pairs_runtime": 1.0393, |
|
"eval_xsum-pairs_samples_per_second": 192.428, |
|
"eval_xsum-pairs_steps_per_second": 4.811, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_compression-pairs_loss": 0.29557526111602783, |
|
"eval_compression-pairs_runtime": 0.2372, |
|
"eval_compression-pairs_samples_per_second": 843.266, |
|
"eval_compression-pairs_steps_per_second": 21.082, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_sciq_pairs_loss": 0.5607883930206299, |
|
"eval_sciq_pairs_runtime": 9.0381, |
|
"eval_sciq_pairs_samples_per_second": 22.128, |
|
"eval_sciq_pairs_steps_per_second": 0.553, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_qasc_pairs_loss": 0.8776007294654846, |
|
"eval_qasc_pairs_runtime": 1.2111, |
|
"eval_qasc_pairs_samples_per_second": 165.141, |
|
"eval_qasc_pairs_steps_per_second": 4.129, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_openbookqa_pairs_loss": 2.414658784866333, |
|
"eval_openbookqa_pairs_runtime": 1.052, |
|
"eval_openbookqa_pairs_samples_per_second": 190.106, |
|
"eval_openbookqa_pairs_steps_per_second": 4.753, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_msmarco_pairs_loss": 1.615893840789795, |
|
"eval_msmarco_pairs_runtime": 2.5183, |
|
"eval_msmarco_pairs_samples_per_second": 79.419, |
|
"eval_msmarco_pairs_steps_per_second": 1.985, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_nq_pairs_loss": 1.5788501501083374, |
|
"eval_nq_pairs_runtime": 5.6429, |
|
"eval_nq_pairs_samples_per_second": 35.443, |
|
"eval_nq_pairs_steps_per_second": 0.886, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_trivia_pairs_loss": 1.8346160650253296, |
|
"eval_trivia_pairs_runtime": 9.0522, |
|
"eval_trivia_pairs_samples_per_second": 22.094, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_quora_pairs_loss": 0.23968417942523956, |
|
"eval_quora_pairs_runtime": 0.6179, |
|
"eval_quora_pairs_samples_per_second": 323.693, |
|
"eval_quora_pairs_steps_per_second": 8.092, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_gooaq_pairs_loss": 1.3264899253845215, |
|
"eval_gooaq_pairs_runtime": 1.549, |
|
"eval_gooaq_pairs_samples_per_second": 129.112, |
|
"eval_gooaq_pairs_steps_per_second": 3.228, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8263801756587202, |
|
"eval_mrpc_pairs_loss": 0.14705294370651245, |
|
"eval_mrpc_pairs_runtime": 0.2494, |
|
"eval_mrpc_pairs_samples_per_second": 801.787, |
|
"eval_mrpc_pairs_steps_per_second": 20.045, |
|
"step": 5269 |
|
}, |
|
{ |
|
"epoch": 0.8281053952321205, |
|
"grad_norm": 17.636714935302734, |
|
"learning_rate": 3.459527948478686e-05, |
|
"loss": 2.3021, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.835633626097867, |
|
"grad_norm": 23.402650833129883, |
|
"learning_rate": 3.4512843672698696e-05, |
|
"loss": 1.5502, |
|
"step": 5328 |
|
}, |
|
{ |
|
"epoch": 0.8431618569636136, |
|
"grad_norm": 13.210539817810059, |
|
"learning_rate": 3.4422889182801225e-05, |
|
"loss": 1.7324, |
|
"step": 5376 |
|
}, |
|
{ |
|
"epoch": 0.8506900878293601, |
|
"grad_norm": 14.795612335205078, |
|
"learning_rate": 3.4325455769696324e-05, |
|
"loss": 1.8119, |
|
"step": 5424 |
|
}, |
|
{ |
|
"epoch": 0.8582183186951067, |
|
"grad_norm": 14.047534942626953, |
|
"learning_rate": 3.422058649323072e-05, |
|
"loss": 1.8507, |
|
"step": 5472 |
|
}, |
|
{ |
|
"epoch": 0.8657465495608532, |
|
"grad_norm": 0.7366377711296082, |
|
"learning_rate": 3.4108327699466066e-05, |
|
"loss": 1.7362, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.8732747804265998, |
|
"grad_norm": 16.555519104003906, |
|
"learning_rate": 3.398872900019673e-05, |
|
"loss": 2.082, |
|
"step": 5568 |
|
}, |
|
{ |
|
"epoch": 0.8808030112923463, |
|
"grad_norm": 16.52071189880371, |
|
"learning_rate": 3.386184325102423e-05, |
|
"loss": 2.1483, |
|
"step": 5616 |
|
}, |
|
{ |
|
"epoch": 0.8883312421580929, |
|
"grad_norm": 16.03848648071289, |
|
"learning_rate": 3.372772652799824e-05, |
|
"loss": 1.3961, |
|
"step": 5664 |
|
}, |
|
{ |
|
"epoch": 0.8958594730238394, |
|
"grad_norm": 15.494946479797363, |
|
"learning_rate": 3.358643810283421e-05, |
|
"loss": 1.6331, |
|
"step": 5712 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_nli-pairs_loss": 1.4454108476638794, |
|
"eval_nli-pairs_runtime": 4.0041, |
|
"eval_nli-pairs_samples_per_second": 49.949, |
|
"eval_nli-pairs_steps_per_second": 1.249, |
|
"eval_sts-test_pearson_cosine": 0.7644735043371528, |
|
"eval_sts-test_pearson_dot": 0.5461512421131185, |
|
"eval_sts-test_pearson_euclidean": 0.7520132891230207, |
|
"eval_sts-test_pearson_manhattan": 0.7535418655995262, |
|
"eval_sts-test_pearson_max": 0.7644735043371528, |
|
"eval_sts-test_spearman_cosine": 0.7630644782411757, |
|
"eval_sts-test_spearman_dot": 0.5239487411838791, |
|
"eval_sts-test_spearman_euclidean": 0.7392793315112096, |
|
"eval_sts-test_spearman_manhattan": 0.7426354353655322, |
|
"eval_sts-test_spearman_max": 0.7630644782411757, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_vitaminc-pairs_loss": 5.328937530517578, |
|
"eval_vitaminc-pairs_runtime": 1.4411, |
|
"eval_vitaminc-pairs_samples_per_second": 115.19, |
|
"eval_vitaminc-pairs_steps_per_second": 2.776, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_sts-label_loss": 4.186919212341309, |
|
"eval_sts-label_runtime": 0.4046, |
|
"eval_sts-label_samples_per_second": 494.357, |
|
"eval_sts-label_steps_per_second": 12.359, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_qnli-contrastive_loss": 0.35569697618484497, |
|
"eval_qnli-contrastive_runtime": 0.2807, |
|
"eval_qnli-contrastive_samples_per_second": 712.531, |
|
"eval_qnli-contrastive_steps_per_second": 17.813, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_scitail-pairs-qa_loss": 0.18789875507354736, |
|
"eval_scitail-pairs-qa_runtime": 1.1135, |
|
"eval_scitail-pairs-qa_samples_per_second": 179.622, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.491, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_scitail-pairs-pos_loss": 0.715129017829895, |
|
"eval_scitail-pairs-pos_runtime": 2.3885, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.736, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.093, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_xsum-pairs_loss": 0.624691367149353, |
|
"eval_xsum-pairs_runtime": 1.0408, |
|
"eval_xsum-pairs_samples_per_second": 192.169, |
|
"eval_xsum-pairs_steps_per_second": 4.804, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_compression-pairs_loss": 0.30371707677841187, |
|
"eval_compression-pairs_runtime": 0.2581, |
|
"eval_compression-pairs_samples_per_second": 774.8, |
|
"eval_compression-pairs_steps_per_second": 19.37, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_sciq_pairs_loss": 0.5236299633979797, |
|
"eval_sciq_pairs_runtime": 9.205, |
|
"eval_sciq_pairs_samples_per_second": 21.727, |
|
"eval_sciq_pairs_steps_per_second": 0.543, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_qasc_pairs_loss": 0.8543006777763367, |
|
"eval_qasc_pairs_runtime": 1.238, |
|
"eval_qasc_pairs_samples_per_second": 161.556, |
|
"eval_qasc_pairs_steps_per_second": 4.039, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_openbookqa_pairs_loss": 2.3740031719207764, |
|
"eval_openbookqa_pairs_runtime": 1.1145, |
|
"eval_openbookqa_pairs_samples_per_second": 179.452, |
|
"eval_openbookqa_pairs_steps_per_second": 4.486, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_msmarco_pairs_loss": 1.4328840970993042, |
|
"eval_msmarco_pairs_runtime": 2.5725, |
|
"eval_msmarco_pairs_samples_per_second": 77.746, |
|
"eval_msmarco_pairs_steps_per_second": 1.944, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_nq_pairs_loss": 1.5343101024627686, |
|
"eval_nq_pairs_runtime": 5.7416, |
|
"eval_nq_pairs_samples_per_second": 34.833, |
|
"eval_nq_pairs_steps_per_second": 0.871, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_trivia_pairs_loss": 1.7511711120605469, |
|
"eval_trivia_pairs_runtime": 9.1035, |
|
"eval_trivia_pairs_samples_per_second": 21.97, |
|
"eval_trivia_pairs_steps_per_second": 0.549, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_quora_pairs_loss": 0.2826410233974457, |
|
"eval_quora_pairs_runtime": 0.6444, |
|
"eval_quora_pairs_samples_per_second": 310.354, |
|
"eval_quora_pairs_steps_per_second": 7.759, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_gooaq_pairs_loss": 1.2128998041152954, |
|
"eval_gooaq_pairs_runtime": 1.5919, |
|
"eval_gooaq_pairs_samples_per_second": 125.634, |
|
"eval_gooaq_pairs_steps_per_second": 3.141, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.9015056461731493, |
|
"eval_mrpc_pairs_loss": 0.16468097269535065, |
|
"eval_mrpc_pairs_runtime": 0.2496, |
|
"eval_mrpc_pairs_samples_per_second": 801.385, |
|
"eval_mrpc_pairs_steps_per_second": 20.035, |
|
"step": 5748 |
|
}, |
|
{ |
|
"epoch": 0.903387703889586, |
|
"grad_norm": 2.899136543273926, |
|
"learning_rate": 3.3438040416718773e-05, |
|
"loss": 1.9863, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.9109159347553325, |
|
"grad_norm": 14.919694900512695, |
|
"learning_rate": 3.3282599052714414e-05, |
|
"loss": 1.6917, |
|
"step": 5808 |
|
}, |
|
{ |
|
"epoch": 0.918444165621079, |
|
"grad_norm": 1.626105785369873, |
|
"learning_rate": 3.312018270677559e-05, |
|
"loss": 1.7409, |
|
"step": 5856 |
|
}, |
|
{ |
|
"epoch": 0.9259723964868256, |
|
"grad_norm": 15.8577299118042, |
|
"learning_rate": 3.295086315738918e-05, |
|
"loss": 1.4397, |
|
"step": 5904 |
|
}, |
|
{ |
|
"epoch": 0.9335006273525721, |
|
"grad_norm": 10.255402565002441, |
|
"learning_rate": 3.277471523385255e-05, |
|
"loss": 1.2977, |
|
"step": 5952 |
|
}, |
|
{ |
|
"epoch": 0.9410288582183187, |
|
"grad_norm": 31.09028434753418, |
|
"learning_rate": 3.259181678320349e-05, |
|
"loss": 1.5429, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.9485570890840652, |
|
"grad_norm": 15.244925498962402, |
|
"learning_rate": 3.2402248635816294e-05, |
|
"loss": 1.6939, |
|
"step": 6048 |
|
}, |
|
{ |
|
"epoch": 0.9560853199498118, |
|
"grad_norm": 19.404956817626953, |
|
"learning_rate": 3.2206094569679564e-05, |
|
"loss": 1.5573, |
|
"step": 6096 |
|
}, |
|
{ |
|
"epoch": 0.9636135508155583, |
|
"grad_norm": 16.607341766357422, |
|
"learning_rate": 3.200344127337121e-05, |
|
"loss": 1.5873, |
|
"step": 6144 |
|
}, |
|
{ |
|
"epoch": 0.9711417816813049, |
|
"grad_norm": 1.1432667970657349, |
|
"learning_rate": 3.179437830774722e-05, |
|
"loss": 1.8802, |
|
"step": 6192 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_nli-pairs_loss": 1.3268091678619385, |
|
"eval_nli-pairs_runtime": 4.0761, |
|
"eval_nli-pairs_samples_per_second": 49.066, |
|
"eval_nli-pairs_steps_per_second": 1.227, |
|
"eval_sts-test_pearson_cosine": 0.7536875808596679, |
|
"eval_sts-test_pearson_dot": 0.5240503602011544, |
|
"eval_sts-test_pearson_euclidean": 0.7440772899082505, |
|
"eval_sts-test_pearson_manhattan": 0.7465814532083983, |
|
"eval_sts-test_pearson_max": 0.7536875808596679, |
|
"eval_sts-test_spearman_cosine": 0.7584726286921011, |
|
"eval_sts-test_spearman_dot": 0.49490205708473545, |
|
"eval_sts-test_spearman_euclidean": 0.7307844981527315, |
|
"eval_sts-test_spearman_manhattan": 0.7362283105144983, |
|
"eval_sts-test_spearman_max": 0.7584726286921011, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_vitaminc-pairs_loss": 5.498671054840088, |
|
"eval_vitaminc-pairs_runtime": 1.4379, |
|
"eval_vitaminc-pairs_samples_per_second": 115.449, |
|
"eval_vitaminc-pairs_steps_per_second": 2.782, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_sts-label_loss": 4.3516716957092285, |
|
"eval_sts-label_runtime": 0.3976, |
|
"eval_sts-label_samples_per_second": 503.027, |
|
"eval_sts-label_steps_per_second": 12.576, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_qnli-contrastive_loss": 0.30157506465911865, |
|
"eval_qnli-contrastive_runtime": 0.2787, |
|
"eval_qnli-contrastive_samples_per_second": 717.742, |
|
"eval_qnli-contrastive_steps_per_second": 17.944, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_scitail-pairs-qa_loss": 0.14389516413211823, |
|
"eval_scitail-pairs-qa_runtime": 1.0465, |
|
"eval_scitail-pairs-qa_samples_per_second": 191.121, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.778, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_scitail-pairs-pos_loss": 0.6142529249191284, |
|
"eval_scitail-pairs-pos_runtime": 2.3525, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.018, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.125, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_xsum-pairs_loss": 0.5579215288162231, |
|
"eval_xsum-pairs_runtime": 1.0381, |
|
"eval_xsum-pairs_samples_per_second": 192.655, |
|
"eval_xsum-pairs_steps_per_second": 4.816, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_compression-pairs_loss": 0.2538767158985138, |
|
"eval_compression-pairs_runtime": 0.2367, |
|
"eval_compression-pairs_samples_per_second": 844.99, |
|
"eval_compression-pairs_steps_per_second": 21.125, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_sciq_pairs_loss": 0.43967145681381226, |
|
"eval_sciq_pairs_runtime": 9.0489, |
|
"eval_sciq_pairs_samples_per_second": 22.102, |
|
"eval_sciq_pairs_steps_per_second": 0.553, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_qasc_pairs_loss": 0.7632485032081604, |
|
"eval_qasc_pairs_runtime": 1.2216, |
|
"eval_qasc_pairs_samples_per_second": 163.713, |
|
"eval_qasc_pairs_steps_per_second": 4.093, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_openbookqa_pairs_loss": 2.370297908782959, |
|
"eval_openbookqa_pairs_runtime": 1.051, |
|
"eval_openbookqa_pairs_samples_per_second": 190.297, |
|
"eval_openbookqa_pairs_steps_per_second": 4.757, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_msmarco_pairs_loss": 1.4191588163375854, |
|
"eval_msmarco_pairs_runtime": 2.5294, |
|
"eval_msmarco_pairs_samples_per_second": 79.071, |
|
"eval_msmarco_pairs_steps_per_second": 1.977, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_nq_pairs_loss": 1.3847519159317017, |
|
"eval_nq_pairs_runtime": 5.6464, |
|
"eval_nq_pairs_samples_per_second": 35.421, |
|
"eval_nq_pairs_steps_per_second": 0.886, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_trivia_pairs_loss": 1.6301060914993286, |
|
"eval_trivia_pairs_runtime": 9.06, |
|
"eval_trivia_pairs_samples_per_second": 22.075, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_quora_pairs_loss": 0.25898078083992004, |
|
"eval_quora_pairs_runtime": 0.6161, |
|
"eval_quora_pairs_samples_per_second": 324.63, |
|
"eval_quora_pairs_steps_per_second": 8.116, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_gooaq_pairs_loss": 1.1585972309112549, |
|
"eval_gooaq_pairs_runtime": 1.547, |
|
"eval_gooaq_pairs_samples_per_second": 129.286, |
|
"eval_gooaq_pairs_steps_per_second": 3.232, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9766311166875784, |
|
"eval_mrpc_pairs_loss": 0.12749388813972473, |
|
"eval_mrpc_pairs_runtime": 0.243, |
|
"eval_mrpc_pairs_samples_per_second": 822.939, |
|
"eval_mrpc_pairs_steps_per_second": 20.573, |
|
"step": 6227 |
|
}, |
|
{ |
|
"epoch": 0.9786700125470514, |
|
"grad_norm": 20.923500061035156, |
|
"learning_rate": 3.157899806636098e-05, |
|
"loss": 1.9813, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.986198243412798, |
|
"grad_norm": 24.244462966918945, |
|
"learning_rate": 3.13573957346308e-05, |
|
"loss": 2.2932, |
|
"step": 6288 |
|
}, |
|
{ |
|
"epoch": 0.9937264742785445, |
|
"grad_norm": 16.324562072753906, |
|
"learning_rate": 3.112966924777352e-05, |
|
"loss": 1.6308, |
|
"step": 6336 |
|
}, |
|
{ |
|
"epoch": 1.001254705144291, |
|
"grad_norm": 14.824076652526855, |
|
"learning_rate": 3.0895919247522884e-05, |
|
"loss": 1.497, |
|
"step": 6384 |
|
}, |
|
{ |
|
"epoch": 1.0087829360100375, |
|
"grad_norm": 18.46307373046875, |
|
"learning_rate": 3.065624903765184e-05, |
|
"loss": 1.758, |
|
"step": 6432 |
|
}, |
|
{ |
|
"epoch": 1.0163111668757843, |
|
"grad_norm": 16.654727935791016, |
|
"learning_rate": 3.0410764538318303e-05, |
|
"loss": 1.6188, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 1.0238393977415308, |
|
"grad_norm": 2.175520896911621, |
|
"learning_rate": 3.0159574239254692e-05, |
|
"loss": 2.1126, |
|
"step": 6528 |
|
}, |
|
{ |
|
"epoch": 1.0313676286072773, |
|
"grad_norm": 20.126880645751953, |
|
"learning_rate": 2.990278915182182e-05, |
|
"loss": 1.6129, |
|
"step": 6576 |
|
}, |
|
{ |
|
"epoch": 1.0388958594730238, |
|
"grad_norm": 24.355375289916992, |
|
"learning_rate": 2.964052275994841e-05, |
|
"loss": 1.4584, |
|
"step": 6624 |
|
}, |
|
{ |
|
"epoch": 1.0464240903387705, |
|
"grad_norm": 11.541900634765625, |
|
"learning_rate": 2.9372890969977852e-05, |
|
"loss": 1.5975, |
|
"step": 6672 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_nli-pairs_loss": 1.3095624446868896, |
|
"eval_nli-pairs_runtime": 4.0202, |
|
"eval_nli-pairs_samples_per_second": 49.749, |
|
"eval_nli-pairs_steps_per_second": 1.244, |
|
"eval_sts-test_pearson_cosine": 0.7778731779702309, |
|
"eval_sts-test_pearson_dot": 0.5593371858287987, |
|
"eval_sts-test_pearson_euclidean": 0.7587435793478869, |
|
"eval_sts-test_pearson_manhattan": 0.7610239287161901, |
|
"eval_sts-test_pearson_max": 0.7778731779702309, |
|
"eval_sts-test_spearman_cosine": 0.7782906514134392, |
|
"eval_sts-test_spearman_dot": 0.5371251012528334, |
|
"eval_sts-test_spearman_euclidean": 0.7452348039865185, |
|
"eval_sts-test_spearman_manhattan": 0.7492626260972672, |
|
"eval_sts-test_spearman_max": 0.7782906514134392, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_vitaminc-pairs_loss": 5.074347496032715, |
|
"eval_vitaminc-pairs_runtime": 1.4418, |
|
"eval_vitaminc-pairs_samples_per_second": 115.134, |
|
"eval_vitaminc-pairs_steps_per_second": 2.774, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_sts-label_loss": 4.242542266845703, |
|
"eval_sts-label_runtime": 0.414, |
|
"eval_sts-label_samples_per_second": 483.143, |
|
"eval_sts-label_steps_per_second": 12.079, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_qnli-contrastive_loss": 0.26250946521759033, |
|
"eval_qnli-contrastive_runtime": 0.2857, |
|
"eval_qnli-contrastive_samples_per_second": 700.035, |
|
"eval_qnli-contrastive_steps_per_second": 17.501, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_scitail-pairs-qa_loss": 0.15288515388965607, |
|
"eval_scitail-pairs-qa_runtime": 1.1069, |
|
"eval_scitail-pairs-qa_samples_per_second": 180.69, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.517, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_scitail-pairs-pos_loss": 0.5382486581802368, |
|
"eval_scitail-pairs-pos_runtime": 2.3648, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.574, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.114, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_xsum-pairs_loss": 0.48308631777763367, |
|
"eval_xsum-pairs_runtime": 1.0411, |
|
"eval_xsum-pairs_samples_per_second": 192.11, |
|
"eval_xsum-pairs_steps_per_second": 4.803, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_compression-pairs_loss": 0.23988038301467896, |
|
"eval_compression-pairs_runtime": 0.2389, |
|
"eval_compression-pairs_samples_per_second": 837.159, |
|
"eval_compression-pairs_steps_per_second": 20.929, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_sciq_pairs_loss": 0.3883107304573059, |
|
"eval_sciq_pairs_runtime": 9.1094, |
|
"eval_sciq_pairs_samples_per_second": 21.955, |
|
"eval_sciq_pairs_steps_per_second": 0.549, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_qasc_pairs_loss": 0.6684954762458801, |
|
"eval_qasc_pairs_runtime": 1.2249, |
|
"eval_qasc_pairs_samples_per_second": 163.273, |
|
"eval_qasc_pairs_steps_per_second": 4.082, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_openbookqa_pairs_loss": 2.076920986175537, |
|
"eval_openbookqa_pairs_runtime": 1.0566, |
|
"eval_openbookqa_pairs_samples_per_second": 189.291, |
|
"eval_openbookqa_pairs_steps_per_second": 4.732, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_msmarco_pairs_loss": 1.3436123132705688, |
|
"eval_msmarco_pairs_runtime": 2.5384, |
|
"eval_msmarco_pairs_samples_per_second": 78.791, |
|
"eval_msmarco_pairs_steps_per_second": 1.97, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_nq_pairs_loss": 1.2744060754776, |
|
"eval_nq_pairs_runtime": 5.6694, |
|
"eval_nq_pairs_samples_per_second": 35.277, |
|
"eval_nq_pairs_steps_per_second": 0.882, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_trivia_pairs_loss": 1.356447458267212, |
|
"eval_trivia_pairs_runtime": 9.0814, |
|
"eval_trivia_pairs_samples_per_second": 22.023, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_quora_pairs_loss": 0.20534881949424744, |
|
"eval_quora_pairs_runtime": 0.6243, |
|
"eval_quora_pairs_samples_per_second": 320.363, |
|
"eval_quora_pairs_steps_per_second": 8.009, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_gooaq_pairs_loss": 1.131415843963623, |
|
"eval_gooaq_pairs_runtime": 1.6004, |
|
"eval_gooaq_pairs_samples_per_second": 124.966, |
|
"eval_gooaq_pairs_steps_per_second": 3.124, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.0517565872020076, |
|
"eval_mrpc_pairs_loss": 0.12877897918224335, |
|
"eval_mrpc_pairs_runtime": 0.2425, |
|
"eval_mrpc_pairs_samples_per_second": 824.742, |
|
"eval_mrpc_pairs_steps_per_second": 20.619, |
|
"step": 6706 |
|
}, |
|
{ |
|
"epoch": 1.053952321204517, |
|
"grad_norm": 12.653462409973145, |
|
"learning_rate": 2.9100012059444395e-05, |
|
"loss": 1.6933, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 1.0614805520702635, |
|
"grad_norm": 13.93374252319336, |
|
"learning_rate": 2.8822006624801445e-05, |
|
"loss": 1.2931, |
|
"step": 6768 |
|
}, |
|
{ |
|
"epoch": 1.06900878293601, |
|
"grad_norm": 17.680423736572266, |
|
"learning_rate": 2.8538997528125016e-05, |
|
"loss": 1.8077, |
|
"step": 6816 |
|
}, |
|
{ |
|
"epoch": 1.0765370138017567, |
|
"grad_norm": 14.7294921875, |
|
"learning_rate": 2.8251109842815857e-05, |
|
"loss": 1.4748, |
|
"step": 6864 |
|
}, |
|
{ |
|
"epoch": 1.0840652446675032, |
|
"grad_norm": 26.121625900268555, |
|
"learning_rate": 2.795847079832445e-05, |
|
"loss": 1.225, |
|
"step": 6912 |
|
}, |
|
{ |
|
"epoch": 1.0915934755332497, |
|
"grad_norm": 5.665552616119385, |
|
"learning_rate": 2.7661209723922966e-05, |
|
"loss": 1.7627, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 1.0991217063989962, |
|
"grad_norm": 1.2186944484710693, |
|
"learning_rate": 2.7359457991549365e-05, |
|
"loss": 1.5613, |
|
"step": 7008 |
|
}, |
|
{ |
|
"epoch": 1.1066499372647427, |
|
"grad_norm": 4.569064617156982, |
|
"learning_rate": 2.7053348957748744e-05, |
|
"loss": 1.6548, |
|
"step": 7056 |
|
}, |
|
{ |
|
"epoch": 1.1141781681304894, |
|
"grad_norm": 11.330352783203125, |
|
"learning_rate": 2.6743017904737505e-05, |
|
"loss": 1.4033, |
|
"step": 7104 |
|
}, |
|
{ |
|
"epoch": 1.121706398996236, |
|
"grad_norm": 27.65351676940918, |
|
"learning_rate": 2.6428601980616635e-05, |
|
"loss": 1.3999, |
|
"step": 7152 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_nli-pairs_loss": 1.212507724761963, |
|
"eval_nli-pairs_runtime": 4.0093, |
|
"eval_nli-pairs_samples_per_second": 49.884, |
|
"eval_nli-pairs_steps_per_second": 1.247, |
|
"eval_sts-test_pearson_cosine": 0.771726417661154, |
|
"eval_sts-test_pearson_dot": 0.5582118135503709, |
|
"eval_sts-test_pearson_euclidean": 0.7582436737188462, |
|
"eval_sts-test_pearson_manhattan": 0.7602002352027541, |
|
"eval_sts-test_pearson_max": 0.771726417661154, |
|
"eval_sts-test_spearman_cosine": 0.7760137659698252, |
|
"eval_sts-test_spearman_dot": 0.5314151368436631, |
|
"eval_sts-test_spearman_euclidean": 0.7447358001087161, |
|
"eval_sts-test_spearman_manhattan": 0.7487681272112986, |
|
"eval_sts-test_spearman_max": 0.7760137659698252, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_vitaminc-pairs_loss": 4.9936699867248535, |
|
"eval_vitaminc-pairs_runtime": 1.4575, |
|
"eval_vitaminc-pairs_samples_per_second": 113.89, |
|
"eval_vitaminc-pairs_steps_per_second": 2.744, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_sts-label_loss": 4.1703057289123535, |
|
"eval_sts-label_runtime": 0.4168, |
|
"eval_sts-label_samples_per_second": 479.874, |
|
"eval_sts-label_steps_per_second": 11.997, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_qnli-contrastive_loss": 0.31846657395362854, |
|
"eval_qnli-contrastive_runtime": 0.2838, |
|
"eval_qnli-contrastive_samples_per_second": 704.663, |
|
"eval_qnli-contrastive_steps_per_second": 17.617, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_scitail-pairs-qa_loss": 0.1417759656906128, |
|
"eval_scitail-pairs-qa_runtime": 1.0581, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.022, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.726, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_scitail-pairs-pos_loss": 0.5202088356018066, |
|
"eval_scitail-pairs-pos_runtime": 2.4081, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.054, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.076, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_xsum-pairs_loss": 0.44685080647468567, |
|
"eval_xsum-pairs_runtime": 1.0416, |
|
"eval_xsum-pairs_samples_per_second": 192.009, |
|
"eval_xsum-pairs_steps_per_second": 4.8, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_compression-pairs_loss": 0.22411338984966278, |
|
"eval_compression-pairs_runtime": 0.2438, |
|
"eval_compression-pairs_samples_per_second": 820.201, |
|
"eval_compression-pairs_steps_per_second": 20.505, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_sciq_pairs_loss": 0.37388285994529724, |
|
"eval_sciq_pairs_runtime": 9.0875, |
|
"eval_sciq_pairs_samples_per_second": 22.008, |
|
"eval_sciq_pairs_steps_per_second": 0.55, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_qasc_pairs_loss": 0.6410768032073975, |
|
"eval_qasc_pairs_runtime": 1.2169, |
|
"eval_qasc_pairs_samples_per_second": 164.348, |
|
"eval_qasc_pairs_steps_per_second": 4.109, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_openbookqa_pairs_loss": 2.043297052383423, |
|
"eval_openbookqa_pairs_runtime": 1.0552, |
|
"eval_openbookqa_pairs_samples_per_second": 189.539, |
|
"eval_openbookqa_pairs_steps_per_second": 4.738, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_msmarco_pairs_loss": 1.2795078754425049, |
|
"eval_msmarco_pairs_runtime": 2.5343, |
|
"eval_msmarco_pairs_samples_per_second": 78.919, |
|
"eval_msmarco_pairs_steps_per_second": 1.973, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_nq_pairs_loss": 1.1771754026412964, |
|
"eval_nq_pairs_runtime": 5.6589, |
|
"eval_nq_pairs_samples_per_second": 35.343, |
|
"eval_nq_pairs_steps_per_second": 0.884, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_trivia_pairs_loss": 1.305577039718628, |
|
"eval_trivia_pairs_runtime": 9.0872, |
|
"eval_trivia_pairs_samples_per_second": 22.009, |
|
"eval_trivia_pairs_steps_per_second": 0.55, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_quora_pairs_loss": 0.2645386755466461, |
|
"eval_quora_pairs_runtime": 0.6429, |
|
"eval_quora_pairs_samples_per_second": 311.086, |
|
"eval_quora_pairs_steps_per_second": 7.777, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_gooaq_pairs_loss": 1.0562756061553955, |
|
"eval_gooaq_pairs_runtime": 1.5772, |
|
"eval_gooaq_pairs_samples_per_second": 126.809, |
|
"eval_gooaq_pairs_steps_per_second": 3.17, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1268820577164367, |
|
"eval_mrpc_pairs_loss": 0.1197453960776329, |
|
"eval_mrpc_pairs_runtime": 0.2393, |
|
"eval_mrpc_pairs_samples_per_second": 835.714, |
|
"eval_mrpc_pairs_steps_per_second": 20.893, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 1.1292346298619824, |
|
"grad_norm": 26.328977584838867, |
|
"learning_rate": 2.611024013876024e-05, |
|
"loss": 1.707, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 1.136762860727729, |
|
"grad_norm": 14.730170249938965, |
|
"learning_rate": 2.578807307640633e-05, |
|
"loss": 1.2812, |
|
"step": 7248 |
|
}, |
|
{ |
|
"epoch": 1.1442910915934754, |
|
"grad_norm": 16.989944458007812, |
|
"learning_rate": 2.546224317247695e-05, |
|
"loss": 2.0209, |
|
"step": 7296 |
|
}, |
|
{ |
|
"epoch": 1.1518193224592221, |
|
"grad_norm": 15.942720413208008, |
|
"learning_rate": 2.5132894424654982e-05, |
|
"loss": 1.465, |
|
"step": 7344 |
|
}, |
|
{ |
|
"epoch": 1.1593475533249686, |
|
"grad_norm": 13.42844009399414, |
|
"learning_rate": 2.4800172385745665e-05, |
|
"loss": 1.6181, |
|
"step": 7392 |
|
}, |
|
{ |
|
"epoch": 1.1668757841907151, |
|
"grad_norm": 2.6255691051483154, |
|
"learning_rate": 2.446422409935082e-05, |
|
"loss": 1.3386, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 1.1744040150564616, |
|
"grad_norm": 11.230644226074219, |
|
"learning_rate": 2.412519803488417e-05, |
|
"loss": 1.4015, |
|
"step": 7488 |
|
}, |
|
{ |
|
"epoch": 1.1819322459222084, |
|
"grad_norm": 17.679582595825195, |
|
"learning_rate": 2.3783244021956605e-05, |
|
"loss": 1.5057, |
|
"step": 7536 |
|
}, |
|
{ |
|
"epoch": 1.1894604767879549, |
|
"grad_norm": 12.440255165100098, |
|
"learning_rate": 2.3438513184160302e-05, |
|
"loss": 1.2441, |
|
"step": 7584 |
|
}, |
|
{ |
|
"epoch": 1.1969887076537014, |
|
"grad_norm": 17.67916488647461, |
|
"learning_rate": 2.309115787228094e-05, |
|
"loss": 1.1718, |
|
"step": 7632 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_nli-pairs_loss": 1.1914833784103394, |
|
"eval_nli-pairs_runtime": 4.1213, |
|
"eval_nli-pairs_samples_per_second": 48.528, |
|
"eval_nli-pairs_steps_per_second": 1.213, |
|
"eval_sts-test_pearson_cosine": 0.7800804097388158, |
|
"eval_sts-test_pearson_dot": 0.5541152451346555, |
|
"eval_sts-test_pearson_euclidean": 0.7555650256629914, |
|
"eval_sts-test_pearson_manhattan": 0.7570661501424889, |
|
"eval_sts-test_pearson_max": 0.7800804097388158, |
|
"eval_sts-test_spearman_cosine": 0.7837125392806782, |
|
"eval_sts-test_spearman_dot": 0.5299991823870257, |
|
"eval_sts-test_spearman_euclidean": 0.7420646288254193, |
|
"eval_sts-test_spearman_manhattan": 0.7454823370503758, |
|
"eval_sts-test_spearman_max": 0.7837125392806782, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_vitaminc-pairs_loss": 4.815649509429932, |
|
"eval_vitaminc-pairs_runtime": 1.4965, |
|
"eval_vitaminc-pairs_samples_per_second": 110.928, |
|
"eval_vitaminc-pairs_steps_per_second": 2.673, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_sts-label_loss": 4.200761795043945, |
|
"eval_sts-label_runtime": 0.4104, |
|
"eval_sts-label_samples_per_second": 487.308, |
|
"eval_sts-label_steps_per_second": 12.183, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_qnli-contrastive_loss": 0.21287916600704193, |
|
"eval_qnli-contrastive_runtime": 0.2849, |
|
"eval_qnli-contrastive_samples_per_second": 702.086, |
|
"eval_qnli-contrastive_steps_per_second": 17.552, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_scitail-pairs-qa_loss": 0.12968133389949799, |
|
"eval_scitail-pairs-qa_runtime": 1.0998, |
|
"eval_scitail-pairs-qa_samples_per_second": 181.845, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.546, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_scitail-pairs-pos_loss": 0.4889526665210724, |
|
"eval_scitail-pairs-pos_runtime": 2.3956, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.486, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.087, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_xsum-pairs_loss": 0.42897719144821167, |
|
"eval_xsum-pairs_runtime": 1.0492, |
|
"eval_xsum-pairs_samples_per_second": 190.623, |
|
"eval_xsum-pairs_steps_per_second": 4.766, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_compression-pairs_loss": 0.20111607015132904, |
|
"eval_compression-pairs_runtime": 0.2549, |
|
"eval_compression-pairs_samples_per_second": 784.74, |
|
"eval_compression-pairs_steps_per_second": 19.618, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_sciq_pairs_loss": 0.36706623435020447, |
|
"eval_sciq_pairs_runtime": 9.1959, |
|
"eval_sciq_pairs_samples_per_second": 21.749, |
|
"eval_sciq_pairs_steps_per_second": 0.544, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_qasc_pairs_loss": 0.5133731961250305, |
|
"eval_qasc_pairs_runtime": 1.2209, |
|
"eval_qasc_pairs_samples_per_second": 163.813, |
|
"eval_qasc_pairs_steps_per_second": 4.095, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_openbookqa_pairs_loss": 1.9764889478683472, |
|
"eval_openbookqa_pairs_runtime": 1.1177, |
|
"eval_openbookqa_pairs_samples_per_second": 178.939, |
|
"eval_openbookqa_pairs_steps_per_second": 4.473, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_msmarco_pairs_loss": 1.1972062587738037, |
|
"eval_msmarco_pairs_runtime": 2.5812, |
|
"eval_msmarco_pairs_samples_per_second": 77.485, |
|
"eval_msmarco_pairs_steps_per_second": 1.937, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_nq_pairs_loss": 1.1358052492141724, |
|
"eval_nq_pairs_runtime": 5.7952, |
|
"eval_nq_pairs_samples_per_second": 34.511, |
|
"eval_nq_pairs_steps_per_second": 0.863, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_trivia_pairs_loss": 1.3059653043746948, |
|
"eval_trivia_pairs_runtime": 9.1651, |
|
"eval_trivia_pairs_samples_per_second": 21.822, |
|
"eval_trivia_pairs_steps_per_second": 0.546, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_quora_pairs_loss": 0.23538345098495483, |
|
"eval_quora_pairs_runtime": 0.6553, |
|
"eval_quora_pairs_samples_per_second": 305.181, |
|
"eval_quora_pairs_steps_per_second": 7.63, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_gooaq_pairs_loss": 1.0195775032043457, |
|
"eval_gooaq_pairs_runtime": 1.5615, |
|
"eval_gooaq_pairs_samples_per_second": 128.081, |
|
"eval_gooaq_pairs_steps_per_second": 3.202, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2020075282308658, |
|
"eval_mrpc_pairs_loss": 0.10825151205062866, |
|
"eval_mrpc_pairs_runtime": 0.2544, |
|
"eval_mrpc_pairs_samples_per_second": 786.113, |
|
"eval_mrpc_pairs_steps_per_second": 19.653, |
|
"step": 7664 |
|
}, |
|
{ |
|
"epoch": 1.2045169385194479, |
|
"grad_norm": 11.216833114624023, |
|
"learning_rate": 2.2741331596967632e-05, |
|
"loss": 1.7562, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 1.2120451693851946, |
|
"grad_norm": 0.7245489358901978, |
|
"learning_rate": 2.238918896089022e-05, |
|
"loss": 1.2954, |
|
"step": 7728 |
|
}, |
|
{ |
|
"epoch": 1.219573400250941, |
|
"grad_norm": 18.958086013793945, |
|
"learning_rate": 2.2034885590413966e-05, |
|
"loss": 1.7821, |
|
"step": 7776 |
|
}, |
|
{ |
|
"epoch": 1.2271016311166876, |
|
"grad_norm": 1.1195874214172363, |
|
"learning_rate": 2.1678578066821842e-05, |
|
"loss": 1.4494, |
|
"step": 7824 |
|
}, |
|
{ |
|
"epoch": 1.234629861982434, |
|
"grad_norm": 10.63464641571045, |
|
"learning_rate": 2.1320423857114807e-05, |
|
"loss": 1.1674, |
|
"step": 7872 |
|
}, |
|
{ |
|
"epoch": 1.2421580928481806, |
|
"grad_norm": 20.8350830078125, |
|
"learning_rate": 2.0960581244420636e-05, |
|
"loss": 1.1511, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 1.2496863237139273, |
|
"grad_norm": 20.231365203857422, |
|
"learning_rate": 2.0599209258042067e-05, |
|
"loss": 1.1629, |
|
"step": 7968 |
|
}, |
|
{ |
|
"epoch": 1.2572145545796738, |
|
"grad_norm": 2.3765740394592285, |
|
"learning_rate": 2.023646760317521e-05, |
|
"loss": 1.167, |
|
"step": 8016 |
|
}, |
|
{ |
|
"epoch": 1.2647427854454203, |
|
"grad_norm": 0.45109543204307556, |
|
"learning_rate": 1.9872516590329177e-05, |
|
"loss": 1.5169, |
|
"step": 8064 |
|
}, |
|
{ |
|
"epoch": 1.272271016311167, |
|
"grad_norm": 19.704111099243164, |
|
"learning_rate": 1.9507517064478322e-05, |
|
"loss": 1.1595, |
|
"step": 8112 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_nli-pairs_loss": 1.117584466934204, |
|
"eval_nli-pairs_runtime": 4.0331, |
|
"eval_nli-pairs_samples_per_second": 49.59, |
|
"eval_nli-pairs_steps_per_second": 1.24, |
|
"eval_sts-test_pearson_cosine": 0.7735358515290777, |
|
"eval_sts-test_pearson_dot": 0.560523750110175, |
|
"eval_sts-test_pearson_euclidean": 0.7469529952792658, |
|
"eval_sts-test_pearson_manhattan": 0.7479614812850265, |
|
"eval_sts-test_pearson_max": 0.7735358515290777, |
|
"eval_sts-test_spearman_cosine": 0.7789103785997893, |
|
"eval_sts-test_spearman_dot": 0.534930979573433, |
|
"eval_sts-test_spearman_euclidean": 0.7320475561187462, |
|
"eval_sts-test_spearman_manhattan": 0.7359660045883268, |
|
"eval_sts-test_spearman_max": 0.7789103785997893, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_vitaminc-pairs_loss": 4.841164588928223, |
|
"eval_vitaminc-pairs_runtime": 1.4299, |
|
"eval_vitaminc-pairs_samples_per_second": 116.092, |
|
"eval_vitaminc-pairs_steps_per_second": 2.797, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_sts-label_loss": 4.0971784591674805, |
|
"eval_sts-label_runtime": 0.4026, |
|
"eval_sts-label_samples_per_second": 496.749, |
|
"eval_sts-label_steps_per_second": 12.419, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_qnli-contrastive_loss": 0.2050691843032837, |
|
"eval_qnli-contrastive_runtime": 0.2817, |
|
"eval_qnli-contrastive_samples_per_second": 710.023, |
|
"eval_qnli-contrastive_steps_per_second": 17.751, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_scitail-pairs-qa_loss": 0.11168017983436584, |
|
"eval_scitail-pairs-qa_runtime": 1.056, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.398, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.735, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_scitail-pairs-pos_loss": 0.495766818523407, |
|
"eval_scitail-pairs-pos_runtime": 2.4272, |
|
"eval_scitail-pairs-pos_samples_per_second": 82.4, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.06, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_xsum-pairs_loss": 0.43334144353866577, |
|
"eval_xsum-pairs_runtime": 1.0522, |
|
"eval_xsum-pairs_samples_per_second": 190.086, |
|
"eval_xsum-pairs_steps_per_second": 4.752, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_compression-pairs_loss": 0.1837506741285324, |
|
"eval_compression-pairs_runtime": 0.2435, |
|
"eval_compression-pairs_samples_per_second": 821.487, |
|
"eval_compression-pairs_steps_per_second": 20.537, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_sciq_pairs_loss": 0.35108524560928345, |
|
"eval_sciq_pairs_runtime": 9.102, |
|
"eval_sciq_pairs_samples_per_second": 21.973, |
|
"eval_sciq_pairs_steps_per_second": 0.549, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_qasc_pairs_loss": 0.486527681350708, |
|
"eval_qasc_pairs_runtime": 1.2312, |
|
"eval_qasc_pairs_samples_per_second": 162.45, |
|
"eval_qasc_pairs_steps_per_second": 4.061, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_openbookqa_pairs_loss": 1.9944250583648682, |
|
"eval_openbookqa_pairs_runtime": 1.0997, |
|
"eval_openbookqa_pairs_samples_per_second": 181.872, |
|
"eval_openbookqa_pairs_steps_per_second": 4.547, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_msmarco_pairs_loss": 1.1277974843978882, |
|
"eval_msmarco_pairs_runtime": 2.5357, |
|
"eval_msmarco_pairs_samples_per_second": 78.874, |
|
"eval_msmarco_pairs_steps_per_second": 1.972, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_nq_pairs_loss": 1.092708945274353, |
|
"eval_nq_pairs_runtime": 5.6524, |
|
"eval_nq_pairs_samples_per_second": 35.383, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_trivia_pairs_loss": 1.3160641193389893, |
|
"eval_trivia_pairs_runtime": 9.0422, |
|
"eval_trivia_pairs_samples_per_second": 22.118, |
|
"eval_trivia_pairs_steps_per_second": 0.553, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_quora_pairs_loss": 0.1377909630537033, |
|
"eval_quora_pairs_runtime": 0.6252, |
|
"eval_quora_pairs_samples_per_second": 319.913, |
|
"eval_quora_pairs_steps_per_second": 7.998, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_gooaq_pairs_loss": 0.9788944721221924, |
|
"eval_gooaq_pairs_runtime": 1.5536, |
|
"eval_gooaq_pairs_samples_per_second": 128.731, |
|
"eval_gooaq_pairs_steps_per_second": 3.218, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2771329987452948, |
|
"eval_mrpc_pairs_loss": 0.09496909379959106, |
|
"eval_mrpc_pairs_runtime": 0.2386, |
|
"eval_mrpc_pairs_samples_per_second": 838.384, |
|
"eval_mrpc_pairs_steps_per_second": 20.96, |
|
"step": 8143 |
|
}, |
|
{ |
|
"epoch": 1.2797992471769133, |
|
"grad_norm": 10.626935958862305, |
|
"learning_rate": 1.914163033397814e-05, |
|
"loss": 1.4598, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 1.28732747804266, |
|
"grad_norm": 70.21989440917969, |
|
"learning_rate": 1.8775018099276515e-05, |
|
"loss": 1.2038, |
|
"step": 8208 |
|
}, |
|
{ |
|
"epoch": 1.2948557089084065, |
|
"grad_norm": 13.136975288391113, |
|
"learning_rate": 1.8407842381451562e-05, |
|
"loss": 1.2381, |
|
"step": 8256 |
|
}, |
|
{ |
|
"epoch": 1.302383939774153, |
|
"grad_norm": 14.311907768249512, |
|
"learning_rate": 1.804026545060783e-05, |
|
"loss": 0.9591, |
|
"step": 8304 |
|
}, |
|
{ |
|
"epoch": 1.3099121706398997, |
|
"grad_norm": 11.372726440429688, |
|
"learning_rate": 1.767244975416249e-05, |
|
"loss": 1.3451, |
|
"step": 8352 |
|
}, |
|
{ |
|
"epoch": 1.3174404015056462, |
|
"grad_norm": 12.1253080368042, |
|
"learning_rate": 1.7304557845053075e-05, |
|
"loss": 1.2669, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 1.3249686323713927, |
|
"grad_norm": 10.53577995300293, |
|
"learning_rate": 1.693675230989863e-05, |
|
"loss": 1.4754, |
|
"step": 8448 |
|
}, |
|
{ |
|
"epoch": 1.3324968632371392, |
|
"grad_norm": 16.669851303100586, |
|
"learning_rate": 1.6576849494355775e-05, |
|
"loss": 1.2094, |
|
"step": 8496 |
|
}, |
|
{ |
|
"epoch": 1.3400250941028857, |
|
"grad_norm": 0.5680477023124695, |
|
"learning_rate": 1.6209694016709658e-05, |
|
"loss": 1.4406, |
|
"step": 8544 |
|
}, |
|
{ |
|
"epoch": 1.3475533249686324, |
|
"grad_norm": 13.915863990783691, |
|
"learning_rate": 1.5843108778533816e-05, |
|
"loss": 1.3063, |
|
"step": 8592 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_nli-pairs_loss": 1.1088637113571167, |
|
"eval_nli-pairs_runtime": 4.0091, |
|
"eval_nli-pairs_samples_per_second": 49.887, |
|
"eval_nli-pairs_steps_per_second": 1.247, |
|
"eval_sts-test_pearson_cosine": 0.7694392469514715, |
|
"eval_sts-test_pearson_dot": 0.5341434503970798, |
|
"eval_sts-test_pearson_euclidean": 0.7443691249920426, |
|
"eval_sts-test_pearson_manhattan": 0.7461066741203818, |
|
"eval_sts-test_pearson_max": 0.7694392469514715, |
|
"eval_sts-test_spearman_cosine": 0.7756586193959978, |
|
"eval_sts-test_spearman_dot": 0.5051721749088769, |
|
"eval_sts-test_spearman_euclidean": 0.73008171628055, |
|
"eval_sts-test_spearman_manhattan": 0.7347429927090328, |
|
"eval_sts-test_spearman_max": 0.7756586193959978, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_vitaminc-pairs_loss": 4.769205093383789, |
|
"eval_vitaminc-pairs_runtime": 1.4375, |
|
"eval_vitaminc-pairs_samples_per_second": 115.477, |
|
"eval_vitaminc-pairs_steps_per_second": 2.783, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_sts-label_loss": 4.07755184173584, |
|
"eval_sts-label_runtime": 0.4049, |
|
"eval_sts-label_samples_per_second": 493.895, |
|
"eval_sts-label_steps_per_second": 12.347, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_qnli-contrastive_loss": 0.17923730611801147, |
|
"eval_qnli-contrastive_runtime": 0.2806, |
|
"eval_qnli-contrastive_samples_per_second": 712.745, |
|
"eval_qnli-contrastive_steps_per_second": 17.819, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_scitail-pairs-qa_loss": 0.10800071805715561, |
|
"eval_scitail-pairs-qa_runtime": 1.1053, |
|
"eval_scitail-pairs-qa_samples_per_second": 180.952, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.524, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_scitail-pairs-pos_loss": 0.48035934567451477, |
|
"eval_scitail-pairs-pos_runtime": 2.3848, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.864, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.097, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_xsum-pairs_loss": 0.408091276884079, |
|
"eval_xsum-pairs_runtime": 1.043, |
|
"eval_xsum-pairs_samples_per_second": 191.746, |
|
"eval_xsum-pairs_steps_per_second": 4.794, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_compression-pairs_loss": 0.17393065989017487, |
|
"eval_compression-pairs_runtime": 0.2388, |
|
"eval_compression-pairs_samples_per_second": 837.593, |
|
"eval_compression-pairs_steps_per_second": 20.94, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_sciq_pairs_loss": 0.3497403562068939, |
|
"eval_sciq_pairs_runtime": 9.1128, |
|
"eval_sciq_pairs_samples_per_second": 21.947, |
|
"eval_sciq_pairs_steps_per_second": 0.549, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_qasc_pairs_loss": 0.4620342552661896, |
|
"eval_qasc_pairs_runtime": 1.2208, |
|
"eval_qasc_pairs_samples_per_second": 163.822, |
|
"eval_qasc_pairs_steps_per_second": 4.096, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_openbookqa_pairs_loss": 1.9002811908721924, |
|
"eval_openbookqa_pairs_runtime": 1.0593, |
|
"eval_openbookqa_pairs_samples_per_second": 188.81, |
|
"eval_openbookqa_pairs_steps_per_second": 4.72, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_msmarco_pairs_loss": 1.0928370952606201, |
|
"eval_msmarco_pairs_runtime": 2.5227, |
|
"eval_msmarco_pairs_samples_per_second": 79.281, |
|
"eval_msmarco_pairs_steps_per_second": 1.982, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_nq_pairs_loss": 1.0306727886199951, |
|
"eval_nq_pairs_runtime": 5.6685, |
|
"eval_nq_pairs_samples_per_second": 35.283, |
|
"eval_nq_pairs_steps_per_second": 0.882, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_trivia_pairs_loss": 1.3031796216964722, |
|
"eval_trivia_pairs_runtime": 9.0827, |
|
"eval_trivia_pairs_samples_per_second": 22.02, |
|
"eval_trivia_pairs_steps_per_second": 0.55, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_quora_pairs_loss": 0.16622459888458252, |
|
"eval_quora_pairs_runtime": 0.6233, |
|
"eval_quora_pairs_samples_per_second": 320.863, |
|
"eval_quora_pairs_steps_per_second": 8.022, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_gooaq_pairs_loss": 0.9280920624732971, |
|
"eval_gooaq_pairs_runtime": 1.5586, |
|
"eval_gooaq_pairs_samples_per_second": 128.317, |
|
"eval_gooaq_pairs_steps_per_second": 3.208, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.352258469259724, |
|
"eval_mrpc_pairs_loss": 0.09042897820472717, |
|
"eval_mrpc_pairs_runtime": 0.2477, |
|
"eval_mrpc_pairs_samples_per_second": 807.373, |
|
"eval_mrpc_pairs_steps_per_second": 20.184, |
|
"step": 8622 |
|
}, |
|
{ |
|
"epoch": 1.355081555834379, |
|
"grad_norm": 10.317242622375488, |
|
"learning_rate": 1.5477255788975485e-05, |
|
"loss": 1.1662, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 1.3626097867001254, |
|
"grad_norm": 14.411606788635254, |
|
"learning_rate": 1.5112296733571015e-05, |
|
"loss": 1.1592, |
|
"step": 8688 |
|
}, |
|
{ |
|
"epoch": 1.370138017565872, |
|
"grad_norm": 14.328046798706055, |
|
"learning_rate": 1.474839290279036e-05, |
|
"loss": 1.1742, |
|
"step": 8736 |
|
}, |
|
{ |
|
"epoch": 1.3776662484316184, |
|
"grad_norm": 15.689784049987793, |
|
"learning_rate": 1.4385705120756155e-05, |
|
"loss": 1.2304, |
|
"step": 8784 |
|
}, |
|
{ |
|
"epoch": 1.3851944792973652, |
|
"grad_norm": 1.1622222661972046, |
|
"learning_rate": 1.4024393674168905e-05, |
|
"loss": 1.0637, |
|
"step": 8832 |
|
}, |
|
{ |
|
"epoch": 1.3927227101631117, |
|
"grad_norm": 55.88763427734375, |
|
"learning_rate": 1.3664618241469671e-05, |
|
"loss": 1.8133, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 1.4002509410288582, |
|
"grad_norm": 11.148764610290527, |
|
"learning_rate": 1.3306537822271584e-05, |
|
"loss": 1.682, |
|
"step": 8928 |
|
}, |
|
{ |
|
"epoch": 1.4077791718946049, |
|
"grad_norm": 34.30131149291992, |
|
"learning_rate": 1.2950310667091337e-05, |
|
"loss": 1.1823, |
|
"step": 8976 |
|
}, |
|
{ |
|
"epoch": 1.4153074027603514, |
|
"grad_norm": 1.3407506942749023, |
|
"learning_rate": 1.2596094207411785e-05, |
|
"loss": 1.2953, |
|
"step": 9024 |
|
}, |
|
{ |
|
"epoch": 1.4228356336260979, |
|
"grad_norm": 0.3961174190044403, |
|
"learning_rate": 1.2244044986106427e-05, |
|
"loss": 0.9769, |
|
"step": 9072 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_nli-pairs_loss": 1.065342664718628, |
|
"eval_nli-pairs_runtime": 4.0707, |
|
"eval_nli-pairs_samples_per_second": 49.132, |
|
"eval_nli-pairs_steps_per_second": 1.228, |
|
"eval_sts-test_pearson_cosine": 0.7731513422471518, |
|
"eval_sts-test_pearson_dot": 0.5353916353979121, |
|
"eval_sts-test_pearson_euclidean": 0.7469853367274143, |
|
"eval_sts-test_pearson_manhattan": 0.7481039749789392, |
|
"eval_sts-test_pearson_max": 0.7731513422471518, |
|
"eval_sts-test_spearman_cosine": 0.7792255681202213, |
|
"eval_sts-test_spearman_dot": 0.5083392134944579, |
|
"eval_sts-test_spearman_euclidean": 0.7335576317363027, |
|
"eval_sts-test_spearman_manhattan": 0.7370337922992453, |
|
"eval_sts-test_spearman_max": 0.7792255681202213, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_vitaminc-pairs_loss": 4.8438720703125, |
|
"eval_vitaminc-pairs_runtime": 1.4283, |
|
"eval_vitaminc-pairs_samples_per_second": 116.218, |
|
"eval_vitaminc-pairs_steps_per_second": 2.8, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_sts-label_loss": 4.092858791351318, |
|
"eval_sts-label_runtime": 0.407, |
|
"eval_sts-label_samples_per_second": 491.445, |
|
"eval_sts-label_steps_per_second": 12.286, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_qnli-contrastive_loss": 0.15653812885284424, |
|
"eval_qnli-contrastive_runtime": 0.2813, |
|
"eval_qnli-contrastive_samples_per_second": 711.049, |
|
"eval_qnli-contrastive_steps_per_second": 17.776, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_scitail-pairs-qa_loss": 0.10110773146152496, |
|
"eval_scitail-pairs-qa_runtime": 1.0598, |
|
"eval_scitail-pairs-qa_samples_per_second": 188.718, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.718, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_scitail-pairs-pos_loss": 0.46982836723327637, |
|
"eval_scitail-pairs-pos_runtime": 2.3471, |
|
"eval_scitail-pairs-pos_samples_per_second": 85.213, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.13, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_xsum-pairs_loss": 0.3937150537967682, |
|
"eval_xsum-pairs_runtime": 1.0402, |
|
"eval_xsum-pairs_samples_per_second": 192.278, |
|
"eval_xsum-pairs_steps_per_second": 4.807, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_compression-pairs_loss": 0.16437682509422302, |
|
"eval_compression-pairs_runtime": 0.2405, |
|
"eval_compression-pairs_samples_per_second": 831.709, |
|
"eval_compression-pairs_steps_per_second": 20.793, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_sciq_pairs_loss": 0.3407261073589325, |
|
"eval_sciq_pairs_runtime": 9.0549, |
|
"eval_sciq_pairs_samples_per_second": 22.088, |
|
"eval_sciq_pairs_steps_per_second": 0.552, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_qasc_pairs_loss": 0.4374461770057678, |
|
"eval_qasc_pairs_runtime": 1.2131, |
|
"eval_qasc_pairs_samples_per_second": 164.864, |
|
"eval_qasc_pairs_steps_per_second": 4.122, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_openbookqa_pairs_loss": 1.935410737991333, |
|
"eval_openbookqa_pairs_runtime": 1.0551, |
|
"eval_openbookqa_pairs_samples_per_second": 189.556, |
|
"eval_openbookqa_pairs_steps_per_second": 4.739, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_msmarco_pairs_loss": 1.043725609779358, |
|
"eval_msmarco_pairs_runtime": 2.5259, |
|
"eval_msmarco_pairs_samples_per_second": 79.179, |
|
"eval_msmarco_pairs_steps_per_second": 1.979, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_nq_pairs_loss": 0.9651347398757935, |
|
"eval_nq_pairs_runtime": 5.6774, |
|
"eval_nq_pairs_samples_per_second": 35.227, |
|
"eval_nq_pairs_steps_per_second": 0.881, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_trivia_pairs_loss": 1.2430963516235352, |
|
"eval_trivia_pairs_runtime": 9.0564, |
|
"eval_trivia_pairs_samples_per_second": 22.084, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_quora_pairs_loss": 0.21621538698673248, |
|
"eval_quora_pairs_runtime": 0.6252, |
|
"eval_quora_pairs_samples_per_second": 319.908, |
|
"eval_quora_pairs_steps_per_second": 7.998, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_gooaq_pairs_loss": 0.8927494287490845, |
|
"eval_gooaq_pairs_runtime": 1.5775, |
|
"eval_gooaq_pairs_samples_per_second": 126.779, |
|
"eval_gooaq_pairs_steps_per_second": 3.169, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.427383939774153, |
|
"eval_mrpc_pairs_loss": 0.08728782832622528, |
|
"eval_mrpc_pairs_runtime": 0.2422, |
|
"eval_mrpc_pairs_samples_per_second": 825.773, |
|
"eval_mrpc_pairs_steps_per_second": 20.644, |
|
"step": 9101 |
|
}, |
|
{ |
|
"epoch": 1.4303638644918444, |
|
"grad_norm": 10.624916076660156, |
|
"learning_rate": 1.1894318588256689e-05, |
|
"loss": 1.2641, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 1.4378920953575909, |
|
"grad_norm": 16.24142074584961, |
|
"learning_rate": 1.1547069572392431e-05, |
|
"loss": 1.4688, |
|
"step": 9168 |
|
}, |
|
{ |
|
"epoch": 1.4454203262233376, |
|
"grad_norm": 18.161853790283203, |
|
"learning_rate": 1.120245140218616e-05, |
|
"loss": 1.2061, |
|
"step": 9216 |
|
}, |
|
{ |
|
"epoch": 1.452948557089084, |
|
"grad_norm": 9.601255416870117, |
|
"learning_rate": 1.0860616378631118e-05, |
|
"loss": 1.0768, |
|
"step": 9264 |
|
}, |
|
{ |
|
"epoch": 1.4604767879548306, |
|
"grad_norm": 10.885648727416992, |
|
"learning_rate": 1.0521715572733149e-05, |
|
"loss": 0.9741, |
|
"step": 9312 |
|
}, |
|
{ |
|
"epoch": 1.468005018820577, |
|
"grad_norm": 4.2853007316589355, |
|
"learning_rate": 1.018589875874622e-05, |
|
"loss": 0.9977, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 1.4755332496863236, |
|
"grad_norm": 14.765129089355469, |
|
"learning_rate": 9.853314347980975e-06, |
|
"loss": 1.308, |
|
"step": 9408 |
|
}, |
|
{ |
|
"epoch": 1.4830614805520703, |
|
"grad_norm": 0.6372986435890198, |
|
"learning_rate": 9.524109323215664e-06, |
|
"loss": 1.3859, |
|
"step": 9456 |
|
}, |
|
{ |
|
"epoch": 1.4905897114178168, |
|
"grad_norm": 5.864046573638916, |
|
"learning_rate": 9.19842917373838e-06, |
|
"loss": 0.9261, |
|
"step": 9504 |
|
}, |
|
{ |
|
"epoch": 1.4981179422835633, |
|
"grad_norm": 2.619722366333008, |
|
"learning_rate": 8.87641783104935e-06, |
|
"loss": 1.0473, |
|
"step": 9552 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_nli-pairs_loss": 1.0287954807281494, |
|
"eval_nli-pairs_runtime": 4.1819, |
|
"eval_nli-pairs_samples_per_second": 47.825, |
|
"eval_nli-pairs_steps_per_second": 1.196, |
|
"eval_sts-test_pearson_cosine": 0.772936269755451, |
|
"eval_sts-test_pearson_dot": 0.5408568837341375, |
|
"eval_sts-test_pearson_euclidean": 0.7431310400735589, |
|
"eval_sts-test_pearson_manhattan": 0.7441695396815142, |
|
"eval_sts-test_pearson_max": 0.772936269755451, |
|
"eval_sts-test_spearman_cosine": 0.7808770614234607, |
|
"eval_sts-test_spearman_dot": 0.5146216887404633, |
|
"eval_sts-test_spearman_euclidean": 0.7291900713010305, |
|
"eval_sts-test_spearman_manhattan": 0.732990460571053, |
|
"eval_sts-test_spearman_max": 0.7808770614234607, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_vitaminc-pairs_loss": 4.766480445861816, |
|
"eval_vitaminc-pairs_runtime": 1.5156, |
|
"eval_vitaminc-pairs_samples_per_second": 109.528, |
|
"eval_vitaminc-pairs_steps_per_second": 2.639, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_sts-label_loss": 4.150259494781494, |
|
"eval_sts-label_runtime": 0.4493, |
|
"eval_sts-label_samples_per_second": 445.153, |
|
"eval_sts-label_steps_per_second": 11.129, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_qnli-contrastive_loss": 0.1541556417942047, |
|
"eval_qnli-contrastive_runtime": 0.296, |
|
"eval_qnli-contrastive_samples_per_second": 675.668, |
|
"eval_qnli-contrastive_steps_per_second": 16.892, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_scitail-pairs-qa_loss": 0.1008567065000534, |
|
"eval_scitail-pairs-qa_runtime": 1.0997, |
|
"eval_scitail-pairs-qa_samples_per_second": 181.866, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.547, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_scitail-pairs-pos_loss": 0.47411486506462097, |
|
"eval_scitail-pairs-pos_runtime": 2.4599, |
|
"eval_scitail-pairs-pos_samples_per_second": 81.304, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.033, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_xsum-pairs_loss": 0.3825688064098358, |
|
"eval_xsum-pairs_runtime": 1.0466, |
|
"eval_xsum-pairs_samples_per_second": 191.088, |
|
"eval_xsum-pairs_steps_per_second": 4.777, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_compression-pairs_loss": 0.16390098631381989, |
|
"eval_compression-pairs_runtime": 0.2542, |
|
"eval_compression-pairs_samples_per_second": 786.663, |
|
"eval_compression-pairs_steps_per_second": 19.667, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_sciq_pairs_loss": 0.3327814042568207, |
|
"eval_sciq_pairs_runtime": 9.1285, |
|
"eval_sciq_pairs_samples_per_second": 21.909, |
|
"eval_sciq_pairs_steps_per_second": 0.548, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_qasc_pairs_loss": 0.42448753118515015, |
|
"eval_qasc_pairs_runtime": 1.238, |
|
"eval_qasc_pairs_samples_per_second": 161.554, |
|
"eval_qasc_pairs_steps_per_second": 4.039, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_openbookqa_pairs_loss": 1.9015610218048096, |
|
"eval_openbookqa_pairs_runtime": 1.0684, |
|
"eval_openbookqa_pairs_samples_per_second": 187.196, |
|
"eval_openbookqa_pairs_steps_per_second": 4.68, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_msmarco_pairs_loss": 1.049147605895996, |
|
"eval_msmarco_pairs_runtime": 2.5566, |
|
"eval_msmarco_pairs_samples_per_second": 78.23, |
|
"eval_msmarco_pairs_steps_per_second": 1.956, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_nq_pairs_loss": 0.938487708568573, |
|
"eval_nq_pairs_runtime": 5.7298, |
|
"eval_nq_pairs_samples_per_second": 34.905, |
|
"eval_nq_pairs_steps_per_second": 0.873, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_trivia_pairs_loss": 1.1729891300201416, |
|
"eval_trivia_pairs_runtime": 9.1105, |
|
"eval_trivia_pairs_samples_per_second": 21.953, |
|
"eval_trivia_pairs_steps_per_second": 0.549, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_quora_pairs_loss": 0.24502086639404297, |
|
"eval_quora_pairs_runtime": 0.6539, |
|
"eval_quora_pairs_samples_per_second": 305.842, |
|
"eval_quora_pairs_steps_per_second": 7.646, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_gooaq_pairs_loss": 0.8770759105682373, |
|
"eval_gooaq_pairs_runtime": 1.6351, |
|
"eval_gooaq_pairs_samples_per_second": 122.318, |
|
"eval_gooaq_pairs_steps_per_second": 3.058, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.502509410288582, |
|
"eval_mrpc_pairs_loss": 0.08623871207237244, |
|
"eval_mrpc_pairs_runtime": 0.2404, |
|
"eval_mrpc_pairs_samples_per_second": 832.077, |
|
"eval_mrpc_pairs_steps_per_second": 20.802, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.50564617314931, |
|
"grad_norm": 11.285247802734375, |
|
"learning_rate": 8.558217605251645e-06, |
|
"loss": 1.1806, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 1.5131744040150563, |
|
"grad_norm": 10.830084800720215, |
|
"learning_rate": 8.24396912215848e-06, |
|
"loss": 1.1989, |
|
"step": 9648 |
|
}, |
|
{ |
|
"epoch": 1.520702634880803, |
|
"grad_norm": 9.796110153198242, |
|
"learning_rate": 7.933811261144869e-06, |
|
"loss": 1.0403, |
|
"step": 9696 |
|
}, |
|
{ |
|
"epoch": 1.5282308657465495, |
|
"grad_norm": 0.9950674176216125, |
|
"learning_rate": 7.627881093771088e-06, |
|
"loss": 1.0811, |
|
"step": 9744 |
|
}, |
|
{ |
|
"epoch": 1.535759096612296, |
|
"grad_norm": 11.353304862976074, |
|
"learning_rate": 7.326313823205104e-06, |
|
"loss": 1.3524, |
|
"step": 9792 |
|
}, |
|
{ |
|
"epoch": 1.5432873274780428, |
|
"grad_norm": 12.985937118530273, |
|
"learning_rate": 7.029242724470705e-06, |
|
"loss": 0.9578, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 1.5508155583437893, |
|
"grad_norm": 8.885215759277344, |
|
"learning_rate": 6.736799085547775e-06, |
|
"loss": 1.2745, |
|
"step": 9888 |
|
}, |
|
{ |
|
"epoch": 1.5583437892095358, |
|
"grad_norm": 0.3098331391811371, |
|
"learning_rate": 6.4491121493507095e-06, |
|
"loss": 1.0615, |
|
"step": 9936 |
|
}, |
|
{ |
|
"epoch": 1.5658720200752823, |
|
"grad_norm": 14.668235778808594, |
|
"learning_rate": 6.1663090566106425e-06, |
|
"loss": 0.9778, |
|
"step": 9984 |
|
}, |
|
{ |
|
"epoch": 1.5734002509410288, |
|
"grad_norm": 14.634328842163086, |
|
"learning_rate": 5.888514789686705e-06, |
|
"loss": 1.017, |
|
"step": 10032 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_nli-pairs_loss": 1.018608808517456, |
|
"eval_nli-pairs_runtime": 3.9852, |
|
"eval_nli-pairs_samples_per_second": 50.186, |
|
"eval_nli-pairs_steps_per_second": 1.255, |
|
"eval_sts-test_pearson_cosine": 0.7763527186202379, |
|
"eval_sts-test_pearson_dot": 0.5518277089774203, |
|
"eval_sts-test_pearson_euclidean": 0.7484152647439328, |
|
"eval_sts-test_pearson_manhattan": 0.7486588645110376, |
|
"eval_sts-test_pearson_max": 0.7763527186202379, |
|
"eval_sts-test_spearman_cosine": 0.7862296389332735, |
|
"eval_sts-test_spearman_dot": 0.5246898972209683, |
|
"eval_sts-test_spearman_euclidean": 0.7340067918071436, |
|
"eval_sts-test_spearman_manhattan": 0.7368940903638915, |
|
"eval_sts-test_spearman_max": 0.7862296389332735, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_vitaminc-pairs_loss": 4.802022457122803, |
|
"eval_vitaminc-pairs_runtime": 1.4418, |
|
"eval_vitaminc-pairs_samples_per_second": 115.137, |
|
"eval_vitaminc-pairs_steps_per_second": 2.774, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_sts-label_loss": 4.154313087463379, |
|
"eval_sts-label_runtime": 0.4056, |
|
"eval_sts-label_samples_per_second": 493.08, |
|
"eval_sts-label_steps_per_second": 12.327, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_qnli-contrastive_loss": 0.1918843388557434, |
|
"eval_qnli-contrastive_runtime": 0.2808, |
|
"eval_qnli-contrastive_samples_per_second": 712.295, |
|
"eval_qnli-contrastive_steps_per_second": 17.807, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_scitail-pairs-qa_loss": 0.10166393220424652, |
|
"eval_scitail-pairs-qa_runtime": 1.0544, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.685, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.742, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_scitail-pairs-pos_loss": 0.4371533691883087, |
|
"eval_scitail-pairs-pos_runtime": 2.4374, |
|
"eval_scitail-pairs-pos_samples_per_second": 82.054, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.051, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_xsum-pairs_loss": 0.3739396035671234, |
|
"eval_xsum-pairs_runtime": 1.0423, |
|
"eval_xsum-pairs_samples_per_second": 191.881, |
|
"eval_xsum-pairs_steps_per_second": 4.797, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_compression-pairs_loss": 0.16310901939868927, |
|
"eval_compression-pairs_runtime": 0.2376, |
|
"eval_compression-pairs_samples_per_second": 841.702, |
|
"eval_compression-pairs_steps_per_second": 21.043, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_sciq_pairs_loss": 0.3224416971206665, |
|
"eval_sciq_pairs_runtime": 9.0722, |
|
"eval_sciq_pairs_samples_per_second": 22.045, |
|
"eval_sciq_pairs_steps_per_second": 0.551, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_qasc_pairs_loss": 0.4264788031578064, |
|
"eval_qasc_pairs_runtime": 1.2568, |
|
"eval_qasc_pairs_samples_per_second": 159.131, |
|
"eval_qasc_pairs_steps_per_second": 3.978, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_openbookqa_pairs_loss": 1.824275016784668, |
|
"eval_openbookqa_pairs_runtime": 1.0491, |
|
"eval_openbookqa_pairs_samples_per_second": 190.648, |
|
"eval_openbookqa_pairs_steps_per_second": 4.766, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_msmarco_pairs_loss": 1.0266730785369873, |
|
"eval_msmarco_pairs_runtime": 2.5174, |
|
"eval_msmarco_pairs_samples_per_second": 79.447, |
|
"eval_msmarco_pairs_steps_per_second": 1.986, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_nq_pairs_loss": 0.9473356604576111, |
|
"eval_nq_pairs_runtime": 5.6496, |
|
"eval_nq_pairs_samples_per_second": 35.401, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_trivia_pairs_loss": 1.104791522026062, |
|
"eval_trivia_pairs_runtime": 9.073, |
|
"eval_trivia_pairs_samples_per_second": 22.044, |
|
"eval_trivia_pairs_steps_per_second": 0.551, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_quora_pairs_loss": 0.2634066939353943, |
|
"eval_quora_pairs_runtime": 0.6209, |
|
"eval_quora_pairs_samples_per_second": 322.127, |
|
"eval_quora_pairs_steps_per_second": 8.053, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_gooaq_pairs_loss": 0.8539897203445435, |
|
"eval_gooaq_pairs_runtime": 1.5537, |
|
"eval_gooaq_pairs_samples_per_second": 128.725, |
|
"eval_gooaq_pairs_steps_per_second": 3.218, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5776348808030112, |
|
"eval_mrpc_pairs_loss": 0.08671045303344727, |
|
"eval_mrpc_pairs_runtime": 0.237, |
|
"eval_mrpc_pairs_samples_per_second": 843.81, |
|
"eval_mrpc_pairs_steps_per_second": 21.095, |
|
"step": 10059 |
|
}, |
|
{ |
|
"epoch": 1.5809284818067755, |
|
"grad_norm": 4.454440593719482, |
|
"learning_rate": 5.615852117331175e-06, |
|
"loss": 1.3413, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 1.588456712672522, |
|
"grad_norm": 0.9061813950538635, |
|
"learning_rate": 5.348441540432878e-06, |
|
"loss": 1.1091, |
|
"step": 10128 |
|
}, |
|
{ |
|
"epoch": 1.5959849435382685, |
|
"grad_norm": 5.169127464294434, |
|
"learning_rate": 5.086401238762887e-06, |
|
"loss": 1.3101, |
|
"step": 10176 |
|
}, |
|
{ |
|
"epoch": 1.6035131744040152, |
|
"grad_norm": 4.494983196258545, |
|
"learning_rate": 4.8298470187459895e-06, |
|
"loss": 1.1126, |
|
"step": 10224 |
|
}, |
|
{ |
|
"epoch": 1.6110414052697615, |
|
"grad_norm": 0.4734310507774353, |
|
"learning_rate": 4.578892262281069e-06, |
|
"loss": 0.7667, |
|
"step": 10272 |
|
}, |
|
{ |
|
"epoch": 1.6185696361355082, |
|
"grad_norm": 0.4003587067127228, |
|
"learning_rate": 4.333647876632947e-06, |
|
"loss": 1.0281, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 1.6260978670012547, |
|
"grad_norm": 1.0558795928955078, |
|
"learning_rate": 4.094222245417886e-06, |
|
"loss": 1.2234, |
|
"step": 10368 |
|
}, |
|
{ |
|
"epoch": 1.6336260978670012, |
|
"grad_norm": 1.4687271118164062, |
|
"learning_rate": 3.860721180704409e-06, |
|
"loss": 0.9402, |
|
"step": 10416 |
|
}, |
|
{ |
|
"epoch": 1.641154328732748, |
|
"grad_norm": 16.27718734741211, |
|
"learning_rate": 3.6332478762505653e-06, |
|
"loss": 0.9698, |
|
"step": 10464 |
|
}, |
|
{ |
|
"epoch": 1.6486825595984942, |
|
"grad_norm": 12.507079124450684, |
|
"learning_rate": 3.411902861898354e-06, |
|
"loss": 1.1373, |
|
"step": 10512 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_nli-pairs_loss": 1.0013196468353271, |
|
"eval_nli-pairs_runtime": 4.0021, |
|
"eval_nli-pairs_samples_per_second": 49.974, |
|
"eval_nli-pairs_steps_per_second": 1.249, |
|
"eval_sts-test_pearson_cosine": 0.7756730383098025, |
|
"eval_sts-test_pearson_dot": 0.547730552868467, |
|
"eval_sts-test_pearson_euclidean": 0.7481152916902115, |
|
"eval_sts-test_pearson_manhattan": 0.7486290345837316, |
|
"eval_sts-test_pearson_max": 0.7756730383098025, |
|
"eval_sts-test_spearman_cosine": 0.7844179279938138, |
|
"eval_sts-test_spearman_dot": 0.5204811914426826, |
|
"eval_sts-test_spearman_euclidean": 0.7335913694810947, |
|
"eval_sts-test_spearman_manhattan": 0.7368797089125716, |
|
"eval_sts-test_spearman_max": 0.7844179279938138, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_vitaminc-pairs_loss": 4.780285835266113, |
|
"eval_vitaminc-pairs_runtime": 1.4992, |
|
"eval_vitaminc-pairs_samples_per_second": 110.723, |
|
"eval_vitaminc-pairs_steps_per_second": 2.668, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_sts-label_loss": 4.135310173034668, |
|
"eval_sts-label_runtime": 0.4056, |
|
"eval_sts-label_samples_per_second": 493.077, |
|
"eval_sts-label_steps_per_second": 12.327, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_qnli-contrastive_loss": 0.18042831122875214, |
|
"eval_qnli-contrastive_runtime": 0.2833, |
|
"eval_qnli-contrastive_samples_per_second": 706.004, |
|
"eval_qnli-contrastive_steps_per_second": 17.65, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_scitail-pairs-qa_loss": 0.09958843886852264, |
|
"eval_scitail-pairs-qa_runtime": 1.0443, |
|
"eval_scitail-pairs-qa_samples_per_second": 191.524, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.788, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_scitail-pairs-pos_loss": 0.4427280128002167, |
|
"eval_scitail-pairs-pos_runtime": 2.3552, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.918, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.123, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_xsum-pairs_loss": 0.3688097298145294, |
|
"eval_xsum-pairs_runtime": 1.041, |
|
"eval_xsum-pairs_samples_per_second": 192.117, |
|
"eval_xsum-pairs_steps_per_second": 4.803, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_compression-pairs_loss": 0.15796488523483276, |
|
"eval_compression-pairs_runtime": 0.2407, |
|
"eval_compression-pairs_samples_per_second": 830.774, |
|
"eval_compression-pairs_steps_per_second": 20.769, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_sciq_pairs_loss": 0.3223775029182434, |
|
"eval_sciq_pairs_runtime": 9.094, |
|
"eval_sciq_pairs_samples_per_second": 21.992, |
|
"eval_sciq_pairs_steps_per_second": 0.55, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_qasc_pairs_loss": 0.4077293872833252, |
|
"eval_qasc_pairs_runtime": 1.2194, |
|
"eval_qasc_pairs_samples_per_second": 164.013, |
|
"eval_qasc_pairs_steps_per_second": 4.1, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_openbookqa_pairs_loss": 1.8293620347976685, |
|
"eval_openbookqa_pairs_runtime": 1.0475, |
|
"eval_openbookqa_pairs_samples_per_second": 190.939, |
|
"eval_openbookqa_pairs_steps_per_second": 4.773, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_msmarco_pairs_loss": 1.0322593450546265, |
|
"eval_msmarco_pairs_runtime": 2.5279, |
|
"eval_msmarco_pairs_samples_per_second": 79.117, |
|
"eval_msmarco_pairs_steps_per_second": 1.978, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_nq_pairs_loss": 0.9389599561691284, |
|
"eval_nq_pairs_runtime": 5.6466, |
|
"eval_nq_pairs_samples_per_second": 35.42, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_trivia_pairs_loss": 1.1042495965957642, |
|
"eval_trivia_pairs_runtime": 9.0489, |
|
"eval_trivia_pairs_samples_per_second": 22.102, |
|
"eval_trivia_pairs_steps_per_second": 0.553, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_quora_pairs_loss": 0.17337936162948608, |
|
"eval_quora_pairs_runtime": 0.6556, |
|
"eval_quora_pairs_samples_per_second": 305.082, |
|
"eval_quora_pairs_steps_per_second": 7.627, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_gooaq_pairs_loss": 0.8494808673858643, |
|
"eval_gooaq_pairs_runtime": 1.5517, |
|
"eval_gooaq_pairs_samples_per_second": 128.894, |
|
"eval_gooaq_pairs_steps_per_second": 3.222, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.6527603513174403, |
|
"eval_mrpc_pairs_loss": 0.08429060131311417, |
|
"eval_mrpc_pairs_runtime": 0.2427, |
|
"eval_mrpc_pairs_samples_per_second": 824.181, |
|
"eval_mrpc_pairs_steps_per_second": 20.605, |
|
"step": 10538 |
|
}, |
|
{ |
|
"epoch": 1.656210790464241, |
|
"grad_norm": 9.579873085021973, |
|
"learning_rate": 3.196783959145439e-06, |
|
"loss": 1.433, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 1.6637390213299874, |
|
"grad_norm": 20.199016571044922, |
|
"learning_rate": 2.9879862379138003e-06, |
|
"loss": 0.7482, |
|
"step": 10608 |
|
}, |
|
{ |
|
"epoch": 1.671267252195734, |
|
"grad_norm": 0.24710102379322052, |
|
"learning_rate": 2.7897522915539458e-06, |
|
"loss": 1.1968, |
|
"step": 10656 |
|
}, |
|
{ |
|
"epoch": 1.6787954830614806, |
|
"grad_norm": 9.337010383605957, |
|
"learning_rate": 2.5937345579013605e-06, |
|
"loss": 1.2424, |
|
"step": 10704 |
|
}, |
|
{ |
|
"epoch": 1.6863237139272271, |
|
"grad_norm": 13.432022094726562, |
|
"learning_rate": 2.4043045181907913e-06, |
|
"loss": 1.379, |
|
"step": 10752 |
|
}, |
|
{ |
|
"epoch": 1.6938519447929736, |
|
"grad_norm": 8.947489738464355, |
|
"learning_rate": 2.2215458893753594e-06, |
|
"loss": 1.4127, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 1.7013801756587204, |
|
"grad_norm": 16.45635414123535, |
|
"learning_rate": 2.045539440036334e-06, |
|
"loss": 1.203, |
|
"step": 10848 |
|
}, |
|
{ |
|
"epoch": 1.7089084065244666, |
|
"grad_norm": 15.3703031539917, |
|
"learning_rate": 1.8763629546881908e-06, |
|
"loss": 0.7367, |
|
"step": 10896 |
|
}, |
|
{ |
|
"epoch": 1.7164366373902133, |
|
"grad_norm": 4.174514293670654, |
|
"learning_rate": 1.714091199402421e-06, |
|
"loss": 0.9207, |
|
"step": 10944 |
|
}, |
|
{ |
|
"epoch": 1.7239648682559598, |
|
"grad_norm": 2.0591108798980713, |
|
"learning_rate": 1.558795888765306e-06, |
|
"loss": 1.0067, |
|
"step": 10992 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_nli-pairs_loss": 0.9888688325881958, |
|
"eval_nli-pairs_runtime": 3.9929, |
|
"eval_nli-pairs_samples_per_second": 50.089, |
|
"eval_nli-pairs_steps_per_second": 1.252, |
|
"eval_sts-test_pearson_cosine": 0.7760455578346633, |
|
"eval_sts-test_pearson_dot": 0.5470527605127045, |
|
"eval_sts-test_pearson_euclidean": 0.7462734713816128, |
|
"eval_sts-test_pearson_manhattan": 0.7469300157882987, |
|
"eval_sts-test_pearson_max": 0.7760455578346633, |
|
"eval_sts-test_spearman_cosine": 0.7843247414472263, |
|
"eval_sts-test_spearman_dot": 0.5195687627478929, |
|
"eval_sts-test_spearman_euclidean": 0.7314459897716622, |
|
"eval_sts-test_spearman_manhattan": 0.7349029838686436, |
|
"eval_sts-test_spearman_max": 0.7843247414472263, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_vitaminc-pairs_loss": 4.805673122406006, |
|
"eval_vitaminc-pairs_runtime": 1.4386, |
|
"eval_vitaminc-pairs_samples_per_second": 115.388, |
|
"eval_vitaminc-pairs_steps_per_second": 2.78, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_sts-label_loss": 4.089193344116211, |
|
"eval_sts-label_runtime": 0.4049, |
|
"eval_sts-label_samples_per_second": 493.975, |
|
"eval_sts-label_steps_per_second": 12.349, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_qnli-contrastive_loss": 0.16640476882457733, |
|
"eval_qnli-contrastive_runtime": 0.2871, |
|
"eval_qnli-contrastive_samples_per_second": 696.519, |
|
"eval_qnli-contrastive_steps_per_second": 17.413, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_scitail-pairs-qa_loss": 0.09589708596467972, |
|
"eval_scitail-pairs-qa_runtime": 1.0968, |
|
"eval_scitail-pairs-qa_samples_per_second": 182.341, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.559, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_scitail-pairs-pos_loss": 0.44937801361083984, |
|
"eval_scitail-pairs-pos_runtime": 2.4011, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.296, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.082, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_xsum-pairs_loss": 0.3645179867744446, |
|
"eval_xsum-pairs_runtime": 1.0406, |
|
"eval_xsum-pairs_samples_per_second": 192.196, |
|
"eval_xsum-pairs_steps_per_second": 4.805, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_compression-pairs_loss": 0.15408411622047424, |
|
"eval_compression-pairs_runtime": 0.2368, |
|
"eval_compression-pairs_samples_per_second": 844.756, |
|
"eval_compression-pairs_steps_per_second": 21.119, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_sciq_pairs_loss": 0.3195297122001648, |
|
"eval_sciq_pairs_runtime": 9.0709, |
|
"eval_sciq_pairs_samples_per_second": 22.049, |
|
"eval_sciq_pairs_steps_per_second": 0.551, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_qasc_pairs_loss": 0.40569430589675903, |
|
"eval_qasc_pairs_runtime": 1.2166, |
|
"eval_qasc_pairs_samples_per_second": 164.395, |
|
"eval_qasc_pairs_steps_per_second": 4.11, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_openbookqa_pairs_loss": 1.8376811742782593, |
|
"eval_openbookqa_pairs_runtime": 1.0469, |
|
"eval_openbookqa_pairs_samples_per_second": 191.033, |
|
"eval_openbookqa_pairs_steps_per_second": 4.776, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_msmarco_pairs_loss": 1.0271800756454468, |
|
"eval_msmarco_pairs_runtime": 2.5127, |
|
"eval_msmarco_pairs_samples_per_second": 79.597, |
|
"eval_msmarco_pairs_steps_per_second": 1.99, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_nq_pairs_loss": 0.9162300229072571, |
|
"eval_nq_pairs_runtime": 5.647, |
|
"eval_nq_pairs_samples_per_second": 35.417, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_trivia_pairs_loss": 1.0903488397598267, |
|
"eval_trivia_pairs_runtime": 9.0283, |
|
"eval_trivia_pairs_samples_per_second": 22.153, |
|
"eval_trivia_pairs_steps_per_second": 0.554, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_quora_pairs_loss": 0.19430270791053772, |
|
"eval_quora_pairs_runtime": 0.6239, |
|
"eval_quora_pairs_samples_per_second": 320.54, |
|
"eval_quora_pairs_steps_per_second": 8.013, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_gooaq_pairs_loss": 0.8347020149230957, |
|
"eval_gooaq_pairs_runtime": 1.5623, |
|
"eval_gooaq_pairs_samples_per_second": 128.013, |
|
"eval_gooaq_pairs_steps_per_second": 3.2, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7278858218318696, |
|
"eval_mrpc_pairs_loss": 0.08172078430652618, |
|
"eval_mrpc_pairs_runtime": 0.2448, |
|
"eval_mrpc_pairs_samples_per_second": 816.98, |
|
"eval_mrpc_pairs_steps_per_second": 20.425, |
|
"step": 11017 |
|
}, |
|
{ |
|
"epoch": 1.7314930991217063, |
|
"grad_norm": 16.512481689453125, |
|
"learning_rate": 1.410545654184303e-06, |
|
"loss": 0.8782, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 1.739021329987453, |
|
"grad_norm": 15.815868377685547, |
|
"learning_rate": 1.2694060135569684e-06, |
|
"loss": 1.1617, |
|
"step": 11088 |
|
}, |
|
{ |
|
"epoch": 1.7465495608531993, |
|
"grad_norm": 0.39903897047042847, |
|
"learning_rate": 1.135439342315913e-06, |
|
"loss": 1.0122, |
|
"step": 11136 |
|
}, |
|
{ |
|
"epoch": 1.754077791718946, |
|
"grad_norm": 7.398995876312256, |
|
"learning_rate": 1.0087048458625138e-06, |
|
"loss": 0.6694, |
|
"step": 11184 |
|
}, |
|
{ |
|
"epoch": 1.7616060225846926, |
|
"grad_norm": 10.535834312438965, |
|
"learning_rate": 8.892585334016114e-07, |
|
"loss": 1.1209, |
|
"step": 11232 |
|
}, |
|
{ |
|
"epoch": 1.769134253450439, |
|
"grad_norm": 14.994894027709961, |
|
"learning_rate": 7.771531931887249e-07, |
|
"loss": 1.1, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 1.7766624843161858, |
|
"grad_norm": 29.789087295532227, |
|
"learning_rate": 6.724383692007457e-07, |
|
"loss": 0.9796, |
|
"step": 11328 |
|
}, |
|
{ |
|
"epoch": 1.7841907151819323, |
|
"grad_norm": 16.97585105895996, |
|
"learning_rate": 5.751603392404059e-07, |
|
"loss": 0.9533, |
|
"step": 11376 |
|
}, |
|
{ |
|
"epoch": 1.7917189460476788, |
|
"grad_norm": 36.554412841796875, |
|
"learning_rate": 4.853620944842135e-07, |
|
"loss": 0.9207, |
|
"step": 11424 |
|
}, |
|
{ |
|
"epoch": 1.7992471769134255, |
|
"grad_norm": 12.20507526397705, |
|
"learning_rate": 4.030833204828687e-07, |
|
"loss": 1.0267, |
|
"step": 11472 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_nli-pairs_loss": 0.9834117293357849, |
|
"eval_nli-pairs_runtime": 4.454, |
|
"eval_nli-pairs_samples_per_second": 44.904, |
|
"eval_nli-pairs_steps_per_second": 1.123, |
|
"eval_sts-test_pearson_cosine": 0.7757193524880511, |
|
"eval_sts-test_pearson_dot": 0.5461517523115149, |
|
"eval_sts-test_pearson_euclidean": 0.7464325603201095, |
|
"eval_sts-test_pearson_manhattan": 0.7470492681562052, |
|
"eval_sts-test_pearson_max": 0.7757193524880511, |
|
"eval_sts-test_spearman_cosine": 0.7841895178722113, |
|
"eval_sts-test_spearman_dot": 0.5190518591730641, |
|
"eval_sts-test_spearman_euclidean": 0.7317065051907755, |
|
"eval_sts-test_spearman_manhattan": 0.7348866670601053, |
|
"eval_sts-test_spearman_max": 0.7841895178722113, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_vitaminc-pairs_loss": 4.804275989532471, |
|
"eval_vitaminc-pairs_runtime": 1.4573, |
|
"eval_vitaminc-pairs_samples_per_second": 113.908, |
|
"eval_vitaminc-pairs_steps_per_second": 2.745, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_sts-label_loss": 4.131558418273926, |
|
"eval_sts-label_runtime": 0.4133, |
|
"eval_sts-label_samples_per_second": 483.878, |
|
"eval_sts-label_steps_per_second": 12.097, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_qnli-contrastive_loss": 0.16553626954555511, |
|
"eval_qnli-contrastive_runtime": 0.2817, |
|
"eval_qnli-contrastive_samples_per_second": 709.973, |
|
"eval_qnli-contrastive_steps_per_second": 17.749, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_scitail-pairs-qa_loss": 0.09527866542339325, |
|
"eval_scitail-pairs-qa_runtime": 1.0708, |
|
"eval_scitail-pairs-qa_samples_per_second": 186.771, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.669, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_scitail-pairs-pos_loss": 0.446532279253006, |
|
"eval_scitail-pairs-pos_runtime": 2.3976, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.418, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.085, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_xsum-pairs_loss": 0.36277005076408386, |
|
"eval_xsum-pairs_runtime": 1.0497, |
|
"eval_xsum-pairs_samples_per_second": 190.534, |
|
"eval_xsum-pairs_steps_per_second": 4.763, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_compression-pairs_loss": 0.15356417000293732, |
|
"eval_compression-pairs_runtime": 0.2565, |
|
"eval_compression-pairs_samples_per_second": 779.676, |
|
"eval_compression-pairs_steps_per_second": 19.492, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_sciq_pairs_loss": 0.3197508752346039, |
|
"eval_sciq_pairs_runtime": 9.2253, |
|
"eval_sciq_pairs_samples_per_second": 21.679, |
|
"eval_sciq_pairs_steps_per_second": 0.542, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_qasc_pairs_loss": 0.40177223086357117, |
|
"eval_qasc_pairs_runtime": 1.2313, |
|
"eval_qasc_pairs_samples_per_second": 162.427, |
|
"eval_qasc_pairs_steps_per_second": 4.061, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_openbookqa_pairs_loss": 1.8343558311462402, |
|
"eval_openbookqa_pairs_runtime": 1.0933, |
|
"eval_openbookqa_pairs_samples_per_second": 182.939, |
|
"eval_openbookqa_pairs_steps_per_second": 4.573, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_msmarco_pairs_loss": 1.0230004787445068, |
|
"eval_msmarco_pairs_runtime": 2.5513, |
|
"eval_msmarco_pairs_samples_per_second": 78.392, |
|
"eval_msmarco_pairs_steps_per_second": 1.96, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_nq_pairs_loss": 0.9057186841964722, |
|
"eval_nq_pairs_runtime": 5.7628, |
|
"eval_nq_pairs_samples_per_second": 34.705, |
|
"eval_nq_pairs_steps_per_second": 0.868, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_trivia_pairs_loss": 1.0895284414291382, |
|
"eval_trivia_pairs_runtime": 9.1178, |
|
"eval_trivia_pairs_samples_per_second": 21.935, |
|
"eval_trivia_pairs_steps_per_second": 0.548, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_quora_pairs_loss": 0.1568813920021057, |
|
"eval_quora_pairs_runtime": 0.6394, |
|
"eval_quora_pairs_samples_per_second": 312.772, |
|
"eval_quora_pairs_steps_per_second": 7.819, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_gooaq_pairs_loss": 0.8383786082267761, |
|
"eval_gooaq_pairs_runtime": 1.5566, |
|
"eval_gooaq_pairs_samples_per_second": 128.488, |
|
"eval_gooaq_pairs_steps_per_second": 3.212, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8030112923462986, |
|
"eval_mrpc_pairs_loss": 0.08112096786499023, |
|
"eval_mrpc_pairs_runtime": 0.2439, |
|
"eval_mrpc_pairs_samples_per_second": 819.882, |
|
"eval_mrpc_pairs_steps_per_second": 20.497, |
|
"step": 11496 |
|
}, |
|
{ |
|
"epoch": 1.8067754077791718, |
|
"grad_norm": 24.458271026611328, |
|
"learning_rate": 3.2836037962258505e-07, |
|
"loss": 0.8635, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 1.8143036386449185, |
|
"grad_norm": 6.1133222579956055, |
|
"learning_rate": 2.612262950550589e-07, |
|
"loss": 0.9252, |
|
"step": 11568 |
|
}, |
|
{ |
|
"epoch": 1.821831869510665, |
|
"grad_norm": 10.278335571289062, |
|
"learning_rate": 2.0171073610316003e-07, |
|
"loss": 1.32, |
|
"step": 11616 |
|
}, |
|
{ |
|
"epoch": 1.8293601003764115, |
|
"grad_norm": 14.742138862609863, |
|
"learning_rate": 1.4984000514884242e-07, |
|
"loss": 1.3209, |
|
"step": 11664 |
|
}, |
|
{ |
|
"epoch": 1.8368883312421582, |
|
"grad_norm": 7.676764011383057, |
|
"learning_rate": 1.056370260090303e-07, |
|
"loss": 0.8251, |
|
"step": 11712 |
|
}, |
|
{ |
|
"epoch": 1.8444165621079045, |
|
"grad_norm": 1.2442536354064941, |
|
"learning_rate": 6.912133380464951e-08, |
|
"loss": 0.8179, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 1.8519447929736512, |
|
"grad_norm": 15.847275733947754, |
|
"learning_rate": 4.030906632723735e-08, |
|
"loss": 0.8666, |
|
"step": 11808 |
|
}, |
|
{ |
|
"epoch": 1.8594730238393977, |
|
"grad_norm": 0.36177679896354675, |
|
"learning_rate": 1.9212956906992372e-08, |
|
"loss": 0.9244, |
|
"step": 11856 |
|
}, |
|
{ |
|
"epoch": 1.8670012547051442, |
|
"grad_norm": 8.72006607055664, |
|
"learning_rate": 5.842328785392869e-09, |
|
"loss": 0.9468, |
|
"step": 11904 |
|
}, |
|
{ |
|
"epoch": 1.874529485570891, |
|
"grad_norm": 15.35616397857666, |
|
"learning_rate": 2.0309099486470748e-10, |
|
"loss": 1.2313, |
|
"step": 11952 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_nli-pairs_loss": 1.0213488340377808, |
|
"eval_nli-pairs_runtime": 3.9888, |
|
"eval_nli-pairs_samples_per_second": 50.14, |
|
"eval_nli-pairs_steps_per_second": 1.254, |
|
"eval_sts-test_pearson_cosine": 0.7751861574983592, |
|
"eval_sts-test_pearson_dot": 0.5406109395741264, |
|
"eval_sts-test_pearson_euclidean": 0.7432969762219094, |
|
"eval_sts-test_pearson_manhattan": 0.7450333218883929, |
|
"eval_sts-test_pearson_max": 0.7751861574983592, |
|
"eval_sts-test_spearman_cosine": 0.7794274058735736, |
|
"eval_sts-test_spearman_dot": 0.5144540377132975, |
|
"eval_sts-test_spearman_euclidean": 0.727450852531023, |
|
"eval_sts-test_spearman_manhattan": 0.7318298848990707, |
|
"eval_sts-test_spearman_max": 0.7794274058735736, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_vitaminc-pairs_loss": 4.819454193115234, |
|
"eval_vitaminc-pairs_runtime": 1.4385, |
|
"eval_vitaminc-pairs_samples_per_second": 115.395, |
|
"eval_vitaminc-pairs_steps_per_second": 2.781, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_sts-label_loss": 4.0922932624816895, |
|
"eval_sts-label_runtime": 0.403, |
|
"eval_sts-label_samples_per_second": 496.282, |
|
"eval_sts-label_steps_per_second": 12.407, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_qnli-contrastive_loss": 0.14450308680534363, |
|
"eval_qnli-contrastive_runtime": 0.2805, |
|
"eval_qnli-contrastive_samples_per_second": 712.965, |
|
"eval_qnli-contrastive_steps_per_second": 17.824, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_scitail-pairs-qa_loss": 0.0972411260008812, |
|
"eval_scitail-pairs-qa_runtime": 1.0535, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.845, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.746, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_scitail-pairs-pos_loss": 0.47073617577552795, |
|
"eval_scitail-pairs-pos_runtime": 2.4056, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.141, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.079, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_xsum-pairs_loss": 0.3865247964859009, |
|
"eval_xsum-pairs_runtime": 1.0467, |
|
"eval_xsum-pairs_samples_per_second": 191.083, |
|
"eval_xsum-pairs_steps_per_second": 4.777, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_compression-pairs_loss": 0.15423807501792908, |
|
"eval_compression-pairs_runtime": 0.2378, |
|
"eval_compression-pairs_samples_per_second": 841.003, |
|
"eval_compression-pairs_steps_per_second": 21.025, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_sciq_pairs_loss": 0.3360276520252228, |
|
"eval_sciq_pairs_runtime": 9.0552, |
|
"eval_sciq_pairs_samples_per_second": 22.087, |
|
"eval_sciq_pairs_steps_per_second": 0.552, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_qasc_pairs_loss": 0.40982764959335327, |
|
"eval_qasc_pairs_runtime": 1.2436, |
|
"eval_qasc_pairs_samples_per_second": 160.828, |
|
"eval_qasc_pairs_steps_per_second": 4.021, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_openbookqa_pairs_loss": 1.8729889392852783, |
|
"eval_openbookqa_pairs_runtime": 1.106, |
|
"eval_openbookqa_pairs_samples_per_second": 180.839, |
|
"eval_openbookqa_pairs_steps_per_second": 4.521, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_msmarco_pairs_loss": 1.0765292644500732, |
|
"eval_msmarco_pairs_runtime": 2.5194, |
|
"eval_msmarco_pairs_samples_per_second": 79.385, |
|
"eval_msmarco_pairs_steps_per_second": 1.985, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_nq_pairs_loss": 0.9723155498504639, |
|
"eval_nq_pairs_runtime": 5.6285, |
|
"eval_nq_pairs_samples_per_second": 35.534, |
|
"eval_nq_pairs_steps_per_second": 0.888, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_trivia_pairs_loss": 1.2097830772399902, |
|
"eval_trivia_pairs_runtime": 9.0412, |
|
"eval_trivia_pairs_samples_per_second": 22.121, |
|
"eval_trivia_pairs_steps_per_second": 0.553, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_quora_pairs_loss": 0.1777983158826828, |
|
"eval_quora_pairs_runtime": 0.6175, |
|
"eval_quora_pairs_samples_per_second": 323.895, |
|
"eval_quora_pairs_steps_per_second": 8.097, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_gooaq_pairs_loss": 0.8855485320091248, |
|
"eval_gooaq_pairs_runtime": 1.5417, |
|
"eval_gooaq_pairs_samples_per_second": 129.724, |
|
"eval_gooaq_pairs_steps_per_second": 3.243, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8781367628607277, |
|
"eval_mrpc_pairs_loss": 0.0837549939751625, |
|
"eval_mrpc_pairs_runtime": 0.2368, |
|
"eval_mrpc_pairs_samples_per_second": 844.455, |
|
"eval_mrpc_pairs_steps_per_second": 21.111, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 1.8820577164366374, |
|
"grad_norm": 6.167259216308594, |
|
"learning_rate": 3.4997702264252654e-05, |
|
"loss": 1.0019, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.889585947302384, |
|
"grad_norm": 0.49681809544563293, |
|
"learning_rate": 3.498787466266714e-05, |
|
"loss": 0.8415, |
|
"step": 12048 |
|
}, |
|
{ |
|
"epoch": 1.8971141781681304, |
|
"grad_norm": 11.792595863342285, |
|
"learning_rate": 3.4970318447471354e-05, |
|
"loss": 1.0103, |
|
"step": 12096 |
|
}, |
|
{ |
|
"epoch": 1.904642409033877, |
|
"grad_norm": 0.8563029766082764, |
|
"learning_rate": 3.494504137748141e-05, |
|
"loss": 1.1249, |
|
"step": 12144 |
|
}, |
|
{ |
|
"epoch": 1.9121706398996237, |
|
"grad_norm": 36.67128372192383, |
|
"learning_rate": 3.4912054623677835e-05, |
|
"loss": 1.2655, |
|
"step": 12192 |
|
}, |
|
{ |
|
"epoch": 1.9196988707653702, |
|
"grad_norm": 15.591288566589355, |
|
"learning_rate": 3.4871372764268616e-05, |
|
"loss": 0.8125, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 1.9272271016311167, |
|
"grad_norm": 0.8200180530548096, |
|
"learning_rate": 3.482301377824655e-05, |
|
"loss": 0.8559, |
|
"step": 12288 |
|
}, |
|
{ |
|
"epoch": 1.9347553324968634, |
|
"grad_norm": 0.6030488014221191, |
|
"learning_rate": 3.476699903744353e-05, |
|
"loss": 0.8416, |
|
"step": 12336 |
|
}, |
|
{ |
|
"epoch": 1.9422835633626097, |
|
"grad_norm": 17.357330322265625, |
|
"learning_rate": 3.470335329708547e-05, |
|
"loss": 1.0393, |
|
"step": 12384 |
|
}, |
|
{ |
|
"epoch": 1.9498117942283564, |
|
"grad_norm": 15.783160209655762, |
|
"learning_rate": 3.463210468485197e-05, |
|
"loss": 0.946, |
|
"step": 12432 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_nli-pairs_loss": 1.0603748559951782, |
|
"eval_nli-pairs_runtime": 4.0941, |
|
"eval_nli-pairs_samples_per_second": 48.85, |
|
"eval_nli-pairs_steps_per_second": 1.221, |
|
"eval_sts-test_pearson_cosine": 0.7721603126678054, |
|
"eval_sts-test_pearson_dot": 0.5349835988482088, |
|
"eval_sts-test_pearson_euclidean": 0.742188046420877, |
|
"eval_sts-test_pearson_manhattan": 0.740692747387156, |
|
"eval_sts-test_pearson_max": 0.7721603126678054, |
|
"eval_sts-test_spearman_cosine": 0.7864643989345994, |
|
"eval_sts-test_spearman_dot": 0.5098554969670107, |
|
"eval_sts-test_spearman_euclidean": 0.7318205135578197, |
|
"eval_sts-test_spearman_manhattan": 0.7339408946246045, |
|
"eval_sts-test_spearman_max": 0.7864643989345994, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_vitaminc-pairs_loss": 4.802134990692139, |
|
"eval_vitaminc-pairs_runtime": 1.4328, |
|
"eval_vitaminc-pairs_samples_per_second": 115.859, |
|
"eval_vitaminc-pairs_steps_per_second": 2.792, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_sts-label_loss": 4.164713382720947, |
|
"eval_sts-label_runtime": 0.4062, |
|
"eval_sts-label_samples_per_second": 492.426, |
|
"eval_sts-label_steps_per_second": 12.311, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_qnli-contrastive_loss": 0.1906559020280838, |
|
"eval_qnli-contrastive_runtime": 0.285, |
|
"eval_qnli-contrastive_samples_per_second": 701.782, |
|
"eval_qnli-contrastive_steps_per_second": 17.545, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_scitail-pairs-qa_loss": 0.09206719696521759, |
|
"eval_scitail-pairs-qa_runtime": 1.0605, |
|
"eval_scitail-pairs-qa_samples_per_second": 188.59, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.715, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_scitail-pairs-pos_loss": 0.4270685613155365, |
|
"eval_scitail-pairs-pos_runtime": 2.3609, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.714, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.118, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_xsum-pairs_loss": 0.386574923992157, |
|
"eval_xsum-pairs_runtime": 1.0761, |
|
"eval_xsum-pairs_samples_per_second": 185.848, |
|
"eval_xsum-pairs_steps_per_second": 4.646, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_compression-pairs_loss": 0.15844617784023285, |
|
"eval_compression-pairs_runtime": 0.2411, |
|
"eval_compression-pairs_samples_per_second": 829.616, |
|
"eval_compression-pairs_steps_per_second": 20.74, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_sciq_pairs_loss": 0.32742640376091003, |
|
"eval_sciq_pairs_runtime": 9.0687, |
|
"eval_sciq_pairs_samples_per_second": 22.054, |
|
"eval_sciq_pairs_steps_per_second": 0.551, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_qasc_pairs_loss": 0.44627976417541504, |
|
"eval_qasc_pairs_runtime": 1.2115, |
|
"eval_qasc_pairs_samples_per_second": 165.083, |
|
"eval_qasc_pairs_steps_per_second": 4.127, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_openbookqa_pairs_loss": 1.8450264930725098, |
|
"eval_openbookqa_pairs_runtime": 1.0468, |
|
"eval_openbookqa_pairs_samples_per_second": 191.053, |
|
"eval_openbookqa_pairs_steps_per_second": 4.776, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_msmarco_pairs_loss": 1.0895458459854126, |
|
"eval_msmarco_pairs_runtime": 2.5207, |
|
"eval_msmarco_pairs_samples_per_second": 79.343, |
|
"eval_msmarco_pairs_steps_per_second": 1.984, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_nq_pairs_loss": 0.9783583283424377, |
|
"eval_nq_pairs_runtime": 5.6317, |
|
"eval_nq_pairs_samples_per_second": 35.513, |
|
"eval_nq_pairs_steps_per_second": 0.888, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_trivia_pairs_loss": 1.1956011056900024, |
|
"eval_trivia_pairs_runtime": 9.0547, |
|
"eval_trivia_pairs_samples_per_second": 22.088, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_quora_pairs_loss": 0.2143821269273758, |
|
"eval_quora_pairs_runtime": 0.6206, |
|
"eval_quora_pairs_samples_per_second": 322.293, |
|
"eval_quora_pairs_steps_per_second": 8.057, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_gooaq_pairs_loss": 0.890216052532196, |
|
"eval_gooaq_pairs_runtime": 1.545, |
|
"eval_gooaq_pairs_samples_per_second": 129.45, |
|
"eval_gooaq_pairs_steps_per_second": 3.236, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9532622333751568, |
|
"eval_mrpc_pairs_loss": 0.08200729638338089, |
|
"eval_mrpc_pairs_runtime": 0.2402, |
|
"eval_mrpc_pairs_samples_per_second": 832.672, |
|
"eval_mrpc_pairs_steps_per_second": 20.817, |
|
"step": 12454 |
|
}, |
|
{ |
|
"epoch": 1.9573400250941029, |
|
"grad_norm": 9.203104972839355, |
|
"learning_rate": 3.455328468844549e-05, |
|
"loss": 1.0748, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 1.9648682559598494, |
|
"grad_norm": 1.351884126663208, |
|
"learning_rate": 3.4466928141675676e-05, |
|
"loss": 0.8413, |
|
"step": 12528 |
|
}, |
|
{ |
|
"epoch": 1.972396486825596, |
|
"grad_norm": 13.26588249206543, |
|
"learning_rate": 3.4375104727660065e-05, |
|
"loss": 1.2138, |
|
"step": 12576 |
|
}, |
|
{ |
|
"epoch": 1.9799247176913424, |
|
"grad_norm": 7.214226245880127, |
|
"learning_rate": 3.427394779454246e-05, |
|
"loss": 1.2554, |
|
"step": 12624 |
|
}, |
|
{ |
|
"epoch": 1.987452948557089, |
|
"grad_norm": 9.377580642700195, |
|
"learning_rate": 3.416537776156432e-05, |
|
"loss": 1.4695, |
|
"step": 12672 |
|
}, |
|
{ |
|
"epoch": 1.9949811794228356, |
|
"grad_norm": 14.431234359741211, |
|
"learning_rate": 3.4049442610304357e-05, |
|
"loss": 0.9206, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 2.002509410288582, |
|
"grad_norm": 39.02935028076172, |
|
"learning_rate": 3.392619357729103e-05, |
|
"loss": 0.9354, |
|
"step": 12768 |
|
}, |
|
{ |
|
"epoch": 2.010037641154329, |
|
"grad_norm": 0.6344715356826782, |
|
"learning_rate": 3.379568513135902e-05, |
|
"loss": 0.8492, |
|
"step": 12816 |
|
}, |
|
{ |
|
"epoch": 2.017565872020075, |
|
"grad_norm": 12.117080688476562, |
|
"learning_rate": 3.365797494957723e-05, |
|
"loss": 0.9937, |
|
"step": 12864 |
|
}, |
|
{ |
|
"epoch": 2.025094102885822, |
|
"grad_norm": 8.92581844329834, |
|
"learning_rate": 3.351312389175882e-05, |
|
"loss": 1.423, |
|
"step": 12912 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_nli-pairs_loss": 1.1507288217544556, |
|
"eval_nli-pairs_runtime": 3.9823, |
|
"eval_nli-pairs_samples_per_second": 50.222, |
|
"eval_nli-pairs_steps_per_second": 1.256, |
|
"eval_sts-test_pearson_cosine": 0.7722624362390709, |
|
"eval_sts-test_pearson_dot": 0.5303764984983903, |
|
"eval_sts-test_pearson_euclidean": 0.7411643880467264, |
|
"eval_sts-test_pearson_manhattan": 0.7391579789846456, |
|
"eval_sts-test_pearson_max": 0.7722624362390709, |
|
"eval_sts-test_spearman_cosine": 0.7821947805835431, |
|
"eval_sts-test_spearman_dot": 0.5068928313854526, |
|
"eval_sts-test_spearman_euclidean": 0.7301088773642002, |
|
"eval_sts-test_spearman_manhattan": 0.7326196614978934, |
|
"eval_sts-test_spearman_max": 0.7821947805835431, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_vitaminc-pairs_loss": 4.279026985168457, |
|
"eval_vitaminc-pairs_runtime": 1.4397, |
|
"eval_vitaminc-pairs_samples_per_second": 115.303, |
|
"eval_vitaminc-pairs_steps_per_second": 2.778, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_sts-label_loss": 4.452165603637695, |
|
"eval_sts-label_runtime": 0.4088, |
|
"eval_sts-label_samples_per_second": 489.236, |
|
"eval_sts-label_steps_per_second": 12.231, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_qnli-contrastive_loss": 0.17014659941196442, |
|
"eval_qnli-contrastive_runtime": 0.2882, |
|
"eval_qnli-contrastive_samples_per_second": 693.965, |
|
"eval_qnli-contrastive_steps_per_second": 17.349, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_scitail-pairs-qa_loss": 0.09296510368585587, |
|
"eval_scitail-pairs-qa_runtime": 1.0809, |
|
"eval_scitail-pairs-qa_samples_per_second": 185.023, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.626, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_scitail-pairs-pos_loss": 0.4562944173812866, |
|
"eval_scitail-pairs-pos_runtime": 2.376, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.175, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.104, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_xsum-pairs_loss": 0.3622417151927948, |
|
"eval_xsum-pairs_runtime": 1.0416, |
|
"eval_xsum-pairs_samples_per_second": 192.012, |
|
"eval_xsum-pairs_steps_per_second": 4.8, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_compression-pairs_loss": 0.15833701193332672, |
|
"eval_compression-pairs_runtime": 0.2407, |
|
"eval_compression-pairs_samples_per_second": 831.007, |
|
"eval_compression-pairs_steps_per_second": 20.775, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_sciq_pairs_loss": 0.3300960958003998, |
|
"eval_sciq_pairs_runtime": 9.1169, |
|
"eval_sciq_pairs_samples_per_second": 21.937, |
|
"eval_sciq_pairs_steps_per_second": 0.548, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_qasc_pairs_loss": 0.4440248906612396, |
|
"eval_qasc_pairs_runtime": 1.2186, |
|
"eval_qasc_pairs_samples_per_second": 164.116, |
|
"eval_qasc_pairs_steps_per_second": 4.103, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_openbookqa_pairs_loss": 1.9062250852584839, |
|
"eval_openbookqa_pairs_runtime": 1.054, |
|
"eval_openbookqa_pairs_samples_per_second": 189.748, |
|
"eval_openbookqa_pairs_steps_per_second": 4.744, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_msmarco_pairs_loss": 1.0610954761505127, |
|
"eval_msmarco_pairs_runtime": 2.5303, |
|
"eval_msmarco_pairs_samples_per_second": 79.042, |
|
"eval_msmarco_pairs_steps_per_second": 1.976, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_nq_pairs_loss": 0.9248062968254089, |
|
"eval_nq_pairs_runtime": 5.6477, |
|
"eval_nq_pairs_samples_per_second": 35.413, |
|
"eval_nq_pairs_steps_per_second": 0.885, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_trivia_pairs_loss": 1.1339099407196045, |
|
"eval_trivia_pairs_runtime": 9.046, |
|
"eval_trivia_pairs_samples_per_second": 22.109, |
|
"eval_trivia_pairs_steps_per_second": 0.553, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_quora_pairs_loss": 0.197435200214386, |
|
"eval_quora_pairs_runtime": 0.6181, |
|
"eval_quora_pairs_samples_per_second": 323.589, |
|
"eval_quora_pairs_steps_per_second": 8.09, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_gooaq_pairs_loss": 0.8682229518890381, |
|
"eval_gooaq_pairs_runtime": 1.5695, |
|
"eval_gooaq_pairs_samples_per_second": 127.427, |
|
"eval_gooaq_pairs_steps_per_second": 3.186, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.028387703889586, |
|
"eval_mrpc_pairs_loss": 0.08277301490306854, |
|
"eval_mrpc_pairs_runtime": 0.2472, |
|
"eval_mrpc_pairs_samples_per_second": 809.045, |
|
"eval_mrpc_pairs_steps_per_second": 20.226, |
|
"step": 12933 |
|
}, |
|
{ |
|
"epoch": 2.0326223337515685, |
|
"grad_norm": 8.625104904174805, |
|
"learning_rate": 3.3361195973564865e-05, |
|
"loss": 0.9107, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 2.040150564617315, |
|
"grad_norm": 11.012100219726562, |
|
"learning_rate": 3.320225833821311e-05, |
|
"loss": 0.8694, |
|
"step": 13008 |
|
}, |
|
{ |
|
"epoch": 2.0476787954830615, |
|
"grad_norm": 18.39250946044922, |
|
"learning_rate": 3.303638122680463e-05, |
|
"loss": 1.2955, |
|
"step": 13056 |
|
}, |
|
{ |
|
"epoch": 2.055207026348808, |
|
"grad_norm": 0.33739256858825684, |
|
"learning_rate": 3.286363794728137e-05, |
|
"loss": 0.7578, |
|
"step": 13104 |
|
}, |
|
{ |
|
"epoch": 2.0627352572145545, |
|
"grad_norm": 11.637800216674805, |
|
"learning_rate": 3.268410484202841e-05, |
|
"loss": 0.8352, |
|
"step": 13152 |
|
}, |
|
{ |
|
"epoch": 2.0702634880803013, |
|
"grad_norm": 3.6323413848876953, |
|
"learning_rate": 3.249786125413502e-05, |
|
"loss": 1.1223, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 2.0777917189460475, |
|
"grad_norm": 2.320704936981201, |
|
"learning_rate": 3.230498949232972e-05, |
|
"loss": 0.8278, |
|
"step": 13248 |
|
}, |
|
{ |
|
"epoch": 2.0853199498117942, |
|
"grad_norm": 19.040969848632812, |
|
"learning_rate": 3.2105574794604686e-05, |
|
"loss": 0.9415, |
|
"step": 13296 |
|
}, |
|
{ |
|
"epoch": 2.092848180677541, |
|
"grad_norm": 0.5109882950782776, |
|
"learning_rate": 3.18997052905455e-05, |
|
"loss": 1.017, |
|
"step": 13344 |
|
}, |
|
{ |
|
"epoch": 2.1003764115432872, |
|
"grad_norm": 15.141865730285645, |
|
"learning_rate": 3.1687471962383085e-05, |
|
"loss": 1.0765, |
|
"step": 13392 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_nli-pairs_loss": 1.051226019859314, |
|
"eval_nli-pairs_runtime": 4.0486, |
|
"eval_nli-pairs_samples_per_second": 49.4, |
|
"eval_nli-pairs_steps_per_second": 1.235, |
|
"eval_sts-test_pearson_cosine": 0.7812575172994014, |
|
"eval_sts-test_pearson_dot": 0.5319105750599413, |
|
"eval_sts-test_pearson_euclidean": 0.750747218773846, |
|
"eval_sts-test_pearson_manhattan": 0.7509656763912702, |
|
"eval_sts-test_pearson_max": 0.7812575172994014, |
|
"eval_sts-test_spearman_cosine": 0.7910115404740932, |
|
"eval_sts-test_spearman_dot": 0.510004063994186, |
|
"eval_sts-test_spearman_euclidean": 0.7384195619199833, |
|
"eval_sts-test_spearman_manhattan": 0.7411514762626671, |
|
"eval_sts-test_spearman_max": 0.7910115404740932, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_vitaminc-pairs_loss": 4.327915668487549, |
|
"eval_vitaminc-pairs_runtime": 1.4479, |
|
"eval_vitaminc-pairs_samples_per_second": 114.649, |
|
"eval_vitaminc-pairs_steps_per_second": 2.763, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_sts-label_loss": 4.202516078948975, |
|
"eval_sts-label_runtime": 0.4274, |
|
"eval_sts-label_samples_per_second": 467.911, |
|
"eval_sts-label_steps_per_second": 11.698, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_qnli-contrastive_loss": 0.17492428421974182, |
|
"eval_qnli-contrastive_runtime": 0.2825, |
|
"eval_qnli-contrastive_samples_per_second": 707.881, |
|
"eval_qnli-contrastive_steps_per_second": 17.697, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_scitail-pairs-qa_loss": 0.09003904461860657, |
|
"eval_scitail-pairs-qa_runtime": 1.0939, |
|
"eval_scitail-pairs-qa_samples_per_second": 182.836, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.571, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_scitail-pairs-pos_loss": 0.46629810333251953, |
|
"eval_scitail-pairs-pos_runtime": 2.4148, |
|
"eval_scitail-pairs-pos_samples_per_second": 82.823, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.071, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_xsum-pairs_loss": 0.34408459067344666, |
|
"eval_xsum-pairs_runtime": 1.0742, |
|
"eval_xsum-pairs_samples_per_second": 186.18, |
|
"eval_xsum-pairs_steps_per_second": 4.655, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_compression-pairs_loss": 0.1467471718788147, |
|
"eval_compression-pairs_runtime": 0.2637, |
|
"eval_compression-pairs_samples_per_second": 758.571, |
|
"eval_compression-pairs_steps_per_second": 18.964, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_sciq_pairs_loss": 0.31176942586898804, |
|
"eval_sciq_pairs_runtime": 9.1255, |
|
"eval_sciq_pairs_samples_per_second": 21.917, |
|
"eval_sciq_pairs_steps_per_second": 0.548, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_qasc_pairs_loss": 0.41609370708465576, |
|
"eval_qasc_pairs_runtime": 1.2726, |
|
"eval_qasc_pairs_samples_per_second": 157.157, |
|
"eval_qasc_pairs_steps_per_second": 3.929, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_openbookqa_pairs_loss": 1.8419994115829468, |
|
"eval_openbookqa_pairs_runtime": 1.1105, |
|
"eval_openbookqa_pairs_samples_per_second": 180.091, |
|
"eval_openbookqa_pairs_steps_per_second": 4.502, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_msmarco_pairs_loss": 1.1004538536071777, |
|
"eval_msmarco_pairs_runtime": 2.5657, |
|
"eval_msmarco_pairs_samples_per_second": 77.95, |
|
"eval_msmarco_pairs_steps_per_second": 1.949, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_nq_pairs_loss": 0.9002810716629028, |
|
"eval_nq_pairs_runtime": 5.6938, |
|
"eval_nq_pairs_samples_per_second": 35.126, |
|
"eval_nq_pairs_steps_per_second": 0.878, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_trivia_pairs_loss": 1.086058259010315, |
|
"eval_trivia_pairs_runtime": 9.1706, |
|
"eval_trivia_pairs_samples_per_second": 21.809, |
|
"eval_trivia_pairs_steps_per_second": 0.545, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_quora_pairs_loss": 0.1935713142156601, |
|
"eval_quora_pairs_runtime": 0.6501, |
|
"eval_quora_pairs_samples_per_second": 307.657, |
|
"eval_quora_pairs_steps_per_second": 7.691, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_gooaq_pairs_loss": 0.883912980556488, |
|
"eval_gooaq_pairs_runtime": 1.5812, |
|
"eval_gooaq_pairs_samples_per_second": 126.486, |
|
"eval_gooaq_pairs_steps_per_second": 3.162, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.103513174404015, |
|
"eval_mrpc_pairs_loss": 0.08029545843601227, |
|
"eval_mrpc_pairs_runtime": 0.2872, |
|
"eval_mrpc_pairs_samples_per_second": 696.382, |
|
"eval_mrpc_pairs_steps_per_second": 17.41, |
|
"step": 13412 |
|
}, |
|
{ |
|
"epoch": 2.107904642409034, |
|
"grad_norm": 0.8797653317451477, |
|
"learning_rate": 3.146896860478492e-05, |
|
"loss": 1.0399, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 2.1154328732747802, |
|
"grad_norm": 0.8809025287628174, |
|
"learning_rate": 3.1244291783403247e-05, |
|
"loss": 0.9078, |
|
"step": 13488 |
|
}, |
|
{ |
|
"epoch": 2.122961104140527, |
|
"grad_norm": 12.174256324768066, |
|
"learning_rate": 3.101354079219879e-05, |
|
"loss": 0.9414, |
|
"step": 13536 |
|
}, |
|
{ |
|
"epoch": 2.1304893350062737, |
|
"grad_norm": 7.465297698974609, |
|
"learning_rate": 3.07768176095586e-05, |
|
"loss": 0.9909, |
|
"step": 13584 |
|
}, |
|
{ |
|
"epoch": 2.13801756587202, |
|
"grad_norm": 13.727421760559082, |
|
"learning_rate": 3.053422685322763e-05, |
|
"loss": 1.1089, |
|
"step": 13632 |
|
}, |
|
{ |
|
"epoch": 2.1455457967377667, |
|
"grad_norm": 15.126276969909668, |
|
"learning_rate": 3.0285875734073832e-05, |
|
"loss": 1.315, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 2.1530740276035134, |
|
"grad_norm": 7.151478290557861, |
|
"learning_rate": 3.0031874008707226e-05, |
|
"loss": 0.8123, |
|
"step": 13728 |
|
}, |
|
{ |
|
"epoch": 2.1606022584692597, |
|
"grad_norm": 9.124503135681152, |
|
"learning_rate": 2.977233393097396e-05, |
|
"loss": 1.2184, |
|
"step": 13776 |
|
}, |
|
{ |
|
"epoch": 2.1681304893350064, |
|
"grad_norm": 14.512338638305664, |
|
"learning_rate": 2.9507370202346677e-05, |
|
"loss": 0.7468, |
|
"step": 13824 |
|
}, |
|
{ |
|
"epoch": 2.1756587202007527, |
|
"grad_norm": 7.921672344207764, |
|
"learning_rate": 2.9237099921233195e-05, |
|
"loss": 0.794, |
|
"step": 13872 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_nli-pairs_loss": 0.9923555254936218, |
|
"eval_nli-pairs_runtime": 4.0616, |
|
"eval_nli-pairs_samples_per_second": 49.242, |
|
"eval_nli-pairs_steps_per_second": 1.231, |
|
"eval_sts-test_pearson_cosine": 0.7764138225504212, |
|
"eval_sts-test_pearson_dot": 0.5280954278502723, |
|
"eval_sts-test_pearson_euclidean": 0.7424878084852355, |
|
"eval_sts-test_pearson_manhattan": 0.7418329788550313, |
|
"eval_sts-test_pearson_max": 0.7764138225504212, |
|
"eval_sts-test_spearman_cosine": 0.7855803224985884, |
|
"eval_sts-test_spearman_dot": 0.5061382013250837, |
|
"eval_sts-test_spearman_euclidean": 0.7273293251792637, |
|
"eval_sts-test_spearman_manhattan": 0.7299928706510171, |
|
"eval_sts-test_spearman_max": 0.7855803224985884, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_vitaminc-pairs_loss": 4.433829307556152, |
|
"eval_vitaminc-pairs_runtime": 1.4393, |
|
"eval_vitaminc-pairs_samples_per_second": 115.33, |
|
"eval_vitaminc-pairs_steps_per_second": 2.779, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_sts-label_loss": 4.203037738800049, |
|
"eval_sts-label_runtime": 0.4039, |
|
"eval_sts-label_samples_per_second": 495.17, |
|
"eval_sts-label_steps_per_second": 12.379, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_qnli-contrastive_loss": 0.14975911378860474, |
|
"eval_qnli-contrastive_runtime": 0.2846, |
|
"eval_qnli-contrastive_samples_per_second": 702.864, |
|
"eval_qnli-contrastive_steps_per_second": 17.572, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_scitail-pairs-qa_loss": 0.08012403547763824, |
|
"eval_scitail-pairs-qa_runtime": 1.0534, |
|
"eval_scitail-pairs-qa_samples_per_second": 189.868, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.747, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_scitail-pairs-pos_loss": 0.42602407932281494, |
|
"eval_scitail-pairs-pos_runtime": 2.3645, |
|
"eval_scitail-pairs-pos_samples_per_second": 84.586, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.115, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_xsum-pairs_loss": 0.3360922336578369, |
|
"eval_xsum-pairs_runtime": 1.0464, |
|
"eval_xsum-pairs_samples_per_second": 191.13, |
|
"eval_xsum-pairs_steps_per_second": 4.778, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_compression-pairs_loss": 0.1300394982099533, |
|
"eval_compression-pairs_runtime": 0.2447, |
|
"eval_compression-pairs_samples_per_second": 817.251, |
|
"eval_compression-pairs_steps_per_second": 20.431, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_sciq_pairs_loss": 0.3116128444671631, |
|
"eval_sciq_pairs_runtime": 9.0757, |
|
"eval_sciq_pairs_samples_per_second": 22.037, |
|
"eval_sciq_pairs_steps_per_second": 0.551, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_qasc_pairs_loss": 0.361092746257782, |
|
"eval_qasc_pairs_runtime": 1.2089, |
|
"eval_qasc_pairs_samples_per_second": 165.434, |
|
"eval_qasc_pairs_steps_per_second": 4.136, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_openbookqa_pairs_loss": 1.8503968715667725, |
|
"eval_openbookqa_pairs_runtime": 1.0605, |
|
"eval_openbookqa_pairs_samples_per_second": 188.596, |
|
"eval_openbookqa_pairs_steps_per_second": 4.715, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_msmarco_pairs_loss": 1.0155786275863647, |
|
"eval_msmarco_pairs_runtime": 2.5332, |
|
"eval_msmarco_pairs_samples_per_second": 78.953, |
|
"eval_msmarco_pairs_steps_per_second": 1.974, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_nq_pairs_loss": 0.8908740878105164, |
|
"eval_nq_pairs_runtime": 5.6668, |
|
"eval_nq_pairs_samples_per_second": 35.294, |
|
"eval_nq_pairs_steps_per_second": 0.882, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_trivia_pairs_loss": 1.1117126941680908, |
|
"eval_trivia_pairs_runtime": 9.1174, |
|
"eval_trivia_pairs_samples_per_second": 21.936, |
|
"eval_trivia_pairs_steps_per_second": 0.548, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_quora_pairs_loss": 0.10165992379188538, |
|
"eval_quora_pairs_runtime": 0.6259, |
|
"eval_quora_pairs_samples_per_second": 319.533, |
|
"eval_quora_pairs_steps_per_second": 7.988, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_gooaq_pairs_loss": 0.8515159487724304, |
|
"eval_gooaq_pairs_runtime": 1.555, |
|
"eval_gooaq_pairs_samples_per_second": 128.615, |
|
"eval_gooaq_pairs_steps_per_second": 3.215, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1786386449184443, |
|
"eval_mrpc_pairs_loss": 0.0670856311917305, |
|
"eval_mrpc_pairs_runtime": 0.2435, |
|
"eval_mrpc_pairs_samples_per_second": 821.504, |
|
"eval_mrpc_pairs_steps_per_second": 20.538, |
|
"step": 13891 |
|
}, |
|
{ |
|
"epoch": 2.1831869510664994, |
|
"grad_norm": 7.595925807952881, |
|
"learning_rate": 2.896164253122592e-05, |
|
"loss": 0.8892, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 2.190715181932246, |
|
"grad_norm": 0.5041866898536682, |
|
"learning_rate": 2.8681119768314744e-05, |
|
"loss": 0.7443, |
|
"step": 13968 |
|
}, |
|
{ |
|
"epoch": 2.1982434127979924, |
|
"grad_norm": 10.744353294372559, |
|
"learning_rate": 2.8395655607086955e-05, |
|
"loss": 0.6776, |
|
"step": 14016 |
|
}, |
|
{ |
|
"epoch": 2.205771643663739, |
|
"grad_norm": 13.672101974487305, |
|
"learning_rate": 2.810537620593767e-05, |
|
"loss": 1.2239, |
|
"step": 14064 |
|
}, |
|
{ |
|
"epoch": 2.2132998745294854, |
|
"grad_norm": 3.225550413131714, |
|
"learning_rate": 2.7810409851315294e-05, |
|
"loss": 0.8225, |
|
"step": 14112 |
|
}, |
|
{ |
|
"epoch": 2.220828105395232, |
|
"grad_norm": 3.210339069366455, |
|
"learning_rate": 2.7510886901026408e-05, |
|
"loss": 1.2052, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 2.228356336260979, |
|
"grad_norm": 6.894575119018555, |
|
"learning_rate": 2.7206939726625295e-05, |
|
"loss": 0.9522, |
|
"step": 14208 |
|
}, |
|
{ |
|
"epoch": 2.235884567126725, |
|
"grad_norm": 11.719073295593262, |
|
"learning_rate": 2.6898702654913484e-05, |
|
"loss": 0.7615, |
|
"step": 14256 |
|
}, |
|
{ |
|
"epoch": 2.243412797992472, |
|
"grad_norm": 15.548043251037598, |
|
"learning_rate": 2.6586311908575173e-05, |
|
"loss": 0.8145, |
|
"step": 14304 |
|
}, |
|
{ |
|
"epoch": 2.250941028858218, |
|
"grad_norm": 17.0067081451416, |
|
"learning_rate": 2.626990554597484e-05, |
|
"loss": 0.6474, |
|
"step": 14352 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_nli-pairs_loss": 0.9902753829956055, |
|
"eval_nli-pairs_runtime": 3.9871, |
|
"eval_nli-pairs_samples_per_second": 50.162, |
|
"eval_nli-pairs_steps_per_second": 1.254, |
|
"eval_sts-test_pearson_cosine": 0.7808536106776984, |
|
"eval_sts-test_pearson_dot": 0.5357771815123735, |
|
"eval_sts-test_pearson_euclidean": 0.7468127931633377, |
|
"eval_sts-test_pearson_manhattan": 0.7458146588384492, |
|
"eval_sts-test_pearson_max": 0.7808536106776984, |
|
"eval_sts-test_spearman_cosine": 0.7912070672899862, |
|
"eval_sts-test_spearman_dot": 0.5140789847704963, |
|
"eval_sts-test_spearman_euclidean": 0.734115394908759, |
|
"eval_sts-test_spearman_manhattan": 0.7357129898281574, |
|
"eval_sts-test_spearman_max": 0.7912070672899862, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_vitaminc-pairs_loss": 4.326738357543945, |
|
"eval_vitaminc-pairs_runtime": 1.4474, |
|
"eval_vitaminc-pairs_samples_per_second": 114.69, |
|
"eval_vitaminc-pairs_steps_per_second": 2.764, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_sts-label_loss": 4.177000522613525, |
|
"eval_sts-label_runtime": 0.418, |
|
"eval_sts-label_samples_per_second": 478.456, |
|
"eval_sts-label_steps_per_second": 11.961, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_qnli-contrastive_loss": 0.17705124616622925, |
|
"eval_qnli-contrastive_runtime": 0.2935, |
|
"eval_qnli-contrastive_samples_per_second": 681.421, |
|
"eval_qnli-contrastive_steps_per_second": 17.036, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_scitail-pairs-qa_loss": 0.08105171471834183, |
|
"eval_scitail-pairs-qa_runtime": 1.0726, |
|
"eval_scitail-pairs-qa_samples_per_second": 186.462, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.662, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_scitail-pairs-pos_loss": 0.413583904504776, |
|
"eval_scitail-pairs-pos_runtime": 2.3897, |
|
"eval_scitail-pairs-pos_samples_per_second": 83.693, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.092, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_xsum-pairs_loss": 0.30970945954322815, |
|
"eval_xsum-pairs_runtime": 1.0443, |
|
"eval_xsum-pairs_samples_per_second": 191.522, |
|
"eval_xsum-pairs_steps_per_second": 4.788, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_compression-pairs_loss": 0.13123837113380432, |
|
"eval_compression-pairs_runtime": 0.2457, |
|
"eval_compression-pairs_samples_per_second": 813.839, |
|
"eval_compression-pairs_steps_per_second": 20.346, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_sciq_pairs_loss": 0.30473247170448303, |
|
"eval_sciq_pairs_runtime": 9.12, |
|
"eval_sciq_pairs_samples_per_second": 21.93, |
|
"eval_sciq_pairs_steps_per_second": 0.548, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_qasc_pairs_loss": 0.35160771012306213, |
|
"eval_qasc_pairs_runtime": 1.2422, |
|
"eval_qasc_pairs_samples_per_second": 160.998, |
|
"eval_qasc_pairs_steps_per_second": 4.025, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_openbookqa_pairs_loss": 1.7501661777496338, |
|
"eval_openbookqa_pairs_runtime": 1.0571, |
|
"eval_openbookqa_pairs_samples_per_second": 189.201, |
|
"eval_openbookqa_pairs_steps_per_second": 4.73, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_msmarco_pairs_loss": 0.9768362641334534, |
|
"eval_msmarco_pairs_runtime": 2.529, |
|
"eval_msmarco_pairs_samples_per_second": 79.083, |
|
"eval_msmarco_pairs_steps_per_second": 1.977, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_nq_pairs_loss": 0.8456315398216248, |
|
"eval_nq_pairs_runtime": 5.6662, |
|
"eval_nq_pairs_samples_per_second": 35.297, |
|
"eval_nq_pairs_steps_per_second": 0.882, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_trivia_pairs_loss": 1.0666593313217163, |
|
"eval_trivia_pairs_runtime": 9.0944, |
|
"eval_trivia_pairs_samples_per_second": 21.992, |
|
"eval_trivia_pairs_steps_per_second": 0.55, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_quora_pairs_loss": 0.21501043438911438, |
|
"eval_quora_pairs_runtime": 0.6293, |
|
"eval_quora_pairs_samples_per_second": 317.812, |
|
"eval_quora_pairs_steps_per_second": 7.945, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_gooaq_pairs_loss": 0.8077111840248108, |
|
"eval_gooaq_pairs_runtime": 1.5722, |
|
"eval_gooaq_pairs_samples_per_second": 127.211, |
|
"eval_gooaq_pairs_steps_per_second": 3.18, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.2537641154328734, |
|
"eval_mrpc_pairs_loss": 0.07021843641996384, |
|
"eval_mrpc_pairs_runtime": 0.2501, |
|
"eval_mrpc_pairs_samples_per_second": 799.6, |
|
"eval_mrpc_pairs_steps_per_second": 19.99, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.258469259723965, |
|
"grad_norm": 5.989051342010498, |
|
"learning_rate": 2.59496234001435e-05, |
|
"loss": 0.7813, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 2.2659974905897116, |
|
"grad_norm": 15.973469734191895, |
|
"learning_rate": 2.5625607016980774e-05, |
|
"loss": 0.9862, |
|
"step": 14448 |
|
}, |
|
{ |
|
"epoch": 2.273525721455458, |
|
"grad_norm": 16.60488510131836, |
|
"learning_rate": 2.5297999592699854e-05, |
|
"loss": 0.8707, |
|
"step": 14496 |
|
}, |
|
{ |
|
"epoch": 2.2810539523212046, |
|
"grad_norm": 13.824111938476562, |
|
"learning_rate": 2.496694591054328e-05, |
|
"loss": 0.9914, |
|
"step": 14544 |
|
}, |
|
{ |
|
"epoch": 2.288582183186951, |
|
"grad_norm": 0.27927935123443604, |
|
"learning_rate": 2.4632592276797123e-05, |
|
"loss": 1.0876, |
|
"step": 14592 |
|
}, |
|
{ |
|
"epoch": 2.2961104140526976, |
|
"grad_norm": 26.05866050720215, |
|
"learning_rate": 2.4295086456132303e-05, |
|
"loss": 0.792, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 2.3036386449184443, |
|
"grad_norm": 13.912291526794434, |
|
"learning_rate": 2.3954577606301227e-05, |
|
"loss": 0.6299, |
|
"step": 14688 |
|
}, |
|
{ |
|
"epoch": 2.3111668757841906, |
|
"grad_norm": 16.938758850097656, |
|
"learning_rate": 2.3611216212218842e-05, |
|
"loss": 0.9561, |
|
"step": 14736 |
|
}, |
|
{ |
|
"epoch": 2.3186951066499373, |
|
"grad_norm": 9.285428047180176, |
|
"learning_rate": 2.3265154019457132e-05, |
|
"loss": 0.8283, |
|
"step": 14784 |
|
}, |
|
{ |
|
"epoch": 2.326223337515684, |
|
"grad_norm": 1.6328561305999756, |
|
"learning_rate": 2.29165439671825e-05, |
|
"loss": 1.0737, |
|
"step": 14832 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_nli-pairs_loss": 0.9611924290657043, |
|
"eval_nli-pairs_runtime": 3.9911, |
|
"eval_nli-pairs_samples_per_second": 50.111, |
|
"eval_nli-pairs_steps_per_second": 1.253, |
|
"eval_sts-test_pearson_cosine": 0.7845481404056492, |
|
"eval_sts-test_pearson_dot": 0.5385486676413874, |
|
"eval_sts-test_pearson_euclidean": 0.7401100148283691, |
|
"eval_sts-test_pearson_manhattan": 0.7381113824263572, |
|
"eval_sts-test_pearson_max": 0.7845481404056492, |
|
"eval_sts-test_spearman_cosine": 0.794691588986042, |
|
"eval_sts-test_spearman_dot": 0.5190592879825671, |
|
"eval_sts-test_spearman_euclidean": 0.7278460500778233, |
|
"eval_sts-test_spearman_manhattan": 0.7287580068840045, |
|
"eval_sts-test_spearman_max": 0.794691588986042, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_vitaminc-pairs_loss": 4.501389503479004, |
|
"eval_vitaminc-pairs_runtime": 1.4538, |
|
"eval_vitaminc-pairs_samples_per_second": 114.183, |
|
"eval_vitaminc-pairs_steps_per_second": 2.751, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_sts-label_loss": 4.10113000869751, |
|
"eval_sts-label_runtime": 0.4117, |
|
"eval_sts-label_samples_per_second": 485.82, |
|
"eval_sts-label_steps_per_second": 12.146, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_qnli-contrastive_loss": 0.11201102286577225, |
|
"eval_qnli-contrastive_runtime": 0.293, |
|
"eval_qnli-contrastive_samples_per_second": 682.573, |
|
"eval_qnli-contrastive_steps_per_second": 17.064, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_scitail-pairs-qa_loss": 0.0739838182926178, |
|
"eval_scitail-pairs-qa_runtime": 1.0651, |
|
"eval_scitail-pairs-qa_samples_per_second": 187.778, |
|
"eval_scitail-pairs-qa_steps_per_second": 4.694, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_scitail-pairs-pos_loss": 0.42078059911727905, |
|
"eval_scitail-pairs-pos_runtime": 2.4326, |
|
"eval_scitail-pairs-pos_samples_per_second": 82.217, |
|
"eval_scitail-pairs-pos_steps_per_second": 2.055, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_xsum-pairs_loss": 0.3173944354057312, |
|
"eval_xsum-pairs_runtime": 1.0505, |
|
"eval_xsum-pairs_samples_per_second": 190.393, |
|
"eval_xsum-pairs_steps_per_second": 4.76, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_compression-pairs_loss": 0.12480150908231735, |
|
"eval_compression-pairs_runtime": 0.2424, |
|
"eval_compression-pairs_samples_per_second": 825.217, |
|
"eval_compression-pairs_steps_per_second": 20.63, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_sciq_pairs_loss": 0.32480019330978394, |
|
"eval_sciq_pairs_runtime": 9.0956, |
|
"eval_sciq_pairs_samples_per_second": 21.989, |
|
"eval_sciq_pairs_steps_per_second": 0.55, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_qasc_pairs_loss": 0.3498300313949585, |
|
"eval_qasc_pairs_runtime": 1.2463, |
|
"eval_qasc_pairs_samples_per_second": 160.475, |
|
"eval_qasc_pairs_steps_per_second": 4.012, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_openbookqa_pairs_loss": 1.8068655729293823, |
|
"eval_openbookqa_pairs_runtime": 1.0596, |
|
"eval_openbookqa_pairs_samples_per_second": 188.748, |
|
"eval_openbookqa_pairs_steps_per_second": 4.719, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_msmarco_pairs_loss": 0.8900260925292969, |
|
"eval_msmarco_pairs_runtime": 2.5231, |
|
"eval_msmarco_pairs_samples_per_second": 79.268, |
|
"eval_msmarco_pairs_steps_per_second": 1.982, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_nq_pairs_loss": 0.8023759722709656, |
|
"eval_nq_pairs_runtime": 5.6432, |
|
"eval_nq_pairs_samples_per_second": 35.441, |
|
"eval_nq_pairs_steps_per_second": 0.886, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_trivia_pairs_loss": 1.0682170391082764, |
|
"eval_trivia_pairs_runtime": 9.0648, |
|
"eval_trivia_pairs_samples_per_second": 22.063, |
|
"eval_trivia_pairs_steps_per_second": 0.552, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_quora_pairs_loss": 0.20286118984222412, |
|
"eval_quora_pairs_runtime": 0.623, |
|
"eval_quora_pairs_samples_per_second": 321.01, |
|
"eval_quora_pairs_steps_per_second": 8.025, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_gooaq_pairs_loss": 0.7672585844993591, |
|
"eval_gooaq_pairs_runtime": 1.552, |
|
"eval_gooaq_pairs_samples_per_second": 128.866, |
|
"eval_gooaq_pairs_steps_per_second": 3.222, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3288895859473024, |
|
"eval_mrpc_pairs_loss": 0.06376803666353226, |
|
"eval_mrpc_pairs_runtime": 0.2417, |
|
"eval_mrpc_pairs_samples_per_second": 827.633, |
|
"eval_mrpc_pairs_steps_per_second": 20.691, |
|
"step": 14849 |
|
}, |
|
{ |
|
"epoch": 2.3337515683814303, |
|
"grad_norm": 7.665156364440918, |
|
"learning_rate": 2.2565540120565625e-05, |
|
"loss": 0.9617, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 2.341279799247177, |
|
"grad_norm": 3.3739352226257324, |
|
"learning_rate": 2.2212297602693703e-05, |
|
"loss": 0.9771, |
|
"step": 14928 |
|
}, |
|
{ |
|
"epoch": 2.3488080301129233, |
|
"grad_norm": 0.46657243371009827, |
|
"learning_rate": 2.185697252601516e-05, |
|
"loss": 0.8833, |
|
"step": 14976 |
|
}, |
|
{ |
|
"epoch": 2.35633626097867, |
|
"grad_norm": 7.030867099761963, |
|
"learning_rate": 2.1499721923347103e-05, |
|
"loss": 0.8536, |
|
"step": 15024 |
|
}, |
|
{ |
|
"epoch": 2.3638644918444167, |
|
"grad_norm": 0.3481261730194092, |
|
"learning_rate": 2.114070367847608e-05, |
|
"loss": 0.8807, |
|
"step": 15072 |
|
}, |
|
{ |
|
"epoch": 2.371392722710163, |
|
"grad_norm": 0.4611968696117401, |
|
"learning_rate": 2.0780076456382662e-05, |
|
"loss": 0.8687, |
|
"step": 15120 |
|
}, |
|
{ |
|
"epoch": 2.3789209535759097, |
|
"grad_norm": 1.0353537797927856, |
|
"learning_rate": 2.041799963312086e-05, |
|
"loss": 1.087, |
|
"step": 15168 |
|
}, |
|
{ |
|
"epoch": 2.3864491844416564, |
|
"grad_norm": 13.026145935058594, |
|
"learning_rate": 2.0054633225383283e-05, |
|
"loss": 0.753, |
|
"step": 15216 |
|
}, |
|
{ |
|
"epoch": 2.3939774153074027, |
|
"grad_norm": 0.20962007343769073, |
|
"learning_rate": 1.9690137819783136e-05, |
|
"loss": 1.294, |
|
"step": 15264 |
|
} |
|
], |
|
"logging_steps": 48, |
|
"max_steps": 19128, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 1913, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 42, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|