|
{ |
|
"best_metric": 0.4959826756763837, |
|
"best_model_checkpoint": "/fsx/anton/cosmopedia/edu_score/bert_snowflake_regression_4/checkpoint-27000", |
|
"epoch": 16.383495145631066, |
|
"eval_steps": 1000, |
|
"global_step": 27000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06067961165048544, |
|
"grad_norm": 0.5638211965560913, |
|
"learning_rate": 0.0002990898058252427, |
|
"loss": 0.4753, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.12135922330097088, |
|
"grad_norm": 0.47830212116241455, |
|
"learning_rate": 0.0002981796116504854, |
|
"loss": 0.357, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1820388349514563, |
|
"grad_norm": 0.6941384077072144, |
|
"learning_rate": 0.0002972694174757281, |
|
"loss": 0.3542, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.24271844660194175, |
|
"grad_norm": 0.459163635969162, |
|
"learning_rate": 0.00029635922330097087, |
|
"loss": 0.3508, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.30339805825242716, |
|
"grad_norm": 1.1585971117019653, |
|
"learning_rate": 0.0002954490291262136, |
|
"loss": 0.3407, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.3640776699029126, |
|
"grad_norm": 0.6505594849586487, |
|
"learning_rate": 0.0002945388349514563, |
|
"loss": 0.3394, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.42475728155339804, |
|
"grad_norm": 0.9804072976112366, |
|
"learning_rate": 0.000293628640776699, |
|
"loss": 0.3435, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.4854368932038835, |
|
"grad_norm": 0.5816351175308228, |
|
"learning_rate": 0.0002927184466019417, |
|
"loss": 0.3323, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.5461165048543689, |
|
"grad_norm": 0.6582027673721313, |
|
"learning_rate": 0.00029180825242718447, |
|
"loss": 0.3293, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.6067961165048543, |
|
"grad_norm": 0.8432559370994568, |
|
"learning_rate": 0.0002908980582524271, |
|
"loss": 0.3337, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.6067961165048543, |
|
"eval_accuracy": 0.6516312117268014, |
|
"eval_f1_macro": 0.388617225018415, |
|
"eval_loss": 0.32428401708602905, |
|
"eval_precision": 0.5273989146868351, |
|
"eval_recall": 0.3731209274235363, |
|
"eval_runtime": 63.5791, |
|
"eval_samples_per_second": 737.144, |
|
"eval_steps_per_second": 5.772, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.6674757281553398, |
|
"grad_norm": 0.6387248039245605, |
|
"learning_rate": 0.0002899878640776699, |
|
"loss": 0.3292, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.7281553398058253, |
|
"grad_norm": 2.0111730098724365, |
|
"learning_rate": 0.0002890776699029126, |
|
"loss": 0.3283, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.7888349514563107, |
|
"grad_norm": 1.0319699048995972, |
|
"learning_rate": 0.0002881674757281553, |
|
"loss": 0.3236, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.8495145631067961, |
|
"grad_norm": 1.194286584854126, |
|
"learning_rate": 0.000287257281553398, |
|
"loss": 0.3178, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.9101941747572816, |
|
"grad_norm": 0.4329046308994293, |
|
"learning_rate": 0.00028634708737864073, |
|
"loss": 0.3234, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.970873786407767, |
|
"grad_norm": 0.4490291476249695, |
|
"learning_rate": 0.0002854368932038835, |
|
"loss": 0.3148, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.0315533980582525, |
|
"grad_norm": 1.9341398477554321, |
|
"learning_rate": 0.0002845266990291262, |
|
"loss": 0.3157, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.0922330097087378, |
|
"grad_norm": 0.6705629825592041, |
|
"learning_rate": 0.0002836165048543689, |
|
"loss": 0.3144, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.1529126213592233, |
|
"grad_norm": 0.3708420395851135, |
|
"learning_rate": 0.0002827063106796116, |
|
"loss": 0.3106, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.2135922330097086, |
|
"grad_norm": 0.4166070222854614, |
|
"learning_rate": 0.00028179611650485433, |
|
"loss": 0.3065, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.2135922330097086, |
|
"eval_accuracy": 0.6756993193505025, |
|
"eval_f1_macro": 0.4177625223934091, |
|
"eval_loss": 0.3020932972431183, |
|
"eval_precision": 0.5208365532058327, |
|
"eval_recall": 0.3908095186846785, |
|
"eval_runtime": 63.8686, |
|
"eval_samples_per_second": 733.803, |
|
"eval_steps_per_second": 5.746, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.2742718446601942, |
|
"grad_norm": 0.949810266494751, |
|
"learning_rate": 0.00028088592233009704, |
|
"loss": 0.309, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.3349514563106797, |
|
"grad_norm": 0.6933236718177795, |
|
"learning_rate": 0.00027997572815533975, |
|
"loss": 0.3039, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.395631067961165, |
|
"grad_norm": 0.4874693751335144, |
|
"learning_rate": 0.0002790655339805825, |
|
"loss": 0.3016, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.4563106796116505, |
|
"grad_norm": 0.5307803750038147, |
|
"learning_rate": 0.0002781553398058252, |
|
"loss": 0.295, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.516990291262136, |
|
"grad_norm": 0.7260825634002686, |
|
"learning_rate": 0.00027724514563106793, |
|
"loss": 0.298, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.5776699029126213, |
|
"grad_norm": 1.3546072244644165, |
|
"learning_rate": 0.00027633495145631064, |
|
"loss": 0.2937, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.6383495145631068, |
|
"grad_norm": 0.7695233821868896, |
|
"learning_rate": 0.00027542475728155335, |
|
"loss": 0.2939, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.6990291262135924, |
|
"grad_norm": 0.46857160329818726, |
|
"learning_rate": 0.0002745145631067961, |
|
"loss": 0.2911, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.7597087378640777, |
|
"grad_norm": 0.521542489528656, |
|
"learning_rate": 0.0002736043689320388, |
|
"loss": 0.2936, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.820388349514563, |
|
"grad_norm": 1.1797749996185303, |
|
"learning_rate": 0.00027269417475728154, |
|
"loss": 0.291, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.820388349514563, |
|
"eval_accuracy": 0.6800947361683061, |
|
"eval_f1_macro": 0.43005056999542096, |
|
"eval_loss": 0.2904761731624603, |
|
"eval_precision": 0.5283372179121356, |
|
"eval_recall": 0.40012210038254903, |
|
"eval_runtime": 63.5697, |
|
"eval_samples_per_second": 737.253, |
|
"eval_steps_per_second": 5.773, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.8810679611650487, |
|
"grad_norm": 0.9966709613800049, |
|
"learning_rate": 0.00027178398058252425, |
|
"loss": 0.2965, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.941747572815534, |
|
"grad_norm": 0.40996024012565613, |
|
"learning_rate": 0.00027087378640776696, |
|
"loss": 0.2941, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.0024271844660193, |
|
"grad_norm": 0.5450060367584229, |
|
"learning_rate": 0.00026996359223300967, |
|
"loss": 0.2912, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.063106796116505, |
|
"grad_norm": 0.5307539701461792, |
|
"learning_rate": 0.0002690533980582524, |
|
"loss": 0.2872, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.1237864077669903, |
|
"grad_norm": 0.5863193273544312, |
|
"learning_rate": 0.00026814320388349514, |
|
"loss": 0.2929, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.1844660194174756, |
|
"grad_norm": 0.584078311920166, |
|
"learning_rate": 0.00026723300970873785, |
|
"loss": 0.2879, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.2451456310679614, |
|
"grad_norm": 0.6381602883338928, |
|
"learning_rate": 0.00026632281553398056, |
|
"loss": 0.2892, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.3058252427184467, |
|
"grad_norm": 0.4760149121284485, |
|
"learning_rate": 0.00026541262135922327, |
|
"loss": 0.2863, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.366504854368932, |
|
"grad_norm": 0.4088296890258789, |
|
"learning_rate": 0.000264502427184466, |
|
"loss": 0.2913, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.4271844660194173, |
|
"grad_norm": 1.3476176261901855, |
|
"learning_rate": 0.00026359223300970874, |
|
"loss": 0.2845, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.4271844660194173, |
|
"eval_accuracy": 0.6929182580493738, |
|
"eval_f1_macro": 0.4442664055274559, |
|
"eval_loss": 0.280377060174942, |
|
"eval_precision": 0.5328225535144124, |
|
"eval_recall": 0.41160029395774395, |
|
"eval_runtime": 64.199, |
|
"eval_samples_per_second": 730.027, |
|
"eval_steps_per_second": 5.717, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.487864077669903, |
|
"grad_norm": 0.6318752765655518, |
|
"learning_rate": 0.0002626820388349514, |
|
"loss": 0.2812, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.5485436893203883, |
|
"grad_norm": 0.49435973167419434, |
|
"learning_rate": 0.00026177184466019416, |
|
"loss": 0.2803, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.6092233009708736, |
|
"grad_norm": 0.4300900101661682, |
|
"learning_rate": 0.00026086165048543687, |
|
"loss": 0.2853, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.6699029126213594, |
|
"grad_norm": 0.9545436501502991, |
|
"learning_rate": 0.0002599514563106796, |
|
"loss": 0.2813, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.7305825242718447, |
|
"grad_norm": 0.5803716778755188, |
|
"learning_rate": 0.0002590412621359223, |
|
"loss": 0.2838, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.79126213592233, |
|
"grad_norm": 1.4714713096618652, |
|
"learning_rate": 0.000258131067961165, |
|
"loss": 0.2814, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.8519417475728153, |
|
"grad_norm": 0.6767821311950684, |
|
"learning_rate": 0.00025722087378640777, |
|
"loss": 0.2741, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 2.912621359223301, |
|
"grad_norm": 0.4653462767601013, |
|
"learning_rate": 0.0002563106796116505, |
|
"loss": 0.2783, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.9733009708737863, |
|
"grad_norm": 1.3012775182724, |
|
"learning_rate": 0.0002554004854368932, |
|
"loss": 0.283, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.033980582524272, |
|
"grad_norm": 0.4733451306819916, |
|
"learning_rate": 0.0002544902912621359, |
|
"loss": 0.2767, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.033980582524272, |
|
"eval_accuracy": 0.6949666076343696, |
|
"eval_f1_macro": 0.4524166681181929, |
|
"eval_loss": 0.2772601842880249, |
|
"eval_precision": 0.52914110464024, |
|
"eval_recall": 0.42261704559523156, |
|
"eval_runtime": 63.4257, |
|
"eval_samples_per_second": 738.927, |
|
"eval_steps_per_second": 5.786, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.0946601941747574, |
|
"grad_norm": 0.4103662371635437, |
|
"learning_rate": 0.0002535800970873786, |
|
"loss": 0.2796, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.1553398058252426, |
|
"grad_norm": 0.4195462763309479, |
|
"learning_rate": 0.0002526699029126213, |
|
"loss": 0.2757, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.216019417475728, |
|
"grad_norm": 1.2391552925109863, |
|
"learning_rate": 0.0002517597087378641, |
|
"loss": 0.2783, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.2766990291262137, |
|
"grad_norm": 1.2029412984848022, |
|
"learning_rate": 0.0002508495145631068, |
|
"loss": 0.2772, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.337378640776699, |
|
"grad_norm": 0.5050978660583496, |
|
"learning_rate": 0.0002499393203883495, |
|
"loss": 0.2776, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.3980582524271843, |
|
"grad_norm": 1.0107412338256836, |
|
"learning_rate": 0.0002490291262135922, |
|
"loss": 0.2766, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.45873786407767, |
|
"grad_norm": 0.4374917149543762, |
|
"learning_rate": 0.0002481189320388349, |
|
"loss": 0.2719, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 3.5194174757281553, |
|
"grad_norm": 1.6768765449523926, |
|
"learning_rate": 0.0002472087378640777, |
|
"loss": 0.2803, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.5800970873786406, |
|
"grad_norm": 0.8120823502540588, |
|
"learning_rate": 0.0002462985436893204, |
|
"loss": 0.2723, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 3.6407766990291264, |
|
"grad_norm": 1.3967177867889404, |
|
"learning_rate": 0.0002453883495145631, |
|
"loss": 0.2796, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.6407766990291264, |
|
"eval_accuracy": 0.7001301555465466, |
|
"eval_f1_macro": 0.46180498265852127, |
|
"eval_loss": 0.272257536649704, |
|
"eval_precision": 0.5281578618931982, |
|
"eval_recall": 0.4315295129889904, |
|
"eval_runtime": 63.8213, |
|
"eval_samples_per_second": 734.347, |
|
"eval_steps_per_second": 5.75, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.7014563106796117, |
|
"grad_norm": 0.6093985438346863, |
|
"learning_rate": 0.0002444781553398058, |
|
"loss": 0.2744, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 3.762135922330097, |
|
"grad_norm": 0.7282202243804932, |
|
"learning_rate": 0.00024356796116504852, |
|
"loss": 0.2715, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 3.8228155339805827, |
|
"grad_norm": 1.1341967582702637, |
|
"learning_rate": 0.00024265776699029123, |
|
"loss": 0.2709, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 3.883495145631068, |
|
"grad_norm": 0.8576841354370117, |
|
"learning_rate": 0.00024174757281553394, |
|
"loss": 0.275, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 3.9441747572815533, |
|
"grad_norm": 0.5656840205192566, |
|
"learning_rate": 0.00024083737864077668, |
|
"loss": 0.2676, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.004854368932039, |
|
"grad_norm": 0.6544743180274963, |
|
"learning_rate": 0.0002399271844660194, |
|
"loss": 0.2734, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 4.065533980582524, |
|
"grad_norm": 1.5159205198287964, |
|
"learning_rate": 0.0002390169902912621, |
|
"loss": 0.2666, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 4.12621359223301, |
|
"grad_norm": 0.9112799763679504, |
|
"learning_rate": 0.00023810679611650483, |
|
"loss": 0.2646, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 4.186893203883495, |
|
"grad_norm": 0.6971092224121094, |
|
"learning_rate": 0.00023719660194174754, |
|
"loss": 0.2681, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 4.247572815533981, |
|
"grad_norm": 0.5126680731773376, |
|
"learning_rate": 0.00023628640776699028, |
|
"loss": 0.2669, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.247572815533981, |
|
"eval_accuracy": 0.7024985597541981, |
|
"eval_f1_macro": 0.4678440617843855, |
|
"eval_loss": 0.26851192116737366, |
|
"eval_precision": 0.5330566032345198, |
|
"eval_recall": 0.43620949446199525, |
|
"eval_runtime": 63.9009, |
|
"eval_samples_per_second": 733.432, |
|
"eval_steps_per_second": 5.743, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.308252427184466, |
|
"grad_norm": 0.40426695346832275, |
|
"learning_rate": 0.000235376213592233, |
|
"loss": 0.2722, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 4.368932038834951, |
|
"grad_norm": 0.4849171042442322, |
|
"learning_rate": 0.0002344660194174757, |
|
"loss": 0.273, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 4.429611650485437, |
|
"grad_norm": 0.7887744307518005, |
|
"learning_rate": 0.00023355582524271844, |
|
"loss": 0.2711, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 4.490291262135923, |
|
"grad_norm": 1.282893419265747, |
|
"learning_rate": 0.00023264563106796115, |
|
"loss": 0.2633, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 4.550970873786408, |
|
"grad_norm": 0.4942808449268341, |
|
"learning_rate": 0.00023173543689320386, |
|
"loss": 0.2655, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 4.611650485436893, |
|
"grad_norm": 0.4098761975765228, |
|
"learning_rate": 0.0002308252427184466, |
|
"loss": 0.2719, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 4.672330097087379, |
|
"grad_norm": 1.1584999561309814, |
|
"learning_rate": 0.0002299150485436893, |
|
"loss": 0.2646, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 4.733009708737864, |
|
"grad_norm": 0.5956116318702698, |
|
"learning_rate": 0.00022900485436893204, |
|
"loss": 0.2681, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 4.793689320388349, |
|
"grad_norm": 0.7886734008789062, |
|
"learning_rate": 0.00022809466019417472, |
|
"loss": 0.2648, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 4.854368932038835, |
|
"grad_norm": 0.7086551785469055, |
|
"learning_rate": 0.00022718446601941746, |
|
"loss": 0.2667, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.854368932038835, |
|
"eval_accuracy": 0.7028826253013848, |
|
"eval_f1_macro": 0.4680422447785217, |
|
"eval_loss": 0.2663130462169647, |
|
"eval_precision": 0.5388642351614853, |
|
"eval_recall": 0.4370831407609595, |
|
"eval_runtime": 63.9774, |
|
"eval_samples_per_second": 732.556, |
|
"eval_steps_per_second": 5.736, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.915048543689321, |
|
"grad_norm": 0.6959311366081238, |
|
"learning_rate": 0.0002262742718446602, |
|
"loss": 0.2662, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 4.975728155339806, |
|
"grad_norm": 1.318164587020874, |
|
"learning_rate": 0.0002253640776699029, |
|
"loss": 0.2679, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 5.036407766990291, |
|
"grad_norm": 0.5080001354217529, |
|
"learning_rate": 0.00022445388349514561, |
|
"loss": 0.2662, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 5.097087378640777, |
|
"grad_norm": 0.40693196654319763, |
|
"learning_rate": 0.00022354368932038832, |
|
"loss": 0.2641, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 5.157766990291262, |
|
"grad_norm": 0.42492175102233887, |
|
"learning_rate": 0.00022263349514563106, |
|
"loss": 0.2669, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 5.218446601941747, |
|
"grad_norm": 0.9580904841423035, |
|
"learning_rate": 0.00022172330097087374, |
|
"loss": 0.2613, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 5.279126213592233, |
|
"grad_norm": 0.573049008846283, |
|
"learning_rate": 0.00022081310679611648, |
|
"loss": 0.26, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 5.339805825242719, |
|
"grad_norm": 0.7355623841285706, |
|
"learning_rate": 0.00021990291262135922, |
|
"loss": 0.2682, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 5.400485436893204, |
|
"grad_norm": 0.5867640972137451, |
|
"learning_rate": 0.00021899271844660193, |
|
"loss": 0.2677, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 5.461165048543689, |
|
"grad_norm": 1.283672571182251, |
|
"learning_rate": 0.00021808252427184464, |
|
"loss": 0.2582, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.461165048543689, |
|
"eval_accuracy": 0.7030319841252907, |
|
"eval_f1_macro": 0.476191462667983, |
|
"eval_loss": 0.2642744779586792, |
|
"eval_precision": 0.5389130008234221, |
|
"eval_recall": 0.4475692410266268, |
|
"eval_runtime": 63.527, |
|
"eval_samples_per_second": 737.749, |
|
"eval_steps_per_second": 5.777, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.521844660194175, |
|
"grad_norm": 0.4764248728752136, |
|
"learning_rate": 0.00021717233009708735, |
|
"loss": 0.2596, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 5.58252427184466, |
|
"grad_norm": 0.7470163106918335, |
|
"learning_rate": 0.00021626213592233008, |
|
"loss": 0.2632, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 5.643203883495145, |
|
"grad_norm": 0.7628265023231506, |
|
"learning_rate": 0.00021535194174757282, |
|
"loss": 0.2583, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 5.703883495145631, |
|
"grad_norm": 0.4257694184780121, |
|
"learning_rate": 0.0002144417475728155, |
|
"loss": 0.2627, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 5.764563106796117, |
|
"grad_norm": 0.5351702570915222, |
|
"learning_rate": 0.00021353155339805824, |
|
"loss": 0.2578, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 5.825242718446602, |
|
"grad_norm": 1.9470020532608032, |
|
"learning_rate": 0.00021262135922330095, |
|
"loss": 0.2599, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 5.885922330097087, |
|
"grad_norm": 0.41918280720710754, |
|
"learning_rate": 0.00021171116504854369, |
|
"loss": 0.2604, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 5.946601941747573, |
|
"grad_norm": 0.7866085767745972, |
|
"learning_rate": 0.00021080097087378637, |
|
"loss": 0.2645, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 6.007281553398058, |
|
"grad_norm": 0.773632287979126, |
|
"learning_rate": 0.0002098907766990291, |
|
"loss": 0.2594, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 6.067961165048544, |
|
"grad_norm": 0.905390202999115, |
|
"learning_rate": 0.00020898058252427184, |
|
"loss": 0.2587, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 6.067961165048544, |
|
"eval_accuracy": 0.7069153135468453, |
|
"eval_f1_macro": 0.47329615730761665, |
|
"eval_loss": 0.26157495379447937, |
|
"eval_precision": 0.5382757148939651, |
|
"eval_recall": 0.4417990410758896, |
|
"eval_runtime": 64.0102, |
|
"eval_samples_per_second": 732.181, |
|
"eval_steps_per_second": 5.733, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 6.128640776699029, |
|
"grad_norm": 0.7645794749259949, |
|
"learning_rate": 0.00020807038834951455, |
|
"loss": 0.2584, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 6.189320388349515, |
|
"grad_norm": 0.80213862657547, |
|
"learning_rate": 0.00020716019417475726, |
|
"loss": 0.2576, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"grad_norm": 0.49435296654701233, |
|
"learning_rate": 0.00020624999999999997, |
|
"loss": 0.2578, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 6.310679611650485, |
|
"grad_norm": 1.2024767398834229, |
|
"learning_rate": 0.0002053398058252427, |
|
"loss": 0.2628, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 6.371359223300971, |
|
"grad_norm": 0.43948036432266235, |
|
"learning_rate": 0.0002044296116504854, |
|
"loss": 0.2555, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 6.432038834951456, |
|
"grad_norm": 1.0257009267807007, |
|
"learning_rate": 0.00020351941747572813, |
|
"loss": 0.2582, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 6.492718446601942, |
|
"grad_norm": 1.2218210697174072, |
|
"learning_rate": 0.00020260922330097086, |
|
"loss": 0.2603, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 6.553398058252427, |
|
"grad_norm": 0.43541598320007324, |
|
"learning_rate": 0.00020169902912621357, |
|
"loss": 0.2529, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 6.614077669902913, |
|
"grad_norm": 0.4533527195453644, |
|
"learning_rate": 0.00020078883495145628, |
|
"loss": 0.255, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 6.674757281553398, |
|
"grad_norm": 0.714527428150177, |
|
"learning_rate": 0.000199878640776699, |
|
"loss": 0.2524, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 6.674757281553398, |
|
"eval_accuracy": 0.7088356412827789, |
|
"eval_f1_macro": 0.471743421411692, |
|
"eval_loss": 0.2603248357772827, |
|
"eval_precision": 0.5487634473468738, |
|
"eval_recall": 0.43766522368406974, |
|
"eval_runtime": 64.8962, |
|
"eval_samples_per_second": 722.184, |
|
"eval_steps_per_second": 5.655, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 6.735436893203883, |
|
"grad_norm": 0.6983916163444519, |
|
"learning_rate": 0.00019896844660194173, |
|
"loss": 0.2585, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 6.796116504854369, |
|
"grad_norm": 0.8627694845199585, |
|
"learning_rate": 0.00019805825242718447, |
|
"loss": 0.2557, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 6.856796116504855, |
|
"grad_norm": 0.40715476870536804, |
|
"learning_rate": 0.00019714805825242715, |
|
"loss": 0.2588, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 6.91747572815534, |
|
"grad_norm": 1.0038883686065674, |
|
"learning_rate": 0.0001962378640776699, |
|
"loss": 0.2542, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 6.978155339805825, |
|
"grad_norm": 0.941238522529602, |
|
"learning_rate": 0.0001953276699029126, |
|
"loss": 0.2597, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 7.038834951456311, |
|
"grad_norm": 0.6316859722137451, |
|
"learning_rate": 0.00019441747572815533, |
|
"loss": 0.2576, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 7.099514563106796, |
|
"grad_norm": 0.5387765169143677, |
|
"learning_rate": 0.00019350728155339804, |
|
"loss": 0.2525, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 7.160194174757281, |
|
"grad_norm": 0.9042656421661377, |
|
"learning_rate": 0.00019259708737864075, |
|
"loss": 0.2552, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 7.220873786407767, |
|
"grad_norm": 0.6641530990600586, |
|
"learning_rate": 0.0001916868932038835, |
|
"loss": 0.2495, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 7.281553398058253, |
|
"grad_norm": 1.050493836402893, |
|
"learning_rate": 0.0001907766990291262, |
|
"loss": 0.2538, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 7.281553398058253, |
|
"eval_accuracy": 0.7098171421255894, |
|
"eval_f1_macro": 0.4800779183645662, |
|
"eval_loss": 0.25889766216278076, |
|
"eval_precision": 0.5445483450890162, |
|
"eval_recall": 0.44845043483484776, |
|
"eval_runtime": 63.7915, |
|
"eval_samples_per_second": 734.69, |
|
"eval_steps_per_second": 5.753, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 7.342233009708738, |
|
"grad_norm": 0.7460100054740906, |
|
"learning_rate": 0.0001898665048543689, |
|
"loss": 0.2576, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 7.402912621359223, |
|
"grad_norm": 1.8915176391601562, |
|
"learning_rate": 0.00018895631067961165, |
|
"loss": 0.2563, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 7.463592233009709, |
|
"grad_norm": 0.6642426252365112, |
|
"learning_rate": 0.00018804611650485436, |
|
"loss": 0.2515, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 7.524271844660194, |
|
"grad_norm": 0.7037068009376526, |
|
"learning_rate": 0.00018713592233009707, |
|
"loss": 0.2525, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 7.584951456310679, |
|
"grad_norm": 0.44775909185409546, |
|
"learning_rate": 0.00018622572815533978, |
|
"loss": 0.2557, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 7.645631067961165, |
|
"grad_norm": 1.2254287004470825, |
|
"learning_rate": 0.0001853155339805825, |
|
"loss": 0.252, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 7.706310679611651, |
|
"grad_norm": 0.7916478514671326, |
|
"learning_rate": 0.00018440533980582525, |
|
"loss": 0.2503, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 7.766990291262136, |
|
"grad_norm": 0.5165425539016724, |
|
"learning_rate": 0.00018349514563106793, |
|
"loss": 0.259, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 7.827669902912621, |
|
"grad_norm": 0.45590564608573914, |
|
"learning_rate": 0.00018258495145631067, |
|
"loss": 0.2536, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 7.888349514563107, |
|
"grad_norm": 0.7460587620735168, |
|
"learning_rate": 0.00018167475728155338, |
|
"loss": 0.2536, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 7.888349514563107, |
|
"eval_accuracy": 0.7118014807860542, |
|
"eval_f1_macro": 0.48153346170307537, |
|
"eval_loss": 0.2587451934814453, |
|
"eval_precision": 0.5428621211876972, |
|
"eval_recall": 0.44971967808318786, |
|
"eval_runtime": 63.3555, |
|
"eval_samples_per_second": 739.747, |
|
"eval_steps_per_second": 5.793, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 7.949029126213592, |
|
"grad_norm": 0.5832739472389221, |
|
"learning_rate": 0.00018076456310679612, |
|
"loss": 0.2549, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 8.009708737864077, |
|
"grad_norm": 0.991656482219696, |
|
"learning_rate": 0.0001798543689320388, |
|
"loss": 0.2516, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 8.070388349514563, |
|
"grad_norm": 0.5339873433113098, |
|
"learning_rate": 0.00017894417475728154, |
|
"loss": 0.2492, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 8.131067961165048, |
|
"grad_norm": 0.5186127424240112, |
|
"learning_rate": 0.00017803398058252427, |
|
"loss": 0.2521, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 8.191747572815533, |
|
"grad_norm": 0.7012351751327515, |
|
"learning_rate": 0.00017712378640776698, |
|
"loss": 0.247, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 8.25242718446602, |
|
"grad_norm": 0.5204672813415527, |
|
"learning_rate": 0.0001762135922330097, |
|
"loss": 0.2528, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 8.313106796116505, |
|
"grad_norm": 0.46670639514923096, |
|
"learning_rate": 0.0001753033980582524, |
|
"loss": 0.2484, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 8.37378640776699, |
|
"grad_norm": 0.8041334748268127, |
|
"learning_rate": 0.00017439320388349514, |
|
"loss": 0.2534, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 8.434466019417476, |
|
"grad_norm": 0.482832133769989, |
|
"learning_rate": 0.00017348300970873787, |
|
"loss": 0.2463, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 8.495145631067961, |
|
"grad_norm": 0.9541172981262207, |
|
"learning_rate": 0.00017257281553398056, |
|
"loss": 0.2527, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 8.495145631067961, |
|
"eval_accuracy": 0.7112040454904304, |
|
"eval_f1_macro": 0.4703790685018183, |
|
"eval_loss": 0.2594839036464691, |
|
"eval_precision": 0.5469933734414426, |
|
"eval_recall": 0.43675078934194306, |
|
"eval_runtime": 64.5706, |
|
"eval_samples_per_second": 725.825, |
|
"eval_steps_per_second": 5.684, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 8.555825242718447, |
|
"grad_norm": 0.41729798913002014, |
|
"learning_rate": 0.0001716626213592233, |
|
"loss": 0.2566, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 8.616504854368932, |
|
"grad_norm": 1.1643282175064087, |
|
"learning_rate": 0.000170752427184466, |
|
"loss": 0.2492, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 8.677184466019417, |
|
"grad_norm": 0.5015943050384521, |
|
"learning_rate": 0.00016984223300970874, |
|
"loss": 0.2484, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 8.737864077669903, |
|
"grad_norm": 0.6088297963142395, |
|
"learning_rate": 0.00016893203883495142, |
|
"loss": 0.2481, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 8.798543689320388, |
|
"grad_norm": 0.6608941555023193, |
|
"learning_rate": 0.00016802184466019416, |
|
"loss": 0.2509, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 8.859223300970873, |
|
"grad_norm": 0.506313145160675, |
|
"learning_rate": 0.0001671116504854369, |
|
"loss": 0.2488, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 8.919902912621358, |
|
"grad_norm": 0.579694390296936, |
|
"learning_rate": 0.00016620145631067958, |
|
"loss": 0.252, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 8.980582524271846, |
|
"grad_norm": 1.048781156539917, |
|
"learning_rate": 0.00016529126213592232, |
|
"loss": 0.2569, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 9.04126213592233, |
|
"grad_norm": 0.4762475788593292, |
|
"learning_rate": 0.00016438106796116503, |
|
"loss": 0.247, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 9.101941747572816, |
|
"grad_norm": 0.5784612894058228, |
|
"learning_rate": 0.00016347087378640776, |
|
"loss": 0.249, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 9.101941747572816, |
|
"eval_accuracy": 0.7119295026351163, |
|
"eval_f1_macro": 0.48462198319681066, |
|
"eval_loss": 0.2577269971370697, |
|
"eval_precision": 0.6220453203620933, |
|
"eval_recall": 0.45259946630146386, |
|
"eval_runtime": 63.7117, |
|
"eval_samples_per_second": 735.611, |
|
"eval_steps_per_second": 5.76, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 9.162621359223301, |
|
"grad_norm": 1.0214685201644897, |
|
"learning_rate": 0.00016256067961165045, |
|
"loss": 0.2449, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 9.223300970873787, |
|
"grad_norm": 0.5309824347496033, |
|
"learning_rate": 0.00016165048543689318, |
|
"loss": 0.2456, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 9.283980582524272, |
|
"grad_norm": 2.5317320823669434, |
|
"learning_rate": 0.00016074029126213592, |
|
"loss": 0.2428, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 9.344660194174757, |
|
"grad_norm": 0.588573157787323, |
|
"learning_rate": 0.00015983009708737863, |
|
"loss": 0.2451, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 9.405339805825243, |
|
"grad_norm": 0.687719464302063, |
|
"learning_rate": 0.00015891990291262134, |
|
"loss": 0.2503, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 9.466019417475728, |
|
"grad_norm": 1.0443726778030396, |
|
"learning_rate": 0.00015800970873786405, |
|
"loss": 0.2485, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 9.526699029126213, |
|
"grad_norm": 1.1230727434158325, |
|
"learning_rate": 0.00015709951456310679, |
|
"loss": 0.2479, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 9.587378640776699, |
|
"grad_norm": 0.6641112565994263, |
|
"learning_rate": 0.00015618932038834952, |
|
"loss": 0.2489, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 9.648058252427184, |
|
"grad_norm": 0.4918789863586426, |
|
"learning_rate": 0.0001552791262135922, |
|
"loss": 0.2481, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 9.70873786407767, |
|
"grad_norm": 0.7862838506698608, |
|
"learning_rate": 0.00015436893203883494, |
|
"loss": 0.2481, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 9.70873786407767, |
|
"eval_accuracy": 0.713487101798707, |
|
"eval_f1_macro": 0.4839261840621368, |
|
"eval_loss": 0.2561655640602112, |
|
"eval_precision": 0.6345013597016592, |
|
"eval_recall": 0.4492915653199398, |
|
"eval_runtime": 63.9745, |
|
"eval_samples_per_second": 732.589, |
|
"eval_steps_per_second": 5.737, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 9.769417475728154, |
|
"grad_norm": 0.8769936561584473, |
|
"learning_rate": 0.00015345873786407765, |
|
"loss": 0.2479, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 9.830097087378642, |
|
"grad_norm": 0.9663941860198975, |
|
"learning_rate": 0.0001525485436893204, |
|
"loss": 0.2548, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 9.890776699029127, |
|
"grad_norm": 0.4475904405117035, |
|
"learning_rate": 0.00015163834951456307, |
|
"loss": 0.2504, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 9.951456310679612, |
|
"grad_norm": 0.9180653691291809, |
|
"learning_rate": 0.0001507281553398058, |
|
"loss": 0.246, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 10.012135922330097, |
|
"grad_norm": 1.1535005569458008, |
|
"learning_rate": 0.00014981796116504852, |
|
"loss": 0.2463, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 10.072815533980583, |
|
"grad_norm": 0.5621392130851746, |
|
"learning_rate": 0.00014890776699029125, |
|
"loss": 0.2432, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 10.133495145631068, |
|
"grad_norm": 0.5198004245758057, |
|
"learning_rate": 0.00014799757281553396, |
|
"loss": 0.2463, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 10.194174757281553, |
|
"grad_norm": 0.7625685930252075, |
|
"learning_rate": 0.00014708737864077667, |
|
"loss": 0.2474, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 10.254854368932039, |
|
"grad_norm": 0.4673856794834137, |
|
"learning_rate": 0.0001461771844660194, |
|
"loss": 0.244, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 10.315533980582524, |
|
"grad_norm": 0.6320896744728088, |
|
"learning_rate": 0.00014526699029126212, |
|
"loss": 0.241, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 10.315533980582524, |
|
"eval_accuracy": 0.7026265816032603, |
|
"eval_f1_macro": 0.4873663248833824, |
|
"eval_loss": 0.2613597810268402, |
|
"eval_precision": 0.5926410330413402, |
|
"eval_recall": 0.46290412629796984, |
|
"eval_runtime": 63.8686, |
|
"eval_samples_per_second": 733.803, |
|
"eval_steps_per_second": 5.746, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 10.37621359223301, |
|
"grad_norm": 1.3767356872558594, |
|
"learning_rate": 0.00014435679611650483, |
|
"loss": 0.2431, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 10.436893203883495, |
|
"grad_norm": 0.6615713834762573, |
|
"learning_rate": 0.00014344660194174757, |
|
"loss": 0.2401, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 10.49757281553398, |
|
"grad_norm": 0.5104607343673706, |
|
"learning_rate": 0.00014253640776699028, |
|
"loss": 0.2467, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 10.558252427184467, |
|
"grad_norm": 1.2690117359161377, |
|
"learning_rate": 0.00014162621359223301, |
|
"loss": 0.2474, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 10.618932038834952, |
|
"grad_norm": 0.5738195180892944, |
|
"learning_rate": 0.00014071601941747572, |
|
"loss": 0.2508, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 10.679611650485437, |
|
"grad_norm": 0.9724490642547607, |
|
"learning_rate": 0.00013980582524271843, |
|
"loss": 0.2457, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 10.740291262135923, |
|
"grad_norm": 0.5149191617965698, |
|
"learning_rate": 0.00013889563106796114, |
|
"loss": 0.2453, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 10.800970873786408, |
|
"grad_norm": 0.4608076512813568, |
|
"learning_rate": 0.00013798543689320388, |
|
"loss": 0.2461, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 10.861650485436893, |
|
"grad_norm": 0.7955921292304993, |
|
"learning_rate": 0.0001370752427184466, |
|
"loss": 0.2427, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 10.922330097087379, |
|
"grad_norm": 0.75131756067276, |
|
"learning_rate": 0.00013616504854368933, |
|
"loss": 0.2441, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 10.922330097087379, |
|
"eval_accuracy": 0.712249557257772, |
|
"eval_f1_macro": 0.48099926256661046, |
|
"eval_loss": 0.2561965882778168, |
|
"eval_precision": 0.6261557755424023, |
|
"eval_recall": 0.44725199588180664, |
|
"eval_runtime": 63.857, |
|
"eval_samples_per_second": 733.937, |
|
"eval_steps_per_second": 5.747, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 10.983009708737864, |
|
"grad_norm": 0.791315495967865, |
|
"learning_rate": 0.00013525485436893204, |
|
"loss": 0.2486, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 11.04368932038835, |
|
"grad_norm": 0.7312906980514526, |
|
"learning_rate": 0.00013434466019417475, |
|
"loss": 0.2431, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 11.104368932038835, |
|
"grad_norm": 0.609377384185791, |
|
"learning_rate": 0.00013343446601941746, |
|
"loss": 0.2399, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 11.16504854368932, |
|
"grad_norm": 0.5242405533790588, |
|
"learning_rate": 0.0001325242718446602, |
|
"loss": 0.2424, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 11.225728155339805, |
|
"grad_norm": 0.6860948204994202, |
|
"learning_rate": 0.0001316140776699029, |
|
"loss": 0.2434, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 11.28640776699029, |
|
"grad_norm": 2.105470657348633, |
|
"learning_rate": 0.0001307038834951456, |
|
"loss": 0.2393, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 11.347087378640778, |
|
"grad_norm": 0.990110456943512, |
|
"learning_rate": 0.00012979368932038835, |
|
"loss": 0.2446, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 11.407766990291263, |
|
"grad_norm": 0.5341915488243103, |
|
"learning_rate": 0.00012888349514563106, |
|
"loss": 0.2449, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 11.468446601941748, |
|
"grad_norm": 1.2902977466583252, |
|
"learning_rate": 0.00012797330097087377, |
|
"loss": 0.245, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 11.529126213592233, |
|
"grad_norm": 0.5241170525550842, |
|
"learning_rate": 0.00012706310679611648, |
|
"loss": 0.2433, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 11.529126213592233, |
|
"eval_accuracy": 0.7136791345723004, |
|
"eval_f1_macro": 0.47974793179174946, |
|
"eval_loss": 0.25500378012657166, |
|
"eval_precision": 0.5467676042682321, |
|
"eval_recall": 0.44681795057266877, |
|
"eval_runtime": 63.4183, |
|
"eval_samples_per_second": 739.014, |
|
"eval_steps_per_second": 5.787, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 11.589805825242719, |
|
"grad_norm": 1.6583833694458008, |
|
"learning_rate": 0.00012615291262135922, |
|
"loss": 0.2444, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 11.650485436893204, |
|
"grad_norm": 0.5068759322166443, |
|
"learning_rate": 0.00012524271844660192, |
|
"loss": 0.2417, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 11.71116504854369, |
|
"grad_norm": 0.8810881972312927, |
|
"learning_rate": 0.00012433252427184466, |
|
"loss": 0.2443, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 11.771844660194175, |
|
"grad_norm": 1.347113847732544, |
|
"learning_rate": 0.00012342233009708737, |
|
"loss": 0.2408, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 11.83252427184466, |
|
"grad_norm": 0.5012452602386475, |
|
"learning_rate": 0.00012251213592233008, |
|
"loss": 0.2443, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 11.893203883495145, |
|
"grad_norm": 0.6998531222343445, |
|
"learning_rate": 0.0001216019417475728, |
|
"loss": 0.2449, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 11.95388349514563, |
|
"grad_norm": 0.9693962335586548, |
|
"learning_rate": 0.00012069174757281553, |
|
"loss": 0.2415, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 12.014563106796116, |
|
"grad_norm": 0.5657163858413696, |
|
"learning_rate": 0.00011978155339805824, |
|
"loss": 0.2416, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 12.075242718446601, |
|
"grad_norm": 0.527268648147583, |
|
"learning_rate": 0.00011887135922330096, |
|
"loss": 0.2377, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 12.135922330097088, |
|
"grad_norm": 0.4967799782752991, |
|
"learning_rate": 0.00011796116504854367, |
|
"loss": 0.2436, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 12.135922330097088, |
|
"eval_accuracy": 0.7124629270062091, |
|
"eval_f1_macro": 0.48898241670534826, |
|
"eval_loss": 0.2549245059490204, |
|
"eval_precision": 0.6280111982698415, |
|
"eval_recall": 0.45813565668329703, |
|
"eval_runtime": 63.6697, |
|
"eval_samples_per_second": 736.095, |
|
"eval_steps_per_second": 5.764, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 12.196601941747574, |
|
"grad_norm": 0.5225503444671631, |
|
"learning_rate": 0.00011705097087378641, |
|
"loss": 0.2405, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 12.257281553398059, |
|
"grad_norm": 0.7009778022766113, |
|
"learning_rate": 0.00011614077669902912, |
|
"loss": 0.2375, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 12.317961165048544, |
|
"grad_norm": 0.5107778310775757, |
|
"learning_rate": 0.00011523058252427184, |
|
"loss": 0.2396, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 12.37864077669903, |
|
"grad_norm": 0.5113606452941895, |
|
"learning_rate": 0.00011432038834951455, |
|
"loss": 0.2379, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 12.439320388349515, |
|
"grad_norm": 0.6248862743377686, |
|
"learning_rate": 0.00011341019417475727, |
|
"loss": 0.2453, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"grad_norm": 0.5648460984230042, |
|
"learning_rate": 0.0001125, |
|
"loss": 0.2391, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 12.560679611650485, |
|
"grad_norm": 0.7846024632453918, |
|
"learning_rate": 0.0001115898058252427, |
|
"loss": 0.2445, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 12.62135922330097, |
|
"grad_norm": 0.8849174976348877, |
|
"learning_rate": 0.00011067961165048543, |
|
"loss": 0.2392, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 12.682038834951456, |
|
"grad_norm": 0.5515842437744141, |
|
"learning_rate": 0.00010976941747572814, |
|
"loss": 0.2429, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 12.742718446601941, |
|
"grad_norm": 0.9129722118377686, |
|
"learning_rate": 0.00010885922330097086, |
|
"loss": 0.2459, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 12.742718446601941, |
|
"eval_accuracy": 0.7137218085219877, |
|
"eval_f1_macro": 0.49319871257798575, |
|
"eval_loss": 0.25367996096611023, |
|
"eval_precision": 0.6236252217705004, |
|
"eval_recall": 0.4640439902470517, |
|
"eval_runtime": 63.865, |
|
"eval_samples_per_second": 733.844, |
|
"eval_steps_per_second": 5.746, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 12.803398058252426, |
|
"grad_norm": 0.8533498048782349, |
|
"learning_rate": 0.00010794902912621357, |
|
"loss": 0.2384, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 12.864077669902912, |
|
"grad_norm": 0.5744712352752686, |
|
"learning_rate": 0.00010703883495145631, |
|
"loss": 0.241, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 12.924757281553399, |
|
"grad_norm": 0.9214051961898804, |
|
"learning_rate": 0.00010612864077669902, |
|
"loss": 0.2437, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 12.985436893203884, |
|
"grad_norm": 0.561357319355011, |
|
"learning_rate": 0.00010521844660194174, |
|
"loss": 0.2394, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 13.04611650485437, |
|
"grad_norm": 0.746048629283905, |
|
"learning_rate": 0.00010430825242718445, |
|
"loss": 0.2367, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 13.106796116504855, |
|
"grad_norm": 0.5449076294898987, |
|
"learning_rate": 0.00010339805825242718, |
|
"loss": 0.2392, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 13.16747572815534, |
|
"grad_norm": 0.7307832837104797, |
|
"learning_rate": 0.00010248786407766989, |
|
"loss": 0.2379, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 13.228155339805825, |
|
"grad_norm": 0.9420238137245178, |
|
"learning_rate": 0.00010157766990291262, |
|
"loss": 0.2348, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 13.28883495145631, |
|
"grad_norm": 0.43917521834373474, |
|
"learning_rate": 0.00010066747572815533, |
|
"loss": 0.2375, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 13.349514563106796, |
|
"grad_norm": 0.5659081339836121, |
|
"learning_rate": 9.975728155339806e-05, |
|
"loss": 0.24, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 13.349514563106796, |
|
"eval_accuracy": 0.7130176883521454, |
|
"eval_f1_macro": 0.4889124257466886, |
|
"eval_loss": 0.254099041223526, |
|
"eval_precision": 0.6244022587227488, |
|
"eval_recall": 0.4579878521374155, |
|
"eval_runtime": 63.8542, |
|
"eval_samples_per_second": 733.968, |
|
"eval_steps_per_second": 5.747, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 13.410194174757281, |
|
"grad_norm": 0.6600030660629272, |
|
"learning_rate": 9.884708737864077e-05, |
|
"loss": 0.2355, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 13.470873786407767, |
|
"grad_norm": 0.49261102080345154, |
|
"learning_rate": 9.793689320388349e-05, |
|
"loss": 0.2388, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 13.531553398058252, |
|
"grad_norm": 0.9407384395599365, |
|
"learning_rate": 9.70266990291262e-05, |
|
"loss": 0.2384, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 13.592233009708737, |
|
"grad_norm": 0.9749634265899658, |
|
"learning_rate": 9.611650485436893e-05, |
|
"loss": 0.2392, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 13.652912621359224, |
|
"grad_norm": 0.5800509452819824, |
|
"learning_rate": 9.520631067961164e-05, |
|
"loss": 0.2373, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 13.71359223300971, |
|
"grad_norm": 0.7015953063964844, |
|
"learning_rate": 9.429611650485437e-05, |
|
"loss": 0.2384, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 13.774271844660195, |
|
"grad_norm": 0.5446757078170776, |
|
"learning_rate": 9.338592233009708e-05, |
|
"loss": 0.2399, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 13.83495145631068, |
|
"grad_norm": 1.1201865673065186, |
|
"learning_rate": 9.247572815533979e-05, |
|
"loss": 0.2423, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 13.895631067961165, |
|
"grad_norm": 0.789397656917572, |
|
"learning_rate": 9.156553398058251e-05, |
|
"loss": 0.2401, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 13.95631067961165, |
|
"grad_norm": 0.9811580777168274, |
|
"learning_rate": 9.065533980582523e-05, |
|
"loss": 0.2429, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 13.95631067961165, |
|
"eval_accuracy": 0.7146819723899546, |
|
"eval_f1_macro": 0.4950660124078677, |
|
"eval_loss": 0.2540215849876404, |
|
"eval_precision": 0.6269585302337652, |
|
"eval_recall": 0.4640707919907739, |
|
"eval_runtime": 63.4303, |
|
"eval_samples_per_second": 738.874, |
|
"eval_steps_per_second": 5.786, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 14.016990291262136, |
|
"grad_norm": 0.6500277519226074, |
|
"learning_rate": 8.974514563106796e-05, |
|
"loss": 0.2399, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 14.077669902912621, |
|
"grad_norm": 0.7648112177848816, |
|
"learning_rate": 8.883495145631067e-05, |
|
"loss": 0.2355, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 14.138349514563107, |
|
"grad_norm": 0.6670911312103271, |
|
"learning_rate": 8.792475728155339e-05, |
|
"loss": 0.2371, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 14.199029126213592, |
|
"grad_norm": 0.8316827416419983, |
|
"learning_rate": 8.70145631067961e-05, |
|
"loss": 0.2355, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 14.259708737864077, |
|
"grad_norm": 0.4973256587982178, |
|
"learning_rate": 8.610436893203884e-05, |
|
"loss": 0.24, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 14.320388349514563, |
|
"grad_norm": 0.6986468434333801, |
|
"learning_rate": 8.519417475728155e-05, |
|
"loss": 0.2319, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 14.381067961165048, |
|
"grad_norm": 0.5316904783248901, |
|
"learning_rate": 8.428398058252427e-05, |
|
"loss": 0.2334, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 14.441747572815533, |
|
"grad_norm": 0.5285237431526184, |
|
"learning_rate": 8.337378640776698e-05, |
|
"loss": 0.2372, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 14.50242718446602, |
|
"grad_norm": 0.7617068290710449, |
|
"learning_rate": 8.24635922330097e-05, |
|
"loss": 0.2434, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 14.563106796116505, |
|
"grad_norm": 0.4870689809322357, |
|
"learning_rate": 8.155339805825241e-05, |
|
"loss": 0.2388, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 14.563106796116505, |
|
"eval_accuracy": 0.7128896665030832, |
|
"eval_f1_macro": 0.48631927316876294, |
|
"eval_loss": 0.2543109357357025, |
|
"eval_precision": 0.6364589670810381, |
|
"eval_recall": 0.451184525833245, |
|
"eval_runtime": 63.5544, |
|
"eval_samples_per_second": 737.432, |
|
"eval_steps_per_second": 5.775, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 14.62378640776699, |
|
"grad_norm": 0.48867112398147583, |
|
"learning_rate": 8.064320388349515e-05, |
|
"loss": 0.2424, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 14.684466019417476, |
|
"grad_norm": 1.3717137575149536, |
|
"learning_rate": 7.973300970873786e-05, |
|
"loss": 0.2373, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 14.745145631067961, |
|
"grad_norm": 0.5244446396827698, |
|
"learning_rate": 7.882281553398058e-05, |
|
"loss": 0.2406, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 14.805825242718447, |
|
"grad_norm": 0.49141696095466614, |
|
"learning_rate": 7.791262135922329e-05, |
|
"loss": 0.2345, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 14.866504854368932, |
|
"grad_norm": 0.5758472681045532, |
|
"learning_rate": 7.700242718446602e-05, |
|
"loss": 0.2372, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 14.927184466019417, |
|
"grad_norm": 0.4706755578517914, |
|
"learning_rate": 7.609223300970873e-05, |
|
"loss": 0.2363, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 14.987864077669903, |
|
"grad_norm": 1.1661343574523926, |
|
"learning_rate": 7.518203883495146e-05, |
|
"loss": 0.2363, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 15.048543689320388, |
|
"grad_norm": 0.6530236005783081, |
|
"learning_rate": 7.427184466019417e-05, |
|
"loss": 0.2371, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 15.109223300970873, |
|
"grad_norm": 0.6235638856887817, |
|
"learning_rate": 7.336165048543688e-05, |
|
"loss": 0.2365, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 15.169902912621358, |
|
"grad_norm": 1.0099172592163086, |
|
"learning_rate": 7.24514563106796e-05, |
|
"loss": 0.2333, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 15.169902912621358, |
|
"eval_accuracy": 0.7117374698615231, |
|
"eval_f1_macro": 0.4930065435769275, |
|
"eval_loss": 0.25366273522377014, |
|
"eval_precision": 0.6254110142417604, |
|
"eval_recall": 0.4636190323809964, |
|
"eval_runtime": 63.6634, |
|
"eval_samples_per_second": 736.168, |
|
"eval_steps_per_second": 5.765, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 15.230582524271846, |
|
"grad_norm": 0.4479919672012329, |
|
"learning_rate": 7.154126213592233e-05, |
|
"loss": 0.2363, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 15.29126213592233, |
|
"grad_norm": 0.5574977993965149, |
|
"learning_rate": 7.063106796116504e-05, |
|
"loss": 0.2323, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 15.351941747572816, |
|
"grad_norm": 0.815531849861145, |
|
"learning_rate": 6.972087378640776e-05, |
|
"loss": 0.2378, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 15.412621359223301, |
|
"grad_norm": 0.47219938039779663, |
|
"learning_rate": 6.881067961165048e-05, |
|
"loss": 0.2363, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 15.473300970873787, |
|
"grad_norm": 1.4584532976150513, |
|
"learning_rate": 6.79004854368932e-05, |
|
"loss": 0.2323, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 15.533980582524272, |
|
"grad_norm": 0.5099394917488098, |
|
"learning_rate": 6.699029126213592e-05, |
|
"loss": 0.2343, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 15.594660194174757, |
|
"grad_norm": 0.624839186668396, |
|
"learning_rate": 6.608009708737864e-05, |
|
"loss": 0.2357, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 15.655339805825243, |
|
"grad_norm": 0.9318602085113525, |
|
"learning_rate": 6.516990291262135e-05, |
|
"loss": 0.2331, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 15.716019417475728, |
|
"grad_norm": 0.6715418100357056, |
|
"learning_rate": 6.425970873786407e-05, |
|
"loss": 0.2365, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 15.776699029126213, |
|
"grad_norm": 0.48603758215904236, |
|
"learning_rate": 6.334951456310678e-05, |
|
"loss": 0.2392, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 15.776699029126213, |
|
"eval_accuracy": 0.7145326135660486, |
|
"eval_f1_macro": 0.49345940490874235, |
|
"eval_loss": 0.2531893253326416, |
|
"eval_precision": 0.6264577154052439, |
|
"eval_recall": 0.4631683363045007, |
|
"eval_runtime": 63.9314, |
|
"eval_samples_per_second": 733.082, |
|
"eval_steps_per_second": 5.741, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 15.837378640776699, |
|
"grad_norm": 1.03753662109375, |
|
"learning_rate": 6.243932038834951e-05, |
|
"loss": 0.234, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 15.898058252427184, |
|
"grad_norm": 0.7353742718696594, |
|
"learning_rate": 6.152912621359223e-05, |
|
"loss": 0.2381, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 15.95873786407767, |
|
"grad_norm": 0.8260138034820557, |
|
"learning_rate": 6.061893203883495e-05, |
|
"loss": 0.2394, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 16.019417475728154, |
|
"grad_norm": 0.7960408926010132, |
|
"learning_rate": 5.9708737864077663e-05, |
|
"loss": 0.2377, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 16.08009708737864, |
|
"grad_norm": 0.5640716552734375, |
|
"learning_rate": 5.879854368932038e-05, |
|
"loss": 0.2374, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 16.140776699029125, |
|
"grad_norm": 0.8281972408294678, |
|
"learning_rate": 5.78883495145631e-05, |
|
"loss": 0.2369, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 16.20145631067961, |
|
"grad_norm": 0.49466079473495483, |
|
"learning_rate": 5.697815533980582e-05, |
|
"loss": 0.2352, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 16.262135922330096, |
|
"grad_norm": 0.5278394818305969, |
|
"learning_rate": 5.6067961165048536e-05, |
|
"loss": 0.2321, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 16.32281553398058, |
|
"grad_norm": 1.5943635702133179, |
|
"learning_rate": 5.515776699029126e-05, |
|
"loss": 0.2368, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 16.383495145631066, |
|
"grad_norm": 0.7204076647758484, |
|
"learning_rate": 5.4247572815533976e-05, |
|
"loss": 0.2344, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 16.383495145631066, |
|
"eval_accuracy": 0.7135511127232381, |
|
"eval_f1_macro": 0.4959826756763837, |
|
"eval_loss": 0.2531072497367859, |
|
"eval_precision": 0.5967354989597543, |
|
"eval_recall": 0.4670103239766295, |
|
"eval_runtime": 63.8776, |
|
"eval_samples_per_second": 733.7, |
|
"eval_steps_per_second": 5.745, |
|
"step": 27000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 32960, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 20, |
|
"save_steps": 1000, |
|
"total_flos": 1.8182410386600591e+18, |
|
"train_batch_size": 256, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|