|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 2187, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0004572473708276177, |
|
"grad_norm": 8.096893602247533, |
|
"learning_rate": 4.5662100456621004e-08, |
|
"loss": 4.0564, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.002286236854138089, |
|
"grad_norm": 9.362922497604572, |
|
"learning_rate": 2.2831050228310502e-07, |
|
"loss": 4.0286, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.004572473708276178, |
|
"grad_norm": 8.436200820951127, |
|
"learning_rate": 4.5662100456621004e-07, |
|
"loss": 4.0215, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.006858710562414266, |
|
"grad_norm": 8.311166463949595, |
|
"learning_rate": 6.849315068493151e-07, |
|
"loss": 3.995, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.009144947416552356, |
|
"grad_norm": 8.314737539002055, |
|
"learning_rate": 9.132420091324201e-07, |
|
"loss": 4.0089, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.011431184270690443, |
|
"grad_norm": 8.207038819761589, |
|
"learning_rate": 1.1415525114155251e-06, |
|
"loss": 4.0363, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.013717421124828532, |
|
"grad_norm": 7.559088257570073, |
|
"learning_rate": 1.3698630136986302e-06, |
|
"loss": 3.9854, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01600365797896662, |
|
"grad_norm": 7.1469550333759315, |
|
"learning_rate": 1.5981735159817353e-06, |
|
"loss": 3.9172, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01828989483310471, |
|
"grad_norm": 5.971779564925808, |
|
"learning_rate": 1.8264840182648401e-06, |
|
"loss": 3.7833, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0205761316872428, |
|
"grad_norm": 5.297232176162252, |
|
"learning_rate": 2.0547945205479454e-06, |
|
"loss": 3.7258, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.022862368541380886, |
|
"grad_norm": 4.217560181118984, |
|
"learning_rate": 2.2831050228310503e-06, |
|
"loss": 3.6069, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.025148605395518976, |
|
"grad_norm": 3.786598130029432, |
|
"learning_rate": 2.511415525114155e-06, |
|
"loss": 3.553, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.027434842249657063, |
|
"grad_norm": 2.6863763381878782, |
|
"learning_rate": 2.7397260273972604e-06, |
|
"loss": 3.4564, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.029721079103795154, |
|
"grad_norm": 2.322206779034821, |
|
"learning_rate": 2.9680365296803653e-06, |
|
"loss": 3.331, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.03200731595793324, |
|
"grad_norm": 1.7328570105860337, |
|
"learning_rate": 3.1963470319634706e-06, |
|
"loss": 3.2806, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03429355281207133, |
|
"grad_norm": 1.4846651112766411, |
|
"learning_rate": 3.4246575342465754e-06, |
|
"loss": 3.2356, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.03657978966620942, |
|
"grad_norm": 1.1848731903024705, |
|
"learning_rate": 3.6529680365296803e-06, |
|
"loss": 3.1934, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.038866026520347506, |
|
"grad_norm": 1.0381920877926754, |
|
"learning_rate": 3.881278538812785e-06, |
|
"loss": 3.1245, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0411522633744856, |
|
"grad_norm": 0.9141218079482407, |
|
"learning_rate": 4.109589041095891e-06, |
|
"loss": 3.0468, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.04343850022862369, |
|
"grad_norm": 1.0161753499176187, |
|
"learning_rate": 4.337899543378996e-06, |
|
"loss": 3.0329, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.04572473708276177, |
|
"grad_norm": 0.8996453649527762, |
|
"learning_rate": 4.566210045662101e-06, |
|
"loss": 2.9958, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04801097393689986, |
|
"grad_norm": 0.8082797335102323, |
|
"learning_rate": 4.7945205479452054e-06, |
|
"loss": 2.9651, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.05029721079103795, |
|
"grad_norm": 0.7090431405422901, |
|
"learning_rate": 5.02283105022831e-06, |
|
"loss": 2.927, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.05258344764517604, |
|
"grad_norm": 1.2265537925663061, |
|
"learning_rate": 5.251141552511416e-06, |
|
"loss": 2.862, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.05486968449931413, |
|
"grad_norm": 0.650571444620453, |
|
"learning_rate": 5.479452054794521e-06, |
|
"loss": 2.857, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.05715592135345222, |
|
"grad_norm": 0.7089485111846239, |
|
"learning_rate": 5.7077625570776266e-06, |
|
"loss": 2.8209, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.05944215820759031, |
|
"grad_norm": 0.5922494361050838, |
|
"learning_rate": 5.936073059360731e-06, |
|
"loss": 2.8037, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.06172839506172839, |
|
"grad_norm": 0.5597217919230902, |
|
"learning_rate": 6.164383561643836e-06, |
|
"loss": 2.7487, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.06401463191586648, |
|
"grad_norm": 0.6045746583730743, |
|
"learning_rate": 6.392694063926941e-06, |
|
"loss": 2.6981, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.06630086877000457, |
|
"grad_norm": 0.6479924774135967, |
|
"learning_rate": 6.621004566210046e-06, |
|
"loss": 2.7036, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.06858710562414266, |
|
"grad_norm": 0.760723993748018, |
|
"learning_rate": 6.849315068493151e-06, |
|
"loss": 2.6821, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.07087334247828075, |
|
"grad_norm": 0.5889973577684341, |
|
"learning_rate": 7.077625570776257e-06, |
|
"loss": 2.6882, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.07315957933241884, |
|
"grad_norm": 0.6201384588278992, |
|
"learning_rate": 7.305936073059361e-06, |
|
"loss": 2.6441, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.07544581618655692, |
|
"grad_norm": 0.645862573214957, |
|
"learning_rate": 7.534246575342466e-06, |
|
"loss": 2.5878, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.07773205304069501, |
|
"grad_norm": 0.880791499233313, |
|
"learning_rate": 7.76255707762557e-06, |
|
"loss": 2.5665, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.0800182898948331, |
|
"grad_norm": 0.7581098091472079, |
|
"learning_rate": 7.990867579908676e-06, |
|
"loss": 2.5423, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.0823045267489712, |
|
"grad_norm": 0.7502504535360037, |
|
"learning_rate": 8.219178082191782e-06, |
|
"loss": 2.5348, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.08459076360310928, |
|
"grad_norm": 0.9587325899501735, |
|
"learning_rate": 8.447488584474887e-06, |
|
"loss": 2.4652, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.08687700045724737, |
|
"grad_norm": 1.0327228370595574, |
|
"learning_rate": 8.675799086757991e-06, |
|
"loss": 2.4066, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.08916323731138547, |
|
"grad_norm": 0.8853835960264104, |
|
"learning_rate": 8.904109589041097e-06, |
|
"loss": 2.3642, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.09144947416552354, |
|
"grad_norm": 1.0446953486337078, |
|
"learning_rate": 9.132420091324201e-06, |
|
"loss": 2.3237, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.09373571101966163, |
|
"grad_norm": 1.1013758488210148, |
|
"learning_rate": 9.360730593607307e-06, |
|
"loss": 2.2331, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.09602194787379972, |
|
"grad_norm": 1.2192543249794923, |
|
"learning_rate": 9.589041095890411e-06, |
|
"loss": 2.1264, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.09830818472793781, |
|
"grad_norm": 1.3533953895273099, |
|
"learning_rate": 9.817351598173517e-06, |
|
"loss": 2.0554, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.1005944215820759, |
|
"grad_norm": 1.1876482609404326, |
|
"learning_rate": 9.999993629265979e-06, |
|
"loss": 1.9859, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.102880658436214, |
|
"grad_norm": 1.1847416528253172, |
|
"learning_rate": 9.999770655279843e-06, |
|
"loss": 1.8986, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.10516689529035209, |
|
"grad_norm": 1.3137466650624998, |
|
"learning_rate": 9.999229160826947e-06, |
|
"loss": 1.8, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.10745313214449016, |
|
"grad_norm": 1.830150495140023, |
|
"learning_rate": 9.998369180404283e-06, |
|
"loss": 1.7138, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.10973936899862825, |
|
"grad_norm": 1.1159850299398295, |
|
"learning_rate": 9.997190768798639e-06, |
|
"loss": 1.6867, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.11202560585276634, |
|
"grad_norm": 0.9727694366367986, |
|
"learning_rate": 9.995694001083103e-06, |
|
"loss": 1.6469, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.11431184270690443, |
|
"grad_norm": 1.135743426814773, |
|
"learning_rate": 9.993878972612276e-06, |
|
"loss": 1.5607, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.11659807956104253, |
|
"grad_norm": 1.0363437963731608, |
|
"learning_rate": 9.991745799016206e-06, |
|
"loss": 1.5332, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.11888431641518062, |
|
"grad_norm": 1.018006180331875, |
|
"learning_rate": 9.989294616193018e-06, |
|
"loss": 1.4962, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.1211705532693187, |
|
"grad_norm": 0.9493951106581935, |
|
"learning_rate": 9.986525580300253e-06, |
|
"loss": 1.4403, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.12345679012345678, |
|
"grad_norm": 4.150830186272059, |
|
"learning_rate": 9.983438867744923e-06, |
|
"loss": 1.4382, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.12574302697759487, |
|
"grad_norm": 0.8458476848705546, |
|
"learning_rate": 9.980034675172274e-06, |
|
"loss": 1.4248, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.12802926383173296, |
|
"grad_norm": 1.8171861028727991, |
|
"learning_rate": 9.976313219453255e-06, |
|
"loss": 1.4055, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.13031550068587106, |
|
"grad_norm": 0.7389926811741014, |
|
"learning_rate": 9.972274737670702e-06, |
|
"loss": 1.4033, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.13260173754000915, |
|
"grad_norm": 0.8834746515415843, |
|
"learning_rate": 9.967919487104237e-06, |
|
"loss": 1.3724, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.13488797439414724, |
|
"grad_norm": 0.8166186304734012, |
|
"learning_rate": 9.963247745213876e-06, |
|
"loss": 1.3721, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.13717421124828533, |
|
"grad_norm": 0.6771475216933378, |
|
"learning_rate": 9.958259809622353e-06, |
|
"loss": 1.3555, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.13946044810242342, |
|
"grad_norm": 0.60525762012324, |
|
"learning_rate": 9.952955998096155e-06, |
|
"loss": 1.36, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.1417466849565615, |
|
"grad_norm": 0.6126617626167846, |
|
"learning_rate": 9.94733664852529e-06, |
|
"loss": 1.353, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.1440329218106996, |
|
"grad_norm": 0.6630794657190928, |
|
"learning_rate": 9.941402118901743e-06, |
|
"loss": 1.3359, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.1463191586648377, |
|
"grad_norm": 0.6758533351396738, |
|
"learning_rate": 9.935152787296689e-06, |
|
"loss": 1.3402, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.14860539551897575, |
|
"grad_norm": 0.739719330356037, |
|
"learning_rate": 9.928589051836392e-06, |
|
"loss": 1.3346, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.15089163237311384, |
|
"grad_norm": 0.7258290118963521, |
|
"learning_rate": 9.921711330676848e-06, |
|
"loss": 1.3356, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.15317786922725193, |
|
"grad_norm": 0.6274092924270468, |
|
"learning_rate": 9.91452006197715e-06, |
|
"loss": 1.3362, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.15546410608139002, |
|
"grad_norm": 0.768028072114212, |
|
"learning_rate": 9.907015703871558e-06, |
|
"loss": 1.3214, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.15775034293552812, |
|
"grad_norm": 0.7738373400419118, |
|
"learning_rate": 9.899198734440335e-06, |
|
"loss": 1.331, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.1600365797896662, |
|
"grad_norm": 0.6855410863811031, |
|
"learning_rate": 9.891069651679273e-06, |
|
"loss": 1.3142, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1623228166438043, |
|
"grad_norm": 0.6405023247699122, |
|
"learning_rate": 9.882628973467972e-06, |
|
"loss": 1.3171, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.1646090534979424, |
|
"grad_norm": 0.6764400756880153, |
|
"learning_rate": 9.873877237536854e-06, |
|
"loss": 1.3189, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.16689529035208048, |
|
"grad_norm": 0.6298462983903607, |
|
"learning_rate": 9.86481500143289e-06, |
|
"loss": 1.3059, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.16918152720621857, |
|
"grad_norm": 0.6606697771559132, |
|
"learning_rate": 9.855442842484101e-06, |
|
"loss": 1.3267, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.17146776406035666, |
|
"grad_norm": 0.5895037669135822, |
|
"learning_rate": 9.84576135776276e-06, |
|
"loss": 1.3057, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.17375400091449475, |
|
"grad_norm": 0.5762405642901876, |
|
"learning_rate": 9.835771164047365e-06, |
|
"loss": 1.3016, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.17604023776863284, |
|
"grad_norm": 0.6301891918568133, |
|
"learning_rate": 9.825472897783344e-06, |
|
"loss": 1.3046, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.17832647462277093, |
|
"grad_norm": 0.6189017845225122, |
|
"learning_rate": 9.814867215042503e-06, |
|
"loss": 1.3089, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.18061271147690902, |
|
"grad_norm": 0.6279515665165573, |
|
"learning_rate": 9.803954791481239e-06, |
|
"loss": 1.3011, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.18289894833104708, |
|
"grad_norm": 0.6380039476156935, |
|
"learning_rate": 9.792736322297489e-06, |
|
"loss": 1.2758, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.18518518518518517, |
|
"grad_norm": 0.7506004279154695, |
|
"learning_rate": 9.781212522186442e-06, |
|
"loss": 1.312, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.18747142203932327, |
|
"grad_norm": 0.7054181242720778, |
|
"learning_rate": 9.769384125295012e-06, |
|
"loss": 1.3112, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.18975765889346136, |
|
"grad_norm": 0.5797880483237029, |
|
"learning_rate": 9.757251885175063e-06, |
|
"loss": 1.2998, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.19204389574759945, |
|
"grad_norm": 0.6040659600524477, |
|
"learning_rate": 9.744816574735405e-06, |
|
"loss": 1.3018, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.19433013260173754, |
|
"grad_norm": 0.7044299546094256, |
|
"learning_rate": 9.732078986192552e-06, |
|
"loss": 1.2818, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.19661636945587563, |
|
"grad_norm": 0.567841572649114, |
|
"learning_rate": 9.719039931020258e-06, |
|
"loss": 1.2733, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.19890260631001372, |
|
"grad_norm": 0.5378351616772565, |
|
"learning_rate": 9.705700239897809e-06, |
|
"loss": 1.2861, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.2011888431641518, |
|
"grad_norm": 0.5372339490006793, |
|
"learning_rate": 9.692060762657118e-06, |
|
"loss": 1.2821, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.2034750800182899, |
|
"grad_norm": 0.6353680076674888, |
|
"learning_rate": 9.678122368228571e-06, |
|
"loss": 1.2643, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.205761316872428, |
|
"grad_norm": 0.6263499547366734, |
|
"learning_rate": 9.66388594458568e-06, |
|
"loss": 1.2826, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.20804755372656608, |
|
"grad_norm": 0.6119180746423146, |
|
"learning_rate": 9.649352398688506e-06, |
|
"loss": 1.2856, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.21033379058070417, |
|
"grad_norm": 0.6640618234127624, |
|
"learning_rate": 9.634522656425885e-06, |
|
"loss": 1.2765, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.21262002743484226, |
|
"grad_norm": 0.6253602428713037, |
|
"learning_rate": 9.619397662556434e-06, |
|
"loss": 1.2661, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.21490626428898033, |
|
"grad_norm": 0.6463257272674591, |
|
"learning_rate": 9.603978380648375e-06, |
|
"loss": 1.2838, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.21719250114311842, |
|
"grad_norm": 0.6916869993480118, |
|
"learning_rate": 9.588265793018141e-06, |
|
"loss": 1.2785, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.2194787379972565, |
|
"grad_norm": 0.578420093141111, |
|
"learning_rate": 9.572260900667794e-06, |
|
"loss": 1.2627, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.2217649748513946, |
|
"grad_norm": 0.6016744117162259, |
|
"learning_rate": 9.555964723221258e-06, |
|
"loss": 1.2672, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.2240512117055327, |
|
"grad_norm": 0.6325422647436533, |
|
"learning_rate": 9.539378298859365e-06, |
|
"loss": 1.2667, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.22633744855967078, |
|
"grad_norm": 0.674420764332063, |
|
"learning_rate": 9.522502684253709e-06, |
|
"loss": 1.2601, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.22862368541380887, |
|
"grad_norm": 0.6942742236531446, |
|
"learning_rate": 9.505338954499332e-06, |
|
"loss": 1.275, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.23090992226794696, |
|
"grad_norm": 0.5661617220667517, |
|
"learning_rate": 9.487888203046232e-06, |
|
"loss": 1.2683, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.23319615912208505, |
|
"grad_norm": 0.6389133947347537, |
|
"learning_rate": 9.4701515416297e-06, |
|
"loss": 1.2659, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.23548239597622314, |
|
"grad_norm": 0.561786602813537, |
|
"learning_rate": 9.452130100199504e-06, |
|
"loss": 1.2664, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.23776863283036123, |
|
"grad_norm": 0.5666699221383189, |
|
"learning_rate": 9.433825026847891e-06, |
|
"loss": 1.2573, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.24005486968449932, |
|
"grad_norm": 0.6718711112993888, |
|
"learning_rate": 9.415237487736452e-06, |
|
"loss": 1.2545, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.2423411065386374, |
|
"grad_norm": 0.5637527283960878, |
|
"learning_rate": 9.396368667021835e-06, |
|
"loss": 1.2723, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.2446273433927755, |
|
"grad_norm": 0.583426898925874, |
|
"learning_rate": 9.377219766780288e-06, |
|
"loss": 1.2473, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.24691358024691357, |
|
"grad_norm": 0.7422622561747031, |
|
"learning_rate": 9.3577920069311e-06, |
|
"loss": 1.2609, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.24919981710105166, |
|
"grad_norm": 0.7536416453907702, |
|
"learning_rate": 9.338086625158867e-06, |
|
"loss": 1.2655, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.25148605395518975, |
|
"grad_norm": 0.5911621999933799, |
|
"learning_rate": 9.318104876834652e-06, |
|
"loss": 1.2652, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.25377229080932784, |
|
"grad_norm": 0.6482915887304207, |
|
"learning_rate": 9.297848034936007e-06, |
|
"loss": 1.2488, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.25605852766346593, |
|
"grad_norm": 0.7813862221549358, |
|
"learning_rate": 9.277317389965871e-06, |
|
"loss": 1.2678, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.258344764517604, |
|
"grad_norm": 0.601959447185496, |
|
"learning_rate": 9.256514249870366e-06, |
|
"loss": 1.2549, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.2606310013717421, |
|
"grad_norm": 0.5439593292691556, |
|
"learning_rate": 9.235439939955458e-06, |
|
"loss": 1.2311, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.2629172382258802, |
|
"grad_norm": 0.6462948109732727, |
|
"learning_rate": 9.214095802802533e-06, |
|
"loss": 1.2605, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.2652034750800183, |
|
"grad_norm": 0.6523908850821281, |
|
"learning_rate": 9.192483198182876e-06, |
|
"loss": 1.2577, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.2674897119341564, |
|
"grad_norm": 0.6285230592028435, |
|
"learning_rate": 9.170603502971017e-06, |
|
"loss": 1.233, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.2697759487882945, |
|
"grad_norm": 0.5990676661488948, |
|
"learning_rate": 9.148458111057043e-06, |
|
"loss": 1.2444, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.27206218564243256, |
|
"grad_norm": 0.5443537881683997, |
|
"learning_rate": 9.12604843325778e-06, |
|
"loss": 1.2282, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.27434842249657065, |
|
"grad_norm": 0.5804764131758829, |
|
"learning_rate": 9.103375897226919e-06, |
|
"loss": 1.253, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.27663465935070874, |
|
"grad_norm": 0.5905170219986889, |
|
"learning_rate": 9.080441947364065e-06, |
|
"loss": 1.2472, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.27892089620484684, |
|
"grad_norm": 0.6003218456115103, |
|
"learning_rate": 9.057248044722718e-06, |
|
"loss": 1.2421, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.2812071330589849, |
|
"grad_norm": 0.5683857920528798, |
|
"learning_rate": 9.033795666917191e-06, |
|
"loss": 1.2551, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.283493369913123, |
|
"grad_norm": 0.5908776822300396, |
|
"learning_rate": 9.010086308028487e-06, |
|
"loss": 1.2375, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.2857796067672611, |
|
"grad_norm": 0.6118010788168986, |
|
"learning_rate": 8.986121478509096e-06, |
|
"loss": 1.2347, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.2880658436213992, |
|
"grad_norm": 0.5787813457678733, |
|
"learning_rate": 8.961902705086785e-06, |
|
"loss": 1.2395, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.2903520804755373, |
|
"grad_norm": 0.6290839595278495, |
|
"learning_rate": 8.937431530667329e-06, |
|
"loss": 1.2263, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.2926383173296754, |
|
"grad_norm": 0.5459763353494508, |
|
"learning_rate": 8.912709514236218e-06, |
|
"loss": 1.2285, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.29492455418381347, |
|
"grad_norm": 0.6301840515917086, |
|
"learning_rate": 8.887738230759334e-06, |
|
"loss": 1.2374, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.2972107910379515, |
|
"grad_norm": 0.5413584040020849, |
|
"learning_rate": 8.862519271082624e-06, |
|
"loss": 1.2505, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.2994970278920896, |
|
"grad_norm": 0.5979355091788396, |
|
"learning_rate": 8.83705424183074e-06, |
|
"loss": 1.2238, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.3017832647462277, |
|
"grad_norm": 0.6873493941298675, |
|
"learning_rate": 8.811344765304698e-06, |
|
"loss": 1.2262, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.3040695016003658, |
|
"grad_norm": 0.6699975954695512, |
|
"learning_rate": 8.785392479378522e-06, |
|
"loss": 1.23, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.30635573845450387, |
|
"grad_norm": 0.6860546025784545, |
|
"learning_rate": 8.759199037394888e-06, |
|
"loss": 1.2424, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.30864197530864196, |
|
"grad_norm": 0.7598573834174616, |
|
"learning_rate": 8.732766108059814e-06, |
|
"loss": 1.2138, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.31092821216278005, |
|
"grad_norm": 0.723323270057115, |
|
"learning_rate": 8.70609537533634e-06, |
|
"loss": 1.2373, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.31321444901691814, |
|
"grad_norm": 0.6170455054157933, |
|
"learning_rate": 8.679188538337248e-06, |
|
"loss": 1.2257, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.31550068587105623, |
|
"grad_norm": 0.7413957440287698, |
|
"learning_rate": 8.652047311216823e-06, |
|
"loss": 1.2075, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.3177869227251943, |
|
"grad_norm": 0.7424365012242525, |
|
"learning_rate": 8.62467342306164e-06, |
|
"loss": 1.2238, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.3200731595793324, |
|
"grad_norm": 0.8566227798899636, |
|
"learning_rate": 8.597068617780419e-06, |
|
"loss": 1.2278, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.3223593964334705, |
|
"grad_norm": 0.647075376724737, |
|
"learning_rate": 8.569234653992916e-06, |
|
"loss": 1.2407, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.3246456332876086, |
|
"grad_norm": 0.6249088936722902, |
|
"learning_rate": 8.541173304917895e-06, |
|
"loss": 1.2231, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.3269318701417467, |
|
"grad_norm": 0.70817264277616, |
|
"learning_rate": 8.512886358260162e-06, |
|
"loss": 1.2345, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.3292181069958848, |
|
"grad_norm": 0.5956107721750036, |
|
"learning_rate": 8.484375616096658e-06, |
|
"loss": 1.225, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.33150434385002286, |
|
"grad_norm": 0.6062042871270218, |
|
"learning_rate": 8.455642894761684e-06, |
|
"loss": 1.2185, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.33379058070416096, |
|
"grad_norm": 0.66611343630398, |
|
"learning_rate": 8.426690024731161e-06, |
|
"loss": 1.2171, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.33607681755829905, |
|
"grad_norm": 0.6006939272932527, |
|
"learning_rate": 8.39751885050603e-06, |
|
"loss": 1.2168, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.33836305441243714, |
|
"grad_norm": 0.5888998376074026, |
|
"learning_rate": 8.36813123049474e-06, |
|
"loss": 1.2447, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.3406492912665752, |
|
"grad_norm": 0.6170255283448466, |
|
"learning_rate": 8.338529036894855e-06, |
|
"loss": 1.2386, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.3429355281207133, |
|
"grad_norm": 0.6592250171561639, |
|
"learning_rate": 8.308714155573785e-06, |
|
"loss": 1.2095, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.3452217649748514, |
|
"grad_norm": 0.5948350472440084, |
|
"learning_rate": 8.278688485948634e-06, |
|
"loss": 1.2204, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.3475080018289895, |
|
"grad_norm": 0.6884759018973265, |
|
"learning_rate": 8.248453940865204e-06, |
|
"loss": 1.2205, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.3497942386831276, |
|
"grad_norm": 0.5629453296642776, |
|
"learning_rate": 8.218012446476128e-06, |
|
"loss": 1.2087, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.3520804755372657, |
|
"grad_norm": 0.5703699859674032, |
|
"learning_rate": 8.187365942118162e-06, |
|
"loss": 1.2038, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.35436671239140377, |
|
"grad_norm": 0.5758055939006159, |
|
"learning_rate": 8.156516380188635e-06, |
|
"loss": 1.2015, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.35665294924554186, |
|
"grad_norm": 0.6814380489670292, |
|
"learning_rate": 8.125465726021068e-06, |
|
"loss": 1.2267, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.35893918609967995, |
|
"grad_norm": 0.58819101648096, |
|
"learning_rate": 8.09421595775997e-06, |
|
"loss": 1.2065, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.36122542295381804, |
|
"grad_norm": 0.599220106737159, |
|
"learning_rate": 8.062769066234807e-06, |
|
"loss": 1.2084, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.3635116598079561, |
|
"grad_norm": 0.5687079813226833, |
|
"learning_rate": 8.031127054833192e-06, |
|
"loss": 1.2311, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.36579789666209417, |
|
"grad_norm": 0.6076443328436887, |
|
"learning_rate": 7.999291939373232e-06, |
|
"loss": 1.209, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.36808413351623226, |
|
"grad_norm": 0.5767468288489239, |
|
"learning_rate": 7.967265747975124e-06, |
|
"loss": 1.2153, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.37037037037037035, |
|
"grad_norm": 0.6275130557605428, |
|
"learning_rate": 7.93505052093194e-06, |
|
"loss": 1.2206, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.37265660722450844, |
|
"grad_norm": 0.5920904031157348, |
|
"learning_rate": 7.90264831057965e-06, |
|
"loss": 1.2149, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.37494284407864653, |
|
"grad_norm": 0.5841477404583847, |
|
"learning_rate": 7.870061181166372e-06, |
|
"loss": 1.2134, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.3772290809327846, |
|
"grad_norm": 0.545565275285448, |
|
"learning_rate": 7.837291208720867e-06, |
|
"loss": 1.2185, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.3795153177869227, |
|
"grad_norm": 0.6183231148929101, |
|
"learning_rate": 7.804340480920274e-06, |
|
"loss": 1.2064, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.3818015546410608, |
|
"grad_norm": 0.5801259298558049, |
|
"learning_rate": 7.771211096957125e-06, |
|
"loss": 1.2049, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.3840877914951989, |
|
"grad_norm": 0.579347207611424, |
|
"learning_rate": 7.737905167405596e-06, |
|
"loss": 1.2185, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.386374028349337, |
|
"grad_norm": 0.6262921976973932, |
|
"learning_rate": 7.704424814087056e-06, |
|
"loss": 1.2137, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.3886602652034751, |
|
"grad_norm": 0.6070706881138944, |
|
"learning_rate": 7.670772169934902e-06, |
|
"loss": 1.2177, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.39094650205761317, |
|
"grad_norm": 0.5688216055326876, |
|
"learning_rate": 7.636949378858647e-06, |
|
"loss": 1.2016, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.39323273891175126, |
|
"grad_norm": 0.6166249078020826, |
|
"learning_rate": 7.602958595607375e-06, |
|
"loss": 1.1957, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.39551897576588935, |
|
"grad_norm": 0.5778886288472463, |
|
"learning_rate": 7.568801985632439e-06, |
|
"loss": 1.2105, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.39780521262002744, |
|
"grad_norm": 0.6732218435967291, |
|
"learning_rate": 7.5344817249495195e-06, |
|
"loss": 1.2047, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.40009144947416553, |
|
"grad_norm": 0.672208759556888, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 1.1854, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.4023776863283036, |
|
"grad_norm": 0.6180565492464766, |
|
"learning_rate": 7.465359007511667e-06, |
|
"loss": 1.185, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.4046639231824417, |
|
"grad_norm": 0.6266745151721254, |
|
"learning_rate": 7.430560954358764e-06, |
|
"loss": 1.2082, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.4069501600365798, |
|
"grad_norm": 0.6163182978581346, |
|
"learning_rate": 7.395608057421406e-06, |
|
"loss": 1.2194, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.4092363968907179, |
|
"grad_norm": 0.6262674693601461, |
|
"learning_rate": 7.360502543444339e-06, |
|
"loss": 1.2188, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.411522633744856, |
|
"grad_norm": 0.5549642780561265, |
|
"learning_rate": 7.325246648895089e-06, |
|
"loss": 1.1986, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.41380887059899407, |
|
"grad_norm": 0.5540368046559051, |
|
"learning_rate": 7.289842619821475e-06, |
|
"loss": 1.2175, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.41609510745313216, |
|
"grad_norm": 0.587023330497459, |
|
"learning_rate": 7.254292711708529e-06, |
|
"loss": 1.2029, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.41838134430727025, |
|
"grad_norm": 0.5513581130094706, |
|
"learning_rate": 7.218599189334799e-06, |
|
"loss": 1.2009, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.42066758116140834, |
|
"grad_norm": 0.7237520794327035, |
|
"learning_rate": 7.182764326628068e-06, |
|
"loss": 1.2063, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.42295381801554643, |
|
"grad_norm": 0.5476819110298711, |
|
"learning_rate": 7.146790406520491e-06, |
|
"loss": 1.2107, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.4252400548696845, |
|
"grad_norm": 0.5753924094787153, |
|
"learning_rate": 7.1106797208031554e-06, |
|
"loss": 1.2133, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.4275262917238226, |
|
"grad_norm": 0.6489054914059448, |
|
"learning_rate": 7.0744345699800755e-06, |
|
"loss": 1.1991, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.42981252857796065, |
|
"grad_norm": 0.6239602498665449, |
|
"learning_rate": 7.038057263121639e-06, |
|
"loss": 1.1937, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.43209876543209874, |
|
"grad_norm": 0.5954140813357963, |
|
"learning_rate": 7.001550117717499e-06, |
|
"loss": 1.2092, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.43438500228623683, |
|
"grad_norm": 0.5953175778315464, |
|
"learning_rate": 6.9649154595289326e-06, |
|
"loss": 1.1957, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.4366712391403749, |
|
"grad_norm": 0.6030938627687562, |
|
"learning_rate": 6.92815562244068e-06, |
|
"loss": 1.1827, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.438957475994513, |
|
"grad_norm": 0.6882999466791362, |
|
"learning_rate": 6.891272948312251e-06, |
|
"loss": 1.2102, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.4412437128486511, |
|
"grad_norm": 0.6080281045836577, |
|
"learning_rate": 6.854269786828741e-06, |
|
"loss": 1.2093, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.4435299497027892, |
|
"grad_norm": 0.756192409869553, |
|
"learning_rate": 6.817148495351131e-06, |
|
"loss": 1.2159, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.4458161865569273, |
|
"grad_norm": 0.5892520162590819, |
|
"learning_rate": 6.779911438766117e-06, |
|
"loss": 1.193, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.4481024234110654, |
|
"grad_norm": 0.6265917897470434, |
|
"learning_rate": 6.742560989335438e-06, |
|
"loss": 1.1951, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.45038866026520347, |
|
"grad_norm": 0.5927415516536023, |
|
"learning_rate": 6.705099526544757e-06, |
|
"loss": 1.1973, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.45267489711934156, |
|
"grad_norm": 0.5602604942191215, |
|
"learning_rate": 6.667529436952064e-06, |
|
"loss": 1.1945, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.45496113397347965, |
|
"grad_norm": 0.751574883051813, |
|
"learning_rate": 6.629853114035643e-06, |
|
"loss": 1.2134, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.45724737082761774, |
|
"grad_norm": 0.6000318274839507, |
|
"learning_rate": 6.5920729580415795e-06, |
|
"loss": 1.2104, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.45953360768175583, |
|
"grad_norm": 0.5783065549399249, |
|
"learning_rate": 6.554191375830861e-06, |
|
"loss": 1.2016, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.4618198445358939, |
|
"grad_norm": 0.5751980188798808, |
|
"learning_rate": 6.516210780726032e-06, |
|
"loss": 1.1794, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.464106081390032, |
|
"grad_norm": 0.6096335885035103, |
|
"learning_rate": 6.478133592357455e-06, |
|
"loss": 1.1816, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.4663923182441701, |
|
"grad_norm": 0.5848690144740822, |
|
"learning_rate": 6.43996223650916e-06, |
|
"loss": 1.1735, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.4686785550983082, |
|
"grad_norm": 0.6273777569367492, |
|
"learning_rate": 6.401699144964306e-06, |
|
"loss": 1.1864, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.4709647919524463, |
|
"grad_norm": 0.5772389229176554, |
|
"learning_rate": 6.3633467553502625e-06, |
|
"loss": 1.1953, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.4732510288065844, |
|
"grad_norm": 0.6320660706578101, |
|
"learning_rate": 6.32490751098331e-06, |
|
"loss": 1.1778, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.47553726566072246, |
|
"grad_norm": 0.628014857385664, |
|
"learning_rate": 6.286383860712982e-06, |
|
"loss": 1.1978, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.47782350251486055, |
|
"grad_norm": 0.6165011857453245, |
|
"learning_rate": 6.247778258766069e-06, |
|
"loss": 1.1783, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.48010973936899864, |
|
"grad_norm": 0.6680859473813631, |
|
"learning_rate": 6.209093164590253e-06, |
|
"loss": 1.1883, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.48239597622313674, |
|
"grad_norm": 0.6230269069079273, |
|
"learning_rate": 6.170331042697425e-06, |
|
"loss": 1.1923, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.4846822130772748, |
|
"grad_norm": 0.6472681484163015, |
|
"learning_rate": 6.131494362506693e-06, |
|
"loss": 1.1826, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.4869684499314129, |
|
"grad_norm": 0.6799978087591872, |
|
"learning_rate": 6.09258559818704e-06, |
|
"loss": 1.1829, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.489254686785551, |
|
"grad_norm": 0.5617426984448537, |
|
"learning_rate": 6.053607228499719e-06, |
|
"loss": 1.1941, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.4915409236396891, |
|
"grad_norm": 0.6444058153599652, |
|
"learning_rate": 6.014561736640334e-06, |
|
"loss": 1.2, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.49382716049382713, |
|
"grad_norm": 0.6016265988080601, |
|
"learning_rate": 5.975451610080643e-06, |
|
"loss": 1.1655, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.4961133973479652, |
|
"grad_norm": 0.7053148286233416, |
|
"learning_rate": 5.936279340410082e-06, |
|
"loss": 1.172, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.4983996342021033, |
|
"grad_norm": 0.5586357561653685, |
|
"learning_rate": 5.8970474231770445e-06, |
|
"loss": 1.1922, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.5006858710562414, |
|
"grad_norm": 0.7895760074140119, |
|
"learning_rate": 5.857758357729892e-06, |
|
"loss": 1.1839, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.5029721079103795, |
|
"grad_norm": 0.7313666592611404, |
|
"learning_rate": 5.8184146470577265e-06, |
|
"loss": 1.1813, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.5052583447645176, |
|
"grad_norm": 0.6067591576327228, |
|
"learning_rate": 5.779018797630934e-06, |
|
"loss": 1.1855, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.5075445816186557, |
|
"grad_norm": 0.6144330199450508, |
|
"learning_rate": 5.739573319241505e-06, |
|
"loss": 1.1924, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.5098308184727938, |
|
"grad_norm": 0.6075048668745815, |
|
"learning_rate": 5.7000807248431466e-06, |
|
"loss": 1.1783, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.5121170553269319, |
|
"grad_norm": 0.6763365315316732, |
|
"learning_rate": 5.66054353039118e-06, |
|
"loss": 1.1873, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.51440329218107, |
|
"grad_norm": 0.652936999197392, |
|
"learning_rate": 5.620964254682267e-06, |
|
"loss": 1.2019, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.516689529035208, |
|
"grad_norm": 0.7510930690144121, |
|
"learning_rate": 5.58134541919394e-06, |
|
"loss": 1.1863, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.5189757658893461, |
|
"grad_norm": 0.7485282723991191, |
|
"learning_rate": 5.5416895479239665e-06, |
|
"loss": 1.1878, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.5212620027434842, |
|
"grad_norm": 0.6650793765929232, |
|
"learning_rate": 5.501999167229554e-06, |
|
"loss": 1.1844, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.5235482395976223, |
|
"grad_norm": 0.6617004106280673, |
|
"learning_rate": 5.4622768056664e-06, |
|
"loss": 1.1819, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.5258344764517604, |
|
"grad_norm": 0.639306148093516, |
|
"learning_rate": 5.42252499382761e-06, |
|
"loss": 1.1844, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.5281207133058985, |
|
"grad_norm": 0.590573720499581, |
|
"learning_rate": 5.38274626418248e-06, |
|
"loss": 1.1848, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.5304069501600366, |
|
"grad_norm": 0.625235396788826, |
|
"learning_rate": 5.3429431509151515e-06, |
|
"loss": 1.1904, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.5326931870141747, |
|
"grad_norm": 0.5840052674712635, |
|
"learning_rate": 5.303118189763187e-06, |
|
"loss": 1.1829, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.5349794238683128, |
|
"grad_norm": 0.5940842973816081, |
|
"learning_rate": 5.263273917856e-06, |
|
"loss": 1.1774, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.5372656607224509, |
|
"grad_norm": 0.5991239115995499, |
|
"learning_rate": 5.22341287355324e-06, |
|
"loss": 1.1857, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.539551897576589, |
|
"grad_norm": 0.6248756548437343, |
|
"learning_rate": 5.183537596283075e-06, |
|
"loss": 1.1799, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.541838134430727, |
|
"grad_norm": 0.6023807247895316, |
|
"learning_rate": 5.143650626380417e-06, |
|
"loss": 1.1858, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.5441243712848651, |
|
"grad_norm": 0.6101959497751839, |
|
"learning_rate": 5.103754504925071e-06, |
|
"loss": 1.1961, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.5464106081390032, |
|
"grad_norm": 0.569676114190435, |
|
"learning_rate": 5.06385177357987e-06, |
|
"loss": 1.1766, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.5486968449931413, |
|
"grad_norm": 0.5819652008689743, |
|
"learning_rate": 5.023944974428739e-06, |
|
"loss": 1.1734, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.5509830818472794, |
|
"grad_norm": 0.5661449507234365, |
|
"learning_rate": 4.9840366498147495e-06, |
|
"loss": 1.1908, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.5532693187014175, |
|
"grad_norm": 0.6109491726102372, |
|
"learning_rate": 4.944129342178156e-06, |
|
"loss": 1.1784, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.5811074689104263, |
|
"learning_rate": 4.90422559389443e-06, |
|
"loss": 1.1746, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.5578417924096937, |
|
"grad_norm": 0.6060458081756667, |
|
"learning_rate": 4.864327947112281e-06, |
|
"loss": 1.195, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.5601280292638318, |
|
"grad_norm": 0.6226718536570417, |
|
"learning_rate": 4.82443894359171e-06, |
|
"loss": 1.1786, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.5624142661179699, |
|
"grad_norm": 0.5995864510713481, |
|
"learning_rate": 4.784561124542088e-06, |
|
"loss": 1.1791, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.5647005029721079, |
|
"grad_norm": 0.5701958838449743, |
|
"learning_rate": 4.744697030460248e-06, |
|
"loss": 1.1647, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.566986739826246, |
|
"grad_norm": 0.6293939505655973, |
|
"learning_rate": 4.7048492009686525e-06, |
|
"loss": 1.1692, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.5692729766803841, |
|
"grad_norm": 0.6850447194966206, |
|
"learning_rate": 4.6650201746535926e-06, |
|
"loss": 1.1673, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.5715592135345222, |
|
"grad_norm": 0.6040120516739561, |
|
"learning_rate": 4.625212488903467e-06, |
|
"loss": 1.1834, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.5738454503886603, |
|
"grad_norm": 0.5686706476550618, |
|
"learning_rate": 4.585428679747133e-06, |
|
"loss": 1.1716, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.5761316872427984, |
|
"grad_norm": 0.5946931657837966, |
|
"learning_rate": 4.545671281692331e-06, |
|
"loss": 1.1705, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.5784179240969365, |
|
"grad_norm": 0.6120143356512502, |
|
"learning_rate": 4.505942827564242e-06, |
|
"loss": 1.1807, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.5807041609510746, |
|
"grad_norm": 0.6341171747185648, |
|
"learning_rate": 4.466245848344106e-06, |
|
"loss": 1.1839, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.5829903978052127, |
|
"grad_norm": 0.6494090868678567, |
|
"learning_rate": 4.426582873007999e-06, |
|
"loss": 1.1684, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.5852766346593508, |
|
"grad_norm": 0.6252524175950205, |
|
"learning_rate": 4.386956428365701e-06, |
|
"loss": 1.1878, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.5875628715134888, |
|
"grad_norm": 0.5911175497758677, |
|
"learning_rate": 4.347369038899744e-06, |
|
"loss": 1.1828, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.5898491083676269, |
|
"grad_norm": 0.5988939599453593, |
|
"learning_rate": 4.307823226604555e-06, |
|
"loss": 1.1735, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.5921353452217649, |
|
"grad_norm": 0.5813355536422021, |
|
"learning_rate": 4.2683215108258145e-06, |
|
"loss": 1.1706, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.594421582075903, |
|
"grad_norm": 0.6208043705991068, |
|
"learning_rate": 4.228866408099945e-06, |
|
"loss": 1.1907, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.5967078189300411, |
|
"grad_norm": 0.6512006631857741, |
|
"learning_rate": 4.189460431993788e-06, |
|
"loss": 1.1951, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.5989940557841792, |
|
"grad_norm": 0.5845471180993255, |
|
"learning_rate": 4.150106092944475e-06, |
|
"loss": 1.1717, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.6012802926383173, |
|
"grad_norm": 0.5949045334275538, |
|
"learning_rate": 4.110805898099492e-06, |
|
"loss": 1.1833, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.6035665294924554, |
|
"grad_norm": 0.5971913414181261, |
|
"learning_rate": 4.071562351156966e-06, |
|
"loss": 1.1786, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.6058527663465935, |
|
"grad_norm": 0.6178601149254982, |
|
"learning_rate": 4.032377952206148e-06, |
|
"loss": 1.1793, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.6081390032007316, |
|
"grad_norm": 0.6046188006147395, |
|
"learning_rate": 3.993255197568154e-06, |
|
"loss": 1.169, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.6104252400548696, |
|
"grad_norm": 0.5919458656130715, |
|
"learning_rate": 3.954196579636918e-06, |
|
"loss": 1.1692, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.6127114769090077, |
|
"grad_norm": 0.5727049539306068, |
|
"learning_rate": 3.91520458672042e-06, |
|
"loss": 1.1747, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.6149977137631458, |
|
"grad_norm": 0.6040809405921704, |
|
"learning_rate": 3.876281702882156e-06, |
|
"loss": 1.1935, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.6172839506172839, |
|
"grad_norm": 0.5747789602798682, |
|
"learning_rate": 3.837430407782896e-06, |
|
"loss": 1.175, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.619570187471422, |
|
"grad_norm": 0.6001909994942644, |
|
"learning_rate": 3.7986531765226965e-06, |
|
"loss": 1.1718, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.6218564243255601, |
|
"grad_norm": 0.5499338552551708, |
|
"learning_rate": 3.759952479483232e-06, |
|
"loss": 1.1615, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.6241426611796982, |
|
"grad_norm": 0.62697610396954, |
|
"learning_rate": 3.7213307821704115e-06, |
|
"loss": 1.1616, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.6264288980338363, |
|
"grad_norm": 0.637904015143814, |
|
"learning_rate": 3.6827905450573022e-06, |
|
"loss": 1.1784, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.6287151348879744, |
|
"grad_norm": 0.6235229612947039, |
|
"learning_rate": 3.6443342234273905e-06, |
|
"loss": 1.1674, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.6310013717421125, |
|
"grad_norm": 0.744429415227132, |
|
"learning_rate": 3.6059642672181537e-06, |
|
"loss": 1.1678, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.6332876085962506, |
|
"grad_norm": 0.5903117671660288, |
|
"learning_rate": 3.5676831208649887e-06, |
|
"loss": 1.1661, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.6355738454503886, |
|
"grad_norm": 0.5977435348831742, |
|
"learning_rate": 3.5294932231454838e-06, |
|
"loss": 1.1655, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.6378600823045267, |
|
"grad_norm": 0.6262251229258455, |
|
"learning_rate": 3.4913970070240388e-06, |
|
"loss": 1.1827, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.6401463191586648, |
|
"grad_norm": 0.6039362156672261, |
|
"learning_rate": 3.4533968994968913e-06, |
|
"loss": 1.162, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.6424325560128029, |
|
"grad_norm": 0.610471777862986, |
|
"learning_rate": 3.41549532143748e-06, |
|
"loss": 1.1719, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.644718792866941, |
|
"grad_norm": 0.6124948412563855, |
|
"learning_rate": 3.3776946874422268e-06, |
|
"loss": 1.161, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.6470050297210791, |
|
"grad_norm": 0.596054515528405, |
|
"learning_rate": 3.3399974056767095e-06, |
|
"loss": 1.1677, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.6492912665752172, |
|
"grad_norm": 0.6199519548446956, |
|
"learning_rate": 3.30240587772224e-06, |
|
"loss": 1.1731, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.6515775034293553, |
|
"grad_norm": 0.6123382818220521, |
|
"learning_rate": 3.2649224984228756e-06, |
|
"loss": 1.1751, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.6538637402834934, |
|
"grad_norm": 0.6521756883889377, |
|
"learning_rate": 3.227549655732843e-06, |
|
"loss": 1.1746, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.6561499771376315, |
|
"grad_norm": 0.6292502440238857, |
|
"learning_rate": 3.19028973056441e-06, |
|
"loss": 1.1796, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.6584362139917695, |
|
"grad_norm": 0.7223300006546375, |
|
"learning_rate": 3.153145096636211e-06, |
|
"loss": 1.1769, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.6607224508459076, |
|
"grad_norm": 0.6123252900962536, |
|
"learning_rate": 3.1161181203220146e-06, |
|
"loss": 1.1798, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.6630086877000457, |
|
"grad_norm": 0.6176590524451245, |
|
"learning_rate": 3.079211160499975e-06, |
|
"loss": 1.1628, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.6652949245541838, |
|
"grad_norm": 0.6851380779593121, |
|
"learning_rate": 3.0424265684023556e-06, |
|
"loss": 1.1621, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.6675811614083219, |
|
"grad_norm": 0.6135186798564677, |
|
"learning_rate": 3.0057666874657365e-06, |
|
"loss": 1.1817, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.66986739826246, |
|
"grad_norm": 0.6162664151552476, |
|
"learning_rate": 2.9692338531817205e-06, |
|
"loss": 1.1621, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.6721536351165981, |
|
"grad_norm": 0.6209879083469707, |
|
"learning_rate": 2.9328303929481507e-06, |
|
"loss": 1.1788, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.6744398719707362, |
|
"grad_norm": 0.6564960801220917, |
|
"learning_rate": 2.8965586259208295e-06, |
|
"loss": 1.1497, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.6767261088248743, |
|
"grad_norm": 0.6100366044161921, |
|
"learning_rate": 2.860420862865787e-06, |
|
"loss": 1.1641, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.6790123456790124, |
|
"grad_norm": 0.6401282278697755, |
|
"learning_rate": 2.82441940601205e-06, |
|
"loss": 1.1647, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.6812985825331505, |
|
"grad_norm": 0.5948814066139619, |
|
"learning_rate": 2.7885565489049948e-06, |
|
"loss": 1.1862, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.6835848193872885, |
|
"grad_norm": 0.575891260626997, |
|
"learning_rate": 2.7528345762602125e-06, |
|
"loss": 1.149, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.6858710562414266, |
|
"grad_norm": 0.6321328549868929, |
|
"learning_rate": 2.7172557638179674e-06, |
|
"loss": 1.1722, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.6881572930955647, |
|
"grad_norm": 0.620537429422375, |
|
"learning_rate": 2.681822378198221e-06, |
|
"loss": 1.1667, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.6904435299497028, |
|
"grad_norm": 0.5916688359774108, |
|
"learning_rate": 2.6465366767562162e-06, |
|
"loss": 1.1742, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.6927297668038409, |
|
"grad_norm": 0.649532932905328, |
|
"learning_rate": 2.611400907438685e-06, |
|
"loss": 1.1664, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.695016003657979, |
|
"grad_norm": 0.5887639490410209, |
|
"learning_rate": 2.5764173086406306e-06, |
|
"loss": 1.1684, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.6973022405121171, |
|
"grad_norm": 0.5909674256777088, |
|
"learning_rate": 2.5415881090627227e-06, |
|
"loss": 1.1681, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.6995884773662552, |
|
"grad_norm": 0.6669572713903603, |
|
"learning_rate": 2.506915527569318e-06, |
|
"loss": 1.1692, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.7018747142203933, |
|
"grad_norm": 0.6291006193664693, |
|
"learning_rate": 2.472401773047107e-06, |
|
"loss": 1.1707, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.7041609510745314, |
|
"grad_norm": 0.6241336853751712, |
|
"learning_rate": 2.438049044264382e-06, |
|
"loss": 1.1763, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.7064471879286695, |
|
"grad_norm": 0.6233093811845397, |
|
"learning_rate": 2.4038595297309712e-06, |
|
"loss": 1.1595, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.7087334247828075, |
|
"grad_norm": 0.6099376654855213, |
|
"learning_rate": 2.3698354075588105e-06, |
|
"loss": 1.1815, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.7110196616369456, |
|
"grad_norm": 0.608739940642273, |
|
"learning_rate": 2.3359788453231723e-06, |
|
"loss": 1.1558, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.7133058984910837, |
|
"grad_norm": 0.6060804682823651, |
|
"learning_rate": 2.3022919999245964e-06, |
|
"loss": 1.1737, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.7155921353452218, |
|
"grad_norm": 0.6554029837627439, |
|
"learning_rate": 2.2687770174514674e-06, |
|
"loss": 1.1763, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.7178783721993599, |
|
"grad_norm": 0.6199763037940721, |
|
"learning_rate": 2.23543603304329e-06, |
|
"loss": 1.1668, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.720164609053498, |
|
"grad_norm": 0.7002533112076955, |
|
"learning_rate": 2.20227117075468e-06, |
|
"loss": 1.1717, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.7224508459076361, |
|
"grad_norm": 0.5685258465602809, |
|
"learning_rate": 2.1692845434200323e-06, |
|
"loss": 1.1793, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.7247370827617741, |
|
"grad_norm": 0.5988803647429354, |
|
"learning_rate": 2.136478252518924e-06, |
|
"loss": 1.1762, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.7270233196159122, |
|
"grad_norm": 0.6220944262982843, |
|
"learning_rate": 2.103854388042243e-06, |
|
"loss": 1.1732, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.7293095564700502, |
|
"grad_norm": 0.5872374752551915, |
|
"learning_rate": 2.071415028359026e-06, |
|
"loss": 1.1653, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.7315957933241883, |
|
"grad_norm": 0.6315378201627972, |
|
"learning_rate": 2.0391622400840665e-06, |
|
"loss": 1.1631, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.7338820301783264, |
|
"grad_norm": 0.6166479295990325, |
|
"learning_rate": 2.0070980779462513e-06, |
|
"loss": 1.1632, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.7361682670324645, |
|
"grad_norm": 0.6082820756952414, |
|
"learning_rate": 1.975224584657648e-06, |
|
"loss": 1.1609, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.7384545038866026, |
|
"grad_norm": 0.5711567863660318, |
|
"learning_rate": 1.943543790783392e-06, |
|
"loss": 1.1629, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.7407407407407407, |
|
"grad_norm": 0.5934876997772376, |
|
"learning_rate": 1.9120577146123125e-06, |
|
"loss": 1.1711, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.7430269775948788, |
|
"grad_norm": 0.6044258229955937, |
|
"learning_rate": 1.8807683620283496e-06, |
|
"loss": 1.1792, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.7453132144490169, |
|
"grad_norm": 0.6414108282805848, |
|
"learning_rate": 1.8496777263827775e-06, |
|
"loss": 1.1909, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.747599451303155, |
|
"grad_norm": 0.5928077840962543, |
|
"learning_rate": 1.8187877883672024e-06, |
|
"loss": 1.177, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.7498856881572931, |
|
"grad_norm": 0.5674967348667851, |
|
"learning_rate": 1.7881005158873826e-06, |
|
"loss": 1.1698, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.7521719250114312, |
|
"grad_norm": 0.6190325214784786, |
|
"learning_rate": 1.757617863937865e-06, |
|
"loss": 1.1564, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.7544581618655692, |
|
"grad_norm": 0.5994621485851359, |
|
"learning_rate": 1.7273417744774323e-06, |
|
"loss": 1.1682, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.7567443987197073, |
|
"grad_norm": 0.6486512119864596, |
|
"learning_rate": 1.6972741763053835e-06, |
|
"loss": 1.1695, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.7590306355738454, |
|
"grad_norm": 0.6124244446703457, |
|
"learning_rate": 1.6674169849386606e-06, |
|
"loss": 1.1735, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.7613168724279835, |
|
"grad_norm": 0.6215393083401685, |
|
"learning_rate": 1.6377721024898214e-06, |
|
"loss": 1.1611, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.7636031092821216, |
|
"grad_norm": 0.6379465283211975, |
|
"learning_rate": 1.608341417545849e-06, |
|
"loss": 1.1481, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.7658893461362597, |
|
"grad_norm": 0.5646658898706897, |
|
"learning_rate": 1.5791268050478487e-06, |
|
"loss": 1.1732, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.7681755829903978, |
|
"grad_norm": 0.6028441016085894, |
|
"learning_rate": 1.5501301261715896e-06, |
|
"loss": 1.1703, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.7704618198445359, |
|
"grad_norm": 0.6313316478647917, |
|
"learning_rate": 1.5213532282089466e-06, |
|
"loss": 1.1631, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.772748056698674, |
|
"grad_norm": 0.600237347487572, |
|
"learning_rate": 1.4927979444502028e-06, |
|
"loss": 1.1642, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.7750342935528121, |
|
"grad_norm": 0.5957448361281138, |
|
"learning_rate": 1.4644660940672628e-06, |
|
"loss": 1.1668, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.7773205304069501, |
|
"grad_norm": 0.5872437663700951, |
|
"learning_rate": 1.4363594819977606e-06, |
|
"loss": 1.1707, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.7796067672610882, |
|
"grad_norm": 0.7075549655922131, |
|
"learning_rate": 1.4084798988300684e-06, |
|
"loss": 1.1723, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.7818930041152263, |
|
"grad_norm": 0.6203199463017092, |
|
"learning_rate": 1.3808291206892232e-06, |
|
"loss": 1.1668, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.7841792409693644, |
|
"grad_norm": 0.5759538308213393, |
|
"learning_rate": 1.3534089091237757e-06, |
|
"loss": 1.1598, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.7864654778235025, |
|
"grad_norm": 0.5942123152988342, |
|
"learning_rate": 1.3262210109935719e-06, |
|
"loss": 1.1699, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.7887517146776406, |
|
"grad_norm": 0.6597153339968819, |
|
"learning_rate": 1.2992671583584587e-06, |
|
"loss": 1.163, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.7910379515317787, |
|
"grad_norm": 0.5994756887911626, |
|
"learning_rate": 1.2725490683679458e-06, |
|
"loss": 1.1797, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.7933241883859168, |
|
"grad_norm": 0.5942174681280669, |
|
"learning_rate": 1.2460684431518055e-06, |
|
"loss": 1.1649, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.7956104252400549, |
|
"grad_norm": 0.5884403788886147, |
|
"learning_rate": 1.2198269697116416e-06, |
|
"loss": 1.1627, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.797896662094193, |
|
"grad_norm": 0.5917506875732326, |
|
"learning_rate": 1.1938263198134087e-06, |
|
"loss": 1.1729, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.8001828989483311, |
|
"grad_norm": 0.5689945244963683, |
|
"learning_rate": 1.168068149880912e-06, |
|
"loss": 1.1639, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.8024691358024691, |
|
"grad_norm": 0.5945700377730089, |
|
"learning_rate": 1.1425541008902852e-06, |
|
"loss": 1.1616, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.8047553726566072, |
|
"grad_norm": 0.5960318855848052, |
|
"learning_rate": 1.1172857982654445e-06, |
|
"loss": 1.1796, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.8070416095107453, |
|
"grad_norm": 0.606906781862042, |
|
"learning_rate": 1.092264851774536e-06, |
|
"loss": 1.1524, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.8093278463648834, |
|
"grad_norm": 0.6686014083887466, |
|
"learning_rate": 1.067492855427385e-06, |
|
"loss": 1.1681, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.8116140832190215, |
|
"grad_norm": 0.6637295349703526, |
|
"learning_rate": 1.0429713873739505e-06, |
|
"loss": 1.1603, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.8139003200731596, |
|
"grad_norm": 0.5937746781646984, |
|
"learning_rate": 1.0187020098037759e-06, |
|
"loss": 1.1577, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.8161865569272977, |
|
"grad_norm": 0.6154438358761861, |
|
"learning_rate": 9.946862688464753e-07, |
|
"loss": 1.1596, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.8184727937814358, |
|
"grad_norm": 0.6511739287376433, |
|
"learning_rate": 9.709256944732343e-07, |
|
"loss": 1.1707, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.8207590306355739, |
|
"grad_norm": 0.6174881374069865, |
|
"learning_rate": 9.474218003993275e-07, |
|
"loss": 1.1775, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.823045267489712, |
|
"grad_norm": 0.5791204684491382, |
|
"learning_rate": 9.241760839877023e-07, |
|
"loss": 1.1571, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.82533150434385, |
|
"grad_norm": 0.6464260391976697, |
|
"learning_rate": 9.011900261535767e-07, |
|
"loss": 1.1713, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.8276177411979881, |
|
"grad_norm": 0.6102288143326278, |
|
"learning_rate": 8.784650912700909e-07, |
|
"loss": 1.1654, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.8299039780521262, |
|
"grad_norm": 0.6226743471510658, |
|
"learning_rate": 8.560027270750276e-07, |
|
"loss": 1.1655, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.8321902149062643, |
|
"grad_norm": 0.6079710775307922, |
|
"learning_rate": 8.338043645785698e-07, |
|
"loss": 1.1669, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.8344764517604024, |
|
"grad_norm": 0.6077180347148399, |
|
"learning_rate": 8.118714179721404e-07, |
|
"loss": 1.1529, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.8367626886145405, |
|
"grad_norm": 0.6420590181680129, |
|
"learning_rate": 7.902052845383112e-07, |
|
"loss": 1.1662, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.8390489254686786, |
|
"grad_norm": 0.5675937752707487, |
|
"learning_rate": 7.6880734456178e-07, |
|
"loss": 1.1638, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.8413351623228167, |
|
"grad_norm": 0.5963600943686237, |
|
"learning_rate": 7.476789612414414e-07, |
|
"loss": 1.1648, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.8436213991769548, |
|
"grad_norm": 0.6248451529177521, |
|
"learning_rate": 7.268214806035423e-07, |
|
"loss": 1.1704, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.8459076360310929, |
|
"grad_norm": 0.6582130785897107, |
|
"learning_rate": 7.062362314159211e-07, |
|
"loss": 1.1716, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.848193872885231, |
|
"grad_norm": 0.6104979563533071, |
|
"learning_rate": 6.859245251033697e-07, |
|
"loss": 1.1551, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.850480109739369, |
|
"grad_norm": 0.6291505363028616, |
|
"learning_rate": 6.658876556640781e-07, |
|
"loss": 1.1606, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.8527663465935071, |
|
"grad_norm": 0.626351910055198, |
|
"learning_rate": 6.461268995871967e-07, |
|
"loss": 1.1648, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.8550525834476452, |
|
"grad_norm": 0.5991977091276379, |
|
"learning_rate": 6.266435157715222e-07, |
|
"loss": 1.1403, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.8573388203017832, |
|
"grad_norm": 0.6133109082285381, |
|
"learning_rate": 6.074387454452891e-07, |
|
"loss": 1.1578, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.8596250571559213, |
|
"grad_norm": 0.6062420232877472, |
|
"learning_rate": 5.885138120870965e-07, |
|
"loss": 1.1422, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.8619112940100594, |
|
"grad_norm": 0.5920619164293491, |
|
"learning_rate": 5.698699213479697e-07, |
|
"loss": 1.1503, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.8641975308641975, |
|
"grad_norm": 0.6179934405963249, |
|
"learning_rate": 5.515082609745465e-07, |
|
"loss": 1.1728, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.8664837677183356, |
|
"grad_norm": 0.6191884681224713, |
|
"learning_rate": 5.334300007334065e-07, |
|
"loss": 1.1514, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.8687700045724737, |
|
"grad_norm": 0.6148818189812965, |
|
"learning_rate": 5.156362923365587e-07, |
|
"loss": 1.1772, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.8710562414266118, |
|
"grad_norm": 0.5927964681781609, |
|
"learning_rate": 4.981282693680584e-07, |
|
"loss": 1.1747, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.8733424782807498, |
|
"grad_norm": 0.630038523819453, |
|
"learning_rate": 4.80907047211796e-07, |
|
"loss": 1.1638, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.8756287151348879, |
|
"grad_norm": 0.5822419290829026, |
|
"learning_rate": 4.639737229804403e-07, |
|
"loss": 1.1667, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.877914951989026, |
|
"grad_norm": 0.6169634205827448, |
|
"learning_rate": 4.473293754455399e-07, |
|
"loss": 1.1695, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.8802011888431641, |
|
"grad_norm": 0.5892947845386679, |
|
"learning_rate": 4.3097506496880325e-07, |
|
"loss": 1.1684, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.8824874256973022, |
|
"grad_norm": 0.6796811793089527, |
|
"learning_rate": 4.149118334345403e-07, |
|
"loss": 1.1604, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.8847736625514403, |
|
"grad_norm": 0.5951100132603444, |
|
"learning_rate": 3.9914070418329123e-07, |
|
"loss": 1.1632, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.8870598994055784, |
|
"grad_norm": 0.6710610553022762, |
|
"learning_rate": 3.836626819466338e-07, |
|
"loss": 1.1455, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.8893461362597165, |
|
"grad_norm": 0.6128779790737046, |
|
"learning_rate": 3.684787527831707e-07, |
|
"loss": 1.1609, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.8916323731138546, |
|
"grad_norm": 0.5800567298586133, |
|
"learning_rate": 3.53589884015712e-07, |
|
"loss": 1.1636, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.8939186099679927, |
|
"grad_norm": 0.5600191099569565, |
|
"learning_rate": 3.3899702416965166e-07, |
|
"loss": 1.1721, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.8962048468221308, |
|
"grad_norm": 0.5964683215562515, |
|
"learning_rate": 3.247011029125391e-07, |
|
"loss": 1.1508, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.8984910836762688, |
|
"grad_norm": 0.6125213377358303, |
|
"learning_rate": 3.1070303099485055e-07, |
|
"loss": 1.1716, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.9007773205304069, |
|
"grad_norm": 0.5812964318078312, |
|
"learning_rate": 2.9700370019197287e-07, |
|
"loss": 1.1495, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.903063557384545, |
|
"grad_norm": 0.5947330421470328, |
|
"learning_rate": 2.8360398324738415e-07, |
|
"loss": 1.1446, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.9053497942386831, |
|
"grad_norm": 0.5936630268160432, |
|
"learning_rate": 2.7050473381706186e-07, |
|
"loss": 1.1519, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.9076360310928212, |
|
"grad_norm": 0.6228979256825669, |
|
"learning_rate": 2.577067864150906e-07, |
|
"loss": 1.1688, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.9099222679469593, |
|
"grad_norm": 0.6500515468078818, |
|
"learning_rate": 2.452109563605065e-07, |
|
"loss": 1.1718, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.9122085048010974, |
|
"grad_norm": 0.568112374463465, |
|
"learning_rate": 2.330180397253473e-07, |
|
"loss": 1.169, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.9144947416552355, |
|
"grad_norm": 0.6014335143268985, |
|
"learning_rate": 2.2112881328394287e-07, |
|
"loss": 1.1556, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.9167809785093736, |
|
"grad_norm": 0.5814781144236604, |
|
"learning_rate": 2.0954403446342753e-07, |
|
"loss": 1.1688, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.9190672153635117, |
|
"grad_norm": 0.6269697024329176, |
|
"learning_rate": 1.9826444129548317e-07, |
|
"loss": 1.1791, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.9213534522176497, |
|
"grad_norm": 0.5793724546294099, |
|
"learning_rate": 1.8729075236932903e-07, |
|
"loss": 1.1736, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.9236396890717878, |
|
"grad_norm": 0.5757028817840649, |
|
"learning_rate": 1.7662366678593502e-07, |
|
"loss": 1.1674, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.9259259259259259, |
|
"grad_norm": 0.6383512892284545, |
|
"learning_rate": 1.6626386411348783e-07, |
|
"loss": 1.1725, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.928212162780064, |
|
"grad_norm": 0.6064267969457637, |
|
"learning_rate": 1.56212004344099e-07, |
|
"loss": 1.1596, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.9304983996342021, |
|
"grad_norm": 0.6046327277263103, |
|
"learning_rate": 1.4646872785175182e-07, |
|
"loss": 1.1616, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.9327846364883402, |
|
"grad_norm": 0.611959733363112, |
|
"learning_rate": 1.3703465535151505e-07, |
|
"loss": 1.1614, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.9350708733424783, |
|
"grad_norm": 0.6153837948383357, |
|
"learning_rate": 1.2791038785999243e-07, |
|
"loss": 1.1494, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.9373571101966164, |
|
"grad_norm": 0.5507733416769363, |
|
"learning_rate": 1.1909650665703265e-07, |
|
"loss": 1.1331, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.9396433470507545, |
|
"grad_norm": 0.5787602661155832, |
|
"learning_rate": 1.1059357324870456e-07, |
|
"loss": 1.1548, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.9419295839048926, |
|
"grad_norm": 0.5848374134615248, |
|
"learning_rate": 1.024021293315175e-07, |
|
"loss": 1.1628, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.9442158207590307, |
|
"grad_norm": 0.585861722501522, |
|
"learning_rate": 9.452269675791603e-08, |
|
"loss": 1.1424, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.9465020576131687, |
|
"grad_norm": 0.5870866242087308, |
|
"learning_rate": 8.69557775030344e-08, |
|
"loss": 1.181, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.9487882944673068, |
|
"grad_norm": 0.5917858310575264, |
|
"learning_rate": 7.970185363271432e-08, |
|
"loss": 1.1564, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.9510745313214449, |
|
"grad_norm": 0.6272259568011471, |
|
"learning_rate": 7.276138727279669e-08, |
|
"loss": 1.1659, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.953360768175583, |
|
"grad_norm": 0.607366888512829, |
|
"learning_rate": 6.613482057968023e-08, |
|
"loss": 1.1612, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.9556470050297211, |
|
"grad_norm": 0.61579614820576, |
|
"learning_rate": 5.982257571215178e-08, |
|
"loss": 1.1644, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.9579332418838592, |
|
"grad_norm": 0.6162342496797737, |
|
"learning_rate": 5.382505480449274e-08, |
|
"loss": 1.1439, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.9602194787379973, |
|
"grad_norm": 0.5880335959078453, |
|
"learning_rate": 4.814263994086077e-08, |
|
"loss": 1.1405, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.9625057155921354, |
|
"grad_norm": 0.5978901392727579, |
|
"learning_rate": 4.2775693130948094e-08, |
|
"loss": 1.1792, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.9647919524462735, |
|
"grad_norm": 0.5725207858399001, |
|
"learning_rate": 3.772455628691829e-08, |
|
"loss": 1.1679, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.9670781893004116, |
|
"grad_norm": 0.6126681514493614, |
|
"learning_rate": 3.2989551201624836e-08, |
|
"loss": 1.1621, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.9693644261545497, |
|
"grad_norm": 0.6026354249744876, |
|
"learning_rate": 2.857097952810972e-08, |
|
"loss": 1.1728, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.9716506630086877, |
|
"grad_norm": 0.5876159431495082, |
|
"learning_rate": 2.4469122760388264e-08, |
|
"loss": 1.1552, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.9739368998628258, |
|
"grad_norm": 0.5795939734314318, |
|
"learning_rate": 2.0684242215511797e-08, |
|
"loss": 1.1586, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.9762231367169639, |
|
"grad_norm": 0.6100064497073957, |
|
"learning_rate": 1.7216579016925415e-08, |
|
"loss": 1.1585, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 0.978509373571102, |
|
"grad_norm": 0.6410024148442394, |
|
"learning_rate": 1.4066354079101396e-08, |
|
"loss": 1.1576, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.9807956104252401, |
|
"grad_norm": 0.5946394925998356, |
|
"learning_rate": 1.1233768093468766e-08, |
|
"loss": 1.1565, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 0.9830818472793782, |
|
"grad_norm": 0.5993080705042445, |
|
"learning_rate": 8.719001515627434e-09, |
|
"loss": 1.1649, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.9853680841335163, |
|
"grad_norm": 0.5857680491868433, |
|
"learning_rate": 6.5222145538501595e-09, |
|
"loss": 1.176, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 0.9876543209876543, |
|
"grad_norm": 0.6157142971328977, |
|
"learning_rate": 4.643547158878492e-09, |
|
"loss": 1.146, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.9899405578417924, |
|
"grad_norm": 0.6005659801135901, |
|
"learning_rate": 3.0831190150054646e-09, |
|
"loss": 1.1607, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 0.9922267946959304, |
|
"grad_norm": 0.5963682235084494, |
|
"learning_rate": 1.8410295324505778e-09, |
|
"loss": 1.1668, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.9945130315500685, |
|
"grad_norm": 0.649218390898171, |
|
"learning_rate": 9.173578410281992e-10, |
|
"loss": 1.1602, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.9967992684042066, |
|
"grad_norm": 0.612662110275474, |
|
"learning_rate": 3.1216278510493027e-10, |
|
"loss": 1.1596, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.9990855052583447, |
|
"grad_norm": 0.6025732837303296, |
|
"learning_rate": 2.548291985149387e-11, |
|
"loss": 1.147, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_runtime": 4.0833, |
|
"eval_samples_per_second": 2.449, |
|
"eval_steps_per_second": 0.735, |
|
"step": 2187 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 2187, |
|
"total_flos": 9703359095242752.0, |
|
"train_loss": 0.0, |
|
"train_runtime": 0.009, |
|
"train_samples_per_second": 3870652.356, |
|
"train_steps_per_second": 241991.844 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 2187, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 9703359095242752.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|