|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.5159621490167481, |
|
"eval_steps": 500, |
|
"global_step": 1500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.003439747660111654, |
|
"grad_norm": 4.688905239105225, |
|
"learning_rate": 1.7199862401100792e-07, |
|
"loss": 13.1365, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.006879495320223308, |
|
"grad_norm": 4.80711030960083, |
|
"learning_rate": 3.4399724802201585e-07, |
|
"loss": 13.154, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.010319242980334962, |
|
"grad_norm": 4.933587074279785, |
|
"learning_rate": 5.159958720330238e-07, |
|
"loss": 13.1537, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.013758990640446616, |
|
"grad_norm": 4.939642906188965, |
|
"learning_rate": 6.879944960440317e-07, |
|
"loss": 13.1529, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01719873830055827, |
|
"grad_norm": 5.070457935333252, |
|
"learning_rate": 8.599931200550396e-07, |
|
"loss": 13.1224, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.020638485960669924, |
|
"grad_norm": 5.011481285095215, |
|
"learning_rate": 1.0319917440660475e-06, |
|
"loss": 13.0346, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02407823362078158, |
|
"grad_norm": 5.395399570465088, |
|
"learning_rate": 1.2039903680770555e-06, |
|
"loss": 12.9762, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.027517981280893233, |
|
"grad_norm": 5.811434745788574, |
|
"learning_rate": 1.3759889920880634e-06, |
|
"loss": 12.853, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03095772894100489, |
|
"grad_norm": 6.088072776794434, |
|
"learning_rate": 1.5479876160990713e-06, |
|
"loss": 12.6852, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03439747660111654, |
|
"grad_norm": 6.066171646118164, |
|
"learning_rate": 1.7199862401100792e-06, |
|
"loss": 12.4486, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.037837224261228194, |
|
"grad_norm": 6.659308910369873, |
|
"learning_rate": 1.891984864121087e-06, |
|
"loss": 12.1745, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04127697192133985, |
|
"grad_norm": 7.230446815490723, |
|
"learning_rate": 2.063983488132095e-06, |
|
"loss": 11.8162, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.04471671958145151, |
|
"grad_norm": 7.676384449005127, |
|
"learning_rate": 2.2359821121431026e-06, |
|
"loss": 11.3668, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04815646724156316, |
|
"grad_norm": 8.068939208984375, |
|
"learning_rate": 2.407980736154111e-06, |
|
"loss": 10.9014, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.05159621490167481, |
|
"grad_norm": 8.794692039489746, |
|
"learning_rate": 2.5799793601651184e-06, |
|
"loss": 10.327, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.055035962561786465, |
|
"grad_norm": 9.102154731750488, |
|
"learning_rate": 2.7519779841761268e-06, |
|
"loss": 9.6304, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05847571022189812, |
|
"grad_norm": 8.208027839660645, |
|
"learning_rate": 2.9239766081871343e-06, |
|
"loss": 8.8888, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.06191545788200978, |
|
"grad_norm": 6.106681823730469, |
|
"learning_rate": 3.0959752321981426e-06, |
|
"loss": 8.1024, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06535520554212143, |
|
"grad_norm": 3.8156211376190186, |
|
"learning_rate": 3.2679738562091506e-06, |
|
"loss": 7.5017, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.06879495320223308, |
|
"grad_norm": 3.4960334300994873, |
|
"learning_rate": 3.4399724802201585e-06, |
|
"loss": 7.0312, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07223470086234474, |
|
"grad_norm": 3.1875991821289062, |
|
"learning_rate": 3.611971104231166e-06, |
|
"loss": 6.7019, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.07567444852245639, |
|
"grad_norm": 3.0836124420166016, |
|
"learning_rate": 3.783969728242174e-06, |
|
"loss": 6.3059, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.07911419618256804, |
|
"grad_norm": 2.8337793350219727, |
|
"learning_rate": 3.955968352253183e-06, |
|
"loss": 5.9064, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.0825539438426797, |
|
"grad_norm": 2.593134641647339, |
|
"learning_rate": 4.12796697626419e-06, |
|
"loss": 5.5292, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.08599369150279136, |
|
"grad_norm": 2.23457670211792, |
|
"learning_rate": 4.299965600275198e-06, |
|
"loss": 5.1508, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.08943343916290301, |
|
"grad_norm": 2.220684051513672, |
|
"learning_rate": 4.471964224286205e-06, |
|
"loss": 4.9064, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.09287318682301467, |
|
"grad_norm": 2.066509962081909, |
|
"learning_rate": 4.643962848297214e-06, |
|
"loss": 4.5587, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.09631293448312632, |
|
"grad_norm": 1.9317891597747803, |
|
"learning_rate": 4.815961472308222e-06, |
|
"loss": 4.2968, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.09975268214323797, |
|
"grad_norm": 1.852779746055603, |
|
"learning_rate": 4.987960096319229e-06, |
|
"loss": 3.9759, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.10319242980334963, |
|
"grad_norm": 1.7298812866210938, |
|
"learning_rate": 5.159958720330237e-06, |
|
"loss": 3.6811, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.10663217746346128, |
|
"grad_norm": 1.5647941827774048, |
|
"learning_rate": 5.331957344341246e-06, |
|
"loss": 3.4349, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.11007192512357293, |
|
"grad_norm": 1.423771619796753, |
|
"learning_rate": 5.5039559683522536e-06, |
|
"loss": 3.1899, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.11351167278368458, |
|
"grad_norm": 1.2160062789916992, |
|
"learning_rate": 5.675954592363261e-06, |
|
"loss": 3.0158, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.11695142044379624, |
|
"grad_norm": 1.162079095840454, |
|
"learning_rate": 5.8479532163742686e-06, |
|
"loss": 2.8798, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.1203911681039079, |
|
"grad_norm": 0.9763687252998352, |
|
"learning_rate": 6.019951840385278e-06, |
|
"loss": 2.7573, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.12383091576401956, |
|
"grad_norm": 0.9514108300209045, |
|
"learning_rate": 6.191950464396285e-06, |
|
"loss": 2.6355, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.1272706634241312, |
|
"grad_norm": 0.8850705623626709, |
|
"learning_rate": 6.363949088407294e-06, |
|
"loss": 2.5371, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.13071041108424286, |
|
"grad_norm": 0.8824434280395508, |
|
"learning_rate": 6.535947712418301e-06, |
|
"loss": 2.4404, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.13415015874435451, |
|
"grad_norm": 0.9136425256729126, |
|
"learning_rate": 6.7079463364293095e-06, |
|
"loss": 2.3809, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.13758990640446617, |
|
"grad_norm": 1.0447123050689697, |
|
"learning_rate": 6.879944960440317e-06, |
|
"loss": 2.3013, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.14102965406457782, |
|
"grad_norm": 1.0016720294952393, |
|
"learning_rate": 7.0519435844513245e-06, |
|
"loss": 2.2526, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.14446940172468947, |
|
"grad_norm": 0.9885135293006897, |
|
"learning_rate": 7.223942208462332e-06, |
|
"loss": 2.2099, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.14790914938480113, |
|
"grad_norm": 0.9515292048454285, |
|
"learning_rate": 7.39594083247334e-06, |
|
"loss": 2.1602, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.15134889704491278, |
|
"grad_norm": 0.8287838101387024, |
|
"learning_rate": 7.567939456484348e-06, |
|
"loss": 2.1251, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.15478864470502443, |
|
"grad_norm": 1.160489559173584, |
|
"learning_rate": 7.739938080495356e-06, |
|
"loss": 2.0836, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.15822839236513608, |
|
"grad_norm": 1.1019588708877563, |
|
"learning_rate": 7.911936704506365e-06, |
|
"loss": 2.0474, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.16166814002524774, |
|
"grad_norm": 0.9389374256134033, |
|
"learning_rate": 8.083935328517373e-06, |
|
"loss": 2.013, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.1651078876853594, |
|
"grad_norm": 0.9914370775222778, |
|
"learning_rate": 8.25593395252838e-06, |
|
"loss": 1.9829, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.16854763534547107, |
|
"grad_norm": 0.8827924132347107, |
|
"learning_rate": 8.427932576539388e-06, |
|
"loss": 1.9574, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.17198738300558272, |
|
"grad_norm": 1.0819416046142578, |
|
"learning_rate": 8.599931200550395e-06, |
|
"loss": 1.9333, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.17542713066569438, |
|
"grad_norm": 1.006886601448059, |
|
"learning_rate": 8.771929824561403e-06, |
|
"loss": 1.8845, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.17886687832580603, |
|
"grad_norm": 0.8748846054077148, |
|
"learning_rate": 8.94392844857241e-06, |
|
"loss": 1.8757, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.18230662598591768, |
|
"grad_norm": 0.9551517963409424, |
|
"learning_rate": 9.11592707258342e-06, |
|
"loss": 1.8478, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.18574637364602933, |
|
"grad_norm": 1.4100078344345093, |
|
"learning_rate": 9.287925696594429e-06, |
|
"loss": 1.8186, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.189186121306141, |
|
"grad_norm": 1.5030393600463867, |
|
"learning_rate": 9.459924320605436e-06, |
|
"loss": 1.7939, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.19262586896625264, |
|
"grad_norm": 1.0008896589279175, |
|
"learning_rate": 9.631922944616444e-06, |
|
"loss": 1.7711, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.1960656166263643, |
|
"grad_norm": 1.5640122890472412, |
|
"learning_rate": 9.803921568627451e-06, |
|
"loss": 1.7555, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.19950536428647594, |
|
"grad_norm": 1.041014313697815, |
|
"learning_rate": 9.975920192638459e-06, |
|
"loss": 1.7305, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.2029451119465876, |
|
"grad_norm": 1.2513008117675781, |
|
"learning_rate": 1.0147918816649466e-05, |
|
"loss": 1.7083, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.20638485960669925, |
|
"grad_norm": 1.2905975580215454, |
|
"learning_rate": 1.0319917440660474e-05, |
|
"loss": 1.7061, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2098246072668109, |
|
"grad_norm": 1.0839684009552002, |
|
"learning_rate": 1.0491916064671483e-05, |
|
"loss": 1.6722, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.21326435492692256, |
|
"grad_norm": 1.2286721467971802, |
|
"learning_rate": 1.0663914688682492e-05, |
|
"loss": 1.6444, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.2167041025870342, |
|
"grad_norm": 1.3250325918197632, |
|
"learning_rate": 1.08359133126935e-05, |
|
"loss": 1.6229, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.22014385024714586, |
|
"grad_norm": 1.176580548286438, |
|
"learning_rate": 1.1007911936704507e-05, |
|
"loss": 1.6143, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.22358359790725751, |
|
"grad_norm": 1.4954670667648315, |
|
"learning_rate": 1.1179910560715515e-05, |
|
"loss": 1.597, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.22702334556736917, |
|
"grad_norm": 1.0406630039215088, |
|
"learning_rate": 1.1351909184726522e-05, |
|
"loss": 1.5877, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.23046309322748082, |
|
"grad_norm": 1.5156970024108887, |
|
"learning_rate": 1.152390780873753e-05, |
|
"loss": 1.5601, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.23390284088759247, |
|
"grad_norm": 1.3882994651794434, |
|
"learning_rate": 1.1695906432748537e-05, |
|
"loss": 1.5456, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.23734258854770415, |
|
"grad_norm": 1.5137856006622314, |
|
"learning_rate": 1.1867905056759546e-05, |
|
"loss": 1.5101, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.2407823362078158, |
|
"grad_norm": 1.2775577306747437, |
|
"learning_rate": 1.2039903680770556e-05, |
|
"loss": 1.5042, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.24422208386792746, |
|
"grad_norm": 1.3952692747116089, |
|
"learning_rate": 1.2211902304781563e-05, |
|
"loss": 1.4864, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.2476618315280391, |
|
"grad_norm": 1.5909626483917236, |
|
"learning_rate": 1.238390092879257e-05, |
|
"loss": 1.4703, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.25110157918815074, |
|
"grad_norm": 1.4731673002243042, |
|
"learning_rate": 1.255589955280358e-05, |
|
"loss": 1.4434, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2545413268482624, |
|
"grad_norm": 1.370367169380188, |
|
"learning_rate": 1.2727898176814587e-05, |
|
"loss": 1.4522, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.25798107450837404, |
|
"grad_norm": 1.7441356182098389, |
|
"learning_rate": 1.2899896800825595e-05, |
|
"loss": 1.4124, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.2614208221684857, |
|
"grad_norm": 1.6607742309570312, |
|
"learning_rate": 1.3071895424836602e-05, |
|
"loss": 1.409, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.26486056982859735, |
|
"grad_norm": 1.423419713973999, |
|
"learning_rate": 1.324389404884761e-05, |
|
"loss": 1.3912, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.26830031748870903, |
|
"grad_norm": 1.6139755249023438, |
|
"learning_rate": 1.3415892672858619e-05, |
|
"loss": 1.3753, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.2717400651488207, |
|
"grad_norm": 1.9674487113952637, |
|
"learning_rate": 1.3587891296869626e-05, |
|
"loss": 1.3718, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.27517981280893233, |
|
"grad_norm": 2.2793004512786865, |
|
"learning_rate": 1.3759889920880634e-05, |
|
"loss": 1.3426, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.278619560469044, |
|
"grad_norm": 1.6088732481002808, |
|
"learning_rate": 1.3931888544891641e-05, |
|
"loss": 1.3336, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.28205930812915564, |
|
"grad_norm": 2.0331807136535645, |
|
"learning_rate": 1.4103887168902649e-05, |
|
"loss": 1.3224, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.2854990557892673, |
|
"grad_norm": 1.7430800199508667, |
|
"learning_rate": 1.4275885792913656e-05, |
|
"loss": 1.3212, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.28893880344937894, |
|
"grad_norm": 1.4452030658721924, |
|
"learning_rate": 1.4447884416924664e-05, |
|
"loss": 1.3141, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.2923785511094906, |
|
"grad_norm": 1.6565144062042236, |
|
"learning_rate": 1.4619883040935673e-05, |
|
"loss": 1.2885, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.29581829876960225, |
|
"grad_norm": 1.3976999521255493, |
|
"learning_rate": 1.479188166494668e-05, |
|
"loss": 1.2906, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.29925804642971393, |
|
"grad_norm": 2.229381561279297, |
|
"learning_rate": 1.4963880288957688e-05, |
|
"loss": 1.267, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.30269779408982556, |
|
"grad_norm": 1.6807219982147217, |
|
"learning_rate": 1.5135878912968696e-05, |
|
"loss": 1.2696, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.30613754174993724, |
|
"grad_norm": 1.2693103551864624, |
|
"learning_rate": 1.5307877536979705e-05, |
|
"loss": 1.2572, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.30957728941004886, |
|
"grad_norm": 1.9241502285003662, |
|
"learning_rate": 1.5479876160990712e-05, |
|
"loss": 1.2568, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.31301703707016054, |
|
"grad_norm": 1.660744547843933, |
|
"learning_rate": 1.5651874785001723e-05, |
|
"loss": 1.2521, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.31645678473027217, |
|
"grad_norm": 1.706275463104248, |
|
"learning_rate": 1.582387340901273e-05, |
|
"loss": 1.2357, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.31989653239038385, |
|
"grad_norm": 1.8540180921554565, |
|
"learning_rate": 1.5995872033023738e-05, |
|
"loss": 1.2167, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.3233362800504955, |
|
"grad_norm": 1.8119511604309082, |
|
"learning_rate": 1.6167870657034746e-05, |
|
"loss": 1.2544, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.32677602771060715, |
|
"grad_norm": 1.9702472686767578, |
|
"learning_rate": 1.6339869281045753e-05, |
|
"loss": 1.2094, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.3302157753707188, |
|
"grad_norm": 2.0417630672454834, |
|
"learning_rate": 1.651186790505676e-05, |
|
"loss": 1.1975, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.33365552303083046, |
|
"grad_norm": 1.4115933179855347, |
|
"learning_rate": 1.6683866529067768e-05, |
|
"loss": 1.1897, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.33709527069094214, |
|
"grad_norm": 2.0552473068237305, |
|
"learning_rate": 1.6855865153078776e-05, |
|
"loss": 1.2015, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.34053501835105376, |
|
"grad_norm": 2.383172035217285, |
|
"learning_rate": 1.7027863777089783e-05, |
|
"loss": 1.1849, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.34397476601116544, |
|
"grad_norm": 2.141005516052246, |
|
"learning_rate": 1.719986240110079e-05, |
|
"loss": 1.1892, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.34741451367127707, |
|
"grad_norm": 2.242393732070923, |
|
"learning_rate": 1.7371861025111798e-05, |
|
"loss": 1.1812, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.35085426133138875, |
|
"grad_norm": 2.2587263584136963, |
|
"learning_rate": 1.7543859649122806e-05, |
|
"loss": 1.1722, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.3542940089915004, |
|
"grad_norm": 2.2541704177856445, |
|
"learning_rate": 1.7715858273133813e-05, |
|
"loss": 1.1677, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.35773375665161206, |
|
"grad_norm": 1.5090018510818481, |
|
"learning_rate": 1.788785689714482e-05, |
|
"loss": 1.1496, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.3611735043117237, |
|
"grad_norm": 2.1420912742614746, |
|
"learning_rate": 1.805985552115583e-05, |
|
"loss": 1.152, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.36461325197183536, |
|
"grad_norm": 1.6955997943878174, |
|
"learning_rate": 1.823185414516684e-05, |
|
"loss": 1.1514, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.368052999631947, |
|
"grad_norm": 2.603424549102783, |
|
"learning_rate": 1.840385276917785e-05, |
|
"loss": 1.1543, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.37149274729205867, |
|
"grad_norm": 1.6074223518371582, |
|
"learning_rate": 1.8575851393188857e-05, |
|
"loss": 1.1632, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.3749324949521703, |
|
"grad_norm": 2.813610553741455, |
|
"learning_rate": 1.8747850017199865e-05, |
|
"loss": 1.1383, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.378372242612282, |
|
"grad_norm": 2.169351816177368, |
|
"learning_rate": 1.8919848641210872e-05, |
|
"loss": 1.1316, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.3818119902723936, |
|
"grad_norm": 2.653705596923828, |
|
"learning_rate": 1.909184726522188e-05, |
|
"loss": 1.1433, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.3852517379325053, |
|
"grad_norm": 3.17232084274292, |
|
"learning_rate": 1.9263845889232888e-05, |
|
"loss": 1.1385, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.3886914855926169, |
|
"grad_norm": 2.2379255294799805, |
|
"learning_rate": 1.9435844513243895e-05, |
|
"loss": 1.1205, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.3921312332527286, |
|
"grad_norm": 2.5751757621765137, |
|
"learning_rate": 1.9607843137254903e-05, |
|
"loss": 1.1125, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.3955709809128402, |
|
"grad_norm": 1.6813164949417114, |
|
"learning_rate": 1.977984176126591e-05, |
|
"loss": 1.1147, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.3990107285729519, |
|
"grad_norm": 2.8093101978302, |
|
"learning_rate": 1.9951840385276918e-05, |
|
"loss": 1.1057, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.40245047623306357, |
|
"grad_norm": 2.059581995010376, |
|
"learning_rate": 2.0123839009287925e-05, |
|
"loss": 1.1037, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.4058902238931752, |
|
"grad_norm": 1.9385766983032227, |
|
"learning_rate": 2.0295837633298933e-05, |
|
"loss": 1.0987, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.4093299715532869, |
|
"grad_norm": 1.7843406200408936, |
|
"learning_rate": 2.046783625730994e-05, |
|
"loss": 1.0917, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.4127697192133985, |
|
"grad_norm": 1.7654945850372314, |
|
"learning_rate": 2.0639834881320948e-05, |
|
"loss": 1.1, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.4162094668735102, |
|
"grad_norm": 2.3094961643218994, |
|
"learning_rate": 2.081183350533196e-05, |
|
"loss": 1.0891, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.4196492145336218, |
|
"grad_norm": 2.654541492462158, |
|
"learning_rate": 2.0983832129342966e-05, |
|
"loss": 1.0909, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.4230889621937335, |
|
"grad_norm": 2.7142221927642822, |
|
"learning_rate": 2.1155830753353977e-05, |
|
"loss": 1.1068, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.4265287098538451, |
|
"grad_norm": 1.7891792058944702, |
|
"learning_rate": 2.1327829377364984e-05, |
|
"loss": 1.076, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.4299684575139568, |
|
"grad_norm": 1.8982404470443726, |
|
"learning_rate": 2.1499828001375992e-05, |
|
"loss": 1.0703, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.4334082051740684, |
|
"grad_norm": 2.5560953617095947, |
|
"learning_rate": 2.1671826625387e-05, |
|
"loss": 1.0696, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.4368479528341801, |
|
"grad_norm": 2.6159865856170654, |
|
"learning_rate": 2.1843825249398007e-05, |
|
"loss": 1.069, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.4402877004942917, |
|
"grad_norm": 2.1036102771759033, |
|
"learning_rate": 2.2015823873409014e-05, |
|
"loss": 1.054, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.4437274481544034, |
|
"grad_norm": 1.6638193130493164, |
|
"learning_rate": 2.2187822497420022e-05, |
|
"loss": 1.0554, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.44716719581451503, |
|
"grad_norm": 1.7467694282531738, |
|
"learning_rate": 2.235982112143103e-05, |
|
"loss": 1.0356, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.4506069434746267, |
|
"grad_norm": 1.9570752382278442, |
|
"learning_rate": 2.2531819745442037e-05, |
|
"loss": 1.0501, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.45404669113473833, |
|
"grad_norm": 2.062814950942993, |
|
"learning_rate": 2.2703818369453044e-05, |
|
"loss": 1.041, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.45748643879485, |
|
"grad_norm": 3.205482006072998, |
|
"learning_rate": 2.2875816993464052e-05, |
|
"loss": 1.0428, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.46092618645496164, |
|
"grad_norm": 3.798652172088623, |
|
"learning_rate": 2.304781561747506e-05, |
|
"loss": 1.055, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.4643659341150733, |
|
"grad_norm": 2.6361844539642334, |
|
"learning_rate": 2.3219814241486067e-05, |
|
"loss": 1.051, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.46780568177518494, |
|
"grad_norm": 3.9871771335601807, |
|
"learning_rate": 2.3391812865497074e-05, |
|
"loss": 1.0576, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.4712454294352966, |
|
"grad_norm": 2.9001245498657227, |
|
"learning_rate": 2.3563811489508085e-05, |
|
"loss": 1.0521, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.4746851770954083, |
|
"grad_norm": 3.8362972736358643, |
|
"learning_rate": 2.3735810113519093e-05, |
|
"loss": 1.0392, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.47812492475551993, |
|
"grad_norm": 3.222538709640503, |
|
"learning_rate": 2.39078087375301e-05, |
|
"loss": 1.0449, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.4815646724156316, |
|
"grad_norm": 4.081332683563232, |
|
"learning_rate": 2.407980736154111e-05, |
|
"loss": 1.0402, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.48500442007574324, |
|
"grad_norm": 3.8474042415618896, |
|
"learning_rate": 2.425180598555212e-05, |
|
"loss": 1.0199, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.4884441677358549, |
|
"grad_norm": 3.6466848850250244, |
|
"learning_rate": 2.4423804609563126e-05, |
|
"loss": 1.0352, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.49188391539596654, |
|
"grad_norm": 3.44462251663208, |
|
"learning_rate": 2.4595803233574134e-05, |
|
"loss": 1.0476, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.4953236630560782, |
|
"grad_norm": 2.647538661956787, |
|
"learning_rate": 2.476780185758514e-05, |
|
"loss": 1.0094, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.49876341071618985, |
|
"grad_norm": 2.3128774166107178, |
|
"learning_rate": 2.493980048159615e-05, |
|
"loss": 1.0291, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.5022031583763015, |
|
"grad_norm": 3.0225725173950195, |
|
"learning_rate": 2.511179910560716e-05, |
|
"loss": 1.0321, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.5056429060364132, |
|
"grad_norm": 2.3091073036193848, |
|
"learning_rate": 2.5283797729618164e-05, |
|
"loss": 1.0078, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.5090826536965248, |
|
"grad_norm": 2.2101924419403076, |
|
"learning_rate": 2.5455796353629174e-05, |
|
"loss": 1.0316, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.5125224013566365, |
|
"grad_norm": 2.288010597229004, |
|
"learning_rate": 2.562779497764018e-05, |
|
"loss": 1.0158, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.5159621490167481, |
|
"grad_norm": 2.550391674041748, |
|
"learning_rate": 2.579979360165119e-05, |
|
"loss": 1.0131, |
|
"step": 1500 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 14535, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.63721472e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|