|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.774406103693428, |
|
"eval_steps": 1000, |
|
"global_step": 8000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0034680076296167853, |
|
"grad_norm": 0.7528897523880005, |
|
"learning_rate": 0.00019982, |
|
"loss": 4.9086, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.006936015259233571, |
|
"grad_norm": 0.17007838189601898, |
|
"learning_rate": 0.00019962000000000002, |
|
"loss": 0.1029, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.010404022888850355, |
|
"grad_norm": 0.12162350863218307, |
|
"learning_rate": 0.00019942, |
|
"loss": 0.0687, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.013872030518467141, |
|
"grad_norm": 0.08915918320417404, |
|
"learning_rate": 0.00019922, |
|
"loss": 0.0636, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.017340038148083926, |
|
"grad_norm": 0.11356709897518158, |
|
"learning_rate": 0.00019902, |
|
"loss": 0.0623, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02080804577770071, |
|
"grad_norm": 0.09400122612714767, |
|
"learning_rate": 0.00019882, |
|
"loss": 0.0561, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.024276053407317495, |
|
"grad_norm": 0.1102517619729042, |
|
"learning_rate": 0.00019862000000000002, |
|
"loss": 0.0603, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.027744061036934282, |
|
"grad_norm": 0.09301582723855972, |
|
"learning_rate": 0.00019842000000000001, |
|
"loss": 0.0559, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.031212068666551067, |
|
"grad_norm": 0.08400452882051468, |
|
"learning_rate": 0.00019822, |
|
"loss": 0.0645, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03468007629616785, |
|
"grad_norm": 0.10444998741149902, |
|
"learning_rate": 0.00019802, |
|
"loss": 0.0644, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.038148083925784636, |
|
"grad_norm": 0.06524047255516052, |
|
"learning_rate": 0.00019782, |
|
"loss": 0.0599, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04161609155540142, |
|
"grad_norm": 0.0640910267829895, |
|
"learning_rate": 0.00019762, |
|
"loss": 0.0542, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.045084099185018205, |
|
"grad_norm": 0.07332012802362442, |
|
"learning_rate": 0.00019742000000000002, |
|
"loss": 0.0567, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.04855210681463499, |
|
"grad_norm": 4.02905797958374, |
|
"learning_rate": 0.00019722, |
|
"loss": 0.0592, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.05202011444425178, |
|
"grad_norm": 0.09805350750684738, |
|
"learning_rate": 0.00019702, |
|
"loss": 0.0657, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.055488122073868565, |
|
"grad_norm": 0.056836508214473724, |
|
"learning_rate": 0.00019682, |
|
"loss": 0.0523, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.05895612970348535, |
|
"grad_norm": 0.08670804649591446, |
|
"learning_rate": 0.00019662, |
|
"loss": 0.0585, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.062424137333102134, |
|
"grad_norm": 0.08405158668756485, |
|
"learning_rate": 0.00019642, |
|
"loss": 0.0537, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.06589214496271892, |
|
"grad_norm": 0.0825357437133789, |
|
"learning_rate": 0.00019622000000000002, |
|
"loss": 0.055, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.0693601525923357, |
|
"grad_norm": 0.06040720269083977, |
|
"learning_rate": 0.00019602, |
|
"loss": 0.0554, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07282816022195249, |
|
"grad_norm": 0.08198798447847366, |
|
"learning_rate": 0.00019582, |
|
"loss": 0.0599, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.07629616785156927, |
|
"grad_norm": 0.05527138710021973, |
|
"learning_rate": 0.00019562, |
|
"loss": 0.0533, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.07976417548118606, |
|
"grad_norm": 0.07315631955862045, |
|
"learning_rate": 0.00019542, |
|
"loss": 0.0528, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.08323218311080284, |
|
"grad_norm": 0.08420062810182571, |
|
"learning_rate": 0.00019522, |
|
"loss": 0.0518, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.08670019074041962, |
|
"grad_norm": 0.0529639795422554, |
|
"learning_rate": 0.00019502, |
|
"loss": 0.0568, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.09016819837003641, |
|
"grad_norm": 0.09339221566915512, |
|
"learning_rate": 0.00019482, |
|
"loss": 0.0557, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.0936362059996532, |
|
"grad_norm": 0.06453025341033936, |
|
"learning_rate": 0.00019462, |
|
"loss": 0.062, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.09710421362926998, |
|
"grad_norm": 0.07090363651514053, |
|
"learning_rate": 0.00019442, |
|
"loss": 0.0609, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.10057222125888678, |
|
"grad_norm": 0.1122497022151947, |
|
"learning_rate": 0.00019422, |
|
"loss": 0.0541, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.10404022888850356, |
|
"grad_norm": 0.09782398492097855, |
|
"learning_rate": 0.00019402, |
|
"loss": 0.0553, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.10750823651812035, |
|
"grad_norm": 0.06216060370206833, |
|
"learning_rate": 0.00019382, |
|
"loss": 0.0483, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.11097624414773713, |
|
"grad_norm": 0.08817891031503677, |
|
"learning_rate": 0.00019362, |
|
"loss": 0.058, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.11444425177735391, |
|
"grad_norm": 0.07251620292663574, |
|
"learning_rate": 0.00019342, |
|
"loss": 0.0477, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.1179122594069707, |
|
"grad_norm": 0.059537626802921295, |
|
"learning_rate": 0.00019322, |
|
"loss": 0.0559, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.12138026703658748, |
|
"grad_norm": 0.06329932063817978, |
|
"learning_rate": 0.00019302, |
|
"loss": 0.0557, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.12484827466620427, |
|
"grad_norm": 0.06601905822753906, |
|
"learning_rate": 0.00019282000000000001, |
|
"loss": 0.0575, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.12831628229582104, |
|
"grad_norm": 0.0639985054731369, |
|
"learning_rate": 0.00019262, |
|
"loss": 0.0519, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.13178428992543784, |
|
"grad_norm": 0.08187698572874069, |
|
"learning_rate": 0.00019242, |
|
"loss": 0.0545, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.13525229755505463, |
|
"grad_norm": 0.06790990382432938, |
|
"learning_rate": 0.00019222, |
|
"loss": 0.0474, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.1387203051846714, |
|
"grad_norm": 0.04906002804636955, |
|
"learning_rate": 0.00019202, |
|
"loss": 0.0579, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.1421883128142882, |
|
"grad_norm": 0.06125594303011894, |
|
"learning_rate": 0.00019182, |
|
"loss": 0.0421, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.14565632044390497, |
|
"grad_norm": 0.06252908706665039, |
|
"learning_rate": 0.00019162, |
|
"loss": 0.0577, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.14912432807352177, |
|
"grad_norm": 0.09376012533903122, |
|
"learning_rate": 0.00019142, |
|
"loss": 0.0546, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.15259233570313854, |
|
"grad_norm": 0.05025137588381767, |
|
"learning_rate": 0.00019122, |
|
"loss": 0.0505, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.15606034333275534, |
|
"grad_norm": 0.07034559547901154, |
|
"learning_rate": 0.00019102, |
|
"loss": 0.0495, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1595283509623721, |
|
"grad_norm": 0.07475865632295609, |
|
"learning_rate": 0.00019082, |
|
"loss": 0.0533, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.1629963585919889, |
|
"grad_norm": 0.07459491491317749, |
|
"learning_rate": 0.00019062, |
|
"loss": 0.0551, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.16646436622160568, |
|
"grad_norm": 0.052578963339328766, |
|
"learning_rate": 0.00019042, |
|
"loss": 0.0564, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.16993237385122248, |
|
"grad_norm": 0.05785336345434189, |
|
"learning_rate": 0.00019022, |
|
"loss": 0.0549, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.17340038148083925, |
|
"grad_norm": 0.06621215492486954, |
|
"learning_rate": 0.00019002, |
|
"loss": 0.0528, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.17686838911045605, |
|
"grad_norm": 0.07456778734922409, |
|
"learning_rate": 0.00018982000000000002, |
|
"loss": 0.0496, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.18033639674007282, |
|
"grad_norm": 0.06371001899242401, |
|
"learning_rate": 0.00018962000000000002, |
|
"loss": 0.0509, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.18380440436968962, |
|
"grad_norm": 0.052791863679885864, |
|
"learning_rate": 0.00018942, |
|
"loss": 0.0504, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.1872724119993064, |
|
"grad_norm": 0.06491260975599289, |
|
"learning_rate": 0.00018922, |
|
"loss": 0.0527, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.19074041962892319, |
|
"grad_norm": 0.07835149765014648, |
|
"learning_rate": 0.00018902000000000003, |
|
"loss": 0.0564, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.19420842725853996, |
|
"grad_norm": 0.05977100506424904, |
|
"learning_rate": 0.00018882000000000003, |
|
"loss": 0.0477, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.19767643488815675, |
|
"grad_norm": 0.07109620422124863, |
|
"learning_rate": 0.00018862000000000002, |
|
"loss": 0.0537, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.20114444251777355, |
|
"grad_norm": 0.10347943007946014, |
|
"learning_rate": 0.00018842000000000002, |
|
"loss": 0.0542, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.20461245014739032, |
|
"grad_norm": 0.06503281742334366, |
|
"learning_rate": 0.00018822, |
|
"loss": 0.0484, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.20808045777700712, |
|
"grad_norm": 0.07903438806533813, |
|
"learning_rate": 0.00018802, |
|
"loss": 0.0597, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2115484654066239, |
|
"grad_norm": 0.0713895708322525, |
|
"learning_rate": 0.00018782000000000003, |
|
"loss": 0.0498, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.2150164730362407, |
|
"grad_norm": 0.061313629150390625, |
|
"learning_rate": 0.00018762000000000002, |
|
"loss": 0.053, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.21848448066585746, |
|
"grad_norm": 0.07045572996139526, |
|
"learning_rate": 0.00018742000000000002, |
|
"loss": 0.0541, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.22195248829547426, |
|
"grad_norm": 0.07118247449398041, |
|
"learning_rate": 0.00018722, |
|
"loss": 0.0586, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.22542049592509103, |
|
"grad_norm": 0.05364071577787399, |
|
"learning_rate": 0.00018702, |
|
"loss": 0.0497, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.22888850355470783, |
|
"grad_norm": 0.07208040356636047, |
|
"learning_rate": 0.00018682000000000003, |
|
"loss": 0.0559, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.2323565111843246, |
|
"grad_norm": 0.07200731337070465, |
|
"learning_rate": 0.00018662000000000003, |
|
"loss": 0.048, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.2358245188139414, |
|
"grad_norm": 0.05730220302939415, |
|
"learning_rate": 0.00018642000000000002, |
|
"loss": 0.0495, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.23929252644355817, |
|
"grad_norm": 0.06378819793462753, |
|
"learning_rate": 0.00018622000000000002, |
|
"loss": 0.0552, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.24276053407317497, |
|
"grad_norm": 0.05866115912795067, |
|
"learning_rate": 0.00018602, |
|
"loss": 0.0541, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.24622854170279174, |
|
"grad_norm": 0.07053161412477493, |
|
"learning_rate": 0.00018582, |
|
"loss": 0.0533, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.24969654933240854, |
|
"grad_norm": 0.07450433820486069, |
|
"learning_rate": 0.00018562000000000003, |
|
"loss": 0.0527, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.25316455696202533, |
|
"grad_norm": 0.07172481715679169, |
|
"learning_rate": 0.00018542000000000002, |
|
"loss": 0.0583, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.2566325645916421, |
|
"grad_norm": 0.0536239892244339, |
|
"learning_rate": 0.00018522000000000002, |
|
"loss": 0.0486, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.2601005722212589, |
|
"grad_norm": 0.06321065127849579, |
|
"learning_rate": 0.00018502000000000001, |
|
"loss": 0.0412, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.2635685798508757, |
|
"grad_norm": 0.056946441531181335, |
|
"learning_rate": 0.00018482, |
|
"loss": 0.051, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.26703658748049247, |
|
"grad_norm": 0.05819573253393173, |
|
"learning_rate": 0.00018462, |
|
"loss": 0.0517, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.27050459511010927, |
|
"grad_norm": 0.05857665091753006, |
|
"learning_rate": 0.00018442000000000003, |
|
"loss": 0.0529, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.273972602739726, |
|
"grad_norm": 0.06014329940080643, |
|
"learning_rate": 0.00018422000000000002, |
|
"loss": 0.0544, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.2774406103693428, |
|
"grad_norm": 0.09187959134578705, |
|
"learning_rate": 0.00018402000000000002, |
|
"loss": 0.0439, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.2809086179989596, |
|
"grad_norm": 0.056131429970264435, |
|
"learning_rate": 0.00018382, |
|
"loss": 0.0389, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.2843766256285764, |
|
"grad_norm": 0.04884343966841698, |
|
"learning_rate": 0.00018362, |
|
"loss": 0.0475, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.28784463325819315, |
|
"grad_norm": 0.09032488614320755, |
|
"learning_rate": 0.00018342, |
|
"loss": 0.056, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.29131264088780995, |
|
"grad_norm": 0.09926522523164749, |
|
"learning_rate": 0.00018322000000000002, |
|
"loss": 0.0515, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.29478064851742675, |
|
"grad_norm": 0.04553750529885292, |
|
"learning_rate": 0.00018302000000000002, |
|
"loss": 0.0505, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.29824865614704354, |
|
"grad_norm": 0.058485984802246094, |
|
"learning_rate": 0.00018282000000000001, |
|
"loss": 0.0453, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.3017166637766603, |
|
"grad_norm": 0.052825070917606354, |
|
"learning_rate": 0.00018262, |
|
"loss": 0.053, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.3051846714062771, |
|
"grad_norm": 0.07626510411500931, |
|
"learning_rate": 0.00018242, |
|
"loss": 0.0474, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.3086526790358939, |
|
"grad_norm": 0.07244163751602173, |
|
"learning_rate": 0.00018222, |
|
"loss": 0.0456, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.3121206866655107, |
|
"grad_norm": 0.09651289880275726, |
|
"learning_rate": 0.00018202000000000002, |
|
"loss": 0.0499, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.3155886942951274, |
|
"grad_norm": 0.0737752839922905, |
|
"learning_rate": 0.00018182000000000002, |
|
"loss": 0.0517, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.3190567019247442, |
|
"grad_norm": 0.06631263345479965, |
|
"learning_rate": 0.00018162, |
|
"loss": 0.0499, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.322524709554361, |
|
"grad_norm": 0.10696552693843842, |
|
"learning_rate": 0.00018142, |
|
"loss": 0.0524, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.3259927171839778, |
|
"grad_norm": 0.06753025203943253, |
|
"learning_rate": 0.00018122, |
|
"loss": 0.0458, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.32946072481359456, |
|
"grad_norm": 0.05970798432826996, |
|
"learning_rate": 0.00018102000000000003, |
|
"loss": 0.0471, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.33292873244321136, |
|
"grad_norm": 0.05324438214302063, |
|
"learning_rate": 0.00018082000000000002, |
|
"loss": 0.0468, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.33639674007282816, |
|
"grad_norm": 0.0749637559056282, |
|
"learning_rate": 0.00018062000000000002, |
|
"loss": 0.0507, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.33986474770244496, |
|
"grad_norm": 0.07317759841680527, |
|
"learning_rate": 0.00018042, |
|
"loss": 0.0482, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.34333275533206176, |
|
"grad_norm": 0.11602938175201416, |
|
"learning_rate": 0.00018022, |
|
"loss": 0.0558, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.3468007629616785, |
|
"grad_norm": 0.07047244161367416, |
|
"learning_rate": 0.00018002, |
|
"loss": 0.0547, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.3468007629616785, |
|
"eval_loss": 0.062305010855197906, |
|
"eval_runtime": 714.4647, |
|
"eval_samples_per_second": 12.852, |
|
"eval_steps_per_second": 1.607, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.3502687705912953, |
|
"grad_norm": 0.07715447247028351, |
|
"learning_rate": 0.00017982000000000002, |
|
"loss": 0.0445, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.3537367782209121, |
|
"grad_norm": 0.08189492672681808, |
|
"learning_rate": 0.00017962000000000002, |
|
"loss": 0.0529, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.3572047858505289, |
|
"grad_norm": 0.06011577695608139, |
|
"learning_rate": 0.00017942, |
|
"loss": 0.0469, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.36067279348014564, |
|
"grad_norm": 0.06397314369678497, |
|
"learning_rate": 0.00017922, |
|
"loss": 0.0542, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.36414080110976244, |
|
"grad_norm": 0.06121763586997986, |
|
"learning_rate": 0.00017902, |
|
"loss": 0.0474, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.36760880873937923, |
|
"grad_norm": 0.0765228345990181, |
|
"learning_rate": 0.00017882, |
|
"loss": 0.0476, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.37107681636899603, |
|
"grad_norm": 0.0813635066151619, |
|
"learning_rate": 0.00017862000000000002, |
|
"loss": 0.0488, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.3745448239986128, |
|
"grad_norm": 0.06827688962221146, |
|
"learning_rate": 0.00017842000000000002, |
|
"loss": 0.0431, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.3780128316282296, |
|
"grad_norm": 0.06176091730594635, |
|
"learning_rate": 0.00017822, |
|
"loss": 0.0513, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.38148083925784637, |
|
"grad_norm": 0.07062922418117523, |
|
"learning_rate": 0.00017802, |
|
"loss": 0.0499, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.38494884688746317, |
|
"grad_norm": 0.059431031346321106, |
|
"learning_rate": 0.00017782, |
|
"loss": 0.0556, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.3884168545170799, |
|
"grad_norm": 0.06391894072294235, |
|
"learning_rate": 0.00017762, |
|
"loss": 0.0535, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.3918848621466967, |
|
"grad_norm": 0.08487355709075928, |
|
"learning_rate": 0.00017742000000000002, |
|
"loss": 0.0509, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.3953528697763135, |
|
"grad_norm": 0.06291911005973816, |
|
"learning_rate": 0.00017722000000000001, |
|
"loss": 0.0462, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.3988208774059303, |
|
"grad_norm": 0.06936580687761307, |
|
"learning_rate": 0.00017702, |
|
"loss": 0.0465, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.4022888850355471, |
|
"grad_norm": 0.06751543283462524, |
|
"learning_rate": 0.00017682, |
|
"loss": 0.0553, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.40575689266516385, |
|
"grad_norm": 0.08026771247386932, |
|
"learning_rate": 0.00017662, |
|
"loss": 0.0503, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.40922490029478065, |
|
"grad_norm": 0.05316636711359024, |
|
"learning_rate": 0.00017642, |
|
"loss": 0.0426, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.41269290792439745, |
|
"grad_norm": 0.15491995215415955, |
|
"learning_rate": 0.00017622000000000002, |
|
"loss": 0.0481, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.41616091555401424, |
|
"grad_norm": 0.059617578983306885, |
|
"learning_rate": 0.00017602, |
|
"loss": 0.0443, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.419628923183631, |
|
"grad_norm": 0.08931437879800797, |
|
"learning_rate": 0.00017582, |
|
"loss": 0.0462, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.4230969308132478, |
|
"grad_norm": 1.4069609642028809, |
|
"learning_rate": 0.00017562, |
|
"loss": 0.0577, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.4265649384428646, |
|
"grad_norm": 7.056313514709473, |
|
"learning_rate": 0.00017542, |
|
"loss": 0.3501, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.4300329460724814, |
|
"grad_norm": 0.8197808265686035, |
|
"learning_rate": 0.00017522000000000002, |
|
"loss": 0.1137, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.4335009537020981, |
|
"grad_norm": 0.4104197025299072, |
|
"learning_rate": 0.00017502000000000001, |
|
"loss": 0.1134, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.4369689613317149, |
|
"grad_norm": 1.3801881074905396, |
|
"learning_rate": 0.00017482, |
|
"loss": 0.0869, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.4404369689613317, |
|
"grad_norm": 3.6437034606933594, |
|
"learning_rate": 0.00017462, |
|
"loss": 0.0544, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.4439049765909485, |
|
"grad_norm": 1.1371792554855347, |
|
"learning_rate": 0.00017442, |
|
"loss": 0.0929, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.44737298422056526, |
|
"grad_norm": 2.9385204315185547, |
|
"learning_rate": 0.00017422, |
|
"loss": 0.1166, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.45084099185018206, |
|
"grad_norm": 0.8243001699447632, |
|
"learning_rate": 0.00017402000000000002, |
|
"loss": 0.0907, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.45430899947979886, |
|
"grad_norm": 1.3375756740570068, |
|
"learning_rate": 0.00017382, |
|
"loss": 0.0641, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.45777700710941566, |
|
"grad_norm": 0.7172356843948364, |
|
"learning_rate": 0.00017362, |
|
"loss": 0.0575, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.4612450147390324, |
|
"grad_norm": 0.18255668878555298, |
|
"learning_rate": 0.00017342, |
|
"loss": 0.0628, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.4647130223686492, |
|
"grad_norm": 0.11856569349765778, |
|
"learning_rate": 0.00017322, |
|
"loss": 0.1111, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.468181029998266, |
|
"grad_norm": 0.16621063649654388, |
|
"learning_rate": 0.00017302, |
|
"loss": 0.0446, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.4716490376278828, |
|
"grad_norm": 0.10158076882362366, |
|
"learning_rate": 0.00017282000000000002, |
|
"loss": 0.0532, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.4751170452574996, |
|
"grad_norm": 0.08783379942178726, |
|
"learning_rate": 0.00017262, |
|
"loss": 0.0491, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.47858505288711634, |
|
"grad_norm": 0.08198387920856476, |
|
"learning_rate": 0.00017242, |
|
"loss": 0.0485, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.48205306051673313, |
|
"grad_norm": 0.09546195715665817, |
|
"learning_rate": 0.00017222, |
|
"loss": 0.0486, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.48552106814634993, |
|
"grad_norm": 0.13258413970470428, |
|
"learning_rate": 0.00017202, |
|
"loss": 0.0544, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.48898907577596673, |
|
"grad_norm": 0.09508573263883591, |
|
"learning_rate": 0.00017182, |
|
"loss": 0.0486, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.4924570834055835, |
|
"grad_norm": 0.07258226722478867, |
|
"learning_rate": 0.00017162000000000001, |
|
"loss": 0.0448, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.49592509103520027, |
|
"grad_norm": 0.05399150773882866, |
|
"learning_rate": 0.00017142, |
|
"loss": 0.0436, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.49939309866481707, |
|
"grad_norm": 0.06922592222690582, |
|
"learning_rate": 0.00017122, |
|
"loss": 0.0451, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.5028611062944338, |
|
"grad_norm": 0.05059856176376343, |
|
"learning_rate": 0.00017102, |
|
"loss": 0.0413, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.5063291139240507, |
|
"grad_norm": 0.06339547783136368, |
|
"learning_rate": 0.00017082, |
|
"loss": 0.0495, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.5097971215536674, |
|
"grad_norm": 0.0521874763071537, |
|
"learning_rate": 0.00017062, |
|
"loss": 0.0489, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.5132651291832842, |
|
"grad_norm": 0.15334908664226532, |
|
"learning_rate": 0.00017042, |
|
"loss": 0.04, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.516733136812901, |
|
"grad_norm": 0.07494404166936874, |
|
"learning_rate": 0.00017022, |
|
"loss": 0.0537, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.5202011444425177, |
|
"grad_norm": 0.052238237112760544, |
|
"learning_rate": 0.00017002, |
|
"loss": 0.0492, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.5236691520721346, |
|
"grad_norm": 0.06761351227760315, |
|
"learning_rate": 0.00016982, |
|
"loss": 0.051, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.5271371597017513, |
|
"grad_norm": 0.07580805569887161, |
|
"learning_rate": 0.00016962, |
|
"loss": 0.0498, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.5306051673313681, |
|
"grad_norm": 0.07397795468568802, |
|
"learning_rate": 0.00016942000000000001, |
|
"loss": 0.0471, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.5340731749609849, |
|
"grad_norm": 0.04779529199004173, |
|
"learning_rate": 0.00016922, |
|
"loss": 0.0465, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.5375411825906017, |
|
"grad_norm": 0.08508727699518204, |
|
"learning_rate": 0.00016902, |
|
"loss": 0.046, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.5410091902202185, |
|
"grad_norm": 0.06656762957572937, |
|
"learning_rate": 0.00016882, |
|
"loss": 0.0433, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.5444771978498353, |
|
"grad_norm": 0.05622195079922676, |
|
"learning_rate": 0.00016862, |
|
"loss": 0.0455, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.547945205479452, |
|
"grad_norm": 0.06422954052686691, |
|
"learning_rate": 0.00016842, |
|
"loss": 0.0428, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.5514132131090689, |
|
"grad_norm": 0.08891351521015167, |
|
"learning_rate": 0.00016822, |
|
"loss": 0.0495, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.5548812207386856, |
|
"grad_norm": 0.08472294360399246, |
|
"learning_rate": 0.00016802, |
|
"loss": 0.0525, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.5583492283683024, |
|
"grad_norm": 0.08518495410680771, |
|
"learning_rate": 0.00016782, |
|
"loss": 0.0518, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.5618172359979192, |
|
"grad_norm": 0.1859533041715622, |
|
"learning_rate": 0.00016762, |
|
"loss": 0.0478, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.565285243627536, |
|
"grad_norm": 0.05245356634259224, |
|
"learning_rate": 0.00016742, |
|
"loss": 0.0418, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.5687532512571528, |
|
"grad_norm": 0.05982668697834015, |
|
"learning_rate": 0.00016722, |
|
"loss": 0.0458, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.5722212588867696, |
|
"grad_norm": 0.0750059187412262, |
|
"learning_rate": 0.00016702, |
|
"loss": 0.0489, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.5756892665163863, |
|
"grad_norm": 0.061683133244514465, |
|
"learning_rate": 0.00016682, |
|
"loss": 0.0461, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.5791572741460032, |
|
"grad_norm": 0.06833604723215103, |
|
"learning_rate": 0.00016662, |
|
"loss": 0.0466, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.5826252817756199, |
|
"grad_norm": 0.05507722124457359, |
|
"learning_rate": 0.00016642, |
|
"loss": 0.0416, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.5860932894052366, |
|
"grad_norm": 0.09453442692756653, |
|
"learning_rate": 0.00016622, |
|
"loss": 0.0418, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.5895612970348535, |
|
"grad_norm": 0.0574457123875618, |
|
"learning_rate": 0.00016601999999999999, |
|
"loss": 0.0495, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.5930293046644702, |
|
"grad_norm": 0.0829281136393547, |
|
"learning_rate": 0.00016582, |
|
"loss": 0.0435, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.5964973122940871, |
|
"grad_norm": 0.0569952018558979, |
|
"learning_rate": 0.00016562, |
|
"loss": 0.0476, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.5999653199237038, |
|
"grad_norm": 0.09291055798530579, |
|
"learning_rate": 0.00016542, |
|
"loss": 0.0446, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.6034333275533206, |
|
"grad_norm": 0.03767919912934303, |
|
"learning_rate": 0.00016522, |
|
"loss": 0.0426, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.6069013351829374, |
|
"grad_norm": 0.07064680010080338, |
|
"learning_rate": 0.00016502, |
|
"loss": 0.0491, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.6103693428125542, |
|
"grad_norm": 0.05687623471021652, |
|
"learning_rate": 0.00016482, |
|
"loss": 0.0454, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.613837350442171, |
|
"grad_norm": 0.04933289438486099, |
|
"learning_rate": 0.00016462, |
|
"loss": 0.0385, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.6173053580717878, |
|
"grad_norm": 0.0763295516371727, |
|
"learning_rate": 0.00016442000000000003, |
|
"loss": 0.0448, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.6207733657014045, |
|
"grad_norm": 0.04926356300711632, |
|
"learning_rate": 0.00016422000000000002, |
|
"loss": 0.0484, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.6242413733310214, |
|
"grad_norm": 0.07130167633295059, |
|
"learning_rate": 0.00016402000000000002, |
|
"loss": 0.0403, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.6277093809606381, |
|
"grad_norm": 0.06025327742099762, |
|
"learning_rate": 0.00016382000000000001, |
|
"loss": 0.0483, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.6311773885902549, |
|
"grad_norm": 0.06522911787033081, |
|
"learning_rate": 0.00016362, |
|
"loss": 0.0433, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.6346453962198717, |
|
"grad_norm": 0.09565310180187225, |
|
"learning_rate": 0.00016342, |
|
"loss": 0.0539, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.6381134038494884, |
|
"grad_norm": 0.08908990025520325, |
|
"learning_rate": 0.00016322000000000003, |
|
"loss": 0.0479, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.6415814114791053, |
|
"grad_norm": 0.05405285581946373, |
|
"learning_rate": 0.00016302000000000002, |
|
"loss": 0.0416, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.645049419108722, |
|
"grad_norm": 0.0656275674700737, |
|
"learning_rate": 0.00016282000000000002, |
|
"loss": 0.0455, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.6485174267383388, |
|
"grad_norm": 0.30592814087867737, |
|
"learning_rate": 0.00016262, |
|
"loss": 0.0471, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.6519854343679556, |
|
"grad_norm": 0.06433047354221344, |
|
"learning_rate": 0.00016242, |
|
"loss": 0.045, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.6554534419975724, |
|
"grad_norm": 0.05807631090283394, |
|
"learning_rate": 0.00016222000000000003, |
|
"loss": 0.0502, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.6589214496271891, |
|
"grad_norm": 0.09822454303503036, |
|
"learning_rate": 0.00016202000000000002, |
|
"loss": 0.0465, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.662389457256806, |
|
"grad_norm": 0.06036192551255226, |
|
"learning_rate": 0.00016182000000000002, |
|
"loss": 0.0455, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.6658574648864227, |
|
"grad_norm": 0.05637621134519577, |
|
"learning_rate": 0.00016162000000000001, |
|
"loss": 0.0459, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.6693254725160396, |
|
"grad_norm": 0.062302861362695694, |
|
"learning_rate": 0.00016142, |
|
"loss": 0.0461, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.6727934801456563, |
|
"grad_norm": 0.05843142420053482, |
|
"learning_rate": 0.00016122, |
|
"loss": 0.0455, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.6762614877752731, |
|
"grad_norm": 0.052565112709999084, |
|
"learning_rate": 0.00016102000000000003, |
|
"loss": 0.0432, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.6797294954048899, |
|
"grad_norm": 0.059093691408634186, |
|
"learning_rate": 0.00016082000000000002, |
|
"loss": 0.0433, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.6831975030345067, |
|
"grad_norm": 0.07197156548500061, |
|
"learning_rate": 0.00016062000000000002, |
|
"loss": 0.0465, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.6866655106641235, |
|
"grad_norm": 0.039968665689229965, |
|
"learning_rate": 0.00016042, |
|
"loss": 0.0381, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.6901335182937403, |
|
"grad_norm": 0.05672089383006096, |
|
"learning_rate": 0.00016022, |
|
"loss": 0.0422, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.693601525923357, |
|
"grad_norm": 0.10138531029224396, |
|
"learning_rate": 0.00016002, |
|
"loss": 0.0464, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.693601525923357, |
|
"eval_loss": 0.06039171665906906, |
|
"eval_runtime": 710.8114, |
|
"eval_samples_per_second": 12.918, |
|
"eval_steps_per_second": 1.615, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.6970695335529739, |
|
"grad_norm": 0.0607437863945961, |
|
"learning_rate": 0.00015982000000000002, |
|
"loss": 0.0481, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.7005375411825906, |
|
"grad_norm": 0.06681676208972931, |
|
"learning_rate": 0.00015962000000000002, |
|
"loss": 0.049, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.7040055488122073, |
|
"grad_norm": 0.07225602120161057, |
|
"learning_rate": 0.00015942000000000002, |
|
"loss": 0.0465, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.7074735564418242, |
|
"grad_norm": 0.056885506957769394, |
|
"learning_rate": 0.00015922, |
|
"loss": 0.0452, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.7109415640714409, |
|
"grad_norm": 0.05572199076414108, |
|
"learning_rate": 0.00015902, |
|
"loss": 0.0423, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.7144095717010578, |
|
"grad_norm": 0.0516788586974144, |
|
"learning_rate": 0.00015882, |
|
"loss": 0.0404, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.7178775793306745, |
|
"grad_norm": 0.05214313417673111, |
|
"learning_rate": 0.00015862000000000002, |
|
"loss": 0.0428, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.7213455869602913, |
|
"grad_norm": 0.06317329406738281, |
|
"learning_rate": 0.00015842000000000002, |
|
"loss": 0.0447, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.7248135945899081, |
|
"grad_norm": 0.07574247568845749, |
|
"learning_rate": 0.00015822, |
|
"loss": 0.0395, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.7282816022195249, |
|
"grad_norm": 0.06367363035678864, |
|
"learning_rate": 0.00015802, |
|
"loss": 0.0439, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.7317496098491416, |
|
"grad_norm": 0.06190785393118858, |
|
"learning_rate": 0.00015782, |
|
"loss": 0.0404, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.7352176174787585, |
|
"grad_norm": 0.07103675603866577, |
|
"learning_rate": 0.00015762, |
|
"loss": 0.0468, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.7386856251083752, |
|
"grad_norm": 0.04957522451877594, |
|
"learning_rate": 0.00015742000000000002, |
|
"loss": 0.0453, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.7421536327379921, |
|
"grad_norm": 0.07643826305866241, |
|
"learning_rate": 0.00015722000000000002, |
|
"loss": 0.0404, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.7456216403676088, |
|
"grad_norm": 0.04660920053720474, |
|
"learning_rate": 0.00015702, |
|
"loss": 0.0373, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.7490896479972255, |
|
"grad_norm": 0.06206243112683296, |
|
"learning_rate": 0.00015682, |
|
"loss": 0.0406, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.7525576556268424, |
|
"grad_norm": 0.07452013343572617, |
|
"learning_rate": 0.00015662, |
|
"loss": 0.043, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.7560256632564591, |
|
"grad_norm": 0.04379798844456673, |
|
"learning_rate": 0.00015642000000000002, |
|
"loss": 0.0494, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.759493670886076, |
|
"grad_norm": 0.05709415674209595, |
|
"learning_rate": 0.00015622000000000002, |
|
"loss": 0.041, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.7629616785156927, |
|
"grad_norm": 0.07468123733997345, |
|
"learning_rate": 0.00015602000000000001, |
|
"loss": 0.0403, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.7664296861453095, |
|
"grad_norm": 0.08100239932537079, |
|
"learning_rate": 0.00015582, |
|
"loss": 0.0481, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.7698976937749263, |
|
"grad_norm": 0.07269110530614853, |
|
"learning_rate": 0.00015562, |
|
"loss": 0.0394, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.7733657014045431, |
|
"grad_norm": 0.060352034866809845, |
|
"learning_rate": 0.00015542, |
|
"loss": 0.04, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.7768337090341598, |
|
"grad_norm": 0.05698138475418091, |
|
"learning_rate": 0.00015522000000000002, |
|
"loss": 0.0383, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.7803017166637767, |
|
"grad_norm": 0.06030441075563431, |
|
"learning_rate": 0.00015502000000000002, |
|
"loss": 0.0456, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.7837697242933934, |
|
"grad_norm": 0.08272086083889008, |
|
"learning_rate": 0.00015482, |
|
"loss": 0.0443, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.7872377319230103, |
|
"grad_norm": 0.06746231019496918, |
|
"learning_rate": 0.00015462, |
|
"loss": 0.04, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.790705739552627, |
|
"grad_norm": 0.08844996243715286, |
|
"learning_rate": 0.00015442, |
|
"loss": 0.0418, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.7941737471822438, |
|
"grad_norm": 0.04906987398862839, |
|
"learning_rate": 0.00015422, |
|
"loss": 0.0452, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.7976417548118606, |
|
"grad_norm": 0.053966376930475235, |
|
"learning_rate": 0.00015402000000000002, |
|
"loss": 0.0404, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.8011097624414774, |
|
"grad_norm": 0.07681586593389511, |
|
"learning_rate": 0.00015382000000000001, |
|
"loss": 0.05, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.8045777700710942, |
|
"grad_norm": 0.059431836009025574, |
|
"learning_rate": 0.00015362, |
|
"loss": 0.0496, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.808045777700711, |
|
"grad_norm": 0.058449339121580124, |
|
"learning_rate": 0.00015342, |
|
"loss": 0.0419, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.8115137853303277, |
|
"grad_norm": 0.05404651537537575, |
|
"learning_rate": 0.00015322, |
|
"loss": 0.0382, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.8149817929599446, |
|
"grad_norm": 0.06095472350716591, |
|
"learning_rate": 0.00015302, |
|
"loss": 0.0487, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.8184498005895613, |
|
"grad_norm": 0.07456117123365402, |
|
"learning_rate": 0.00015282000000000002, |
|
"loss": 0.0386, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.821917808219178, |
|
"grad_norm": 0.06098613142967224, |
|
"learning_rate": 0.00015262, |
|
"loss": 0.0425, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.8253858158487949, |
|
"grad_norm": 0.07008852809667587, |
|
"learning_rate": 0.00015242, |
|
"loss": 0.0457, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.8288538234784116, |
|
"grad_norm": 0.06413611769676208, |
|
"learning_rate": 0.00015222, |
|
"loss": 0.0433, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.8323218311080285, |
|
"grad_norm": 0.05168429762125015, |
|
"learning_rate": 0.00015202, |
|
"loss": 0.0441, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.8357898387376452, |
|
"grad_norm": 0.0708162784576416, |
|
"learning_rate": 0.00015182, |
|
"loss": 0.0393, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.839257846367262, |
|
"grad_norm": 0.05856487527489662, |
|
"learning_rate": 0.00015162000000000002, |
|
"loss": 0.0437, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.8427258539968788, |
|
"grad_norm": 0.06069020926952362, |
|
"learning_rate": 0.00015142, |
|
"loss": 0.0484, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.8461938616264956, |
|
"grad_norm": 0.0641227513551712, |
|
"learning_rate": 0.00015122, |
|
"loss": 0.046, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.8496618692561123, |
|
"grad_norm": 0.06422239542007446, |
|
"learning_rate": 0.00015102, |
|
"loss": 0.043, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.8531298768857292, |
|
"grad_norm": 0.07488572597503662, |
|
"learning_rate": 0.00015082, |
|
"loss": 0.0517, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.8565978845153459, |
|
"grad_norm": 0.06784242391586304, |
|
"learning_rate": 0.00015062000000000002, |
|
"loss": 0.0437, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.8600658921449628, |
|
"grad_norm": 0.07483550906181335, |
|
"learning_rate": 0.00015042, |
|
"loss": 0.0491, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.8635338997745795, |
|
"grad_norm": 0.06362838298082352, |
|
"learning_rate": 0.00015022, |
|
"loss": 0.043, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.8670019074041962, |
|
"grad_norm": 0.06175532937049866, |
|
"learning_rate": 0.00015002, |
|
"loss": 0.0425, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.8704699150338131, |
|
"grad_norm": 0.08987358957529068, |
|
"learning_rate": 0.00014982, |
|
"loss": 0.0479, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.8739379226634298, |
|
"grad_norm": 0.05209062620997429, |
|
"learning_rate": 0.00014962, |
|
"loss": 0.045, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.8774059302930467, |
|
"grad_norm": 0.055044736713171005, |
|
"learning_rate": 0.00014942000000000002, |
|
"loss": 0.0469, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.8808739379226634, |
|
"grad_norm": 0.0646812692284584, |
|
"learning_rate": 0.00014922, |
|
"loss": 0.0426, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.8843419455522802, |
|
"grad_norm": 0.048389263451099396, |
|
"learning_rate": 0.00014902, |
|
"loss": 0.0417, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.887809953181897, |
|
"grad_norm": 0.07316736876964569, |
|
"learning_rate": 0.00014882, |
|
"loss": 0.0412, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.8912779608115138, |
|
"grad_norm": 0.06801818311214447, |
|
"learning_rate": 0.00014862, |
|
"loss": 0.0446, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.8947459684411305, |
|
"grad_norm": 0.08190831542015076, |
|
"learning_rate": 0.00014842, |
|
"loss": 0.0423, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.8982139760707474, |
|
"grad_norm": 0.05850045010447502, |
|
"learning_rate": 0.00014822000000000001, |
|
"loss": 0.0418, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.9016819837003641, |
|
"grad_norm": 0.07633431255817413, |
|
"learning_rate": 0.00014802, |
|
"loss": 0.0439, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.905149991329981, |
|
"grad_norm": 0.07250861078500748, |
|
"learning_rate": 0.00014782, |
|
"loss": 0.0456, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.9086179989595977, |
|
"grad_norm": 0.05381698161363602, |
|
"learning_rate": 0.00014762, |
|
"loss": 0.0398, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.9120860065892145, |
|
"grad_norm": 0.05847073718905449, |
|
"learning_rate": 0.00014742, |
|
"loss": 0.0426, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.9155540142188313, |
|
"grad_norm": 0.04121188446879387, |
|
"learning_rate": 0.00014722, |
|
"loss": 0.0379, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.9190220218484481, |
|
"grad_norm": 0.05500589683651924, |
|
"learning_rate": 0.00014702, |
|
"loss": 0.0429, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.9224900294780648, |
|
"grad_norm": 0.07371719181537628, |
|
"learning_rate": 0.00014682, |
|
"loss": 0.0404, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.9259580371076817, |
|
"grad_norm": 0.07182417809963226, |
|
"learning_rate": 0.00014662, |
|
"loss": 0.0425, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.9294260447372984, |
|
"grad_norm": 0.07057616114616394, |
|
"learning_rate": 0.00014642, |
|
"loss": 0.0447, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.9328940523669152, |
|
"grad_norm": 0.06059495732188225, |
|
"learning_rate": 0.00014622, |
|
"loss": 0.0467, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.936362059996532, |
|
"grad_norm": 0.05471622198820114, |
|
"learning_rate": 0.00014602, |
|
"loss": 0.0429, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.9398300676261487, |
|
"grad_norm": 0.2795173227787018, |
|
"learning_rate": 0.00014582, |
|
"loss": 0.0441, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.9432980752557656, |
|
"grad_norm": 0.05505786091089249, |
|
"learning_rate": 0.00014562, |
|
"loss": 0.0481, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.9467660828853823, |
|
"grad_norm": 0.06604549288749695, |
|
"learning_rate": 0.00014542, |
|
"loss": 0.0436, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.9502340905149992, |
|
"grad_norm": 0.04876833036541939, |
|
"learning_rate": 0.00014522, |
|
"loss": 0.0427, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.9537020981446159, |
|
"grad_norm": 0.06586755067110062, |
|
"learning_rate": 0.00014502, |
|
"loss": 0.0376, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.9571701057742327, |
|
"grad_norm": 0.05991548299789429, |
|
"learning_rate": 0.00014482, |
|
"loss": 0.0425, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.9606381134038495, |
|
"grad_norm": 0.05464167147874832, |
|
"learning_rate": 0.00014462, |
|
"loss": 0.0475, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.9641061210334663, |
|
"grad_norm": 0.08428128808736801, |
|
"learning_rate": 0.00014442, |
|
"loss": 0.0407, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.967574128663083, |
|
"grad_norm": 0.05811979994177818, |
|
"learning_rate": 0.00014422, |
|
"loss": 0.0417, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.9710421362926999, |
|
"grad_norm": 0.06777170300483704, |
|
"learning_rate": 0.00014402, |
|
"loss": 0.0394, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.9745101439223166, |
|
"grad_norm": 0.07404989749193192, |
|
"learning_rate": 0.00014382, |
|
"loss": 0.0451, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.9779781515519335, |
|
"grad_norm": 0.08595024049282074, |
|
"learning_rate": 0.00014362, |
|
"loss": 0.0518, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.9814461591815502, |
|
"grad_norm": 0.07936517149209976, |
|
"learning_rate": 0.00014342, |
|
"loss": 0.0368, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.984914166811167, |
|
"grad_norm": 0.08811389654874802, |
|
"learning_rate": 0.00014322, |
|
"loss": 0.0444, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.9883821744407838, |
|
"grad_norm": 0.06864507496356964, |
|
"learning_rate": 0.00014302, |
|
"loss": 0.0445, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.9918501820704005, |
|
"grad_norm": 0.05267275124788284, |
|
"learning_rate": 0.00014282, |
|
"loss": 0.0485, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.9953181897000173, |
|
"grad_norm": 0.051028452813625336, |
|
"learning_rate": 0.00014261999999999999, |
|
"loss": 0.0416, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.9987861973296341, |
|
"grad_norm": 0.047300126403570175, |
|
"learning_rate": 0.00014242, |
|
"loss": 0.0443, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.0022542049592509, |
|
"grad_norm": 0.06848949193954468, |
|
"learning_rate": 0.00014222, |
|
"loss": 0.0422, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.0057222125888676, |
|
"grad_norm": 0.05861698463559151, |
|
"learning_rate": 0.00014202, |
|
"loss": 0.0298, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.0091902202184846, |
|
"grad_norm": 0.0671234130859375, |
|
"learning_rate": 0.00014182, |
|
"loss": 0.0271, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.0126582278481013, |
|
"grad_norm": 0.07007008045911789, |
|
"learning_rate": 0.00014162, |
|
"loss": 0.0279, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.016126235477718, |
|
"grad_norm": 0.07409070432186127, |
|
"learning_rate": 0.00014141999999999998, |
|
"loss": 0.0332, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.0195942431073348, |
|
"grad_norm": 0.0820993110537529, |
|
"learning_rate": 0.00014122, |
|
"loss": 0.032, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.0230622507369516, |
|
"grad_norm": 0.0683741420507431, |
|
"learning_rate": 0.00014102, |
|
"loss": 0.0285, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.0265302583665683, |
|
"grad_norm": 0.09559917449951172, |
|
"learning_rate": 0.00014082, |
|
"loss": 0.0328, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.0299982659961853, |
|
"grad_norm": 0.07927672564983368, |
|
"learning_rate": 0.00014062, |
|
"loss": 0.0317, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.033466273625802, |
|
"grad_norm": 0.06615123897790909, |
|
"learning_rate": 0.00014042, |
|
"loss": 0.0289, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.0369342812554188, |
|
"grad_norm": 0.051615212112665176, |
|
"learning_rate": 0.00014022, |
|
"loss": 0.0274, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.0404022888850355, |
|
"grad_norm": 0.060853827744722366, |
|
"learning_rate": 0.00014002, |
|
"loss": 0.033, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.0404022888850355, |
|
"eval_loss": 0.060293715447187424, |
|
"eval_runtime": 713.1843, |
|
"eval_samples_per_second": 12.875, |
|
"eval_steps_per_second": 1.61, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.0438702965146522, |
|
"grad_norm": 0.05449477955698967, |
|
"learning_rate": 0.00013982000000000003, |
|
"loss": 0.0308, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.0473383041442692, |
|
"grad_norm": 0.06479578465223312, |
|
"learning_rate": 0.00013962000000000002, |
|
"loss": 0.0343, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.050806311773886, |
|
"grad_norm": 0.060166362673044205, |
|
"learning_rate": 0.00013942000000000002, |
|
"loss": 0.0366, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.0542743194035027, |
|
"grad_norm": 0.07008329033851624, |
|
"learning_rate": 0.00013922, |
|
"loss": 0.0327, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.0577423270331194, |
|
"grad_norm": 0.07188612222671509, |
|
"learning_rate": 0.00013902, |
|
"loss": 0.0317, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.0612103346627362, |
|
"grad_norm": 0.06554035097360611, |
|
"learning_rate": 0.00013882000000000003, |
|
"loss": 0.031, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.0646783422923531, |
|
"grad_norm": 0.0675990879535675, |
|
"learning_rate": 0.00013862000000000002, |
|
"loss": 0.0341, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.0681463499219699, |
|
"grad_norm": 0.061157677322626114, |
|
"learning_rate": 0.00013842000000000002, |
|
"loss": 0.0301, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.0716143575515866, |
|
"grad_norm": 0.07951588183641434, |
|
"learning_rate": 0.00013822000000000001, |
|
"loss": 0.0289, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.0750823651812034, |
|
"grad_norm": 0.1063622459769249, |
|
"learning_rate": 0.00013802, |
|
"loss": 0.0325, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.0785503728108201, |
|
"grad_norm": 0.07916730642318726, |
|
"learning_rate": 0.00013782, |
|
"loss": 0.0338, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.082018380440437, |
|
"grad_norm": 0.0792151466012001, |
|
"learning_rate": 0.00013762000000000003, |
|
"loss": 0.0287, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.0854863880700538, |
|
"grad_norm": 0.0631512925028801, |
|
"learning_rate": 0.00013742000000000002, |
|
"loss": 0.031, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.0889543956996706, |
|
"grad_norm": 0.07032682001590729, |
|
"learning_rate": 0.00013722000000000002, |
|
"loss": 0.0284, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.0924224033292873, |
|
"grad_norm": 0.07017088681459427, |
|
"learning_rate": 0.00013702, |
|
"loss": 0.0292, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.095890410958904, |
|
"grad_norm": 0.06100435182452202, |
|
"learning_rate": 0.00013682, |
|
"loss": 0.0359, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.0993584185885208, |
|
"grad_norm": 0.05606581270694733, |
|
"learning_rate": 0.00013662, |
|
"loss": 0.032, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.1028264262181378, |
|
"grad_norm": 0.07687368988990784, |
|
"learning_rate": 0.00013642000000000003, |
|
"loss": 0.0353, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.1062944338477545, |
|
"grad_norm": 0.05796977877616882, |
|
"learning_rate": 0.00013622000000000002, |
|
"loss": 0.0286, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.1097624414773712, |
|
"grad_norm": 0.08586996048688889, |
|
"learning_rate": 0.00013602000000000002, |
|
"loss": 0.0285, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.113230449106988, |
|
"grad_norm": 0.08199802041053772, |
|
"learning_rate": 0.00013582, |
|
"loss": 0.0346, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.1166984567366047, |
|
"grad_norm": 0.06581319123506546, |
|
"learning_rate": 0.00013562, |
|
"loss": 0.029, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.1201664643662217, |
|
"grad_norm": 0.06755177676677704, |
|
"learning_rate": 0.00013542, |
|
"loss": 0.0323, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.1236344719958384, |
|
"grad_norm": 0.07359416782855988, |
|
"learning_rate": 0.00013522000000000002, |
|
"loss": 0.031, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.1271024796254552, |
|
"grad_norm": 0.07140175998210907, |
|
"learning_rate": 0.00013502000000000002, |
|
"loss": 0.0341, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.130570487255072, |
|
"grad_norm": 0.07570434361696243, |
|
"learning_rate": 0.00013482000000000001, |
|
"loss": 0.0307, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.1340384948846887, |
|
"grad_norm": 0.0646577849984169, |
|
"learning_rate": 0.00013462, |
|
"loss": 0.0353, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.1375065025143056, |
|
"grad_norm": 0.07016121596097946, |
|
"learning_rate": 0.00013442, |
|
"loss": 0.0367, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.1409745101439224, |
|
"grad_norm": 0.06299825757741928, |
|
"learning_rate": 0.00013422, |
|
"loss": 0.0355, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.1444425177735391, |
|
"grad_norm": 0.07232199609279633, |
|
"learning_rate": 0.00013402000000000002, |
|
"loss": 0.0302, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.1479105254031559, |
|
"grad_norm": 0.06672387570142746, |
|
"learning_rate": 0.00013382000000000002, |
|
"loss": 0.0314, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.1513785330327726, |
|
"grad_norm": 0.06597165018320084, |
|
"learning_rate": 0.00013362, |
|
"loss": 0.0316, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.1548465406623896, |
|
"grad_norm": 0.07945774495601654, |
|
"learning_rate": 0.00013342, |
|
"loss": 0.0304, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.1583145482920063, |
|
"grad_norm": 0.0883309543132782, |
|
"learning_rate": 0.00013322, |
|
"loss": 0.0313, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.161782555921623, |
|
"grad_norm": 0.1610005646944046, |
|
"learning_rate": 0.00013302000000000002, |
|
"loss": 0.0379, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.1652505635512398, |
|
"grad_norm": 0.08350630104541779, |
|
"learning_rate": 0.00013282000000000002, |
|
"loss": 0.0283, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.1687185711808565, |
|
"grad_norm": 0.08908521384000778, |
|
"learning_rate": 0.00013262000000000001, |
|
"loss": 0.0271, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.1721865788104733, |
|
"grad_norm": 0.06909502297639847, |
|
"learning_rate": 0.00013242, |
|
"loss": 0.0312, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.1756545864400902, |
|
"grad_norm": 0.0837428942322731, |
|
"learning_rate": 0.00013222, |
|
"loss": 0.0351, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.179122594069707, |
|
"grad_norm": 0.06636606156826019, |
|
"learning_rate": 0.00013202, |
|
"loss": 0.0284, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.1825906016993237, |
|
"grad_norm": 0.0990837961435318, |
|
"learning_rate": 0.00013182000000000002, |
|
"loss": 0.0301, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.1860586093289405, |
|
"grad_norm": 0.08313869684934616, |
|
"learning_rate": 0.00013162000000000002, |
|
"loss": 0.0335, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.1895266169585572, |
|
"grad_norm": 0.07330479472875595, |
|
"learning_rate": 0.00013142, |
|
"loss": 0.0328, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.1929946245881742, |
|
"grad_norm": 0.05378459393978119, |
|
"learning_rate": 0.00013122, |
|
"loss": 0.0306, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.196462632217791, |
|
"grad_norm": 0.06030990183353424, |
|
"learning_rate": 0.00013102, |
|
"loss": 0.0303, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.1999306398474077, |
|
"grad_norm": 0.07298003882169724, |
|
"learning_rate": 0.00013082, |
|
"loss": 0.0312, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.2033986474770244, |
|
"grad_norm": 0.06307482719421387, |
|
"learning_rate": 0.00013062000000000002, |
|
"loss": 0.0324, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.2068666551066412, |
|
"grad_norm": 0.06007950380444527, |
|
"learning_rate": 0.00013042000000000002, |
|
"loss": 0.0339, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.2103346627362581, |
|
"grad_norm": 0.07138363271951675, |
|
"learning_rate": 0.00013022, |
|
"loss": 0.0371, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.2138026703658749, |
|
"grad_norm": 0.06266158819198608, |
|
"learning_rate": 0.00013002, |
|
"loss": 0.0329, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.2172706779954916, |
|
"grad_norm": 0.06397438049316406, |
|
"learning_rate": 0.00012982, |
|
"loss": 0.0296, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.2207386856251083, |
|
"grad_norm": 0.061814188957214355, |
|
"learning_rate": 0.00012962, |
|
"loss": 0.0306, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.224206693254725, |
|
"grad_norm": 0.060092389583587646, |
|
"learning_rate": 0.00012942000000000002, |
|
"loss": 0.0323, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.227674700884342, |
|
"grad_norm": 0.10667088627815247, |
|
"learning_rate": 0.00012922, |
|
"loss": 0.0298, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.2311427085139588, |
|
"grad_norm": 0.09048482030630112, |
|
"learning_rate": 0.00012902, |
|
"loss": 0.0319, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.2346107161435755, |
|
"grad_norm": 0.09124518185853958, |
|
"learning_rate": 0.00012882, |
|
"loss": 0.0361, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.2380787237731923, |
|
"grad_norm": 0.05594000220298767, |
|
"learning_rate": 0.00012862, |
|
"loss": 0.0341, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.241546731402809, |
|
"grad_norm": 0.06354895979166031, |
|
"learning_rate": 0.00012842, |
|
"loss": 0.0281, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.2450147390324258, |
|
"grad_norm": 0.059312548488378525, |
|
"learning_rate": 0.00012822000000000002, |
|
"loss": 0.0276, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.2484827466620427, |
|
"grad_norm": 0.06291409581899643, |
|
"learning_rate": 0.00012802, |
|
"loss": 0.0315, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.2519507542916595, |
|
"grad_norm": 0.05183565244078636, |
|
"learning_rate": 0.00012782, |
|
"loss": 0.0294, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.2554187619212762, |
|
"grad_norm": 0.06372030079364777, |
|
"learning_rate": 0.00012762, |
|
"loss": 0.03, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.258886769550893, |
|
"grad_norm": 0.1327325701713562, |
|
"learning_rate": 0.00012742, |
|
"loss": 0.0366, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.26235477718051, |
|
"grad_norm": 0.07337796688079834, |
|
"learning_rate": 0.00012722000000000002, |
|
"loss": 0.0324, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.2658227848101267, |
|
"grad_norm": 0.06626396626234055, |
|
"learning_rate": 0.00012702000000000001, |
|
"loss": 0.0322, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.2692907924397434, |
|
"grad_norm": 0.07255198061466217, |
|
"learning_rate": 0.00012682, |
|
"loss": 0.036, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.2727588000693602, |
|
"grad_norm": 0.0766686201095581, |
|
"learning_rate": 0.00012662, |
|
"loss": 0.0315, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.276226807698977, |
|
"grad_norm": 0.06377042084932327, |
|
"learning_rate": 0.00012642, |
|
"loss": 0.0322, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.2796948153285936, |
|
"grad_norm": 0.0723329707980156, |
|
"learning_rate": 0.00012622, |
|
"loss": 0.0356, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.2831628229582104, |
|
"grad_norm": 0.0876326933503151, |
|
"learning_rate": 0.00012602000000000002, |
|
"loss": 0.0346, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.2866308305878273, |
|
"grad_norm": 0.07476814091205597, |
|
"learning_rate": 0.00012582, |
|
"loss": 0.0316, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.290098838217444, |
|
"grad_norm": 0.07503268122673035, |
|
"learning_rate": 0.00012562, |
|
"loss": 0.0318, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.2935668458470608, |
|
"grad_norm": 0.10708837956190109, |
|
"learning_rate": 0.00012542, |
|
"loss": 0.0349, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.2970348534766776, |
|
"grad_norm": 0.08280046284198761, |
|
"learning_rate": 0.00012522, |
|
"loss": 0.0324, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.3005028611062945, |
|
"grad_norm": 0.06630868464708328, |
|
"learning_rate": 0.00012502, |
|
"loss": 0.0305, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.3039708687359113, |
|
"grad_norm": 0.09755595773458481, |
|
"learning_rate": 0.00012482000000000001, |
|
"loss": 0.0338, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.307438876365528, |
|
"grad_norm": 0.08033673465251923, |
|
"learning_rate": 0.00012462, |
|
"loss": 0.0312, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.3109068839951448, |
|
"grad_norm": 0.0811261385679245, |
|
"learning_rate": 0.00012442, |
|
"loss": 0.0322, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.3143748916247615, |
|
"grad_norm": 0.06264316290616989, |
|
"learning_rate": 0.00012422, |
|
"loss": 0.0292, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.3178428992543783, |
|
"grad_norm": 0.07748369127511978, |
|
"learning_rate": 0.00012402, |
|
"loss": 0.0302, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.3213109068839952, |
|
"grad_norm": 0.0690523013472557, |
|
"learning_rate": 0.00012382, |
|
"loss": 0.0336, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.324778914513612, |
|
"grad_norm": 0.09423090517520905, |
|
"learning_rate": 0.00012362, |
|
"loss": 0.0295, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.3282469221432287, |
|
"grad_norm": 0.08562049269676208, |
|
"learning_rate": 0.00012342, |
|
"loss": 0.0292, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.3317149297728454, |
|
"grad_norm": 0.05059509724378586, |
|
"learning_rate": 0.00012322, |
|
"loss": 0.0266, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.3351829374024624, |
|
"grad_norm": 0.06358881294727325, |
|
"learning_rate": 0.00012302, |
|
"loss": 0.0309, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.3386509450320792, |
|
"grad_norm": 0.1298878937959671, |
|
"learning_rate": 0.00012282, |
|
"loss": 0.0356, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.342118952661696, |
|
"grad_norm": 0.06405311822891235, |
|
"learning_rate": 0.00012262, |
|
"loss": 0.0317, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.3455869602913126, |
|
"grad_norm": 0.07615106552839279, |
|
"learning_rate": 0.00012242, |
|
"loss": 0.0285, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.3490549679209294, |
|
"grad_norm": 0.08331302553415298, |
|
"learning_rate": 0.00012222, |
|
"loss": 0.0359, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.3525229755505461, |
|
"grad_norm": 0.06869524717330933, |
|
"learning_rate": 0.00012202, |
|
"loss": 0.0319, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.3559909831801629, |
|
"grad_norm": 0.08540484309196472, |
|
"learning_rate": 0.00012182, |
|
"loss": 0.0328, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.3594589908097798, |
|
"grad_norm": 0.07091011852025986, |
|
"learning_rate": 0.00012162, |
|
"loss": 0.0368, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.3629269984393966, |
|
"grad_norm": 0.075434111058712, |
|
"learning_rate": 0.00012142, |
|
"loss": 0.0336, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.3663950060690133, |
|
"grad_norm": 0.06716951727867126, |
|
"learning_rate": 0.00012122, |
|
"loss": 0.0335, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.36986301369863, |
|
"grad_norm": 0.09082087874412537, |
|
"learning_rate": 0.00012102, |
|
"loss": 0.0345, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.373331021328247, |
|
"grad_norm": 0.0903453379869461, |
|
"learning_rate": 0.00012082, |
|
"loss": 0.0415, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.3767990289578638, |
|
"grad_norm": 0.052235305309295654, |
|
"learning_rate": 0.00012062, |
|
"loss": 0.0282, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.3802670365874805, |
|
"grad_norm": 0.07253699749708176, |
|
"learning_rate": 0.00012042, |
|
"loss": 0.0316, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.3837350442170973, |
|
"grad_norm": 0.0548410601913929, |
|
"learning_rate": 0.00012022, |
|
"loss": 0.0303, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.387203051846714, |
|
"grad_norm": 0.08785740286111832, |
|
"learning_rate": 0.00012001999999999999, |
|
"loss": 0.0337, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.387203051846714, |
|
"eval_loss": 0.060144323855638504, |
|
"eval_runtime": 708.9302, |
|
"eval_samples_per_second": 12.952, |
|
"eval_steps_per_second": 1.619, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.3906710594763307, |
|
"grad_norm": 0.0650157481431961, |
|
"learning_rate": 0.00011982, |
|
"loss": 0.0327, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.3941390671059477, |
|
"grad_norm": 0.06679214537143707, |
|
"learning_rate": 0.00011962, |
|
"loss": 0.0352, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.3976070747355644, |
|
"grad_norm": 0.08368890732526779, |
|
"learning_rate": 0.00011942, |
|
"loss": 0.0346, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.4010750823651812, |
|
"grad_norm": 0.09027834981679916, |
|
"learning_rate": 0.00011922, |
|
"loss": 0.0304, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.404543089994798, |
|
"grad_norm": 0.0602988600730896, |
|
"learning_rate": 0.00011902, |
|
"loss": 0.0278, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.408011097624415, |
|
"grad_norm": 0.08348573744297028, |
|
"learning_rate": 0.00011882, |
|
"loss": 0.0316, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.4114791052540316, |
|
"grad_norm": 0.065241239964962, |
|
"learning_rate": 0.00011862, |
|
"loss": 0.0308, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.4149471128836484, |
|
"grad_norm": 0.058722469955682755, |
|
"learning_rate": 0.00011842, |
|
"loss": 0.0312, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.4184151205132651, |
|
"grad_norm": 0.06701633334159851, |
|
"learning_rate": 0.00011822, |
|
"loss": 0.0303, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.4218831281428819, |
|
"grad_norm": 0.06035483255982399, |
|
"learning_rate": 0.00011802, |
|
"loss": 0.0263, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.4253511357724986, |
|
"grad_norm": 0.07209423929452896, |
|
"learning_rate": 0.00011782, |
|
"loss": 0.0299, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.4288191434021154, |
|
"grad_norm": 0.08608460426330566, |
|
"learning_rate": 0.00011762, |
|
"loss": 0.0299, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.4322871510317323, |
|
"grad_norm": 0.06970153748989105, |
|
"learning_rate": 0.00011742, |
|
"loss": 0.0345, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.435755158661349, |
|
"grad_norm": 0.05995609238743782, |
|
"learning_rate": 0.00011721999999999999, |
|
"loss": 0.0284, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.4392231662909658, |
|
"grad_norm": 0.03222940117120743, |
|
"learning_rate": 0.00011702, |
|
"loss": 0.0293, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.4426911739205825, |
|
"grad_norm": 0.07076498866081238, |
|
"learning_rate": 0.00011682, |
|
"loss": 0.0362, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.4461591815501995, |
|
"grad_norm": 0.07425186783075333, |
|
"learning_rate": 0.00011661999999999999, |
|
"loss": 0.0281, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.4496271891798163, |
|
"grad_norm": 0.04875819757580757, |
|
"learning_rate": 0.00011642, |
|
"loss": 0.0361, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.453095196809433, |
|
"grad_norm": 0.06577154994010925, |
|
"learning_rate": 0.00011622, |
|
"loss": 0.0291, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.4565632044390497, |
|
"grad_norm": 0.08174604177474976, |
|
"learning_rate": 0.00011601999999999999, |
|
"loss": 0.0284, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.4600312120686665, |
|
"grad_norm": 0.08212857693433762, |
|
"learning_rate": 0.00011582, |
|
"loss": 0.03, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.4634992196982832, |
|
"grad_norm": 0.06090838089585304, |
|
"learning_rate": 0.00011562, |
|
"loss": 0.028, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.4669672273279002, |
|
"grad_norm": 0.07029874622821808, |
|
"learning_rate": 0.00011541999999999999, |
|
"loss": 0.033, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.470435234957517, |
|
"grad_norm": 0.10370688140392303, |
|
"learning_rate": 0.00011522, |
|
"loss": 0.0341, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.4739032425871337, |
|
"grad_norm": 0.06497831642627716, |
|
"learning_rate": 0.00011501999999999999, |
|
"loss": 0.0355, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.4773712502167504, |
|
"grad_norm": 0.047859255224466324, |
|
"learning_rate": 0.00011482000000000002, |
|
"loss": 0.0312, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.4808392578463674, |
|
"grad_norm": 0.044814929366111755, |
|
"learning_rate": 0.00011462000000000001, |
|
"loss": 0.0325, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.4843072654759841, |
|
"grad_norm": 0.08687663078308105, |
|
"learning_rate": 0.00011442000000000002, |
|
"loss": 0.0346, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.4877752731056009, |
|
"grad_norm": 0.06687606126070023, |
|
"learning_rate": 0.00011422000000000001, |
|
"loss": 0.0318, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.4912432807352176, |
|
"grad_norm": 0.0769667997956276, |
|
"learning_rate": 0.00011402000000000001, |
|
"loss": 0.0314, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.4947112883648344, |
|
"grad_norm": 0.07912110537290573, |
|
"learning_rate": 0.00011382000000000002, |
|
"loss": 0.0269, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.498179295994451, |
|
"grad_norm": 0.06801219284534454, |
|
"learning_rate": 0.00011362000000000001, |
|
"loss": 0.032, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.5016473036240678, |
|
"grad_norm": 0.07353610545396805, |
|
"learning_rate": 0.00011342000000000001, |
|
"loss": 0.0328, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.5051153112536848, |
|
"grad_norm": 0.05926644429564476, |
|
"learning_rate": 0.00011322000000000002, |
|
"loss": 0.0268, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.5085833188833015, |
|
"grad_norm": 0.07942460477352142, |
|
"learning_rate": 0.00011302000000000001, |
|
"loss": 0.0299, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.5120513265129183, |
|
"grad_norm": 0.09032566100358963, |
|
"learning_rate": 0.00011282000000000002, |
|
"loss": 0.0344, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.5155193341425353, |
|
"grad_norm": 0.08992986381053925, |
|
"learning_rate": 0.00011262000000000002, |
|
"loss": 0.0282, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.518987341772152, |
|
"grad_norm": 0.10055962204933167, |
|
"learning_rate": 0.00011242000000000001, |
|
"loss": 0.0368, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.5224553494017687, |
|
"grad_norm": 0.06706701964139938, |
|
"learning_rate": 0.00011222000000000002, |
|
"loss": 0.0348, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.5259233570313855, |
|
"grad_norm": 0.07412678748369217, |
|
"learning_rate": 0.00011202000000000002, |
|
"loss": 0.0322, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.5293913646610022, |
|
"grad_norm": 0.0761900544166565, |
|
"learning_rate": 0.00011182000000000001, |
|
"loss": 0.0334, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.532859372290619, |
|
"grad_norm": 0.06172578036785126, |
|
"learning_rate": 0.00011162000000000002, |
|
"loss": 0.0326, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.5363273799202357, |
|
"grad_norm": 0.06953331083059311, |
|
"learning_rate": 0.00011142000000000001, |
|
"loss": 0.0298, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.5397953875498525, |
|
"grad_norm": 0.07618329674005508, |
|
"learning_rate": 0.00011122000000000001, |
|
"loss": 0.0283, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.5432633951794694, |
|
"grad_norm": 0.07265307009220123, |
|
"learning_rate": 0.00011102000000000002, |
|
"loss": 0.032, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.5467314028090862, |
|
"grad_norm": 0.07409724593162537, |
|
"learning_rate": 0.00011082000000000001, |
|
"loss": 0.0303, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.5501994104387031, |
|
"grad_norm": 0.05352557823061943, |
|
"learning_rate": 0.00011062000000000001, |
|
"loss": 0.0304, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.5536674180683199, |
|
"grad_norm": 0.10606401413679123, |
|
"learning_rate": 0.00011042000000000002, |
|
"loss": 0.0338, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.5571354256979366, |
|
"grad_norm": 0.07364092022180557, |
|
"learning_rate": 0.00011022000000000001, |
|
"loss": 0.0308, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.5606034333275534, |
|
"grad_norm": 0.08737417310476303, |
|
"learning_rate": 0.00011002000000000001, |
|
"loss": 0.0299, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.56407144095717, |
|
"grad_norm": 0.07062090188264847, |
|
"learning_rate": 0.00010982000000000002, |
|
"loss": 0.0314, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.5675394485867868, |
|
"grad_norm": 0.0711718276143074, |
|
"learning_rate": 0.00010962000000000001, |
|
"loss": 0.0322, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.5710074562164036, |
|
"grad_norm": 0.08718711882829666, |
|
"learning_rate": 0.00010942, |
|
"loss": 0.0297, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.5744754638460203, |
|
"grad_norm": 0.06502439081668854, |
|
"learning_rate": 0.00010922000000000001, |
|
"loss": 0.0385, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.5779434714756373, |
|
"grad_norm": 0.07162761688232422, |
|
"learning_rate": 0.00010902000000000001, |
|
"loss": 0.0301, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.581411479105254, |
|
"grad_norm": 0.07721313089132309, |
|
"learning_rate": 0.00010882, |
|
"loss": 0.0284, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.5848794867348708, |
|
"grad_norm": 0.07071566581726074, |
|
"learning_rate": 0.00010862000000000001, |
|
"loss": 0.0352, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.5883474943644877, |
|
"grad_norm": 0.1029210165143013, |
|
"learning_rate": 0.00010842000000000001, |
|
"loss": 0.0319, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.5918155019941045, |
|
"grad_norm": 0.05683687701821327, |
|
"learning_rate": 0.00010822, |
|
"loss": 0.0343, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.5952835096237212, |
|
"grad_norm": 0.05821290984749794, |
|
"learning_rate": 0.00010802000000000001, |
|
"loss": 0.039, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.598751517253338, |
|
"grad_norm": 0.07931312173604965, |
|
"learning_rate": 0.00010782000000000001, |
|
"loss": 0.0294, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.6022195248829547, |
|
"grad_norm": 0.06197603419423103, |
|
"learning_rate": 0.00010762, |
|
"loss": 0.0322, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.6056875325125715, |
|
"grad_norm": 0.06312838196754456, |
|
"learning_rate": 0.00010742000000000001, |
|
"loss": 0.0298, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.6091555401421882, |
|
"grad_norm": 0.07855828106403351, |
|
"learning_rate": 0.00010722000000000001, |
|
"loss": 0.0303, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.612623547771805, |
|
"grad_norm": 0.055718790739774704, |
|
"learning_rate": 0.00010702000000000002, |
|
"loss": 0.0311, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.616091555401422, |
|
"grad_norm": 0.07305306941270828, |
|
"learning_rate": 0.00010682000000000001, |
|
"loss": 0.0303, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.6195595630310387, |
|
"grad_norm": 0.07300154864788055, |
|
"learning_rate": 0.00010662, |
|
"loss": 0.0301, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.6230275706606556, |
|
"grad_norm": 0.06121309846639633, |
|
"learning_rate": 0.00010642000000000001, |
|
"loss": 0.029, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.6264955782902724, |
|
"grad_norm": 0.060993146151304245, |
|
"learning_rate": 0.00010622000000000001, |
|
"loss": 0.0297, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.629963585919889, |
|
"grad_norm": 0.07691816985607147, |
|
"learning_rate": 0.00010602, |
|
"loss": 0.0339, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.6334315935495058, |
|
"grad_norm": 0.07278670370578766, |
|
"learning_rate": 0.00010582000000000001, |
|
"loss": 0.0262, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.6368996011791226, |
|
"grad_norm": 0.05704551190137863, |
|
"learning_rate": 0.00010562000000000001, |
|
"loss": 0.028, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.6403676088087393, |
|
"grad_norm": 0.07973553240299225, |
|
"learning_rate": 0.00010542, |
|
"loss": 0.0328, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.643835616438356, |
|
"grad_norm": 0.05720138177275658, |
|
"learning_rate": 0.00010522000000000001, |
|
"loss": 0.0309, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.6473036240679728, |
|
"grad_norm": 0.07787197083234787, |
|
"learning_rate": 0.00010502000000000001, |
|
"loss": 0.0294, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.6507716316975898, |
|
"grad_norm": 0.08834118396043777, |
|
"learning_rate": 0.00010482, |
|
"loss": 0.0301, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.6542396393272065, |
|
"grad_norm": 0.08383214473724365, |
|
"learning_rate": 0.00010462000000000001, |
|
"loss": 0.0322, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.6577076469568233, |
|
"grad_norm": 0.08266714960336685, |
|
"learning_rate": 0.00010442, |
|
"loss": 0.0323, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.6611756545864402, |
|
"grad_norm": 0.06535809487104416, |
|
"learning_rate": 0.00010422, |
|
"loss": 0.0334, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.664643662216057, |
|
"grad_norm": 0.07224865257740021, |
|
"learning_rate": 0.00010402000000000001, |
|
"loss": 0.0312, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.6681116698456737, |
|
"grad_norm": 0.07816470414400101, |
|
"learning_rate": 0.00010382, |
|
"loss": 0.0342, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.6715796774752905, |
|
"grad_norm": 0.48284128308296204, |
|
"learning_rate": 0.00010362, |
|
"loss": 0.0387, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.6750476851049072, |
|
"grad_norm": 0.15331751108169556, |
|
"learning_rate": 0.00010342000000000001, |
|
"loss": 0.0289, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.678515692734524, |
|
"grad_norm": 0.08506326377391815, |
|
"learning_rate": 0.00010322, |
|
"loss": 0.0299, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.6819837003641407, |
|
"grad_norm": 0.06808125227689743, |
|
"learning_rate": 0.00010302, |
|
"loss": 0.0266, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.6854517079937574, |
|
"grad_norm": 0.07349207252264023, |
|
"learning_rate": 0.00010282000000000001, |
|
"loss": 0.0297, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.6889197156233744, |
|
"grad_norm": 0.3372306525707245, |
|
"learning_rate": 0.00010262, |
|
"loss": 0.0342, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.6923877232529911, |
|
"grad_norm": 0.10550106316804886, |
|
"learning_rate": 0.00010242, |
|
"loss": 0.0338, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.695855730882608, |
|
"grad_norm": 0.0708487331867218, |
|
"learning_rate": 0.00010222000000000001, |
|
"loss": 0.0294, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.6993237385122248, |
|
"grad_norm": 0.5124090313911438, |
|
"learning_rate": 0.00010202, |
|
"loss": 0.0321, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.7027917461418416, |
|
"grad_norm": 0.08598774671554565, |
|
"learning_rate": 0.00010182, |
|
"loss": 0.0326, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.7062597537714583, |
|
"grad_norm": 0.06315886229276657, |
|
"learning_rate": 0.00010162, |
|
"loss": 0.0358, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.709727761401075, |
|
"grad_norm": 0.2513497769832611, |
|
"learning_rate": 0.00010142, |
|
"loss": 0.0352, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.7131957690306918, |
|
"grad_norm": 0.06378067284822464, |
|
"learning_rate": 0.00010122000000000001, |
|
"loss": 0.0277, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.7166637766603086, |
|
"grad_norm": 0.08322855085134506, |
|
"learning_rate": 0.00010102, |
|
"loss": 0.0303, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.7201317842899253, |
|
"grad_norm": 0.1097235381603241, |
|
"learning_rate": 0.00010082, |
|
"loss": 0.0328, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.7235997919195423, |
|
"grad_norm": 0.06668414920568466, |
|
"learning_rate": 0.00010062000000000001, |
|
"loss": 0.034, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.727067799549159, |
|
"grad_norm": 0.08046268671751022, |
|
"learning_rate": 0.00010042, |
|
"loss": 0.0315, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.7305358071787758, |
|
"grad_norm": 0.06834772229194641, |
|
"learning_rate": 0.00010022, |
|
"loss": 0.0304, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.7340038148083927, |
|
"grad_norm": 0.06567910313606262, |
|
"learning_rate": 0.00010002000000000001, |
|
"loss": 0.0338, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.7340038148083927, |
|
"eval_loss": 0.059129249304533005, |
|
"eval_runtime": 713.2468, |
|
"eval_samples_per_second": 12.874, |
|
"eval_steps_per_second": 1.61, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.7374718224380095, |
|
"grad_norm": 0.0854811891913414, |
|
"learning_rate": 9.982e-05, |
|
"loss": 0.0308, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.7409398300676262, |
|
"grad_norm": 0.07272527366876602, |
|
"learning_rate": 9.962e-05, |
|
"loss": 0.0333, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.744407837697243, |
|
"grad_norm": 0.0846826583147049, |
|
"learning_rate": 9.942000000000001e-05, |
|
"loss": 0.0342, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.7478758453268597, |
|
"grad_norm": 0.06776320934295654, |
|
"learning_rate": 9.922e-05, |
|
"loss": 0.0333, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.7513438529564764, |
|
"grad_norm": 0.07815729081630707, |
|
"learning_rate": 9.902e-05, |
|
"loss": 0.0355, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.7548118605860932, |
|
"grad_norm": 0.13363681733608246, |
|
"learning_rate": 9.882e-05, |
|
"loss": 0.032, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.75827986821571, |
|
"grad_norm": 0.05876624956727028, |
|
"learning_rate": 9.862e-05, |
|
"loss": 0.033, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.7617478758453269, |
|
"grad_norm": 0.10773160308599472, |
|
"learning_rate": 9.842e-05, |
|
"loss": 0.0309, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.7652158834749436, |
|
"grad_norm": 0.05928561091423035, |
|
"learning_rate": 9.822e-05, |
|
"loss": 0.0288, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.7686838911045606, |
|
"grad_norm": 0.058999065309762955, |
|
"learning_rate": 9.802e-05, |
|
"loss": 0.0251, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.7721518987341773, |
|
"grad_norm": 0.0768052414059639, |
|
"learning_rate": 9.782e-05, |
|
"loss": 0.0292, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.775619906363794, |
|
"grad_norm": 0.1691245585680008, |
|
"learning_rate": 9.762e-05, |
|
"loss": 0.0327, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.7790879139934108, |
|
"grad_norm": 0.08563978224992752, |
|
"learning_rate": 9.742e-05, |
|
"loss": 0.0321, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.7825559216230276, |
|
"grad_norm": 0.07125357538461685, |
|
"learning_rate": 9.722e-05, |
|
"loss": 0.0281, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.7860239292526443, |
|
"grad_norm": 0.10519967973232269, |
|
"learning_rate": 9.702e-05, |
|
"loss": 0.03, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.789491936882261, |
|
"grad_norm": 0.08440076559782028, |
|
"learning_rate": 9.682e-05, |
|
"loss": 0.0309, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.7929599445118778, |
|
"grad_norm": 0.09944937378168106, |
|
"learning_rate": 9.661999999999999e-05, |
|
"loss": 0.0323, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.7964279521414948, |
|
"grad_norm": 0.07523104548454285, |
|
"learning_rate": 9.642e-05, |
|
"loss": 0.026, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.7998959597711115, |
|
"grad_norm": 0.09478747099637985, |
|
"learning_rate": 9.622000000000001e-05, |
|
"loss": 0.0304, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.8033639674007282, |
|
"grad_norm": 0.0627417042851448, |
|
"learning_rate": 9.602e-05, |
|
"loss": 0.0353, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.8068319750303452, |
|
"grad_norm": 0.06294772773981094, |
|
"learning_rate": 9.582000000000001e-05, |
|
"loss": 0.0301, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.810299982659962, |
|
"grad_norm": 0.06453125923871994, |
|
"learning_rate": 9.562000000000001e-05, |
|
"loss": 0.0372, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.8137679902895787, |
|
"grad_norm": 0.08603645861148834, |
|
"learning_rate": 9.542e-05, |
|
"loss": 0.0313, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.8172359979191954, |
|
"grad_norm": 0.0609930120408535, |
|
"learning_rate": 9.522000000000001e-05, |
|
"loss": 0.0337, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.8207040055488122, |
|
"grad_norm": 0.06789145618677139, |
|
"learning_rate": 9.502000000000001e-05, |
|
"loss": 0.0291, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.824172013178429, |
|
"grad_norm": 0.062086399644613266, |
|
"learning_rate": 9.482e-05, |
|
"loss": 0.0284, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.8276400208080457, |
|
"grad_norm": 0.0755184143781662, |
|
"learning_rate": 9.462000000000001e-05, |
|
"loss": 0.0311, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.8311080284376624, |
|
"grad_norm": 0.06724268943071365, |
|
"learning_rate": 9.442000000000001e-05, |
|
"loss": 0.0288, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.8345760360672794, |
|
"grad_norm": 0.08856779336929321, |
|
"learning_rate": 9.422e-05, |
|
"loss": 0.0313, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.8380440436968961, |
|
"grad_norm": 0.08593250811100006, |
|
"learning_rate": 9.402000000000001e-05, |
|
"loss": 0.0299, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.841512051326513, |
|
"grad_norm": 0.07944291085004807, |
|
"learning_rate": 9.382e-05, |
|
"loss": 0.0309, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.8449800589561298, |
|
"grad_norm": 0.06297358870506287, |
|
"learning_rate": 9.362e-05, |
|
"loss": 0.0322, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.8484480665857466, |
|
"grad_norm": 0.0893145203590393, |
|
"learning_rate": 9.342000000000001e-05, |
|
"loss": 0.0297, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.8519160742153633, |
|
"grad_norm": 0.06626788526773453, |
|
"learning_rate": 9.322e-05, |
|
"loss": 0.0359, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.85538408184498, |
|
"grad_norm": 0.10941380262374878, |
|
"learning_rate": 9.302e-05, |
|
"loss": 0.0322, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.8588520894745968, |
|
"grad_norm": 0.08435889333486557, |
|
"learning_rate": 9.282000000000001e-05, |
|
"loss": 0.0361, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.8623200971042135, |
|
"grad_norm": 0.08986232429742813, |
|
"learning_rate": 9.262e-05, |
|
"loss": 0.0288, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.8657881047338303, |
|
"grad_norm": 0.0800371989607811, |
|
"learning_rate": 9.242000000000001e-05, |
|
"loss": 0.0339, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.8692561123634472, |
|
"grad_norm": 0.08191009610891342, |
|
"learning_rate": 9.222000000000001e-05, |
|
"loss": 0.0284, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.872724119993064, |
|
"grad_norm": 0.10277281701564789, |
|
"learning_rate": 9.202e-05, |
|
"loss": 0.0307, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.8761921276226807, |
|
"grad_norm": 0.08141244202852249, |
|
"learning_rate": 9.182000000000001e-05, |
|
"loss": 0.0327, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.8796601352522977, |
|
"grad_norm": 0.07554444670677185, |
|
"learning_rate": 9.162000000000001e-05, |
|
"loss": 0.031, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.8831281428819144, |
|
"grad_norm": 0.0898871198296547, |
|
"learning_rate": 9.142e-05, |
|
"loss": 0.0306, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.8865961505115312, |
|
"grad_norm": 0.06314833462238312, |
|
"learning_rate": 9.122000000000001e-05, |
|
"loss": 0.0348, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.890064158141148, |
|
"grad_norm": 0.06320305913686752, |
|
"learning_rate": 9.102e-05, |
|
"loss": 0.0277, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.8935321657707647, |
|
"grad_norm": 0.09586924314498901, |
|
"learning_rate": 9.082e-05, |
|
"loss": 0.0361, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.8970001734003814, |
|
"grad_norm": 0.0577247217297554, |
|
"learning_rate": 9.062000000000001e-05, |
|
"loss": 0.0295, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.9004681810299981, |
|
"grad_norm": 0.06483156979084015, |
|
"learning_rate": 9.042e-05, |
|
"loss": 0.0311, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.903936188659615, |
|
"grad_norm": 0.0748353898525238, |
|
"learning_rate": 9.022e-05, |
|
"loss": 0.0324, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.9074041962892319, |
|
"grad_norm": 0.09499184042215347, |
|
"learning_rate": 9.002000000000001e-05, |
|
"loss": 0.0307, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.9108722039188486, |
|
"grad_norm": 0.07709678262472153, |
|
"learning_rate": 8.982e-05, |
|
"loss": 0.0337, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.9143402115484656, |
|
"grad_norm": 0.06375749409198761, |
|
"learning_rate": 8.962e-05, |
|
"loss": 0.0231, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.9178082191780823, |
|
"grad_norm": 0.04287609085440636, |
|
"learning_rate": 8.942000000000001e-05, |
|
"loss": 0.0243, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.921276226807699, |
|
"grad_norm": 0.07707373052835464, |
|
"learning_rate": 8.922e-05, |
|
"loss": 0.0304, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.9247442344373158, |
|
"grad_norm": 0.06724567711353302, |
|
"learning_rate": 8.902e-05, |
|
"loss": 0.0323, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.9282122420669325, |
|
"grad_norm": 0.06895706802606583, |
|
"learning_rate": 8.882000000000001e-05, |
|
"loss": 0.0346, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.9316802496965493, |
|
"grad_norm": 0.0870724767446518, |
|
"learning_rate": 8.862e-05, |
|
"loss": 0.0326, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.935148257326166, |
|
"grad_norm": 0.061669524759054184, |
|
"learning_rate": 8.842e-05, |
|
"loss": 0.0287, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.9386162649557828, |
|
"grad_norm": 0.06450454145669937, |
|
"learning_rate": 8.822e-05, |
|
"loss": 0.0295, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.9420842725853997, |
|
"grad_norm": 0.07439760863780975, |
|
"learning_rate": 8.802e-05, |
|
"loss": 0.0325, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.9455522802150165, |
|
"grad_norm": 0.08620608597993851, |
|
"learning_rate": 8.782e-05, |
|
"loss": 0.0306, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.9490202878446332, |
|
"grad_norm": 0.08417027443647385, |
|
"learning_rate": 8.762e-05, |
|
"loss": 0.0339, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.9524882954742502, |
|
"grad_norm": 0.0670836940407753, |
|
"learning_rate": 8.742e-05, |
|
"loss": 0.0318, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.955956303103867, |
|
"grad_norm": 0.08096006512641907, |
|
"learning_rate": 8.722e-05, |
|
"loss": 0.0269, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.9594243107334837, |
|
"grad_norm": 0.04559866338968277, |
|
"learning_rate": 8.702e-05, |
|
"loss": 0.0271, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.9628923183631004, |
|
"grad_norm": 0.08112025260925293, |
|
"learning_rate": 8.682e-05, |
|
"loss": 0.036, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.9663603259927172, |
|
"grad_norm": 0.07936326414346695, |
|
"learning_rate": 8.662000000000001e-05, |
|
"loss": 0.0323, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.969828333622334, |
|
"grad_norm": 0.09269768744707108, |
|
"learning_rate": 8.642e-05, |
|
"loss": 0.0318, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.9732963412519506, |
|
"grad_norm": 0.06654026359319687, |
|
"learning_rate": 8.622e-05, |
|
"loss": 0.0291, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.9767643488815674, |
|
"grad_norm": 0.0971643254160881, |
|
"learning_rate": 8.602e-05, |
|
"loss": 0.031, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.9802323565111843, |
|
"grad_norm": 0.06067187711596489, |
|
"learning_rate": 8.582e-05, |
|
"loss": 0.0306, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.983700364140801, |
|
"grad_norm": 0.0945192202925682, |
|
"learning_rate": 8.562e-05, |
|
"loss": 0.0328, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.987168371770418, |
|
"grad_norm": 0.07627417147159576, |
|
"learning_rate": 8.542e-05, |
|
"loss": 0.0352, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.9906363794000348, |
|
"grad_norm": 0.08669853955507278, |
|
"learning_rate": 8.522e-05, |
|
"loss": 0.0337, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 1.9941043870296515, |
|
"grad_norm": 0.06610149890184402, |
|
"learning_rate": 8.502e-05, |
|
"loss": 0.0288, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 1.9975723946592683, |
|
"grad_norm": 0.06989070028066635, |
|
"learning_rate": 8.482e-05, |
|
"loss": 0.0306, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 2.001040402288885, |
|
"grad_norm": 0.04675092548131943, |
|
"learning_rate": 8.462e-05, |
|
"loss": 0.0245, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 2.0045084099185018, |
|
"grad_norm": 0.053521353751420975, |
|
"learning_rate": 8.442e-05, |
|
"loss": 0.0158, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 2.0079764175481185, |
|
"grad_norm": 0.06453324854373932, |
|
"learning_rate": 8.422e-05, |
|
"loss": 0.0131, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 2.0114444251777353, |
|
"grad_norm": 0.0763096958398819, |
|
"learning_rate": 8.402e-05, |
|
"loss": 0.0112, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.014912432807352, |
|
"grad_norm": 0.0837683454155922, |
|
"learning_rate": 8.382e-05, |
|
"loss": 0.0135, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 2.018380440436969, |
|
"grad_norm": 0.0868675634264946, |
|
"learning_rate": 8.362000000000002e-05, |
|
"loss": 0.0116, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 2.021848448066586, |
|
"grad_norm": 0.06839966773986816, |
|
"learning_rate": 8.342000000000001e-05, |
|
"loss": 0.011, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 2.0253164556962027, |
|
"grad_norm": 0.10859765112400055, |
|
"learning_rate": 8.322e-05, |
|
"loss": 0.0132, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 2.0287844633258194, |
|
"grad_norm": 0.08824854344129562, |
|
"learning_rate": 8.302000000000001e-05, |
|
"loss": 0.0129, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.032252470955436, |
|
"grad_norm": 0.0745101124048233, |
|
"learning_rate": 8.282000000000001e-05, |
|
"loss": 0.013, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 2.035720478585053, |
|
"grad_norm": 0.06870684772729874, |
|
"learning_rate": 8.262e-05, |
|
"loss": 0.0126, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 2.0391884862146696, |
|
"grad_norm": 0.08353777974843979, |
|
"learning_rate": 8.242000000000001e-05, |
|
"loss": 0.0127, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 2.0426564938442864, |
|
"grad_norm": 0.07282493263483047, |
|
"learning_rate": 8.222000000000001e-05, |
|
"loss": 0.014, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 2.046124501473903, |
|
"grad_norm": 0.07777410745620728, |
|
"learning_rate": 8.202e-05, |
|
"loss": 0.0108, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.04959250910352, |
|
"grad_norm": 0.06361842900514603, |
|
"learning_rate": 8.182000000000001e-05, |
|
"loss": 0.0132, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 2.0530605167331366, |
|
"grad_norm": 0.09724973887205124, |
|
"learning_rate": 8.162000000000001e-05, |
|
"loss": 0.0139, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 2.056528524362754, |
|
"grad_norm": 0.09359490126371384, |
|
"learning_rate": 8.142e-05, |
|
"loss": 0.013, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 2.0599965319923705, |
|
"grad_norm": 0.06116607412695885, |
|
"learning_rate": 8.122000000000001e-05, |
|
"loss": 0.0143, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 2.0634645396219873, |
|
"grad_norm": 0.06787212938070297, |
|
"learning_rate": 8.102000000000001e-05, |
|
"loss": 0.0137, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.066932547251604, |
|
"grad_norm": 0.06523068249225616, |
|
"learning_rate": 8.082e-05, |
|
"loss": 0.0116, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 2.0704005548812208, |
|
"grad_norm": 0.09076279401779175, |
|
"learning_rate": 8.062000000000001e-05, |
|
"loss": 0.01, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 2.0738685625108375, |
|
"grad_norm": 0.07997199892997742, |
|
"learning_rate": 8.042e-05, |
|
"loss": 0.0132, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 2.0773365701404543, |
|
"grad_norm": 0.0561593659222126, |
|
"learning_rate": 8.022e-05, |
|
"loss": 0.0117, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 2.080804577770071, |
|
"grad_norm": 0.08588268607854843, |
|
"learning_rate": 8.002000000000001e-05, |
|
"loss": 0.0125, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.080804577770071, |
|
"eval_loss": 0.06557230651378632, |
|
"eval_runtime": 707.733, |
|
"eval_samples_per_second": 12.974, |
|
"eval_steps_per_second": 1.622, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.0842725853996877, |
|
"grad_norm": 0.09114642441272736, |
|
"learning_rate": 7.982e-05, |
|
"loss": 0.0107, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 2.0877405930293045, |
|
"grad_norm": 0.06732100248336792, |
|
"learning_rate": 7.962e-05, |
|
"loss": 0.012, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 2.0912086006589217, |
|
"grad_norm": 0.08021605014801025, |
|
"learning_rate": 7.942000000000001e-05, |
|
"loss": 0.0123, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 2.0946766082885384, |
|
"grad_norm": 0.06761088222265244, |
|
"learning_rate": 7.922e-05, |
|
"loss": 0.0117, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 2.098144615918155, |
|
"grad_norm": 0.11337202787399292, |
|
"learning_rate": 7.902e-05, |
|
"loss": 0.0131, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.101612623547772, |
|
"grad_norm": 0.09859013557434082, |
|
"learning_rate": 7.882000000000001e-05, |
|
"loss": 0.0139, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 2.1050806311773886, |
|
"grad_norm": 0.08745191991329193, |
|
"learning_rate": 7.862e-05, |
|
"loss": 0.0139, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 2.1085486388070054, |
|
"grad_norm": 0.04903840273618698, |
|
"learning_rate": 7.842e-05, |
|
"loss": 0.0142, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 2.112016646436622, |
|
"grad_norm": 0.07992135733366013, |
|
"learning_rate": 7.822e-05, |
|
"loss": 0.0123, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 2.115484654066239, |
|
"grad_norm": 0.08518462628126144, |
|
"learning_rate": 7.802e-05, |
|
"loss": 0.0138, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.1189526616958556, |
|
"grad_norm": 0.08434431999921799, |
|
"learning_rate": 7.782000000000001e-05, |
|
"loss": 0.0155, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 2.1224206693254724, |
|
"grad_norm": 0.08734823763370514, |
|
"learning_rate": 7.762e-05, |
|
"loss": 0.0125, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 2.125888676955089, |
|
"grad_norm": 0.08129347115755081, |
|
"learning_rate": 7.742e-05, |
|
"loss": 0.0118, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 2.1293566845847063, |
|
"grad_norm": 0.09434698522090912, |
|
"learning_rate": 7.722000000000001e-05, |
|
"loss": 0.0128, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 2.132824692214323, |
|
"grad_norm": 0.08226180821657181, |
|
"learning_rate": 7.702e-05, |
|
"loss": 0.0159, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.1362926998439398, |
|
"grad_norm": 0.06351976841688156, |
|
"learning_rate": 7.682e-05, |
|
"loss": 0.0135, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 2.1397607074735565, |
|
"grad_norm": 0.07568191736936569, |
|
"learning_rate": 7.662000000000001e-05, |
|
"loss": 0.0122, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 2.1432287151031733, |
|
"grad_norm": 0.08052569627761841, |
|
"learning_rate": 7.642e-05, |
|
"loss": 0.0123, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 2.14669672273279, |
|
"grad_norm": 0.08767958730459213, |
|
"learning_rate": 7.622e-05, |
|
"loss": 0.0145, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 2.1501647303624067, |
|
"grad_norm": 0.09768091887235641, |
|
"learning_rate": 7.602000000000001e-05, |
|
"loss": 0.0113, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.1536327379920235, |
|
"grad_norm": 0.09293138980865479, |
|
"learning_rate": 7.582e-05, |
|
"loss": 0.0111, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 2.1571007456216402, |
|
"grad_norm": 0.07943341881036758, |
|
"learning_rate": 7.562e-05, |
|
"loss": 0.0133, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 2.160568753251257, |
|
"grad_norm": 0.08316273242235184, |
|
"learning_rate": 7.542e-05, |
|
"loss": 0.0142, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 2.164036760880874, |
|
"grad_norm": 0.05390129238367081, |
|
"learning_rate": 7.522e-05, |
|
"loss": 0.0128, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 2.167504768510491, |
|
"grad_norm": 0.05505843088030815, |
|
"learning_rate": 7.502e-05, |
|
"loss": 0.0124, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.1709727761401076, |
|
"grad_norm": 0.08427543193101883, |
|
"learning_rate": 7.482e-05, |
|
"loss": 0.0136, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 2.1744407837697244, |
|
"grad_norm": 0.06911098957061768, |
|
"learning_rate": 7.462e-05, |
|
"loss": 0.0117, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 2.177908791399341, |
|
"grad_norm": 0.08595360815525055, |
|
"learning_rate": 7.442e-05, |
|
"loss": 0.0155, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 2.181376799028958, |
|
"grad_norm": 0.0827205702662468, |
|
"learning_rate": 7.422e-05, |
|
"loss": 0.0127, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 2.1848448066585746, |
|
"grad_norm": 0.12057662755250931, |
|
"learning_rate": 7.402e-05, |
|
"loss": 0.0113, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.1883128142881914, |
|
"grad_norm": 0.10641255229711533, |
|
"learning_rate": 7.382e-05, |
|
"loss": 0.0123, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 2.191780821917808, |
|
"grad_norm": 0.11188361793756485, |
|
"learning_rate": 7.362e-05, |
|
"loss": 0.013, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 2.195248829547425, |
|
"grad_norm": 0.06386271864175797, |
|
"learning_rate": 7.342e-05, |
|
"loss": 0.0114, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 2.1987168371770416, |
|
"grad_norm": 0.09285665303468704, |
|
"learning_rate": 7.322e-05, |
|
"loss": 0.0156, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 2.2021848448066588, |
|
"grad_norm": 0.0885651484131813, |
|
"learning_rate": 7.302e-05, |
|
"loss": 0.013, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.2056528524362755, |
|
"grad_norm": 0.09753404557704926, |
|
"learning_rate": 7.282e-05, |
|
"loss": 0.0135, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 2.2091208600658923, |
|
"grad_norm": 0.06870284676551819, |
|
"learning_rate": 7.261999999999999e-05, |
|
"loss": 0.0163, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 2.212588867695509, |
|
"grad_norm": 0.07846927642822266, |
|
"learning_rate": 7.242e-05, |
|
"loss": 0.0147, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 2.2160568753251257, |
|
"grad_norm": 0.06510089337825775, |
|
"learning_rate": 7.222e-05, |
|
"loss": 0.0129, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 2.2195248829547425, |
|
"grad_norm": 0.13771073520183563, |
|
"learning_rate": 7.202e-05, |
|
"loss": 0.0151, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.2229928905843592, |
|
"grad_norm": 0.08157498389482498, |
|
"learning_rate": 7.182e-05, |
|
"loss": 0.0136, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 2.226460898213976, |
|
"grad_norm": 0.09444098174571991, |
|
"learning_rate": 7.162e-05, |
|
"loss": 0.016, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 2.2299289058435927, |
|
"grad_norm": 0.0773581713438034, |
|
"learning_rate": 7.142e-05, |
|
"loss": 0.0132, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 2.2333969134732095, |
|
"grad_norm": 0.10038639605045319, |
|
"learning_rate": 7.122000000000001e-05, |
|
"loss": 0.0142, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 2.2368649211028266, |
|
"grad_norm": 0.10728047043085098, |
|
"learning_rate": 7.102000000000001e-05, |
|
"loss": 0.0152, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.2403329287324434, |
|
"grad_norm": 0.10695026069879532, |
|
"learning_rate": 7.082e-05, |
|
"loss": 0.0163, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 2.24380093636206, |
|
"grad_norm": 0.06996133178472519, |
|
"learning_rate": 7.062000000000001e-05, |
|
"loss": 0.0149, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 2.247268943991677, |
|
"grad_norm": 0.04395058751106262, |
|
"learning_rate": 7.042000000000001e-05, |
|
"loss": 0.0127, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 2.2507369516212936, |
|
"grad_norm": 0.06262753158807755, |
|
"learning_rate": 7.022e-05, |
|
"loss": 0.014, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 2.2542049592509104, |
|
"grad_norm": 0.073272705078125, |
|
"learning_rate": 7.002000000000001e-05, |
|
"loss": 0.0121, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.257672966880527, |
|
"grad_norm": 0.11415940523147583, |
|
"learning_rate": 6.982e-05, |
|
"loss": 0.0136, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 2.261140974510144, |
|
"grad_norm": 0.09325289726257324, |
|
"learning_rate": 6.962e-05, |
|
"loss": 0.0121, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 2.2646089821397606, |
|
"grad_norm": 0.07223498821258545, |
|
"learning_rate": 6.942000000000001e-05, |
|
"loss": 0.0143, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 2.2680769897693773, |
|
"grad_norm": 0.08595094084739685, |
|
"learning_rate": 6.922e-05, |
|
"loss": 0.0137, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 2.271544997398994, |
|
"grad_norm": 0.08120746910572052, |
|
"learning_rate": 6.902000000000001e-05, |
|
"loss": 0.0101, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 2.2750130050286113, |
|
"grad_norm": 0.1085987389087677, |
|
"learning_rate": 6.882000000000001e-05, |
|
"loss": 0.0149, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 2.278481012658228, |
|
"grad_norm": 0.07946083694696426, |
|
"learning_rate": 6.862e-05, |
|
"loss": 0.0133, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 2.2819490202878447, |
|
"grad_norm": 0.05504854768514633, |
|
"learning_rate": 6.842000000000001e-05, |
|
"loss": 0.0139, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 2.2854170279174615, |
|
"grad_norm": 0.07158561050891876, |
|
"learning_rate": 6.822000000000001e-05, |
|
"loss": 0.0135, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 2.2888850355470782, |
|
"grad_norm": 0.06974880397319794, |
|
"learning_rate": 6.802e-05, |
|
"loss": 0.0123, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.292353043176695, |
|
"grad_norm": 0.08541780710220337, |
|
"learning_rate": 6.782000000000001e-05, |
|
"loss": 0.0144, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 2.2958210508063117, |
|
"grad_norm": 0.10203000158071518, |
|
"learning_rate": 6.762e-05, |
|
"loss": 0.0138, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 2.2992890584359285, |
|
"grad_norm": 0.09905651211738586, |
|
"learning_rate": 6.742e-05, |
|
"loss": 0.0145, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 2.302757066065545, |
|
"grad_norm": 0.04088059067726135, |
|
"learning_rate": 6.722000000000001e-05, |
|
"loss": 0.011, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 2.306225073695162, |
|
"grad_norm": 0.10128574818372726, |
|
"learning_rate": 6.702e-05, |
|
"loss": 0.0159, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 2.309693081324779, |
|
"grad_norm": 0.08455543220043182, |
|
"learning_rate": 6.682e-05, |
|
"loss": 0.0131, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 2.313161088954396, |
|
"grad_norm": 0.1268063485622406, |
|
"learning_rate": 6.662000000000001e-05, |
|
"loss": 0.0129, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 2.3166290965840126, |
|
"grad_norm": 0.11783897876739502, |
|
"learning_rate": 6.642e-05, |
|
"loss": 0.0106, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 2.3200971042136294, |
|
"grad_norm": 0.09608971327543259, |
|
"learning_rate": 6.622e-05, |
|
"loss": 0.0137, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 2.323565111843246, |
|
"grad_norm": 0.07495573163032532, |
|
"learning_rate": 6.602000000000001e-05, |
|
"loss": 0.0131, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.327033119472863, |
|
"grad_norm": 0.07772886753082275, |
|
"learning_rate": 6.582e-05, |
|
"loss": 0.0124, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 2.3305011271024796, |
|
"grad_norm": 0.07365009188652039, |
|
"learning_rate": 6.562e-05, |
|
"loss": 0.0112, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.3339691347320963, |
|
"grad_norm": 0.0697893276810646, |
|
"learning_rate": 6.542000000000001e-05, |
|
"loss": 0.0105, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 2.337437142361713, |
|
"grad_norm": 0.09057148545980453, |
|
"learning_rate": 6.522e-05, |
|
"loss": 0.0134, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 2.34090514999133, |
|
"grad_norm": 0.09601489454507828, |
|
"learning_rate": 6.502e-05, |
|
"loss": 0.014, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.3443731576209466, |
|
"grad_norm": 0.11969607323408127, |
|
"learning_rate": 6.482e-05, |
|
"loss": 0.0116, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 2.3478411652505637, |
|
"grad_norm": 0.103757843375206, |
|
"learning_rate": 6.462e-05, |
|
"loss": 0.0088, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 2.3513091728801805, |
|
"grad_norm": 0.09077152609825134, |
|
"learning_rate": 6.442e-05, |
|
"loss": 0.0111, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.3547771805097972, |
|
"grad_norm": 0.06362780928611755, |
|
"learning_rate": 6.422e-05, |
|
"loss": 0.01, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 2.358245188139414, |
|
"grad_norm": 0.10334885120391846, |
|
"learning_rate": 6.402e-05, |
|
"loss": 0.0144, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.3617131957690307, |
|
"grad_norm": 0.0676029622554779, |
|
"learning_rate": 6.382e-05, |
|
"loss": 0.0131, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 2.3651812033986475, |
|
"grad_norm": 0.06794017553329468, |
|
"learning_rate": 6.362e-05, |
|
"loss": 0.0136, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 2.368649211028264, |
|
"grad_norm": 0.11005677282810211, |
|
"learning_rate": 6.342e-05, |
|
"loss": 0.0141, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 2.372117218657881, |
|
"grad_norm": 0.07998326420783997, |
|
"learning_rate": 6.322000000000001e-05, |
|
"loss": 0.0119, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.3755852262874977, |
|
"grad_norm": 0.07724535465240479, |
|
"learning_rate": 6.302e-05, |
|
"loss": 0.0149, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.3790532339171144, |
|
"grad_norm": 0.07563537359237671, |
|
"learning_rate": 6.282e-05, |
|
"loss": 0.0145, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 2.3825212415467316, |
|
"grad_norm": 0.048786722123622894, |
|
"learning_rate": 6.262000000000001e-05, |
|
"loss": 0.0122, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 2.3859892491763484, |
|
"grad_norm": 0.09792380034923553, |
|
"learning_rate": 6.242e-05, |
|
"loss": 0.0133, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 2.389457256805965, |
|
"grad_norm": 0.06414589285850525, |
|
"learning_rate": 6.222e-05, |
|
"loss": 0.0119, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 2.392925264435582, |
|
"grad_norm": 0.0844031348824501, |
|
"learning_rate": 6.202e-05, |
|
"loss": 0.0139, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.3963932720651986, |
|
"grad_norm": 0.0652434229850769, |
|
"learning_rate": 6.182e-05, |
|
"loss": 0.01, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 2.3998612796948153, |
|
"grad_norm": 0.07981958985328674, |
|
"learning_rate": 6.162e-05, |
|
"loss": 0.0105, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 2.403329287324432, |
|
"grad_norm": 0.064891017973423, |
|
"learning_rate": 6.142e-05, |
|
"loss": 0.0138, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 2.406797294954049, |
|
"grad_norm": 0.07090698927640915, |
|
"learning_rate": 6.122e-05, |
|
"loss": 0.0103, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 2.4102653025836656, |
|
"grad_norm": 0.09227363765239716, |
|
"learning_rate": 6.102e-05, |
|
"loss": 0.0146, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.4137333102132823, |
|
"grad_norm": 0.056207556277513504, |
|
"learning_rate": 6.082e-05, |
|
"loss": 0.0143, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.417201317842899, |
|
"grad_norm": 0.08221688121557236, |
|
"learning_rate": 6.062e-05, |
|
"loss": 0.0158, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 2.4206693254725162, |
|
"grad_norm": 0.08280789107084274, |
|
"learning_rate": 6.042e-05, |
|
"loss": 0.0139, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 2.424137333102133, |
|
"grad_norm": 0.06660090386867523, |
|
"learning_rate": 6.0219999999999996e-05, |
|
"loss": 0.013, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 2.4276053407317497, |
|
"grad_norm": 0.10185576230287552, |
|
"learning_rate": 6.002e-05, |
|
"loss": 0.0153, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.4276053407317497, |
|
"eval_loss": 0.0646032989025116, |
|
"eval_runtime": 713.661, |
|
"eval_samples_per_second": 12.866, |
|
"eval_steps_per_second": 1.609, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.4310733483613665, |
|
"grad_norm": 0.11863771080970764, |
|
"learning_rate": 5.982e-05, |
|
"loss": 0.0149, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 2.434541355990983, |
|
"grad_norm": 0.08282611519098282, |
|
"learning_rate": 5.9619999999999995e-05, |
|
"loss": 0.0114, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.4380093636206, |
|
"grad_norm": 0.10327205806970596, |
|
"learning_rate": 5.942e-05, |
|
"loss": 0.0113, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 2.4414773712502167, |
|
"grad_norm": 0.07424825429916382, |
|
"learning_rate": 5.922e-05, |
|
"loss": 0.0149, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 2.4449453788798334, |
|
"grad_norm": 0.07107970863580704, |
|
"learning_rate": 5.902e-05, |
|
"loss": 0.0133, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.44841338650945, |
|
"grad_norm": 0.08729968965053558, |
|
"learning_rate": 5.8819999999999996e-05, |
|
"loss": 0.0136, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 2.451881394139067, |
|
"grad_norm": 0.06300070136785507, |
|
"learning_rate": 5.862000000000001e-05, |
|
"loss": 0.01, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 2.455349401768684, |
|
"grad_norm": 0.107483871281147, |
|
"learning_rate": 5.8420000000000006e-05, |
|
"loss": 0.0124, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.458817409398301, |
|
"grad_norm": 0.06871318072080612, |
|
"learning_rate": 5.822000000000001e-05, |
|
"loss": 0.013, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 2.4622854170279176, |
|
"grad_norm": 0.10566007345914841, |
|
"learning_rate": 5.802000000000001e-05, |
|
"loss": 0.0133, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.4657534246575343, |
|
"grad_norm": 0.04511050507426262, |
|
"learning_rate": 5.7820000000000005e-05, |
|
"loss": 0.012, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 2.469221432287151, |
|
"grad_norm": 0.09614109992980957, |
|
"learning_rate": 5.762000000000001e-05, |
|
"loss": 0.0115, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 2.472689439916768, |
|
"grad_norm": 0.07863055169582367, |
|
"learning_rate": 5.742000000000001e-05, |
|
"loss": 0.0141, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 2.4761574475463846, |
|
"grad_norm": 0.09857816249132156, |
|
"learning_rate": 5.7220000000000004e-05, |
|
"loss": 0.0131, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.4796254551760013, |
|
"grad_norm": 0.11649773269891739, |
|
"learning_rate": 5.7020000000000006e-05, |
|
"loss": 0.014, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.483093462805618, |
|
"grad_norm": 0.09917131811380386, |
|
"learning_rate": 5.682000000000001e-05, |
|
"loss": 0.0119, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 2.486561470435235, |
|
"grad_norm": 0.090948186814785, |
|
"learning_rate": 5.6620000000000003e-05, |
|
"loss": 0.0131, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 2.4900294780648515, |
|
"grad_norm": 0.10430017858743668, |
|
"learning_rate": 5.6420000000000005e-05, |
|
"loss": 0.0122, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 2.4934974856944687, |
|
"grad_norm": 0.08023589104413986, |
|
"learning_rate": 5.622000000000001e-05, |
|
"loss": 0.0146, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 2.4969654933240855, |
|
"grad_norm": 0.058440957218408585, |
|
"learning_rate": 5.602000000000001e-05, |
|
"loss": 0.0125, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.500433500953702, |
|
"grad_norm": 0.07056768983602524, |
|
"learning_rate": 5.5820000000000004e-05, |
|
"loss": 0.0122, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 2.503901508583319, |
|
"grad_norm": 0.09822002053260803, |
|
"learning_rate": 5.5620000000000006e-05, |
|
"loss": 0.0111, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 2.5073695162129357, |
|
"grad_norm": 0.04473882168531418, |
|
"learning_rate": 5.542000000000001e-05, |
|
"loss": 0.0126, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 2.5108375238425524, |
|
"grad_norm": 0.07200278341770172, |
|
"learning_rate": 5.522e-05, |
|
"loss": 0.0123, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 2.514305531472169, |
|
"grad_norm": 0.054528553038835526, |
|
"learning_rate": 5.5020000000000005e-05, |
|
"loss": 0.0139, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 2.517773539101786, |
|
"grad_norm": 0.06125443056225777, |
|
"learning_rate": 5.482000000000001e-05, |
|
"loss": 0.0136, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.5212415467314027, |
|
"grad_norm": 0.0841405987739563, |
|
"learning_rate": 5.462e-05, |
|
"loss": 0.012, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 2.52470955436102, |
|
"grad_norm": 0.05345413088798523, |
|
"learning_rate": 5.4420000000000004e-05, |
|
"loss": 0.0124, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 2.5281775619906366, |
|
"grad_norm": 0.1266576498746872, |
|
"learning_rate": 5.4220000000000006e-05, |
|
"loss": 0.0115, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 2.5316455696202533, |
|
"grad_norm": 0.06762862950563431, |
|
"learning_rate": 5.402e-05, |
|
"loss": 0.0144, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.53511357724987, |
|
"grad_norm": 0.06965386867523193, |
|
"learning_rate": 5.382e-05, |
|
"loss": 0.0107, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.538581584879487, |
|
"grad_norm": 0.13501428067684174, |
|
"learning_rate": 5.3620000000000005e-05, |
|
"loss": 0.0122, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.5420495925091036, |
|
"grad_norm": 0.0995524674654007, |
|
"learning_rate": 5.342e-05, |
|
"loss": 0.0129, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.5455176001387203, |
|
"grad_norm": 0.10781200975179672, |
|
"learning_rate": 5.322e-05, |
|
"loss": 0.0123, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.548985607768337, |
|
"grad_norm": 0.08666220307350159, |
|
"learning_rate": 5.3020000000000004e-05, |
|
"loss": 0.0128, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.552453615397954, |
|
"grad_norm": 0.0817803218960762, |
|
"learning_rate": 5.2820000000000006e-05, |
|
"loss": 0.0133, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.5559216230275705, |
|
"grad_norm": 0.08346541970968246, |
|
"learning_rate": 5.262e-05, |
|
"loss": 0.0142, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.5593896306571873, |
|
"grad_norm": 0.08062835782766342, |
|
"learning_rate": 5.242e-05, |
|
"loss": 0.0108, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.562857638286804, |
|
"grad_norm": 0.07476343959569931, |
|
"learning_rate": 5.2220000000000005e-05, |
|
"loss": 0.0129, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.5663256459164208, |
|
"grad_norm": 0.0925203412771225, |
|
"learning_rate": 5.202e-05, |
|
"loss": 0.0137, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.569793653546038, |
|
"grad_norm": 0.06427571177482605, |
|
"learning_rate": 5.182e-05, |
|
"loss": 0.012, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.5732616611756547, |
|
"grad_norm": 0.08569345623254776, |
|
"learning_rate": 5.1620000000000004e-05, |
|
"loss": 0.0103, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.5767296688052714, |
|
"grad_norm": 0.09565524011850357, |
|
"learning_rate": 5.142e-05, |
|
"loss": 0.0099, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.580197676434888, |
|
"grad_norm": 0.09814833849668503, |
|
"learning_rate": 5.122e-05, |
|
"loss": 0.0125, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.583665684064505, |
|
"grad_norm": 0.09652476012706757, |
|
"learning_rate": 5.102e-05, |
|
"loss": 0.0139, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.5871336916941217, |
|
"grad_norm": 0.09088584780693054, |
|
"learning_rate": 5.082e-05, |
|
"loss": 0.0129, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.5906016993237384, |
|
"grad_norm": 0.0785548985004425, |
|
"learning_rate": 5.062e-05, |
|
"loss": 0.0119, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.594069706953355, |
|
"grad_norm": 0.0708330050110817, |
|
"learning_rate": 5.042e-05, |
|
"loss": 0.0121, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.5975377145829723, |
|
"grad_norm": 0.09456871449947357, |
|
"learning_rate": 5.0220000000000004e-05, |
|
"loss": 0.0127, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.601005722212589, |
|
"grad_norm": 0.05804060399532318, |
|
"learning_rate": 5.002e-05, |
|
"loss": 0.012, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.604473729842206, |
|
"grad_norm": 0.08347133547067642, |
|
"learning_rate": 4.982e-05, |
|
"loss": 0.0121, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.6079417374718226, |
|
"grad_norm": 0.10097778588533401, |
|
"learning_rate": 4.962e-05, |
|
"loss": 0.0116, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.6114097451014393, |
|
"grad_norm": 0.07667971402406693, |
|
"learning_rate": 4.942e-05, |
|
"loss": 0.0144, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.614877752731056, |
|
"grad_norm": 0.09747060388326645, |
|
"learning_rate": 4.9220000000000006e-05, |
|
"loss": 0.0115, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.618345760360673, |
|
"grad_norm": 0.07479839771986008, |
|
"learning_rate": 4.902e-05, |
|
"loss": 0.0127, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.6218137679902895, |
|
"grad_norm": 0.11268935352563858, |
|
"learning_rate": 4.8820000000000004e-05, |
|
"loss": 0.0143, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.6252817756199063, |
|
"grad_norm": 0.0846814289689064, |
|
"learning_rate": 4.8620000000000005e-05, |
|
"loss": 0.0133, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.628749783249523, |
|
"grad_norm": 0.10621094703674316, |
|
"learning_rate": 4.842000000000001e-05, |
|
"loss": 0.013, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.6322177908791398, |
|
"grad_norm": 0.07603222131729126, |
|
"learning_rate": 4.822e-05, |
|
"loss": 0.0124, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.6356857985087565, |
|
"grad_norm": 0.08150995522737503, |
|
"learning_rate": 4.8020000000000004e-05, |
|
"loss": 0.013, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.6391538061383732, |
|
"grad_norm": 0.08962032943964005, |
|
"learning_rate": 4.7820000000000006e-05, |
|
"loss": 0.0119, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.6426218137679904, |
|
"grad_norm": 0.07465488463640213, |
|
"learning_rate": 4.762e-05, |
|
"loss": 0.0116, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.646089821397607, |
|
"grad_norm": 0.1040426716208458, |
|
"learning_rate": 4.742e-05, |
|
"loss": 0.0115, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.649557829027224, |
|
"grad_norm": 0.14476630091667175, |
|
"learning_rate": 4.7220000000000005e-05, |
|
"loss": 0.0114, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.6530258366568407, |
|
"grad_norm": 0.07727625221014023, |
|
"learning_rate": 4.702e-05, |
|
"loss": 0.0122, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.6564938442864574, |
|
"grad_norm": 0.0987255796790123, |
|
"learning_rate": 4.682e-05, |
|
"loss": 0.0132, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.659961851916074, |
|
"grad_norm": 0.07335236668586731, |
|
"learning_rate": 4.6620000000000004e-05, |
|
"loss": 0.0126, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.663429859545691, |
|
"grad_norm": 0.09524571150541306, |
|
"learning_rate": 4.642e-05, |
|
"loss": 0.0114, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.6668978671753076, |
|
"grad_norm": 0.11405625939369202, |
|
"learning_rate": 4.622e-05, |
|
"loss": 0.0144, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.670365874804925, |
|
"grad_norm": 0.09738898277282715, |
|
"learning_rate": 4.602e-05, |
|
"loss": 0.0139, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.6738338824345416, |
|
"grad_norm": 0.08633222430944443, |
|
"learning_rate": 4.5820000000000005e-05, |
|
"loss": 0.0136, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.6773018900641583, |
|
"grad_norm": 0.09597263485193253, |
|
"learning_rate": 4.562e-05, |
|
"loss": 0.015, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.680769897693775, |
|
"grad_norm": 0.0945490151643753, |
|
"learning_rate": 4.542e-05, |
|
"loss": 0.012, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.684237905323392, |
|
"grad_norm": 0.0953177884221077, |
|
"learning_rate": 4.5220000000000004e-05, |
|
"loss": 0.0121, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.6877059129530085, |
|
"grad_norm": 0.09284798800945282, |
|
"learning_rate": 4.502e-05, |
|
"loss": 0.0128, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.6911739205826253, |
|
"grad_norm": 0.10929513722658157, |
|
"learning_rate": 4.482e-05, |
|
"loss": 0.0122, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.694641928212242, |
|
"grad_norm": 0.10034655779600143, |
|
"learning_rate": 4.462e-05, |
|
"loss": 0.0137, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.6981099358418588, |
|
"grad_norm": 0.06497249752283096, |
|
"learning_rate": 4.442e-05, |
|
"loss": 0.0116, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.7015779434714755, |
|
"grad_norm": 0.061711184680461884, |
|
"learning_rate": 4.422e-05, |
|
"loss": 0.0118, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.7050459511010922, |
|
"grad_norm": 0.08474565297365189, |
|
"learning_rate": 4.402e-05, |
|
"loss": 0.0113, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.708513958730709, |
|
"grad_norm": 0.06698320806026459, |
|
"learning_rate": 4.382e-05, |
|
"loss": 0.0116, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.7119819663603257, |
|
"grad_norm": 0.11040494590997696, |
|
"learning_rate": 4.362e-05, |
|
"loss": 0.014, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.715449973989943, |
|
"grad_norm": 0.077610544860363, |
|
"learning_rate": 4.342e-05, |
|
"loss": 0.0148, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.7189179816195597, |
|
"grad_norm": 0.07504422217607498, |
|
"learning_rate": 4.3219999999999996e-05, |
|
"loss": 0.0109, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.7223859892491764, |
|
"grad_norm": 0.083879753947258, |
|
"learning_rate": 4.3020000000000005e-05, |
|
"loss": 0.0136, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.725853996878793, |
|
"grad_norm": 0.07689789682626724, |
|
"learning_rate": 4.282000000000001e-05, |
|
"loss": 0.013, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.72932200450841, |
|
"grad_norm": 0.07872592657804489, |
|
"learning_rate": 4.262e-05, |
|
"loss": 0.0123, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.7327900121380266, |
|
"grad_norm": 0.10761665552854538, |
|
"learning_rate": 4.2420000000000004e-05, |
|
"loss": 0.0141, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.7362580197676434, |
|
"grad_norm": 0.07873303443193436, |
|
"learning_rate": 4.2220000000000006e-05, |
|
"loss": 0.0114, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.73972602739726, |
|
"grad_norm": 0.05685030296444893, |
|
"learning_rate": 4.202e-05, |
|
"loss": 0.0102, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.7431940350268773, |
|
"grad_norm": 0.10359474271535873, |
|
"learning_rate": 4.182e-05, |
|
"loss": 0.0131, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.746662042656494, |
|
"grad_norm": 0.07630922645330429, |
|
"learning_rate": 4.1620000000000005e-05, |
|
"loss": 0.0115, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.750130050286111, |
|
"grad_norm": 0.09103643894195557, |
|
"learning_rate": 4.142000000000001e-05, |
|
"loss": 0.012, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.7535980579157275, |
|
"grad_norm": 0.08436847478151321, |
|
"learning_rate": 4.122e-05, |
|
"loss": 0.0115, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.7570660655453443, |
|
"grad_norm": 0.09147916734218597, |
|
"learning_rate": 4.1020000000000004e-05, |
|
"loss": 0.0164, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.760534073174961, |
|
"grad_norm": 0.1505836397409439, |
|
"learning_rate": 4.0820000000000006e-05, |
|
"loss": 0.0136, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.7640020808045778, |
|
"grad_norm": 0.06953336298465729, |
|
"learning_rate": 4.062e-05, |
|
"loss": 0.0114, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.7674700884341945, |
|
"grad_norm": 0.08808320015668869, |
|
"learning_rate": 4.042e-05, |
|
"loss": 0.0095, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.7709380960638113, |
|
"grad_norm": 0.10280350595712662, |
|
"learning_rate": 4.0220000000000005e-05, |
|
"loss": 0.0095, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.774406103693428, |
|
"grad_norm": 0.11173941940069199, |
|
"learning_rate": 4.002e-05, |
|
"loss": 0.0128, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.774406103693428, |
|
"eval_loss": 0.06569719314575195, |
|
"eval_runtime": 714.8383, |
|
"eval_samples_per_second": 12.845, |
|
"eval_steps_per_second": 1.606, |
|
"step": 8000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 10000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6.09781455160793e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|