{ "best_metric": 0.23256993293762207, "best_model_checkpoint": "finetuned-fake-food/checkpoint-6700", "epoch": 4.547751389590703, "eval_steps": 100, "global_step": 9000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005053057099545225, "grad_norm": 1.6020498275756836, "learning_rate": 0.0001997777777777778, "loss": 0.6467, "step": 10 }, { "epoch": 0.01010611419909045, "grad_norm": 1.8525471687316895, "learning_rate": 0.00019955555555555558, "loss": 0.6495, "step": 20 }, { "epoch": 0.015159171298635674, "grad_norm": 1.8445143699645996, "learning_rate": 0.00019933333333333334, "loss": 0.5474, "step": 30 }, { "epoch": 0.0202122283981809, "grad_norm": 4.539891242980957, "learning_rate": 0.00019911111111111111, "loss": 0.6233, "step": 40 }, { "epoch": 0.025265285497726123, "grad_norm": 0.8423069715499878, "learning_rate": 0.0001988888888888889, "loss": 0.6967, "step": 50 }, { "epoch": 0.03031834259727135, "grad_norm": 1.808174967765808, "learning_rate": 0.00019866666666666668, "loss": 0.6394, "step": 60 }, { "epoch": 0.035371399696816574, "grad_norm": 2.431649684906006, "learning_rate": 0.00019844444444444445, "loss": 0.5102, "step": 70 }, { "epoch": 0.0404244567963618, "grad_norm": 3.6550588607788086, "learning_rate": 0.00019822222222222225, "loss": 0.605, "step": 80 }, { "epoch": 0.045477513895907026, "grad_norm": 6.30785608291626, "learning_rate": 0.00019800000000000002, "loss": 0.5917, "step": 90 }, { "epoch": 0.050530570995452245, "grad_norm": 0.64481121301651, "learning_rate": 0.00019777777777777778, "loss": 0.5991, "step": 100 }, { "epoch": 0.050530570995452245, "eval_accuracy": 0.7028284998209811, "eval_loss": 0.6128867864608765, "eval_runtime": 26.3295, "eval_samples_per_second": 106.079, "eval_steps_per_second": 13.293, "step": 100 }, { "epoch": 0.05558362809499747, "grad_norm": 0.5892785787582397, "learning_rate": 0.00019755555555555555, "loss": 0.5369, "step": 110 }, { "epoch": 0.0606366851945427, "grad_norm": 1.9678053855895996, "learning_rate": 0.00019733333333333335, "loss": 0.6281, "step": 120 }, { "epoch": 0.06568974229408793, "grad_norm": 5.071497917175293, "learning_rate": 0.00019711111111111112, "loss": 0.5774, "step": 130 }, { "epoch": 0.07074279939363315, "grad_norm": 2.3490710258483887, "learning_rate": 0.0001968888888888889, "loss": 0.5453, "step": 140 }, { "epoch": 0.07579585649317837, "grad_norm": 2.5878753662109375, "learning_rate": 0.00019666666666666666, "loss": 0.444, "step": 150 }, { "epoch": 0.0808489135927236, "grad_norm": 5.393430233001709, "learning_rate": 0.00019644444444444445, "loss": 0.6578, "step": 160 }, { "epoch": 0.08590197069226882, "grad_norm": 3.0397963523864746, "learning_rate": 0.00019622222222222225, "loss": 0.6167, "step": 170 }, { "epoch": 0.09095502779181405, "grad_norm": 1.22067129611969, "learning_rate": 0.000196, "loss": 0.5448, "step": 180 }, { "epoch": 0.09600808489135927, "grad_norm": 1.5877925157546997, "learning_rate": 0.0001957777777777778, "loss": 0.3992, "step": 190 }, { "epoch": 0.10106114199090449, "grad_norm": 6.1517815589904785, "learning_rate": 0.00019555555555555556, "loss": 0.6593, "step": 200 }, { "epoch": 0.10106114199090449, "eval_accuracy": 0.8363766559255281, "eval_loss": 0.43375319242477417, "eval_runtime": 23.7559, "eval_samples_per_second": 117.571, "eval_steps_per_second": 14.733, "step": 200 }, { "epoch": 0.10611419909044972, "grad_norm": 1.4394699335098267, "learning_rate": 0.00019533333333333336, "loss": 0.484, "step": 210 }, { "epoch": 0.11116725618999494, "grad_norm": 3.380403995513916, "learning_rate": 0.0001951111111111111, "loss": 0.5274, "step": 220 }, { "epoch": 0.11622031328954018, "grad_norm": 5.558537483215332, "learning_rate": 0.0001948888888888889, "loss": 0.5731, "step": 230 }, { "epoch": 0.1212733703890854, "grad_norm": 3.260127305984497, "learning_rate": 0.0001946666666666667, "loss": 0.6198, "step": 240 }, { "epoch": 0.12632642748863063, "grad_norm": 1.4037119150161743, "learning_rate": 0.00019444444444444446, "loss": 0.4793, "step": 250 }, { "epoch": 0.13137948458817586, "grad_norm": 2.3021624088287354, "learning_rate": 0.00019422222222222223, "loss": 0.5727, "step": 260 }, { "epoch": 0.13643254168772107, "grad_norm": 3.1738462448120117, "learning_rate": 0.000194, "loss": 0.3962, "step": 270 }, { "epoch": 0.1414855987872663, "grad_norm": 0.7951613068580627, "learning_rate": 0.0001937777777777778, "loss": 0.8582, "step": 280 }, { "epoch": 0.14653865588681153, "grad_norm": 1.5663028955459595, "learning_rate": 0.00019355555555555557, "loss": 0.5187, "step": 290 }, { "epoch": 0.15159171298635674, "grad_norm": 3.4048633575439453, "learning_rate": 0.00019333333333333333, "loss": 0.4908, "step": 300 }, { "epoch": 0.15159171298635674, "eval_accuracy": 0.8098818474758325, "eval_loss": 0.4489921033382416, "eval_runtime": 22.7942, "eval_samples_per_second": 122.531, "eval_steps_per_second": 15.355, "step": 300 }, { "epoch": 0.15664477008590197, "grad_norm": 2.1454992294311523, "learning_rate": 0.0001931111111111111, "loss": 0.4886, "step": 310 }, { "epoch": 0.1616978271854472, "grad_norm": 7.062505722045898, "learning_rate": 0.0001928888888888889, "loss": 0.6002, "step": 320 }, { "epoch": 0.1667508842849924, "grad_norm": 2.0749993324279785, "learning_rate": 0.0001926666666666667, "loss": 0.5287, "step": 330 }, { "epoch": 0.17180394138453764, "grad_norm": 1.076101541519165, "learning_rate": 0.00019244444444444444, "loss": 0.4165, "step": 340 }, { "epoch": 0.17685699848408287, "grad_norm": 2.960982322692871, "learning_rate": 0.00019222222222222224, "loss": 0.5904, "step": 350 }, { "epoch": 0.1819100555836281, "grad_norm": 1.387323021888733, "learning_rate": 0.000192, "loss": 0.5503, "step": 360 }, { "epoch": 0.1869631126831733, "grad_norm": 1.7477341890335083, "learning_rate": 0.0001917777777777778, "loss": 0.6334, "step": 370 }, { "epoch": 0.19201616978271854, "grad_norm": 1.803114652633667, "learning_rate": 0.00019155555555555554, "loss": 0.3982, "step": 380 }, { "epoch": 0.19706922688226378, "grad_norm": 8.091989517211914, "learning_rate": 0.00019133333333333334, "loss": 0.457, "step": 390 }, { "epoch": 0.20212228398180898, "grad_norm": 2.163998603820801, "learning_rate": 0.00019111111111111114, "loss": 0.4756, "step": 400 }, { "epoch": 0.20212228398180898, "eval_accuracy": 0.7003222341568206, "eval_loss": 0.7639068961143494, "eval_runtime": 23.9048, "eval_samples_per_second": 116.838, "eval_steps_per_second": 14.641, "step": 400 }, { "epoch": 0.20717534108135421, "grad_norm": 1.778550148010254, "learning_rate": 0.0001908888888888889, "loss": 0.603, "step": 410 }, { "epoch": 0.21222839818089945, "grad_norm": 1.4883553981781006, "learning_rate": 0.00019066666666666668, "loss": 0.5976, "step": 420 }, { "epoch": 0.21728145528044468, "grad_norm": 0.9844093918800354, "learning_rate": 0.00019044444444444444, "loss": 0.5234, "step": 430 }, { "epoch": 0.22233451237998988, "grad_norm": 0.9249500632286072, "learning_rate": 0.00019022222222222224, "loss": 0.4766, "step": 440 }, { "epoch": 0.22738756947953512, "grad_norm": 2.516437530517578, "learning_rate": 0.00019, "loss": 0.532, "step": 450 }, { "epoch": 0.23244062657908035, "grad_norm": 1.32094144821167, "learning_rate": 0.00018977777777777778, "loss": 0.62, "step": 460 }, { "epoch": 0.23749368367862556, "grad_norm": 5.317640781402588, "learning_rate": 0.00018955555555555558, "loss": 0.5053, "step": 470 }, { "epoch": 0.2425467407781708, "grad_norm": 0.9952796101570129, "learning_rate": 0.00018933333333333335, "loss": 0.3515, "step": 480 }, { "epoch": 0.24759979787771602, "grad_norm": 5.339252948760986, "learning_rate": 0.00018911111111111112, "loss": 0.7594, "step": 490 }, { "epoch": 0.25265285497726125, "grad_norm": 1.5515260696411133, "learning_rate": 0.00018888888888888888, "loss": 0.547, "step": 500 }, { "epoch": 0.25265285497726125, "eval_accuracy": 0.8335123523093448, "eval_loss": 0.42532363533973694, "eval_runtime": 22.6359, "eval_samples_per_second": 123.388, "eval_steps_per_second": 15.462, "step": 500 }, { "epoch": 0.2577059120768065, "grad_norm": 1.3761298656463623, "learning_rate": 0.00018866666666666668, "loss": 0.4438, "step": 510 }, { "epoch": 0.2627589691763517, "grad_norm": 4.520216464996338, "learning_rate": 0.00018844444444444445, "loss": 0.5003, "step": 520 }, { "epoch": 0.2678120262758969, "grad_norm": 1.277355670928955, "learning_rate": 0.00018822222222222222, "loss": 0.472, "step": 530 }, { "epoch": 0.27286508337544213, "grad_norm": 0.9009717106819153, "learning_rate": 0.000188, "loss": 0.5756, "step": 540 }, { "epoch": 0.27791814047498736, "grad_norm": 1.265140414237976, "learning_rate": 0.00018777777777777779, "loss": 0.5466, "step": 550 }, { "epoch": 0.2829711975745326, "grad_norm": 0.7211174368858337, "learning_rate": 0.00018755555555555558, "loss": 0.4687, "step": 560 }, { "epoch": 0.28802425467407783, "grad_norm": 0.7686963081359863, "learning_rate": 0.00018733333333333335, "loss": 0.571, "step": 570 }, { "epoch": 0.29307731177362306, "grad_norm": 1.9314404726028442, "learning_rate": 0.00018711111111111112, "loss": 0.532, "step": 580 }, { "epoch": 0.2981303688731683, "grad_norm": 2.081638813018799, "learning_rate": 0.0001868888888888889, "loss": 0.5143, "step": 590 }, { "epoch": 0.30318342597271347, "grad_norm": 3.3979947566986084, "learning_rate": 0.0001866666666666667, "loss": 0.4702, "step": 600 }, { "epoch": 0.30318342597271347, "eval_accuracy": 0.8446115288220551, "eval_loss": 0.38639551401138306, "eval_runtime": 23.3543, "eval_samples_per_second": 119.593, "eval_steps_per_second": 14.987, "step": 600 }, { "epoch": 0.3082364830722587, "grad_norm": 1.4857347011566162, "learning_rate": 0.00018644444444444446, "loss": 0.5053, "step": 610 }, { "epoch": 0.31328954017180394, "grad_norm": 0.7065760493278503, "learning_rate": 0.00018622222222222223, "loss": 0.4728, "step": 620 }, { "epoch": 0.31834259727134917, "grad_norm": 2.9684479236602783, "learning_rate": 0.00018600000000000002, "loss": 0.4999, "step": 630 }, { "epoch": 0.3233956543708944, "grad_norm": 2.780421018600464, "learning_rate": 0.0001857777777777778, "loss": 0.3192, "step": 640 }, { "epoch": 0.32844871147043964, "grad_norm": 0.8906468749046326, "learning_rate": 0.00018555555555555556, "loss": 0.7761, "step": 650 }, { "epoch": 0.3335017685699848, "grad_norm": 2.108656644821167, "learning_rate": 0.00018533333333333333, "loss": 0.6522, "step": 660 }, { "epoch": 0.33855482566953005, "grad_norm": 2.96596622467041, "learning_rate": 0.00018511111111111113, "loss": 0.4721, "step": 670 }, { "epoch": 0.3436078827690753, "grad_norm": 2.0396857261657715, "learning_rate": 0.0001848888888888889, "loss": 0.4571, "step": 680 }, { "epoch": 0.3486609398686205, "grad_norm": 2.564168930053711, "learning_rate": 0.00018466666666666666, "loss": 0.4676, "step": 690 }, { "epoch": 0.35371399696816574, "grad_norm": 2.480363607406616, "learning_rate": 0.00018444444444444446, "loss": 0.5099, "step": 700 }, { "epoch": 0.35371399696816574, "eval_accuracy": 0.7755102040816326, "eval_loss": 0.48194998502731323, "eval_runtime": 21.4702, "eval_samples_per_second": 130.088, "eval_steps_per_second": 16.302, "step": 700 }, { "epoch": 0.358767054067711, "grad_norm": 2.140857696533203, "learning_rate": 0.00018422222222222223, "loss": 0.5115, "step": 710 }, { "epoch": 0.3638201111672562, "grad_norm": 2.32558012008667, "learning_rate": 0.00018400000000000003, "loss": 0.4627, "step": 720 }, { "epoch": 0.3688731682668014, "grad_norm": 3.254412889480591, "learning_rate": 0.00018377777777777777, "loss": 0.4797, "step": 730 }, { "epoch": 0.3739262253663466, "grad_norm": 1.9636939764022827, "learning_rate": 0.00018355555555555557, "loss": 0.5634, "step": 740 }, { "epoch": 0.37897928246589185, "grad_norm": 1.8012185096740723, "learning_rate": 0.00018333333333333334, "loss": 0.4575, "step": 750 }, { "epoch": 0.3840323395654371, "grad_norm": 5.658529758453369, "learning_rate": 0.00018311111111111113, "loss": 0.5222, "step": 760 }, { "epoch": 0.3890853966649823, "grad_norm": 4.113037109375, "learning_rate": 0.00018288888888888887, "loss": 0.4088, "step": 770 }, { "epoch": 0.39413845376452755, "grad_norm": 4.239760398864746, "learning_rate": 0.00018266666666666667, "loss": 0.425, "step": 780 }, { "epoch": 0.3991915108640728, "grad_norm": 5.495861053466797, "learning_rate": 0.00018244444444444447, "loss": 0.4213, "step": 790 }, { "epoch": 0.40424456796361796, "grad_norm": 2.3264899253845215, "learning_rate": 0.00018222222222222224, "loss": 0.5484, "step": 800 }, { "epoch": 0.40424456796361796, "eval_accuracy": 0.8263515932688865, "eval_loss": 0.3940279483795166, "eval_runtime": 22.3098, "eval_samples_per_second": 125.191, "eval_steps_per_second": 15.688, "step": 800 }, { "epoch": 0.4092976250631632, "grad_norm": 3.747011184692383, "learning_rate": 0.000182, "loss": 0.3838, "step": 810 }, { "epoch": 0.41435068216270843, "grad_norm": 4.610471248626709, "learning_rate": 0.00018177777777777778, "loss": 0.5677, "step": 820 }, { "epoch": 0.41940373926225366, "grad_norm": 1.2806724309921265, "learning_rate": 0.00018155555555555557, "loss": 0.5668, "step": 830 }, { "epoch": 0.4244567963617989, "grad_norm": 2.143632650375366, "learning_rate": 0.00018133333333333334, "loss": 0.4984, "step": 840 }, { "epoch": 0.4295098534613441, "grad_norm": 2.1807491779327393, "learning_rate": 0.0001811111111111111, "loss": 0.4222, "step": 850 }, { "epoch": 0.43456291056088936, "grad_norm": 3.1031441688537598, "learning_rate": 0.0001808888888888889, "loss": 0.4854, "step": 860 }, { "epoch": 0.43961596766043454, "grad_norm": 1.3116328716278076, "learning_rate": 0.00018066666666666668, "loss": 0.414, "step": 870 }, { "epoch": 0.44466902475997977, "grad_norm": 2.7630324363708496, "learning_rate": 0.00018044444444444447, "loss": 0.4963, "step": 880 }, { "epoch": 0.449722081859525, "grad_norm": 1.7805945873260498, "learning_rate": 0.00018022222222222221, "loss": 0.5202, "step": 890 }, { "epoch": 0.45477513895907024, "grad_norm": 2.960754632949829, "learning_rate": 0.00018, "loss": 0.6263, "step": 900 }, { "epoch": 0.45477513895907024, "eval_accuracy": 0.7117794486215538, "eval_loss": 0.6219449043273926, "eval_runtime": 22.1628, "eval_samples_per_second": 126.022, "eval_steps_per_second": 15.792, "step": 900 }, { "epoch": 0.45982819605861547, "grad_norm": 1.2690253257751465, "learning_rate": 0.00017977777777777778, "loss": 0.5016, "step": 910 }, { "epoch": 0.4648812531581607, "grad_norm": 2.105319023132324, "learning_rate": 0.00017955555555555558, "loss": 0.455, "step": 920 }, { "epoch": 0.46993431025770593, "grad_norm": 0.5122328400611877, "learning_rate": 0.00017933333333333332, "loss": 0.5045, "step": 930 }, { "epoch": 0.4749873673572511, "grad_norm": 1.700193166732788, "learning_rate": 0.00017911111111111112, "loss": 0.568, "step": 940 }, { "epoch": 0.48004042445679634, "grad_norm": 1.9073634147644043, "learning_rate": 0.0001788888888888889, "loss": 0.4445, "step": 950 }, { "epoch": 0.4850934815563416, "grad_norm": 2.1367106437683105, "learning_rate": 0.00017866666666666668, "loss": 0.5335, "step": 960 }, { "epoch": 0.4901465386558868, "grad_norm": 2.9629406929016113, "learning_rate": 0.00017844444444444445, "loss": 0.4809, "step": 970 }, { "epoch": 0.49519959575543204, "grad_norm": 3.2274210453033447, "learning_rate": 0.00017822222222222222, "loss": 0.4762, "step": 980 }, { "epoch": 0.5002526528549772, "grad_norm": 2.4677717685699463, "learning_rate": 0.00017800000000000002, "loss": 0.5479, "step": 990 }, { "epoch": 0.5053057099545225, "grad_norm": 1.7024959325790405, "learning_rate": 0.00017777777777777779, "loss": 0.5453, "step": 1000 }, { "epoch": 0.5053057099545225, "eval_accuracy": 0.7887576083064805, "eval_loss": 0.454845130443573, "eval_runtime": 22.2317, "eval_samples_per_second": 125.631, "eval_steps_per_second": 15.743, "step": 1000 }, { "epoch": 0.5103587670540677, "grad_norm": 2.923689126968384, "learning_rate": 0.00017755555555555556, "loss": 0.5531, "step": 1010 }, { "epoch": 0.515411824153613, "grad_norm": 0.9824873208999634, "learning_rate": 0.00017733333333333335, "loss": 0.4836, "step": 1020 }, { "epoch": 0.5204648812531582, "grad_norm": 2.362745523452759, "learning_rate": 0.00017711111111111112, "loss": 0.4541, "step": 1030 }, { "epoch": 0.5255179383527034, "grad_norm": 0.6012383699417114, "learning_rate": 0.0001768888888888889, "loss": 0.3289, "step": 1040 }, { "epoch": 0.5305709954522486, "grad_norm": 3.0225465297698975, "learning_rate": 0.00017666666666666666, "loss": 0.4928, "step": 1050 }, { "epoch": 0.5356240525517938, "grad_norm": 1.9450099468231201, "learning_rate": 0.00017644444444444446, "loss": 0.5408, "step": 1060 }, { "epoch": 0.5406771096513391, "grad_norm": 1.4540809392929077, "learning_rate": 0.00017622222222222223, "loss": 0.4319, "step": 1070 }, { "epoch": 0.5457301667508843, "grad_norm": 1.9152274131774902, "learning_rate": 0.00017600000000000002, "loss": 0.4556, "step": 1080 }, { "epoch": 0.5507832238504295, "grad_norm": 4.454532146453857, "learning_rate": 0.0001757777777777778, "loss": 0.4623, "step": 1090 }, { "epoch": 0.5558362809499747, "grad_norm": 2.6191518306732178, "learning_rate": 0.00017555555555555556, "loss": 0.5431, "step": 1100 }, { "epoch": 0.5558362809499747, "eval_accuracy": 0.8084496956677408, "eval_loss": 0.4210474193096161, "eval_runtime": 20.9097, "eval_samples_per_second": 133.574, "eval_steps_per_second": 16.739, "step": 1100 }, { "epoch": 0.56088933804952, "grad_norm": 3.4355955123901367, "learning_rate": 0.00017533333333333336, "loss": 0.6591, "step": 1110 }, { "epoch": 0.5659423951490652, "grad_norm": 1.8346270322799683, "learning_rate": 0.00017511111111111113, "loss": 0.4865, "step": 1120 }, { "epoch": 0.5709954522486104, "grad_norm": 0.6583804488182068, "learning_rate": 0.0001748888888888889, "loss": 0.4366, "step": 1130 }, { "epoch": 0.5760485093481557, "grad_norm": 2.2707302570343018, "learning_rate": 0.00017466666666666667, "loss": 0.5579, "step": 1140 }, { "epoch": 0.5811015664477008, "grad_norm": 3.1430397033691406, "learning_rate": 0.00017444444444444446, "loss": 0.622, "step": 1150 }, { "epoch": 0.5861546235472461, "grad_norm": 1.2443580627441406, "learning_rate": 0.00017422222222222223, "loss": 0.583, "step": 1160 }, { "epoch": 0.5912076806467913, "grad_norm": 0.9249362349510193, "learning_rate": 0.000174, "loss": 0.5084, "step": 1170 }, { "epoch": 0.5962607377463366, "grad_norm": 1.8132987022399902, "learning_rate": 0.0001737777777777778, "loss": 0.4207, "step": 1180 }, { "epoch": 0.6013137948458818, "grad_norm": 1.5119869709014893, "learning_rate": 0.00017355555555555557, "loss": 0.436, "step": 1190 }, { "epoch": 0.6063668519454269, "grad_norm": 4.935205936431885, "learning_rate": 0.00017333333333333334, "loss": 0.5678, "step": 1200 }, { "epoch": 0.6063668519454269, "eval_accuracy": 0.8037952022914429, "eval_loss": 0.4945639967918396, "eval_runtime": 22.3966, "eval_samples_per_second": 124.706, "eval_steps_per_second": 15.627, "step": 1200 }, { "epoch": 0.6114199090449722, "grad_norm": 3.2047390937805176, "learning_rate": 0.0001731111111111111, "loss": 0.3757, "step": 1210 }, { "epoch": 0.6164729661445174, "grad_norm": 1.1126270294189453, "learning_rate": 0.0001728888888888889, "loss": 0.4233, "step": 1220 }, { "epoch": 0.6215260232440627, "grad_norm": 2.307159185409546, "learning_rate": 0.00017266666666666667, "loss": 0.697, "step": 1230 }, { "epoch": 0.6265790803436079, "grad_norm": 1.7499977350234985, "learning_rate": 0.00017244444444444444, "loss": 0.3902, "step": 1240 }, { "epoch": 0.631632137443153, "grad_norm": 1.4569963216781616, "learning_rate": 0.00017222222222222224, "loss": 0.4809, "step": 1250 }, { "epoch": 0.6366851945426983, "grad_norm": 2.1392791271209717, "learning_rate": 0.000172, "loss": 0.4767, "step": 1260 }, { "epoch": 0.6417382516422435, "grad_norm": 1.7471693754196167, "learning_rate": 0.0001717777777777778, "loss": 0.4493, "step": 1270 }, { "epoch": 0.6467913087417888, "grad_norm": 1.5734494924545288, "learning_rate": 0.00017155555555555555, "loss": 0.5824, "step": 1280 }, { "epoch": 0.651844365841334, "grad_norm": 0.9970535635948181, "learning_rate": 0.00017133333333333334, "loss": 0.3455, "step": 1290 }, { "epoch": 0.6568974229408793, "grad_norm": 4.512068271636963, "learning_rate": 0.0001711111111111111, "loss": 0.3266, "step": 1300 }, { "epoch": 0.6568974229408793, "eval_accuracy": 0.8263515932688865, "eval_loss": 0.4538181722164154, "eval_runtime": 22.3661, "eval_samples_per_second": 124.877, "eval_steps_per_second": 15.649, "step": 1300 }, { "epoch": 0.6619504800404244, "grad_norm": 3.7386505603790283, "learning_rate": 0.0001708888888888889, "loss": 0.6928, "step": 1310 }, { "epoch": 0.6670035371399696, "grad_norm": 1.3968713283538818, "learning_rate": 0.00017066666666666668, "loss": 0.4317, "step": 1320 }, { "epoch": 0.6720565942395149, "grad_norm": 2.0303447246551514, "learning_rate": 0.00017044444444444445, "loss": 0.4526, "step": 1330 }, { "epoch": 0.6771096513390601, "grad_norm": 1.145485281944275, "learning_rate": 0.00017022222222222224, "loss": 0.5742, "step": 1340 }, { "epoch": 0.6821627084386054, "grad_norm": 1.014075756072998, "learning_rate": 0.00017, "loss": 0.4812, "step": 1350 }, { "epoch": 0.6872157655381506, "grad_norm": 0.7424379587173462, "learning_rate": 0.00016977777777777778, "loss": 0.3391, "step": 1360 }, { "epoch": 0.6922688226376958, "grad_norm": 1.990348219871521, "learning_rate": 0.00016955555555555555, "loss": 0.477, "step": 1370 }, { "epoch": 0.697321879737241, "grad_norm": 2.611443281173706, "learning_rate": 0.00016933333333333335, "loss": 0.436, "step": 1380 }, { "epoch": 0.7023749368367862, "grad_norm": 1.5179691314697266, "learning_rate": 0.00016911111111111112, "loss": 0.6461, "step": 1390 }, { "epoch": 0.7074279939363315, "grad_norm": 1.9120951890945435, "learning_rate": 0.00016888888888888889, "loss": 0.4225, "step": 1400 }, { "epoch": 0.7074279939363315, "eval_accuracy": 0.8088077336197637, "eval_loss": 0.4366118609905243, "eval_runtime": 22.4189, "eval_samples_per_second": 124.582, "eval_steps_per_second": 15.612, "step": 1400 }, { "epoch": 0.7124810510358767, "grad_norm": 2.0365612506866455, "learning_rate": 0.00016866666666666668, "loss": 0.4125, "step": 1410 }, { "epoch": 0.717534108135422, "grad_norm": 1.036264181137085, "learning_rate": 0.00016844444444444445, "loss": 0.3497, "step": 1420 }, { "epoch": 0.7225871652349671, "grad_norm": 4.05681037902832, "learning_rate": 0.00016822222222222225, "loss": 0.435, "step": 1430 }, { "epoch": 0.7276402223345124, "grad_norm": 1.6648637056350708, "learning_rate": 0.000168, "loss": 0.398, "step": 1440 }, { "epoch": 0.7326932794340576, "grad_norm": 3.195124864578247, "learning_rate": 0.0001677777777777778, "loss": 0.6537, "step": 1450 }, { "epoch": 0.7377463365336028, "grad_norm": 1.0564615726470947, "learning_rate": 0.00016755555555555556, "loss": 0.3518, "step": 1460 }, { "epoch": 0.7427993936331481, "grad_norm": 1.3577885627746582, "learning_rate": 0.00016733333333333335, "loss": 0.3378, "step": 1470 }, { "epoch": 0.7478524507326932, "grad_norm": 2.538241147994995, "learning_rate": 0.00016711111111111112, "loss": 0.4399, "step": 1480 }, { "epoch": 0.7529055078322385, "grad_norm": 2.032485246658325, "learning_rate": 0.0001668888888888889, "loss": 0.3973, "step": 1490 }, { "epoch": 0.7579585649317837, "grad_norm": 0.5845909714698792, "learning_rate": 0.0001666666666666667, "loss": 0.32, "step": 1500 }, { "epoch": 0.7579585649317837, "eval_accuracy": 0.7883995703544576, "eval_loss": 0.5585851669311523, "eval_runtime": 21.3123, "eval_samples_per_second": 131.051, "eval_steps_per_second": 16.422, "step": 1500 }, { "epoch": 0.763011622031329, "grad_norm": 1.060084342956543, "learning_rate": 0.00016644444444444446, "loss": 0.4507, "step": 1510 }, { "epoch": 0.7680646791308742, "grad_norm": 3.0685956478118896, "learning_rate": 0.00016622222222222223, "loss": 0.4476, "step": 1520 }, { "epoch": 0.7731177362304194, "grad_norm": 1.285022258758545, "learning_rate": 0.000166, "loss": 0.4391, "step": 1530 }, { "epoch": 0.7781707933299646, "grad_norm": 0.9301333427429199, "learning_rate": 0.0001657777777777778, "loss": 0.5981, "step": 1540 }, { "epoch": 0.7832238504295098, "grad_norm": 1.7216811180114746, "learning_rate": 0.00016555555555555556, "loss": 0.4924, "step": 1550 }, { "epoch": 0.7882769075290551, "grad_norm": 3.7065885066986084, "learning_rate": 0.00016533333333333333, "loss": 0.4895, "step": 1560 }, { "epoch": 0.7933299646286003, "grad_norm": 1.7684383392333984, "learning_rate": 0.00016511111111111113, "loss": 0.5223, "step": 1570 }, { "epoch": 0.7983830217281456, "grad_norm": 2.3638968467712402, "learning_rate": 0.0001648888888888889, "loss": 0.4228, "step": 1580 }, { "epoch": 0.8034360788276907, "grad_norm": 2.4584741592407227, "learning_rate": 0.00016466666666666667, "loss": 0.5006, "step": 1590 }, { "epoch": 0.8084891359272359, "grad_norm": 1.8796658515930176, "learning_rate": 0.00016444444444444444, "loss": 0.473, "step": 1600 }, { "epoch": 0.8084891359272359, "eval_accuracy": 0.7973505191550304, "eval_loss": 0.48050203919410706, "eval_runtime": 21.4795, "eval_samples_per_second": 130.031, "eval_steps_per_second": 16.295, "step": 1600 }, { "epoch": 0.8135421930267812, "grad_norm": 2.43819260597229, "learning_rate": 0.00016422222222222223, "loss": 0.4502, "step": 1610 }, { "epoch": 0.8185952501263264, "grad_norm": 0.9898656606674194, "learning_rate": 0.000164, "loss": 0.4386, "step": 1620 }, { "epoch": 0.8236483072258717, "grad_norm": 3.075253963470459, "learning_rate": 0.0001637777777777778, "loss": 0.4411, "step": 1630 }, { "epoch": 0.8287013643254169, "grad_norm": 1.5815861225128174, "learning_rate": 0.0001635777777777778, "loss": 0.5317, "step": 1640 }, { "epoch": 0.8337544214249621, "grad_norm": 2.222179889678955, "learning_rate": 0.00016335555555555556, "loss": 0.3035, "step": 1650 }, { "epoch": 0.8388074785245073, "grad_norm": 1.270073652267456, "learning_rate": 0.00016313333333333333, "loss": 0.4034, "step": 1660 }, { "epoch": 0.8438605356240525, "grad_norm": 1.8750139474868774, "learning_rate": 0.00016291111111111113, "loss": 0.5429, "step": 1670 }, { "epoch": 0.8489135927235978, "grad_norm": 2.5823447704315186, "learning_rate": 0.0001626888888888889, "loss": 0.5483, "step": 1680 }, { "epoch": 0.853966649823143, "grad_norm": 1.3424440622329712, "learning_rate": 0.00016246666666666667, "loss": 0.4951, "step": 1690 }, { "epoch": 0.8590197069226883, "grad_norm": 1.6336854696273804, "learning_rate": 0.00016224444444444444, "loss": 0.4557, "step": 1700 }, { "epoch": 0.8590197069226883, "eval_accuracy": 0.8370927318295739, "eval_loss": 0.3707003891468048, "eval_runtime": 22.3815, "eval_samples_per_second": 124.791, "eval_steps_per_second": 15.638, "step": 1700 }, { "epoch": 0.8640727640222334, "grad_norm": 1.8110884428024292, "learning_rate": 0.00016202222222222223, "loss": 0.4014, "step": 1710 }, { "epoch": 0.8691258211217787, "grad_norm": 2.2485573291778564, "learning_rate": 0.00016180000000000003, "loss": 0.5139, "step": 1720 }, { "epoch": 0.8741788782213239, "grad_norm": 1.4717490673065186, "learning_rate": 0.00016157777777777777, "loss": 0.427, "step": 1730 }, { "epoch": 0.8792319353208691, "grad_norm": 1.3247179985046387, "learning_rate": 0.00016135555555555557, "loss": 0.5735, "step": 1740 }, { "epoch": 0.8842849924204144, "grad_norm": 1.9749815464019775, "learning_rate": 0.00016113333333333334, "loss": 0.4366, "step": 1750 }, { "epoch": 0.8893380495199595, "grad_norm": 1.1374492645263672, "learning_rate": 0.00016091111111111113, "loss": 0.3669, "step": 1760 }, { "epoch": 0.8943911066195048, "grad_norm": 0.9348354339599609, "learning_rate": 0.00016068888888888888, "loss": 0.6215, "step": 1770 }, { "epoch": 0.89944416371905, "grad_norm": 3.336740255355835, "learning_rate": 0.00016046666666666667, "loss": 0.4449, "step": 1780 }, { "epoch": 0.9044972208185953, "grad_norm": 1.1256966590881348, "learning_rate": 0.00016024444444444444, "loss": 0.3973, "step": 1790 }, { "epoch": 0.9095502779181405, "grad_norm": 2.256025552749634, "learning_rate": 0.00016002222222222224, "loss": 0.408, "step": 1800 }, { "epoch": 0.9095502779181405, "eval_accuracy": 0.7998567848191909, "eval_loss": 0.49676695466041565, "eval_runtime": 21.7521, "eval_samples_per_second": 128.401, "eval_steps_per_second": 16.09, "step": 1800 }, { "epoch": 0.9146033350176856, "grad_norm": 1.9306086301803589, "learning_rate": 0.0001598, "loss": 0.4096, "step": 1810 }, { "epoch": 0.9196563921172309, "grad_norm": 1.706842303276062, "learning_rate": 0.00015957777777777778, "loss": 0.5705, "step": 1820 }, { "epoch": 0.9247094492167761, "grad_norm": 1.115983486175537, "learning_rate": 0.00015935555555555557, "loss": 0.4, "step": 1830 }, { "epoch": 0.9297625063163214, "grad_norm": 3.9791793823242188, "learning_rate": 0.00015913333333333334, "loss": 0.5316, "step": 1840 }, { "epoch": 0.9348155634158666, "grad_norm": 1.2531650066375732, "learning_rate": 0.0001589111111111111, "loss": 0.4081, "step": 1850 }, { "epoch": 0.9398686205154119, "grad_norm": 8.854723930358887, "learning_rate": 0.00015868888888888888, "loss": 0.4802, "step": 1860 }, { "epoch": 0.944921677614957, "grad_norm": 1.6594269275665283, "learning_rate": 0.00015846666666666668, "loss": 0.3648, "step": 1870 }, { "epoch": 0.9499747347145022, "grad_norm": 2.8492445945739746, "learning_rate": 0.00015824444444444448, "loss": 0.5144, "step": 1880 }, { "epoch": 0.9550277918140475, "grad_norm": 0.9804765582084656, "learning_rate": 0.00015802222222222222, "loss": 0.3473, "step": 1890 }, { "epoch": 0.9600808489135927, "grad_norm": 2.4545788764953613, "learning_rate": 0.00015780000000000001, "loss": 0.4979, "step": 1900 }, { "epoch": 0.9600808489135927, "eval_accuracy": 0.7898317221625493, "eval_loss": 0.44323283433914185, "eval_runtime": 22.292, "eval_samples_per_second": 125.291, "eval_steps_per_second": 15.701, "step": 1900 }, { "epoch": 0.965133906013138, "grad_norm": 3.2131950855255127, "learning_rate": 0.00015757777777777778, "loss": 0.4122, "step": 1910 }, { "epoch": 0.9701869631126832, "grad_norm": 1.4557360410690308, "learning_rate": 0.00015735555555555558, "loss": 0.4336, "step": 1920 }, { "epoch": 0.9752400202122284, "grad_norm": 5.198827743530273, "learning_rate": 0.00015713333333333332, "loss": 0.5444, "step": 1930 }, { "epoch": 0.9802930773117736, "grad_norm": 1.1999950408935547, "learning_rate": 0.00015691111111111112, "loss": 0.395, "step": 1940 }, { "epoch": 0.9853461344113188, "grad_norm": 7.783042907714844, "learning_rate": 0.00015668888888888891, "loss": 0.3799, "step": 1950 }, { "epoch": 0.9903991915108641, "grad_norm": 1.327025294303894, "learning_rate": 0.00015646666666666668, "loss": 0.395, "step": 1960 }, { "epoch": 0.9954522486104093, "grad_norm": 2.7618377208709717, "learning_rate": 0.00015624444444444445, "loss": 0.4848, "step": 1970 }, { "epoch": 1.0005053057099544, "grad_norm": 1.5084019899368286, "learning_rate": 0.00015602222222222222, "loss": 0.4566, "step": 1980 }, { "epoch": 1.0055583628094997, "grad_norm": 8.516404151916504, "learning_rate": 0.00015580000000000002, "loss": 0.3587, "step": 1990 }, { "epoch": 1.010611419909045, "grad_norm": 2.565399646759033, "learning_rate": 0.0001555777777777778, "loss": 0.4115, "step": 2000 }, { "epoch": 1.010611419909045, "eval_accuracy": 0.8392409595417114, "eval_loss": 0.37224915623664856, "eval_runtime": 21.6376, "eval_samples_per_second": 129.081, "eval_steps_per_second": 16.176, "step": 2000 }, { "epoch": 1.0156644770085903, "grad_norm": 2.0563793182373047, "learning_rate": 0.00015535555555555556, "loss": 0.3663, "step": 2010 }, { "epoch": 1.0207175341081354, "grad_norm": 1.1219370365142822, "learning_rate": 0.00015513333333333333, "loss": 0.2725, "step": 2020 }, { "epoch": 1.0257705912076807, "grad_norm": 5.346210956573486, "learning_rate": 0.00015491111111111112, "loss": 0.5517, "step": 2030 }, { "epoch": 1.030823648307226, "grad_norm": 2.1599607467651367, "learning_rate": 0.0001546888888888889, "loss": 0.4699, "step": 2040 }, { "epoch": 1.035876705406771, "grad_norm": 2.0968916416168213, "learning_rate": 0.00015446666666666666, "loss": 0.5148, "step": 2050 }, { "epoch": 1.0409297625063163, "grad_norm": 1.5473324060440063, "learning_rate": 0.00015424444444444446, "loss": 0.4261, "step": 2060 }, { "epoch": 1.0459828196058616, "grad_norm": 2.198204755783081, "learning_rate": 0.00015402222222222223, "loss": 0.4995, "step": 2070 }, { "epoch": 1.0510358767054067, "grad_norm": 3.301290988922119, "learning_rate": 0.0001538, "loss": 0.4424, "step": 2080 }, { "epoch": 1.056088933804952, "grad_norm": 2.5087785720825195, "learning_rate": 0.00015357777777777777, "loss": 0.4171, "step": 2090 }, { "epoch": 1.0611419909044972, "grad_norm": 2.4859185218811035, "learning_rate": 0.00015335555555555556, "loss": 0.3421, "step": 2100 }, { "epoch": 1.0611419909044972, "eval_accuracy": 0.7400644468313641, "eval_loss": 0.5449945330619812, "eval_runtime": 21.8771, "eval_samples_per_second": 127.668, "eval_steps_per_second": 15.998, "step": 2100 }, { "epoch": 1.0661950480040425, "grad_norm": 2.9975504875183105, "learning_rate": 0.00015313333333333336, "loss": 0.3988, "step": 2110 }, { "epoch": 1.0712481051035876, "grad_norm": 1.1956052780151367, "learning_rate": 0.00015291111111111113, "loss": 0.4364, "step": 2120 }, { "epoch": 1.0763011622031329, "grad_norm": 1.7462997436523438, "learning_rate": 0.0001526888888888889, "loss": 0.412, "step": 2130 }, { "epoch": 1.0813542193026782, "grad_norm": 1.935174584388733, "learning_rate": 0.00015246666666666667, "loss": 0.4162, "step": 2140 }, { "epoch": 1.0864072764022232, "grad_norm": 2.8826425075531006, "learning_rate": 0.00015224444444444446, "loss": 0.4252, "step": 2150 }, { "epoch": 1.0914603335017685, "grad_norm": 3.358872652053833, "learning_rate": 0.00015202222222222223, "loss": 0.5634, "step": 2160 }, { "epoch": 1.0965133906013138, "grad_norm": 1.1377031803131104, "learning_rate": 0.0001518, "loss": 0.4235, "step": 2170 }, { "epoch": 1.101566447700859, "grad_norm": 1.375770926475525, "learning_rate": 0.00015157777777777777, "loss": 0.3415, "step": 2180 }, { "epoch": 1.1066195048004042, "grad_norm": 3.0093822479248047, "learning_rate": 0.00015135555555555557, "loss": 0.453, "step": 2190 }, { "epoch": 1.1116725618999495, "grad_norm": 13.947113037109375, "learning_rate": 0.00015113333333333334, "loss": 0.5165, "step": 2200 }, { "epoch": 1.1116725618999495, "eval_accuracy": 0.798782670963122, "eval_loss": 0.4611000120639801, "eval_runtime": 21.9329, "eval_samples_per_second": 127.343, "eval_steps_per_second": 15.958, "step": 2200 }, { "epoch": 1.1167256189994947, "grad_norm": 1.700982689857483, "learning_rate": 0.0001509111111111111, "loss": 0.3809, "step": 2210 }, { "epoch": 1.1217786760990398, "grad_norm": 2.483978748321533, "learning_rate": 0.0001506888888888889, "loss": 0.5038, "step": 2220 }, { "epoch": 1.126831733198585, "grad_norm": 1.1441247463226318, "learning_rate": 0.00015046666666666667, "loss": 0.3939, "step": 2230 }, { "epoch": 1.1318847902981304, "grad_norm": 2.9879062175750732, "learning_rate": 0.00015024444444444444, "loss": 0.4299, "step": 2240 }, { "epoch": 1.1369378473976757, "grad_norm": 1.0201929807662964, "learning_rate": 0.0001500222222222222, "loss": 0.3327, "step": 2250 }, { "epoch": 1.1419909044972207, "grad_norm": 3.08019757270813, "learning_rate": 0.0001498, "loss": 0.3942, "step": 2260 }, { "epoch": 1.147043961596766, "grad_norm": 0.6657310128211975, "learning_rate": 0.0001495777777777778, "loss": 0.4098, "step": 2270 }, { "epoch": 1.1520970186963113, "grad_norm": 3.8270938396453857, "learning_rate": 0.00014935555555555555, "loss": 0.3656, "step": 2280 }, { "epoch": 1.1571500757958564, "grad_norm": 0.9307794570922852, "learning_rate": 0.00014913333333333334, "loss": 0.4363, "step": 2290 }, { "epoch": 1.1622031328954017, "grad_norm": 2.5703837871551514, "learning_rate": 0.00014891111111111111, "loss": 0.4066, "step": 2300 }, { "epoch": 1.1622031328954017, "eval_accuracy": 0.8725384890798424, "eval_loss": 0.3225650489330292, "eval_runtime": 22.2715, "eval_samples_per_second": 125.407, "eval_steps_per_second": 15.715, "step": 2300 }, { "epoch": 1.167256189994947, "grad_norm": 1.6910845041275024, "learning_rate": 0.0001486888888888889, "loss": 0.4286, "step": 2310 }, { "epoch": 1.1723092470944922, "grad_norm": 1.9290587902069092, "learning_rate": 0.00014846666666666665, "loss": 0.2911, "step": 2320 }, { "epoch": 1.1773623041940373, "grad_norm": 5.606022357940674, "learning_rate": 0.00014824444444444445, "loss": 0.4133, "step": 2330 }, { "epoch": 1.1824153612935826, "grad_norm": 13.100260734558105, "learning_rate": 0.00014802222222222225, "loss": 0.4488, "step": 2340 }, { "epoch": 1.187468418393128, "grad_norm": 0.7320961356163025, "learning_rate": 0.00014780000000000001, "loss": 0.5085, "step": 2350 }, { "epoch": 1.192521475492673, "grad_norm": 1.5058331489562988, "learning_rate": 0.00014757777777777778, "loss": 0.4258, "step": 2360 }, { "epoch": 1.1975745325922182, "grad_norm": 1.7382888793945312, "learning_rate": 0.00014735555555555555, "loss": 0.3538, "step": 2370 }, { "epoch": 1.2026275896917635, "grad_norm": 2.7547197341918945, "learning_rate": 0.00014713333333333335, "loss": 0.356, "step": 2380 }, { "epoch": 1.2076806467913088, "grad_norm": 1.4990507364273071, "learning_rate": 0.00014691111111111112, "loss": 0.296, "step": 2390 }, { "epoch": 1.2127337038908539, "grad_norm": 4.220057964324951, "learning_rate": 0.0001466888888888889, "loss": 0.5085, "step": 2400 }, { "epoch": 1.2127337038908539, "eval_accuracy": 0.7762262799856785, "eval_loss": 0.5857972502708435, "eval_runtime": 21.357, "eval_samples_per_second": 130.777, "eval_steps_per_second": 16.388, "step": 2400 }, { "epoch": 1.2177867609903992, "grad_norm": 2.0989768505096436, "learning_rate": 0.00014646666666666666, "loss": 0.4585, "step": 2410 }, { "epoch": 1.2228398180899445, "grad_norm": 2.374248743057251, "learning_rate": 0.00014624444444444445, "loss": 0.4377, "step": 2420 }, { "epoch": 1.2278928751894895, "grad_norm": 0.8370972275733948, "learning_rate": 0.00014602222222222225, "loss": 0.5153, "step": 2430 }, { "epoch": 1.2329459322890348, "grad_norm": 0.8592082262039185, "learning_rate": 0.0001458, "loss": 0.3292, "step": 2440 }, { "epoch": 1.23799898938858, "grad_norm": 2.7484936714172363, "learning_rate": 0.0001455777777777778, "loss": 0.4038, "step": 2450 }, { "epoch": 1.2430520464881254, "grad_norm": 2.1439602375030518, "learning_rate": 0.00014535555555555556, "loss": 0.4894, "step": 2460 }, { "epoch": 1.2481051035876705, "grad_norm": 7.989200592041016, "learning_rate": 0.00014513333333333336, "loss": 0.5044, "step": 2470 }, { "epoch": 1.2531581606872157, "grad_norm": 1.8504892587661743, "learning_rate": 0.0001449111111111111, "loss": 0.3693, "step": 2480 }, { "epoch": 1.258211217786761, "grad_norm": 3.3306427001953125, "learning_rate": 0.0001446888888888889, "loss": 0.3577, "step": 2490 }, { "epoch": 1.263264274886306, "grad_norm": 2.0794196128845215, "learning_rate": 0.0001444666666666667, "loss": 0.4814, "step": 2500 }, { "epoch": 1.263264274886306, "eval_accuracy": 0.7765843179377014, "eval_loss": 0.39810770750045776, "eval_runtime": 21.1156, "eval_samples_per_second": 132.272, "eval_steps_per_second": 16.575, "step": 2500 }, { "epoch": 1.2683173319858514, "grad_norm": 4.378689289093018, "learning_rate": 0.00014424444444444446, "loss": 0.5345, "step": 2510 }, { "epoch": 1.2733703890853967, "grad_norm": 2.1999661922454834, "learning_rate": 0.00014402222222222223, "loss": 0.4487, "step": 2520 }, { "epoch": 1.278423446184942, "grad_norm": 1.7317285537719727, "learning_rate": 0.0001438, "loss": 0.4158, "step": 2530 }, { "epoch": 1.283476503284487, "grad_norm": 1.6528956890106201, "learning_rate": 0.0001435777777777778, "loss": 0.423, "step": 2540 }, { "epoch": 1.2885295603840323, "grad_norm": 2.0606424808502197, "learning_rate": 0.00014335555555555556, "loss": 0.3426, "step": 2550 }, { "epoch": 1.2935826174835776, "grad_norm": 3.3886213302612305, "learning_rate": 0.00014313333333333333, "loss": 0.2466, "step": 2560 }, { "epoch": 1.2986356745831227, "grad_norm": 4.556110858917236, "learning_rate": 0.00014291111111111113, "loss": 0.5412, "step": 2570 }, { "epoch": 1.303688731682668, "grad_norm": 3.602874517440796, "learning_rate": 0.0001426888888888889, "loss": 0.4321, "step": 2580 }, { "epoch": 1.3087417887822133, "grad_norm": 2.471771717071533, "learning_rate": 0.00014246666666666667, "loss": 0.4233, "step": 2590 }, { "epoch": 1.3137948458817585, "grad_norm": 1.1459356546401978, "learning_rate": 0.00014224444444444444, "loss": 0.4554, "step": 2600 }, { "epoch": 1.3137948458817585, "eval_accuracy": 0.7815968492660222, "eval_loss": 0.5076118111610413, "eval_runtime": 22.225, "eval_samples_per_second": 125.669, "eval_steps_per_second": 15.748, "step": 2600 }, { "epoch": 1.3188479029813036, "grad_norm": 1.9474142789840698, "learning_rate": 0.00014202222222222224, "loss": 0.4251, "step": 2610 }, { "epoch": 1.323900960080849, "grad_norm": 1.4904980659484863, "learning_rate": 0.0001418, "loss": 0.4869, "step": 2620 }, { "epoch": 1.3289540171803942, "grad_norm": 1.6901966333389282, "learning_rate": 0.00014157777777777777, "loss": 0.3536, "step": 2630 }, { "epoch": 1.3340070742799393, "grad_norm": 3.12516188621521, "learning_rate": 0.00014135555555555554, "loss": 0.4814, "step": 2640 }, { "epoch": 1.3390601313794845, "grad_norm": 1.6590672731399536, "learning_rate": 0.00014113333333333334, "loss": 0.4346, "step": 2650 }, { "epoch": 1.3441131884790298, "grad_norm": 1.1147915124893188, "learning_rate": 0.00014091111111111114, "loss": 0.3464, "step": 2660 }, { "epoch": 1.3491662455785751, "grad_norm": 0.9573747515678406, "learning_rate": 0.0001406888888888889, "loss": 0.389, "step": 2670 }, { "epoch": 1.3542193026781202, "grad_norm": 1.9805030822753906, "learning_rate": 0.00014046666666666667, "loss": 0.3141, "step": 2680 }, { "epoch": 1.3592723597776655, "grad_norm": 0.7358626127243042, "learning_rate": 0.00014024444444444444, "loss": 0.4402, "step": 2690 }, { "epoch": 1.3643254168772108, "grad_norm": 2.9765231609344482, "learning_rate": 0.00014002222222222224, "loss": 0.2816, "step": 2700 }, { "epoch": 1.3643254168772108, "eval_accuracy": 0.8127461510920158, "eval_loss": 0.47318604588508606, "eval_runtime": 21.8028, "eval_samples_per_second": 128.103, "eval_steps_per_second": 16.053, "step": 2700 }, { "epoch": 1.3693784739767558, "grad_norm": 1.0896739959716797, "learning_rate": 0.0001398, "loss": 0.3198, "step": 2710 }, { "epoch": 1.3744315310763011, "grad_norm": 2.42411732673645, "learning_rate": 0.00013957777777777778, "loss": 0.377, "step": 2720 }, { "epoch": 1.3794845881758464, "grad_norm": 2.3906593322753906, "learning_rate": 0.00013935555555555558, "loss": 0.3727, "step": 2730 }, { "epoch": 1.3845376452753917, "grad_norm": 12.783044815063477, "learning_rate": 0.00013913333333333335, "loss": 0.5074, "step": 2740 }, { "epoch": 1.3895907023749368, "grad_norm": 1.8081468343734741, "learning_rate": 0.00013891111111111111, "loss": 0.3761, "step": 2750 }, { "epoch": 1.394643759474482, "grad_norm": 2.222616672515869, "learning_rate": 0.00013868888888888888, "loss": 0.3254, "step": 2760 }, { "epoch": 1.3996968165740273, "grad_norm": 2.0407073497772217, "learning_rate": 0.00013846666666666668, "loss": 0.2819, "step": 2770 }, { "epoch": 1.4047498736735724, "grad_norm": 7.398078441619873, "learning_rate": 0.00013824444444444445, "loss": 0.4261, "step": 2780 }, { "epoch": 1.4098029307731177, "grad_norm": 59.964385986328125, "learning_rate": 0.00013802222222222222, "loss": 0.4918, "step": 2790 }, { "epoch": 1.414855987872663, "grad_norm": 2.23128080368042, "learning_rate": 0.0001378, "loss": 0.2516, "step": 2800 }, { "epoch": 1.414855987872663, "eval_accuracy": 0.807375581811672, "eval_loss": 0.43152865767478943, "eval_runtime": 21.3822, "eval_samples_per_second": 130.623, "eval_steps_per_second": 16.369, "step": 2800 }, { "epoch": 1.4199090449722083, "grad_norm": 1.753738284111023, "learning_rate": 0.00013757777777777778, "loss": 0.6053, "step": 2810 }, { "epoch": 1.4249621020717533, "grad_norm": 3.4239344596862793, "learning_rate": 0.00013735555555555558, "loss": 0.4158, "step": 2820 }, { "epoch": 1.4300151591712986, "grad_norm": 3.337329387664795, "learning_rate": 0.00013713333333333332, "loss": 0.3406, "step": 2830 }, { "epoch": 1.435068216270844, "grad_norm": 0.9226961731910706, "learning_rate": 0.00013691111111111112, "loss": 0.3624, "step": 2840 }, { "epoch": 1.440121273370389, "grad_norm": 5.681606292724609, "learning_rate": 0.0001366888888888889, "loss": 0.4399, "step": 2850 }, { "epoch": 1.4451743304699343, "grad_norm": 2.9148125648498535, "learning_rate": 0.00013646666666666669, "loss": 0.4295, "step": 2860 }, { "epoch": 1.4502273875694796, "grad_norm": 2.7697062492370605, "learning_rate": 0.00013624444444444443, "loss": 0.565, "step": 2870 }, { "epoch": 1.4552804446690248, "grad_norm": 3.508700132369995, "learning_rate": 0.00013602222222222222, "loss": 0.295, "step": 2880 }, { "epoch": 1.46033350176857, "grad_norm": 1.9089268445968628, "learning_rate": 0.00013580000000000002, "loss": 0.3559, "step": 2890 }, { "epoch": 1.4653865588681152, "grad_norm": 0.7833030819892883, "learning_rate": 0.0001355777777777778, "loss": 0.2903, "step": 2900 }, { "epoch": 1.4653865588681152, "eval_accuracy": 0.8557107053347655, "eval_loss": 0.38446640968322754, "eval_runtime": 22.6761, "eval_samples_per_second": 123.169, "eval_steps_per_second": 15.435, "step": 2900 }, { "epoch": 1.4704396159676605, "grad_norm": 4.780989170074463, "learning_rate": 0.00013535555555555556, "loss": 0.3312, "step": 2910 }, { "epoch": 1.4754926730672056, "grad_norm": 1.7012922763824463, "learning_rate": 0.00013513333333333333, "loss": 0.6287, "step": 2920 }, { "epoch": 1.4805457301667508, "grad_norm": 2.220500946044922, "learning_rate": 0.00013491111111111113, "loss": 0.3873, "step": 2930 }, { "epoch": 1.4855987872662961, "grad_norm": 2.506354331970215, "learning_rate": 0.0001346888888888889, "loss": 0.6636, "step": 2940 }, { "epoch": 1.4906518443658414, "grad_norm": 1.7330349683761597, "learning_rate": 0.00013446666666666666, "loss": 0.3755, "step": 2950 }, { "epoch": 1.4957049014653865, "grad_norm": 1.4615668058395386, "learning_rate": 0.00013424444444444446, "loss": 0.4543, "step": 2960 }, { "epoch": 1.5007579585649318, "grad_norm": 1.5931336879730225, "learning_rate": 0.00013402222222222223, "loss": 0.3237, "step": 2970 }, { "epoch": 1.505811015664477, "grad_norm": 2.628242254257202, "learning_rate": 0.00013380000000000003, "loss": 0.4424, "step": 2980 }, { "epoch": 1.5108640727640221, "grad_norm": 2.077012538909912, "learning_rate": 0.00013357777777777777, "loss": 0.5038, "step": 2990 }, { "epoch": 1.5159171298635674, "grad_norm": 1.190704345703125, "learning_rate": 0.00013335555555555557, "loss": 0.3493, "step": 3000 }, { "epoch": 1.5159171298635674, "eval_accuracy": 0.7977085571070534, "eval_loss": 0.49209100008010864, "eval_runtime": 21.3741, "eval_samples_per_second": 130.672, "eval_steps_per_second": 16.375, "step": 3000 }, { "epoch": 1.5209701869631127, "grad_norm": 2.0057621002197266, "learning_rate": 0.00013313333333333333, "loss": 0.3376, "step": 3010 }, { "epoch": 1.526023244062658, "grad_norm": 1.6876704692840576, "learning_rate": 0.00013291111111111113, "loss": 0.4106, "step": 3020 }, { "epoch": 1.5310763011622033, "grad_norm": 0.5925251841545105, "learning_rate": 0.00013268888888888887, "loss": 0.2921, "step": 3030 }, { "epoch": 1.5361293582617483, "grad_norm": 4.319688320159912, "learning_rate": 0.00013246666666666667, "loss": 0.5481, "step": 3040 }, { "epoch": 1.5411824153612936, "grad_norm": 1.275367259979248, "learning_rate": 0.00013224444444444447, "loss": 0.4418, "step": 3050 }, { "epoch": 1.5462354724608387, "grad_norm": 5.050511837005615, "learning_rate": 0.00013202222222222224, "loss": 0.4144, "step": 3060 }, { "epoch": 1.551288529560384, "grad_norm": 6.44945764541626, "learning_rate": 0.0001318, "loss": 0.351, "step": 3070 }, { "epoch": 1.5563415866599293, "grad_norm": 3.6111042499542236, "learning_rate": 0.00013157777777777777, "loss": 0.4933, "step": 3080 }, { "epoch": 1.5613946437594746, "grad_norm": 1.7661700248718262, "learning_rate": 0.00013135555555555557, "loss": 0.4354, "step": 3090 }, { "epoch": 1.5664477008590199, "grad_norm": 1.1507279872894287, "learning_rate": 0.00013113333333333334, "loss": 0.4251, "step": 3100 }, { "epoch": 1.5664477008590199, "eval_accuracy": 0.8231292517006803, "eval_loss": 0.38550040125846863, "eval_runtime": 22.5736, "eval_samples_per_second": 123.729, "eval_steps_per_second": 15.505, "step": 3100 }, { "epoch": 1.571500757958565, "grad_norm": 2.0319905281066895, "learning_rate": 0.0001309111111111111, "loss": 0.3939, "step": 3110 }, { "epoch": 1.5765538150581102, "grad_norm": 1.942826747894287, "learning_rate": 0.0001306888888888889, "loss": 0.456, "step": 3120 }, { "epoch": 1.5816068721576553, "grad_norm": 5.354785442352295, "learning_rate": 0.00013046666666666668, "loss": 0.4848, "step": 3130 }, { "epoch": 1.5866599292572006, "grad_norm": 2.888082265853882, "learning_rate": 0.00013024444444444445, "loss": 0.4821, "step": 3140 }, { "epoch": 1.5917129863567459, "grad_norm": 1.3641563653945923, "learning_rate": 0.00013002222222222221, "loss": 0.3901, "step": 3150 }, { "epoch": 1.5967660434562911, "grad_norm": 3.273465871810913, "learning_rate": 0.0001298, "loss": 0.4677, "step": 3160 }, { "epoch": 1.6018191005558364, "grad_norm": 2.5433409214019775, "learning_rate": 0.00012957777777777778, "loss": 0.402, "step": 3170 }, { "epoch": 1.6068721576553815, "grad_norm": 0.8629383444786072, "learning_rate": 0.00012935555555555558, "loss": 0.3301, "step": 3180 }, { "epoch": 1.6119252147549268, "grad_norm": 1.8332223892211914, "learning_rate": 0.00012913333333333335, "loss": 0.5108, "step": 3190 }, { "epoch": 1.6169782718544718, "grad_norm": 4.568374156951904, "learning_rate": 0.00012891111111111112, "loss": 0.3356, "step": 3200 }, { "epoch": 1.6169782718544718, "eval_accuracy": 0.832796276405299, "eval_loss": 0.4012071490287781, "eval_runtime": 22.657, "eval_samples_per_second": 123.273, "eval_steps_per_second": 15.448, "step": 3200 }, { "epoch": 1.6220313289540171, "grad_norm": 1.0055279731750488, "learning_rate": 0.0001286888888888889, "loss": 0.4794, "step": 3210 }, { "epoch": 1.6270843860535624, "grad_norm": 0.6549146771430969, "learning_rate": 0.00012846666666666668, "loss": 0.4712, "step": 3220 }, { "epoch": 1.6321374431531077, "grad_norm": 3.6750988960266113, "learning_rate": 0.00012824444444444445, "loss": 0.309, "step": 3230 }, { "epoch": 1.637190500252653, "grad_norm": 5.386397361755371, "learning_rate": 0.00012802222222222222, "loss": 0.483, "step": 3240 }, { "epoch": 1.642243557352198, "grad_norm": 0.7565990686416626, "learning_rate": 0.00012780000000000002, "loss": 0.4779, "step": 3250 }, { "epoch": 1.6472966144517434, "grad_norm": 1.376508116722107, "learning_rate": 0.00012757777777777779, "loss": 0.4482, "step": 3260 }, { "epoch": 1.6523496715512884, "grad_norm": 0.6989770531654358, "learning_rate": 0.00012735555555555556, "loss": 0.3291, "step": 3270 }, { "epoch": 1.6574027286508337, "grad_norm": 7.5308451652526855, "learning_rate": 0.00012713333333333335, "loss": 0.3636, "step": 3280 }, { "epoch": 1.662455785750379, "grad_norm": 0.882866382598877, "learning_rate": 0.00012691111111111112, "loss": 0.4644, "step": 3290 }, { "epoch": 1.6675088428499243, "grad_norm": 0.7267002463340759, "learning_rate": 0.0001266888888888889, "loss": 0.3597, "step": 3300 }, { "epoch": 1.6675088428499243, "eval_accuracy": 0.849624060150376, "eval_loss": 0.33080559968948364, "eval_runtime": 22.2243, "eval_samples_per_second": 125.673, "eval_steps_per_second": 15.749, "step": 3300 }, { "epoch": 1.6725618999494696, "grad_norm": 2.6613428592681885, "learning_rate": 0.00012646666666666666, "loss": 0.3901, "step": 3310 }, { "epoch": 1.6776149570490146, "grad_norm": 0.5789041519165039, "learning_rate": 0.00012624444444444446, "loss": 0.389, "step": 3320 }, { "epoch": 1.68266801414856, "grad_norm": 3.9312431812286377, "learning_rate": 0.00012602222222222223, "loss": 0.454, "step": 3330 }, { "epoch": 1.687721071248105, "grad_norm": 0.7821488976478577, "learning_rate": 0.0001258, "loss": 0.4258, "step": 3340 }, { "epoch": 1.6927741283476503, "grad_norm": 2.7562386989593506, "learning_rate": 0.0001255777777777778, "loss": 0.5281, "step": 3350 }, { "epoch": 1.6978271854471956, "grad_norm": 2.04308819770813, "learning_rate": 0.00012535555555555556, "loss": 0.4504, "step": 3360 }, { "epoch": 1.7028802425467409, "grad_norm": 1.4415785074234009, "learning_rate": 0.00012513333333333336, "loss": 0.3671, "step": 3370 }, { "epoch": 1.7079332996462862, "grad_norm": 1.168853521347046, "learning_rate": 0.0001249111111111111, "loss": 0.458, "step": 3380 }, { "epoch": 1.7129863567458312, "grad_norm": 1.5632737874984741, "learning_rate": 0.0001246888888888889, "loss": 0.396, "step": 3390 }, { "epoch": 1.7180394138453765, "grad_norm": 4.223479270935059, "learning_rate": 0.00012446666666666667, "loss": 0.257, "step": 3400 }, { "epoch": 1.7180394138453765, "eval_accuracy": 0.8138202649480845, "eval_loss": 0.41035306453704834, "eval_runtime": 21.6736, "eval_samples_per_second": 128.867, "eval_steps_per_second": 16.149, "step": 3400 }, { "epoch": 1.7230924709449216, "grad_norm": 1.055281162261963, "learning_rate": 0.00012424444444444446, "loss": 0.5326, "step": 3410 }, { "epoch": 1.7281455280444669, "grad_norm": 1.696380376815796, "learning_rate": 0.0001240222222222222, "loss": 0.5247, "step": 3420 }, { "epoch": 1.7331985851440121, "grad_norm": 9.939696311950684, "learning_rate": 0.0001238, "loss": 0.3881, "step": 3430 }, { "epoch": 1.7382516422435574, "grad_norm": 1.0340977907180786, "learning_rate": 0.0001235777777777778, "loss": 0.4003, "step": 3440 }, { "epoch": 1.7433046993431027, "grad_norm": 0.7816683650016785, "learning_rate": 0.00012335555555555557, "loss": 0.4034, "step": 3450 }, { "epoch": 1.7483577564426478, "grad_norm": 1.9582682847976685, "learning_rate": 0.00012313333333333334, "loss": 0.3969, "step": 3460 }, { "epoch": 1.753410813542193, "grad_norm": 2.5059309005737305, "learning_rate": 0.0001229111111111111, "loss": 0.3695, "step": 3470 }, { "epoch": 1.7584638706417381, "grad_norm": 2.4241933822631836, "learning_rate": 0.0001226888888888889, "loss": 0.3496, "step": 3480 }, { "epoch": 1.7635169277412834, "grad_norm": 2.603628158569336, "learning_rate": 0.00012246666666666667, "loss": 0.3538, "step": 3490 }, { "epoch": 1.7685699848408287, "grad_norm": 3.012014389038086, "learning_rate": 0.00012224444444444444, "loss": 0.3709, "step": 3500 }, { "epoch": 1.7685699848408287, "eval_accuracy": 0.8879341210168278, "eval_loss": 0.2768643796443939, "eval_runtime": 22.0942, "eval_samples_per_second": 126.413, "eval_steps_per_second": 15.841, "step": 3500 }, { "epoch": 1.773623041940374, "grad_norm": 1.784348964691162, "learning_rate": 0.00012202222222222224, "loss": 0.4173, "step": 3510 }, { "epoch": 1.7786760990399193, "grad_norm": 0.6317536234855652, "learning_rate": 0.0001218, "loss": 0.3496, "step": 3520 }, { "epoch": 1.7837291561394644, "grad_norm": 2.0189199447631836, "learning_rate": 0.00012157777777777779, "loss": 0.3708, "step": 3530 }, { "epoch": 1.7887822132390097, "grad_norm": 1.4992175102233887, "learning_rate": 0.00012135555555555556, "loss": 0.3357, "step": 3540 }, { "epoch": 1.7938352703385547, "grad_norm": 0.9670875668525696, "learning_rate": 0.00012113333333333334, "loss": 0.3449, "step": 3550 }, { "epoch": 1.7988883274381, "grad_norm": 2.3772332668304443, "learning_rate": 0.00012091111111111111, "loss": 0.4378, "step": 3560 }, { "epoch": 1.8039413845376453, "grad_norm": 2.047031879425049, "learning_rate": 0.0001206888888888889, "loss": 0.4959, "step": 3570 }, { "epoch": 1.8089944416371906, "grad_norm": 5.458591461181641, "learning_rate": 0.00012046666666666668, "loss": 0.444, "step": 3580 }, { "epoch": 1.8140474987367359, "grad_norm": 3.6228652000427246, "learning_rate": 0.00012024444444444445, "loss": 0.4622, "step": 3590 }, { "epoch": 1.819100555836281, "grad_norm": 4.836859703063965, "learning_rate": 0.00012002222222222224, "loss": 0.3393, "step": 3600 }, { "epoch": 1.819100555836281, "eval_accuracy": 0.8643036161833154, "eval_loss": 0.3412274420261383, "eval_runtime": 21.2864, "eval_samples_per_second": 131.21, "eval_steps_per_second": 16.442, "step": 3600 }, { "epoch": 1.8241536129358262, "grad_norm": 1.5120463371276855, "learning_rate": 0.0001198, "loss": 0.4104, "step": 3610 }, { "epoch": 1.8292066700353713, "grad_norm": 4.8899946212768555, "learning_rate": 0.0001195777777777778, "loss": 0.3979, "step": 3620 }, { "epoch": 1.8342597271349166, "grad_norm": 3.557809352874756, "learning_rate": 0.00011935555555555555, "loss": 0.4674, "step": 3630 }, { "epoch": 1.8393127842344619, "grad_norm": 4.010376930236816, "learning_rate": 0.00011913333333333335, "loss": 0.405, "step": 3640 }, { "epoch": 1.8443658413340072, "grad_norm": 2.8860723972320557, "learning_rate": 0.0001189111111111111, "loss": 0.3503, "step": 3650 }, { "epoch": 1.8494188984335524, "grad_norm": 4.228562355041504, "learning_rate": 0.0001186888888888889, "loss": 0.3722, "step": 3660 }, { "epoch": 1.8544719555330975, "grad_norm": 4.062005519866943, "learning_rate": 0.00011846666666666668, "loss": 0.4213, "step": 3670 }, { "epoch": 1.8595250126326428, "grad_norm": 1.2301322221755981, "learning_rate": 0.00011824444444444445, "loss": 0.3977, "step": 3680 }, { "epoch": 1.8645780697321879, "grad_norm": 2.740429162979126, "learning_rate": 0.00011802222222222223, "loss": 0.4617, "step": 3690 }, { "epoch": 1.8696311268317332, "grad_norm": 1.4782252311706543, "learning_rate": 0.0001178, "loss": 0.4151, "step": 3700 }, { "epoch": 1.8696311268317332, "eval_accuracy": 0.87468671679198, "eval_loss": 0.30783456563949585, "eval_runtime": 22.3441, "eval_samples_per_second": 124.999, "eval_steps_per_second": 15.664, "step": 3700 }, { "epoch": 1.8746841839312784, "grad_norm": 2.2572925090789795, "learning_rate": 0.00011757777777777779, "loss": 0.3715, "step": 3710 }, { "epoch": 1.8797372410308237, "grad_norm": 2.1371333599090576, "learning_rate": 0.00011735555555555556, "loss": 0.5212, "step": 3720 }, { "epoch": 1.884790298130369, "grad_norm": 2.4311232566833496, "learning_rate": 0.00011713333333333334, "loss": 0.4655, "step": 3730 }, { "epoch": 1.889843355229914, "grad_norm": 1.8637980222702026, "learning_rate": 0.00011691111111111112, "loss": 0.4577, "step": 3740 }, { "epoch": 1.8948964123294592, "grad_norm": 2.4759740829467773, "learning_rate": 0.00011668888888888889, "loss": 0.5329, "step": 3750 }, { "epoch": 1.8999494694290044, "grad_norm": 2.1878087520599365, "learning_rate": 0.00011646666666666667, "loss": 0.4461, "step": 3760 }, { "epoch": 1.9050025265285497, "grad_norm": 1.3572431802749634, "learning_rate": 0.00011624444444444444, "loss": 0.3531, "step": 3770 }, { "epoch": 1.910055583628095, "grad_norm": 0.6816936135292053, "learning_rate": 0.00011602222222222223, "loss": 0.3259, "step": 3780 }, { "epoch": 1.9151086407276403, "grad_norm": 0.9463779926300049, "learning_rate": 0.0001158, "loss": 0.2646, "step": 3790 }, { "epoch": 1.9201616978271856, "grad_norm": 0.36243578791618347, "learning_rate": 0.00011557777777777778, "loss": 0.3043, "step": 3800 }, { "epoch": 1.9201616978271856, "eval_accuracy": 0.8650196920873613, "eval_loss": 0.34238138794898987, "eval_runtime": 21.2793, "eval_samples_per_second": 131.254, "eval_steps_per_second": 16.448, "step": 3800 }, { "epoch": 1.9252147549267307, "grad_norm": 0.9268227219581604, "learning_rate": 0.00011535555555555555, "loss": 0.3243, "step": 3810 }, { "epoch": 1.9302678120262757, "grad_norm": 3.1845664978027344, "learning_rate": 0.00011513333333333333, "loss": 0.4937, "step": 3820 }, { "epoch": 1.935320869125821, "grad_norm": 5.107193470001221, "learning_rate": 0.00011491111111111113, "loss": 0.5765, "step": 3830 }, { "epoch": 1.9403739262253663, "grad_norm": 1.0507038831710815, "learning_rate": 0.0001146888888888889, "loss": 0.419, "step": 3840 }, { "epoch": 1.9454269833249116, "grad_norm": 1.0803500413894653, "learning_rate": 0.00011446666666666668, "loss": 0.3792, "step": 3850 }, { "epoch": 1.9504800404244569, "grad_norm": 1.9741244316101074, "learning_rate": 0.00011424444444444445, "loss": 0.3598, "step": 3860 }, { "epoch": 1.9555330975240022, "grad_norm": 3.1313271522521973, "learning_rate": 0.00011402222222222223, "loss": 0.3459, "step": 3870 }, { "epoch": 1.9605861546235472, "grad_norm": 5.064655780792236, "learning_rate": 0.0001138, "loss": 0.4616, "step": 3880 }, { "epoch": 1.9656392117230923, "grad_norm": 0.6265459060668945, "learning_rate": 0.00011357777777777778, "loss": 0.5146, "step": 3890 }, { "epoch": 1.9706922688226376, "grad_norm": 1.0022975206375122, "learning_rate": 0.00011335555555555557, "loss": 0.3302, "step": 3900 }, { "epoch": 1.9706922688226376, "eval_accuracy": 0.8335123523093448, "eval_loss": 0.3512967824935913, "eval_runtime": 32.7928, "eval_samples_per_second": 85.171, "eval_steps_per_second": 10.673, "step": 3900 }, { "epoch": 1.9757453259221829, "grad_norm": 2.7722008228302, "learning_rate": 0.00011313333333333334, "loss": 0.5051, "step": 3910 }, { "epoch": 1.9807983830217282, "grad_norm": 1.5756025314331055, "learning_rate": 0.00011291111111111112, "loss": 0.3682, "step": 3920 }, { "epoch": 1.9858514401212735, "grad_norm": 2.518623113632202, "learning_rate": 0.00011268888888888889, "loss": 0.4204, "step": 3930 }, { "epoch": 1.9909044972208187, "grad_norm": 2.619114637374878, "learning_rate": 0.00011246666666666667, "loss": 0.3214, "step": 3940 }, { "epoch": 1.9959575543203638, "grad_norm": 2.4030561447143555, "learning_rate": 0.00011224444444444444, "loss": 0.4218, "step": 3950 }, { "epoch": 2.001010611419909, "grad_norm": 1.155045986175537, "learning_rate": 0.00011204444444444444, "loss": 0.4088, "step": 3960 }, { "epoch": 2.006063668519454, "grad_norm": 5.908751487731934, "learning_rate": 0.00011182222222222223, "loss": 0.2628, "step": 3970 }, { "epoch": 2.0111167256189995, "grad_norm": 5.073615074157715, "learning_rate": 0.00011160000000000002, "loss": 0.3461, "step": 3980 }, { "epoch": 2.0161697827185447, "grad_norm": 1.5512999296188354, "learning_rate": 0.00011137777777777779, "loss": 0.4615, "step": 3990 }, { "epoch": 2.02122283981809, "grad_norm": 1.2998409271240234, "learning_rate": 0.00011115555555555557, "loss": 0.4033, "step": 4000 }, { "epoch": 2.02122283981809, "eval_accuracy": 0.8510562119584676, "eval_loss": 0.3371362090110779, "eval_runtime": 23.5333, "eval_samples_per_second": 118.683, "eval_steps_per_second": 14.873, "step": 4000 }, { "epoch": 2.0262758969176353, "grad_norm": 1.6033413410186768, "learning_rate": 0.00011093333333333334, "loss": 0.2972, "step": 4010 }, { "epoch": 2.0313289540171806, "grad_norm": 3.248300313949585, "learning_rate": 0.00011071111111111112, "loss": 0.3059, "step": 4020 }, { "epoch": 2.0363820111167255, "grad_norm": 2.213425636291504, "learning_rate": 0.00011048888888888889, "loss": 0.3951, "step": 4030 }, { "epoch": 2.0414350682162707, "grad_norm": 3.888850450515747, "learning_rate": 0.00011026666666666667, "loss": 0.5878, "step": 4040 }, { "epoch": 2.046488125315816, "grad_norm": 2.1447207927703857, "learning_rate": 0.00011004444444444444, "loss": 0.3561, "step": 4050 }, { "epoch": 2.0515411824153613, "grad_norm": 0.9198641777038574, "learning_rate": 0.00010982222222222222, "loss": 0.2937, "step": 4060 }, { "epoch": 2.0565942395149066, "grad_norm": 3.6353166103363037, "learning_rate": 0.00010960000000000001, "loss": 0.4323, "step": 4070 }, { "epoch": 2.061647296614452, "grad_norm": 2.254685878753662, "learning_rate": 0.00010937777777777778, "loss": 0.3451, "step": 4080 }, { "epoch": 2.0667003537139967, "grad_norm": 5.627804756164551, "learning_rate": 0.00010915555555555556, "loss": 0.3757, "step": 4090 }, { "epoch": 2.071753410813542, "grad_norm": 2.4160265922546387, "learning_rate": 0.00010893333333333333, "loss": 0.3386, "step": 4100 }, { "epoch": 2.071753410813542, "eval_accuracy": 0.8395989974937343, "eval_loss": 0.34022802114486694, "eval_runtime": 28.7275, "eval_samples_per_second": 97.224, "eval_steps_per_second": 12.183, "step": 4100 }, { "epoch": 2.0768064679130873, "grad_norm": 2.111335515975952, "learning_rate": 0.00010871111111111113, "loss": 0.3273, "step": 4110 }, { "epoch": 2.0818595250126326, "grad_norm": 0.5963095426559448, "learning_rate": 0.00010848888888888888, "loss": 0.3177, "step": 4120 }, { "epoch": 2.086912582112178, "grad_norm": 0.9041915535926819, "learning_rate": 0.00010826666666666668, "loss": 0.3696, "step": 4130 }, { "epoch": 2.091965639211723, "grad_norm": 0.9016739726066589, "learning_rate": 0.00010804444444444446, "loss": 0.4326, "step": 4140 }, { "epoch": 2.0970186963112685, "grad_norm": 3.7320876121520996, "learning_rate": 0.00010782222222222223, "loss": 0.5018, "step": 4150 }, { "epoch": 2.1020717534108133, "grad_norm": 4.937158584594727, "learning_rate": 0.00010760000000000001, "loss": 0.3579, "step": 4160 }, { "epoch": 2.1071248105103586, "grad_norm": 1.5423624515533447, "learning_rate": 0.00010737777777777778, "loss": 0.3822, "step": 4170 }, { "epoch": 2.112177867609904, "grad_norm": 1.0011184215545654, "learning_rate": 0.00010715555555555557, "loss": 0.2305, "step": 4180 }, { "epoch": 2.117230924709449, "grad_norm": 2.009866952896118, "learning_rate": 0.00010693333333333333, "loss": 0.306, "step": 4190 }, { "epoch": 2.1222839818089945, "grad_norm": 1.8194949626922607, "learning_rate": 0.00010671111111111112, "loss": 0.3661, "step": 4200 }, { "epoch": 2.1222839818089945, "eval_accuracy": 0.8560687432867884, "eval_loss": 0.3276958465576172, "eval_runtime": 21.7445, "eval_samples_per_second": 128.446, "eval_steps_per_second": 16.096, "step": 4200 }, { "epoch": 2.1273370389085398, "grad_norm": 0.9683951735496521, "learning_rate": 0.0001064888888888889, "loss": 0.4137, "step": 4210 }, { "epoch": 2.132390096008085, "grad_norm": 0.4518190920352936, "learning_rate": 0.00010626666666666667, "loss": 0.2384, "step": 4220 }, { "epoch": 2.13744315310763, "grad_norm": 3.257291555404663, "learning_rate": 0.00010604444444444445, "loss": 0.5985, "step": 4230 }, { "epoch": 2.142496210207175, "grad_norm": 2.4535927772521973, "learning_rate": 0.00010582222222222222, "loss": 0.4149, "step": 4240 }, { "epoch": 2.1475492673067205, "grad_norm": 1.2677336931228638, "learning_rate": 0.0001056, "loss": 0.4937, "step": 4250 }, { "epoch": 2.1526023244062658, "grad_norm": 0.7350954413414001, "learning_rate": 0.00010537777777777777, "loss": 0.3773, "step": 4260 }, { "epoch": 2.157655381505811, "grad_norm": 3.036113977432251, "learning_rate": 0.00010515555555555556, "loss": 0.3251, "step": 4270 }, { "epoch": 2.1627084386053563, "grad_norm": 3.8023736476898193, "learning_rate": 0.00010493333333333333, "loss": 0.2999, "step": 4280 }, { "epoch": 2.1677614957049016, "grad_norm": 11.649937629699707, "learning_rate": 0.00010471111111111111, "loss": 0.3958, "step": 4290 }, { "epoch": 2.1728145528044465, "grad_norm": 1.797082781791687, "learning_rate": 0.0001044888888888889, "loss": 0.2914, "step": 4300 }, { "epoch": 2.1728145528044465, "eval_accuracy": 0.8650196920873613, "eval_loss": 0.30653056502342224, "eval_runtime": 21.9529, "eval_samples_per_second": 127.227, "eval_steps_per_second": 15.943, "step": 4300 }, { "epoch": 2.1778676099039918, "grad_norm": 7.0801215171813965, "learning_rate": 0.00010426666666666666, "loss": 0.4043, "step": 4310 }, { "epoch": 2.182920667003537, "grad_norm": 3.196516275405884, "learning_rate": 0.00010404444444444446, "loss": 0.3812, "step": 4320 }, { "epoch": 2.1879737241030823, "grad_norm": 1.0163555145263672, "learning_rate": 0.00010382222222222221, "loss": 0.4174, "step": 4330 }, { "epoch": 2.1930267812026276, "grad_norm": 1.1031217575073242, "learning_rate": 0.00010360000000000001, "loss": 0.3089, "step": 4340 }, { "epoch": 2.198079838302173, "grad_norm": 2.9304659366607666, "learning_rate": 0.00010337777777777777, "loss": 0.3677, "step": 4350 }, { "epoch": 2.203132895401718, "grad_norm": 2.1115951538085938, "learning_rate": 0.00010315555555555556, "loss": 0.526, "step": 4360 }, { "epoch": 2.208185952501263, "grad_norm": 2.1388587951660156, "learning_rate": 0.00010293333333333335, "loss": 0.443, "step": 4370 }, { "epoch": 2.2132390096008083, "grad_norm": 1.8565598726272583, "learning_rate": 0.00010271111111111112, "loss": 0.4296, "step": 4380 }, { "epoch": 2.2182920667003536, "grad_norm": 1.2174099683761597, "learning_rate": 0.0001024888888888889, "loss": 0.3909, "step": 4390 }, { "epoch": 2.223345123799899, "grad_norm": 1.8065009117126465, "learning_rate": 0.00010226666666666667, "loss": 0.4444, "step": 4400 }, { "epoch": 2.223345123799899, "eval_accuracy": 0.849266022198353, "eval_loss": 0.32065466046333313, "eval_runtime": 22.1045, "eval_samples_per_second": 126.354, "eval_steps_per_second": 15.834, "step": 4400 }, { "epoch": 2.228398180899444, "grad_norm": 5.9587860107421875, "learning_rate": 0.00010204444444444445, "loss": 0.3872, "step": 4410 }, { "epoch": 2.2334512379989895, "grad_norm": 3.405024528503418, "learning_rate": 0.00010182222222222222, "loss": 0.5316, "step": 4420 }, { "epoch": 2.2385042950985348, "grad_norm": 1.4730502367019653, "learning_rate": 0.0001016, "loss": 0.2845, "step": 4430 }, { "epoch": 2.2435573521980796, "grad_norm": 1.6922245025634766, "learning_rate": 0.0001013777777777778, "loss": 0.4203, "step": 4440 }, { "epoch": 2.248610409297625, "grad_norm": 1.6515165567398071, "learning_rate": 0.00010115555555555556, "loss": 0.5056, "step": 4450 }, { "epoch": 2.25366346639717, "grad_norm": 4.3611273765563965, "learning_rate": 0.00010093333333333335, "loss": 0.3796, "step": 4460 }, { "epoch": 2.2587165234967155, "grad_norm": 1.1580610275268555, "learning_rate": 0.00010071111111111111, "loss": 0.2071, "step": 4470 }, { "epoch": 2.2637695805962608, "grad_norm": 2.572770833969116, "learning_rate": 0.0001004888888888889, "loss": 0.331, "step": 4480 }, { "epoch": 2.268822637695806, "grad_norm": 3.23252534866333, "learning_rate": 0.00010026666666666666, "loss": 0.4169, "step": 4490 }, { "epoch": 2.2738756947953513, "grad_norm": 1.696787714958191, "learning_rate": 0.00010004444444444446, "loss": 0.2922, "step": 4500 }, { "epoch": 2.2738756947953513, "eval_accuracy": 0.8686000716075905, "eval_loss": 0.29682785272598267, "eval_runtime": 22.5202, "eval_samples_per_second": 124.022, "eval_steps_per_second": 15.542, "step": 4500 }, { "epoch": 2.278928751894896, "grad_norm": 4.435757160186768, "learning_rate": 9.982222222222223e-05, "loss": 0.3005, "step": 4510 }, { "epoch": 2.2839818089944415, "grad_norm": 2.5016374588012695, "learning_rate": 9.960000000000001e-05, "loss": 0.3691, "step": 4520 }, { "epoch": 2.2890348660939868, "grad_norm": 0.4370728135108948, "learning_rate": 9.937777777777778e-05, "loss": 0.3493, "step": 4530 }, { "epoch": 2.294087923193532, "grad_norm": 3.3483691215515137, "learning_rate": 9.915555555555556e-05, "loss": 0.3222, "step": 4540 }, { "epoch": 2.2991409802930773, "grad_norm": 1.3733609914779663, "learning_rate": 9.893333333333333e-05, "loss": 0.3031, "step": 4550 }, { "epoch": 2.3041940373926226, "grad_norm": 1.2431563138961792, "learning_rate": 9.871111111111113e-05, "loss": 0.479, "step": 4560 }, { "epoch": 2.309247094492168, "grad_norm": 1.2833112478256226, "learning_rate": 9.84888888888889e-05, "loss": 0.3924, "step": 4570 }, { "epoch": 2.3143001515917128, "grad_norm": 2.7992002964019775, "learning_rate": 9.826666666666668e-05, "loss": 0.4434, "step": 4580 }, { "epoch": 2.319353208691258, "grad_norm": 1.8905996084213257, "learning_rate": 9.804444444444445e-05, "loss": 0.3713, "step": 4590 }, { "epoch": 2.3244062657908033, "grad_norm": 6.806291580200195, "learning_rate": 9.782222222222223e-05, "loss": 0.3464, "step": 4600 }, { "epoch": 2.3244062657908033, "eval_accuracy": 0.8070175438596491, "eval_loss": 0.415149450302124, "eval_runtime": 21.375, "eval_samples_per_second": 130.666, "eval_steps_per_second": 16.374, "step": 4600 }, { "epoch": 2.3294593228903486, "grad_norm": 0.6903018355369568, "learning_rate": 9.76e-05, "loss": 0.324, "step": 4610 }, { "epoch": 2.334512379989894, "grad_norm": 1.1575100421905518, "learning_rate": 9.737777777777778e-05, "loss": 0.5926, "step": 4620 }, { "epoch": 2.339565437089439, "grad_norm": 1.9170591831207275, "learning_rate": 9.715555555555555e-05, "loss": 0.3373, "step": 4630 }, { "epoch": 2.3446184941889845, "grad_norm": 5.118298053741455, "learning_rate": 9.693333333333335e-05, "loss": 0.4884, "step": 4640 }, { "epoch": 2.3496715512885293, "grad_norm": 1.2098969221115112, "learning_rate": 9.671111111111112e-05, "loss": 0.3405, "step": 4650 }, { "epoch": 2.3547246083880746, "grad_norm": 1.0907807350158691, "learning_rate": 9.64888888888889e-05, "loss": 0.3382, "step": 4660 }, { "epoch": 2.35977766548762, "grad_norm": 1.0323749780654907, "learning_rate": 9.626666666666667e-05, "loss": 0.3487, "step": 4670 }, { "epoch": 2.364830722587165, "grad_norm": 4.246807098388672, "learning_rate": 9.604444444444445e-05, "loss": 0.4093, "step": 4680 }, { "epoch": 2.3698837796867105, "grad_norm": 1.1187255382537842, "learning_rate": 9.582222222222222e-05, "loss": 0.2646, "step": 4690 }, { "epoch": 2.374936836786256, "grad_norm": 0.5659494400024414, "learning_rate": 9.56e-05, "loss": 0.2684, "step": 4700 }, { "epoch": 2.374936836786256, "eval_accuracy": 0.8385248836376656, "eval_loss": 0.3810117244720459, "eval_runtime": 22.0316, "eval_samples_per_second": 126.772, "eval_steps_per_second": 15.886, "step": 4700 }, { "epoch": 2.379989893885801, "grad_norm": 2.3157734870910645, "learning_rate": 9.537777777777778e-05, "loss": 0.3789, "step": 4710 }, { "epoch": 2.385042950985346, "grad_norm": 3.5712051391601562, "learning_rate": 9.515555555555556e-05, "loss": 0.261, "step": 4720 }, { "epoch": 2.390096008084891, "grad_norm": 1.3805081844329834, "learning_rate": 9.493333333333334e-05, "loss": 0.3808, "step": 4730 }, { "epoch": 2.3951490651844365, "grad_norm": 3.2516157627105713, "learning_rate": 9.471111111111111e-05, "loss": 0.3226, "step": 4740 }, { "epoch": 2.4002021222839818, "grad_norm": 3.301288604736328, "learning_rate": 9.44888888888889e-05, "loss": 0.352, "step": 4750 }, { "epoch": 2.405255179383527, "grad_norm": 4.863332271575928, "learning_rate": 9.426666666666666e-05, "loss": 0.4049, "step": 4760 }, { "epoch": 2.4103082364830724, "grad_norm": 3.177022695541382, "learning_rate": 9.404444444444445e-05, "loss": 0.3898, "step": 4770 }, { "epoch": 2.4153612935826176, "grad_norm": 1.482892394065857, "learning_rate": 9.382222222222223e-05, "loss": 0.414, "step": 4780 }, { "epoch": 2.4204143506821625, "grad_norm": 0.5814030766487122, "learning_rate": 9.360000000000001e-05, "loss": 0.2165, "step": 4790 }, { "epoch": 2.4254674077817078, "grad_norm": 1.0302190780639648, "learning_rate": 9.337777777777778e-05, "loss": 0.3779, "step": 4800 }, { "epoch": 2.4254674077817078, "eval_accuracy": 0.8514142499104905, "eval_loss": 0.3367854058742523, "eval_runtime": 21.7291, "eval_samples_per_second": 128.537, "eval_steps_per_second": 16.107, "step": 4800 }, { "epoch": 2.430520464881253, "grad_norm": 0.9444165229797363, "learning_rate": 9.315555555555556e-05, "loss": 0.3039, "step": 4810 }, { "epoch": 2.4355735219807984, "grad_norm": 1.5407205820083618, "learning_rate": 9.293333333333333e-05, "loss": 0.3338, "step": 4820 }, { "epoch": 2.4406265790803436, "grad_norm": 3.1662657260894775, "learning_rate": 9.271111111111112e-05, "loss": 0.3524, "step": 4830 }, { "epoch": 2.445679636179889, "grad_norm": 6.594492435455322, "learning_rate": 9.248888888888889e-05, "loss": 0.3431, "step": 4840 }, { "epoch": 2.450732693279434, "grad_norm": 6.407169818878174, "learning_rate": 9.226666666666667e-05, "loss": 0.4499, "step": 4850 }, { "epoch": 2.455785750378979, "grad_norm": 1.0439250469207764, "learning_rate": 9.204444444444444e-05, "loss": 0.3844, "step": 4860 }, { "epoch": 2.4608388074785243, "grad_norm": 0.4750136435031891, "learning_rate": 9.182222222222223e-05, "loss": 0.3028, "step": 4870 }, { "epoch": 2.4658918645780696, "grad_norm": 1.3194782733917236, "learning_rate": 9.16e-05, "loss": 0.3545, "step": 4880 }, { "epoch": 2.470944921677615, "grad_norm": 3.070331335067749, "learning_rate": 9.137777777777779e-05, "loss": 0.3891, "step": 4890 }, { "epoch": 2.47599797877716, "grad_norm": 3.481339931488037, "learning_rate": 9.115555555555556e-05, "loss": 0.4462, "step": 4900 }, { "epoch": 2.47599797877716, "eval_accuracy": 0.8965270318653777, "eval_loss": 0.2676783800125122, "eval_runtime": 21.1536, "eval_samples_per_second": 132.034, "eval_steps_per_second": 16.546, "step": 4900 }, { "epoch": 2.4810510358767055, "grad_norm": 1.1255896091461182, "learning_rate": 9.093333333333334e-05, "loss": 0.4646, "step": 4910 }, { "epoch": 2.486104092976251, "grad_norm": 3.072685956954956, "learning_rate": 9.071111111111111e-05, "loss": 0.4594, "step": 4920 }, { "epoch": 2.4911571500757956, "grad_norm": 1.3503284454345703, "learning_rate": 9.048888888888889e-05, "loss": 0.3707, "step": 4930 }, { "epoch": 2.496210207175341, "grad_norm": 3.355848789215088, "learning_rate": 9.026666666666666e-05, "loss": 0.3026, "step": 4940 }, { "epoch": 2.501263264274886, "grad_norm": 6.09469747543335, "learning_rate": 9.004444444444446e-05, "loss": 0.4533, "step": 4950 }, { "epoch": 2.5063163213744315, "grad_norm": 1.0013080835342407, "learning_rate": 8.982222222222223e-05, "loss": 0.3047, "step": 4960 }, { "epoch": 2.511369378473977, "grad_norm": 2.1469695568084717, "learning_rate": 8.960000000000001e-05, "loss": 0.2558, "step": 4970 }, { "epoch": 2.516422435573522, "grad_norm": 2.7564926147460938, "learning_rate": 8.937777777777778e-05, "loss": 0.418, "step": 4980 }, { "epoch": 2.5214754926730674, "grad_norm": 0.4358915686607361, "learning_rate": 8.915555555555556e-05, "loss": 0.3014, "step": 4990 }, { "epoch": 2.526528549772612, "grad_norm": 4.178025245666504, "learning_rate": 8.893333333333333e-05, "loss": 0.3766, "step": 5000 }, { "epoch": 2.526528549772612, "eval_accuracy": 0.8438954529180093, "eval_loss": 0.3731708228588104, "eval_runtime": 22.0748, "eval_samples_per_second": 126.524, "eval_steps_per_second": 15.855, "step": 5000 }, { "epoch": 2.5315816068721575, "grad_norm": 3.0973637104034424, "learning_rate": 8.871111111111111e-05, "loss": 0.4647, "step": 5010 }, { "epoch": 2.536634663971703, "grad_norm": 0.8293595910072327, "learning_rate": 8.848888888888888e-05, "loss": 0.3309, "step": 5020 }, { "epoch": 2.541687721071248, "grad_norm": 1.5221765041351318, "learning_rate": 8.826666666666668e-05, "loss": 0.3401, "step": 5030 }, { "epoch": 2.5467407781707934, "grad_norm": 0.9221659898757935, "learning_rate": 8.804444444444445e-05, "loss": 0.2232, "step": 5040 }, { "epoch": 2.5517938352703387, "grad_norm": 1.1780564785003662, "learning_rate": 8.782222222222223e-05, "loss": 0.234, "step": 5050 }, { "epoch": 2.556846892369884, "grad_norm": 1.6946532726287842, "learning_rate": 8.76e-05, "loss": 0.2024, "step": 5060 }, { "epoch": 2.561899949469429, "grad_norm": 11.388461112976074, "learning_rate": 8.737777777777778e-05, "loss": 0.2402, "step": 5070 }, { "epoch": 2.566953006568974, "grad_norm": 4.736606597900391, "learning_rate": 8.715555555555555e-05, "loss": 0.2916, "step": 5080 }, { "epoch": 2.5720060636685194, "grad_norm": 2.6253232955932617, "learning_rate": 8.693333333333334e-05, "loss": 0.4113, "step": 5090 }, { "epoch": 2.5770591207680646, "grad_norm": 1.1557193994522095, "learning_rate": 8.671111111111112e-05, "loss": 0.4971, "step": 5100 }, { "epoch": 2.5770591207680646, "eval_accuracy": 0.8617973505191551, "eval_loss": 0.32662326097488403, "eval_runtime": 22.3316, "eval_samples_per_second": 125.069, "eval_steps_per_second": 15.673, "step": 5100 }, { "epoch": 2.58211217786761, "grad_norm": 1.1730996370315552, "learning_rate": 8.64888888888889e-05, "loss": 0.5626, "step": 5110 }, { "epoch": 2.5871652349671552, "grad_norm": 3.1787731647491455, "learning_rate": 8.626666666666667e-05, "loss": 0.4468, "step": 5120 }, { "epoch": 2.5922182920667005, "grad_norm": 1.4849531650543213, "learning_rate": 8.604444444444445e-05, "loss": 0.3166, "step": 5130 }, { "epoch": 2.5972713491662454, "grad_norm": 2.90570330619812, "learning_rate": 8.582222222222222e-05, "loss": 0.2753, "step": 5140 }, { "epoch": 2.6023244062657906, "grad_norm": 6.187442302703857, "learning_rate": 8.560000000000001e-05, "loss": 0.2136, "step": 5150 }, { "epoch": 2.607377463365336, "grad_norm": 3.46016788482666, "learning_rate": 8.537777777777778e-05, "loss": 0.4352, "step": 5160 }, { "epoch": 2.6124305204648812, "grad_norm": 3.486663579940796, "learning_rate": 8.515555555555556e-05, "loss": 0.4136, "step": 5170 }, { "epoch": 2.6174835775644265, "grad_norm": 8.44055461883545, "learning_rate": 8.493333333333334e-05, "loss": 0.4653, "step": 5180 }, { "epoch": 2.622536634663972, "grad_norm": 0.8385331034660339, "learning_rate": 8.471111111111113e-05, "loss": 0.3393, "step": 5190 }, { "epoch": 2.627589691763517, "grad_norm": 0.9159512519836426, "learning_rate": 8.44888888888889e-05, "loss": 0.3795, "step": 5200 }, { "epoch": 2.627589691763517, "eval_accuracy": 0.8607232366630863, "eval_loss": 0.3379737436771393, "eval_runtime": 21.6774, "eval_samples_per_second": 128.844, "eval_steps_per_second": 16.146, "step": 5200 }, { "epoch": 2.632642748863062, "grad_norm": 3.4703867435455322, "learning_rate": 8.426666666666668e-05, "loss": 0.4484, "step": 5210 }, { "epoch": 2.637695805962607, "grad_norm": 1.2379624843597412, "learning_rate": 8.404444444444445e-05, "loss": 0.3456, "step": 5220 }, { "epoch": 2.6427488630621525, "grad_norm": 1.6584135293960571, "learning_rate": 8.382222222222223e-05, "loss": 0.3585, "step": 5230 }, { "epoch": 2.647801920161698, "grad_norm": 2.661447763442993, "learning_rate": 8.36e-05, "loss": 0.3476, "step": 5240 }, { "epoch": 2.652854977261243, "grad_norm": 2.410435438156128, "learning_rate": 8.337777777777778e-05, "loss": 0.2709, "step": 5250 }, { "epoch": 2.6579080343607884, "grad_norm": 0.6268301010131836, "learning_rate": 8.315555555555557e-05, "loss": 0.4133, "step": 5260 }, { "epoch": 2.6629610914603337, "grad_norm": 0.8858219385147095, "learning_rate": 8.293333333333333e-05, "loss": 0.3238, "step": 5270 }, { "epoch": 2.6680141485598785, "grad_norm": 3.1612367630004883, "learning_rate": 8.271111111111112e-05, "loss": 0.5068, "step": 5280 }, { "epoch": 2.673067205659424, "grad_norm": 2.118220806121826, "learning_rate": 8.248888888888889e-05, "loss": 0.3381, "step": 5290 }, { "epoch": 2.678120262758969, "grad_norm": 1.6003226041793823, "learning_rate": 8.226666666666667e-05, "loss": 0.4205, "step": 5300 }, { "epoch": 2.678120262758969, "eval_accuracy": 0.8617973505191551, "eval_loss": 0.34362542629241943, "eval_runtime": 21.8792, "eval_samples_per_second": 127.655, "eval_steps_per_second": 15.997, "step": 5300 }, { "epoch": 2.6831733198585144, "grad_norm": 3.5239148139953613, "learning_rate": 8.204444444444445e-05, "loss": 0.4689, "step": 5310 }, { "epoch": 2.6882263769580597, "grad_norm": 0.9401977062225342, "learning_rate": 8.182222222222222e-05, "loss": 0.2116, "step": 5320 }, { "epoch": 2.693279434057605, "grad_norm": 3.3830580711364746, "learning_rate": 8.16e-05, "loss": 0.2812, "step": 5330 }, { "epoch": 2.6983324911571502, "grad_norm": 4.928273677825928, "learning_rate": 8.137777777777779e-05, "loss": 0.2851, "step": 5340 }, { "epoch": 2.703385548256695, "grad_norm": 4.011064529418945, "learning_rate": 8.115555555555556e-05, "loss": 0.518, "step": 5350 }, { "epoch": 2.7084386053562404, "grad_norm": 2.41815185546875, "learning_rate": 8.093333333333334e-05, "loss": 0.3202, "step": 5360 }, { "epoch": 2.7134916624557857, "grad_norm": 2.0322022438049316, "learning_rate": 8.071111111111111e-05, "loss": 0.4735, "step": 5370 }, { "epoch": 2.718544719555331, "grad_norm": 2.1056594848632812, "learning_rate": 8.048888888888889e-05, "loss": 0.4202, "step": 5380 }, { "epoch": 2.7235977766548762, "grad_norm": 2.228013038635254, "learning_rate": 8.026666666666666e-05, "loss": 0.4005, "step": 5390 }, { "epoch": 2.7286508337544215, "grad_norm": 2.2472286224365234, "learning_rate": 8.004444444444444e-05, "loss": 0.3652, "step": 5400 }, { "epoch": 2.7286508337544215, "eval_accuracy": 0.8517722878625135, "eval_loss": 0.34825843572616577, "eval_runtime": 23.475, "eval_samples_per_second": 118.978, "eval_steps_per_second": 14.909, "step": 5400 }, { "epoch": 2.733703890853967, "grad_norm": 2.0401411056518555, "learning_rate": 7.982222222222223e-05, "loss": 0.3914, "step": 5410 }, { "epoch": 2.7387569479535117, "grad_norm": 2.2407047748565674, "learning_rate": 7.960000000000001e-05, "loss": 0.4178, "step": 5420 }, { "epoch": 2.743810005053057, "grad_norm": 4.68412446975708, "learning_rate": 7.937777777777778e-05, "loss": 0.3795, "step": 5430 }, { "epoch": 2.7488630621526022, "grad_norm": 6.206188201904297, "learning_rate": 7.915555555555556e-05, "loss": 0.4226, "step": 5440 }, { "epoch": 2.7539161192521475, "grad_norm": 2.395399808883667, "learning_rate": 7.893333333333333e-05, "loss": 0.3077, "step": 5450 }, { "epoch": 2.758969176351693, "grad_norm": 0.5152959227561951, "learning_rate": 7.871111111111111e-05, "loss": 0.2805, "step": 5460 }, { "epoch": 2.764022233451238, "grad_norm": 4.421482086181641, "learning_rate": 7.848888888888888e-05, "loss": 0.2435, "step": 5470 }, { "epoch": 2.7690752905507834, "grad_norm": 1.4925546646118164, "learning_rate": 7.826666666666667e-05, "loss": 0.309, "step": 5480 }, { "epoch": 2.7741283476503282, "grad_norm": 1.8580783605575562, "learning_rate": 7.804444444444445e-05, "loss": 0.3518, "step": 5490 }, { "epoch": 2.7791814047498735, "grad_norm": 2.6896092891693115, "learning_rate": 7.782222222222223e-05, "loss": 0.3999, "step": 5500 }, { "epoch": 2.7791814047498735, "eval_accuracy": 0.8907984246330111, "eval_loss": 0.26029789447784424, "eval_runtime": 34.9298, "eval_samples_per_second": 79.96, "eval_steps_per_second": 10.02, "step": 5500 }, { "epoch": 2.784234461849419, "grad_norm": 2.076117992401123, "learning_rate": 7.76e-05, "loss": 0.306, "step": 5510 }, { "epoch": 2.789287518948964, "grad_norm": 1.225536823272705, "learning_rate": 7.737777777777779e-05, "loss": 0.3124, "step": 5520 }, { "epoch": 2.7943405760485094, "grad_norm": 3.0649378299713135, "learning_rate": 7.715555555555555e-05, "loss": 0.4085, "step": 5530 }, { "epoch": 2.7993936331480547, "grad_norm": 3.2014713287353516, "learning_rate": 7.693333333333334e-05, "loss": 0.3938, "step": 5540 }, { "epoch": 2.8044466902476, "grad_norm": 2.591061592102051, "learning_rate": 7.671111111111111e-05, "loss": 0.4353, "step": 5550 }, { "epoch": 2.809499747347145, "grad_norm": 2.042320966720581, "learning_rate": 7.648888888888889e-05, "loss": 0.2404, "step": 5560 }, { "epoch": 2.81455280444669, "grad_norm": 3.255751371383667, "learning_rate": 7.626666666666667e-05, "loss": 0.4042, "step": 5570 }, { "epoch": 2.8196058615462354, "grad_norm": 0.9873941540718079, "learning_rate": 7.604444444444446e-05, "loss": 0.2521, "step": 5580 }, { "epoch": 2.8246589186457807, "grad_norm": 2.1552438735961914, "learning_rate": 7.582222222222223e-05, "loss": 0.4721, "step": 5590 }, { "epoch": 2.829711975745326, "grad_norm": 3.4882593154907227, "learning_rate": 7.560000000000001e-05, "loss": 0.2909, "step": 5600 }, { "epoch": 2.829711975745326, "eval_accuracy": 0.8693161475116362, "eval_loss": 0.3080480694770813, "eval_runtime": 22.1351, "eval_samples_per_second": 126.18, "eval_steps_per_second": 15.812, "step": 5600 }, { "epoch": 2.8347650328448712, "grad_norm": 1.2430808544158936, "learning_rate": 7.537777777777778e-05, "loss": 0.2978, "step": 5610 }, { "epoch": 2.8398180899444165, "grad_norm": 3.880941390991211, "learning_rate": 7.515555555555556e-05, "loss": 0.4236, "step": 5620 }, { "epoch": 2.8448711470439614, "grad_norm": 3.4325268268585205, "learning_rate": 7.493333333333333e-05, "loss": 0.3418, "step": 5630 }, { "epoch": 2.8499242041435067, "grad_norm": 2.5943145751953125, "learning_rate": 7.471111111111111e-05, "loss": 0.2879, "step": 5640 }, { "epoch": 2.854977261243052, "grad_norm": 4.398190021514893, "learning_rate": 7.44888888888889e-05, "loss": 0.3009, "step": 5650 }, { "epoch": 2.8600303183425972, "grad_norm": 3.223385810852051, "learning_rate": 7.426666666666668e-05, "loss": 0.3087, "step": 5660 }, { "epoch": 2.8650833754421425, "grad_norm": 1.4026299715042114, "learning_rate": 7.404444444444445e-05, "loss": 0.4915, "step": 5670 }, { "epoch": 2.870136432541688, "grad_norm": 5.0744194984436035, "learning_rate": 7.382222222222223e-05, "loss": 0.3001, "step": 5680 }, { "epoch": 2.875189489641233, "grad_norm": 2.3227591514587402, "learning_rate": 7.36e-05, "loss": 0.2285, "step": 5690 }, { "epoch": 2.880242546740778, "grad_norm": 2.202967643737793, "learning_rate": 7.337777777777778e-05, "loss": 0.3703, "step": 5700 }, { "epoch": 2.880242546740778, "eval_accuracy": 0.8807733619763695, "eval_loss": 0.29504504799842834, "eval_runtime": 22.1907, "eval_samples_per_second": 125.864, "eval_steps_per_second": 15.772, "step": 5700 }, { "epoch": 2.8852956038403232, "grad_norm": 3.5945940017700195, "learning_rate": 7.315555555555555e-05, "loss": 0.2504, "step": 5710 }, { "epoch": 2.8903486609398685, "grad_norm": 0.8813655376434326, "learning_rate": 7.293333333333334e-05, "loss": 0.2745, "step": 5720 }, { "epoch": 2.895401718039414, "grad_norm": 1.0525150299072266, "learning_rate": 7.271111111111112e-05, "loss": 0.4289, "step": 5730 }, { "epoch": 2.900454775138959, "grad_norm": 1.6976814270019531, "learning_rate": 7.24888888888889e-05, "loss": 0.3154, "step": 5740 }, { "epoch": 2.9055078322385044, "grad_norm": 0.7887847423553467, "learning_rate": 7.226666666666667e-05, "loss": 0.3418, "step": 5750 }, { "epoch": 2.9105608893380497, "grad_norm": 4.849771499633789, "learning_rate": 7.204444444444445e-05, "loss": 0.317, "step": 5760 }, { "epoch": 2.9156139464375945, "grad_norm": 3.2878639698028564, "learning_rate": 7.182222222222222e-05, "loss": 0.4522, "step": 5770 }, { "epoch": 2.92066700353714, "grad_norm": 1.5125163793563843, "learning_rate": 7.16e-05, "loss": 0.366, "step": 5780 }, { "epoch": 2.925720060636685, "grad_norm": 1.905968189239502, "learning_rate": 7.137777777777778e-05, "loss": 0.2858, "step": 5790 }, { "epoch": 2.9307731177362304, "grad_norm": 3.730903387069702, "learning_rate": 7.115555555555556e-05, "loss": 0.4048, "step": 5800 }, { "epoch": 2.9307731177362304, "eval_accuracy": 0.8499820981023989, "eval_loss": 0.3190869987010956, "eval_runtime": 32.5094, "eval_samples_per_second": 85.913, "eval_steps_per_second": 10.766, "step": 5800 }, { "epoch": 2.9358261748357757, "grad_norm": 4.303518772125244, "learning_rate": 7.093333333333334e-05, "loss": 0.3473, "step": 5810 }, { "epoch": 2.940879231935321, "grad_norm": 2.2424232959747314, "learning_rate": 7.071111111111111e-05, "loss": 0.3791, "step": 5820 }, { "epoch": 2.9459322890348663, "grad_norm": 3.953787326812744, "learning_rate": 7.048888888888889e-05, "loss": 0.2901, "step": 5830 }, { "epoch": 2.950985346134411, "grad_norm": 3.499569892883301, "learning_rate": 7.026666666666668e-05, "loss": 0.4007, "step": 5840 }, { "epoch": 2.9560384032339564, "grad_norm": 2.771538019180298, "learning_rate": 7.004444444444445e-05, "loss": 0.262, "step": 5850 }, { "epoch": 2.9610914603335017, "grad_norm": 0.8818975687026978, "learning_rate": 6.982222222222223e-05, "loss": 0.222, "step": 5860 }, { "epoch": 2.966144517433047, "grad_norm": 2.2992067337036133, "learning_rate": 6.96e-05, "loss": 0.327, "step": 5870 }, { "epoch": 2.9711975745325923, "grad_norm": 0.2194339632987976, "learning_rate": 6.937777777777778e-05, "loss": 0.3098, "step": 5880 }, { "epoch": 2.9762506316321375, "grad_norm": 1.6648062467575073, "learning_rate": 6.915555555555556e-05, "loss": 0.341, "step": 5890 }, { "epoch": 2.981303688731683, "grad_norm": 3.647399663925171, "learning_rate": 6.893333333333333e-05, "loss": 0.3333, "step": 5900 }, { "epoch": 2.981303688731683, "eval_accuracy": 0.8442534908700322, "eval_loss": 0.3772529363632202, "eval_runtime": 23.329, "eval_samples_per_second": 119.722, "eval_steps_per_second": 15.003, "step": 5900 }, { "epoch": 2.9863567458312277, "grad_norm": 1.5647292137145996, "learning_rate": 6.871111111111112e-05, "loss": 0.428, "step": 5910 }, { "epoch": 2.991409802930773, "grad_norm": 1.4988956451416016, "learning_rate": 6.848888888888889e-05, "loss": 0.287, "step": 5920 }, { "epoch": 2.9964628600303183, "grad_norm": 3.3127992153167725, "learning_rate": 6.826666666666667e-05, "loss": 0.3751, "step": 5930 }, { "epoch": 3.0015159171298635, "grad_norm": 2.675394296646118, "learning_rate": 6.804444444444444e-05, "loss": 0.3207, "step": 5940 }, { "epoch": 3.006568974229409, "grad_norm": 2.956672191619873, "learning_rate": 6.782222222222222e-05, "loss": 0.2395, "step": 5950 }, { "epoch": 3.011622031328954, "grad_norm": 1.2761588096618652, "learning_rate": 6.76e-05, "loss": 0.3996, "step": 5960 }, { "epoch": 3.0166750884284994, "grad_norm": 3.3987114429473877, "learning_rate": 6.737777777777779e-05, "loss": 0.3969, "step": 5970 }, { "epoch": 3.0217281455280443, "grad_norm": 0.618013322353363, "learning_rate": 6.715555555555556e-05, "loss": 0.3171, "step": 5980 }, { "epoch": 3.0267812026275895, "grad_norm": 1.8241897821426392, "learning_rate": 6.693333333333334e-05, "loss": 0.2785, "step": 5990 }, { "epoch": 3.031834259727135, "grad_norm": 0.8514062762260437, "learning_rate": 6.671111111111111e-05, "loss": 0.2917, "step": 6000 }, { "epoch": 3.031834259727135, "eval_accuracy": 0.8431793770139635, "eval_loss": 0.37310540676116943, "eval_runtime": 22.445, "eval_samples_per_second": 124.437, "eval_steps_per_second": 15.594, "step": 6000 }, { "epoch": 3.03688731682668, "grad_norm": 3.3822669982910156, "learning_rate": 6.648888888888889e-05, "loss": 0.3108, "step": 6010 }, { "epoch": 3.0419403739262254, "grad_norm": 1.4962226152420044, "learning_rate": 6.626666666666666e-05, "loss": 0.3088, "step": 6020 }, { "epoch": 3.0469934310257707, "grad_norm": 1.0340983867645264, "learning_rate": 6.604444444444444e-05, "loss": 0.2422, "step": 6030 }, { "epoch": 3.052046488125316, "grad_norm": 0.6716813445091248, "learning_rate": 6.582222222222223e-05, "loss": 0.1821, "step": 6040 }, { "epoch": 3.057099545224861, "grad_norm": 4.281582355499268, "learning_rate": 6.560000000000001e-05, "loss": 0.3793, "step": 6050 }, { "epoch": 3.062152602324406, "grad_norm": 8.729735374450684, "learning_rate": 6.537777777777778e-05, "loss": 0.1702, "step": 6060 }, { "epoch": 3.0672056594239514, "grad_norm": 7.878430366516113, "learning_rate": 6.515555555555556e-05, "loss": 0.4737, "step": 6070 }, { "epoch": 3.0722587165234967, "grad_norm": 0.1650269776582718, "learning_rate": 6.493333333333333e-05, "loss": 0.1508, "step": 6080 }, { "epoch": 3.077311773623042, "grad_norm": 0.7151690125465393, "learning_rate": 6.471111111111111e-05, "loss": 0.3404, "step": 6090 }, { "epoch": 3.0823648307225873, "grad_norm": 2.156454086303711, "learning_rate": 6.448888888888888e-05, "loss": 0.4204, "step": 6100 }, { "epoch": 3.0823648307225873, "eval_accuracy": 0.8528464017185822, "eval_loss": 0.37828850746154785, "eval_runtime": 22.0672, "eval_samples_per_second": 126.568, "eval_steps_per_second": 15.861, "step": 6100 }, { "epoch": 3.0874178878221326, "grad_norm": 7.122072219848633, "learning_rate": 6.426666666666668e-05, "loss": 0.4055, "step": 6110 }, { "epoch": 3.0924709449216774, "grad_norm": 1.1863442659378052, "learning_rate": 6.404444444444445e-05, "loss": 0.4349, "step": 6120 }, { "epoch": 3.0975240020212227, "grad_norm": 0.818795382976532, "learning_rate": 6.382222222222223e-05, "loss": 0.3389, "step": 6130 }, { "epoch": 3.102577059120768, "grad_norm": 1.5806599855422974, "learning_rate": 6.36e-05, "loss": 0.2334, "step": 6140 }, { "epoch": 3.1076301162203133, "grad_norm": 1.5807373523712158, "learning_rate": 6.337777777777778e-05, "loss": 0.2308, "step": 6150 }, { "epoch": 3.1126831733198586, "grad_norm": 2.569695472717285, "learning_rate": 6.315555555555555e-05, "loss": 0.2191, "step": 6160 }, { "epoch": 3.117736230419404, "grad_norm": 5.298694610595703, "learning_rate": 6.293333333333334e-05, "loss": 0.3194, "step": 6170 }, { "epoch": 3.122789287518949, "grad_norm": 1.8455535173416138, "learning_rate": 6.27111111111111e-05, "loss": 0.3616, "step": 6180 }, { "epoch": 3.127842344618494, "grad_norm": 7.147807598114014, "learning_rate": 6.24888888888889e-05, "loss": 0.3347, "step": 6190 }, { "epoch": 3.1328954017180393, "grad_norm": 3.828178644180298, "learning_rate": 6.226666666666667e-05, "loss": 0.3832, "step": 6200 }, { "epoch": 3.1328954017180393, "eval_accuracy": 0.8693161475116362, "eval_loss": 0.3008694350719452, "eval_runtime": 22.3667, "eval_samples_per_second": 124.873, "eval_steps_per_second": 15.648, "step": 6200 }, { "epoch": 3.1379484588175846, "grad_norm": 4.436605453491211, "learning_rate": 6.204444444444445e-05, "loss": 0.4343, "step": 6210 }, { "epoch": 3.14300151591713, "grad_norm": 2.5253171920776367, "learning_rate": 6.182222222222222e-05, "loss": 0.2627, "step": 6220 }, { "epoch": 3.148054573016675, "grad_norm": 2.0070149898529053, "learning_rate": 6.16e-05, "loss": 0.2449, "step": 6230 }, { "epoch": 3.1531076301162204, "grad_norm": 0.8311755061149597, "learning_rate": 6.137777777777778e-05, "loss": 0.3659, "step": 6240 }, { "epoch": 3.1581606872157657, "grad_norm": 4.4285478591918945, "learning_rate": 6.115555555555556e-05, "loss": 0.4016, "step": 6250 }, { "epoch": 3.1632137443153105, "grad_norm": 2.689814805984497, "learning_rate": 6.093333333333333e-05, "loss": 0.3232, "step": 6260 }, { "epoch": 3.168266801414856, "grad_norm": 2.877394914627075, "learning_rate": 6.071111111111112e-05, "loss": 0.2843, "step": 6270 }, { "epoch": 3.173319858514401, "grad_norm": 1.4330358505249023, "learning_rate": 6.0488888888888894e-05, "loss": 0.2416, "step": 6280 }, { "epoch": 3.1783729156139464, "grad_norm": 3.4811785221099854, "learning_rate": 6.026666666666667e-05, "loss": 0.2397, "step": 6290 }, { "epoch": 3.1834259727134917, "grad_norm": 3.5192840099334717, "learning_rate": 6.0044444444444446e-05, "loss": 0.32, "step": 6300 }, { "epoch": 3.1834259727134917, "eval_accuracy": 0.8367346938775511, "eval_loss": 0.36899709701538086, "eval_runtime": 22.1835, "eval_samples_per_second": 125.904, "eval_steps_per_second": 15.777, "step": 6300 }, { "epoch": 3.188479029813037, "grad_norm": 2.0998518466949463, "learning_rate": 5.982222222222222e-05, "loss": 0.3383, "step": 6310 }, { "epoch": 3.1935320869125823, "grad_norm": 7.645174503326416, "learning_rate": 5.96e-05, "loss": 0.2519, "step": 6320 }, { "epoch": 3.198585144012127, "grad_norm": 0.5995309352874756, "learning_rate": 5.9377777777777775e-05, "loss": 0.286, "step": 6330 }, { "epoch": 3.2036382011116724, "grad_norm": 1.9291926622390747, "learning_rate": 5.915555555555555e-05, "loss": 0.3522, "step": 6340 }, { "epoch": 3.2086912582112177, "grad_norm": 2.8859310150146484, "learning_rate": 5.893333333333334e-05, "loss": 0.2833, "step": 6350 }, { "epoch": 3.213744315310763, "grad_norm": 2.0815603733062744, "learning_rate": 5.871111111111112e-05, "loss": 0.2778, "step": 6360 }, { "epoch": 3.2187973724103083, "grad_norm": 5.9688801765441895, "learning_rate": 5.848888888888889e-05, "loss": 0.2712, "step": 6370 }, { "epoch": 3.2238504295098536, "grad_norm": 1.616219401359558, "learning_rate": 5.826666666666667e-05, "loss": 0.2327, "step": 6380 }, { "epoch": 3.228903486609399, "grad_norm": 0.7303557395935059, "learning_rate": 5.8044444444444445e-05, "loss": 0.3687, "step": 6390 }, { "epoch": 3.2339565437089437, "grad_norm": 3.5262770652770996, "learning_rate": 5.782222222222222e-05, "loss": 0.3761, "step": 6400 }, { "epoch": 3.2339565437089437, "eval_accuracy": 0.8392409595417114, "eval_loss": 0.3397537171840668, "eval_runtime": 21.8836, "eval_samples_per_second": 127.63, "eval_steps_per_second": 15.994, "step": 6400 }, { "epoch": 3.239009600808489, "grad_norm": 2.527420997619629, "learning_rate": 5.76e-05, "loss": 0.3856, "step": 6410 }, { "epoch": 3.2440626579080343, "grad_norm": 1.4119328260421753, "learning_rate": 5.737777777777779e-05, "loss": 0.3655, "step": 6420 }, { "epoch": 3.2491157150075796, "grad_norm": 2.6743695735931396, "learning_rate": 5.715555555555556e-05, "loss": 0.2527, "step": 6430 }, { "epoch": 3.254168772107125, "grad_norm": 1.3962702751159668, "learning_rate": 5.693333333333334e-05, "loss": 0.3161, "step": 6440 }, { "epoch": 3.25922182920667, "grad_norm": 2.3100271224975586, "learning_rate": 5.6711111111111116e-05, "loss": 0.3051, "step": 6450 }, { "epoch": 3.2642748863062154, "grad_norm": 1.7482351064682007, "learning_rate": 5.648888888888889e-05, "loss": 0.2152, "step": 6460 }, { "epoch": 3.2693279434057603, "grad_norm": 4.773191928863525, "learning_rate": 5.626666666666667e-05, "loss": 0.291, "step": 6470 }, { "epoch": 3.2743810005053056, "grad_norm": 1.4123096466064453, "learning_rate": 5.6044444444444444e-05, "loss": 0.3263, "step": 6480 }, { "epoch": 3.279434057604851, "grad_norm": 8.474733352661133, "learning_rate": 5.582222222222222e-05, "loss": 0.4064, "step": 6490 }, { "epoch": 3.284487114704396, "grad_norm": 4.797526836395264, "learning_rate": 5.560000000000001e-05, "loss": 0.4041, "step": 6500 }, { "epoch": 3.284487114704396, "eval_accuracy": 0.8761188686000716, "eval_loss": 0.2725840210914612, "eval_runtime": 23.1102, "eval_samples_per_second": 120.856, "eval_steps_per_second": 15.145, "step": 6500 }, { "epoch": 3.2895401718039414, "grad_norm": 0.610005795955658, "learning_rate": 5.5377777777777786e-05, "loss": 0.2767, "step": 6510 }, { "epoch": 3.2945932289034867, "grad_norm": 2.132481813430786, "learning_rate": 5.515555555555556e-05, "loss": 0.3442, "step": 6520 }, { "epoch": 3.299646286003032, "grad_norm": 1.465570330619812, "learning_rate": 5.493333333333334e-05, "loss": 0.3322, "step": 6530 }, { "epoch": 3.304699343102577, "grad_norm": 3.551342487335205, "learning_rate": 5.4711111111111114e-05, "loss": 0.283, "step": 6540 }, { "epoch": 3.309752400202122, "grad_norm": 2.26155686378479, "learning_rate": 5.448888888888889e-05, "loss": 0.4365, "step": 6550 }, { "epoch": 3.3148054573016674, "grad_norm": 1.0264147520065308, "learning_rate": 5.4266666666666667e-05, "loss": 0.3945, "step": 6560 }, { "epoch": 3.3198585144012127, "grad_norm": 0.975284218788147, "learning_rate": 5.404444444444444e-05, "loss": 0.3536, "step": 6570 }, { "epoch": 3.324911571500758, "grad_norm": 0.632727324962616, "learning_rate": 5.382222222222223e-05, "loss": 0.2341, "step": 6580 }, { "epoch": 3.3299646286003033, "grad_norm": 0.36360910534858704, "learning_rate": 5.360000000000001e-05, "loss": 0.1528, "step": 6590 }, { "epoch": 3.3350176856998486, "grad_norm": 0.28347837924957275, "learning_rate": 5.3377777777777785e-05, "loss": 0.3373, "step": 6600 }, { "epoch": 3.3350176856998486, "eval_accuracy": 0.828499820981024, "eval_loss": 0.3734741508960724, "eval_runtime": 22.6698, "eval_samples_per_second": 123.203, "eval_steps_per_second": 15.439, "step": 6600 }, { "epoch": 3.3400707427993934, "grad_norm": 2.1117758750915527, "learning_rate": 5.315555555555556e-05, "loss": 0.4591, "step": 6610 }, { "epoch": 3.3451237998989387, "grad_norm": 0.6761242151260376, "learning_rate": 5.293333333333334e-05, "loss": 0.3138, "step": 6620 }, { "epoch": 3.350176856998484, "grad_norm": 0.20823055505752563, "learning_rate": 5.271111111111111e-05, "loss": 0.3087, "step": 6630 }, { "epoch": 3.3552299140980293, "grad_norm": 3.516846179962158, "learning_rate": 5.248888888888889e-05, "loss": 0.317, "step": 6640 }, { "epoch": 3.3602829711975746, "grad_norm": 4.570127010345459, "learning_rate": 5.2266666666666665e-05, "loss": 0.3218, "step": 6650 }, { "epoch": 3.36533602829712, "grad_norm": 0.7695002555847168, "learning_rate": 5.204444444444445e-05, "loss": 0.33, "step": 6660 }, { "epoch": 3.370389085396665, "grad_norm": 0.1837574988603592, "learning_rate": 5.1822222222222224e-05, "loss": 0.3042, "step": 6670 }, { "epoch": 3.37544214249621, "grad_norm": 0.43669673800468445, "learning_rate": 5.16e-05, "loss": 0.2121, "step": 6680 }, { "epoch": 3.3804951995957553, "grad_norm": 5.660565376281738, "learning_rate": 5.1377777777777784e-05, "loss": 0.45, "step": 6690 }, { "epoch": 3.3855482566953006, "grad_norm": 0.8151216506958008, "learning_rate": 5.115555555555556e-05, "loss": 0.2869, "step": 6700 }, { "epoch": 3.3855482566953006, "eval_accuracy": 0.8986752595775153, "eval_loss": 0.23256993293762207, "eval_runtime": 22.5495, "eval_samples_per_second": 123.861, "eval_steps_per_second": 15.521, "step": 6700 }, { "epoch": 3.390601313794846, "grad_norm": 0.24948927760124207, "learning_rate": 5.0933333333333336e-05, "loss": 0.2391, "step": 6710 }, { "epoch": 3.395654370894391, "grad_norm": 1.697418212890625, "learning_rate": 5.071111111111111e-05, "loss": 0.2439, "step": 6720 }, { "epoch": 3.4007074279939364, "grad_norm": 5.376853942871094, "learning_rate": 5.0488888888888895e-05, "loss": 0.3189, "step": 6730 }, { "epoch": 3.4057604850934817, "grad_norm": 1.7416326999664307, "learning_rate": 5.026666666666667e-05, "loss": 0.3712, "step": 6740 }, { "epoch": 3.4108135421930266, "grad_norm": 2.053687572479248, "learning_rate": 5.004444444444445e-05, "loss": 0.3047, "step": 6750 }, { "epoch": 3.415866599292572, "grad_norm": 0.6359758377075195, "learning_rate": 4.982222222222222e-05, "loss": 0.3863, "step": 6760 }, { "epoch": 3.420919656392117, "grad_norm": 0.6436170339584351, "learning_rate": 4.96e-05, "loss": 0.3256, "step": 6770 }, { "epoch": 3.4259727134916624, "grad_norm": 1.9105366468429565, "learning_rate": 4.9377777777777776e-05, "loss": 0.2984, "step": 6780 }, { "epoch": 3.4310257705912077, "grad_norm": 6.483319282531738, "learning_rate": 4.915555555555556e-05, "loss": 0.1794, "step": 6790 }, { "epoch": 3.436078827690753, "grad_norm": 3.084446430206299, "learning_rate": 4.8933333333333335e-05, "loss": 0.3381, "step": 6800 }, { "epoch": 3.436078827690753, "eval_accuracy": 0.8933046902971715, "eval_loss": 0.25619128346443176, "eval_runtime": 23.4504, "eval_samples_per_second": 119.102, "eval_steps_per_second": 14.925, "step": 6800 }, { "epoch": 3.4411318847902983, "grad_norm": 1.4437936544418335, "learning_rate": 4.871111111111111e-05, "loss": 0.2922, "step": 6810 }, { "epoch": 3.446184941889843, "grad_norm": 1.9302983283996582, "learning_rate": 4.848888888888889e-05, "loss": 0.2633, "step": 6820 }, { "epoch": 3.4512379989893884, "grad_norm": 3.164926767349243, "learning_rate": 4.826666666666667e-05, "loss": 0.3575, "step": 6830 }, { "epoch": 3.4562910560889337, "grad_norm": 1.095962405204773, "learning_rate": 4.8044444444444446e-05, "loss": 0.2745, "step": 6840 }, { "epoch": 3.461344113188479, "grad_norm": 3.9413673877716064, "learning_rate": 4.782222222222222e-05, "loss": 0.2937, "step": 6850 }, { "epoch": 3.4663971702880243, "grad_norm": 2.4038617610931396, "learning_rate": 4.76e-05, "loss": 0.3215, "step": 6860 }, { "epoch": 3.4714502273875696, "grad_norm": 0.2632848620414734, "learning_rate": 4.737777777777778e-05, "loss": 0.2838, "step": 6870 }, { "epoch": 3.476503284487115, "grad_norm": 0.4174223840236664, "learning_rate": 4.715555555555556e-05, "loss": 0.3127, "step": 6880 }, { "epoch": 3.4815563415866597, "grad_norm": 2.04986834526062, "learning_rate": 4.6933333333333333e-05, "loss": 0.3236, "step": 6890 }, { "epoch": 3.486609398686205, "grad_norm": 0.6955283284187317, "learning_rate": 4.671111111111111e-05, "loss": 0.2193, "step": 6900 }, { "epoch": 3.486609398686205, "eval_accuracy": 0.891156462585034, "eval_loss": 0.26051437854766846, "eval_runtime": 23.393, "eval_samples_per_second": 119.394, "eval_steps_per_second": 14.962, "step": 6900 }, { "epoch": 3.4916624557857503, "grad_norm": 2.3382861614227295, "learning_rate": 4.651111111111111e-05, "loss": 0.2941, "step": 6910 }, { "epoch": 3.4967155128852956, "grad_norm": 0.3210231363773346, "learning_rate": 4.6288888888888894e-05, "loss": 0.2989, "step": 6920 }, { "epoch": 3.501768569984841, "grad_norm": 2.856682777404785, "learning_rate": 4.606666666666667e-05, "loss": 0.293, "step": 6930 }, { "epoch": 3.506821627084386, "grad_norm": 0.7344805598258972, "learning_rate": 4.584444444444445e-05, "loss": 0.228, "step": 6940 }, { "epoch": 3.5118746841839314, "grad_norm": 1.6938222646713257, "learning_rate": 4.562222222222222e-05, "loss": 0.3437, "step": 6950 }, { "epoch": 3.5169277412834763, "grad_norm": 7.076164245605469, "learning_rate": 4.5400000000000006e-05, "loss": 0.4092, "step": 6960 }, { "epoch": 3.5219807983830216, "grad_norm": 1.6173063516616821, "learning_rate": 4.517777777777778e-05, "loss": 0.3343, "step": 6970 }, { "epoch": 3.527033855482567, "grad_norm": 3.324890613555908, "learning_rate": 4.495555555555556e-05, "loss": 0.2106, "step": 6980 }, { "epoch": 3.532086912582112, "grad_norm": 2.447829484939575, "learning_rate": 4.473333333333334e-05, "loss": 0.2544, "step": 6990 }, { "epoch": 3.5371399696816574, "grad_norm": 1.924780011177063, "learning_rate": 4.451111111111112e-05, "loss": 0.2685, "step": 7000 }, { "epoch": 3.5371399696816574, "eval_accuracy": 0.8822055137844611, "eval_loss": 0.259235680103302, "eval_runtime": 22.4964, "eval_samples_per_second": 124.153, "eval_steps_per_second": 15.558, "step": 7000 }, { "epoch": 3.5421930267812027, "grad_norm": 3.632363796234131, "learning_rate": 4.428888888888889e-05, "loss": 0.4105, "step": 7010 }, { "epoch": 3.547246083880748, "grad_norm": 1.6931264400482178, "learning_rate": 4.406666666666667e-05, "loss": 0.2914, "step": 7020 }, { "epoch": 3.552299140980293, "grad_norm": 0.9192767143249512, "learning_rate": 4.384444444444445e-05, "loss": 0.3363, "step": 7030 }, { "epoch": 3.557352198079838, "grad_norm": 3.0867137908935547, "learning_rate": 4.362222222222223e-05, "loss": 0.2812, "step": 7040 }, { "epoch": 3.5624052551793834, "grad_norm": 2.4090187549591064, "learning_rate": 4.3400000000000005e-05, "loss": 0.3179, "step": 7050 }, { "epoch": 3.5674583122789287, "grad_norm": 2.5159058570861816, "learning_rate": 4.317777777777778e-05, "loss": 0.3918, "step": 7060 }, { "epoch": 3.572511369378474, "grad_norm": 1.5174832344055176, "learning_rate": 4.295555555555556e-05, "loss": 0.3291, "step": 7070 }, { "epoch": 3.5775644264780193, "grad_norm": 2.3133010864257812, "learning_rate": 4.273333333333333e-05, "loss": 0.4057, "step": 7080 }, { "epoch": 3.5826174835775646, "grad_norm": 8.338386535644531, "learning_rate": 4.2511111111111116e-05, "loss": 0.2886, "step": 7090 }, { "epoch": 3.5876705406771094, "grad_norm": 7.558538913726807, "learning_rate": 4.228888888888889e-05, "loss": 0.2867, "step": 7100 }, { "epoch": 3.5876705406771094, "eval_accuracy": 0.8635875402792696, "eval_loss": 0.3182123899459839, "eval_runtime": 22.649, "eval_samples_per_second": 123.317, "eval_steps_per_second": 15.453, "step": 7100 }, { "epoch": 3.5927235977766547, "grad_norm": 0.4639037549495697, "learning_rate": 4.206666666666667e-05, "loss": 0.2857, "step": 7110 }, { "epoch": 3.5977766548762, "grad_norm": 4.637019157409668, "learning_rate": 4.1844444444444444e-05, "loss": 0.2121, "step": 7120 }, { "epoch": 3.6028297119757453, "grad_norm": 3.765474557876587, "learning_rate": 4.162222222222222e-05, "loss": 0.2615, "step": 7130 }, { "epoch": 3.6078827690752906, "grad_norm": 0.471752405166626, "learning_rate": 4.14e-05, "loss": 0.2469, "step": 7140 }, { "epoch": 3.612935826174836, "grad_norm": 0.5334476232528687, "learning_rate": 4.117777777777778e-05, "loss": 0.2722, "step": 7150 }, { "epoch": 3.617988883274381, "grad_norm": 5.297542572021484, "learning_rate": 4.0955555555555556e-05, "loss": 0.2744, "step": 7160 }, { "epoch": 3.623041940373926, "grad_norm": 2.4122414588928223, "learning_rate": 4.073333333333333e-05, "loss": 0.445, "step": 7170 }, { "epoch": 3.6280949974734713, "grad_norm": 0.47642749547958374, "learning_rate": 4.051111111111111e-05, "loss": 0.1814, "step": 7180 }, { "epoch": 3.6331480545730166, "grad_norm": 1.3290168046951294, "learning_rate": 4.028888888888889e-05, "loss": 0.3884, "step": 7190 }, { "epoch": 3.638201111672562, "grad_norm": 1.3797765970230103, "learning_rate": 4.006666666666667e-05, "loss": 0.318, "step": 7200 }, { "epoch": 3.638201111672562, "eval_accuracy": 0.874328678839957, "eval_loss": 0.29881954193115234, "eval_runtime": 21.6349, "eval_samples_per_second": 129.097, "eval_steps_per_second": 16.178, "step": 7200 }, { "epoch": 3.643254168772107, "grad_norm": 0.7499635815620422, "learning_rate": 3.984444444444444e-05, "loss": 0.3527, "step": 7210 }, { "epoch": 3.6483072258716525, "grad_norm": 3.2787723541259766, "learning_rate": 3.962222222222222e-05, "loss": 0.2758, "step": 7220 }, { "epoch": 3.6533602829711977, "grad_norm": 0.3964061439037323, "learning_rate": 3.94e-05, "loss": 0.3183, "step": 7230 }, { "epoch": 3.6584133400707426, "grad_norm": 1.4817743301391602, "learning_rate": 3.917777777777778e-05, "loss": 0.3821, "step": 7240 }, { "epoch": 3.663466397170288, "grad_norm": 1.7741836309432983, "learning_rate": 3.8955555555555555e-05, "loss": 0.3728, "step": 7250 }, { "epoch": 3.668519454269833, "grad_norm": 8.853656768798828, "learning_rate": 3.873333333333333e-05, "loss": 0.1927, "step": 7260 }, { "epoch": 3.6735725113693785, "grad_norm": 0.8767987489700317, "learning_rate": 3.8511111111111114e-05, "loss": 0.2607, "step": 7270 }, { "epoch": 3.6786255684689237, "grad_norm": 0.57962566614151, "learning_rate": 3.828888888888889e-05, "loss": 0.3245, "step": 7280 }, { "epoch": 3.683678625568469, "grad_norm": 2.1979587078094482, "learning_rate": 3.8066666666666666e-05, "loss": 0.3073, "step": 7290 }, { "epoch": 3.6887316826680143, "grad_norm": 3.9322900772094727, "learning_rate": 3.784444444444445e-05, "loss": 0.3088, "step": 7300 }, { "epoch": 3.6887316826680143, "eval_accuracy": 0.8768349445041175, "eval_loss": 0.2870176434516907, "eval_runtime": 22.5248, "eval_samples_per_second": 123.997, "eval_steps_per_second": 15.538, "step": 7300 }, { "epoch": 3.693784739767559, "grad_norm": 2.439561128616333, "learning_rate": 3.7622222222222225e-05, "loss": 0.253, "step": 7310 }, { "epoch": 3.6988377968671045, "grad_norm": 1.8486074209213257, "learning_rate": 3.74e-05, "loss": 0.2314, "step": 7320 }, { "epoch": 3.7038908539666497, "grad_norm": 2.2029669284820557, "learning_rate": 3.717777777777778e-05, "loss": 0.2555, "step": 7330 }, { "epoch": 3.708943911066195, "grad_norm": 1.3750510215759277, "learning_rate": 3.695555555555556e-05, "loss": 0.2541, "step": 7340 }, { "epoch": 3.7139969681657403, "grad_norm": 1.9020977020263672, "learning_rate": 3.6733333333333336e-05, "loss": 0.4256, "step": 7350 }, { "epoch": 3.7190500252652856, "grad_norm": 3.492295026779175, "learning_rate": 3.651111111111111e-05, "loss": 0.2445, "step": 7360 }, { "epoch": 3.724103082364831, "grad_norm": 3.168321132659912, "learning_rate": 3.628888888888889e-05, "loss": 0.3001, "step": 7370 }, { "epoch": 3.7291561394643757, "grad_norm": 1.113955020904541, "learning_rate": 3.606666666666667e-05, "loss": 0.4109, "step": 7380 }, { "epoch": 3.734209196563921, "grad_norm": 0.5008397102355957, "learning_rate": 3.584444444444445e-05, "loss": 0.4284, "step": 7390 }, { "epoch": 3.7392622536634663, "grad_norm": 1.0201084613800049, "learning_rate": 3.5622222222222224e-05, "loss": 0.3531, "step": 7400 }, { "epoch": 3.7392622536634663, "eval_accuracy": 0.8696741854636592, "eval_loss": 0.2923896908760071, "eval_runtime": 21.9275, "eval_samples_per_second": 127.375, "eval_steps_per_second": 15.962, "step": 7400 }, { "epoch": 3.7443153107630116, "grad_norm": 11.952394485473633, "learning_rate": 3.54e-05, "loss": 0.2287, "step": 7410 }, { "epoch": 3.749368367862557, "grad_norm": 2.617206335067749, "learning_rate": 3.517777777777778e-05, "loss": 0.2518, "step": 7420 }, { "epoch": 3.754421424962102, "grad_norm": 2.063462018966675, "learning_rate": 3.495555555555556e-05, "loss": 0.1718, "step": 7430 }, { "epoch": 3.7594744820616475, "grad_norm": 4.8402252197265625, "learning_rate": 3.4733333333333335e-05, "loss": 0.3585, "step": 7440 }, { "epoch": 3.7645275391611923, "grad_norm": 0.1775069236755371, "learning_rate": 3.451111111111111e-05, "loss": 0.1578, "step": 7450 }, { "epoch": 3.7695805962607376, "grad_norm": 2.416515588760376, "learning_rate": 3.4288888888888894e-05, "loss": 0.3968, "step": 7460 }, { "epoch": 3.774633653360283, "grad_norm": 1.1179537773132324, "learning_rate": 3.406666666666667e-05, "loss": 0.3176, "step": 7470 }, { "epoch": 3.779686710459828, "grad_norm": 3.3910624980926514, "learning_rate": 3.3844444444444446e-05, "loss": 0.1796, "step": 7480 }, { "epoch": 3.7847397675593735, "grad_norm": 6.630885124206543, "learning_rate": 3.362222222222222e-05, "loss": 0.1927, "step": 7490 }, { "epoch": 3.7897928246589188, "grad_norm": 0.6718080043792725, "learning_rate": 3.3400000000000005e-05, "loss": 0.2605, "step": 7500 }, { "epoch": 3.7897928246589188, "eval_accuracy": 0.870390261367705, "eval_loss": 0.2942241132259369, "eval_runtime": 21.6662, "eval_samples_per_second": 128.91, "eval_steps_per_second": 16.154, "step": 7500 }, { "epoch": 3.794845881758464, "grad_norm": 3.3835859298706055, "learning_rate": 3.317777777777778e-05, "loss": 0.3342, "step": 7510 }, { "epoch": 3.799898938858009, "grad_norm": 0.3456265330314636, "learning_rate": 3.295555555555556e-05, "loss": 0.2156, "step": 7520 }, { "epoch": 3.804951995957554, "grad_norm": 0.5506961345672607, "learning_rate": 3.2733333333333334e-05, "loss": 0.316, "step": 7530 }, { "epoch": 3.8100050530570995, "grad_norm": 0.2967621088027954, "learning_rate": 3.251111111111112e-05, "loss": 0.2192, "step": 7540 }, { "epoch": 3.8150581101566448, "grad_norm": 2.0887911319732666, "learning_rate": 3.228888888888889e-05, "loss": 0.3395, "step": 7550 }, { "epoch": 3.82011116725619, "grad_norm": 1.0434236526489258, "learning_rate": 3.206666666666667e-05, "loss": 0.2203, "step": 7560 }, { "epoch": 3.8251642243557353, "grad_norm": 1.0598390102386475, "learning_rate": 3.1844444444444445e-05, "loss": 0.3126, "step": 7570 }, { "epoch": 3.8302172814552806, "grad_norm": 0.7207527160644531, "learning_rate": 3.162222222222223e-05, "loss": 0.3728, "step": 7580 }, { "epoch": 3.8352703385548255, "grad_norm": 1.1682573556900024, "learning_rate": 3.1400000000000004e-05, "loss": 0.3624, "step": 7590 }, { "epoch": 3.8403233956543708, "grad_norm": 5.3315348625183105, "learning_rate": 3.117777777777778e-05, "loss": 0.419, "step": 7600 }, { "epoch": 3.8403233956543708, "eval_accuracy": 0.8485499462943072, "eval_loss": 0.3634319305419922, "eval_runtime": 21.5469, "eval_samples_per_second": 129.624, "eval_steps_per_second": 16.244, "step": 7600 }, { "epoch": 3.845376452753916, "grad_norm": 0.7724176645278931, "learning_rate": 3.0955555555555557e-05, "loss": 0.2619, "step": 7610 }, { "epoch": 3.8504295098534613, "grad_norm": 4.522735118865967, "learning_rate": 3.073333333333334e-05, "loss": 0.4648, "step": 7620 }, { "epoch": 3.8554825669530066, "grad_norm": 1.0167016983032227, "learning_rate": 3.0511111111111112e-05, "loss": 0.2751, "step": 7630 }, { "epoch": 3.860535624052552, "grad_norm": 3.7538797855377197, "learning_rate": 3.028888888888889e-05, "loss": 0.1539, "step": 7640 }, { "epoch": 3.865588681152097, "grad_norm": 8.2957763671875, "learning_rate": 3.006666666666667e-05, "loss": 0.2429, "step": 7650 }, { "epoch": 3.870641738251642, "grad_norm": 1.827283501625061, "learning_rate": 2.9844444444444447e-05, "loss": 0.2768, "step": 7660 }, { "epoch": 3.8756947953511873, "grad_norm": 0.781579852104187, "learning_rate": 2.9622222222222224e-05, "loss": 0.3652, "step": 7670 }, { "epoch": 3.8807478524507326, "grad_norm": 5.0481719970703125, "learning_rate": 2.94e-05, "loss": 0.2989, "step": 7680 }, { "epoch": 3.885800909550278, "grad_norm": 6.631250381469727, "learning_rate": 2.9177777777777783e-05, "loss": 0.2492, "step": 7690 }, { "epoch": 3.890853966649823, "grad_norm": 5.6653876304626465, "learning_rate": 2.895555555555556e-05, "loss": 0.264, "step": 7700 }, { "epoch": 3.890853966649823, "eval_accuracy": 0.8628714643752238, "eval_loss": 0.29964956641197205, "eval_runtime": 22.1093, "eval_samples_per_second": 126.327, "eval_steps_per_second": 15.83, "step": 7700 }, { "epoch": 3.8959070237493685, "grad_norm": 0.4400680959224701, "learning_rate": 2.8733333333333335e-05, "loss": 0.1466, "step": 7710 }, { "epoch": 3.9009600808489138, "grad_norm": 1.1072263717651367, "learning_rate": 2.851111111111111e-05, "loss": 0.2983, "step": 7720 }, { "epoch": 3.9060131379484586, "grad_norm": 0.5539777874946594, "learning_rate": 2.8288888888888894e-05, "loss": 0.2874, "step": 7730 }, { "epoch": 3.911066195048004, "grad_norm": 4.213361740112305, "learning_rate": 2.806666666666667e-05, "loss": 0.2965, "step": 7740 }, { "epoch": 3.916119252147549, "grad_norm": 3.2600481510162354, "learning_rate": 2.7844444444444446e-05, "loss": 0.3162, "step": 7750 }, { "epoch": 3.9211723092470945, "grad_norm": 1.8276543617248535, "learning_rate": 2.7622222222222222e-05, "loss": 0.3039, "step": 7760 }, { "epoch": 3.9262253663466398, "grad_norm": 0.9345902800559998, "learning_rate": 2.7400000000000002e-05, "loss": 0.3426, "step": 7770 }, { "epoch": 3.931278423446185, "grad_norm": 3.4985907077789307, "learning_rate": 2.717777777777778e-05, "loss": 0.2709, "step": 7780 }, { "epoch": 3.9363314805457303, "grad_norm": 0.544694721698761, "learning_rate": 2.6955555555555558e-05, "loss": 0.408, "step": 7790 }, { "epoch": 3.941384537645275, "grad_norm": 4.879567623138428, "learning_rate": 2.6733333333333334e-05, "loss": 0.2349, "step": 7800 }, { "epoch": 3.941384537645275, "eval_accuracy": 0.8936627282491945, "eval_loss": 0.24169301986694336, "eval_runtime": 21.8782, "eval_samples_per_second": 127.662, "eval_steps_per_second": 15.998, "step": 7800 }, { "epoch": 3.9464375947448205, "grad_norm": 2.6091549396514893, "learning_rate": 2.6511111111111113e-05, "loss": 0.264, "step": 7810 }, { "epoch": 3.9514906518443658, "grad_norm": 0.46488797664642334, "learning_rate": 2.628888888888889e-05, "loss": 0.2789, "step": 7820 }, { "epoch": 3.956543708943911, "grad_norm": 1.9423531293869019, "learning_rate": 2.6066666666666666e-05, "loss": 0.2615, "step": 7830 }, { "epoch": 3.9615967660434563, "grad_norm": 7.142584323883057, "learning_rate": 2.5844444444444442e-05, "loss": 0.249, "step": 7840 }, { "epoch": 3.9666498231430016, "grad_norm": 5.641912460327148, "learning_rate": 2.5622222222222225e-05, "loss": 0.4524, "step": 7850 }, { "epoch": 3.971702880242547, "grad_norm": 1.00862717628479, "learning_rate": 2.54e-05, "loss": 0.4576, "step": 7860 }, { "epoch": 3.9767559373420918, "grad_norm": 2.411088466644287, "learning_rate": 2.5177777777777777e-05, "loss": 0.2272, "step": 7870 }, { "epoch": 3.981808994441637, "grad_norm": 0.4181084930896759, "learning_rate": 2.4955555555555556e-05, "loss": 0.308, "step": 7880 }, { "epoch": 3.9868620515411823, "grad_norm": 1.7667807340621948, "learning_rate": 2.4733333333333333e-05, "loss": 0.2388, "step": 7890 }, { "epoch": 3.9919151086407276, "grad_norm": 0.6423434019088745, "learning_rate": 2.4511111111111112e-05, "loss": 0.2726, "step": 7900 }, { "epoch": 3.9919151086407276, "eval_accuracy": 0.8517722878625135, "eval_loss": 0.322764128446579, "eval_runtime": 21.7641, "eval_samples_per_second": 128.33, "eval_steps_per_second": 16.081, "step": 7900 }, { "epoch": 3.996968165740273, "grad_norm": 0.9120103716850281, "learning_rate": 2.4288888888888888e-05, "loss": 0.2801, "step": 7910 }, { "epoch": 4.002021222839818, "grad_norm": 2.4902210235595703, "learning_rate": 2.4066666666666668e-05, "loss": 0.264, "step": 7920 }, { "epoch": 4.0070742799393635, "grad_norm": 2.0421929359436035, "learning_rate": 2.3844444444444444e-05, "loss": 0.2189, "step": 7930 }, { "epoch": 4.012127337038908, "grad_norm": 2.8844339847564697, "learning_rate": 2.3622222222222223e-05, "loss": 0.3575, "step": 7940 }, { "epoch": 4.017180394138454, "grad_norm": 4.857579708099365, "learning_rate": 2.3400000000000003e-05, "loss": 0.1671, "step": 7950 }, { "epoch": 4.022233451237999, "grad_norm": 2.1485273838043213, "learning_rate": 2.317777777777778e-05, "loss": 0.2341, "step": 7960 }, { "epoch": 4.027286508337545, "grad_norm": 0.5558630228042603, "learning_rate": 2.295555555555556e-05, "loss": 0.2864, "step": 7970 }, { "epoch": 4.0323395654370895, "grad_norm": 2.4777448177337646, "learning_rate": 2.2733333333333335e-05, "loss": 0.1765, "step": 7980 }, { "epoch": 4.037392622536634, "grad_norm": 2.453984022140503, "learning_rate": 2.2511111111111114e-05, "loss": 0.2702, "step": 7990 }, { "epoch": 4.04244567963618, "grad_norm": 3.945199728012085, "learning_rate": 2.228888888888889e-05, "loss": 0.3398, "step": 8000 }, { "epoch": 4.04244567963618, "eval_accuracy": 0.8897243107769424, "eval_loss": 0.2683921456336975, "eval_runtime": 22.293, "eval_samples_per_second": 125.286, "eval_steps_per_second": 15.7, "step": 8000 }, { "epoch": 4.047498736735725, "grad_norm": 6.585222244262695, "learning_rate": 2.206666666666667e-05, "loss": 0.2531, "step": 8010 }, { "epoch": 4.052551793835271, "grad_norm": 0.7060272693634033, "learning_rate": 2.1844444444444446e-05, "loss": 0.1186, "step": 8020 }, { "epoch": 4.0576048509348155, "grad_norm": 2.3023953437805176, "learning_rate": 2.1622222222222226e-05, "loss": 0.31, "step": 8030 }, { "epoch": 4.062657908034361, "grad_norm": 4.076186656951904, "learning_rate": 2.1400000000000002e-05, "loss": 0.1671, "step": 8040 }, { "epoch": 4.067710965133906, "grad_norm": 2.7586264610290527, "learning_rate": 2.117777777777778e-05, "loss": 0.3629, "step": 8050 }, { "epoch": 4.072764022233451, "grad_norm": 2.3629884719848633, "learning_rate": 2.0955555555555557e-05, "loss": 0.3269, "step": 8060 }, { "epoch": 4.077817079332997, "grad_norm": 0.2406623214483261, "learning_rate": 2.0733333333333334e-05, "loss": 0.2697, "step": 8070 }, { "epoch": 4.0828701364325415, "grad_norm": 1.6978013515472412, "learning_rate": 2.0511111111111113e-05, "loss": 0.2613, "step": 8080 }, { "epoch": 4.087923193532087, "grad_norm": 3.5561013221740723, "learning_rate": 2.028888888888889e-05, "loss": 0.2489, "step": 8090 }, { "epoch": 4.092976250631632, "grad_norm": 12.138863563537598, "learning_rate": 2.0066666666666665e-05, "loss": 0.1933, "step": 8100 }, { "epoch": 4.092976250631632, "eval_accuracy": 0.8918725384890799, "eval_loss": 0.2656717300415039, "eval_runtime": 21.8937, "eval_samples_per_second": 127.571, "eval_steps_per_second": 15.986, "step": 8100 }, { "epoch": 4.098029307731178, "grad_norm": 0.09319577366113663, "learning_rate": 1.9844444444444445e-05, "loss": 0.2288, "step": 8110 }, { "epoch": 4.103082364830723, "grad_norm": 0.20730702579021454, "learning_rate": 1.962222222222222e-05, "loss": 0.1315, "step": 8120 }, { "epoch": 4.1081354219302675, "grad_norm": 0.4735649824142456, "learning_rate": 1.94e-05, "loss": 0.2496, "step": 8130 }, { "epoch": 4.113188479029813, "grad_norm": 0.8919284343719482, "learning_rate": 1.9177777777777777e-05, "loss": 0.2327, "step": 8140 }, { "epoch": 4.118241536129358, "grad_norm": 2.047518014907837, "learning_rate": 1.8955555555555556e-05, "loss": 0.3802, "step": 8150 }, { "epoch": 4.123294593228904, "grad_norm": 2.4581847190856934, "learning_rate": 1.8733333333333332e-05, "loss": 0.5422, "step": 8160 }, { "epoch": 4.128347650328449, "grad_norm": 29.212650299072266, "learning_rate": 1.8511111111111112e-05, "loss": 0.3291, "step": 8170 }, { "epoch": 4.1334007074279935, "grad_norm": 2.1183395385742188, "learning_rate": 1.8288888888888888e-05, "loss": 0.1113, "step": 8180 }, { "epoch": 4.138453764527539, "grad_norm": 0.6621558666229248, "learning_rate": 1.8066666666666668e-05, "loss": 0.279, "step": 8190 }, { "epoch": 4.143506821627084, "grad_norm": 2.743483304977417, "learning_rate": 1.7844444444444444e-05, "loss": 0.435, "step": 8200 }, { "epoch": 4.143506821627084, "eval_accuracy": 0.8972431077694235, "eval_loss": 0.24550849199295044, "eval_runtime": 22.5812, "eval_samples_per_second": 123.687, "eval_steps_per_second": 15.5, "step": 8200 }, { "epoch": 4.14855987872663, "grad_norm": 2.269160747528076, "learning_rate": 1.7622222222222223e-05, "loss": 0.2591, "step": 8210 }, { "epoch": 4.153612935826175, "grad_norm": 4.895859241485596, "learning_rate": 1.74e-05, "loss": 0.2851, "step": 8220 }, { "epoch": 4.15866599292572, "grad_norm": 2.937652349472046, "learning_rate": 1.717777777777778e-05, "loss": 0.3393, "step": 8230 }, { "epoch": 4.163719050025265, "grad_norm": 2.60981822013855, "learning_rate": 1.6955555555555555e-05, "loss": 0.3561, "step": 8240 }, { "epoch": 4.168772107124811, "grad_norm": 2.735541582107544, "learning_rate": 1.6733333333333335e-05, "loss": 0.2485, "step": 8250 }, { "epoch": 4.173825164224356, "grad_norm": 3.868619680404663, "learning_rate": 1.651111111111111e-05, "loss": 0.2773, "step": 8260 }, { "epoch": 4.178878221323901, "grad_norm": 3.7380282878875732, "learning_rate": 1.628888888888889e-05, "loss": 0.3329, "step": 8270 }, { "epoch": 4.183931278423446, "grad_norm": 0.8130131959915161, "learning_rate": 1.606666666666667e-05, "loss": 0.2427, "step": 8280 }, { "epoch": 4.188984335522991, "grad_norm": 0.9071109294891357, "learning_rate": 1.5844444444444446e-05, "loss": 0.3555, "step": 8290 }, { "epoch": 4.194037392622537, "grad_norm": 2.2046091556549072, "learning_rate": 1.5622222222222225e-05, "loss": 0.2373, "step": 8300 }, { "epoch": 4.194037392622537, "eval_accuracy": 0.8689581095596133, "eval_loss": 0.29286259412765503, "eval_runtime": 22.3207, "eval_samples_per_second": 125.13, "eval_steps_per_second": 15.68, "step": 8300 }, { "epoch": 4.199090449722082, "grad_norm": 0.9894067645072937, "learning_rate": 1.54e-05, "loss": 0.2124, "step": 8310 }, { "epoch": 4.204143506821627, "grad_norm": 6.283033847808838, "learning_rate": 1.517777777777778e-05, "loss": 0.1809, "step": 8320 }, { "epoch": 4.209196563921172, "grad_norm": 1.6928883790969849, "learning_rate": 1.4955555555555556e-05, "loss": 0.2545, "step": 8330 }, { "epoch": 4.214249621020717, "grad_norm": 3.008721351623535, "learning_rate": 1.4733333333333335e-05, "loss": 0.1915, "step": 8340 }, { "epoch": 4.219302678120263, "grad_norm": 0.14142248034477234, "learning_rate": 1.4511111111111111e-05, "loss": 0.3754, "step": 8350 }, { "epoch": 4.224355735219808, "grad_norm": 1.5255669355392456, "learning_rate": 1.428888888888889e-05, "loss": 0.2875, "step": 8360 }, { "epoch": 4.2294087923193535, "grad_norm": 2.080801010131836, "learning_rate": 1.4066666666666667e-05, "loss": 0.325, "step": 8370 }, { "epoch": 4.234461849418898, "grad_norm": 4.126662254333496, "learning_rate": 1.3844444444444446e-05, "loss": 0.2149, "step": 8380 }, { "epoch": 4.239514906518444, "grad_norm": 2.3410325050354004, "learning_rate": 1.3622222222222223e-05, "loss": 0.342, "step": 8390 }, { "epoch": 4.244567963617989, "grad_norm": 2.614257335662842, "learning_rate": 1.3400000000000002e-05, "loss": 0.3151, "step": 8400 }, { "epoch": 4.244567963617989, "eval_accuracy": 0.8761188686000716, "eval_loss": 0.27450793981552124, "eval_runtime": 22.3598, "eval_samples_per_second": 124.912, "eval_steps_per_second": 15.653, "step": 8400 }, { "epoch": 4.249621020717534, "grad_norm": 3.3928370475769043, "learning_rate": 1.3177777777777778e-05, "loss": 0.1989, "step": 8410 }, { "epoch": 4.2546740778170795, "grad_norm": 1.081748604774475, "learning_rate": 1.2955555555555556e-05, "loss": 0.1584, "step": 8420 }, { "epoch": 4.259727134916624, "grad_norm": 1.3160364627838135, "learning_rate": 1.2733333333333334e-05, "loss": 0.2706, "step": 8430 }, { "epoch": 4.26478019201617, "grad_norm": 0.348531574010849, "learning_rate": 1.2511111111111112e-05, "loss": 0.2824, "step": 8440 }, { "epoch": 4.269833249115715, "grad_norm": 0.4771299660205841, "learning_rate": 1.228888888888889e-05, "loss": 0.2507, "step": 8450 }, { "epoch": 4.27488630621526, "grad_norm": 4.751070022583008, "learning_rate": 1.2066666666666667e-05, "loss": 0.2523, "step": 8460 }, { "epoch": 4.2799393633148055, "grad_norm": 5.439723968505859, "learning_rate": 1.1844444444444445e-05, "loss": 0.2357, "step": 8470 }, { "epoch": 4.28499242041435, "grad_norm": 4.030198574066162, "learning_rate": 1.1622222222222223e-05, "loss": 0.362, "step": 8480 }, { "epoch": 4.290045477513896, "grad_norm": 5.852336883544922, "learning_rate": 1.1400000000000001e-05, "loss": 0.3724, "step": 8490 }, { "epoch": 4.295098534613441, "grad_norm": 0.802811861038208, "learning_rate": 1.1177777777777779e-05, "loss": 0.2258, "step": 8500 }, { "epoch": 4.295098534613441, "eval_accuracy": 0.8922305764411027, "eval_loss": 0.2485956847667694, "eval_runtime": 22.0752, "eval_samples_per_second": 126.522, "eval_steps_per_second": 15.855, "step": 8500 }, { "epoch": 4.300151591712987, "grad_norm": 3.8925795555114746, "learning_rate": 1.0955555555555557e-05, "loss": 0.1691, "step": 8510 }, { "epoch": 4.3052046488125315, "grad_norm": 3.1423988342285156, "learning_rate": 1.0733333333333334e-05, "loss": 0.2413, "step": 8520 }, { "epoch": 4.310257705912077, "grad_norm": 3.6486198902130127, "learning_rate": 1.0511111111111112e-05, "loss": 0.2423, "step": 8530 }, { "epoch": 4.315310763011622, "grad_norm": 0.24655957520008087, "learning_rate": 1.028888888888889e-05, "loss": 0.3108, "step": 8540 }, { "epoch": 4.320363820111167, "grad_norm": 3.491961717605591, "learning_rate": 1.0066666666666668e-05, "loss": 0.3822, "step": 8550 }, { "epoch": 4.325416877210713, "grad_norm": 3.5754337310791016, "learning_rate": 9.844444444444446e-06, "loss": 0.2048, "step": 8560 }, { "epoch": 4.3304699343102575, "grad_norm": 0.7345679998397827, "learning_rate": 9.622222222222222e-06, "loss": 0.3607, "step": 8570 }, { "epoch": 4.335522991409803, "grad_norm": 1.6499830484390259, "learning_rate": 9.4e-06, "loss": 0.2995, "step": 8580 }, { "epoch": 4.340576048509348, "grad_norm": 0.2334776073694229, "learning_rate": 9.177777777777778e-06, "loss": 0.2736, "step": 8590 }, { "epoch": 4.345629105608893, "grad_norm": 0.7128919363021851, "learning_rate": 8.955555555555555e-06, "loss": 0.2592, "step": 8600 }, { "epoch": 4.345629105608893, "eval_accuracy": 0.8800572860723237, "eval_loss": 0.2695930600166321, "eval_runtime": 30.0494, "eval_samples_per_second": 92.947, "eval_steps_per_second": 11.647, "step": 8600 }, { "epoch": 4.350682162708439, "grad_norm": 0.7076693177223206, "learning_rate": 8.733333333333333e-06, "loss": 0.2665, "step": 8610 }, { "epoch": 4.3557352198079835, "grad_norm": 3.5077743530273438, "learning_rate": 8.511111111111111e-06, "loss": 0.3848, "step": 8620 }, { "epoch": 4.360788276907529, "grad_norm": 3.9970624446868896, "learning_rate": 8.288888888888889e-06, "loss": 0.3145, "step": 8630 }, { "epoch": 4.365841334007074, "grad_norm": 4.022054672241211, "learning_rate": 8.066666666666667e-06, "loss": 0.3285, "step": 8640 }, { "epoch": 4.37089439110662, "grad_norm": 1.1614007949829102, "learning_rate": 7.844444444444445e-06, "loss": 0.1111, "step": 8650 }, { "epoch": 4.375947448206165, "grad_norm": 0.6075769662857056, "learning_rate": 7.6222222222222225e-06, "loss": 0.4019, "step": 8660 }, { "epoch": 4.38100050530571, "grad_norm": 1.6628307104110718, "learning_rate": 7.4e-06, "loss": 0.2537, "step": 8670 }, { "epoch": 4.386053562405255, "grad_norm": 4.175094127655029, "learning_rate": 7.177777777777778e-06, "loss": 0.4741, "step": 8680 }, { "epoch": 4.3911066195048, "grad_norm": 4.80354642868042, "learning_rate": 6.955555555555555e-06, "loss": 0.2563, "step": 8690 }, { "epoch": 4.396159676604346, "grad_norm": 2.0704190731048584, "learning_rate": 6.733333333333333e-06, "loss": 0.2301, "step": 8700 }, { "epoch": 4.396159676604346, "eval_accuracy": 0.8811313999283924, "eval_loss": 0.2719084620475769, "eval_runtime": 22.3616, "eval_samples_per_second": 124.902, "eval_steps_per_second": 15.652, "step": 8700 }, { "epoch": 4.401212733703891, "grad_norm": 0.8465782999992371, "learning_rate": 6.511111111111111e-06, "loss": 0.2421, "step": 8710 }, { "epoch": 4.406265790803436, "grad_norm": 0.5483872294425964, "learning_rate": 6.288888888888889e-06, "loss": 0.2298, "step": 8720 }, { "epoch": 4.411318847902981, "grad_norm": 5.510003089904785, "learning_rate": 6.066666666666667e-06, "loss": 0.2665, "step": 8730 }, { "epoch": 4.416371905002526, "grad_norm": 1.7184778451919556, "learning_rate": 5.844444444444445e-06, "loss": 0.3181, "step": 8740 }, { "epoch": 4.421424962102072, "grad_norm": 2.2276451587677, "learning_rate": 5.622222222222222e-06, "loss": 0.1919, "step": 8750 }, { "epoch": 4.426478019201617, "grad_norm": 2.302110433578491, "learning_rate": 5.4e-06, "loss": 0.2329, "step": 8760 }, { "epoch": 4.431531076301162, "grad_norm": 0.6475743651390076, "learning_rate": 5.177777777777778e-06, "loss": 0.2842, "step": 8770 }, { "epoch": 4.436584133400707, "grad_norm": 0.36002904176712036, "learning_rate": 4.955555555555556e-06, "loss": 0.3685, "step": 8780 }, { "epoch": 4.441637190500253, "grad_norm": 0.969427227973938, "learning_rate": 4.7333333333333335e-06, "loss": 0.1987, "step": 8790 }, { "epoch": 4.446690247599798, "grad_norm": 1.3879456520080566, "learning_rate": 4.511111111111111e-06, "loss": 0.1388, "step": 8800 }, { "epoch": 4.446690247599798, "eval_accuracy": 0.8879341210168278, "eval_loss": 0.26173341274261475, "eval_runtime": 21.9537, "eval_samples_per_second": 127.222, "eval_steps_per_second": 15.943, "step": 8800 }, { "epoch": 4.4517433046993435, "grad_norm": 0.42694732546806335, "learning_rate": 4.288888888888889e-06, "loss": 0.2792, "step": 8810 }, { "epoch": 4.456796361798888, "grad_norm": 4.3826584815979, "learning_rate": 4.066666666666666e-06, "loss": 0.2924, "step": 8820 }, { "epoch": 4.461849418898433, "grad_norm": 0.43947136402130127, "learning_rate": 3.844444444444445e-06, "loss": 0.1683, "step": 8830 }, { "epoch": 4.466902475997979, "grad_norm": 2.782113790512085, "learning_rate": 3.6222222222222226e-06, "loss": 0.2266, "step": 8840 }, { "epoch": 4.471955533097524, "grad_norm": 4.082117080688477, "learning_rate": 3.4000000000000005e-06, "loss": 0.2909, "step": 8850 }, { "epoch": 4.4770085901970695, "grad_norm": 2.0453526973724365, "learning_rate": 3.1777777777777783e-06, "loss": 0.4169, "step": 8860 }, { "epoch": 4.482061647296614, "grad_norm": 1.80349862575531, "learning_rate": 2.9555555555555557e-06, "loss": 0.3351, "step": 8870 }, { "epoch": 4.487114704396159, "grad_norm": 4.107021808624268, "learning_rate": 2.7333333333333336e-06, "loss": 0.3326, "step": 8880 }, { "epoch": 4.492167761495705, "grad_norm": 0.3571154773235321, "learning_rate": 2.5111111111111114e-06, "loss": 0.2998, "step": 8890 }, { "epoch": 4.49722081859525, "grad_norm": 1.5004687309265137, "learning_rate": 2.2888888888888892e-06, "loss": 0.3242, "step": 8900 }, { "epoch": 4.49722081859525, "eval_accuracy": 0.8915145005370569, "eval_loss": 0.2542950510978699, "eval_runtime": 21.383, "eval_samples_per_second": 130.618, "eval_steps_per_second": 16.368, "step": 8900 }, { "epoch": 4.5022738756947955, "grad_norm": 2.2130260467529297, "learning_rate": 2.0666666666666666e-06, "loss": 0.2592, "step": 8910 }, { "epoch": 4.50732693279434, "grad_norm": 0.3435809910297394, "learning_rate": 1.8444444444444445e-06, "loss": 0.1236, "step": 8920 }, { "epoch": 4.512379989893886, "grad_norm": 0.44044575095176697, "learning_rate": 1.622222222222222e-06, "loss": 0.2928, "step": 8930 }, { "epoch": 4.517433046993431, "grad_norm": 3.092510938644409, "learning_rate": 1.4000000000000001e-06, "loss": 0.1894, "step": 8940 }, { "epoch": 4.522486104092977, "grad_norm": 2.642177104949951, "learning_rate": 1.1777777777777778e-06, "loss": 0.1698, "step": 8950 }, { "epoch": 4.5275391611925215, "grad_norm": 2.354233503341675, "learning_rate": 9.555555555555556e-07, "loss": 0.3587, "step": 8960 }, { "epoch": 4.532592218292066, "grad_norm": 3.531097888946533, "learning_rate": 7.333333333333333e-07, "loss": 0.2285, "step": 8970 }, { "epoch": 4.537645275391612, "grad_norm": 2.694009304046631, "learning_rate": 5.111111111111112e-07, "loss": 0.1547, "step": 8980 }, { "epoch": 4.542698332491157, "grad_norm": 1.5706273317337036, "learning_rate": 2.888888888888889e-07, "loss": 0.3346, "step": 8990 }, { "epoch": 4.547751389590703, "grad_norm": 0.9025068283081055, "learning_rate": 6.666666666666667e-08, "loss": 0.1693, "step": 9000 }, { "epoch": 4.547751389590703, "eval_accuracy": 0.8879341210168278, "eval_loss": 0.26023030281066895, "eval_runtime": 22.2029, "eval_samples_per_second": 125.794, "eval_steps_per_second": 15.764, "step": 9000 }, { "epoch": 4.547751389590703, "step": 9000, "total_flos": 5.577253476541415e+18, "train_loss": 0.37995564444859825, "train_runtime": 4981.1969, "train_samples_per_second": 14.454, "train_steps_per_second": 1.807 } ], "logging_steps": 10, "max_steps": 9000, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.577253476541415e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }