{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 9676, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0010334849111202976, "grad_norm": 3.4666478633880615, "learning_rate": 0.0006200909466721785, "loss": 7.9994, "step": 10 }, { "epoch": 0.002066969822240595, "grad_norm": 1.0015512704849243, "learning_rate": 0.001240181893344357, "loss": 7.0943, "step": 20 }, { "epoch": 0.0031004547333608927, "grad_norm": 0.8696095943450928, "learning_rate": 0.0018602728400165356, "loss": 6.3072, "step": 30 }, { "epoch": 0.00413393964448119, "grad_norm": 0.32349398732185364, "learning_rate": 0.002480363786688714, "loss": 5.9694, "step": 40 }, { "epoch": 0.005167424555601488, "grad_norm": 1.6740758419036865, "learning_rate": 0.0031004547333608927, "loss": 5.8536, "step": 50 }, { "epoch": 0.0062009094667217855, "grad_norm": 1.6459004878997803, "learning_rate": 0.003720545680033071, "loss": 5.7039, "step": 60 }, { "epoch": 0.007234394377842084, "grad_norm": 0.48763737082481384, "learning_rate": 0.0043406366267052495, "loss": 5.4789, "step": 70 }, { "epoch": 0.00826787928896238, "grad_norm": 0.4977681338787079, "learning_rate": 0.004960727573377428, "loss": 5.269, "step": 80 }, { "epoch": 0.009301364200082679, "grad_norm": 0.5395390391349792, "learning_rate": 0.005580818520049607, "loss": 5.0768, "step": 90 }, { "epoch": 0.010334849111202976, "grad_norm": 0.4573260247707367, "learning_rate": 0.006, "loss": 4.9493, "step": 100 }, { "epoch": 0.011368334022323274, "grad_norm": 0.5551472902297974, "learning_rate": 0.006, "loss": 4.7982, "step": 110 }, { "epoch": 0.012401818933443571, "grad_norm": 0.4980376958847046, "learning_rate": 0.006, "loss": 4.6285, "step": 120 }, { "epoch": 0.01343530384456387, "grad_norm": 0.4951136112213135, "learning_rate": 0.006, "loss": 4.5176, "step": 130 }, { "epoch": 0.014468788755684168, "grad_norm": 0.5360985398292542, "learning_rate": 0.006, "loss": 4.3572, "step": 140 }, { "epoch": 0.015502273666804465, "grad_norm": 0.4469200670719147, "learning_rate": 0.006, "loss": 4.214, "step": 150 }, { "epoch": 0.01653575857792476, "grad_norm": 0.6157914400100708, "learning_rate": 0.006, "loss": 4.0854, "step": 160 }, { "epoch": 0.01756924348904506, "grad_norm": 0.4493155777454376, "learning_rate": 0.006, "loss": 3.9583, "step": 170 }, { "epoch": 0.018602728400165358, "grad_norm": 0.6523879766464233, "learning_rate": 0.006, "loss": 3.8434, "step": 180 }, { "epoch": 0.019636213311285657, "grad_norm": 0.7103180885314941, "learning_rate": 0.006, "loss": 3.7414, "step": 190 }, { "epoch": 0.02066969822240595, "grad_norm": 0.44575124979019165, "learning_rate": 0.006, "loss": 3.6829, "step": 200 }, { "epoch": 0.02170318313352625, "grad_norm": 0.5272648334503174, "learning_rate": 0.006, "loss": 3.6056, "step": 210 }, { "epoch": 0.02273666804464655, "grad_norm": 0.4514292776584625, "learning_rate": 0.006, "loss": 3.5279, "step": 220 }, { "epoch": 0.023770152955766847, "grad_norm": 0.39017120003700256, "learning_rate": 0.006, "loss": 3.4585, "step": 230 }, { "epoch": 0.024803637866887142, "grad_norm": 0.4593961536884308, "learning_rate": 0.006, "loss": 3.4018, "step": 240 }, { "epoch": 0.02583712277800744, "grad_norm": 0.48860013484954834, "learning_rate": 0.006, "loss": 3.3523, "step": 250 }, { "epoch": 0.02687060768912774, "grad_norm": 0.36506471037864685, "learning_rate": 0.006, "loss": 3.3124, "step": 260 }, { "epoch": 0.027904092600248037, "grad_norm": 0.3506007194519043, "learning_rate": 0.006, "loss": 3.2678, "step": 270 }, { "epoch": 0.028937577511368336, "grad_norm": 0.3703238070011139, "learning_rate": 0.006, "loss": 3.2231, "step": 280 }, { "epoch": 0.02997106242248863, "grad_norm": 0.45603325963020325, "learning_rate": 0.006, "loss": 3.1847, "step": 290 }, { "epoch": 0.03100454733360893, "grad_norm": 0.37536856532096863, "learning_rate": 0.006, "loss": 3.1585, "step": 300 }, { "epoch": 0.032038032244729224, "grad_norm": 0.3810868561267853, "learning_rate": 0.006, "loss": 3.1063, "step": 310 }, { "epoch": 0.03307151715584952, "grad_norm": 0.4568467438220978, "learning_rate": 0.006, "loss": 3.0649, "step": 320 }, { "epoch": 0.03410500206696982, "grad_norm": 0.36088162660598755, "learning_rate": 0.006, "loss": 3.0495, "step": 330 }, { "epoch": 0.03513848697809012, "grad_norm": 0.3663162887096405, "learning_rate": 0.006, "loss": 3.0188, "step": 340 }, { "epoch": 0.03617197188921042, "grad_norm": 0.5975127220153809, "learning_rate": 0.006, "loss": 2.994, "step": 350 }, { "epoch": 0.037205456800330716, "grad_norm": 0.3454865515232086, "learning_rate": 0.006, "loss": 2.9814, "step": 360 }, { "epoch": 0.038238941711451015, "grad_norm": 0.4206918179988861, "learning_rate": 0.006, "loss": 2.9306, "step": 370 }, { "epoch": 0.03927242662257131, "grad_norm": 0.3579295575618744, "learning_rate": 0.006, "loss": 2.9165, "step": 380 }, { "epoch": 0.04030591153369161, "grad_norm": 0.3658403158187866, "learning_rate": 0.006, "loss": 2.9103, "step": 390 }, { "epoch": 0.0413393964448119, "grad_norm": 0.3459131717681885, "learning_rate": 0.006, "loss": 2.8628, "step": 400 }, { "epoch": 0.0423728813559322, "grad_norm": 0.4726301431655884, "learning_rate": 0.006, "loss": 2.8334, "step": 410 }, { "epoch": 0.0434063662670525, "grad_norm": 0.3717091381549835, "learning_rate": 0.006, "loss": 2.8188, "step": 420 }, { "epoch": 0.0444398511781728, "grad_norm": 0.3234303295612335, "learning_rate": 0.006, "loss": 2.8076, "step": 430 }, { "epoch": 0.0454733360892931, "grad_norm": 0.4717100262641907, "learning_rate": 0.006, "loss": 2.7871, "step": 440 }, { "epoch": 0.046506821000413395, "grad_norm": 0.35768577456474304, "learning_rate": 0.006, "loss": 2.7601, "step": 450 }, { "epoch": 0.047540305911533694, "grad_norm": 0.3499695062637329, "learning_rate": 0.006, "loss": 2.73, "step": 460 }, { "epoch": 0.04857379082265399, "grad_norm": 0.4336409866809845, "learning_rate": 0.006, "loss": 2.7235, "step": 470 }, { "epoch": 0.049607275733774284, "grad_norm": 0.39916467666625977, "learning_rate": 0.006, "loss": 2.7007, "step": 480 }, { "epoch": 0.05064076064489458, "grad_norm": 0.3249908983707428, "learning_rate": 0.006, "loss": 2.6705, "step": 490 }, { "epoch": 0.05167424555601488, "grad_norm": 0.36379745602607727, "learning_rate": 0.006, "loss": 2.6393, "step": 500 }, { "epoch": 0.05270773046713518, "grad_norm": 0.3417607843875885, "learning_rate": 0.006, "loss": 2.6767, "step": 510 }, { "epoch": 0.05374121537825548, "grad_norm": 0.2848694920539856, "learning_rate": 0.006, "loss": 2.6244, "step": 520 }, { "epoch": 0.054774700289375776, "grad_norm": 0.2725778818130493, "learning_rate": 0.006, "loss": 2.5704, "step": 530 }, { "epoch": 0.055808185200496074, "grad_norm": 0.2887585461139679, "learning_rate": 0.006, "loss": 2.5508, "step": 540 }, { "epoch": 0.05684167011161637, "grad_norm": 0.422413170337677, "learning_rate": 0.006, "loss": 2.5433, "step": 550 }, { "epoch": 0.05787515502273667, "grad_norm": 0.5642419457435608, "learning_rate": 0.006, "loss": 2.5267, "step": 560 }, { "epoch": 0.05890863993385696, "grad_norm": 0.33949220180511475, "learning_rate": 0.006, "loss": 2.4942, "step": 570 }, { "epoch": 0.05994212484497726, "grad_norm": 0.36603158712387085, "learning_rate": 0.006, "loss": 2.4755, "step": 580 }, { "epoch": 0.06097560975609756, "grad_norm": 0.3936167061328888, "learning_rate": 0.006, "loss": 2.4358, "step": 590 }, { "epoch": 0.06200909466721786, "grad_norm": 0.46108072996139526, "learning_rate": 0.006, "loss": 2.4312, "step": 600 }, { "epoch": 0.06304257957833816, "grad_norm": 0.3659123480319977, "learning_rate": 0.006, "loss": 2.4023, "step": 610 }, { "epoch": 0.06407606448945845, "grad_norm": 0.4888301193714142, "learning_rate": 0.006, "loss": 2.3887, "step": 620 }, { "epoch": 0.06510954940057875, "grad_norm": 0.33769381046295166, "learning_rate": 0.006, "loss": 2.3645, "step": 630 }, { "epoch": 0.06614303431169904, "grad_norm": 0.4479112923145294, "learning_rate": 0.006, "loss": 2.3123, "step": 640 }, { "epoch": 0.06717651922281935, "grad_norm": 0.3715905249118805, "learning_rate": 0.006, "loss": 2.3009, "step": 650 }, { "epoch": 0.06821000413393964, "grad_norm": 0.43582767248153687, "learning_rate": 0.006, "loss": 2.3256, "step": 660 }, { "epoch": 0.06924348904505995, "grad_norm": 0.295462042093277, "learning_rate": 0.006, "loss": 2.281, "step": 670 }, { "epoch": 0.07027697395618024, "grad_norm": 0.2555060386657715, "learning_rate": 0.006, "loss": 2.2711, "step": 680 }, { "epoch": 0.07131045886730054, "grad_norm": 0.2812485992908478, "learning_rate": 0.006, "loss": 2.2392, "step": 690 }, { "epoch": 0.07234394377842084, "grad_norm": 0.2797698974609375, "learning_rate": 0.006, "loss": 2.2316, "step": 700 }, { "epoch": 0.07337742868954113, "grad_norm": 0.4201503396034241, "learning_rate": 0.006, "loss": 2.1894, "step": 710 }, { "epoch": 0.07441091360066143, "grad_norm": 0.48455363512039185, "learning_rate": 0.006, "loss": 2.1682, "step": 720 }, { "epoch": 0.07544439851178172, "grad_norm": 0.4906412959098816, "learning_rate": 0.006, "loss": 2.1594, "step": 730 }, { "epoch": 0.07647788342290203, "grad_norm": 0.3227013051509857, "learning_rate": 0.006, "loss": 2.148, "step": 740 }, { "epoch": 0.07751136833402232, "grad_norm": 0.24952304363250732, "learning_rate": 0.006, "loss": 2.1323, "step": 750 }, { "epoch": 0.07854485324514263, "grad_norm": 0.4688262641429901, "learning_rate": 0.006, "loss": 2.0781, "step": 760 }, { "epoch": 0.07957833815626292, "grad_norm": 0.3051510453224182, "learning_rate": 0.006, "loss": 2.1005, "step": 770 }, { "epoch": 0.08061182306738322, "grad_norm": 0.303608775138855, "learning_rate": 0.006, "loss": 2.0619, "step": 780 }, { "epoch": 0.08164530797850351, "grad_norm": 0.41049811244010925, "learning_rate": 0.006, "loss": 2.069, "step": 790 }, { "epoch": 0.0826787928896238, "grad_norm": 0.25236430764198303, "learning_rate": 0.006, "loss": 2.0536, "step": 800 }, { "epoch": 0.08371227780074411, "grad_norm": 0.2919696569442749, "learning_rate": 0.006, "loss": 2.0221, "step": 810 }, { "epoch": 0.0847457627118644, "grad_norm": 0.3965113162994385, "learning_rate": 0.006, "loss": 2.0008, "step": 820 }, { "epoch": 0.08577924762298471, "grad_norm": 0.3784347176551819, "learning_rate": 0.006, "loss": 2.0041, "step": 830 }, { "epoch": 0.086812732534105, "grad_norm": 0.32382646203041077, "learning_rate": 0.006, "loss": 1.9987, "step": 840 }, { "epoch": 0.0878462174452253, "grad_norm": 0.28474798798561096, "learning_rate": 0.006, "loss": 1.9859, "step": 850 }, { "epoch": 0.0888797023563456, "grad_norm": 0.35125505924224854, "learning_rate": 0.006, "loss": 1.9719, "step": 860 }, { "epoch": 0.08991318726746589, "grad_norm": 0.2959253787994385, "learning_rate": 0.006, "loss": 1.9691, "step": 870 }, { "epoch": 0.0909466721785862, "grad_norm": 0.2681068181991577, "learning_rate": 0.006, "loss": 1.946, "step": 880 }, { "epoch": 0.09198015708970649, "grad_norm": 0.2882836163043976, "learning_rate": 0.006, "loss": 1.9227, "step": 890 }, { "epoch": 0.09301364200082679, "grad_norm": 0.2792555093765259, "learning_rate": 0.006, "loss": 1.9067, "step": 900 }, { "epoch": 0.09404712691194708, "grad_norm": 0.31801220774650574, "learning_rate": 0.006, "loss": 1.8997, "step": 910 }, { "epoch": 0.09508061182306739, "grad_norm": 0.3104709982872009, "learning_rate": 0.006, "loss": 1.909, "step": 920 }, { "epoch": 0.09611409673418768, "grad_norm": 0.4227132499217987, "learning_rate": 0.006, "loss": 1.8898, "step": 930 }, { "epoch": 0.09714758164530798, "grad_norm": 0.2799486815929413, "learning_rate": 0.006, "loss": 1.8799, "step": 940 }, { "epoch": 0.09818106655642828, "grad_norm": 0.25304150581359863, "learning_rate": 0.006, "loss": 1.8542, "step": 950 }, { "epoch": 0.09921455146754857, "grad_norm": 0.40398165583610535, "learning_rate": 0.006, "loss": 1.8495, "step": 960 }, { "epoch": 0.10024803637866887, "grad_norm": 0.24255605041980743, "learning_rate": 0.006, "loss": 1.8279, "step": 970 }, { "epoch": 0.10128152128978916, "grad_norm": 0.251412957906723, "learning_rate": 0.006, "loss": 1.8448, "step": 980 }, { "epoch": 0.10231500620090947, "grad_norm": 0.33690279722213745, "learning_rate": 0.006, "loss": 1.8345, "step": 990 }, { "epoch": 0.10334849111202976, "grad_norm": 0.3986241817474365, "learning_rate": 0.006, "loss": 1.8272, "step": 1000 }, { "epoch": 0.10438197602315007, "grad_norm": 0.275290310382843, "learning_rate": 0.006, "loss": 1.8177, "step": 1010 }, { "epoch": 0.10541546093427036, "grad_norm": 0.24022069573402405, "learning_rate": 0.006, "loss": 1.8, "step": 1020 }, { "epoch": 0.10644894584539066, "grad_norm": 0.22507286071777344, "learning_rate": 0.006, "loss": 1.7825, "step": 1030 }, { "epoch": 0.10748243075651095, "grad_norm": 0.2965947985649109, "learning_rate": 0.006, "loss": 1.778, "step": 1040 }, { "epoch": 0.10851591566763125, "grad_norm": 0.3291703164577484, "learning_rate": 0.006, "loss": 1.7864, "step": 1050 }, { "epoch": 0.10954940057875155, "grad_norm": 0.2484293133020401, "learning_rate": 0.006, "loss": 1.7649, "step": 1060 }, { "epoch": 0.11058288548987184, "grad_norm": 0.30289486050605774, "learning_rate": 0.006, "loss": 1.7756, "step": 1070 }, { "epoch": 0.11161637040099215, "grad_norm": 0.2765989601612091, "learning_rate": 0.006, "loss": 1.7458, "step": 1080 }, { "epoch": 0.11264985531211244, "grad_norm": 0.24408799409866333, "learning_rate": 0.006, "loss": 1.7598, "step": 1090 }, { "epoch": 0.11368334022323275, "grad_norm": 0.2399667352437973, "learning_rate": 0.006, "loss": 1.754, "step": 1100 }, { "epoch": 0.11471682513435304, "grad_norm": 0.2846289873123169, "learning_rate": 0.006, "loss": 1.7406, "step": 1110 }, { "epoch": 0.11575031004547334, "grad_norm": 0.26227328181266785, "learning_rate": 0.006, "loss": 1.7336, "step": 1120 }, { "epoch": 0.11678379495659363, "grad_norm": 0.2673957347869873, "learning_rate": 0.006, "loss": 1.7281, "step": 1130 }, { "epoch": 0.11781727986771393, "grad_norm": 0.3353629410266876, "learning_rate": 0.006, "loss": 1.7169, "step": 1140 }, { "epoch": 0.11885076477883423, "grad_norm": 0.21159133315086365, "learning_rate": 0.006, "loss": 1.7016, "step": 1150 }, { "epoch": 0.11988424968995452, "grad_norm": 0.24120958149433136, "learning_rate": 0.006, "loss": 1.7033, "step": 1160 }, { "epoch": 0.12091773460107483, "grad_norm": 0.26428887248039246, "learning_rate": 0.006, "loss": 1.6867, "step": 1170 }, { "epoch": 0.12195121951219512, "grad_norm": 0.27754953503608704, "learning_rate": 0.006, "loss": 1.6901, "step": 1180 }, { "epoch": 0.12298470442331542, "grad_norm": 0.36125853657722473, "learning_rate": 0.006, "loss": 1.6808, "step": 1190 }, { "epoch": 0.12401818933443572, "grad_norm": 0.23145093023777008, "learning_rate": 0.006, "loss": 1.6946, "step": 1200 }, { "epoch": 0.12505167424555602, "grad_norm": 0.22397373616695404, "learning_rate": 0.006, "loss": 1.6619, "step": 1210 }, { "epoch": 0.1260851591566763, "grad_norm": 0.2219894528388977, "learning_rate": 0.006, "loss": 1.6483, "step": 1220 }, { "epoch": 0.1271186440677966, "grad_norm": 0.2552240788936615, "learning_rate": 0.006, "loss": 1.6574, "step": 1230 }, { "epoch": 0.1281521289789169, "grad_norm": 0.4382542073726654, "learning_rate": 0.006, "loss": 1.663, "step": 1240 }, { "epoch": 0.12918561389003722, "grad_norm": 0.2551495134830475, "learning_rate": 0.006, "loss": 1.6543, "step": 1250 }, { "epoch": 0.1302190988011575, "grad_norm": 0.21785561740398407, "learning_rate": 0.006, "loss": 1.6479, "step": 1260 }, { "epoch": 0.1312525837122778, "grad_norm": 0.25551608204841614, "learning_rate": 0.006, "loss": 1.6303, "step": 1270 }, { "epoch": 0.1322860686233981, "grad_norm": 0.30908089876174927, "learning_rate": 0.006, "loss": 1.6388, "step": 1280 }, { "epoch": 0.1333195535345184, "grad_norm": 0.39932873845100403, "learning_rate": 0.006, "loss": 1.6192, "step": 1290 }, { "epoch": 0.1343530384456387, "grad_norm": 0.2735249102115631, "learning_rate": 0.006, "loss": 1.6327, "step": 1300 }, { "epoch": 0.135386523356759, "grad_norm": 0.23229070007801056, "learning_rate": 0.006, "loss": 1.6278, "step": 1310 }, { "epoch": 0.13642000826787928, "grad_norm": 0.29770010709762573, "learning_rate": 0.006, "loss": 1.6137, "step": 1320 }, { "epoch": 0.13745349317899958, "grad_norm": 0.2619315981864929, "learning_rate": 0.006, "loss": 1.6071, "step": 1330 }, { "epoch": 0.1384869780901199, "grad_norm": 0.2108081877231598, "learning_rate": 0.006, "loss": 1.6189, "step": 1340 }, { "epoch": 0.13952046300124019, "grad_norm": 0.2384573072195053, "learning_rate": 0.006, "loss": 1.6035, "step": 1350 }, { "epoch": 0.14055394791236048, "grad_norm": 0.2051675170660019, "learning_rate": 0.006, "loss": 1.5866, "step": 1360 }, { "epoch": 0.14158743282348077, "grad_norm": 0.3068487346172333, "learning_rate": 0.006, "loss": 1.5727, "step": 1370 }, { "epoch": 0.1426209177346011, "grad_norm": 0.24623431265354156, "learning_rate": 0.006, "loss": 1.5786, "step": 1380 }, { "epoch": 0.14365440264572138, "grad_norm": 0.3030165731906891, "learning_rate": 0.006, "loss": 1.5793, "step": 1390 }, { "epoch": 0.14468788755684167, "grad_norm": 0.236822709441185, "learning_rate": 0.006, "loss": 1.5737, "step": 1400 }, { "epoch": 0.14572137246796196, "grad_norm": 0.4476490318775177, "learning_rate": 0.006, "loss": 1.5743, "step": 1410 }, { "epoch": 0.14675485737908225, "grad_norm": 0.3274894952774048, "learning_rate": 0.006, "loss": 1.5579, "step": 1420 }, { "epoch": 0.14778834229020257, "grad_norm": 0.2623043358325958, "learning_rate": 0.006, "loss": 1.5666, "step": 1430 }, { "epoch": 0.14882182720132286, "grad_norm": 0.23737861216068268, "learning_rate": 0.006, "loss": 1.553, "step": 1440 }, { "epoch": 0.14985531211244316, "grad_norm": 0.23686368763446808, "learning_rate": 0.006, "loss": 1.5598, "step": 1450 }, { "epoch": 0.15088879702356345, "grad_norm": 0.2711767256259918, "learning_rate": 0.006, "loss": 1.5508, "step": 1460 }, { "epoch": 0.15192228193468377, "grad_norm": 0.1919260323047638, "learning_rate": 0.006, "loss": 1.5364, "step": 1470 }, { "epoch": 0.15295576684580406, "grad_norm": 0.24973885715007782, "learning_rate": 0.006, "loss": 1.5444, "step": 1480 }, { "epoch": 0.15398925175692435, "grad_norm": 0.2096351981163025, "learning_rate": 0.006, "loss": 1.5422, "step": 1490 }, { "epoch": 0.15502273666804464, "grad_norm": 0.28911715745925903, "learning_rate": 0.006, "loss": 1.5562, "step": 1500 }, { "epoch": 0.15605622157916493, "grad_norm": 0.2608921229839325, "learning_rate": 0.006, "loss": 1.5153, "step": 1510 }, { "epoch": 0.15708970649028525, "grad_norm": 0.2605489194393158, "learning_rate": 0.006, "loss": 1.5262, "step": 1520 }, { "epoch": 0.15812319140140554, "grad_norm": 0.2948496341705322, "learning_rate": 0.006, "loss": 1.5318, "step": 1530 }, { "epoch": 0.15915667631252584, "grad_norm": 0.33066225051879883, "learning_rate": 0.006, "loss": 1.5138, "step": 1540 }, { "epoch": 0.16019016122364613, "grad_norm": 0.2134758085012436, "learning_rate": 0.006, "loss": 1.5199, "step": 1550 }, { "epoch": 0.16122364613476645, "grad_norm": 0.22920167446136475, "learning_rate": 0.006, "loss": 1.5084, "step": 1560 }, { "epoch": 0.16225713104588674, "grad_norm": 0.22044587135314941, "learning_rate": 0.006, "loss": 1.5254, "step": 1570 }, { "epoch": 0.16329061595700703, "grad_norm": 0.28658127784729004, "learning_rate": 0.006, "loss": 1.5149, "step": 1580 }, { "epoch": 0.16432410086812732, "grad_norm": 0.1932067573070526, "learning_rate": 0.006, "loss": 1.4943, "step": 1590 }, { "epoch": 0.1653575857792476, "grad_norm": 0.19694995880126953, "learning_rate": 0.006, "loss": 1.4871, "step": 1600 }, { "epoch": 0.16639107069036793, "grad_norm": 0.2092551738023758, "learning_rate": 0.006, "loss": 1.481, "step": 1610 }, { "epoch": 0.16742455560148822, "grad_norm": 0.20461459457874298, "learning_rate": 0.006, "loss": 1.4958, "step": 1620 }, { "epoch": 0.16845804051260851, "grad_norm": 0.2160821110010147, "learning_rate": 0.006, "loss": 1.5013, "step": 1630 }, { "epoch": 0.1694915254237288, "grad_norm": 0.23302793502807617, "learning_rate": 0.006, "loss": 1.4991, "step": 1640 }, { "epoch": 0.1705250103348491, "grad_norm": 0.20844736695289612, "learning_rate": 0.006, "loss": 1.4765, "step": 1650 }, { "epoch": 0.17155849524596942, "grad_norm": 0.18730531632900238, "learning_rate": 0.006, "loss": 1.4732, "step": 1660 }, { "epoch": 0.1725919801570897, "grad_norm": 0.2798612415790558, "learning_rate": 0.006, "loss": 1.4809, "step": 1670 }, { "epoch": 0.17362546506821, "grad_norm": 0.17376913130283356, "learning_rate": 0.006, "loss": 1.4847, "step": 1680 }, { "epoch": 0.1746589499793303, "grad_norm": 0.1984642595052719, "learning_rate": 0.006, "loss": 1.4677, "step": 1690 }, { "epoch": 0.1756924348904506, "grad_norm": 0.18918636441230774, "learning_rate": 0.006, "loss": 1.4623, "step": 1700 }, { "epoch": 0.1767259198015709, "grad_norm": 0.22821608185768127, "learning_rate": 0.006, "loss": 1.4836, "step": 1710 }, { "epoch": 0.1777594047126912, "grad_norm": 0.2023961991071701, "learning_rate": 0.006, "loss": 1.4665, "step": 1720 }, { "epoch": 0.17879288962381149, "grad_norm": 0.21143494546413422, "learning_rate": 0.006, "loss": 1.4637, "step": 1730 }, { "epoch": 0.17982637453493178, "grad_norm": 0.18367299437522888, "learning_rate": 0.006, "loss": 1.4564, "step": 1740 }, { "epoch": 0.1808598594460521, "grad_norm": 0.18500660359859467, "learning_rate": 0.006, "loss": 1.4485, "step": 1750 }, { "epoch": 0.1818933443571724, "grad_norm": 0.208240807056427, "learning_rate": 0.006, "loss": 1.449, "step": 1760 }, { "epoch": 0.18292682926829268, "grad_norm": 0.28708386421203613, "learning_rate": 0.006, "loss": 1.4514, "step": 1770 }, { "epoch": 0.18396031417941297, "grad_norm": 0.23758843541145325, "learning_rate": 0.006, "loss": 1.4376, "step": 1780 }, { "epoch": 0.1849937990905333, "grad_norm": 0.23937605321407318, "learning_rate": 0.006, "loss": 1.4406, "step": 1790 }, { "epoch": 0.18602728400165358, "grad_norm": 0.44142553210258484, "learning_rate": 0.006, "loss": 1.4331, "step": 1800 }, { "epoch": 0.18706076891277387, "grad_norm": 0.1951988786458969, "learning_rate": 0.006, "loss": 1.4443, "step": 1810 }, { "epoch": 0.18809425382389416, "grad_norm": 0.22376669943332672, "learning_rate": 0.006, "loss": 1.4345, "step": 1820 }, { "epoch": 0.18912773873501446, "grad_norm": 0.24400733411312103, "learning_rate": 0.006, "loss": 1.4405, "step": 1830 }, { "epoch": 0.19016122364613477, "grad_norm": 0.20870748162269592, "learning_rate": 0.006, "loss": 1.4291, "step": 1840 }, { "epoch": 0.19119470855725507, "grad_norm": 0.2239631861448288, "learning_rate": 0.006, "loss": 1.4269, "step": 1850 }, { "epoch": 0.19222819346837536, "grad_norm": 0.24455626308918, "learning_rate": 0.006, "loss": 1.4146, "step": 1860 }, { "epoch": 0.19326167837949565, "grad_norm": 0.25358930230140686, "learning_rate": 0.006, "loss": 1.4431, "step": 1870 }, { "epoch": 0.19429516329061597, "grad_norm": 0.1898830235004425, "learning_rate": 0.006, "loss": 1.4222, "step": 1880 }, { "epoch": 0.19532864820173626, "grad_norm": 0.21167783439159393, "learning_rate": 0.006, "loss": 1.4251, "step": 1890 }, { "epoch": 0.19636213311285655, "grad_norm": 0.2603727877140045, "learning_rate": 0.006, "loss": 1.426, "step": 1900 }, { "epoch": 0.19739561802397684, "grad_norm": 0.193504199385643, "learning_rate": 0.006, "loss": 1.4041, "step": 1910 }, { "epoch": 0.19842910293509713, "grad_norm": 0.22818297147750854, "learning_rate": 0.006, "loss": 1.4204, "step": 1920 }, { "epoch": 0.19946258784621745, "grad_norm": 0.20980095863342285, "learning_rate": 0.006, "loss": 1.405, "step": 1930 }, { "epoch": 0.20049607275733775, "grad_norm": 0.17794106900691986, "learning_rate": 0.006, "loss": 1.3871, "step": 1940 }, { "epoch": 0.20152955766845804, "grad_norm": 0.2572697103023529, "learning_rate": 0.006, "loss": 1.4001, "step": 1950 }, { "epoch": 0.20256304257957833, "grad_norm": 0.2605222463607788, "learning_rate": 0.006, "loss": 1.4029, "step": 1960 }, { "epoch": 0.20359652749069865, "grad_norm": 0.27022120356559753, "learning_rate": 0.006, "loss": 1.4009, "step": 1970 }, { "epoch": 0.20463001240181894, "grad_norm": 0.20254716277122498, "learning_rate": 0.006, "loss": 1.3946, "step": 1980 }, { "epoch": 0.20566349731293923, "grad_norm": 0.19156356155872345, "learning_rate": 0.006, "loss": 1.3848, "step": 1990 }, { "epoch": 0.20669698222405952, "grad_norm": 0.182766854763031, "learning_rate": 0.006, "loss": 1.3964, "step": 2000 }, { "epoch": 0.2077304671351798, "grad_norm": 0.1684897243976593, "learning_rate": 0.006, "loss": 1.3872, "step": 2010 }, { "epoch": 0.20876395204630013, "grad_norm": 0.19054663181304932, "learning_rate": 0.006, "loss": 1.3817, "step": 2020 }, { "epoch": 0.20979743695742042, "grad_norm": 0.17960067093372345, "learning_rate": 0.006, "loss": 1.389, "step": 2030 }, { "epoch": 0.21083092186854072, "grad_norm": 0.23440410196781158, "learning_rate": 0.006, "loss": 1.3707, "step": 2040 }, { "epoch": 0.211864406779661, "grad_norm": 0.21306714415550232, "learning_rate": 0.006, "loss": 1.3789, "step": 2050 }, { "epoch": 0.21289789169078133, "grad_norm": 0.24979044497013092, "learning_rate": 0.006, "loss": 1.3755, "step": 2060 }, { "epoch": 0.21393137660190162, "grad_norm": 0.2036939263343811, "learning_rate": 0.006, "loss": 1.3825, "step": 2070 }, { "epoch": 0.2149648615130219, "grad_norm": 0.21026568114757538, "learning_rate": 0.006, "loss": 1.3701, "step": 2080 }, { "epoch": 0.2159983464241422, "grad_norm": 0.18213844299316406, "learning_rate": 0.006, "loss": 1.3835, "step": 2090 }, { "epoch": 0.2170318313352625, "grad_norm": 0.23943543434143066, "learning_rate": 0.006, "loss": 1.3681, "step": 2100 }, { "epoch": 0.2180653162463828, "grad_norm": 0.27890509366989136, "learning_rate": 0.006, "loss": 1.3746, "step": 2110 }, { "epoch": 0.2190988011575031, "grad_norm": 0.1875576227903366, "learning_rate": 0.006, "loss": 1.3833, "step": 2120 }, { "epoch": 0.2201322860686234, "grad_norm": 0.21590693295001984, "learning_rate": 0.006, "loss": 1.3679, "step": 2130 }, { "epoch": 0.2211657709797437, "grad_norm": 0.16515956819057465, "learning_rate": 0.006, "loss": 1.3628, "step": 2140 }, { "epoch": 0.222199255890864, "grad_norm": 0.3273829221725464, "learning_rate": 0.006, "loss": 1.3605, "step": 2150 }, { "epoch": 0.2232327408019843, "grad_norm": 0.19835293292999268, "learning_rate": 0.006, "loss": 1.3599, "step": 2160 }, { "epoch": 0.2242662257131046, "grad_norm": 0.20204949378967285, "learning_rate": 0.006, "loss": 1.3548, "step": 2170 }, { "epoch": 0.22529971062422488, "grad_norm": 0.20704525709152222, "learning_rate": 0.006, "loss": 1.3625, "step": 2180 }, { "epoch": 0.22633319553534517, "grad_norm": 0.16702768206596375, "learning_rate": 0.006, "loss": 1.3475, "step": 2190 }, { "epoch": 0.2273666804464655, "grad_norm": 0.2254660427570343, "learning_rate": 0.006, "loss": 1.3508, "step": 2200 }, { "epoch": 0.22840016535758578, "grad_norm": 0.17011414468288422, "learning_rate": 0.006, "loss": 1.3416, "step": 2210 }, { "epoch": 0.22943365026870607, "grad_norm": 0.24636393785476685, "learning_rate": 0.006, "loss": 1.3487, "step": 2220 }, { "epoch": 0.23046713517982637, "grad_norm": 0.2143375277519226, "learning_rate": 0.006, "loss": 1.345, "step": 2230 }, { "epoch": 0.23150062009094668, "grad_norm": 0.1868714541196823, "learning_rate": 0.006, "loss": 1.3359, "step": 2240 }, { "epoch": 0.23253410500206698, "grad_norm": 0.1683160364627838, "learning_rate": 0.006, "loss": 1.3399, "step": 2250 }, { "epoch": 0.23356758991318727, "grad_norm": 0.199097141623497, "learning_rate": 0.006, "loss": 1.3568, "step": 2260 }, { "epoch": 0.23460107482430756, "grad_norm": 0.18377502262592316, "learning_rate": 0.006, "loss": 1.3488, "step": 2270 }, { "epoch": 0.23563455973542785, "grad_norm": 0.20199266076087952, "learning_rate": 0.006, "loss": 1.344, "step": 2280 }, { "epoch": 0.23666804464654817, "grad_norm": 0.271793931722641, "learning_rate": 0.006, "loss": 1.3169, "step": 2290 }, { "epoch": 0.23770152955766846, "grad_norm": 0.17339195311069489, "learning_rate": 0.006, "loss": 1.3328, "step": 2300 }, { "epoch": 0.23873501446878875, "grad_norm": 0.17549367249011993, "learning_rate": 0.006, "loss": 1.3354, "step": 2310 }, { "epoch": 0.23976849937990904, "grad_norm": 0.1800873875617981, "learning_rate": 0.006, "loss": 1.322, "step": 2320 }, { "epoch": 0.24080198429102936, "grad_norm": 0.22731319069862366, "learning_rate": 0.006, "loss": 1.3235, "step": 2330 }, { "epoch": 0.24183546920214966, "grad_norm": 0.22910168766975403, "learning_rate": 0.006, "loss": 1.3243, "step": 2340 }, { "epoch": 0.24286895411326995, "grad_norm": 0.1611885130405426, "learning_rate": 0.006, "loss": 1.3352, "step": 2350 }, { "epoch": 0.24390243902439024, "grad_norm": 0.19426490366458893, "learning_rate": 0.006, "loss": 1.3091, "step": 2360 }, { "epoch": 0.24493592393551053, "grad_norm": 0.19100475311279297, "learning_rate": 0.006, "loss": 1.3198, "step": 2370 }, { "epoch": 0.24596940884663085, "grad_norm": 0.2597349286079407, "learning_rate": 0.006, "loss": 1.302, "step": 2380 }, { "epoch": 0.24700289375775114, "grad_norm": 0.16532281041145325, "learning_rate": 0.006, "loss": 1.3182, "step": 2390 }, { "epoch": 0.24803637866887143, "grad_norm": 0.2065669149160385, "learning_rate": 0.006, "loss": 1.3182, "step": 2400 }, { "epoch": 0.24906986357999172, "grad_norm": 0.1802499145269394, "learning_rate": 0.006, "loss": 1.3206, "step": 2410 }, { "epoch": 0.25010334849111204, "grad_norm": 0.2926013469696045, "learning_rate": 0.006, "loss": 1.3089, "step": 2420 }, { "epoch": 0.2511368334022323, "grad_norm": 0.28276434540748596, "learning_rate": 0.006, "loss": 1.307, "step": 2430 }, { "epoch": 0.2521703183133526, "grad_norm": 0.18156251311302185, "learning_rate": 0.006, "loss": 1.3227, "step": 2440 }, { "epoch": 0.25320380322447295, "grad_norm": 0.18208877742290497, "learning_rate": 0.006, "loss": 1.3065, "step": 2450 }, { "epoch": 0.2542372881355932, "grad_norm": 0.18455766141414642, "learning_rate": 0.006, "loss": 1.3019, "step": 2460 }, { "epoch": 0.25527077304671353, "grad_norm": 0.23398499190807343, "learning_rate": 0.006, "loss": 1.3126, "step": 2470 }, { "epoch": 0.2563042579578338, "grad_norm": 0.1919931322336197, "learning_rate": 0.006, "loss": 1.2982, "step": 2480 }, { "epoch": 0.2573377428689541, "grad_norm": 0.24925625324249268, "learning_rate": 0.006, "loss": 1.32, "step": 2490 }, { "epoch": 0.25837122778007443, "grad_norm": 0.2440209984779358, "learning_rate": 0.006, "loss": 1.2904, "step": 2500 }, { "epoch": 0.2594047126911947, "grad_norm": 0.16576959192752838, "learning_rate": 0.006, "loss": 1.3065, "step": 2510 }, { "epoch": 0.260438197602315, "grad_norm": 0.18232128024101257, "learning_rate": 0.006, "loss": 1.3048, "step": 2520 }, { "epoch": 0.2614716825134353, "grad_norm": 0.23745986819267273, "learning_rate": 0.006, "loss": 1.292, "step": 2530 }, { "epoch": 0.2625051674245556, "grad_norm": 0.22818538546562195, "learning_rate": 0.006, "loss": 1.2964, "step": 2540 }, { "epoch": 0.2635386523356759, "grad_norm": 0.1811428666114807, "learning_rate": 0.006, "loss": 1.2968, "step": 2550 }, { "epoch": 0.2645721372467962, "grad_norm": 0.2562517821788788, "learning_rate": 0.006, "loss": 1.2985, "step": 2560 }, { "epoch": 0.2656056221579165, "grad_norm": 0.16540707647800446, "learning_rate": 0.006, "loss": 1.2962, "step": 2570 }, { "epoch": 0.2666391070690368, "grad_norm": 0.17198024690151215, "learning_rate": 0.006, "loss": 1.2835, "step": 2580 }, { "epoch": 0.2676725919801571, "grad_norm": 0.15511544048786163, "learning_rate": 0.006, "loss": 1.2689, "step": 2590 }, { "epoch": 0.2687060768912774, "grad_norm": 0.1916809380054474, "learning_rate": 0.006, "loss": 1.2822, "step": 2600 }, { "epoch": 0.26973956180239766, "grad_norm": 0.17753975093364716, "learning_rate": 0.006, "loss": 1.2843, "step": 2610 }, { "epoch": 0.270773046713518, "grad_norm": 0.18664464354515076, "learning_rate": 0.006, "loss": 1.2876, "step": 2620 }, { "epoch": 0.2718065316246383, "grad_norm": 0.283917635679245, "learning_rate": 0.006, "loss": 1.2745, "step": 2630 }, { "epoch": 0.27284001653575857, "grad_norm": 0.1494530886411667, "learning_rate": 0.006, "loss": 1.2848, "step": 2640 }, { "epoch": 0.2738735014468789, "grad_norm": 0.2093777358531952, "learning_rate": 0.006, "loss": 1.2786, "step": 2650 }, { "epoch": 0.27490698635799915, "grad_norm": 0.19870586693286896, "learning_rate": 0.006, "loss": 1.2895, "step": 2660 }, { "epoch": 0.27594047126911947, "grad_norm": 0.18807923793792725, "learning_rate": 0.006, "loss": 1.2784, "step": 2670 }, { "epoch": 0.2769739561802398, "grad_norm": 0.16479447484016418, "learning_rate": 0.006, "loss": 1.2815, "step": 2680 }, { "epoch": 0.27800744109136005, "grad_norm": 0.19209875166416168, "learning_rate": 0.006, "loss": 1.2791, "step": 2690 }, { "epoch": 0.27904092600248037, "grad_norm": 0.20269426703453064, "learning_rate": 0.006, "loss": 1.2828, "step": 2700 }, { "epoch": 0.28007441091360064, "grad_norm": 0.1880909502506256, "learning_rate": 0.006, "loss": 1.2786, "step": 2710 }, { "epoch": 0.28110789582472095, "grad_norm": 0.16020233929157257, "learning_rate": 0.006, "loss": 1.2637, "step": 2720 }, { "epoch": 0.2821413807358413, "grad_norm": 0.17531338334083557, "learning_rate": 0.006, "loss": 1.264, "step": 2730 }, { "epoch": 0.28317486564696154, "grad_norm": 0.16038616001605988, "learning_rate": 0.006, "loss": 1.2692, "step": 2740 }, { "epoch": 0.28420835055808186, "grad_norm": 0.19676966965198517, "learning_rate": 0.006, "loss": 1.271, "step": 2750 }, { "epoch": 0.2852418354692022, "grad_norm": 0.20371529459953308, "learning_rate": 0.006, "loss": 1.2637, "step": 2760 }, { "epoch": 0.28627532038032244, "grad_norm": 0.18698766827583313, "learning_rate": 0.006, "loss": 1.2776, "step": 2770 }, { "epoch": 0.28730880529144276, "grad_norm": 0.20747795701026917, "learning_rate": 0.006, "loss": 1.2573, "step": 2780 }, { "epoch": 0.288342290202563, "grad_norm": 0.2328963726758957, "learning_rate": 0.006, "loss": 1.2568, "step": 2790 }, { "epoch": 0.28937577511368334, "grad_norm": 0.1655196100473404, "learning_rate": 0.006, "loss": 1.2614, "step": 2800 }, { "epoch": 0.29040926002480366, "grad_norm": 0.1615329086780548, "learning_rate": 0.006, "loss": 1.2712, "step": 2810 }, { "epoch": 0.2914427449359239, "grad_norm": 0.15835103392601013, "learning_rate": 0.006, "loss": 1.2524, "step": 2820 }, { "epoch": 0.29247622984704424, "grad_norm": 0.20250938832759857, "learning_rate": 0.006, "loss": 1.264, "step": 2830 }, { "epoch": 0.2935097147581645, "grad_norm": 0.17166636884212494, "learning_rate": 0.006, "loss": 1.2563, "step": 2840 }, { "epoch": 0.2945431996692848, "grad_norm": 0.2327214777469635, "learning_rate": 0.006, "loss": 1.2546, "step": 2850 }, { "epoch": 0.29557668458040515, "grad_norm": 0.19193734228610992, "learning_rate": 0.006, "loss": 1.2647, "step": 2860 }, { "epoch": 0.2966101694915254, "grad_norm": 0.21454569697380066, "learning_rate": 0.006, "loss": 1.2573, "step": 2870 }, { "epoch": 0.29764365440264573, "grad_norm": 0.18797141313552856, "learning_rate": 0.006, "loss": 1.2661, "step": 2880 }, { "epoch": 0.298677139313766, "grad_norm": 0.16613776981830597, "learning_rate": 0.006, "loss": 1.2593, "step": 2890 }, { "epoch": 0.2997106242248863, "grad_norm": 0.15257228910923004, "learning_rate": 0.006, "loss": 1.2534, "step": 2900 }, { "epoch": 0.30074410913600663, "grad_norm": 0.1718466728925705, "learning_rate": 0.006, "loss": 1.2526, "step": 2910 }, { "epoch": 0.3017775940471269, "grad_norm": 0.17519408464431763, "learning_rate": 0.006, "loss": 1.236, "step": 2920 }, { "epoch": 0.3028110789582472, "grad_norm": 0.1739313006401062, "learning_rate": 0.006, "loss": 1.2416, "step": 2930 }, { "epoch": 0.30384456386936753, "grad_norm": 0.2000485360622406, "learning_rate": 0.006, "loss": 1.2525, "step": 2940 }, { "epoch": 0.3048780487804878, "grad_norm": 0.21297138929367065, "learning_rate": 0.006, "loss": 1.2312, "step": 2950 }, { "epoch": 0.3059115336916081, "grad_norm": 0.1601925939321518, "learning_rate": 0.006, "loss": 1.2575, "step": 2960 }, { "epoch": 0.3069450186027284, "grad_norm": 0.22686167061328888, "learning_rate": 0.006, "loss": 1.2421, "step": 2970 }, { "epoch": 0.3079785035138487, "grad_norm": 0.2552824318408966, "learning_rate": 0.006, "loss": 1.255, "step": 2980 }, { "epoch": 0.309011988424969, "grad_norm": 0.1520908921957016, "learning_rate": 0.006, "loss": 1.2335, "step": 2990 }, { "epoch": 0.3100454733360893, "grad_norm": 0.18067720532417297, "learning_rate": 0.006, "loss": 1.2278, "step": 3000 }, { "epoch": 0.3110789582472096, "grad_norm": 0.2269776612520218, "learning_rate": 0.006, "loss": 1.2401, "step": 3010 }, { "epoch": 0.31211244315832987, "grad_norm": 0.17168264091014862, "learning_rate": 0.006, "loss": 1.245, "step": 3020 }, { "epoch": 0.3131459280694502, "grad_norm": 0.20345386862754822, "learning_rate": 0.006, "loss": 1.2286, "step": 3030 }, { "epoch": 0.3141794129805705, "grad_norm": 0.17593909800052643, "learning_rate": 0.006, "loss": 1.2439, "step": 3040 }, { "epoch": 0.31521289789169077, "grad_norm": 0.1948511153459549, "learning_rate": 0.006, "loss": 1.2513, "step": 3050 }, { "epoch": 0.3162463828028111, "grad_norm": 0.1758977472782135, "learning_rate": 0.006, "loss": 1.2428, "step": 3060 }, { "epoch": 0.31727986771393135, "grad_norm": 0.14619344472885132, "learning_rate": 0.006, "loss": 1.2295, "step": 3070 }, { "epoch": 0.31831335262505167, "grad_norm": 0.15746112167835236, "learning_rate": 0.006, "loss": 1.2133, "step": 3080 }, { "epoch": 0.319346837536172, "grad_norm": 0.14768971502780914, "learning_rate": 0.006, "loss": 1.2234, "step": 3090 }, { "epoch": 0.32038032244729225, "grad_norm": 0.15294450521469116, "learning_rate": 0.006, "loss": 1.2255, "step": 3100 }, { "epoch": 0.3214138073584126, "grad_norm": 0.18775102496147156, "learning_rate": 0.006, "loss": 1.2378, "step": 3110 }, { "epoch": 0.3224472922695329, "grad_norm": 0.15432560443878174, "learning_rate": 0.006, "loss": 1.2359, "step": 3120 }, { "epoch": 0.32348077718065316, "grad_norm": 0.17162322998046875, "learning_rate": 0.006, "loss": 1.2257, "step": 3130 }, { "epoch": 0.3245142620917735, "grad_norm": 0.21141092479228973, "learning_rate": 0.006, "loss": 1.2301, "step": 3140 }, { "epoch": 0.32554774700289374, "grad_norm": 0.1810036152601242, "learning_rate": 0.006, "loss": 1.2204, "step": 3150 }, { "epoch": 0.32658123191401406, "grad_norm": 0.20127083361148834, "learning_rate": 0.006, "loss": 1.2263, "step": 3160 }, { "epoch": 0.3276147168251344, "grad_norm": 0.1795072853565216, "learning_rate": 0.006, "loss": 1.2224, "step": 3170 }, { "epoch": 0.32864820173625464, "grad_norm": 0.26130935549736023, "learning_rate": 0.006, "loss": 1.2236, "step": 3180 }, { "epoch": 0.32968168664737496, "grad_norm": 0.17055651545524597, "learning_rate": 0.006, "loss": 1.2229, "step": 3190 }, { "epoch": 0.3307151715584952, "grad_norm": 0.24257224798202515, "learning_rate": 0.006, "loss": 1.218, "step": 3200 }, { "epoch": 0.33174865646961554, "grad_norm": 0.1574934422969818, "learning_rate": 0.006, "loss": 1.2248, "step": 3210 }, { "epoch": 0.33278214138073586, "grad_norm": 0.18338268995285034, "learning_rate": 0.006, "loss": 1.211, "step": 3220 }, { "epoch": 0.3338156262918561, "grad_norm": 0.23949076235294342, "learning_rate": 0.006, "loss": 1.2107, "step": 3230 }, { "epoch": 0.33484911120297645, "grad_norm": 0.23181311786174774, "learning_rate": 0.006, "loss": 1.2157, "step": 3240 }, { "epoch": 0.3358825961140967, "grad_norm": 0.1727987825870514, "learning_rate": 0.006, "loss": 1.2113, "step": 3250 }, { "epoch": 0.33691608102521703, "grad_norm": 0.17050251364707947, "learning_rate": 0.006, "loss": 1.2196, "step": 3260 }, { "epoch": 0.33794956593633735, "grad_norm": 0.1836351752281189, "learning_rate": 0.006, "loss": 1.2164, "step": 3270 }, { "epoch": 0.3389830508474576, "grad_norm": 0.16653425991535187, "learning_rate": 0.006, "loss": 1.2108, "step": 3280 }, { "epoch": 0.34001653575857793, "grad_norm": 0.19799716770648956, "learning_rate": 0.006, "loss": 1.2072, "step": 3290 }, { "epoch": 0.3410500206696982, "grad_norm": 0.15341606736183167, "learning_rate": 0.006, "loss": 1.2157, "step": 3300 }, { "epoch": 0.3420835055808185, "grad_norm": 0.15945138037204742, "learning_rate": 0.006, "loss": 1.2065, "step": 3310 }, { "epoch": 0.34311699049193883, "grad_norm": 0.17721377313137054, "learning_rate": 0.006, "loss": 1.2056, "step": 3320 }, { "epoch": 0.3441504754030591, "grad_norm": 0.1358070969581604, "learning_rate": 0.006, "loss": 1.2171, "step": 3330 }, { "epoch": 0.3451839603141794, "grad_norm": 0.17187251150608063, "learning_rate": 0.006, "loss": 1.2046, "step": 3340 }, { "epoch": 0.34621744522529974, "grad_norm": 0.14732858538627625, "learning_rate": 0.006, "loss": 1.2025, "step": 3350 }, { "epoch": 0.34725093013642, "grad_norm": 0.224739208817482, "learning_rate": 0.006, "loss": 1.2213, "step": 3360 }, { "epoch": 0.3482844150475403, "grad_norm": 0.20575636625289917, "learning_rate": 0.006, "loss": 1.1922, "step": 3370 }, { "epoch": 0.3493178999586606, "grad_norm": 0.15750053524971008, "learning_rate": 0.006, "loss": 1.209, "step": 3380 }, { "epoch": 0.3503513848697809, "grad_norm": 0.27751949429512024, "learning_rate": 0.006, "loss": 1.2079, "step": 3390 }, { "epoch": 0.3513848697809012, "grad_norm": 0.14315839111804962, "learning_rate": 0.006, "loss": 1.1934, "step": 3400 }, { "epoch": 0.3524183546920215, "grad_norm": 0.15051007270812988, "learning_rate": 0.006, "loss": 1.2067, "step": 3410 }, { "epoch": 0.3534518396031418, "grad_norm": 0.15940986573696136, "learning_rate": 0.006, "loss": 1.2097, "step": 3420 }, { "epoch": 0.35448532451426207, "grad_norm": 0.1412818878889084, "learning_rate": 0.006, "loss": 1.1959, "step": 3430 }, { "epoch": 0.3555188094253824, "grad_norm": 0.1425037980079651, "learning_rate": 0.006, "loss": 1.1901, "step": 3440 }, { "epoch": 0.3565522943365027, "grad_norm": 0.15428400039672852, "learning_rate": 0.006, "loss": 1.2037, "step": 3450 }, { "epoch": 0.35758577924762297, "grad_norm": 0.1565396934747696, "learning_rate": 0.006, "loss": 1.2066, "step": 3460 }, { "epoch": 0.3586192641587433, "grad_norm": 0.1757258176803589, "learning_rate": 0.006, "loss": 1.1947, "step": 3470 }, { "epoch": 0.35965274906986355, "grad_norm": 0.23932603001594543, "learning_rate": 0.006, "loss": 1.1993, "step": 3480 }, { "epoch": 0.3606862339809839, "grad_norm": 0.17783917486667633, "learning_rate": 0.006, "loss": 1.1881, "step": 3490 }, { "epoch": 0.3617197188921042, "grad_norm": 0.13933712244033813, "learning_rate": 0.006, "loss": 1.1969, "step": 3500 }, { "epoch": 0.36275320380322446, "grad_norm": 0.13872799277305603, "learning_rate": 0.006, "loss": 1.1958, "step": 3510 }, { "epoch": 0.3637866887143448, "grad_norm": 0.18496429920196533, "learning_rate": 0.006, "loss": 1.1978, "step": 3520 }, { "epoch": 0.3648201736254651, "grad_norm": 0.2182753086090088, "learning_rate": 0.006, "loss": 1.1957, "step": 3530 }, { "epoch": 0.36585365853658536, "grad_norm": 0.15641410648822784, "learning_rate": 0.006, "loss": 1.1828, "step": 3540 }, { "epoch": 0.3668871434477057, "grad_norm": 0.15196652710437775, "learning_rate": 0.006, "loss": 1.1901, "step": 3550 }, { "epoch": 0.36792062835882594, "grad_norm": 0.15420852601528168, "learning_rate": 0.006, "loss": 1.1966, "step": 3560 }, { "epoch": 0.36895411326994626, "grad_norm": 0.1419685333967209, "learning_rate": 0.006, "loss": 1.1835, "step": 3570 }, { "epoch": 0.3699875981810666, "grad_norm": 0.20714856684207916, "learning_rate": 0.006, "loss": 1.1969, "step": 3580 }, { "epoch": 0.37102108309218684, "grad_norm": 0.14852945506572723, "learning_rate": 0.006, "loss": 1.1979, "step": 3590 }, { "epoch": 0.37205456800330716, "grad_norm": 0.1534539759159088, "learning_rate": 0.006, "loss": 1.1814, "step": 3600 }, { "epoch": 0.3730880529144274, "grad_norm": 0.1536993384361267, "learning_rate": 0.006, "loss": 1.1935, "step": 3610 }, { "epoch": 0.37412153782554775, "grad_norm": 0.14678268134593964, "learning_rate": 0.006, "loss": 1.1879, "step": 3620 }, { "epoch": 0.37515502273666806, "grad_norm": 0.15194503962993622, "learning_rate": 0.006, "loss": 1.1851, "step": 3630 }, { "epoch": 0.37618850764778833, "grad_norm": 0.1853533834218979, "learning_rate": 0.006, "loss": 1.177, "step": 3640 }, { "epoch": 0.37722199255890865, "grad_norm": 0.19894924759864807, "learning_rate": 0.006, "loss": 1.1792, "step": 3650 }, { "epoch": 0.3782554774700289, "grad_norm": 0.14289256930351257, "learning_rate": 0.006, "loss": 1.1916, "step": 3660 }, { "epoch": 0.37928896238114923, "grad_norm": 0.15364837646484375, "learning_rate": 0.006, "loss": 1.1773, "step": 3670 }, { "epoch": 0.38032244729226955, "grad_norm": 0.26057279109954834, "learning_rate": 0.006, "loss": 1.1772, "step": 3680 }, { "epoch": 0.3813559322033898, "grad_norm": 0.14714686572551727, "learning_rate": 0.006, "loss": 1.183, "step": 3690 }, { "epoch": 0.38238941711451013, "grad_norm": 0.13475820422172546, "learning_rate": 0.006, "loss": 1.1734, "step": 3700 }, { "epoch": 0.38342290202563045, "grad_norm": 0.169892817735672, "learning_rate": 0.006, "loss": 1.1833, "step": 3710 }, { "epoch": 0.3844563869367507, "grad_norm": 0.15819117426872253, "learning_rate": 0.006, "loss": 1.1693, "step": 3720 }, { "epoch": 0.38548987184787104, "grad_norm": 0.2074589729309082, "learning_rate": 0.006, "loss": 1.1773, "step": 3730 }, { "epoch": 0.3865233567589913, "grad_norm": 0.1474793255329132, "learning_rate": 0.006, "loss": 1.1746, "step": 3740 }, { "epoch": 0.3875568416701116, "grad_norm": 0.20919297635555267, "learning_rate": 0.006, "loss": 1.1644, "step": 3750 }, { "epoch": 0.38859032658123194, "grad_norm": 0.1802784502506256, "learning_rate": 0.006, "loss": 1.1785, "step": 3760 }, { "epoch": 0.3896238114923522, "grad_norm": 0.13777178525924683, "learning_rate": 0.006, "loss": 1.1762, "step": 3770 }, { "epoch": 0.3906572964034725, "grad_norm": 0.19703027606010437, "learning_rate": 0.006, "loss": 1.1715, "step": 3780 }, { "epoch": 0.3916907813145928, "grad_norm": 0.13620315492153168, "learning_rate": 0.006, "loss": 1.1729, "step": 3790 }, { "epoch": 0.3927242662257131, "grad_norm": 0.19151124358177185, "learning_rate": 0.006, "loss": 1.165, "step": 3800 }, { "epoch": 0.3937577511368334, "grad_norm": 0.14211703836917877, "learning_rate": 0.006, "loss": 1.1705, "step": 3810 }, { "epoch": 0.3947912360479537, "grad_norm": 0.13506080210208893, "learning_rate": 0.006, "loss": 1.1656, "step": 3820 }, { "epoch": 0.395824720959074, "grad_norm": 0.15365293622016907, "learning_rate": 0.006, "loss": 1.1642, "step": 3830 }, { "epoch": 0.39685820587019427, "grad_norm": 0.1530192494392395, "learning_rate": 0.006, "loss": 1.1621, "step": 3840 }, { "epoch": 0.3978916907813146, "grad_norm": 0.1344013661146164, "learning_rate": 0.006, "loss": 1.1694, "step": 3850 }, { "epoch": 0.3989251756924349, "grad_norm": 0.1469976007938385, "learning_rate": 0.006, "loss": 1.1706, "step": 3860 }, { "epoch": 0.39995866060355517, "grad_norm": 0.14535875618457794, "learning_rate": 0.006, "loss": 1.1641, "step": 3870 }, { "epoch": 0.4009921455146755, "grad_norm": 0.22596335411071777, "learning_rate": 0.006, "loss": 1.1835, "step": 3880 }, { "epoch": 0.4020256304257958, "grad_norm": 0.214060440659523, "learning_rate": 0.006, "loss": 1.1688, "step": 3890 }, { "epoch": 0.4030591153369161, "grad_norm": 0.14078359305858612, "learning_rate": 0.006, "loss": 1.158, "step": 3900 }, { "epoch": 0.4040926002480364, "grad_norm": 0.15023833513259888, "learning_rate": 0.006, "loss": 1.1599, "step": 3910 }, { "epoch": 0.40512608515915666, "grad_norm": 0.13650792837142944, "learning_rate": 0.006, "loss": 1.1512, "step": 3920 }, { "epoch": 0.406159570070277, "grad_norm": 0.1389831155538559, "learning_rate": 0.006, "loss": 1.1624, "step": 3930 }, { "epoch": 0.4071930549813973, "grad_norm": 0.1548847109079361, "learning_rate": 0.006, "loss": 1.1822, "step": 3940 }, { "epoch": 0.40822653989251756, "grad_norm": 0.14826175570487976, "learning_rate": 0.006, "loss": 1.1539, "step": 3950 }, { "epoch": 0.4092600248036379, "grad_norm": 0.27681705355644226, "learning_rate": 0.006, "loss": 1.1543, "step": 3960 }, { "epoch": 0.41029350971475814, "grad_norm": 0.17979800701141357, "learning_rate": 0.006, "loss": 1.1583, "step": 3970 }, { "epoch": 0.41132699462587846, "grad_norm": 0.14845605194568634, "learning_rate": 0.006, "loss": 1.1559, "step": 3980 }, { "epoch": 0.4123604795369988, "grad_norm": 0.18015886843204498, "learning_rate": 0.006, "loss": 1.1663, "step": 3990 }, { "epoch": 0.41339396444811904, "grad_norm": 0.2010691910982132, "learning_rate": 0.006, "loss": 1.1572, "step": 4000 }, { "epoch": 0.41442744935923936, "grad_norm": 0.14041297137737274, "learning_rate": 0.006, "loss": 1.1621, "step": 4010 }, { "epoch": 0.4154609342703596, "grad_norm": 0.13786040246486664, "learning_rate": 0.006, "loss": 1.157, "step": 4020 }, { "epoch": 0.41649441918147995, "grad_norm": 0.17166124284267426, "learning_rate": 0.006, "loss": 1.178, "step": 4030 }, { "epoch": 0.41752790409260027, "grad_norm": 0.13718323409557343, "learning_rate": 0.006, "loss": 1.1514, "step": 4040 }, { "epoch": 0.41856138900372053, "grad_norm": 0.15878678858280182, "learning_rate": 0.006, "loss": 1.1718, "step": 4050 }, { "epoch": 0.41959487391484085, "grad_norm": 0.14852339029312134, "learning_rate": 0.006, "loss": 1.1559, "step": 4060 }, { "epoch": 0.42062835882596117, "grad_norm": 0.17088137567043304, "learning_rate": 0.006, "loss": 1.1527, "step": 4070 }, { "epoch": 0.42166184373708143, "grad_norm": 0.15074481070041656, "learning_rate": 0.006, "loss": 1.1519, "step": 4080 }, { "epoch": 0.42269532864820175, "grad_norm": 0.13103771209716797, "learning_rate": 0.006, "loss": 1.1475, "step": 4090 }, { "epoch": 0.423728813559322, "grad_norm": 0.1428011655807495, "learning_rate": 0.006, "loss": 1.1528, "step": 4100 }, { "epoch": 0.42476229847044233, "grad_norm": 0.14280404150485992, "learning_rate": 0.006, "loss": 1.1542, "step": 4110 }, { "epoch": 0.42579578338156265, "grad_norm": 0.14426454901695251, "learning_rate": 0.006, "loss": 1.1632, "step": 4120 }, { "epoch": 0.4268292682926829, "grad_norm": 0.13187375664710999, "learning_rate": 0.006, "loss": 1.1564, "step": 4130 }, { "epoch": 0.42786275320380324, "grad_norm": 0.1555539071559906, "learning_rate": 0.006, "loss": 1.1497, "step": 4140 }, { "epoch": 0.4288962381149235, "grad_norm": 0.14076194167137146, "learning_rate": 0.006, "loss": 1.1431, "step": 4150 }, { "epoch": 0.4299297230260438, "grad_norm": 0.14553508162498474, "learning_rate": 0.006, "loss": 1.1444, "step": 4160 }, { "epoch": 0.43096320793716414, "grad_norm": 0.14131340384483337, "learning_rate": 0.006, "loss": 1.1504, "step": 4170 }, { "epoch": 0.4319966928482844, "grad_norm": 0.1349923461675644, "learning_rate": 0.006, "loss": 1.1499, "step": 4180 }, { "epoch": 0.4330301777594047, "grad_norm": 0.1801183521747589, "learning_rate": 0.006, "loss": 1.1458, "step": 4190 }, { "epoch": 0.434063662670525, "grad_norm": 0.16793182492256165, "learning_rate": 0.006, "loss": 1.1527, "step": 4200 }, { "epoch": 0.4350971475816453, "grad_norm": 0.19141018390655518, "learning_rate": 0.006, "loss": 1.1451, "step": 4210 }, { "epoch": 0.4361306324927656, "grad_norm": 0.1311262547969818, "learning_rate": 0.006, "loss": 1.1433, "step": 4220 }, { "epoch": 0.4371641174038859, "grad_norm": 0.1293700635433197, "learning_rate": 0.006, "loss": 1.1251, "step": 4230 }, { "epoch": 0.4381976023150062, "grad_norm": 0.13121871650218964, "learning_rate": 0.006, "loss": 1.1392, "step": 4240 }, { "epoch": 0.43923108722612647, "grad_norm": 0.13999488949775696, "learning_rate": 0.006, "loss": 1.135, "step": 4250 }, { "epoch": 0.4402645721372468, "grad_norm": 0.14969255030155182, "learning_rate": 0.006, "loss": 1.1399, "step": 4260 }, { "epoch": 0.4412980570483671, "grad_norm": 0.14516979455947876, "learning_rate": 0.006, "loss": 1.1385, "step": 4270 }, { "epoch": 0.4423315419594874, "grad_norm": 0.13377705216407776, "learning_rate": 0.006, "loss": 1.1409, "step": 4280 }, { "epoch": 0.4433650268706077, "grad_norm": 0.1788671314716339, "learning_rate": 0.006, "loss": 1.1368, "step": 4290 }, { "epoch": 0.444398511781728, "grad_norm": 0.1459660530090332, "learning_rate": 0.006, "loss": 1.1399, "step": 4300 }, { "epoch": 0.4454319966928483, "grad_norm": 0.14903658628463745, "learning_rate": 0.006, "loss": 1.1347, "step": 4310 }, { "epoch": 0.4464654816039686, "grad_norm": 0.16043029725551605, "learning_rate": 0.006, "loss": 1.1395, "step": 4320 }, { "epoch": 0.44749896651508886, "grad_norm": 0.1458473652601242, "learning_rate": 0.006, "loss": 1.1442, "step": 4330 }, { "epoch": 0.4485324514262092, "grad_norm": 0.1839747577905655, "learning_rate": 0.006, "loss": 1.1278, "step": 4340 }, { "epoch": 0.4495659363373295, "grad_norm": 0.14207588136196136, "learning_rate": 0.006, "loss": 1.1355, "step": 4350 }, { "epoch": 0.45059942124844976, "grad_norm": 0.1426732838153839, "learning_rate": 0.006, "loss": 1.1345, "step": 4360 }, { "epoch": 0.4516329061595701, "grad_norm": 0.19898781180381775, "learning_rate": 0.006, "loss": 1.128, "step": 4370 }, { "epoch": 0.45266639107069034, "grad_norm": 0.1501811444759369, "learning_rate": 0.006, "loss": 1.1294, "step": 4380 }, { "epoch": 0.45369987598181066, "grad_norm": 0.13546693325042725, "learning_rate": 0.006, "loss": 1.1459, "step": 4390 }, { "epoch": 0.454733360892931, "grad_norm": 0.1429995894432068, "learning_rate": 0.006, "loss": 1.1401, "step": 4400 }, { "epoch": 0.45576684580405125, "grad_norm": 0.16272728145122528, "learning_rate": 0.006, "loss": 1.1388, "step": 4410 }, { "epoch": 0.45680033071517157, "grad_norm": 0.13995374739170074, "learning_rate": 0.006, "loss": 1.1348, "step": 4420 }, { "epoch": 0.45783381562629183, "grad_norm": 0.13635413348674774, "learning_rate": 0.006, "loss": 1.13, "step": 4430 }, { "epoch": 0.45886730053741215, "grad_norm": 0.12863661348819733, "learning_rate": 0.006, "loss": 1.1289, "step": 4440 }, { "epoch": 0.45990078544853247, "grad_norm": 0.14227396249771118, "learning_rate": 0.006, "loss": 1.1304, "step": 4450 }, { "epoch": 0.46093427035965273, "grad_norm": 0.21000246703624725, "learning_rate": 0.006, "loss": 1.132, "step": 4460 }, { "epoch": 0.46196775527077305, "grad_norm": 0.15686342120170593, "learning_rate": 0.006, "loss": 1.1456, "step": 4470 }, { "epoch": 0.46300124018189337, "grad_norm": 0.13623683154582977, "learning_rate": 0.006, "loss": 1.1229, "step": 4480 }, { "epoch": 0.46403472509301363, "grad_norm": 0.13092273473739624, "learning_rate": 0.006, "loss": 1.1199, "step": 4490 }, { "epoch": 0.46506821000413395, "grad_norm": 0.13690875470638275, "learning_rate": 0.006, "loss": 1.1095, "step": 4500 }, { "epoch": 0.4661016949152542, "grad_norm": 0.12770096957683563, "learning_rate": 0.006, "loss": 1.1251, "step": 4510 }, { "epoch": 0.46713517982637454, "grad_norm": 0.12008727341890335, "learning_rate": 0.006, "loss": 1.1151, "step": 4520 }, { "epoch": 0.46816866473749486, "grad_norm": 0.13290061056613922, "learning_rate": 0.006, "loss": 1.1278, "step": 4530 }, { "epoch": 0.4692021496486151, "grad_norm": 0.16319867968559265, "learning_rate": 0.006, "loss": 1.1261, "step": 4540 }, { "epoch": 0.47023563455973544, "grad_norm": 0.13294318318367004, "learning_rate": 0.006, "loss": 1.1314, "step": 4550 }, { "epoch": 0.4712691194708557, "grad_norm": 0.15630632638931274, "learning_rate": 0.006, "loss": 1.1423, "step": 4560 }, { "epoch": 0.472302604381976, "grad_norm": 0.16916899383068085, "learning_rate": 0.006, "loss": 1.1385, "step": 4570 }, { "epoch": 0.47333608929309634, "grad_norm": 0.2240544855594635, "learning_rate": 0.006, "loss": 1.1307, "step": 4580 }, { "epoch": 0.4743695742042166, "grad_norm": 0.12742292881011963, "learning_rate": 0.006, "loss": 1.1278, "step": 4590 }, { "epoch": 0.4754030591153369, "grad_norm": 0.13861480355262756, "learning_rate": 0.006, "loss": 1.1248, "step": 4600 }, { "epoch": 0.4764365440264572, "grad_norm": 0.16158758103847504, "learning_rate": 0.006, "loss": 1.1437, "step": 4610 }, { "epoch": 0.4774700289375775, "grad_norm": 0.13849808275699615, "learning_rate": 0.006, "loss": 1.1345, "step": 4620 }, { "epoch": 0.4785035138486978, "grad_norm": 0.1381804645061493, "learning_rate": 0.006, "loss": 1.124, "step": 4630 }, { "epoch": 0.4795369987598181, "grad_norm": 0.1629112809896469, "learning_rate": 0.006, "loss": 1.1198, "step": 4640 }, { "epoch": 0.4805704836709384, "grad_norm": 0.1835983693599701, "learning_rate": 0.006, "loss": 1.1198, "step": 4650 }, { "epoch": 0.48160396858205873, "grad_norm": 0.13495475053787231, "learning_rate": 0.006, "loss": 1.1266, "step": 4660 }, { "epoch": 0.482637453493179, "grad_norm": 0.15339604020118713, "learning_rate": 0.006, "loss": 1.1111, "step": 4670 }, { "epoch": 0.4836709384042993, "grad_norm": 0.14230936765670776, "learning_rate": 0.006, "loss": 1.1002, "step": 4680 }, { "epoch": 0.4847044233154196, "grad_norm": 0.13988140225410461, "learning_rate": 0.006, "loss": 1.1132, "step": 4690 }, { "epoch": 0.4857379082265399, "grad_norm": 0.17499862611293793, "learning_rate": 0.006, "loss": 1.1223, "step": 4700 }, { "epoch": 0.4867713931376602, "grad_norm": 0.15850752592086792, "learning_rate": 0.006, "loss": 1.1199, "step": 4710 }, { "epoch": 0.4878048780487805, "grad_norm": 0.12476572394371033, "learning_rate": 0.006, "loss": 1.1252, "step": 4720 }, { "epoch": 0.4888383629599008, "grad_norm": 0.14426189661026, "learning_rate": 0.006, "loss": 1.1132, "step": 4730 }, { "epoch": 0.48987184787102106, "grad_norm": 0.14036618173122406, "learning_rate": 0.006, "loss": 1.1226, "step": 4740 }, { "epoch": 0.4909053327821414, "grad_norm": 0.1434011608362198, "learning_rate": 0.006, "loss": 1.1123, "step": 4750 }, { "epoch": 0.4919388176932617, "grad_norm": 0.19470463693141937, "learning_rate": 0.006, "loss": 1.1321, "step": 4760 }, { "epoch": 0.49297230260438196, "grad_norm": 0.12864422798156738, "learning_rate": 0.006, "loss": 1.1176, "step": 4770 }, { "epoch": 0.4940057875155023, "grad_norm": 0.16909734904766083, "learning_rate": 0.006, "loss": 1.1164, "step": 4780 }, { "epoch": 0.49503927242662255, "grad_norm": 0.14817242324352264, "learning_rate": 0.006, "loss": 1.1086, "step": 4790 }, { "epoch": 0.49607275733774286, "grad_norm": 0.16644009947776794, "learning_rate": 0.006, "loss": 1.1142, "step": 4800 }, { "epoch": 0.4971062422488632, "grad_norm": 0.16777795553207397, "learning_rate": 0.006, "loss": 1.1077, "step": 4810 }, { "epoch": 0.49813972715998345, "grad_norm": 0.1416415572166443, "learning_rate": 0.006, "loss": 1.122, "step": 4820 }, { "epoch": 0.49917321207110377, "grad_norm": 0.1440482884645462, "learning_rate": 0.006, "loss": 1.1129, "step": 4830 }, { "epoch": 0.5002066969822241, "grad_norm": 0.1519850343465805, "learning_rate": 0.006, "loss": 1.1144, "step": 4840 }, { "epoch": 0.5012401818933444, "grad_norm": 0.24164853990077972, "learning_rate": 0.006, "loss": 1.1235, "step": 4850 }, { "epoch": 0.5022736668044646, "grad_norm": 0.16202867031097412, "learning_rate": 0.006, "loss": 1.1054, "step": 4860 }, { "epoch": 0.5033071517155849, "grad_norm": 0.13450877368450165, "learning_rate": 0.006, "loss": 1.1112, "step": 4870 }, { "epoch": 0.5043406366267053, "grad_norm": 0.1484173834323883, "learning_rate": 0.006, "loss": 1.103, "step": 4880 }, { "epoch": 0.5053741215378256, "grad_norm": 0.14877624809741974, "learning_rate": 0.006, "loss": 1.1353, "step": 4890 }, { "epoch": 0.5064076064489459, "grad_norm": 0.14426067471504211, "learning_rate": 0.006, "loss": 1.1172, "step": 4900 }, { "epoch": 0.5074410913600661, "grad_norm": 0.14011584222316742, "learning_rate": 0.006, "loss": 1.1171, "step": 4910 }, { "epoch": 0.5084745762711864, "grad_norm": 0.20109255611896515, "learning_rate": 0.006, "loss": 1.1097, "step": 4920 }, { "epoch": 0.5095080611823067, "grad_norm": 0.1302032619714737, "learning_rate": 0.006, "loss": 1.1092, "step": 4930 }, { "epoch": 0.5105415460934271, "grad_norm": 0.13182514905929565, "learning_rate": 0.006, "loss": 1.1043, "step": 4940 }, { "epoch": 0.5115750310045474, "grad_norm": 0.13346487283706665, "learning_rate": 0.006, "loss": 1.1133, "step": 4950 }, { "epoch": 0.5126085159156676, "grad_norm": 0.13681770861148834, "learning_rate": 0.006, "loss": 1.1153, "step": 4960 }, { "epoch": 0.5136420008267879, "grad_norm": 0.14637672901153564, "learning_rate": 0.006, "loss": 1.0984, "step": 4970 }, { "epoch": 0.5146754857379082, "grad_norm": 0.14188942313194275, "learning_rate": 0.006, "loss": 1.1033, "step": 4980 }, { "epoch": 0.5157089706490285, "grad_norm": 0.1467074155807495, "learning_rate": 0.006, "loss": 1.095, "step": 4990 }, { "epoch": 0.5167424555601489, "grad_norm": 0.15010304749011993, "learning_rate": 0.006, "loss": 1.1029, "step": 5000 }, { "epoch": 0.5177759404712691, "grad_norm": 0.13308066129684448, "learning_rate": 0.006, "loss": 1.1004, "step": 5010 }, { "epoch": 0.5188094253823894, "grad_norm": 0.13878673315048218, "learning_rate": 0.006, "loss": 1.1055, "step": 5020 }, { "epoch": 0.5198429102935097, "grad_norm": 0.14125706255435944, "learning_rate": 0.006, "loss": 1.0927, "step": 5030 }, { "epoch": 0.52087639520463, "grad_norm": 0.1262088418006897, "learning_rate": 0.006, "loss": 1.1112, "step": 5040 }, { "epoch": 0.5219098801157503, "grad_norm": 0.20926526188850403, "learning_rate": 0.006, "loss": 1.1169, "step": 5050 }, { "epoch": 0.5229433650268706, "grad_norm": 0.2341676652431488, "learning_rate": 0.006, "loss": 1.1198, "step": 5060 }, { "epoch": 0.5239768499379909, "grad_norm": 0.15894795954227448, "learning_rate": 0.006, "loss": 1.1017, "step": 5070 }, { "epoch": 0.5250103348491112, "grad_norm": 0.14738234877586365, "learning_rate": 0.006, "loss": 1.1021, "step": 5080 }, { "epoch": 0.5260438197602315, "grad_norm": 0.14371953904628754, "learning_rate": 0.006, "loss": 1.0978, "step": 5090 }, { "epoch": 0.5270773046713518, "grad_norm": 0.13194578886032104, "learning_rate": 0.006, "loss": 1.1073, "step": 5100 }, { "epoch": 0.5281107895824722, "grad_norm": 0.1411438137292862, "learning_rate": 0.006, "loss": 1.0931, "step": 5110 }, { "epoch": 0.5291442744935924, "grad_norm": 0.1347801685333252, "learning_rate": 0.006, "loss": 1.0896, "step": 5120 }, { "epoch": 0.5301777594047127, "grad_norm": 0.12243802100419998, "learning_rate": 0.006, "loss": 1.0912, "step": 5130 }, { "epoch": 0.531211244315833, "grad_norm": 0.15618611872196198, "learning_rate": 0.006, "loss": 1.0925, "step": 5140 }, { "epoch": 0.5322447292269533, "grad_norm": 0.133057102560997, "learning_rate": 0.006, "loss": 1.0937, "step": 5150 }, { "epoch": 0.5332782141380736, "grad_norm": 0.14488716423511505, "learning_rate": 0.006, "loss": 1.1057, "step": 5160 }, { "epoch": 0.5343116990491938, "grad_norm": 0.15518608689308167, "learning_rate": 0.006, "loss": 1.1073, "step": 5170 }, { "epoch": 0.5353451839603142, "grad_norm": 0.14580237865447998, "learning_rate": 0.006, "loss": 1.1064, "step": 5180 }, { "epoch": 0.5363786688714345, "grad_norm": 0.12775352597236633, "learning_rate": 0.006, "loss": 1.1043, "step": 5190 }, { "epoch": 0.5374121537825548, "grad_norm": 0.15589575469493866, "learning_rate": 0.006, "loss": 1.0995, "step": 5200 }, { "epoch": 0.5384456386936751, "grad_norm": 0.1348804533481598, "learning_rate": 0.006, "loss": 1.0959, "step": 5210 }, { "epoch": 0.5394791236047953, "grad_norm": 0.1302793323993683, "learning_rate": 0.006, "loss": 1.0833, "step": 5220 }, { "epoch": 0.5405126085159156, "grad_norm": 0.1520976573228836, "learning_rate": 0.006, "loss": 1.0991, "step": 5230 }, { "epoch": 0.541546093427036, "grad_norm": 0.13820461928844452, "learning_rate": 0.006, "loss": 1.1039, "step": 5240 }, { "epoch": 0.5425795783381563, "grad_norm": 0.13032610714435577, "learning_rate": 0.006, "loss": 1.0947, "step": 5250 }, { "epoch": 0.5436130632492766, "grad_norm": 0.1633882224559784, "learning_rate": 0.006, "loss": 1.1009, "step": 5260 }, { "epoch": 0.5446465481603968, "grad_norm": 0.2063656449317932, "learning_rate": 0.006, "loss": 1.0947, "step": 5270 }, { "epoch": 0.5456800330715171, "grad_norm": 0.13242821395397186, "learning_rate": 0.006, "loss": 1.0904, "step": 5280 }, { "epoch": 0.5467135179826375, "grad_norm": 0.15254896879196167, "learning_rate": 0.006, "loss": 1.0955, "step": 5290 }, { "epoch": 0.5477470028937578, "grad_norm": 0.12627582252025604, "learning_rate": 0.006, "loss": 1.1101, "step": 5300 }, { "epoch": 0.5487804878048781, "grad_norm": 0.132407084107399, "learning_rate": 0.006, "loss": 1.0947, "step": 5310 }, { "epoch": 0.5498139727159983, "grad_norm": 0.1298685371875763, "learning_rate": 0.006, "loss": 1.0853, "step": 5320 }, { "epoch": 0.5508474576271186, "grad_norm": 0.25235456228256226, "learning_rate": 0.006, "loss": 1.1048, "step": 5330 }, { "epoch": 0.5518809425382389, "grad_norm": 0.15979771316051483, "learning_rate": 0.006, "loss": 1.0945, "step": 5340 }, { "epoch": 0.5529144274493593, "grad_norm": 0.12209047377109528, "learning_rate": 0.006, "loss": 1.0918, "step": 5350 }, { "epoch": 0.5539479123604796, "grad_norm": 0.14376087486743927, "learning_rate": 0.006, "loss": 1.1055, "step": 5360 }, { "epoch": 0.5549813972715998, "grad_norm": 0.12132777273654938, "learning_rate": 0.006, "loss": 1.1009, "step": 5370 }, { "epoch": 0.5560148821827201, "grad_norm": 0.12317829579114914, "learning_rate": 0.006, "loss": 1.0873, "step": 5380 }, { "epoch": 0.5570483670938404, "grad_norm": 0.15228794515132904, "learning_rate": 0.006, "loss": 1.0883, "step": 5390 }, { "epoch": 0.5580818520049607, "grad_norm": 0.11975687742233276, "learning_rate": 0.006, "loss": 1.1075, "step": 5400 }, { "epoch": 0.5591153369160811, "grad_norm": 0.15219765901565552, "learning_rate": 0.006, "loss": 1.087, "step": 5410 }, { "epoch": 0.5601488218272013, "grad_norm": 0.12764790654182434, "learning_rate": 0.006, "loss": 1.0928, "step": 5420 }, { "epoch": 0.5611823067383216, "grad_norm": 0.18155109882354736, "learning_rate": 0.006, "loss": 1.0924, "step": 5430 }, { "epoch": 0.5622157916494419, "grad_norm": 0.16299748420715332, "learning_rate": 0.006, "loss": 1.0923, "step": 5440 }, { "epoch": 0.5632492765605622, "grad_norm": 0.1478397250175476, "learning_rate": 0.006, "loss": 1.0838, "step": 5450 }, { "epoch": 0.5642827614716825, "grad_norm": 0.14866343140602112, "learning_rate": 0.006, "loss": 1.0861, "step": 5460 }, { "epoch": 0.5653162463828028, "grad_norm": 0.15744392573833466, "learning_rate": 0.006, "loss": 1.0981, "step": 5470 }, { "epoch": 0.5663497312939231, "grad_norm": 0.1575409471988678, "learning_rate": 0.006, "loss": 1.0972, "step": 5480 }, { "epoch": 0.5673832162050434, "grad_norm": 0.2215876430273056, "learning_rate": 0.006, "loss": 1.0926, "step": 5490 }, { "epoch": 0.5684167011161637, "grad_norm": 0.2060500532388687, "learning_rate": 0.006, "loss": 1.0885, "step": 5500 }, { "epoch": 0.569450186027284, "grad_norm": 0.19035674631595612, "learning_rate": 0.006, "loss": 1.0852, "step": 5510 }, { "epoch": 0.5704836709384044, "grad_norm": 0.15188254415988922, "learning_rate": 0.006, "loss": 1.0883, "step": 5520 }, { "epoch": 0.5715171558495246, "grad_norm": 0.1428184062242508, "learning_rate": 0.006, "loss": 1.0777, "step": 5530 }, { "epoch": 0.5725506407606449, "grad_norm": 0.12388137727975845, "learning_rate": 0.006, "loss": 1.0884, "step": 5540 }, { "epoch": 0.5735841256717652, "grad_norm": 0.1246667355298996, "learning_rate": 0.006, "loss": 1.0807, "step": 5550 }, { "epoch": 0.5746176105828855, "grad_norm": 0.15170446038246155, "learning_rate": 0.006, "loss": 1.0908, "step": 5560 }, { "epoch": 0.5756510954940058, "grad_norm": 0.12245498597621918, "learning_rate": 0.006, "loss": 1.0915, "step": 5570 }, { "epoch": 0.576684580405126, "grad_norm": 0.12406784296035767, "learning_rate": 0.006, "loss": 1.0821, "step": 5580 }, { "epoch": 0.5777180653162464, "grad_norm": 0.173844575881958, "learning_rate": 0.006, "loss": 1.0879, "step": 5590 }, { "epoch": 0.5787515502273667, "grad_norm": 0.1362677961587906, "learning_rate": 0.006, "loss": 1.0929, "step": 5600 }, { "epoch": 0.579785035138487, "grad_norm": 0.13046661019325256, "learning_rate": 0.006, "loss": 1.0891, "step": 5610 }, { "epoch": 0.5808185200496073, "grad_norm": 0.13165681064128876, "learning_rate": 0.006, "loss": 1.0842, "step": 5620 }, { "epoch": 0.5818520049607275, "grad_norm": 0.1174607053399086, "learning_rate": 0.006, "loss": 1.0827, "step": 5630 }, { "epoch": 0.5828854898718479, "grad_norm": 0.12023981660604477, "learning_rate": 0.006, "loss": 1.0725, "step": 5640 }, { "epoch": 0.5839189747829682, "grad_norm": 0.16525237262248993, "learning_rate": 0.006, "loss": 1.0812, "step": 5650 }, { "epoch": 0.5849524596940885, "grad_norm": 0.14079253375530243, "learning_rate": 0.006, "loss": 1.0859, "step": 5660 }, { "epoch": 0.5859859446052088, "grad_norm": 0.1366300880908966, "learning_rate": 0.006, "loss": 1.0847, "step": 5670 }, { "epoch": 0.587019429516329, "grad_norm": 0.12389519810676575, "learning_rate": 0.006, "loss": 1.0897, "step": 5680 }, { "epoch": 0.5880529144274493, "grad_norm": 0.1193692535161972, "learning_rate": 0.006, "loss": 1.0727, "step": 5690 }, { "epoch": 0.5890863993385697, "grad_norm": 0.14679639041423798, "learning_rate": 0.006, "loss": 1.0849, "step": 5700 }, { "epoch": 0.59011988424969, "grad_norm": 0.13050967454910278, "learning_rate": 0.006, "loss": 1.0753, "step": 5710 }, { "epoch": 0.5911533691608103, "grad_norm": 0.12082899361848831, "learning_rate": 0.006, "loss": 1.078, "step": 5720 }, { "epoch": 0.5921868540719305, "grad_norm": 0.1468031257390976, "learning_rate": 0.006, "loss": 1.0866, "step": 5730 }, { "epoch": 0.5932203389830508, "grad_norm": 0.12028497457504272, "learning_rate": 0.006, "loss": 1.0726, "step": 5740 }, { "epoch": 0.5942538238941711, "grad_norm": 0.14379620552062988, "learning_rate": 0.006, "loss": 1.088, "step": 5750 }, { "epoch": 0.5952873088052915, "grad_norm": 0.13077853620052338, "learning_rate": 0.006, "loss": 1.0743, "step": 5760 }, { "epoch": 0.5963207937164118, "grad_norm": 0.14271710813045502, "learning_rate": 0.006, "loss": 1.078, "step": 5770 }, { "epoch": 0.597354278627532, "grad_norm": 0.1357925534248352, "learning_rate": 0.006, "loss": 1.0777, "step": 5780 }, { "epoch": 0.5983877635386523, "grad_norm": 0.12296663224697113, "learning_rate": 0.006, "loss": 1.0561, "step": 5790 }, { "epoch": 0.5994212484497726, "grad_norm": 0.1592247039079666, "learning_rate": 0.006, "loss": 1.0731, "step": 5800 }, { "epoch": 0.600454733360893, "grad_norm": 0.24744771420955658, "learning_rate": 0.006, "loss": 1.0703, "step": 5810 }, { "epoch": 0.6014882182720133, "grad_norm": 0.18649180233478546, "learning_rate": 0.006, "loss": 1.0663, "step": 5820 }, { "epoch": 0.6025217031831335, "grad_norm": 0.15970060229301453, "learning_rate": 0.006, "loss": 1.0814, "step": 5830 }, { "epoch": 0.6035551880942538, "grad_norm": 0.15505217015743256, "learning_rate": 0.006, "loss": 1.0819, "step": 5840 }, { "epoch": 0.6045886730053741, "grad_norm": 0.1281612068414688, "learning_rate": 0.006, "loss": 1.073, "step": 5850 }, { "epoch": 0.6056221579164944, "grad_norm": 0.13322196900844574, "learning_rate": 0.006, "loss": 1.0851, "step": 5860 }, { "epoch": 0.6066556428276147, "grad_norm": 0.1385001540184021, "learning_rate": 0.006, "loss": 1.0714, "step": 5870 }, { "epoch": 0.6076891277387351, "grad_norm": 0.12118349224328995, "learning_rate": 0.006, "loss": 1.0744, "step": 5880 }, { "epoch": 0.6087226126498553, "grad_norm": 0.12691213190555573, "learning_rate": 0.006, "loss": 1.0664, "step": 5890 }, { "epoch": 0.6097560975609756, "grad_norm": 0.14859186112880707, "learning_rate": 0.006, "loss": 1.0821, "step": 5900 }, { "epoch": 0.6107895824720959, "grad_norm": 0.14851497113704681, "learning_rate": 0.006, "loss": 1.08, "step": 5910 }, { "epoch": 0.6118230673832162, "grad_norm": 0.1358513981103897, "learning_rate": 0.006, "loss": 1.0747, "step": 5920 }, { "epoch": 0.6128565522943366, "grad_norm": 0.14147254824638367, "learning_rate": 0.006, "loss": 1.0805, "step": 5930 }, { "epoch": 0.6138900372054568, "grad_norm": 0.13193708658218384, "learning_rate": 0.006, "loss": 1.079, "step": 5940 }, { "epoch": 0.6149235221165771, "grad_norm": 0.1255359649658203, "learning_rate": 0.006, "loss": 1.0785, "step": 5950 }, { "epoch": 0.6159570070276974, "grad_norm": 0.19004510343074799, "learning_rate": 0.006, "loss": 1.0728, "step": 5960 }, { "epoch": 0.6169904919388177, "grad_norm": 0.2386738657951355, "learning_rate": 0.006, "loss": 1.0836, "step": 5970 }, { "epoch": 0.618023976849938, "grad_norm": 0.14890602231025696, "learning_rate": 0.006, "loss": 1.0791, "step": 5980 }, { "epoch": 0.6190574617610582, "grad_norm": 0.15145020186901093, "learning_rate": 0.006, "loss": 1.0717, "step": 5990 }, { "epoch": 0.6200909466721786, "grad_norm": 0.12245503813028336, "learning_rate": 0.006, "loss": 1.064, "step": 6000 }, { "epoch": 0.6211244315832989, "grad_norm": 0.13569168746471405, "learning_rate": 0.006, "loss": 1.0717, "step": 6010 }, { "epoch": 0.6221579164944192, "grad_norm": 0.14238642156124115, "learning_rate": 0.006, "loss": 1.072, "step": 6020 }, { "epoch": 0.6231914014055395, "grad_norm": 0.13332600891590118, "learning_rate": 0.006, "loss": 1.0821, "step": 6030 }, { "epoch": 0.6242248863166597, "grad_norm": 0.12775711715221405, "learning_rate": 0.006, "loss": 1.0665, "step": 6040 }, { "epoch": 0.62525837122778, "grad_norm": 0.12378199398517609, "learning_rate": 0.006, "loss": 1.0594, "step": 6050 }, { "epoch": 0.6262918561389004, "grad_norm": 0.127638041973114, "learning_rate": 0.006, "loss": 1.064, "step": 6060 }, { "epoch": 0.6273253410500207, "grad_norm": 0.1450446993112564, "learning_rate": 0.006, "loss": 1.0718, "step": 6070 }, { "epoch": 0.628358825961141, "grad_norm": 0.14722521603107452, "learning_rate": 0.006, "loss": 1.0786, "step": 6080 }, { "epoch": 0.6293923108722612, "grad_norm": 0.11357062309980392, "learning_rate": 0.006, "loss": 1.0713, "step": 6090 }, { "epoch": 0.6304257957833815, "grad_norm": 0.1290910542011261, "learning_rate": 0.006, "loss": 1.0692, "step": 6100 }, { "epoch": 0.6314592806945019, "grad_norm": 0.12275319546461105, "learning_rate": 0.006, "loss": 1.0599, "step": 6110 }, { "epoch": 0.6324927656056222, "grad_norm": 0.14953821897506714, "learning_rate": 0.006, "loss": 1.0735, "step": 6120 }, { "epoch": 0.6335262505167425, "grad_norm": 0.15243026614189148, "learning_rate": 0.006, "loss": 1.0641, "step": 6130 }, { "epoch": 0.6345597354278627, "grad_norm": 0.17082957923412323, "learning_rate": 0.006, "loss": 1.0669, "step": 6140 }, { "epoch": 0.635593220338983, "grad_norm": 0.14030581712722778, "learning_rate": 0.006, "loss": 1.0659, "step": 6150 }, { "epoch": 0.6366267052501033, "grad_norm": 0.1736230105161667, "learning_rate": 0.006, "loss": 1.0698, "step": 6160 }, { "epoch": 0.6376601901612237, "grad_norm": 0.12367543578147888, "learning_rate": 0.006, "loss": 1.0657, "step": 6170 }, { "epoch": 0.638693675072344, "grad_norm": 0.14306339621543884, "learning_rate": 0.006, "loss": 1.0731, "step": 6180 }, { "epoch": 0.6397271599834642, "grad_norm": 0.1417074203491211, "learning_rate": 0.006, "loss": 1.0704, "step": 6190 }, { "epoch": 0.6407606448945845, "grad_norm": 0.12904278934001923, "learning_rate": 0.006, "loss": 1.0619, "step": 6200 }, { "epoch": 0.6417941298057048, "grad_norm": 0.13394799828529358, "learning_rate": 0.006, "loss": 1.054, "step": 6210 }, { "epoch": 0.6428276147168251, "grad_norm": 0.13006740808486938, "learning_rate": 0.006, "loss": 1.058, "step": 6220 }, { "epoch": 0.6438610996279455, "grad_norm": 0.1411658525466919, "learning_rate": 0.006, "loss": 1.0688, "step": 6230 }, { "epoch": 0.6448945845390658, "grad_norm": 0.12172463536262512, "learning_rate": 0.006, "loss": 1.0599, "step": 6240 }, { "epoch": 0.645928069450186, "grad_norm": 0.1132531389594078, "learning_rate": 0.006, "loss": 1.0493, "step": 6250 }, { "epoch": 0.6469615543613063, "grad_norm": 0.15614502131938934, "learning_rate": 0.006, "loss": 1.0637, "step": 6260 }, { "epoch": 0.6479950392724266, "grad_norm": 0.14655426144599915, "learning_rate": 0.006, "loss": 1.0537, "step": 6270 }, { "epoch": 0.649028524183547, "grad_norm": 0.1248323991894722, "learning_rate": 0.006, "loss": 1.0735, "step": 6280 }, { "epoch": 0.6500620090946673, "grad_norm": 0.13113942742347717, "learning_rate": 0.006, "loss": 1.0561, "step": 6290 }, { "epoch": 0.6510954940057875, "grad_norm": 0.11488605290651321, "learning_rate": 0.006, "loss": 1.0626, "step": 6300 }, { "epoch": 0.6521289789169078, "grad_norm": 0.13470590114593506, "learning_rate": 0.006, "loss": 1.0629, "step": 6310 }, { "epoch": 0.6531624638280281, "grad_norm": 0.13424867391586304, "learning_rate": 0.006, "loss": 1.0634, "step": 6320 }, { "epoch": 0.6541959487391484, "grad_norm": 0.13190287351608276, "learning_rate": 0.006, "loss": 1.0697, "step": 6330 }, { "epoch": 0.6552294336502688, "grad_norm": 0.12076553702354431, "learning_rate": 0.006, "loss": 1.0559, "step": 6340 }, { "epoch": 0.656262918561389, "grad_norm": 0.12332470715045929, "learning_rate": 0.006, "loss": 1.0511, "step": 6350 }, { "epoch": 0.6572964034725093, "grad_norm": 0.13250276446342468, "learning_rate": 0.006, "loss": 1.0772, "step": 6360 }, { "epoch": 0.6583298883836296, "grad_norm": 0.12431011348962784, "learning_rate": 0.006, "loss": 1.0574, "step": 6370 }, { "epoch": 0.6593633732947499, "grad_norm": 0.11310591548681259, "learning_rate": 0.006, "loss": 1.0548, "step": 6380 }, { "epoch": 0.6603968582058702, "grad_norm": 0.12900187075138092, "learning_rate": 0.006, "loss": 1.0624, "step": 6390 }, { "epoch": 0.6614303431169904, "grad_norm": 0.21367403864860535, "learning_rate": 0.006, "loss": 1.0681, "step": 6400 }, { "epoch": 0.6624638280281108, "grad_norm": 0.17459499835968018, "learning_rate": 0.006, "loss": 1.0547, "step": 6410 }, { "epoch": 0.6634973129392311, "grad_norm": 0.12943004071712494, "learning_rate": 0.006, "loss": 1.0659, "step": 6420 }, { "epoch": 0.6645307978503514, "grad_norm": 0.12108403444290161, "learning_rate": 0.006, "loss": 1.0721, "step": 6430 }, { "epoch": 0.6655642827614717, "grad_norm": 0.14485575258731842, "learning_rate": 0.006, "loss": 1.0502, "step": 6440 }, { "epoch": 0.6665977676725919, "grad_norm": 0.12145304679870605, "learning_rate": 0.006, "loss": 1.0489, "step": 6450 }, { "epoch": 0.6676312525837123, "grad_norm": 0.12879317998886108, "learning_rate": 0.006, "loss": 1.0603, "step": 6460 }, { "epoch": 0.6686647374948326, "grad_norm": 0.13428542017936707, "learning_rate": 0.006, "loss": 1.0538, "step": 6470 }, { "epoch": 0.6696982224059529, "grad_norm": 0.13670314848423004, "learning_rate": 0.006, "loss": 1.053, "step": 6480 }, { "epoch": 0.6707317073170732, "grad_norm": 0.12726275622844696, "learning_rate": 0.006, "loss": 1.0555, "step": 6490 }, { "epoch": 0.6717651922281934, "grad_norm": 0.1293756067752838, "learning_rate": 0.006, "loss": 1.0579, "step": 6500 }, { "epoch": 0.6727986771393137, "grad_norm": 0.15011049807071686, "learning_rate": 0.006, "loss": 1.0528, "step": 6510 }, { "epoch": 0.6738321620504341, "grad_norm": 0.1232418343424797, "learning_rate": 0.006, "loss": 1.0474, "step": 6520 }, { "epoch": 0.6748656469615544, "grad_norm": 0.13466301560401917, "learning_rate": 0.006, "loss": 1.0561, "step": 6530 }, { "epoch": 0.6758991318726747, "grad_norm": 0.18297189474105835, "learning_rate": 0.006, "loss": 1.0569, "step": 6540 }, { "epoch": 0.6769326167837949, "grad_norm": 0.13041065633296967, "learning_rate": 0.006, "loss": 1.0614, "step": 6550 }, { "epoch": 0.6779661016949152, "grad_norm": 0.15428952872753143, "learning_rate": 0.006, "loss": 1.0526, "step": 6560 }, { "epoch": 0.6789995866060355, "grad_norm": 0.1226574033498764, "learning_rate": 0.006, "loss": 1.0476, "step": 6570 }, { "epoch": 0.6800330715171559, "grad_norm": 0.1592113971710205, "learning_rate": 0.006, "loss": 1.057, "step": 6580 }, { "epoch": 0.6810665564282762, "grad_norm": 0.131056547164917, "learning_rate": 0.006, "loss": 1.0643, "step": 6590 }, { "epoch": 0.6821000413393964, "grad_norm": 0.17470133304595947, "learning_rate": 0.006, "loss": 1.0545, "step": 6600 }, { "epoch": 0.6831335262505167, "grad_norm": 0.14145781099796295, "learning_rate": 0.006, "loss": 1.0459, "step": 6610 }, { "epoch": 0.684167011161637, "grad_norm": 0.11432395130395889, "learning_rate": 0.006, "loss": 1.0454, "step": 6620 }, { "epoch": 0.6852004960727573, "grad_norm": 0.12126456946134567, "learning_rate": 0.006, "loss": 1.0467, "step": 6630 }, { "epoch": 0.6862339809838777, "grad_norm": 0.12609295547008514, "learning_rate": 0.006, "loss": 1.0734, "step": 6640 }, { "epoch": 0.687267465894998, "grad_norm": 0.15360839664936066, "learning_rate": 0.006, "loss": 1.0424, "step": 6650 }, { "epoch": 0.6883009508061182, "grad_norm": 0.13050951063632965, "learning_rate": 0.006, "loss": 1.0532, "step": 6660 }, { "epoch": 0.6893344357172385, "grad_norm": 0.11345363408327103, "learning_rate": 0.006, "loss": 1.0616, "step": 6670 }, { "epoch": 0.6903679206283588, "grad_norm": 0.12678933143615723, "learning_rate": 0.006, "loss": 1.0472, "step": 6680 }, { "epoch": 0.6914014055394792, "grad_norm": 0.12220112234354019, "learning_rate": 0.006, "loss": 1.048, "step": 6690 }, { "epoch": 0.6924348904505995, "grad_norm": 0.12634249031543732, "learning_rate": 0.006, "loss": 1.0489, "step": 6700 }, { "epoch": 0.6934683753617197, "grad_norm": 0.14149808883666992, "learning_rate": 0.006, "loss": 1.0548, "step": 6710 }, { "epoch": 0.69450186027284, "grad_norm": 0.12240725010633469, "learning_rate": 0.006, "loss": 1.0466, "step": 6720 }, { "epoch": 0.6955353451839603, "grad_norm": 0.1353718340396881, "learning_rate": 0.006, "loss": 1.0423, "step": 6730 }, { "epoch": 0.6965688300950806, "grad_norm": 0.18347500264644623, "learning_rate": 0.006, "loss": 1.0485, "step": 6740 }, { "epoch": 0.697602315006201, "grad_norm": 0.12348928302526474, "learning_rate": 0.006, "loss": 1.0573, "step": 6750 }, { "epoch": 0.6986357999173212, "grad_norm": 0.1428414285182953, "learning_rate": 0.006, "loss": 1.0412, "step": 6760 }, { "epoch": 0.6996692848284415, "grad_norm": 0.1410830169916153, "learning_rate": 0.006, "loss": 1.0569, "step": 6770 }, { "epoch": 0.7007027697395618, "grad_norm": 0.14577074348926544, "learning_rate": 0.005935620400122873, "loss": 1.0546, "step": 6780 }, { "epoch": 0.7017362546506821, "grad_norm": 0.14377078413963318, "learning_rate": 0.005842197295969981, "loss": 1.045, "step": 6790 }, { "epoch": 0.7027697395618024, "grad_norm": 0.12980251014232635, "learning_rate": 0.005750244615429315, "loss": 1.0595, "step": 6800 }, { "epoch": 0.7038032244729227, "grad_norm": 0.13310031592845917, "learning_rate": 0.005659739214915371, "loss": 1.0492, "step": 6810 }, { "epoch": 0.704836709384043, "grad_norm": 0.12173794209957123, "learning_rate": 0.005570658315108789, "loss": 1.0409, "step": 6820 }, { "epoch": 0.7058701942951633, "grad_norm": 0.17054332792758942, "learning_rate": 0.005482979495223033, "loss": 1.0505, "step": 6830 }, { "epoch": 0.7069036792062836, "grad_norm": 0.18403039872646332, "learning_rate": 0.005396680687361299, "loss": 1.0462, "step": 6840 }, { "epoch": 0.7079371641174039, "grad_norm": 0.13148552179336548, "learning_rate": 0.005311740170962234, "loss": 1.049, "step": 6850 }, { "epoch": 0.7089706490285241, "grad_norm": 0.13557471334934235, "learning_rate": 0.005228136567333095, "loss": 1.0348, "step": 6860 }, { "epoch": 0.7100041339396445, "grad_norm": 0.11231327056884766, "learning_rate": 0.005145848834268932, "loss": 1.0309, "step": 6870 }, { "epoch": 0.7110376188507648, "grad_norm": 0.14308036863803864, "learning_rate": 0.005064856260756483, "loss": 1.024, "step": 6880 }, { "epoch": 0.7120711037618851, "grad_norm": 0.11384549736976624, "learning_rate": 0.0049851384617614035, "loss": 1.0255, "step": 6890 }, { "epoch": 0.7131045886730054, "grad_norm": 0.1373758465051651, "learning_rate": 0.004906675373097566, "loss": 1.0248, "step": 6900 }, { "epoch": 0.7141380735841256, "grad_norm": 0.13972637057304382, "learning_rate": 0.00482944724637709, "loss": 1.0197, "step": 6910 }, { "epoch": 0.7151715584952459, "grad_norm": 0.11777855455875397, "learning_rate": 0.004753434644039878, "loss": 1.0197, "step": 6920 }, { "epoch": 0.7162050434063663, "grad_norm": 0.10751146823167801, "learning_rate": 0.004678618434461362, "loss": 1.0272, "step": 6930 }, { "epoch": 0.7172385283174866, "grad_norm": 0.11851785331964493, "learning_rate": 0.004604979787137271, "loss": 1.0109, "step": 6940 }, { "epoch": 0.7182720132286069, "grad_norm": 0.1341877430677414, "learning_rate": 0.004532500167944173, "loss": 1.0106, "step": 6950 }, { "epoch": 0.7193054981397271, "grad_norm": 0.11008955538272858, "learning_rate": 0.004461161334474619, "loss": 1.0152, "step": 6960 }, { "epoch": 0.7203389830508474, "grad_norm": 0.16443827748298645, "learning_rate": 0.0043909453314457106, "loss": 1.0118, "step": 6970 }, { "epoch": 0.7213724679619677, "grad_norm": 0.11731505393981934, "learning_rate": 0.004321834486179928, "loss": 1.0081, "step": 6980 }, { "epoch": 0.7224059528730881, "grad_norm": 0.11518187075853348, "learning_rate": 0.004253811404157096, "loss": 1.004, "step": 6990 }, { "epoch": 0.7234394377842084, "grad_norm": 0.10781227797269821, "learning_rate": 0.004186858964636351, "loss": 0.9879, "step": 7000 }, { "epoch": 0.7244729226953287, "grad_norm": 0.12845131754875183, "learning_rate": 0.004120960316347017, "loss": 1.0148, "step": 7010 }, { "epoch": 0.7255064076064489, "grad_norm": 0.11902609467506409, "learning_rate": 0.004056098873247311, "loss": 1.0012, "step": 7020 }, { "epoch": 0.7265398925175692, "grad_norm": 0.11711709946393967, "learning_rate": 0.003992258310349796, "loss": 1.0134, "step": 7030 }, { "epoch": 0.7275733774286895, "grad_norm": 0.12396284192800522, "learning_rate": 0.003929422559612546, "loss": 1.0023, "step": 7040 }, { "epoch": 0.7286068623398099, "grad_norm": 0.10830879211425781, "learning_rate": 0.0038675758058949712, "loss": 1.0126, "step": 7050 }, { "epoch": 0.7296403472509302, "grad_norm": 0.11904972046613693, "learning_rate": 0.003806702482977311, "loss": 0.9993, "step": 7060 }, { "epoch": 0.7306738321620504, "grad_norm": 0.14758719503879547, "learning_rate": 0.0037467872696427614, "loss": 0.9919, "step": 7070 }, { "epoch": 0.7317073170731707, "grad_norm": 0.10925702750682831, "learning_rate": 0.0036878150858212825, "loss": 0.9991, "step": 7080 }, { "epoch": 0.732740801984291, "grad_norm": 0.1237645223736763, "learning_rate": 0.0036297710887940875, "loss": 1.004, "step": 7090 }, { "epoch": 0.7337742868954114, "grad_norm": 0.12318721413612366, "learning_rate": 0.003572640669457879, "loss": 0.9833, "step": 7100 }, { "epoch": 0.7348077718065317, "grad_norm": 0.10309872031211853, "learning_rate": 0.003516409448647884, "loss": 0.9919, "step": 7110 }, { "epoch": 0.7358412567176519, "grad_norm": 0.10170971602201462, "learning_rate": 0.003461063273518752, "loss": 0.9908, "step": 7120 }, { "epoch": 0.7368747416287722, "grad_norm": 0.13282230496406555, "learning_rate": 0.0034065882139824315, "loss": 0.9849, "step": 7130 }, { "epoch": 0.7379082265398925, "grad_norm": 0.10541505366563797, "learning_rate": 0.0033529705592020968, "loss": 0.9893, "step": 7140 }, { "epoch": 0.7389417114510128, "grad_norm": 0.10669828951358795, "learning_rate": 0.0033001968141412714, "loss": 0.9841, "step": 7150 }, { "epoch": 0.7399751963621332, "grad_norm": 0.11185015738010406, "learning_rate": 0.0032482536961672545, "loss": 0.9888, "step": 7160 }, { "epoch": 0.7410086812732534, "grad_norm": 0.11848310381174088, "learning_rate": 0.003197128131708016, "loss": 0.9753, "step": 7170 }, { "epoch": 0.7420421661843737, "grad_norm": 0.12162330746650696, "learning_rate": 0.0031468072529617073, "loss": 0.9896, "step": 7180 }, { "epoch": 0.743075651095494, "grad_norm": 0.10726076364517212, "learning_rate": 0.0030972783946579593, "loss": 0.967, "step": 7190 }, { "epoch": 0.7441091360066143, "grad_norm": 0.11569386720657349, "learning_rate": 0.003048529090870162, "loss": 0.9691, "step": 7200 }, { "epoch": 0.7451426209177346, "grad_norm": 0.11116993427276611, "learning_rate": 0.0030005470718779117, "loss": 0.9783, "step": 7210 }, { "epoch": 0.7461761058288549, "grad_norm": 0.10824961215257645, "learning_rate": 0.002953320261078841, "loss": 0.9589, "step": 7220 }, { "epoch": 0.7472095907399752, "grad_norm": 0.1097089946269989, "learning_rate": 0.00290683677194906, "loss": 0.9707, "step": 7230 }, { "epoch": 0.7482430756510955, "grad_norm": 0.13413402438163757, "learning_rate": 0.002861084905051434, "loss": 0.9583, "step": 7240 }, { "epoch": 0.7492765605622158, "grad_norm": 0.10966306924819946, "learning_rate": 0.0028160531450909495, "loss": 0.9656, "step": 7250 }, { "epoch": 0.7503100454733361, "grad_norm": 0.10541026294231415, "learning_rate": 0.0027717301580164277, "loss": 0.9568, "step": 7260 }, { "epoch": 0.7513435303844563, "grad_norm": 0.10642149299383163, "learning_rate": 0.002728104788167857, "loss": 0.973, "step": 7270 }, { "epoch": 0.7523770152955767, "grad_norm": 0.10307744145393372, "learning_rate": 0.0026851660554686207, "loss": 0.973, "step": 7280 }, { "epoch": 0.753410500206697, "grad_norm": 0.10917676985263824, "learning_rate": 0.002642903152661921, "loss": 0.9739, "step": 7290 }, { "epoch": 0.7544439851178173, "grad_norm": 0.12075357139110565, "learning_rate": 0.002601305442590699, "loss": 0.9668, "step": 7300 }, { "epoch": 0.7554774700289376, "grad_norm": 0.10885104537010193, "learning_rate": 0.0025603624555203654, "loss": 0.9732, "step": 7310 }, { "epoch": 0.7565109549400578, "grad_norm": 0.11787907034158707, "learning_rate": 0.0025200638865036733, "loss": 0.9597, "step": 7320 }, { "epoch": 0.7575444398511781, "grad_norm": 0.11566518247127533, "learning_rate": 0.0024803995927870632, "loss": 0.9595, "step": 7330 }, { "epoch": 0.7585779247622985, "grad_norm": 0.09844059497117996, "learning_rate": 0.002441359591257831, "loss": 0.9561, "step": 7340 }, { "epoch": 0.7596114096734188, "grad_norm": 0.1304013878107071, "learning_rate": 0.002402934055931478, "loss": 0.9454, "step": 7350 }, { "epoch": 0.7606448945845391, "grad_norm": 0.1041470617055893, "learning_rate": 0.002365113315478606, "loss": 0.9453, "step": 7360 }, { "epoch": 0.7616783794956593, "grad_norm": 0.10921091586351395, "learning_rate": 0.00232788785079074, "loss": 0.9644, "step": 7370 }, { "epoch": 0.7627118644067796, "grad_norm": 0.12734107673168182, "learning_rate": 0.0022912482925844615, "loss": 0.9491, "step": 7380 }, { "epoch": 0.7637453493179, "grad_norm": 0.11103381216526031, "learning_rate": 0.0022551854190432517, "loss": 0.9379, "step": 7390 }, { "epoch": 0.7647788342290203, "grad_norm": 0.10525085031986237, "learning_rate": 0.0022196901534964524, "loss": 0.956, "step": 7400 }, { "epoch": 0.7658123191401406, "grad_norm": 0.10325931012630463, "learning_rate": 0.0021847535621347543, "loss": 0.9478, "step": 7410 }, { "epoch": 0.7668458040512609, "grad_norm": 0.10702349990606308, "learning_rate": 0.0021503668517616494, "loss": 0.9408, "step": 7420 }, { "epoch": 0.7678792889623811, "grad_norm": 0.10343906283378601, "learning_rate": 0.002116521367580266, "loss": 0.9505, "step": 7430 }, { "epoch": 0.7689127738735014, "grad_norm": 0.11631717532873154, "learning_rate": 0.0020832085910150425, "loss": 0.9483, "step": 7440 }, { "epoch": 0.7699462587846218, "grad_norm": 0.114792600274086, "learning_rate": 0.0020504201375676876, "loss": 0.9425, "step": 7450 }, { "epoch": 0.7709797436957421, "grad_norm": 0.10078386217355728, "learning_rate": 0.0020181477547068817, "loss": 0.955, "step": 7460 }, { "epoch": 0.7720132286068624, "grad_norm": 0.10332197695970535, "learning_rate": 0.0019863833197911983, "loss": 0.9488, "step": 7470 }, { "epoch": 0.7730467135179826, "grad_norm": 0.1078406572341919, "learning_rate": 0.001955118838024712, "loss": 0.9536, "step": 7480 }, { "epoch": 0.7740801984291029, "grad_norm": 0.1059098169207573, "learning_rate": 0.0019243464404447924, "loss": 0.9396, "step": 7490 }, { "epoch": 0.7751136833402232, "grad_norm": 0.10873686522245407, "learning_rate": 0.0018940583819415564, "loss": 0.9412, "step": 7500 }, { "epoch": 0.7761471682513436, "grad_norm": 0.10201577097177505, "learning_rate": 0.0018642470393085062, "loss": 0.9406, "step": 7510 }, { "epoch": 0.7771806531624639, "grad_norm": 0.10363386571407318, "learning_rate": 0.0018349049093238407, "loss": 0.9349, "step": 7520 }, { "epoch": 0.7782141380735841, "grad_norm": 0.10290543735027313, "learning_rate": 0.001806024606861967, "loss": 0.9433, "step": 7530 }, { "epoch": 0.7792476229847044, "grad_norm": 0.09915654361248016, "learning_rate": 0.0017775988630347405, "loss": 0.9333, "step": 7540 }, { "epoch": 0.7802811078958247, "grad_norm": 0.10667795687913895, "learning_rate": 0.0017496205233619542, "loss": 0.9282, "step": 7550 }, { "epoch": 0.781314592806945, "grad_norm": 0.0981353297829628, "learning_rate": 0.001722082545970627, "loss": 0.9443, "step": 7560 }, { "epoch": 0.7823480777180654, "grad_norm": 0.09341371059417725, "learning_rate": 0.0016949779998226353, "loss": 0.9255, "step": 7570 }, { "epoch": 0.7833815626291856, "grad_norm": 0.09437743574380875, "learning_rate": 0.0016683000629702357, "loss": 0.9374, "step": 7580 }, { "epoch": 0.7844150475403059, "grad_norm": 0.10137518495321274, "learning_rate": 0.001642042020839051, "loss": 0.9434, "step": 7590 }, { "epoch": 0.7854485324514262, "grad_norm": 0.10156513750553131, "learning_rate": 0.001616197264538076, "loss": 0.9282, "step": 7600 }, { "epoch": 0.7864820173625465, "grad_norm": 0.10588687658309937, "learning_rate": 0.0015907592891962843, "loss": 0.9329, "step": 7610 }, { "epoch": 0.7875155022736668, "grad_norm": 0.10155037045478821, "learning_rate": 0.0015657216923254177, "loss": 0.9268, "step": 7620 }, { "epoch": 0.788548987184787, "grad_norm": 0.10379229485988617, "learning_rate": 0.0015410781722085428, "loss": 0.9437, "step": 7630 }, { "epoch": 0.7895824720959074, "grad_norm": 0.10199972987174988, "learning_rate": 0.0015168225263139693, "loss": 0.9425, "step": 7640 }, { "epoch": 0.7906159570070277, "grad_norm": 0.10469642281532288, "learning_rate": 0.0014929486497341341, "loss": 0.924, "step": 7650 }, { "epoch": 0.791649441918148, "grad_norm": 0.09616528451442719, "learning_rate": 0.0014694505336490583, "loss": 0.931, "step": 7660 }, { "epoch": 0.7926829268292683, "grad_norm": 0.10645470768213272, "learning_rate": 0.001446322263813983, "loss": 0.9266, "step": 7670 }, { "epoch": 0.7937164117403885, "grad_norm": 0.10266025364398956, "learning_rate": 0.0014235580190708143, "loss": 0.9251, "step": 7680 }, { "epoch": 0.7947498966515089, "grad_norm": 0.10014299303293228, "learning_rate": 0.0014011520698829947, "loss": 0.9268, "step": 7690 }, { "epoch": 0.7957833815626292, "grad_norm": 0.10017547011375427, "learning_rate": 0.0013790987768934346, "loss": 0.9345, "step": 7700 }, { "epoch": 0.7968168664737495, "grad_norm": 0.09650063514709473, "learning_rate": 0.0013573925895051418, "loss": 0.919, "step": 7710 }, { "epoch": 0.7978503513848698, "grad_norm": 0.11645365506410599, "learning_rate": 0.00133602804448419, "loss": 0.9256, "step": 7720 }, { "epoch": 0.79888383629599, "grad_norm": 0.09640100598335266, "learning_rate": 0.0013149997645846788, "loss": 0.9252, "step": 7730 }, { "epoch": 0.7999173212071103, "grad_norm": 0.09893805533647537, "learning_rate": 0.0012943024571953315, "loss": 0.9209, "step": 7740 }, { "epoch": 0.8009508061182307, "grad_norm": 0.10975757986307144, "learning_rate": 0.001273930913007398, "loss": 0.9259, "step": 7750 }, { "epoch": 0.801984291029351, "grad_norm": 0.09872964024543762, "learning_rate": 0.0012538800047035218, "loss": 0.9139, "step": 7760 }, { "epoch": 0.8030177759404713, "grad_norm": 0.10294859856367111, "learning_rate": 0.0012341446856672469, "loss": 0.9219, "step": 7770 }, { "epoch": 0.8040512608515916, "grad_norm": 0.09699690341949463, "learning_rate": 0.0012147199887128318, "loss": 0.9171, "step": 7780 }, { "epoch": 0.8050847457627118, "grad_norm": 0.10195200890302658, "learning_rate": 0.0011956010248350593, "loss": 0.9181, "step": 7790 }, { "epoch": 0.8061182306738321, "grad_norm": 0.10153238475322723, "learning_rate": 0.0011767829819787205, "loss": 0.9122, "step": 7800 }, { "epoch": 0.8071517155849525, "grad_norm": 0.09719067811965942, "learning_rate": 0.0011582611238274692, "loss": 0.9261, "step": 7810 }, { "epoch": 0.8081852004960728, "grad_norm": 0.09960606694221497, "learning_rate": 0.0011400307886117367, "loss": 0.9159, "step": 7820 }, { "epoch": 0.8092186854071931, "grad_norm": 0.10018911957740784, "learning_rate": 0.001122087387935411, "loss": 0.9043, "step": 7830 }, { "epoch": 0.8102521703183133, "grad_norm": 0.09987188875675201, "learning_rate": 0.0011044264056209812, "loss": 0.923, "step": 7840 }, { "epoch": 0.8112856552294336, "grad_norm": 0.10447081923484802, "learning_rate": 0.0010870433965728626, "loss": 0.912, "step": 7850 }, { "epoch": 0.812319140140554, "grad_norm": 0.09891889244318008, "learning_rate": 0.0010699339856586065, "loss": 0.9231, "step": 7860 }, { "epoch": 0.8133526250516743, "grad_norm": 0.11543388664722443, "learning_rate": 0.0010530938666077263, "loss": 0.9189, "step": 7870 }, { "epoch": 0.8143861099627946, "grad_norm": 0.10272625833749771, "learning_rate": 0.0010365188009278474, "loss": 0.9218, "step": 7880 }, { "epoch": 0.8154195948739148, "grad_norm": 0.11380596458911896, "learning_rate": 0.0010202046168379232, "loss": 0.9089, "step": 7890 }, { "epoch": 0.8164530797850351, "grad_norm": 0.10323087871074677, "learning_rate": 0.0010041472082182382, "loss": 0.9053, "step": 7900 }, { "epoch": 0.8174865646961554, "grad_norm": 0.10377652198076248, "learning_rate": 0.0009883425335769375, "loss": 0.9167, "step": 7910 }, { "epoch": 0.8185200496072758, "grad_norm": 0.09777701646089554, "learning_rate": 0.0009727866150328237, "loss": 0.9134, "step": 7920 }, { "epoch": 0.8195535345183961, "grad_norm": 0.10352311283349991, "learning_rate": 0.0009574755373141626, "loss": 0.9203, "step": 7930 }, { "epoch": 0.8205870194295163, "grad_norm": 0.1033698320388794, "learning_rate": 0.0009424054467732486, "loss": 0.916, "step": 7940 }, { "epoch": 0.8216205043406366, "grad_norm": 0.10148289799690247, "learning_rate": 0.0009275725504164789, "loss": 0.9055, "step": 7950 }, { "epoch": 0.8226539892517569, "grad_norm": 0.1043614000082016, "learning_rate": 0.0009129731149496945, "loss": 0.9103, "step": 7960 }, { "epoch": 0.8236874741628772, "grad_norm": 0.10638295859098434, "learning_rate": 0.0008986034658385467, "loss": 0.9095, "step": 7970 }, { "epoch": 0.8247209590739976, "grad_norm": 0.09755829721689224, "learning_rate": 0.0008844599863836532, "loss": 0.911, "step": 7980 }, { "epoch": 0.8257544439851178, "grad_norm": 0.09516286849975586, "learning_rate": 0.0008705391168103099, "loss": 0.8951, "step": 7990 }, { "epoch": 0.8267879288962381, "grad_norm": 0.09857280552387238, "learning_rate": 0.0008568373533725312, "loss": 0.9129, "step": 8000 }, { "epoch": 0.8278214138073584, "grad_norm": 0.09658730775117874, "learning_rate": 0.0008433512474711912, "loss": 0.9123, "step": 8010 }, { "epoch": 0.8288548987184787, "grad_norm": 0.09653860330581665, "learning_rate": 0.0008300774047860455, "loss": 0.8981, "step": 8020 }, { "epoch": 0.829888383629599, "grad_norm": 0.10416875779628754, "learning_rate": 0.0008170124844214137, "loss": 0.9173, "step": 8030 }, { "epoch": 0.8309218685407193, "grad_norm": 0.09936217218637466, "learning_rate": 0.0008041531980653093, "loss": 0.9129, "step": 8040 }, { "epoch": 0.8319553534518396, "grad_norm": 0.10089723765850067, "learning_rate": 0.0007914963091618035, "loss": 0.9138, "step": 8050 }, { "epoch": 0.8329888383629599, "grad_norm": 0.09607098996639252, "learning_rate": 0.0007790386320964167, "loss": 0.8986, "step": 8060 }, { "epoch": 0.8340223232740802, "grad_norm": 0.09406198561191559, "learning_rate": 0.0007667770313943293, "loss": 0.9062, "step": 8070 }, { "epoch": 0.8350558081852005, "grad_norm": 0.1027892455458641, "learning_rate": 0.0007547084209312148, "loss": 0.9082, "step": 8080 }, { "epoch": 0.8360892930963207, "grad_norm": 0.09503350406885147, "learning_rate": 0.0007428297631564922, "loss": 0.915, "step": 8090 }, { "epoch": 0.8371227780074411, "grad_norm": 0.11514163017272949, "learning_rate": 0.000731138068328804, "loss": 0.8963, "step": 8100 }, { "epoch": 0.8381562629185614, "grad_norm": 0.09855664521455765, "learning_rate": 0.0007196303937635279, "loss": 0.8902, "step": 8110 }, { "epoch": 0.8391897478296817, "grad_norm": 0.0985339879989624, "learning_rate": 0.000708303843092133, "loss": 0.9087, "step": 8120 }, { "epoch": 0.840223232740802, "grad_norm": 0.10324376076459885, "learning_rate": 0.0006971555655331906, "loss": 0.9129, "step": 8130 }, { "epoch": 0.8412567176519223, "grad_norm": 0.10233034193515778, "learning_rate": 0.0006861827551748618, "loss": 0.8974, "step": 8140 }, { "epoch": 0.8422902025630425, "grad_norm": 0.09102931618690491, "learning_rate": 0.0006753826502686767, "loss": 0.8945, "step": 8150 }, { "epoch": 0.8433236874741629, "grad_norm": 0.09884581714868546, "learning_rate": 0.0006647525325344291, "loss": 0.9076, "step": 8160 }, { "epoch": 0.8443571723852832, "grad_norm": 0.10044873505830765, "learning_rate": 0.0006542897264760131, "loss": 0.897, "step": 8170 }, { "epoch": 0.8453906572964035, "grad_norm": 0.10729029029607773, "learning_rate": 0.000643991598708026, "loss": 0.9061, "step": 8180 }, { "epoch": 0.8464241422075238, "grad_norm": 0.09865026921033859, "learning_rate": 0.0006338555572929715, "loss": 0.9198, "step": 8190 }, { "epoch": 0.847457627118644, "grad_norm": 0.0979701355099678, "learning_rate": 0.0006238790510888947, "loss": 0.9027, "step": 8200 }, { "epoch": 0.8484911120297643, "grad_norm": 0.10607071965932846, "learning_rate": 0.0006140595691072842, "loss": 0.9005, "step": 8210 }, { "epoch": 0.8495245969408847, "grad_norm": 0.09750446677207947, "learning_rate": 0.0006043946398810819, "loss": 0.8955, "step": 8220 }, { "epoch": 0.850558081852005, "grad_norm": 0.10503144562244415, "learning_rate": 0.0005948818308426382, "loss": 0.8987, "step": 8230 }, { "epoch": 0.8515915667631253, "grad_norm": 0.09300017356872559, "learning_rate": 0.0005855187477114587, "loss": 0.8943, "step": 8240 }, { "epoch": 0.8526250516742455, "grad_norm": 0.09706971049308777, "learning_rate": 0.0005763030338915879, "loss": 0.8942, "step": 8250 }, { "epoch": 0.8536585365853658, "grad_norm": 0.11037931591272354, "learning_rate": 0.000567232369878477, "loss": 0.8958, "step": 8260 }, { "epoch": 0.8546920214964862, "grad_norm": 0.09634577482938766, "learning_rate": 0.0005583044726751868, "loss": 0.8976, "step": 8270 }, { "epoch": 0.8557255064076065, "grad_norm": 0.10242386907339096, "learning_rate": 0.0005495170952177806, "loss": 0.8902, "step": 8280 }, { "epoch": 0.8567589913187268, "grad_norm": 0.10059966146945953, "learning_rate": 0.00054086802580976, "loss": 0.8896, "step": 8290 }, { "epoch": 0.857792476229847, "grad_norm": 0.09923160821199417, "learning_rate": 0.0005323550875654025, "loss": 0.9073, "step": 8300 }, { "epoch": 0.8588259611409673, "grad_norm": 0.09592054039239883, "learning_rate": 0.000523976137861861, "loss": 0.896, "step": 8310 }, { "epoch": 0.8598594460520876, "grad_norm": 0.09789422899484634, "learning_rate": 0.0005157290677998864, "loss": 0.8878, "step": 8320 }, { "epoch": 0.860892930963208, "grad_norm": 0.09961704909801483, "learning_rate": 0.0005076118016730388, "loss": 0.9077, "step": 8330 }, { "epoch": 0.8619264158743283, "grad_norm": 0.10076375305652618, "learning_rate": 0.0004996222964452523, "loss": 0.8979, "step": 8340 }, { "epoch": 0.8629599007854485, "grad_norm": 0.0950293093919754, "learning_rate": 0.000491758541236623, "loss": 0.8966, "step": 8350 }, { "epoch": 0.8639933856965688, "grad_norm": 0.10502886027097702, "learning_rate": 0.0004840185568172905, "loss": 0.9094, "step": 8360 }, { "epoch": 0.8650268706076891, "grad_norm": 0.09783428907394409, "learning_rate": 0.0004764003951092846, "loss": 0.904, "step": 8370 }, { "epoch": 0.8660603555188094, "grad_norm": 0.10145466774702072, "learning_rate": 0.00046890213869621406, "loss": 0.8912, "step": 8380 }, { "epoch": 0.8670938404299298, "grad_norm": 0.09995938092470169, "learning_rate": 0.000461521900340671, "loss": 0.91, "step": 8390 }, { "epoch": 0.86812732534105, "grad_norm": 0.09882347285747528, "learning_rate": 0.0004542578225092324, "loss": 0.8935, "step": 8400 }, { "epoch": 0.8691608102521703, "grad_norm": 0.1028551384806633, "learning_rate": 0.00044710807690493667, "loss": 0.9005, "step": 8410 }, { "epoch": 0.8701942951632906, "grad_norm": 0.09993860870599747, "learning_rate": 0.00044007086400711984, "loss": 0.8883, "step": 8420 }, { "epoch": 0.8712277800744109, "grad_norm": 0.09587843716144562, "learning_rate": 0.0004331444126184934, "loss": 0.896, "step": 8430 }, { "epoch": 0.8722612649855312, "grad_norm": 0.10601107031106949, "learning_rate": 0.0004263269794193516, "loss": 0.9027, "step": 8440 }, { "epoch": 0.8732947498966515, "grad_norm": 0.09365765005350113, "learning_rate": 0.00041961684852879507, "loss": 0.8928, "step": 8450 }, { "epoch": 0.8743282348077718, "grad_norm": 0.10556492209434509, "learning_rate": 0.0004130123310728602, "loss": 0.8872, "step": 8460 }, { "epoch": 0.8753617197188921, "grad_norm": 0.09618480503559113, "learning_rate": 0.0004065117647594465, "loss": 0.8917, "step": 8470 }, { "epoch": 0.8763952046300124, "grad_norm": 0.100033700466156, "learning_rate": 0.0004001135134599342, "loss": 0.8949, "step": 8480 }, { "epoch": 0.8774286895411327, "grad_norm": 0.09465377777814865, "learning_rate": 0.00039381596679738596, "loss": 0.8931, "step": 8490 }, { "epoch": 0.8784621744522529, "grad_norm": 0.09475496411323547, "learning_rate": 0.00038761753974123146, "loss": 0.8856, "step": 8500 }, { "epoch": 0.8794956593633733, "grad_norm": 0.09431330114603043, "learning_rate": 0.0003815166722083307, "loss": 0.8921, "step": 8510 }, { "epoch": 0.8805291442744936, "grad_norm": 0.11336002498865128, "learning_rate": 0.00037551182867031637, "loss": 0.8944, "step": 8520 }, { "epoch": 0.8815626291856139, "grad_norm": 0.09444057941436768, "learning_rate": 0.00036960149776711637, "loss": 0.8907, "step": 8530 }, { "epoch": 0.8825961140967342, "grad_norm": 0.09592239558696747, "learning_rate": 0.0003637841919265597, "loss": 0.8993, "step": 8540 }, { "epoch": 0.8836295990078545, "grad_norm": 0.09695941209793091, "learning_rate": 0.0003580584469899685, "loss": 0.8948, "step": 8550 }, { "epoch": 0.8846630839189747, "grad_norm": 0.10035042464733124, "learning_rate": 0.0003524228218436444, "loss": 0.8897, "step": 8560 }, { "epoch": 0.8856965688300951, "grad_norm": 0.11594710499048233, "learning_rate": 0.0003468758980561539, "loss": 0.896, "step": 8570 }, { "epoch": 0.8867300537412154, "grad_norm": 0.10346604883670807, "learning_rate": 0.00034141627952132355, "loss": 0.8885, "step": 8580 }, { "epoch": 0.8877635386523357, "grad_norm": 0.09101969003677368, "learning_rate": 0.0003360425921068532, "loss": 0.8867, "step": 8590 }, { "epoch": 0.888797023563456, "grad_norm": 0.09395504742860794, "learning_rate": 0.00033075348330846094, "loss": 0.8871, "step": 8600 }, { "epoch": 0.8898305084745762, "grad_norm": 0.09610182791948318, "learning_rate": 0.00032554762190947073, "loss": 0.8913, "step": 8610 }, { "epoch": 0.8908639933856966, "grad_norm": 0.11420192569494247, "learning_rate": 0.0003204236976457585, "loss": 0.8897, "step": 8620 }, { "epoch": 0.8918974782968169, "grad_norm": 0.10842256993055344, "learning_rate": 0.00031538042087597126, "loss": 0.8887, "step": 8630 }, { "epoch": 0.8929309632079372, "grad_norm": 0.10306430608034134, "learning_rate": 0.00031041652225693745, "loss": 0.8882, "step": 8640 }, { "epoch": 0.8939644481190575, "grad_norm": 0.09563726931810379, "learning_rate": 0.00030553075242418533, "loss": 0.8884, "step": 8650 }, { "epoch": 0.8949979330301777, "grad_norm": 0.09344979375600815, "learning_rate": 0.0003007218816774904, "loss": 0.8853, "step": 8660 }, { "epoch": 0.896031417941298, "grad_norm": 0.09084004163742065, "learning_rate": 0.0002959886996713722, "loss": 0.8844, "step": 8670 }, { "epoch": 0.8970649028524184, "grad_norm": 0.0981561467051506, "learning_rate": 0.0002913300151104616, "loss": 0.8987, "step": 8680 }, { "epoch": 0.8980983877635387, "grad_norm": 0.09380181133747101, "learning_rate": 0.00028674465544966445, "loss": 0.8881, "step": 8690 }, { "epoch": 0.899131872674659, "grad_norm": 0.10047593712806702, "learning_rate": 0.00028223146659904236, "loss": 0.8949, "step": 8700 }, { "epoch": 0.9001653575857792, "grad_norm": 0.10739541053771973, "learning_rate": 0.0002777893126333406, "loss": 0.8969, "step": 8710 }, { "epoch": 0.9011988424968995, "grad_norm": 0.09939514100551605, "learning_rate": 0.0002734170755060864, "loss": 0.8808, "step": 8720 }, { "epoch": 0.9022323274080198, "grad_norm": 0.09798012673854828, "learning_rate": 0.0002691136547681875, "loss": 0.8898, "step": 8730 }, { "epoch": 0.9032658123191402, "grad_norm": 0.09667760878801346, "learning_rate": 0.00026487796729096044, "loss": 0.8996, "step": 8740 }, { "epoch": 0.9042992972302605, "grad_norm": 0.09401620179414749, "learning_rate": 0.0002607089469935173, "loss": 0.8797, "step": 8750 }, { "epoch": 0.9053327821413807, "grad_norm": 0.09637662023305893, "learning_rate": 0.0002566055445744437, "loss": 0.885, "step": 8760 }, { "epoch": 0.906366267052501, "grad_norm": 0.0979810431599617, "learning_rate": 0.00025256672724770034, "loss": 0.8765, "step": 8770 }, { "epoch": 0.9073997519636213, "grad_norm": 0.09577486664056778, "learning_rate": 0.0002485914784826802, "loss": 0.878, "step": 8780 }, { "epoch": 0.9084332368747416, "grad_norm": 0.09690549224615097, "learning_rate": 0.0002446787977483582, "loss": 0.8951, "step": 8790 }, { "epoch": 0.909466721785862, "grad_norm": 0.09132607281208038, "learning_rate": 0.00024082770026146767, "loss": 0.8922, "step": 8800 }, { "epoch": 0.9105002066969822, "grad_norm": 0.10059286653995514, "learning_rate": 0.00023703721673863946, "loss": 0.8961, "step": 8810 }, { "epoch": 0.9115336916081025, "grad_norm": 0.09476503729820251, "learning_rate": 0.00023330639315244326, "loss": 0.8891, "step": 8820 }, { "epoch": 0.9125671765192228, "grad_norm": 0.097599096596241, "learning_rate": 0.00022963429049126816, "loss": 0.8877, "step": 8830 }, { "epoch": 0.9136006614303431, "grad_norm": 0.09217467904090881, "learning_rate": 0.00022601998452298264, "loss": 0.893, "step": 8840 }, { "epoch": 0.9146341463414634, "grad_norm": 0.11072806268930435, "learning_rate": 0.00022246256556231452, "loss": 0.8943, "step": 8850 }, { "epoch": 0.9156676312525837, "grad_norm": 0.10245074331760406, "learning_rate": 0.00021896113824189194, "loss": 0.9008, "step": 8860 }, { "epoch": 0.916701116163704, "grad_norm": 0.0995858684182167, "learning_rate": 0.00021551482128688842, "loss": 0.8831, "step": 8870 }, { "epoch": 0.9177346010748243, "grad_norm": 0.09560049325227737, "learning_rate": 0.0002121227472932145, "loss": 0.8978, "step": 8880 }, { "epoch": 0.9187680859859446, "grad_norm": 0.09800171852111816, "learning_rate": 0.00020878406250920078, "loss": 0.8993, "step": 8890 }, { "epoch": 0.9198015708970649, "grad_norm": 0.09587955474853516, "learning_rate": 0.00020549792662071688, "loss": 0.8947, "step": 8900 }, { "epoch": 0.9208350558081853, "grad_norm": 0.09761986881494522, "learning_rate": 0.00020226351253967265, "loss": 0.8968, "step": 8910 }, { "epoch": 0.9218685407193055, "grad_norm": 0.10232323408126831, "learning_rate": 0.00019908000619584844, "loss": 0.8743, "step": 8920 }, { "epoch": 0.9229020256304258, "grad_norm": 0.09598308801651001, "learning_rate": 0.00019594660633200127, "loss": 0.8886, "step": 8930 }, { "epoch": 0.9239355105415461, "grad_norm": 0.09552961587905884, "learning_rate": 0.0001928625243021967, "loss": 0.898, "step": 8940 }, { "epoch": 0.9249689954526664, "grad_norm": 0.09763569384813309, "learning_rate": 0.000189826983873314, "loss": 0.8872, "step": 8950 }, { "epoch": 0.9260024803637867, "grad_norm": 0.10132598876953125, "learning_rate": 0.00018683922102967604, "loss": 0.8844, "step": 8960 }, { "epoch": 0.927035965274907, "grad_norm": 0.10523135215044022, "learning_rate": 0.00018389848378075436, "loss": 0.8886, "step": 8970 }, { "epoch": 0.9280694501860273, "grad_norm": 0.10074016451835632, "learning_rate": 0.0001810040319719, "loss": 0.8868, "step": 8980 }, { "epoch": 0.9291029350971476, "grad_norm": 0.09313800185918808, "learning_rate": 0.00017815513709805425, "loss": 0.8944, "step": 8990 }, { "epoch": 0.9301364200082679, "grad_norm": 0.09687028080224991, "learning_rate": 0.00017535108212039094, "loss": 0.8764, "step": 9000 }, { "epoch": 0.9311699049193882, "grad_norm": 0.09959317743778229, "learning_rate": 0.00017259116128584488, "loss": 0.8879, "step": 9010 }, { "epoch": 0.9322033898305084, "grad_norm": 0.09753499180078506, "learning_rate": 0.0001698746799494807, "loss": 0.8878, "step": 9020 }, { "epoch": 0.9332368747416288, "grad_norm": 0.10160617530345917, "learning_rate": 0.00016720095439965764, "loss": 0.8878, "step": 9030 }, { "epoch": 0.9342703596527491, "grad_norm": 0.09427917003631592, "learning_rate": 0.00016456931168594587, "loss": 0.8917, "step": 9040 }, { "epoch": 0.9353038445638694, "grad_norm": 0.09126557409763336, "learning_rate": 0.0001619790894497517, "loss": 0.8756, "step": 9050 }, { "epoch": 0.9363373294749897, "grad_norm": 0.10263457894325256, "learning_rate": 0.0001594296357576083, "loss": 0.8968, "step": 9060 }, { "epoch": 0.9373708143861099, "grad_norm": 0.09004776179790497, "learning_rate": 0.00015692030893709055, "loss": 0.8747, "step": 9070 }, { "epoch": 0.9384042992972302, "grad_norm": 0.09706980735063553, "learning_rate": 0.0001544504774153122, "loss": 0.8782, "step": 9080 }, { "epoch": 0.9394377842083506, "grad_norm": 0.11443266272544861, "learning_rate": 0.00015201951955996554, "loss": 0.8954, "step": 9090 }, { "epoch": 0.9404712691194709, "grad_norm": 0.09983173757791519, "learning_rate": 0.00014962682352286227, "loss": 0.8839, "step": 9100 }, { "epoch": 0.9415047540305912, "grad_norm": 0.09485773742198944, "learning_rate": 0.00014727178708593758, "loss": 0.8873, "step": 9110 }, { "epoch": 0.9425382389417114, "grad_norm": 0.09533528983592987, "learning_rate": 0.00014495381750967773, "loss": 0.8803, "step": 9120 }, { "epoch": 0.9435717238528317, "grad_norm": 0.09723920375108719, "learning_rate": 0.00014267233138393328, "loss": 0.894, "step": 9130 }, { "epoch": 0.944605208763952, "grad_norm": 0.10340782254934311, "learning_rate": 0.00014042675448108058, "loss": 0.8876, "step": 9140 }, { "epoch": 0.9456386936750724, "grad_norm": 0.10088169574737549, "learning_rate": 0.00013821652161149427, "loss": 0.8797, "step": 9150 }, { "epoch": 0.9466721785861927, "grad_norm": 0.09791452437639236, "learning_rate": 0.0001360410764812946, "loss": 0.8897, "step": 9160 }, { "epoch": 0.9477056634973129, "grad_norm": 0.0989762470126152, "learning_rate": 0.00013389987155233373, "loss": 0.885, "step": 9170 }, { "epoch": 0.9487391484084332, "grad_norm": 0.09686373174190521, "learning_rate": 0.0001317923679043859, "loss": 0.88, "step": 9180 }, { "epoch": 0.9497726333195535, "grad_norm": 0.094807930290699, "learning_rate": 0.00012971803509950622, "loss": 0.8877, "step": 9190 }, { "epoch": 0.9508061182306738, "grad_norm": 0.0951213538646698, "learning_rate": 0.000127676351048525, "loss": 0.8858, "step": 9200 }, { "epoch": 0.9518396031417942, "grad_norm": 0.09922289848327637, "learning_rate": 0.00012566680187964274, "loss": 0.8826, "step": 9210 }, { "epoch": 0.9528730880529144, "grad_norm": 0.09398606419563293, "learning_rate": 0.0001236888818090938, "loss": 0.8817, "step": 9220 }, { "epoch": 0.9539065729640347, "grad_norm": 0.1025305688381195, "learning_rate": 0.0001217420930138456, "loss": 0.8809, "step": 9230 }, { "epoch": 0.954940057875155, "grad_norm": 0.0998004898428917, "learning_rate": 0.00011982594550630144, "loss": 0.8878, "step": 9240 }, { "epoch": 0.9559735427862753, "grad_norm": 0.09735240042209625, "learning_rate": 0.00011793995701097545, "loss": 0.8713, "step": 9250 }, { "epoch": 0.9570070276973957, "grad_norm": 0.09462594240903854, "learning_rate": 0.00011608365284310853, "loss": 0.8932, "step": 9260 }, { "epoch": 0.958040512608516, "grad_norm": 0.0939006358385086, "learning_rate": 0.00011425656578919491, "loss": 0.8865, "step": 9270 }, { "epoch": 0.9590739975196362, "grad_norm": 0.10327833890914917, "learning_rate": 0.00011245823598938918, "loss": 0.885, "step": 9280 }, { "epoch": 0.9601074824307565, "grad_norm": 0.10130874067544937, "learning_rate": 0.00011068821082176394, "loss": 0.8737, "step": 9290 }, { "epoch": 0.9611409673418768, "grad_norm": 0.0958571583032608, "learning_rate": 0.00010894604478838941, "loss": 0.8836, "step": 9300 }, { "epoch": 0.9621744522529971, "grad_norm": 0.10654882341623306, "learning_rate": 0.00010723129940320598, "loss": 0.8808, "step": 9310 }, { "epoch": 0.9632079371641175, "grad_norm": 0.10343246906995773, "learning_rate": 0.00010554354308166149, "loss": 0.8838, "step": 9320 }, { "epoch": 0.9642414220752377, "grad_norm": 0.09667105227708817, "learning_rate": 0.00010388235103208581, "loss": 0.8944, "step": 9330 }, { "epoch": 0.965274906986358, "grad_norm": 0.09733343869447708, "learning_rate": 0.00010224730514877476, "loss": 0.8807, "step": 9340 }, { "epoch": 0.9663083918974783, "grad_norm": 0.09809593111276627, "learning_rate": 0.0001006379939067572, "loss": 0.8824, "step": 9350 }, { "epoch": 0.9673418768085986, "grad_norm": 0.09950044006109238, "learning_rate": 9.905401225821806e-05, "loss": 0.8895, "step": 9360 }, { "epoch": 0.9683753617197189, "grad_norm": 0.09585009515285492, "learning_rate": 9.749496153055196e-05, "loss": 0.871, "step": 9370 }, { "epoch": 0.9694088466308391, "grad_norm": 0.09470534324645996, "learning_rate": 9.596044932602107e-05, "loss": 0.8805, "step": 9380 }, { "epoch": 0.9704423315419595, "grad_norm": 0.09828654676675797, "learning_rate": 9.445008942299262e-05, "loss": 0.8796, "step": 9390 }, { "epoch": 0.9714758164530798, "grad_norm": 0.09540653228759766, "learning_rate": 9.296350167873062e-05, "loss": 0.8769, "step": 9400 }, { "epoch": 0.9725093013642001, "grad_norm": 0.09175027906894684, "learning_rate": 9.150031193371763e-05, "loss": 0.8823, "step": 9410 }, { "epoch": 0.9735427862753204, "grad_norm": 0.09416258335113525, "learning_rate": 9.00601519174826e-05, "loss": 0.8826, "step": 9420 }, { "epoch": 0.9745762711864406, "grad_norm": 0.102226123213768, "learning_rate": 8.864265915591077e-05, "loss": 0.8813, "step": 9430 }, { "epoch": 0.975609756097561, "grad_norm": 0.1030995100736618, "learning_rate": 8.724747688001238e-05, "loss": 0.89, "step": 9440 }, { "epoch": 0.9766432410086813, "grad_norm": 0.09412838518619537, "learning_rate": 8.587425393612774e-05, "loss": 0.8869, "step": 9450 }, { "epoch": 0.9776767259198016, "grad_norm": 0.09783417731523514, "learning_rate": 8.45226446975449e-05, "loss": 0.8745, "step": 9460 }, { "epoch": 0.9787102108309219, "grad_norm": 0.09457999467849731, "learning_rate": 8.31923089775091e-05, "loss": 0.8854, "step": 9470 }, { "epoch": 0.9797436957420421, "grad_norm": 0.0977085754275322, "learning_rate": 8.188291194360097e-05, "loss": 0.8743, "step": 9480 }, { "epoch": 0.9807771806531624, "grad_norm": 0.09691467881202698, "learning_rate": 8.059412403346259e-05, "loss": 0.8798, "step": 9490 }, { "epoch": 0.9818106655642828, "grad_norm": 0.09877024590969086, "learning_rate": 7.932562087184977e-05, "loss": 0.8785, "step": 9500 }, { "epoch": 0.9828441504754031, "grad_norm": 0.09752603620290756, "learning_rate": 7.807708318899015e-05, "loss": 0.8787, "step": 9510 }, { "epoch": 0.9838776353865234, "grad_norm": 0.10295706242322922, "learning_rate": 7.684819674022599e-05, "loss": 0.8782, "step": 9520 }, { "epoch": 0.9849111202976436, "grad_norm": 0.09535824507474899, "learning_rate": 7.563865222692196e-05, "loss": 0.8861, "step": 9530 }, { "epoch": 0.9859446052087639, "grad_norm": 0.09367913007736206, "learning_rate": 7.444814521861768e-05, "loss": 0.8791, "step": 9540 }, { "epoch": 0.9869780901198842, "grad_norm": 0.10222353786230087, "learning_rate": 7.327637607640558e-05, "loss": 0.884, "step": 9550 }, { "epoch": 0.9880115750310046, "grad_norm": 0.1052263155579567, "learning_rate": 7.212304987751476e-05, "loss": 0.8766, "step": 9560 }, { "epoch": 0.9890450599421249, "grad_norm": 0.09064412862062454, "learning_rate": 7.098787634108176e-05, "loss": 0.8846, "step": 9570 }, { "epoch": 0.9900785448532451, "grad_norm": 0.09217038005590439, "learning_rate": 6.98705697550898e-05, "loss": 0.878, "step": 9580 }, { "epoch": 0.9911120297643654, "grad_norm": 0.08925709873437881, "learning_rate": 6.87708489044578e-05, "loss": 0.8659, "step": 9590 }, { "epoch": 0.9921455146754857, "grad_norm": 0.09898435324430466, "learning_rate": 6.768843700026141e-05, "loss": 0.8891, "step": 9600 }, { "epoch": 0.993178999586606, "grad_norm": 0.09614672511816025, "learning_rate": 6.662306161006783e-05, "loss": 0.8828, "step": 9610 }, { "epoch": 0.9942124844977264, "grad_norm": 0.09297246485948563, "learning_rate": 6.557445458936731e-05, "loss": 0.8905, "step": 9620 }, { "epoch": 0.9952459694088466, "grad_norm": 0.10573872923851013, "learning_rate": 6.45423520140839e-05, "loss": 0.8876, "step": 9630 }, { "epoch": 0.9962794543199669, "grad_norm": 0.09352509677410126, "learning_rate": 6.352649411414819e-05, "loss": 0.8856, "step": 9640 }, { "epoch": 0.9973129392310872, "grad_norm": 0.0931333377957344, "learning_rate": 6.252662520811583e-05, "loss": 0.8848, "step": 9650 }, { "epoch": 0.9983464241422075, "grad_norm": 0.10176722705364227, "learning_rate": 6.154249363881505e-05, "loss": 0.8792, "step": 9660 }, { "epoch": 0.9993799090533279, "grad_norm": 0.09999672323465347, "learning_rate": 6.057385171000694e-05, "loss": 0.8789, "step": 9670 } ], "logging_steps": 10, "max_steps": 9676, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9.24620217896713e+16, "train_batch_size": 1024, "trial_name": null, "trial_params": null }