llama3-8b-closedqa-gpt4o-100k / trainer_state.json
chansung's picture
Model save
edd93d9 verified
raw
history blame
206 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.991416309012875,
"eval_steps": 500,
"global_step": 5820,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0017167381974248926,
"grad_norm": 6.089517593383789,
"learning_rate": 3.436426116838488e-07,
"loss": 2.0708,
"step": 1
},
{
"epoch": 0.008583690987124463,
"grad_norm": 6.525425434112549,
"learning_rate": 1.7182130584192443e-06,
"loss": 2.0164,
"step": 5
},
{
"epoch": 0.017167381974248927,
"grad_norm": 5.859015464782715,
"learning_rate": 3.4364261168384886e-06,
"loss": 2.0159,
"step": 10
},
{
"epoch": 0.02575107296137339,
"grad_norm": 4.282891750335693,
"learning_rate": 5.154639175257732e-06,
"loss": 1.9357,
"step": 15
},
{
"epoch": 0.034334763948497854,
"grad_norm": 2.575316905975342,
"learning_rate": 6.872852233676977e-06,
"loss": 1.8673,
"step": 20
},
{
"epoch": 0.04291845493562232,
"grad_norm": 1.4356714487075806,
"learning_rate": 8.591065292096221e-06,
"loss": 1.7557,
"step": 25
},
{
"epoch": 0.05150214592274678,
"grad_norm": 0.756867527961731,
"learning_rate": 1.0309278350515464e-05,
"loss": 1.6729,
"step": 30
},
{
"epoch": 0.060085836909871244,
"grad_norm": 0.5709408521652222,
"learning_rate": 1.2027491408934708e-05,
"loss": 1.6029,
"step": 35
},
{
"epoch": 0.06866952789699571,
"grad_norm": 0.6336522698402405,
"learning_rate": 1.3745704467353954e-05,
"loss": 1.5371,
"step": 40
},
{
"epoch": 0.07725321888412018,
"grad_norm": 0.48585963249206543,
"learning_rate": 1.5463917525773197e-05,
"loss": 1.4468,
"step": 45
},
{
"epoch": 0.08583690987124463,
"grad_norm": 0.45667070150375366,
"learning_rate": 1.7182130584192442e-05,
"loss": 1.3687,
"step": 50
},
{
"epoch": 0.0944206008583691,
"grad_norm": 0.4194663465023041,
"learning_rate": 1.8900343642611683e-05,
"loss": 1.3212,
"step": 55
},
{
"epoch": 0.10300429184549356,
"grad_norm": 0.3849167227745056,
"learning_rate": 2.0618556701030927e-05,
"loss": 1.2499,
"step": 60
},
{
"epoch": 0.11158798283261803,
"grad_norm": 0.32962682843208313,
"learning_rate": 2.2336769759450175e-05,
"loss": 1.2345,
"step": 65
},
{
"epoch": 0.12017167381974249,
"grad_norm": 0.32320886850357056,
"learning_rate": 2.4054982817869417e-05,
"loss": 1.1963,
"step": 70
},
{
"epoch": 0.12875536480686695,
"grad_norm": 0.3570818603038788,
"learning_rate": 2.5773195876288658e-05,
"loss": 1.1658,
"step": 75
},
{
"epoch": 0.13733905579399142,
"grad_norm": 0.3045337498188019,
"learning_rate": 2.749140893470791e-05,
"loss": 1.1628,
"step": 80
},
{
"epoch": 0.1459227467811159,
"grad_norm": 0.2675187289714813,
"learning_rate": 2.920962199312715e-05,
"loss": 1.153,
"step": 85
},
{
"epoch": 0.15450643776824036,
"grad_norm": 0.2941209077835083,
"learning_rate": 3.0927835051546395e-05,
"loss": 1.1233,
"step": 90
},
{
"epoch": 0.1630901287553648,
"grad_norm": 0.30070310831069946,
"learning_rate": 3.2646048109965636e-05,
"loss": 1.1199,
"step": 95
},
{
"epoch": 0.17167381974248927,
"grad_norm": 0.2994473874568939,
"learning_rate": 3.4364261168384884e-05,
"loss": 1.1063,
"step": 100
},
{
"epoch": 0.18025751072961374,
"grad_norm": 0.301921010017395,
"learning_rate": 3.6082474226804125e-05,
"loss": 1.0991,
"step": 105
},
{
"epoch": 0.1888412017167382,
"grad_norm": 0.31418925523757935,
"learning_rate": 3.7800687285223366e-05,
"loss": 1.0939,
"step": 110
},
{
"epoch": 0.19742489270386265,
"grad_norm": 0.31536900997161865,
"learning_rate": 3.9518900343642614e-05,
"loss": 1.0961,
"step": 115
},
{
"epoch": 0.20600858369098712,
"grad_norm": 0.2873052954673767,
"learning_rate": 4.1237113402061855e-05,
"loss": 1.0807,
"step": 120
},
{
"epoch": 0.2145922746781116,
"grad_norm": 0.34555503726005554,
"learning_rate": 4.2955326460481096e-05,
"loss": 1.0645,
"step": 125
},
{
"epoch": 0.22317596566523606,
"grad_norm": 0.313761442899704,
"learning_rate": 4.467353951890035e-05,
"loss": 1.0513,
"step": 130
},
{
"epoch": 0.2317596566523605,
"grad_norm": 0.32121869921684265,
"learning_rate": 4.639175257731959e-05,
"loss": 1.0735,
"step": 135
},
{
"epoch": 0.24034334763948498,
"grad_norm": 0.3088555335998535,
"learning_rate": 4.810996563573883e-05,
"loss": 1.0552,
"step": 140
},
{
"epoch": 0.24892703862660945,
"grad_norm": 0.35492023825645447,
"learning_rate": 4.982817869415808e-05,
"loss": 1.0546,
"step": 145
},
{
"epoch": 0.2575107296137339,
"grad_norm": 0.30434370040893555,
"learning_rate": 5.1546391752577315e-05,
"loss": 1.0422,
"step": 150
},
{
"epoch": 0.26609442060085836,
"grad_norm": 0.30399689078330994,
"learning_rate": 5.326460481099656e-05,
"loss": 1.0465,
"step": 155
},
{
"epoch": 0.27467811158798283,
"grad_norm": 0.30653682351112366,
"learning_rate": 5.498281786941582e-05,
"loss": 1.0404,
"step": 160
},
{
"epoch": 0.2832618025751073,
"grad_norm": 0.3112322986125946,
"learning_rate": 5.670103092783505e-05,
"loss": 1.0369,
"step": 165
},
{
"epoch": 0.2918454935622318,
"grad_norm": 0.3165334165096283,
"learning_rate": 5.84192439862543e-05,
"loss": 1.0433,
"step": 170
},
{
"epoch": 0.30042918454935624,
"grad_norm": 0.34382325410842896,
"learning_rate": 6.013745704467354e-05,
"loss": 1.042,
"step": 175
},
{
"epoch": 0.3090128755364807,
"grad_norm": 0.3302488625049591,
"learning_rate": 6.185567010309279e-05,
"loss": 1.0166,
"step": 180
},
{
"epoch": 0.31759656652360513,
"grad_norm": 0.3078051209449768,
"learning_rate": 6.357388316151203e-05,
"loss": 1.0138,
"step": 185
},
{
"epoch": 0.3261802575107296,
"grad_norm": 0.30805814266204834,
"learning_rate": 6.529209621993127e-05,
"loss": 0.9951,
"step": 190
},
{
"epoch": 0.33476394849785407,
"grad_norm": 0.4036194980144501,
"learning_rate": 6.701030927835051e-05,
"loss": 1.013,
"step": 195
},
{
"epoch": 0.34334763948497854,
"grad_norm": 0.4557146430015564,
"learning_rate": 6.872852233676977e-05,
"loss": 0.9959,
"step": 200
},
{
"epoch": 0.351931330472103,
"grad_norm": 0.36137068271636963,
"learning_rate": 7.044673539518901e-05,
"loss": 1.0068,
"step": 205
},
{
"epoch": 0.3605150214592275,
"grad_norm": 0.3014100193977356,
"learning_rate": 7.216494845360825e-05,
"loss": 0.9923,
"step": 210
},
{
"epoch": 0.36909871244635195,
"grad_norm": 0.290464848279953,
"learning_rate": 7.38831615120275e-05,
"loss": 0.9969,
"step": 215
},
{
"epoch": 0.3776824034334764,
"grad_norm": 0.31709638237953186,
"learning_rate": 7.560137457044673e-05,
"loss": 1.0221,
"step": 220
},
{
"epoch": 0.38626609442060084,
"grad_norm": 0.2746679484844208,
"learning_rate": 7.731958762886599e-05,
"loss": 0.9829,
"step": 225
},
{
"epoch": 0.3948497854077253,
"grad_norm": 0.28260111808776855,
"learning_rate": 7.903780068728523e-05,
"loss": 0.9868,
"step": 230
},
{
"epoch": 0.4034334763948498,
"grad_norm": 0.3063802421092987,
"learning_rate": 8.075601374570447e-05,
"loss": 0.9789,
"step": 235
},
{
"epoch": 0.41201716738197425,
"grad_norm": 0.29451537132263184,
"learning_rate": 8.247422680412371e-05,
"loss": 1.0036,
"step": 240
},
{
"epoch": 0.4206008583690987,
"grad_norm": 0.29433488845825195,
"learning_rate": 8.419243986254296e-05,
"loss": 0.9936,
"step": 245
},
{
"epoch": 0.4291845493562232,
"grad_norm": 0.2725401520729065,
"learning_rate": 8.591065292096219e-05,
"loss": 0.9786,
"step": 250
},
{
"epoch": 0.43776824034334766,
"grad_norm": 0.2759503722190857,
"learning_rate": 8.762886597938145e-05,
"loss": 0.9656,
"step": 255
},
{
"epoch": 0.44635193133047213,
"grad_norm": 0.2762455642223358,
"learning_rate": 8.93470790378007e-05,
"loss": 0.9837,
"step": 260
},
{
"epoch": 0.45493562231759654,
"grad_norm": 0.2803398668766022,
"learning_rate": 9.106529209621993e-05,
"loss": 0.9757,
"step": 265
},
{
"epoch": 0.463519313304721,
"grad_norm": 0.29107633233070374,
"learning_rate": 9.278350515463918e-05,
"loss": 0.9692,
"step": 270
},
{
"epoch": 0.4721030042918455,
"grad_norm": 0.32915282249450684,
"learning_rate": 9.450171821305843e-05,
"loss": 0.9911,
"step": 275
},
{
"epoch": 0.48068669527896996,
"grad_norm": 0.2902511954307556,
"learning_rate": 9.621993127147767e-05,
"loss": 0.9635,
"step": 280
},
{
"epoch": 0.4892703862660944,
"grad_norm": 0.28177133202552795,
"learning_rate": 9.793814432989691e-05,
"loss": 0.9704,
"step": 285
},
{
"epoch": 0.4978540772532189,
"grad_norm": 0.26041337847709656,
"learning_rate": 9.965635738831616e-05,
"loss": 0.9668,
"step": 290
},
{
"epoch": 0.5064377682403434,
"grad_norm": 0.27237018942832947,
"learning_rate": 0.00010137457044673539,
"loss": 0.9751,
"step": 295
},
{
"epoch": 0.5150214592274678,
"grad_norm": 0.2928680181503296,
"learning_rate": 0.00010309278350515463,
"loss": 0.9828,
"step": 300
},
{
"epoch": 0.5236051502145923,
"grad_norm": 0.3540053367614746,
"learning_rate": 0.0001048109965635739,
"loss": 0.9725,
"step": 305
},
{
"epoch": 0.5321888412017167,
"grad_norm": 0.34892937541007996,
"learning_rate": 0.00010652920962199313,
"loss": 0.9688,
"step": 310
},
{
"epoch": 0.5407725321888412,
"grad_norm": 0.2727091908454895,
"learning_rate": 0.00010824742268041237,
"loss": 0.9797,
"step": 315
},
{
"epoch": 0.5493562231759657,
"grad_norm": 0.2613857090473175,
"learning_rate": 0.00010996563573883164,
"loss": 0.9751,
"step": 320
},
{
"epoch": 0.5579399141630901,
"grad_norm": 0.24695193767547607,
"learning_rate": 0.00011168384879725086,
"loss": 0.9661,
"step": 325
},
{
"epoch": 0.5665236051502146,
"grad_norm": 0.25507402420043945,
"learning_rate": 0.0001134020618556701,
"loss": 0.9703,
"step": 330
},
{
"epoch": 0.575107296137339,
"grad_norm": 0.2496771663427353,
"learning_rate": 0.00011512027491408935,
"loss": 0.9766,
"step": 335
},
{
"epoch": 0.5836909871244635,
"grad_norm": 0.2503701448440552,
"learning_rate": 0.0001168384879725086,
"loss": 0.9506,
"step": 340
},
{
"epoch": 0.592274678111588,
"grad_norm": 0.26991888880729675,
"learning_rate": 0.00011855670103092784,
"loss": 0.9587,
"step": 345
},
{
"epoch": 0.6008583690987125,
"grad_norm": 0.24292829632759094,
"learning_rate": 0.00012027491408934708,
"loss": 0.951,
"step": 350
},
{
"epoch": 0.6094420600858369,
"grad_norm": 0.26761141419410706,
"learning_rate": 0.00012199312714776634,
"loss": 0.9525,
"step": 355
},
{
"epoch": 0.6180257510729614,
"grad_norm": 0.24770408868789673,
"learning_rate": 0.00012371134020618558,
"loss": 0.9606,
"step": 360
},
{
"epoch": 0.6266094420600858,
"grad_norm": 0.24938061833381653,
"learning_rate": 0.00012542955326460482,
"loss": 0.9577,
"step": 365
},
{
"epoch": 0.6351931330472103,
"grad_norm": 0.22758124768733978,
"learning_rate": 0.00012714776632302406,
"loss": 0.9693,
"step": 370
},
{
"epoch": 0.6437768240343348,
"grad_norm": 0.24254348874092102,
"learning_rate": 0.0001288659793814433,
"loss": 0.9492,
"step": 375
},
{
"epoch": 0.6523605150214592,
"grad_norm": 0.37186160683631897,
"learning_rate": 0.00013058419243986254,
"loss": 0.9722,
"step": 380
},
{
"epoch": 0.6609442060085837,
"grad_norm": 0.23567767441272736,
"learning_rate": 0.00013230240549828178,
"loss": 0.9567,
"step": 385
},
{
"epoch": 0.6695278969957081,
"grad_norm": 0.2288963794708252,
"learning_rate": 0.00013402061855670103,
"loss": 0.9564,
"step": 390
},
{
"epoch": 0.6781115879828327,
"grad_norm": 0.24152550101280212,
"learning_rate": 0.0001357388316151203,
"loss": 0.9532,
"step": 395
},
{
"epoch": 0.6866952789699571,
"grad_norm": 0.261593759059906,
"learning_rate": 0.00013745704467353953,
"loss": 0.9471,
"step": 400
},
{
"epoch": 0.6952789699570815,
"grad_norm": 0.27105116844177246,
"learning_rate": 0.00013917525773195878,
"loss": 0.9639,
"step": 405
},
{
"epoch": 0.703862660944206,
"grad_norm": 0.361182302236557,
"learning_rate": 0.00014089347079037802,
"loss": 0.9516,
"step": 410
},
{
"epoch": 0.7124463519313304,
"grad_norm": 0.2614869475364685,
"learning_rate": 0.00014261168384879726,
"loss": 0.9615,
"step": 415
},
{
"epoch": 0.721030042918455,
"grad_norm": 0.2598520517349243,
"learning_rate": 0.0001443298969072165,
"loss": 0.9587,
"step": 420
},
{
"epoch": 0.7296137339055794,
"grad_norm": 0.26025694608688354,
"learning_rate": 0.00014604810996563574,
"loss": 0.9483,
"step": 425
},
{
"epoch": 0.7381974248927039,
"grad_norm": 0.27435532212257385,
"learning_rate": 0.000147766323024055,
"loss": 0.9555,
"step": 430
},
{
"epoch": 0.7467811158798283,
"grad_norm": 0.2179042547941208,
"learning_rate": 0.00014948453608247422,
"loss": 0.9719,
"step": 435
},
{
"epoch": 0.7553648068669528,
"grad_norm": 0.25120726227760315,
"learning_rate": 0.00015120274914089346,
"loss": 0.9601,
"step": 440
},
{
"epoch": 0.7639484978540773,
"grad_norm": 0.21970221400260925,
"learning_rate": 0.0001529209621993127,
"loss": 0.9545,
"step": 445
},
{
"epoch": 0.7725321888412017,
"grad_norm": 0.2047254890203476,
"learning_rate": 0.00015463917525773197,
"loss": 0.9448,
"step": 450
},
{
"epoch": 0.7811158798283262,
"grad_norm": 0.21874375641345978,
"learning_rate": 0.00015635738831615121,
"loss": 0.9426,
"step": 455
},
{
"epoch": 0.7896995708154506,
"grad_norm": 0.21705736219882965,
"learning_rate": 0.00015807560137457046,
"loss": 0.941,
"step": 460
},
{
"epoch": 0.7982832618025751,
"grad_norm": 0.20229893922805786,
"learning_rate": 0.0001597938144329897,
"loss": 0.9495,
"step": 465
},
{
"epoch": 0.8068669527896996,
"grad_norm": 0.21079690754413605,
"learning_rate": 0.00016151202749140894,
"loss": 0.9488,
"step": 470
},
{
"epoch": 0.8154506437768241,
"grad_norm": 0.21350346505641937,
"learning_rate": 0.00016323024054982818,
"loss": 0.941,
"step": 475
},
{
"epoch": 0.8240343347639485,
"grad_norm": 0.22696025669574738,
"learning_rate": 0.00016494845360824742,
"loss": 0.9468,
"step": 480
},
{
"epoch": 0.8326180257510729,
"grad_norm": 0.2032315880060196,
"learning_rate": 0.0001666666666666667,
"loss": 0.9649,
"step": 485
},
{
"epoch": 0.8412017167381974,
"grad_norm": 0.24577978253364563,
"learning_rate": 0.00016838487972508593,
"loss": 0.9457,
"step": 490
},
{
"epoch": 0.8497854077253219,
"grad_norm": 0.2154797613620758,
"learning_rate": 0.00017010309278350517,
"loss": 0.9322,
"step": 495
},
{
"epoch": 0.8583690987124464,
"grad_norm": 0.20580855011940002,
"learning_rate": 0.00017182130584192438,
"loss": 0.9417,
"step": 500
},
{
"epoch": 0.8669527896995708,
"grad_norm": 0.2095131129026413,
"learning_rate": 0.00017353951890034365,
"loss": 0.9447,
"step": 505
},
{
"epoch": 0.8755364806866953,
"grad_norm": 0.20167525112628937,
"learning_rate": 0.0001752577319587629,
"loss": 0.9504,
"step": 510
},
{
"epoch": 0.8841201716738197,
"grad_norm": 0.21665619313716888,
"learning_rate": 0.00017697594501718214,
"loss": 0.9386,
"step": 515
},
{
"epoch": 0.8927038626609443,
"grad_norm": 0.2125951647758484,
"learning_rate": 0.0001786941580756014,
"loss": 0.9373,
"step": 520
},
{
"epoch": 0.9012875536480687,
"grad_norm": 0.19751432538032532,
"learning_rate": 0.00018041237113402062,
"loss": 0.935,
"step": 525
},
{
"epoch": 0.9098712446351931,
"grad_norm": 0.20792262256145477,
"learning_rate": 0.00018213058419243986,
"loss": 0.9624,
"step": 530
},
{
"epoch": 0.9184549356223176,
"grad_norm": 0.2099096179008484,
"learning_rate": 0.0001838487972508591,
"loss": 0.9392,
"step": 535
},
{
"epoch": 0.927038626609442,
"grad_norm": 0.21597731113433838,
"learning_rate": 0.00018556701030927837,
"loss": 0.942,
"step": 540
},
{
"epoch": 0.9356223175965666,
"grad_norm": 0.2026844620704651,
"learning_rate": 0.0001872852233676976,
"loss": 0.9579,
"step": 545
},
{
"epoch": 0.944206008583691,
"grad_norm": 0.20321713387966156,
"learning_rate": 0.00018900343642611685,
"loss": 0.9519,
"step": 550
},
{
"epoch": 0.9527896995708155,
"grad_norm": 0.20575563609600067,
"learning_rate": 0.0001907216494845361,
"loss": 0.9363,
"step": 555
},
{
"epoch": 0.9613733905579399,
"grad_norm": 0.21118000149726868,
"learning_rate": 0.00019243986254295533,
"loss": 0.9493,
"step": 560
},
{
"epoch": 0.9699570815450643,
"grad_norm": 0.2216077297925949,
"learning_rate": 0.00019415807560137457,
"loss": 0.9499,
"step": 565
},
{
"epoch": 0.9785407725321889,
"grad_norm": 0.20766399800777435,
"learning_rate": 0.00019587628865979381,
"loss": 0.9359,
"step": 570
},
{
"epoch": 0.9871244635193133,
"grad_norm": 0.30319133400917053,
"learning_rate": 0.00019759450171821308,
"loss": 0.9483,
"step": 575
},
{
"epoch": 0.9957081545064378,
"grad_norm": 0.21939022839069366,
"learning_rate": 0.00019931271477663232,
"loss": 0.9422,
"step": 580
},
{
"epoch": 0.9991416309012876,
"eval_loss": 1.976241946220398,
"eval_runtime": 0.3946,
"eval_samples_per_second": 15.206,
"eval_steps_per_second": 2.534,
"step": 582
},
{
"epoch": 1.0042918454935623,
"grad_norm": 0.21883882582187653,
"learning_rate": 0.00019999983812448848,
"loss": 0.915,
"step": 585
},
{
"epoch": 1.0128755364806867,
"grad_norm": 0.21552623808383942,
"learning_rate": 0.0001999988488871492,
"loss": 0.9101,
"step": 590
},
{
"epoch": 1.0214592274678111,
"grad_norm": 2.067782402038574,
"learning_rate": 0.00019999696035219593,
"loss": 0.9213,
"step": 595
},
{
"epoch": 1.0300429184549356,
"grad_norm": 0.22093655169010162,
"learning_rate": 0.00019999417253661235,
"loss": 0.9244,
"step": 600
},
{
"epoch": 1.0386266094420602,
"grad_norm": 0.2211354523897171,
"learning_rate": 0.00019999048546546954,
"loss": 0.8949,
"step": 605
},
{
"epoch": 1.0472103004291846,
"grad_norm": 0.21562980115413666,
"learning_rate": 0.00019998589917192568,
"loss": 0.9248,
"step": 610
},
{
"epoch": 1.055793991416309,
"grad_norm": 0.22060342133045197,
"learning_rate": 0.00019998041369722556,
"loss": 0.907,
"step": 615
},
{
"epoch": 1.0643776824034334,
"grad_norm": 0.20447732508182526,
"learning_rate": 0.00019997402909070059,
"loss": 0.9155,
"step": 620
},
{
"epoch": 1.0729613733905579,
"grad_norm": 0.21326489746570587,
"learning_rate": 0.000199966745409768,
"loss": 0.9148,
"step": 625
},
{
"epoch": 1.0815450643776825,
"grad_norm": 0.21152488887310028,
"learning_rate": 0.0001999585627199305,
"loss": 0.8954,
"step": 630
},
{
"epoch": 1.090128755364807,
"grad_norm": 0.21141602098941803,
"learning_rate": 0.0001999494810947757,
"loss": 0.9091,
"step": 635
},
{
"epoch": 1.0987124463519313,
"grad_norm": 0.21263065934181213,
"learning_rate": 0.00019993950061597535,
"loss": 0.9065,
"step": 640
},
{
"epoch": 1.1072961373390557,
"grad_norm": 0.21794655919075012,
"learning_rate": 0.00019992862137328474,
"loss": 0.9029,
"step": 645
},
{
"epoch": 1.1158798283261802,
"grad_norm": 0.21535712480545044,
"learning_rate": 0.00019991684346454172,
"loss": 0.9139,
"step": 650
},
{
"epoch": 1.1244635193133048,
"grad_norm": 0.1971653550863266,
"learning_rate": 0.00019990416699566598,
"loss": 0.8918,
"step": 655
},
{
"epoch": 1.1330472103004292,
"grad_norm": 0.2062826007604599,
"learning_rate": 0.000199890592080658,
"loss": 0.9188,
"step": 660
},
{
"epoch": 1.1416309012875536,
"grad_norm": 0.2263791412115097,
"learning_rate": 0.0001998761188415981,
"loss": 0.904,
"step": 665
},
{
"epoch": 1.150214592274678,
"grad_norm": 0.19670893251895905,
"learning_rate": 0.00019986074740864526,
"loss": 0.9165,
"step": 670
},
{
"epoch": 1.1587982832618025,
"grad_norm": 0.21204271912574768,
"learning_rate": 0.000199844477920036,
"loss": 0.8874,
"step": 675
},
{
"epoch": 1.167381974248927,
"grad_norm": 0.19298429787158966,
"learning_rate": 0.00019982731052208309,
"loss": 0.9102,
"step": 680
},
{
"epoch": 1.1759656652360515,
"grad_norm": 0.21324272453784943,
"learning_rate": 0.00019980924536917437,
"loss": 0.9119,
"step": 685
},
{
"epoch": 1.184549356223176,
"grad_norm": 0.192140594124794,
"learning_rate": 0.00019979028262377118,
"loss": 0.8957,
"step": 690
},
{
"epoch": 1.1931330472103003,
"grad_norm": 0.20233942568302155,
"learning_rate": 0.00019977042245640698,
"loss": 0.8969,
"step": 695
},
{
"epoch": 1.201716738197425,
"grad_norm": 0.21660216152668,
"learning_rate": 0.00019974966504568583,
"loss": 0.9064,
"step": 700
},
{
"epoch": 1.2103004291845494,
"grad_norm": 0.2243824154138565,
"learning_rate": 0.0001997280105782808,
"loss": 0.9166,
"step": 705
},
{
"epoch": 1.2188841201716738,
"grad_norm": 0.20581458508968353,
"learning_rate": 0.00019970545924893226,
"loss": 0.9064,
"step": 710
},
{
"epoch": 1.2274678111587982,
"grad_norm": 0.19686444103717804,
"learning_rate": 0.00019968201126044604,
"loss": 0.9126,
"step": 715
},
{
"epoch": 1.2360515021459229,
"grad_norm": 0.21029411256313324,
"learning_rate": 0.00019965766682369186,
"loss": 0.892,
"step": 720
},
{
"epoch": 1.2446351931330473,
"grad_norm": 0.21291205286979675,
"learning_rate": 0.0001996324261576011,
"loss": 0.8936,
"step": 725
},
{
"epoch": 1.2532188841201717,
"grad_norm": 0.2174995094537735,
"learning_rate": 0.00019960628948916518,
"loss": 0.9157,
"step": 730
},
{
"epoch": 1.261802575107296,
"grad_norm": 0.2011156976222992,
"learning_rate": 0.0001995792570534331,
"loss": 0.9161,
"step": 735
},
{
"epoch": 1.2703862660944205,
"grad_norm": 0.20497067272663116,
"learning_rate": 0.00019955132909350984,
"loss": 0.8999,
"step": 740
},
{
"epoch": 1.2789699570815452,
"grad_norm": 0.19909746944904327,
"learning_rate": 0.0001995225058605537,
"loss": 0.9142,
"step": 745
},
{
"epoch": 1.2875536480686696,
"grad_norm": 0.22116069495677948,
"learning_rate": 0.0001994927876137743,
"loss": 0.8992,
"step": 750
},
{
"epoch": 1.296137339055794,
"grad_norm": 0.22861087322235107,
"learning_rate": 0.00019946217462043025,
"loss": 0.898,
"step": 755
},
{
"epoch": 1.3047210300429184,
"grad_norm": 0.20132282376289368,
"learning_rate": 0.0001994306671558266,
"loss": 0.9147,
"step": 760
},
{
"epoch": 1.3133047210300428,
"grad_norm": 0.21803739666938782,
"learning_rate": 0.00019939826550331252,
"loss": 0.9,
"step": 765
},
{
"epoch": 1.3218884120171674,
"grad_norm": 0.19455976784229279,
"learning_rate": 0.0001993649699542786,
"loss": 0.9126,
"step": 770
},
{
"epoch": 1.3304721030042919,
"grad_norm": 0.18571655452251434,
"learning_rate": 0.0001993307808081544,
"loss": 0.9006,
"step": 775
},
{
"epoch": 1.3390557939914163,
"grad_norm": 0.20103998482227325,
"learning_rate": 0.00019929569837240564,
"loss": 0.8881,
"step": 780
},
{
"epoch": 1.3476394849785407,
"grad_norm": 0.19315999746322632,
"learning_rate": 0.00019925972296253145,
"loss": 0.901,
"step": 785
},
{
"epoch": 1.356223175965665,
"grad_norm": 0.2066372036933899,
"learning_rate": 0.00019922285490206156,
"loss": 0.888,
"step": 790
},
{
"epoch": 1.3648068669527897,
"grad_norm": 0.20879539847373962,
"learning_rate": 0.00019918509452255338,
"loss": 0.901,
"step": 795
},
{
"epoch": 1.3733905579399142,
"grad_norm": 0.20333191752433777,
"learning_rate": 0.000199146442163589,
"loss": 0.9099,
"step": 800
},
{
"epoch": 1.3819742489270386,
"grad_norm": 0.1949775516986847,
"learning_rate": 0.00019910689817277216,
"loss": 0.904,
"step": 805
},
{
"epoch": 1.3905579399141632,
"grad_norm": 0.20540495216846466,
"learning_rate": 0.00019906646290572514,
"loss": 0.8965,
"step": 810
},
{
"epoch": 1.3991416309012876,
"grad_norm": 0.19921506941318512,
"learning_rate": 0.00019902513672608553,
"loss": 0.8991,
"step": 815
},
{
"epoch": 1.407725321888412,
"grad_norm": 0.21238817274570465,
"learning_rate": 0.0001989829200055029,
"loss": 0.9026,
"step": 820
},
{
"epoch": 1.4163090128755365,
"grad_norm": 0.2081788033246994,
"learning_rate": 0.00019893981312363562,
"loss": 0.9052,
"step": 825
},
{
"epoch": 1.4248927038626609,
"grad_norm": 0.20578624308109283,
"learning_rate": 0.00019889581646814728,
"loss": 0.9038,
"step": 830
},
{
"epoch": 1.4334763948497855,
"grad_norm": 0.2119644731283188,
"learning_rate": 0.00019885093043470336,
"loss": 0.8936,
"step": 835
},
{
"epoch": 1.44206008583691,
"grad_norm": 0.19631995260715485,
"learning_rate": 0.0001988051554269675,
"loss": 0.9059,
"step": 840
},
{
"epoch": 1.4506437768240343,
"grad_norm": 0.22262215614318848,
"learning_rate": 0.00019875849185659798,
"loss": 0.9172,
"step": 845
},
{
"epoch": 1.4592274678111588,
"grad_norm": 0.19081105291843414,
"learning_rate": 0.00019871094014324404,
"loss": 0.9025,
"step": 850
},
{
"epoch": 1.4678111587982832,
"grad_norm": 0.18824172019958496,
"learning_rate": 0.000198662500714542,
"loss": 0.9141,
"step": 855
},
{
"epoch": 1.4763948497854078,
"grad_norm": 0.20280902087688446,
"learning_rate": 0.0001986131740061115,
"loss": 0.8889,
"step": 860
},
{
"epoch": 1.4849785407725322,
"grad_norm": 0.19314704835414886,
"learning_rate": 0.00019856296046155157,
"loss": 0.8919,
"step": 865
},
{
"epoch": 1.4935622317596566,
"grad_norm": 0.1936980038881302,
"learning_rate": 0.00019851186053243666,
"loss": 0.9015,
"step": 870
},
{
"epoch": 1.5021459227467813,
"grad_norm": 0.21349290013313293,
"learning_rate": 0.00019845987467831242,
"loss": 0.9068,
"step": 875
},
{
"epoch": 1.5107296137339055,
"grad_norm": 0.1915241926908493,
"learning_rate": 0.00019840700336669183,
"loss": 0.9148,
"step": 880
},
{
"epoch": 1.51931330472103,
"grad_norm": 0.1982114166021347,
"learning_rate": 0.00019835324707305076,
"loss": 0.9043,
"step": 885
},
{
"epoch": 1.5278969957081545,
"grad_norm": 0.18504977226257324,
"learning_rate": 0.0001982986062808239,
"loss": 0.8926,
"step": 890
},
{
"epoch": 1.536480686695279,
"grad_norm": 0.22229517996311188,
"learning_rate": 0.0001982430814814002,
"loss": 0.8849,
"step": 895
},
{
"epoch": 1.5450643776824036,
"grad_norm": 0.21088634431362152,
"learning_rate": 0.00019818667317411865,
"loss": 0.9075,
"step": 900
},
{
"epoch": 1.5536480686695278,
"grad_norm": 0.20124419033527374,
"learning_rate": 0.0001981293818662636,
"loss": 0.8914,
"step": 905
},
{
"epoch": 1.5622317596566524,
"grad_norm": 0.19154104590415955,
"learning_rate": 0.0001980712080730604,
"loss": 0.8816,
"step": 910
},
{
"epoch": 1.5708154506437768,
"grad_norm": 0.1901169866323471,
"learning_rate": 0.00019801215231767056,
"loss": 0.9051,
"step": 915
},
{
"epoch": 1.5793991416309012,
"grad_norm": 0.19122549891471863,
"learning_rate": 0.00019795221513118722,
"loss": 0.8965,
"step": 920
},
{
"epoch": 1.5879828326180259,
"grad_norm": 0.192024827003479,
"learning_rate": 0.00019789139705263026,
"loss": 0.8958,
"step": 925
},
{
"epoch": 1.59656652360515,
"grad_norm": 0.19915080070495605,
"learning_rate": 0.0001978296986289415,
"loss": 0.8924,
"step": 930
},
{
"epoch": 1.6051502145922747,
"grad_norm": 0.19441018998622894,
"learning_rate": 0.0001977671204149798,
"loss": 0.8997,
"step": 935
},
{
"epoch": 1.613733905579399,
"grad_norm": 0.20060202479362488,
"learning_rate": 0.000197703662973516,
"loss": 0.8808,
"step": 940
},
{
"epoch": 1.6223175965665235,
"grad_norm": 0.20653241872787476,
"learning_rate": 0.00019763932687522794,
"loss": 0.9032,
"step": 945
},
{
"epoch": 1.6309012875536482,
"grad_norm": 0.1972658932209015,
"learning_rate": 0.00019757411269869527,
"loss": 0.9093,
"step": 950
},
{
"epoch": 1.6394849785407726,
"grad_norm": 0.18751849234104156,
"learning_rate": 0.0001975080210303943,
"loss": 0.8842,
"step": 955
},
{
"epoch": 1.648068669527897,
"grad_norm": 0.1898711621761322,
"learning_rate": 0.00019744105246469263,
"loss": 0.8914,
"step": 960
},
{
"epoch": 1.6566523605150214,
"grad_norm": 0.1935146301984787,
"learning_rate": 0.0001973732076038439,
"loss": 0.881,
"step": 965
},
{
"epoch": 1.6652360515021458,
"grad_norm": 0.19017855823040009,
"learning_rate": 0.00019730448705798239,
"loss": 0.9005,
"step": 970
},
{
"epoch": 1.6738197424892705,
"grad_norm": 0.1929733008146286,
"learning_rate": 0.00019723489144511742,
"loss": 0.8898,
"step": 975
},
{
"epoch": 1.6824034334763949,
"grad_norm": 0.1935940533876419,
"learning_rate": 0.0001971644213911279,
"loss": 0.8889,
"step": 980
},
{
"epoch": 1.6909871244635193,
"grad_norm": 0.19845978915691376,
"learning_rate": 0.0001970930775297566,
"loss": 0.8891,
"step": 985
},
{
"epoch": 1.699570815450644,
"grad_norm": 0.18910686671733856,
"learning_rate": 0.00019702086050260456,
"loss": 0.8909,
"step": 990
},
{
"epoch": 1.7081545064377681,
"grad_norm": 0.20300810039043427,
"learning_rate": 0.00019694777095912534,
"loss": 0.9012,
"step": 995
},
{
"epoch": 1.7167381974248928,
"grad_norm": 0.21887531876564026,
"learning_rate": 0.0001968738095566189,
"loss": 0.9116,
"step": 1000
},
{
"epoch": 1.7253218884120172,
"grad_norm": 0.21735869348049164,
"learning_rate": 0.00019679897696022608,
"loss": 0.8873,
"step": 1005
},
{
"epoch": 1.7339055793991416,
"grad_norm": 0.1856721192598343,
"learning_rate": 0.0001967232738429224,
"loss": 0.8889,
"step": 1010
},
{
"epoch": 1.7424892703862662,
"grad_norm": 0.2046109437942505,
"learning_rate": 0.000196646700885512,
"loss": 0.89,
"step": 1015
},
{
"epoch": 1.7510729613733904,
"grad_norm": 0.19593974947929382,
"learning_rate": 0.0001965692587766216,
"loss": 0.9065,
"step": 1020
},
{
"epoch": 1.759656652360515,
"grad_norm": 0.18540222942829132,
"learning_rate": 0.00019649094821269425,
"loss": 0.8877,
"step": 1025
},
{
"epoch": 1.7682403433476395,
"grad_norm": 0.19899272918701172,
"learning_rate": 0.00019641176989798305,
"loss": 0.8965,
"step": 1030
},
{
"epoch": 1.7768240343347639,
"grad_norm": 0.18957588076591492,
"learning_rate": 0.00019633172454454497,
"loss": 0.8876,
"step": 1035
},
{
"epoch": 1.7854077253218885,
"grad_norm": 0.20278845727443695,
"learning_rate": 0.0001962508128722342,
"loss": 0.8982,
"step": 1040
},
{
"epoch": 1.7939914163090127,
"grad_norm": 0.1874280869960785,
"learning_rate": 0.00019616903560869584,
"loss": 0.9049,
"step": 1045
},
{
"epoch": 1.8025751072961373,
"grad_norm": 0.19025950133800507,
"learning_rate": 0.0001960863934893594,
"loss": 0.8901,
"step": 1050
},
{
"epoch": 1.8111587982832618,
"grad_norm": 0.20806211233139038,
"learning_rate": 0.00019600288725743194,
"loss": 0.9015,
"step": 1055
},
{
"epoch": 1.8197424892703862,
"grad_norm": 0.2008458375930786,
"learning_rate": 0.00019591851766389176,
"loss": 0.9029,
"step": 1060
},
{
"epoch": 1.8283261802575108,
"grad_norm": 0.1986788511276245,
"learning_rate": 0.00019583328546748127,
"loss": 0.8942,
"step": 1065
},
{
"epoch": 1.8369098712446352,
"grad_norm": 0.18976636230945587,
"learning_rate": 0.00019574719143470044,
"loss": 0.8962,
"step": 1070
},
{
"epoch": 1.8454935622317596,
"grad_norm": 0.19317425787448883,
"learning_rate": 0.00019566023633979976,
"loss": 0.8918,
"step": 1075
},
{
"epoch": 1.8540772532188843,
"grad_norm": 0.1889304369688034,
"learning_rate": 0.00019557242096477327,
"loss": 0.8934,
"step": 1080
},
{
"epoch": 1.8626609442060085,
"grad_norm": 0.18771173059940338,
"learning_rate": 0.00019548374609935172,
"loss": 0.8782,
"step": 1085
},
{
"epoch": 1.871244635193133,
"grad_norm": 0.18727517127990723,
"learning_rate": 0.00019539421254099519,
"loss": 0.9014,
"step": 1090
},
{
"epoch": 1.8798283261802575,
"grad_norm": 0.19307033717632294,
"learning_rate": 0.0001953038210948861,
"loss": 0.896,
"step": 1095
},
{
"epoch": 1.888412017167382,
"grad_norm": 0.1863000988960266,
"learning_rate": 0.00019521257257392192,
"loss": 0.8855,
"step": 1100
},
{
"epoch": 1.8969957081545066,
"grad_norm": 0.1884726732969284,
"learning_rate": 0.0001951204677987079,
"loss": 0.8902,
"step": 1105
},
{
"epoch": 1.9055793991416308,
"grad_norm": 0.20304642617702484,
"learning_rate": 0.00019502750759754962,
"loss": 0.8892,
"step": 1110
},
{
"epoch": 1.9141630901287554,
"grad_norm": 0.1887015998363495,
"learning_rate": 0.00019493369280644554,
"loss": 0.8946,
"step": 1115
},
{
"epoch": 1.9227467811158798,
"grad_norm": 0.18979288637638092,
"learning_rate": 0.00019483902426907954,
"loss": 0.8825,
"step": 1120
},
{
"epoch": 1.9313304721030042,
"grad_norm": 0.18896907567977905,
"learning_rate": 0.00019474350283681338,
"loss": 0.887,
"step": 1125
},
{
"epoch": 1.9399141630901289,
"grad_norm": 0.17926710844039917,
"learning_rate": 0.00019464712936867885,
"loss": 0.8832,
"step": 1130
},
{
"epoch": 1.948497854077253,
"grad_norm": 0.19314360618591309,
"learning_rate": 0.00019454990473137028,
"loss": 0.89,
"step": 1135
},
{
"epoch": 1.9570815450643777,
"grad_norm": 0.19700467586517334,
"learning_rate": 0.00019445182979923654,
"loss": 0.8844,
"step": 1140
},
{
"epoch": 1.9656652360515021,
"grad_norm": 0.20681554079055786,
"learning_rate": 0.00019435290545427328,
"loss": 0.896,
"step": 1145
},
{
"epoch": 1.9742489270386265,
"grad_norm": 0.1876552402973175,
"learning_rate": 0.0001942531325861151,
"loss": 0.886,
"step": 1150
},
{
"epoch": 1.9828326180257512,
"grad_norm": 0.18256564438343048,
"learning_rate": 0.0001941525120920273,
"loss": 0.9008,
"step": 1155
},
{
"epoch": 1.9914163090128756,
"grad_norm": 0.19153741002082825,
"learning_rate": 0.00019405104487689798,
"loss": 0.8804,
"step": 1160
},
{
"epoch": 2.0,
"grad_norm": 0.19448307156562805,
"learning_rate": 0.0001939487318532299,
"loss": 0.8939,
"step": 1165
},
{
"epoch": 2.0,
"eval_loss": 2.0231504440307617,
"eval_runtime": 0.3938,
"eval_samples_per_second": 15.238,
"eval_steps_per_second": 2.54,
"step": 1165
},
{
"epoch": 2.0085836909871246,
"grad_norm": 0.2327311784029007,
"learning_rate": 0.00019384557394113228,
"loss": 0.8256,
"step": 1170
},
{
"epoch": 2.017167381974249,
"grad_norm": 0.23671980202198029,
"learning_rate": 0.00019374157206831236,
"loss": 0.8223,
"step": 1175
},
{
"epoch": 2.0257510729613735,
"grad_norm": 0.23530033230781555,
"learning_rate": 0.00019363672717006734,
"loss": 0.8109,
"step": 1180
},
{
"epoch": 2.0343347639484977,
"grad_norm": 0.21592716872692108,
"learning_rate": 0.00019353104018927567,
"loss": 0.8071,
"step": 1185
},
{
"epoch": 2.0429184549356223,
"grad_norm": 0.22698843479156494,
"learning_rate": 0.0001934245120763889,
"loss": 0.8047,
"step": 1190
},
{
"epoch": 2.051502145922747,
"grad_norm": 0.20773455500602722,
"learning_rate": 0.0001933171437894227,
"loss": 0.8081,
"step": 1195
},
{
"epoch": 2.060085836909871,
"grad_norm": 0.21392963826656342,
"learning_rate": 0.00019320893629394873,
"loss": 0.8267,
"step": 1200
},
{
"epoch": 2.0686695278969958,
"grad_norm": 0.1993769258260727,
"learning_rate": 0.00019309989056308556,
"loss": 0.8122,
"step": 1205
},
{
"epoch": 2.0772532188841204,
"grad_norm": 0.21093840897083282,
"learning_rate": 0.00019299000757749016,
"loss": 0.8135,
"step": 1210
},
{
"epoch": 2.0858369098712446,
"grad_norm": 0.21615874767303467,
"learning_rate": 0.00019287928832534897,
"loss": 0.8098,
"step": 1215
},
{
"epoch": 2.0944206008583692,
"grad_norm": 0.21068502962589264,
"learning_rate": 0.00019276773380236904,
"loss": 0.813,
"step": 1220
},
{
"epoch": 2.1030042918454934,
"grad_norm": 0.21265622973442078,
"learning_rate": 0.00019265534501176906,
"loss": 0.8034,
"step": 1225
},
{
"epoch": 2.111587982832618,
"grad_norm": 0.22396954894065857,
"learning_rate": 0.00019254212296427044,
"loss": 0.8221,
"step": 1230
},
{
"epoch": 2.1201716738197427,
"grad_norm": 0.2072274535894394,
"learning_rate": 0.00019242806867808798,
"loss": 0.8235,
"step": 1235
},
{
"epoch": 2.128755364806867,
"grad_norm": 0.20798753201961517,
"learning_rate": 0.00019231318317892106,
"loss": 0.8137,
"step": 1240
},
{
"epoch": 2.1373390557939915,
"grad_norm": 0.2030133157968521,
"learning_rate": 0.00019219746749994405,
"loss": 0.8154,
"step": 1245
},
{
"epoch": 2.1459227467811157,
"grad_norm": 0.22628700733184814,
"learning_rate": 0.0001920809226817973,
"loss": 0.8261,
"step": 1250
},
{
"epoch": 2.1545064377682404,
"grad_norm": 0.21634644269943237,
"learning_rate": 0.00019196354977257766,
"loss": 0.8221,
"step": 1255
},
{
"epoch": 2.163090128755365,
"grad_norm": 0.2259581983089447,
"learning_rate": 0.00019184534982782904,
"loss": 0.8287,
"step": 1260
},
{
"epoch": 2.171673819742489,
"grad_norm": 0.23607933521270752,
"learning_rate": 0.00019172632391053294,
"loss": 0.8218,
"step": 1265
},
{
"epoch": 2.180257510729614,
"grad_norm": 0.20960725843906403,
"learning_rate": 0.0001916064730910989,
"loss": 0.8233,
"step": 1270
},
{
"epoch": 2.188841201716738,
"grad_norm": 0.19818070530891418,
"learning_rate": 0.00019148579844735497,
"loss": 0.8253,
"step": 1275
},
{
"epoch": 2.1974248927038627,
"grad_norm": 0.2142871767282486,
"learning_rate": 0.00019136430106453777,
"loss": 0.8289,
"step": 1280
},
{
"epoch": 2.2060085836909873,
"grad_norm": 0.21934735774993896,
"learning_rate": 0.0001912419820352829,
"loss": 0.8191,
"step": 1285
},
{
"epoch": 2.2145922746781115,
"grad_norm": 0.21653762459754944,
"learning_rate": 0.00019111884245961522,
"loss": 0.8194,
"step": 1290
},
{
"epoch": 2.223175965665236,
"grad_norm": 0.21233248710632324,
"learning_rate": 0.00019099488344493873,
"loss": 0.8247,
"step": 1295
},
{
"epoch": 2.2317596566523603,
"grad_norm": 0.23292584717273712,
"learning_rate": 0.00019087010610602668,
"loss": 0.8197,
"step": 1300
},
{
"epoch": 2.240343347639485,
"grad_norm": 0.20501044392585754,
"learning_rate": 0.00019074451156501164,
"loss": 0.8152,
"step": 1305
},
{
"epoch": 2.2489270386266096,
"grad_norm": 0.23035867512226105,
"learning_rate": 0.00019061810095137533,
"loss": 0.8168,
"step": 1310
},
{
"epoch": 2.257510729613734,
"grad_norm": 0.21323524415493011,
"learning_rate": 0.00019049087540193847,
"loss": 0.8118,
"step": 1315
},
{
"epoch": 2.2660944206008584,
"grad_norm": 0.20477545261383057,
"learning_rate": 0.00019036283606085053,
"loss": 0.8164,
"step": 1320
},
{
"epoch": 2.274678111587983,
"grad_norm": 0.21431773900985718,
"learning_rate": 0.00019023398407957956,
"loss": 0.8258,
"step": 1325
},
{
"epoch": 2.2832618025751072,
"grad_norm": 0.21260547637939453,
"learning_rate": 0.00019010432061690165,
"loss": 0.8166,
"step": 1330
},
{
"epoch": 2.291845493562232,
"grad_norm": 0.20846493542194366,
"learning_rate": 0.00018997384683889067,
"loss": 0.8066,
"step": 1335
},
{
"epoch": 2.300429184549356,
"grad_norm": 0.20466403663158417,
"learning_rate": 0.00018984256391890765,
"loss": 0.8251,
"step": 1340
},
{
"epoch": 2.3090128755364807,
"grad_norm": 0.20920304954051971,
"learning_rate": 0.0001897104730375904,
"loss": 0.8164,
"step": 1345
},
{
"epoch": 2.317596566523605,
"grad_norm": 0.22407568991184235,
"learning_rate": 0.00018957757538284273,
"loss": 0.8156,
"step": 1350
},
{
"epoch": 2.3261802575107295,
"grad_norm": 0.22706876695156097,
"learning_rate": 0.00018944387214982382,
"loss": 0.8348,
"step": 1355
},
{
"epoch": 2.334763948497854,
"grad_norm": 0.20964165031909943,
"learning_rate": 0.00018930936454093753,
"loss": 0.8258,
"step": 1360
},
{
"epoch": 2.3433476394849784,
"grad_norm": 0.23025156557559967,
"learning_rate": 0.00018917405376582145,
"loss": 0.8205,
"step": 1365
},
{
"epoch": 2.351931330472103,
"grad_norm": 0.23171818256378174,
"learning_rate": 0.0001890379410413362,
"loss": 0.8224,
"step": 1370
},
{
"epoch": 2.3605150214592276,
"grad_norm": 0.2035280019044876,
"learning_rate": 0.0001889010275915543,
"loss": 0.8363,
"step": 1375
},
{
"epoch": 2.369098712446352,
"grad_norm": 0.2576422691345215,
"learning_rate": 0.00018876331464774945,
"loss": 0.8216,
"step": 1380
},
{
"epoch": 2.3776824034334765,
"grad_norm": 0.21184222400188446,
"learning_rate": 0.00018862480344838495,
"loss": 0.8161,
"step": 1385
},
{
"epoch": 2.3862660944206007,
"grad_norm": 0.22491346299648285,
"learning_rate": 0.00018848549523910313,
"loss": 0.8261,
"step": 1390
},
{
"epoch": 2.3948497854077253,
"grad_norm": 0.21227188408374786,
"learning_rate": 0.0001883453912727138,
"loss": 0.8377,
"step": 1395
},
{
"epoch": 2.40343347639485,
"grad_norm": 0.21044416725635529,
"learning_rate": 0.0001882044928091831,
"loss": 0.819,
"step": 1400
},
{
"epoch": 2.412017167381974,
"grad_norm": 0.20745404064655304,
"learning_rate": 0.00018806280111562215,
"loss": 0.8265,
"step": 1405
},
{
"epoch": 2.4206008583690988,
"grad_norm": 0.2179802805185318,
"learning_rate": 0.00018792031746627563,
"loss": 0.8382,
"step": 1410
},
{
"epoch": 2.429184549356223,
"grad_norm": 0.20480507612228394,
"learning_rate": 0.00018777704314251032,
"loss": 0.8312,
"step": 1415
},
{
"epoch": 2.4377682403433476,
"grad_norm": 0.23306381702423096,
"learning_rate": 0.00018763297943280368,
"loss": 0.8161,
"step": 1420
},
{
"epoch": 2.4463519313304722,
"grad_norm": 0.21607355773448944,
"learning_rate": 0.00018748812763273208,
"loss": 0.8197,
"step": 1425
},
{
"epoch": 2.4549356223175964,
"grad_norm": 0.21942569315433502,
"learning_rate": 0.0001873424890449593,
"loss": 0.8291,
"step": 1430
},
{
"epoch": 2.463519313304721,
"grad_norm": 0.2144131362438202,
"learning_rate": 0.00018719606497922476,
"loss": 0.8203,
"step": 1435
},
{
"epoch": 2.4721030042918457,
"grad_norm": 0.21602974832057953,
"learning_rate": 0.0001870488567523318,
"loss": 0.8154,
"step": 1440
},
{
"epoch": 2.48068669527897,
"grad_norm": 0.2094966620206833,
"learning_rate": 0.0001869008656881357,
"loss": 0.8197,
"step": 1445
},
{
"epoch": 2.4892703862660945,
"grad_norm": 0.21330519020557404,
"learning_rate": 0.00018675209311753185,
"loss": 0.8325,
"step": 1450
},
{
"epoch": 2.4978540772532187,
"grad_norm": 0.22934697568416595,
"learning_rate": 0.00018660254037844388,
"loss": 0.8238,
"step": 1455
},
{
"epoch": 2.5064377682403434,
"grad_norm": 0.23202557861804962,
"learning_rate": 0.00018645220881581144,
"loss": 0.8277,
"step": 1460
},
{
"epoch": 2.5150214592274676,
"grad_norm": 0.22423741221427917,
"learning_rate": 0.0001863010997815783,
"loss": 0.8205,
"step": 1465
},
{
"epoch": 2.523605150214592,
"grad_norm": 0.2139664888381958,
"learning_rate": 0.00018614921463468002,
"loss": 0.833,
"step": 1470
},
{
"epoch": 2.532188841201717,
"grad_norm": 0.22042877972126007,
"learning_rate": 0.00018599655474103182,
"loss": 0.8281,
"step": 1475
},
{
"epoch": 2.540772532188841,
"grad_norm": 0.21639470756053925,
"learning_rate": 0.0001858431214735163,
"loss": 0.8353,
"step": 1480
},
{
"epoch": 2.5493562231759657,
"grad_norm": 0.21406595408916473,
"learning_rate": 0.00018568891621197103,
"loss": 0.8351,
"step": 1485
},
{
"epoch": 2.5579399141630903,
"grad_norm": 0.20837725698947906,
"learning_rate": 0.00018553394034317622,
"loss": 0.8251,
"step": 1490
},
{
"epoch": 2.5665236051502145,
"grad_norm": 0.21612149477005005,
"learning_rate": 0.0001853781952608422,
"loss": 0.846,
"step": 1495
},
{
"epoch": 2.575107296137339,
"grad_norm": 0.21887291967868805,
"learning_rate": 0.00018522168236559695,
"loss": 0.8388,
"step": 1500
},
{
"epoch": 2.5836909871244638,
"grad_norm": 0.20973001420497894,
"learning_rate": 0.00018506440306497335,
"loss": 0.839,
"step": 1505
},
{
"epoch": 2.592274678111588,
"grad_norm": 0.21462783217430115,
"learning_rate": 0.00018490635877339666,
"loss": 0.8276,
"step": 1510
},
{
"epoch": 2.6008583690987126,
"grad_norm": 0.210985004901886,
"learning_rate": 0.00018474755091217186,
"loss": 0.8221,
"step": 1515
},
{
"epoch": 2.609442060085837,
"grad_norm": 0.20986580848693848,
"learning_rate": 0.00018458798090947065,
"loss": 0.8234,
"step": 1520
},
{
"epoch": 2.6180257510729614,
"grad_norm": 0.22892533242702484,
"learning_rate": 0.00018442765020031877,
"loss": 0.8242,
"step": 1525
},
{
"epoch": 2.6266094420600856,
"grad_norm": 0.2284938395023346,
"learning_rate": 0.0001842665602265831,
"loss": 0.8161,
"step": 1530
},
{
"epoch": 2.6351931330472103,
"grad_norm": 0.22317782044410706,
"learning_rate": 0.00018410471243695856,
"loss": 0.8284,
"step": 1535
},
{
"epoch": 2.643776824034335,
"grad_norm": 0.21049915254116058,
"learning_rate": 0.00018394210828695523,
"loss": 0.8183,
"step": 1540
},
{
"epoch": 2.652360515021459,
"grad_norm": 0.21207213401794434,
"learning_rate": 0.0001837787492388852,
"loss": 0.8287,
"step": 1545
},
{
"epoch": 2.6609442060085837,
"grad_norm": 0.2118200659751892,
"learning_rate": 0.0001836146367618494,
"loss": 0.8204,
"step": 1550
},
{
"epoch": 2.6695278969957084,
"grad_norm": 0.22095955908298492,
"learning_rate": 0.00018344977233172437,
"loss": 0.8335,
"step": 1555
},
{
"epoch": 2.6781115879828326,
"grad_norm": 0.21252469718456268,
"learning_rate": 0.00018328415743114912,
"loss": 0.8191,
"step": 1560
},
{
"epoch": 2.686695278969957,
"grad_norm": 0.20323017239570618,
"learning_rate": 0.0001831177935495116,
"loss": 0.8231,
"step": 1565
},
{
"epoch": 2.6952789699570814,
"grad_norm": 0.21805858612060547,
"learning_rate": 0.00018295068218293547,
"loss": 0.8341,
"step": 1570
},
{
"epoch": 2.703862660944206,
"grad_norm": 0.21013419330120087,
"learning_rate": 0.00018278282483426658,
"loss": 0.839,
"step": 1575
},
{
"epoch": 2.71244635193133,
"grad_norm": 0.21768461167812347,
"learning_rate": 0.0001826142230130594,
"loss": 0.8356,
"step": 1580
},
{
"epoch": 2.721030042918455,
"grad_norm": 0.21069899201393127,
"learning_rate": 0.00018244487823556357,
"loss": 0.8188,
"step": 1585
},
{
"epoch": 2.7296137339055795,
"grad_norm": 0.21788835525512695,
"learning_rate": 0.00018227479202471015,
"loss": 0.8408,
"step": 1590
},
{
"epoch": 2.7381974248927037,
"grad_norm": 0.21380050480365753,
"learning_rate": 0.00018210396591009795,
"loss": 0.8358,
"step": 1595
},
{
"epoch": 2.7467811158798283,
"grad_norm": 0.21521276235580444,
"learning_rate": 0.00018193240142797988,
"loss": 0.8328,
"step": 1600
},
{
"epoch": 2.755364806866953,
"grad_norm": 0.20885252952575684,
"learning_rate": 0.000181760100121249,
"loss": 0.8238,
"step": 1605
},
{
"epoch": 2.763948497854077,
"grad_norm": 0.21117731928825378,
"learning_rate": 0.00018158706353942463,
"loss": 0.8301,
"step": 1610
},
{
"epoch": 2.772532188841202,
"grad_norm": 0.22012095153331757,
"learning_rate": 0.0001814132932386386,
"loss": 0.8357,
"step": 1615
},
{
"epoch": 2.7811158798283264,
"grad_norm": 0.22017072141170502,
"learning_rate": 0.00018123879078162097,
"loss": 0.8323,
"step": 1620
},
{
"epoch": 2.7896995708154506,
"grad_norm": 0.2259422242641449,
"learning_rate": 0.00018106355773768638,
"loss": 0.848,
"step": 1625
},
{
"epoch": 2.7982832618025753,
"grad_norm": 0.21191255748271942,
"learning_rate": 0.0001808875956827194,
"loss": 0.823,
"step": 1630
},
{
"epoch": 2.8068669527896994,
"grad_norm": 0.21371833980083466,
"learning_rate": 0.00018071090619916093,
"loss": 0.8194,
"step": 1635
},
{
"epoch": 2.815450643776824,
"grad_norm": 0.22189456224441528,
"learning_rate": 0.00018053349087599353,
"loss": 0.8329,
"step": 1640
},
{
"epoch": 2.8240343347639483,
"grad_norm": 0.20956319570541382,
"learning_rate": 0.00018035535130872732,
"loss": 0.8293,
"step": 1645
},
{
"epoch": 2.832618025751073,
"grad_norm": 0.21734033524990082,
"learning_rate": 0.0001801764890993856,
"loss": 0.8334,
"step": 1650
},
{
"epoch": 2.8412017167381975,
"grad_norm": 0.2138412892818451,
"learning_rate": 0.00017999690585649052,
"loss": 0.8354,
"step": 1655
},
{
"epoch": 2.8497854077253217,
"grad_norm": 0.21562372148036957,
"learning_rate": 0.00017981660319504845,
"loss": 0.8384,
"step": 1660
},
{
"epoch": 2.8583690987124464,
"grad_norm": 0.21281686425209045,
"learning_rate": 0.0001796355827365356,
"loss": 0.8312,
"step": 1665
},
{
"epoch": 2.866952789699571,
"grad_norm": 0.21461673080921173,
"learning_rate": 0.00017945384610888341,
"loss": 0.8344,
"step": 1670
},
{
"epoch": 2.875536480686695,
"grad_norm": 0.20743022859096527,
"learning_rate": 0.00017927139494646377,
"loss": 0.8215,
"step": 1675
},
{
"epoch": 2.88412017167382,
"grad_norm": 0.21129368245601654,
"learning_rate": 0.00017908823089007457,
"loss": 0.8274,
"step": 1680
},
{
"epoch": 2.8927038626609445,
"grad_norm": 0.2333795428276062,
"learning_rate": 0.00017890435558692475,
"loss": 0.8307,
"step": 1685
},
{
"epoch": 2.9012875536480687,
"grad_norm": 0.21824228763580322,
"learning_rate": 0.0001787197706906196,
"loss": 0.8498,
"step": 1690
},
{
"epoch": 2.909871244635193,
"grad_norm": 0.21459732949733734,
"learning_rate": 0.0001785344778611457,
"loss": 0.8265,
"step": 1695
},
{
"epoch": 2.9184549356223175,
"grad_norm": 0.20637935400009155,
"learning_rate": 0.00017834847876485629,
"loss": 0.8309,
"step": 1700
},
{
"epoch": 2.927038626609442,
"grad_norm": 0.2137777954339981,
"learning_rate": 0.0001781617750744561,
"loss": 0.8345,
"step": 1705
},
{
"epoch": 2.9356223175965663,
"grad_norm": 0.23476457595825195,
"learning_rate": 0.00017797436846898619,
"loss": 0.8335,
"step": 1710
},
{
"epoch": 2.944206008583691,
"grad_norm": 0.20995980501174927,
"learning_rate": 0.00017778626063380917,
"loss": 0.8209,
"step": 1715
},
{
"epoch": 2.9527896995708156,
"grad_norm": 0.2296920269727707,
"learning_rate": 0.00017759745326059379,
"loss": 0.8426,
"step": 1720
},
{
"epoch": 2.96137339055794,
"grad_norm": 0.20545101165771484,
"learning_rate": 0.00017740794804729969,
"loss": 0.8324,
"step": 1725
},
{
"epoch": 2.9699570815450644,
"grad_norm": 0.21105705201625824,
"learning_rate": 0.00017721774669816252,
"loss": 0.8212,
"step": 1730
},
{
"epoch": 2.978540772532189,
"grad_norm": 0.21741057932376862,
"learning_rate": 0.000177026850923678,
"loss": 0.8333,
"step": 1735
},
{
"epoch": 2.9871244635193133,
"grad_norm": 0.22390629351139069,
"learning_rate": 0.00017683526244058716,
"loss": 0.8364,
"step": 1740
},
{
"epoch": 2.995708154506438,
"grad_norm": 0.21623565256595612,
"learning_rate": 0.00017664298297186042,
"loss": 0.8255,
"step": 1745
},
{
"epoch": 2.9991416309012875,
"eval_loss": 2.1085665225982666,
"eval_runtime": 0.3945,
"eval_samples_per_second": 15.208,
"eval_steps_per_second": 2.535,
"step": 1747
},
{
"epoch": 3.004291845493562,
"grad_norm": 0.21203316748142242,
"learning_rate": 0.00017645001424668237,
"loss": 0.7739,
"step": 1750
},
{
"epoch": 3.0128755364806867,
"grad_norm": 0.23032112419605255,
"learning_rate": 0.00017625635800043617,
"loss": 0.741,
"step": 1755
},
{
"epoch": 3.0214592274678114,
"grad_norm": 0.24847178161144257,
"learning_rate": 0.00017606201597468782,
"loss": 0.7348,
"step": 1760
},
{
"epoch": 3.0300429184549356,
"grad_norm": 0.24480335414409637,
"learning_rate": 0.00017586698991717064,
"loss": 0.7212,
"step": 1765
},
{
"epoch": 3.03862660944206,
"grad_norm": 0.24489726126194,
"learning_rate": 0.00017567128158176953,
"loss": 0.7312,
"step": 1770
},
{
"epoch": 3.0472103004291844,
"grad_norm": 0.24028155207633972,
"learning_rate": 0.00017547489272850511,
"loss": 0.7271,
"step": 1775
},
{
"epoch": 3.055793991416309,
"grad_norm": 0.24730311334133148,
"learning_rate": 0.00017527782512351804,
"loss": 0.7344,
"step": 1780
},
{
"epoch": 3.0643776824034337,
"grad_norm": 0.23651528358459473,
"learning_rate": 0.00017508008053905295,
"loss": 0.7297,
"step": 1785
},
{
"epoch": 3.072961373390558,
"grad_norm": 0.23505684733390808,
"learning_rate": 0.0001748816607534426,
"loss": 0.7214,
"step": 1790
},
{
"epoch": 3.0815450643776825,
"grad_norm": 0.2424248605966568,
"learning_rate": 0.00017468256755109199,
"loss": 0.721,
"step": 1795
},
{
"epoch": 3.0901287553648067,
"grad_norm": 0.243468776345253,
"learning_rate": 0.00017448280272246212,
"loss": 0.7203,
"step": 1800
},
{
"epoch": 3.0987124463519313,
"grad_norm": 0.2453926056623459,
"learning_rate": 0.000174282368064054,
"loss": 0.7326,
"step": 1805
},
{
"epoch": 3.107296137339056,
"grad_norm": 0.24548988044261932,
"learning_rate": 0.00017408126537839252,
"loss": 0.7345,
"step": 1810
},
{
"epoch": 3.11587982832618,
"grad_norm": 0.2359829694032669,
"learning_rate": 0.00017387949647401012,
"loss": 0.748,
"step": 1815
},
{
"epoch": 3.124463519313305,
"grad_norm": 0.25862741470336914,
"learning_rate": 0.00017367706316543063,
"loss": 0.7338,
"step": 1820
},
{
"epoch": 3.133047210300429,
"grad_norm": 0.25267720222473145,
"learning_rate": 0.00017347396727315296,
"loss": 0.7403,
"step": 1825
},
{
"epoch": 3.1416309012875536,
"grad_norm": 0.2458384782075882,
"learning_rate": 0.00017327021062363458,
"loss": 0.7432,
"step": 1830
},
{
"epoch": 3.1502145922746783,
"grad_norm": 0.2578388750553131,
"learning_rate": 0.0001730657950492753,
"loss": 0.7447,
"step": 1835
},
{
"epoch": 3.1587982832618025,
"grad_norm": 0.23755429685115814,
"learning_rate": 0.00017286072238840067,
"loss": 0.7389,
"step": 1840
},
{
"epoch": 3.167381974248927,
"grad_norm": 0.24692735075950623,
"learning_rate": 0.0001726549944852455,
"loss": 0.7584,
"step": 1845
},
{
"epoch": 3.1759656652360517,
"grad_norm": 0.24396221339702606,
"learning_rate": 0.00017244861318993713,
"loss": 0.7386,
"step": 1850
},
{
"epoch": 3.184549356223176,
"grad_norm": 0.25548049807548523,
"learning_rate": 0.00017224158035847905,
"loss": 0.738,
"step": 1855
},
{
"epoch": 3.1931330472103006,
"grad_norm": 0.2472919523715973,
"learning_rate": 0.000172033897852734,
"loss": 0.7519,
"step": 1860
},
{
"epoch": 3.2017167381974247,
"grad_norm": 0.245948925614357,
"learning_rate": 0.0001718255675404073,
"loss": 0.7461,
"step": 1865
},
{
"epoch": 3.2103004291845494,
"grad_norm": 0.2513918876647949,
"learning_rate": 0.00017161659129503003,
"loss": 0.7458,
"step": 1870
},
{
"epoch": 3.218884120171674,
"grad_norm": 0.24049414694309235,
"learning_rate": 0.0001714069709959422,
"loss": 0.7344,
"step": 1875
},
{
"epoch": 3.227467811158798,
"grad_norm": 0.25180676579475403,
"learning_rate": 0.00017119670852827588,
"loss": 0.7378,
"step": 1880
},
{
"epoch": 3.236051502145923,
"grad_norm": 0.2704819440841675,
"learning_rate": 0.0001709858057829382,
"loss": 0.7491,
"step": 1885
},
{
"epoch": 3.244635193133047,
"grad_norm": 0.2382296621799469,
"learning_rate": 0.00017077426465659433,
"loss": 0.7433,
"step": 1890
},
{
"epoch": 3.2532188841201717,
"grad_norm": 0.25334346294403076,
"learning_rate": 0.00017056208705165045,
"loss": 0.7505,
"step": 1895
},
{
"epoch": 3.2618025751072963,
"grad_norm": 0.2550380527973175,
"learning_rate": 0.0001703492748762367,
"loss": 0.7531,
"step": 1900
},
{
"epoch": 3.2703862660944205,
"grad_norm": 0.257135808467865,
"learning_rate": 0.00017013583004418993,
"loss": 0.7453,
"step": 1905
},
{
"epoch": 3.278969957081545,
"grad_norm": 0.27000248432159424,
"learning_rate": 0.0001699217544750365,
"loss": 0.7512,
"step": 1910
},
{
"epoch": 3.2875536480686693,
"grad_norm": 0.25268518924713135,
"learning_rate": 0.00016970705009397504,
"loss": 0.7397,
"step": 1915
},
{
"epoch": 3.296137339055794,
"grad_norm": 0.26630303263664246,
"learning_rate": 0.00016949171883185918,
"loss": 0.7436,
"step": 1920
},
{
"epoch": 3.3047210300429186,
"grad_norm": 0.24609267711639404,
"learning_rate": 0.0001692757626251801,
"loss": 0.7402,
"step": 1925
},
{
"epoch": 3.313304721030043,
"grad_norm": 0.26151612401008606,
"learning_rate": 0.00016905918341604922,
"loss": 0.7424,
"step": 1930
},
{
"epoch": 3.3218884120171674,
"grad_norm": 0.2529394030570984,
"learning_rate": 0.00016884198315218055,
"loss": 0.7566,
"step": 1935
},
{
"epoch": 3.3304721030042916,
"grad_norm": 0.2545251250267029,
"learning_rate": 0.0001686241637868734,
"loss": 0.762,
"step": 1940
},
{
"epoch": 3.3390557939914163,
"grad_norm": 0.249998539686203,
"learning_rate": 0.00016840572727899462,
"loss": 0.7399,
"step": 1945
},
{
"epoch": 3.347639484978541,
"grad_norm": 0.24969734251499176,
"learning_rate": 0.0001681866755929612,
"loss": 0.748,
"step": 1950
},
{
"epoch": 3.356223175965665,
"grad_norm": 0.262955904006958,
"learning_rate": 0.00016796701069872238,
"loss": 0.754,
"step": 1955
},
{
"epoch": 3.3648068669527897,
"grad_norm": 0.2457767277956009,
"learning_rate": 0.00016774673457174206,
"loss": 0.7443,
"step": 1960
},
{
"epoch": 3.3733905579399144,
"grad_norm": 0.2644675374031067,
"learning_rate": 0.00016752584919298093,
"loss": 0.7519,
"step": 1965
},
{
"epoch": 3.3819742489270386,
"grad_norm": 0.2620808482170105,
"learning_rate": 0.0001673043565488789,
"loss": 0.7556,
"step": 1970
},
{
"epoch": 3.390557939914163,
"grad_norm": 0.2510511577129364,
"learning_rate": 0.00016708225863133693,
"loss": 0.7556,
"step": 1975
},
{
"epoch": 3.3991416309012874,
"grad_norm": 0.2542615830898285,
"learning_rate": 0.0001668595574376992,
"loss": 0.7502,
"step": 1980
},
{
"epoch": 3.407725321888412,
"grad_norm": 0.25436341762542725,
"learning_rate": 0.0001666362549707354,
"loss": 0.7505,
"step": 1985
},
{
"epoch": 3.4163090128755362,
"grad_norm": 0.24043235182762146,
"learning_rate": 0.00016641235323862236,
"loss": 0.7433,
"step": 1990
},
{
"epoch": 3.424892703862661,
"grad_norm": 0.25933003425598145,
"learning_rate": 0.00016618785425492617,
"loss": 0.7595,
"step": 1995
},
{
"epoch": 3.4334763948497855,
"grad_norm": 0.24922600388526917,
"learning_rate": 0.00016596276003858412,
"loss": 0.7489,
"step": 2000
},
{
"epoch": 3.4420600858369097,
"grad_norm": 0.23722489178180695,
"learning_rate": 0.0001657370726138864,
"loss": 0.7447,
"step": 2005
},
{
"epoch": 3.4506437768240343,
"grad_norm": 0.273787260055542,
"learning_rate": 0.000165510794010458,
"loss": 0.75,
"step": 2010
},
{
"epoch": 3.459227467811159,
"grad_norm": 0.23869618773460388,
"learning_rate": 0.0001652839262632404,
"loss": 0.7463,
"step": 2015
},
{
"epoch": 3.467811158798283,
"grad_norm": 0.264568030834198,
"learning_rate": 0.0001650564714124734,
"loss": 0.7566,
"step": 2020
},
{
"epoch": 3.476394849785408,
"grad_norm": 0.2636789083480835,
"learning_rate": 0.0001648284315036765,
"loss": 0.749,
"step": 2025
},
{
"epoch": 3.484978540772532,
"grad_norm": 0.24988381564617157,
"learning_rate": 0.0001645998085876308,
"loss": 0.7523,
"step": 2030
},
{
"epoch": 3.4935622317596566,
"grad_norm": 0.26084083318710327,
"learning_rate": 0.00016437060472036046,
"loss": 0.7541,
"step": 2035
},
{
"epoch": 3.5021459227467813,
"grad_norm": 0.2548128664493561,
"learning_rate": 0.000164140821963114,
"loss": 0.7593,
"step": 2040
},
{
"epoch": 3.5107296137339055,
"grad_norm": 0.2459845244884491,
"learning_rate": 0.00016391046238234616,
"loss": 0.7485,
"step": 2045
},
{
"epoch": 3.51931330472103,
"grad_norm": 0.26049911975860596,
"learning_rate": 0.00016367952804969895,
"loss": 0.7492,
"step": 2050
},
{
"epoch": 3.5278969957081543,
"grad_norm": 0.2775178551673889,
"learning_rate": 0.00016344802104198324,
"loss": 0.7534,
"step": 2055
},
{
"epoch": 3.536480686695279,
"grad_norm": 0.26411354541778564,
"learning_rate": 0.00016321594344115997,
"loss": 0.7597,
"step": 2060
},
{
"epoch": 3.5450643776824036,
"grad_norm": 0.26565033197402954,
"learning_rate": 0.00016298329733432153,
"loss": 0.7659,
"step": 2065
},
{
"epoch": 3.5536480686695278,
"grad_norm": 0.2576376795768738,
"learning_rate": 0.00016275008481367287,
"loss": 0.7632,
"step": 2070
},
{
"epoch": 3.5622317596566524,
"grad_norm": 0.25273096561431885,
"learning_rate": 0.00016251630797651276,
"loss": 0.7604,
"step": 2075
},
{
"epoch": 3.570815450643777,
"grad_norm": 0.25641995668411255,
"learning_rate": 0.0001622819689252149,
"loss": 0.7575,
"step": 2080
},
{
"epoch": 3.5793991416309012,
"grad_norm": 0.24130167067050934,
"learning_rate": 0.0001620470697672091,
"loss": 0.7512,
"step": 2085
},
{
"epoch": 3.587982832618026,
"grad_norm": 0.2561969757080078,
"learning_rate": 0.00016181161261496216,
"loss": 0.7555,
"step": 2090
},
{
"epoch": 3.59656652360515,
"grad_norm": 0.2512848675251007,
"learning_rate": 0.000161575599585959,
"loss": 0.7556,
"step": 2095
},
{
"epoch": 3.6051502145922747,
"grad_norm": 0.2550983130931854,
"learning_rate": 0.00016133903280268362,
"loss": 0.7673,
"step": 2100
},
{
"epoch": 3.613733905579399,
"grad_norm": 0.2565702795982361,
"learning_rate": 0.00016110191439259997,
"loss": 0.7662,
"step": 2105
},
{
"epoch": 3.6223175965665235,
"grad_norm": 0.24696961045265198,
"learning_rate": 0.00016086424648813273,
"loss": 0.742,
"step": 2110
},
{
"epoch": 3.630901287553648,
"grad_norm": 0.2504982054233551,
"learning_rate": 0.00016062603122664833,
"loss": 0.7514,
"step": 2115
},
{
"epoch": 3.6394849785407724,
"grad_norm": 0.2710771858692169,
"learning_rate": 0.00016038727075043562,
"loss": 0.7681,
"step": 2120
},
{
"epoch": 3.648068669527897,
"grad_norm": 0.26008063554763794,
"learning_rate": 0.0001601479672066865,
"loss": 0.7655,
"step": 2125
},
{
"epoch": 3.6566523605150216,
"grad_norm": 0.25275853276252747,
"learning_rate": 0.00015990812274747692,
"loss": 0.761,
"step": 2130
},
{
"epoch": 3.665236051502146,
"grad_norm": 0.26592886447906494,
"learning_rate": 0.00015966773952974715,
"loss": 0.7529,
"step": 2135
},
{
"epoch": 3.6738197424892705,
"grad_norm": 0.24650226533412933,
"learning_rate": 0.0001594268197152826,
"loss": 0.7538,
"step": 2140
},
{
"epoch": 3.682403433476395,
"grad_norm": 0.25551459193229675,
"learning_rate": 0.00015918536547069435,
"loss": 0.7719,
"step": 2145
},
{
"epoch": 3.6909871244635193,
"grad_norm": 0.24978633224964142,
"learning_rate": 0.0001589433789673997,
"loss": 0.7545,
"step": 2150
},
{
"epoch": 3.699570815450644,
"grad_norm": 0.2471482902765274,
"learning_rate": 0.0001587008623816025,
"loss": 0.749,
"step": 2155
},
{
"epoch": 3.708154506437768,
"grad_norm": 0.24815160036087036,
"learning_rate": 0.00015845781789427377,
"loss": 0.7506,
"step": 2160
},
{
"epoch": 3.7167381974248928,
"grad_norm": 0.26376664638519287,
"learning_rate": 0.00015821424769113193,
"loss": 0.7569,
"step": 2165
},
{
"epoch": 3.725321888412017,
"grad_norm": 0.25356897711753845,
"learning_rate": 0.0001579701539626232,
"loss": 0.7707,
"step": 2170
},
{
"epoch": 3.7339055793991416,
"grad_norm": 0.25035008788108826,
"learning_rate": 0.00015772553890390197,
"loss": 0.76,
"step": 2175
},
{
"epoch": 3.742489270386266,
"grad_norm": 0.2481870949268341,
"learning_rate": 0.0001574804047148109,
"loss": 0.7521,
"step": 2180
},
{
"epoch": 3.7510729613733904,
"grad_norm": 0.25502651929855347,
"learning_rate": 0.00015723475359986127,
"loss": 0.7713,
"step": 2185
},
{
"epoch": 3.759656652360515,
"grad_norm": 0.25871410965919495,
"learning_rate": 0.0001569885877682132,
"loss": 0.7668,
"step": 2190
},
{
"epoch": 3.7682403433476397,
"grad_norm": 0.25364378094673157,
"learning_rate": 0.00015674190943365556,
"loss": 0.754,
"step": 2195
},
{
"epoch": 3.776824034334764,
"grad_norm": 0.2515285909175873,
"learning_rate": 0.0001564947208145863,
"loss": 0.7689,
"step": 2200
},
{
"epoch": 3.7854077253218885,
"grad_norm": 0.24017582833766937,
"learning_rate": 0.00015624702413399231,
"loss": 0.7718,
"step": 2205
},
{
"epoch": 3.7939914163090127,
"grad_norm": 0.26583361625671387,
"learning_rate": 0.00015599882161942966,
"loss": 0.7668,
"step": 2210
},
{
"epoch": 3.8025751072961373,
"grad_norm": 0.2555334270000458,
"learning_rate": 0.00015575011550300323,
"loss": 0.7507,
"step": 2215
},
{
"epoch": 3.8111587982832615,
"grad_norm": 0.2571168839931488,
"learning_rate": 0.000155500908021347,
"loss": 0.752,
"step": 2220
},
{
"epoch": 3.819742489270386,
"grad_norm": 0.26001662015914917,
"learning_rate": 0.0001552512014156037,
"loss": 0.7633,
"step": 2225
},
{
"epoch": 3.828326180257511,
"grad_norm": 0.25432994961738586,
"learning_rate": 0.00015500099793140475,
"loss": 0.758,
"step": 2230
},
{
"epoch": 3.836909871244635,
"grad_norm": 0.24695011973381042,
"learning_rate": 0.00015475029981884992,
"loss": 0.7674,
"step": 2235
},
{
"epoch": 3.8454935622317596,
"grad_norm": 0.24964170157909393,
"learning_rate": 0.00015449910933248743,
"loss": 0.7611,
"step": 2240
},
{
"epoch": 3.8540772532188843,
"grad_norm": 0.2555537223815918,
"learning_rate": 0.00015424742873129324,
"loss": 0.7657,
"step": 2245
},
{
"epoch": 3.8626609442060085,
"grad_norm": 0.27087053656578064,
"learning_rate": 0.00015399526027865107,
"loss": 0.7645,
"step": 2250
},
{
"epoch": 3.871244635193133,
"grad_norm": 0.2602386176586151,
"learning_rate": 0.00015374260624233195,
"loss": 0.7604,
"step": 2255
},
{
"epoch": 3.8798283261802577,
"grad_norm": 0.24609869718551636,
"learning_rate": 0.00015348946889447368,
"loss": 0.7596,
"step": 2260
},
{
"epoch": 3.888412017167382,
"grad_norm": 0.2546113133430481,
"learning_rate": 0.0001532358505115607,
"loss": 0.7594,
"step": 2265
},
{
"epoch": 3.8969957081545066,
"grad_norm": 0.25102177262306213,
"learning_rate": 0.0001529817533744032,
"loss": 0.7606,
"step": 2270
},
{
"epoch": 3.9055793991416308,
"grad_norm": 0.2544861435890198,
"learning_rate": 0.00015272717976811708,
"loss": 0.7535,
"step": 2275
},
{
"epoch": 3.9141630901287554,
"grad_norm": 0.2670022249221802,
"learning_rate": 0.000152472131982103,
"loss": 0.7609,
"step": 2280
},
{
"epoch": 3.9227467811158796,
"grad_norm": 0.2539633810520172,
"learning_rate": 0.00015221661231002605,
"loss": 0.7556,
"step": 2285
},
{
"epoch": 3.9313304721030042,
"grad_norm": 0.2580619156360626,
"learning_rate": 0.00015196062304979497,
"loss": 0.7717,
"step": 2290
},
{
"epoch": 3.939914163090129,
"grad_norm": 0.24921616911888123,
"learning_rate": 0.00015170416650354157,
"loss": 0.7642,
"step": 2295
},
{
"epoch": 3.948497854077253,
"grad_norm": 0.2630159556865692,
"learning_rate": 0.00015144724497760003,
"loss": 0.7522,
"step": 2300
},
{
"epoch": 3.9570815450643777,
"grad_norm": 0.2687895894050598,
"learning_rate": 0.00015118986078248612,
"loss": 0.7623,
"step": 2305
},
{
"epoch": 3.9656652360515023,
"grad_norm": 0.27267009019851685,
"learning_rate": 0.00015093201623287631,
"loss": 0.7755,
"step": 2310
},
{
"epoch": 3.9742489270386265,
"grad_norm": 0.2584875524044037,
"learning_rate": 0.00015067371364758727,
"loss": 0.7582,
"step": 2315
},
{
"epoch": 3.982832618025751,
"grad_norm": 0.25810128450393677,
"learning_rate": 0.00015041495534955467,
"loss": 0.7738,
"step": 2320
},
{
"epoch": 3.991416309012876,
"grad_norm": 0.42070698738098145,
"learning_rate": 0.00015015574366581257,
"loss": 0.7609,
"step": 2325
},
{
"epoch": 4.0,
"grad_norm": 0.2550402581691742,
"learning_rate": 0.0001498960809274722,
"loss": 0.7584,
"step": 2330
},
{
"epoch": 4.0,
"eval_loss": 2.2541472911834717,
"eval_runtime": 0.3945,
"eval_samples_per_second": 15.209,
"eval_steps_per_second": 2.535,
"step": 2330
},
{
"epoch": 4.008583690987124,
"grad_norm": 0.34891819953918457,
"learning_rate": 0.00014963596946970128,
"loss": 0.6641,
"step": 2335
},
{
"epoch": 4.017167381974249,
"grad_norm": 0.26585909724235535,
"learning_rate": 0.0001493754116317029,
"loss": 0.657,
"step": 2340
},
{
"epoch": 4.0257510729613735,
"grad_norm": 0.3037340044975281,
"learning_rate": 0.00014911440975669447,
"loss": 0.6571,
"step": 2345
},
{
"epoch": 4.034334763948498,
"grad_norm": 0.3071967363357544,
"learning_rate": 0.00014885296619188658,
"loss": 0.6607,
"step": 2350
},
{
"epoch": 4.042918454935623,
"grad_norm": 0.28058749437332153,
"learning_rate": 0.00014859108328846204,
"loss": 0.6558,
"step": 2355
},
{
"epoch": 4.051502145922747,
"grad_norm": 0.27246275544166565,
"learning_rate": 0.00014832876340155476,
"loss": 0.6464,
"step": 2360
},
{
"epoch": 4.060085836909871,
"grad_norm": 0.30583012104034424,
"learning_rate": 0.00014806600889022824,
"loss": 0.6602,
"step": 2365
},
{
"epoch": 4.068669527896995,
"grad_norm": 0.2768241763114929,
"learning_rate": 0.0001478028221174548,
"loss": 0.6601,
"step": 2370
},
{
"epoch": 4.07725321888412,
"grad_norm": 0.2900557219982147,
"learning_rate": 0.00014753920545009408,
"loss": 0.6612,
"step": 2375
},
{
"epoch": 4.085836909871245,
"grad_norm": 0.28369641304016113,
"learning_rate": 0.00014727516125887175,
"loss": 0.6497,
"step": 2380
},
{
"epoch": 4.094420600858369,
"grad_norm": 0.2893315553665161,
"learning_rate": 0.0001470106919183582,
"loss": 0.6615,
"step": 2385
},
{
"epoch": 4.103004291845494,
"grad_norm": 0.2940422296524048,
"learning_rate": 0.00014674579980694736,
"loss": 0.6583,
"step": 2390
},
{
"epoch": 4.111587982832618,
"grad_norm": 0.28732019662857056,
"learning_rate": 0.00014648048730683507,
"loss": 0.6717,
"step": 2395
},
{
"epoch": 4.120171673819742,
"grad_norm": 0.31053388118743896,
"learning_rate": 0.0001462147568039977,
"loss": 0.654,
"step": 2400
},
{
"epoch": 4.128755364806867,
"grad_norm": 0.29770082235336304,
"learning_rate": 0.00014594861068817095,
"loss": 0.6641,
"step": 2405
},
{
"epoch": 4.1373390557939915,
"grad_norm": 0.28309738636016846,
"learning_rate": 0.00014568205135282795,
"loss": 0.6715,
"step": 2410
},
{
"epoch": 4.145922746781116,
"grad_norm": 0.29497766494750977,
"learning_rate": 0.00014541508119515808,
"loss": 0.6622,
"step": 2415
},
{
"epoch": 4.154506437768241,
"grad_norm": 0.3100745975971222,
"learning_rate": 0.00014514770261604522,
"loss": 0.6714,
"step": 2420
},
{
"epoch": 4.163090128755365,
"grad_norm": 0.29295915365219116,
"learning_rate": 0.00014487991802004623,
"loss": 0.6617,
"step": 2425
},
{
"epoch": 4.171673819742489,
"grad_norm": 0.2942890524864197,
"learning_rate": 0.0001446117298153693,
"loss": 0.6627,
"step": 2430
},
{
"epoch": 4.180257510729613,
"grad_norm": 0.2925000786781311,
"learning_rate": 0.0001443431404138524,
"loss": 0.6747,
"step": 2435
},
{
"epoch": 4.1888412017167385,
"grad_norm": 0.31883692741394043,
"learning_rate": 0.00014407415223094132,
"loss": 0.6645,
"step": 2440
},
{
"epoch": 4.197424892703863,
"grad_norm": 0.29616445302963257,
"learning_rate": 0.00014380476768566824,
"loss": 0.6653,
"step": 2445
},
{
"epoch": 4.206008583690987,
"grad_norm": 0.2958747148513794,
"learning_rate": 0.00014353498920062987,
"loss": 0.6663,
"step": 2450
},
{
"epoch": 4.214592274678112,
"grad_norm": 0.2944903075695038,
"learning_rate": 0.00014326481920196556,
"loss": 0.6584,
"step": 2455
},
{
"epoch": 4.223175965665236,
"grad_norm": 0.30497610569000244,
"learning_rate": 0.00014299426011933568,
"loss": 0.6734,
"step": 2460
},
{
"epoch": 4.23175965665236,
"grad_norm": 0.2795341908931732,
"learning_rate": 0.0001427233143858996,
"loss": 0.6664,
"step": 2465
},
{
"epoch": 4.240343347639485,
"grad_norm": 0.2952185273170471,
"learning_rate": 0.00014245198443829383,
"loss": 0.675,
"step": 2470
},
{
"epoch": 4.24892703862661,
"grad_norm": 0.29675596952438354,
"learning_rate": 0.0001421802727166103,
"loss": 0.6726,
"step": 2475
},
{
"epoch": 4.257510729613734,
"grad_norm": 0.2960766553878784,
"learning_rate": 0.0001419081816643741,
"loss": 0.6657,
"step": 2480
},
{
"epoch": 4.266094420600858,
"grad_norm": 0.2973078191280365,
"learning_rate": 0.00014163571372852177,
"loss": 0.6781,
"step": 2485
},
{
"epoch": 4.274678111587983,
"grad_norm": 0.293087363243103,
"learning_rate": 0.00014136287135937915,
"loss": 0.6715,
"step": 2490
},
{
"epoch": 4.283261802575107,
"grad_norm": 0.30738070607185364,
"learning_rate": 0.00014108965701063942,
"loss": 0.6692,
"step": 2495
},
{
"epoch": 4.291845493562231,
"grad_norm": 0.29339906573295593,
"learning_rate": 0.0001408160731393409,
"loss": 0.6632,
"step": 2500
},
{
"epoch": 4.3004291845493565,
"grad_norm": 0.3105657696723938,
"learning_rate": 0.00014054212220584525,
"loss": 0.6768,
"step": 2505
},
{
"epoch": 4.309012875536481,
"grad_norm": 0.29471009969711304,
"learning_rate": 0.00014026780667381498,
"loss": 0.6703,
"step": 2510
},
{
"epoch": 4.317596566523605,
"grad_norm": 0.30494722723960876,
"learning_rate": 0.0001399931290101915,
"loss": 0.6725,
"step": 2515
},
{
"epoch": 4.32618025751073,
"grad_norm": 0.2980051040649414,
"learning_rate": 0.00013971809168517298,
"loss": 0.6652,
"step": 2520
},
{
"epoch": 4.334763948497854,
"grad_norm": 0.2986336350440979,
"learning_rate": 0.00013944269717219198,
"loss": 0.6579,
"step": 2525
},
{
"epoch": 4.343347639484978,
"grad_norm": 0.2928684949874878,
"learning_rate": 0.00013916694794789325,
"loss": 0.6797,
"step": 2530
},
{
"epoch": 4.3519313304721035,
"grad_norm": 0.30945298075675964,
"learning_rate": 0.00013889084649211156,
"loss": 0.6781,
"step": 2535
},
{
"epoch": 4.360515021459228,
"grad_norm": 0.29915961623191833,
"learning_rate": 0.0001386143952878493,
"loss": 0.6802,
"step": 2540
},
{
"epoch": 4.369098712446352,
"grad_norm": 0.309627503156662,
"learning_rate": 0.0001383375968212542,
"loss": 0.6728,
"step": 2545
},
{
"epoch": 4.377682403433476,
"grad_norm": 0.292521595954895,
"learning_rate": 0.00013806045358159683,
"loss": 0.6739,
"step": 2550
},
{
"epoch": 4.386266094420601,
"grad_norm": 0.30214038491249084,
"learning_rate": 0.00013778296806124852,
"loss": 0.6821,
"step": 2555
},
{
"epoch": 4.394849785407725,
"grad_norm": 0.30407920479774475,
"learning_rate": 0.0001375051427556586,
"loss": 0.6723,
"step": 2560
},
{
"epoch": 4.4034334763948495,
"grad_norm": 0.3059447109699249,
"learning_rate": 0.00013722698016333218,
"loss": 0.6787,
"step": 2565
},
{
"epoch": 4.412017167381975,
"grad_norm": 0.2976439893245697,
"learning_rate": 0.00013694848278580763,
"loss": 0.6675,
"step": 2570
},
{
"epoch": 4.420600858369099,
"grad_norm": 0.30373451113700867,
"learning_rate": 0.00013666965312763405,
"loss": 0.6743,
"step": 2575
},
{
"epoch": 4.429184549356223,
"grad_norm": 0.2918217182159424,
"learning_rate": 0.00013639049369634876,
"loss": 0.6753,
"step": 2580
},
{
"epoch": 4.437768240343348,
"grad_norm": 0.2900952696800232,
"learning_rate": 0.00013611100700245478,
"loss": 0.6732,
"step": 2585
},
{
"epoch": 4.446351931330472,
"grad_norm": 0.307317852973938,
"learning_rate": 0.00013583119555939815,
"loss": 0.6825,
"step": 2590
},
{
"epoch": 4.454935622317596,
"grad_norm": 0.298483282327652,
"learning_rate": 0.00013555106188354557,
"loss": 0.6744,
"step": 2595
},
{
"epoch": 4.463519313304721,
"grad_norm": 0.3024675250053406,
"learning_rate": 0.00013527060849416144,
"loss": 0.6786,
"step": 2600
},
{
"epoch": 4.472103004291846,
"grad_norm": 0.30732661485671997,
"learning_rate": 0.00013498983791338545,
"loss": 0.6745,
"step": 2605
},
{
"epoch": 4.48068669527897,
"grad_norm": 0.29272985458374023,
"learning_rate": 0.00013470875266620978,
"loss": 0.6782,
"step": 2610
},
{
"epoch": 4.489270386266094,
"grad_norm": 0.3057718873023987,
"learning_rate": 0.00013442735528045643,
"loss": 0.6844,
"step": 2615
},
{
"epoch": 4.497854077253219,
"grad_norm": 0.31071603298187256,
"learning_rate": 0.00013414564828675456,
"loss": 0.6827,
"step": 2620
},
{
"epoch": 4.506437768240343,
"grad_norm": 0.32284069061279297,
"learning_rate": 0.00013386363421851756,
"loss": 0.6809,
"step": 2625
},
{
"epoch": 4.515021459227468,
"grad_norm": 0.3142790198326111,
"learning_rate": 0.00013358131561192046,
"loss": 0.6859,
"step": 2630
},
{
"epoch": 4.523605150214593,
"grad_norm": 0.30643147230148315,
"learning_rate": 0.00013329869500587694,
"loss": 0.6742,
"step": 2635
},
{
"epoch": 4.532188841201717,
"grad_norm": 0.2984931468963623,
"learning_rate": 0.00013301577494201664,
"loss": 0.677,
"step": 2640
},
{
"epoch": 4.540772532188841,
"grad_norm": 0.30939051508903503,
"learning_rate": 0.00013273255796466223,
"loss": 0.6742,
"step": 2645
},
{
"epoch": 4.549356223175966,
"grad_norm": 0.3024001717567444,
"learning_rate": 0.00013244904662080653,
"loss": 0.6718,
"step": 2650
},
{
"epoch": 4.55793991416309,
"grad_norm": 0.294514536857605,
"learning_rate": 0.0001321652434600897,
"loss": 0.6787,
"step": 2655
},
{
"epoch": 4.5665236051502145,
"grad_norm": 0.2984825372695923,
"learning_rate": 0.00013188115103477606,
"loss": 0.6905,
"step": 2660
},
{
"epoch": 4.575107296137339,
"grad_norm": 0.298879474401474,
"learning_rate": 0.00013159677189973152,
"loss": 0.6882,
"step": 2665
},
{
"epoch": 4.583690987124464,
"grad_norm": 0.30825379490852356,
"learning_rate": 0.00013131210861240026,
"loss": 0.6752,
"step": 2670
},
{
"epoch": 4.592274678111588,
"grad_norm": 0.3206503093242645,
"learning_rate": 0.00013102716373278192,
"loss": 0.6827,
"step": 2675
},
{
"epoch": 4.600858369098712,
"grad_norm": 0.3049548268318176,
"learning_rate": 0.00013074193982340847,
"loss": 0.6871,
"step": 2680
},
{
"epoch": 4.609442060085837,
"grad_norm": 0.29572340846061707,
"learning_rate": 0.00013045643944932126,
"loss": 0.6796,
"step": 2685
},
{
"epoch": 4.618025751072961,
"grad_norm": 0.2995782792568207,
"learning_rate": 0.00013017066517804793,
"loss": 0.6827,
"step": 2690
},
{
"epoch": 4.626609442060086,
"grad_norm": 0.31238484382629395,
"learning_rate": 0.00012988461957957922,
"loss": 0.6755,
"step": 2695
},
{
"epoch": 4.63519313304721,
"grad_norm": 0.29801589250564575,
"learning_rate": 0.00012959830522634596,
"loss": 0.6825,
"step": 2700
},
{
"epoch": 4.643776824034335,
"grad_norm": 0.318678081035614,
"learning_rate": 0.000129311724693196,
"loss": 0.6898,
"step": 2705
},
{
"epoch": 4.652360515021459,
"grad_norm": 0.3065158426761627,
"learning_rate": 0.00012902488055737093,
"loss": 0.6765,
"step": 2710
},
{
"epoch": 4.660944206008583,
"grad_norm": 0.30059394240379333,
"learning_rate": 0.00012873777539848283,
"loss": 0.6938,
"step": 2715
},
{
"epoch": 4.669527896995708,
"grad_norm": 0.30598220229148865,
"learning_rate": 0.00012845041179849128,
"loss": 0.686,
"step": 2720
},
{
"epoch": 4.6781115879828326,
"grad_norm": 0.30764105916023254,
"learning_rate": 0.00012816279234168008,
"loss": 0.6886,
"step": 2725
},
{
"epoch": 4.686695278969957,
"grad_norm": 0.302541583776474,
"learning_rate": 0.0001278749196146339,
"loss": 0.6848,
"step": 2730
},
{
"epoch": 4.695278969957082,
"grad_norm": 0.2911517322063446,
"learning_rate": 0.00012758679620621503,
"loss": 0.6859,
"step": 2735
},
{
"epoch": 4.703862660944206,
"grad_norm": 0.2968668043613434,
"learning_rate": 0.00012729842470754032,
"loss": 0.68,
"step": 2740
},
{
"epoch": 4.71244635193133,
"grad_norm": 0.30458763241767883,
"learning_rate": 0.00012700980771195762,
"loss": 0.6785,
"step": 2745
},
{
"epoch": 4.721030042918455,
"grad_norm": 0.3035382032394409,
"learning_rate": 0.00012672094781502252,
"loss": 0.6896,
"step": 2750
},
{
"epoch": 4.7296137339055795,
"grad_norm": 0.2893989682197571,
"learning_rate": 0.00012643184761447512,
"loss": 0.6786,
"step": 2755
},
{
"epoch": 4.738197424892704,
"grad_norm": 0.3313075006008148,
"learning_rate": 0.00012614250971021657,
"loss": 0.6859,
"step": 2760
},
{
"epoch": 4.746781115879829,
"grad_norm": 0.3178950250148773,
"learning_rate": 0.00012585293670428564,
"loss": 0.6855,
"step": 2765
},
{
"epoch": 4.755364806866953,
"grad_norm": 0.28977376222610474,
"learning_rate": 0.00012556313120083546,
"loss": 0.6761,
"step": 2770
},
{
"epoch": 4.763948497854077,
"grad_norm": 0.30826762318611145,
"learning_rate": 0.00012527309580611005,
"loss": 0.677,
"step": 2775
},
{
"epoch": 4.772532188841201,
"grad_norm": 0.3121355473995209,
"learning_rate": 0.0001249828331284207,
"loss": 0.6854,
"step": 2780
},
{
"epoch": 4.781115879828326,
"grad_norm": 0.30394992232322693,
"learning_rate": 0.00012469234577812296,
"loss": 0.6927,
"step": 2785
},
{
"epoch": 4.789699570815451,
"grad_norm": 0.28112301230430603,
"learning_rate": 0.0001244016363675926,
"loss": 0.6917,
"step": 2790
},
{
"epoch": 4.798283261802575,
"grad_norm": 0.2926720678806305,
"learning_rate": 0.00012411070751120254,
"loss": 0.6703,
"step": 2795
},
{
"epoch": 4.8068669527897,
"grad_norm": 0.296567440032959,
"learning_rate": 0.00012381956182529918,
"loss": 0.6831,
"step": 2800
},
{
"epoch": 4.815450643776824,
"grad_norm": 0.29968711733818054,
"learning_rate": 0.00012352820192817877,
"loss": 0.68,
"step": 2805
},
{
"epoch": 4.824034334763948,
"grad_norm": 0.31971994042396545,
"learning_rate": 0.0001232366304400642,
"loss": 0.6936,
"step": 2810
},
{
"epoch": 4.8326180257510725,
"grad_norm": 0.29575350880622864,
"learning_rate": 0.000122944849983081,
"loss": 0.689,
"step": 2815
},
{
"epoch": 4.8412017167381975,
"grad_norm": 0.33436283469200134,
"learning_rate": 0.00012265286318123415,
"loss": 0.6905,
"step": 2820
},
{
"epoch": 4.849785407725322,
"grad_norm": 0.27712202072143555,
"learning_rate": 0.00012236067266038414,
"loss": 0.6856,
"step": 2825
},
{
"epoch": 4.858369098712446,
"grad_norm": 0.32512983679771423,
"learning_rate": 0.00012206828104822363,
"loss": 0.6906,
"step": 2830
},
{
"epoch": 4.866952789699571,
"grad_norm": 0.3128452003002167,
"learning_rate": 0.00012177569097425368,
"loss": 0.6814,
"step": 2835
},
{
"epoch": 4.875536480686695,
"grad_norm": 0.28740525245666504,
"learning_rate": 0.00012148290506976012,
"loss": 0.683,
"step": 2840
},
{
"epoch": 4.884120171673819,
"grad_norm": 0.3094848394393921,
"learning_rate": 0.00012118992596778995,
"loss": 0.6751,
"step": 2845
},
{
"epoch": 4.8927038626609445,
"grad_norm": 0.28997689485549927,
"learning_rate": 0.00012089675630312754,
"loss": 0.6918,
"step": 2850
},
{
"epoch": 4.901287553648069,
"grad_norm": 0.29608073830604553,
"learning_rate": 0.00012060339871227101,
"loss": 0.687,
"step": 2855
},
{
"epoch": 4.909871244635193,
"grad_norm": 0.2982884347438812,
"learning_rate": 0.00012030985583340861,
"loss": 0.6886,
"step": 2860
},
{
"epoch": 4.918454935622318,
"grad_norm": 0.31985583901405334,
"learning_rate": 0.00012001613030639478,
"loss": 0.6787,
"step": 2865
},
{
"epoch": 4.927038626609442,
"grad_norm": 0.31187719106674194,
"learning_rate": 0.00011972222477272663,
"loss": 0.6944,
"step": 2870
},
{
"epoch": 4.935622317596566,
"grad_norm": 0.28702715039253235,
"learning_rate": 0.00011942814187552005,
"loss": 0.6849,
"step": 2875
},
{
"epoch": 4.944206008583691,
"grad_norm": 0.28225070238113403,
"learning_rate": 0.00011913388425948584,
"loss": 0.6791,
"step": 2880
},
{
"epoch": 4.952789699570816,
"grad_norm": 0.3197003901004791,
"learning_rate": 0.00011883945457090633,
"loss": 0.6905,
"step": 2885
},
{
"epoch": 4.96137339055794,
"grad_norm": 0.30299967527389526,
"learning_rate": 0.00011854485545761108,
"loss": 0.6794,
"step": 2890
},
{
"epoch": 4.969957081545064,
"grad_norm": 0.3089299499988556,
"learning_rate": 0.00011825008956895338,
"loss": 0.6815,
"step": 2895
},
{
"epoch": 4.978540772532189,
"grad_norm": 0.31284767389297485,
"learning_rate": 0.0001179551595557864,
"loss": 0.6878,
"step": 2900
},
{
"epoch": 4.987124463519313,
"grad_norm": 0.2950330078601837,
"learning_rate": 0.00011766006807043921,
"loss": 0.6765,
"step": 2905
},
{
"epoch": 4.9957081545064375,
"grad_norm": 0.31097424030303955,
"learning_rate": 0.00011736481776669306,
"loss": 0.6928,
"step": 2910
},
{
"epoch": 4.9991416309012875,
"eval_loss": 2.442364454269409,
"eval_runtime": 0.3946,
"eval_samples_per_second": 15.205,
"eval_steps_per_second": 2.534,
"step": 2912
},
{
"epoch": 5.0042918454935625,
"grad_norm": 0.24456113576889038,
"learning_rate": 0.00011706941129975741,
"loss": 0.6479,
"step": 2915
},
{
"epoch": 5.012875536480687,
"grad_norm": 0.3765704333782196,
"learning_rate": 0.00011677385132624621,
"loss": 0.5981,
"step": 2920
},
{
"epoch": 5.021459227467811,
"grad_norm": 0.30039721727371216,
"learning_rate": 0.00011647814050415378,
"loss": 0.5899,
"step": 2925
},
{
"epoch": 5.030042918454936,
"grad_norm": 0.3503376245498657,
"learning_rate": 0.00011618228149283117,
"loss": 0.586,
"step": 2930
},
{
"epoch": 5.03862660944206,
"grad_norm": 0.315405935049057,
"learning_rate": 0.000115886276952962,
"loss": 0.5842,
"step": 2935
},
{
"epoch": 5.047210300429184,
"grad_norm": 0.3637129068374634,
"learning_rate": 0.00011559012954653865,
"loss": 0.5894,
"step": 2940
},
{
"epoch": 5.055793991416309,
"grad_norm": 0.3133433759212494,
"learning_rate": 0.00011529384193683838,
"loss": 0.5889,
"step": 2945
},
{
"epoch": 5.064377682403434,
"grad_norm": 0.3363387882709503,
"learning_rate": 0.00011499741678839928,
"loss": 0.5992,
"step": 2950
},
{
"epoch": 5.072961373390558,
"grad_norm": 0.33112606406211853,
"learning_rate": 0.00011470085676699627,
"loss": 0.584,
"step": 2955
},
{
"epoch": 5.081545064377682,
"grad_norm": 0.33626094460487366,
"learning_rate": 0.00011440416453961728,
"loss": 0.5907,
"step": 2960
},
{
"epoch": 5.090128755364807,
"grad_norm": 0.32099804282188416,
"learning_rate": 0.00011410734277443915,
"loss": 0.5875,
"step": 2965
},
{
"epoch": 5.098712446351931,
"grad_norm": 0.33036282658576965,
"learning_rate": 0.00011381039414080365,
"loss": 0.5883,
"step": 2970
},
{
"epoch": 5.1072961373390555,
"grad_norm": 0.33885952830314636,
"learning_rate": 0.00011351332130919348,
"loss": 0.5857,
"step": 2975
},
{
"epoch": 5.115879828326181,
"grad_norm": 0.31977617740631104,
"learning_rate": 0.00011321612695120832,
"loss": 0.5855,
"step": 2980
},
{
"epoch": 5.124463519313305,
"grad_norm": 0.3337278664112091,
"learning_rate": 0.00011291881373954065,
"loss": 0.5969,
"step": 2985
},
{
"epoch": 5.133047210300429,
"grad_norm": 0.33656007051467896,
"learning_rate": 0.00011262138434795191,
"loss": 0.5811,
"step": 2990
},
{
"epoch": 5.141630901287554,
"grad_norm": 0.34739845991134644,
"learning_rate": 0.00011232384145124831,
"loss": 0.5932,
"step": 2995
},
{
"epoch": 5.150214592274678,
"grad_norm": 0.33286526799201965,
"learning_rate": 0.0001120261877252568,
"loss": 0.5983,
"step": 3000
},
{
"epoch": 5.1587982832618025,
"grad_norm": 0.3316696584224701,
"learning_rate": 0.00011172842584680107,
"loss": 0.5976,
"step": 3005
},
{
"epoch": 5.167381974248927,
"grad_norm": 0.3135989308357239,
"learning_rate": 0.00011143055849367738,
"loss": 0.5978,
"step": 3010
},
{
"epoch": 5.175965665236052,
"grad_norm": 0.3531875014305115,
"learning_rate": 0.00011113258834463063,
"loss": 0.5965,
"step": 3015
},
{
"epoch": 5.184549356223176,
"grad_norm": 0.3395566940307617,
"learning_rate": 0.00011083451807933008,
"loss": 0.5933,
"step": 3020
},
{
"epoch": 5.1931330472103,
"grad_norm": 0.3231445252895355,
"learning_rate": 0.00011053635037834532,
"loss": 0.6071,
"step": 3025
},
{
"epoch": 5.201716738197425,
"grad_norm": 0.33721092343330383,
"learning_rate": 0.00011023808792312227,
"loss": 0.6049,
"step": 3030
},
{
"epoch": 5.210300429184549,
"grad_norm": 0.32114890217781067,
"learning_rate": 0.00010993973339595896,
"loss": 0.6038,
"step": 3035
},
{
"epoch": 5.218884120171674,
"grad_norm": 0.3454916477203369,
"learning_rate": 0.00010964128947998142,
"loss": 0.5998,
"step": 3040
},
{
"epoch": 5.227467811158799,
"grad_norm": 0.31621354818344116,
"learning_rate": 0.00010934275885911956,
"loss": 0.5977,
"step": 3045
},
{
"epoch": 5.236051502145923,
"grad_norm": 0.3424486517906189,
"learning_rate": 0.00010904414421808303,
"loss": 0.5923,
"step": 3050
},
{
"epoch": 5.244635193133047,
"grad_norm": 0.3184787333011627,
"learning_rate": 0.00010874544824233705,
"loss": 0.5986,
"step": 3055
},
{
"epoch": 5.253218884120171,
"grad_norm": 0.3448573052883148,
"learning_rate": 0.00010844667361807842,
"loss": 0.5931,
"step": 3060
},
{
"epoch": 5.261802575107296,
"grad_norm": 0.314815878868103,
"learning_rate": 0.00010814782303221105,
"loss": 0.6008,
"step": 3065
},
{
"epoch": 5.2703862660944205,
"grad_norm": 0.3566027581691742,
"learning_rate": 0.00010784889917232206,
"loss": 0.585,
"step": 3070
},
{
"epoch": 5.278969957081545,
"grad_norm": 0.31585681438446045,
"learning_rate": 0.0001075499047266576,
"loss": 0.5927,
"step": 3075
},
{
"epoch": 5.28755364806867,
"grad_norm": 0.33874404430389404,
"learning_rate": 0.00010725084238409848,
"loss": 0.6047,
"step": 3080
},
{
"epoch": 5.296137339055794,
"grad_norm": 0.3274739682674408,
"learning_rate": 0.00010695171483413619,
"loss": 0.5919,
"step": 3085
},
{
"epoch": 5.304721030042918,
"grad_norm": 0.3326583802700043,
"learning_rate": 0.00010665252476684864,
"loss": 0.5982,
"step": 3090
},
{
"epoch": 5.313304721030043,
"grad_norm": 0.34962788224220276,
"learning_rate": 0.00010635327487287591,
"loss": 0.5999,
"step": 3095
},
{
"epoch": 5.3218884120171674,
"grad_norm": 0.33242544531822205,
"learning_rate": 0.00010605396784339612,
"loss": 0.6037,
"step": 3100
},
{
"epoch": 5.330472103004292,
"grad_norm": 0.38023149967193604,
"learning_rate": 0.00010575460637010128,
"loss": 0.6068,
"step": 3105
},
{
"epoch": 5.339055793991417,
"grad_norm": 0.3179317116737366,
"learning_rate": 0.00010545519314517291,
"loss": 0.5993,
"step": 3110
},
{
"epoch": 5.347639484978541,
"grad_norm": 0.343841016292572,
"learning_rate": 0.00010515573086125805,
"loss": 0.6029,
"step": 3115
},
{
"epoch": 5.356223175965665,
"grad_norm": 0.3552517890930176,
"learning_rate": 0.00010485622221144484,
"loss": 0.599,
"step": 3120
},
{
"epoch": 5.364806866952789,
"grad_norm": 0.3441089391708374,
"learning_rate": 0.00010455666988923842,
"loss": 0.598,
"step": 3125
},
{
"epoch": 5.373390557939914,
"grad_norm": 0.35372617840766907,
"learning_rate": 0.00010425707658853672,
"loss": 0.597,
"step": 3130
},
{
"epoch": 5.381974248927039,
"grad_norm": 0.33165809512138367,
"learning_rate": 0.00010395744500360612,
"loss": 0.6077,
"step": 3135
},
{
"epoch": 5.390557939914163,
"grad_norm": 0.3485681414604187,
"learning_rate": 0.00010365777782905735,
"loss": 0.5956,
"step": 3140
},
{
"epoch": 5.399141630901288,
"grad_norm": 0.3294559717178345,
"learning_rate": 0.00010335807775982116,
"loss": 0.6076,
"step": 3145
},
{
"epoch": 5.407725321888412,
"grad_norm": 0.3176850378513336,
"learning_rate": 0.00010305834749112421,
"loss": 0.594,
"step": 3150
},
{
"epoch": 5.416309012875536,
"grad_norm": 0.3580548167228699,
"learning_rate": 0.00010275858971846463,
"loss": 0.6029,
"step": 3155
},
{
"epoch": 5.424892703862661,
"grad_norm": 0.32710379362106323,
"learning_rate": 0.00010245880713758793,
"loss": 0.6063,
"step": 3160
},
{
"epoch": 5.4334763948497855,
"grad_norm": 0.37140390276908875,
"learning_rate": 0.00010215900244446279,
"loss": 0.6108,
"step": 3165
},
{
"epoch": 5.44206008583691,
"grad_norm": 0.3271103501319885,
"learning_rate": 0.00010185917833525669,
"loss": 0.6086,
"step": 3170
},
{
"epoch": 5.450643776824034,
"grad_norm": 0.3177226781845093,
"learning_rate": 0.00010155933750631172,
"loss": 0.5976,
"step": 3175
},
{
"epoch": 5.459227467811159,
"grad_norm": 0.33881354331970215,
"learning_rate": 0.00010125948265412033,
"loss": 0.5975,
"step": 3180
},
{
"epoch": 5.467811158798283,
"grad_norm": 0.35869210958480835,
"learning_rate": 0.0001009596164753011,
"loss": 0.6032,
"step": 3185
},
{
"epoch": 5.476394849785407,
"grad_norm": 0.34474968910217285,
"learning_rate": 0.00010065974166657448,
"loss": 0.6014,
"step": 3190
},
{
"epoch": 5.484978540772532,
"grad_norm": 0.3329688012599945,
"learning_rate": 0.00010035986092473847,
"loss": 0.6042,
"step": 3195
},
{
"epoch": 5.493562231759657,
"grad_norm": 0.34499508142471313,
"learning_rate": 0.00010005997694664451,
"loss": 0.5998,
"step": 3200
},
{
"epoch": 5.502145922746781,
"grad_norm": 0.3572762608528137,
"learning_rate": 9.976009242917307e-05,
"loss": 0.6031,
"step": 3205
},
{
"epoch": 5.510729613733906,
"grad_norm": 0.33951064944267273,
"learning_rate": 9.946021006920959e-05,
"loss": 0.6046,
"step": 3210
},
{
"epoch": 5.51931330472103,
"grad_norm": 0.3542894423007965,
"learning_rate": 9.91603325636199e-05,
"loss": 0.6063,
"step": 3215
},
{
"epoch": 5.527896995708154,
"grad_norm": 0.3496350646018982,
"learning_rate": 9.886046260922634e-05,
"loss": 0.6059,
"step": 3220
},
{
"epoch": 5.536480686695279,
"grad_norm": 0.36511626839637756,
"learning_rate": 9.856060290278337e-05,
"loss": 0.6011,
"step": 3225
},
{
"epoch": 5.545064377682404,
"grad_norm": 0.3340092897415161,
"learning_rate": 9.826075614095311e-05,
"loss": 0.6098,
"step": 3230
},
{
"epoch": 5.553648068669528,
"grad_norm": 0.35916590690612793,
"learning_rate": 9.796092502028145e-05,
"loss": 0.6047,
"step": 3235
},
{
"epoch": 5.562231759656653,
"grad_norm": 0.3391835689544678,
"learning_rate": 9.766111223717352e-05,
"loss": 0.61,
"step": 3240
},
{
"epoch": 5.570815450643777,
"grad_norm": 0.3280404508113861,
"learning_rate": 9.736132048786954e-05,
"loss": 0.6088,
"step": 3245
},
{
"epoch": 5.579399141630901,
"grad_norm": 0.31755104660987854,
"learning_rate": 9.706155246842062e-05,
"loss": 0.6051,
"step": 3250
},
{
"epoch": 5.587982832618025,
"grad_norm": 0.33683615922927856,
"learning_rate": 9.676181087466444e-05,
"loss": 0.6071,
"step": 3255
},
{
"epoch": 5.5965665236051505,
"grad_norm": 0.34338319301605225,
"learning_rate": 9.646209840220098e-05,
"loss": 0.6083,
"step": 3260
},
{
"epoch": 5.605150214592275,
"grad_norm": 0.35656723380088806,
"learning_rate": 9.616241774636845e-05,
"loss": 0.6032,
"step": 3265
},
{
"epoch": 5.613733905579399,
"grad_norm": 0.33608657121658325,
"learning_rate": 9.586277160221884e-05,
"loss": 0.6176,
"step": 3270
},
{
"epoch": 5.622317596566524,
"grad_norm": 0.35035207867622375,
"learning_rate": 9.556316266449377e-05,
"loss": 0.6037,
"step": 3275
},
{
"epoch": 5.630901287553648,
"grad_norm": 0.3345491290092468,
"learning_rate": 9.526359362760032e-05,
"loss": 0.6022,
"step": 3280
},
{
"epoch": 5.639484978540772,
"grad_norm": 0.3335653841495514,
"learning_rate": 9.496406718558665e-05,
"loss": 0.6091,
"step": 3285
},
{
"epoch": 5.6480686695278965,
"grad_norm": 0.3456047773361206,
"learning_rate": 9.466458603211796e-05,
"loss": 0.615,
"step": 3290
},
{
"epoch": 5.656652360515022,
"grad_norm": 0.3243827223777771,
"learning_rate": 9.436515286045214e-05,
"loss": 0.6058,
"step": 3295
},
{
"epoch": 5.665236051502146,
"grad_norm": 0.329574853181839,
"learning_rate": 9.406577036341548e-05,
"loss": 0.6054,
"step": 3300
},
{
"epoch": 5.67381974248927,
"grad_norm": 0.3418329954147339,
"learning_rate": 9.376644123337867e-05,
"loss": 0.6098,
"step": 3305
},
{
"epoch": 5.682403433476395,
"grad_norm": 0.36660462617874146,
"learning_rate": 9.346716816223245e-05,
"loss": 0.6187,
"step": 3310
},
{
"epoch": 5.690987124463519,
"grad_norm": 0.3241427540779114,
"learning_rate": 9.316795384136333e-05,
"loss": 0.6121,
"step": 3315
},
{
"epoch": 5.6995708154506435,
"grad_norm": 0.36070680618286133,
"learning_rate": 9.286880096162956e-05,
"loss": 0.6095,
"step": 3320
},
{
"epoch": 5.708154506437769,
"grad_norm": 0.33510082960128784,
"learning_rate": 9.256971221333685e-05,
"loss": 0.6126,
"step": 3325
},
{
"epoch": 5.716738197424893,
"grad_norm": 0.3640751838684082,
"learning_rate": 9.227069028621406e-05,
"loss": 0.6199,
"step": 3330
},
{
"epoch": 5.725321888412017,
"grad_norm": 0.34373095631599426,
"learning_rate": 9.197173786938926e-05,
"loss": 0.6091,
"step": 3335
},
{
"epoch": 5.733905579399142,
"grad_norm": 0.33689820766448975,
"learning_rate": 9.167285765136533e-05,
"loss": 0.6032,
"step": 3340
},
{
"epoch": 5.742489270386266,
"grad_norm": 0.33981624245643616,
"learning_rate": 9.137405231999593e-05,
"loss": 0.6069,
"step": 3345
},
{
"epoch": 5.75107296137339,
"grad_norm": 0.34156525135040283,
"learning_rate": 9.107532456246114e-05,
"loss": 0.6006,
"step": 3350
},
{
"epoch": 5.7596566523605155,
"grad_norm": 0.34921032190322876,
"learning_rate": 9.077667706524348e-05,
"loss": 0.6165,
"step": 3355
},
{
"epoch": 5.76824034334764,
"grad_norm": 0.3511927127838135,
"learning_rate": 9.047811251410376e-05,
"loss": 0.6149,
"step": 3360
},
{
"epoch": 5.776824034334764,
"grad_norm": 0.3529740571975708,
"learning_rate": 9.01796335940567e-05,
"loss": 0.6088,
"step": 3365
},
{
"epoch": 5.785407725321888,
"grad_norm": 0.33018723130226135,
"learning_rate": 8.9881242989347e-05,
"loss": 0.6089,
"step": 3370
},
{
"epoch": 5.793991416309013,
"grad_norm": 0.33075249195098877,
"learning_rate": 8.95829433834252e-05,
"loss": 0.6107,
"step": 3375
},
{
"epoch": 5.802575107296137,
"grad_norm": 0.35009533166885376,
"learning_rate": 8.928473745892339e-05,
"loss": 0.6092,
"step": 3380
},
{
"epoch": 5.8111587982832615,
"grad_norm": 0.337410569190979,
"learning_rate": 8.898662789763115e-05,
"loss": 0.6049,
"step": 3385
},
{
"epoch": 5.819742489270387,
"grad_norm": 0.332481324672699,
"learning_rate": 8.868861738047158e-05,
"loss": 0.625,
"step": 3390
},
{
"epoch": 5.828326180257511,
"grad_norm": 0.34015268087387085,
"learning_rate": 8.839070858747697e-05,
"loss": 0.615,
"step": 3395
},
{
"epoch": 5.836909871244635,
"grad_norm": 0.35073375701904297,
"learning_rate": 8.809290419776488e-05,
"loss": 0.6038,
"step": 3400
},
{
"epoch": 5.845493562231759,
"grad_norm": 0.3357756733894348,
"learning_rate": 8.779520688951383e-05,
"loss": 0.6077,
"step": 3405
},
{
"epoch": 5.854077253218884,
"grad_norm": 0.33854493498802185,
"learning_rate": 8.749761933993945e-05,
"loss": 0.6058,
"step": 3410
},
{
"epoch": 5.8626609442060085,
"grad_norm": 0.3343227505683899,
"learning_rate": 8.720014422527034e-05,
"loss": 0.6113,
"step": 3415
},
{
"epoch": 5.871244635193133,
"grad_norm": 0.35862088203430176,
"learning_rate": 8.690278422072384e-05,
"loss": 0.606,
"step": 3420
},
{
"epoch": 5.879828326180258,
"grad_norm": 0.37902429699897766,
"learning_rate": 8.660554200048215e-05,
"loss": 0.6108,
"step": 3425
},
{
"epoch": 5.888412017167382,
"grad_norm": 0.3685941994190216,
"learning_rate": 8.630842023766831e-05,
"loss": 0.6138,
"step": 3430
},
{
"epoch": 5.896995708154506,
"grad_norm": 0.3437183201313019,
"learning_rate": 8.601142160432193e-05,
"loss": 0.6156,
"step": 3435
},
{
"epoch": 5.905579399141631,
"grad_norm": 0.34756121039390564,
"learning_rate": 8.571454877137539e-05,
"loss": 0.6076,
"step": 3440
},
{
"epoch": 5.914163090128755,
"grad_norm": 0.33176571130752563,
"learning_rate": 8.541780440862977e-05,
"loss": 0.6065,
"step": 3445
},
{
"epoch": 5.92274678111588,
"grad_norm": 0.3526177704334259,
"learning_rate": 8.512119118473067e-05,
"loss": 0.6125,
"step": 3450
},
{
"epoch": 5.931330472103005,
"grad_norm": 0.3659914433956146,
"learning_rate": 8.482471176714454e-05,
"loss": 0.6095,
"step": 3455
},
{
"epoch": 5.939914163090129,
"grad_norm": 0.34757936000823975,
"learning_rate": 8.45283688221344e-05,
"loss": 0.6134,
"step": 3460
},
{
"epoch": 5.948497854077253,
"grad_norm": 0.34589987993240356,
"learning_rate": 8.423216501473585e-05,
"loss": 0.6074,
"step": 3465
},
{
"epoch": 5.957081545064378,
"grad_norm": 0.34556832909584045,
"learning_rate": 8.393610300873345e-05,
"loss": 0.609,
"step": 3470
},
{
"epoch": 5.965665236051502,
"grad_norm": 0.3431447148323059,
"learning_rate": 8.364018546663634e-05,
"loss": 0.6157,
"step": 3475
},
{
"epoch": 5.9742489270386265,
"grad_norm": 0.35311102867126465,
"learning_rate": 8.334441504965455e-05,
"loss": 0.6143,
"step": 3480
},
{
"epoch": 5.982832618025751,
"grad_norm": 0.3422008156776428,
"learning_rate": 8.304879441767504e-05,
"loss": 0.6046,
"step": 3485
},
{
"epoch": 5.991416309012876,
"grad_norm": 0.3697339594364166,
"learning_rate": 8.275332622923769e-05,
"loss": 0.6141,
"step": 3490
},
{
"epoch": 6.0,
"grad_norm": 0.332302987575531,
"learning_rate": 8.245801314151139e-05,
"loss": 0.6102,
"step": 3495
},
{
"epoch": 6.0,
"eval_loss": 2.7089035511016846,
"eval_runtime": 0.3939,
"eval_samples_per_second": 15.232,
"eval_steps_per_second": 2.539,
"step": 3495
},
{
"epoch": 6.008583690987124,
"grad_norm": 0.3818419575691223,
"learning_rate": 8.216285781027036e-05,
"loss": 0.5334,
"step": 3500
},
{
"epoch": 6.017167381974249,
"grad_norm": 0.35173356533050537,
"learning_rate": 8.186786288986992e-05,
"loss": 0.5409,
"step": 3505
},
{
"epoch": 6.0257510729613735,
"grad_norm": 0.348001092672348,
"learning_rate": 8.157303103322296e-05,
"loss": 0.5294,
"step": 3510
},
{
"epoch": 6.034334763948498,
"grad_norm": 0.38012921810150146,
"learning_rate": 8.127836489177584e-05,
"loss": 0.5289,
"step": 3515
},
{
"epoch": 6.042918454935623,
"grad_norm": 0.3508910536766052,
"learning_rate": 8.098386711548458e-05,
"loss": 0.5375,
"step": 3520
},
{
"epoch": 6.051502145922747,
"grad_norm": 0.3329296410083771,
"learning_rate": 8.068954035279121e-05,
"loss": 0.5298,
"step": 3525
},
{
"epoch": 6.060085836909871,
"grad_norm": 0.3630905747413635,
"learning_rate": 8.039538725059976e-05,
"loss": 0.5349,
"step": 3530
},
{
"epoch": 6.068669527896995,
"grad_norm": 0.32074230909347534,
"learning_rate": 8.010141045425244e-05,
"loss": 0.5376,
"step": 3535
},
{
"epoch": 6.07725321888412,
"grad_norm": 0.3476736545562744,
"learning_rate": 7.980761260750607e-05,
"loss": 0.5279,
"step": 3540
},
{
"epoch": 6.085836909871245,
"grad_norm": 0.340426504611969,
"learning_rate": 7.951399635250806e-05,
"loss": 0.5323,
"step": 3545
},
{
"epoch": 6.094420600858369,
"grad_norm": 0.3367501497268677,
"learning_rate": 7.922056432977275e-05,
"loss": 0.5486,
"step": 3550
},
{
"epoch": 6.103004291845494,
"grad_norm": 0.3827115595340729,
"learning_rate": 7.892731917815774e-05,
"loss": 0.528,
"step": 3555
},
{
"epoch": 6.111587982832618,
"grad_norm": 0.33026212453842163,
"learning_rate": 7.863426353484002e-05,
"loss": 0.5303,
"step": 3560
},
{
"epoch": 6.120171673819742,
"grad_norm": 0.3674776256084442,
"learning_rate": 7.834140003529238e-05,
"loss": 0.5355,
"step": 3565
},
{
"epoch": 6.128755364806867,
"grad_norm": 0.3238033056259155,
"learning_rate": 7.804873131325954e-05,
"loss": 0.5382,
"step": 3570
},
{
"epoch": 6.1373390557939915,
"grad_norm": 0.3594464063644409,
"learning_rate": 7.775626000073463e-05,
"loss": 0.5319,
"step": 3575
},
{
"epoch": 6.145922746781116,
"grad_norm": 0.3666881322860718,
"learning_rate": 7.74639887279355e-05,
"loss": 0.5387,
"step": 3580
},
{
"epoch": 6.154506437768241,
"grad_norm": 0.34055057168006897,
"learning_rate": 7.7171920123281e-05,
"loss": 0.5296,
"step": 3585
},
{
"epoch": 6.163090128755365,
"grad_norm": 0.3717866837978363,
"learning_rate": 7.688005681336729e-05,
"loss": 0.5371,
"step": 3590
},
{
"epoch": 6.171673819742489,
"grad_norm": 0.3607926368713379,
"learning_rate": 7.658840142294444e-05,
"loss": 0.5446,
"step": 3595
},
{
"epoch": 6.180257510729613,
"grad_norm": 0.3458268940448761,
"learning_rate": 7.629695657489257e-05,
"loss": 0.5393,
"step": 3600
},
{
"epoch": 6.1888412017167385,
"grad_norm": 0.3347238600254059,
"learning_rate": 7.600572489019842e-05,
"loss": 0.5254,
"step": 3605
},
{
"epoch": 6.197424892703863,
"grad_norm": 0.3612549901008606,
"learning_rate": 7.571470898793173e-05,
"loss": 0.5275,
"step": 3610
},
{
"epoch": 6.206008583690987,
"grad_norm": 0.3335767686367035,
"learning_rate": 7.542391148522168e-05,
"loss": 0.5304,
"step": 3615
},
{
"epoch": 6.214592274678112,
"grad_norm": 0.3704369068145752,
"learning_rate": 7.513333499723343e-05,
"loss": 0.5404,
"step": 3620
},
{
"epoch": 6.223175965665236,
"grad_norm": 0.3659282922744751,
"learning_rate": 7.484298213714442e-05,
"loss": 0.5357,
"step": 3625
},
{
"epoch": 6.23175965665236,
"grad_norm": 0.34985628724098206,
"learning_rate": 7.455285551612105e-05,
"loss": 0.5411,
"step": 3630
},
{
"epoch": 6.240343347639485,
"grad_norm": 0.3511213958263397,
"learning_rate": 7.426295774329524e-05,
"loss": 0.5361,
"step": 3635
},
{
"epoch": 6.24892703862661,
"grad_norm": 0.36643365025520325,
"learning_rate": 7.397329142574063e-05,
"loss": 0.5256,
"step": 3640
},
{
"epoch": 6.257510729613734,
"grad_norm": 0.3581530451774597,
"learning_rate": 7.36838591684495e-05,
"loss": 0.5364,
"step": 3645
},
{
"epoch": 6.266094420600858,
"grad_norm": 0.3588225245475769,
"learning_rate": 7.339466357430928e-05,
"loss": 0.5433,
"step": 3650
},
{
"epoch": 6.274678111587983,
"grad_norm": 0.3423435389995575,
"learning_rate": 7.310570724407892e-05,
"loss": 0.5481,
"step": 3655
},
{
"epoch": 6.283261802575107,
"grad_norm": 0.37222719192504883,
"learning_rate": 7.281699277636572e-05,
"loss": 0.5418,
"step": 3660
},
{
"epoch": 6.291845493562231,
"grad_norm": 0.37203842401504517,
"learning_rate": 7.252852276760193e-05,
"loss": 0.5401,
"step": 3665
},
{
"epoch": 6.3004291845493565,
"grad_norm": 0.3703368306159973,
"learning_rate": 7.224029981202122e-05,
"loss": 0.5381,
"step": 3670
},
{
"epoch": 6.309012875536481,
"grad_norm": 0.36878761649131775,
"learning_rate": 7.195232650163575e-05,
"loss": 0.5257,
"step": 3675
},
{
"epoch": 6.317596566523605,
"grad_norm": 0.3529357314109802,
"learning_rate": 7.16646054262123e-05,
"loss": 0.542,
"step": 3680
},
{
"epoch": 6.32618025751073,
"grad_norm": 0.3673364520072937,
"learning_rate": 7.137713917324945e-05,
"loss": 0.5351,
"step": 3685
},
{
"epoch": 6.334763948497854,
"grad_norm": 0.3675953149795532,
"learning_rate": 7.108993032795418e-05,
"loss": 0.5455,
"step": 3690
},
{
"epoch": 6.343347639484978,
"grad_norm": 0.3677637279033661,
"learning_rate": 7.080298147321844e-05,
"loss": 0.5352,
"step": 3695
},
{
"epoch": 6.3519313304721035,
"grad_norm": 0.36320698261260986,
"learning_rate": 7.051629518959614e-05,
"loss": 0.5445,
"step": 3700
},
{
"epoch": 6.360515021459228,
"grad_norm": 0.36490681767463684,
"learning_rate": 7.022987405527997e-05,
"loss": 0.5379,
"step": 3705
},
{
"epoch": 6.369098712446352,
"grad_norm": 0.37636715173721313,
"learning_rate": 6.994372064607788e-05,
"loss": 0.5416,
"step": 3710
},
{
"epoch": 6.377682403433476,
"grad_norm": 0.3601493239402771,
"learning_rate": 6.96578375353903e-05,
"loss": 0.544,
"step": 3715
},
{
"epoch": 6.386266094420601,
"grad_norm": 0.3615502715110779,
"learning_rate": 6.93722272941869e-05,
"loss": 0.5368,
"step": 3720
},
{
"epoch": 6.394849785407725,
"grad_norm": 0.3495427668094635,
"learning_rate": 6.908689249098321e-05,
"loss": 0.5414,
"step": 3725
},
{
"epoch": 6.4034334763948495,
"grad_norm": 0.36156705021858215,
"learning_rate": 6.880183569181795e-05,
"loss": 0.5379,
"step": 3730
},
{
"epoch": 6.412017167381975,
"grad_norm": 0.38569867610931396,
"learning_rate": 6.85170594602296e-05,
"loss": 0.5445,
"step": 3735
},
{
"epoch": 6.420600858369099,
"grad_norm": 0.37031883001327515,
"learning_rate": 6.823256635723343e-05,
"loss": 0.55,
"step": 3740
},
{
"epoch": 6.429184549356223,
"grad_norm": 0.33991584181785583,
"learning_rate": 6.794835894129865e-05,
"loss": 0.5427,
"step": 3745
},
{
"epoch": 6.437768240343348,
"grad_norm": 0.35758277773857117,
"learning_rate": 6.766443976832517e-05,
"loss": 0.5393,
"step": 3750
},
{
"epoch": 6.446351931330472,
"grad_norm": 0.37792807817459106,
"learning_rate": 6.738081139162072e-05,
"loss": 0.5466,
"step": 3755
},
{
"epoch": 6.454935622317596,
"grad_norm": 0.3743073046207428,
"learning_rate": 6.709747636187789e-05,
"loss": 0.539,
"step": 3760
},
{
"epoch": 6.463519313304721,
"grad_norm": 0.397246390581131,
"learning_rate": 6.68144372271512e-05,
"loss": 0.5417,
"step": 3765
},
{
"epoch": 6.472103004291846,
"grad_norm": 0.3632808029651642,
"learning_rate": 6.653169653283406e-05,
"loss": 0.5403,
"step": 3770
},
{
"epoch": 6.48068669527897,
"grad_norm": 0.3583681583404541,
"learning_rate": 6.624925682163614e-05,
"loss": 0.5462,
"step": 3775
},
{
"epoch": 6.489270386266094,
"grad_norm": 0.3529057502746582,
"learning_rate": 6.59671206335602e-05,
"loss": 0.5425,
"step": 3780
},
{
"epoch": 6.497854077253219,
"grad_norm": 0.3624202311038971,
"learning_rate": 6.568529050587953e-05,
"loss": 0.5462,
"step": 3785
},
{
"epoch": 6.506437768240343,
"grad_norm": 0.3762538433074951,
"learning_rate": 6.540376897311489e-05,
"loss": 0.538,
"step": 3790
},
{
"epoch": 6.515021459227468,
"grad_norm": 0.37663355469703674,
"learning_rate": 6.512255856701177e-05,
"loss": 0.5432,
"step": 3795
},
{
"epoch": 6.523605150214593,
"grad_norm": 0.37055134773254395,
"learning_rate": 6.484166181651785e-05,
"loss": 0.5405,
"step": 3800
},
{
"epoch": 6.532188841201717,
"grad_norm": 0.3634713590145111,
"learning_rate": 6.456108124775999e-05,
"loss": 0.5442,
"step": 3805
},
{
"epoch": 6.540772532188841,
"grad_norm": 0.3575882017612457,
"learning_rate": 6.428081938402149e-05,
"loss": 0.5395,
"step": 3810
},
{
"epoch": 6.549356223175966,
"grad_norm": 0.3856394290924072,
"learning_rate": 6.400087874571973e-05,
"loss": 0.5417,
"step": 3815
},
{
"epoch": 6.55793991416309,
"grad_norm": 0.3542211651802063,
"learning_rate": 6.372126185038313e-05,
"loss": 0.5337,
"step": 3820
},
{
"epoch": 6.5665236051502145,
"grad_norm": 0.3997708559036255,
"learning_rate": 6.344197121262868e-05,
"loss": 0.5431,
"step": 3825
},
{
"epoch": 6.575107296137339,
"grad_norm": 0.3704608082771301,
"learning_rate": 6.316300934413935e-05,
"loss": 0.5356,
"step": 3830
},
{
"epoch": 6.583690987124464,
"grad_norm": 0.3824236989021301,
"learning_rate": 6.288437875364141e-05,
"loss": 0.5406,
"step": 3835
},
{
"epoch": 6.592274678111588,
"grad_norm": 0.3561914563179016,
"learning_rate": 6.260608194688206e-05,
"loss": 0.5405,
"step": 3840
},
{
"epoch": 6.600858369098712,
"grad_norm": 0.3756065368652344,
"learning_rate": 6.232812142660658e-05,
"loss": 0.5365,
"step": 3845
},
{
"epoch": 6.609442060085837,
"grad_norm": 0.3645598292350769,
"learning_rate": 6.205049969253605e-05,
"loss": 0.5358,
"step": 3850
},
{
"epoch": 6.618025751072961,
"grad_norm": 0.37949660420417786,
"learning_rate": 6.17732192413449e-05,
"loss": 0.5452,
"step": 3855
},
{
"epoch": 6.626609442060086,
"grad_norm": 0.36608970165252686,
"learning_rate": 6.149628256663827e-05,
"loss": 0.545,
"step": 3860
},
{
"epoch": 6.63519313304721,
"grad_norm": 0.3485977351665497,
"learning_rate": 6.121969215892972e-05,
"loss": 0.5445,
"step": 3865
},
{
"epoch": 6.643776824034335,
"grad_norm": 0.38575315475463867,
"learning_rate": 6.0943450505618917e-05,
"loss": 0.5448,
"step": 3870
},
{
"epoch": 6.652360515021459,
"grad_norm": 0.3650740087032318,
"learning_rate": 6.066756009096896e-05,
"loss": 0.5489,
"step": 3875
},
{
"epoch": 6.660944206008583,
"grad_norm": 0.3552764058113098,
"learning_rate": 6.039202339608432e-05,
"loss": 0.5461,
"step": 3880
},
{
"epoch": 6.669527896995708,
"grad_norm": 0.3777913749217987,
"learning_rate": 6.01168428988885e-05,
"loss": 0.5437,
"step": 3885
},
{
"epoch": 6.6781115879828326,
"grad_norm": 0.37254467606544495,
"learning_rate": 5.9842021074101605e-05,
"loss": 0.5351,
"step": 3890
},
{
"epoch": 6.686695278969957,
"grad_norm": 0.36322537064552307,
"learning_rate": 5.956756039321825e-05,
"loss": 0.5503,
"step": 3895
},
{
"epoch": 6.695278969957082,
"grad_norm": 0.3798597753047943,
"learning_rate": 5.929346332448511e-05,
"loss": 0.5336,
"step": 3900
},
{
"epoch": 6.703862660944206,
"grad_norm": 0.3622066080570221,
"learning_rate": 5.901973233287901e-05,
"loss": 0.5472,
"step": 3905
},
{
"epoch": 6.71244635193133,
"grad_norm": 0.37123680114746094,
"learning_rate": 5.874636988008457e-05,
"loss": 0.5376,
"step": 3910
},
{
"epoch": 6.721030042918455,
"grad_norm": 0.3789604902267456,
"learning_rate": 5.847337842447209e-05,
"loss": 0.5472,
"step": 3915
},
{
"epoch": 6.7296137339055795,
"grad_norm": 0.3710649311542511,
"learning_rate": 5.820076042107545e-05,
"loss": 0.5459,
"step": 3920
},
{
"epoch": 6.738197424892704,
"grad_norm": 0.41028717160224915,
"learning_rate": 5.792851832157014e-05,
"loss": 0.5415,
"step": 3925
},
{
"epoch": 6.746781115879829,
"grad_norm": 0.3710199296474457,
"learning_rate": 5.765665457425102e-05,
"loss": 0.5376,
"step": 3930
},
{
"epoch": 6.755364806866953,
"grad_norm": 0.37828171253204346,
"learning_rate": 5.7385171624010346e-05,
"loss": 0.5474,
"step": 3935
},
{
"epoch": 6.763948497854077,
"grad_norm": 0.35286852717399597,
"learning_rate": 5.711407191231602e-05,
"loss": 0.5435,
"step": 3940
},
{
"epoch": 6.772532188841201,
"grad_norm": 0.39667871594429016,
"learning_rate": 5.684335787718932e-05,
"loss": 0.5471,
"step": 3945
},
{
"epoch": 6.781115879828326,
"grad_norm": 0.3569738268852234,
"learning_rate": 5.657303195318311e-05,
"loss": 0.5362,
"step": 3950
},
{
"epoch": 6.789699570815451,
"grad_norm": 0.3528185784816742,
"learning_rate": 5.630309657135997e-05,
"loss": 0.5383,
"step": 3955
},
{
"epoch": 6.798283261802575,
"grad_norm": 0.3892223834991455,
"learning_rate": 5.6033554159270294e-05,
"loss": 0.5446,
"step": 3960
},
{
"epoch": 6.8068669527897,
"grad_norm": 0.3695877492427826,
"learning_rate": 5.576440714093046e-05,
"loss": 0.5488,
"step": 3965
},
{
"epoch": 6.815450643776824,
"grad_norm": 0.3762911856174469,
"learning_rate": 5.549565793680105e-05,
"loss": 0.5398,
"step": 3970
},
{
"epoch": 6.824034334763948,
"grad_norm": 0.36472398042678833,
"learning_rate": 5.522730896376506e-05,
"loss": 0.5457,
"step": 3975
},
{
"epoch": 6.8326180257510725,
"grad_norm": 0.37586814165115356,
"learning_rate": 5.495936263510617e-05,
"loss": 0.5396,
"step": 3980
},
{
"epoch": 6.8412017167381975,
"grad_norm": 0.37699511647224426,
"learning_rate": 5.4691821360487086e-05,
"loss": 0.5394,
"step": 3985
},
{
"epoch": 6.849785407725322,
"grad_norm": 0.3839593529701233,
"learning_rate": 5.4424687545927776e-05,
"loss": 0.5429,
"step": 3990
},
{
"epoch": 6.858369098712446,
"grad_norm": 0.3746870756149292,
"learning_rate": 5.415796359378393e-05,
"loss": 0.538,
"step": 3995
},
{
"epoch": 6.866952789699571,
"grad_norm": 0.36862820386886597,
"learning_rate": 5.389165190272527e-05,
"loss": 0.5433,
"step": 4000
},
{
"epoch": 6.875536480686695,
"grad_norm": 0.3722948729991913,
"learning_rate": 5.362575486771414e-05,
"loss": 0.5533,
"step": 4005
},
{
"epoch": 6.884120171673819,
"grad_norm": 0.3654981255531311,
"learning_rate": 5.3360274879983654e-05,
"loss": 0.5345,
"step": 4010
},
{
"epoch": 6.8927038626609445,
"grad_norm": 0.35018405318260193,
"learning_rate": 5.3095214327016474e-05,
"loss": 0.5484,
"step": 4015
},
{
"epoch": 6.901287553648069,
"grad_norm": 0.37176111340522766,
"learning_rate": 5.283057559252341e-05,
"loss": 0.5422,
"step": 4020
},
{
"epoch": 6.909871244635193,
"grad_norm": 0.37949976325035095,
"learning_rate": 5.256636105642154e-05,
"loss": 0.5501,
"step": 4025
},
{
"epoch": 6.918454935622318,
"grad_norm": 0.35853028297424316,
"learning_rate": 5.2302573094813266e-05,
"loss": 0.5427,
"step": 4030
},
{
"epoch": 6.927038626609442,
"grad_norm": 0.370491623878479,
"learning_rate": 5.2039214079964836e-05,
"loss": 0.5426,
"step": 4035
},
{
"epoch": 6.935622317596566,
"grad_norm": 0.3726717233657837,
"learning_rate": 5.177628638028472e-05,
"loss": 0.5447,
"step": 4040
},
{
"epoch": 6.944206008583691,
"grad_norm": 0.37951403856277466,
"learning_rate": 5.1513792360302696e-05,
"loss": 0.5369,
"step": 4045
},
{
"epoch": 6.952789699570816,
"grad_norm": 0.3583022654056549,
"learning_rate": 5.12517343806485e-05,
"loss": 0.5586,
"step": 4050
},
{
"epoch": 6.96137339055794,
"grad_norm": 0.37766262888908386,
"learning_rate": 5.099011479803025e-05,
"loss": 0.537,
"step": 4055
},
{
"epoch": 6.969957081545064,
"grad_norm": 0.36432990431785583,
"learning_rate": 5.0728935965213834e-05,
"loss": 0.5462,
"step": 4060
},
{
"epoch": 6.978540772532189,
"grad_norm": 0.37999647855758667,
"learning_rate": 5.0468200231001286e-05,
"loss": 0.5474,
"step": 4065
},
{
"epoch": 6.987124463519313,
"grad_norm": 0.3890798091888428,
"learning_rate": 5.020790994020972e-05,
"loss": 0.5395,
"step": 4070
},
{
"epoch": 6.9957081545064375,
"grad_norm": 0.3609655797481537,
"learning_rate": 4.994806743365057e-05,
"loss": 0.5466,
"step": 4075
},
{
"epoch": 6.9991416309012875,
"eval_loss": 3.0554237365722656,
"eval_runtime": 0.3942,
"eval_samples_per_second": 15.219,
"eval_steps_per_second": 2.537,
"step": 4077
},
{
"epoch": 7.0042918454935625,
"grad_norm": 0.2814909517765045,
"learning_rate": 4.96886750481082e-05,
"loss": 0.5172,
"step": 4080
},
{
"epoch": 7.012875536480687,
"grad_norm": 0.39267781376838684,
"learning_rate": 4.942973511631889e-05,
"loss": 0.4929,
"step": 4085
},
{
"epoch": 7.021459227467811,
"grad_norm": 0.35972246527671814,
"learning_rate": 4.9171249966950175e-05,
"loss": 0.4891,
"step": 4090
},
{
"epoch": 7.030042918454936,
"grad_norm": 0.31733205914497375,
"learning_rate": 4.8913221924579554e-05,
"loss": 0.4899,
"step": 4095
},
{
"epoch": 7.03862660944206,
"grad_norm": 0.37763702869415283,
"learning_rate": 4.8655653309673776e-05,
"loss": 0.4899,
"step": 4100
},
{
"epoch": 7.047210300429184,
"grad_norm": 0.39104217290878296,
"learning_rate": 4.839854643856792e-05,
"loss": 0.4843,
"step": 4105
},
{
"epoch": 7.055793991416309,
"grad_norm": 0.3271881937980652,
"learning_rate": 4.814190362344454e-05,
"loss": 0.4903,
"step": 4110
},
{
"epoch": 7.064377682403434,
"grad_norm": 0.35317346453666687,
"learning_rate": 4.788572717231293e-05,
"loss": 0.4916,
"step": 4115
},
{
"epoch": 7.072961373390558,
"grad_norm": 0.37199559807777405,
"learning_rate": 4.763001938898832e-05,
"loss": 0.4865,
"step": 4120
},
{
"epoch": 7.081545064377682,
"grad_norm": 0.36147797107696533,
"learning_rate": 4.7374782573071176e-05,
"loss": 0.4884,
"step": 4125
},
{
"epoch": 7.090128755364807,
"grad_norm": 0.3491626977920532,
"learning_rate": 4.712001901992652e-05,
"loss": 0.4926,
"step": 4130
},
{
"epoch": 7.098712446351931,
"grad_norm": 0.36010846495628357,
"learning_rate": 4.686573102066326e-05,
"loss": 0.4942,
"step": 4135
},
{
"epoch": 7.1072961373390555,
"grad_norm": 0.34614065289497375,
"learning_rate": 4.661192086211366e-05,
"loss": 0.4888,
"step": 4140
},
{
"epoch": 7.115879828326181,
"grad_norm": 0.37029707431793213,
"learning_rate": 4.6358590826812664e-05,
"loss": 0.493,
"step": 4145
},
{
"epoch": 7.124463519313305,
"grad_norm": 0.35328662395477295,
"learning_rate": 4.610574319297748e-05,
"loss": 0.4949,
"step": 4150
},
{
"epoch": 7.133047210300429,
"grad_norm": 0.33476021885871887,
"learning_rate": 4.585338023448702e-05,
"loss": 0.488,
"step": 4155
},
{
"epoch": 7.141630901287554,
"grad_norm": 0.36379748582839966,
"learning_rate": 4.560150422086147e-05,
"loss": 0.4927,
"step": 4160
},
{
"epoch": 7.150214592274678,
"grad_norm": 0.3896268308162689,
"learning_rate": 4.535011741724184e-05,
"loss": 0.4917,
"step": 4165
},
{
"epoch": 7.1587982832618025,
"grad_norm": 0.3854301869869232,
"learning_rate": 4.5099222084369805e-05,
"loss": 0.4853,
"step": 4170
},
{
"epoch": 7.167381974248927,
"grad_norm": 0.36009612679481506,
"learning_rate": 4.4848820478566966e-05,
"loss": 0.4905,
"step": 4175
},
{
"epoch": 7.175965665236052,
"grad_norm": 0.3766346573829651,
"learning_rate": 4.45989148517149e-05,
"loss": 0.4936,
"step": 4180
},
{
"epoch": 7.184549356223176,
"grad_norm": 0.3666467070579529,
"learning_rate": 4.4349507451234894e-05,
"loss": 0.487,
"step": 4185
},
{
"epoch": 7.1931330472103,
"grad_norm": 0.3508441150188446,
"learning_rate": 4.410060052006758e-05,
"loss": 0.4916,
"step": 4190
},
{
"epoch": 7.201716738197425,
"grad_norm": 0.3494192957878113,
"learning_rate": 4.3852196296652706e-05,
"loss": 0.4901,
"step": 4195
},
{
"epoch": 7.210300429184549,
"grad_norm": 0.35362881422042847,
"learning_rate": 4.360429701490934e-05,
"loss": 0.4933,
"step": 4200
},
{
"epoch": 7.218884120171674,
"grad_norm": 0.35061484575271606,
"learning_rate": 4.335690490421548e-05,
"loss": 0.4883,
"step": 4205
},
{
"epoch": 7.227467811158799,
"grad_norm": 0.3576537072658539,
"learning_rate": 4.311002218938798e-05,
"loss": 0.4896,
"step": 4210
},
{
"epoch": 7.236051502145923,
"grad_norm": 0.35517919063568115,
"learning_rate": 4.286365109066285e-05,
"loss": 0.4873,
"step": 4215
},
{
"epoch": 7.244635193133047,
"grad_norm": 0.3709685206413269,
"learning_rate": 4.261779382367499e-05,
"loss": 0.495,
"step": 4220
},
{
"epoch": 7.253218884120171,
"grad_norm": 0.39842909574508667,
"learning_rate": 4.237245259943837e-05,
"loss": 0.4957,
"step": 4225
},
{
"epoch": 7.261802575107296,
"grad_norm": 0.3722572922706604,
"learning_rate": 4.212762962432619e-05,
"loss": 0.4978,
"step": 4230
},
{
"epoch": 7.2703862660944205,
"grad_norm": 0.3434411287307739,
"learning_rate": 4.188332710005094e-05,
"loss": 0.4925,
"step": 4235
},
{
"epoch": 7.278969957081545,
"grad_norm": 0.3870338797569275,
"learning_rate": 4.1639547223644706e-05,
"loss": 0.4802,
"step": 4240
},
{
"epoch": 7.28755364806867,
"grad_norm": 0.3743104040622711,
"learning_rate": 4.139629218743931e-05,
"loss": 0.4847,
"step": 4245
},
{
"epoch": 7.296137339055794,
"grad_norm": 0.3608282506465912,
"learning_rate": 4.11535641790467e-05,
"loss": 0.486,
"step": 4250
},
{
"epoch": 7.304721030042918,
"grad_norm": 0.3679661452770233,
"learning_rate": 4.091136538133916e-05,
"loss": 0.4942,
"step": 4255
},
{
"epoch": 7.313304721030043,
"grad_norm": 0.3837164044380188,
"learning_rate": 4.06696979724298e-05,
"loss": 0.4881,
"step": 4260
},
{
"epoch": 7.3218884120171674,
"grad_norm": 0.37015727162361145,
"learning_rate": 4.042856412565287e-05,
"loss": 0.4875,
"step": 4265
},
{
"epoch": 7.330472103004292,
"grad_norm": 0.3824974596500397,
"learning_rate": 4.0187966009544255e-05,
"loss": 0.4895,
"step": 4270
},
{
"epoch": 7.339055793991417,
"grad_norm": 0.356283038854599,
"learning_rate": 3.994790578782198e-05,
"loss": 0.4961,
"step": 4275
},
{
"epoch": 7.347639484978541,
"grad_norm": 0.3656464219093323,
"learning_rate": 3.970838561936675e-05,
"loss": 0.5015,
"step": 4280
},
{
"epoch": 7.356223175965665,
"grad_norm": 0.3856269419193268,
"learning_rate": 3.9469407658202514e-05,
"loss": 0.4941,
"step": 4285
},
{
"epoch": 7.364806866952789,
"grad_norm": 0.4023449122905731,
"learning_rate": 3.9230974053477086e-05,
"loss": 0.4943,
"step": 4290
},
{
"epoch": 7.373390557939914,
"grad_norm": 0.34391605854034424,
"learning_rate": 3.899308694944298e-05,
"loss": 0.5006,
"step": 4295
},
{
"epoch": 7.381974248927039,
"grad_norm": 0.3965080976486206,
"learning_rate": 3.875574848543774e-05,
"loss": 0.4925,
"step": 4300
},
{
"epoch": 7.390557939914163,
"grad_norm": 0.37056249380111694,
"learning_rate": 3.85189607958651e-05,
"loss": 0.5052,
"step": 4305
},
{
"epoch": 7.399141630901288,
"grad_norm": 0.3915135860443115,
"learning_rate": 3.8282726010175715e-05,
"loss": 0.4885,
"step": 4310
},
{
"epoch": 7.407725321888412,
"grad_norm": 0.3784487247467041,
"learning_rate": 3.804704625284774e-05,
"loss": 0.4902,
"step": 4315
},
{
"epoch": 7.416309012875536,
"grad_norm": 0.36971473693847656,
"learning_rate": 3.7811923643367974e-05,
"loss": 0.4971,
"step": 4320
},
{
"epoch": 7.424892703862661,
"grad_norm": 0.36764466762542725,
"learning_rate": 3.757736029621292e-05,
"loss": 0.4873,
"step": 4325
},
{
"epoch": 7.4334763948497855,
"grad_norm": 0.3773200809955597,
"learning_rate": 3.734335832082927e-05,
"loss": 0.5019,
"step": 4330
},
{
"epoch": 7.44206008583691,
"grad_norm": 0.34619271755218506,
"learning_rate": 3.710991982161555e-05,
"loss": 0.4919,
"step": 4335
},
{
"epoch": 7.450643776824034,
"grad_norm": 0.33658042550086975,
"learning_rate": 3.687704689790277e-05,
"loss": 0.4883,
"step": 4340
},
{
"epoch": 7.459227467811159,
"grad_norm": 0.36298757791519165,
"learning_rate": 3.66447416439356e-05,
"loss": 0.5003,
"step": 4345
},
{
"epoch": 7.467811158798283,
"grad_norm": 0.35422852635383606,
"learning_rate": 3.641300614885378e-05,
"loss": 0.4923,
"step": 4350
},
{
"epoch": 7.476394849785407,
"grad_norm": 0.3848954737186432,
"learning_rate": 3.618184249667308e-05,
"loss": 0.4977,
"step": 4355
},
{
"epoch": 7.484978540772532,
"grad_norm": 0.3532540798187256,
"learning_rate": 3.595125276626653e-05,
"loss": 0.5032,
"step": 4360
},
{
"epoch": 7.493562231759657,
"grad_norm": 0.35986649990081787,
"learning_rate": 3.5721239031346066e-05,
"loss": 0.4964,
"step": 4365
},
{
"epoch": 7.502145922746781,
"grad_norm": 0.3666352927684784,
"learning_rate": 3.549180336044352e-05,
"loss": 0.4992,
"step": 4370
},
{
"epoch": 7.510729613733906,
"grad_norm": 0.3380297124385834,
"learning_rate": 3.526294781689206e-05,
"loss": 0.4817,
"step": 4375
},
{
"epoch": 7.51931330472103,
"grad_norm": 0.36128494143486023,
"learning_rate": 3.503467445880789e-05,
"loss": 0.483,
"step": 4380
},
{
"epoch": 7.527896995708154,
"grad_norm": 0.3538447916507721,
"learning_rate": 3.480698533907152e-05,
"loss": 0.4921,
"step": 4385
},
{
"epoch": 7.536480686695279,
"grad_norm": 0.36427024006843567,
"learning_rate": 3.457988250530931e-05,
"loss": 0.4993,
"step": 4390
},
{
"epoch": 7.545064377682404,
"grad_norm": 0.37009111046791077,
"learning_rate": 3.435336799987514e-05,
"loss": 0.4961,
"step": 4395
},
{
"epoch": 7.553648068669528,
"grad_norm": 0.36786022782325745,
"learning_rate": 3.412744385983201e-05,
"loss": 0.4894,
"step": 4400
},
{
"epoch": 7.562231759656653,
"grad_norm": 0.3731597363948822,
"learning_rate": 3.390211211693369e-05,
"loss": 0.4916,
"step": 4405
},
{
"epoch": 7.570815450643777,
"grad_norm": 0.35336822271347046,
"learning_rate": 3.367737479760652e-05,
"loss": 0.4891,
"step": 4410
},
{
"epoch": 7.579399141630901,
"grad_norm": 0.35434237122535706,
"learning_rate": 3.3453233922931094e-05,
"loss": 0.4937,
"step": 4415
},
{
"epoch": 7.587982832618025,
"grad_norm": 0.3567320704460144,
"learning_rate": 3.322969150862416e-05,
"loss": 0.4979,
"step": 4420
},
{
"epoch": 7.5965665236051505,
"grad_norm": 0.3649292588233948,
"learning_rate": 3.300674956502047e-05,
"loss": 0.4925,
"step": 4425
},
{
"epoch": 7.605150214592275,
"grad_norm": 0.39201802015304565,
"learning_rate": 3.2784410097054666e-05,
"loss": 0.4866,
"step": 4430
},
{
"epoch": 7.613733905579399,
"grad_norm": 0.3701328933238983,
"learning_rate": 3.25626751042433e-05,
"loss": 0.4876,
"step": 4435
},
{
"epoch": 7.622317596566524,
"grad_norm": 0.3631632328033447,
"learning_rate": 3.2341546580666796e-05,
"loss": 0.4944,
"step": 4440
},
{
"epoch": 7.630901287553648,
"grad_norm": 0.3858960270881653,
"learning_rate": 3.212102651495167e-05,
"loss": 0.4971,
"step": 4445
},
{
"epoch": 7.639484978540772,
"grad_norm": 0.37257277965545654,
"learning_rate": 3.1901116890252345e-05,
"loss": 0.4971,
"step": 4450
},
{
"epoch": 7.6480686695278965,
"grad_norm": 0.37199750542640686,
"learning_rate": 3.1681819684233605e-05,
"loss": 0.4989,
"step": 4455
},
{
"epoch": 7.656652360515022,
"grad_norm": 0.3896372318267822,
"learning_rate": 3.146313686905279e-05,
"loss": 0.4939,
"step": 4460
},
{
"epoch": 7.665236051502146,
"grad_norm": 0.3580029010772705,
"learning_rate": 3.124507041134177e-05,
"loss": 0.4945,
"step": 4465
},
{
"epoch": 7.67381974248927,
"grad_norm": 0.36499252915382385,
"learning_rate": 3.102762227218957e-05,
"loss": 0.4912,
"step": 4470
},
{
"epoch": 7.682403433476395,
"grad_norm": 0.3598448634147644,
"learning_rate": 3.081079440712473e-05,
"loss": 0.4994,
"step": 4475
},
{
"epoch": 7.690987124463519,
"grad_norm": 0.3923290967941284,
"learning_rate": 3.059458876609742e-05,
"loss": 0.4894,
"step": 4480
},
{
"epoch": 7.6995708154506435,
"grad_norm": 0.3650890588760376,
"learning_rate": 3.0379007293462192e-05,
"loss": 0.4905,
"step": 4485
},
{
"epoch": 7.708154506437769,
"grad_norm": 0.38070616126060486,
"learning_rate": 3.0164051927960492e-05,
"loss": 0.4996,
"step": 4490
},
{
"epoch": 7.716738197424893,
"grad_norm": 0.3461267054080963,
"learning_rate": 2.994972460270291e-05,
"loss": 0.4939,
"step": 4495
},
{
"epoch": 7.725321888412017,
"grad_norm": 0.36452245712280273,
"learning_rate": 2.9736027245152275e-05,
"loss": 0.5021,
"step": 4500
},
{
"epoch": 7.733905579399142,
"grad_norm": 0.4071807861328125,
"learning_rate": 2.9522961777105897e-05,
"loss": 0.5019,
"step": 4505
},
{
"epoch": 7.742489270386266,
"grad_norm": 0.36440128087997437,
"learning_rate": 2.9310530114678502e-05,
"loss": 0.5024,
"step": 4510
},
{
"epoch": 7.75107296137339,
"grad_norm": 0.3590448796749115,
"learning_rate": 2.9098734168284968e-05,
"loss": 0.4874,
"step": 4515
},
{
"epoch": 7.7596566523605155,
"grad_norm": 0.3638148903846741,
"learning_rate": 2.8887575842623093e-05,
"loss": 0.483,
"step": 4520
},
{
"epoch": 7.76824034334764,
"grad_norm": 0.36555618047714233,
"learning_rate": 2.867705703665654e-05,
"loss": 0.4917,
"step": 4525
},
{
"epoch": 7.776824034334764,
"grad_norm": 0.3763795793056488,
"learning_rate": 2.8467179643597697e-05,
"loss": 0.4886,
"step": 4530
},
{
"epoch": 7.785407725321888,
"grad_norm": 0.3643328845500946,
"learning_rate": 2.8257945550890665e-05,
"loss": 0.4981,
"step": 4535
},
{
"epoch": 7.793991416309013,
"grad_norm": 0.3772119879722595,
"learning_rate": 2.8049356640194314e-05,
"loss": 0.4868,
"step": 4540
},
{
"epoch": 7.802575107296137,
"grad_norm": 0.3641767203807831,
"learning_rate": 2.784141478736534e-05,
"loss": 0.4928,
"step": 4545
},
{
"epoch": 7.8111587982832615,
"grad_norm": 0.3673217296600342,
"learning_rate": 2.7634121862441386e-05,
"loss": 0.4922,
"step": 4550
},
{
"epoch": 7.819742489270387,
"grad_norm": 0.3594400882720947,
"learning_rate": 2.742747972962424e-05,
"loss": 0.5024,
"step": 4555
},
{
"epoch": 7.828326180257511,
"grad_norm": 0.36666861176490784,
"learning_rate": 2.722149024726307e-05,
"loss": 0.5001,
"step": 4560
},
{
"epoch": 7.836909871244635,
"grad_norm": 0.3865159749984741,
"learning_rate": 2.7016155267837684e-05,
"loss": 0.4909,
"step": 4565
},
{
"epoch": 7.845493562231759,
"grad_norm": 0.3859226107597351,
"learning_rate": 2.6811476637941922e-05,
"loss": 0.4917,
"step": 4570
},
{
"epoch": 7.854077253218884,
"grad_norm": 0.37502434849739075,
"learning_rate": 2.660745619826701e-05,
"loss": 0.4934,
"step": 4575
},
{
"epoch": 7.8626609442060085,
"grad_norm": 0.3713277280330658,
"learning_rate": 2.6404095783585002e-05,
"loss": 0.5048,
"step": 4580
},
{
"epoch": 7.871244635193133,
"grad_norm": 0.39273905754089355,
"learning_rate": 2.6201397222732316e-05,
"loss": 0.4937,
"step": 4585
},
{
"epoch": 7.879828326180258,
"grad_norm": 0.377205491065979,
"learning_rate": 2.599936233859326e-05,
"loss": 0.4989,
"step": 4590
},
{
"epoch": 7.888412017167382,
"grad_norm": 0.3574148714542389,
"learning_rate": 2.5797992948083592e-05,
"loss": 0.492,
"step": 4595
},
{
"epoch": 7.896995708154506,
"grad_norm": 0.3615160286426544,
"learning_rate": 2.5597290862134405e-05,
"loss": 0.4859,
"step": 4600
},
{
"epoch": 7.905579399141631,
"grad_norm": 0.37071695923805237,
"learning_rate": 2.5397257885675397e-05,
"loss": 0.4884,
"step": 4605
},
{
"epoch": 7.914163090128755,
"grad_norm": 0.36150577664375305,
"learning_rate": 2.5197895817619153e-05,
"loss": 0.4903,
"step": 4610
},
{
"epoch": 7.92274678111588,
"grad_norm": 0.3787161409854889,
"learning_rate": 2.499920645084465e-05,
"loss": 0.498,
"step": 4615
},
{
"epoch": 7.931330472103005,
"grad_norm": 0.36254122853279114,
"learning_rate": 2.480119157218108e-05,
"loss": 0.4968,
"step": 4620
},
{
"epoch": 7.939914163090129,
"grad_norm": 0.3832210302352905,
"learning_rate": 2.4603852962392125e-05,
"loss": 0.4936,
"step": 4625
},
{
"epoch": 7.948497854077253,
"grad_norm": 0.39253130555152893,
"learning_rate": 2.4407192396159627e-05,
"loss": 0.4941,
"step": 4630
},
{
"epoch": 7.957081545064378,
"grad_norm": 0.3705868422985077,
"learning_rate": 2.4211211642067623e-05,
"loss": 0.4864,
"step": 4635
},
{
"epoch": 7.965665236051502,
"grad_norm": 0.38986867666244507,
"learning_rate": 2.401591246258673e-05,
"loss": 0.4971,
"step": 4640
},
{
"epoch": 7.9742489270386265,
"grad_norm": 0.3880539536476135,
"learning_rate": 2.3821296614058054e-05,
"loss": 0.4966,
"step": 4645
},
{
"epoch": 7.982832618025751,
"grad_norm": 0.3790036141872406,
"learning_rate": 2.3627365846677306e-05,
"loss": 0.5004,
"step": 4650
},
{
"epoch": 7.991416309012876,
"grad_norm": 0.3554070293903351,
"learning_rate": 2.3434121904479434e-05,
"loss": 0.4865,
"step": 4655
},
{
"epoch": 8.0,
"grad_norm": 0.3582840859889984,
"learning_rate": 2.3241566525322554e-05,
"loss": 0.5038,
"step": 4660
},
{
"epoch": 8.0,
"eval_loss": 3.4052770137786865,
"eval_runtime": 0.394,
"eval_samples_per_second": 15.23,
"eval_steps_per_second": 2.538,
"step": 4660
},
{
"epoch": 8.008583690987125,
"grad_norm": 0.29518163204193115,
"learning_rate": 2.304970144087255e-05,
"loss": 0.4553,
"step": 4665
},
{
"epoch": 8.017167381974248,
"grad_norm": 0.3456011414527893,
"learning_rate": 2.2858528376587407e-05,
"loss": 0.4638,
"step": 4670
},
{
"epoch": 8.025751072961373,
"grad_norm": 0.3549324572086334,
"learning_rate": 2.2668049051701713e-05,
"loss": 0.461,
"step": 4675
},
{
"epoch": 8.034334763948499,
"grad_norm": 0.31662818789482117,
"learning_rate": 2.247826517921121e-05,
"loss": 0.4662,
"step": 4680
},
{
"epoch": 8.042918454935622,
"grad_norm": 0.3052162230014801,
"learning_rate": 2.2289178465857397e-05,
"loss": 0.4645,
"step": 4685
},
{
"epoch": 8.051502145922747,
"grad_norm": 0.34132641553878784,
"learning_rate": 2.2100790612112133e-05,
"loss": 0.461,
"step": 4690
},
{
"epoch": 8.060085836909872,
"grad_norm": 0.3659987449645996,
"learning_rate": 2.19131033121624e-05,
"loss": 0.4602,
"step": 4695
},
{
"epoch": 8.068669527896995,
"grad_norm": 0.3580094575881958,
"learning_rate": 2.1726118253895034e-05,
"loss": 0.4593,
"step": 4700
},
{
"epoch": 8.07725321888412,
"grad_norm": 0.32578280568122864,
"learning_rate": 2.1539837118881567e-05,
"loss": 0.4593,
"step": 4705
},
{
"epoch": 8.085836909871245,
"grad_norm": 0.3422725200653076,
"learning_rate": 2.135426158236309e-05,
"loss": 0.4624,
"step": 4710
},
{
"epoch": 8.094420600858369,
"grad_norm": 0.33877745270729065,
"learning_rate": 2.116939331323514e-05,
"loss": 0.465,
"step": 4715
},
{
"epoch": 8.103004291845494,
"grad_norm": 0.3325134813785553,
"learning_rate": 2.098523397403288e-05,
"loss": 0.4585,
"step": 4720
},
{
"epoch": 8.111587982832617,
"grad_norm": 0.3371487259864807,
"learning_rate": 2.080178522091585e-05,
"loss": 0.4631,
"step": 4725
},
{
"epoch": 8.120171673819742,
"grad_norm": 0.3465471565723419,
"learning_rate": 2.0619048703653266e-05,
"loss": 0.469,
"step": 4730
},
{
"epoch": 8.128755364806867,
"grad_norm": 0.3354833424091339,
"learning_rate": 2.04370260656093e-05,
"loss": 0.4656,
"step": 4735
},
{
"epoch": 8.13733905579399,
"grad_norm": 0.335443913936615,
"learning_rate": 2.025571894372794e-05,
"loss": 0.4591,
"step": 4740
},
{
"epoch": 8.145922746781116,
"grad_norm": 0.35047757625579834,
"learning_rate": 2.0075128968518573e-05,
"loss": 0.4656,
"step": 4745
},
{
"epoch": 8.15450643776824,
"grad_norm": 0.373524934053421,
"learning_rate": 1.989525776404132e-05,
"loss": 0.4612,
"step": 4750
},
{
"epoch": 8.163090128755364,
"grad_norm": 0.3468015491962433,
"learning_rate": 1.9716106947892164e-05,
"loss": 0.4594,
"step": 4755
},
{
"epoch": 8.17167381974249,
"grad_norm": 0.3522886335849762,
"learning_rate": 1.9537678131188674e-05,
"loss": 0.4635,
"step": 4760
},
{
"epoch": 8.180257510729614,
"grad_norm": 0.350538045167923,
"learning_rate": 1.9359972918555492e-05,
"loss": 0.4615,
"step": 4765
},
{
"epoch": 8.188841201716738,
"grad_norm": 0.31984084844589233,
"learning_rate": 1.9182992908109644e-05,
"loss": 0.4623,
"step": 4770
},
{
"epoch": 8.197424892703863,
"grad_norm": 0.34820571541786194,
"learning_rate": 1.900673969144653e-05,
"loss": 0.465,
"step": 4775
},
{
"epoch": 8.206008583690988,
"grad_norm": 0.3526110053062439,
"learning_rate": 1.883121485362538e-05,
"loss": 0.4608,
"step": 4780
},
{
"epoch": 8.214592274678111,
"grad_norm": 0.3859311044216156,
"learning_rate": 1.865641997315496e-05,
"loss": 0.4705,
"step": 4785
},
{
"epoch": 8.223175965665236,
"grad_norm": 0.3409660756587982,
"learning_rate": 1.8482356621979645e-05,
"loss": 0.4647,
"step": 4790
},
{
"epoch": 8.231759656652361,
"grad_norm": 0.34335795044898987,
"learning_rate": 1.8309026365464998e-05,
"loss": 0.4647,
"step": 4795
},
{
"epoch": 8.240343347639485,
"grad_norm": 0.33586952090263367,
"learning_rate": 1.813643076238375e-05,
"loss": 0.4626,
"step": 4800
},
{
"epoch": 8.24892703862661,
"grad_norm": 0.343476265668869,
"learning_rate": 1.7964571364902005e-05,
"loss": 0.4604,
"step": 4805
},
{
"epoch": 8.257510729613735,
"grad_norm": 0.3527016341686249,
"learning_rate": 1.779344971856497e-05,
"loss": 0.4645,
"step": 4810
},
{
"epoch": 8.266094420600858,
"grad_norm": 0.34603551030158997,
"learning_rate": 1.7623067362283243e-05,
"loss": 0.4641,
"step": 4815
},
{
"epoch": 8.274678111587983,
"grad_norm": 0.3567690849304199,
"learning_rate": 1.7453425828318936e-05,
"loss": 0.4622,
"step": 4820
},
{
"epoch": 8.283261802575108,
"grad_norm": 0.3398036062717438,
"learning_rate": 1.728452664227187e-05,
"loss": 0.457,
"step": 4825
},
{
"epoch": 8.291845493562231,
"grad_norm": 0.359521746635437,
"learning_rate": 1.7116371323065883e-05,
"loss": 0.4616,
"step": 4830
},
{
"epoch": 8.300429184549357,
"grad_norm": 0.3150378465652466,
"learning_rate": 1.694896138293516e-05,
"loss": 0.4578,
"step": 4835
},
{
"epoch": 8.309012875536482,
"grad_norm": 0.3591357469558716,
"learning_rate": 1.6782298327410616e-05,
"loss": 0.4604,
"step": 4840
},
{
"epoch": 8.317596566523605,
"grad_norm": 0.33606967329978943,
"learning_rate": 1.66163836553064e-05,
"loss": 0.4521,
"step": 4845
},
{
"epoch": 8.32618025751073,
"grad_norm": 0.3668070435523987,
"learning_rate": 1.6451218858706374e-05,
"loss": 0.4737,
"step": 4850
},
{
"epoch": 8.334763948497853,
"grad_norm": 0.36258599162101746,
"learning_rate": 1.628680542295069e-05,
"loss": 0.4691,
"step": 4855
},
{
"epoch": 8.343347639484978,
"grad_norm": 0.3564538061618805,
"learning_rate": 1.6123144826622504e-05,
"loss": 0.4634,
"step": 4860
},
{
"epoch": 8.351931330472103,
"grad_norm": 0.36181172728538513,
"learning_rate": 1.5960238541534578e-05,
"loss": 0.4555,
"step": 4865
},
{
"epoch": 8.360515021459227,
"grad_norm": 0.36802351474761963,
"learning_rate": 1.579808803271612e-05,
"loss": 0.4605,
"step": 4870
},
{
"epoch": 8.369098712446352,
"grad_norm": 0.37159237265586853,
"learning_rate": 1.563669475839956e-05,
"loss": 0.46,
"step": 4875
},
{
"epoch": 8.377682403433477,
"grad_norm": 0.36820727586746216,
"learning_rate": 1.5476060170007457e-05,
"loss": 0.467,
"step": 4880
},
{
"epoch": 8.3862660944206,
"grad_norm": 0.3330000340938568,
"learning_rate": 1.531618571213953e-05,
"loss": 0.469,
"step": 4885
},
{
"epoch": 8.394849785407725,
"grad_norm": 0.38085103034973145,
"learning_rate": 1.5157072822559437e-05,
"loss": 0.4644,
"step": 4890
},
{
"epoch": 8.40343347639485,
"grad_norm": 0.35326817631721497,
"learning_rate": 1.4998722932182074e-05,
"loss": 0.4659,
"step": 4895
},
{
"epoch": 8.412017167381974,
"grad_norm": 0.3420933187007904,
"learning_rate": 1.4841137465060672e-05,
"loss": 0.4673,
"step": 4900
},
{
"epoch": 8.420600858369099,
"grad_norm": 0.3507622480392456,
"learning_rate": 1.4684317838373884e-05,
"loss": 0.4721,
"step": 4905
},
{
"epoch": 8.429184549356224,
"grad_norm": 0.35186630487442017,
"learning_rate": 1.4528265462413038e-05,
"loss": 0.4667,
"step": 4910
},
{
"epoch": 8.437768240343347,
"grad_norm": 0.3655546009540558,
"learning_rate": 1.4372981740569646e-05,
"loss": 0.4675,
"step": 4915
},
{
"epoch": 8.446351931330472,
"grad_norm": 0.3504914343357086,
"learning_rate": 1.4218468069322578e-05,
"loss": 0.4657,
"step": 4920
},
{
"epoch": 8.454935622317597,
"grad_norm": 0.3535081446170807,
"learning_rate": 1.4064725838225568e-05,
"loss": 0.4672,
"step": 4925
},
{
"epoch": 8.46351931330472,
"grad_norm": 0.38395631313323975,
"learning_rate": 1.3911756429894763e-05,
"loss": 0.4684,
"step": 4930
},
{
"epoch": 8.472103004291846,
"grad_norm": 0.3384489417076111,
"learning_rate": 1.3759561219996242e-05,
"loss": 0.4515,
"step": 4935
},
{
"epoch": 8.48068669527897,
"grad_norm": 0.3759305477142334,
"learning_rate": 1.3608141577233636e-05,
"loss": 0.4604,
"step": 4940
},
{
"epoch": 8.489270386266094,
"grad_norm": 0.3741336464881897,
"learning_rate": 1.345749886333586e-05,
"loss": 0.4683,
"step": 4945
},
{
"epoch": 8.49785407725322,
"grad_norm": 0.3483313322067261,
"learning_rate": 1.3307634433044846e-05,
"loss": 0.4639,
"step": 4950
},
{
"epoch": 8.506437768240342,
"grad_norm": 0.36218151450157166,
"learning_rate": 1.3158549634103357e-05,
"loss": 0.466,
"step": 4955
},
{
"epoch": 8.515021459227468,
"grad_norm": 0.363930344581604,
"learning_rate": 1.3010245807242849e-05,
"loss": 0.4617,
"step": 4960
},
{
"epoch": 8.523605150214593,
"grad_norm": 0.35775625705718994,
"learning_rate": 1.2862724286171467e-05,
"loss": 0.4717,
"step": 4965
},
{
"epoch": 8.532188841201716,
"grad_norm": 0.3388819098472595,
"learning_rate": 1.2715986397561997e-05,
"loss": 0.467,
"step": 4970
},
{
"epoch": 8.540772532188841,
"grad_norm": 0.3473096787929535,
"learning_rate": 1.2570033461039954e-05,
"loss": 0.4569,
"step": 4975
},
{
"epoch": 8.549356223175966,
"grad_norm": 0.36242905259132385,
"learning_rate": 1.2424866789171729e-05,
"loss": 0.4631,
"step": 4980
},
{
"epoch": 8.55793991416309,
"grad_norm": 0.33919695019721985,
"learning_rate": 1.2280487687452768e-05,
"loss": 0.4658,
"step": 4985
},
{
"epoch": 8.566523605150214,
"grad_norm": 0.36114802956581116,
"learning_rate": 1.2136897454295837e-05,
"loss": 0.4615,
"step": 4990
},
{
"epoch": 8.57510729613734,
"grad_norm": 0.3717144727706909,
"learning_rate": 1.199409738101933e-05,
"loss": 0.4604,
"step": 4995
},
{
"epoch": 8.583690987124463,
"grad_norm": 0.3811343014240265,
"learning_rate": 1.1852088751835689e-05,
"loss": 0.4623,
"step": 5000
},
{
"epoch": 8.592274678111588,
"grad_norm": 0.35531142354011536,
"learning_rate": 1.1710872843839804e-05,
"loss": 0.4609,
"step": 5005
},
{
"epoch": 8.600858369098713,
"grad_norm": 0.3563953936100006,
"learning_rate": 1.1570450926997655e-05,
"loss": 0.4699,
"step": 5010
},
{
"epoch": 8.609442060085836,
"grad_norm": 0.3635469377040863,
"learning_rate": 1.1430824264134654e-05,
"loss": 0.4632,
"step": 5015
},
{
"epoch": 8.618025751072961,
"grad_norm": 0.3603283762931824,
"learning_rate": 1.1291994110924509e-05,
"loss": 0.4671,
"step": 5020
},
{
"epoch": 8.626609442060087,
"grad_norm": 0.35889148712158203,
"learning_rate": 1.1153961715877914e-05,
"loss": 0.4586,
"step": 5025
},
{
"epoch": 8.63519313304721,
"grad_norm": 0.38485071063041687,
"learning_rate": 1.1016728320331093e-05,
"loss": 0.4698,
"step": 5030
},
{
"epoch": 8.643776824034335,
"grad_norm": 0.3366287052631378,
"learning_rate": 1.0880295158434983e-05,
"loss": 0.4598,
"step": 5035
},
{
"epoch": 8.65236051502146,
"grad_norm": 0.3784838914871216,
"learning_rate": 1.0744663457143878e-05,
"loss": 0.4637,
"step": 5040
},
{
"epoch": 8.660944206008583,
"grad_norm": 0.35765987634658813,
"learning_rate": 1.0609834436204403e-05,
"loss": 0.462,
"step": 5045
},
{
"epoch": 8.669527896995708,
"grad_norm": 0.37458154559135437,
"learning_rate": 1.0475809308144747e-05,
"loss": 0.4613,
"step": 5050
},
{
"epoch": 8.678111587982833,
"grad_norm": 0.374141126871109,
"learning_rate": 1.0342589278263559e-05,
"loss": 0.4614,
"step": 5055
},
{
"epoch": 8.686695278969957,
"grad_norm": 0.34101325273513794,
"learning_rate": 1.0210175544619116e-05,
"loss": 0.4627,
"step": 5060
},
{
"epoch": 8.695278969957082,
"grad_norm": 0.345047265291214,
"learning_rate": 1.0078569298018758e-05,
"loss": 0.4708,
"step": 5065
},
{
"epoch": 8.703862660944207,
"grad_norm": 0.3726472854614258,
"learning_rate": 9.947771722007915e-06,
"loss": 0.464,
"step": 5070
},
{
"epoch": 8.71244635193133,
"grad_norm": 0.3675495386123657,
"learning_rate": 9.817783992859564e-06,
"loss": 0.4633,
"step": 5075
},
{
"epoch": 8.721030042918455,
"grad_norm": 0.32659244537353516,
"learning_rate": 9.688607279563766e-06,
"loss": 0.4685,
"step": 5080
},
{
"epoch": 8.729613733905579,
"grad_norm": 0.3733295798301697,
"learning_rate": 9.560242743816972e-06,
"loss": 0.4532,
"step": 5085
},
{
"epoch": 8.738197424892704,
"grad_norm": 0.35878074169158936,
"learning_rate": 9.432691540011674e-06,
"loss": 0.4678,
"step": 5090
},
{
"epoch": 8.746781115879829,
"grad_norm": 0.3598923087120056,
"learning_rate": 9.305954815226014e-06,
"loss": 0.4715,
"step": 5095
},
{
"epoch": 8.755364806866952,
"grad_norm": 0.34524357318878174,
"learning_rate": 9.180033709213454e-06,
"loss": 0.463,
"step": 5100
},
{
"epoch": 8.763948497854077,
"grad_norm": 0.34148141741752625,
"learning_rate": 9.054929354392527e-06,
"loss": 0.4693,
"step": 5105
},
{
"epoch": 8.772532188841202,
"grad_norm": 0.35487231612205505,
"learning_rate": 8.93064287583667e-06,
"loss": 0.4625,
"step": 5110
},
{
"epoch": 8.781115879828326,
"grad_norm": 0.36163830757141113,
"learning_rate": 8.807175391264067e-06,
"loss": 0.4619,
"step": 5115
},
{
"epoch": 8.78969957081545,
"grad_norm": 0.34637895226478577,
"learning_rate": 8.684528011027659e-06,
"loss": 0.4612,
"step": 5120
},
{
"epoch": 8.798283261802576,
"grad_norm": 0.3432014584541321,
"learning_rate": 8.562701838105115e-06,
"loss": 0.4666,
"step": 5125
},
{
"epoch": 8.806866952789699,
"grad_norm": 0.34569093585014343,
"learning_rate": 8.441697968088891e-06,
"loss": 0.4659,
"step": 5130
},
{
"epoch": 8.815450643776824,
"grad_norm": 0.3551480770111084,
"learning_rate": 8.321517489176433e-06,
"loss": 0.4619,
"step": 5135
},
{
"epoch": 8.82403433476395,
"grad_norm": 0.35777968168258667,
"learning_rate": 8.202161482160353e-06,
"loss": 0.4583,
"step": 5140
},
{
"epoch": 8.832618025751072,
"grad_norm": 0.3783648908138275,
"learning_rate": 8.083631020418791e-06,
"loss": 0.4596,
"step": 5145
},
{
"epoch": 8.841201716738198,
"grad_norm": 0.33539873361587524,
"learning_rate": 7.965927169905551e-06,
"loss": 0.4711,
"step": 5150
},
{
"epoch": 8.849785407725323,
"grad_norm": 0.36662939190864563,
"learning_rate": 7.84905098914076e-06,
"loss": 0.4665,
"step": 5155
},
{
"epoch": 8.858369098712446,
"grad_norm": 0.34115639328956604,
"learning_rate": 7.733003529201278e-06,
"loss": 0.4581,
"step": 5160
},
{
"epoch": 8.866952789699571,
"grad_norm": 0.3474951386451721,
"learning_rate": 7.617785833711077e-06,
"loss": 0.4662,
"step": 5165
},
{
"epoch": 8.875536480686696,
"grad_norm": 0.34105169773101807,
"learning_rate": 7.503398938832107e-06,
"loss": 0.4575,
"step": 5170
},
{
"epoch": 8.88412017167382,
"grad_norm": 0.381610631942749,
"learning_rate": 7.389843873254843e-06,
"loss": 0.4616,
"step": 5175
},
{
"epoch": 8.892703862660944,
"grad_norm": 0.3617483079433441,
"learning_rate": 7.277121658189001e-06,
"loss": 0.4629,
"step": 5180
},
{
"epoch": 8.901287553648068,
"grad_norm": 0.3416938781738281,
"learning_rate": 7.165233307354446e-06,
"loss": 0.465,
"step": 5185
},
{
"epoch": 8.909871244635193,
"grad_norm": 0.35436323285102844,
"learning_rate": 7.054179826972074e-06,
"loss": 0.4628,
"step": 5190
},
{
"epoch": 8.918454935622318,
"grad_norm": 0.35174670815467834,
"learning_rate": 6.943962215754618e-06,
"loss": 0.4704,
"step": 5195
},
{
"epoch": 8.927038626609441,
"grad_norm": 0.3784787356853485,
"learning_rate": 6.834581464897871e-06,
"loss": 0.4683,
"step": 5200
},
{
"epoch": 8.935622317596566,
"grad_norm": 0.34359362721443176,
"learning_rate": 6.726038558071656e-06,
"loss": 0.4634,
"step": 5205
},
{
"epoch": 8.944206008583691,
"grad_norm": 0.35282644629478455,
"learning_rate": 6.618334471410925e-06,
"loss": 0.4608,
"step": 5210
},
{
"epoch": 8.952789699570815,
"grad_norm": 0.3536522388458252,
"learning_rate": 6.511470173507161e-06,
"loss": 0.4631,
"step": 5215
},
{
"epoch": 8.96137339055794,
"grad_norm": 0.34291592240333557,
"learning_rate": 6.405446625399481e-06,
"loss": 0.4628,
"step": 5220
},
{
"epoch": 8.969957081545065,
"grad_norm": 0.33180317282676697,
"learning_rate": 6.300264780566112e-06,
"loss": 0.4615,
"step": 5225
},
{
"epoch": 8.978540772532188,
"grad_norm": 0.3489115536212921,
"learning_rate": 6.195925584915752e-06,
"loss": 0.4596,
"step": 5230
},
{
"epoch": 8.987124463519313,
"grad_norm": 0.34033530950546265,
"learning_rate": 6.0924299767791126e-06,
"loss": 0.47,
"step": 5235
},
{
"epoch": 8.995708154506438,
"grad_norm": 0.37230873107910156,
"learning_rate": 5.989778886900432e-06,
"loss": 0.4624,
"step": 5240
},
{
"epoch": 8.999141630901288,
"eval_loss": 3.695244073867798,
"eval_runtime": 0.3944,
"eval_samples_per_second": 15.212,
"eval_steps_per_second": 2.535,
"step": 5242
},
{
"epoch": 9.004291845493562,
"grad_norm": 0.28615859150886536,
"learning_rate": 5.887973238429145e-06,
"loss": 0.4573,
"step": 5245
},
{
"epoch": 9.012875536480687,
"grad_norm": 0.3141264319419861,
"learning_rate": 5.787013946911546e-06,
"loss": 0.4503,
"step": 5250
},
{
"epoch": 9.021459227467812,
"grad_norm": 0.32362473011016846,
"learning_rate": 5.686901920282606e-06,
"loss": 0.4558,
"step": 5255
},
{
"epoch": 9.030042918454935,
"grad_norm": 0.32775941491127014,
"learning_rate": 5.587638058857736e-06,
"loss": 0.445,
"step": 5260
},
{
"epoch": 9.03862660944206,
"grad_norm": 0.33696043491363525,
"learning_rate": 5.48922325532476e-06,
"loss": 0.4521,
"step": 5265
},
{
"epoch": 9.047210300429185,
"grad_norm": 0.3470819294452667,
"learning_rate": 5.391658394735855e-06,
"loss": 0.4513,
"step": 5270
},
{
"epoch": 9.055793991416309,
"grad_norm": 0.3222349286079407,
"learning_rate": 5.2949443544995644e-06,
"loss": 0.4488,
"step": 5275
},
{
"epoch": 9.064377682403434,
"grad_norm": 0.33785441517829895,
"learning_rate": 5.199082004372957e-06,
"loss": 0.4493,
"step": 5280
},
{
"epoch": 9.072961373390559,
"grad_norm": 0.3577852249145508,
"learning_rate": 5.104072206453802e-06,
"loss": 0.4615,
"step": 5285
},
{
"epoch": 9.081545064377682,
"grad_norm": 0.32605546712875366,
"learning_rate": 5.009915815172772e-06,
"loss": 0.4482,
"step": 5290
},
{
"epoch": 9.090128755364807,
"grad_norm": 0.320216566324234,
"learning_rate": 4.916613677285786e-06,
"loss": 0.4518,
"step": 5295
},
{
"epoch": 9.098712446351932,
"grad_norm": 0.323912650346756,
"learning_rate": 4.8241666318664115e-06,
"loss": 0.4442,
"step": 5300
},
{
"epoch": 9.107296137339056,
"grad_norm": 0.342655748128891,
"learning_rate": 4.732575510298276e-06,
"loss": 0.4437,
"step": 5305
},
{
"epoch": 9.11587982832618,
"grad_norm": 0.34046629071235657,
"learning_rate": 4.641841136267666e-06,
"loss": 0.4497,
"step": 5310
},
{
"epoch": 9.124463519313304,
"grad_norm": 0.3281947374343872,
"learning_rate": 4.551964325756031e-06,
"loss": 0.4569,
"step": 5315
},
{
"epoch": 9.133047210300429,
"grad_norm": 0.3604039251804352,
"learning_rate": 4.462945887032632e-06,
"loss": 0.451,
"step": 5320
},
{
"epoch": 9.141630901287554,
"grad_norm": 0.3501492738723755,
"learning_rate": 4.374786620647442e-06,
"loss": 0.448,
"step": 5325
},
{
"epoch": 9.150214592274677,
"grad_norm": 0.3506092429161072,
"learning_rate": 4.287487319423756e-06,
"loss": 0.4459,
"step": 5330
},
{
"epoch": 9.158798283261802,
"grad_norm": 0.3382214307785034,
"learning_rate": 4.20104876845111e-06,
"loss": 0.452,
"step": 5335
},
{
"epoch": 9.167381974248928,
"grad_norm": 0.3224546015262604,
"learning_rate": 4.115471745078314e-06,
"loss": 0.4535,
"step": 5340
},
{
"epoch": 9.17596566523605,
"grad_norm": 0.3321012854576111,
"learning_rate": 4.03075701890635e-06,
"loss": 0.4477,
"step": 5345
},
{
"epoch": 9.184549356223176,
"grad_norm": 0.32435712218284607,
"learning_rate": 3.946905351781472e-06,
"loss": 0.4494,
"step": 5350
},
{
"epoch": 9.193133047210301,
"grad_norm": 0.33920931816101074,
"learning_rate": 3.863917497788438e-06,
"loss": 0.456,
"step": 5355
},
{
"epoch": 9.201716738197424,
"grad_norm": 0.33260124921798706,
"learning_rate": 3.7817942032436048e-06,
"loss": 0.4471,
"step": 5360
},
{
"epoch": 9.21030042918455,
"grad_norm": 0.3275390863418579,
"learning_rate": 3.700536206688321e-06,
"loss": 0.4493,
"step": 5365
},
{
"epoch": 9.218884120171674,
"grad_norm": 0.35647067427635193,
"learning_rate": 3.620144238882206e-06,
"loss": 0.4491,
"step": 5370
},
{
"epoch": 9.227467811158798,
"grad_norm": 0.3307458162307739,
"learning_rate": 3.5406190227966427e-06,
"loss": 0.4504,
"step": 5375
},
{
"epoch": 9.236051502145923,
"grad_norm": 0.35020336508750916,
"learning_rate": 3.4619612736082273e-06,
"loss": 0.4577,
"step": 5380
},
{
"epoch": 9.244635193133048,
"grad_norm": 0.33766666054725647,
"learning_rate": 3.3841716986923624e-06,
"loss": 0.4531,
"step": 5385
},
{
"epoch": 9.253218884120171,
"grad_norm": 0.33843091130256653,
"learning_rate": 3.3072509976169065e-06,
"loss": 0.4564,
"step": 5390
},
{
"epoch": 9.261802575107296,
"grad_norm": 0.3248330056667328,
"learning_rate": 3.2311998621358363e-06,
"loss": 0.4526,
"step": 5395
},
{
"epoch": 9.270386266094421,
"grad_norm": 0.3351515829563141,
"learning_rate": 3.1560189761830728e-06,
"loss": 0.4544,
"step": 5400
},
{
"epoch": 9.278969957081545,
"grad_norm": 0.3289077877998352,
"learning_rate": 3.0817090158663185e-06,
"loss": 0.4449,
"step": 5405
},
{
"epoch": 9.28755364806867,
"grad_norm": 0.32089975476264954,
"learning_rate": 3.008270649460965e-06,
"loss": 0.4496,
"step": 5410
},
{
"epoch": 9.296137339055793,
"grad_norm": 0.2968757748603821,
"learning_rate": 2.9357045374040825e-06,
"loss": 0.4458,
"step": 5415
},
{
"epoch": 9.304721030042918,
"grad_norm": 0.34240734577178955,
"learning_rate": 2.8640113322885185e-06,
"loss": 0.4469,
"step": 5420
},
{
"epoch": 9.313304721030043,
"grad_norm": 0.33385157585144043,
"learning_rate": 2.7931916788569545e-06,
"loss": 0.4527,
"step": 5425
},
{
"epoch": 9.321888412017167,
"grad_norm": 0.34486281871795654,
"learning_rate": 2.723246213996178e-06,
"loss": 0.4542,
"step": 5430
},
{
"epoch": 9.330472103004292,
"grad_norm": 0.3246801495552063,
"learning_rate": 2.654175566731365e-06,
"loss": 0.4574,
"step": 5435
},
{
"epoch": 9.339055793991417,
"grad_norm": 0.33539149165153503,
"learning_rate": 2.5859803582202968e-06,
"loss": 0.4457,
"step": 5440
},
{
"epoch": 9.34763948497854,
"grad_norm": 0.33203625679016113,
"learning_rate": 2.518661201747918e-06,
"loss": 0.4567,
"step": 5445
},
{
"epoch": 9.356223175965665,
"grad_norm": 0.32282063364982605,
"learning_rate": 2.452218702720821e-06,
"loss": 0.4427,
"step": 5450
},
{
"epoch": 9.36480686695279,
"grad_norm": 0.333141028881073,
"learning_rate": 2.3866534586616364e-06,
"loss": 0.4548,
"step": 5455
},
{
"epoch": 9.373390557939913,
"grad_norm": 0.3323938250541687,
"learning_rate": 2.3219660592038285e-06,
"loss": 0.4558,
"step": 5460
},
{
"epoch": 9.381974248927039,
"grad_norm": 0.33186817169189453,
"learning_rate": 2.258157086086388e-06,
"loss": 0.4499,
"step": 5465
},
{
"epoch": 9.390557939914164,
"grad_norm": 0.33666694164276123,
"learning_rate": 2.1952271131484236e-06,
"loss": 0.4533,
"step": 5470
},
{
"epoch": 9.399141630901287,
"grad_norm": 0.3561409115791321,
"learning_rate": 2.133176706324236e-06,
"loss": 0.4574,
"step": 5475
},
{
"epoch": 9.407725321888412,
"grad_norm": 0.3282804489135742,
"learning_rate": 2.0720064236380842e-06,
"loss": 0.4511,
"step": 5480
},
{
"epoch": 9.416309012875537,
"grad_norm": 0.3417915403842926,
"learning_rate": 2.0117168151991606e-06,
"loss": 0.4517,
"step": 5485
},
{
"epoch": 9.42489270386266,
"grad_norm": 0.35541415214538574,
"learning_rate": 1.9523084231967358e-06,
"loss": 0.4498,
"step": 5490
},
{
"epoch": 9.433476394849786,
"grad_norm": 0.33606576919555664,
"learning_rate": 1.893781781895232e-06,
"loss": 0.4466,
"step": 5495
},
{
"epoch": 9.44206008583691,
"grad_norm": 0.333290159702301,
"learning_rate": 1.8361374176293467e-06,
"loss": 0.4514,
"step": 5500
},
{
"epoch": 9.450643776824034,
"grad_norm": 0.3518344461917877,
"learning_rate": 1.7793758487994694e-06,
"loss": 0.4566,
"step": 5505
},
{
"epoch": 9.459227467811159,
"grad_norm": 0.44788244366645813,
"learning_rate": 1.7234975858669178e-06,
"loss": 0.4564,
"step": 5510
},
{
"epoch": 9.467811158798284,
"grad_norm": 0.34256601333618164,
"learning_rate": 1.6685031313493416e-06,
"loss": 0.4493,
"step": 5515
},
{
"epoch": 9.476394849785407,
"grad_norm": 0.33245575428009033,
"learning_rate": 1.6143929798162704e-06,
"loss": 0.4479,
"step": 5520
},
{
"epoch": 9.484978540772532,
"grad_norm": 0.34188759326934814,
"learning_rate": 1.5611676178845958e-06,
"loss": 0.4459,
"step": 5525
},
{
"epoch": 9.493562231759658,
"grad_norm": 0.32970142364501953,
"learning_rate": 1.5088275242142402e-06,
"loss": 0.45,
"step": 5530
},
{
"epoch": 9.50214592274678,
"grad_norm": 0.34352561831474304,
"learning_rate": 1.4573731695038395e-06,
"loss": 0.452,
"step": 5535
},
{
"epoch": 9.510729613733906,
"grad_norm": 0.35988888144493103,
"learning_rate": 1.4068050164864898e-06,
"loss": 0.4497,
"step": 5540
},
{
"epoch": 9.51931330472103,
"grad_norm": 0.32545995712280273,
"learning_rate": 1.3571235199256405e-06,
"loss": 0.4515,
"step": 5545
},
{
"epoch": 9.527896995708154,
"grad_norm": 0.3102465569972992,
"learning_rate": 1.30832912661093e-06,
"loss": 0.4405,
"step": 5550
},
{
"epoch": 9.53648068669528,
"grad_norm": 0.32020366191864014,
"learning_rate": 1.2604222753542339e-06,
"loss": 0.4479,
"step": 5555
},
{
"epoch": 9.545064377682403,
"grad_norm": 0.357705295085907,
"learning_rate": 1.2134033969856907e-06,
"loss": 0.4435,
"step": 5560
},
{
"epoch": 9.553648068669528,
"grad_norm": 0.3494960367679596,
"learning_rate": 1.1672729143497929e-06,
"loss": 0.4502,
"step": 5565
},
{
"epoch": 9.562231759656653,
"grad_norm": 0.33992525935173035,
"learning_rate": 1.1220312423016687e-06,
"loss": 0.4597,
"step": 5570
},
{
"epoch": 9.570815450643776,
"grad_norm": 0.3502410352230072,
"learning_rate": 1.0776787877032736e-06,
"loss": 0.4532,
"step": 5575
},
{
"epoch": 9.579399141630901,
"grad_norm": 0.3116472065448761,
"learning_rate": 1.034215949419748e-06,
"loss": 0.4447,
"step": 5580
},
{
"epoch": 9.587982832618026,
"grad_norm": 0.31818586587905884,
"learning_rate": 9.916431183158881e-07,
"loss": 0.449,
"step": 5585
},
{
"epoch": 9.59656652360515,
"grad_norm": 0.31284070014953613,
"learning_rate": 9.499606772525371e-07,
"loss": 0.4426,
"step": 5590
},
{
"epoch": 9.605150214592275,
"grad_norm": 0.35043418407440186,
"learning_rate": 9.091690010831988e-07,
"loss": 0.4521,
"step": 5595
},
{
"epoch": 9.6137339055794,
"grad_norm": 0.352905809879303,
"learning_rate": 8.692684566506959e-07,
"loss": 0.4451,
"step": 5600
},
{
"epoch": 9.622317596566523,
"grad_norm": 0.31967219710350037,
"learning_rate": 8.30259402783784e-07,
"loss": 0.4576,
"step": 5605
},
{
"epoch": 9.630901287553648,
"grad_norm": 0.3698691129684448,
"learning_rate": 7.921421902939874e-07,
"loss": 0.4494,
"step": 5610
},
{
"epoch": 9.639484978540773,
"grad_norm": 0.32335957884788513,
"learning_rate": 7.54917161972446e-07,
"loss": 0.4464,
"step": 5615
},
{
"epoch": 9.648068669527897,
"grad_norm": 0.3521655201911926,
"learning_rate": 7.185846525867956e-07,
"loss": 0.4571,
"step": 5620
},
{
"epoch": 9.656652360515022,
"grad_norm": 0.3388623893260956,
"learning_rate": 6.831449888781926e-07,
"loss": 0.453,
"step": 5625
},
{
"epoch": 9.665236051502147,
"grad_norm": 0.3277793228626251,
"learning_rate": 6.485984895583608e-07,
"loss": 0.4486,
"step": 5630
},
{
"epoch": 9.67381974248927,
"grad_norm": 0.32895511388778687,
"learning_rate": 6.149454653067044e-07,
"loss": 0.4509,
"step": 5635
},
{
"epoch": 9.682403433476395,
"grad_norm": 0.34782108664512634,
"learning_rate": 5.821862187675775e-07,
"loss": 0.4537,
"step": 5640
},
{
"epoch": 9.690987124463518,
"grad_norm": 0.33518868684768677,
"learning_rate": 5.503210445474638e-07,
"loss": 0.4543,
"step": 5645
},
{
"epoch": 9.699570815450643,
"grad_norm": 0.33902445435523987,
"learning_rate": 5.193502292124341e-07,
"loss": 0.4487,
"step": 5650
},
{
"epoch": 9.708154506437769,
"grad_norm": 0.3404318690299988,
"learning_rate": 4.892740512854932e-07,
"loss": 0.4597,
"step": 5655
},
{
"epoch": 9.716738197424892,
"grad_norm": 0.33036890625953674,
"learning_rate": 4.600927812441036e-07,
"loss": 0.4472,
"step": 5660
},
{
"epoch": 9.725321888412017,
"grad_norm": 0.34414026141166687,
"learning_rate": 4.318066815177435e-07,
"loss": 0.4452,
"step": 5665
},
{
"epoch": 9.733905579399142,
"grad_norm": 0.33726632595062256,
"learning_rate": 4.044160064855751e-07,
"loss": 0.45,
"step": 5670
},
{
"epoch": 9.742489270386265,
"grad_norm": 0.31507617235183716,
"learning_rate": 3.779210024741131e-07,
"loss": 0.4436,
"step": 5675
},
{
"epoch": 9.75107296137339,
"grad_norm": 0.35751578211784363,
"learning_rate": 3.523219077550488e-07,
"loss": 0.4514,
"step": 5680
},
{
"epoch": 9.759656652360515,
"grad_norm": 0.3282091021537781,
"learning_rate": 3.2761895254306287e-07,
"loss": 0.4472,
"step": 5685
},
{
"epoch": 9.768240343347639,
"grad_norm": 0.3441978693008423,
"learning_rate": 3.038123589938047e-07,
"loss": 0.4516,
"step": 5690
},
{
"epoch": 9.776824034334764,
"grad_norm": 0.33709728717803955,
"learning_rate": 2.8090234120188295e-07,
"loss": 0.4508,
"step": 5695
},
{
"epoch": 9.785407725321889,
"grad_norm": 0.32524409890174866,
"learning_rate": 2.588891051988895e-07,
"loss": 0.445,
"step": 5700
},
{
"epoch": 9.793991416309012,
"grad_norm": 0.3647370934486389,
"learning_rate": 2.3777284895162288e-07,
"loss": 0.444,
"step": 5705
},
{
"epoch": 9.802575107296137,
"grad_norm": 0.3221174478530884,
"learning_rate": 2.1755376236025637e-07,
"loss": 0.4478,
"step": 5710
},
{
"epoch": 9.811158798283262,
"grad_norm": 0.35065704584121704,
"learning_rate": 1.9823202725665068e-07,
"loss": 0.4538,
"step": 5715
},
{
"epoch": 9.819742489270386,
"grad_norm": 0.3380087912082672,
"learning_rate": 1.7980781740268848e-07,
"loss": 0.4477,
"step": 5720
},
{
"epoch": 9.82832618025751,
"grad_norm": 0.3177869915962219,
"learning_rate": 1.622812984887867e-07,
"loss": 0.4496,
"step": 5725
},
{
"epoch": 9.836909871244636,
"grad_norm": 0.3140113353729248,
"learning_rate": 1.4565262813230894e-07,
"loss": 0.4485,
"step": 5730
},
{
"epoch": 9.84549356223176,
"grad_norm": 0.32537147402763367,
"learning_rate": 1.2992195587619993e-07,
"loss": 0.4433,
"step": 5735
},
{
"epoch": 9.854077253218884,
"grad_norm": 0.3178805410861969,
"learning_rate": 1.1508942318767535e-07,
"loss": 0.4465,
"step": 5740
},
{
"epoch": 9.86266094420601,
"grad_norm": 0.33153483271598816,
"learning_rate": 1.0115516345686749e-07,
"loss": 0.4567,
"step": 5745
},
{
"epoch": 9.871244635193133,
"grad_norm": 0.31883201003074646,
"learning_rate": 8.811930199568163e-08,
"loss": 0.449,
"step": 5750
},
{
"epoch": 9.879828326180258,
"grad_norm": 0.3389532268047333,
"learning_rate": 7.598195603666369e-08,
"loss": 0.4533,
"step": 5755
},
{
"epoch": 9.888412017167383,
"grad_norm": 0.3354385197162628,
"learning_rate": 6.474323473194543e-08,
"loss": 0.4558,
"step": 5760
},
{
"epoch": 9.896995708154506,
"grad_norm": 0.3773539364337921,
"learning_rate": 5.4403239152212013e-08,
"loss": 0.4524,
"step": 5765
},
{
"epoch": 9.905579399141631,
"grad_norm": 0.30916520953178406,
"learning_rate": 4.4962062285902607e-08,
"loss": 0.4474,
"step": 5770
},
{
"epoch": 9.914163090128756,
"grad_norm": 0.3748643100261688,
"learning_rate": 3.6419789038244504e-08,
"loss": 0.4449,
"step": 5775
},
{
"epoch": 9.92274678111588,
"grad_norm": 0.33963683247566223,
"learning_rate": 2.877649623059808e-08,
"loss": 0.4491,
"step": 5780
},
{
"epoch": 9.931330472103005,
"grad_norm": 0.3399483859539032,
"learning_rate": 2.2032252599690773e-08,
"loss": 0.4523,
"step": 5785
},
{
"epoch": 9.939914163090128,
"grad_norm": 0.3306158781051636,
"learning_rate": 1.6187118797061917e-08,
"loss": 0.4551,
"step": 5790
},
{
"epoch": 9.948497854077253,
"grad_norm": 0.3417421579360962,
"learning_rate": 1.1241147388452167e-08,
"loss": 0.4527,
"step": 5795
},
{
"epoch": 9.957081545064378,
"grad_norm": 0.34274721145629883,
"learning_rate": 7.194382853370485e-09,
"loss": 0.4442,
"step": 5800
},
{
"epoch": 9.965665236051501,
"grad_norm": 0.36596229672431946,
"learning_rate": 4.046861584705575e-09,
"loss": 0.4545,
"step": 5805
},
{
"epoch": 9.974248927038627,
"grad_norm": 0.3520822823047638,
"learning_rate": 1.798611888370605e-09,
"loss": 0.439,
"step": 5810
},
{
"epoch": 9.982832618025752,
"grad_norm": 0.3437098562717438,
"learning_rate": 4.4965398303675745e-10,
"loss": 0.4463,
"step": 5815
},
{
"epoch": 9.991416309012875,
"grad_norm": 0.3176514208316803,
"learning_rate": 0.0,
"loss": 0.454,
"step": 5820
},
{
"epoch": 9.991416309012875,
"eval_loss": 3.8423588275909424,
"eval_runtime": 0.4223,
"eval_samples_per_second": 14.208,
"eval_steps_per_second": 2.368,
"step": 5820
},
{
"epoch": 9.991416309012875,
"step": 5820,
"total_flos": 8.683561975386472e+18,
"train_loss": 0.6765515476977293,
"train_runtime": 24420.5896,
"train_samples_per_second": 7.63,
"train_steps_per_second": 0.238
}
],
"logging_steps": 5,
"max_steps": 5820,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.683561975386472e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}