gemma-2-2b-it-lora-magicoder / trainer_state.json
alvarobartt's picture
alvarobartt HF staff
Model save
452a9b9 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9987096774193547,
"eval_steps": 500,
"global_step": 3486,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008602150537634409,
"grad_norm": 1.435989601453093,
"learning_rate": 5.730659025787966e-06,
"loss": 1.2534,
"step": 10
},
{
"epoch": 0.017204301075268817,
"grad_norm": 0.956311142058486,
"learning_rate": 1.1461318051575932e-05,
"loss": 1.1941,
"step": 20
},
{
"epoch": 0.025806451612903226,
"grad_norm": 0.5955272986463238,
"learning_rate": 1.7191977077363898e-05,
"loss": 1.0416,
"step": 30
},
{
"epoch": 0.034408602150537634,
"grad_norm": 0.5371220864077184,
"learning_rate": 2.2922636103151864e-05,
"loss": 0.9252,
"step": 40
},
{
"epoch": 0.043010752688172046,
"grad_norm": 0.3298242735866968,
"learning_rate": 2.8653295128939826e-05,
"loss": 0.8582,
"step": 50
},
{
"epoch": 0.05161290322580645,
"grad_norm": 0.3017426075394883,
"learning_rate": 3.4383954154727795e-05,
"loss": 0.7998,
"step": 60
},
{
"epoch": 0.060215053763440864,
"grad_norm": 0.3375526814194791,
"learning_rate": 4.011461318051576e-05,
"loss": 0.7638,
"step": 70
},
{
"epoch": 0.06881720430107527,
"grad_norm": 0.2705319023822148,
"learning_rate": 4.584527220630373e-05,
"loss": 0.731,
"step": 80
},
{
"epoch": 0.07741935483870968,
"grad_norm": 0.2857694097360681,
"learning_rate": 5.157593123209169e-05,
"loss": 0.7083,
"step": 90
},
{
"epoch": 0.08602150537634409,
"grad_norm": 0.3098635398347818,
"learning_rate": 5.730659025787965e-05,
"loss": 0.6841,
"step": 100
},
{
"epoch": 0.09462365591397849,
"grad_norm": 0.3198686339975185,
"learning_rate": 6.303724928366761e-05,
"loss": 0.6606,
"step": 110
},
{
"epoch": 0.1032258064516129,
"grad_norm": 0.3190855908911574,
"learning_rate": 6.876790830945559e-05,
"loss": 0.6557,
"step": 120
},
{
"epoch": 0.11182795698924732,
"grad_norm": 0.3381628025718819,
"learning_rate": 7.449856733524355e-05,
"loss": 0.649,
"step": 130
},
{
"epoch": 0.12043010752688173,
"grad_norm": 0.31598591910637613,
"learning_rate": 8.022922636103152e-05,
"loss": 0.6497,
"step": 140
},
{
"epoch": 0.12903225806451613,
"grad_norm": 0.3128921107543011,
"learning_rate": 8.595988538681948e-05,
"loss": 0.6392,
"step": 150
},
{
"epoch": 0.13763440860215054,
"grad_norm": 0.32601075629670023,
"learning_rate": 9.169054441260745e-05,
"loss": 0.6329,
"step": 160
},
{
"epoch": 0.14623655913978495,
"grad_norm": 0.3810140632755748,
"learning_rate": 9.742120343839543e-05,
"loss": 0.6246,
"step": 170
},
{
"epoch": 0.15483870967741936,
"grad_norm": 0.32108686980228696,
"learning_rate": 0.00010315186246418338,
"loss": 0.6341,
"step": 180
},
{
"epoch": 0.16344086021505377,
"grad_norm": 0.35140482524130334,
"learning_rate": 0.00010888252148997136,
"loss": 0.6341,
"step": 190
},
{
"epoch": 0.17204301075268819,
"grad_norm": 0.29928696988494263,
"learning_rate": 0.0001146131805157593,
"loss": 0.6256,
"step": 200
},
{
"epoch": 0.18064516129032257,
"grad_norm": 0.29203351329015653,
"learning_rate": 0.0001203438395415473,
"loss": 0.6206,
"step": 210
},
{
"epoch": 0.18924731182795698,
"grad_norm": 0.269908780529402,
"learning_rate": 0.00012607449856733523,
"loss": 0.6201,
"step": 220
},
{
"epoch": 0.1978494623655914,
"grad_norm": 0.2918951896599236,
"learning_rate": 0.0001318051575931232,
"loss": 0.6181,
"step": 230
},
{
"epoch": 0.2064516129032258,
"grad_norm": 0.2708453980927917,
"learning_rate": 0.00013753581661891118,
"loss": 0.6053,
"step": 240
},
{
"epoch": 0.21505376344086022,
"grad_norm": 0.2662594416135698,
"learning_rate": 0.00014326647564469916,
"loss": 0.6033,
"step": 250
},
{
"epoch": 0.22365591397849463,
"grad_norm": 0.2776129880103733,
"learning_rate": 0.0001489971346704871,
"loss": 0.618,
"step": 260
},
{
"epoch": 0.23225806451612904,
"grad_norm": 0.279068656400945,
"learning_rate": 0.00015472779369627508,
"loss": 0.6111,
"step": 270
},
{
"epoch": 0.24086021505376345,
"grad_norm": 0.2524990953581738,
"learning_rate": 0.00016045845272206303,
"loss": 0.5888,
"step": 280
},
{
"epoch": 0.24946236559139784,
"grad_norm": 0.27628464942088166,
"learning_rate": 0.000166189111747851,
"loss": 0.6051,
"step": 290
},
{
"epoch": 0.25806451612903225,
"grad_norm": 0.2738432329342113,
"learning_rate": 0.00017191977077363896,
"loss": 0.6025,
"step": 300
},
{
"epoch": 0.26666666666666666,
"grad_norm": 0.25607005690614404,
"learning_rate": 0.00017765042979942693,
"loss": 0.6035,
"step": 310
},
{
"epoch": 0.2752688172043011,
"grad_norm": 0.247801611216806,
"learning_rate": 0.0001833810888252149,
"loss": 0.6118,
"step": 320
},
{
"epoch": 0.2838709677419355,
"grad_norm": 0.24971595808191485,
"learning_rate": 0.00018911174785100288,
"loss": 0.6003,
"step": 330
},
{
"epoch": 0.2924731182795699,
"grad_norm": 0.2411785853482113,
"learning_rate": 0.00019484240687679086,
"loss": 0.6019,
"step": 340
},
{
"epoch": 0.3010752688172043,
"grad_norm": 0.24693597842223552,
"learning_rate": 0.0001999999498534943,
"loss": 0.5912,
"step": 350
},
{
"epoch": 0.3096774193548387,
"grad_norm": 0.23649676383007667,
"learning_rate": 0.0001999939323336644,
"loss": 0.6032,
"step": 360
},
{
"epoch": 0.31827956989247314,
"grad_norm": 0.23340962816367694,
"learning_rate": 0.0001999778862042167,
"loss": 0.5867,
"step": 370
},
{
"epoch": 0.32688172043010755,
"grad_norm": 0.228292747503904,
"learning_rate": 0.0001999518130744525,
"loss": 0.5869,
"step": 380
},
{
"epoch": 0.33548387096774196,
"grad_norm": 0.23546242335669995,
"learning_rate": 0.0001999157155593029,
"loss": 0.585,
"step": 390
},
{
"epoch": 0.34408602150537637,
"grad_norm": 0.23135215802858206,
"learning_rate": 0.0001998695972790664,
"loss": 0.5929,
"step": 400
},
{
"epoch": 0.35268817204301073,
"grad_norm": 0.22524299635343834,
"learning_rate": 0.00019981346285904595,
"loss": 0.5976,
"step": 410
},
{
"epoch": 0.36129032258064514,
"grad_norm": 0.21756192058074578,
"learning_rate": 0.00019974731792908506,
"loss": 0.587,
"step": 420
},
{
"epoch": 0.36989247311827955,
"grad_norm": 0.23763473038875538,
"learning_rate": 0.000199671169123003,
"loss": 0.5934,
"step": 430
},
{
"epoch": 0.37849462365591396,
"grad_norm": 0.23701431523985933,
"learning_rate": 0.00019958502407792963,
"loss": 0.6009,
"step": 440
},
{
"epoch": 0.3870967741935484,
"grad_norm": 0.22134697304917678,
"learning_rate": 0.00019948889143353948,
"loss": 0.5994,
"step": 450
},
{
"epoch": 0.3956989247311828,
"grad_norm": 0.2268467235014266,
"learning_rate": 0.00019938278083118517,
"loss": 0.5982,
"step": 460
},
{
"epoch": 0.4043010752688172,
"grad_norm": 0.2247581060586749,
"learning_rate": 0.00019926670291293055,
"loss": 0.5782,
"step": 470
},
{
"epoch": 0.4129032258064516,
"grad_norm": 0.21243655986167898,
"learning_rate": 0.00019914066932048317,
"loss": 0.5874,
"step": 480
},
{
"epoch": 0.421505376344086,
"grad_norm": 0.21385338375703783,
"learning_rate": 0.00019900469269402708,
"loss": 0.5813,
"step": 490
},
{
"epoch": 0.43010752688172044,
"grad_norm": 0.21415772228830113,
"learning_rate": 0.00019885878667095472,
"loss": 0.5934,
"step": 500
},
{
"epoch": 0.43870967741935485,
"grad_norm": 0.23153816801674154,
"learning_rate": 0.00019870296588449948,
"loss": 0.5766,
"step": 510
},
{
"epoch": 0.44731182795698926,
"grad_norm": 0.21768913976009108,
"learning_rate": 0.00019853724596226795,
"loss": 0.5964,
"step": 520
},
{
"epoch": 0.4559139784946237,
"grad_norm": 0.21481929780656356,
"learning_rate": 0.00019836164352467263,
"loss": 0.5748,
"step": 530
},
{
"epoch": 0.4645161290322581,
"grad_norm": 0.29328457757364856,
"learning_rate": 0.00019817617618326516,
"loss": 0.5863,
"step": 540
},
{
"epoch": 0.4731182795698925,
"grad_norm": 0.2140715456768234,
"learning_rate": 0.00019798086253896974,
"loss": 0.5787,
"step": 550
},
{
"epoch": 0.4817204301075269,
"grad_norm": 0.22386132897333755,
"learning_rate": 0.00019777572218021789,
"loss": 0.5863,
"step": 560
},
{
"epoch": 0.49032258064516127,
"grad_norm": 0.23289220288123677,
"learning_rate": 0.00019756077568098372,
"loss": 0.5638,
"step": 570
},
{
"epoch": 0.4989247311827957,
"grad_norm": 0.21115266625450407,
"learning_rate": 0.00019733604459872055,
"loss": 0.5744,
"step": 580
},
{
"epoch": 0.5075268817204301,
"grad_norm": 0.2207246383657776,
"learning_rate": 0.00019710155147219886,
"loss": 0.5789,
"step": 590
},
{
"epoch": 0.5161290322580645,
"grad_norm": 0.2127412426650313,
"learning_rate": 0.00019685731981924592,
"loss": 0.5764,
"step": 600
},
{
"epoch": 0.524731182795699,
"grad_norm": 0.2256297779595538,
"learning_rate": 0.00019660337413438697,
"loss": 0.5798,
"step": 610
},
{
"epoch": 0.5333333333333333,
"grad_norm": 0.22243638154383857,
"learning_rate": 0.00019633973988638877,
"loss": 0.5767,
"step": 620
},
{
"epoch": 0.5419354838709678,
"grad_norm": 0.21425852582870433,
"learning_rate": 0.00019606644351570518,
"loss": 0.5736,
"step": 630
},
{
"epoch": 0.5505376344086022,
"grad_norm": 0.20506928092936336,
"learning_rate": 0.00019578351243182545,
"loss": 0.5766,
"step": 640
},
{
"epoch": 0.5591397849462365,
"grad_norm": 0.2175808639953357,
"learning_rate": 0.0001954909750105252,
"loss": 0.5883,
"step": 650
},
{
"epoch": 0.567741935483871,
"grad_norm": 0.21972174494018643,
"learning_rate": 0.00019518886059102062,
"loss": 0.5774,
"step": 660
},
{
"epoch": 0.5763440860215053,
"grad_norm": 0.2222552681428519,
"learning_rate": 0.0001948771994730259,
"loss": 0.5694,
"step": 670
},
{
"epoch": 0.5849462365591398,
"grad_norm": 0.21552094411185982,
"learning_rate": 0.0001945560229137145,
"loss": 0.5657,
"step": 680
},
{
"epoch": 0.5935483870967742,
"grad_norm": 0.21243053577901072,
"learning_rate": 0.0001942253631245842,
"loss": 0.5776,
"step": 690
},
{
"epoch": 0.6021505376344086,
"grad_norm": 0.21257424919590467,
"learning_rate": 0.00019388525326822665,
"loss": 0.5764,
"step": 700
},
{
"epoch": 0.610752688172043,
"grad_norm": 0.21485742957719725,
"learning_rate": 0.0001935357274550013,
"loss": 0.5821,
"step": 710
},
{
"epoch": 0.6193548387096774,
"grad_norm": 0.2184569847068338,
"learning_rate": 0.00019317682073961444,
"loss": 0.5674,
"step": 720
},
{
"epoch": 0.6279569892473118,
"grad_norm": 0.20912150886711225,
"learning_rate": 0.00019280856911760354,
"loss": 0.559,
"step": 730
},
{
"epoch": 0.6365591397849463,
"grad_norm": 0.21433879312995557,
"learning_rate": 0.00019243100952172723,
"loss": 0.5615,
"step": 740
},
{
"epoch": 0.6451612903225806,
"grad_norm": 0.22214855400737676,
"learning_rate": 0.00019204417981826091,
"loss": 0.5761,
"step": 750
},
{
"epoch": 0.6537634408602151,
"grad_norm": 0.2196194509755129,
"learning_rate": 0.00019164811880319958,
"loss": 0.5718,
"step": 760
},
{
"epoch": 0.6623655913978495,
"grad_norm": 0.2089872234950888,
"learning_rate": 0.00019124286619836637,
"loss": 0.5819,
"step": 770
},
{
"epoch": 0.6709677419354839,
"grad_norm": 0.20958335214398927,
"learning_rate": 0.00019082846264742917,
"loss": 0.5663,
"step": 780
},
{
"epoch": 0.6795698924731183,
"grad_norm": 0.22250020784958888,
"learning_rate": 0.00019040494971182413,
"loss": 0.572,
"step": 790
},
{
"epoch": 0.6881720430107527,
"grad_norm": 0.2162793985278103,
"learning_rate": 0.00018997236986658753,
"loss": 0.5578,
"step": 800
},
{
"epoch": 0.6967741935483871,
"grad_norm": 0.21650281403806723,
"learning_rate": 0.00018953076649609564,
"loss": 0.5698,
"step": 810
},
{
"epoch": 0.7053763440860215,
"grad_norm": 0.21882852535131322,
"learning_rate": 0.000189080183889714,
"loss": 0.5712,
"step": 820
},
{
"epoch": 0.7139784946236559,
"grad_norm": 0.21577728486568806,
"learning_rate": 0.00018862066723735512,
"loss": 0.5676,
"step": 830
},
{
"epoch": 0.7225806451612903,
"grad_norm": 0.20731869739137426,
"learning_rate": 0.00018815226262494655,
"loss": 0.567,
"step": 840
},
{
"epoch": 0.7311827956989247,
"grad_norm": 0.21966746890763253,
"learning_rate": 0.00018767501702980874,
"loss": 0.5738,
"step": 850
},
{
"epoch": 0.7397849462365591,
"grad_norm": 0.2199440815107056,
"learning_rate": 0.00018718897831594355,
"loss": 0.5691,
"step": 860
},
{
"epoch": 0.7483870967741936,
"grad_norm": 0.21722437739899045,
"learning_rate": 0.00018669419522923393,
"loss": 0.5753,
"step": 870
},
{
"epoch": 0.7569892473118279,
"grad_norm": 0.2218916153172574,
"learning_rate": 0.00018619071739255506,
"loss": 0.5572,
"step": 880
},
{
"epoch": 0.7655913978494624,
"grad_norm": 0.20612913226427296,
"learning_rate": 0.00018567859530079753,
"loss": 0.567,
"step": 890
},
{
"epoch": 0.7741935483870968,
"grad_norm": 0.2296891108499081,
"learning_rate": 0.00018515788031580317,
"loss": 0.5735,
"step": 900
},
{
"epoch": 0.7827956989247312,
"grad_norm": 0.20791733581224098,
"learning_rate": 0.0001846286246612138,
"loss": 0.56,
"step": 910
},
{
"epoch": 0.7913978494623656,
"grad_norm": 0.2130914822501822,
"learning_rate": 0.00018409088141723364,
"loss": 0.5588,
"step": 920
},
{
"epoch": 0.8,
"grad_norm": 0.21931892060132505,
"learning_rate": 0.00018354470451530574,
"loss": 0.5718,
"step": 930
},
{
"epoch": 0.8086021505376344,
"grad_norm": 0.20510223609898656,
"learning_rate": 0.00018299014873270314,
"loss": 0.5693,
"step": 940
},
{
"epoch": 0.8172043010752689,
"grad_norm": 0.20754250893604087,
"learning_rate": 0.00018242726968703505,
"loss": 0.5657,
"step": 950
},
{
"epoch": 0.8258064516129032,
"grad_norm": 0.2119047697743998,
"learning_rate": 0.00018185612383066893,
"loss": 0.5523,
"step": 960
},
{
"epoch": 0.8344086021505376,
"grad_norm": 0.21665086623022414,
"learning_rate": 0.00018127676844506874,
"loss": 0.5564,
"step": 970
},
{
"epoch": 0.843010752688172,
"grad_norm": 0.22240006490076028,
"learning_rate": 0.00018068926163505,
"loss": 0.5661,
"step": 980
},
{
"epoch": 0.8516129032258064,
"grad_norm": 0.20224509449303205,
"learning_rate": 0.00018009366232295235,
"loss": 0.5709,
"step": 990
},
{
"epoch": 0.8602150537634409,
"grad_norm": 0.20943438346218987,
"learning_rate": 0.00017949003024273015,
"loss": 0.5567,
"step": 1000
},
{
"epoch": 0.8688172043010752,
"grad_norm": 0.2119577392060311,
"learning_rate": 0.0001788784259339616,
"loss": 0.5726,
"step": 1010
},
{
"epoch": 0.8774193548387097,
"grad_norm": 0.21671277914581005,
"learning_rate": 0.0001782589107357771,
"loss": 0.5721,
"step": 1020
},
{
"epoch": 0.886021505376344,
"grad_norm": 0.21042958546072746,
"learning_rate": 0.00017763154678070733,
"loss": 0.5682,
"step": 1030
},
{
"epoch": 0.8946236559139785,
"grad_norm": 0.21134103487495276,
"learning_rate": 0.0001769963969884521,
"loss": 0.5578,
"step": 1040
},
{
"epoch": 0.9032258064516129,
"grad_norm": 0.2138330925142033,
"learning_rate": 0.0001763535250595696,
"loss": 0.5622,
"step": 1050
},
{
"epoch": 0.9118279569892473,
"grad_norm": 0.20191022832768318,
"learning_rate": 0.00017570299546908812,
"loss": 0.5543,
"step": 1060
},
{
"epoch": 0.9204301075268817,
"grad_norm": 0.20993539174391937,
"learning_rate": 0.0001750448734600394,
"loss": 0.5687,
"step": 1070
},
{
"epoch": 0.9290322580645162,
"grad_norm": 0.21448241595075324,
"learning_rate": 0.0001743792250369155,
"loss": 0.5628,
"step": 1080
},
{
"epoch": 0.9376344086021505,
"grad_norm": 0.21562115673441815,
"learning_rate": 0.00017370611695904895,
"loss": 0.5707,
"step": 1090
},
{
"epoch": 0.946236559139785,
"grad_norm": 0.20800257947078069,
"learning_rate": 0.00017302561673391732,
"loss": 0.5682,
"step": 1100
},
{
"epoch": 0.9548387096774194,
"grad_norm": 0.2142950321860787,
"learning_rate": 0.00017233779261037268,
"loss": 0.5644,
"step": 1110
},
{
"epoch": 0.9634408602150538,
"grad_norm": 0.22988301024268806,
"learning_rate": 0.00017164271357179698,
"loss": 0.5747,
"step": 1120
},
{
"epoch": 0.9720430107526882,
"grad_norm": 0.2044546345524181,
"learning_rate": 0.00017094044932918336,
"loss": 0.5627,
"step": 1130
},
{
"epoch": 0.9806451612903225,
"grad_norm": 0.20801733651601362,
"learning_rate": 0.00017023107031414477,
"loss": 0.5655,
"step": 1140
},
{
"epoch": 0.989247311827957,
"grad_norm": 0.21348629669321226,
"learning_rate": 0.00016951464767185013,
"loss": 0.5569,
"step": 1150
},
{
"epoch": 0.9978494623655914,
"grad_norm": 0.21098078403488563,
"learning_rate": 0.0001687912532538892,
"loss": 0.5686,
"step": 1160
},
{
"epoch": 1.0064516129032257,
"grad_norm": 0.23232328451409845,
"learning_rate": 0.00016806095961106632,
"loss": 0.5352,
"step": 1170
},
{
"epoch": 1.0150537634408603,
"grad_norm": 0.21849412389861325,
"learning_rate": 0.00016732383998612407,
"loss": 0.5166,
"step": 1180
},
{
"epoch": 1.0236559139784946,
"grad_norm": 0.22255257432573036,
"learning_rate": 0.00016657996830639774,
"loss": 0.5286,
"step": 1190
},
{
"epoch": 1.032258064516129,
"grad_norm": 0.23162311614487038,
"learning_rate": 0.0001658294191764009,
"loss": 0.5209,
"step": 1200
},
{
"epoch": 1.0408602150537634,
"grad_norm": 0.23679830183057837,
"learning_rate": 0.0001650722678703432,
"loss": 0.5155,
"step": 1210
},
{
"epoch": 1.049462365591398,
"grad_norm": 0.24431579126710756,
"learning_rate": 0.00016430859032458086,
"loss": 0.5239,
"step": 1220
},
{
"epoch": 1.0580645161290323,
"grad_norm": 0.21925666538005387,
"learning_rate": 0.00016353846313000098,
"loss": 0.5142,
"step": 1230
},
{
"epoch": 1.0666666666666667,
"grad_norm": 0.23676074055821078,
"learning_rate": 0.00016276196352434,
"loss": 0.5267,
"step": 1240
},
{
"epoch": 1.075268817204301,
"grad_norm": 0.23532815378000083,
"learning_rate": 0.00016197916938443733,
"loss": 0.5211,
"step": 1250
},
{
"epoch": 1.0838709677419356,
"grad_norm": 0.23216395268938977,
"learning_rate": 0.00016119015921842503,
"loss": 0.5161,
"step": 1260
},
{
"epoch": 1.09247311827957,
"grad_norm": 0.23822646088799643,
"learning_rate": 0.0001603950121578539,
"loss": 0.5077,
"step": 1270
},
{
"epoch": 1.1010752688172043,
"grad_norm": 0.2290007557059369,
"learning_rate": 0.00015959380794975734,
"loss": 0.5156,
"step": 1280
},
{
"epoch": 1.1096774193548387,
"grad_norm": 0.22484506558131512,
"learning_rate": 0.00015878662694865327,
"loss": 0.5205,
"step": 1290
},
{
"epoch": 1.118279569892473,
"grad_norm": 0.22971689352325197,
"learning_rate": 0.00015797355010848519,
"loss": 0.5218,
"step": 1300
},
{
"epoch": 1.1268817204301076,
"grad_norm": 0.23491219834804952,
"learning_rate": 0.00015715465897450317,
"loss": 0.5248,
"step": 1310
},
{
"epoch": 1.135483870967742,
"grad_norm": 0.23555948179310204,
"learning_rate": 0.00015633003567508557,
"loss": 0.524,
"step": 1320
},
{
"epoch": 1.1440860215053763,
"grad_norm": 0.2275140567256555,
"learning_rate": 0.00015549976291350204,
"loss": 0.517,
"step": 1330
},
{
"epoch": 1.1526881720430107,
"grad_norm": 0.22712302146976637,
"learning_rate": 0.00015466392395961908,
"loss": 0.5243,
"step": 1340
},
{
"epoch": 1.1612903225806452,
"grad_norm": 0.22097671231695554,
"learning_rate": 0.0001538226026415489,
"loss": 0.5118,
"step": 1350
},
{
"epoch": 1.1698924731182796,
"grad_norm": 0.22706370204760404,
"learning_rate": 0.00015297588333724183,
"loss": 0.5235,
"step": 1360
},
{
"epoch": 1.178494623655914,
"grad_norm": 0.2309770159886724,
"learning_rate": 0.00015212385096602415,
"loss": 0.5202,
"step": 1370
},
{
"epoch": 1.1870967741935483,
"grad_norm": 0.23885726185135214,
"learning_rate": 0.0001512665909800811,
"loss": 0.5201,
"step": 1380
},
{
"epoch": 1.1956989247311829,
"grad_norm": 0.23447118491147034,
"learning_rate": 0.00015040418935588682,
"loss": 0.5103,
"step": 1390
},
{
"epoch": 1.2043010752688172,
"grad_norm": 0.24148993682260503,
"learning_rate": 0.0001495367325855816,
"loss": 0.5229,
"step": 1400
},
{
"epoch": 1.2129032258064516,
"grad_norm": 0.22483070355440393,
"learning_rate": 0.00014866430766829743,
"loss": 0.5132,
"step": 1410
},
{
"epoch": 1.221505376344086,
"grad_norm": 0.2254845362019036,
"learning_rate": 0.00014778700210143242,
"loss": 0.5231,
"step": 1420
},
{
"epoch": 1.2301075268817203,
"grad_norm": 0.23952864821588327,
"learning_rate": 0.00014690490387187584,
"loss": 0.5157,
"step": 1430
},
{
"epoch": 1.238709677419355,
"grad_norm": 0.2257464446574845,
"learning_rate": 0.00014601810144718345,
"loss": 0.537,
"step": 1440
},
{
"epoch": 1.2473118279569892,
"grad_norm": 0.23617654028249638,
"learning_rate": 0.0001451266837667051,
"loss": 0.5187,
"step": 1450
},
{
"epoch": 1.2559139784946236,
"grad_norm": 0.23682708876074962,
"learning_rate": 0.00014423074023266463,
"loss": 0.52,
"step": 1460
},
{
"epoch": 1.2645161290322582,
"grad_norm": 0.22896552654447108,
"learning_rate": 0.00014333036070119363,
"loss": 0.5163,
"step": 1470
},
{
"epoch": 1.2731182795698925,
"grad_norm": 0.24027250250612717,
"learning_rate": 0.00014242563547331956,
"loss": 0.517,
"step": 1480
},
{
"epoch": 1.281720430107527,
"grad_norm": 0.23395780934976781,
"learning_rate": 0.00014151665528590925,
"loss": 0.5215,
"step": 1490
},
{
"epoch": 1.2903225806451613,
"grad_norm": 0.24728067879450083,
"learning_rate": 0.0001406035113025687,
"loss": 0.5208,
"step": 1500
},
{
"epoch": 1.2989247311827956,
"grad_norm": 0.22394688366493132,
"learning_rate": 0.0001396862951045001,
"loss": 0.5274,
"step": 1510
},
{
"epoch": 1.3075268817204302,
"grad_norm": 0.23212309124084482,
"learning_rate": 0.00013876509868131692,
"loss": 0.5172,
"step": 1520
},
{
"epoch": 1.3161290322580645,
"grad_norm": 0.23245525514987117,
"learning_rate": 0.0001378400144218181,
"loss": 0.521,
"step": 1530
},
{
"epoch": 1.324731182795699,
"grad_norm": 0.24340497246468923,
"learning_rate": 0.00013691113510472212,
"loss": 0.533,
"step": 1540
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.2321928544760521,
"learning_rate": 0.00013597855388936217,
"loss": 0.5283,
"step": 1550
},
{
"epoch": 1.3419354838709676,
"grad_norm": 0.23326523784187517,
"learning_rate": 0.00013504236430634286,
"loss": 0.5416,
"step": 1560
},
{
"epoch": 1.3505376344086022,
"grad_norm": 0.2262845978211386,
"learning_rate": 0.0001341026602481597,
"loss": 0.5231,
"step": 1570
},
{
"epoch": 1.3591397849462366,
"grad_norm": 0.2371091538942524,
"learning_rate": 0.00013315953595978287,
"loss": 0.5328,
"step": 1580
},
{
"epoch": 1.367741935483871,
"grad_norm": 0.2404491874213034,
"learning_rate": 0.00013221308602920468,
"loss": 0.5273,
"step": 1590
},
{
"epoch": 1.3763440860215055,
"grad_norm": 0.2382492165329928,
"learning_rate": 0.00013126340537795343,
"loss": 0.529,
"step": 1600
},
{
"epoch": 1.3849462365591398,
"grad_norm": 0.22921228109841923,
"learning_rate": 0.0001303105892515734,
"loss": 0.5198,
"step": 1610
},
{
"epoch": 1.3935483870967742,
"grad_norm": 0.22823020260902263,
"learning_rate": 0.00012935473321007255,
"loss": 0.5215,
"step": 1620
},
{
"epoch": 1.4021505376344086,
"grad_norm": 0.23341543795313835,
"learning_rate": 0.0001283959331183386,
"loss": 0.5179,
"step": 1630
},
{
"epoch": 1.410752688172043,
"grad_norm": 0.23483174318348113,
"learning_rate": 0.00012743428513652442,
"loss": 0.5242,
"step": 1640
},
{
"epoch": 1.4193548387096775,
"grad_norm": 0.23015296236307278,
"learning_rate": 0.00012646988571040398,
"loss": 0.522,
"step": 1650
},
{
"epoch": 1.4279569892473118,
"grad_norm": 0.24213814226406005,
"learning_rate": 0.00012550283156169965,
"loss": 0.5238,
"step": 1660
},
{
"epoch": 1.4365591397849462,
"grad_norm": 0.23239320809561437,
"learning_rate": 0.0001245332196783817,
"loss": 0.5156,
"step": 1670
},
{
"epoch": 1.4451612903225808,
"grad_norm": 0.23180542434985962,
"learning_rate": 0.00012356114730494123,
"loss": 0.5184,
"step": 1680
},
{
"epoch": 1.453763440860215,
"grad_norm": 0.23588148083202282,
"learning_rate": 0.00012258671193263716,
"loss": 0.5155,
"step": 1690
},
{
"epoch": 1.4623655913978495,
"grad_norm": 0.23435001337205247,
"learning_rate": 0.00012161001128971879,
"loss": 0.5293,
"step": 1700
},
{
"epoch": 1.4709677419354839,
"grad_norm": 0.22453229677185074,
"learning_rate": 0.00012063114333162438,
"loss": 0.5151,
"step": 1710
},
{
"epoch": 1.4795698924731182,
"grad_norm": 0.2513082837299489,
"learning_rate": 0.00011965020623115688,
"loss": 0.5416,
"step": 1720
},
{
"epoch": 1.4881720430107528,
"grad_norm": 0.23485488435380322,
"learning_rate": 0.0001186672983686381,
"loss": 0.5114,
"step": 1730
},
{
"epoch": 1.4967741935483871,
"grad_norm": 0.2339236059203922,
"learning_rate": 0.00011768251832204187,
"loss": 0.5189,
"step": 1740
},
{
"epoch": 1.5053763440860215,
"grad_norm": 0.2363738308894046,
"learning_rate": 0.00011669596485710741,
"loss": 0.5236,
"step": 1750
},
{
"epoch": 1.513978494623656,
"grad_norm": 0.23458902400188686,
"learning_rate": 0.00011570773691743397,
"loss": 0.5284,
"step": 1760
},
{
"epoch": 1.5225806451612902,
"grad_norm": 0.23836545061013703,
"learning_rate": 0.0001147179336145575,
"loss": 0.5121,
"step": 1770
},
{
"epoch": 1.5311827956989248,
"grad_norm": 0.23790373487214236,
"learning_rate": 0.00011372665421801056,
"loss": 0.5165,
"step": 1780
},
{
"epoch": 1.5397849462365591,
"grad_norm": 0.23929874324218836,
"learning_rate": 0.0001127339981453664,
"loss": 0.5257,
"step": 1790
},
{
"epoch": 1.5483870967741935,
"grad_norm": 0.23994815774490452,
"learning_rate": 0.00011174006495226812,
"loss": 0.5125,
"step": 1800
},
{
"epoch": 1.556989247311828,
"grad_norm": 0.22986043036890028,
"learning_rate": 0.00011074495432244397,
"loss": 0.5142,
"step": 1810
},
{
"epoch": 1.5655913978494622,
"grad_norm": 0.2276160657387438,
"learning_rate": 0.0001097487660577099,
"loss": 0.5198,
"step": 1820
},
{
"epoch": 1.5741935483870968,
"grad_norm": 0.23590750768735258,
"learning_rate": 0.00010875160006796024,
"loss": 0.5203,
"step": 1830
},
{
"epoch": 1.5827956989247312,
"grad_norm": 0.24238494113010198,
"learning_rate": 0.00010775355636114755,
"loss": 0.5288,
"step": 1840
},
{
"epoch": 1.5913978494623655,
"grad_norm": 0.23079601762605023,
"learning_rate": 0.00010675473503325245,
"loss": 0.516,
"step": 1850
},
{
"epoch": 1.6,
"grad_norm": 0.23473655518093983,
"learning_rate": 0.00010575523625824488,
"loss": 0.5245,
"step": 1860
},
{
"epoch": 1.6086021505376344,
"grad_norm": 0.2297499982621768,
"learning_rate": 0.00010475516027803751,
"loss": 0.5162,
"step": 1870
},
{
"epoch": 1.6172043010752688,
"grad_norm": 0.23292165198339548,
"learning_rate": 0.00010375460739243215,
"loss": 0.5246,
"step": 1880
},
{
"epoch": 1.6258064516129034,
"grad_norm": 0.24340787369255004,
"learning_rate": 0.00010275367794906044,
"loss": 0.5099,
"step": 1890
},
{
"epoch": 1.6344086021505375,
"grad_norm": 0.24374482359627925,
"learning_rate": 0.00010175247233331989,
"loss": 0.5097,
"step": 1900
},
{
"epoch": 1.643010752688172,
"grad_norm": 0.23616305839556445,
"learning_rate": 0.00010075109095830584,
"loss": 0.5231,
"step": 1910
},
{
"epoch": 1.6516129032258065,
"grad_norm": 0.23231972508384582,
"learning_rate": 9.974963425474106e-05,
"loss": 0.5213,
"step": 1920
},
{
"epoch": 1.6602150537634408,
"grad_norm": 0.24023523508198666,
"learning_rate": 9.874820266090303e-05,
"loss": 0.5145,
"step": 1930
},
{
"epoch": 1.6688172043010754,
"grad_norm": 0.24186555744758984,
"learning_rate": 9.774689661255106e-05,
"loss": 0.5256,
"step": 1940
},
{
"epoch": 1.6774193548387095,
"grad_norm": 0.2372332084270041,
"learning_rate": 9.67458165328531e-05,
"loss": 0.5094,
"step": 1950
},
{
"epoch": 1.686021505376344,
"grad_norm": 0.2350639603238352,
"learning_rate": 9.574506282231433e-05,
"loss": 0.5119,
"step": 1960
},
{
"epoch": 1.6946236559139785,
"grad_norm": 0.23559422057113213,
"learning_rate": 9.474473584870757e-05,
"loss": 0.5155,
"step": 1970
},
{
"epoch": 1.7032258064516128,
"grad_norm": 0.2336022634685686,
"learning_rate": 9.374493593700723e-05,
"loss": 0.5113,
"step": 1980
},
{
"epoch": 1.7118279569892474,
"grad_norm": 0.23539681381001779,
"learning_rate": 9.274576335932767e-05,
"loss": 0.5189,
"step": 1990
},
{
"epoch": 1.7204301075268817,
"grad_norm": 0.23650024951926094,
"learning_rate": 9.174731832486648e-05,
"loss": 0.5082,
"step": 2000
},
{
"epoch": 1.729032258064516,
"grad_norm": 0.23916533165546086,
"learning_rate": 9.074970096985427e-05,
"loss": 0.5284,
"step": 2010
},
{
"epoch": 1.7376344086021507,
"grad_norm": 0.23608187656307292,
"learning_rate": 8.975301134751202e-05,
"loss": 0.5172,
"step": 2020
},
{
"epoch": 1.7462365591397848,
"grad_norm": 0.23344322072994478,
"learning_rate": 8.87573494180163e-05,
"loss": 0.511,
"step": 2030
},
{
"epoch": 1.7548387096774194,
"grad_norm": 0.24030220361622895,
"learning_rate": 8.77628150384741e-05,
"loss": 0.5254,
"step": 2040
},
{
"epoch": 1.7634408602150538,
"grad_norm": 0.23404925065530963,
"learning_rate": 8.676950795290802e-05,
"loss": 0.5189,
"step": 2050
},
{
"epoch": 1.772043010752688,
"grad_norm": 0.23053178874827976,
"learning_rate": 8.57775277822526e-05,
"loss": 0.5252,
"step": 2060
},
{
"epoch": 1.7806451612903227,
"grad_norm": 0.23781516582822665,
"learning_rate": 8.478697401436323e-05,
"loss": 0.5146,
"step": 2070
},
{
"epoch": 1.789247311827957,
"grad_norm": 0.2276955393985772,
"learning_rate": 8.379794599403836e-05,
"loss": 0.5169,
"step": 2080
},
{
"epoch": 1.7978494623655914,
"grad_norm": 0.23946966468353473,
"learning_rate": 8.281054291305566e-05,
"loss": 0.5181,
"step": 2090
},
{
"epoch": 1.8064516129032258,
"grad_norm": 0.22844184267941248,
"learning_rate": 8.182486380022426e-05,
"loss": 0.5009,
"step": 2100
},
{
"epoch": 1.8150537634408601,
"grad_norm": 0.2386146449957716,
"learning_rate": 8.084100751145277e-05,
"loss": 0.5167,
"step": 2110
},
{
"epoch": 1.8236559139784947,
"grad_norm": 0.22996861067382443,
"learning_rate": 7.985907271983467e-05,
"loss": 0.5172,
"step": 2120
},
{
"epoch": 1.832258064516129,
"grad_norm": 0.23122069068423765,
"learning_rate": 7.887915790575241e-05,
"loss": 0.512,
"step": 2130
},
{
"epoch": 1.8408602150537634,
"grad_norm": 0.226577342689351,
"learning_rate": 7.790136134700042e-05,
"loss": 0.5181,
"step": 2140
},
{
"epoch": 1.849462365591398,
"grad_norm": 0.24205636496242264,
"learning_rate": 7.692578110892876e-05,
"loss": 0.5206,
"step": 2150
},
{
"epoch": 1.8580645161290321,
"grad_norm": 0.24349873078348327,
"learning_rate": 7.595251503460778e-05,
"loss": 0.5226,
"step": 2160
},
{
"epoch": 1.8666666666666667,
"grad_norm": 0.23758597762780134,
"learning_rate": 7.498166073501529e-05,
"loss": 0.5244,
"step": 2170
},
{
"epoch": 1.875268817204301,
"grad_norm": 0.2284958254909026,
"learning_rate": 7.401331557924707e-05,
"loss": 0.505,
"step": 2180
},
{
"epoch": 1.8838709677419354,
"grad_norm": 0.24598875919420438,
"learning_rate": 7.304757668475122e-05,
"loss": 0.5317,
"step": 2190
},
{
"epoch": 1.89247311827957,
"grad_norm": 0.24113280579774263,
"learning_rate": 7.208454090758832e-05,
"loss": 0.5178,
"step": 2200
},
{
"epoch": 1.9010752688172043,
"grad_norm": 0.2362286661628526,
"learning_rate": 7.112430483271746e-05,
"loss": 0.5202,
"step": 2210
},
{
"epoch": 1.9096774193548387,
"grad_norm": 0.23413327678165968,
"learning_rate": 7.016696476430931e-05,
"loss": 0.4944,
"step": 2220
},
{
"epoch": 1.9182795698924733,
"grad_norm": 0.23501629015109748,
"learning_rate": 6.921261671608791e-05,
"loss": 0.5153,
"step": 2230
},
{
"epoch": 1.9268817204301074,
"grad_norm": 0.27066815871872996,
"learning_rate": 6.826135640170101e-05,
"loss": 0.5252,
"step": 2240
},
{
"epoch": 1.935483870967742,
"grad_norm": 0.2439503871040344,
"learning_rate": 6.731327922512074e-05,
"loss": 0.5112,
"step": 2250
},
{
"epoch": 1.9440860215053763,
"grad_norm": 0.23161029568643865,
"learning_rate": 6.636848027107544e-05,
"loss": 0.5307,
"step": 2260
},
{
"epoch": 1.9526881720430107,
"grad_norm": 0.24210935717094484,
"learning_rate": 6.54270542955134e-05,
"loss": 0.5011,
"step": 2270
},
{
"epoch": 1.9612903225806453,
"grad_norm": 0.24063211639329316,
"learning_rate": 6.44890957160994e-05,
"loss": 0.5234,
"step": 2280
},
{
"epoch": 1.9698924731182794,
"grad_norm": 0.23615678793407338,
"learning_rate": 6.355469860274574e-05,
"loss": 0.5046,
"step": 2290
},
{
"epoch": 1.978494623655914,
"grad_norm": 0.22842985809994715,
"learning_rate": 6.262395666817724e-05,
"loss": 0.5012,
"step": 2300
},
{
"epoch": 1.9870967741935484,
"grad_norm": 0.23784853454359128,
"learning_rate": 6.169696325853312e-05,
"loss": 0.519,
"step": 2310
},
{
"epoch": 1.9956989247311827,
"grad_norm": 0.2342469521451372,
"learning_rate": 6.077381134400462e-05,
"loss": 0.5002,
"step": 2320
},
{
"epoch": 2.0043010752688173,
"grad_norm": 0.22833969780513486,
"learning_rate": 5.985459350951121e-05,
"loss": 0.4845,
"step": 2330
},
{
"epoch": 2.0129032258064514,
"grad_norm": 0.2544744796985262,
"learning_rate": 5.893940194541492e-05,
"loss": 0.464,
"step": 2340
},
{
"epoch": 2.021505376344086,
"grad_norm": 0.24382320744489763,
"learning_rate": 5.802832843827419e-05,
"loss": 0.455,
"step": 2350
},
{
"epoch": 2.0301075268817206,
"grad_norm": 0.24697875997895144,
"learning_rate": 5.712146436163863e-05,
"loss": 0.4598,
"step": 2360
},
{
"epoch": 2.0387096774193547,
"grad_norm": 0.2513303656347981,
"learning_rate": 5.6218900666884975e-05,
"loss": 0.4515,
"step": 2370
},
{
"epoch": 2.0473118279569893,
"grad_norm": 0.26236895102262536,
"learning_rate": 5.5320727874095014e-05,
"loss": 0.4598,
"step": 2380
},
{
"epoch": 2.055913978494624,
"grad_norm": 0.25978280232463796,
"learning_rate": 5.4427036062977744e-05,
"loss": 0.451,
"step": 2390
},
{
"epoch": 2.064516129032258,
"grad_norm": 0.26162376545024385,
"learning_rate": 5.3537914863834374e-05,
"loss": 0.4563,
"step": 2400
},
{
"epoch": 2.0731182795698926,
"grad_norm": 0.27200603174186766,
"learning_rate": 5.265345344856979e-05,
"loss": 0.4595,
"step": 2410
},
{
"epoch": 2.0817204301075267,
"grad_norm": 0.2516133186216482,
"learning_rate": 5.1773740521748793e-05,
"loss": 0.4543,
"step": 2420
},
{
"epoch": 2.0903225806451613,
"grad_norm": 0.2559266512535566,
"learning_rate": 5.089886431169999e-05,
"loss": 0.4576,
"step": 2430
},
{
"epoch": 2.098924731182796,
"grad_norm": 0.26643890557668787,
"learning_rate": 5.0028912561667104e-05,
"loss": 0.4651,
"step": 2440
},
{
"epoch": 2.10752688172043,
"grad_norm": 0.2594035563202495,
"learning_rate": 4.916397252100892e-05,
"loss": 0.4581,
"step": 2450
},
{
"epoch": 2.1161290322580646,
"grad_norm": 0.2589087763537241,
"learning_rate": 4.830413093644913e-05,
"loss": 0.4508,
"step": 2460
},
{
"epoch": 2.1247311827956987,
"grad_norm": 0.25564217019549307,
"learning_rate": 4.744947404337605e-05,
"loss": 0.4498,
"step": 2470
},
{
"epoch": 2.1333333333333333,
"grad_norm": 0.25846858564022296,
"learning_rate": 4.660008755719397e-05,
"loss": 0.4439,
"step": 2480
},
{
"epoch": 2.141935483870968,
"grad_norm": 0.2615309859057683,
"learning_rate": 4.5756056664726554e-05,
"loss": 0.4621,
"step": 2490
},
{
"epoch": 2.150537634408602,
"grad_norm": 0.2751137583558904,
"learning_rate": 4.491746601567343e-05,
"loss": 0.4542,
"step": 2500
},
{
"epoch": 2.1591397849462366,
"grad_norm": 0.2640093266145403,
"learning_rate": 4.408439971412013e-05,
"loss": 0.4709,
"step": 2510
},
{
"epoch": 2.167741935483871,
"grad_norm": 0.26804533689921306,
"learning_rate": 4.325694131010346e-05,
"loss": 0.4545,
"step": 2520
},
{
"epoch": 2.1763440860215053,
"grad_norm": 0.26953837584010415,
"learning_rate": 4.243517379123193e-05,
"loss": 0.447,
"step": 2530
},
{
"epoch": 2.18494623655914,
"grad_norm": 0.26331494847521747,
"learning_rate": 4.161917957436271e-05,
"loss": 0.4454,
"step": 2540
},
{
"epoch": 2.193548387096774,
"grad_norm": 0.26650178799391877,
"learning_rate": 4.080904049733607e-05,
"loss": 0.4452,
"step": 2550
},
{
"epoch": 2.2021505376344086,
"grad_norm": 0.2596397199813554,
"learning_rate": 4.0004837810767294e-05,
"loss": 0.455,
"step": 2560
},
{
"epoch": 2.210752688172043,
"grad_norm": 0.2855359315189352,
"learning_rate": 3.9206652169898364e-05,
"loss": 0.4619,
"step": 2570
},
{
"epoch": 2.2193548387096773,
"grad_norm": 0.2715169650377671,
"learning_rate": 3.841456362650837e-05,
"loss": 0.4615,
"step": 2580
},
{
"epoch": 2.227956989247312,
"grad_norm": 0.2628755048951637,
"learning_rate": 3.7628651620885444e-05,
"loss": 0.4645,
"step": 2590
},
{
"epoch": 2.236559139784946,
"grad_norm": 0.28592361434576347,
"learning_rate": 3.6848994973859105e-05,
"loss": 0.4692,
"step": 2600
},
{
"epoch": 2.2451612903225806,
"grad_norm": 0.26764492546621405,
"learning_rate": 3.607567187889538e-05,
"loss": 0.4563,
"step": 2610
},
{
"epoch": 2.253763440860215,
"grad_norm": 0.271421407542708,
"learning_rate": 3.5308759894254496e-05,
"loss": 0.4671,
"step": 2620
},
{
"epoch": 2.2623655913978493,
"grad_norm": 0.2653346076557134,
"learning_rate": 3.45483359352125e-05,
"loss": 0.4535,
"step": 2630
},
{
"epoch": 2.270967741935484,
"grad_norm": 0.2672124362703016,
"learning_rate": 3.379447626634712e-05,
"loss": 0.4523,
"step": 2640
},
{
"epoch": 2.279569892473118,
"grad_norm": 0.2815570906050588,
"learning_rate": 3.304725649388919e-05,
"loss": 0.468,
"step": 2650
},
{
"epoch": 2.2881720430107526,
"grad_norm": 0.27533793807613155,
"learning_rate": 3.230675155813979e-05,
"loss": 0.4587,
"step": 2660
},
{
"epoch": 2.296774193548387,
"grad_norm": 0.2718010567194757,
"learning_rate": 3.1573035725954344e-05,
"loss": 0.4518,
"step": 2670
},
{
"epoch": 2.3053763440860213,
"grad_norm": 0.26881948505806064,
"learning_rate": 3.084618258329443e-05,
"loss": 0.4511,
"step": 2680
},
{
"epoch": 2.313978494623656,
"grad_norm": 0.2735905144552066,
"learning_rate": 3.012626502784729e-05,
"loss": 0.4437,
"step": 2690
},
{
"epoch": 2.3225806451612905,
"grad_norm": 0.2753569674239171,
"learning_rate": 2.9413355261715192e-05,
"loss": 0.457,
"step": 2700
},
{
"epoch": 2.3311827956989246,
"grad_norm": 0.27264140236415646,
"learning_rate": 2.87075247841738e-05,
"loss": 0.4562,
"step": 2710
},
{
"epoch": 2.339784946236559,
"grad_norm": 0.2750362048193341,
"learning_rate": 2.8008844384501566e-05,
"loss": 0.4578,
"step": 2720
},
{
"epoch": 2.3483870967741938,
"grad_norm": 0.2743768587791329,
"learning_rate": 2.7317384134879965e-05,
"loss": 0.4569,
"step": 2730
},
{
"epoch": 2.356989247311828,
"grad_norm": 0.27994254494622967,
"learning_rate": 2.6633213383365906e-05,
"loss": 0.4554,
"step": 2740
},
{
"epoch": 2.3655913978494625,
"grad_norm": 0.273412382006339,
"learning_rate": 2.595640074693664e-05,
"loss": 0.4521,
"step": 2750
},
{
"epoch": 2.3741935483870966,
"grad_norm": 0.26693176903025,
"learning_rate": 2.5287014104607975e-05,
"loss": 0.4472,
"step": 2760
},
{
"epoch": 2.382795698924731,
"grad_norm": 0.26563176753886897,
"learning_rate": 2.4625120590626595e-05,
"loss": 0.4487,
"step": 2770
},
{
"epoch": 2.3913978494623658,
"grad_norm": 0.27008804180981577,
"learning_rate": 2.397078658773699e-05,
"loss": 0.4611,
"step": 2780
},
{
"epoch": 2.4,
"grad_norm": 0.2624172113992881,
"learning_rate": 2.3324077720523785e-05,
"loss": 0.4461,
"step": 2790
},
{
"epoch": 2.4086021505376345,
"grad_norm": 0.27677007427727796,
"learning_rate": 2.2685058848830076e-05,
"loss": 0.4536,
"step": 2800
},
{
"epoch": 2.4172043010752686,
"grad_norm": 0.27569949386731046,
"learning_rate": 2.2053794061252675e-05,
"loss": 0.4474,
"step": 2810
},
{
"epoch": 2.425806451612903,
"grad_norm": 0.267987384900051,
"learning_rate": 2.1430346668714175e-05,
"loss": 0.4487,
"step": 2820
},
{
"epoch": 2.434408602150538,
"grad_norm": 0.2669869480120869,
"learning_rate": 2.0814779198113687e-05,
"loss": 0.455,
"step": 2830
},
{
"epoch": 2.443010752688172,
"grad_norm": 0.26354591329030985,
"learning_rate": 2.020715338605581e-05,
"loss": 0.4616,
"step": 2840
},
{
"epoch": 2.4516129032258065,
"grad_norm": 0.2785720489510887,
"learning_rate": 1.9607530172658715e-05,
"loss": 0.4419,
"step": 2850
},
{
"epoch": 2.4602150537634406,
"grad_norm": 0.27851326696395934,
"learning_rate": 1.9015969695442704e-05,
"loss": 0.4572,
"step": 2860
},
{
"epoch": 2.468817204301075,
"grad_norm": 0.27236418612561275,
"learning_rate": 1.8432531283298458e-05,
"loss": 0.453,
"step": 2870
},
{
"epoch": 2.47741935483871,
"grad_norm": 0.2816067593156165,
"learning_rate": 1.7857273450537227e-05,
"loss": 0.4607,
"step": 2880
},
{
"epoch": 2.486021505376344,
"grad_norm": 0.27424768533392857,
"learning_rate": 1.7290253891022e-05,
"loss": 0.4385,
"step": 2890
},
{
"epoch": 2.4946236559139785,
"grad_norm": 0.2766562755277204,
"learning_rate": 1.673152947238139e-05,
"loss": 0.4567,
"step": 2900
},
{
"epoch": 2.5032258064516126,
"grad_norm": 0.2614813998070769,
"learning_rate": 1.618115623030625e-05,
"loss": 0.4428,
"step": 2910
},
{
"epoch": 2.511827956989247,
"grad_norm": 0.272212168117098,
"learning_rate": 1.5639189362929695e-05,
"loss": 0.453,
"step": 2920
},
{
"epoch": 2.520430107526882,
"grad_norm": 0.2747575768340182,
"learning_rate": 1.5105683225291211e-05,
"loss": 0.4586,
"step": 2930
},
{
"epoch": 2.5290322580645164,
"grad_norm": 0.2711344401379175,
"learning_rate": 1.4580691323885209e-05,
"loss": 0.4506,
"step": 2940
},
{
"epoch": 2.5376344086021505,
"grad_norm": 0.27821711929840565,
"learning_rate": 1.4064266311294793e-05,
"loss": 0.4669,
"step": 2950
},
{
"epoch": 2.546236559139785,
"grad_norm": 0.2807569627250837,
"learning_rate": 1.3556459980911085e-05,
"loss": 0.4567,
"step": 2960
},
{
"epoch": 2.554838709677419,
"grad_norm": 0.27478294823892596,
"learning_rate": 1.305732326173882e-05,
"loss": 0.4629,
"step": 2970
},
{
"epoch": 2.563440860215054,
"grad_norm": 0.2694458161205595,
"learning_rate": 1.2566906213288388e-05,
"loss": 0.4518,
"step": 2980
},
{
"epoch": 2.5720430107526884,
"grad_norm": 0.26379184313364523,
"learning_rate": 1.2085258020555556e-05,
"loss": 0.4489,
"step": 2990
},
{
"epoch": 2.5806451612903225,
"grad_norm": 0.27978762666210216,
"learning_rate": 1.1612426989088232e-05,
"loss": 0.4493,
"step": 3000
},
{
"epoch": 2.589247311827957,
"grad_norm": 0.28005079300575625,
"learning_rate": 1.1148460540142125e-05,
"loss": 0.4501,
"step": 3010
},
{
"epoch": 2.5978494623655912,
"grad_norm": 0.2765800997942907,
"learning_rate": 1.0693405205924579e-05,
"loss": 0.4599,
"step": 3020
},
{
"epoch": 2.606451612903226,
"grad_norm": 0.26856411845690764,
"learning_rate": 1.0247306624927789e-05,
"loss": 0.4489,
"step": 3030
},
{
"epoch": 2.6150537634408604,
"grad_norm": 0.2758438510594412,
"learning_rate": 9.810209537351645e-06,
"loss": 0.4427,
"step": 3040
},
{
"epoch": 2.6236559139784945,
"grad_norm": 0.2808673768423207,
"learning_rate": 9.382157780616606e-06,
"loss": 0.4369,
"step": 3050
},
{
"epoch": 2.632258064516129,
"grad_norm": 0.26355013415952777,
"learning_rate": 8.963194284967202e-06,
"loss": 0.4543,
"step": 3060
},
{
"epoch": 2.6408602150537632,
"grad_norm": 0.26682463219677827,
"learning_rate": 8.553361069166388e-06,
"loss": 0.4448,
"step": 3070
},
{
"epoch": 2.649462365591398,
"grad_norm": 0.2824919005760602,
"learning_rate": 8.15269923628147e-06,
"loss": 0.463,
"step": 3080
},
{
"epoch": 2.6580645161290324,
"grad_norm": 0.2606472077942361,
"learning_rate": 7.761248969561729e-06,
"loss": 0.4435,
"step": 3090
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.2716672911173004,
"learning_rate": 7.379049528408433e-06,
"loss": 0.4436,
"step": 3100
},
{
"epoch": 2.675268817204301,
"grad_norm": 0.2692189798747163,
"learning_rate": 7.00613924443726e-06,
"loss": 0.456,
"step": 3110
},
{
"epoch": 2.6838709677419352,
"grad_norm": 0.2751219149255623,
"learning_rate": 6.642555517634197e-06,
"loss": 0.4486,
"step": 3120
},
{
"epoch": 2.69247311827957,
"grad_norm": 0.26558006981794996,
"learning_rate": 6.288334812604324e-06,
"loss": 0.448,
"step": 3130
},
{
"epoch": 2.7010752688172044,
"grad_norm": 0.2859552474881587,
"learning_rate": 5.943512654914951e-06,
"loss": 0.4357,
"step": 3140
},
{
"epoch": 2.709677419354839,
"grad_norm": 0.27548323875619696,
"learning_rate": 5.6081236275325355e-06,
"loss": 0.4624,
"step": 3150
},
{
"epoch": 2.718279569892473,
"grad_norm": 0.2806542534666262,
"learning_rate": 5.282201367354245e-06,
"loss": 0.4505,
"step": 3160
},
{
"epoch": 2.7268817204301077,
"grad_norm": 0.26475293341781525,
"learning_rate": 4.965778561834644e-06,
"loss": 0.4386,
"step": 3170
},
{
"epoch": 2.735483870967742,
"grad_norm": 0.28440947626379526,
"learning_rate": 4.658886945707164e-06,
"loss": 0.4542,
"step": 3180
},
{
"epoch": 2.7440860215053764,
"grad_norm": 0.2696410246213196,
"learning_rate": 4.361557297801499e-06,
"loss": 0.4416,
"step": 3190
},
{
"epoch": 2.752688172043011,
"grad_norm": 0.27708040093077324,
"learning_rate": 4.073819437956694e-06,
"loss": 0.4471,
"step": 3200
},
{
"epoch": 2.761290322580645,
"grad_norm": 0.2615573424318419,
"learning_rate": 3.7957022240304173e-06,
"loss": 0.4505,
"step": 3210
},
{
"epoch": 2.7698924731182797,
"grad_norm": 0.2776744589099636,
"learning_rate": 3.5272335490047937e-06,
"loss": 0.4416,
"step": 3220
},
{
"epoch": 2.778494623655914,
"grad_norm": 0.28362883428775265,
"learning_rate": 3.2684403381889272e-06,
"loss": 0.4562,
"step": 3230
},
{
"epoch": 2.7870967741935484,
"grad_norm": 0.2659089658104915,
"learning_rate": 3.019348546518508e-06,
"loss": 0.4542,
"step": 3240
},
{
"epoch": 2.795698924731183,
"grad_norm": 0.27817617897785124,
"learning_rate": 2.7799831559527258e-06,
"loss": 0.4535,
"step": 3250
},
{
"epoch": 2.804301075268817,
"grad_norm": 0.2767257855534767,
"learning_rate": 2.550368172968809e-06,
"loss": 0.4438,
"step": 3260
},
{
"epoch": 2.8129032258064517,
"grad_norm": 0.2659520383981123,
"learning_rate": 2.3305266261542945e-06,
"loss": 0.4648,
"step": 3270
},
{
"epoch": 2.821505376344086,
"grad_norm": 0.26860959961537617,
"learning_rate": 2.1204805638975646e-06,
"loss": 0.4478,
"step": 3280
},
{
"epoch": 2.8301075268817204,
"grad_norm": 0.2772290842479072,
"learning_rate": 1.9202510521763696e-06,
"loss": 0.4556,
"step": 3290
},
{
"epoch": 2.838709677419355,
"grad_norm": 0.2773729267266774,
"learning_rate": 1.7298581724452978e-06,
"loss": 0.4546,
"step": 3300
},
{
"epoch": 2.847311827956989,
"grad_norm": 0.288243164539903,
"learning_rate": 1.5493210196216079e-06,
"loss": 0.457,
"step": 3310
},
{
"epoch": 2.8559139784946237,
"grad_norm": 0.27792696429990915,
"learning_rate": 1.378657700170205e-06,
"loss": 0.4585,
"step": 3320
},
{
"epoch": 2.864516129032258,
"grad_norm": 0.27882422779333554,
"learning_rate": 1.2178853302877159e-06,
"loss": 0.4486,
"step": 3330
},
{
"epoch": 2.8731182795698924,
"grad_norm": 0.28133695329057534,
"learning_rate": 1.0670200341858394e-06,
"loss": 0.4502,
"step": 3340
},
{
"epoch": 2.881720430107527,
"grad_norm": 0.2859131700306455,
"learning_rate": 9.260769424742633e-07,
"loss": 0.4452,
"step": 3350
},
{
"epoch": 2.8903225806451616,
"grad_norm": 0.26707708683830167,
"learning_rate": 7.950701906431324e-07,
"loss": 0.4498,
"step": 3360
},
{
"epoch": 2.8989247311827957,
"grad_norm": 0.2716338526967156,
"learning_rate": 6.740129176453725e-07,
"loss": 0.4568,
"step": 3370
},
{
"epoch": 2.90752688172043,
"grad_norm": 0.29044410591630965,
"learning_rate": 5.629172645789882e-07,
"loss": 0.4517,
"step": 3380
},
{
"epoch": 2.9161290322580644,
"grad_norm": 0.27054199049752875,
"learning_rate": 4.617943734694152e-07,
"loss": 0.4559,
"step": 3390
},
{
"epoch": 2.924731182795699,
"grad_norm": 0.27515263341966606,
"learning_rate": 3.7065438615198066e-07,
"loss": 0.4576,
"step": 3400
},
{
"epoch": 2.9333333333333336,
"grad_norm": 0.2653013119464242,
"learning_rate": 2.8950644325485e-07,
"loss": 0.449,
"step": 3410
},
{
"epoch": 2.9419354838709677,
"grad_norm": 0.2791971697586357,
"learning_rate": 2.183586832822493e-07,
"loss": 0.4562,
"step": 3420
},
{
"epoch": 2.9505376344086023,
"grad_norm": 0.2668878384341789,
"learning_rate": 1.572182417982515e-07,
"loss": 0.447,
"step": 3430
},
{
"epoch": 2.9591397849462364,
"grad_norm": 0.2759015749441583,
"learning_rate": 1.0609125071109338e-07,
"loss": 0.4567,
"step": 3440
},
{
"epoch": 2.967741935483871,
"grad_norm": 0.283562302360861,
"learning_rate": 6.49828376582673e-08,
"loss": 0.4503,
"step": 3450
},
{
"epoch": 2.9763440860215056,
"grad_norm": 0.28059419780497596,
"learning_rate": 3.3897125492188266e-08,
"loss": 0.4525,
"step": 3460
},
{
"epoch": 2.9849462365591397,
"grad_norm": 0.2757143204657877,
"learning_rate": 1.2837231866746902e-08,
"loss": 0.4455,
"step": 3470
},
{
"epoch": 2.9935483870967743,
"grad_norm": 0.26729797929321086,
"learning_rate": 1.8052689246150779e-09,
"loss": 0.4443,
"step": 3480
},
{
"epoch": 2.9987096774193547,
"step": 3486,
"total_flos": 6514990399881216.0,
"train_loss": 0.5279938033170147,
"train_runtime": 9440.7969,
"train_samples_per_second": 5.91,
"train_steps_per_second": 0.369
}
],
"logging_steps": 10,
"max_steps": 3486,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6514990399881216.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}