bzhang0426's picture
Upload 12 files
d9812af verified
raw
history blame contribute delete
No virus
88.2 kB
{
"best_metric": 0.3018401563167572,
"best_model_checkpoint": "../../saves/LLaMA3-70B-qlora-bnb/lora/sft/AG_16000-3/checkpoint-3500",
"epoch": 2.9411764705882355,
"eval_steps": 100,
"global_step": 4500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006535947712418301,
"grad_norm": 13.670208930969238,
"learning_rate": 8.714596949891069e-07,
"loss": 2.4115,
"step": 10
},
{
"epoch": 0.013071895424836602,
"grad_norm": 31.536911010742188,
"learning_rate": 2.6143790849673204e-06,
"loss": 2.6282,
"step": 20
},
{
"epoch": 0.0196078431372549,
"grad_norm": 14.474563598632812,
"learning_rate": 4.7930283224400875e-06,
"loss": 2.5659,
"step": 30
},
{
"epoch": 0.026143790849673203,
"grad_norm": 14.182409286499023,
"learning_rate": 6.971677559912855e-06,
"loss": 2.3685,
"step": 40
},
{
"epoch": 0.032679738562091505,
"grad_norm": 34.63951873779297,
"learning_rate": 8.932461873638345e-06,
"loss": 2.351,
"step": 50
},
{
"epoch": 0.0392156862745098,
"grad_norm": 23.15449333190918,
"learning_rate": 1.1111111111111112e-05,
"loss": 2.1125,
"step": 60
},
{
"epoch": 0.0457516339869281,
"grad_norm": 32.01618576049805,
"learning_rate": 1.328976034858388e-05,
"loss": 1.5225,
"step": 70
},
{
"epoch": 0.05228758169934641,
"grad_norm": 8.76276969909668,
"learning_rate": 1.5468409586056645e-05,
"loss": 1.1141,
"step": 80
},
{
"epoch": 0.058823529411764705,
"grad_norm": 12.724370956420898,
"learning_rate": 1.7647058823529414e-05,
"loss": 0.654,
"step": 90
},
{
"epoch": 0.06535947712418301,
"grad_norm": 13.92119312286377,
"learning_rate": 1.982570806100218e-05,
"loss": 0.4776,
"step": 100
},
{
"epoch": 0.06535947712418301,
"eval_loss": 0.460627019405365,
"eval_runtime": 7467.6581,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 100
},
{
"epoch": 0.0718954248366013,
"grad_norm": 13.412103652954102,
"learning_rate": 2.2004357298474944e-05,
"loss": 0.4607,
"step": 110
},
{
"epoch": 0.0784313725490196,
"grad_norm": 16.269054412841797,
"learning_rate": 2.4183006535947712e-05,
"loss": 0.3912,
"step": 120
},
{
"epoch": 0.08496732026143791,
"grad_norm": 5.597168922424316,
"learning_rate": 2.636165577342048e-05,
"loss": 0.3446,
"step": 130
},
{
"epoch": 0.0915032679738562,
"grad_norm": 7.308394432067871,
"learning_rate": 2.854030501089325e-05,
"loss": 0.3572,
"step": 140
},
{
"epoch": 0.09803921568627451,
"grad_norm": 8.47480583190918,
"learning_rate": 3.0718954248366014e-05,
"loss": 0.3224,
"step": 150
},
{
"epoch": 0.10457516339869281,
"grad_norm": 6.073367595672607,
"learning_rate": 3.289760348583878e-05,
"loss": 0.4026,
"step": 160
},
{
"epoch": 0.1111111111111111,
"grad_norm": 6.276689052581787,
"learning_rate": 3.507625272331155e-05,
"loss": 0.367,
"step": 170
},
{
"epoch": 0.11764705882352941,
"grad_norm": 11.574933052062988,
"learning_rate": 3.725490196078432e-05,
"loss": 0.3921,
"step": 180
},
{
"epoch": 0.12418300653594772,
"grad_norm": 11.751296043395996,
"learning_rate": 3.943355119825709e-05,
"loss": 0.3898,
"step": 190
},
{
"epoch": 0.13071895424836602,
"grad_norm": 9.689138412475586,
"learning_rate": 4.161220043572985e-05,
"loss": 0.3675,
"step": 200
},
{
"epoch": 0.13071895424836602,
"eval_loss": 0.3592655658721924,
"eval_runtime": 7465.6804,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 200
},
{
"epoch": 0.13725490196078433,
"grad_norm": 5.775482177734375,
"learning_rate": 4.379084967320262e-05,
"loss": 0.3693,
"step": 210
},
{
"epoch": 0.1437908496732026,
"grad_norm": 12.238544464111328,
"learning_rate": 4.5969498910675387e-05,
"loss": 0.4207,
"step": 220
},
{
"epoch": 0.1503267973856209,
"grad_norm": 6.162591457366943,
"learning_rate": 4.814814814814815e-05,
"loss": 0.3702,
"step": 230
},
{
"epoch": 0.1568627450980392,
"grad_norm": 5.743127346038818,
"learning_rate": 5.032679738562092e-05,
"loss": 0.3505,
"step": 240
},
{
"epoch": 0.16339869281045752,
"grad_norm": 12.115300178527832,
"learning_rate": 5.250544662309368e-05,
"loss": 0.3654,
"step": 250
},
{
"epoch": 0.16993464052287582,
"grad_norm": 12.006166458129883,
"learning_rate": 5.4684095860566454e-05,
"loss": 0.352,
"step": 260
},
{
"epoch": 0.17647058823529413,
"grad_norm": 3.973567247390747,
"learning_rate": 5.6862745098039215e-05,
"loss": 0.36,
"step": 270
},
{
"epoch": 0.1830065359477124,
"grad_norm": 5.307390213012695,
"learning_rate": 5.904139433551199e-05,
"loss": 0.3475,
"step": 280
},
{
"epoch": 0.1895424836601307,
"grad_norm": 5.818578243255615,
"learning_rate": 6.122004357298475e-05,
"loss": 0.4053,
"step": 290
},
{
"epoch": 0.19607843137254902,
"grad_norm": 5.835134506225586,
"learning_rate": 6.339869281045751e-05,
"loss": 0.3761,
"step": 300
},
{
"epoch": 0.19607843137254902,
"eval_loss": 0.4034684896469116,
"eval_runtime": 7466.244,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 300
},
{
"epoch": 0.20261437908496732,
"grad_norm": 3.4767305850982666,
"learning_rate": 6.557734204793029e-05,
"loss": 0.3709,
"step": 310
},
{
"epoch": 0.20915032679738563,
"grad_norm": 5.347959995269775,
"learning_rate": 6.775599128540305e-05,
"loss": 0.35,
"step": 320
},
{
"epoch": 0.21568627450980393,
"grad_norm": 5.1962480545043945,
"learning_rate": 6.993464052287581e-05,
"loss": 0.3892,
"step": 330
},
{
"epoch": 0.2222222222222222,
"grad_norm": 3.1436469554901123,
"learning_rate": 7.211328976034859e-05,
"loss": 0.3538,
"step": 340
},
{
"epoch": 0.22875816993464052,
"grad_norm": 2.677011489868164,
"learning_rate": 7.429193899782135e-05,
"loss": 0.3533,
"step": 350
},
{
"epoch": 0.23529411764705882,
"grad_norm": 11.516694068908691,
"learning_rate": 7.647058823529411e-05,
"loss": 0.4018,
"step": 360
},
{
"epoch": 0.24183006535947713,
"grad_norm": 10.96320629119873,
"learning_rate": 7.864923747276689e-05,
"loss": 0.43,
"step": 370
},
{
"epoch": 0.24836601307189543,
"grad_norm": 6.594156265258789,
"learning_rate": 8.082788671023965e-05,
"loss": 0.3737,
"step": 380
},
{
"epoch": 0.2549019607843137,
"grad_norm": 4.763483047485352,
"learning_rate": 8.300653594771242e-05,
"loss": 0.3629,
"step": 390
},
{
"epoch": 0.26143790849673204,
"grad_norm": 3.6535167694091797,
"learning_rate": 8.518518518518518e-05,
"loss": 0.3465,
"step": 400
},
{
"epoch": 0.26143790849673204,
"eval_loss": 0.35558873414993286,
"eval_runtime": 7465.9696,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 400
},
{
"epoch": 0.2679738562091503,
"grad_norm": 10.275704383850098,
"learning_rate": 8.736383442265795e-05,
"loss": 0.3435,
"step": 410
},
{
"epoch": 0.27450980392156865,
"grad_norm": 6.834077835083008,
"learning_rate": 8.954248366013072e-05,
"loss": 0.3828,
"step": 420
},
{
"epoch": 0.28104575163398693,
"grad_norm": 13.24137020111084,
"learning_rate": 9.172113289760348e-05,
"loss": 0.3704,
"step": 430
},
{
"epoch": 0.2875816993464052,
"grad_norm": 9.123649597167969,
"learning_rate": 9.389978213507626e-05,
"loss": 0.3569,
"step": 440
},
{
"epoch": 0.29411764705882354,
"grad_norm": 6.183751106262207,
"learning_rate": 9.607843137254903e-05,
"loss": 0.4278,
"step": 450
},
{
"epoch": 0.3006535947712418,
"grad_norm": 7.0605645179748535,
"learning_rate": 9.82570806100218e-05,
"loss": 0.3747,
"step": 460
},
{
"epoch": 0.30718954248366015,
"grad_norm": 4.283732891082764,
"learning_rate": 9.999994216519553e-05,
"loss": 0.3289,
"step": 470
},
{
"epoch": 0.3137254901960784,
"grad_norm": 12.275751113891602,
"learning_rate": 9.999791796108715e-05,
"loss": 0.4004,
"step": 480
},
{
"epoch": 0.3202614379084967,
"grad_norm": 6.7879958152771,
"learning_rate": 9.999300215054801e-05,
"loss": 0.3768,
"step": 490
},
{
"epoch": 0.32679738562091504,
"grad_norm": 3.2305805683135986,
"learning_rate": 9.998519501788174e-05,
"loss": 0.394,
"step": 500
},
{
"epoch": 0.32679738562091504,
"eval_loss": 0.3434564173221588,
"eval_runtime": 7464.9222,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 500
},
{
"epoch": 0.3333333333333333,
"grad_norm": 3.8374526500701904,
"learning_rate": 9.997449701461023e-05,
"loss": 0.3625,
"step": 510
},
{
"epoch": 0.33986928104575165,
"grad_norm": 5.077846050262451,
"learning_rate": 9.996090875944755e-05,
"loss": 0.359,
"step": 520
},
{
"epoch": 0.3464052287581699,
"grad_norm": 3.4323623180389404,
"learning_rate": 9.994443103826414e-05,
"loss": 0.356,
"step": 530
},
{
"epoch": 0.35294117647058826,
"grad_norm": 4.085783004760742,
"learning_rate": 9.992506480404138e-05,
"loss": 0.3463,
"step": 540
},
{
"epoch": 0.35947712418300654,
"grad_norm": 7.318523406982422,
"learning_rate": 9.990281117681645e-05,
"loss": 0.3896,
"step": 550
},
{
"epoch": 0.3660130718954248,
"grad_norm": 16.51464080810547,
"learning_rate": 9.987767144361759e-05,
"loss": 0.343,
"step": 560
},
{
"epoch": 0.37254901960784315,
"grad_norm": 2.0446882247924805,
"learning_rate": 9.98496470583896e-05,
"loss": 0.3291,
"step": 570
},
{
"epoch": 0.3790849673202614,
"grad_norm": 2.331265687942505,
"learning_rate": 9.981873964190987e-05,
"loss": 0.3571,
"step": 580
},
{
"epoch": 0.38562091503267976,
"grad_norm": 6.242280006408691,
"learning_rate": 9.978495098169445e-05,
"loss": 0.3624,
"step": 590
},
{
"epoch": 0.39215686274509803,
"grad_norm": 1.5557068586349487,
"learning_rate": 9.974828303189491e-05,
"loss": 0.3579,
"step": 600
},
{
"epoch": 0.39215686274509803,
"eval_loss": 0.341349333524704,
"eval_runtime": 7465.0518,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 600
},
{
"epoch": 0.39869281045751637,
"grad_norm": 4.937715530395508,
"learning_rate": 9.970873791318512e-05,
"loss": 0.3576,
"step": 610
},
{
"epoch": 0.40522875816993464,
"grad_norm": 4.85018253326416,
"learning_rate": 9.966631791263872e-05,
"loss": 0.3567,
"step": 620
},
{
"epoch": 0.4117647058823529,
"grad_norm": 4.347261905670166,
"learning_rate": 9.96210254835968e-05,
"loss": 0.3372,
"step": 630
},
{
"epoch": 0.41830065359477125,
"grad_norm": 4.017812728881836,
"learning_rate": 9.9572863245526e-05,
"loss": 0.3363,
"step": 640
},
{
"epoch": 0.42483660130718953,
"grad_norm": 7.451604843139648,
"learning_rate": 9.952183398386706e-05,
"loss": 0.3269,
"step": 650
},
{
"epoch": 0.43137254901960786,
"grad_norm": 12.005084991455078,
"learning_rate": 9.946794064987371e-05,
"loss": 0.3242,
"step": 660
},
{
"epoch": 0.43790849673202614,
"grad_norm": 3.397099494934082,
"learning_rate": 9.941118636044193e-05,
"loss": 0.383,
"step": 670
},
{
"epoch": 0.4444444444444444,
"grad_norm": 2.081789493560791,
"learning_rate": 9.935157439792982e-05,
"loss": 0.3354,
"step": 680
},
{
"epoch": 0.45098039215686275,
"grad_norm": 3.097064733505249,
"learning_rate": 9.928910820996756e-05,
"loss": 0.3281,
"step": 690
},
{
"epoch": 0.45751633986928103,
"grad_norm": 5.233646869659424,
"learning_rate": 9.922379140925826e-05,
"loss": 0.3524,
"step": 700
},
{
"epoch": 0.45751633986928103,
"eval_loss": 0.3694455921649933,
"eval_runtime": 7465.0105,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 700
},
{
"epoch": 0.46405228758169936,
"grad_norm": 4.417613983154297,
"learning_rate": 9.915562777336879e-05,
"loss": 0.3924,
"step": 710
},
{
"epoch": 0.47058823529411764,
"grad_norm": 2.472074508666992,
"learning_rate": 9.908462124451152e-05,
"loss": 0.3488,
"step": 720
},
{
"epoch": 0.477124183006536,
"grad_norm": 2.744296073913574,
"learning_rate": 9.901077592931612e-05,
"loss": 0.3291,
"step": 730
},
{
"epoch": 0.48366013071895425,
"grad_norm": 3.440641403198242,
"learning_rate": 9.893409609859222e-05,
"loss": 0.3901,
"step": 740
},
{
"epoch": 0.49019607843137253,
"grad_norm": 2.749112844467163,
"learning_rate": 9.88545861870823e-05,
"loss": 0.3535,
"step": 750
},
{
"epoch": 0.49673202614379086,
"grad_norm": 4.953378200531006,
"learning_rate": 9.877225079320526e-05,
"loss": 0.3361,
"step": 760
},
{
"epoch": 0.5032679738562091,
"grad_norm": 2.971996307373047,
"learning_rate": 9.868709467879051e-05,
"loss": 0.3772,
"step": 770
},
{
"epoch": 0.5098039215686274,
"grad_norm": 3.6259024143218994,
"learning_rate": 9.859912276880247e-05,
"loss": 0.3476,
"step": 780
},
{
"epoch": 0.5163398692810458,
"grad_norm": 3.4774110317230225,
"learning_rate": 9.850834015105583e-05,
"loss": 0.3622,
"step": 790
},
{
"epoch": 0.5228758169934641,
"grad_norm": 5.781759262084961,
"learning_rate": 9.841475207592122e-05,
"loss": 0.3554,
"step": 800
},
{
"epoch": 0.5228758169934641,
"eval_loss": 0.351721853017807,
"eval_runtime": 7466.3074,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 800
},
{
"epoch": 0.5294117647058824,
"grad_norm": 2.1952409744262695,
"learning_rate": 9.831836395602163e-05,
"loss": 0.3345,
"step": 810
},
{
"epoch": 0.5359477124183006,
"grad_norm": 5.838998317718506,
"learning_rate": 9.821918136591934e-05,
"loss": 0.3387,
"step": 820
},
{
"epoch": 0.5424836601307189,
"grad_norm": 2.8472695350646973,
"learning_rate": 9.811721004179352e-05,
"loss": 0.3351,
"step": 830
},
{
"epoch": 0.5490196078431373,
"grad_norm": 4.2800068855285645,
"learning_rate": 9.801245588110848e-05,
"loss": 0.3373,
"step": 840
},
{
"epoch": 0.5555555555555556,
"grad_norm": 2.831066370010376,
"learning_rate": 9.790492494227258e-05,
"loss": 0.3718,
"step": 850
},
{
"epoch": 0.5620915032679739,
"grad_norm": 2.9960384368896484,
"learning_rate": 9.779462344428789e-05,
"loss": 0.3391,
"step": 860
},
{
"epoch": 0.5686274509803921,
"grad_norm": 2.505591869354248,
"learning_rate": 9.768155776639044e-05,
"loss": 0.356,
"step": 870
},
{
"epoch": 0.5751633986928104,
"grad_norm": 5.350845813751221,
"learning_rate": 9.756573444768133e-05,
"loss": 0.3377,
"step": 880
},
{
"epoch": 0.5816993464052288,
"grad_norm": 3.700345516204834,
"learning_rate": 9.744716018674862e-05,
"loss": 0.3306,
"step": 890
},
{
"epoch": 0.5882352941176471,
"grad_norm": 3.254387140274048,
"learning_rate": 9.732584184127973e-05,
"loss": 0.3378,
"step": 900
},
{
"epoch": 0.5882352941176471,
"eval_loss": 0.3639741837978363,
"eval_runtime": 7465.3231,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 900
},
{
"epoch": 0.5947712418300654,
"grad_norm": 5.645087242126465,
"learning_rate": 9.7201786427665e-05,
"loss": 0.3604,
"step": 910
},
{
"epoch": 0.6013071895424836,
"grad_norm": 6.0515456199646,
"learning_rate": 9.707500112059183e-05,
"loss": 0.3789,
"step": 920
},
{
"epoch": 0.6078431372549019,
"grad_norm": 7.369515419006348,
"learning_rate": 9.694549325262974e-05,
"loss": 0.3571,
"step": 930
},
{
"epoch": 0.6143790849673203,
"grad_norm": 7.788597106933594,
"learning_rate": 9.681327031380629e-05,
"loss": 0.3424,
"step": 940
},
{
"epoch": 0.6209150326797386,
"grad_norm": 4.71685791015625,
"learning_rate": 9.667833995117391e-05,
"loss": 0.3474,
"step": 950
},
{
"epoch": 0.6274509803921569,
"grad_norm": 4.404577255249023,
"learning_rate": 9.654070996836765e-05,
"loss": 0.3431,
"step": 960
},
{
"epoch": 0.6339869281045751,
"grad_norm": 12.022799491882324,
"learning_rate": 9.640038832515381e-05,
"loss": 0.3246,
"step": 970
},
{
"epoch": 0.6405228758169934,
"grad_norm": 3.9037022590637207,
"learning_rate": 9.625738313696966e-05,
"loss": 0.3394,
"step": 980
},
{
"epoch": 0.6470588235294118,
"grad_norm": 3.2752866744995117,
"learning_rate": 9.611170267445401e-05,
"loss": 0.376,
"step": 990
},
{
"epoch": 0.6535947712418301,
"grad_norm": 4.711609840393066,
"learning_rate": 9.596335536296897e-05,
"loss": 0.3245,
"step": 1000
},
{
"epoch": 0.6535947712418301,
"eval_loss": 0.32997554540634155,
"eval_runtime": 7465.3556,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1000
},
{
"epoch": 0.6601307189542484,
"grad_norm": 1.1813262701034546,
"learning_rate": 9.581234978211257e-05,
"loss": 0.3446,
"step": 1010
},
{
"epoch": 0.6666666666666666,
"grad_norm": 3.696199417114258,
"learning_rate": 9.565869466522265e-05,
"loss": 0.3484,
"step": 1020
},
{
"epoch": 0.673202614379085,
"grad_norm": 1.4563554525375366,
"learning_rate": 9.550239889887179e-05,
"loss": 0.3155,
"step": 1030
},
{
"epoch": 0.6797385620915033,
"grad_norm": 1.6707898378372192,
"learning_rate": 9.534347152235317e-05,
"loss": 0.3364,
"step": 1040
},
{
"epoch": 0.6862745098039216,
"grad_norm": 3.146296977996826,
"learning_rate": 9.518192172715807e-05,
"loss": 0.3286,
"step": 1050
},
{
"epoch": 0.6928104575163399,
"grad_norm": 8.924692153930664,
"learning_rate": 9.501775885644405e-05,
"loss": 0.3337,
"step": 1060
},
{
"epoch": 0.6993464052287581,
"grad_norm": 4.0824480056762695,
"learning_rate": 9.485099240449474e-05,
"loss": 0.328,
"step": 1070
},
{
"epoch": 0.7058823529411765,
"grad_norm": 2.718278408050537,
"learning_rate": 9.468163201617062e-05,
"loss": 0.3186,
"step": 1080
},
{
"epoch": 0.7124183006535948,
"grad_norm": 3.190894603729248,
"learning_rate": 9.450968748635133e-05,
"loss": 0.3397,
"step": 1090
},
{
"epoch": 0.7189542483660131,
"grad_norm": 3.299405813217163,
"learning_rate": 9.433516875936916e-05,
"loss": 0.4178,
"step": 1100
},
{
"epoch": 0.7189542483660131,
"eval_loss": 0.3399566113948822,
"eval_runtime": 7464.7626,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1100
},
{
"epoch": 0.7254901960784313,
"grad_norm": 2.694213628768921,
"learning_rate": 9.415808592843383e-05,
"loss": 0.3442,
"step": 1110
},
{
"epoch": 0.7320261437908496,
"grad_norm": 1.782759428024292,
"learning_rate": 9.397844923504885e-05,
"loss": 0.2938,
"step": 1120
},
{
"epoch": 0.738562091503268,
"grad_norm": 2.5469982624053955,
"learning_rate": 9.37962690684192e-05,
"loss": 0.3099,
"step": 1130
},
{
"epoch": 0.7450980392156863,
"grad_norm": 4.2174906730651855,
"learning_rate": 9.361155596485046e-05,
"loss": 0.3695,
"step": 1140
},
{
"epoch": 0.7516339869281046,
"grad_norm": 2.9205563068389893,
"learning_rate": 9.342432060713942e-05,
"loss": 0.3253,
"step": 1150
},
{
"epoch": 0.7581699346405228,
"grad_norm": 1.3839212656021118,
"learning_rate": 9.323457382395628e-05,
"loss": 0.3306,
"step": 1160
},
{
"epoch": 0.7647058823529411,
"grad_norm": 3.993194341659546,
"learning_rate": 9.304232658921839e-05,
"loss": 0.3048,
"step": 1170
},
{
"epoch": 0.7712418300653595,
"grad_norm": 3.2434442043304443,
"learning_rate": 9.284759002145552e-05,
"loss": 0.3235,
"step": 1180
},
{
"epoch": 0.7777777777777778,
"grad_norm": 6.1412153244018555,
"learning_rate": 9.26503753831669e-05,
"loss": 0.2977,
"step": 1190
},
{
"epoch": 0.7843137254901961,
"grad_norm": 6.04712438583374,
"learning_rate": 9.245069408016977e-05,
"loss": 0.2912,
"step": 1200
},
{
"epoch": 0.7843137254901961,
"eval_loss": 0.34279727935791016,
"eval_runtime": 7466.0217,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1200
},
{
"epoch": 0.7908496732026143,
"grad_norm": 1.9226710796356201,
"learning_rate": 9.224855766093985e-05,
"loss": 0.3681,
"step": 1210
},
{
"epoch": 0.7973856209150327,
"grad_norm": 2.343475341796875,
"learning_rate": 9.204397781594331e-05,
"loss": 0.3081,
"step": 1220
},
{
"epoch": 0.803921568627451,
"grad_norm": 4.996776103973389,
"learning_rate": 9.183696637696077e-05,
"loss": 0.3268,
"step": 1230
},
{
"epoch": 0.8104575163398693,
"grad_norm": 3.761845111846924,
"learning_rate": 9.162753531640292e-05,
"loss": 1.001,
"step": 1240
},
{
"epoch": 0.8169934640522876,
"grad_norm": 13.737248420715332,
"learning_rate": 9.141569674661817e-05,
"loss": 0.3769,
"step": 1250
},
{
"epoch": 0.8235294117647058,
"grad_norm": 8.417398452758789,
"learning_rate": 9.120146291919204e-05,
"loss": 0.402,
"step": 1260
},
{
"epoch": 0.8300653594771242,
"grad_norm": 3.1971797943115234,
"learning_rate": 9.098484622423882e-05,
"loss": 0.3705,
"step": 1270
},
{
"epoch": 0.8366013071895425,
"grad_norm": 8.737996101379395,
"learning_rate": 9.076585918968468e-05,
"loss": 0.3536,
"step": 1280
},
{
"epoch": 0.8431372549019608,
"grad_norm": 3.097640037536621,
"learning_rate": 9.054451448054335e-05,
"loss": 0.3847,
"step": 1290
},
{
"epoch": 0.8496732026143791,
"grad_norm": 3.883129596710205,
"learning_rate": 9.03208248981836e-05,
"loss": 0.3905,
"step": 1300
},
{
"epoch": 0.8496732026143791,
"eval_loss": 0.36324241757392883,
"eval_runtime": 7459.1059,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1300
},
{
"epoch": 0.8562091503267973,
"grad_norm": 3.2262656688690186,
"learning_rate": 9.009480337958883e-05,
"loss": 0.351,
"step": 1310
},
{
"epoch": 0.8627450980392157,
"grad_norm": 2.727170705795288,
"learning_rate": 8.986646299660889e-05,
"loss": 0.3583,
"step": 1320
},
{
"epoch": 0.869281045751634,
"grad_norm": 6.688016414642334,
"learning_rate": 8.963581695520408e-05,
"loss": 0.3513,
"step": 1330
},
{
"epoch": 0.8758169934640523,
"grad_norm": 1.6296730041503906,
"learning_rate": 8.940287859468139e-05,
"loss": 0.3563,
"step": 1340
},
{
"epoch": 0.8823529411764706,
"grad_norm": 1.2997907400131226,
"learning_rate": 8.916766138692303e-05,
"loss": 0.3741,
"step": 1350
},
{
"epoch": 0.8888888888888888,
"grad_norm": 7.427231788635254,
"learning_rate": 8.893017893560727e-05,
"loss": 0.3419,
"step": 1360
},
{
"epoch": 0.8954248366013072,
"grad_norm": 1.4134552478790283,
"learning_rate": 8.869044497542172e-05,
"loss": 0.3585,
"step": 1370
},
{
"epoch": 0.9019607843137255,
"grad_norm": 1.9037864208221436,
"learning_rate": 8.844847337126893e-05,
"loss": 0.3563,
"step": 1380
},
{
"epoch": 0.9084967320261438,
"grad_norm": 2.263465404510498,
"learning_rate": 8.820427811746456e-05,
"loss": 0.3429,
"step": 1390
},
{
"epoch": 0.9150326797385621,
"grad_norm": 3.114661455154419,
"learning_rate": 8.795787333692807e-05,
"loss": 0.3389,
"step": 1400
},
{
"epoch": 0.9150326797385621,
"eval_loss": 0.35517746210098267,
"eval_runtime": 7467.6922,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1400
},
{
"epoch": 0.9215686274509803,
"grad_norm": 2.826345682144165,
"learning_rate": 8.770927328036575e-05,
"loss": 0.3097,
"step": 1410
},
{
"epoch": 0.9281045751633987,
"grad_norm": 2.172060012817383,
"learning_rate": 8.745849232544681e-05,
"loss": 0.3014,
"step": 1420
},
{
"epoch": 0.934640522875817,
"grad_norm": 2.7672016620635986,
"learning_rate": 8.720554497597159e-05,
"loss": 0.3057,
"step": 1430
},
{
"epoch": 0.9411764705882353,
"grad_norm": 6.170320510864258,
"learning_rate": 8.695044586103296e-05,
"loss": 0.3334,
"step": 1440
},
{
"epoch": 0.9477124183006536,
"grad_norm": 3.521531581878662,
"learning_rate": 8.669320973417006e-05,
"loss": 0.3057,
"step": 1450
},
{
"epoch": 0.954248366013072,
"grad_norm": 3.1676554679870605,
"learning_rate": 8.643385147251515e-05,
"loss": 0.2932,
"step": 1460
},
{
"epoch": 0.9607843137254902,
"grad_norm": 2.741800546646118,
"learning_rate": 8.617238607593319e-05,
"loss": 0.352,
"step": 1470
},
{
"epoch": 0.9673202614379085,
"grad_norm": 3.750753164291382,
"learning_rate": 8.590882866615432e-05,
"loss": 0.329,
"step": 1480
},
{
"epoch": 0.9738562091503268,
"grad_norm": 4.675954818725586,
"learning_rate": 8.564319448589926e-05,
"loss": 0.2993,
"step": 1490
},
{
"epoch": 0.9803921568627451,
"grad_norm": 3.0706264972686768,
"learning_rate": 8.537549889799781e-05,
"loss": 0.3422,
"step": 1500
},
{
"epoch": 0.9803921568627451,
"eval_loss": 0.35487601161003113,
"eval_runtime": 7465.2366,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1500
},
{
"epoch": 0.9869281045751634,
"grad_norm": 4.867375373840332,
"learning_rate": 8.510575738450032e-05,
"loss": 0.326,
"step": 1510
},
{
"epoch": 0.9934640522875817,
"grad_norm": 4.509498119354248,
"learning_rate": 8.483398554578232e-05,
"loss": 0.2879,
"step": 1520
},
{
"epoch": 1.0,
"grad_norm": 5.889903545379639,
"learning_rate": 8.456019909964224e-05,
"loss": 0.2913,
"step": 1530
},
{
"epoch": 1.0065359477124183,
"grad_norm": 3.161592960357666,
"learning_rate": 8.428441388039238e-05,
"loss": 0.3207,
"step": 1540
},
{
"epoch": 1.0130718954248366,
"grad_norm": 6.485883712768555,
"learning_rate": 8.400664583794319e-05,
"loss": 0.3892,
"step": 1550
},
{
"epoch": 1.0196078431372548,
"grad_norm": 2.0741546154022217,
"learning_rate": 8.372691103688079e-05,
"loss": 0.2883,
"step": 1560
},
{
"epoch": 1.026143790849673,
"grad_norm": 1.5439857244491577,
"learning_rate": 8.34452256555378e-05,
"loss": 0.3029,
"step": 1570
},
{
"epoch": 1.0326797385620916,
"grad_norm": 4.525278568267822,
"learning_rate": 8.316160598505784e-05,
"loss": 0.2775,
"step": 1580
},
{
"epoch": 1.0392156862745099,
"grad_norm": 3.212484121322632,
"learning_rate": 8.28760684284532e-05,
"loss": 0.3194,
"step": 1590
},
{
"epoch": 1.0457516339869282,
"grad_norm": 3.7706832885742188,
"learning_rate": 8.25886294996562e-05,
"loss": 0.2597,
"step": 1600
},
{
"epoch": 1.0457516339869282,
"eval_loss": 0.35163000226020813,
"eval_runtime": 7465.7169,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1600
},
{
"epoch": 1.0522875816993464,
"grad_norm": 3.6558139324188232,
"learning_rate": 8.22993058225642e-05,
"loss": 0.3301,
"step": 1610
},
{
"epoch": 1.0588235294117647,
"grad_norm": 2.815463066101074,
"learning_rate": 8.200811413007807e-05,
"loss": 0.2903,
"step": 1620
},
{
"epoch": 1.065359477124183,
"grad_norm": 6.215457439422607,
"learning_rate": 8.171507126313451e-05,
"loss": 0.3399,
"step": 1630
},
{
"epoch": 1.0718954248366013,
"grad_norm": 1.9962519407272339,
"learning_rate": 8.142019416973199e-05,
"loss": 0.3062,
"step": 1640
},
{
"epoch": 1.0784313725490196,
"grad_norm": 1.8813626766204834,
"learning_rate": 8.112349990395065e-05,
"loss": 0.3419,
"step": 1650
},
{
"epoch": 1.0849673202614378,
"grad_norm": 5.976492881774902,
"learning_rate": 8.082500562496596e-05,
"loss": 0.3135,
"step": 1660
},
{
"epoch": 1.091503267973856,
"grad_norm": 2.2381093502044678,
"learning_rate": 8.052472859605631e-05,
"loss": 0.3222,
"step": 1670
},
{
"epoch": 1.0980392156862746,
"grad_norm": 3.5524046421051025,
"learning_rate": 8.02226861836046e-05,
"loss": 0.2798,
"step": 1680
},
{
"epoch": 1.1045751633986929,
"grad_norm": 5.400121688842773,
"learning_rate": 7.991889585609387e-05,
"loss": 0.3218,
"step": 1690
},
{
"epoch": 1.1111111111111112,
"grad_norm": 37.609107971191406,
"learning_rate": 7.961337518309704e-05,
"loss": 0.3235,
"step": 1700
},
{
"epoch": 1.1111111111111112,
"eval_loss": 0.325325071811676,
"eval_runtime": 7465.0207,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1700
},
{
"epoch": 1.1176470588235294,
"grad_norm": 2.548970937728882,
"learning_rate": 7.930614183426074e-05,
"loss": 0.325,
"step": 1710
},
{
"epoch": 1.1241830065359477,
"grad_norm": 2.6757702827453613,
"learning_rate": 7.89972135782834e-05,
"loss": 0.2955,
"step": 1720
},
{
"epoch": 1.130718954248366,
"grad_norm": 2.190845012664795,
"learning_rate": 7.868660828188765e-05,
"loss": 0.3062,
"step": 1730
},
{
"epoch": 1.1372549019607843,
"grad_norm": 2.709989547729492,
"learning_rate": 7.837434390878698e-05,
"loss": 0.3231,
"step": 1740
},
{
"epoch": 1.1437908496732025,
"grad_norm": 2.8716742992401123,
"learning_rate": 7.806043851864674e-05,
"loss": 0.2853,
"step": 1750
},
{
"epoch": 1.1503267973856208,
"grad_norm": 4.456597805023193,
"learning_rate": 7.774491026603985e-05,
"loss": 0.3226,
"step": 1760
},
{
"epoch": 1.156862745098039,
"grad_norm": 13.620016098022461,
"learning_rate": 7.742777739939666e-05,
"loss": 0.2897,
"step": 1770
},
{
"epoch": 1.1633986928104576,
"grad_norm": 2.6009891033172607,
"learning_rate": 7.710905825994962e-05,
"loss": 0.2482,
"step": 1780
},
{
"epoch": 1.1699346405228759,
"grad_norm": 7.119050025939941,
"learning_rate": 7.678877128067261e-05,
"loss": 0.3552,
"step": 1790
},
{
"epoch": 1.1764705882352942,
"grad_norm": 2.4954652786254883,
"learning_rate": 7.646693498521471e-05,
"loss": 0.3148,
"step": 1800
},
{
"epoch": 1.1764705882352942,
"eval_loss": 0.31464409828186035,
"eval_runtime": 7466.0649,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1800
},
{
"epoch": 1.1830065359477124,
"grad_norm": 2.542064905166626,
"learning_rate": 7.614356798682904e-05,
"loss": 0.2973,
"step": 1810
},
{
"epoch": 1.1895424836601307,
"grad_norm": 2.8179709911346436,
"learning_rate": 7.581868898729618e-05,
"loss": 0.3049,
"step": 1820
},
{
"epoch": 1.196078431372549,
"grad_norm": 1.8893694877624512,
"learning_rate": 7.549231677584262e-05,
"loss": 0.2867,
"step": 1830
},
{
"epoch": 1.2026143790849673,
"grad_norm": 3.376964807510376,
"learning_rate": 7.516447022805407e-05,
"loss": 0.3599,
"step": 1840
},
{
"epoch": 1.2091503267973855,
"grad_norm": 1.8863316774368286,
"learning_rate": 7.483516830478379e-05,
"loss": 0.3283,
"step": 1850
},
{
"epoch": 1.215686274509804,
"grad_norm": 5.017312049865723,
"learning_rate": 7.450443005105601e-05,
"loss": 0.3335,
"step": 1860
},
{
"epoch": 1.2222222222222223,
"grad_norm": 1.3233927488327026,
"learning_rate": 7.417227459496445e-05,
"loss": 0.3031,
"step": 1870
},
{
"epoch": 1.2287581699346406,
"grad_norm": 2.100111484527588,
"learning_rate": 7.383872114656611e-05,
"loss": 0.3251,
"step": 1880
},
{
"epoch": 1.2352941176470589,
"grad_norm": 3.839590311050415,
"learning_rate": 7.35037889967702e-05,
"loss": 0.2726,
"step": 1890
},
{
"epoch": 1.2418300653594772,
"grad_norm": 3.1495704650878906,
"learning_rate": 7.31674975162225e-05,
"loss": 0.3502,
"step": 1900
},
{
"epoch": 1.2418300653594772,
"eval_loss": 0.3091997504234314,
"eval_runtime": 7466.4145,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 1900
},
{
"epoch": 1.2483660130718954,
"grad_norm": 5.134574890136719,
"learning_rate": 7.282986615418503e-05,
"loss": 0.2902,
"step": 1910
},
{
"epoch": 1.2549019607843137,
"grad_norm": 2.6781771183013916,
"learning_rate": 7.249091443741126e-05,
"loss": 0.2741,
"step": 1920
},
{
"epoch": 1.261437908496732,
"grad_norm": 6.247425079345703,
"learning_rate": 7.215066196901676e-05,
"loss": 0.3026,
"step": 1930
},
{
"epoch": 1.2679738562091503,
"grad_norm": 3.1725962162017822,
"learning_rate": 7.180912842734548e-05,
"loss": 0.3574,
"step": 1940
},
{
"epoch": 1.2745098039215685,
"grad_norm": 5.736270427703857,
"learning_rate": 7.146633356483161e-05,
"loss": 0.3166,
"step": 1950
},
{
"epoch": 1.2810457516339868,
"grad_norm": 1.851462960243225,
"learning_rate": 7.11222972068573e-05,
"loss": 0.2737,
"step": 1960
},
{
"epoch": 1.287581699346405,
"grad_norm": 2.9439609050750732,
"learning_rate": 7.077703925060594e-05,
"loss": 0.3556,
"step": 1970
},
{
"epoch": 1.2941176470588236,
"grad_norm": 3.4385366439819336,
"learning_rate": 7.043057966391157e-05,
"loss": 0.3269,
"step": 1980
},
{
"epoch": 1.3006535947712419,
"grad_norm": 3.6838247776031494,
"learning_rate": 7.008293848410396e-05,
"loss": 0.3342,
"step": 1990
},
{
"epoch": 1.3071895424836601,
"grad_norm": 1.6612685918807983,
"learning_rate": 6.973413581684972e-05,
"loss": 0.3009,
"step": 2000
},
{
"epoch": 1.3071895424836601,
"eval_loss": 0.33713921904563904,
"eval_runtime": 7466.3969,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2000
},
{
"epoch": 1.3137254901960784,
"grad_norm": 6.528134822845459,
"learning_rate": 6.945427096600644e-05,
"loss": 0.3152,
"step": 2010
},
{
"epoch": 1.3202614379084967,
"grad_norm": 1.521199345588684,
"learning_rate": 6.913856280078387e-05,
"loss": 0.3429,
"step": 2020
},
{
"epoch": 1.326797385620915,
"grad_norm": 1.8718252182006836,
"learning_rate": 6.878672506701743e-05,
"loss": 0.3259,
"step": 2030
},
{
"epoch": 1.3333333333333333,
"grad_norm": 2.5089643001556396,
"learning_rate": 6.846914154947518e-05,
"loss": 0.3482,
"step": 2040
},
{
"epoch": 1.3398692810457518,
"grad_norm": 3.0759692192077637,
"learning_rate": 6.811525687459402e-05,
"loss": 0.3382,
"step": 2050
},
{
"epoch": 1.34640522875817,
"grad_norm": 4.5103044509887695,
"learning_rate": 6.776032451222086e-05,
"loss": 0.3476,
"step": 2060
},
{
"epoch": 1.3529411764705883,
"grad_norm": 7.423872470855713,
"learning_rate": 6.740436498970452e-05,
"loss": 0.3198,
"step": 2070
},
{
"epoch": 1.3594771241830066,
"grad_norm": 2.7317986488342285,
"learning_rate": 6.704739889379915e-05,
"loss": 0.279,
"step": 2080
},
{
"epoch": 1.3660130718954249,
"grad_norm": 2.618377923965454,
"learning_rate": 6.668944686947354e-05,
"loss": 0.328,
"step": 2090
},
{
"epoch": 1.3725490196078431,
"grad_norm": 19.447980880737305,
"learning_rate": 6.636646418781669e-05,
"loss": 0.3247,
"step": 2100
},
{
"epoch": 1.3725490196078431,
"eval_loss": 0.3066463768482208,
"eval_runtime": 7466.6542,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2100
},
{
"epoch": 1.3790849673202614,
"grad_norm": 1.8498194217681885,
"learning_rate": 6.600669597963677e-05,
"loss": 0.2986,
"step": 2110
},
{
"epoch": 1.3856209150326797,
"grad_norm": 1.9933720827102661,
"learning_rate": 6.564600203160773e-05,
"loss": 0.3237,
"step": 2120
},
{
"epoch": 1.392156862745098,
"grad_norm": 3.7571139335632324,
"learning_rate": 6.5284403204297e-05,
"loss": 0.2923,
"step": 2130
},
{
"epoch": 1.3986928104575163,
"grad_norm": 10.47822380065918,
"learning_rate": 6.49219204106053e-05,
"loss": 0.3003,
"step": 2140
},
{
"epoch": 1.4052287581699345,
"grad_norm": 2.708207607269287,
"learning_rate": 6.459494743075649e-05,
"loss": 0.269,
"step": 2150
},
{
"epoch": 1.4117647058823528,
"grad_norm": 7.0187153816223145,
"learning_rate": 6.423084289810528e-05,
"loss": 0.2916,
"step": 2160
},
{
"epoch": 1.4183006535947713,
"grad_norm": 4.117656707763672,
"learning_rate": 6.38659153312455e-05,
"loss": 0.3612,
"step": 2170
},
{
"epoch": 1.4248366013071896,
"grad_norm": 4.099621295928955,
"learning_rate": 6.350018583559398e-05,
"loss": 0.3231,
"step": 2180
},
{
"epoch": 1.4313725490196079,
"grad_norm": 3.927912712097168,
"learning_rate": 6.313367556294672e-05,
"loss": 0.2905,
"step": 2190
},
{
"epoch": 1.4379084967320261,
"grad_norm": 2.0472779273986816,
"learning_rate": 6.276640571025562e-05,
"loss": 0.3047,
"step": 2200
},
{
"epoch": 1.4379084967320261,
"eval_loss": 0.310788094997406,
"eval_runtime": 7464.5268,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2200
},
{
"epoch": 1.4444444444444444,
"grad_norm": 3.7081167697906494,
"learning_rate": 6.23983975184025e-05,
"loss": 0.2777,
"step": 2210
},
{
"epoch": 1.4509803921568627,
"grad_norm": 2.031965494155884,
"learning_rate": 6.202967227097073e-05,
"loss": 0.2724,
"step": 2220
},
{
"epoch": 1.457516339869281,
"grad_norm": 2.594900369644165,
"learning_rate": 6.166025129301422e-05,
"loss": 0.3403,
"step": 2230
},
{
"epoch": 1.4640522875816995,
"grad_norm": 10.884799003601074,
"learning_rate": 6.129015594982416e-05,
"loss": 0.2926,
"step": 2240
},
{
"epoch": 1.4705882352941178,
"grad_norm": 3.915910243988037,
"learning_rate": 6.091940764569331e-05,
"loss": 0.3651,
"step": 2250
},
{
"epoch": 1.477124183006536,
"grad_norm": 16.43703269958496,
"learning_rate": 6.054802782267815e-05,
"loss": 0.3281,
"step": 2260
},
{
"epoch": 1.4836601307189543,
"grad_norm": 2.157759666442871,
"learning_rate": 6.0176037959358775e-05,
"loss": 0.3234,
"step": 2270
},
{
"epoch": 1.4901960784313726,
"grad_norm": 8.77623176574707,
"learning_rate": 5.980345956959663e-05,
"loss": 0.2683,
"step": 2280
},
{
"epoch": 1.4967320261437909,
"grad_norm": 10.245627403259277,
"learning_rate": 5.943031420129036e-05,
"loss": 0.3416,
"step": 2290
},
{
"epoch": 1.5032679738562091,
"grad_norm": 2.9165868759155273,
"learning_rate": 5.905662343512952e-05,
"loss": 0.3082,
"step": 2300
},
{
"epoch": 1.5032679738562091,
"eval_loss": 0.30869877338409424,
"eval_runtime": 7463.6998,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2300
},
{
"epoch": 1.5098039215686274,
"grad_norm": 2.973310947418213,
"learning_rate": 5.868240888334653e-05,
"loss": 0.3351,
"step": 2310
},
{
"epoch": 1.5163398692810457,
"grad_norm": 1.646573543548584,
"learning_rate": 5.8307692188466656e-05,
"loss": 0.303,
"step": 2320
},
{
"epoch": 1.522875816993464,
"grad_norm": 3.7888035774230957,
"learning_rate": 5.7932495022056454e-05,
"loss": 0.3052,
"step": 2330
},
{
"epoch": 1.5294117647058822,
"grad_norm": 4.961613655090332,
"learning_rate": 5.755683908347026e-05,
"loss": 0.3076,
"step": 2340
},
{
"epoch": 1.5359477124183005,
"grad_norm": 2.823336124420166,
"learning_rate": 5.7180746098595325e-05,
"loss": 0.2848,
"step": 2350
},
{
"epoch": 1.5424836601307188,
"grad_norm": 3.311382293701172,
"learning_rate": 5.680423781859524e-05,
"loss": 0.3028,
"step": 2360
},
{
"epoch": 1.5490196078431373,
"grad_norm": 1.4395604133605957,
"learning_rate": 5.6427336018652024e-05,
"loss": 0.2899,
"step": 2370
},
{
"epoch": 1.5555555555555556,
"grad_norm": 6.176699638366699,
"learning_rate": 5.6050062496706744e-05,
"loss": 0.3159,
"step": 2380
},
{
"epoch": 1.5620915032679739,
"grad_norm": 3.6819286346435547,
"learning_rate": 5.5672439072198804e-05,
"loss": 0.3084,
"step": 2390
},
{
"epoch": 1.5686274509803921,
"grad_norm": 1.9665669202804565,
"learning_rate": 5.529448758480408e-05,
"loss": 0.3224,
"step": 2400
},
{
"epoch": 1.5686274509803921,
"eval_loss": 0.30697837471961975,
"eval_runtime": 7464.764,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2400
},
{
"epoch": 1.5751633986928104,
"grad_norm": 1.5518563985824585,
"learning_rate": 5.491622989317181e-05,
"loss": 0.3173,
"step": 2410
},
{
"epoch": 1.581699346405229,
"grad_norm": 2.218691110610962,
"learning_rate": 5.4537687873660427e-05,
"loss": 0.3152,
"step": 2420
},
{
"epoch": 1.5882352941176472,
"grad_norm": 7.945935249328613,
"learning_rate": 5.415888341907232e-05,
"loss": 0.2788,
"step": 2430
},
{
"epoch": 1.5947712418300655,
"grad_norm": 2.2534475326538086,
"learning_rate": 5.3779838437387696e-05,
"loss": 0.326,
"step": 2440
},
{
"epoch": 1.6013071895424837,
"grad_norm": 2.623176097869873,
"learning_rate": 5.340057485049751e-05,
"loss": 0.2889,
"step": 2450
},
{
"epoch": 1.607843137254902,
"grad_norm": 1.6333633661270142,
"learning_rate": 5.302111459293569e-05,
"loss": 0.3171,
"step": 2460
},
{
"epoch": 1.6143790849673203,
"grad_norm": 7.508179664611816,
"learning_rate": 5.2641479610610503e-05,
"loss": 0.3068,
"step": 2470
},
{
"epoch": 1.6209150326797386,
"grad_norm": 4.800639629364014,
"learning_rate": 5.226169185953532e-05,
"loss": 0.2964,
"step": 2480
},
{
"epoch": 1.6274509803921569,
"grad_norm": 3.6136810779571533,
"learning_rate": 5.1881773304558856e-05,
"loss": 0.3202,
"step": 2490
},
{
"epoch": 1.6339869281045751,
"grad_norm": 1.746603012084961,
"learning_rate": 5.153975292780853e-05,
"loss": 0.2849,
"step": 2500
},
{
"epoch": 1.6339869281045751,
"eval_loss": 0.316631942987442,
"eval_runtime": 7464.5174,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2500
},
{
"epoch": 1.6405228758169934,
"grad_norm": 4.562976837158203,
"learning_rate": 5.1159646384647754e-05,
"loss": 0.2793,
"step": 2510
},
{
"epoch": 1.6470588235294117,
"grad_norm": 4.200897216796875,
"learning_rate": 5.0779472773875356e-05,
"loss": 0.3339,
"step": 2520
},
{
"epoch": 1.65359477124183,
"grad_norm": 2.1885018348693848,
"learning_rate": 5.0399254082656075e-05,
"loss": 0.3523,
"step": 2530
},
{
"epoch": 1.6601307189542482,
"grad_norm": 3.327165365219116,
"learning_rate": 5.0019012300761826e-05,
"loss": 0.3005,
"step": 2540
},
{
"epoch": 1.6666666666666665,
"grad_norm": 2.69273042678833,
"learning_rate": 4.963876941929997e-05,
"loss": 0.3143,
"step": 2550
},
{
"epoch": 1.673202614379085,
"grad_norm": 2.155317783355713,
"learning_rate": 4.9258547429441454e-05,
"loss": 0.3309,
"step": 2560
},
{
"epoch": 1.6797385620915033,
"grad_norm": 3.1310834884643555,
"learning_rate": 4.887836832114898e-05,
"loss": 0.3638,
"step": 2570
},
{
"epoch": 1.6862745098039216,
"grad_norm": 5.145089626312256,
"learning_rate": 4.8498254081905217e-05,
"loss": 0.3092,
"step": 2580
},
{
"epoch": 1.6928104575163399,
"grad_norm": 1.412678599357605,
"learning_rate": 4.811822669544115e-05,
"loss": 0.3013,
"step": 2590
},
{
"epoch": 1.6993464052287581,
"grad_norm": 50.4953727722168,
"learning_rate": 4.7738308140464685e-05,
"loss": 0.319,
"step": 2600
},
{
"epoch": 1.6993464052287581,
"eval_loss": 0.30734336376190186,
"eval_runtime": 7463.6236,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2600
},
{
"epoch": 1.7058823529411766,
"grad_norm": 2.8010504245758057,
"learning_rate": 4.7358520389389515e-05,
"loss": 0.2828,
"step": 2610
},
{
"epoch": 1.712418300653595,
"grad_norm": 3.2057955265045166,
"learning_rate": 4.697888540706431e-05,
"loss": 0.2932,
"step": 2620
},
{
"epoch": 1.7189542483660132,
"grad_norm": 5.941930294036865,
"learning_rate": 4.65994251495025e-05,
"loss": 0.3581,
"step": 2630
},
{
"epoch": 1.7254901960784315,
"grad_norm": 6.0155158042907715,
"learning_rate": 4.622016156261232e-05,
"loss": 0.287,
"step": 2640
},
{
"epoch": 1.7320261437908497,
"grad_norm": 2.266997814178467,
"learning_rate": 4.5841116580927695e-05,
"loss": 0.3033,
"step": 2650
},
{
"epoch": 1.738562091503268,
"grad_norm": 2.44047212600708,
"learning_rate": 4.546231212633959e-05,
"loss": 0.3032,
"step": 2660
},
{
"epoch": 1.7450980392156863,
"grad_norm": 10.935256004333496,
"learning_rate": 4.5083770106828203e-05,
"loss": 0.3318,
"step": 2670
},
{
"epoch": 1.7516339869281046,
"grad_norm": 8.668164253234863,
"learning_rate": 4.470551241519594e-05,
"loss": 0.2884,
"step": 2680
},
{
"epoch": 1.7581699346405228,
"grad_norm": 5.217986106872559,
"learning_rate": 4.432756092780122e-05,
"loss": 0.3411,
"step": 2690
},
{
"epoch": 1.7647058823529411,
"grad_norm": 2.592409610748291,
"learning_rate": 4.394993750329328e-05,
"loss": 0.319,
"step": 2700
},
{
"epoch": 1.7647058823529411,
"eval_loss": 0.31023845076560974,
"eval_runtime": 7463.1227,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2700
},
{
"epoch": 1.7712418300653594,
"grad_norm": 6.279813766479492,
"learning_rate": 4.3572663981347974e-05,
"loss": 0.3047,
"step": 2710
},
{
"epoch": 1.7777777777777777,
"grad_norm": 5.978330612182617,
"learning_rate": 4.3195762181404764e-05,
"loss": 0.3,
"step": 2720
},
{
"epoch": 1.784313725490196,
"grad_norm": 3.543389320373535,
"learning_rate": 4.281925390140469e-05,
"loss": 0.3143,
"step": 2730
},
{
"epoch": 1.7908496732026142,
"grad_norm": 2.8691015243530273,
"learning_rate": 4.2443160916529735e-05,
"loss": 0.3302,
"step": 2740
},
{
"epoch": 1.7973856209150327,
"grad_norm": 3.5990967750549316,
"learning_rate": 4.206750497794355e-05,
"loss": 0.3109,
"step": 2750
},
{
"epoch": 1.803921568627451,
"grad_norm": 2.575509786605835,
"learning_rate": 4.169230781153334e-05,
"loss": 0.2668,
"step": 2760
},
{
"epoch": 1.8104575163398693,
"grad_norm": 5.235141754150391,
"learning_rate": 4.131759111665349e-05,
"loss": 0.3043,
"step": 2770
},
{
"epoch": 1.8169934640522876,
"grad_norm": 9.062365531921387,
"learning_rate": 4.09433765648705e-05,
"loss": 0.3003,
"step": 2780
},
{
"epoch": 1.8235294117647058,
"grad_norm": 2.82456636428833,
"learning_rate": 4.056968579870965e-05,
"loss": 0.2778,
"step": 2790
},
{
"epoch": 1.8300653594771243,
"grad_norm": 6.167715549468994,
"learning_rate": 4.0196540430403377e-05,
"loss": 0.3419,
"step": 2800
},
{
"epoch": 1.8300653594771243,
"eval_loss": 0.3168606758117676,
"eval_runtime": 7462.498,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2800
},
{
"epoch": 1.8366013071895426,
"grad_norm": 4.378960609436035,
"learning_rate": 3.982396204064124e-05,
"loss": 0.2729,
"step": 2810
},
{
"epoch": 1.843137254901961,
"grad_norm": 2.968966245651245,
"learning_rate": 3.945197217732186e-05,
"loss": 0.3031,
"step": 2820
},
{
"epoch": 1.8496732026143792,
"grad_norm": 1.4191527366638184,
"learning_rate": 3.908059235430671e-05,
"loss": 0.3135,
"step": 2830
},
{
"epoch": 1.8562091503267975,
"grad_norm": 3.671318769454956,
"learning_rate": 3.870984405017586e-05,
"loss": 0.2538,
"step": 2840
},
{
"epoch": 1.8627450980392157,
"grad_norm": 7.276266098022461,
"learning_rate": 3.8339748706985786e-05,
"loss": 0.2591,
"step": 2850
},
{
"epoch": 1.869281045751634,
"grad_norm": 4.304380416870117,
"learning_rate": 3.797032772902929e-05,
"loss": 0.3706,
"step": 2860
},
{
"epoch": 1.8758169934640523,
"grad_norm": 6.998644828796387,
"learning_rate": 3.760160248159752e-05,
"loss": 0.325,
"step": 2870
},
{
"epoch": 1.8823529411764706,
"grad_norm": 8.353042602539062,
"learning_rate": 3.723359428974439e-05,
"loss": 0.3581,
"step": 2880
},
{
"epoch": 1.8888888888888888,
"grad_norm": 5.707284450531006,
"learning_rate": 3.686632443705328e-05,
"loss": 0.3016,
"step": 2890
},
{
"epoch": 1.8954248366013071,
"grad_norm": 2.3612542152404785,
"learning_rate": 3.649981416440603e-05,
"loss": 0.323,
"step": 2900
},
{
"epoch": 1.8954248366013071,
"eval_loss": 0.30723679065704346,
"eval_runtime": 7462.5684,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 2900
},
{
"epoch": 1.9019607843137254,
"grad_norm": 9.104491233825684,
"learning_rate": 3.613408466875452e-05,
"loss": 0.2923,
"step": 2910
},
{
"epoch": 1.9084967320261437,
"grad_norm": 2.203709602355957,
"learning_rate": 3.576915710189475e-05,
"loss": 0.3133,
"step": 2920
},
{
"epoch": 1.915032679738562,
"grad_norm": 2.8168318271636963,
"learning_rate": 3.540505256924354e-05,
"loss": 0.3283,
"step": 2930
},
{
"epoch": 1.9215686274509802,
"grad_norm": 2.40824294090271,
"learning_rate": 3.5041792128617927e-05,
"loss": 0.296,
"step": 2940
},
{
"epoch": 1.9281045751633987,
"grad_norm": 1.6007869243621826,
"learning_rate": 3.467939678901726e-05,
"loss": 0.2885,
"step": 2950
},
{
"epoch": 1.934640522875817,
"grad_norm": 2.6064846515655518,
"learning_rate": 3.43178875094082e-05,
"loss": 0.2826,
"step": 2960
},
{
"epoch": 1.9411764705882353,
"grad_norm": 3.1505980491638184,
"learning_rate": 3.395728519751249e-05,
"loss": 0.3147,
"step": 2970
},
{
"epoch": 1.9477124183006536,
"grad_norm": 2.0151422023773193,
"learning_rate": 3.3597610708597845e-05,
"loss": 0.2811,
"step": 2980
},
{
"epoch": 1.954248366013072,
"grad_norm": 5.63386869430542,
"learning_rate": 3.3238884844271776e-05,
"loss": 0.3105,
"step": 2990
},
{
"epoch": 1.9607843137254903,
"grad_norm": 7.039790153503418,
"learning_rate": 3.288112835127849e-05,
"loss": 0.2829,
"step": 3000
},
{
"epoch": 1.9607843137254903,
"eval_loss": 0.31224942207336426,
"eval_runtime": 7463.0695,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3000
},
{
"epoch": 1.9673202614379086,
"grad_norm": 1.942665457725525,
"learning_rate": 3.2524361920299105e-05,
"loss": 0.3209,
"step": 3010
},
{
"epoch": 1.973856209150327,
"grad_norm": 2.1431195735931396,
"learning_rate": 3.216860618475493e-05,
"loss": 0.2936,
"step": 3020
},
{
"epoch": 1.9803921568627452,
"grad_norm": 12.91400146484375,
"learning_rate": 3.181388171961415e-05,
"loss": 0.3471,
"step": 3030
},
{
"epoch": 1.9869281045751634,
"grad_norm": 6.381521224975586,
"learning_rate": 3.1460209040201965e-05,
"loss": 0.3261,
"step": 3040
},
{
"epoch": 1.9934640522875817,
"grad_norm": 4.691803932189941,
"learning_rate": 3.110760860101397e-05,
"loss": 0.2674,
"step": 3050
},
{
"epoch": 2.0,
"grad_norm": 3.9562430381774902,
"learning_rate": 3.0756100794533305e-05,
"loss": 0.2879,
"step": 3060
},
{
"epoch": 2.0065359477124183,
"grad_norm": 4.3201141357421875,
"learning_rate": 3.0405705950051144e-05,
"loss": 0.2914,
"step": 3070
},
{
"epoch": 2.0130718954248366,
"grad_norm": 3.7568459510803223,
"learning_rate": 3.0056444332491063e-05,
"loss": 0.3116,
"step": 3080
},
{
"epoch": 2.019607843137255,
"grad_norm": 2.2305781841278076,
"learning_rate": 2.970833614123696e-05,
"loss": 0.2957,
"step": 3090
},
{
"epoch": 2.026143790849673,
"grad_norm": 2.8348355293273926,
"learning_rate": 2.936140150896485e-05,
"loss": 0.3093,
"step": 3100
},
{
"epoch": 2.026143790849673,
"eval_loss": 0.30744266510009766,
"eval_runtime": 7463.4558,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3100
},
{
"epoch": 2.0326797385620914,
"grad_norm": 2.012289524078369,
"learning_rate": 2.901566050047855e-05,
"loss": 0.2565,
"step": 3110
},
{
"epoch": 2.0392156862745097,
"grad_norm": 3.6523494720458984,
"learning_rate": 2.8671133111549155e-05,
"loss": 0.2881,
"step": 3120
},
{
"epoch": 2.045751633986928,
"grad_norm": 2.896991491317749,
"learning_rate": 2.8327839267758648e-05,
"loss": 0.3225,
"step": 3130
},
{
"epoch": 2.052287581699346,
"grad_norm": 3.325443744659424,
"learning_rate": 2.7985798823347507e-05,
"loss": 0.2898,
"step": 3140
},
{
"epoch": 2.0588235294117645,
"grad_norm": 2.0076019763946533,
"learning_rate": 2.7645031560066437e-05,
"loss": 0.2988,
"step": 3150
},
{
"epoch": 2.065359477124183,
"grad_norm": 1.9376704692840576,
"learning_rate": 2.730555718603228e-05,
"loss": 0.2642,
"step": 3160
},
{
"epoch": 2.0718954248366015,
"grad_norm": 4.068740367889404,
"learning_rate": 2.6967395334588234e-05,
"loss": 0.2956,
"step": 3170
},
{
"epoch": 2.0784313725490198,
"grad_norm": 2.9397168159484863,
"learning_rate": 2.6630565563168418e-05,
"loss": 0.317,
"step": 3180
},
{
"epoch": 2.084967320261438,
"grad_norm": 2.2718117237091064,
"learning_rate": 2.6295087352166643e-05,
"loss": 0.2847,
"step": 3190
},
{
"epoch": 2.0915032679738563,
"grad_norm": 3.948641777038574,
"learning_rate": 2.599432858399628e-05,
"loss": 0.2799,
"step": 3200
},
{
"epoch": 2.0915032679738563,
"eval_loss": 0.306661993265152,
"eval_runtime": 7463.0196,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3200
},
{
"epoch": 2.0980392156862746,
"grad_norm": 2.652907133102417,
"learning_rate": 2.5661471725414854e-05,
"loss": 0.2944,
"step": 3210
},
{
"epoch": 2.104575163398693,
"grad_norm": 2.062485694885254,
"learning_rate": 2.533002247434479e-05,
"loss": 0.2397,
"step": 3220
},
{
"epoch": 2.111111111111111,
"grad_norm": 1.6863266229629517,
"learning_rate": 2.500000000000001e-05,
"loss": 0.2616,
"step": 3230
},
{
"epoch": 2.1176470588235294,
"grad_norm": 4.490372180938721,
"learning_rate": 2.4671423389077498e-05,
"loss": 0.3018,
"step": 3240
},
{
"epoch": 2.1241830065359477,
"grad_norm": 2.2945055961608887,
"learning_rate": 2.434431164465336e-05,
"loss": 0.2904,
"step": 3250
},
{
"epoch": 2.130718954248366,
"grad_norm": 1.8708246946334839,
"learning_rate": 2.401868368508387e-05,
"loss": 0.2651,
"step": 3260
},
{
"epoch": 2.1372549019607843,
"grad_norm": 5.456343173980713,
"learning_rate": 2.369455834291125e-05,
"loss": 0.2885,
"step": 3270
},
{
"epoch": 2.1437908496732025,
"grad_norm": 2.5962464809417725,
"learning_rate": 2.3371954363774552e-05,
"loss": 0.2936,
"step": 3280
},
{
"epoch": 2.150326797385621,
"grad_norm": 3.175043821334839,
"learning_rate": 2.3050890405325532e-05,
"loss": 0.3827,
"step": 3290
},
{
"epoch": 2.156862745098039,
"grad_norm": 1.3506354093551636,
"learning_rate": 2.2731385036149516e-05,
"loss": 0.2489,
"step": 3300
},
{
"epoch": 2.156862745098039,
"eval_loss": 0.30653849244117737,
"eval_runtime": 7462.3372,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3300
},
{
"epoch": 2.1633986928104574,
"grad_norm": 3.176468849182129,
"learning_rate": 2.2413456734691595e-05,
"loss": 0.3081,
"step": 3310
},
{
"epoch": 2.1699346405228757,
"grad_norm": 3.331033229827881,
"learning_rate": 2.2097123888187827e-05,
"loss": 0.3057,
"step": 3320
},
{
"epoch": 2.176470588235294,
"grad_norm": 3.275542736053467,
"learning_rate": 2.1782404791601903e-05,
"loss": 0.2979,
"step": 3330
},
{
"epoch": 2.183006535947712,
"grad_norm": 1.958678960800171,
"learning_rate": 2.1469317646567013e-05,
"loss": 0.2879,
"step": 3340
},
{
"epoch": 2.189542483660131,
"grad_norm": 2.5814414024353027,
"learning_rate": 2.1157880560333197e-05,
"loss": 0.2733,
"step": 3350
},
{
"epoch": 2.196078431372549,
"grad_norm": 3.4207053184509277,
"learning_rate": 2.0848111544720117e-05,
"loss": 0.3109,
"step": 3360
},
{
"epoch": 2.2026143790849675,
"grad_norm": 2.685720205307007,
"learning_rate": 2.0540028515075328e-05,
"loss": 0.3195,
"step": 3370
},
{
"epoch": 2.2091503267973858,
"grad_norm": 3.908967971801758,
"learning_rate": 2.023364928923817e-05,
"loss": 0.2651,
"step": 3380
},
{
"epoch": 2.215686274509804,
"grad_norm": 2.786430597305298,
"learning_rate": 1.9928991586509276e-05,
"loss": 0.2687,
"step": 3390
},
{
"epoch": 2.2222222222222223,
"grad_norm": 3.6411643028259277,
"learning_rate": 1.9626073026625818e-05,
"loss": 0.2846,
"step": 3400
},
{
"epoch": 2.2222222222222223,
"eval_loss": 0.3210596442222595,
"eval_runtime": 7462.9136,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3400
},
{
"epoch": 2.2287581699346406,
"grad_norm": 2.521268129348755,
"learning_rate": 1.9324911128742407e-05,
"loss": 0.3536,
"step": 3410
},
{
"epoch": 2.235294117647059,
"grad_norm": 2.3333940505981445,
"learning_rate": 1.905538176438527e-05,
"loss": 0.2823,
"step": 3420
},
{
"epoch": 2.241830065359477,
"grad_norm": 3.3351223468780518,
"learning_rate": 1.8757605424895984e-05,
"loss": 0.2645,
"step": 3430
},
{
"epoch": 2.2483660130718954,
"grad_norm": 2.780357599258423,
"learning_rate": 1.8461635974828296e-05,
"loss": 0.275,
"step": 3440
},
{
"epoch": 2.2549019607843137,
"grad_norm": 3.99027419090271,
"learning_rate": 1.8167490531438284e-05,
"loss": 0.352,
"step": 3450
},
{
"epoch": 2.261437908496732,
"grad_norm": 3.214839458465576,
"learning_rate": 1.7875186106491448e-05,
"loss": 0.2764,
"step": 3460
},
{
"epoch": 2.2679738562091503,
"grad_norm": 3.3453619480133057,
"learning_rate": 1.7584739605278832e-05,
"loss": 0.3093,
"step": 3470
},
{
"epoch": 2.2745098039215685,
"grad_norm": 1.7691558599472046,
"learning_rate": 1.7296167825639326e-05,
"loss": 0.2948,
"step": 3480
},
{
"epoch": 2.281045751633987,
"grad_norm": 2.7585806846618652,
"learning_rate": 1.7009487456988164e-05,
"loss": 0.288,
"step": 3490
},
{
"epoch": 2.287581699346405,
"grad_norm": 3.013615846633911,
"learning_rate": 1.6724715079351687e-05,
"loss": 0.2841,
"step": 3500
},
{
"epoch": 2.287581699346405,
"eval_loss": 0.3018401563167572,
"eval_runtime": 7463.5622,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3500
},
{
"epoch": 2.2941176470588234,
"grad_norm": 5.529947280883789,
"learning_rate": 1.6441867162408514e-05,
"loss": 0.3024,
"step": 3510
},
{
"epoch": 2.3006535947712417,
"grad_norm": 2.649116277694702,
"learning_rate": 1.6160960064536908e-05,
"loss": 0.319,
"step": 3520
},
{
"epoch": 2.30718954248366,
"grad_norm": 5.389304161071777,
"learning_rate": 1.5882010031868777e-05,
"loss": 0.296,
"step": 3530
},
{
"epoch": 2.313725490196078,
"grad_norm": 2.242767095565796,
"learning_rate": 1.5605033197350095e-05,
"loss": 0.2631,
"step": 3540
},
{
"epoch": 2.3202614379084965,
"grad_norm": 6.1497111320495605,
"learning_rate": 1.5330045579807802e-05,
"loss": 0.2901,
"step": 3550
},
{
"epoch": 2.326797385620915,
"grad_norm": NaN,
"learning_rate": 1.5084270651401933e-05,
"loss": 0.3096,
"step": 3560
},
{
"epoch": 2.3333333333333335,
"grad_norm": 2.6111481189727783,
"learning_rate": 1.4813106265206484e-05,
"loss": 0.315,
"step": 3570
},
{
"epoch": 2.3398692810457518,
"grad_norm": 1.8103219270706177,
"learning_rate": 1.4543976896714795e-05,
"loss": 0.3009,
"step": 3580
},
{
"epoch": 2.34640522875817,
"grad_norm": 4.367152690887451,
"learning_rate": 1.4276898110899262e-05,
"loss": 0.2815,
"step": 3590
},
{
"epoch": 2.3529411764705883,
"grad_norm": 5.697513580322266,
"learning_rate": 1.4011885354137754e-05,
"loss": 0.3119,
"step": 3600
},
{
"epoch": 2.3529411764705883,
"eval_loss": 0.30415767431259155,
"eval_runtime": 7463.8583,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3600
},
{
"epoch": 2.3594771241830066,
"grad_norm": 2.6390066146850586,
"learning_rate": 1.3748953953320371e-05,
"loss": 0.3065,
"step": 3610
},
{
"epoch": 2.366013071895425,
"grad_norm": 1.9461013078689575,
"learning_rate": 1.3488119114962884e-05,
"loss": 0.2627,
"step": 3620
},
{
"epoch": 2.372549019607843,
"grad_norm": 2.6425626277923584,
"learning_rate": 1.322939592432738e-05,
"loss": 0.2935,
"step": 3630
},
{
"epoch": 2.3790849673202614,
"grad_norm": 3.9877243041992188,
"learning_rate": 1.297279934454978e-05,
"loss": 0.2721,
"step": 3640
},
{
"epoch": 2.3856209150326797,
"grad_norm": 2.479541778564453,
"learning_rate": 1.271834421577443e-05,
"loss": 0.2705,
"step": 3650
},
{
"epoch": 2.392156862745098,
"grad_norm": 3.324650526046753,
"learning_rate": 1.2466045254295871e-05,
"loss": 0.2933,
"step": 3660
},
{
"epoch": 2.3986928104575163,
"grad_norm": 3.598247766494751,
"learning_rate": 1.2215917051707676e-05,
"loss": 0.2712,
"step": 3670
},
{
"epoch": 2.4052287581699345,
"grad_norm": 1.403212070465088,
"learning_rate": 1.1967974074058647e-05,
"loss": 0.3371,
"step": 3680
},
{
"epoch": 2.411764705882353,
"grad_norm": 20.876041412353516,
"learning_rate": 1.1722230661016043e-05,
"loss": 0.2962,
"step": 3690
},
{
"epoch": 2.418300653594771,
"grad_norm": 2.889564037322998,
"learning_rate": 1.147870102503636e-05,
"loss": 0.2978,
"step": 3700
},
{
"epoch": 2.418300653594771,
"eval_loss": 0.30770230293273926,
"eval_runtime": 7462.6718,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3700
},
{
"epoch": 2.4248366013071894,
"grad_norm": 2.3619654178619385,
"learning_rate": 1.1237399250543307e-05,
"loss": 0.2689,
"step": 3710
},
{
"epoch": 2.431372549019608,
"grad_norm": 3.7691731452941895,
"learning_rate": 1.0998339293113252e-05,
"loss": 0.274,
"step": 3720
},
{
"epoch": 2.4379084967320264,
"grad_norm": 2.3495209217071533,
"learning_rate": 1.0761534978668152e-05,
"loss": 0.2776,
"step": 3730
},
{
"epoch": 2.4444444444444446,
"grad_norm": 1.8520164489746094,
"learning_rate": 1.0527000002675853e-05,
"loss": 0.2815,
"step": 3740
},
{
"epoch": 2.450980392156863,
"grad_norm": 2.9609193801879883,
"learning_rate": 1.0294747929358073e-05,
"loss": 0.291,
"step": 3750
},
{
"epoch": 2.457516339869281,
"grad_norm": 3.36863112449646,
"learning_rate": 1.0064792190905914e-05,
"loss": 0.2957,
"step": 3760
},
{
"epoch": 2.4640522875816995,
"grad_norm": 2.343014717102051,
"learning_rate": 9.837146086703014e-06,
"loss": 0.2504,
"step": 3770
},
{
"epoch": 2.4705882352941178,
"grad_norm": 3.688849687576294,
"learning_rate": 9.61182278255639e-06,
"loss": 0.2479,
"step": 3780
},
{
"epoch": 2.477124183006536,
"grad_norm": 2.4627552032470703,
"learning_rate": 9.388835309934984e-06,
"loss": 0.3038,
"step": 3790
},
{
"epoch": 2.4836601307189543,
"grad_norm": 2.27083683013916,
"learning_rate": 9.168196565216008e-06,
"loss": 0.3068,
"step": 3800
},
{
"epoch": 2.4836601307189543,
"eval_loss": 0.3110058605670929,
"eval_runtime": 7463.183,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3800
},
{
"epoch": 2.4901960784313726,
"grad_norm": 2.4768147468566895,
"learning_rate": 8.949919308939082e-06,
"loss": 0.2489,
"step": 3810
},
{
"epoch": 2.496732026143791,
"grad_norm": 14.81464958190918,
"learning_rate": 8.734016165068226e-06,
"loss": 0.2852,
"step": 3820
},
{
"epoch": 2.503267973856209,
"grad_norm": 5.01676607131958,
"learning_rate": 8.520499620261802e-06,
"loss": 0.2511,
"step": 3830
},
{
"epoch": 2.5098039215686274,
"grad_norm": 3.20218825340271,
"learning_rate": 8.309382023150286e-06,
"loss": 0.245,
"step": 3840
},
{
"epoch": 2.5163398692810457,
"grad_norm": 3.2415478229522705,
"learning_rate": 8.100675583622137e-06,
"loss": 0.2736,
"step": 3850
},
{
"epoch": 2.522875816993464,
"grad_norm": 4.451015472412109,
"learning_rate": 7.894392372117627e-06,
"loss": 0.3537,
"step": 3860
},
{
"epoch": 2.5294117647058822,
"grad_norm": 3.3448715209960938,
"learning_rate": 7.690544318930765e-06,
"loss": 0.2854,
"step": 3870
},
{
"epoch": 2.5359477124183005,
"grad_norm": 2.145322561264038,
"learning_rate": 7.489143213519301e-06,
"loss": 0.2923,
"step": 3880
},
{
"epoch": 2.542483660130719,
"grad_norm": 2.5505619049072266,
"learning_rate": 7.290200703822897e-06,
"loss": 0.2938,
"step": 3890
},
{
"epoch": 2.549019607843137,
"grad_norm": 2.9079599380493164,
"learning_rate": 7.093728295589475e-06,
"loss": 0.3237,
"step": 3900
},
{
"epoch": 2.549019607843137,
"eval_loss": 0.30904993414878845,
"eval_runtime": 7464.5061,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 3900
},
{
"epoch": 2.5555555555555554,
"grad_norm": 1.6857258081436157,
"learning_rate": 6.899737351709757e-06,
"loss": 0.2478,
"step": 3910
},
{
"epoch": 2.5620915032679736,
"grad_norm": 4.590330600738525,
"learning_rate": 6.70823909156017e-06,
"loss": 0.2728,
"step": 3920
},
{
"epoch": 2.568627450980392,
"grad_norm": 1.7133220434188843,
"learning_rate": 6.51924459035389e-06,
"loss": 0.2144,
"step": 3930
},
{
"epoch": 2.57516339869281,
"grad_norm": 5.041657447814941,
"learning_rate": 6.332764778500361e-06,
"loss": 0.3144,
"step": 3940
},
{
"epoch": 2.581699346405229,
"grad_norm": 3.0704329013824463,
"learning_rate": 6.148810440973168e-06,
"loss": 0.2795,
"step": 3950
},
{
"epoch": 2.588235294117647,
"grad_norm": 3.937464714050293,
"learning_rate": 5.967392216686207e-06,
"loss": 0.2751,
"step": 3960
},
{
"epoch": 2.5947712418300655,
"grad_norm": 2.519258975982666,
"learning_rate": 5.788520597878477e-06,
"loss": 0.3139,
"step": 3970
},
{
"epoch": 2.6013071895424837,
"grad_norm": 1.8165565729141235,
"learning_rate": 5.6122059295072085e-06,
"loss": 0.2162,
"step": 3980
},
{
"epoch": 2.607843137254902,
"grad_norm": 4.116787433624268,
"learning_rate": 5.438458408649588e-06,
"loss": 0.3408,
"step": 3990
},
{
"epoch": 2.6143790849673203,
"grad_norm": 2.385406494140625,
"learning_rate": 5.26728808391303e-06,
"loss": 0.2811,
"step": 4000
},
{
"epoch": 2.6143790849673203,
"eval_loss": 0.31030258536338806,
"eval_runtime": 7463.2353,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4000
},
{
"epoch": 2.6209150326797386,
"grad_norm": 3.6135222911834717,
"learning_rate": 5.098704854853986e-06,
"loss": 0.2767,
"step": 4010
},
{
"epoch": 2.627450980392157,
"grad_norm": 3.542980670928955,
"learning_rate": 4.932718471405445e-06,
"loss": 0.2658,
"step": 4020
},
{
"epoch": 2.633986928104575,
"grad_norm": 3.6396703720092773,
"learning_rate": 4.769338533313011e-06,
"loss": 0.3349,
"step": 4030
},
{
"epoch": 2.6405228758169934,
"grad_norm": 2.3764588832855225,
"learning_rate": 4.608574489579759e-06,
"loss": 0.3037,
"step": 4040
},
{
"epoch": 2.6470588235294117,
"grad_norm": 2.5128707885742188,
"learning_rate": 4.4504356379196925e-06,
"loss": 0.2574,
"step": 4050
},
{
"epoch": 2.65359477124183,
"grad_norm": 1.1791373491287231,
"learning_rate": 4.294931124220069e-06,
"loss": 0.2541,
"step": 4060
},
{
"epoch": 2.6601307189542482,
"grad_norm": 3.077044725418091,
"learning_rate": 4.142069942012422e-06,
"loss": 0.3129,
"step": 4070
},
{
"epoch": 2.6666666666666665,
"grad_norm": 1.5024341344833374,
"learning_rate": 3.991860931952435e-06,
"loss": 0.2796,
"step": 4080
},
{
"epoch": 2.6732026143790852,
"grad_norm": 3.4516193866729736,
"learning_rate": 3.844312781308629e-06,
"loss": 0.3117,
"step": 4090
},
{
"epoch": 2.6797385620915035,
"grad_norm": 7.4698381423950195,
"learning_rate": 3.69943402345998e-06,
"loss": 0.2585,
"step": 4100
},
{
"epoch": 2.6797385620915035,
"eval_loss": 0.3094639480113983,
"eval_runtime": 7463.506,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4100
},
{
"epoch": 2.686274509803922,
"grad_norm": 4.485745429992676,
"learning_rate": 3.5572330374023465e-06,
"loss": 0.2691,
"step": 4110
},
{
"epoch": 2.69281045751634,
"grad_norm": 2.9982385635375977,
"learning_rate": 3.4177180472639014e-06,
"loss": 0.3215,
"step": 4120
},
{
"epoch": 2.6993464052287583,
"grad_norm": 3.644282341003418,
"learning_rate": 3.280897121829485e-06,
"loss": 0.2615,
"step": 4130
},
{
"epoch": 2.7058823529411766,
"grad_norm": 2.4241082668304443,
"learning_rate": 3.146778174073972e-06,
"loss": 0.2638,
"step": 4140
},
{
"epoch": 2.712418300653595,
"grad_norm": 3.6460297107696533,
"learning_rate": 3.0153689607045845e-06,
"loss": 0.2446,
"step": 4150
},
{
"epoch": 2.718954248366013,
"grad_norm": 3.8150696754455566,
"learning_rate": 2.886677081712319e-06,
"loss": 0.3122,
"step": 4160
},
{
"epoch": 2.7254901960784315,
"grad_norm": 2.6598055362701416,
"learning_rate": 2.7607099799323965e-06,
"loss": 0.3118,
"step": 4170
},
{
"epoch": 2.7320261437908497,
"grad_norm": 4.361299514770508,
"learning_rate": 2.637474940613821e-06,
"loss": 0.2776,
"step": 4180
},
{
"epoch": 2.738562091503268,
"grad_norm": 11.310445785522461,
"learning_rate": 2.5169790909980174e-06,
"loss": 0.3527,
"step": 4190
},
{
"epoch": 2.7450980392156863,
"grad_norm": 3.9927046298980713,
"learning_rate": 2.3992293999066385e-06,
"loss": 0.2911,
"step": 4200
},
{
"epoch": 2.7450980392156863,
"eval_loss": 0.31082937121391296,
"eval_runtime": 7463.9458,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4200
},
{
"epoch": 2.7516339869281046,
"grad_norm": 2.656670570373535,
"learning_rate": 2.2842326773385438e-06,
"loss": 0.2853,
"step": 4210
},
{
"epoch": 2.758169934640523,
"grad_norm": 9.812518119812012,
"learning_rate": 2.17199557407593e-06,
"loss": 0.2622,
"step": 4220
},
{
"epoch": 2.764705882352941,
"grad_norm": 4.411550521850586,
"learning_rate": 2.0625245812996797e-06,
"loss": 0.2961,
"step": 4230
},
{
"epoch": 2.7712418300653594,
"grad_norm": 4.210736274719238,
"learning_rate": 1.955826030213964e-06,
"loss": 0.3131,
"step": 4240
},
{
"epoch": 2.7777777777777777,
"grad_norm": 10.433571815490723,
"learning_rate": 1.8519060916800668e-06,
"loss": 0.2771,
"step": 4250
},
{
"epoch": 2.784313725490196,
"grad_norm": 2.862966775894165,
"learning_rate": 1.7507707758595172e-06,
"loss": 0.2763,
"step": 4260
},
{
"epoch": 2.7908496732026142,
"grad_norm": 6.950551509857178,
"learning_rate": 1.6524259318664692e-06,
"loss": 0.3239,
"step": 4270
},
{
"epoch": 2.7973856209150325,
"grad_norm": 4.454069137573242,
"learning_rate": 1.5568772474294224e-06,
"loss": 0.3074,
"step": 4280
},
{
"epoch": 2.803921568627451,
"grad_norm": 3.450611114501953,
"learning_rate": 1.4641302485622953e-06,
"loss": 0.3235,
"step": 4290
},
{
"epoch": 2.810457516339869,
"grad_norm": 3.067033290863037,
"learning_rate": 1.3741902992448152e-06,
"loss": 0.3288,
"step": 4300
},
{
"epoch": 2.810457516339869,
"eval_loss": 0.3071349561214447,
"eval_runtime": 7463.7911,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4300
},
{
"epoch": 2.8169934640522873,
"grad_norm": 4.035493850708008,
"learning_rate": 1.2870626011123044e-06,
"loss": 0.2968,
"step": 4310
},
{
"epoch": 2.8235294117647056,
"grad_norm": 1.9869184494018555,
"learning_rate": 1.2027521931548214e-06,
"loss": 0.3015,
"step": 4320
},
{
"epoch": 2.8300653594771243,
"grad_norm": 3.0594305992126465,
"learning_rate": 1.1212639514257829e-06,
"loss": 0.3119,
"step": 4330
},
{
"epoch": 2.8366013071895426,
"grad_norm": 3.2473435401916504,
"learning_rate": 1.042602588759889e-06,
"loss": 0.3069,
"step": 4340
},
{
"epoch": 2.843137254901961,
"grad_norm": 4.640011310577393,
"learning_rate": 9.667726545006417e-07,
"loss": 0.3186,
"step": 4350
},
{
"epoch": 2.849673202614379,
"grad_norm": 1.6823493242263794,
"learning_rate": 8.937785342371662e-07,
"loss": 0.2628,
"step": 4360
},
{
"epoch": 2.8562091503267975,
"grad_norm": 2.6158905029296875,
"learning_rate": 8.236244495505974e-07,
"loss": 0.296,
"step": 4370
},
{
"epoch": 2.8627450980392157,
"grad_norm": 1.7890446186065674,
"learning_rate": 7.563144577699355e-07,
"loss": 0.3062,
"step": 4380
},
{
"epoch": 2.869281045751634,
"grad_norm": 11.531109809875488,
"learning_rate": 6.918524517373903e-07,
"loss": 0.3105,
"step": 4390
},
{
"epoch": 2.8758169934640523,
"grad_norm": 2.951564073562622,
"learning_rate": 6.302421595832164e-07,
"loss": 0.2993,
"step": 4400
},
{
"epoch": 2.8758169934640523,
"eval_loss": 0.3057985007762909,
"eval_runtime": 7464.7786,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4400
},
{
"epoch": 2.8823529411764706,
"grad_norm": 5.375591278076172,
"learning_rate": 5.714871445101356e-07,
"loss": 0.2902,
"step": 4410
},
{
"epoch": 2.888888888888889,
"grad_norm": 2.2659952640533447,
"learning_rate": 5.155908045872349e-07,
"loss": 0.2578,
"step": 4420
},
{
"epoch": 2.895424836601307,
"grad_norm": 3.643691062927246,
"learning_rate": 4.625563725534465e-07,
"loss": 0.26,
"step": 4430
},
{
"epoch": 2.9019607843137254,
"grad_norm": 3.761317253112793,
"learning_rate": 4.1238691563058594e-07,
"loss": 0.3239,
"step": 4440
},
{
"epoch": 2.9084967320261437,
"grad_norm": 2.697275161743164,
"learning_rate": 3.6508533534596046e-07,
"loss": 0.2802,
"step": 4450
},
{
"epoch": 2.915032679738562,
"grad_norm": 4.7579450607299805,
"learning_rate": 3.20654367364559e-07,
"loss": 0.2347,
"step": 4460
},
{
"epoch": 2.9215686274509802,
"grad_norm": 14.727649688720703,
"learning_rate": 2.7909658133083994e-07,
"loss": 0.3004,
"step": 4470
},
{
"epoch": 2.928104575163399,
"grad_norm": 10.497050285339355,
"learning_rate": 2.404143807201109e-07,
"loss": 0.3049,
"step": 4480
},
{
"epoch": 2.9346405228758172,
"grad_norm": 3.075261116027832,
"learning_rate": 2.0461000269953456e-07,
"loss": 0.3362,
"step": 4490
},
{
"epoch": 2.9411764705882355,
"grad_norm": 2.50616192817688,
"learning_rate": 1.7168551799872645e-07,
"loss": 0.322,
"step": 4500
},
{
"epoch": 2.9411764705882355,
"eval_loss": 0.3057360351085663,
"eval_runtime": 7463.9401,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.182,
"step": 4500
}
],
"logging_steps": 10,
"max_steps": 4590,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"total_flos": 1.4730845079567021e+20,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}