lop_jun2024 / trainer_state.json
obulikrish's picture
lop_jun24
551d626
raw
history blame contribute delete
No virus
176 kB
{
"best_metric": 0.44340217113494873,
"best_model_checkpoint": "saved_model/lop_jun2024/checkpoint-10982",
"epoch": 2.999931707983337,
"eval_steps": 500,
"global_step": 10982,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": NaN,
"learning_rate": 0.0,
"loss": 74.2157,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 13.99621295928955,
"learning_rate": 2.5e-06,
"loss": 74.5237,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 14.391114234924316,
"learning_rate": 6.5000000000000004e-06,
"loss": 74.2338,
"step": 20
},
{
"epoch": 0.01,
"grad_norm": 15.539421081542969,
"learning_rate": 1.1500000000000002e-05,
"loss": 72.8847,
"step": 30
},
{
"epoch": 0.01,
"grad_norm": 16.487953186035156,
"learning_rate": 1.65e-05,
"loss": 70.7699,
"step": 40
},
{
"epoch": 0.01,
"grad_norm": 21.840412139892578,
"learning_rate": 2.15e-05,
"loss": 66.4777,
"step": 50
},
{
"epoch": 0.02,
"grad_norm": 23.74176025390625,
"learning_rate": 2.6000000000000002e-05,
"loss": 59.438,
"step": 60
},
{
"epoch": 0.02,
"grad_norm": 26.7894287109375,
"learning_rate": 3.1e-05,
"loss": 45.7214,
"step": 70
},
{
"epoch": 0.02,
"grad_norm": 21.767683029174805,
"learning_rate": 3.55e-05,
"loss": 27.7672,
"step": 80
},
{
"epoch": 0.02,
"grad_norm": 15.309523582458496,
"learning_rate": 4.05e-05,
"loss": 12.9366,
"step": 90
},
{
"epoch": 0.03,
"grad_norm": 7.069633960723877,
"learning_rate": 4.55e-05,
"loss": 4.342,
"step": 100
},
{
"epoch": 0.03,
"grad_norm": 10.602043151855469,
"learning_rate": 5.05e-05,
"loss": 1.7191,
"step": 110
},
{
"epoch": 0.03,
"grad_norm": 8.00312614440918,
"learning_rate": 5.550000000000001e-05,
"loss": 1.3934,
"step": 120
},
{
"epoch": 0.04,
"grad_norm": 7.026750087738037,
"learning_rate": 6.05e-05,
"loss": 1.3728,
"step": 130
},
{
"epoch": 0.04,
"grad_norm": 10.686494827270508,
"learning_rate": 6.55e-05,
"loss": 1.2595,
"step": 140
},
{
"epoch": 0.04,
"grad_norm": 7.127823352813721,
"learning_rate": 7.05e-05,
"loss": 1.2226,
"step": 150
},
{
"epoch": 0.04,
"grad_norm": 6.253384590148926,
"learning_rate": 7.55e-05,
"loss": 1.2934,
"step": 160
},
{
"epoch": 0.05,
"grad_norm": 7.56700325012207,
"learning_rate": 8.05e-05,
"loss": 1.2571,
"step": 170
},
{
"epoch": 0.05,
"grad_norm": 9.671475410461426,
"learning_rate": 8.55e-05,
"loss": 1.2005,
"step": 180
},
{
"epoch": 0.05,
"grad_norm": 7.772515296936035,
"learning_rate": 9.05e-05,
"loss": 1.253,
"step": 190
},
{
"epoch": 0.05,
"grad_norm": 6.593009948730469,
"learning_rate": 9.55e-05,
"loss": 1.2269,
"step": 200
},
{
"epoch": 0.06,
"grad_norm": 9.796707153320312,
"learning_rate": 9.999725274725276e-05,
"loss": 1.2425,
"step": 210
},
{
"epoch": 0.06,
"grad_norm": 5.557741641998291,
"learning_rate": 9.996978021978023e-05,
"loss": 1.2383,
"step": 220
},
{
"epoch": 0.06,
"grad_norm": 7.321559906005859,
"learning_rate": 9.994230769230771e-05,
"loss": 1.2026,
"step": 230
},
{
"epoch": 0.07,
"grad_norm": 3.337256908416748,
"learning_rate": 9.991483516483518e-05,
"loss": 1.1289,
"step": 240
},
{
"epoch": 0.07,
"grad_norm": 4.283127307891846,
"learning_rate": 9.988736263736264e-05,
"loss": 1.0357,
"step": 250
},
{
"epoch": 0.07,
"grad_norm": 4.2405242919921875,
"learning_rate": 9.985989010989013e-05,
"loss": 1.0067,
"step": 260
},
{
"epoch": 0.07,
"grad_norm": 6.622239112854004,
"learning_rate": 9.98324175824176e-05,
"loss": 0.9816,
"step": 270
},
{
"epoch": 0.08,
"grad_norm": 5.521809101104736,
"learning_rate": 9.980494505494506e-05,
"loss": 0.9578,
"step": 280
},
{
"epoch": 0.08,
"grad_norm": 3.7347898483276367,
"learning_rate": 9.977747252747254e-05,
"loss": 0.8762,
"step": 290
},
{
"epoch": 0.08,
"grad_norm": 6.452709674835205,
"learning_rate": 9.975000000000001e-05,
"loss": 0.8188,
"step": 300
},
{
"epoch": 0.08,
"grad_norm": 3.2089035511016846,
"learning_rate": 9.972252747252748e-05,
"loss": 0.8247,
"step": 310
},
{
"epoch": 0.09,
"grad_norm": 4.455322742462158,
"learning_rate": 9.969505494505496e-05,
"loss": 0.7599,
"step": 320
},
{
"epoch": 0.09,
"grad_norm": 4.537048816680908,
"learning_rate": 9.966758241758242e-05,
"loss": 0.7968,
"step": 330
},
{
"epoch": 0.09,
"grad_norm": 3.754340648651123,
"learning_rate": 9.964010989010988e-05,
"loss": 0.72,
"step": 340
},
{
"epoch": 0.1,
"grad_norm": 4.538580894470215,
"learning_rate": 9.961263736263737e-05,
"loss": 0.7207,
"step": 350
},
{
"epoch": 0.1,
"grad_norm": 4.3613457679748535,
"learning_rate": 9.958516483516483e-05,
"loss": 0.7235,
"step": 360
},
{
"epoch": 0.1,
"grad_norm": 2.2669098377227783,
"learning_rate": 9.95576923076923e-05,
"loss": 0.7018,
"step": 370
},
{
"epoch": 0.1,
"grad_norm": 5.329736232757568,
"learning_rate": 9.953021978021978e-05,
"loss": 0.7193,
"step": 380
},
{
"epoch": 0.11,
"grad_norm": 2.5049781799316406,
"learning_rate": 9.950274725274725e-05,
"loss": 0.6948,
"step": 390
},
{
"epoch": 0.11,
"grad_norm": 3.5907931327819824,
"learning_rate": 9.947527472527472e-05,
"loss": 0.6637,
"step": 400
},
{
"epoch": 0.11,
"grad_norm": 3.264726161956787,
"learning_rate": 9.94478021978022e-05,
"loss": 0.6588,
"step": 410
},
{
"epoch": 0.11,
"grad_norm": 3.763380765914917,
"learning_rate": 9.942032967032967e-05,
"loss": 0.6479,
"step": 420
},
{
"epoch": 0.12,
"grad_norm": 1.9857630729675293,
"learning_rate": 9.939285714285714e-05,
"loss": 0.6481,
"step": 430
},
{
"epoch": 0.12,
"grad_norm": 2.649977922439575,
"learning_rate": 9.936538461538462e-05,
"loss": 0.6284,
"step": 440
},
{
"epoch": 0.12,
"grad_norm": 1.4232739210128784,
"learning_rate": 9.933791208791209e-05,
"loss": 0.6214,
"step": 450
},
{
"epoch": 0.13,
"grad_norm": 5.327674865722656,
"learning_rate": 9.931043956043956e-05,
"loss": 0.633,
"step": 460
},
{
"epoch": 0.13,
"grad_norm": 3.673241376876831,
"learning_rate": 9.928296703296704e-05,
"loss": 0.6792,
"step": 470
},
{
"epoch": 0.13,
"grad_norm": 3.6698083877563477,
"learning_rate": 9.92554945054945e-05,
"loss": 0.6297,
"step": 480
},
{
"epoch": 0.13,
"grad_norm": 2.056175470352173,
"learning_rate": 9.922802197802197e-05,
"loss": 0.6172,
"step": 490
},
{
"epoch": 0.14,
"grad_norm": 1.345395803451538,
"learning_rate": 9.920054945054946e-05,
"loss": 0.605,
"step": 500
},
{
"epoch": 0.14,
"grad_norm": 1.7540231943130493,
"learning_rate": 9.917307692307692e-05,
"loss": 0.6288,
"step": 510
},
{
"epoch": 0.14,
"grad_norm": 4.050562858581543,
"learning_rate": 9.914560439560439e-05,
"loss": 0.6281,
"step": 520
},
{
"epoch": 0.14,
"grad_norm": 5.210700511932373,
"learning_rate": 9.911813186813187e-05,
"loss": 0.6138,
"step": 530
},
{
"epoch": 0.15,
"grad_norm": 3.6927013397216797,
"learning_rate": 9.909065934065934e-05,
"loss": 0.6033,
"step": 540
},
{
"epoch": 0.15,
"grad_norm": 3.4765613079071045,
"learning_rate": 9.906318681318681e-05,
"loss": 0.5991,
"step": 550
},
{
"epoch": 0.15,
"grad_norm": 1.8127527236938477,
"learning_rate": 9.903571428571429e-05,
"loss": 0.5994,
"step": 560
},
{
"epoch": 0.16,
"grad_norm": 1.6926054954528809,
"learning_rate": 9.900824175824176e-05,
"loss": 0.5988,
"step": 570
},
{
"epoch": 0.16,
"grad_norm": 2.3790557384490967,
"learning_rate": 9.898076923076923e-05,
"loss": 0.6147,
"step": 580
},
{
"epoch": 0.16,
"grad_norm": 3.558159589767456,
"learning_rate": 9.895329670329671e-05,
"loss": 0.5858,
"step": 590
},
{
"epoch": 0.16,
"grad_norm": 3.5203194618225098,
"learning_rate": 9.892582417582418e-05,
"loss": 0.6036,
"step": 600
},
{
"epoch": 0.17,
"grad_norm": 2.4314634799957275,
"learning_rate": 9.889835164835165e-05,
"loss": 0.5789,
"step": 610
},
{
"epoch": 0.17,
"grad_norm": 2.956555128097534,
"learning_rate": 9.887087912087913e-05,
"loss": 0.6021,
"step": 620
},
{
"epoch": 0.17,
"grad_norm": 2.231612205505371,
"learning_rate": 9.88434065934066e-05,
"loss": 0.5825,
"step": 630
},
{
"epoch": 0.17,
"grad_norm": 2.1588950157165527,
"learning_rate": 9.881593406593406e-05,
"loss": 0.5924,
"step": 640
},
{
"epoch": 0.18,
"grad_norm": 2.8479809761047363,
"learning_rate": 9.878846153846155e-05,
"loss": 0.584,
"step": 650
},
{
"epoch": 0.18,
"grad_norm": 2.2029404640197754,
"learning_rate": 9.876098901098901e-05,
"loss": 0.5855,
"step": 660
},
{
"epoch": 0.18,
"grad_norm": 2.243802070617676,
"learning_rate": 9.873351648351648e-05,
"loss": 0.584,
"step": 670
},
{
"epoch": 0.19,
"grad_norm": 2.8837687969207764,
"learning_rate": 9.870604395604396e-05,
"loss": 0.5867,
"step": 680
},
{
"epoch": 0.19,
"grad_norm": 2.0768215656280518,
"learning_rate": 9.867857142857143e-05,
"loss": 0.5823,
"step": 690
},
{
"epoch": 0.19,
"grad_norm": 2.950639247894287,
"learning_rate": 9.86510989010989e-05,
"loss": 0.5832,
"step": 700
},
{
"epoch": 0.19,
"grad_norm": 1.0848056077957153,
"learning_rate": 9.862362637362638e-05,
"loss": 0.5595,
"step": 710
},
{
"epoch": 0.2,
"grad_norm": 1.293261170387268,
"learning_rate": 9.859615384615385e-05,
"loss": 0.5926,
"step": 720
},
{
"epoch": 0.2,
"grad_norm": 2.286271095275879,
"learning_rate": 9.856868131868132e-05,
"loss": 0.5658,
"step": 730
},
{
"epoch": 0.2,
"grad_norm": 2.9983866214752197,
"learning_rate": 9.85412087912088e-05,
"loss": 0.5749,
"step": 740
},
{
"epoch": 0.2,
"grad_norm": 2.712332248687744,
"learning_rate": 9.851373626373627e-05,
"loss": 0.5757,
"step": 750
},
{
"epoch": 0.21,
"grad_norm": 3.5591819286346436,
"learning_rate": 9.848626373626374e-05,
"loss": 0.5706,
"step": 760
},
{
"epoch": 0.21,
"grad_norm": 3.00175142288208,
"learning_rate": 9.845879120879122e-05,
"loss": 0.5654,
"step": 770
},
{
"epoch": 0.21,
"grad_norm": 1.5290532112121582,
"learning_rate": 9.843131868131869e-05,
"loss": 0.578,
"step": 780
},
{
"epoch": 0.22,
"grad_norm": 1.8737398386001587,
"learning_rate": 9.840384615384615e-05,
"loss": 0.5598,
"step": 790
},
{
"epoch": 0.22,
"grad_norm": 1.190427541732788,
"learning_rate": 9.837637362637364e-05,
"loss": 0.5507,
"step": 800
},
{
"epoch": 0.22,
"grad_norm": 2.8329577445983887,
"learning_rate": 9.83489010989011e-05,
"loss": 0.5394,
"step": 810
},
{
"epoch": 0.22,
"grad_norm": 2.6253738403320312,
"learning_rate": 9.832142857142857e-05,
"loss": 0.5404,
"step": 820
},
{
"epoch": 0.23,
"grad_norm": 2.1394851207733154,
"learning_rate": 9.829395604395605e-05,
"loss": 0.5616,
"step": 830
},
{
"epoch": 0.23,
"grad_norm": 1.5708709955215454,
"learning_rate": 9.826648351648352e-05,
"loss": 0.5606,
"step": 840
},
{
"epoch": 0.23,
"grad_norm": 1.521190881729126,
"learning_rate": 9.823901098901099e-05,
"loss": 0.5605,
"step": 850
},
{
"epoch": 0.23,
"grad_norm": 1.7570327520370483,
"learning_rate": 9.821153846153847e-05,
"loss": 0.5475,
"step": 860
},
{
"epoch": 0.24,
"grad_norm": 2.039400577545166,
"learning_rate": 9.818406593406594e-05,
"loss": 0.5451,
"step": 870
},
{
"epoch": 0.24,
"grad_norm": 1.9855557680130005,
"learning_rate": 9.815659340659341e-05,
"loss": 0.5378,
"step": 880
},
{
"epoch": 0.24,
"grad_norm": 1.3518325090408325,
"learning_rate": 9.812912087912089e-05,
"loss": 0.5447,
"step": 890
},
{
"epoch": 0.25,
"grad_norm": 2.36753511428833,
"learning_rate": 9.810164835164836e-05,
"loss": 0.5593,
"step": 900
},
{
"epoch": 0.25,
"grad_norm": 1.611568570137024,
"learning_rate": 9.807417582417583e-05,
"loss": 0.536,
"step": 910
},
{
"epoch": 0.25,
"grad_norm": 1.3011342287063599,
"learning_rate": 9.804670329670331e-05,
"loss": 0.5508,
"step": 920
},
{
"epoch": 0.25,
"grad_norm": 1.6673119068145752,
"learning_rate": 9.801923076923078e-05,
"loss": 0.5427,
"step": 930
},
{
"epoch": 0.26,
"grad_norm": 1.4966050386428833,
"learning_rate": 9.799175824175824e-05,
"loss": 0.5364,
"step": 940
},
{
"epoch": 0.26,
"grad_norm": 1.2000831365585327,
"learning_rate": 9.796428571428573e-05,
"loss": 0.5452,
"step": 950
},
{
"epoch": 0.26,
"grad_norm": 1.9247843027114868,
"learning_rate": 9.79368131868132e-05,
"loss": 0.529,
"step": 960
},
{
"epoch": 0.26,
"grad_norm": 1.6240330934524536,
"learning_rate": 9.790934065934066e-05,
"loss": 0.5523,
"step": 970
},
{
"epoch": 0.27,
"grad_norm": 1.4320908784866333,
"learning_rate": 9.788186813186814e-05,
"loss": 0.5567,
"step": 980
},
{
"epoch": 0.27,
"grad_norm": 1.7884244918823242,
"learning_rate": 9.785439560439561e-05,
"loss": 0.5244,
"step": 990
},
{
"epoch": 0.27,
"grad_norm": 1.9070675373077393,
"learning_rate": 9.782692307692308e-05,
"loss": 0.5354,
"step": 1000
},
{
"epoch": 0.28,
"grad_norm": 1.0304174423217773,
"learning_rate": 9.779945054945056e-05,
"loss": 0.5443,
"step": 1010
},
{
"epoch": 0.28,
"grad_norm": 2.184297800064087,
"learning_rate": 9.777197802197803e-05,
"loss": 0.5201,
"step": 1020
},
{
"epoch": 0.28,
"grad_norm": 1.2110681533813477,
"learning_rate": 9.77445054945055e-05,
"loss": 0.5363,
"step": 1030
},
{
"epoch": 0.28,
"grad_norm": 2.25685977935791,
"learning_rate": 9.771703296703298e-05,
"loss": 0.5392,
"step": 1040
},
{
"epoch": 0.29,
"grad_norm": 2.558457851409912,
"learning_rate": 9.768956043956045e-05,
"loss": 0.5466,
"step": 1050
},
{
"epoch": 0.29,
"grad_norm": 1.5898990631103516,
"learning_rate": 9.766208791208792e-05,
"loss": 0.5386,
"step": 1060
},
{
"epoch": 0.29,
"grad_norm": 2.6933228969573975,
"learning_rate": 9.76346153846154e-05,
"loss": 0.5565,
"step": 1070
},
{
"epoch": 0.3,
"grad_norm": 1.440042495727539,
"learning_rate": 9.760714285714287e-05,
"loss": 0.5507,
"step": 1080
},
{
"epoch": 0.3,
"grad_norm": 1.9476529359817505,
"learning_rate": 9.757967032967033e-05,
"loss": 0.5276,
"step": 1090
},
{
"epoch": 0.3,
"grad_norm": 3.142998218536377,
"learning_rate": 9.755219780219782e-05,
"loss": 0.5238,
"step": 1100
},
{
"epoch": 0.3,
"grad_norm": 2.13360857963562,
"learning_rate": 9.752472527472528e-05,
"loss": 0.5326,
"step": 1110
},
{
"epoch": 0.31,
"grad_norm": 1.2551095485687256,
"learning_rate": 9.749725274725275e-05,
"loss": 0.5569,
"step": 1120
},
{
"epoch": 0.31,
"grad_norm": 1.0858782529830933,
"learning_rate": 9.746978021978023e-05,
"loss": 0.5425,
"step": 1130
},
{
"epoch": 0.31,
"grad_norm": 1.3724364042282104,
"learning_rate": 9.74423076923077e-05,
"loss": 0.5381,
"step": 1140
},
{
"epoch": 0.31,
"grad_norm": 1.2211335897445679,
"learning_rate": 9.741483516483517e-05,
"loss": 0.5147,
"step": 1150
},
{
"epoch": 0.32,
"grad_norm": 1.806498646736145,
"learning_rate": 9.738736263736264e-05,
"loss": 0.5218,
"step": 1160
},
{
"epoch": 0.32,
"grad_norm": 0.7817394733428955,
"learning_rate": 9.735989010989012e-05,
"loss": 0.539,
"step": 1170
},
{
"epoch": 0.32,
"grad_norm": 2.054344415664673,
"learning_rate": 9.733241758241759e-05,
"loss": 0.5305,
"step": 1180
},
{
"epoch": 0.33,
"grad_norm": 0.8021188974380493,
"learning_rate": 9.730494505494506e-05,
"loss": 0.519,
"step": 1190
},
{
"epoch": 0.33,
"grad_norm": 1.833899736404419,
"learning_rate": 9.727747252747254e-05,
"loss": 0.5589,
"step": 1200
},
{
"epoch": 0.33,
"grad_norm": 2.6109519004821777,
"learning_rate": 9.725e-05,
"loss": 0.5333,
"step": 1210
},
{
"epoch": 0.33,
"grad_norm": 2.0866036415100098,
"learning_rate": 9.722252747252747e-05,
"loss": 0.5436,
"step": 1220
},
{
"epoch": 0.34,
"grad_norm": 2.150935649871826,
"learning_rate": 9.719505494505496e-05,
"loss": 0.5202,
"step": 1230
},
{
"epoch": 0.34,
"grad_norm": 1.280685544013977,
"learning_rate": 9.716758241758242e-05,
"loss": 0.5146,
"step": 1240
},
{
"epoch": 0.34,
"grad_norm": 0.7753365635871887,
"learning_rate": 9.714010989010989e-05,
"loss": 0.5106,
"step": 1250
},
{
"epoch": 0.34,
"grad_norm": 0.8706464767456055,
"learning_rate": 9.711263736263737e-05,
"loss": 0.4952,
"step": 1260
},
{
"epoch": 0.35,
"grad_norm": 1.3467265367507935,
"learning_rate": 9.708516483516484e-05,
"loss": 0.5105,
"step": 1270
},
{
"epoch": 0.35,
"grad_norm": 2.5618174076080322,
"learning_rate": 9.705769230769231e-05,
"loss": 0.5065,
"step": 1280
},
{
"epoch": 0.35,
"grad_norm": 1.3450369834899902,
"learning_rate": 9.703021978021979e-05,
"loss": 0.5201,
"step": 1290
},
{
"epoch": 0.36,
"grad_norm": 1.3272931575775146,
"learning_rate": 9.700274725274726e-05,
"loss": 0.5189,
"step": 1300
},
{
"epoch": 0.36,
"grad_norm": 2.324739694595337,
"learning_rate": 9.697527472527473e-05,
"loss": 0.5298,
"step": 1310
},
{
"epoch": 0.36,
"grad_norm": 1.375444769859314,
"learning_rate": 9.694780219780221e-05,
"loss": 0.527,
"step": 1320
},
{
"epoch": 0.36,
"grad_norm": 1.1685923337936401,
"learning_rate": 9.692032967032968e-05,
"loss": 0.5303,
"step": 1330
},
{
"epoch": 0.37,
"grad_norm": 1.6014115810394287,
"learning_rate": 9.689285714285715e-05,
"loss": 0.5399,
"step": 1340
},
{
"epoch": 0.37,
"grad_norm": 2.3533401489257812,
"learning_rate": 9.686538461538463e-05,
"loss": 0.5185,
"step": 1350
},
{
"epoch": 0.37,
"grad_norm": 1.6813533306121826,
"learning_rate": 9.68379120879121e-05,
"loss": 0.4986,
"step": 1360
},
{
"epoch": 0.37,
"grad_norm": 1.1102067232131958,
"learning_rate": 9.681043956043956e-05,
"loss": 0.5313,
"step": 1370
},
{
"epoch": 0.38,
"grad_norm": 1.576412320137024,
"learning_rate": 9.678296703296705e-05,
"loss": 0.5087,
"step": 1380
},
{
"epoch": 0.38,
"grad_norm": 1.5507354736328125,
"learning_rate": 9.675549450549451e-05,
"loss": 0.532,
"step": 1390
},
{
"epoch": 0.38,
"grad_norm": 1.4840365648269653,
"learning_rate": 9.672802197802198e-05,
"loss": 0.5191,
"step": 1400
},
{
"epoch": 0.39,
"grad_norm": 1.8767573833465576,
"learning_rate": 9.670054945054946e-05,
"loss": 0.5252,
"step": 1410
},
{
"epoch": 0.39,
"grad_norm": 1.0486218929290771,
"learning_rate": 9.667307692307693e-05,
"loss": 0.5152,
"step": 1420
},
{
"epoch": 0.39,
"grad_norm": 2.066174030303955,
"learning_rate": 9.66456043956044e-05,
"loss": 0.5163,
"step": 1430
},
{
"epoch": 0.39,
"grad_norm": 1.7480829954147339,
"learning_rate": 9.661813186813188e-05,
"loss": 0.5292,
"step": 1440
},
{
"epoch": 0.4,
"grad_norm": 1.6554828882217407,
"learning_rate": 9.659065934065935e-05,
"loss": 0.5301,
"step": 1450
},
{
"epoch": 0.4,
"grad_norm": 1.0723321437835693,
"learning_rate": 9.656318681318682e-05,
"loss": 0.5118,
"step": 1460
},
{
"epoch": 0.4,
"grad_norm": 1.7398673295974731,
"learning_rate": 9.653571428571429e-05,
"loss": 0.5268,
"step": 1470
},
{
"epoch": 0.4,
"grad_norm": 1.2732940912246704,
"learning_rate": 9.650824175824175e-05,
"loss": 0.5085,
"step": 1480
},
{
"epoch": 0.41,
"grad_norm": 1.0260119438171387,
"learning_rate": 9.648076923076924e-05,
"loss": 0.5243,
"step": 1490
},
{
"epoch": 0.41,
"grad_norm": 0.8816061019897461,
"learning_rate": 9.64532967032967e-05,
"loss": 0.5158,
"step": 1500
},
{
"epoch": 0.41,
"grad_norm": 2.7940518856048584,
"learning_rate": 9.642582417582417e-05,
"loss": 0.5051,
"step": 1510
},
{
"epoch": 0.42,
"grad_norm": 1.1889257431030273,
"learning_rate": 9.639835164835165e-05,
"loss": 0.5111,
"step": 1520
},
{
"epoch": 0.42,
"grad_norm": 1.8643608093261719,
"learning_rate": 9.637087912087912e-05,
"loss": 0.5136,
"step": 1530
},
{
"epoch": 0.42,
"grad_norm": 1.658339023590088,
"learning_rate": 9.634340659340659e-05,
"loss": 0.5132,
"step": 1540
},
{
"epoch": 0.42,
"grad_norm": 0.6262873411178589,
"learning_rate": 9.631593406593407e-05,
"loss": 0.5047,
"step": 1550
},
{
"epoch": 0.43,
"grad_norm": 1.2276610136032104,
"learning_rate": 9.628846153846154e-05,
"loss": 0.4997,
"step": 1560
},
{
"epoch": 0.43,
"grad_norm": 1.3661067485809326,
"learning_rate": 9.626098901098901e-05,
"loss": 0.5113,
"step": 1570
},
{
"epoch": 0.43,
"grad_norm": 1.423560380935669,
"learning_rate": 9.623351648351649e-05,
"loss": 0.5167,
"step": 1580
},
{
"epoch": 0.43,
"grad_norm": 1.391201376914978,
"learning_rate": 9.620604395604396e-05,
"loss": 0.5057,
"step": 1590
},
{
"epoch": 0.44,
"grad_norm": 1.5331265926361084,
"learning_rate": 9.617857142857143e-05,
"loss": 0.5105,
"step": 1600
},
{
"epoch": 0.44,
"grad_norm": 2.0285804271698,
"learning_rate": 9.615109890109891e-05,
"loss": 0.5009,
"step": 1610
},
{
"epoch": 0.44,
"grad_norm": 1.470106840133667,
"learning_rate": 9.612362637362638e-05,
"loss": 0.5082,
"step": 1620
},
{
"epoch": 0.45,
"grad_norm": 0.8231136798858643,
"learning_rate": 9.609615384615384e-05,
"loss": 0.4893,
"step": 1630
},
{
"epoch": 0.45,
"grad_norm": 1.4825439453125,
"learning_rate": 9.606868131868133e-05,
"loss": 0.5132,
"step": 1640
},
{
"epoch": 0.45,
"grad_norm": 1.8054287433624268,
"learning_rate": 9.60412087912088e-05,
"loss": 0.4878,
"step": 1650
},
{
"epoch": 0.45,
"grad_norm": 0.9411340355873108,
"learning_rate": 9.601373626373626e-05,
"loss": 0.4908,
"step": 1660
},
{
"epoch": 0.46,
"grad_norm": 1.754782795906067,
"learning_rate": 9.598626373626374e-05,
"loss": 0.5206,
"step": 1670
},
{
"epoch": 0.46,
"grad_norm": 1.1740814447402954,
"learning_rate": 9.595879120879121e-05,
"loss": 0.514,
"step": 1680
},
{
"epoch": 0.46,
"grad_norm": 1.0887936353683472,
"learning_rate": 9.593131868131868e-05,
"loss": 0.5095,
"step": 1690
},
{
"epoch": 0.46,
"grad_norm": 2.0167980194091797,
"learning_rate": 9.590384615384616e-05,
"loss": 0.5066,
"step": 1700
},
{
"epoch": 0.47,
"grad_norm": 1.6176742315292358,
"learning_rate": 9.587637362637363e-05,
"loss": 0.5016,
"step": 1710
},
{
"epoch": 0.47,
"grad_norm": 1.0261584520339966,
"learning_rate": 9.58489010989011e-05,
"loss": 0.5032,
"step": 1720
},
{
"epoch": 0.47,
"grad_norm": 1.0735365152359009,
"learning_rate": 9.582142857142858e-05,
"loss": 0.4894,
"step": 1730
},
{
"epoch": 0.48,
"grad_norm": 1.4293698072433472,
"learning_rate": 9.579395604395605e-05,
"loss": 0.5111,
"step": 1740
},
{
"epoch": 0.48,
"grad_norm": 1.296971321105957,
"learning_rate": 9.576648351648352e-05,
"loss": 0.5078,
"step": 1750
},
{
"epoch": 0.48,
"grad_norm": 0.7727321982383728,
"learning_rate": 9.5739010989011e-05,
"loss": 0.4923,
"step": 1760
},
{
"epoch": 0.48,
"grad_norm": 2.6028566360473633,
"learning_rate": 9.571153846153847e-05,
"loss": 0.4994,
"step": 1770
},
{
"epoch": 0.49,
"grad_norm": 1.2606744766235352,
"learning_rate": 9.568406593406593e-05,
"loss": 0.4885,
"step": 1780
},
{
"epoch": 0.49,
"grad_norm": 1.812680721282959,
"learning_rate": 9.565659340659342e-05,
"loss": 0.5101,
"step": 1790
},
{
"epoch": 0.49,
"grad_norm": 1.2432329654693604,
"learning_rate": 9.562912087912088e-05,
"loss": 0.5012,
"step": 1800
},
{
"epoch": 0.49,
"grad_norm": 1.365795373916626,
"learning_rate": 9.560164835164835e-05,
"loss": 0.5093,
"step": 1810
},
{
"epoch": 0.5,
"grad_norm": 1.4509589672088623,
"learning_rate": 9.557417582417583e-05,
"loss": 0.4921,
"step": 1820
},
{
"epoch": 0.5,
"grad_norm": 1.0986443758010864,
"learning_rate": 9.55467032967033e-05,
"loss": 0.5003,
"step": 1830
},
{
"epoch": 0.5,
"grad_norm": 1.46748685836792,
"learning_rate": 9.551923076923077e-05,
"loss": 0.4925,
"step": 1840
},
{
"epoch": 0.51,
"grad_norm": 1.3228458166122437,
"learning_rate": 9.549175824175825e-05,
"loss": 0.5032,
"step": 1850
},
{
"epoch": 0.51,
"grad_norm": 0.8542389273643494,
"learning_rate": 9.546428571428572e-05,
"loss": 0.4896,
"step": 1860
},
{
"epoch": 0.51,
"grad_norm": 1.7039434909820557,
"learning_rate": 9.543681318681319e-05,
"loss": 0.4852,
"step": 1870
},
{
"epoch": 0.51,
"grad_norm": 1.1597633361816406,
"learning_rate": 9.540934065934067e-05,
"loss": 0.5015,
"step": 1880
},
{
"epoch": 0.52,
"grad_norm": 2.178933620452881,
"learning_rate": 9.538186813186814e-05,
"loss": 0.5298,
"step": 1890
},
{
"epoch": 0.52,
"grad_norm": 1.1454273462295532,
"learning_rate": 9.53543956043956e-05,
"loss": 0.5046,
"step": 1900
},
{
"epoch": 0.52,
"grad_norm": 0.7407316565513611,
"learning_rate": 9.532692307692309e-05,
"loss": 0.486,
"step": 1910
},
{
"epoch": 0.52,
"grad_norm": 1.7394108772277832,
"learning_rate": 9.529945054945056e-05,
"loss": 0.4977,
"step": 1920
},
{
"epoch": 0.53,
"grad_norm": 0.9852740168571472,
"learning_rate": 9.527197802197802e-05,
"loss": 0.4871,
"step": 1930
},
{
"epoch": 0.53,
"grad_norm": 1.4632673263549805,
"learning_rate": 9.52445054945055e-05,
"loss": 0.5082,
"step": 1940
},
{
"epoch": 0.53,
"grad_norm": 1.4848504066467285,
"learning_rate": 9.521703296703297e-05,
"loss": 0.5067,
"step": 1950
},
{
"epoch": 0.54,
"grad_norm": 1.5140033960342407,
"learning_rate": 9.518956043956044e-05,
"loss": 0.5063,
"step": 1960
},
{
"epoch": 0.54,
"grad_norm": 2.4344804286956787,
"learning_rate": 9.516208791208791e-05,
"loss": 0.5099,
"step": 1970
},
{
"epoch": 0.54,
"grad_norm": 0.8636808395385742,
"learning_rate": 9.513461538461539e-05,
"loss": 0.4902,
"step": 1980
},
{
"epoch": 0.54,
"grad_norm": 1.4363267421722412,
"learning_rate": 9.510714285714286e-05,
"loss": 0.498,
"step": 1990
},
{
"epoch": 0.55,
"grad_norm": 0.7588983774185181,
"learning_rate": 9.507967032967033e-05,
"loss": 0.4884,
"step": 2000
},
{
"epoch": 0.55,
"grad_norm": 1.4090018272399902,
"learning_rate": 9.505219780219781e-05,
"loss": 0.5082,
"step": 2010
},
{
"epoch": 0.55,
"grad_norm": 2.2110331058502197,
"learning_rate": 9.502472527472528e-05,
"loss": 0.5064,
"step": 2020
},
{
"epoch": 0.55,
"grad_norm": 1.2035874128341675,
"learning_rate": 9.499725274725275e-05,
"loss": 0.4952,
"step": 2030
},
{
"epoch": 0.56,
"grad_norm": 1.597834825515747,
"learning_rate": 9.496978021978023e-05,
"loss": 0.4956,
"step": 2040
},
{
"epoch": 0.56,
"grad_norm": 1.4878921508789062,
"learning_rate": 9.49423076923077e-05,
"loss": 0.4865,
"step": 2050
},
{
"epoch": 0.56,
"grad_norm": 1.6747124195098877,
"learning_rate": 9.491483516483516e-05,
"loss": 0.4935,
"step": 2060
},
{
"epoch": 0.57,
"grad_norm": 1.3186380863189697,
"learning_rate": 9.488736263736264e-05,
"loss": 0.4938,
"step": 2070
},
{
"epoch": 0.57,
"grad_norm": 2.2865145206451416,
"learning_rate": 9.485989010989011e-05,
"loss": 0.4985,
"step": 2080
},
{
"epoch": 0.57,
"grad_norm": 2.515855550765991,
"learning_rate": 9.483241758241758e-05,
"loss": 0.5049,
"step": 2090
},
{
"epoch": 0.57,
"grad_norm": 1.2123080492019653,
"learning_rate": 9.480494505494506e-05,
"loss": 0.5004,
"step": 2100
},
{
"epoch": 0.58,
"grad_norm": 1.0919526815414429,
"learning_rate": 9.477747252747253e-05,
"loss": 0.4854,
"step": 2110
},
{
"epoch": 0.58,
"grad_norm": 1.3105583190917969,
"learning_rate": 9.475e-05,
"loss": 0.4866,
"step": 2120
},
{
"epoch": 0.58,
"grad_norm": 1.745622158050537,
"learning_rate": 9.472252747252748e-05,
"loss": 0.4892,
"step": 2130
},
{
"epoch": 0.58,
"grad_norm": 0.6633902192115784,
"learning_rate": 9.469505494505495e-05,
"loss": 0.507,
"step": 2140
},
{
"epoch": 0.59,
"grad_norm": 1.0530184507369995,
"learning_rate": 9.466758241758242e-05,
"loss": 0.5035,
"step": 2150
},
{
"epoch": 0.59,
"grad_norm": 1.2395071983337402,
"learning_rate": 9.46401098901099e-05,
"loss": 0.4926,
"step": 2160
},
{
"epoch": 0.59,
"grad_norm": 1.139224886894226,
"learning_rate": 9.461263736263737e-05,
"loss": 0.4861,
"step": 2170
},
{
"epoch": 0.6,
"grad_norm": 0.844968318939209,
"learning_rate": 9.458516483516484e-05,
"loss": 0.4855,
"step": 2180
},
{
"epoch": 0.6,
"grad_norm": 1.2388585805892944,
"learning_rate": 9.455769230769232e-05,
"loss": 0.4887,
"step": 2190
},
{
"epoch": 0.6,
"grad_norm": 1.4267388582229614,
"learning_rate": 9.453021978021978e-05,
"loss": 0.5031,
"step": 2200
},
{
"epoch": 0.6,
"grad_norm": 0.7797860503196716,
"learning_rate": 9.450274725274725e-05,
"loss": 0.5014,
"step": 2210
},
{
"epoch": 0.61,
"grad_norm": 1.3723804950714111,
"learning_rate": 9.447527472527473e-05,
"loss": 0.4887,
"step": 2220
},
{
"epoch": 0.61,
"grad_norm": 1.396022081375122,
"learning_rate": 9.44478021978022e-05,
"loss": 0.4869,
"step": 2230
},
{
"epoch": 0.61,
"grad_norm": 1.3555471897125244,
"learning_rate": 9.442032967032967e-05,
"loss": 0.4879,
"step": 2240
},
{
"epoch": 0.61,
"grad_norm": 0.9702646136283875,
"learning_rate": 9.439285714285715e-05,
"loss": 0.4894,
"step": 2250
},
{
"epoch": 0.62,
"grad_norm": 0.7753974199295044,
"learning_rate": 9.436538461538462e-05,
"loss": 0.5092,
"step": 2260
},
{
"epoch": 0.62,
"grad_norm": 1.020336389541626,
"learning_rate": 9.433791208791209e-05,
"loss": 0.4961,
"step": 2270
},
{
"epoch": 0.62,
"grad_norm": 1.2176239490509033,
"learning_rate": 9.431043956043957e-05,
"loss": 0.4827,
"step": 2280
},
{
"epoch": 0.63,
"grad_norm": 0.8337396383285522,
"learning_rate": 9.428296703296704e-05,
"loss": 0.4931,
"step": 2290
},
{
"epoch": 0.63,
"grad_norm": 0.8999798893928528,
"learning_rate": 9.425549450549451e-05,
"loss": 0.4912,
"step": 2300
},
{
"epoch": 0.63,
"grad_norm": 1.6992242336273193,
"learning_rate": 9.422802197802199e-05,
"loss": 0.4835,
"step": 2310
},
{
"epoch": 0.63,
"grad_norm": 0.7739867568016052,
"learning_rate": 9.420054945054946e-05,
"loss": 0.4783,
"step": 2320
},
{
"epoch": 0.64,
"grad_norm": 1.2350486516952515,
"learning_rate": 9.417307692307692e-05,
"loss": 0.4904,
"step": 2330
},
{
"epoch": 0.64,
"grad_norm": 1.1423838138580322,
"learning_rate": 9.41456043956044e-05,
"loss": 0.4884,
"step": 2340
},
{
"epoch": 0.64,
"grad_norm": 1.0562171936035156,
"learning_rate": 9.411813186813187e-05,
"loss": 0.5082,
"step": 2350
},
{
"epoch": 0.64,
"grad_norm": 0.7696706056594849,
"learning_rate": 9.409065934065934e-05,
"loss": 0.4889,
"step": 2360
},
{
"epoch": 0.65,
"grad_norm": 0.5980774760246277,
"learning_rate": 9.406318681318682e-05,
"loss": 0.48,
"step": 2370
},
{
"epoch": 0.65,
"grad_norm": 1.413926124572754,
"learning_rate": 9.403571428571429e-05,
"loss": 0.4971,
"step": 2380
},
{
"epoch": 0.65,
"grad_norm": 0.8020157217979431,
"learning_rate": 9.400824175824176e-05,
"loss": 0.4926,
"step": 2390
},
{
"epoch": 0.66,
"grad_norm": 0.6620104312896729,
"learning_rate": 9.398076923076924e-05,
"loss": 0.4941,
"step": 2400
},
{
"epoch": 0.66,
"grad_norm": 1.9766381978988647,
"learning_rate": 9.395329670329671e-05,
"loss": 0.4867,
"step": 2410
},
{
"epoch": 0.66,
"grad_norm": 1.3235608339309692,
"learning_rate": 9.392582417582418e-05,
"loss": 0.5061,
"step": 2420
},
{
"epoch": 0.66,
"grad_norm": 0.8266246914863586,
"learning_rate": 9.389835164835166e-05,
"loss": 0.4842,
"step": 2430
},
{
"epoch": 0.67,
"grad_norm": 1.1611332893371582,
"learning_rate": 9.387087912087913e-05,
"loss": 0.4795,
"step": 2440
},
{
"epoch": 0.67,
"grad_norm": 1.750346064567566,
"learning_rate": 9.38434065934066e-05,
"loss": 0.4859,
"step": 2450
},
{
"epoch": 0.67,
"grad_norm": 0.7080681324005127,
"learning_rate": 9.381593406593408e-05,
"loss": 0.4762,
"step": 2460
},
{
"epoch": 0.67,
"grad_norm": 1.3531988859176636,
"learning_rate": 9.378846153846155e-05,
"loss": 0.4921,
"step": 2470
},
{
"epoch": 0.68,
"grad_norm": 0.8100531697273254,
"learning_rate": 9.376098901098901e-05,
"loss": 0.4912,
"step": 2480
},
{
"epoch": 0.68,
"grad_norm": 1.6417909860610962,
"learning_rate": 9.37335164835165e-05,
"loss": 0.4868,
"step": 2490
},
{
"epoch": 0.68,
"grad_norm": 2.7357239723205566,
"learning_rate": 9.370604395604396e-05,
"loss": 0.4837,
"step": 2500
},
{
"epoch": 0.69,
"grad_norm": 0.7493315935134888,
"learning_rate": 9.367857142857143e-05,
"loss": 0.494,
"step": 2510
},
{
"epoch": 0.69,
"grad_norm": 1.4594992399215698,
"learning_rate": 9.365109890109891e-05,
"loss": 0.4967,
"step": 2520
},
{
"epoch": 0.69,
"grad_norm": 1.2568315267562866,
"learning_rate": 9.362362637362638e-05,
"loss": 0.4865,
"step": 2530
},
{
"epoch": 0.69,
"grad_norm": 0.8711192607879639,
"learning_rate": 9.359615384615385e-05,
"loss": 0.4915,
"step": 2540
},
{
"epoch": 0.7,
"grad_norm": 1.6470551490783691,
"learning_rate": 9.356868131868133e-05,
"loss": 0.4736,
"step": 2550
},
{
"epoch": 0.7,
"grad_norm": 2.410200357437134,
"learning_rate": 9.35412087912088e-05,
"loss": 0.5078,
"step": 2560
},
{
"epoch": 0.7,
"grad_norm": 1.0611746311187744,
"learning_rate": 9.351373626373627e-05,
"loss": 0.4986,
"step": 2570
},
{
"epoch": 0.7,
"grad_norm": 0.820321798324585,
"learning_rate": 9.348626373626375e-05,
"loss": 0.4788,
"step": 2580
},
{
"epoch": 0.71,
"grad_norm": 0.7887302041053772,
"learning_rate": 9.345879120879122e-05,
"loss": 0.4878,
"step": 2590
},
{
"epoch": 0.71,
"grad_norm": 1.2939894199371338,
"learning_rate": 9.343131868131869e-05,
"loss": 0.4896,
"step": 2600
},
{
"epoch": 0.71,
"grad_norm": 0.7468405961990356,
"learning_rate": 9.340384615384615e-05,
"loss": 0.4932,
"step": 2610
},
{
"epoch": 0.72,
"grad_norm": 1.63861882686615,
"learning_rate": 9.337637362637362e-05,
"loss": 0.4739,
"step": 2620
},
{
"epoch": 0.72,
"grad_norm": 1.1842321157455444,
"learning_rate": 9.33489010989011e-05,
"loss": 0.4845,
"step": 2630
},
{
"epoch": 0.72,
"grad_norm": 1.0613666772842407,
"learning_rate": 9.332142857142857e-05,
"loss": 0.4806,
"step": 2640
},
{
"epoch": 0.72,
"grad_norm": 0.9317752718925476,
"learning_rate": 9.329395604395604e-05,
"loss": 0.4803,
"step": 2650
},
{
"epoch": 0.73,
"grad_norm": 1.378394603729248,
"learning_rate": 9.326648351648352e-05,
"loss": 0.4942,
"step": 2660
},
{
"epoch": 0.73,
"grad_norm": 0.748068630695343,
"learning_rate": 9.323901098901099e-05,
"loss": 0.4763,
"step": 2670
},
{
"epoch": 0.73,
"grad_norm": 1.4142990112304688,
"learning_rate": 9.321153846153846e-05,
"loss": 0.4913,
"step": 2680
},
{
"epoch": 0.73,
"grad_norm": 0.6746355891227722,
"learning_rate": 9.318406593406594e-05,
"loss": 0.4781,
"step": 2690
},
{
"epoch": 0.74,
"grad_norm": 1.146001935005188,
"learning_rate": 9.315659340659341e-05,
"loss": 0.4789,
"step": 2700
},
{
"epoch": 0.74,
"grad_norm": 1.0291472673416138,
"learning_rate": 9.312912087912088e-05,
"loss": 0.4813,
"step": 2710
},
{
"epoch": 0.74,
"grad_norm": 1.7905545234680176,
"learning_rate": 9.310164835164836e-05,
"loss": 0.475,
"step": 2720
},
{
"epoch": 0.75,
"grad_norm": 0.8916068077087402,
"learning_rate": 9.307417582417583e-05,
"loss": 0.4891,
"step": 2730
},
{
"epoch": 0.75,
"grad_norm": 1.0133769512176514,
"learning_rate": 9.30467032967033e-05,
"loss": 0.4857,
"step": 2740
},
{
"epoch": 0.75,
"grad_norm": 1.6718567609786987,
"learning_rate": 9.301923076923078e-05,
"loss": 0.4915,
"step": 2750
},
{
"epoch": 0.75,
"grad_norm": 1.2999911308288574,
"learning_rate": 9.299175824175824e-05,
"loss": 0.4872,
"step": 2760
},
{
"epoch": 0.76,
"grad_norm": 1.4812780618667603,
"learning_rate": 9.296428571428571e-05,
"loss": 0.4886,
"step": 2770
},
{
"epoch": 0.76,
"grad_norm": 1.3728208541870117,
"learning_rate": 9.293681318681318e-05,
"loss": 0.5005,
"step": 2780
},
{
"epoch": 0.76,
"grad_norm": 1.2773633003234863,
"learning_rate": 9.290934065934066e-05,
"loss": 0.4796,
"step": 2790
},
{
"epoch": 0.76,
"grad_norm": 0.8197434544563293,
"learning_rate": 9.288186813186813e-05,
"loss": 0.4817,
"step": 2800
},
{
"epoch": 0.77,
"grad_norm": 0.830235481262207,
"learning_rate": 9.28543956043956e-05,
"loss": 0.4956,
"step": 2810
},
{
"epoch": 0.77,
"grad_norm": 3.1286449432373047,
"learning_rate": 9.282692307692308e-05,
"loss": 0.4822,
"step": 2820
},
{
"epoch": 0.77,
"grad_norm": 0.9349818825721741,
"learning_rate": 9.279945054945055e-05,
"loss": 0.4924,
"step": 2830
},
{
"epoch": 0.78,
"grad_norm": 1.8471746444702148,
"learning_rate": 9.277197802197802e-05,
"loss": 0.4771,
"step": 2840
},
{
"epoch": 0.78,
"grad_norm": 0.9180948734283447,
"learning_rate": 9.27445054945055e-05,
"loss": 0.4855,
"step": 2850
},
{
"epoch": 0.78,
"grad_norm": 0.7180835604667664,
"learning_rate": 9.271703296703297e-05,
"loss": 0.4801,
"step": 2860
},
{
"epoch": 0.78,
"grad_norm": 1.5345978736877441,
"learning_rate": 9.268956043956043e-05,
"loss": 0.4791,
"step": 2870
},
{
"epoch": 0.79,
"grad_norm": 1.521747350692749,
"learning_rate": 9.266208791208792e-05,
"loss": 0.4793,
"step": 2880
},
{
"epoch": 0.79,
"grad_norm": 1.2364057302474976,
"learning_rate": 9.263461538461538e-05,
"loss": 0.4756,
"step": 2890
},
{
"epoch": 0.79,
"grad_norm": 0.5741211771965027,
"learning_rate": 9.260714285714285e-05,
"loss": 0.4849,
"step": 2900
},
{
"epoch": 0.79,
"grad_norm": 1.7251524925231934,
"learning_rate": 9.257967032967033e-05,
"loss": 0.4762,
"step": 2910
},
{
"epoch": 0.8,
"grad_norm": 2.704491138458252,
"learning_rate": 9.25521978021978e-05,
"loss": 0.4777,
"step": 2920
},
{
"epoch": 0.8,
"grad_norm": 0.9178886413574219,
"learning_rate": 9.252472527472527e-05,
"loss": 0.5077,
"step": 2930
},
{
"epoch": 0.8,
"grad_norm": 1.2007173299789429,
"learning_rate": 9.249725274725275e-05,
"loss": 0.4916,
"step": 2940
},
{
"epoch": 0.81,
"grad_norm": 0.8556911945343018,
"learning_rate": 9.246978021978022e-05,
"loss": 0.477,
"step": 2950
},
{
"epoch": 0.81,
"grad_norm": 0.7754554152488708,
"learning_rate": 9.244230769230769e-05,
"loss": 0.4859,
"step": 2960
},
{
"epoch": 0.81,
"grad_norm": 2.333521842956543,
"learning_rate": 9.241483516483517e-05,
"loss": 0.4811,
"step": 2970
},
{
"epoch": 0.81,
"grad_norm": 0.9892757534980774,
"learning_rate": 9.238736263736264e-05,
"loss": 0.4777,
"step": 2980
},
{
"epoch": 0.82,
"grad_norm": 1.0326083898544312,
"learning_rate": 9.23598901098901e-05,
"loss": 0.4891,
"step": 2990
},
{
"epoch": 0.82,
"grad_norm": 1.4555460214614868,
"learning_rate": 9.233241758241759e-05,
"loss": 0.4802,
"step": 3000
},
{
"epoch": 0.82,
"grad_norm": 0.8390610814094543,
"learning_rate": 9.230494505494506e-05,
"loss": 0.4835,
"step": 3010
},
{
"epoch": 0.82,
"grad_norm": 1.0490790605545044,
"learning_rate": 9.227747252747252e-05,
"loss": 0.4806,
"step": 3020
},
{
"epoch": 0.83,
"grad_norm": 0.8793150186538696,
"learning_rate": 9.225e-05,
"loss": 0.4846,
"step": 3030
},
{
"epoch": 0.83,
"grad_norm": 1.2842731475830078,
"learning_rate": 9.222252747252747e-05,
"loss": 0.4715,
"step": 3040
},
{
"epoch": 0.83,
"grad_norm": 0.7668898701667786,
"learning_rate": 9.219505494505494e-05,
"loss": 0.4749,
"step": 3050
},
{
"epoch": 0.84,
"grad_norm": 0.8144903182983398,
"learning_rate": 9.216758241758242e-05,
"loss": 0.4705,
"step": 3060
},
{
"epoch": 0.84,
"grad_norm": 0.6727710962295532,
"learning_rate": 9.214010989010989e-05,
"loss": 0.4826,
"step": 3070
},
{
"epoch": 0.84,
"grad_norm": 0.8702958822250366,
"learning_rate": 9.211263736263736e-05,
"loss": 0.4679,
"step": 3080
},
{
"epoch": 0.84,
"grad_norm": 0.9223732352256775,
"learning_rate": 9.208516483516484e-05,
"loss": 0.4875,
"step": 3090
},
{
"epoch": 0.85,
"grad_norm": 0.7086817622184753,
"learning_rate": 9.205769230769231e-05,
"loss": 0.4771,
"step": 3100
},
{
"epoch": 0.85,
"grad_norm": 1.000447154045105,
"learning_rate": 9.203021978021978e-05,
"loss": 0.4674,
"step": 3110
},
{
"epoch": 0.85,
"grad_norm": 1.6982853412628174,
"learning_rate": 9.200274725274726e-05,
"loss": 0.4853,
"step": 3120
},
{
"epoch": 0.86,
"grad_norm": 0.8975854516029358,
"learning_rate": 9.197527472527473e-05,
"loss": 0.4757,
"step": 3130
},
{
"epoch": 0.86,
"grad_norm": 1.638816237449646,
"learning_rate": 9.19478021978022e-05,
"loss": 0.4862,
"step": 3140
},
{
"epoch": 0.86,
"grad_norm": 1.0290194749832153,
"learning_rate": 9.192032967032968e-05,
"loss": 0.4758,
"step": 3150
},
{
"epoch": 0.86,
"grad_norm": 1.0205775499343872,
"learning_rate": 9.189285714285715e-05,
"loss": 0.4723,
"step": 3160
},
{
"epoch": 0.87,
"grad_norm": 1.2193140983581543,
"learning_rate": 9.186538461538461e-05,
"loss": 0.4869,
"step": 3170
},
{
"epoch": 0.87,
"grad_norm": 1.0402013063430786,
"learning_rate": 9.18379120879121e-05,
"loss": 0.4911,
"step": 3180
},
{
"epoch": 0.87,
"grad_norm": 1.1915889978408813,
"learning_rate": 9.181043956043956e-05,
"loss": 0.4757,
"step": 3190
},
{
"epoch": 0.87,
"grad_norm": 1.2371594905853271,
"learning_rate": 9.178296703296703e-05,
"loss": 0.4811,
"step": 3200
},
{
"epoch": 0.88,
"grad_norm": 0.9872990846633911,
"learning_rate": 9.175549450549451e-05,
"loss": 0.4784,
"step": 3210
},
{
"epoch": 0.88,
"grad_norm": 0.9502705335617065,
"learning_rate": 9.172802197802198e-05,
"loss": 0.4716,
"step": 3220
},
{
"epoch": 0.88,
"grad_norm": 1.9108206033706665,
"learning_rate": 9.170054945054945e-05,
"loss": 0.4828,
"step": 3230
},
{
"epoch": 0.89,
"grad_norm": 0.5662224888801575,
"learning_rate": 9.167307692307693e-05,
"loss": 0.485,
"step": 3240
},
{
"epoch": 0.89,
"grad_norm": 0.9142247438430786,
"learning_rate": 9.16456043956044e-05,
"loss": 0.4851,
"step": 3250
},
{
"epoch": 0.89,
"grad_norm": 1.817579984664917,
"learning_rate": 9.161813186813187e-05,
"loss": 0.4807,
"step": 3260
},
{
"epoch": 0.89,
"grad_norm": 1.3611973524093628,
"learning_rate": 9.159065934065935e-05,
"loss": 0.4815,
"step": 3270
},
{
"epoch": 0.9,
"grad_norm": 1.4995872974395752,
"learning_rate": 9.156318681318682e-05,
"loss": 0.4779,
"step": 3280
},
{
"epoch": 0.9,
"grad_norm": 0.8315975069999695,
"learning_rate": 9.153571428571429e-05,
"loss": 0.4732,
"step": 3290
},
{
"epoch": 0.9,
"grad_norm": 1.285447359085083,
"learning_rate": 9.150824175824177e-05,
"loss": 0.4769,
"step": 3300
},
{
"epoch": 0.9,
"grad_norm": 1.265435814857483,
"learning_rate": 9.148076923076924e-05,
"loss": 0.4884,
"step": 3310
},
{
"epoch": 0.91,
"grad_norm": 0.7437517046928406,
"learning_rate": 9.14532967032967e-05,
"loss": 0.4807,
"step": 3320
},
{
"epoch": 0.91,
"grad_norm": 0.8307008743286133,
"learning_rate": 9.142582417582419e-05,
"loss": 0.4723,
"step": 3330
},
{
"epoch": 0.91,
"grad_norm": 1.2148619890213013,
"learning_rate": 9.139835164835165e-05,
"loss": 0.461,
"step": 3340
},
{
"epoch": 0.92,
"grad_norm": 0.5682019591331482,
"learning_rate": 9.137087912087912e-05,
"loss": 0.4673,
"step": 3350
},
{
"epoch": 0.92,
"grad_norm": 0.582708477973938,
"learning_rate": 9.13434065934066e-05,
"loss": 0.4657,
"step": 3360
},
{
"epoch": 0.92,
"grad_norm": 1.0724083185195923,
"learning_rate": 9.131593406593407e-05,
"loss": 0.4634,
"step": 3370
},
{
"epoch": 0.92,
"grad_norm": 1.511168360710144,
"learning_rate": 9.128846153846154e-05,
"loss": 0.4888,
"step": 3380
},
{
"epoch": 0.93,
"grad_norm": 1.5445438623428345,
"learning_rate": 9.126098901098902e-05,
"loss": 0.4757,
"step": 3390
},
{
"epoch": 0.93,
"grad_norm": 0.9609503746032715,
"learning_rate": 9.123351648351649e-05,
"loss": 0.4813,
"step": 3400
},
{
"epoch": 0.93,
"grad_norm": 2.094918966293335,
"learning_rate": 9.120604395604396e-05,
"loss": 0.4963,
"step": 3410
},
{
"epoch": 0.93,
"grad_norm": 1.6844788789749146,
"learning_rate": 9.117857142857144e-05,
"loss": 0.4902,
"step": 3420
},
{
"epoch": 0.94,
"grad_norm": 1.2747464179992676,
"learning_rate": 9.115109890109891e-05,
"loss": 0.4884,
"step": 3430
},
{
"epoch": 0.94,
"grad_norm": 1.4035594463348389,
"learning_rate": 9.112362637362638e-05,
"loss": 0.482,
"step": 3440
},
{
"epoch": 0.94,
"grad_norm": 0.6047748923301697,
"learning_rate": 9.109615384615386e-05,
"loss": 0.4846,
"step": 3450
},
{
"epoch": 0.95,
"grad_norm": 1.6141401529312134,
"learning_rate": 9.106868131868133e-05,
"loss": 0.4744,
"step": 3460
},
{
"epoch": 0.95,
"grad_norm": 0.8329553008079529,
"learning_rate": 9.10412087912088e-05,
"loss": 0.4738,
"step": 3470
},
{
"epoch": 0.95,
"grad_norm": 0.7500654458999634,
"learning_rate": 9.101373626373628e-05,
"loss": 0.4763,
"step": 3480
},
{
"epoch": 0.95,
"grad_norm": 0.8617835640907288,
"learning_rate": 9.098626373626374e-05,
"loss": 0.4738,
"step": 3490
},
{
"epoch": 0.96,
"grad_norm": 1.0339410305023193,
"learning_rate": 9.095879120879121e-05,
"loss": 0.4726,
"step": 3500
},
{
"epoch": 0.96,
"grad_norm": 1.13742995262146,
"learning_rate": 9.09313186813187e-05,
"loss": 0.4827,
"step": 3510
},
{
"epoch": 0.96,
"grad_norm": 1.611344575881958,
"learning_rate": 9.090384615384616e-05,
"loss": 0.4844,
"step": 3520
},
{
"epoch": 0.96,
"grad_norm": 1.056773066520691,
"learning_rate": 9.087637362637363e-05,
"loss": 0.4799,
"step": 3530
},
{
"epoch": 0.97,
"grad_norm": 1.2290446758270264,
"learning_rate": 9.084890109890111e-05,
"loss": 0.4764,
"step": 3540
},
{
"epoch": 0.97,
"grad_norm": 0.7461050748825073,
"learning_rate": 9.082142857142858e-05,
"loss": 0.4664,
"step": 3550
},
{
"epoch": 0.97,
"grad_norm": 0.5372506380081177,
"learning_rate": 9.079395604395605e-05,
"loss": 0.4795,
"step": 3560
},
{
"epoch": 0.98,
"grad_norm": 1.658233642578125,
"learning_rate": 9.076648351648353e-05,
"loss": 0.4655,
"step": 3570
},
{
"epoch": 0.98,
"grad_norm": 1.7918599843978882,
"learning_rate": 9.0739010989011e-05,
"loss": 0.4777,
"step": 3580
},
{
"epoch": 0.98,
"grad_norm": 0.9504883289337158,
"learning_rate": 9.071153846153847e-05,
"loss": 0.4791,
"step": 3590
},
{
"epoch": 0.98,
"grad_norm": 1.5631986856460571,
"learning_rate": 9.068406593406595e-05,
"loss": 0.4807,
"step": 3600
},
{
"epoch": 0.99,
"grad_norm": 0.9991289377212524,
"learning_rate": 9.065659340659342e-05,
"loss": 0.4679,
"step": 3610
},
{
"epoch": 0.99,
"grad_norm": 1.130434274673462,
"learning_rate": 9.062912087912088e-05,
"loss": 0.4781,
"step": 3620
},
{
"epoch": 0.99,
"grad_norm": 0.639451265335083,
"learning_rate": 9.060164835164837e-05,
"loss": 0.4815,
"step": 3630
},
{
"epoch": 0.99,
"grad_norm": 1.1776291131973267,
"learning_rate": 9.057417582417583e-05,
"loss": 0.4983,
"step": 3640
},
{
"epoch": 1.0,
"grad_norm": 1.4073742628097534,
"learning_rate": 9.05467032967033e-05,
"loss": 0.4711,
"step": 3650
},
{
"epoch": 1.0,
"grad_norm": 1.308344841003418,
"learning_rate": 9.051923076923078e-05,
"loss": 0.4847,
"step": 3660
},
{
"epoch": 1.0,
"eval_loss": 0.44810301065444946,
"eval_runtime": 1256.2985,
"eval_samples_per_second": 232.516,
"eval_steps_per_second": 3.634,
"step": 3660
},
{
"epoch": 1.0,
"grad_norm": 0.6774665117263794,
"learning_rate": 9.049175824175825e-05,
"loss": 0.4791,
"step": 3670
},
{
"epoch": 1.01,
"grad_norm": 0.6826792359352112,
"learning_rate": 9.046428571428572e-05,
"loss": 0.4775,
"step": 3680
},
{
"epoch": 1.01,
"grad_norm": 0.7845907211303711,
"learning_rate": 9.04368131868132e-05,
"loss": 0.4765,
"step": 3690
},
{
"epoch": 1.01,
"grad_norm": 0.9099870324134827,
"learning_rate": 9.040934065934067e-05,
"loss": 0.4689,
"step": 3700
},
{
"epoch": 1.01,
"grad_norm": 1.374972939491272,
"learning_rate": 9.038186813186814e-05,
"loss": 0.469,
"step": 3710
},
{
"epoch": 1.02,
"grad_norm": 0.8698770403862,
"learning_rate": 9.035439560439562e-05,
"loss": 0.4717,
"step": 3720
},
{
"epoch": 1.02,
"grad_norm": 0.8832442164421082,
"learning_rate": 9.032692307692309e-05,
"loss": 0.4653,
"step": 3730
},
{
"epoch": 1.02,
"grad_norm": 0.722259521484375,
"learning_rate": 9.029945054945054e-05,
"loss": 0.4818,
"step": 3740
},
{
"epoch": 1.02,
"grad_norm": 1.1362375020980835,
"learning_rate": 9.027197802197802e-05,
"loss": 0.4674,
"step": 3750
},
{
"epoch": 1.03,
"grad_norm": 1.0420037508010864,
"learning_rate": 9.024450549450549e-05,
"loss": 0.4806,
"step": 3760
},
{
"epoch": 1.03,
"grad_norm": 0.9857313632965088,
"learning_rate": 9.021703296703296e-05,
"loss": 0.478,
"step": 3770
},
{
"epoch": 1.03,
"grad_norm": 1.1428401470184326,
"learning_rate": 9.018956043956044e-05,
"loss": 0.4742,
"step": 3780
},
{
"epoch": 1.04,
"grad_norm": 1.5837888717651367,
"learning_rate": 9.016208791208791e-05,
"loss": 0.4841,
"step": 3790
},
{
"epoch": 1.04,
"grad_norm": 1.9123620986938477,
"learning_rate": 9.013461538461538e-05,
"loss": 0.4707,
"step": 3800
},
{
"epoch": 1.04,
"grad_norm": 0.9019024968147278,
"learning_rate": 9.010714285714286e-05,
"loss": 0.4731,
"step": 3810
},
{
"epoch": 1.04,
"grad_norm": 0.9601519107818604,
"learning_rate": 9.007967032967033e-05,
"loss": 0.4644,
"step": 3820
},
{
"epoch": 1.05,
"grad_norm": 0.4769773483276367,
"learning_rate": 9.00521978021978e-05,
"loss": 0.4755,
"step": 3830
},
{
"epoch": 1.05,
"grad_norm": 1.2578529119491577,
"learning_rate": 9.002472527472528e-05,
"loss": 0.4703,
"step": 3840
},
{
"epoch": 1.05,
"grad_norm": 1.1451194286346436,
"learning_rate": 8.999725274725275e-05,
"loss": 0.4748,
"step": 3850
},
{
"epoch": 1.05,
"grad_norm": 0.4718180000782013,
"learning_rate": 8.996978021978021e-05,
"loss": 0.4673,
"step": 3860
},
{
"epoch": 1.06,
"grad_norm": 1.0323463678359985,
"learning_rate": 8.99423076923077e-05,
"loss": 0.4611,
"step": 3870
},
{
"epoch": 1.06,
"grad_norm": 0.905623197555542,
"learning_rate": 8.991483516483516e-05,
"loss": 0.4651,
"step": 3880
},
{
"epoch": 1.06,
"grad_norm": 0.6803371906280518,
"learning_rate": 8.988736263736263e-05,
"loss": 0.4693,
"step": 3890
},
{
"epoch": 1.07,
"grad_norm": 0.7653062343597412,
"learning_rate": 8.985989010989011e-05,
"loss": 0.4676,
"step": 3900
},
{
"epoch": 1.07,
"grad_norm": 0.6493901610374451,
"learning_rate": 8.983241758241758e-05,
"loss": 0.4796,
"step": 3910
},
{
"epoch": 1.07,
"grad_norm": 1.3194290399551392,
"learning_rate": 8.980494505494505e-05,
"loss": 0.4709,
"step": 3920
},
{
"epoch": 1.07,
"grad_norm": 1.3907560110092163,
"learning_rate": 8.977747252747253e-05,
"loss": 0.4791,
"step": 3930
},
{
"epoch": 1.08,
"grad_norm": 0.9326291084289551,
"learning_rate": 8.975e-05,
"loss": 0.4733,
"step": 3940
},
{
"epoch": 1.08,
"grad_norm": 0.6711992025375366,
"learning_rate": 8.972252747252747e-05,
"loss": 0.4607,
"step": 3950
},
{
"epoch": 1.08,
"grad_norm": 1.0052268505096436,
"learning_rate": 8.969505494505495e-05,
"loss": 0.4565,
"step": 3960
},
{
"epoch": 1.08,
"grad_norm": 1.893130898475647,
"learning_rate": 8.966758241758242e-05,
"loss": 0.4678,
"step": 3970
},
{
"epoch": 1.09,
"grad_norm": 1.019831657409668,
"learning_rate": 8.964010989010989e-05,
"loss": 0.4929,
"step": 3980
},
{
"epoch": 1.09,
"grad_norm": 1.096341848373413,
"learning_rate": 8.961263736263737e-05,
"loss": 0.4812,
"step": 3990
},
{
"epoch": 1.09,
"grad_norm": 1.3617221117019653,
"learning_rate": 8.958516483516484e-05,
"loss": 0.479,
"step": 4000
},
{
"epoch": 1.1,
"grad_norm": 0.6594845056533813,
"learning_rate": 8.95576923076923e-05,
"loss": 0.4784,
"step": 4010
},
{
"epoch": 1.1,
"grad_norm": 0.7108995914459229,
"learning_rate": 8.953021978021979e-05,
"loss": 0.4624,
"step": 4020
},
{
"epoch": 1.1,
"grad_norm": 0.4914230704307556,
"learning_rate": 8.950274725274725e-05,
"loss": 0.4573,
"step": 4030
},
{
"epoch": 1.1,
"grad_norm": 0.738325834274292,
"learning_rate": 8.947527472527472e-05,
"loss": 0.4598,
"step": 4040
},
{
"epoch": 1.11,
"grad_norm": 1.5622313022613525,
"learning_rate": 8.94478021978022e-05,
"loss": 0.4769,
"step": 4050
},
{
"epoch": 1.11,
"grad_norm": 0.84259033203125,
"learning_rate": 8.942032967032967e-05,
"loss": 0.4587,
"step": 4060
},
{
"epoch": 1.11,
"grad_norm": 1.1515220403671265,
"learning_rate": 8.939285714285714e-05,
"loss": 0.4698,
"step": 4070
},
{
"epoch": 1.11,
"grad_norm": 1.2757478952407837,
"learning_rate": 8.936538461538462e-05,
"loss": 0.4662,
"step": 4080
},
{
"epoch": 1.12,
"grad_norm": 1.413041591644287,
"learning_rate": 8.933791208791209e-05,
"loss": 0.4739,
"step": 4090
},
{
"epoch": 1.12,
"grad_norm": 1.3950985670089722,
"learning_rate": 8.931043956043956e-05,
"loss": 0.4652,
"step": 4100
},
{
"epoch": 1.12,
"grad_norm": 1.1247773170471191,
"learning_rate": 8.928296703296704e-05,
"loss": 0.4704,
"step": 4110
},
{
"epoch": 1.13,
"grad_norm": 0.7197868824005127,
"learning_rate": 8.925549450549451e-05,
"loss": 0.466,
"step": 4120
},
{
"epoch": 1.13,
"grad_norm": 1.606062889099121,
"learning_rate": 8.922802197802198e-05,
"loss": 0.4697,
"step": 4130
},
{
"epoch": 1.13,
"grad_norm": 0.8834720849990845,
"learning_rate": 8.920054945054946e-05,
"loss": 0.4787,
"step": 4140
},
{
"epoch": 1.13,
"grad_norm": 0.6989302635192871,
"learning_rate": 8.917307692307693e-05,
"loss": 0.4553,
"step": 4150
},
{
"epoch": 1.14,
"grad_norm": 1.1178115606307983,
"learning_rate": 8.91456043956044e-05,
"loss": 0.4679,
"step": 4160
},
{
"epoch": 1.14,
"grad_norm": 0.8493313789367676,
"learning_rate": 8.911813186813188e-05,
"loss": 0.4915,
"step": 4170
},
{
"epoch": 1.14,
"grad_norm": 0.6862457990646362,
"learning_rate": 8.909065934065934e-05,
"loss": 0.4667,
"step": 4180
},
{
"epoch": 1.14,
"grad_norm": 0.5265287160873413,
"learning_rate": 8.906318681318681e-05,
"loss": 0.4675,
"step": 4190
},
{
"epoch": 1.15,
"grad_norm": 1.6895486116409302,
"learning_rate": 8.903571428571429e-05,
"loss": 0.4694,
"step": 4200
},
{
"epoch": 1.15,
"grad_norm": 0.9672292470932007,
"learning_rate": 8.900824175824176e-05,
"loss": 0.4788,
"step": 4210
},
{
"epoch": 1.15,
"grad_norm": 0.9348291158676147,
"learning_rate": 8.898076923076923e-05,
"loss": 0.461,
"step": 4220
},
{
"epoch": 1.16,
"grad_norm": 0.5047799348831177,
"learning_rate": 8.895329670329671e-05,
"loss": 0.467,
"step": 4230
},
{
"epoch": 1.16,
"grad_norm": 1.6985957622528076,
"learning_rate": 8.892582417582418e-05,
"loss": 0.4581,
"step": 4240
},
{
"epoch": 1.16,
"grad_norm": 0.8573766946792603,
"learning_rate": 8.889835164835165e-05,
"loss": 0.4624,
"step": 4250
},
{
"epoch": 1.16,
"grad_norm": 0.6403746008872986,
"learning_rate": 8.887087912087913e-05,
"loss": 0.4574,
"step": 4260
},
{
"epoch": 1.17,
"grad_norm": 0.9546878933906555,
"learning_rate": 8.88434065934066e-05,
"loss": 0.4803,
"step": 4270
},
{
"epoch": 1.17,
"grad_norm": 0.6992652416229248,
"learning_rate": 8.881593406593407e-05,
"loss": 0.4696,
"step": 4280
},
{
"epoch": 1.17,
"grad_norm": 1.035543441772461,
"learning_rate": 8.878846153846155e-05,
"loss": 0.4787,
"step": 4290
},
{
"epoch": 1.17,
"grad_norm": 1.5563474893569946,
"learning_rate": 8.876098901098902e-05,
"loss": 0.4708,
"step": 4300
},
{
"epoch": 1.18,
"grad_norm": 1.7203930616378784,
"learning_rate": 8.873351648351648e-05,
"loss": 0.4725,
"step": 4310
},
{
"epoch": 1.18,
"grad_norm": 1.4295490980148315,
"learning_rate": 8.870604395604396e-05,
"loss": 0.464,
"step": 4320
},
{
"epoch": 1.18,
"grad_norm": 1.1832698583602905,
"learning_rate": 8.867857142857143e-05,
"loss": 0.475,
"step": 4330
},
{
"epoch": 1.19,
"grad_norm": 0.9679028391838074,
"learning_rate": 8.86510989010989e-05,
"loss": 0.4656,
"step": 4340
},
{
"epoch": 1.19,
"grad_norm": 0.9646287560462952,
"learning_rate": 8.862362637362638e-05,
"loss": 0.4707,
"step": 4350
},
{
"epoch": 1.19,
"grad_norm": 1.2284992933273315,
"learning_rate": 8.859615384615385e-05,
"loss": 0.4812,
"step": 4360
},
{
"epoch": 1.19,
"grad_norm": 0.7886459231376648,
"learning_rate": 8.856868131868132e-05,
"loss": 0.4641,
"step": 4370
},
{
"epoch": 1.2,
"grad_norm": 0.8464663028717041,
"learning_rate": 8.85412087912088e-05,
"loss": 0.4685,
"step": 4380
},
{
"epoch": 1.2,
"grad_norm": 1.0549769401550293,
"learning_rate": 8.851373626373627e-05,
"loss": 0.4573,
"step": 4390
},
{
"epoch": 1.2,
"grad_norm": 1.3452584743499756,
"learning_rate": 8.848626373626374e-05,
"loss": 0.4823,
"step": 4400
},
{
"epoch": 1.2,
"grad_norm": 1.5250259637832642,
"learning_rate": 8.845879120879122e-05,
"loss": 0.4762,
"step": 4410
},
{
"epoch": 1.21,
"grad_norm": 0.7843578457832336,
"learning_rate": 8.843131868131869e-05,
"loss": 0.4714,
"step": 4420
},
{
"epoch": 1.21,
"grad_norm": 1.0666967630386353,
"learning_rate": 8.840384615384616e-05,
"loss": 0.4743,
"step": 4430
},
{
"epoch": 1.21,
"grad_norm": 1.4348440170288086,
"learning_rate": 8.837637362637364e-05,
"loss": 0.4879,
"step": 4440
},
{
"epoch": 1.22,
"grad_norm": 1.036281704902649,
"learning_rate": 8.83489010989011e-05,
"loss": 0.469,
"step": 4450
},
{
"epoch": 1.22,
"grad_norm": 0.7014042139053345,
"learning_rate": 8.832142857142857e-05,
"loss": 0.4716,
"step": 4460
},
{
"epoch": 1.22,
"grad_norm": 1.593533992767334,
"learning_rate": 8.829395604395605e-05,
"loss": 0.4745,
"step": 4470
},
{
"epoch": 1.22,
"grad_norm": 0.4499157965183258,
"learning_rate": 8.826648351648352e-05,
"loss": 0.4764,
"step": 4480
},
{
"epoch": 1.23,
"grad_norm": 0.7330986857414246,
"learning_rate": 8.823901098901099e-05,
"loss": 0.4733,
"step": 4490
},
{
"epoch": 1.23,
"grad_norm": 0.7915944457054138,
"learning_rate": 8.821153846153847e-05,
"loss": 0.4605,
"step": 4500
},
{
"epoch": 1.23,
"grad_norm": 0.9147601127624512,
"learning_rate": 8.818406593406594e-05,
"loss": 0.4729,
"step": 4510
},
{
"epoch": 1.23,
"grad_norm": 1.1145246028900146,
"learning_rate": 8.815659340659341e-05,
"loss": 0.4757,
"step": 4520
},
{
"epoch": 1.24,
"grad_norm": 1.0525473356246948,
"learning_rate": 8.812912087912089e-05,
"loss": 0.4585,
"step": 4530
},
{
"epoch": 1.24,
"grad_norm": 0.7515787482261658,
"learning_rate": 8.810164835164836e-05,
"loss": 0.4557,
"step": 4540
},
{
"epoch": 1.24,
"grad_norm": 0.9737600088119507,
"learning_rate": 8.807417582417583e-05,
"loss": 0.4647,
"step": 4550
},
{
"epoch": 1.25,
"grad_norm": 1.1443486213684082,
"learning_rate": 8.804670329670331e-05,
"loss": 0.4685,
"step": 4560
},
{
"epoch": 1.25,
"grad_norm": 1.4745419025421143,
"learning_rate": 8.801923076923078e-05,
"loss": 0.4736,
"step": 4570
},
{
"epoch": 1.25,
"grad_norm": 1.1491049528121948,
"learning_rate": 8.799175824175824e-05,
"loss": 0.4604,
"step": 4580
},
{
"epoch": 1.25,
"grad_norm": 0.742324948310852,
"learning_rate": 8.796428571428573e-05,
"loss": 0.457,
"step": 4590
},
{
"epoch": 1.26,
"grad_norm": 1.1914622783660889,
"learning_rate": 8.79368131868132e-05,
"loss": 0.4638,
"step": 4600
},
{
"epoch": 1.26,
"grad_norm": 0.6555135846138,
"learning_rate": 8.790934065934066e-05,
"loss": 0.4806,
"step": 4610
},
{
"epoch": 1.26,
"grad_norm": 0.8627590537071228,
"learning_rate": 8.788186813186814e-05,
"loss": 0.4803,
"step": 4620
},
{
"epoch": 1.26,
"grad_norm": 1.0868836641311646,
"learning_rate": 8.785439560439561e-05,
"loss": 0.4598,
"step": 4630
},
{
"epoch": 1.27,
"grad_norm": 1.291483998298645,
"learning_rate": 8.782692307692308e-05,
"loss": 0.4707,
"step": 4640
},
{
"epoch": 1.27,
"grad_norm": 0.9766128659248352,
"learning_rate": 8.779945054945056e-05,
"loss": 0.4707,
"step": 4650
},
{
"epoch": 1.27,
"grad_norm": 1.0301162004470825,
"learning_rate": 8.777197802197803e-05,
"loss": 0.4821,
"step": 4660
},
{
"epoch": 1.28,
"grad_norm": 1.0474865436553955,
"learning_rate": 8.77445054945055e-05,
"loss": 0.4603,
"step": 4670
},
{
"epoch": 1.28,
"grad_norm": 1.1218842267990112,
"learning_rate": 8.771703296703298e-05,
"loss": 0.4827,
"step": 4680
},
{
"epoch": 1.28,
"grad_norm": 1.0587611198425293,
"learning_rate": 8.768956043956045e-05,
"loss": 0.4746,
"step": 4690
},
{
"epoch": 1.28,
"grad_norm": 0.8249474763870239,
"learning_rate": 8.766208791208792e-05,
"loss": 0.4648,
"step": 4700
},
{
"epoch": 1.29,
"grad_norm": 0.6225512027740479,
"learning_rate": 8.76346153846154e-05,
"loss": 0.4582,
"step": 4710
},
{
"epoch": 1.29,
"grad_norm": 1.2126622200012207,
"learning_rate": 8.760714285714287e-05,
"loss": 0.4617,
"step": 4720
},
{
"epoch": 1.29,
"grad_norm": 1.6850072145462036,
"learning_rate": 8.757967032967033e-05,
"loss": 0.4793,
"step": 4730
},
{
"epoch": 1.29,
"grad_norm": 0.5053130984306335,
"learning_rate": 8.755219780219782e-05,
"loss": 0.4648,
"step": 4740
},
{
"epoch": 1.3,
"grad_norm": 0.4924371838569641,
"learning_rate": 8.752472527472528e-05,
"loss": 0.4694,
"step": 4750
},
{
"epoch": 1.3,
"grad_norm": 0.7182402014732361,
"learning_rate": 8.749725274725275e-05,
"loss": 0.4611,
"step": 4760
},
{
"epoch": 1.3,
"grad_norm": 0.5569953322410583,
"learning_rate": 8.746978021978023e-05,
"loss": 0.4632,
"step": 4770
},
{
"epoch": 1.31,
"grad_norm": 1.0777251720428467,
"learning_rate": 8.74423076923077e-05,
"loss": 0.4735,
"step": 4780
},
{
"epoch": 1.31,
"grad_norm": 0.6310242414474487,
"learning_rate": 8.741483516483517e-05,
"loss": 0.4567,
"step": 4790
},
{
"epoch": 1.31,
"grad_norm": 0.94218510389328,
"learning_rate": 8.738736263736265e-05,
"loss": 0.4626,
"step": 4800
},
{
"epoch": 1.31,
"grad_norm": 0.4554530084133148,
"learning_rate": 8.735989010989012e-05,
"loss": 0.4681,
"step": 4810
},
{
"epoch": 1.32,
"grad_norm": 1.2381364107131958,
"learning_rate": 8.733241758241759e-05,
"loss": 0.467,
"step": 4820
},
{
"epoch": 1.32,
"grad_norm": 1.1469919681549072,
"learning_rate": 8.730494505494507e-05,
"loss": 0.4643,
"step": 4830
},
{
"epoch": 1.32,
"grad_norm": 0.7159097194671631,
"learning_rate": 8.727747252747254e-05,
"loss": 0.4731,
"step": 4840
},
{
"epoch": 1.32,
"grad_norm": 0.8695803880691528,
"learning_rate": 8.725e-05,
"loss": 0.4646,
"step": 4850
},
{
"epoch": 1.33,
"grad_norm": 0.8535722494125366,
"learning_rate": 8.722252747252749e-05,
"loss": 0.4612,
"step": 4860
},
{
"epoch": 1.33,
"grad_norm": 1.231641173362732,
"learning_rate": 8.719505494505496e-05,
"loss": 0.48,
"step": 4870
},
{
"epoch": 1.33,
"grad_norm": 1.0003236532211304,
"learning_rate": 8.716758241758241e-05,
"loss": 0.4741,
"step": 4880
},
{
"epoch": 1.34,
"grad_norm": 1.8520704507827759,
"learning_rate": 8.714010989010989e-05,
"loss": 0.4693,
"step": 4890
},
{
"epoch": 1.34,
"grad_norm": 1.0661410093307495,
"learning_rate": 8.711263736263736e-05,
"loss": 0.4848,
"step": 4900
},
{
"epoch": 1.34,
"grad_norm": 0.8280912637710571,
"learning_rate": 8.708516483516483e-05,
"loss": 0.4703,
"step": 4910
},
{
"epoch": 1.34,
"grad_norm": 1.169047474861145,
"learning_rate": 8.705769230769231e-05,
"loss": 0.4777,
"step": 4920
},
{
"epoch": 1.35,
"grad_norm": 0.7568297982215881,
"learning_rate": 8.703021978021978e-05,
"loss": 0.4752,
"step": 4930
},
{
"epoch": 1.35,
"grad_norm": 1.5538333654403687,
"learning_rate": 8.700274725274725e-05,
"loss": 0.4569,
"step": 4940
},
{
"epoch": 1.35,
"grad_norm": 0.6673538684844971,
"learning_rate": 8.697527472527473e-05,
"loss": 0.4659,
"step": 4950
},
{
"epoch": 1.35,
"grad_norm": 0.9370101690292358,
"learning_rate": 8.69478021978022e-05,
"loss": 0.4644,
"step": 4960
},
{
"epoch": 1.36,
"grad_norm": 1.748659610748291,
"learning_rate": 8.692032967032966e-05,
"loss": 0.458,
"step": 4970
},
{
"epoch": 1.36,
"grad_norm": 0.8026989102363586,
"learning_rate": 8.689285714285715e-05,
"loss": 0.4696,
"step": 4980
},
{
"epoch": 1.36,
"grad_norm": 1.7652183771133423,
"learning_rate": 8.686538461538461e-05,
"loss": 0.4676,
"step": 4990
},
{
"epoch": 1.37,
"grad_norm": 0.6286242008209229,
"learning_rate": 8.683791208791208e-05,
"loss": 0.4638,
"step": 5000
},
{
"epoch": 1.37,
"grad_norm": 2.5077078342437744,
"learning_rate": 8.681043956043956e-05,
"loss": 0.4774,
"step": 5010
},
{
"epoch": 1.37,
"grad_norm": 0.6874992251396179,
"learning_rate": 8.678296703296703e-05,
"loss": 0.4621,
"step": 5020
},
{
"epoch": 1.37,
"grad_norm": 1.2562603950500488,
"learning_rate": 8.67554945054945e-05,
"loss": 0.4745,
"step": 5030
},
{
"epoch": 1.38,
"grad_norm": 1.0717673301696777,
"learning_rate": 8.672802197802198e-05,
"loss": 0.4726,
"step": 5040
},
{
"epoch": 1.38,
"grad_norm": 0.744522750377655,
"learning_rate": 8.670054945054945e-05,
"loss": 0.4823,
"step": 5050
},
{
"epoch": 1.38,
"grad_norm": 0.879859447479248,
"learning_rate": 8.667307692307692e-05,
"loss": 0.4651,
"step": 5060
},
{
"epoch": 1.38,
"grad_norm": 0.6319553852081299,
"learning_rate": 8.66456043956044e-05,
"loss": 0.4714,
"step": 5070
},
{
"epoch": 1.39,
"grad_norm": 0.8513883948326111,
"learning_rate": 8.661813186813187e-05,
"loss": 0.4644,
"step": 5080
},
{
"epoch": 1.39,
"grad_norm": 1.7870076894760132,
"learning_rate": 8.659065934065934e-05,
"loss": 0.4714,
"step": 5090
},
{
"epoch": 1.39,
"grad_norm": 0.8447625041007996,
"learning_rate": 8.656318681318682e-05,
"loss": 0.4717,
"step": 5100
},
{
"epoch": 1.4,
"grad_norm": 0.7471101880073547,
"learning_rate": 8.653571428571429e-05,
"loss": 0.4594,
"step": 5110
},
{
"epoch": 1.4,
"grad_norm": 0.7438103556632996,
"learning_rate": 8.650824175824175e-05,
"loss": 0.4622,
"step": 5120
},
{
"epoch": 1.4,
"grad_norm": 0.8779565095901489,
"learning_rate": 8.648076923076924e-05,
"loss": 0.4639,
"step": 5130
},
{
"epoch": 1.4,
"grad_norm": 1.4103574752807617,
"learning_rate": 8.64532967032967e-05,
"loss": 0.4568,
"step": 5140
},
{
"epoch": 1.41,
"grad_norm": 1.2644096612930298,
"learning_rate": 8.642582417582417e-05,
"loss": 0.4696,
"step": 5150
},
{
"epoch": 1.41,
"grad_norm": 0.7385484576225281,
"learning_rate": 8.639835164835165e-05,
"loss": 0.4638,
"step": 5160
},
{
"epoch": 1.41,
"grad_norm": 0.7698835730552673,
"learning_rate": 8.637087912087912e-05,
"loss": 0.4762,
"step": 5170
},
{
"epoch": 1.42,
"grad_norm": 1.049194574356079,
"learning_rate": 8.634340659340659e-05,
"loss": 0.4588,
"step": 5180
},
{
"epoch": 1.42,
"grad_norm": 0.7376847267150879,
"learning_rate": 8.631593406593407e-05,
"loss": 0.46,
"step": 5190
},
{
"epoch": 1.42,
"grad_norm": 1.1413201093673706,
"learning_rate": 8.628846153846154e-05,
"loss": 0.4662,
"step": 5200
},
{
"epoch": 1.42,
"grad_norm": 0.9843841195106506,
"learning_rate": 8.626098901098901e-05,
"loss": 0.4663,
"step": 5210
},
{
"epoch": 1.43,
"grad_norm": 0.6186099052429199,
"learning_rate": 8.623351648351649e-05,
"loss": 0.4683,
"step": 5220
},
{
"epoch": 1.43,
"grad_norm": 1.4082993268966675,
"learning_rate": 8.620604395604396e-05,
"loss": 0.4639,
"step": 5230
},
{
"epoch": 1.43,
"grad_norm": 0.9625537395477295,
"learning_rate": 8.617857142857143e-05,
"loss": 0.4646,
"step": 5240
},
{
"epoch": 1.43,
"grad_norm": 1.0915727615356445,
"learning_rate": 8.615109890109891e-05,
"loss": 0.4669,
"step": 5250
},
{
"epoch": 1.44,
"grad_norm": 0.9645349383354187,
"learning_rate": 8.612362637362638e-05,
"loss": 0.4572,
"step": 5260
},
{
"epoch": 1.44,
"grad_norm": 0.7993157505989075,
"learning_rate": 8.609615384615384e-05,
"loss": 0.4623,
"step": 5270
},
{
"epoch": 1.44,
"grad_norm": 1.3415577411651611,
"learning_rate": 8.606868131868133e-05,
"loss": 0.4569,
"step": 5280
},
{
"epoch": 1.45,
"grad_norm": 1.1217197179794312,
"learning_rate": 8.60412087912088e-05,
"loss": 0.4632,
"step": 5290
},
{
"epoch": 1.45,
"grad_norm": 0.6021980047225952,
"learning_rate": 8.601373626373626e-05,
"loss": 0.4688,
"step": 5300
},
{
"epoch": 1.45,
"grad_norm": 0.9672883749008179,
"learning_rate": 8.598626373626374e-05,
"loss": 0.4594,
"step": 5310
},
{
"epoch": 1.45,
"grad_norm": 0.8534507155418396,
"learning_rate": 8.595879120879121e-05,
"loss": 0.4496,
"step": 5320
},
{
"epoch": 1.46,
"grad_norm": 1.6172966957092285,
"learning_rate": 8.593131868131868e-05,
"loss": 0.4751,
"step": 5330
},
{
"epoch": 1.46,
"grad_norm": 1.378061056137085,
"learning_rate": 8.590384615384616e-05,
"loss": 0.4597,
"step": 5340
},
{
"epoch": 1.46,
"grad_norm": 0.8239535093307495,
"learning_rate": 8.587637362637363e-05,
"loss": 0.4549,
"step": 5350
},
{
"epoch": 1.46,
"grad_norm": 0.6672975420951843,
"learning_rate": 8.58489010989011e-05,
"loss": 0.4694,
"step": 5360
},
{
"epoch": 1.47,
"grad_norm": 1.042051076889038,
"learning_rate": 8.582142857142858e-05,
"loss": 0.4667,
"step": 5370
},
{
"epoch": 1.47,
"grad_norm": 1.6398990154266357,
"learning_rate": 8.579395604395605e-05,
"loss": 0.4657,
"step": 5380
},
{
"epoch": 1.47,
"grad_norm": 0.5243840217590332,
"learning_rate": 8.576648351648352e-05,
"loss": 0.4733,
"step": 5390
},
{
"epoch": 1.48,
"grad_norm": 1.3996922969818115,
"learning_rate": 8.5739010989011e-05,
"loss": 0.4747,
"step": 5400
},
{
"epoch": 1.48,
"grad_norm": 0.7381965517997742,
"learning_rate": 8.571153846153847e-05,
"loss": 0.4694,
"step": 5410
},
{
"epoch": 1.48,
"grad_norm": 1.1151881217956543,
"learning_rate": 8.568406593406593e-05,
"loss": 0.4658,
"step": 5420
},
{
"epoch": 1.48,
"grad_norm": 1.0418914556503296,
"learning_rate": 8.565659340659342e-05,
"loss": 0.4568,
"step": 5430
},
{
"epoch": 1.49,
"grad_norm": 0.7290151119232178,
"learning_rate": 8.562912087912088e-05,
"loss": 0.4642,
"step": 5440
},
{
"epoch": 1.49,
"grad_norm": 0.8483603000640869,
"learning_rate": 8.560164835164835e-05,
"loss": 0.4912,
"step": 5450
},
{
"epoch": 1.49,
"grad_norm": 1.1545953750610352,
"learning_rate": 8.557417582417583e-05,
"loss": 0.4738,
"step": 5460
},
{
"epoch": 1.49,
"grad_norm": 1.0104776620864868,
"learning_rate": 8.55467032967033e-05,
"loss": 0.4751,
"step": 5470
},
{
"epoch": 1.5,
"grad_norm": 0.9386792778968811,
"learning_rate": 8.551923076923077e-05,
"loss": 0.4549,
"step": 5480
},
{
"epoch": 1.5,
"grad_norm": 0.6779844760894775,
"learning_rate": 8.549175824175825e-05,
"loss": 0.4682,
"step": 5490
},
{
"epoch": 1.5,
"grad_norm": 0.7957237958908081,
"learning_rate": 8.546428571428572e-05,
"loss": 0.4561,
"step": 5500
},
{
"epoch": 1.51,
"grad_norm": 0.8393499255180359,
"learning_rate": 8.543681318681319e-05,
"loss": 0.4659,
"step": 5510
},
{
"epoch": 1.51,
"grad_norm": 0.9744943380355835,
"learning_rate": 8.540934065934067e-05,
"loss": 0.4545,
"step": 5520
},
{
"epoch": 1.51,
"grad_norm": 1.1456732749938965,
"learning_rate": 8.538186813186814e-05,
"loss": 0.4614,
"step": 5530
},
{
"epoch": 1.51,
"grad_norm": 0.8066684603691101,
"learning_rate": 8.53543956043956e-05,
"loss": 0.475,
"step": 5540
},
{
"epoch": 1.52,
"grad_norm": 1.4929931163787842,
"learning_rate": 8.532692307692309e-05,
"loss": 0.4695,
"step": 5550
},
{
"epoch": 1.52,
"grad_norm": 0.7223564386367798,
"learning_rate": 8.529945054945056e-05,
"loss": 0.4692,
"step": 5560
},
{
"epoch": 1.52,
"grad_norm": 1.3222299814224243,
"learning_rate": 8.527197802197802e-05,
"loss": 0.4533,
"step": 5570
},
{
"epoch": 1.52,
"grad_norm": 1.1262595653533936,
"learning_rate": 8.52445054945055e-05,
"loss": 0.4637,
"step": 5580
},
{
"epoch": 1.53,
"grad_norm": 0.9097437858581543,
"learning_rate": 8.521703296703297e-05,
"loss": 0.4692,
"step": 5590
},
{
"epoch": 1.53,
"grad_norm": 1.5121077299118042,
"learning_rate": 8.518956043956044e-05,
"loss": 0.4591,
"step": 5600
},
{
"epoch": 1.53,
"grad_norm": 1.1432225704193115,
"learning_rate": 8.516208791208792e-05,
"loss": 0.4723,
"step": 5610
},
{
"epoch": 1.54,
"grad_norm": 1.3745416402816772,
"learning_rate": 8.513461538461539e-05,
"loss": 0.4725,
"step": 5620
},
{
"epoch": 1.54,
"grad_norm": 0.8508789539337158,
"learning_rate": 8.510714285714286e-05,
"loss": 0.4607,
"step": 5630
},
{
"epoch": 1.54,
"grad_norm": 1.3097485303878784,
"learning_rate": 8.507967032967034e-05,
"loss": 0.4635,
"step": 5640
},
{
"epoch": 1.54,
"grad_norm": 1.29444420337677,
"learning_rate": 8.505219780219781e-05,
"loss": 0.4628,
"step": 5650
},
{
"epoch": 1.55,
"grad_norm": 0.5336117744445801,
"learning_rate": 8.502472527472528e-05,
"loss": 0.4568,
"step": 5660
},
{
"epoch": 1.55,
"grad_norm": 0.7139732837677002,
"learning_rate": 8.499725274725276e-05,
"loss": 0.4574,
"step": 5670
},
{
"epoch": 1.55,
"grad_norm": 0.7436287999153137,
"learning_rate": 8.496978021978023e-05,
"loss": 0.455,
"step": 5680
},
{
"epoch": 1.55,
"grad_norm": 0.9436907172203064,
"learning_rate": 8.49423076923077e-05,
"loss": 0.4678,
"step": 5690
},
{
"epoch": 1.56,
"grad_norm": 0.9827747344970703,
"learning_rate": 8.491483516483518e-05,
"loss": 0.4614,
"step": 5700
},
{
"epoch": 1.56,
"grad_norm": 0.7332632541656494,
"learning_rate": 8.488736263736265e-05,
"loss": 0.4675,
"step": 5710
},
{
"epoch": 1.56,
"grad_norm": 1.1933000087738037,
"learning_rate": 8.485989010989011e-05,
"loss": 0.4528,
"step": 5720
},
{
"epoch": 1.57,
"grad_norm": 1.1396616697311401,
"learning_rate": 8.48324175824176e-05,
"loss": 0.4663,
"step": 5730
},
{
"epoch": 1.57,
"grad_norm": 1.4335391521453857,
"learning_rate": 8.480494505494506e-05,
"loss": 0.4658,
"step": 5740
},
{
"epoch": 1.57,
"grad_norm": 0.85484778881073,
"learning_rate": 8.477747252747253e-05,
"loss": 0.4536,
"step": 5750
},
{
"epoch": 1.57,
"grad_norm": 1.206037163734436,
"learning_rate": 8.475000000000001e-05,
"loss": 0.4665,
"step": 5760
},
{
"epoch": 1.58,
"grad_norm": 1.0052751302719116,
"learning_rate": 8.472252747252748e-05,
"loss": 0.4657,
"step": 5770
},
{
"epoch": 1.58,
"grad_norm": 0.6977716088294983,
"learning_rate": 8.469505494505495e-05,
"loss": 0.4594,
"step": 5780
},
{
"epoch": 1.58,
"grad_norm": 0.975903332233429,
"learning_rate": 8.466758241758243e-05,
"loss": 0.4739,
"step": 5790
},
{
"epoch": 1.58,
"grad_norm": 1.0114597082138062,
"learning_rate": 8.46401098901099e-05,
"loss": 0.4732,
"step": 5800
},
{
"epoch": 1.59,
"grad_norm": 0.5334266424179077,
"learning_rate": 8.461263736263737e-05,
"loss": 0.4585,
"step": 5810
},
{
"epoch": 1.59,
"grad_norm": 1.6815012693405151,
"learning_rate": 8.458516483516485e-05,
"loss": 0.4737,
"step": 5820
},
{
"epoch": 1.59,
"grad_norm": 1.1055243015289307,
"learning_rate": 8.455769230769232e-05,
"loss": 0.4608,
"step": 5830
},
{
"epoch": 1.6,
"grad_norm": 1.3854832649230957,
"learning_rate": 8.453021978021979e-05,
"loss": 0.4659,
"step": 5840
},
{
"epoch": 1.6,
"grad_norm": 1.0487163066864014,
"learning_rate": 8.450274725274727e-05,
"loss": 0.4519,
"step": 5850
},
{
"epoch": 1.6,
"grad_norm": 1.0301164388656616,
"learning_rate": 8.447527472527474e-05,
"loss": 0.4598,
"step": 5860
},
{
"epoch": 1.6,
"grad_norm": 0.7901486754417419,
"learning_rate": 8.44478021978022e-05,
"loss": 0.4727,
"step": 5870
},
{
"epoch": 1.61,
"grad_norm": 0.6557927131652832,
"learning_rate": 8.442032967032969e-05,
"loss": 0.4523,
"step": 5880
},
{
"epoch": 1.61,
"grad_norm": 0.8954731225967407,
"learning_rate": 8.439285714285715e-05,
"loss": 0.4672,
"step": 5890
},
{
"epoch": 1.61,
"grad_norm": 0.6038153171539307,
"learning_rate": 8.436538461538462e-05,
"loss": 0.4604,
"step": 5900
},
{
"epoch": 1.61,
"grad_norm": 0.9118047952651978,
"learning_rate": 8.433791208791209e-05,
"loss": 0.4592,
"step": 5910
},
{
"epoch": 1.62,
"grad_norm": 1.125707745552063,
"learning_rate": 8.431043956043957e-05,
"loss": 0.4713,
"step": 5920
},
{
"epoch": 1.62,
"grad_norm": 0.7669263482093811,
"learning_rate": 8.428296703296704e-05,
"loss": 0.465,
"step": 5930
},
{
"epoch": 1.62,
"grad_norm": 1.0012739896774292,
"learning_rate": 8.425549450549451e-05,
"loss": 0.4661,
"step": 5940
},
{
"epoch": 1.63,
"grad_norm": 0.650155782699585,
"learning_rate": 8.422802197802199e-05,
"loss": 0.4649,
"step": 5950
},
{
"epoch": 1.63,
"grad_norm": 1.224027156829834,
"learning_rate": 8.420054945054946e-05,
"loss": 0.4486,
"step": 5960
},
{
"epoch": 1.63,
"grad_norm": 1.350738763809204,
"learning_rate": 8.417307692307693e-05,
"loss": 0.4714,
"step": 5970
},
{
"epoch": 1.63,
"grad_norm": 1.238620400428772,
"learning_rate": 8.414560439560441e-05,
"loss": 0.4644,
"step": 5980
},
{
"epoch": 1.64,
"grad_norm": 1.4456909894943237,
"learning_rate": 8.411813186813188e-05,
"loss": 0.4674,
"step": 5990
},
{
"epoch": 1.64,
"grad_norm": 0.9819515943527222,
"learning_rate": 8.409065934065934e-05,
"loss": 0.4681,
"step": 6000
},
{
"epoch": 1.64,
"grad_norm": 0.7437455058097839,
"learning_rate": 8.406318681318683e-05,
"loss": 0.4701,
"step": 6010
},
{
"epoch": 1.64,
"grad_norm": 0.7628225088119507,
"learning_rate": 8.403571428571428e-05,
"loss": 0.4608,
"step": 6020
},
{
"epoch": 1.65,
"grad_norm": 1.0726768970489502,
"learning_rate": 8.400824175824176e-05,
"loss": 0.461,
"step": 6030
},
{
"epoch": 1.65,
"grad_norm": 0.799857497215271,
"learning_rate": 8.398076923076923e-05,
"loss": 0.4587,
"step": 6040
},
{
"epoch": 1.65,
"grad_norm": 1.242092251777649,
"learning_rate": 8.39532967032967e-05,
"loss": 0.4569,
"step": 6050
},
{
"epoch": 1.66,
"grad_norm": 0.971057116985321,
"learning_rate": 8.392582417582418e-05,
"loss": 0.4537,
"step": 6060
},
{
"epoch": 1.66,
"grad_norm": 1.1566013097763062,
"learning_rate": 8.389835164835165e-05,
"loss": 0.4638,
"step": 6070
},
{
"epoch": 1.66,
"grad_norm": 1.2802190780639648,
"learning_rate": 8.387087912087912e-05,
"loss": 0.4663,
"step": 6080
},
{
"epoch": 1.66,
"grad_norm": 1.0269036293029785,
"learning_rate": 8.38434065934066e-05,
"loss": 0.4628,
"step": 6090
},
{
"epoch": 1.67,
"grad_norm": 0.7012946605682373,
"learning_rate": 8.381593406593407e-05,
"loss": 0.4513,
"step": 6100
},
{
"epoch": 1.67,
"grad_norm": 0.5895891189575195,
"learning_rate": 8.378846153846153e-05,
"loss": 0.4584,
"step": 6110
},
{
"epoch": 1.67,
"grad_norm": 0.9521509408950806,
"learning_rate": 8.376098901098902e-05,
"loss": 0.4587,
"step": 6120
},
{
"epoch": 1.67,
"grad_norm": 1.409681797027588,
"learning_rate": 8.373351648351648e-05,
"loss": 0.4592,
"step": 6130
},
{
"epoch": 1.68,
"grad_norm": 1.0833555459976196,
"learning_rate": 8.370604395604395e-05,
"loss": 0.4562,
"step": 6140
},
{
"epoch": 1.68,
"grad_norm": 0.8994373679161072,
"learning_rate": 8.367857142857143e-05,
"loss": 0.4593,
"step": 6150
},
{
"epoch": 1.68,
"grad_norm": 1.2591134309768677,
"learning_rate": 8.36510989010989e-05,
"loss": 0.4632,
"step": 6160
},
{
"epoch": 1.69,
"grad_norm": 0.8163917064666748,
"learning_rate": 8.362362637362637e-05,
"loss": 0.4767,
"step": 6170
},
{
"epoch": 1.69,
"grad_norm": 0.6897978186607361,
"learning_rate": 8.359615384615385e-05,
"loss": 0.4569,
"step": 6180
},
{
"epoch": 1.69,
"grad_norm": 1.0527364015579224,
"learning_rate": 8.356868131868132e-05,
"loss": 0.4618,
"step": 6190
},
{
"epoch": 1.69,
"grad_norm": 1.4118075370788574,
"learning_rate": 8.354120879120879e-05,
"loss": 0.4542,
"step": 6200
},
{
"epoch": 1.7,
"grad_norm": 0.9567866325378418,
"learning_rate": 8.351373626373627e-05,
"loss": 0.4673,
"step": 6210
},
{
"epoch": 1.7,
"grad_norm": 1.0207067728042603,
"learning_rate": 8.348626373626374e-05,
"loss": 0.4538,
"step": 6220
},
{
"epoch": 1.7,
"grad_norm": 0.5897815227508545,
"learning_rate": 8.34587912087912e-05,
"loss": 0.4577,
"step": 6230
},
{
"epoch": 1.7,
"grad_norm": 0.5948817133903503,
"learning_rate": 8.343131868131869e-05,
"loss": 0.4629,
"step": 6240
},
{
"epoch": 1.71,
"grad_norm": 1.6685130596160889,
"learning_rate": 8.340384615384616e-05,
"loss": 0.461,
"step": 6250
},
{
"epoch": 1.71,
"grad_norm": 1.502945899963379,
"learning_rate": 8.337637362637362e-05,
"loss": 0.4649,
"step": 6260
},
{
"epoch": 1.71,
"grad_norm": 1.5422593355178833,
"learning_rate": 8.33489010989011e-05,
"loss": 0.4494,
"step": 6270
},
{
"epoch": 1.72,
"grad_norm": 0.999934732913971,
"learning_rate": 8.332142857142857e-05,
"loss": 0.464,
"step": 6280
},
{
"epoch": 1.72,
"grad_norm": 0.6580365896224976,
"learning_rate": 8.329395604395604e-05,
"loss": 0.4382,
"step": 6290
},
{
"epoch": 1.72,
"grad_norm": 0.9826914668083191,
"learning_rate": 8.326648351648352e-05,
"loss": 0.4646,
"step": 6300
},
{
"epoch": 1.72,
"grad_norm": 1.6006591320037842,
"learning_rate": 8.323901098901099e-05,
"loss": 0.4709,
"step": 6310
},
{
"epoch": 1.73,
"grad_norm": 0.46454253792762756,
"learning_rate": 8.321153846153846e-05,
"loss": 0.4572,
"step": 6320
},
{
"epoch": 1.73,
"grad_norm": 1.1660053730010986,
"learning_rate": 8.318406593406594e-05,
"loss": 0.4537,
"step": 6330
},
{
"epoch": 1.73,
"grad_norm": 0.7951260805130005,
"learning_rate": 8.315659340659341e-05,
"loss": 0.4491,
"step": 6340
},
{
"epoch": 1.73,
"grad_norm": 0.9338926672935486,
"learning_rate": 8.312912087912088e-05,
"loss": 0.4608,
"step": 6350
},
{
"epoch": 1.74,
"grad_norm": 1.2365031242370605,
"learning_rate": 8.310164835164836e-05,
"loss": 0.4582,
"step": 6360
},
{
"epoch": 1.74,
"grad_norm": 0.7614923119544983,
"learning_rate": 8.307417582417583e-05,
"loss": 0.4665,
"step": 6370
},
{
"epoch": 1.74,
"grad_norm": 0.9391283392906189,
"learning_rate": 8.30467032967033e-05,
"loss": 0.4593,
"step": 6380
},
{
"epoch": 1.75,
"grad_norm": 1.4633766412734985,
"learning_rate": 8.301923076923078e-05,
"loss": 0.4611,
"step": 6390
},
{
"epoch": 1.75,
"grad_norm": 1.159623146057129,
"learning_rate": 8.299175824175825e-05,
"loss": 0.4622,
"step": 6400
},
{
"epoch": 1.75,
"grad_norm": 1.0221785306930542,
"learning_rate": 8.296428571428571e-05,
"loss": 0.4483,
"step": 6410
},
{
"epoch": 1.75,
"grad_norm": 1.5070372819900513,
"learning_rate": 8.29368131868132e-05,
"loss": 0.4677,
"step": 6420
},
{
"epoch": 1.76,
"grad_norm": 0.9225922226905823,
"learning_rate": 8.290934065934066e-05,
"loss": 0.4711,
"step": 6430
},
{
"epoch": 1.76,
"grad_norm": 0.9344289302825928,
"learning_rate": 8.288186813186813e-05,
"loss": 0.4591,
"step": 6440
},
{
"epoch": 1.76,
"grad_norm": 0.5448782444000244,
"learning_rate": 8.285439560439561e-05,
"loss": 0.4635,
"step": 6450
},
{
"epoch": 1.76,
"grad_norm": 0.465524822473526,
"learning_rate": 8.282692307692308e-05,
"loss": 0.4581,
"step": 6460
},
{
"epoch": 1.77,
"grad_norm": 1.3583608865737915,
"learning_rate": 8.279945054945055e-05,
"loss": 0.4718,
"step": 6470
},
{
"epoch": 1.77,
"grad_norm": 1.3173762559890747,
"learning_rate": 8.277197802197803e-05,
"loss": 0.4577,
"step": 6480
},
{
"epoch": 1.77,
"grad_norm": 1.3845728635787964,
"learning_rate": 8.27445054945055e-05,
"loss": 0.4623,
"step": 6490
},
{
"epoch": 1.78,
"grad_norm": 1.0992422103881836,
"learning_rate": 8.271703296703297e-05,
"loss": 0.4543,
"step": 6500
},
{
"epoch": 1.78,
"grad_norm": 0.5556319355964661,
"learning_rate": 8.268956043956045e-05,
"loss": 0.4613,
"step": 6510
},
{
"epoch": 1.78,
"grad_norm": 0.7095728516578674,
"learning_rate": 8.266208791208792e-05,
"loss": 0.455,
"step": 6520
},
{
"epoch": 1.78,
"grad_norm": 1.1960381269454956,
"learning_rate": 8.263461538461539e-05,
"loss": 0.4545,
"step": 6530
},
{
"epoch": 1.79,
"grad_norm": 1.0696423053741455,
"learning_rate": 8.260714285714287e-05,
"loss": 0.4666,
"step": 6540
},
{
"epoch": 1.79,
"grad_norm": 0.5211924314498901,
"learning_rate": 8.257967032967034e-05,
"loss": 0.4535,
"step": 6550
},
{
"epoch": 1.79,
"grad_norm": 1.1284767389297485,
"learning_rate": 8.25521978021978e-05,
"loss": 0.4633,
"step": 6560
},
{
"epoch": 1.79,
"grad_norm": 1.3033208847045898,
"learning_rate": 8.252472527472529e-05,
"loss": 0.4812,
"step": 6570
},
{
"epoch": 1.8,
"grad_norm": 1.2627853155136108,
"learning_rate": 8.249725274725275e-05,
"loss": 0.4676,
"step": 6580
},
{
"epoch": 1.8,
"grad_norm": 0.5073243379592896,
"learning_rate": 8.246978021978022e-05,
"loss": 0.4638,
"step": 6590
},
{
"epoch": 1.8,
"grad_norm": 0.6492868661880493,
"learning_rate": 8.24423076923077e-05,
"loss": 0.46,
"step": 6600
},
{
"epoch": 1.81,
"grad_norm": 0.8350780010223389,
"learning_rate": 8.241483516483517e-05,
"loss": 0.4549,
"step": 6610
},
{
"epoch": 1.81,
"grad_norm": 1.3048648834228516,
"learning_rate": 8.238736263736264e-05,
"loss": 0.47,
"step": 6620
},
{
"epoch": 1.81,
"grad_norm": 1.2212404012680054,
"learning_rate": 8.235989010989012e-05,
"loss": 0.4623,
"step": 6630
},
{
"epoch": 1.81,
"grad_norm": 0.9617910981178284,
"learning_rate": 8.233241758241759e-05,
"loss": 0.4566,
"step": 6640
},
{
"epoch": 1.82,
"grad_norm": 0.5445905923843384,
"learning_rate": 8.230494505494506e-05,
"loss": 0.458,
"step": 6650
},
{
"epoch": 1.82,
"grad_norm": 0.6512438654899597,
"learning_rate": 8.227747252747254e-05,
"loss": 0.4516,
"step": 6660
},
{
"epoch": 1.82,
"grad_norm": 1.2203713655471802,
"learning_rate": 8.225000000000001e-05,
"loss": 0.4566,
"step": 6670
},
{
"epoch": 1.82,
"grad_norm": 0.5803031325340271,
"learning_rate": 8.222252747252748e-05,
"loss": 0.4622,
"step": 6680
},
{
"epoch": 1.83,
"grad_norm": 0.5457717180252075,
"learning_rate": 8.219505494505496e-05,
"loss": 0.4509,
"step": 6690
},
{
"epoch": 1.83,
"grad_norm": 0.4855049252510071,
"learning_rate": 8.216758241758242e-05,
"loss": 0.4517,
"step": 6700
},
{
"epoch": 1.83,
"grad_norm": 0.49980971217155457,
"learning_rate": 8.214010989010989e-05,
"loss": 0.4592,
"step": 6710
},
{
"epoch": 1.84,
"grad_norm": 0.8323667049407959,
"learning_rate": 8.211263736263736e-05,
"loss": 0.4515,
"step": 6720
},
{
"epoch": 1.84,
"grad_norm": 0.6058827638626099,
"learning_rate": 8.208516483516484e-05,
"loss": 0.4437,
"step": 6730
},
{
"epoch": 1.84,
"grad_norm": 1.3227874040603638,
"learning_rate": 8.205769230769231e-05,
"loss": 0.4497,
"step": 6740
},
{
"epoch": 1.84,
"grad_norm": 1.14075767993927,
"learning_rate": 8.203021978021978e-05,
"loss": 0.455,
"step": 6750
},
{
"epoch": 1.85,
"grad_norm": 1.0633773803710938,
"learning_rate": 8.200274725274726e-05,
"loss": 0.4582,
"step": 6760
},
{
"epoch": 1.85,
"grad_norm": 0.671051025390625,
"learning_rate": 8.197527472527473e-05,
"loss": 0.4584,
"step": 6770
},
{
"epoch": 1.85,
"grad_norm": 0.8227337002754211,
"learning_rate": 8.19478021978022e-05,
"loss": 0.4564,
"step": 6780
},
{
"epoch": 1.85,
"grad_norm": 1.3290925025939941,
"learning_rate": 8.192032967032968e-05,
"loss": 0.4527,
"step": 6790
},
{
"epoch": 1.86,
"grad_norm": 1.160757303237915,
"learning_rate": 8.189285714285715e-05,
"loss": 0.4417,
"step": 6800
},
{
"epoch": 1.86,
"grad_norm": 0.8233458399772644,
"learning_rate": 8.186538461538462e-05,
"loss": 0.4459,
"step": 6810
},
{
"epoch": 1.86,
"grad_norm": 0.7575914859771729,
"learning_rate": 8.18379120879121e-05,
"loss": 0.4636,
"step": 6820
},
{
"epoch": 1.87,
"grad_norm": 0.35609036684036255,
"learning_rate": 8.181043956043956e-05,
"loss": 0.4479,
"step": 6830
},
{
"epoch": 1.87,
"grad_norm": 1.0687757730484009,
"learning_rate": 8.178296703296703e-05,
"loss": 0.4494,
"step": 6840
},
{
"epoch": 1.87,
"grad_norm": 0.936469316482544,
"learning_rate": 8.175549450549451e-05,
"loss": 0.4605,
"step": 6850
},
{
"epoch": 1.87,
"grad_norm": 0.6257254481315613,
"learning_rate": 8.172802197802198e-05,
"loss": 0.4559,
"step": 6860
},
{
"epoch": 1.88,
"grad_norm": 1.0161064863204956,
"learning_rate": 8.170054945054945e-05,
"loss": 0.4677,
"step": 6870
},
{
"epoch": 1.88,
"grad_norm": 1.0407649278640747,
"learning_rate": 8.167307692307693e-05,
"loss": 0.4542,
"step": 6880
},
{
"epoch": 1.88,
"grad_norm": 0.7439064383506775,
"learning_rate": 8.16456043956044e-05,
"loss": 0.4747,
"step": 6890
},
{
"epoch": 1.88,
"grad_norm": 0.37432214617729187,
"learning_rate": 8.161813186813187e-05,
"loss": 0.4595,
"step": 6900
},
{
"epoch": 1.89,
"grad_norm": 0.9039118885993958,
"learning_rate": 8.159065934065935e-05,
"loss": 0.4701,
"step": 6910
},
{
"epoch": 1.89,
"grad_norm": 0.5294177532196045,
"learning_rate": 8.156318681318682e-05,
"loss": 0.4533,
"step": 6920
},
{
"epoch": 1.89,
"grad_norm": 0.8023412227630615,
"learning_rate": 8.153571428571429e-05,
"loss": 0.4632,
"step": 6930
},
{
"epoch": 1.9,
"grad_norm": 1.1089890003204346,
"learning_rate": 8.150824175824177e-05,
"loss": 0.4627,
"step": 6940
},
{
"epoch": 1.9,
"grad_norm": 0.5587044954299927,
"learning_rate": 8.148076923076924e-05,
"loss": 0.4613,
"step": 6950
},
{
"epoch": 1.9,
"grad_norm": 0.7494495511054993,
"learning_rate": 8.14532967032967e-05,
"loss": 0.4478,
"step": 6960
},
{
"epoch": 1.9,
"grad_norm": 0.7425660490989685,
"learning_rate": 8.142582417582419e-05,
"loss": 0.4585,
"step": 6970
},
{
"epoch": 1.91,
"grad_norm": 0.9916179180145264,
"learning_rate": 8.139835164835165e-05,
"loss": 0.464,
"step": 6980
},
{
"epoch": 1.91,
"grad_norm": 0.9130895137786865,
"learning_rate": 8.137087912087912e-05,
"loss": 0.4628,
"step": 6990
},
{
"epoch": 1.91,
"grad_norm": 0.7386452555656433,
"learning_rate": 8.13434065934066e-05,
"loss": 0.457,
"step": 7000
},
{
"epoch": 1.91,
"grad_norm": 1.5726503133773804,
"learning_rate": 8.131593406593407e-05,
"loss": 0.461,
"step": 7010
},
{
"epoch": 1.92,
"grad_norm": 0.6793839335441589,
"learning_rate": 8.128846153846154e-05,
"loss": 0.4521,
"step": 7020
},
{
"epoch": 1.92,
"grad_norm": 1.5302358865737915,
"learning_rate": 8.126098901098902e-05,
"loss": 0.4598,
"step": 7030
},
{
"epoch": 1.92,
"grad_norm": 0.9924741387367249,
"learning_rate": 8.123351648351649e-05,
"loss": 0.4599,
"step": 7040
},
{
"epoch": 1.93,
"grad_norm": 0.7319537997245789,
"learning_rate": 8.120604395604396e-05,
"loss": 0.4585,
"step": 7050
},
{
"epoch": 1.93,
"grad_norm": 1.067428469657898,
"learning_rate": 8.117857142857144e-05,
"loss": 0.4509,
"step": 7060
},
{
"epoch": 1.93,
"grad_norm": 1.2225853204727173,
"learning_rate": 8.115109890109891e-05,
"loss": 0.4591,
"step": 7070
},
{
"epoch": 1.93,
"grad_norm": 0.7233381867408752,
"learning_rate": 8.112362637362638e-05,
"loss": 0.4568,
"step": 7080
},
{
"epoch": 1.94,
"grad_norm": 1.3318099975585938,
"learning_rate": 8.109615384615386e-05,
"loss": 0.4715,
"step": 7090
},
{
"epoch": 1.94,
"grad_norm": 1.4362515211105347,
"learning_rate": 8.106868131868133e-05,
"loss": 0.4539,
"step": 7100
},
{
"epoch": 1.94,
"grad_norm": 0.8498636484146118,
"learning_rate": 8.10412087912088e-05,
"loss": 0.4531,
"step": 7110
},
{
"epoch": 1.94,
"grad_norm": 0.5961604118347168,
"learning_rate": 8.101373626373628e-05,
"loss": 0.4561,
"step": 7120
},
{
"epoch": 1.95,
"grad_norm": 1.013180136680603,
"learning_rate": 8.098626373626374e-05,
"loss": 0.4581,
"step": 7130
},
{
"epoch": 1.95,
"grad_norm": 0.9159004092216492,
"learning_rate": 8.095879120879121e-05,
"loss": 0.4498,
"step": 7140
},
{
"epoch": 1.95,
"grad_norm": 0.9847846031188965,
"learning_rate": 8.093131868131868e-05,
"loss": 0.4584,
"step": 7150
},
{
"epoch": 1.96,
"grad_norm": 1.5007630586624146,
"learning_rate": 8.090384615384615e-05,
"loss": 0.4524,
"step": 7160
},
{
"epoch": 1.96,
"grad_norm": 0.8937559723854065,
"learning_rate": 8.087637362637363e-05,
"loss": 0.4701,
"step": 7170
},
{
"epoch": 1.96,
"grad_norm": 0.9551745653152466,
"learning_rate": 8.08489010989011e-05,
"loss": 0.4412,
"step": 7180
},
{
"epoch": 1.96,
"grad_norm": 0.7207891345024109,
"learning_rate": 8.082142857142857e-05,
"loss": 0.4577,
"step": 7190
},
{
"epoch": 1.97,
"grad_norm": 0.5330765843391418,
"learning_rate": 8.079395604395605e-05,
"loss": 0.4531,
"step": 7200
},
{
"epoch": 1.97,
"grad_norm": 1.0517854690551758,
"learning_rate": 8.076648351648352e-05,
"loss": 0.4339,
"step": 7210
},
{
"epoch": 1.97,
"grad_norm": 1.2942813634872437,
"learning_rate": 8.073901098901098e-05,
"loss": 0.4575,
"step": 7220
},
{
"epoch": 1.98,
"grad_norm": 0.9348652362823486,
"learning_rate": 8.071153846153847e-05,
"loss": 0.4472,
"step": 7230
},
{
"epoch": 1.98,
"grad_norm": 1.095439076423645,
"learning_rate": 8.068406593406593e-05,
"loss": 0.4503,
"step": 7240
},
{
"epoch": 1.98,
"grad_norm": 0.5965198278427124,
"learning_rate": 8.06565934065934e-05,
"loss": 0.4499,
"step": 7250
},
{
"epoch": 1.98,
"grad_norm": 0.5533403158187866,
"learning_rate": 8.062912087912088e-05,
"loss": 0.4676,
"step": 7260
},
{
"epoch": 1.99,
"grad_norm": 1.1052013635635376,
"learning_rate": 8.060164835164835e-05,
"loss": 0.4635,
"step": 7270
},
{
"epoch": 1.99,
"grad_norm": 0.9924907088279724,
"learning_rate": 8.057417582417582e-05,
"loss": 0.4606,
"step": 7280
},
{
"epoch": 1.99,
"grad_norm": 0.5882344245910645,
"learning_rate": 8.05467032967033e-05,
"loss": 0.4456,
"step": 7290
},
{
"epoch": 1.99,
"grad_norm": 1.0022166967391968,
"learning_rate": 8.051923076923077e-05,
"loss": 0.4505,
"step": 7300
},
{
"epoch": 2.0,
"grad_norm": 0.8352431058883667,
"learning_rate": 8.049175824175824e-05,
"loss": 0.4567,
"step": 7310
},
{
"epoch": 2.0,
"grad_norm": 1.0245552062988281,
"learning_rate": 8.046428571428572e-05,
"loss": 0.4509,
"step": 7320
},
{
"epoch": 2.0,
"eval_loss": 0.44436022639274597,
"eval_runtime": 1246.3805,
"eval_samples_per_second": 234.366,
"eval_steps_per_second": 3.663,
"step": 7321
},
{
"epoch": 2.0,
"grad_norm": 0.6890380382537842,
"learning_rate": 8.043681318681319e-05,
"loss": 0.4559,
"step": 7330
},
{
"epoch": 2.01,
"grad_norm": 0.8071466088294983,
"learning_rate": 8.040934065934066e-05,
"loss": 0.4494,
"step": 7340
},
{
"epoch": 2.01,
"grad_norm": 0.5615882873535156,
"learning_rate": 8.038186813186814e-05,
"loss": 0.444,
"step": 7350
},
{
"epoch": 2.01,
"grad_norm": 1.312966227531433,
"learning_rate": 8.03543956043956e-05,
"loss": 0.4549,
"step": 7360
},
{
"epoch": 2.01,
"grad_norm": 0.7778192162513733,
"learning_rate": 8.032692307692307e-05,
"loss": 0.4561,
"step": 7370
},
{
"epoch": 2.02,
"grad_norm": 0.9343681335449219,
"learning_rate": 8.029945054945056e-05,
"loss": 0.4565,
"step": 7380
},
{
"epoch": 2.02,
"grad_norm": 0.585110068321228,
"learning_rate": 8.027197802197802e-05,
"loss": 0.4502,
"step": 7390
},
{
"epoch": 2.02,
"grad_norm": 1.0197051763534546,
"learning_rate": 8.024450549450549e-05,
"loss": 0.4595,
"step": 7400
},
{
"epoch": 2.02,
"grad_norm": 1.0265450477600098,
"learning_rate": 8.021703296703297e-05,
"loss": 0.4425,
"step": 7410
},
{
"epoch": 2.03,
"grad_norm": 1.1836851835250854,
"learning_rate": 8.018956043956044e-05,
"loss": 0.4606,
"step": 7420
},
{
"epoch": 2.03,
"grad_norm": 0.9758961796760559,
"learning_rate": 8.016208791208791e-05,
"loss": 0.4494,
"step": 7430
},
{
"epoch": 2.03,
"grad_norm": 1.0412979125976562,
"learning_rate": 8.013461538461539e-05,
"loss": 0.4593,
"step": 7440
},
{
"epoch": 2.04,
"grad_norm": 1.13507878780365,
"learning_rate": 8.010714285714286e-05,
"loss": 0.4598,
"step": 7450
},
{
"epoch": 2.04,
"grad_norm": 1.118143081665039,
"learning_rate": 8.007967032967033e-05,
"loss": 0.4597,
"step": 7460
},
{
"epoch": 2.04,
"grad_norm": 0.6100110411643982,
"learning_rate": 8.005219780219781e-05,
"loss": 0.46,
"step": 7470
},
{
"epoch": 2.04,
"grad_norm": 1.4135487079620361,
"learning_rate": 8.002472527472528e-05,
"loss": 0.455,
"step": 7480
},
{
"epoch": 2.05,
"grad_norm": 0.7120094299316406,
"learning_rate": 7.999725274725275e-05,
"loss": 0.4459,
"step": 7490
},
{
"epoch": 2.05,
"grad_norm": 0.7432197332382202,
"learning_rate": 7.996978021978023e-05,
"loss": 0.4509,
"step": 7500
},
{
"epoch": 2.05,
"grad_norm": 1.1492975950241089,
"learning_rate": 7.99423076923077e-05,
"loss": 0.4638,
"step": 7510
},
{
"epoch": 2.05,
"grad_norm": 1.0090208053588867,
"learning_rate": 7.991483516483516e-05,
"loss": 0.4702,
"step": 7520
},
{
"epoch": 2.06,
"grad_norm": 1.09323251247406,
"learning_rate": 7.988736263736263e-05,
"loss": 0.4518,
"step": 7530
},
{
"epoch": 2.06,
"grad_norm": 1.0353306531906128,
"learning_rate": 7.985989010989011e-05,
"loss": 0.4522,
"step": 7540
},
{
"epoch": 2.06,
"grad_norm": 0.8499657511711121,
"learning_rate": 7.983241758241758e-05,
"loss": 0.4637,
"step": 7550
},
{
"epoch": 2.07,
"grad_norm": 1.0327955484390259,
"learning_rate": 7.980494505494505e-05,
"loss": 0.4444,
"step": 7560
},
{
"epoch": 2.07,
"grad_norm": 1.239608883857727,
"learning_rate": 7.977747252747253e-05,
"loss": 0.4536,
"step": 7570
},
{
"epoch": 2.07,
"grad_norm": 0.6841428279876709,
"learning_rate": 7.975e-05,
"loss": 0.4519,
"step": 7580
},
{
"epoch": 2.07,
"grad_norm": 1.1058799028396606,
"learning_rate": 7.972252747252747e-05,
"loss": 0.4511,
"step": 7590
},
{
"epoch": 2.08,
"grad_norm": 1.1671322584152222,
"learning_rate": 7.969505494505495e-05,
"loss": 0.4512,
"step": 7600
},
{
"epoch": 2.08,
"grad_norm": 0.6706550121307373,
"learning_rate": 7.966758241758242e-05,
"loss": 0.4493,
"step": 7610
},
{
"epoch": 2.08,
"grad_norm": 0.752795398235321,
"learning_rate": 7.964010989010989e-05,
"loss": 0.4494,
"step": 7620
},
{
"epoch": 2.08,
"grad_norm": 0.6255919933319092,
"learning_rate": 7.961263736263737e-05,
"loss": 0.4525,
"step": 7630
},
{
"epoch": 2.09,
"grad_norm": 0.4543440341949463,
"learning_rate": 7.958516483516484e-05,
"loss": 0.4527,
"step": 7640
},
{
"epoch": 2.09,
"grad_norm": 0.6038147807121277,
"learning_rate": 7.95576923076923e-05,
"loss": 0.4431,
"step": 7650
},
{
"epoch": 2.09,
"grad_norm": 1.010972261428833,
"learning_rate": 7.953021978021979e-05,
"loss": 0.451,
"step": 7660
},
{
"epoch": 2.1,
"grad_norm": 1.0627601146697998,
"learning_rate": 7.950274725274725e-05,
"loss": 0.4441,
"step": 7670
},
{
"epoch": 2.1,
"grad_norm": 0.6361072063446045,
"learning_rate": 7.947527472527472e-05,
"loss": 0.4592,
"step": 7680
},
{
"epoch": 2.1,
"grad_norm": 0.897415280342102,
"learning_rate": 7.94478021978022e-05,
"loss": 0.4535,
"step": 7690
},
{
"epoch": 2.1,
"grad_norm": 0.6545829772949219,
"learning_rate": 7.942032967032967e-05,
"loss": 0.4589,
"step": 7700
},
{
"epoch": 2.11,
"grad_norm": 0.9311442375183105,
"learning_rate": 7.939285714285714e-05,
"loss": 0.4417,
"step": 7710
},
{
"epoch": 2.11,
"grad_norm": 1.5107172727584839,
"learning_rate": 7.936538461538462e-05,
"loss": 0.4503,
"step": 7720
},
{
"epoch": 2.11,
"grad_norm": 0.9508922100067139,
"learning_rate": 7.933791208791209e-05,
"loss": 0.4532,
"step": 7730
},
{
"epoch": 2.11,
"grad_norm": 0.8631314039230347,
"learning_rate": 7.931043956043956e-05,
"loss": 0.4551,
"step": 7740
},
{
"epoch": 2.12,
"grad_norm": 0.639757513999939,
"learning_rate": 7.928296703296704e-05,
"loss": 0.4531,
"step": 7750
},
{
"epoch": 2.12,
"grad_norm": 0.8901143074035645,
"learning_rate": 7.925549450549451e-05,
"loss": 0.4563,
"step": 7760
},
{
"epoch": 2.12,
"grad_norm": 1.7677491903305054,
"learning_rate": 7.922802197802198e-05,
"loss": 0.4467,
"step": 7770
},
{
"epoch": 2.13,
"grad_norm": 0.9918643236160278,
"learning_rate": 7.920054945054946e-05,
"loss": 0.4604,
"step": 7780
},
{
"epoch": 2.13,
"grad_norm": 0.7385686635971069,
"learning_rate": 7.917307692307693e-05,
"loss": 0.4484,
"step": 7790
},
{
"epoch": 2.13,
"grad_norm": 0.4740804433822632,
"learning_rate": 7.91456043956044e-05,
"loss": 0.4581,
"step": 7800
},
{
"epoch": 2.13,
"grad_norm": 0.3722800016403198,
"learning_rate": 7.911813186813188e-05,
"loss": 0.4437,
"step": 7810
},
{
"epoch": 2.14,
"grad_norm": 0.5825007557868958,
"learning_rate": 7.909065934065934e-05,
"loss": 0.4495,
"step": 7820
},
{
"epoch": 2.14,
"grad_norm": 0.7577399015426636,
"learning_rate": 7.906318681318681e-05,
"loss": 0.4507,
"step": 7830
},
{
"epoch": 2.14,
"grad_norm": 0.5723931193351746,
"learning_rate": 7.90357142857143e-05,
"loss": 0.4512,
"step": 7840
},
{
"epoch": 2.14,
"grad_norm": 1.0552576780319214,
"learning_rate": 7.900824175824176e-05,
"loss": 0.4493,
"step": 7850
},
{
"epoch": 2.15,
"grad_norm": 1.1965206861495972,
"learning_rate": 7.898076923076923e-05,
"loss": 0.4502,
"step": 7860
},
{
"epoch": 2.15,
"grad_norm": 0.7538771629333496,
"learning_rate": 7.895329670329671e-05,
"loss": 0.4576,
"step": 7870
},
{
"epoch": 2.15,
"grad_norm": 1.2679303884506226,
"learning_rate": 7.892582417582418e-05,
"loss": 0.4464,
"step": 7880
},
{
"epoch": 2.16,
"grad_norm": 0.8973550200462341,
"learning_rate": 7.889835164835165e-05,
"loss": 0.4596,
"step": 7890
},
{
"epoch": 2.16,
"grad_norm": 0.7081739902496338,
"learning_rate": 7.887087912087913e-05,
"loss": 0.4368,
"step": 7900
},
{
"epoch": 2.16,
"grad_norm": 0.7710354328155518,
"learning_rate": 7.88434065934066e-05,
"loss": 0.4463,
"step": 7910
},
{
"epoch": 2.16,
"grad_norm": 0.6749390363693237,
"learning_rate": 7.881593406593407e-05,
"loss": 0.4515,
"step": 7920
},
{
"epoch": 2.17,
"grad_norm": 0.6680644154548645,
"learning_rate": 7.878846153846155e-05,
"loss": 0.4433,
"step": 7930
},
{
"epoch": 2.17,
"grad_norm": 1.5634280443191528,
"learning_rate": 7.876098901098902e-05,
"loss": 0.4702,
"step": 7940
},
{
"epoch": 2.17,
"grad_norm": 0.6959145665168762,
"learning_rate": 7.873351648351648e-05,
"loss": 0.4504,
"step": 7950
},
{
"epoch": 2.17,
"grad_norm": 0.7739858627319336,
"learning_rate": 7.870604395604397e-05,
"loss": 0.4606,
"step": 7960
},
{
"epoch": 2.18,
"grad_norm": 0.5519704818725586,
"learning_rate": 7.867857142857143e-05,
"loss": 0.4658,
"step": 7970
},
{
"epoch": 2.18,
"grad_norm": 0.5425352454185486,
"learning_rate": 7.86510989010989e-05,
"loss": 0.4445,
"step": 7980
},
{
"epoch": 2.18,
"grad_norm": 0.8816283941268921,
"learning_rate": 7.862362637362638e-05,
"loss": 0.4621,
"step": 7990
},
{
"epoch": 2.19,
"grad_norm": 1.0816186666488647,
"learning_rate": 7.859615384615385e-05,
"loss": 0.4622,
"step": 8000
},
{
"epoch": 2.19,
"grad_norm": 0.8483924865722656,
"learning_rate": 7.856868131868132e-05,
"loss": 0.4533,
"step": 8010
},
{
"epoch": 2.19,
"grad_norm": 0.45229572057724,
"learning_rate": 7.85412087912088e-05,
"loss": 0.4626,
"step": 8020
},
{
"epoch": 2.19,
"grad_norm": 0.6593546867370605,
"learning_rate": 7.851373626373627e-05,
"loss": 0.4498,
"step": 8030
},
{
"epoch": 2.2,
"grad_norm": 1.214553952217102,
"learning_rate": 7.848626373626374e-05,
"loss": 0.4497,
"step": 8040
},
{
"epoch": 2.2,
"grad_norm": 0.7151647210121155,
"learning_rate": 7.845879120879122e-05,
"loss": 0.4447,
"step": 8050
},
{
"epoch": 2.2,
"grad_norm": 0.7126444578170776,
"learning_rate": 7.843131868131869e-05,
"loss": 0.4622,
"step": 8060
},
{
"epoch": 2.2,
"grad_norm": 0.8902627229690552,
"learning_rate": 7.840384615384616e-05,
"loss": 0.4439,
"step": 8070
},
{
"epoch": 2.21,
"grad_norm": 0.7599775195121765,
"learning_rate": 7.837637362637364e-05,
"loss": 0.4537,
"step": 8080
},
{
"epoch": 2.21,
"grad_norm": 0.9020054340362549,
"learning_rate": 7.83489010989011e-05,
"loss": 0.4658,
"step": 8090
},
{
"epoch": 2.21,
"grad_norm": 0.8659080266952515,
"learning_rate": 7.832142857142857e-05,
"loss": 0.4553,
"step": 8100
},
{
"epoch": 2.22,
"grad_norm": 0.6685529351234436,
"learning_rate": 7.829395604395606e-05,
"loss": 0.4604,
"step": 8110
},
{
"epoch": 2.22,
"grad_norm": 0.6612065434455872,
"learning_rate": 7.826648351648352e-05,
"loss": 0.4522,
"step": 8120
},
{
"epoch": 2.22,
"grad_norm": 0.7674763202667236,
"learning_rate": 7.823901098901099e-05,
"loss": 0.4523,
"step": 8130
},
{
"epoch": 2.22,
"grad_norm": 0.5680274963378906,
"learning_rate": 7.821153846153847e-05,
"loss": 0.4649,
"step": 8140
},
{
"epoch": 2.23,
"grad_norm": 1.0545960664749146,
"learning_rate": 7.818406593406594e-05,
"loss": 0.4507,
"step": 8150
},
{
"epoch": 2.23,
"grad_norm": 0.5723569393157959,
"learning_rate": 7.815659340659341e-05,
"loss": 0.4418,
"step": 8160
},
{
"epoch": 2.23,
"grad_norm": 0.9119620323181152,
"learning_rate": 7.812912087912089e-05,
"loss": 0.4535,
"step": 8170
},
{
"epoch": 2.23,
"grad_norm": 1.0251758098602295,
"learning_rate": 7.810164835164836e-05,
"loss": 0.4503,
"step": 8180
},
{
"epoch": 2.24,
"grad_norm": 0.6839396357536316,
"learning_rate": 7.807417582417583e-05,
"loss": 0.452,
"step": 8190
},
{
"epoch": 2.24,
"grad_norm": 1.3468687534332275,
"learning_rate": 7.804670329670331e-05,
"loss": 0.4442,
"step": 8200
},
{
"epoch": 2.24,
"grad_norm": 0.9562901854515076,
"learning_rate": 7.801923076923078e-05,
"loss": 0.4596,
"step": 8210
},
{
"epoch": 2.25,
"grad_norm": 1.092677354812622,
"learning_rate": 7.799175824175825e-05,
"loss": 0.4454,
"step": 8220
},
{
"epoch": 2.25,
"grad_norm": 0.7029426097869873,
"learning_rate": 7.796428571428573e-05,
"loss": 0.4463,
"step": 8230
},
{
"epoch": 2.25,
"grad_norm": 0.7543437480926514,
"learning_rate": 7.79368131868132e-05,
"loss": 0.4588,
"step": 8240
},
{
"epoch": 2.25,
"grad_norm": 1.5089532136917114,
"learning_rate": 7.790934065934066e-05,
"loss": 0.4531,
"step": 8250
},
{
"epoch": 2.26,
"grad_norm": 0.6691949963569641,
"learning_rate": 7.788186813186815e-05,
"loss": 0.4494,
"step": 8260
},
{
"epoch": 2.26,
"grad_norm": 1.0210888385772705,
"learning_rate": 7.785439560439561e-05,
"loss": 0.4548,
"step": 8270
},
{
"epoch": 2.26,
"grad_norm": 0.9251496195793152,
"learning_rate": 7.782692307692308e-05,
"loss": 0.4576,
"step": 8280
},
{
"epoch": 2.26,
"grad_norm": 1.2618598937988281,
"learning_rate": 7.779945054945055e-05,
"loss": 0.4478,
"step": 8290
},
{
"epoch": 2.27,
"grad_norm": 1.2252585887908936,
"learning_rate": 7.777197802197802e-05,
"loss": 0.4573,
"step": 8300
},
{
"epoch": 2.27,
"grad_norm": 0.6124866604804993,
"learning_rate": 7.77445054945055e-05,
"loss": 0.4506,
"step": 8310
},
{
"epoch": 2.27,
"grad_norm": 1.0547674894332886,
"learning_rate": 7.771703296703297e-05,
"loss": 0.4418,
"step": 8320
},
{
"epoch": 2.28,
"grad_norm": 1.1824958324432373,
"learning_rate": 7.768956043956044e-05,
"loss": 0.4529,
"step": 8330
},
{
"epoch": 2.28,
"grad_norm": 0.9790034890174866,
"learning_rate": 7.76620879120879e-05,
"loss": 0.455,
"step": 8340
},
{
"epoch": 2.28,
"grad_norm": 0.9925460815429688,
"learning_rate": 7.763461538461539e-05,
"loss": 0.4607,
"step": 8350
},
{
"epoch": 2.28,
"grad_norm": 0.8586925864219666,
"learning_rate": 7.760714285714285e-05,
"loss": 0.4644,
"step": 8360
},
{
"epoch": 2.29,
"grad_norm": 1.4445006847381592,
"learning_rate": 7.757967032967032e-05,
"loss": 0.4513,
"step": 8370
},
{
"epoch": 2.29,
"grad_norm": 0.9150937795639038,
"learning_rate": 7.75521978021978e-05,
"loss": 0.4534,
"step": 8380
},
{
"epoch": 2.29,
"grad_norm": 0.9148111939430237,
"learning_rate": 7.752472527472527e-05,
"loss": 0.4424,
"step": 8390
},
{
"epoch": 2.29,
"grad_norm": 0.9973462224006653,
"learning_rate": 7.749725274725274e-05,
"loss": 0.4523,
"step": 8400
},
{
"epoch": 2.3,
"grad_norm": 1.022926926612854,
"learning_rate": 7.746978021978022e-05,
"loss": 0.444,
"step": 8410
},
{
"epoch": 2.3,
"grad_norm": 0.7249711155891418,
"learning_rate": 7.744230769230769e-05,
"loss": 0.4473,
"step": 8420
},
{
"epoch": 2.3,
"grad_norm": 0.44306471943855286,
"learning_rate": 7.741483516483516e-05,
"loss": 0.4452,
"step": 8430
},
{
"epoch": 2.31,
"grad_norm": 0.9625614881515503,
"learning_rate": 7.738736263736264e-05,
"loss": 0.4373,
"step": 8440
},
{
"epoch": 2.31,
"grad_norm": 0.5116167664527893,
"learning_rate": 7.735989010989011e-05,
"loss": 0.4505,
"step": 8450
},
{
"epoch": 2.31,
"grad_norm": 0.7496402859687805,
"learning_rate": 7.733241758241758e-05,
"loss": 0.4451,
"step": 8460
},
{
"epoch": 2.31,
"grad_norm": 1.3204615116119385,
"learning_rate": 7.730494505494506e-05,
"loss": 0.4576,
"step": 8470
},
{
"epoch": 2.32,
"grad_norm": 0.7599005699157715,
"learning_rate": 7.727747252747253e-05,
"loss": 0.4493,
"step": 8480
},
{
"epoch": 2.32,
"grad_norm": 0.5699270367622375,
"learning_rate": 7.725e-05,
"loss": 0.4568,
"step": 8490
},
{
"epoch": 2.32,
"grad_norm": 0.6626598238945007,
"learning_rate": 7.722252747252748e-05,
"loss": 0.4504,
"step": 8500
},
{
"epoch": 2.32,
"grad_norm": 0.6977862119674683,
"learning_rate": 7.719505494505494e-05,
"loss": 0.4529,
"step": 8510
},
{
"epoch": 2.33,
"grad_norm": 0.883980393409729,
"learning_rate": 7.716758241758241e-05,
"loss": 0.4527,
"step": 8520
},
{
"epoch": 2.33,
"grad_norm": 0.9930477738380432,
"learning_rate": 7.71401098901099e-05,
"loss": 0.4524,
"step": 8530
},
{
"epoch": 2.33,
"grad_norm": 0.5075715780258179,
"learning_rate": 7.711263736263736e-05,
"loss": 0.4436,
"step": 8540
},
{
"epoch": 2.34,
"grad_norm": 0.7683893442153931,
"learning_rate": 7.708516483516483e-05,
"loss": 0.4608,
"step": 8550
},
{
"epoch": 2.34,
"grad_norm": 1.6546239852905273,
"learning_rate": 7.705769230769231e-05,
"loss": 0.4521,
"step": 8560
},
{
"epoch": 2.34,
"grad_norm": 0.8132153153419495,
"learning_rate": 7.703021978021978e-05,
"loss": 0.4542,
"step": 8570
},
{
"epoch": 2.34,
"grad_norm": 0.5142685770988464,
"learning_rate": 7.700274725274725e-05,
"loss": 0.4548,
"step": 8580
},
{
"epoch": 2.35,
"grad_norm": 1.100368618965149,
"learning_rate": 7.697527472527473e-05,
"loss": 0.4602,
"step": 8590
},
{
"epoch": 2.35,
"grad_norm": 1.018351674079895,
"learning_rate": 7.69478021978022e-05,
"loss": 0.4424,
"step": 8600
},
{
"epoch": 2.35,
"grad_norm": 0.557134211063385,
"learning_rate": 7.692032967032967e-05,
"loss": 0.4486,
"step": 8610
},
{
"epoch": 2.35,
"grad_norm": 0.701932966709137,
"learning_rate": 7.689285714285715e-05,
"loss": 0.4543,
"step": 8620
},
{
"epoch": 2.36,
"grad_norm": 1.0351616144180298,
"learning_rate": 7.686538461538462e-05,
"loss": 0.4628,
"step": 8630
},
{
"epoch": 2.36,
"grad_norm": 0.5028539896011353,
"learning_rate": 7.683791208791208e-05,
"loss": 0.4566,
"step": 8640
},
{
"epoch": 2.36,
"grad_norm": 0.7769728302955627,
"learning_rate": 7.681043956043957e-05,
"loss": 0.4429,
"step": 8650
},
{
"epoch": 2.37,
"grad_norm": 0.7966272830963135,
"learning_rate": 7.678296703296703e-05,
"loss": 0.4377,
"step": 8660
},
{
"epoch": 2.37,
"grad_norm": 1.1576814651489258,
"learning_rate": 7.67554945054945e-05,
"loss": 0.4507,
"step": 8670
},
{
"epoch": 2.37,
"grad_norm": 0.8654575943946838,
"learning_rate": 7.672802197802198e-05,
"loss": 0.4598,
"step": 8680
},
{
"epoch": 2.37,
"grad_norm": 0.7794680595397949,
"learning_rate": 7.670054945054945e-05,
"loss": 0.4424,
"step": 8690
},
{
"epoch": 2.38,
"grad_norm": 0.8271937370300293,
"learning_rate": 7.667307692307692e-05,
"loss": 0.4586,
"step": 8700
},
{
"epoch": 2.38,
"grad_norm": 0.8499754071235657,
"learning_rate": 7.66456043956044e-05,
"loss": 0.4476,
"step": 8710
},
{
"epoch": 2.38,
"grad_norm": 0.7553250789642334,
"learning_rate": 7.661813186813187e-05,
"loss": 0.4435,
"step": 8720
},
{
"epoch": 2.38,
"grad_norm": 0.5660779476165771,
"learning_rate": 7.659065934065934e-05,
"loss": 0.4518,
"step": 8730
},
{
"epoch": 2.39,
"grad_norm": 1.4875767230987549,
"learning_rate": 7.656318681318682e-05,
"loss": 0.4542,
"step": 8740
},
{
"epoch": 2.39,
"grad_norm": 1.1859349012374878,
"learning_rate": 7.653846153846153e-05,
"loss": 0.4528,
"step": 8750
},
{
"epoch": 2.39,
"grad_norm": 0.6259100437164307,
"learning_rate": 7.651098901098902e-05,
"loss": 0.4629,
"step": 8760
},
{
"epoch": 2.4,
"grad_norm": 0.4763346314430237,
"learning_rate": 7.648351648351648e-05,
"loss": 0.4479,
"step": 8770
},
{
"epoch": 2.4,
"grad_norm": 0.6478772759437561,
"learning_rate": 7.645604395604395e-05,
"loss": 0.4313,
"step": 8780
},
{
"epoch": 2.4,
"grad_norm": 1.1578466892242432,
"learning_rate": 7.642857142857143e-05,
"loss": 0.4499,
"step": 8790
},
{
"epoch": 2.4,
"grad_norm": 0.5163312554359436,
"learning_rate": 7.64010989010989e-05,
"loss": 0.4479,
"step": 8800
},
{
"epoch": 2.41,
"grad_norm": 1.183361530303955,
"learning_rate": 7.637362637362637e-05,
"loss": 0.4524,
"step": 8810
},
{
"epoch": 2.41,
"grad_norm": 0.6872965097427368,
"learning_rate": 7.634615384615385e-05,
"loss": 0.4445,
"step": 8820
},
{
"epoch": 2.41,
"grad_norm": 0.7339246273040771,
"learning_rate": 7.631868131868132e-05,
"loss": 0.4446,
"step": 8830
},
{
"epoch": 2.41,
"grad_norm": 0.7101340889930725,
"learning_rate": 7.629120879120879e-05,
"loss": 0.45,
"step": 8840
},
{
"epoch": 2.42,
"grad_norm": 0.5717099905014038,
"learning_rate": 7.626373626373627e-05,
"loss": 0.4503,
"step": 8850
},
{
"epoch": 2.42,
"grad_norm": 0.6713473796844482,
"learning_rate": 7.623626373626374e-05,
"loss": 0.4528,
"step": 8860
},
{
"epoch": 2.42,
"grad_norm": 0.9480137228965759,
"learning_rate": 7.62087912087912e-05,
"loss": 0.4569,
"step": 8870
},
{
"epoch": 2.43,
"grad_norm": 0.594673752784729,
"learning_rate": 7.618131868131869e-05,
"loss": 0.4627,
"step": 8880
},
{
"epoch": 2.43,
"grad_norm": 0.8691803216934204,
"learning_rate": 7.615384615384616e-05,
"loss": 0.4459,
"step": 8890
},
{
"epoch": 2.43,
"grad_norm": 0.8882876634597778,
"learning_rate": 7.612637362637362e-05,
"loss": 0.4478,
"step": 8900
},
{
"epoch": 2.43,
"grad_norm": 0.48794516921043396,
"learning_rate": 7.60989010989011e-05,
"loss": 0.4479,
"step": 8910
},
{
"epoch": 2.44,
"grad_norm": 0.4966289699077606,
"learning_rate": 7.607142857142857e-05,
"loss": 0.4395,
"step": 8920
},
{
"epoch": 2.44,
"grad_norm": 0.792293131351471,
"learning_rate": 7.604395604395604e-05,
"loss": 0.4353,
"step": 8930
},
{
"epoch": 2.44,
"grad_norm": 0.6615070104598999,
"learning_rate": 7.601648351648351e-05,
"loss": 0.4478,
"step": 8940
},
{
"epoch": 2.44,
"grad_norm": 0.8686419725418091,
"learning_rate": 7.598901098901099e-05,
"loss": 0.4516,
"step": 8950
},
{
"epoch": 2.45,
"grad_norm": 1.2833001613616943,
"learning_rate": 7.596153846153846e-05,
"loss": 0.4526,
"step": 8960
},
{
"epoch": 2.45,
"grad_norm": 0.7414630055427551,
"learning_rate": 7.593406593406593e-05,
"loss": 0.4549,
"step": 8970
},
{
"epoch": 2.45,
"grad_norm": 0.9745138883590698,
"learning_rate": 7.590659340659341e-05,
"loss": 0.4412,
"step": 8980
},
{
"epoch": 2.46,
"grad_norm": 0.8407401442527771,
"learning_rate": 7.587912087912088e-05,
"loss": 0.4511,
"step": 8990
},
{
"epoch": 2.46,
"grad_norm": 1.1383010149002075,
"learning_rate": 7.585164835164835e-05,
"loss": 0.4556,
"step": 9000
},
{
"epoch": 2.46,
"grad_norm": 1.0419915914535522,
"learning_rate": 7.582417582417583e-05,
"loss": 0.4504,
"step": 9010
},
{
"epoch": 2.46,
"grad_norm": 0.931141197681427,
"learning_rate": 7.57967032967033e-05,
"loss": 0.4464,
"step": 9020
},
{
"epoch": 2.47,
"grad_norm": 0.6264098286628723,
"learning_rate": 7.576923076923076e-05,
"loss": 0.4627,
"step": 9030
},
{
"epoch": 2.47,
"grad_norm": 0.8912045359611511,
"learning_rate": 7.574175824175825e-05,
"loss": 0.4452,
"step": 9040
},
{
"epoch": 2.47,
"grad_norm": 1.3010141849517822,
"learning_rate": 7.571428571428571e-05,
"loss": 0.4595,
"step": 9050
},
{
"epoch": 2.47,
"grad_norm": 0.8275282979011536,
"learning_rate": 7.568681318681318e-05,
"loss": 0.4486,
"step": 9060
},
{
"epoch": 2.48,
"grad_norm": 0.8537049293518066,
"learning_rate": 7.565934065934066e-05,
"loss": 0.4506,
"step": 9070
},
{
"epoch": 2.48,
"grad_norm": 0.5747140645980835,
"learning_rate": 7.563186813186813e-05,
"loss": 0.4488,
"step": 9080
},
{
"epoch": 2.48,
"grad_norm": 0.8508340716362,
"learning_rate": 7.56043956043956e-05,
"loss": 0.4436,
"step": 9090
},
{
"epoch": 2.49,
"grad_norm": 1.1689362525939941,
"learning_rate": 7.557692307692308e-05,
"loss": 0.4586,
"step": 9100
},
{
"epoch": 2.49,
"grad_norm": 1.1048357486724854,
"learning_rate": 7.554945054945055e-05,
"loss": 0.4568,
"step": 9110
},
{
"epoch": 2.49,
"grad_norm": 1.1694691181182861,
"learning_rate": 7.552197802197802e-05,
"loss": 0.4571,
"step": 9120
},
{
"epoch": 2.49,
"grad_norm": 0.9257465600967407,
"learning_rate": 7.54945054945055e-05,
"loss": 0.4544,
"step": 9130
},
{
"epoch": 2.5,
"grad_norm": 1.3826684951782227,
"learning_rate": 7.546703296703297e-05,
"loss": 0.44,
"step": 9140
},
{
"epoch": 2.5,
"grad_norm": 0.8815349340438843,
"learning_rate": 7.543956043956044e-05,
"loss": 0.4438,
"step": 9150
},
{
"epoch": 2.5,
"grad_norm": 0.661222517490387,
"learning_rate": 7.541208791208792e-05,
"loss": 0.4479,
"step": 9160
},
{
"epoch": 2.5,
"grad_norm": 1.222886085510254,
"learning_rate": 7.538461538461539e-05,
"loss": 0.4487,
"step": 9170
},
{
"epoch": 2.51,
"grad_norm": 1.1454687118530273,
"learning_rate": 7.535714285714285e-05,
"loss": 0.4625,
"step": 9180
},
{
"epoch": 2.51,
"grad_norm": 0.49560263752937317,
"learning_rate": 7.532967032967034e-05,
"loss": 0.4518,
"step": 9190
},
{
"epoch": 2.51,
"grad_norm": 0.8997675776481628,
"learning_rate": 7.53021978021978e-05,
"loss": 0.4426,
"step": 9200
},
{
"epoch": 2.52,
"grad_norm": 0.9074673056602478,
"learning_rate": 7.527472527472527e-05,
"loss": 0.4524,
"step": 9210
},
{
"epoch": 2.52,
"grad_norm": 0.7328269481658936,
"learning_rate": 7.524725274725275e-05,
"loss": 0.4512,
"step": 9220
},
{
"epoch": 2.52,
"grad_norm": 0.7607899904251099,
"learning_rate": 7.521978021978022e-05,
"loss": 0.4491,
"step": 9230
},
{
"epoch": 2.52,
"grad_norm": 0.5627162456512451,
"learning_rate": 7.519230769230769e-05,
"loss": 0.4511,
"step": 9240
},
{
"epoch": 2.53,
"grad_norm": 0.9358956813812256,
"learning_rate": 7.516483516483517e-05,
"loss": 0.453,
"step": 9250
},
{
"epoch": 2.53,
"grad_norm": 1.2619479894638062,
"learning_rate": 7.513736263736264e-05,
"loss": 0.4432,
"step": 9260
},
{
"epoch": 2.53,
"grad_norm": 0.9375318884849548,
"learning_rate": 7.510989010989011e-05,
"loss": 0.4539,
"step": 9270
},
{
"epoch": 2.53,
"grad_norm": 0.8454050421714783,
"learning_rate": 7.508241758241759e-05,
"loss": 0.4515,
"step": 9280
},
{
"epoch": 2.54,
"grad_norm": 0.9856541156768799,
"learning_rate": 7.505494505494506e-05,
"loss": 0.445,
"step": 9290
},
{
"epoch": 2.54,
"grad_norm": 0.6597672700881958,
"learning_rate": 7.502747252747253e-05,
"loss": 0.4505,
"step": 9300
},
{
"epoch": 2.54,
"grad_norm": 1.0249258279800415,
"learning_rate": 7.500000000000001e-05,
"loss": 0.4516,
"step": 9310
},
{
"epoch": 2.55,
"grad_norm": 1.003083348274231,
"learning_rate": 7.497252747252748e-05,
"loss": 0.4457,
"step": 9320
},
{
"epoch": 2.55,
"grad_norm": 1.236377477645874,
"learning_rate": 7.494505494505494e-05,
"loss": 0.4573,
"step": 9330
},
{
"epoch": 2.55,
"grad_norm": 1.5360130071640015,
"learning_rate": 7.491758241758242e-05,
"loss": 0.4478,
"step": 9340
},
{
"epoch": 2.55,
"grad_norm": 0.9379085898399353,
"learning_rate": 7.489010989010989e-05,
"loss": 0.4575,
"step": 9350
},
{
"epoch": 2.56,
"grad_norm": 0.7734600901603699,
"learning_rate": 7.486263736263736e-05,
"loss": 0.4571,
"step": 9360
},
{
"epoch": 2.56,
"grad_norm": 0.6792673468589783,
"learning_rate": 7.483516483516484e-05,
"loss": 0.4522,
"step": 9370
},
{
"epoch": 2.56,
"grad_norm": 0.8936319351196289,
"learning_rate": 7.480769230769231e-05,
"loss": 0.4576,
"step": 9380
},
{
"epoch": 2.57,
"grad_norm": 0.7541890144348145,
"learning_rate": 7.478021978021978e-05,
"loss": 0.4456,
"step": 9390
},
{
"epoch": 2.57,
"grad_norm": 0.8661054968833923,
"learning_rate": 7.475274725274726e-05,
"loss": 0.4428,
"step": 9400
},
{
"epoch": 2.57,
"grad_norm": 0.919201672077179,
"learning_rate": 7.472527472527473e-05,
"loss": 0.4507,
"step": 9410
},
{
"epoch": 2.57,
"grad_norm": 1.0688129663467407,
"learning_rate": 7.46978021978022e-05,
"loss": 0.4469,
"step": 9420
},
{
"epoch": 2.58,
"grad_norm": 7.527632236480713,
"learning_rate": 7.467032967032968e-05,
"loss": 0.4512,
"step": 9430
},
{
"epoch": 2.58,
"grad_norm": 0.5668841600418091,
"learning_rate": 7.464285714285715e-05,
"loss": 0.4484,
"step": 9440
},
{
"epoch": 2.58,
"grad_norm": 0.7174908518791199,
"learning_rate": 7.461538461538462e-05,
"loss": 0.4361,
"step": 9450
},
{
"epoch": 2.58,
"grad_norm": 0.693144679069519,
"learning_rate": 7.45879120879121e-05,
"loss": 0.4483,
"step": 9460
},
{
"epoch": 2.59,
"grad_norm": 1.2530773878097534,
"learning_rate": 7.456043956043956e-05,
"loss": 0.4495,
"step": 9470
},
{
"epoch": 2.59,
"grad_norm": 0.7528218626976013,
"learning_rate": 7.453296703296703e-05,
"loss": 0.4429,
"step": 9480
},
{
"epoch": 2.59,
"grad_norm": 0.596477746963501,
"learning_rate": 7.450549450549451e-05,
"loss": 0.4526,
"step": 9490
},
{
"epoch": 2.6,
"grad_norm": 0.5093569755554199,
"learning_rate": 7.447802197802198e-05,
"loss": 0.4419,
"step": 9500
},
{
"epoch": 2.6,
"grad_norm": 1.0681297779083252,
"learning_rate": 7.445054945054945e-05,
"loss": 0.4394,
"step": 9510
},
{
"epoch": 2.6,
"grad_norm": 1.280721664428711,
"learning_rate": 7.442307692307693e-05,
"loss": 0.4425,
"step": 9520
},
{
"epoch": 2.6,
"grad_norm": 0.8306493759155273,
"learning_rate": 7.43956043956044e-05,
"loss": 0.4464,
"step": 9530
},
{
"epoch": 2.61,
"grad_norm": 0.7929884195327759,
"learning_rate": 7.436813186813187e-05,
"loss": 0.4532,
"step": 9540
},
{
"epoch": 2.61,
"grad_norm": 0.9622000455856323,
"learning_rate": 7.434065934065935e-05,
"loss": 0.4516,
"step": 9550
},
{
"epoch": 2.61,
"grad_norm": 0.8442659378051758,
"learning_rate": 7.431318681318682e-05,
"loss": 0.4417,
"step": 9560
},
{
"epoch": 2.61,
"grad_norm": 0.6593378782272339,
"learning_rate": 7.428571428571429e-05,
"loss": 0.4417,
"step": 9570
},
{
"epoch": 2.62,
"grad_norm": 0.8290765881538391,
"learning_rate": 7.425824175824177e-05,
"loss": 0.4471,
"step": 9580
},
{
"epoch": 2.62,
"grad_norm": 0.9711737632751465,
"learning_rate": 7.423076923076924e-05,
"loss": 0.4403,
"step": 9590
},
{
"epoch": 2.62,
"grad_norm": 0.9451545476913452,
"learning_rate": 7.42032967032967e-05,
"loss": 0.4522,
"step": 9600
},
{
"epoch": 2.63,
"grad_norm": 1.1259006261825562,
"learning_rate": 7.417582417582419e-05,
"loss": 0.4557,
"step": 9610
},
{
"epoch": 2.63,
"grad_norm": 0.6328985095024109,
"learning_rate": 7.414835164835165e-05,
"loss": 0.4454,
"step": 9620
},
{
"epoch": 2.63,
"grad_norm": 0.9557839035987854,
"learning_rate": 7.412087912087912e-05,
"loss": 0.4627,
"step": 9630
},
{
"epoch": 2.63,
"grad_norm": 0.6751061081886292,
"learning_rate": 7.40934065934066e-05,
"loss": 0.4521,
"step": 9640
},
{
"epoch": 2.64,
"grad_norm": 0.49840301275253296,
"learning_rate": 7.406593406593407e-05,
"loss": 0.4541,
"step": 9650
},
{
"epoch": 2.64,
"grad_norm": 1.2455097436904907,
"learning_rate": 7.403846153846154e-05,
"loss": 0.4466,
"step": 9660
},
{
"epoch": 2.64,
"grad_norm": 1.1257638931274414,
"learning_rate": 7.401098901098902e-05,
"loss": 0.4492,
"step": 9670
},
{
"epoch": 2.64,
"grad_norm": 0.623653769493103,
"learning_rate": 7.398351648351649e-05,
"loss": 0.4475,
"step": 9680
},
{
"epoch": 2.65,
"grad_norm": 0.6021344661712646,
"learning_rate": 7.395604395604396e-05,
"loss": 0.4393,
"step": 9690
},
{
"epoch": 2.65,
"grad_norm": 0.7180885672569275,
"learning_rate": 7.392857142857144e-05,
"loss": 0.4571,
"step": 9700
},
{
"epoch": 2.65,
"grad_norm": 0.5711954832077026,
"learning_rate": 7.390109890109891e-05,
"loss": 0.4534,
"step": 9710
},
{
"epoch": 2.66,
"grad_norm": 0.8202961683273315,
"learning_rate": 7.387362637362638e-05,
"loss": 0.4568,
"step": 9720
},
{
"epoch": 2.66,
"grad_norm": 0.5622140765190125,
"learning_rate": 7.384615384615386e-05,
"loss": 0.4415,
"step": 9730
},
{
"epoch": 2.66,
"grad_norm": 0.9798756837844849,
"learning_rate": 7.381868131868133e-05,
"loss": 0.4371,
"step": 9740
},
{
"epoch": 2.66,
"grad_norm": 0.6842076778411865,
"learning_rate": 7.37912087912088e-05,
"loss": 0.4495,
"step": 9750
},
{
"epoch": 2.67,
"grad_norm": 0.6150854229927063,
"learning_rate": 7.376373626373628e-05,
"loss": 0.4566,
"step": 9760
},
{
"epoch": 2.67,
"grad_norm": 0.5068964958190918,
"learning_rate": 7.373626373626374e-05,
"loss": 0.4445,
"step": 9770
},
{
"epoch": 2.67,
"grad_norm": 0.8547071218490601,
"learning_rate": 7.370879120879121e-05,
"loss": 0.4398,
"step": 9780
},
{
"epoch": 2.67,
"grad_norm": 0.9119787216186523,
"learning_rate": 7.36813186813187e-05,
"loss": 0.4622,
"step": 9790
},
{
"epoch": 2.68,
"grad_norm": 0.6831972002983093,
"learning_rate": 7.365384615384616e-05,
"loss": 0.4493,
"step": 9800
},
{
"epoch": 2.68,
"grad_norm": 1.850950002670288,
"learning_rate": 7.362637362637363e-05,
"loss": 0.4418,
"step": 9810
},
{
"epoch": 2.68,
"grad_norm": 1.3177289962768555,
"learning_rate": 7.359890109890111e-05,
"loss": 0.4514,
"step": 9820
},
{
"epoch": 2.69,
"grad_norm": 0.8253613114356995,
"learning_rate": 7.357142857142858e-05,
"loss": 0.4536,
"step": 9830
},
{
"epoch": 2.69,
"grad_norm": 0.9882771968841553,
"learning_rate": 7.354395604395605e-05,
"loss": 0.4388,
"step": 9840
},
{
"epoch": 2.69,
"grad_norm": 1.1343753337860107,
"learning_rate": 7.351648351648353e-05,
"loss": 0.4495,
"step": 9850
},
{
"epoch": 2.69,
"grad_norm": 0.9710454344749451,
"learning_rate": 7.3489010989011e-05,
"loss": 0.4546,
"step": 9860
},
{
"epoch": 2.7,
"grad_norm": 1.0125503540039062,
"learning_rate": 7.346153846153847e-05,
"loss": 0.4436,
"step": 9870
},
{
"epoch": 2.7,
"grad_norm": 0.972205400466919,
"learning_rate": 7.343406593406593e-05,
"loss": 0.4603,
"step": 9880
},
{
"epoch": 2.7,
"grad_norm": 0.8217471837997437,
"learning_rate": 7.34065934065934e-05,
"loss": 0.4551,
"step": 9890
},
{
"epoch": 2.7,
"grad_norm": 0.9497684836387634,
"learning_rate": 7.337912087912087e-05,
"loss": 0.4398,
"step": 9900
},
{
"epoch": 2.71,
"grad_norm": 1.2341935634613037,
"learning_rate": 7.335164835164835e-05,
"loss": 0.4412,
"step": 9910
},
{
"epoch": 2.71,
"grad_norm": 1.0097248554229736,
"learning_rate": 7.332417582417582e-05,
"loss": 0.4536,
"step": 9920
},
{
"epoch": 2.71,
"grad_norm": 1.1135015487670898,
"learning_rate": 7.329670329670329e-05,
"loss": 0.4482,
"step": 9930
},
{
"epoch": 2.72,
"grad_norm": 1.0130164623260498,
"learning_rate": 7.326923076923077e-05,
"loss": 0.4499,
"step": 9940
},
{
"epoch": 2.72,
"grad_norm": 0.802757740020752,
"learning_rate": 7.324175824175824e-05,
"loss": 0.4544,
"step": 9950
},
{
"epoch": 2.72,
"grad_norm": 0.5492908954620361,
"learning_rate": 7.321428571428571e-05,
"loss": 0.4457,
"step": 9960
},
{
"epoch": 2.72,
"grad_norm": 1.1192032098770142,
"learning_rate": 7.318681318681319e-05,
"loss": 0.4464,
"step": 9970
},
{
"epoch": 2.73,
"grad_norm": 0.5543531179428101,
"learning_rate": 7.315934065934066e-05,
"loss": 0.456,
"step": 9980
},
{
"epoch": 2.73,
"grad_norm": 0.741788387298584,
"learning_rate": 7.313186813186812e-05,
"loss": 0.4384,
"step": 9990
},
{
"epoch": 2.73,
"grad_norm": 0.6588950157165527,
"learning_rate": 7.31043956043956e-05,
"loss": 0.4606,
"step": 10000
},
{
"epoch": 2.73,
"grad_norm": 0.5929410457611084,
"learning_rate": 7.307692307692307e-05,
"loss": 0.4474,
"step": 10010
},
{
"epoch": 2.74,
"grad_norm": 0.624821662902832,
"learning_rate": 7.304945054945054e-05,
"loss": 0.4486,
"step": 10020
},
{
"epoch": 2.74,
"grad_norm": 0.7971046566963196,
"learning_rate": 7.302197802197802e-05,
"loss": 0.4433,
"step": 10030
},
{
"epoch": 2.74,
"grad_norm": 0.81767737865448,
"learning_rate": 7.299450549450549e-05,
"loss": 0.4583,
"step": 10040
},
{
"epoch": 2.75,
"grad_norm": 0.7773674726486206,
"learning_rate": 7.296703296703296e-05,
"loss": 0.4488,
"step": 10050
},
{
"epoch": 2.75,
"grad_norm": 0.7860414981842041,
"learning_rate": 7.293956043956044e-05,
"loss": 0.4548,
"step": 10060
},
{
"epoch": 2.75,
"grad_norm": 0.627571702003479,
"learning_rate": 7.291208791208791e-05,
"loss": 0.4477,
"step": 10070
},
{
"epoch": 2.75,
"grad_norm": 0.7049760818481445,
"learning_rate": 7.288461538461538e-05,
"loss": 0.4376,
"step": 10080
},
{
"epoch": 2.76,
"grad_norm": 0.5372808575630188,
"learning_rate": 7.285714285714286e-05,
"loss": 0.4552,
"step": 10090
},
{
"epoch": 2.76,
"grad_norm": 0.6712595224380493,
"learning_rate": 7.282967032967033e-05,
"loss": 0.4555,
"step": 10100
},
{
"epoch": 2.76,
"grad_norm": 0.4722815752029419,
"learning_rate": 7.28021978021978e-05,
"loss": 0.4536,
"step": 10110
},
{
"epoch": 2.76,
"grad_norm": 0.48171865940093994,
"learning_rate": 7.277472527472528e-05,
"loss": 0.4542,
"step": 10120
},
{
"epoch": 2.77,
"grad_norm": 0.8346227407455444,
"learning_rate": 7.274725274725275e-05,
"loss": 0.454,
"step": 10130
},
{
"epoch": 2.77,
"grad_norm": 0.42196372151374817,
"learning_rate": 7.271978021978021e-05,
"loss": 0.4514,
"step": 10140
},
{
"epoch": 2.77,
"grad_norm": 0.655756950378418,
"learning_rate": 7.26923076923077e-05,
"loss": 0.458,
"step": 10150
},
{
"epoch": 2.78,
"grad_norm": 1.0113352537155151,
"learning_rate": 7.266483516483516e-05,
"loss": 0.4537,
"step": 10160
},
{
"epoch": 2.78,
"grad_norm": 0.5985142588615417,
"learning_rate": 7.263736263736263e-05,
"loss": 0.4408,
"step": 10170
},
{
"epoch": 2.78,
"grad_norm": 1.50113844871521,
"learning_rate": 7.260989010989011e-05,
"loss": 0.4516,
"step": 10180
},
{
"epoch": 2.78,
"grad_norm": 1.536289095878601,
"learning_rate": 7.258241758241758e-05,
"loss": 0.4465,
"step": 10190
},
{
"epoch": 2.79,
"grad_norm": 0.6726506948471069,
"learning_rate": 7.255494505494505e-05,
"loss": 0.4365,
"step": 10200
},
{
"epoch": 2.79,
"grad_norm": 0.6390563249588013,
"learning_rate": 7.252747252747253e-05,
"loss": 0.4401,
"step": 10210
},
{
"epoch": 2.79,
"grad_norm": 0.7105517983436584,
"learning_rate": 7.25e-05,
"loss": 0.445,
"step": 10220
},
{
"epoch": 2.79,
"grad_norm": 0.9596685767173767,
"learning_rate": 7.247252747252747e-05,
"loss": 0.4367,
"step": 10230
},
{
"epoch": 2.8,
"grad_norm": 1.161080002784729,
"learning_rate": 7.244505494505495e-05,
"loss": 0.4372,
"step": 10240
},
{
"epoch": 2.8,
"grad_norm": 0.9476664662361145,
"learning_rate": 7.241758241758242e-05,
"loss": 0.4416,
"step": 10250
},
{
"epoch": 2.8,
"grad_norm": 0.7320071458816528,
"learning_rate": 7.239010989010989e-05,
"loss": 0.4549,
"step": 10260
},
{
"epoch": 2.81,
"grad_norm": 1.2132612466812134,
"learning_rate": 7.236263736263737e-05,
"loss": 0.4573,
"step": 10270
},
{
"epoch": 2.81,
"grad_norm": 0.5772107839584351,
"learning_rate": 7.233516483516484e-05,
"loss": 0.4502,
"step": 10280
},
{
"epoch": 2.81,
"grad_norm": 1.0478837490081787,
"learning_rate": 7.23076923076923e-05,
"loss": 0.452,
"step": 10290
},
{
"epoch": 2.81,
"grad_norm": 0.6411312222480774,
"learning_rate": 7.228021978021979e-05,
"loss": 0.4424,
"step": 10300
},
{
"epoch": 2.82,
"grad_norm": 0.7831696271896362,
"learning_rate": 7.225274725274725e-05,
"loss": 0.4525,
"step": 10310
},
{
"epoch": 2.82,
"grad_norm": 0.45390376448631287,
"learning_rate": 7.222527472527472e-05,
"loss": 0.4401,
"step": 10320
},
{
"epoch": 2.82,
"grad_norm": 0.8081623911857605,
"learning_rate": 7.21978021978022e-05,
"loss": 0.4396,
"step": 10330
},
{
"epoch": 2.82,
"grad_norm": 1.091840147972107,
"learning_rate": 7.217032967032967e-05,
"loss": 0.4481,
"step": 10340
},
{
"epoch": 2.83,
"grad_norm": 1.3704334497451782,
"learning_rate": 7.214285714285714e-05,
"loss": 0.4563,
"step": 10350
},
{
"epoch": 2.83,
"grad_norm": 0.8897873163223267,
"learning_rate": 7.211538461538462e-05,
"loss": 0.4537,
"step": 10360
},
{
"epoch": 2.83,
"grad_norm": 1.2123398780822754,
"learning_rate": 7.208791208791209e-05,
"loss": 0.4421,
"step": 10370
},
{
"epoch": 2.84,
"grad_norm": 0.7756679058074951,
"learning_rate": 7.206043956043956e-05,
"loss": 0.4455,
"step": 10380
},
{
"epoch": 2.84,
"grad_norm": 0.5030421614646912,
"learning_rate": 7.203296703296704e-05,
"loss": 0.4482,
"step": 10390
},
{
"epoch": 2.84,
"grad_norm": 0.5804952383041382,
"learning_rate": 7.200549450549451e-05,
"loss": 0.4468,
"step": 10400
},
{
"epoch": 2.84,
"grad_norm": 0.5755320191383362,
"learning_rate": 7.197802197802198e-05,
"loss": 0.4473,
"step": 10410
},
{
"epoch": 2.85,
"grad_norm": 0.7598744034767151,
"learning_rate": 7.195054945054946e-05,
"loss": 0.4609,
"step": 10420
},
{
"epoch": 2.85,
"grad_norm": 0.5830199122428894,
"learning_rate": 7.192307692307693e-05,
"loss": 0.4574,
"step": 10430
},
{
"epoch": 2.85,
"grad_norm": 0.6908131837844849,
"learning_rate": 7.18956043956044e-05,
"loss": 0.4419,
"step": 10440
},
{
"epoch": 2.85,
"grad_norm": 1.064139485359192,
"learning_rate": 7.186813186813188e-05,
"loss": 0.4467,
"step": 10450
},
{
"epoch": 2.86,
"grad_norm": 1.5694611072540283,
"learning_rate": 7.184065934065934e-05,
"loss": 0.4638,
"step": 10460
},
{
"epoch": 2.86,
"grad_norm": 0.6071065664291382,
"learning_rate": 7.181318681318681e-05,
"loss": 0.4449,
"step": 10470
},
{
"epoch": 2.86,
"grad_norm": 0.7417973279953003,
"learning_rate": 7.17857142857143e-05,
"loss": 0.4452,
"step": 10480
},
{
"epoch": 2.87,
"grad_norm": 1.0433728694915771,
"learning_rate": 7.175824175824176e-05,
"loss": 0.4413,
"step": 10490
},
{
"epoch": 2.87,
"grad_norm": 0.5444348454475403,
"learning_rate": 7.173076923076923e-05,
"loss": 0.453,
"step": 10500
},
{
"epoch": 2.87,
"grad_norm": 1.0602647066116333,
"learning_rate": 7.170329670329671e-05,
"loss": 0.4489,
"step": 10510
},
{
"epoch": 2.87,
"grad_norm": 0.7549176216125488,
"learning_rate": 7.167582417582418e-05,
"loss": 0.4322,
"step": 10520
},
{
"epoch": 2.88,
"grad_norm": 0.7759855389595032,
"learning_rate": 7.164835164835165e-05,
"loss": 0.4459,
"step": 10530
},
{
"epoch": 2.88,
"grad_norm": 0.6525812149047852,
"learning_rate": 7.162087912087913e-05,
"loss": 0.4533,
"step": 10540
},
{
"epoch": 2.88,
"grad_norm": 0.7948151230812073,
"learning_rate": 7.15934065934066e-05,
"loss": 0.4412,
"step": 10550
},
{
"epoch": 2.88,
"grad_norm": 0.8408129215240479,
"learning_rate": 7.156593406593407e-05,
"loss": 0.4441,
"step": 10560
},
{
"epoch": 2.89,
"grad_norm": 0.8035346865653992,
"learning_rate": 7.153846153846155e-05,
"loss": 0.463,
"step": 10570
},
{
"epoch": 2.89,
"grad_norm": 0.9992803335189819,
"learning_rate": 7.151098901098902e-05,
"loss": 0.4426,
"step": 10580
},
{
"epoch": 2.89,
"grad_norm": 0.7550234794616699,
"learning_rate": 7.148351648351648e-05,
"loss": 0.4455,
"step": 10590
},
{
"epoch": 2.9,
"grad_norm": 0.47757139801979065,
"learning_rate": 7.145604395604397e-05,
"loss": 0.4487,
"step": 10600
},
{
"epoch": 2.9,
"grad_norm": 0.6475192308425903,
"learning_rate": 7.142857142857143e-05,
"loss": 0.4493,
"step": 10610
},
{
"epoch": 2.9,
"grad_norm": 1.0151840448379517,
"learning_rate": 7.14010989010989e-05,
"loss": 0.432,
"step": 10620
},
{
"epoch": 2.9,
"grad_norm": 0.586617648601532,
"learning_rate": 7.137362637362638e-05,
"loss": 0.44,
"step": 10630
},
{
"epoch": 2.91,
"grad_norm": 0.49713996052742004,
"learning_rate": 7.134615384615385e-05,
"loss": 0.4411,
"step": 10640
},
{
"epoch": 2.91,
"grad_norm": 0.8117252588272095,
"learning_rate": 7.131868131868132e-05,
"loss": 0.4505,
"step": 10650
},
{
"epoch": 2.91,
"grad_norm": 0.9269289374351501,
"learning_rate": 7.12912087912088e-05,
"loss": 0.4481,
"step": 10660
},
{
"epoch": 2.91,
"grad_norm": 1.0942814350128174,
"learning_rate": 7.126373626373627e-05,
"loss": 0.4407,
"step": 10670
},
{
"epoch": 2.92,
"grad_norm": 0.6903828382492065,
"learning_rate": 7.123626373626374e-05,
"loss": 0.439,
"step": 10680
},
{
"epoch": 2.92,
"grad_norm": 0.8993271589279175,
"learning_rate": 7.120879120879122e-05,
"loss": 0.4421,
"step": 10690
},
{
"epoch": 2.92,
"grad_norm": 0.8367154002189636,
"learning_rate": 7.118131868131869e-05,
"loss": 0.4492,
"step": 10700
},
{
"epoch": 2.93,
"grad_norm": 0.7720317244529724,
"learning_rate": 7.115384615384616e-05,
"loss": 0.4521,
"step": 10710
},
{
"epoch": 2.93,
"grad_norm": 1.0028263330459595,
"learning_rate": 7.112637362637364e-05,
"loss": 0.4537,
"step": 10720
},
{
"epoch": 2.93,
"grad_norm": 0.663088321685791,
"learning_rate": 7.10989010989011e-05,
"loss": 0.4484,
"step": 10730
},
{
"epoch": 2.93,
"grad_norm": 1.360903263092041,
"learning_rate": 7.107142857142857e-05,
"loss": 0.4663,
"step": 10740
},
{
"epoch": 2.94,
"grad_norm": 1.5924625396728516,
"learning_rate": 7.104395604395606e-05,
"loss": 0.4599,
"step": 10750
},
{
"epoch": 2.94,
"grad_norm": 0.7223337888717651,
"learning_rate": 7.101648351648352e-05,
"loss": 0.4528,
"step": 10760
},
{
"epoch": 2.94,
"grad_norm": 0.8362312316894531,
"learning_rate": 7.098901098901099e-05,
"loss": 0.4469,
"step": 10770
},
{
"epoch": 2.94,
"grad_norm": 0.5594087839126587,
"learning_rate": 7.096153846153847e-05,
"loss": 0.4484,
"step": 10780
},
{
"epoch": 2.95,
"grad_norm": 0.8571014404296875,
"learning_rate": 7.093406593406594e-05,
"loss": 0.4414,
"step": 10790
},
{
"epoch": 2.95,
"grad_norm": 0.9150867462158203,
"learning_rate": 7.090659340659341e-05,
"loss": 0.453,
"step": 10800
},
{
"epoch": 2.95,
"grad_norm": 0.6607587337493896,
"learning_rate": 7.087912087912089e-05,
"loss": 0.4507,
"step": 10810
},
{
"epoch": 2.96,
"grad_norm": 0.7428970336914062,
"learning_rate": 7.085164835164836e-05,
"loss": 0.4498,
"step": 10820
},
{
"epoch": 2.96,
"grad_norm": 0.9267810583114624,
"learning_rate": 7.082417582417583e-05,
"loss": 0.4536,
"step": 10830
},
{
"epoch": 2.96,
"grad_norm": 0.4540146291255951,
"learning_rate": 7.079670329670331e-05,
"loss": 0.4396,
"step": 10840
},
{
"epoch": 2.96,
"grad_norm": 1.3146427869796753,
"learning_rate": 7.076923076923078e-05,
"loss": 0.4439,
"step": 10850
},
{
"epoch": 2.97,
"grad_norm": 0.7954951524734497,
"learning_rate": 7.074175824175825e-05,
"loss": 0.4452,
"step": 10860
},
{
"epoch": 2.97,
"grad_norm": 0.9023503661155701,
"learning_rate": 7.071428571428573e-05,
"loss": 0.4248,
"step": 10870
},
{
"epoch": 2.97,
"grad_norm": 0.7970243692398071,
"learning_rate": 7.06868131868132e-05,
"loss": 0.4546,
"step": 10880
},
{
"epoch": 2.97,
"grad_norm": 0.6860815286636353,
"learning_rate": 7.065934065934066e-05,
"loss": 0.4393,
"step": 10890
},
{
"epoch": 2.98,
"grad_norm": 0.6077113747596741,
"learning_rate": 7.063186813186815e-05,
"loss": 0.4438,
"step": 10900
},
{
"epoch": 2.98,
"grad_norm": 0.5479728579521179,
"learning_rate": 7.060439560439561e-05,
"loss": 0.4474,
"step": 10910
},
{
"epoch": 2.98,
"grad_norm": 0.8500062823295593,
"learning_rate": 7.057692307692308e-05,
"loss": 0.4403,
"step": 10920
},
{
"epoch": 2.99,
"grad_norm": 0.8599127531051636,
"learning_rate": 7.054945054945056e-05,
"loss": 0.4367,
"step": 10930
},
{
"epoch": 2.99,
"grad_norm": 1.3294146060943604,
"learning_rate": 7.052197802197803e-05,
"loss": 0.4352,
"step": 10940
},
{
"epoch": 2.99,
"grad_norm": 0.7617499828338623,
"learning_rate": 7.04945054945055e-05,
"loss": 0.4454,
"step": 10950
},
{
"epoch": 2.99,
"grad_norm": 1.2405683994293213,
"learning_rate": 7.046703296703298e-05,
"loss": 0.4513,
"step": 10960
},
{
"epoch": 3.0,
"grad_norm": 0.768068253993988,
"learning_rate": 7.043956043956045e-05,
"loss": 0.4448,
"step": 10970
},
{
"epoch": 3.0,
"grad_norm": 1.0528051853179932,
"learning_rate": 7.041208791208792e-05,
"loss": 0.4436,
"step": 10980
},
{
"epoch": 3.0,
"eval_loss": 0.44340217113494873,
"eval_runtime": 1239.9169,
"eval_samples_per_second": 235.588,
"eval_steps_per_second": 3.682,
"step": 10982
}
],
"logging_steps": 10,
"max_steps": 36600,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 4.683801644117262e+18,
"train_batch_size": 6,
"trial_name": null,
"trial_params": null
}