Llama-2-7b-ultrachat-syn200k-2e / trainer_state.json
kykim0's picture
Model save
177fb19 verified
raw
history blame
26.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.99936,
"eval_steps": 500,
"global_step": 1562,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0128,
"grad_norm": 3.1302534958568193,
"learning_rate": 1.2738853503184715e-06,
"loss": 0.6569,
"step": 10
},
{
"epoch": 0.0256,
"grad_norm": 3.1468711921047627,
"learning_rate": 2.547770700636943e-06,
"loss": 0.5179,
"step": 20
},
{
"epoch": 0.0384,
"grad_norm": 0.5540977306160393,
"learning_rate": 3.821656050955415e-06,
"loss": 0.2369,
"step": 30
},
{
"epoch": 0.0512,
"grad_norm": 0.5002871919834374,
"learning_rate": 5.095541401273886e-06,
"loss": 0.1819,
"step": 40
},
{
"epoch": 0.064,
"grad_norm": 0.7150745925714665,
"learning_rate": 6.369426751592357e-06,
"loss": 0.1759,
"step": 50
},
{
"epoch": 0.0768,
"grad_norm": 0.7619036468756623,
"learning_rate": 7.64331210191083e-06,
"loss": 0.16,
"step": 60
},
{
"epoch": 0.0896,
"grad_norm": 0.44198933481881236,
"learning_rate": 8.9171974522293e-06,
"loss": 0.1608,
"step": 70
},
{
"epoch": 0.1024,
"grad_norm": 0.37478693328542767,
"learning_rate": 1.0191082802547772e-05,
"loss": 0.1617,
"step": 80
},
{
"epoch": 0.1152,
"grad_norm": 0.3696718954770576,
"learning_rate": 1.1464968152866242e-05,
"loss": 0.1596,
"step": 90
},
{
"epoch": 0.128,
"grad_norm": 0.5203159641478062,
"learning_rate": 1.2738853503184714e-05,
"loss": 0.1602,
"step": 100
},
{
"epoch": 0.1408,
"grad_norm": 0.37773491168605033,
"learning_rate": 1.4012738853503186e-05,
"loss": 0.1485,
"step": 110
},
{
"epoch": 0.1536,
"grad_norm": 0.24203685828143712,
"learning_rate": 1.528662420382166e-05,
"loss": 0.1555,
"step": 120
},
{
"epoch": 0.1664,
"grad_norm": 0.3898599912917371,
"learning_rate": 1.6560509554140128e-05,
"loss": 0.1571,
"step": 130
},
{
"epoch": 0.1792,
"grad_norm": 0.22025246679268023,
"learning_rate": 1.78343949044586e-05,
"loss": 0.1559,
"step": 140
},
{
"epoch": 0.192,
"grad_norm": 0.5943517884905586,
"learning_rate": 1.910828025477707e-05,
"loss": 0.1578,
"step": 150
},
{
"epoch": 0.2048,
"grad_norm": 0.4363975905376457,
"learning_rate": 1.999977501271127e-05,
"loss": 0.1468,
"step": 160
},
{
"epoch": 0.2176,
"grad_norm": 0.4874326999671375,
"learning_rate": 1.9995775520317924e-05,
"loss": 0.1519,
"step": 170
},
{
"epoch": 0.2304,
"grad_norm": 0.32242234734849035,
"learning_rate": 1.998677861199748e-05,
"loss": 0.1537,
"step": 180
},
{
"epoch": 0.2432,
"grad_norm": 0.338978065876802,
"learning_rate": 1.9972788785779404e-05,
"loss": 0.1535,
"step": 190
},
{
"epoch": 0.256,
"grad_norm": 0.5405845227789843,
"learning_rate": 1.9953813035916444e-05,
"loss": 0.1552,
"step": 200
},
{
"epoch": 0.2688,
"grad_norm": 0.3187325098825458,
"learning_rate": 1.9929860849387815e-05,
"loss": 0.1441,
"step": 210
},
{
"epoch": 0.2816,
"grad_norm": 0.181656187619335,
"learning_rate": 1.9900944201156164e-05,
"loss": 0.1489,
"step": 220
},
{
"epoch": 0.2944,
"grad_norm": 0.26409308919505864,
"learning_rate": 1.986707754818064e-05,
"loss": 0.1509,
"step": 230
},
{
"epoch": 0.3072,
"grad_norm": 0.23639855452929004,
"learning_rate": 1.982827782218912e-05,
"loss": 0.1525,
"step": 240
},
{
"epoch": 0.32,
"grad_norm": 0.42132507371171557,
"learning_rate": 1.9784564421213122e-05,
"loss": 0.1513,
"step": 250
},
{
"epoch": 0.3328,
"grad_norm": 0.30911958261458256,
"learning_rate": 1.9735959199889723e-05,
"loss": 0.1415,
"step": 260
},
{
"epoch": 0.3456,
"grad_norm": 0.286475064941514,
"learning_rate": 1.968248645853526e-05,
"loss": 0.1483,
"step": 270
},
{
"epoch": 0.3584,
"grad_norm": 0.3223686350575121,
"learning_rate": 1.9624172930996322e-05,
"loss": 0.1496,
"step": 280
},
{
"epoch": 0.3712,
"grad_norm": 0.33039144824318656,
"learning_rate": 1.956104777128409e-05,
"loss": 0.1496,
"step": 290
},
{
"epoch": 0.384,
"grad_norm": 0.3268436732392019,
"learning_rate": 1.9493142538998713e-05,
"loss": 0.1524,
"step": 300
},
{
"epoch": 0.3968,
"grad_norm": 0.3768976535567336,
"learning_rate": 1.9420491183550983e-05,
"loss": 0.1397,
"step": 310
},
{
"epoch": 0.4096,
"grad_norm": 0.344679435905093,
"learning_rate": 1.934313002718924e-05,
"loss": 0.1459,
"step": 320
},
{
"epoch": 0.4224,
"grad_norm": 0.1867877479920934,
"learning_rate": 1.9261097746839974e-05,
"loss": 0.149,
"step": 330
},
{
"epoch": 0.4352,
"grad_norm": 0.22826415693304572,
"learning_rate": 1.9174435354771167e-05,
"loss": 0.1515,
"step": 340
},
{
"epoch": 0.448,
"grad_norm": 0.2574654153325295,
"learning_rate": 1.9083186178088103e-05,
"loss": 0.1501,
"step": 350
},
{
"epoch": 0.4608,
"grad_norm": 0.23070296857068737,
"learning_rate": 1.898739583707187e-05,
"loss": 0.1383,
"step": 360
},
{
"epoch": 0.4736,
"grad_norm": 0.21129982050815863,
"learning_rate": 1.8887112222371363e-05,
"loss": 0.1437,
"step": 370
},
{
"epoch": 0.4864,
"grad_norm": 0.17026221614799503,
"learning_rate": 1.8782385471060217e-05,
"loss": 0.1495,
"step": 380
},
{
"epoch": 0.4992,
"grad_norm": 0.2103435712105906,
"learning_rate": 1.8673267941570646e-05,
"loss": 0.1495,
"step": 390
},
{
"epoch": 0.512,
"grad_norm": 0.2949165035198057,
"learning_rate": 1.8559814187516692e-05,
"loss": 0.1482,
"step": 400
},
{
"epoch": 0.5248,
"grad_norm": 0.2727792628609804,
"learning_rate": 1.844208093042e-05,
"loss": 0.1385,
"step": 410
},
{
"epoch": 0.5376,
"grad_norm": 0.19011236871362203,
"learning_rate": 1.8320127031351723e-05,
"loss": 0.1447,
"step": 420
},
{
"epoch": 0.5504,
"grad_norm": 0.20224724252134554,
"learning_rate": 1.8194013461504774e-05,
"loss": 0.1457,
"step": 430
},
{
"epoch": 0.5632,
"grad_norm": 0.21758912206027958,
"learning_rate": 1.806380327171111e-05,
"loss": 0.146,
"step": 440
},
{
"epoch": 0.576,
"grad_norm": 0.394244725563091,
"learning_rate": 1.792956156091928e-05,
"loss": 0.1473,
"step": 450
},
{
"epoch": 0.5888,
"grad_norm": 0.3292629665585757,
"learning_rate": 1.7791355443648045e-05,
"loss": 0.1378,
"step": 460
},
{
"epoch": 0.6016,
"grad_norm": 0.19521004688992183,
"learning_rate": 1.7649254016432247e-05,
"loss": 0.1456,
"step": 470
},
{
"epoch": 0.6144,
"grad_norm": 0.2316602793370239,
"learning_rate": 1.750332832327786e-05,
"loss": 0.1451,
"step": 480
},
{
"epoch": 0.6272,
"grad_norm": 0.17573541945620594,
"learning_rate": 1.735365132014329e-05,
"loss": 0.1486,
"step": 490
},
{
"epoch": 0.64,
"grad_norm": 0.3527896812147239,
"learning_rate": 1.7200297838464864e-05,
"loss": 0.1488,
"step": 500
},
{
"epoch": 0.6528,
"grad_norm": 0.21609855998849087,
"learning_rate": 1.7043344547744637e-05,
"loss": 0.138,
"step": 510
},
{
"epoch": 0.6656,
"grad_norm": 0.20743287144243716,
"learning_rate": 1.6882869917219265e-05,
"loss": 0.1432,
"step": 520
},
{
"epoch": 0.6784,
"grad_norm": 0.14306391155148201,
"learning_rate": 1.6718954176629088e-05,
"loss": 0.1458,
"step": 530
},
{
"epoch": 0.6912,
"grad_norm": 0.13082293792565916,
"learning_rate": 1.6551679276107047e-05,
"loss": 0.1468,
"step": 540
},
{
"epoch": 0.704,
"grad_norm": 0.36161924084937497,
"learning_rate": 1.638112884520748e-05,
"loss": 0.1469,
"step": 550
},
{
"epoch": 0.7168,
"grad_norm": 0.28617618689753893,
"learning_rate": 1.620738815109531e-05,
"loss": 0.1371,
"step": 560
},
{
"epoch": 0.7296,
"grad_norm": 0.13750653322717393,
"learning_rate": 1.6030544055916462e-05,
"loss": 0.1423,
"step": 570
},
{
"epoch": 0.7424,
"grad_norm": 0.15234113342700173,
"learning_rate": 1.5850684973370913e-05,
"loss": 0.1473,
"step": 580
},
{
"epoch": 0.7552,
"grad_norm": 0.1674605203651358,
"learning_rate": 1.5667900824510005e-05,
"loss": 0.1465,
"step": 590
},
{
"epoch": 0.768,
"grad_norm": 0.30120726166690853,
"learning_rate": 1.5482282992780155e-05,
"loss": 0.1459,
"step": 600
},
{
"epoch": 0.7808,
"grad_norm": 0.2556669288569797,
"learning_rate": 1.5293924278335475e-05,
"loss": 0.1372,
"step": 610
},
{
"epoch": 0.7936,
"grad_norm": 0.19878411049962907,
"learning_rate": 1.5102918851642035e-05,
"loss": 0.1416,
"step": 620
},
{
"epoch": 0.8064,
"grad_norm": 0.14869095864210355,
"learning_rate": 1.4909362206397114e-05,
"loss": 0.1435,
"step": 630
},
{
"epoch": 0.8192,
"grad_norm": 0.21996626003622385,
"learning_rate": 1.4713351111786824e-05,
"loss": 0.1462,
"step": 640
},
{
"epoch": 0.832,
"grad_norm": 0.29224476727899246,
"learning_rate": 1.4514983564106117e-05,
"loss": 0.1449,
"step": 650
},
{
"epoch": 0.8448,
"grad_norm": 0.2688310061351837,
"learning_rate": 1.431435873776524e-05,
"loss": 0.1374,
"step": 660
},
{
"epoch": 0.8576,
"grad_norm": 0.20038216433613684,
"learning_rate": 1.4111576935707214e-05,
"loss": 0.1426,
"step": 670
},
{
"epoch": 0.8704,
"grad_norm": 0.13687808263034185,
"learning_rate": 1.3906739539261115e-05,
"loss": 0.1457,
"step": 680
},
{
"epoch": 0.8832,
"grad_norm": 0.19032971439937477,
"learning_rate": 1.3699948957456176e-05,
"loss": 0.1455,
"step": 690
},
{
"epoch": 0.896,
"grad_norm": 0.3208508208806956,
"learning_rate": 1.3491308575822111e-05,
"loss": 0.1435,
"step": 700
},
{
"epoch": 0.9088,
"grad_norm": 0.19524278584922394,
"learning_rate": 1.3280922704701231e-05,
"loss": 0.1365,
"step": 710
},
{
"epoch": 0.9216,
"grad_norm": 0.17798035245935892,
"learning_rate": 1.3068896527098205e-05,
"loss": 0.1433,
"step": 720
},
{
"epoch": 0.9344,
"grad_norm": 0.17884600087369845,
"learning_rate": 1.2855336046093513e-05,
"loss": 0.1447,
"step": 730
},
{
"epoch": 0.9472,
"grad_norm": 0.14100954677816485,
"learning_rate": 1.2640348031846897e-05,
"loss": 0.1447,
"step": 740
},
{
"epoch": 0.96,
"grad_norm": 0.33304512484187876,
"learning_rate": 1.2424039968217361e-05,
"loss": 0.145,
"step": 750
},
{
"epoch": 0.9728,
"grad_norm": 0.2096407271139896,
"learning_rate": 1.220651999902627e-05,
"loss": 0.1355,
"step": 760
},
{
"epoch": 0.9856,
"grad_norm": 0.18948626555533854,
"learning_rate": 1.1987896873990572e-05,
"loss": 0.1422,
"step": 770
},
{
"epoch": 0.9984,
"grad_norm": 0.25576476392968256,
"learning_rate": 1.176827989435307e-05,
"loss": 0.1473,
"step": 780
},
{
"epoch": 0.99968,
"eval_loss": 0.13926610350608826,
"eval_runtime": 42.0798,
"eval_samples_per_second": 23.764,
"eval_steps_per_second": 1.497,
"step": 781
},
{
"epoch": 1.0112,
"grad_norm": 0.15072347755987592,
"learning_rate": 1.1547778858236938e-05,
"loss": 0.1361,
"step": 790
},
{
"epoch": 1.024,
"grad_norm": 0.24763730684547336,
"learning_rate": 1.1326504005751875e-05,
"loss": 0.1426,
"step": 800
},
{
"epoch": 1.0368,
"grad_norm": 0.2049805600857785,
"learning_rate": 1.1104565963879233e-05,
"loss": 0.1417,
"step": 810
},
{
"epoch": 1.0496,
"grad_norm": 0.17505933336916096,
"learning_rate": 1.0882075691163749e-05,
"loss": 0.1425,
"step": 820
},
{
"epoch": 1.0624,
"grad_norm": 0.2367359754977405,
"learning_rate": 1.065914442223952e-05,
"loss": 0.1395,
"step": 830
},
{
"epoch": 1.0752,
"grad_norm": 0.2168168491904799,
"learning_rate": 1.0435883612217928e-05,
"loss": 0.1368,
"step": 840
},
{
"epoch": 1.088,
"grad_norm": 0.16667141395128426,
"learning_rate": 1.0212404880965348e-05,
"loss": 0.1397,
"step": 850
},
{
"epoch": 1.1008,
"grad_norm": 0.19793016123521015,
"learning_rate": 9.988819957298487e-06,
"loss": 0.1447,
"step": 860
},
{
"epoch": 1.1136,
"grad_norm": 0.21711053823618134,
"learning_rate": 9.765240623125238e-06,
"loss": 0.1418,
"step": 870
},
{
"epoch": 1.1264,
"grad_norm": 0.24562717005747997,
"learning_rate": 9.541778657559026e-06,
"loss": 0.1371,
"step": 880
},
{
"epoch": 1.1392,
"grad_norm": 0.17839427327356683,
"learning_rate": 9.318545781034499e-06,
"loss": 0.1368,
"step": 890
},
{
"epoch": 1.152,
"grad_norm": 0.18971426870401556,
"learning_rate": 9.095653599452604e-06,
"loss": 0.1411,
"step": 900
},
{
"epoch": 1.1648,
"grad_norm": 0.2028582361588426,
"learning_rate": 8.873213548382895e-06,
"loss": 0.1432,
"step": 910
},
{
"epoch": 1.1776,
"grad_norm": 0.1910575100313919,
"learning_rate": 8.651336837351013e-06,
"loss": 0.1441,
"step": 920
},
{
"epoch": 1.1904,
"grad_norm": 0.21795820010528985,
"learning_rate": 8.430134394239182e-06,
"loss": 0.1389,
"step": 930
},
{
"epoch": 1.2032,
"grad_norm": 0.1702846327798786,
"learning_rate": 8.209716809827475e-06,
"loss": 0.1387,
"step": 940
},
{
"epoch": 1.216,
"grad_norm": 0.19574705900720263,
"learning_rate": 7.990194282503678e-06,
"loss": 0.1408,
"step": 950
},
{
"epoch": 1.2288000000000001,
"grad_norm": 0.15444562246412258,
"learning_rate": 7.771676563169276e-06,
"loss": 0.1424,
"step": 960
},
{
"epoch": 1.2416,
"grad_norm": 0.27192054451029335,
"learning_rate": 7.554272900369184e-06,
"loss": 0.1431,
"step": 970
},
{
"epoch": 1.2544,
"grad_norm": 0.28293536797942354,
"learning_rate": 7.338091985672651e-06,
"loss": 0.1375,
"step": 980
},
{
"epoch": 1.2671999999999999,
"grad_norm": 0.27054236168586154,
"learning_rate": 7.1232418993325836e-06,
"loss": 0.1346,
"step": 990
},
{
"epoch": 1.28,
"grad_norm": 0.25032137575039304,
"learning_rate": 6.909830056250527e-06,
"loss": 0.1401,
"step": 1000
},
{
"epoch": 1.2928,
"grad_norm": 0.1593137714997539,
"learning_rate": 6.697963152274294e-06,
"loss": 0.1415,
"step": 1010
},
{
"epoch": 1.3056,
"grad_norm": 0.17468257527642406,
"learning_rate": 6.487747110855061e-06,
"loss": 0.1421,
"step": 1020
},
{
"epoch": 1.3184,
"grad_norm": 0.198558265817394,
"learning_rate": 6.279287030090664e-06,
"loss": 0.1387,
"step": 1030
},
{
"epoch": 1.3312,
"grad_norm": 0.11504815520373578,
"learning_rate": 6.072687130181493e-06,
"loss": 0.1353,
"step": 1040
},
{
"epoch": 1.3439999999999999,
"grad_norm": 0.13858890875298494,
"learning_rate": 5.868050701325314e-06,
"loss": 0.1398,
"step": 1050
},
{
"epoch": 1.3568,
"grad_norm": 0.13445081822803,
"learning_rate": 5.665480052077051e-06,
"loss": 0.1409,
"step": 1060
},
{
"epoch": 1.3696,
"grad_norm": 0.16055599811743235,
"learning_rate": 5.465076458199332e-06,
"loss": 0.1417,
"step": 1070
},
{
"epoch": 1.3824,
"grad_norm": 0.19169771082374196,
"learning_rate": 5.266940112029406e-06,
"loss": 0.138,
"step": 1080
},
{
"epoch": 1.3952,
"grad_norm": 0.18154527027617387,
"learning_rate": 5.071170072387672e-06,
"loss": 0.1373,
"step": 1090
},
{
"epoch": 1.408,
"grad_norm": 0.13523001447293553,
"learning_rate": 4.877864215052969e-06,
"loss": 0.1382,
"step": 1100
},
{
"epoch": 1.4208,
"grad_norm": 0.14620192480202174,
"learning_rate": 4.687119183829305e-06,
"loss": 0.1403,
"step": 1110
},
{
"epoch": 1.4336,
"grad_norm": 0.14563804992816887,
"learning_rate": 4.4990303422285005e-06,
"loss": 0.1401,
"step": 1120
},
{
"epoch": 1.4464000000000001,
"grad_norm": 0.18048671111150355,
"learning_rate": 4.31369172579296e-06,
"loss": 0.1351,
"step": 1130
},
{
"epoch": 1.4592,
"grad_norm": 0.1274956008805008,
"learning_rate": 4.1311959950823125e-06,
"loss": 0.1357,
"step": 1140
},
{
"epoch": 1.472,
"grad_norm": 0.15186812567018518,
"learning_rate": 3.951634389347522e-06,
"loss": 0.1405,
"step": 1150
},
{
"epoch": 1.4848,
"grad_norm": 0.1743781587386354,
"learning_rate": 3.7750966809155633e-06,
"loss": 0.1411,
"step": 1160
},
{
"epoch": 1.4976,
"grad_norm": 0.16145108253749546,
"learning_rate": 3.6016711303075003e-06,
"loss": 0.1418,
"step": 1170
},
{
"epoch": 1.5104,
"grad_norm": 0.2443925847158833,
"learning_rate": 3.431444442112395e-06,
"loss": 0.1365,
"step": 1180
},
{
"epoch": 1.5232,
"grad_norm": 0.16076378293495353,
"learning_rate": 3.264501721639086e-06,
"loss": 0.1354,
"step": 1190
},
{
"epoch": 1.536,
"grad_norm": 0.20722028172668885,
"learning_rate": 3.1009264323675737e-06,
"loss": 0.1387,
"step": 1200
},
{
"epoch": 1.5488,
"grad_norm": 0.2367805090756499,
"learning_rate": 2.940800354221205e-06,
"loss": 0.1414,
"step": 1210
},
{
"epoch": 1.5615999999999999,
"grad_norm": 0.2045021555576582,
"learning_rate": 2.7842035426805858e-06,
"loss": 0.1435,
"step": 1220
},
{
"epoch": 1.5744,
"grad_norm": 0.3378391898032357,
"learning_rate": 2.6312142887596116e-06,
"loss": 0.137,
"step": 1230
},
{
"epoch": 1.5872000000000002,
"grad_norm": 0.15144211781786535,
"learning_rate": 2.4819090798636504e-06,
"loss": 0.1362,
"step": 1240
},
{
"epoch": 1.6,
"grad_norm": 0.19705372401646207,
"learning_rate": 2.336362561549459e-06,
"loss": 0.1398,
"step": 1250
},
{
"epoch": 1.6128,
"grad_norm": 0.1981712638313628,
"learning_rate": 2.194647500205925e-06,
"loss": 0.1412,
"step": 1260
},
{
"epoch": 1.6256,
"grad_norm": 0.22583463728566275,
"learning_rate": 2.056834746674313e-06,
"loss": 0.1402,
"step": 1270
},
{
"epoch": 1.6383999999999999,
"grad_norm": 0.3166766019266127,
"learning_rate": 1.9229932008261756e-06,
"loss": 0.1366,
"step": 1280
},
{
"epoch": 1.6512,
"grad_norm": 0.13358881231561032,
"learning_rate": 1.7931897771166861e-06,
"loss": 0.1346,
"step": 1290
},
{
"epoch": 1.6640000000000001,
"grad_norm": 0.28514268179697894,
"learning_rate": 1.6674893711305528e-06,
"loss": 0.1399,
"step": 1300
},
{
"epoch": 1.6768,
"grad_norm": 0.08846971622020448,
"learning_rate": 1.5459548271373005e-06,
"loss": 0.1383,
"step": 1310
},
{
"epoch": 1.6896,
"grad_norm": 0.28809034512805015,
"learning_rate": 1.4286469066721055e-06,
"loss": 0.1399,
"step": 1320
},
{
"epoch": 1.7024,
"grad_norm": 0.1598977533627517,
"learning_rate": 1.3156242581578815e-06,
"loss": 0.1351,
"step": 1330
},
{
"epoch": 1.7151999999999998,
"grad_norm": 0.19800584875933322,
"learning_rate": 1.2069433875838498e-06,
"loss": 0.1347,
"step": 1340
},
{
"epoch": 1.728,
"grad_norm": 0.12495858993727499,
"learning_rate": 1.10265863025521e-06,
"loss": 0.1376,
"step": 1350
},
{
"epoch": 1.7408000000000001,
"grad_norm": 0.12691325643260487,
"learning_rate": 1.0028221236280445e-06,
"loss": 0.1408,
"step": 1360
},
{
"epoch": 1.7536,
"grad_norm": 0.14092289840202932,
"learning_rate": 9.074837812430626e-07,
"loss": 0.1404,
"step": 1370
},
{
"epoch": 1.7664,
"grad_norm": 0.1356179011766789,
"learning_rate": 8.166912677711769e-07,
"loss": 0.1344,
"step": 1380
},
{
"epoch": 1.7792,
"grad_norm": 0.11072101415903435,
"learning_rate": 7.304899751834194e-07,
"loss": 0.1363,
"step": 1390
},
{
"epoch": 1.792,
"grad_norm": 0.09364893615869158,
"learning_rate": 6.489230000571067e-07,
"loss": 0.1387,
"step": 1400
},
{
"epoch": 1.8048,
"grad_norm": 0.09719389776054362,
"learning_rate": 5.720311220295727e-07,
"loss": 0.1406,
"step": 1410
},
{
"epoch": 1.8176,
"grad_norm": 0.14149128093688274,
"learning_rate": 4.998527834102873e-07,
"loss": 0.1411,
"step": 1420
},
{
"epoch": 1.8304,
"grad_norm": 0.15889816186833727,
"learning_rate": 4.3242406996150674e-07,
"loss": 0.1345,
"step": 1430
},
{
"epoch": 1.8432,
"grad_norm": 0.0958974593538415,
"learning_rate": 3.697786928571034e-07,
"loss": 0.1341,
"step": 1440
},
{
"epoch": 1.8559999999999999,
"grad_norm": 0.07979518065313038,
"learning_rate": 3.119479718285601e-07,
"loss": 0.1381,
"step": 1450
},
{
"epoch": 1.8688,
"grad_norm": 0.10204011708199633,
"learning_rate": 2.5896081950659737e-07,
"loss": 0.1408,
"step": 1460
},
{
"epoch": 1.8816000000000002,
"grad_norm": 0.11627971006018599,
"learning_rate": 2.108437269662089e-07,
"loss": 0.1433,
"step": 1470
},
{
"epoch": 1.8944,
"grad_norm": 0.15458392923737632,
"learning_rate": 1.6762075048238347e-07,
"loss": 0.1359,
"step": 1480
},
{
"epoch": 1.9072,
"grad_norm": 0.10450160701417593,
"learning_rate": 1.2931349950309535e-07,
"loss": 0.1342,
"step": 1490
},
{
"epoch": 1.92,
"grad_norm": 0.09770183301391486,
"learning_rate": 9.594112584561e-08,
"loss": 0.1381,
"step": 1500
},
{
"epoch": 1.9327999999999999,
"grad_norm": 0.09381477173800809,
"learning_rate": 6.752031412147086e-08,
"loss": 0.1394,
"step": 1510
},
{
"epoch": 1.9456,
"grad_norm": 0.12023221514437896,
"learning_rate": 4.406527339498423e-08,
"loss": 0.1416,
"step": 1520
},
{
"epoch": 1.9584000000000001,
"grad_norm": 0.16776774231570918,
"learning_rate": 2.5587730079362772e-08,
"loss": 0.134,
"step": 1530
},
{
"epoch": 1.9712,
"grad_norm": 0.09797271568610905,
"learning_rate": 1.2096922074060768e-08,
"loss": 0.134,
"step": 1540
},
{
"epoch": 1.984,
"grad_norm": 0.09345305263894858,
"learning_rate": 3.5995941462718677e-09,
"loss": 0.138,
"step": 1550
},
{
"epoch": 1.9968,
"grad_norm": 0.08633576245142834,
"learning_rate": 9.999455885978393e-11,
"loss": 0.1406,
"step": 1560
},
{
"epoch": 1.99936,
"eval_loss": 0.13485482335090637,
"eval_runtime": 42.0832,
"eval_samples_per_second": 23.762,
"eval_steps_per_second": 1.497,
"step": 1562
},
{
"epoch": 1.99936,
"step": 1562,
"total_flos": 534097819484160.0,
"train_loss": 0.14969347108265196,
"train_runtime": 45991.4783,
"train_samples_per_second": 4.349,
"train_steps_per_second": 0.034
}
],
"logging_steps": 10,
"max_steps": 1562,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 781,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 534097819484160.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}