kevin-token-compressor / trainer_state.json
aoxo's picture
Upload 11 files
1afabdd verified
raw
history blame
18.2 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.02284174099749883,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002284174099749883,
"grad_norm": 0.6011635661125183,
"learning_rate": 0.00019995065603657316,
"loss": 2.2538,
"step": 10
},
{
"epoch": 0.0004568348199499766,
"grad_norm": 0.8175456523895264,
"learning_rate": 0.00019980267284282717,
"loss": 2.1126,
"step": 20
},
{
"epoch": 0.0006852522299249648,
"grad_norm": 1.1491563320159912,
"learning_rate": 0.00019955619646030802,
"loss": 2.228,
"step": 30
},
{
"epoch": 0.0009136696398999532,
"grad_norm": 1.0827499628067017,
"learning_rate": 0.0001992114701314478,
"loss": 2.1703,
"step": 40
},
{
"epoch": 0.0011420870498749414,
"grad_norm": 1.157465934753418,
"learning_rate": 0.00019876883405951377,
"loss": 2.0955,
"step": 50
},
{
"epoch": 0.0013705044598499297,
"grad_norm": 1.0857897996902466,
"learning_rate": 0.0001982287250728689,
"loss": 2.1769,
"step": 60
},
{
"epoch": 0.0015989218698249181,
"grad_norm": 1.377300500869751,
"learning_rate": 0.00019759167619387476,
"loss": 2.1335,
"step": 70
},
{
"epoch": 0.0018273392797999064,
"grad_norm": 0.8308928608894348,
"learning_rate": 0.0001968583161128631,
"loss": 2.0746,
"step": 80
},
{
"epoch": 0.0020557566897748945,
"grad_norm": 1.1545639038085938,
"learning_rate": 0.0001960293685676943,
"loss": 2.3118,
"step": 90
},
{
"epoch": 0.0022841740997498828,
"grad_norm": 0.9956186413764954,
"learning_rate": 0.00019510565162951537,
"loss": 2.1476,
"step": 100
},
{
"epoch": 0.002512591509724871,
"grad_norm": 1.1778929233551025,
"learning_rate": 0.00019408807689542257,
"loss": 1.9218,
"step": 110
},
{
"epoch": 0.0027410089196998593,
"grad_norm": 1.2166578769683838,
"learning_rate": 0.00019297764858882514,
"loss": 2.1764,
"step": 120
},
{
"epoch": 0.002969426329674848,
"grad_norm": 0.9831048846244812,
"learning_rate": 0.00019177546256839812,
"loss": 2.1327,
"step": 130
},
{
"epoch": 0.0031978437396498363,
"grad_norm": 0.9150891900062561,
"learning_rate": 0.00019048270524660196,
"loss": 2.0719,
"step": 140
},
{
"epoch": 0.0034262611496248246,
"grad_norm": 1.078007459640503,
"learning_rate": 0.0001891006524188368,
"loss": 2.1048,
"step": 150
},
{
"epoch": 0.003654678559599813,
"grad_norm": 0.9422916173934937,
"learning_rate": 0.00018763066800438636,
"loss": 2.0419,
"step": 160
},
{
"epoch": 0.003883095969574801,
"grad_norm": 1.0269567966461182,
"learning_rate": 0.0001860742027003944,
"loss": 2.0837,
"step": 170
},
{
"epoch": 0.004111513379549789,
"grad_norm": 0.9426142573356628,
"learning_rate": 0.00018443279255020152,
"loss": 2.0422,
"step": 180
},
{
"epoch": 0.004339930789524778,
"grad_norm": 1.018721103668213,
"learning_rate": 0.00018270805742745617,
"loss": 2.1384,
"step": 190
},
{
"epoch": 0.0045683481994997655,
"grad_norm": 1.1313502788543701,
"learning_rate": 0.00018090169943749476,
"loss": 2.1318,
"step": 200
},
{
"epoch": 0.004796765609474754,
"grad_norm": 0.6934864521026611,
"learning_rate": 0.00017901550123756906,
"loss": 2.2722,
"step": 210
},
{
"epoch": 0.005025183019449742,
"grad_norm": 0.8822490572929382,
"learning_rate": 0.00017705132427757895,
"loss": 2.0774,
"step": 220
},
{
"epoch": 0.005253600429424731,
"grad_norm": 1.1275163888931274,
"learning_rate": 0.00017501110696304596,
"loss": 2.1253,
"step": 230
},
{
"epoch": 0.005482017839399719,
"grad_norm": 1.0035189390182495,
"learning_rate": 0.00017289686274214118,
"loss": 2.1624,
"step": 240
},
{
"epoch": 0.005710435249374707,
"grad_norm": 0.8009467124938965,
"learning_rate": 0.00017071067811865476,
"loss": 2.0543,
"step": 250
},
{
"epoch": 0.005938852659349696,
"grad_norm": 0.9582347869873047,
"learning_rate": 0.00016845471059286887,
"loss": 2.1542,
"step": 260
},
{
"epoch": 0.006167270069324684,
"grad_norm": 1.0823897123336792,
"learning_rate": 0.00016613118653236518,
"loss": 2.3268,
"step": 270
},
{
"epoch": 0.006395687479299673,
"grad_norm": 0.9078360795974731,
"learning_rate": 0.000163742398974869,
"loss": 2.0749,
"step": 280
},
{
"epoch": 0.0066241048892746604,
"grad_norm": 1.0934821367263794,
"learning_rate": 0.00016129070536529766,
"loss": 2.2403,
"step": 290
},
{
"epoch": 0.006852522299249649,
"grad_norm": 1.464321494102478,
"learning_rate": 0.00015877852522924732,
"loss": 2.0575,
"step": 300
},
{
"epoch": 0.007080939709224637,
"grad_norm": 0.9363439679145813,
"learning_rate": 0.00015620833778521307,
"loss": 2.3062,
"step": 310
},
{
"epoch": 0.007309357119199626,
"grad_norm": 1.1936748027801514,
"learning_rate": 0.00015358267949789966,
"loss": 2.081,
"step": 320
},
{
"epoch": 0.0075377745291746135,
"grad_norm": 0.8682211637496948,
"learning_rate": 0.00015090414157503714,
"loss": 2.0289,
"step": 330
},
{
"epoch": 0.007766191939149602,
"grad_norm": 1.123822569847107,
"learning_rate": 0.00014817536741017152,
"loss": 2.2525,
"step": 340
},
{
"epoch": 0.00799460934912459,
"grad_norm": 0.8017176389694214,
"learning_rate": 0.00014539904997395468,
"loss": 2.213,
"step": 350
},
{
"epoch": 0.008223026759099578,
"grad_norm": 0.8773420453071594,
"learning_rate": 0.00014257792915650728,
"loss": 2.0703,
"step": 360
},
{
"epoch": 0.008451444169074568,
"grad_norm": 1.1905558109283447,
"learning_rate": 0.00013971478906347806,
"loss": 2.1984,
"step": 370
},
{
"epoch": 0.008679861579049555,
"grad_norm": 0.8345734477043152,
"learning_rate": 0.00013681245526846783,
"loss": 2.0921,
"step": 380
},
{
"epoch": 0.008908278989024543,
"grad_norm": 1.126680612564087,
"learning_rate": 0.00013387379202452917,
"loss": 2.0888,
"step": 390
},
{
"epoch": 0.009136696398999531,
"grad_norm": 1.0113277435302734,
"learning_rate": 0.00013090169943749476,
"loss": 2.0344,
"step": 400
},
{
"epoch": 0.00936511380897452,
"grad_norm": 1.6838539838790894,
"learning_rate": 0.00012789911060392294,
"loss": 2.0585,
"step": 410
},
{
"epoch": 0.009593531218949508,
"grad_norm": 0.9599931240081787,
"learning_rate": 0.0001248689887164855,
"loss": 2.1937,
"step": 420
},
{
"epoch": 0.009821948628924496,
"grad_norm": 1.091261386871338,
"learning_rate": 0.00012181432413965428,
"loss": 2.2075,
"step": 430
},
{
"epoch": 0.010050366038899484,
"grad_norm": 1.618888020515442,
"learning_rate": 0.00011873813145857249,
"loss": 2.0918,
"step": 440
},
{
"epoch": 0.010278783448874474,
"grad_norm": 0.8518305420875549,
"learning_rate": 0.0001156434465040231,
"loss": 2.0674,
"step": 450
},
{
"epoch": 0.010507200858849462,
"grad_norm": 0.823520839214325,
"learning_rate": 0.00011253332335643043,
"loss": 2.118,
"step": 460
},
{
"epoch": 0.01073561826882445,
"grad_norm": 1.0874645709991455,
"learning_rate": 0.00010941083133185146,
"loss": 2.1401,
"step": 470
},
{
"epoch": 0.010964035678799437,
"grad_norm": 1.0509600639343262,
"learning_rate": 0.00010627905195293135,
"loss": 2.2394,
"step": 480
},
{
"epoch": 0.011192453088774427,
"grad_norm": 0.8310794234275818,
"learning_rate": 0.00010314107590781284,
"loss": 2.09,
"step": 490
},
{
"epoch": 0.011420870498749415,
"grad_norm": 0.963942289352417,
"learning_rate": 0.0001,
"loss": 2.0202,
"step": 500
},
{
"epoch": 0.011649287908724403,
"grad_norm": 0.9792200326919556,
"learning_rate": 9.685892409218717e-05,
"loss": 2.2468,
"step": 510
},
{
"epoch": 0.011877705318699392,
"grad_norm": 0.8218494653701782,
"learning_rate": 9.372094804706867e-05,
"loss": 2.3187,
"step": 520
},
{
"epoch": 0.01210612272867438,
"grad_norm": 1.120417833328247,
"learning_rate": 9.058916866814858e-05,
"loss": 2.1024,
"step": 530
},
{
"epoch": 0.012334540138649368,
"grad_norm": 1.0600074529647827,
"learning_rate": 8.746667664356956e-05,
"loss": 2.1056,
"step": 540
},
{
"epoch": 0.012562957548624356,
"grad_norm": 1.0849493741989136,
"learning_rate": 8.435655349597689e-05,
"loss": 2.0486,
"step": 550
},
{
"epoch": 0.012791374958599345,
"grad_norm": 0.9597050547599792,
"learning_rate": 8.126186854142752e-05,
"loss": 2.1524,
"step": 560
},
{
"epoch": 0.013019792368574333,
"grad_norm": 0.6964913010597229,
"learning_rate": 7.818567586034577e-05,
"loss": 2.5253,
"step": 570
},
{
"epoch": 0.013248209778549321,
"grad_norm": 0.8805552124977112,
"learning_rate": 7.513101128351454e-05,
"loss": 1.9732,
"step": 580
},
{
"epoch": 0.013476627188524309,
"grad_norm": 0.8434686660766602,
"learning_rate": 7.210088939607708e-05,
"loss": 2.1449,
"step": 590
},
{
"epoch": 0.013705044598499298,
"grad_norm": 0.7308034896850586,
"learning_rate": 6.909830056250527e-05,
"loss": 1.9493,
"step": 600
},
{
"epoch": 0.013933462008474286,
"grad_norm": 0.7747200131416321,
"learning_rate": 6.612620797547087e-05,
"loss": 2.4289,
"step": 610
},
{
"epoch": 0.014161879418449274,
"grad_norm": 0.9479460716247559,
"learning_rate": 6.318754473153221e-05,
"loss": 2.0873,
"step": 620
},
{
"epoch": 0.014390296828424262,
"grad_norm": 0.8679899573326111,
"learning_rate": 6.0285210936521955e-05,
"loss": 2.1007,
"step": 630
},
{
"epoch": 0.014618714238399251,
"grad_norm": 1.1402413845062256,
"learning_rate": 5.7422070843492734e-05,
"loss": 1.9974,
"step": 640
},
{
"epoch": 0.01484713164837424,
"grad_norm": 0.944284975528717,
"learning_rate": 5.4600950026045326e-05,
"loss": 2.1228,
"step": 650
},
{
"epoch": 0.015075549058349227,
"grad_norm": 1.2122584581375122,
"learning_rate": 5.182463258982846e-05,
"loss": 2.1118,
"step": 660
},
{
"epoch": 0.015303966468324215,
"grad_norm": 0.8726314306259155,
"learning_rate": 4.909585842496287e-05,
"loss": 2.0166,
"step": 670
},
{
"epoch": 0.015532383878299204,
"grad_norm": 0.8935607075691223,
"learning_rate": 4.6417320502100316e-05,
"loss": 2.0992,
"step": 680
},
{
"epoch": 0.01576080128827419,
"grad_norm": 0.7619888186454773,
"learning_rate": 4.379166221478697e-05,
"loss": 2.1272,
"step": 690
},
{
"epoch": 0.01598921869824918,
"grad_norm": 0.9617013931274414,
"learning_rate": 4.12214747707527e-05,
"loss": 2.0397,
"step": 700
},
{
"epoch": 0.01621763610822417,
"grad_norm": 0.7471845746040344,
"learning_rate": 3.8709294634702376e-05,
"loss": 2.1752,
"step": 710
},
{
"epoch": 0.016446053518199156,
"grad_norm": 0.8926049470901489,
"learning_rate": 3.6257601025131026e-05,
"loss": 2.0067,
"step": 720
},
{
"epoch": 0.016674470928174145,
"grad_norm": 0.885032594203949,
"learning_rate": 3.386881346763483e-05,
"loss": 2.0448,
"step": 730
},
{
"epoch": 0.016902888338149135,
"grad_norm": 1.1092376708984375,
"learning_rate": 3.154528940713113e-05,
"loss": 2.1073,
"step": 740
},
{
"epoch": 0.01713130574812412,
"grad_norm": 1.155680537223816,
"learning_rate": 2.9289321881345254e-05,
"loss": 1.978,
"step": 750
},
{
"epoch": 0.01735972315809911,
"grad_norm": 0.829304575920105,
"learning_rate": 2.7103137257858868e-05,
"loss": 1.9944,
"step": 760
},
{
"epoch": 0.0175881405680741,
"grad_norm": 0.8414214849472046,
"learning_rate": 2.4988893036954043e-05,
"loss": 2.5413,
"step": 770
},
{
"epoch": 0.017816557978049086,
"grad_norm": 0.9645959138870239,
"learning_rate": 2.2948675722421086e-05,
"loss": 2.2689,
"step": 780
},
{
"epoch": 0.018044975388024076,
"grad_norm": 0.8575518727302551,
"learning_rate": 2.098449876243096e-05,
"loss": 2.1654,
"step": 790
},
{
"epoch": 0.018273392797999062,
"grad_norm": 0.8750948309898376,
"learning_rate": 1.9098300562505266e-05,
"loss": 2.2528,
"step": 800
},
{
"epoch": 0.01850181020797405,
"grad_norm": 0.7940350770950317,
"learning_rate": 1.7291942572543807e-05,
"loss": 1.9905,
"step": 810
},
{
"epoch": 0.01873022761794904,
"grad_norm": 0.9825338125228882,
"learning_rate": 1.5567207449798515e-05,
"loss": 2.0347,
"step": 820
},
{
"epoch": 0.018958645027924027,
"grad_norm": 0.7255304455757141,
"learning_rate": 1.3925797299605647e-05,
"loss": 2.0235,
"step": 830
},
{
"epoch": 0.019187062437899017,
"grad_norm": 0.9329121708869934,
"learning_rate": 1.2369331995613665e-05,
"loss": 2.0792,
"step": 840
},
{
"epoch": 0.019415479847874006,
"grad_norm": 0.8776724934577942,
"learning_rate": 1.0899347581163221e-05,
"loss": 2.0174,
"step": 850
},
{
"epoch": 0.019643897257848993,
"grad_norm": 0.9908276796340942,
"learning_rate": 9.517294753398064e-06,
"loss": 1.9447,
"step": 860
},
{
"epoch": 0.019872314667823982,
"grad_norm": 0.9467048645019531,
"learning_rate": 8.224537431601886e-06,
"loss": 2.1375,
"step": 870
},
{
"epoch": 0.02010073207779897,
"grad_norm": 1.004712700843811,
"learning_rate": 7.022351411174866e-06,
"loss": 2.3284,
"step": 880
},
{
"epoch": 0.020329149487773958,
"grad_norm": 1.0520464181900024,
"learning_rate": 5.911923104577455e-06,
"loss": 2.0863,
"step": 890
},
{
"epoch": 0.020557566897748947,
"grad_norm": 0.8571903109550476,
"learning_rate": 4.8943483704846475e-06,
"loss": 2.0996,
"step": 900
},
{
"epoch": 0.020785984307723934,
"grad_norm": 0.9216386079788208,
"learning_rate": 3.970631432305694e-06,
"loss": 2.0802,
"step": 910
},
{
"epoch": 0.021014401717698923,
"grad_norm": 0.9897679686546326,
"learning_rate": 3.1416838871368924e-06,
"loss": 2.2645,
"step": 920
},
{
"epoch": 0.021242819127673913,
"grad_norm": 0.94163578748703,
"learning_rate": 2.4083238061252567e-06,
"loss": 2.1278,
"step": 930
},
{
"epoch": 0.0214712365376489,
"grad_norm": 0.8382057547569275,
"learning_rate": 1.771274927131139e-06,
"loss": 2.1074,
"step": 940
},
{
"epoch": 0.02169965394762389,
"grad_norm": 0.9168360829353333,
"learning_rate": 1.231165940486234e-06,
"loss": 1.993,
"step": 950
},
{
"epoch": 0.021928071357598874,
"grad_norm": 1.0946999788284302,
"learning_rate": 7.885298685522235e-07,
"loss": 2.1184,
"step": 960
},
{
"epoch": 0.022156488767573864,
"grad_norm": 0.8011813163757324,
"learning_rate": 4.438035396920004e-07,
"loss": 2.1275,
"step": 970
},
{
"epoch": 0.022384906177548854,
"grad_norm": 1.2118070125579834,
"learning_rate": 1.973271571728441e-07,
"loss": 2.0635,
"step": 980
},
{
"epoch": 0.02261332358752384,
"grad_norm": 0.8423234820365906,
"learning_rate": 4.934396342684e-08,
"loss": 2.049,
"step": 990
},
{
"epoch": 0.02284174099749883,
"grad_norm": 1.2177627086639404,
"learning_rate": 0.0,
"loss": 2.2816,
"step": 1000
}
],
"logging_steps": 10,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 8.681733252309811e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}