borakaragul's picture
Upload folder using huggingface_hub
1278a20 verified
raw
history blame contribute delete
No virus
16.9 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 100,
"global_step": 833,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012004801920768308,
"grad_norm": 0.3538109064102173,
"learning_rate": 9.879951980792317e-05,
"loss": 2.3075,
"step": 10
},
{
"epoch": 0.024009603841536616,
"grad_norm": 0.3431814908981323,
"learning_rate": 9.759903961584634e-05,
"loss": 2.1468,
"step": 20
},
{
"epoch": 0.03601440576230492,
"grad_norm": 0.6265680193901062,
"learning_rate": 9.639855942376951e-05,
"loss": 2.0353,
"step": 30
},
{
"epoch": 0.04801920768307323,
"grad_norm": 0.3244292736053467,
"learning_rate": 9.519807923169268e-05,
"loss": 2.0219,
"step": 40
},
{
"epoch": 0.060024009603841535,
"grad_norm": 0.6210302114486694,
"learning_rate": 9.399759903961585e-05,
"loss": 1.9809,
"step": 50
},
{
"epoch": 0.07202881152460984,
"grad_norm": 0.4695661962032318,
"learning_rate": 9.279711884753903e-05,
"loss": 1.9376,
"step": 60
},
{
"epoch": 0.08403361344537816,
"grad_norm": 0.27379557490348816,
"learning_rate": 9.159663865546218e-05,
"loss": 1.923,
"step": 70
},
{
"epoch": 0.09603841536614646,
"grad_norm": 0.4497829079627991,
"learning_rate": 9.039615846338536e-05,
"loss": 1.9875,
"step": 80
},
{
"epoch": 0.10804321728691477,
"grad_norm": 0.3675350844860077,
"learning_rate": 8.919567827130852e-05,
"loss": 1.953,
"step": 90
},
{
"epoch": 0.12004801920768307,
"grad_norm": 0.293821781873703,
"learning_rate": 8.79951980792317e-05,
"loss": 1.8549,
"step": 100
},
{
"epoch": 0.12004801920768307,
"eval_loss": 1.889587640762329,
"eval_runtime": 39.6181,
"eval_samples_per_second": 25.241,
"eval_steps_per_second": 4.215,
"step": 100
},
{
"epoch": 0.13205282112845138,
"grad_norm": 0.25892049074172974,
"learning_rate": 8.679471788715487e-05,
"loss": 1.8372,
"step": 110
},
{
"epoch": 0.14405762304921968,
"grad_norm": 0.33091431856155396,
"learning_rate": 8.559423769507804e-05,
"loss": 1.9107,
"step": 120
},
{
"epoch": 0.15606242496998798,
"grad_norm": 0.3052417039871216,
"learning_rate": 8.43937575030012e-05,
"loss": 1.8123,
"step": 130
},
{
"epoch": 0.16806722689075632,
"grad_norm": 0.3017849028110504,
"learning_rate": 8.319327731092437e-05,
"loss": 1.8966,
"step": 140
},
{
"epoch": 0.18007202881152462,
"grad_norm": 0.28657111525535583,
"learning_rate": 8.199279711884754e-05,
"loss": 1.8395,
"step": 150
},
{
"epoch": 0.19207683073229292,
"grad_norm": 0.3324660062789917,
"learning_rate": 8.079231692677071e-05,
"loss": 1.8841,
"step": 160
},
{
"epoch": 0.20408163265306123,
"grad_norm": 0.28194233775138855,
"learning_rate": 7.959183673469388e-05,
"loss": 1.8627,
"step": 170
},
{
"epoch": 0.21608643457382953,
"grad_norm": 0.45277833938598633,
"learning_rate": 7.839135654261706e-05,
"loss": 1.8989,
"step": 180
},
{
"epoch": 0.22809123649459784,
"grad_norm": 0.3656202256679535,
"learning_rate": 7.719087635054022e-05,
"loss": 1.911,
"step": 190
},
{
"epoch": 0.24009603841536614,
"grad_norm": 0.33644187450408936,
"learning_rate": 7.599039615846338e-05,
"loss": 1.7805,
"step": 200
},
{
"epoch": 0.24009603841536614,
"eval_loss": 1.8699440956115723,
"eval_runtime": 39.0269,
"eval_samples_per_second": 25.623,
"eval_steps_per_second": 4.279,
"step": 200
},
{
"epoch": 0.25210084033613445,
"grad_norm": 0.24649439752101898,
"learning_rate": 7.478991596638657e-05,
"loss": 1.841,
"step": 210
},
{
"epoch": 0.26410564225690275,
"grad_norm": 0.30980348587036133,
"learning_rate": 7.358943577430972e-05,
"loss": 1.9065,
"step": 220
},
{
"epoch": 0.27611044417767105,
"grad_norm": 0.2731610834598541,
"learning_rate": 7.23889555822329e-05,
"loss": 1.8713,
"step": 230
},
{
"epoch": 0.28811524609843936,
"grad_norm": 0.3211459517478943,
"learning_rate": 7.118847539015606e-05,
"loss": 1.8719,
"step": 240
},
{
"epoch": 0.30012004801920766,
"grad_norm": 0.3356248736381531,
"learning_rate": 6.998799519807924e-05,
"loss": 1.8211,
"step": 250
},
{
"epoch": 0.31212484993997597,
"grad_norm": 0.35724079608917236,
"learning_rate": 6.878751500600241e-05,
"loss": 1.8773,
"step": 260
},
{
"epoch": 0.3241296518607443,
"grad_norm": 0.5718697905540466,
"learning_rate": 6.758703481392558e-05,
"loss": 1.8231,
"step": 270
},
{
"epoch": 0.33613445378151263,
"grad_norm": 0.32961568236351013,
"learning_rate": 6.638655462184874e-05,
"loss": 1.9494,
"step": 280
},
{
"epoch": 0.34813925570228094,
"grad_norm": 0.3693414628505707,
"learning_rate": 6.518607442977191e-05,
"loss": 1.835,
"step": 290
},
{
"epoch": 0.36014405762304924,
"grad_norm": 0.4198884665966034,
"learning_rate": 6.398559423769508e-05,
"loss": 1.8722,
"step": 300
},
{
"epoch": 0.36014405762304924,
"eval_loss": 1.8641259670257568,
"eval_runtime": 39.2717,
"eval_samples_per_second": 25.464,
"eval_steps_per_second": 4.252,
"step": 300
},
{
"epoch": 0.37214885954381755,
"grad_norm": 0.3070141077041626,
"learning_rate": 6.278511404561825e-05,
"loss": 1.8771,
"step": 310
},
{
"epoch": 0.38415366146458585,
"grad_norm": 0.43982160091400146,
"learning_rate": 6.158463385354142e-05,
"loss": 1.8954,
"step": 320
},
{
"epoch": 0.39615846338535415,
"grad_norm": 0.2642657458782196,
"learning_rate": 6.038415366146459e-05,
"loss": 1.8977,
"step": 330
},
{
"epoch": 0.40816326530612246,
"grad_norm": 0.25595635175704956,
"learning_rate": 5.918367346938776e-05,
"loss": 1.8949,
"step": 340
},
{
"epoch": 0.42016806722689076,
"grad_norm": 0.27742770314216614,
"learning_rate": 5.7983193277310935e-05,
"loss": 1.8223,
"step": 350
},
{
"epoch": 0.43217286914765907,
"grad_norm": 0.2831798791885376,
"learning_rate": 5.6782713085234096e-05,
"loss": 1.8477,
"step": 360
},
{
"epoch": 0.44417767106842737,
"grad_norm": 0.3077690005302429,
"learning_rate": 5.558223289315727e-05,
"loss": 1.9376,
"step": 370
},
{
"epoch": 0.4561824729891957,
"grad_norm": 0.5014916658401489,
"learning_rate": 5.438175270108043e-05,
"loss": 1.9241,
"step": 380
},
{
"epoch": 0.468187274909964,
"grad_norm": 0.7313567399978638,
"learning_rate": 5.31812725090036e-05,
"loss": 1.8995,
"step": 390
},
{
"epoch": 0.4801920768307323,
"grad_norm": 0.26589518785476685,
"learning_rate": 5.1980792316926776e-05,
"loss": 1.95,
"step": 400
},
{
"epoch": 0.4801920768307323,
"eval_loss": 1.8611316680908203,
"eval_runtime": 39.4525,
"eval_samples_per_second": 25.347,
"eval_steps_per_second": 4.233,
"step": 400
},
{
"epoch": 0.4921968787515006,
"grad_norm": 0.293312668800354,
"learning_rate": 5.078031212484994e-05,
"loss": 1.8592,
"step": 410
},
{
"epoch": 0.5042016806722689,
"grad_norm": 0.35831350088119507,
"learning_rate": 4.957983193277311e-05,
"loss": 1.9295,
"step": 420
},
{
"epoch": 0.5162064825930373,
"grad_norm": 0.33949533104896545,
"learning_rate": 4.837935174069628e-05,
"loss": 1.8581,
"step": 430
},
{
"epoch": 0.5282112845138055,
"grad_norm": 0.5358735918998718,
"learning_rate": 4.717887154861945e-05,
"loss": 1.8824,
"step": 440
},
{
"epoch": 0.5402160864345739,
"grad_norm": 0.3915724456310272,
"learning_rate": 4.5978391356542624e-05,
"loss": 1.8699,
"step": 450
},
{
"epoch": 0.5522208883553421,
"grad_norm": 0.3336695730686188,
"learning_rate": 4.477791116446579e-05,
"loss": 1.9749,
"step": 460
},
{
"epoch": 0.5642256902761105,
"grad_norm": 0.3887428641319275,
"learning_rate": 4.3577430972388954e-05,
"loss": 1.8134,
"step": 470
},
{
"epoch": 0.5762304921968787,
"grad_norm": 0.26423540711402893,
"learning_rate": 4.237695078031212e-05,
"loss": 1.8495,
"step": 480
},
{
"epoch": 0.5882352941176471,
"grad_norm": 0.30970582365989685,
"learning_rate": 4.11764705882353e-05,
"loss": 1.8386,
"step": 490
},
{
"epoch": 0.6002400960384153,
"grad_norm": 0.38180047273635864,
"learning_rate": 3.9975990396158466e-05,
"loss": 1.8729,
"step": 500
},
{
"epoch": 0.6002400960384153,
"eval_loss": 1.8590781688690186,
"eval_runtime": 39.0438,
"eval_samples_per_second": 25.612,
"eval_steps_per_second": 4.277,
"step": 500
},
{
"epoch": 0.6122448979591837,
"grad_norm": 0.4111509621143341,
"learning_rate": 3.8775510204081634e-05,
"loss": 1.833,
"step": 510
},
{
"epoch": 0.6242496998799519,
"grad_norm": 0.515201210975647,
"learning_rate": 3.75750300120048e-05,
"loss": 1.8744,
"step": 520
},
{
"epoch": 0.6362545018007203,
"grad_norm": 0.2900288701057434,
"learning_rate": 3.637454981992797e-05,
"loss": 1.8364,
"step": 530
},
{
"epoch": 0.6482593037214885,
"grad_norm": 0.2963515520095825,
"learning_rate": 3.517406962785114e-05,
"loss": 1.8469,
"step": 540
},
{
"epoch": 0.6602641056422569,
"grad_norm": 0.41035813093185425,
"learning_rate": 3.3973589435774314e-05,
"loss": 1.9474,
"step": 550
},
{
"epoch": 0.6722689075630253,
"grad_norm": 0.246292382478714,
"learning_rate": 3.277310924369748e-05,
"loss": 1.8128,
"step": 560
},
{
"epoch": 0.6842737094837935,
"grad_norm": 0.3911287188529968,
"learning_rate": 3.157262905162065e-05,
"loss": 1.8663,
"step": 570
},
{
"epoch": 0.6962785114045619,
"grad_norm": 0.3035126328468323,
"learning_rate": 3.037214885954382e-05,
"loss": 1.8921,
"step": 580
},
{
"epoch": 0.7082833133253301,
"grad_norm": 0.2977759540081024,
"learning_rate": 2.917166866746699e-05,
"loss": 1.8288,
"step": 590
},
{
"epoch": 0.7202881152460985,
"grad_norm": 0.3059289753437042,
"learning_rate": 2.797118847539016e-05,
"loss": 1.8482,
"step": 600
},
{
"epoch": 0.7202881152460985,
"eval_loss": 1.8568615913391113,
"eval_runtime": 39.5808,
"eval_samples_per_second": 25.265,
"eval_steps_per_second": 4.219,
"step": 600
},
{
"epoch": 0.7322929171668667,
"grad_norm": 0.36612552404403687,
"learning_rate": 2.6770708283313327e-05,
"loss": 1.8577,
"step": 610
},
{
"epoch": 0.7442977190876351,
"grad_norm": 0.2968533933162689,
"learning_rate": 2.5570228091236498e-05,
"loss": 1.8295,
"step": 620
},
{
"epoch": 0.7563025210084033,
"grad_norm": 0.40536123514175415,
"learning_rate": 2.4369747899159663e-05,
"loss": 1.9242,
"step": 630
},
{
"epoch": 0.7683073229291717,
"grad_norm": 0.3592563271522522,
"learning_rate": 2.3169267707082835e-05,
"loss": 1.8677,
"step": 640
},
{
"epoch": 0.78031212484994,
"grad_norm": 0.3828364908695221,
"learning_rate": 2.1968787515006003e-05,
"loss": 1.9055,
"step": 650
},
{
"epoch": 0.7923169267707083,
"grad_norm": 0.26257479190826416,
"learning_rate": 2.076830732292917e-05,
"loss": 1.9383,
"step": 660
},
{
"epoch": 0.8043217286914766,
"grad_norm": 0.3253800570964813,
"learning_rate": 1.9567827130852343e-05,
"loss": 1.8142,
"step": 670
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.3129618167877197,
"learning_rate": 1.836734693877551e-05,
"loss": 1.8447,
"step": 680
},
{
"epoch": 0.8283313325330132,
"grad_norm": 0.32753467559814453,
"learning_rate": 1.7166866746698683e-05,
"loss": 1.8617,
"step": 690
},
{
"epoch": 0.8403361344537815,
"grad_norm": 0.28356197476387024,
"learning_rate": 1.5966386554621848e-05,
"loss": 1.768,
"step": 700
},
{
"epoch": 0.8403361344537815,
"eval_loss": 1.8553369045257568,
"eval_runtime": 39.3665,
"eval_samples_per_second": 25.402,
"eval_steps_per_second": 4.242,
"step": 700
},
{
"epoch": 0.8523409363745498,
"grad_norm": 0.42871856689453125,
"learning_rate": 1.4765906362545018e-05,
"loss": 1.8756,
"step": 710
},
{
"epoch": 0.8643457382953181,
"grad_norm": 0.32104918360710144,
"learning_rate": 1.3565426170468188e-05,
"loss": 1.8396,
"step": 720
},
{
"epoch": 0.8763505402160864,
"grad_norm": 0.3462695777416229,
"learning_rate": 1.2364945978391356e-05,
"loss": 1.8499,
"step": 730
},
{
"epoch": 0.8883553421368547,
"grad_norm": 0.38381311297416687,
"learning_rate": 1.1164465786314526e-05,
"loss": 1.8353,
"step": 740
},
{
"epoch": 0.9003601440576231,
"grad_norm": 0.48733091354370117,
"learning_rate": 9.963985594237696e-06,
"loss": 1.8566,
"step": 750
},
{
"epoch": 0.9123649459783914,
"grad_norm": 0.35725510120391846,
"learning_rate": 8.763505402160866e-06,
"loss": 1.9016,
"step": 760
},
{
"epoch": 0.9243697478991597,
"grad_norm": 0.2945646047592163,
"learning_rate": 7.563025210084033e-06,
"loss": 1.7727,
"step": 770
},
{
"epoch": 0.936374549819928,
"grad_norm": 0.2922873795032501,
"learning_rate": 6.362545018007203e-06,
"loss": 1.8482,
"step": 780
},
{
"epoch": 0.9483793517406963,
"grad_norm": 0.29431605339050293,
"learning_rate": 5.162064825930372e-06,
"loss": 1.8867,
"step": 790
},
{
"epoch": 0.9603841536614646,
"grad_norm": 0.32220232486724854,
"learning_rate": 3.9615846338535415e-06,
"loss": 1.8904,
"step": 800
},
{
"epoch": 0.9603841536614646,
"eval_loss": 1.854928970336914,
"eval_runtime": 39.3819,
"eval_samples_per_second": 25.392,
"eval_steps_per_second": 4.241,
"step": 800
},
{
"epoch": 0.9723889555822329,
"grad_norm": 0.2991810142993927,
"learning_rate": 2.7611044417767106e-06,
"loss": 1.8523,
"step": 810
},
{
"epoch": 0.9843937575030012,
"grad_norm": 0.3944421112537384,
"learning_rate": 1.56062424969988e-06,
"loss": 1.8502,
"step": 820
},
{
"epoch": 0.9963985594237695,
"grad_norm": 0.3155655264854431,
"learning_rate": 3.601440576230492e-07,
"loss": 1.9278,
"step": 830
}
],
"logging_steps": 10,
"max_steps": 833,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.3760243809940275e+17,
"train_batch_size": 12,
"trial_name": null,
"trial_params": null
}