chatglm3-6b_sa_v0.1 / trainer_state.json
sci-m-wang's picture
Upload 15 files
eb518e3 verified
{
"best_metric": 0.8991448283195496,
"best_model_checkpoint": "../../output/chatglm3-6b/LangGPT/checkpoint-700",
"epoch": 9.0,
"eval_steps": 100,
"global_step": 720,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.125,
"grad_norm": 0.40453147888183594,
"learning_rate": 2.5e-05,
"loss": 1.6827,
"step": 10
},
{
"epoch": 0.25,
"grad_norm": 0.5551838874816895,
"learning_rate": 5e-05,
"loss": 1.6309,
"step": 20
},
{
"epoch": 0.375,
"grad_norm": 0.7859359383583069,
"learning_rate": 4.997482666353287e-05,
"loss": 1.5415,
"step": 30
},
{
"epoch": 0.5,
"grad_norm": 0.597720742225647,
"learning_rate": 4.989935734988098e-05,
"loss": 1.393,
"step": 40
},
{
"epoch": 0.625,
"grad_norm": 0.4020984172821045,
"learning_rate": 4.977374404419837e-05,
"loss": 1.2563,
"step": 50
},
{
"epoch": 0.75,
"grad_norm": 0.35916563868522644,
"learning_rate": 4.959823971496574e-05,
"loss": 1.1963,
"step": 60
},
{
"epoch": 0.875,
"grad_norm": 0.3013848066329956,
"learning_rate": 4.937319780454559e-05,
"loss": 1.1385,
"step": 70
},
{
"epoch": 1.0,
"grad_norm": 0.23849129676818848,
"learning_rate": 4.909907151739633e-05,
"loss": 1.1085,
"step": 80
},
{
"epoch": 1.125,
"grad_norm": 0.22885890305042267,
"learning_rate": 4.877641290737884e-05,
"loss": 1.1025,
"step": 90
},
{
"epoch": 1.25,
"grad_norm": 0.20683708786964417,
"learning_rate": 4.8405871765993433e-05,
"loss": 1.0558,
"step": 100
},
{
"epoch": 1.25,
"eval_loss": 1.0500283241271973,
"eval_runtime": 353.0769,
"eval_samples_per_second": 2.419,
"eval_steps_per_second": 0.606,
"step": 100
},
{
"epoch": 1.375,
"grad_norm": 0.20663395524024963,
"learning_rate": 4.7988194313786275e-05,
"loss": 1.0258,
"step": 110
},
{
"epoch": 1.5,
"grad_norm": 0.18335361778736115,
"learning_rate": 4.752422169756048e-05,
"loss": 1.0261,
"step": 120
},
{
"epoch": 1.625,
"grad_norm": 0.18184833228588104,
"learning_rate": 4.701488829641845e-05,
"loss": 0.9923,
"step": 130
},
{
"epoch": 1.75,
"grad_norm": 0.19089923799037933,
"learning_rate": 4.6461219840046654e-05,
"loss": 0.9835,
"step": 140
},
{
"epoch": 1.875,
"grad_norm": 0.17791251838207245,
"learning_rate": 4.586433134303257e-05,
"loss": 1.0039,
"step": 150
},
{
"epoch": 2.0,
"grad_norm": 0.18376672267913818,
"learning_rate": 4.522542485937369e-05,
"loss": 0.9947,
"step": 160
},
{
"epoch": 2.125,
"grad_norm": 0.20052292943000793,
"learning_rate": 4.454578706170075e-05,
"loss": 0.9821,
"step": 170
},
{
"epoch": 2.25,
"grad_norm": 0.19209513068199158,
"learning_rate": 4.382678665009028e-05,
"loss": 0.9535,
"step": 180
},
{
"epoch": 2.375,
"grad_norm": 0.19733993709087372,
"learning_rate": 4.306987159568479e-05,
"loss": 0.9514,
"step": 190
},
{
"epoch": 2.5,
"grad_norm": 0.18989509344100952,
"learning_rate": 4.227656622467162e-05,
"loss": 0.9566,
"step": 200
},
{
"epoch": 2.5,
"eval_loss": 0.9630343914031982,
"eval_runtime": 353.288,
"eval_samples_per_second": 2.417,
"eval_steps_per_second": 0.606,
"step": 200
},
{
"epoch": 2.625,
"grad_norm": 0.19188831746578217,
"learning_rate": 4.144846814849282e-05,
"loss": 0.9655,
"step": 210
},
{
"epoch": 2.75,
"grad_norm": 0.2034657597541809,
"learning_rate": 4.058724504646834e-05,
"loss": 0.9537,
"step": 220
},
{
"epoch": 2.875,
"grad_norm": 0.20900140702724457,
"learning_rate": 3.969463130731183e-05,
"loss": 0.951,
"step": 230
},
{
"epoch": 3.0,
"grad_norm": 0.231728196144104,
"learning_rate": 3.8772424536302564e-05,
"loss": 0.938,
"step": 240
},
{
"epoch": 3.125,
"grad_norm": 0.21837086975574493,
"learning_rate": 3.782248193514766e-05,
"loss": 0.955,
"step": 250
},
{
"epoch": 3.25,
"grad_norm": 0.2057914286851883,
"learning_rate": 3.6846716561824965e-05,
"loss": 0.9319,
"step": 260
},
{
"epoch": 3.375,
"grad_norm": 0.22230790555477142,
"learning_rate": 3.5847093477938956e-05,
"loss": 0.9385,
"step": 270
},
{
"epoch": 3.5,
"grad_norm": 0.24387766420841217,
"learning_rate": 3.4825625791348096e-05,
"loss": 0.911,
"step": 280
},
{
"epoch": 3.625,
"grad_norm": 0.2634485065937042,
"learning_rate": 3.378437060203357e-05,
"loss": 0.9366,
"step": 290
},
{
"epoch": 3.75,
"grad_norm": 0.22965680062770844,
"learning_rate": 3.272542485937369e-05,
"loss": 0.9082,
"step": 300
},
{
"epoch": 3.75,
"eval_loss": 0.928753137588501,
"eval_runtime": 353.3512,
"eval_samples_per_second": 2.417,
"eval_steps_per_second": 0.606,
"step": 300
},
{
"epoch": 3.875,
"grad_norm": 0.21778391301631927,
"learning_rate": 3.165092113916688e-05,
"loss": 0.9158,
"step": 310
},
{
"epoch": 4.0,
"grad_norm": 0.24541890621185303,
"learning_rate": 3.056302334890786e-05,
"loss": 0.9027,
"step": 320
},
{
"epoch": 4.125,
"grad_norm": 0.25015348196029663,
"learning_rate": 2.9463922369965917e-05,
"loss": 0.9336,
"step": 330
},
{
"epoch": 4.25,
"grad_norm": 0.22015893459320068,
"learning_rate": 2.8355831645441388e-05,
"loss": 0.9161,
"step": 340
},
{
"epoch": 4.375,
"grad_norm": 0.2516670823097229,
"learning_rate": 2.724098272258584e-05,
"loss": 0.8966,
"step": 350
},
{
"epoch": 4.5,
"grad_norm": 0.2541712820529938,
"learning_rate": 2.6121620758762877e-05,
"loss": 0.8954,
"step": 360
},
{
"epoch": 4.625,
"grad_norm": 0.25608915090560913,
"learning_rate": 2.5e-05,
"loss": 0.8815,
"step": 370
},
{
"epoch": 4.75,
"grad_norm": 0.24169643223285675,
"learning_rate": 2.3878379241237136e-05,
"loss": 0.89,
"step": 380
},
{
"epoch": 4.875,
"grad_norm": 0.2623349130153656,
"learning_rate": 2.2759017277414166e-05,
"loss": 0.9196,
"step": 390
},
{
"epoch": 5.0,
"grad_norm": 0.29517388343811035,
"learning_rate": 2.164416835455862e-05,
"loss": 0.8992,
"step": 400
},
{
"epoch": 5.0,
"eval_loss": 0.9107962846755981,
"eval_runtime": 353.3297,
"eval_samples_per_second": 2.417,
"eval_steps_per_second": 0.606,
"step": 400
},
{
"epoch": 5.125,
"grad_norm": 0.2589443027973175,
"learning_rate": 2.0536077630034086e-05,
"loss": 0.888,
"step": 410
},
{
"epoch": 5.25,
"grad_norm": 0.24191297590732574,
"learning_rate": 1.9436976651092144e-05,
"loss": 0.8901,
"step": 420
},
{
"epoch": 5.375,
"grad_norm": 0.27726104855537415,
"learning_rate": 1.8349078860833123e-05,
"loss": 0.9147,
"step": 430
},
{
"epoch": 5.5,
"grad_norm": 0.23908096551895142,
"learning_rate": 1.7274575140626318e-05,
"loss": 0.8925,
"step": 440
},
{
"epoch": 5.625,
"grad_norm": 0.30176234245300293,
"learning_rate": 1.621562939796643e-05,
"loss": 0.9012,
"step": 450
},
{
"epoch": 5.75,
"grad_norm": 0.23645330965518951,
"learning_rate": 1.5174374208651912e-05,
"loss": 0.8808,
"step": 460
},
{
"epoch": 5.875,
"grad_norm": 0.2720588147640228,
"learning_rate": 1.4152906522061048e-05,
"loss": 0.8816,
"step": 470
},
{
"epoch": 6.0,
"grad_norm": 0.2631034553050995,
"learning_rate": 1.3153283438175034e-05,
"loss": 0.8941,
"step": 480
},
{
"epoch": 6.125,
"grad_norm": 0.2486189901828766,
"learning_rate": 1.217751806485235e-05,
"loss": 0.9048,
"step": 490
},
{
"epoch": 6.25,
"grad_norm": 0.2926970422267914,
"learning_rate": 1.122757546369744e-05,
"loss": 0.8874,
"step": 500
},
{
"epoch": 6.25,
"eval_loss": 0.9028034806251526,
"eval_runtime": 353.2345,
"eval_samples_per_second": 2.418,
"eval_steps_per_second": 0.606,
"step": 500
},
{
"epoch": 6.375,
"grad_norm": 0.25221139192581177,
"learning_rate": 1.0305368692688174e-05,
"loss": 0.8738,
"step": 510
},
{
"epoch": 6.5,
"grad_norm": 0.2523793578147888,
"learning_rate": 9.412754953531663e-06,
"loss": 0.8951,
"step": 520
},
{
"epoch": 6.625,
"grad_norm": 0.2493809163570404,
"learning_rate": 8.551531851507186e-06,
"loss": 0.8914,
"step": 530
},
{
"epoch": 6.75,
"grad_norm": 0.2688143253326416,
"learning_rate": 7.723433775328384e-06,
"loss": 0.8818,
"step": 540
},
{
"epoch": 6.875,
"grad_norm": 0.2695543169975281,
"learning_rate": 6.930128404315214e-06,
"loss": 0.8794,
"step": 550
},
{
"epoch": 7.0,
"grad_norm": 0.27596864104270935,
"learning_rate": 6.173213349909729e-06,
"loss": 0.8814,
"step": 560
},
{
"epoch": 7.125,
"grad_norm": 0.27881208062171936,
"learning_rate": 5.454212938299255e-06,
"loss": 0.8909,
"step": 570
},
{
"epoch": 7.25,
"grad_norm": 0.2895490825176239,
"learning_rate": 4.7745751406263165e-06,
"loss": 0.8737,
"step": 580
},
{
"epoch": 7.375,
"grad_norm": 0.25476014614105225,
"learning_rate": 4.135668656967434e-06,
"loss": 0.8937,
"step": 590
},
{
"epoch": 7.5,
"grad_norm": 0.2785739600658417,
"learning_rate": 3.5387801599533475e-06,
"loss": 0.8835,
"step": 600
},
{
"epoch": 7.5,
"eval_loss": 0.899681031703949,
"eval_runtime": 353.2175,
"eval_samples_per_second": 2.418,
"eval_steps_per_second": 0.606,
"step": 600
},
{
"epoch": 7.625,
"grad_norm": 0.27525651454925537,
"learning_rate": 2.98511170358155e-06,
"loss": 0.8841,
"step": 610
},
{
"epoch": 7.75,
"grad_norm": 0.25052082538604736,
"learning_rate": 2.475778302439524e-06,
"loss": 0.8979,
"step": 620
},
{
"epoch": 7.875,
"grad_norm": 0.2501230537891388,
"learning_rate": 2.0118056862137357e-06,
"loss": 0.8696,
"step": 630
},
{
"epoch": 8.0,
"grad_norm": 0.2521611452102661,
"learning_rate": 1.59412823400657e-06,
"loss": 0.8782,
"step": 640
},
{
"epoch": 8.125,
"grad_norm": 0.249056875705719,
"learning_rate": 1.2235870926211619e-06,
"loss": 0.8873,
"step": 650
},
{
"epoch": 8.25,
"grad_norm": 0.27458131313323975,
"learning_rate": 9.009284826036691e-07,
"loss": 0.8737,
"step": 660
},
{
"epoch": 8.375,
"grad_norm": 0.24417945742607117,
"learning_rate": 6.268021954544096e-07,
"loss": 0.8883,
"step": 670
},
{
"epoch": 8.5,
"grad_norm": 0.25331562757492065,
"learning_rate": 4.0176028503425835e-07,
"loss": 0.87,
"step": 680
},
{
"epoch": 8.625,
"grad_norm": 0.25556355714797974,
"learning_rate": 2.262559558016325e-07,
"loss": 0.8746,
"step": 690
},
{
"epoch": 8.75,
"grad_norm": 0.27511876821517944,
"learning_rate": 1.006426501190233e-07,
"loss": 0.8912,
"step": 700
},
{
"epoch": 8.75,
"eval_loss": 0.8991448283195496,
"eval_runtime": 353.267,
"eval_samples_per_second": 2.417,
"eval_steps_per_second": 0.606,
"step": 700
},
{
"epoch": 8.875,
"grad_norm": 0.2639774680137634,
"learning_rate": 2.5173336467135267e-08,
"loss": 0.9007,
"step": 710
},
{
"epoch": 9.0,
"grad_norm": 0.27557557821273804,
"learning_rate": 0.0,
"loss": 0.8796,
"step": 720
},
{
"epoch": 9.0,
"step": 720,
"total_flos": 2.5580424283828716e+18,
"train_loss": 0.9693152533637153,
"train_runtime": 60193.9387,
"train_samples_per_second": 1.148,
"train_steps_per_second": 0.012
}
],
"logging_steps": 10,
"max_steps": 720,
"num_input_tokens_seen": 0,
"num_train_epochs": 9,
"save_steps": 100,
"total_flos": 2.5580424283828716e+18,
"train_batch_size": 12,
"trial_name": null,
"trial_params": null
}