|
{ |
|
"best_metric": 0.8991448283195496, |
|
"best_model_checkpoint": "../../output/chatglm3-6b/LangGPT/checkpoint-700", |
|
"epoch": 9.0, |
|
"eval_steps": 100, |
|
"global_step": 720, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.125, |
|
"grad_norm": 0.40453147888183594, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.6827, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.5551838874816895, |
|
"learning_rate": 5e-05, |
|
"loss": 1.6309, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.375, |
|
"grad_norm": 0.7859359383583069, |
|
"learning_rate": 4.997482666353287e-05, |
|
"loss": 1.5415, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.597720742225647, |
|
"learning_rate": 4.989935734988098e-05, |
|
"loss": 1.393, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 0.4020984172821045, |
|
"learning_rate": 4.977374404419837e-05, |
|
"loss": 1.2563, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.35916563868522644, |
|
"learning_rate": 4.959823971496574e-05, |
|
"loss": 1.1963, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.875, |
|
"grad_norm": 0.3013848066329956, |
|
"learning_rate": 4.937319780454559e-05, |
|
"loss": 1.1385, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.23849129676818848, |
|
"learning_rate": 4.909907151739633e-05, |
|
"loss": 1.1085, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.125, |
|
"grad_norm": 0.22885890305042267, |
|
"learning_rate": 4.877641290737884e-05, |
|
"loss": 1.1025, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 0.20683708786964417, |
|
"learning_rate": 4.8405871765993433e-05, |
|
"loss": 1.0558, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 1.0500283241271973, |
|
"eval_runtime": 353.0769, |
|
"eval_samples_per_second": 2.419, |
|
"eval_steps_per_second": 0.606, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.375, |
|
"grad_norm": 0.20663395524024963, |
|
"learning_rate": 4.7988194313786275e-05, |
|
"loss": 1.0258, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 0.18335361778736115, |
|
"learning_rate": 4.752422169756048e-05, |
|
"loss": 1.0261, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.625, |
|
"grad_norm": 0.18184833228588104, |
|
"learning_rate": 4.701488829641845e-05, |
|
"loss": 0.9923, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 0.19089923799037933, |
|
"learning_rate": 4.6461219840046654e-05, |
|
"loss": 0.9835, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 0.17791251838207245, |
|
"learning_rate": 4.586433134303257e-05, |
|
"loss": 1.0039, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.18376672267913818, |
|
"learning_rate": 4.522542485937369e-05, |
|
"loss": 0.9947, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.125, |
|
"grad_norm": 0.20052292943000793, |
|
"learning_rate": 4.454578706170075e-05, |
|
"loss": 0.9821, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 0.19209513068199158, |
|
"learning_rate": 4.382678665009028e-05, |
|
"loss": 0.9535, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.375, |
|
"grad_norm": 0.19733993709087372, |
|
"learning_rate": 4.306987159568479e-05, |
|
"loss": 0.9514, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 0.18989509344100952, |
|
"learning_rate": 4.227656622467162e-05, |
|
"loss": 0.9566, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 0.9630343914031982, |
|
"eval_runtime": 353.288, |
|
"eval_samples_per_second": 2.417, |
|
"eval_steps_per_second": 0.606, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.625, |
|
"grad_norm": 0.19188831746578217, |
|
"learning_rate": 4.144846814849282e-05, |
|
"loss": 0.9655, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 0.2034657597541809, |
|
"learning_rate": 4.058724504646834e-05, |
|
"loss": 0.9537, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.875, |
|
"grad_norm": 0.20900140702724457, |
|
"learning_rate": 3.969463130731183e-05, |
|
"loss": 0.951, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.231728196144104, |
|
"learning_rate": 3.8772424536302564e-05, |
|
"loss": 0.938, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.125, |
|
"grad_norm": 0.21837086975574493, |
|
"learning_rate": 3.782248193514766e-05, |
|
"loss": 0.955, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 0.2057914286851883, |
|
"learning_rate": 3.6846716561824965e-05, |
|
"loss": 0.9319, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.375, |
|
"grad_norm": 0.22230790555477142, |
|
"learning_rate": 3.5847093477938956e-05, |
|
"loss": 0.9385, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 0.24387766420841217, |
|
"learning_rate": 3.4825625791348096e-05, |
|
"loss": 0.911, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 3.625, |
|
"grad_norm": 0.2634485065937042, |
|
"learning_rate": 3.378437060203357e-05, |
|
"loss": 0.9366, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 0.22965680062770844, |
|
"learning_rate": 3.272542485937369e-05, |
|
"loss": 0.9082, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"eval_loss": 0.928753137588501, |
|
"eval_runtime": 353.3512, |
|
"eval_samples_per_second": 2.417, |
|
"eval_steps_per_second": 0.606, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.875, |
|
"grad_norm": 0.21778391301631927, |
|
"learning_rate": 3.165092113916688e-05, |
|
"loss": 0.9158, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.24541890621185303, |
|
"learning_rate": 3.056302334890786e-05, |
|
"loss": 0.9027, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 4.125, |
|
"grad_norm": 0.25015348196029663, |
|
"learning_rate": 2.9463922369965917e-05, |
|
"loss": 0.9336, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"grad_norm": 0.22015893459320068, |
|
"learning_rate": 2.8355831645441388e-05, |
|
"loss": 0.9161, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.375, |
|
"grad_norm": 0.2516670823097229, |
|
"learning_rate": 2.724098272258584e-05, |
|
"loss": 0.8966, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"grad_norm": 0.2541712820529938, |
|
"learning_rate": 2.6121620758762877e-05, |
|
"loss": 0.8954, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 4.625, |
|
"grad_norm": 0.25608915090560913, |
|
"learning_rate": 2.5e-05, |
|
"loss": 0.8815, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"grad_norm": 0.24169643223285675, |
|
"learning_rate": 2.3878379241237136e-05, |
|
"loss": 0.89, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 4.875, |
|
"grad_norm": 0.2623349130153656, |
|
"learning_rate": 2.2759017277414166e-05, |
|
"loss": 0.9196, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.29517388343811035, |
|
"learning_rate": 2.164416835455862e-05, |
|
"loss": 0.8992, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.9107962846755981, |
|
"eval_runtime": 353.3297, |
|
"eval_samples_per_second": 2.417, |
|
"eval_steps_per_second": 0.606, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 5.125, |
|
"grad_norm": 0.2589443027973175, |
|
"learning_rate": 2.0536077630034086e-05, |
|
"loss": 0.888, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"grad_norm": 0.24191297590732574, |
|
"learning_rate": 1.9436976651092144e-05, |
|
"loss": 0.8901, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.375, |
|
"grad_norm": 0.27726104855537415, |
|
"learning_rate": 1.8349078860833123e-05, |
|
"loss": 0.9147, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"grad_norm": 0.23908096551895142, |
|
"learning_rate": 1.7274575140626318e-05, |
|
"loss": 0.8925, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.625, |
|
"grad_norm": 0.30176234245300293, |
|
"learning_rate": 1.621562939796643e-05, |
|
"loss": 0.9012, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"grad_norm": 0.23645330965518951, |
|
"learning_rate": 1.5174374208651912e-05, |
|
"loss": 0.8808, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 5.875, |
|
"grad_norm": 0.2720588147640228, |
|
"learning_rate": 1.4152906522061048e-05, |
|
"loss": 0.8816, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 0.2631034553050995, |
|
"learning_rate": 1.3153283438175034e-05, |
|
"loss": 0.8941, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 6.125, |
|
"grad_norm": 0.2486189901828766, |
|
"learning_rate": 1.217751806485235e-05, |
|
"loss": 0.9048, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"grad_norm": 0.2926970422267914, |
|
"learning_rate": 1.122757546369744e-05, |
|
"loss": 0.8874, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"eval_loss": 0.9028034806251526, |
|
"eval_runtime": 353.2345, |
|
"eval_samples_per_second": 2.418, |
|
"eval_steps_per_second": 0.606, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 6.375, |
|
"grad_norm": 0.25221139192581177, |
|
"learning_rate": 1.0305368692688174e-05, |
|
"loss": 0.8738, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"grad_norm": 0.2523793578147888, |
|
"learning_rate": 9.412754953531663e-06, |
|
"loss": 0.8951, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 6.625, |
|
"grad_norm": 0.2493809163570404, |
|
"learning_rate": 8.551531851507186e-06, |
|
"loss": 0.8914, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"grad_norm": 0.2688143253326416, |
|
"learning_rate": 7.723433775328384e-06, |
|
"loss": 0.8818, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 6.875, |
|
"grad_norm": 0.2695543169975281, |
|
"learning_rate": 6.930128404315214e-06, |
|
"loss": 0.8794, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 0.27596864104270935, |
|
"learning_rate": 6.173213349909729e-06, |
|
"loss": 0.8814, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 7.125, |
|
"grad_norm": 0.27881208062171936, |
|
"learning_rate": 5.454212938299255e-06, |
|
"loss": 0.8909, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"grad_norm": 0.2895490825176239, |
|
"learning_rate": 4.7745751406263165e-06, |
|
"loss": 0.8737, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 7.375, |
|
"grad_norm": 0.25476014614105225, |
|
"learning_rate": 4.135668656967434e-06, |
|
"loss": 0.8937, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"grad_norm": 0.2785739600658417, |
|
"learning_rate": 3.5387801599533475e-06, |
|
"loss": 0.8835, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"eval_loss": 0.899681031703949, |
|
"eval_runtime": 353.2175, |
|
"eval_samples_per_second": 2.418, |
|
"eval_steps_per_second": 0.606, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 7.625, |
|
"grad_norm": 0.27525651454925537, |
|
"learning_rate": 2.98511170358155e-06, |
|
"loss": 0.8841, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"grad_norm": 0.25052082538604736, |
|
"learning_rate": 2.475778302439524e-06, |
|
"loss": 0.8979, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 7.875, |
|
"grad_norm": 0.2501230537891388, |
|
"learning_rate": 2.0118056862137357e-06, |
|
"loss": 0.8696, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.2521611452102661, |
|
"learning_rate": 1.59412823400657e-06, |
|
"loss": 0.8782, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 8.125, |
|
"grad_norm": 0.249056875705719, |
|
"learning_rate": 1.2235870926211619e-06, |
|
"loss": 0.8873, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"grad_norm": 0.27458131313323975, |
|
"learning_rate": 9.009284826036691e-07, |
|
"loss": 0.8737, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 8.375, |
|
"grad_norm": 0.24417945742607117, |
|
"learning_rate": 6.268021954544096e-07, |
|
"loss": 0.8883, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"grad_norm": 0.25331562757492065, |
|
"learning_rate": 4.0176028503425835e-07, |
|
"loss": 0.87, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 8.625, |
|
"grad_norm": 0.25556355714797974, |
|
"learning_rate": 2.262559558016325e-07, |
|
"loss": 0.8746, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"grad_norm": 0.27511876821517944, |
|
"learning_rate": 1.006426501190233e-07, |
|
"loss": 0.8912, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"eval_loss": 0.8991448283195496, |
|
"eval_runtime": 353.267, |
|
"eval_samples_per_second": 2.417, |
|
"eval_steps_per_second": 0.606, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 8.875, |
|
"grad_norm": 0.2639774680137634, |
|
"learning_rate": 2.5173336467135267e-08, |
|
"loss": 0.9007, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 0.27557557821273804, |
|
"learning_rate": 0.0, |
|
"loss": 0.8796, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"step": 720, |
|
"total_flos": 2.5580424283828716e+18, |
|
"train_loss": 0.9693152533637153, |
|
"train_runtime": 60193.9387, |
|
"train_samples_per_second": 1.148, |
|
"train_steps_per_second": 0.012 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 720, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 9, |
|
"save_steps": 100, |
|
"total_flos": 2.5580424283828716e+18, |
|
"train_batch_size": 12, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|