|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.995488721804511, |
|
"eval_steps": 500, |
|
"global_step": 996, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03007518796992481, |
|
"grad_norm": 5.823914796640075, |
|
"learning_rate": 5e-06, |
|
"loss": 0.801, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06015037593984962, |
|
"grad_norm": 2.3399969132192253, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7172, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09022556390977443, |
|
"grad_norm": 0.8783578572732146, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6923, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12030075187969924, |
|
"grad_norm": 1.7232554735132826, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6834, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.15037593984962405, |
|
"grad_norm": 1.6494035164725938, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6655, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.18045112781954886, |
|
"grad_norm": 0.9942658093552336, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6568, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 0.8609069727317351, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6548, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.24060150375939848, |
|
"grad_norm": 0.8740060444495542, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6454, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.2706766917293233, |
|
"grad_norm": 1.0114201009544495, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6331, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.3007518796992481, |
|
"grad_norm": 0.782353210012559, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6267, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3308270676691729, |
|
"grad_norm": 0.6061308240718153, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6305, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3609022556390977, |
|
"grad_norm": 0.663886451339362, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6246, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.39097744360902253, |
|
"grad_norm": 0.8509285444900951, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6305, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 0.650634026198435, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6291, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.45112781954887216, |
|
"grad_norm": 0.7050446482120244, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6279, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.48120300751879697, |
|
"grad_norm": 0.5260864128397903, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6167, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5112781954887218, |
|
"grad_norm": 0.5159636362174167, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6189, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5413533834586466, |
|
"grad_norm": 0.576027072578746, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6096, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5714285714285714, |
|
"grad_norm": 0.6020714354457843, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6207, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6015037593984962, |
|
"grad_norm": 0.8904744763001984, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6143, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 0.5645606661034147, |
|
"learning_rate": 5e-06, |
|
"loss": 0.619, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.6616541353383458, |
|
"grad_norm": 0.7456282489942, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6229, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.6917293233082706, |
|
"grad_norm": 0.47671003637879944, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6141, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7218045112781954, |
|
"grad_norm": 0.5712162016443497, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6148, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.7518796992481203, |
|
"grad_norm": 0.5787527818544967, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6112, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.7819548872180451, |
|
"grad_norm": 0.7727895820276254, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6124, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8120300751879699, |
|
"grad_norm": 0.6627350412484229, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6101, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 0.49307733921025576, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6101, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.8721804511278195, |
|
"grad_norm": 0.5299740752648865, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6074, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.9022556390977443, |
|
"grad_norm": 0.9821666854504033, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6035, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9323308270676691, |
|
"grad_norm": 0.5537120584191855, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6037, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.9624060150375939, |
|
"grad_norm": 0.5346047989400851, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6128, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.9924812030075187, |
|
"grad_norm": 0.5177704737274331, |
|
"learning_rate": 5e-06, |
|
"loss": 0.6043, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.9984962406015038, |
|
"eval_loss": 0.6096732020378113, |
|
"eval_runtime": 116.9563, |
|
"eval_samples_per_second": 76.593, |
|
"eval_steps_per_second": 0.599, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.0225563909774436, |
|
"grad_norm": 0.4996873605367385, |
|
"learning_rate": 5e-06, |
|
"loss": 0.58, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 0.59979801353788, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5596, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.0827067669172932, |
|
"grad_norm": 0.645810347625552, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5591, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.112781954887218, |
|
"grad_norm": 0.6113621459091085, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5553, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1428571428571428, |
|
"grad_norm": 0.6400931476209645, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5564, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.1729323308270676, |
|
"grad_norm": 0.6192317007821239, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5605, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.2030075187969924, |
|
"grad_norm": 0.5237078678145406, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5519, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.2330827067669172, |
|
"grad_norm": 0.5140051147329998, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5574, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.263157894736842, |
|
"grad_norm": 0.6369697944320184, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5541, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.2932330827067668, |
|
"grad_norm": 0.5985062376730702, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5582, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.3233082706766917, |
|
"grad_norm": 0.48850272868134653, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5532, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.3533834586466165, |
|
"grad_norm": 0.5920840023317582, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5571, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.3834586466165413, |
|
"grad_norm": 0.5287134171137652, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5487, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.413533834586466, |
|
"grad_norm": 0.48569117162380887, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5586, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.443609022556391, |
|
"grad_norm": 0.5524909003831754, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5528, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.4736842105263157, |
|
"grad_norm": 0.45528161895298735, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5624, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.5037593984962405, |
|
"grad_norm": 0.5654083015816137, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5565, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.5338345864661656, |
|
"grad_norm": 0.5087876058685868, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5587, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.5639097744360901, |
|
"grad_norm": 0.4290062049121276, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5493, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.5939849624060152, |
|
"grad_norm": 0.6235091434876753, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5603, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.6240601503759398, |
|
"grad_norm": 0.5314512113272476, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5564, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.6541353383458648, |
|
"grad_norm": 0.5049026603516721, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5605, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.6842105263157894, |
|
"grad_norm": 0.5650921119799542, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5601, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.7142857142857144, |
|
"grad_norm": 0.5193160484708012, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5554, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.744360902255639, |
|
"grad_norm": 0.5213313242863221, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5645, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.774436090225564, |
|
"grad_norm": 0.6327600595776153, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5536, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.8045112781954886, |
|
"grad_norm": 0.5461472827380875, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5625, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.8345864661654137, |
|
"grad_norm": 0.6541180795154581, |
|
"learning_rate": 5e-06, |
|
"loss": 0.554, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.8646616541353382, |
|
"grad_norm": 0.5572692340571848, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5529, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.8947368421052633, |
|
"grad_norm": 0.5724927689201437, |
|
"learning_rate": 5e-06, |
|
"loss": 0.547, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.9248120300751879, |
|
"grad_norm": 0.48899030621990486, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5508, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.954887218045113, |
|
"grad_norm": 0.4783796037203571, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5626, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.9849624060150375, |
|
"grad_norm": 0.4380429796439722, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5579, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.6012639403343201, |
|
"eval_runtime": 116.0368, |
|
"eval_samples_per_second": 77.2, |
|
"eval_steps_per_second": 0.603, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.0150375939849625, |
|
"grad_norm": 0.8195061236201225, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5349, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.045112781954887, |
|
"grad_norm": 0.5727072462132908, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5023, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.075187969924812, |
|
"grad_norm": 0.5716937306059446, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5025, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.1052631578947367, |
|
"grad_norm": 0.5936120169922254, |
|
"learning_rate": 5e-06, |
|
"loss": 0.502, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.1353383458646618, |
|
"grad_norm": 0.5323072505815232, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5005, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.1654135338345863, |
|
"grad_norm": 0.5082968360924967, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5037, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.1954887218045114, |
|
"grad_norm": 0.5227485772254855, |
|
"learning_rate": 5e-06, |
|
"loss": 0.4989, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.225563909774436, |
|
"grad_norm": 0.6533374147188858, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5051, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.255639097744361, |
|
"grad_norm": 0.5912234992209703, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5048, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.2857142857142856, |
|
"grad_norm": 0.5434227115064355, |
|
"learning_rate": 5e-06, |
|
"loss": 0.508, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.3157894736842106, |
|
"grad_norm": 0.5720087264790767, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5099, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.345864661654135, |
|
"grad_norm": 0.6089476678610672, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5024, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.3759398496240602, |
|
"grad_norm": 0.5508640915356684, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5115, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.406015037593985, |
|
"grad_norm": 0.5118656252283122, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5051, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.43609022556391, |
|
"grad_norm": 0.5065404177593222, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5094, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.4661654135338344, |
|
"grad_norm": 0.5181429354716248, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5106, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.4962406015037595, |
|
"grad_norm": 0.5322514457543749, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5109, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.526315789473684, |
|
"grad_norm": 0.6196654734671412, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5061, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.556390977443609, |
|
"grad_norm": 0.5256550356188423, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5096, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.5864661654135337, |
|
"grad_norm": 0.4721653082846406, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5091, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.6165413533834587, |
|
"grad_norm": 0.5461766886120224, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5125, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.6466165413533833, |
|
"grad_norm": 0.8881592761313983, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5133, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.6766917293233083, |
|
"grad_norm": 0.5215263169318092, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5137, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.706766917293233, |
|
"grad_norm": 0.5970816519010325, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5085, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.736842105263158, |
|
"grad_norm": 0.5663757119297533, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5136, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.7669172932330826, |
|
"grad_norm": 0.48531820284070437, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5105, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.7969924812030076, |
|
"grad_norm": 0.5235848494113168, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5106, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.827067669172932, |
|
"grad_norm": 0.6230274670862681, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5109, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.857142857142857, |
|
"grad_norm": 0.5464639670084257, |
|
"learning_rate": 5e-06, |
|
"loss": 0.502, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.887218045112782, |
|
"grad_norm": 0.5388668922788399, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5059, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.917293233082707, |
|
"grad_norm": 0.5634658215547509, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5147, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.9473684210526314, |
|
"grad_norm": 0.5876554762818471, |
|
"learning_rate": 5e-06, |
|
"loss": 0.507, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.9774436090225564, |
|
"grad_norm": 0.5816222818084577, |
|
"learning_rate": 5e-06, |
|
"loss": 0.5116, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.995488721804511, |
|
"eval_loss": 0.605595588684082, |
|
"eval_runtime": 114.1196, |
|
"eval_samples_per_second": 78.497, |
|
"eval_steps_per_second": 0.613, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 2.995488721804511, |
|
"step": 996, |
|
"total_flos": 1667918337146880.0, |
|
"train_loss": 0.5664463261045126, |
|
"train_runtime": 17372.9601, |
|
"train_samples_per_second": 29.39, |
|
"train_steps_per_second": 0.057 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 996, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1667918337146880.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|