{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.995488721804511, "eval_steps": 500, "global_step": 996, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03007518796992481, "grad_norm": 5.823914796640075, "learning_rate": 5e-06, "loss": 0.801, "step": 10 }, { "epoch": 0.06015037593984962, "grad_norm": 2.3399969132192253, "learning_rate": 5e-06, "loss": 0.7172, "step": 20 }, { "epoch": 0.09022556390977443, "grad_norm": 0.8783578572732146, "learning_rate": 5e-06, "loss": 0.6923, "step": 30 }, { "epoch": 0.12030075187969924, "grad_norm": 1.7232554735132826, "learning_rate": 5e-06, "loss": 0.6834, "step": 40 }, { "epoch": 0.15037593984962405, "grad_norm": 1.6494035164725938, "learning_rate": 5e-06, "loss": 0.6655, "step": 50 }, { "epoch": 0.18045112781954886, "grad_norm": 0.9942658093552336, "learning_rate": 5e-06, "loss": 0.6568, "step": 60 }, { "epoch": 0.21052631578947367, "grad_norm": 0.8609069727317351, "learning_rate": 5e-06, "loss": 0.6548, "step": 70 }, { "epoch": 0.24060150375939848, "grad_norm": 0.8740060444495542, "learning_rate": 5e-06, "loss": 0.6454, "step": 80 }, { "epoch": 0.2706766917293233, "grad_norm": 1.0114201009544495, "learning_rate": 5e-06, "loss": 0.6331, "step": 90 }, { "epoch": 0.3007518796992481, "grad_norm": 0.782353210012559, "learning_rate": 5e-06, "loss": 0.6267, "step": 100 }, { "epoch": 0.3308270676691729, "grad_norm": 0.6061308240718153, "learning_rate": 5e-06, "loss": 0.6305, "step": 110 }, { "epoch": 0.3609022556390977, "grad_norm": 0.663886451339362, "learning_rate": 5e-06, "loss": 0.6246, "step": 120 }, { "epoch": 0.39097744360902253, "grad_norm": 0.8509285444900951, "learning_rate": 5e-06, "loss": 0.6305, "step": 130 }, { "epoch": 0.42105263157894735, "grad_norm": 0.650634026198435, "learning_rate": 5e-06, "loss": 0.6291, "step": 140 }, { "epoch": 0.45112781954887216, "grad_norm": 0.7050446482120244, "learning_rate": 5e-06, "loss": 0.6279, "step": 150 }, { "epoch": 0.48120300751879697, "grad_norm": 0.5260864128397903, "learning_rate": 5e-06, "loss": 0.6167, "step": 160 }, { "epoch": 0.5112781954887218, "grad_norm": 0.5159636362174167, "learning_rate": 5e-06, "loss": 0.6189, "step": 170 }, { "epoch": 0.5413533834586466, "grad_norm": 0.576027072578746, "learning_rate": 5e-06, "loss": 0.6096, "step": 180 }, { "epoch": 0.5714285714285714, "grad_norm": 0.6020714354457843, "learning_rate": 5e-06, "loss": 0.6207, "step": 190 }, { "epoch": 0.6015037593984962, "grad_norm": 0.8904744763001984, "learning_rate": 5e-06, "loss": 0.6143, "step": 200 }, { "epoch": 0.631578947368421, "grad_norm": 0.5645606661034147, "learning_rate": 5e-06, "loss": 0.619, "step": 210 }, { "epoch": 0.6616541353383458, "grad_norm": 0.7456282489942, "learning_rate": 5e-06, "loss": 0.6229, "step": 220 }, { "epoch": 0.6917293233082706, "grad_norm": 0.47671003637879944, "learning_rate": 5e-06, "loss": 0.6141, "step": 230 }, { "epoch": 0.7218045112781954, "grad_norm": 0.5712162016443497, "learning_rate": 5e-06, "loss": 0.6148, "step": 240 }, { "epoch": 0.7518796992481203, "grad_norm": 0.5787527818544967, "learning_rate": 5e-06, "loss": 0.6112, "step": 250 }, { "epoch": 0.7819548872180451, "grad_norm": 0.7727895820276254, "learning_rate": 5e-06, "loss": 0.6124, "step": 260 }, { "epoch": 0.8120300751879699, "grad_norm": 0.6627350412484229, "learning_rate": 5e-06, "loss": 0.6101, "step": 270 }, { "epoch": 0.8421052631578947, "grad_norm": 0.49307733921025576, "learning_rate": 5e-06, "loss": 0.6101, "step": 280 }, { "epoch": 0.8721804511278195, "grad_norm": 0.5299740752648865, "learning_rate": 5e-06, "loss": 0.6074, "step": 290 }, { "epoch": 0.9022556390977443, "grad_norm": 0.9821666854504033, "learning_rate": 5e-06, "loss": 0.6035, "step": 300 }, { "epoch": 0.9323308270676691, "grad_norm": 0.5537120584191855, "learning_rate": 5e-06, "loss": 0.6037, "step": 310 }, { "epoch": 0.9624060150375939, "grad_norm": 0.5346047989400851, "learning_rate": 5e-06, "loss": 0.6128, "step": 320 }, { "epoch": 0.9924812030075187, "grad_norm": 0.5177704737274331, "learning_rate": 5e-06, "loss": 0.6043, "step": 330 }, { "epoch": 0.9984962406015038, "eval_loss": 0.6096732020378113, "eval_runtime": 116.9563, "eval_samples_per_second": 76.593, "eval_steps_per_second": 0.599, "step": 332 }, { "epoch": 1.0225563909774436, "grad_norm": 0.4996873605367385, "learning_rate": 5e-06, "loss": 0.58, "step": 340 }, { "epoch": 1.0526315789473684, "grad_norm": 0.59979801353788, "learning_rate": 5e-06, "loss": 0.5596, "step": 350 }, { "epoch": 1.0827067669172932, "grad_norm": 0.645810347625552, "learning_rate": 5e-06, "loss": 0.5591, "step": 360 }, { "epoch": 1.112781954887218, "grad_norm": 0.6113621459091085, "learning_rate": 5e-06, "loss": 0.5553, "step": 370 }, { "epoch": 1.1428571428571428, "grad_norm": 0.6400931476209645, "learning_rate": 5e-06, "loss": 0.5564, "step": 380 }, { "epoch": 1.1729323308270676, "grad_norm": 0.6192317007821239, "learning_rate": 5e-06, "loss": 0.5605, "step": 390 }, { "epoch": 1.2030075187969924, "grad_norm": 0.5237078678145406, "learning_rate": 5e-06, "loss": 0.5519, "step": 400 }, { "epoch": 1.2330827067669172, "grad_norm": 0.5140051147329998, "learning_rate": 5e-06, "loss": 0.5574, "step": 410 }, { "epoch": 1.263157894736842, "grad_norm": 0.6369697944320184, "learning_rate": 5e-06, "loss": 0.5541, "step": 420 }, { "epoch": 1.2932330827067668, "grad_norm": 0.5985062376730702, "learning_rate": 5e-06, "loss": 0.5582, "step": 430 }, { "epoch": 1.3233082706766917, "grad_norm": 0.48850272868134653, "learning_rate": 5e-06, "loss": 0.5532, "step": 440 }, { "epoch": 1.3533834586466165, "grad_norm": 0.5920840023317582, "learning_rate": 5e-06, "loss": 0.5571, "step": 450 }, { "epoch": 1.3834586466165413, "grad_norm": 0.5287134171137652, "learning_rate": 5e-06, "loss": 0.5487, "step": 460 }, { "epoch": 1.413533834586466, "grad_norm": 0.48569117162380887, "learning_rate": 5e-06, "loss": 0.5586, "step": 470 }, { "epoch": 1.443609022556391, "grad_norm": 0.5524909003831754, "learning_rate": 5e-06, "loss": 0.5528, "step": 480 }, { "epoch": 1.4736842105263157, "grad_norm": 0.45528161895298735, "learning_rate": 5e-06, "loss": 0.5624, "step": 490 }, { "epoch": 1.5037593984962405, "grad_norm": 0.5654083015816137, "learning_rate": 5e-06, "loss": 0.5565, "step": 500 }, { "epoch": 1.5338345864661656, "grad_norm": 0.5087876058685868, "learning_rate": 5e-06, "loss": 0.5587, "step": 510 }, { "epoch": 1.5639097744360901, "grad_norm": 0.4290062049121276, "learning_rate": 5e-06, "loss": 0.5493, "step": 520 }, { "epoch": 1.5939849624060152, "grad_norm": 0.6235091434876753, "learning_rate": 5e-06, "loss": 0.5603, "step": 530 }, { "epoch": 1.6240601503759398, "grad_norm": 0.5314512113272476, "learning_rate": 5e-06, "loss": 0.5564, "step": 540 }, { "epoch": 1.6541353383458648, "grad_norm": 0.5049026603516721, "learning_rate": 5e-06, "loss": 0.5605, "step": 550 }, { "epoch": 1.6842105263157894, "grad_norm": 0.5650921119799542, "learning_rate": 5e-06, "loss": 0.5601, "step": 560 }, { "epoch": 1.7142857142857144, "grad_norm": 0.5193160484708012, "learning_rate": 5e-06, "loss": 0.5554, "step": 570 }, { "epoch": 1.744360902255639, "grad_norm": 0.5213313242863221, "learning_rate": 5e-06, "loss": 0.5645, "step": 580 }, { "epoch": 1.774436090225564, "grad_norm": 0.6327600595776153, "learning_rate": 5e-06, "loss": 0.5536, "step": 590 }, { "epoch": 1.8045112781954886, "grad_norm": 0.5461472827380875, "learning_rate": 5e-06, "loss": 0.5625, "step": 600 }, { "epoch": 1.8345864661654137, "grad_norm": 0.6541180795154581, "learning_rate": 5e-06, "loss": 0.554, "step": 610 }, { "epoch": 1.8646616541353382, "grad_norm": 0.5572692340571848, "learning_rate": 5e-06, "loss": 0.5529, "step": 620 }, { "epoch": 1.8947368421052633, "grad_norm": 0.5724927689201437, "learning_rate": 5e-06, "loss": 0.547, "step": 630 }, { "epoch": 1.9248120300751879, "grad_norm": 0.48899030621990486, "learning_rate": 5e-06, "loss": 0.5508, "step": 640 }, { "epoch": 1.954887218045113, "grad_norm": 0.4783796037203571, "learning_rate": 5e-06, "loss": 0.5626, "step": 650 }, { "epoch": 1.9849624060150375, "grad_norm": 0.4380429796439722, "learning_rate": 5e-06, "loss": 0.5579, "step": 660 }, { "epoch": 2.0, "eval_loss": 0.6012639403343201, "eval_runtime": 116.0368, "eval_samples_per_second": 77.2, "eval_steps_per_second": 0.603, "step": 665 }, { "epoch": 2.0150375939849625, "grad_norm": 0.8195061236201225, "learning_rate": 5e-06, "loss": 0.5349, "step": 670 }, { "epoch": 2.045112781954887, "grad_norm": 0.5727072462132908, "learning_rate": 5e-06, "loss": 0.5023, "step": 680 }, { "epoch": 2.075187969924812, "grad_norm": 0.5716937306059446, "learning_rate": 5e-06, "loss": 0.5025, "step": 690 }, { "epoch": 2.1052631578947367, "grad_norm": 0.5936120169922254, "learning_rate": 5e-06, "loss": 0.502, "step": 700 }, { "epoch": 2.1353383458646618, "grad_norm": 0.5323072505815232, "learning_rate": 5e-06, "loss": 0.5005, "step": 710 }, { "epoch": 2.1654135338345863, "grad_norm": 0.5082968360924967, "learning_rate": 5e-06, "loss": 0.5037, "step": 720 }, { "epoch": 2.1954887218045114, "grad_norm": 0.5227485772254855, "learning_rate": 5e-06, "loss": 0.4989, "step": 730 }, { "epoch": 2.225563909774436, "grad_norm": 0.6533374147188858, "learning_rate": 5e-06, "loss": 0.5051, "step": 740 }, { "epoch": 2.255639097744361, "grad_norm": 0.5912234992209703, "learning_rate": 5e-06, "loss": 0.5048, "step": 750 }, { "epoch": 2.2857142857142856, "grad_norm": 0.5434227115064355, "learning_rate": 5e-06, "loss": 0.508, "step": 760 }, { "epoch": 2.3157894736842106, "grad_norm": 0.5720087264790767, "learning_rate": 5e-06, "loss": 0.5099, "step": 770 }, { "epoch": 2.345864661654135, "grad_norm": 0.6089476678610672, "learning_rate": 5e-06, "loss": 0.5024, "step": 780 }, { "epoch": 2.3759398496240602, "grad_norm": 0.5508640915356684, "learning_rate": 5e-06, "loss": 0.5115, "step": 790 }, { "epoch": 2.406015037593985, "grad_norm": 0.5118656252283122, "learning_rate": 5e-06, "loss": 0.5051, "step": 800 }, { "epoch": 2.43609022556391, "grad_norm": 0.5065404177593222, "learning_rate": 5e-06, "loss": 0.5094, "step": 810 }, { "epoch": 2.4661654135338344, "grad_norm": 0.5181429354716248, "learning_rate": 5e-06, "loss": 0.5106, "step": 820 }, { "epoch": 2.4962406015037595, "grad_norm": 0.5322514457543749, "learning_rate": 5e-06, "loss": 0.5109, "step": 830 }, { "epoch": 2.526315789473684, "grad_norm": 0.6196654734671412, "learning_rate": 5e-06, "loss": 0.5061, "step": 840 }, { "epoch": 2.556390977443609, "grad_norm": 0.5256550356188423, "learning_rate": 5e-06, "loss": 0.5096, "step": 850 }, { "epoch": 2.5864661654135337, "grad_norm": 0.4721653082846406, "learning_rate": 5e-06, "loss": 0.5091, "step": 860 }, { "epoch": 2.6165413533834587, "grad_norm": 0.5461766886120224, "learning_rate": 5e-06, "loss": 0.5125, "step": 870 }, { "epoch": 2.6466165413533833, "grad_norm": 0.8881592761313983, "learning_rate": 5e-06, "loss": 0.5133, "step": 880 }, { "epoch": 2.6766917293233083, "grad_norm": 0.5215263169318092, "learning_rate": 5e-06, "loss": 0.5137, "step": 890 }, { "epoch": 2.706766917293233, "grad_norm": 0.5970816519010325, "learning_rate": 5e-06, "loss": 0.5085, "step": 900 }, { "epoch": 2.736842105263158, "grad_norm": 0.5663757119297533, "learning_rate": 5e-06, "loss": 0.5136, "step": 910 }, { "epoch": 2.7669172932330826, "grad_norm": 0.48531820284070437, "learning_rate": 5e-06, "loss": 0.5105, "step": 920 }, { "epoch": 2.7969924812030076, "grad_norm": 0.5235848494113168, "learning_rate": 5e-06, "loss": 0.5106, "step": 930 }, { "epoch": 2.827067669172932, "grad_norm": 0.6230274670862681, "learning_rate": 5e-06, "loss": 0.5109, "step": 940 }, { "epoch": 2.857142857142857, "grad_norm": 0.5464639670084257, "learning_rate": 5e-06, "loss": 0.502, "step": 950 }, { "epoch": 2.887218045112782, "grad_norm": 0.5388668922788399, "learning_rate": 5e-06, "loss": 0.5059, "step": 960 }, { "epoch": 2.917293233082707, "grad_norm": 0.5634658215547509, "learning_rate": 5e-06, "loss": 0.5147, "step": 970 }, { "epoch": 2.9473684210526314, "grad_norm": 0.5876554762818471, "learning_rate": 5e-06, "loss": 0.507, "step": 980 }, { "epoch": 2.9774436090225564, "grad_norm": 0.5816222818084577, "learning_rate": 5e-06, "loss": 0.5116, "step": 990 }, { "epoch": 2.995488721804511, "eval_loss": 0.605595588684082, "eval_runtime": 114.1196, "eval_samples_per_second": 78.497, "eval_steps_per_second": 0.613, "step": 996 }, { "epoch": 2.995488721804511, "step": 996, "total_flos": 1667918337146880.0, "train_loss": 0.5664463261045126, "train_runtime": 17372.9601, "train_samples_per_second": 29.39, "train_steps_per_second": 0.057 } ], "logging_steps": 10, "max_steps": 996, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1667918337146880.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }