{ "best_metric": 0.8894595801697186, "best_model_checkpoint": "tsec_vit_model/checkpoint-2520", "epoch": 10.0, "eval_steps": 500, "global_step": 2800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03571428571428571, "grad_norm": 0.632759690284729, "learning_rate": 1.7857142857142857e-06, "loss": 0.7005, "step": 10 }, { "epoch": 0.07142857142857142, "grad_norm": 0.6090130805969238, "learning_rate": 3.5714285714285714e-06, "loss": 0.6888, "step": 20 }, { "epoch": 0.10714285714285714, "grad_norm": 0.5027034878730774, "learning_rate": 5.357142857142857e-06, "loss": 0.6841, "step": 30 }, { "epoch": 0.14285714285714285, "grad_norm": 0.7158761620521545, "learning_rate": 7.142857142857143e-06, "loss": 0.6711, "step": 40 }, { "epoch": 0.17857142857142858, "grad_norm": 1.0677605867385864, "learning_rate": 8.92857142857143e-06, "loss": 0.6439, "step": 50 }, { "epoch": 0.21428571428571427, "grad_norm": 0.716444194316864, "learning_rate": 1.0714285714285714e-05, "loss": 0.6388, "step": 60 }, { "epoch": 0.25, "grad_norm": 0.9646293520927429, "learning_rate": 1.25e-05, "loss": 0.5942, "step": 70 }, { "epoch": 0.2857142857142857, "grad_norm": 1.5171552896499634, "learning_rate": 1.4285714285714285e-05, "loss": 0.587, "step": 80 }, { "epoch": 0.32142857142857145, "grad_norm": 1.2230862379074097, "learning_rate": 1.6071428571428572e-05, "loss": 0.5882, "step": 90 }, { "epoch": 0.35714285714285715, "grad_norm": 1.0726110935211182, "learning_rate": 1.785714285714286e-05, "loss": 0.5506, "step": 100 }, { "epoch": 0.39285714285714285, "grad_norm": 2.488288640975952, "learning_rate": 1.9642857142857145e-05, "loss": 0.5487, "step": 110 }, { "epoch": 0.42857142857142855, "grad_norm": 2.186500310897827, "learning_rate": 2.1428571428571428e-05, "loss": 0.5138, "step": 120 }, { "epoch": 0.4642857142857143, "grad_norm": 2.18605637550354, "learning_rate": 2.3214285714285715e-05, "loss": 0.5203, "step": 130 }, { "epoch": 0.5, "grad_norm": 1.5495938062667847, "learning_rate": 2.5e-05, "loss": 0.5006, "step": 140 }, { "epoch": 0.5357142857142857, "grad_norm": 1.3939014673233032, "learning_rate": 2.6785714285714288e-05, "loss": 0.5171, "step": 150 }, { "epoch": 0.5714285714285714, "grad_norm": 1.4516574144363403, "learning_rate": 2.857142857142857e-05, "loss": 0.4882, "step": 160 }, { "epoch": 0.6071428571428571, "grad_norm": 1.5352245569229126, "learning_rate": 3.0357142857142857e-05, "loss": 0.4809, "step": 170 }, { "epoch": 0.6428571428571429, "grad_norm": 2.6514480113983154, "learning_rate": 3.2142857142857144e-05, "loss": 0.458, "step": 180 }, { "epoch": 0.6785714285714286, "grad_norm": 2.1452901363372803, "learning_rate": 3.392857142857143e-05, "loss": 0.478, "step": 190 }, { "epoch": 0.7142857142857143, "grad_norm": 1.5536608695983887, "learning_rate": 3.571428571428572e-05, "loss": 0.4851, "step": 200 }, { "epoch": 0.75, "grad_norm": 1.7489731311798096, "learning_rate": 3.7500000000000003e-05, "loss": 0.4608, "step": 210 }, { "epoch": 0.7857142857142857, "grad_norm": 1.6109980344772339, "learning_rate": 3.928571428571429e-05, "loss": 0.4611, "step": 220 }, { "epoch": 0.8214285714285714, "grad_norm": 2.391719102859497, "learning_rate": 4.107142857142857e-05, "loss": 0.3998, "step": 230 }, { "epoch": 0.8571428571428571, "grad_norm": 3.6933391094207764, "learning_rate": 4.2857142857142856e-05, "loss": 0.4624, "step": 240 }, { "epoch": 0.8928571428571429, "grad_norm": 2.0588369369506836, "learning_rate": 4.464285714285715e-05, "loss": 0.4437, "step": 250 }, { "epoch": 0.9285714285714286, "grad_norm": 2.831156015396118, "learning_rate": 4.642857142857143e-05, "loss": 0.4652, "step": 260 }, { "epoch": 0.9642857142857143, "grad_norm": 2.0229790210723877, "learning_rate": 4.8214285714285716e-05, "loss": 0.4037, "step": 270 }, { "epoch": 1.0, "grad_norm": 2.1834700107574463, "learning_rate": 5e-05, "loss": 0.4387, "step": 280 }, { "epoch": 1.0, "eval_accuracy": 0.8150960250111657, "eval_loss": 0.4179099500179291, "eval_runtime": 110.4006, "eval_samples_per_second": 40.561, "eval_steps_per_second": 2.536, "step": 280 }, { "epoch": 1.0357142857142858, "grad_norm": 1.7746831178665161, "learning_rate": 4.9801587301587306e-05, "loss": 0.4069, "step": 290 }, { "epoch": 1.0714285714285714, "grad_norm": 2.6947922706604004, "learning_rate": 4.960317460317461e-05, "loss": 0.4334, "step": 300 }, { "epoch": 1.1071428571428572, "grad_norm": 1.4933383464813232, "learning_rate": 4.940476190476191e-05, "loss": 0.4409, "step": 310 }, { "epoch": 1.1428571428571428, "grad_norm": 1.8830089569091797, "learning_rate": 4.9206349206349204e-05, "loss": 0.4065, "step": 320 }, { "epoch": 1.1785714285714286, "grad_norm": 1.978843092918396, "learning_rate": 4.900793650793651e-05, "loss": 0.4273, "step": 330 }, { "epoch": 1.2142857142857142, "grad_norm": 1.2928149700164795, "learning_rate": 4.880952380952381e-05, "loss": 0.4376, "step": 340 }, { "epoch": 1.25, "grad_norm": 1.5401108264923096, "learning_rate": 4.8611111111111115e-05, "loss": 0.4061, "step": 350 }, { "epoch": 1.2857142857142856, "grad_norm": 2.296520709991455, "learning_rate": 4.841269841269841e-05, "loss": 0.433, "step": 360 }, { "epoch": 1.3214285714285714, "grad_norm": 1.2640879154205322, "learning_rate": 4.8214285714285716e-05, "loss": 0.4142, "step": 370 }, { "epoch": 1.3571428571428572, "grad_norm": 1.3309462070465088, "learning_rate": 4.801587301587302e-05, "loss": 0.4222, "step": 380 }, { "epoch": 1.3928571428571428, "grad_norm": 1.8462568521499634, "learning_rate": 4.781746031746032e-05, "loss": 0.4483, "step": 390 }, { "epoch": 1.4285714285714286, "grad_norm": 1.074401617050171, "learning_rate": 4.761904761904762e-05, "loss": 0.3762, "step": 400 }, { "epoch": 1.4642857142857144, "grad_norm": 1.5691473484039307, "learning_rate": 4.7420634920634924e-05, "loss": 0.415, "step": 410 }, { "epoch": 1.5, "grad_norm": 1.2222189903259277, "learning_rate": 4.722222222222222e-05, "loss": 0.3862, "step": 420 }, { "epoch": 1.5357142857142856, "grad_norm": 1.4210410118103027, "learning_rate": 4.7023809523809525e-05, "loss": 0.352, "step": 430 }, { "epoch": 1.5714285714285714, "grad_norm": 1.632934331893921, "learning_rate": 4.682539682539683e-05, "loss": 0.3873, "step": 440 }, { "epoch": 1.6071428571428572, "grad_norm": 1.3977652788162231, "learning_rate": 4.662698412698413e-05, "loss": 0.3763, "step": 450 }, { "epoch": 1.6428571428571428, "grad_norm": 2.74967622756958, "learning_rate": 4.642857142857143e-05, "loss": 0.3823, "step": 460 }, { "epoch": 1.6785714285714286, "grad_norm": 1.1579408645629883, "learning_rate": 4.623015873015873e-05, "loss": 0.3967, "step": 470 }, { "epoch": 1.7142857142857144, "grad_norm": 2.7291440963745117, "learning_rate": 4.603174603174603e-05, "loss": 0.4201, "step": 480 }, { "epoch": 1.75, "grad_norm": 1.0126214027404785, "learning_rate": 4.5833333333333334e-05, "loss": 0.3598, "step": 490 }, { "epoch": 1.7857142857142856, "grad_norm": 1.6705009937286377, "learning_rate": 4.563492063492064e-05, "loss": 0.366, "step": 500 }, { "epoch": 1.8214285714285714, "grad_norm": 1.6749546527862549, "learning_rate": 4.543650793650794e-05, "loss": 0.3854, "step": 510 }, { "epoch": 1.8571428571428572, "grad_norm": 1.0892480611801147, "learning_rate": 4.523809523809524e-05, "loss": 0.3563, "step": 520 }, { "epoch": 1.8928571428571428, "grad_norm": 1.4169703722000122, "learning_rate": 4.503968253968254e-05, "loss": 0.3903, "step": 530 }, { "epoch": 1.9285714285714286, "grad_norm": 1.393864631652832, "learning_rate": 4.4841269841269846e-05, "loss": 0.3915, "step": 540 }, { "epoch": 1.9642857142857144, "grad_norm": 1.6008644104003906, "learning_rate": 4.464285714285715e-05, "loss": 0.3665, "step": 550 }, { "epoch": 2.0, "grad_norm": 1.0977710485458374, "learning_rate": 4.4444444444444447e-05, "loss": 0.4239, "step": 560 }, { "epoch": 2.0, "eval_accuracy": 0.8398838767306833, "eval_loss": 0.3610800504684448, "eval_runtime": 114.7581, "eval_samples_per_second": 39.021, "eval_steps_per_second": 2.44, "step": 560 }, { "epoch": 2.0357142857142856, "grad_norm": 1.4717609882354736, "learning_rate": 4.4246031746031744e-05, "loss": 0.3609, "step": 570 }, { "epoch": 2.0714285714285716, "grad_norm": 1.0943901538848877, "learning_rate": 4.404761904761905e-05, "loss": 0.3379, "step": 580 }, { "epoch": 2.107142857142857, "grad_norm": 1.8704146146774292, "learning_rate": 4.384920634920635e-05, "loss": 0.364, "step": 590 }, { "epoch": 2.142857142857143, "grad_norm": 1.2977815866470337, "learning_rate": 4.3650793650793655e-05, "loss": 0.3834, "step": 600 }, { "epoch": 2.1785714285714284, "grad_norm": 1.6529649496078491, "learning_rate": 4.345238095238096e-05, "loss": 0.3688, "step": 610 }, { "epoch": 2.2142857142857144, "grad_norm": 1.1914763450622559, "learning_rate": 4.3253968253968256e-05, "loss": 0.3548, "step": 620 }, { "epoch": 2.25, "grad_norm": 1.70374596118927, "learning_rate": 4.305555555555556e-05, "loss": 0.3323, "step": 630 }, { "epoch": 2.2857142857142856, "grad_norm": 1.1331793069839478, "learning_rate": 4.2857142857142856e-05, "loss": 0.3494, "step": 640 }, { "epoch": 2.3214285714285716, "grad_norm": 1.3151099681854248, "learning_rate": 4.265873015873016e-05, "loss": 0.3177, "step": 650 }, { "epoch": 2.357142857142857, "grad_norm": 1.7670910358428955, "learning_rate": 4.2460317460317464e-05, "loss": 0.3145, "step": 660 }, { "epoch": 2.392857142857143, "grad_norm": 1.441829800605774, "learning_rate": 4.226190476190476e-05, "loss": 0.2968, "step": 670 }, { "epoch": 2.4285714285714284, "grad_norm": 1.5992521047592163, "learning_rate": 4.2063492063492065e-05, "loss": 0.353, "step": 680 }, { "epoch": 2.4642857142857144, "grad_norm": 1.076711654663086, "learning_rate": 4.186507936507937e-05, "loss": 0.3461, "step": 690 }, { "epoch": 2.5, "grad_norm": 1.0764169692993164, "learning_rate": 4.166666666666667e-05, "loss": 0.3345, "step": 700 }, { "epoch": 2.5357142857142856, "grad_norm": 1.3735431432724, "learning_rate": 4.1468253968253976e-05, "loss": 0.3612, "step": 710 }, { "epoch": 2.571428571428571, "grad_norm": 1.4703452587127686, "learning_rate": 4.126984126984127e-05, "loss": 0.3444, "step": 720 }, { "epoch": 2.607142857142857, "grad_norm": 1.451257348060608, "learning_rate": 4.107142857142857e-05, "loss": 0.3287, "step": 730 }, { "epoch": 2.642857142857143, "grad_norm": 1.8615336418151855, "learning_rate": 4.0873015873015874e-05, "loss": 0.3419, "step": 740 }, { "epoch": 2.678571428571429, "grad_norm": 2.925323963165283, "learning_rate": 4.067460317460318e-05, "loss": 0.3217, "step": 750 }, { "epoch": 2.7142857142857144, "grad_norm": 1.7260726690292358, "learning_rate": 4.047619047619048e-05, "loss": 0.3584, "step": 760 }, { "epoch": 2.75, "grad_norm": 1.4659258127212524, "learning_rate": 4.027777777777778e-05, "loss": 0.3821, "step": 770 }, { "epoch": 2.7857142857142856, "grad_norm": 1.7859466075897217, "learning_rate": 4.007936507936508e-05, "loss": 0.3595, "step": 780 }, { "epoch": 2.821428571428571, "grad_norm": 1.2320784330368042, "learning_rate": 3.9880952380952386e-05, "loss": 0.3764, "step": 790 }, { "epoch": 2.857142857142857, "grad_norm": 1.4097161293029785, "learning_rate": 3.968253968253968e-05, "loss": 0.3282, "step": 800 }, { "epoch": 2.892857142857143, "grad_norm": 1.185285210609436, "learning_rate": 3.9484126984126986e-05, "loss": 0.3214, "step": 810 }, { "epoch": 2.928571428571429, "grad_norm": 1.3706032037734985, "learning_rate": 3.928571428571429e-05, "loss": 0.3092, "step": 820 }, { "epoch": 2.9642857142857144, "grad_norm": 2.1342084407806396, "learning_rate": 3.908730158730159e-05, "loss": 0.3274, "step": 830 }, { "epoch": 3.0, "grad_norm": 1.438025712966919, "learning_rate": 3.888888888888889e-05, "loss": 0.3148, "step": 840 }, { "epoch": 3.0, "eval_accuracy": 0.8599821348816435, "eval_loss": 0.31560027599334717, "eval_runtime": 116.0276, "eval_samples_per_second": 38.594, "eval_steps_per_second": 2.413, "step": 840 }, { "epoch": 3.0357142857142856, "grad_norm": 1.6471174955368042, "learning_rate": 3.8690476190476195e-05, "loss": 0.3198, "step": 850 }, { "epoch": 3.0714285714285716, "grad_norm": 1.5575629472732544, "learning_rate": 3.84920634920635e-05, "loss": 0.3127, "step": 860 }, { "epoch": 3.107142857142857, "grad_norm": 1.1578933000564575, "learning_rate": 3.8293650793650795e-05, "loss": 0.3416, "step": 870 }, { "epoch": 3.142857142857143, "grad_norm": 1.9167553186416626, "learning_rate": 3.809523809523809e-05, "loss": 0.3626, "step": 880 }, { "epoch": 3.1785714285714284, "grad_norm": 1.038451910018921, "learning_rate": 3.7896825396825396e-05, "loss": 0.336, "step": 890 }, { "epoch": 3.2142857142857144, "grad_norm": 1.2739081382751465, "learning_rate": 3.76984126984127e-05, "loss": 0.2838, "step": 900 }, { "epoch": 3.25, "grad_norm": 1.2061238288879395, "learning_rate": 3.7500000000000003e-05, "loss": 0.3099, "step": 910 }, { "epoch": 3.2857142857142856, "grad_norm": 1.5365681648254395, "learning_rate": 3.730158730158731e-05, "loss": 0.3163, "step": 920 }, { "epoch": 3.3214285714285716, "grad_norm": 2.33838152885437, "learning_rate": 3.7103174603174604e-05, "loss": 0.3109, "step": 930 }, { "epoch": 3.357142857142857, "grad_norm": 1.2280935049057007, "learning_rate": 3.690476190476191e-05, "loss": 0.3104, "step": 940 }, { "epoch": 3.392857142857143, "grad_norm": 1.2506486177444458, "learning_rate": 3.6706349206349205e-05, "loss": 0.339, "step": 950 }, { "epoch": 3.4285714285714284, "grad_norm": 1.0430885553359985, "learning_rate": 3.650793650793651e-05, "loss": 0.3015, "step": 960 }, { "epoch": 3.4642857142857144, "grad_norm": 1.5323641300201416, "learning_rate": 3.630952380952381e-05, "loss": 0.3408, "step": 970 }, { "epoch": 3.5, "grad_norm": 1.0826184749603271, "learning_rate": 3.611111111111111e-05, "loss": 0.3061, "step": 980 }, { "epoch": 3.5357142857142856, "grad_norm": 2.019437313079834, "learning_rate": 3.591269841269841e-05, "loss": 0.3465, "step": 990 }, { "epoch": 3.571428571428571, "grad_norm": 1.5877951383590698, "learning_rate": 3.571428571428572e-05, "loss": 0.3023, "step": 1000 }, { "epoch": 3.607142857142857, "grad_norm": 1.5832830667495728, "learning_rate": 3.551587301587302e-05, "loss": 0.262, "step": 1010 }, { "epoch": 3.642857142857143, "grad_norm": 1.5758150815963745, "learning_rate": 3.5317460317460324e-05, "loss": 0.3337, "step": 1020 }, { "epoch": 3.678571428571429, "grad_norm": 1.255638837814331, "learning_rate": 3.511904761904762e-05, "loss": 0.2857, "step": 1030 }, { "epoch": 3.7142857142857144, "grad_norm": 1.5360593795776367, "learning_rate": 3.492063492063492e-05, "loss": 0.2964, "step": 1040 }, { "epoch": 3.75, "grad_norm": 1.0826270580291748, "learning_rate": 3.472222222222222e-05, "loss": 0.2872, "step": 1050 }, { "epoch": 3.7857142857142856, "grad_norm": 1.6883878707885742, "learning_rate": 3.4523809523809526e-05, "loss": 0.3016, "step": 1060 }, { "epoch": 3.821428571428571, "grad_norm": 1.8018616437911987, "learning_rate": 3.432539682539683e-05, "loss": 0.3115, "step": 1070 }, { "epoch": 3.857142857142857, "grad_norm": 1.9364327192306519, "learning_rate": 3.412698412698413e-05, "loss": 0.2986, "step": 1080 }, { "epoch": 3.892857142857143, "grad_norm": 1.6606508493423462, "learning_rate": 3.392857142857143e-05, "loss": 0.3017, "step": 1090 }, { "epoch": 3.928571428571429, "grad_norm": 2.1849498748779297, "learning_rate": 3.3730158730158734e-05, "loss": 0.2779, "step": 1100 }, { "epoch": 3.9642857142857144, "grad_norm": 1.5579859018325806, "learning_rate": 3.353174603174603e-05, "loss": 0.3199, "step": 1110 }, { "epoch": 4.0, "grad_norm": 2.6287128925323486, "learning_rate": 3.3333333333333335e-05, "loss": 0.2988, "step": 1120 }, { "epoch": 4.0, "eval_accuracy": 0.8729343456900402, "eval_loss": 0.30018478631973267, "eval_runtime": 115.4434, "eval_samples_per_second": 38.79, "eval_steps_per_second": 2.425, "step": 1120 }, { "epoch": 4.035714285714286, "grad_norm": 1.5949360132217407, "learning_rate": 3.313492063492064e-05, "loss": 0.305, "step": 1130 }, { "epoch": 4.071428571428571, "grad_norm": 1.678801417350769, "learning_rate": 3.2936507936507936e-05, "loss": 0.2677, "step": 1140 }, { "epoch": 4.107142857142857, "grad_norm": 1.6015031337738037, "learning_rate": 3.273809523809524e-05, "loss": 0.3167, "step": 1150 }, { "epoch": 4.142857142857143, "grad_norm": 1.525894045829773, "learning_rate": 3.253968253968254e-05, "loss": 0.2745, "step": 1160 }, { "epoch": 4.178571428571429, "grad_norm": 1.2955286502838135, "learning_rate": 3.234126984126985e-05, "loss": 0.2925, "step": 1170 }, { "epoch": 4.214285714285714, "grad_norm": 2.1041815280914307, "learning_rate": 3.2142857142857144e-05, "loss": 0.2782, "step": 1180 }, { "epoch": 4.25, "grad_norm": 1.7325941324234009, "learning_rate": 3.194444444444444e-05, "loss": 0.302, "step": 1190 }, { "epoch": 4.285714285714286, "grad_norm": 1.9786794185638428, "learning_rate": 3.1746031746031745e-05, "loss": 0.2864, "step": 1200 }, { "epoch": 4.321428571428571, "grad_norm": 1.3869291543960571, "learning_rate": 3.154761904761905e-05, "loss": 0.2834, "step": 1210 }, { "epoch": 4.357142857142857, "grad_norm": 2.1556620597839355, "learning_rate": 3.134920634920635e-05, "loss": 0.3011, "step": 1220 }, { "epoch": 4.392857142857143, "grad_norm": 1.6058976650238037, "learning_rate": 3.1150793650793656e-05, "loss": 0.2893, "step": 1230 }, { "epoch": 4.428571428571429, "grad_norm": 1.9528658390045166, "learning_rate": 3.095238095238095e-05, "loss": 0.2725, "step": 1240 }, { "epoch": 4.464285714285714, "grad_norm": 1.5343618392944336, "learning_rate": 3.075396825396826e-05, "loss": 0.2578, "step": 1250 }, { "epoch": 4.5, "grad_norm": 1.0154541730880737, "learning_rate": 3.055555555555556e-05, "loss": 0.2897, "step": 1260 }, { "epoch": 4.535714285714286, "grad_norm": 2.355865001678467, "learning_rate": 3.0357142857142857e-05, "loss": 0.2658, "step": 1270 }, { "epoch": 4.571428571428571, "grad_norm": 1.24091374874115, "learning_rate": 3.0158730158730158e-05, "loss": 0.3198, "step": 1280 }, { "epoch": 4.607142857142857, "grad_norm": 1.0385922193527222, "learning_rate": 2.996031746031746e-05, "loss": 0.285, "step": 1290 }, { "epoch": 4.642857142857143, "grad_norm": 1.032619833946228, "learning_rate": 2.9761904761904762e-05, "loss": 0.2741, "step": 1300 }, { "epoch": 4.678571428571429, "grad_norm": 1.587849736213684, "learning_rate": 2.9563492063492066e-05, "loss": 0.3093, "step": 1310 }, { "epoch": 4.714285714285714, "grad_norm": 1.2873854637145996, "learning_rate": 2.9365079365079366e-05, "loss": 0.299, "step": 1320 }, { "epoch": 4.75, "grad_norm": 1.6315929889678955, "learning_rate": 2.916666666666667e-05, "loss": 0.2776, "step": 1330 }, { "epoch": 4.785714285714286, "grad_norm": 1.8490331172943115, "learning_rate": 2.8968253968253974e-05, "loss": 0.2613, "step": 1340 }, { "epoch": 4.821428571428571, "grad_norm": 1.8352344036102295, "learning_rate": 2.876984126984127e-05, "loss": 0.281, "step": 1350 }, { "epoch": 4.857142857142857, "grad_norm": 1.0744656324386597, "learning_rate": 2.857142857142857e-05, "loss": 0.2622, "step": 1360 }, { "epoch": 4.892857142857143, "grad_norm": 1.2477718591690063, "learning_rate": 2.8373015873015875e-05, "loss": 0.2997, "step": 1370 }, { "epoch": 4.928571428571429, "grad_norm": 1.5114529132843018, "learning_rate": 2.8174603174603175e-05, "loss": 0.2827, "step": 1380 }, { "epoch": 4.964285714285714, "grad_norm": 1.3266760110855103, "learning_rate": 2.797619047619048e-05, "loss": 0.283, "step": 1390 }, { "epoch": 5.0, "grad_norm": 2.2647006511688232, "learning_rate": 2.777777777777778e-05, "loss": 0.2498, "step": 1400 }, { "epoch": 5.0, "eval_accuracy": 0.8693613220187584, "eval_loss": 0.30874186754226685, "eval_runtime": 116.4377, "eval_samples_per_second": 38.458, "eval_steps_per_second": 2.405, "step": 1400 }, { "epoch": 5.035714285714286, "grad_norm": 1.6704837083816528, "learning_rate": 2.7579365079365083e-05, "loss": 0.2973, "step": 1410 }, { "epoch": 5.071428571428571, "grad_norm": 1.9291003942489624, "learning_rate": 2.7380952380952383e-05, "loss": 0.2756, "step": 1420 }, { "epoch": 5.107142857142857, "grad_norm": 2.44765043258667, "learning_rate": 2.718253968253968e-05, "loss": 0.2575, "step": 1430 }, { "epoch": 5.142857142857143, "grad_norm": 2.0583574771881104, "learning_rate": 2.6984126984126984e-05, "loss": 0.256, "step": 1440 }, { "epoch": 5.178571428571429, "grad_norm": 1.8882899284362793, "learning_rate": 2.6785714285714288e-05, "loss": 0.299, "step": 1450 }, { "epoch": 5.214285714285714, "grad_norm": 1.5178027153015137, "learning_rate": 2.6587301587301588e-05, "loss": 0.3209, "step": 1460 }, { "epoch": 5.25, "grad_norm": 1.4342873096466064, "learning_rate": 2.6388888888888892e-05, "loss": 0.2828, "step": 1470 }, { "epoch": 5.285714285714286, "grad_norm": 1.3157316446304321, "learning_rate": 2.6190476190476192e-05, "loss": 0.2655, "step": 1480 }, { "epoch": 5.321428571428571, "grad_norm": 1.0939053297042847, "learning_rate": 2.5992063492063496e-05, "loss": 0.278, "step": 1490 }, { "epoch": 5.357142857142857, "grad_norm": 1.9465833902359009, "learning_rate": 2.5793650793650796e-05, "loss": 0.2652, "step": 1500 }, { "epoch": 5.392857142857143, "grad_norm": 1.5688917636871338, "learning_rate": 2.5595238095238093e-05, "loss": 0.2622, "step": 1510 }, { "epoch": 5.428571428571429, "grad_norm": 1.3724976778030396, "learning_rate": 2.5396825396825397e-05, "loss": 0.2524, "step": 1520 }, { "epoch": 5.464285714285714, "grad_norm": 1.160733699798584, "learning_rate": 2.5198412698412697e-05, "loss": 0.2868, "step": 1530 }, { "epoch": 5.5, "grad_norm": 1.2799668312072754, "learning_rate": 2.5e-05, "loss": 0.2713, "step": 1540 }, { "epoch": 5.535714285714286, "grad_norm": 1.9062435626983643, "learning_rate": 2.4801587301587305e-05, "loss": 0.2527, "step": 1550 }, { "epoch": 5.571428571428571, "grad_norm": 2.5998220443725586, "learning_rate": 2.4603174603174602e-05, "loss": 0.2668, "step": 1560 }, { "epoch": 5.607142857142857, "grad_norm": 2.346576452255249, "learning_rate": 2.4404761904761906e-05, "loss": 0.2557, "step": 1570 }, { "epoch": 5.642857142857143, "grad_norm": 1.9609774351119995, "learning_rate": 2.4206349206349206e-05, "loss": 0.2706, "step": 1580 }, { "epoch": 5.678571428571429, "grad_norm": 1.8994617462158203, "learning_rate": 2.400793650793651e-05, "loss": 0.2878, "step": 1590 }, { "epoch": 5.714285714285714, "grad_norm": 1.677849531173706, "learning_rate": 2.380952380952381e-05, "loss": 0.2761, "step": 1600 }, { "epoch": 5.75, "grad_norm": 1.6345635652542114, "learning_rate": 2.361111111111111e-05, "loss": 0.2231, "step": 1610 }, { "epoch": 5.785714285714286, "grad_norm": 1.6198554039001465, "learning_rate": 2.3412698412698414e-05, "loss": 0.2612, "step": 1620 }, { "epoch": 5.821428571428571, "grad_norm": 2.052764415740967, "learning_rate": 2.3214285714285715e-05, "loss": 0.2738, "step": 1630 }, { "epoch": 5.857142857142857, "grad_norm": 1.0596450567245483, "learning_rate": 2.3015873015873015e-05, "loss": 0.2597, "step": 1640 }, { "epoch": 5.892857142857143, "grad_norm": 1.9113609790802002, "learning_rate": 2.281746031746032e-05, "loss": 0.2894, "step": 1650 }, { "epoch": 5.928571428571429, "grad_norm": 1.7055829763412476, "learning_rate": 2.261904761904762e-05, "loss": 0.226, "step": 1660 }, { "epoch": 5.964285714285714, "grad_norm": 0.969464898109436, "learning_rate": 2.2420634920634923e-05, "loss": 0.2977, "step": 1670 }, { "epoch": 6.0, "grad_norm": 3.1485939025878906, "learning_rate": 2.2222222222222223e-05, "loss": 0.3028, "step": 1680 }, { "epoch": 6.0, "eval_accuracy": 0.8715944618133095, "eval_loss": 0.29658079147338867, "eval_runtime": 113.6842, "eval_samples_per_second": 39.39, "eval_steps_per_second": 2.463, "step": 1680 }, { "epoch": 6.035714285714286, "grad_norm": 1.4179354906082153, "learning_rate": 2.2023809523809524e-05, "loss": 0.2681, "step": 1690 }, { "epoch": 6.071428571428571, "grad_norm": 2.099681854248047, "learning_rate": 2.1825396825396827e-05, "loss": 0.2441, "step": 1700 }, { "epoch": 6.107142857142857, "grad_norm": 1.5406619310379028, "learning_rate": 2.1626984126984128e-05, "loss": 0.2707, "step": 1710 }, { "epoch": 6.142857142857143, "grad_norm": 2.4518625736236572, "learning_rate": 2.1428571428571428e-05, "loss": 0.2942, "step": 1720 }, { "epoch": 6.178571428571429, "grad_norm": 1.1450200080871582, "learning_rate": 2.1230158730158732e-05, "loss": 0.2336, "step": 1730 }, { "epoch": 6.214285714285714, "grad_norm": 1.4170438051223755, "learning_rate": 2.1031746031746032e-05, "loss": 0.2857, "step": 1740 }, { "epoch": 6.25, "grad_norm": 1.2231560945510864, "learning_rate": 2.0833333333333336e-05, "loss": 0.2507, "step": 1750 }, { "epoch": 6.285714285714286, "grad_norm": 1.4597039222717285, "learning_rate": 2.0634920634920636e-05, "loss": 0.2688, "step": 1760 }, { "epoch": 6.321428571428571, "grad_norm": 1.9395854473114014, "learning_rate": 2.0436507936507937e-05, "loss": 0.2715, "step": 1770 }, { "epoch": 6.357142857142857, "grad_norm": 0.9649907946586609, "learning_rate": 2.023809523809524e-05, "loss": 0.2508, "step": 1780 }, { "epoch": 6.392857142857143, "grad_norm": 1.3842933177947998, "learning_rate": 2.003968253968254e-05, "loss": 0.2818, "step": 1790 }, { "epoch": 6.428571428571429, "grad_norm": 1.7550239562988281, "learning_rate": 1.984126984126984e-05, "loss": 0.2553, "step": 1800 }, { "epoch": 6.464285714285714, "grad_norm": 1.04232919216156, "learning_rate": 1.9642857142857145e-05, "loss": 0.2303, "step": 1810 }, { "epoch": 6.5, "grad_norm": 2.306325674057007, "learning_rate": 1.9444444444444445e-05, "loss": 0.257, "step": 1820 }, { "epoch": 6.535714285714286, "grad_norm": 1.704186201095581, "learning_rate": 1.924603174603175e-05, "loss": 0.2844, "step": 1830 }, { "epoch": 6.571428571428571, "grad_norm": 1.3930236101150513, "learning_rate": 1.9047619047619046e-05, "loss": 0.2784, "step": 1840 }, { "epoch": 6.607142857142857, "grad_norm": 1.5494309663772583, "learning_rate": 1.884920634920635e-05, "loss": 0.2735, "step": 1850 }, { "epoch": 6.642857142857143, "grad_norm": 1.5971505641937256, "learning_rate": 1.8650793650793654e-05, "loss": 0.2466, "step": 1860 }, { "epoch": 6.678571428571429, "grad_norm": 1.8419290781021118, "learning_rate": 1.8452380952380954e-05, "loss": 0.2476, "step": 1870 }, { "epoch": 6.714285714285714, "grad_norm": 1.3679064512252808, "learning_rate": 1.8253968253968254e-05, "loss": 0.2277, "step": 1880 }, { "epoch": 6.75, "grad_norm": 1.1586477756500244, "learning_rate": 1.8055555555555555e-05, "loss": 0.2266, "step": 1890 }, { "epoch": 6.785714285714286, "grad_norm": 1.9979654550552368, "learning_rate": 1.785714285714286e-05, "loss": 0.2747, "step": 1900 }, { "epoch": 6.821428571428571, "grad_norm": 1.4875764846801758, "learning_rate": 1.7658730158730162e-05, "loss": 0.2666, "step": 1910 }, { "epoch": 6.857142857142857, "grad_norm": 1.6196482181549072, "learning_rate": 1.746031746031746e-05, "loss": 0.2532, "step": 1920 }, { "epoch": 6.892857142857143, "grad_norm": 2.7135918140411377, "learning_rate": 1.7261904761904763e-05, "loss": 0.2265, "step": 1930 }, { "epoch": 6.928571428571429, "grad_norm": 1.423257827758789, "learning_rate": 1.7063492063492063e-05, "loss": 0.2621, "step": 1940 }, { "epoch": 6.964285714285714, "grad_norm": 2.1509084701538086, "learning_rate": 1.6865079365079367e-05, "loss": 0.2636, "step": 1950 }, { "epoch": 7.0, "grad_norm": 1.562657117843628, "learning_rate": 1.6666666666666667e-05, "loss": 0.2179, "step": 1960 }, { "epoch": 7.0, "eval_accuracy": 0.8807503349709692, "eval_loss": 0.2742190361022949, "eval_runtime": 113.8381, "eval_samples_per_second": 39.337, "eval_steps_per_second": 2.46, "step": 1960 }, { "epoch": 7.035714285714286, "grad_norm": 1.4138425588607788, "learning_rate": 1.6468253968253968e-05, "loss": 0.2251, "step": 1970 }, { "epoch": 7.071428571428571, "grad_norm": 2.6958255767822266, "learning_rate": 1.626984126984127e-05, "loss": 0.2521, "step": 1980 }, { "epoch": 7.107142857142857, "grad_norm": 2.014803647994995, "learning_rate": 1.6071428571428572e-05, "loss": 0.216, "step": 1990 }, { "epoch": 7.142857142857143, "grad_norm": 2.1656410694122314, "learning_rate": 1.5873015873015872e-05, "loss": 0.2085, "step": 2000 }, { "epoch": 7.178571428571429, "grad_norm": 2.207980155944824, "learning_rate": 1.5674603174603176e-05, "loss": 0.263, "step": 2010 }, { "epoch": 7.214285714285714, "grad_norm": 1.978162169456482, "learning_rate": 1.5476190476190476e-05, "loss": 0.2478, "step": 2020 }, { "epoch": 7.25, "grad_norm": 1.2705848217010498, "learning_rate": 1.527777777777778e-05, "loss": 0.2338, "step": 2030 }, { "epoch": 7.285714285714286, "grad_norm": 1.6287261247634888, "learning_rate": 1.5079365079365079e-05, "loss": 0.2063, "step": 2040 }, { "epoch": 7.321428571428571, "grad_norm": 2.0518875122070312, "learning_rate": 1.4880952380952381e-05, "loss": 0.3032, "step": 2050 }, { "epoch": 7.357142857142857, "grad_norm": 2.7869536876678467, "learning_rate": 1.4682539682539683e-05, "loss": 0.2242, "step": 2060 }, { "epoch": 7.392857142857143, "grad_norm": 1.007391095161438, "learning_rate": 1.4484126984126987e-05, "loss": 0.2589, "step": 2070 }, { "epoch": 7.428571428571429, "grad_norm": 1.2812851667404175, "learning_rate": 1.4285714285714285e-05, "loss": 0.2383, "step": 2080 }, { "epoch": 7.464285714285714, "grad_norm": 1.1465330123901367, "learning_rate": 1.4087301587301587e-05, "loss": 0.2284, "step": 2090 }, { "epoch": 7.5, "grad_norm": 2.2567813396453857, "learning_rate": 1.388888888888889e-05, "loss": 0.213, "step": 2100 }, { "epoch": 7.535714285714286, "grad_norm": 1.7949641942977905, "learning_rate": 1.3690476190476192e-05, "loss": 0.2368, "step": 2110 }, { "epoch": 7.571428571428571, "grad_norm": 2.443598985671997, "learning_rate": 1.3492063492063492e-05, "loss": 0.2665, "step": 2120 }, { "epoch": 7.607142857142857, "grad_norm": 1.9699336290359497, "learning_rate": 1.3293650793650794e-05, "loss": 0.2346, "step": 2130 }, { "epoch": 7.642857142857143, "grad_norm": 1.543039083480835, "learning_rate": 1.3095238095238096e-05, "loss": 0.2432, "step": 2140 }, { "epoch": 7.678571428571429, "grad_norm": 1.9814691543579102, "learning_rate": 1.2896825396825398e-05, "loss": 0.2575, "step": 2150 }, { "epoch": 7.714285714285714, "grad_norm": 2.1088602542877197, "learning_rate": 1.2698412698412699e-05, "loss": 0.2346, "step": 2160 }, { "epoch": 7.75, "grad_norm": 1.5648256540298462, "learning_rate": 1.25e-05, "loss": 0.2403, "step": 2170 }, { "epoch": 7.785714285714286, "grad_norm": 1.6079583168029785, "learning_rate": 1.2301587301587301e-05, "loss": 0.2414, "step": 2180 }, { "epoch": 7.821428571428571, "grad_norm": 1.2859593629837036, "learning_rate": 1.2103174603174603e-05, "loss": 0.2756, "step": 2190 }, { "epoch": 7.857142857142857, "grad_norm": 2.072089672088623, "learning_rate": 1.1904761904761905e-05, "loss": 0.2484, "step": 2200 }, { "epoch": 7.892857142857143, "grad_norm": 1.623353362083435, "learning_rate": 1.1706349206349207e-05, "loss": 0.274, "step": 2210 }, { "epoch": 7.928571428571429, "grad_norm": 1.70241379737854, "learning_rate": 1.1507936507936508e-05, "loss": 0.2307, "step": 2220 }, { "epoch": 7.964285714285714, "grad_norm": 1.7186700105667114, "learning_rate": 1.130952380952381e-05, "loss": 0.2223, "step": 2230 }, { "epoch": 8.0, "grad_norm": 2.0665862560272217, "learning_rate": 1.1111111111111112e-05, "loss": 0.2274, "step": 2240 }, { "epoch": 8.0, "eval_accuracy": 0.8814202769093346, "eval_loss": 0.2860513925552368, "eval_runtime": 115.9278, "eval_samples_per_second": 38.627, "eval_steps_per_second": 2.415, "step": 2240 }, { "epoch": 8.035714285714286, "grad_norm": 1.3822689056396484, "learning_rate": 1.0912698412698414e-05, "loss": 0.238, "step": 2250 }, { "epoch": 8.071428571428571, "grad_norm": 1.4226853847503662, "learning_rate": 1.0714285714285714e-05, "loss": 0.266, "step": 2260 }, { "epoch": 8.107142857142858, "grad_norm": 1.9675803184509277, "learning_rate": 1.0515873015873016e-05, "loss": 0.2814, "step": 2270 }, { "epoch": 8.142857142857142, "grad_norm": 1.8133440017700195, "learning_rate": 1.0317460317460318e-05, "loss": 0.2486, "step": 2280 }, { "epoch": 8.178571428571429, "grad_norm": 1.2323780059814453, "learning_rate": 1.011904761904762e-05, "loss": 0.2175, "step": 2290 }, { "epoch": 8.214285714285714, "grad_norm": 1.7063086032867432, "learning_rate": 9.92063492063492e-06, "loss": 0.2125, "step": 2300 }, { "epoch": 8.25, "grad_norm": 1.532769799232483, "learning_rate": 9.722222222222223e-06, "loss": 0.1813, "step": 2310 }, { "epoch": 8.285714285714286, "grad_norm": 1.5620160102844238, "learning_rate": 9.523809523809523e-06, "loss": 0.2355, "step": 2320 }, { "epoch": 8.321428571428571, "grad_norm": 1.6230847835540771, "learning_rate": 9.325396825396827e-06, "loss": 0.2199, "step": 2330 }, { "epoch": 8.357142857142858, "grad_norm": 1.552085041999817, "learning_rate": 9.126984126984127e-06, "loss": 0.2479, "step": 2340 }, { "epoch": 8.392857142857142, "grad_norm": 1.8006685972213745, "learning_rate": 8.92857142857143e-06, "loss": 0.2677, "step": 2350 }, { "epoch": 8.428571428571429, "grad_norm": 1.2057029008865356, "learning_rate": 8.73015873015873e-06, "loss": 0.2022, "step": 2360 }, { "epoch": 8.464285714285714, "grad_norm": 1.4805638790130615, "learning_rate": 8.531746031746032e-06, "loss": 0.2158, "step": 2370 }, { "epoch": 8.5, "grad_norm": 1.441603422164917, "learning_rate": 8.333333333333334e-06, "loss": 0.2166, "step": 2380 }, { "epoch": 8.535714285714286, "grad_norm": 2.171687602996826, "learning_rate": 8.134920634920636e-06, "loss": 0.2332, "step": 2390 }, { "epoch": 8.571428571428571, "grad_norm": 1.5523641109466553, "learning_rate": 7.936507936507936e-06, "loss": 0.2099, "step": 2400 }, { "epoch": 8.607142857142858, "grad_norm": 1.429527759552002, "learning_rate": 7.738095238095238e-06, "loss": 0.1934, "step": 2410 }, { "epoch": 8.642857142857142, "grad_norm": 3.8555209636688232, "learning_rate": 7.5396825396825394e-06, "loss": 0.2813, "step": 2420 }, { "epoch": 8.678571428571429, "grad_norm": 1.118416666984558, "learning_rate": 7.3412698412698415e-06, "loss": 0.2358, "step": 2430 }, { "epoch": 8.714285714285714, "grad_norm": 1.9037988185882568, "learning_rate": 7.142857142857143e-06, "loss": 0.2091, "step": 2440 }, { "epoch": 8.75, "grad_norm": 1.8768919706344604, "learning_rate": 6.944444444444445e-06, "loss": 0.2176, "step": 2450 }, { "epoch": 8.785714285714286, "grad_norm": 1.606123685836792, "learning_rate": 6.746031746031746e-06, "loss": 0.2388, "step": 2460 }, { "epoch": 8.821428571428571, "grad_norm": 2.6942107677459717, "learning_rate": 6.547619047619048e-06, "loss": 0.2182, "step": 2470 }, { "epoch": 8.857142857142858, "grad_norm": 1.449601173400879, "learning_rate": 6.349206349206349e-06, "loss": 0.2102, "step": 2480 }, { "epoch": 8.892857142857142, "grad_norm": 2.651686668395996, "learning_rate": 6.1507936507936505e-06, "loss": 0.2166, "step": 2490 }, { "epoch": 8.928571428571429, "grad_norm": 2.6743762493133545, "learning_rate": 5.9523809523809525e-06, "loss": 0.2468, "step": 2500 }, { "epoch": 8.964285714285714, "grad_norm": 2.29903244972229, "learning_rate": 5.753968253968254e-06, "loss": 0.2111, "step": 2510 }, { "epoch": 9.0, "grad_norm": 1.8269622325897217, "learning_rate": 5.555555555555556e-06, "loss": 0.2195, "step": 2520 }, { "epoch": 9.0, "eval_accuracy": 0.8894595801697186, "eval_loss": 0.26261791586875916, "eval_runtime": 114.2956, "eval_samples_per_second": 39.179, "eval_steps_per_second": 2.45, "step": 2520 }, { "epoch": 9.035714285714286, "grad_norm": 3.011329174041748, "learning_rate": 5.357142857142857e-06, "loss": 0.1827, "step": 2530 }, { "epoch": 9.071428571428571, "grad_norm": 2.3060262203216553, "learning_rate": 5.158730158730159e-06, "loss": 0.1862, "step": 2540 }, { "epoch": 9.107142857142858, "grad_norm": 1.7220100164413452, "learning_rate": 4.96031746031746e-06, "loss": 0.2215, "step": 2550 }, { "epoch": 9.142857142857142, "grad_norm": 2.463092803955078, "learning_rate": 4.7619047619047615e-06, "loss": 0.2228, "step": 2560 }, { "epoch": 9.178571428571429, "grad_norm": 1.204136848449707, "learning_rate": 4.563492063492064e-06, "loss": 0.2439, "step": 2570 }, { "epoch": 9.214285714285714, "grad_norm": 2.263396978378296, "learning_rate": 4.365079365079365e-06, "loss": 0.2266, "step": 2580 }, { "epoch": 9.25, "grad_norm": 2.7832555770874023, "learning_rate": 4.166666666666667e-06, "loss": 0.2089, "step": 2590 }, { "epoch": 9.285714285714286, "grad_norm": 2.0564024448394775, "learning_rate": 3.968253968253968e-06, "loss": 0.1839, "step": 2600 }, { "epoch": 9.321428571428571, "grad_norm": 2.0316998958587646, "learning_rate": 3.7698412698412697e-06, "loss": 0.2424, "step": 2610 }, { "epoch": 9.357142857142858, "grad_norm": 2.229687213897705, "learning_rate": 3.5714285714285714e-06, "loss": 0.2489, "step": 2620 }, { "epoch": 9.392857142857142, "grad_norm": 1.7529199123382568, "learning_rate": 3.373015873015873e-06, "loss": 0.2273, "step": 2630 }, { "epoch": 9.428571428571429, "grad_norm": 1.5242239236831665, "learning_rate": 3.1746031746031746e-06, "loss": 0.2224, "step": 2640 }, { "epoch": 9.464285714285714, "grad_norm": 1.5499508380889893, "learning_rate": 2.9761904761904763e-06, "loss": 0.2338, "step": 2650 }, { "epoch": 9.5, "grad_norm": 2.5902230739593506, "learning_rate": 2.777777777777778e-06, "loss": 0.1909, "step": 2660 }, { "epoch": 9.535714285714286, "grad_norm": 1.3243242502212524, "learning_rate": 2.5793650793650795e-06, "loss": 0.2199, "step": 2670 }, { "epoch": 9.571428571428571, "grad_norm": 1.9745112657546997, "learning_rate": 2.3809523809523808e-06, "loss": 0.2172, "step": 2680 }, { "epoch": 9.607142857142858, "grad_norm": 2.10951828956604, "learning_rate": 2.1825396825396824e-06, "loss": 0.2331, "step": 2690 }, { "epoch": 9.642857142857142, "grad_norm": 2.113539457321167, "learning_rate": 1.984126984126984e-06, "loss": 0.2373, "step": 2700 }, { "epoch": 9.678571428571429, "grad_norm": 1.548854112625122, "learning_rate": 1.7857142857142857e-06, "loss": 0.2035, "step": 2710 }, { "epoch": 9.714285714285714, "grad_norm": 1.797196626663208, "learning_rate": 1.5873015873015873e-06, "loss": 0.2015, "step": 2720 }, { "epoch": 9.75, "grad_norm": 1.9279841184616089, "learning_rate": 1.388888888888889e-06, "loss": 0.2208, "step": 2730 }, { "epoch": 9.785714285714286, "grad_norm": 0.950290858745575, "learning_rate": 1.1904761904761904e-06, "loss": 0.2394, "step": 2740 }, { "epoch": 9.821428571428571, "grad_norm": 1.2573094367980957, "learning_rate": 9.92063492063492e-07, "loss": 0.2491, "step": 2750 }, { "epoch": 9.857142857142858, "grad_norm": 1.8162927627563477, "learning_rate": 7.936507936507937e-07, "loss": 0.2229, "step": 2760 }, { "epoch": 9.892857142857142, "grad_norm": 2.0845260620117188, "learning_rate": 5.952380952380952e-07, "loss": 0.1984, "step": 2770 }, { "epoch": 9.928571428571429, "grad_norm": 1.8501282930374146, "learning_rate": 3.9682539682539683e-07, "loss": 0.2287, "step": 2780 }, { "epoch": 9.964285714285714, "grad_norm": 2.941807746887207, "learning_rate": 1.9841269841269841e-07, "loss": 0.2233, "step": 2790 }, { "epoch": 10.0, "grad_norm": 1.2508630752563477, "learning_rate": 0.0, "loss": 0.1886, "step": 2800 }, { "epoch": 10.0, "eval_accuracy": 0.8865564984368022, "eval_loss": 0.2717145085334778, "eval_runtime": 113.959, "eval_samples_per_second": 39.295, "eval_steps_per_second": 2.457, "step": 2800 }, { "epoch": 10.0, "step": 2800, "total_flos": 1.3877265500181135e+19, "train_loss": 0.3087472263830049, "train_runtime": 6534.1253, "train_samples_per_second": 27.407, "train_steps_per_second": 0.429 } ], "logging_steps": 10, "max_steps": 2800, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3877265500181135e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }