|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9902828495388996, |
|
"eval_steps": 500, |
|
"global_step": 48000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0, |
|
"grad_norm": 13.376683235168457, |
|
"learning_rate": 0.00001, |
|
"loss": 4.2362, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0, |
|
"grad_norm": 2.2011091709136963, |
|
"learning_rate": 0.00002, |
|
"loss": 0.6427, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.1116416454315186, |
|
"learning_rate": 0.00003, |
|
"loss": 0.5047, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 2.325719118118286, |
|
"learning_rate": 0.00004, |
|
"loss": 0.4955, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.2606059312820435, |
|
"learning_rate": 0.00005, |
|
"loss": 0.5104, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.9306057691574097, |
|
"learning_rate": 0.000049895770361259934, |
|
"loss": 0.5024, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1.3757009506225586, |
|
"learning_rate": 0.00004979154072251986, |
|
"loss": 0.4814, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.6575660705566406, |
|
"learning_rate": 0.00004968731108377979, |
|
"loss": 0.4675, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.6257671117782593, |
|
"learning_rate": 0.000049583081445039714, |
|
"loss": 0.4673, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.1650177240371704, |
|
"learning_rate": 0.000049478851806299645, |
|
"loss": 0.4635, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.2291123867034912, |
|
"learning_rate": 0.00004937462216755957, |
|
"loss": 0.4645, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 1.094923973083496, |
|
"learning_rate": 0.0000492703925288195, |
|
"loss": 0.463, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.4243600368499756, |
|
"learning_rate": 0.000049166162890079426, |
|
"loss": 0.4305, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.5323872566223145, |
|
"learning_rate": 0.00004906193325133936, |
|
"loss": 0.4431, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.4820982217788696, |
|
"learning_rate": 0.00004895770361259928, |
|
"loss": 0.4301, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 1.6124228239059448, |
|
"learning_rate": 0.00004885347397385921, |
|
"loss": 0.4537, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.5138214826583862, |
|
"learning_rate": 0.00004874924433511914, |
|
"loss": 0.4101, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 2.3704726696014404, |
|
"learning_rate": 0.00004864501469637907, |
|
"loss": 0.4113, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.4032459259033203, |
|
"learning_rate": 0.00004854078505763899, |
|
"loss": 0.4297, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.3240689039230347, |
|
"learning_rate": 0.000048436555418898924, |
|
"loss": 0.4193, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.411370038986206, |
|
"learning_rate": 0.00004833232578015885, |
|
"loss": 0.4118, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.4880633354187012, |
|
"learning_rate": 0.00004822809614141878, |
|
"loss": 0.4284, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.3598105907440186, |
|
"learning_rate": 0.000048123866502678705, |
|
"loss": 0.4119, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.2027459144592285, |
|
"learning_rate": 0.000048019636863938636, |
|
"loss": 0.4073, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.35155189037323, |
|
"learning_rate": 0.00004791540722519856, |
|
"loss": 0.4134, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1.8489034175872803, |
|
"learning_rate": 0.00004781117758645849, |
|
"loss": 0.4069, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.2636560201644897, |
|
"learning_rate": 0.000047706947947718416, |
|
"loss": 0.4224, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.4396322965621948, |
|
"learning_rate": 0.00004760271830897835, |
|
"loss": 0.3951, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.3303143978118896, |
|
"learning_rate": 0.00004749848867023827, |
|
"loss": 0.4111, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.2054022550582886, |
|
"learning_rate": 0.0000473942590314982, |
|
"loss": 0.4155, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.1086294651031494, |
|
"learning_rate": 0.00004729002939275813, |
|
"loss": 0.3912, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0423736572265625, |
|
"learning_rate": 0.00004718579975401806, |
|
"loss": 0.4174, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1115955114364624, |
|
"learning_rate": 0.000047081570115277984, |
|
"loss": 0.4293, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1089532375335693, |
|
"learning_rate": 0.000046977340476537915, |
|
"loss": 0.3957, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.1033278703689575, |
|
"learning_rate": 0.00004687311083779784, |
|
"loss": 0.3913, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 1.0068538188934326, |
|
"learning_rate": 0.00004676888119905777, |
|
"loss": 0.3931, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.0825397968292236, |
|
"learning_rate": 0.000046664651560317695, |
|
"loss": 0.4043, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1124091148376465, |
|
"learning_rate": 0.000046560421921577626, |
|
"loss": 0.3892, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.8287498354911804, |
|
"learning_rate": 0.00004645619228283755, |
|
"loss": 0.3902, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.3230834007263184, |
|
"learning_rate": 0.00004635196264409748, |
|
"loss": 0.3946, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 1.1308443546295166, |
|
"learning_rate": 0.00004624773300535741, |
|
"loss": 0.388, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.016345739364624, |
|
"learning_rate": 0.00004614350336661734, |
|
"loss": 0.3885, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.7567281723022461, |
|
"learning_rate": 0.00004603927372787726, |
|
"loss": 0.3733, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.0191072225570679, |
|
"learning_rate": 0.000045935044089137194, |
|
"loss": 0.3755, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.9752495288848877, |
|
"learning_rate": 0.00004583081445039712, |
|
"loss": 0.3891, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 1.063784122467041, |
|
"learning_rate": 0.00004572658481165705, |
|
"loss": 0.3502, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.4277328252792358, |
|
"learning_rate": 0.000045622355172916974, |
|
"loss": 0.3832, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.051148533821106, |
|
"learning_rate": 0.000045518125534176906, |
|
"loss": 0.3844, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0719451904296875, |
|
"learning_rate": 0.00004541389589543683, |
|
"loss": 0.3591, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 1.0701349973678589, |
|
"learning_rate": 0.00004530966625669676, |
|
"loss": 0.3549, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9545413255691528, |
|
"learning_rate": 0.000045205436617956686, |
|
"loss": 0.3791, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.2077405452728271, |
|
"learning_rate": 0.00004510120697921662, |
|
"loss": 0.3794, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.0325422286987305, |
|
"learning_rate": 0.00004499697734047654, |
|
"loss": 0.3709, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.231856107711792, |
|
"learning_rate": 0.00004489274770173647, |
|
"loss": 0.375, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1.0496745109558105, |
|
"learning_rate": 0.0000447885180629964, |
|
"loss": 0.3633, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.9745790362358093, |
|
"learning_rate": 0.00004468428842425633, |
|
"loss": 0.3541, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8768660426139832, |
|
"learning_rate": 0.000044580058785516246, |
|
"loss": 0.354, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.2748613357543945, |
|
"learning_rate": 0.00004447582914677618, |
|
"loss": 0.3463, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.6291444301605225, |
|
"learning_rate": 0.0000443715995080361, |
|
"loss": 0.3483, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.1184431314468384, |
|
"learning_rate": 0.000044267369869296034, |
|
"loss": 0.3616, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.042864441871643, |
|
"learning_rate": 0.00004416314023055596, |
|
"loss": 0.3659, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.2588552236557007, |
|
"learning_rate": 0.00004405891059181589, |
|
"loss": 0.3685, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.9592663645744324, |
|
"learning_rate": 0.000043954680953075814, |
|
"loss": 0.3346, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.0394409894943237, |
|
"learning_rate": 0.000043850451314335745, |
|
"loss": 0.3543, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 1.0655689239501953, |
|
"learning_rate": 0.00004374622167559567, |
|
"loss": 0.3608, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.0429097414016724, |
|
"learning_rate": 0.0000436419920368556, |
|
"loss": 0.3752, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.105862021446228, |
|
"learning_rate": 0.000043537762398115525, |
|
"loss": 0.3656, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.396264672279358, |
|
"learning_rate": 0.00004343353275937546, |
|
"loss": 0.3649, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.1206328868865967, |
|
"learning_rate": 0.00004332930312063538, |
|
"loss": 0.3658, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.1256790161132812, |
|
"learning_rate": 0.00004322507348189531, |
|
"loss": 0.3604, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.0484185218811035, |
|
"learning_rate": 0.00004312084384315524, |
|
"loss": 0.3538, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1310747861862183, |
|
"learning_rate": 0.00004301661420441517, |
|
"loss": 0.3454, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.3282368183135986, |
|
"learning_rate": 0.00004291238456567509, |
|
"loss": 0.3531, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1081411838531494, |
|
"learning_rate": 0.000042808154926935024, |
|
"loss": 0.3572, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.0838216543197632, |
|
"learning_rate": 0.00004270392528819495, |
|
"loss": 0.3609, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.0632649660110474, |
|
"learning_rate": 0.00004259969564945488, |
|
"loss": 0.3555, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.2421514987945557, |
|
"learning_rate": 0.000042495466010714804, |
|
"loss": 0.3504, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.0846999883651733, |
|
"learning_rate": 0.000042391236371974736, |
|
"loss": 0.3697, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.0196987390518188, |
|
"learning_rate": 0.00004228700673323466, |
|
"loss": 0.331, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.896594226360321, |
|
"learning_rate": 0.00004218277709449459, |
|
"loss": 0.3501, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.0059070587158203, |
|
"learning_rate": 0.000042078547455754516, |
|
"loss": 0.3414, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9597504734992981, |
|
"learning_rate": 0.00004197431781701445, |
|
"loss": 0.3405, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9826390743255615, |
|
"learning_rate": 0.00004187008817827437, |
|
"loss": 0.3444, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 1.1618746519088745, |
|
"learning_rate": 0.0000417658585395343, |
|
"loss": 0.3428, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.3454729318618774, |
|
"learning_rate": 0.00004166162890079423, |
|
"loss": 0.3546, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.0757421255111694, |
|
"learning_rate": 0.00004155739926205416, |
|
"loss": 0.3409, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.1405227184295654, |
|
"learning_rate": 0.000041453169623314083, |
|
"loss": 0.367, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.336688756942749, |
|
"learning_rate": 0.000041348939984574015, |
|
"loss": 0.353, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 1.0525643825531006, |
|
"learning_rate": 0.00004124471034583394, |
|
"loss": 0.3394, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9103542566299438, |
|
"learning_rate": 0.00004114048070709387, |
|
"loss": 0.3542, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.1108696460723877, |
|
"learning_rate": 0.000041036251068353795, |
|
"loss": 0.3468, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.9561033248901367, |
|
"learning_rate": 0.000040932021429613726, |
|
"loss": 0.3234, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 1.0222822427749634, |
|
"learning_rate": 0.00004082779179087365, |
|
"loss": 0.3557, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.7020508050918579, |
|
"learning_rate": 0.00004072356215213358, |
|
"loss": 0.3387, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.1984186172485352, |
|
"learning_rate": 0.00004061933251339351, |
|
"loss": 0.3396, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 1.0309916734695435, |
|
"learning_rate": 0.00004051510287465344, |
|
"loss": 0.3361, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9544962048530579, |
|
"learning_rate": 0.00004041087323591336, |
|
"loss": 0.3198, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.924387514591217, |
|
"learning_rate": 0.000040306643597173294, |
|
"loss": 0.346, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.9332679510116577, |
|
"learning_rate": 0.00004020241395843322, |
|
"loss": 0.3583, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8733212351799011, |
|
"learning_rate": 0.00004009818431969315, |
|
"loss": 0.3258, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.1735975742340088, |
|
"learning_rate": 0.000039993954680953074, |
|
"loss": 0.3319, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.1871587038040161, |
|
"learning_rate": 0.000039889725042213005, |
|
"loss": 0.3323, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.0501062870025635, |
|
"learning_rate": 0.00003978549540347294, |
|
"loss": 0.3501, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 1.6133193969726562, |
|
"learning_rate": 0.00003968126576473286, |
|
"loss": 0.3353, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.0860315561294556, |
|
"learning_rate": 0.00003957703612599279, |
|
"loss": 0.3247, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.9817527532577515, |
|
"learning_rate": 0.00003947280648725272, |
|
"loss": 0.3486, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.185621738433838, |
|
"learning_rate": 0.00003936857684851265, |
|
"loss": 0.3344, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.211653709411621, |
|
"learning_rate": 0.00003926434720977257, |
|
"loss": 0.3308, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.1859573125839233, |
|
"learning_rate": 0.000039160117571032504, |
|
"loss": 0.3462, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9363047480583191, |
|
"learning_rate": 0.00003905588793229243, |
|
"loss": 0.3331, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.8858780860900879, |
|
"learning_rate": 0.00003895165829355236, |
|
"loss": 0.3501, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.187888264656067, |
|
"learning_rate": 0.000038847428654812284, |
|
"loss": 0.3202, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 1.33687424659729, |
|
"learning_rate": 0.000038743199016072216, |
|
"loss": 0.3488, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.230508804321289, |
|
"learning_rate": 0.00003863896937733214, |
|
"loss": 0.322, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.2062541246414185, |
|
"learning_rate": 0.00003853473973859207, |
|
"loss": 0.3446, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9526596069335938, |
|
"learning_rate": 0.000038430510099851996, |
|
"loss": 0.3167, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8020007014274597, |
|
"learning_rate": 0.00003832628046111193, |
|
"loss": 0.3289, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.8741321563720703, |
|
"learning_rate": 0.00003822205082237185, |
|
"loss": 0.3307, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8475280404090881, |
|
"learning_rate": 0.00003811782118363178, |
|
"loss": 0.3116, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8570387959480286, |
|
"learning_rate": 0.00003801359154489171, |
|
"loss": 0.3345, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.237730860710144, |
|
"learning_rate": 0.00003790936190615164, |
|
"loss": 0.3457, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.1261372566223145, |
|
"learning_rate": 0.00003780513226741156, |
|
"loss": 0.3261, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.7940207719802856, |
|
"learning_rate": 0.000037700902628671495, |
|
"loss": 0.3272, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.0150914192199707, |
|
"learning_rate": 0.00003759667298993142, |
|
"loss": 0.3338, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8289426565170288, |
|
"learning_rate": 0.00003749244335119135, |
|
"loss": 0.3485, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.11201012134552, |
|
"learning_rate": 0.000037388213712451275, |
|
"loss": 0.321, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.7759777903556824, |
|
"learning_rate": 0.000037283984073711206, |
|
"loss": 0.336, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.2184884548187256, |
|
"learning_rate": 0.00003717975443497113, |
|
"loss": 0.3029, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.7660321593284607, |
|
"learning_rate": 0.00003707552479623106, |
|
"loss": 0.336, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.8834217190742493, |
|
"learning_rate": 0.000036971295157490987, |
|
"loss": 0.3184, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.1337840557098389, |
|
"learning_rate": 0.00003686706551875092, |
|
"loss": 0.3216, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.0339350700378418, |
|
"learning_rate": 0.00003676283588001084, |
|
"loss": 0.3162, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1.222301721572876, |
|
"learning_rate": 0.000036658606241270774, |
|
"loss": 0.3344, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.290209412574768, |
|
"learning_rate": 0.0000365543766025307, |
|
"loss": 0.3212, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.372733235359192, |
|
"learning_rate": 0.00003645014696379063, |
|
"loss": 0.3375, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.7824010252952576, |
|
"learning_rate": 0.000036345917325050554, |
|
"loss": 0.3393, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.637102484703064, |
|
"learning_rate": 0.000036241687686310485, |
|
"loss": 0.3318, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.9548413753509521, |
|
"learning_rate": 0.00003613745804757041, |
|
"loss": 0.3041, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.3793456554412842, |
|
"learning_rate": 0.00003603322840883034, |
|
"loss": 0.3208, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.323983073234558, |
|
"learning_rate": 0.000035928998770090266, |
|
"loss": 0.3077, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8947316408157349, |
|
"learning_rate": 0.0000358247691313502, |
|
"loss": 0.3081, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.2341340780258179, |
|
"learning_rate": 0.00003572053949261012, |
|
"loss": 0.3243, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.1560657024383545, |
|
"learning_rate": 0.00003561630985387005, |
|
"loss": 0.3045, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.0063599348068237, |
|
"learning_rate": 0.00003551208021512998, |
|
"loss": 0.2984, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.9587939977645874, |
|
"learning_rate": 0.00003540785057638991, |
|
"loss": 0.3215, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.9009751081466675, |
|
"learning_rate": 0.00003530362093764983, |
|
"loss": 0.3257, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 1.1289176940917969, |
|
"learning_rate": 0.000035199391298909764, |
|
"loss": 0.3151, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.0544753074645996, |
|
"learning_rate": 0.00003509516166016969, |
|
"loss": 0.3123, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 1.6340482234954834, |
|
"learning_rate": 0.00003499093202142962, |
|
"loss": 0.3238, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8671633005142212, |
|
"learning_rate": 0.000034886702382689545, |
|
"loss": 0.3174, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.8871846199035645, |
|
"learning_rate": 0.000034782472743949476, |
|
"loss": 0.3045, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.9014525413513184, |
|
"learning_rate": 0.0000346782431052094, |
|
"loss": 0.3335, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.819088876247406, |
|
"learning_rate": 0.00003457401346646933, |
|
"loss": 0.3222, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.1166553497314453, |
|
"learning_rate": 0.000034469783827729256, |
|
"loss": 0.3303, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.0952930450439453, |
|
"learning_rate": 0.00003436555418898919, |
|
"loss": 0.3095, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 1.186812400817871, |
|
"learning_rate": 0.00003426132455024911, |
|
"loss": 0.3146, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9790433049201965, |
|
"learning_rate": 0.00003415709491150904, |
|
"loss": 0.321, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9488071203231812, |
|
"learning_rate": 0.00003405286527276897, |
|
"loss": 0.321, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9488551616668701, |
|
"learning_rate": 0.00003394863563402889, |
|
"loss": 0.2923, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9801424145698547, |
|
"learning_rate": 0.00003384440599528882, |
|
"loss": 0.3131, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 1.5585353374481201, |
|
"learning_rate": 0.00003374017635654875, |
|
"loss": 0.3274, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 1.0194026231765747, |
|
"learning_rate": 0.00003363594671780867, |
|
"loss": 0.3294, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.89570552110672, |
|
"learning_rate": 0.000033531717079068604, |
|
"loss": 0.3041, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.361720085144043, |
|
"learning_rate": 0.00003342748744032853, |
|
"loss": 0.3001, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.6049641370773315, |
|
"learning_rate": 0.00003332325780158846, |
|
"loss": 0.2993, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.1714413166046143, |
|
"learning_rate": 0.000033219028162848384, |
|
"loss": 0.3252, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.6133936643600464, |
|
"learning_rate": 0.000033114798524108315, |
|
"loss": 0.3162, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9461352229118347, |
|
"learning_rate": 0.00003301056888536824, |
|
"loss": 0.3, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.1674548387527466, |
|
"learning_rate": 0.00003290633924662817, |
|
"loss": 0.3053, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9445935487747192, |
|
"learning_rate": 0.000032802109607888096, |
|
"loss": 0.319, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.7129838466644287, |
|
"learning_rate": 0.00003269787996914803, |
|
"loss": 0.3095, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.9326237440109253, |
|
"learning_rate": 0.00003259365033040795, |
|
"loss": 0.3174, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8677362203598022, |
|
"learning_rate": 0.00003248942069166788, |
|
"loss": 0.3056, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.7561855316162109, |
|
"learning_rate": 0.00003238519105292781, |
|
"loss": 0.2939, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.0813933610916138, |
|
"learning_rate": 0.00003228096141418774, |
|
"loss": 0.296, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.1002167463302612, |
|
"learning_rate": 0.00003217673177544766, |
|
"loss": 0.3142, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.709709882736206, |
|
"learning_rate": 0.000032072502136707594, |
|
"loss": 0.2931, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.699057936668396, |
|
"learning_rate": 0.00003196827249796752, |
|
"loss": 0.3009, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.4942712783813477, |
|
"learning_rate": 0.00003186404285922745, |
|
"loss": 0.3203, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 1.069948434829712, |
|
"learning_rate": 0.000031759813220487375, |
|
"loss": 0.3118, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9980233311653137, |
|
"learning_rate": 0.000031655583581747306, |
|
"loss": 0.2937, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.743611752986908, |
|
"learning_rate": 0.00003155135394300723, |
|
"loss": 0.3064, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8888816237449646, |
|
"learning_rate": 0.00003144712430426716, |
|
"loss": 0.309, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9147290587425232, |
|
"learning_rate": 0.000031342894665527086, |
|
"loss": 0.3104, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1.191470742225647, |
|
"learning_rate": 0.00003123866502678702, |
|
"loss": 0.307, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.8025168776512146, |
|
"learning_rate": 0.00003113443538804694, |
|
"loss": 0.2981, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.0088508129119873, |
|
"learning_rate": 0.000031030205749306873, |
|
"loss": 0.2946, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9163653254508972, |
|
"learning_rate": 0.0000309259761105668, |
|
"loss": 0.314, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.7615305185317993, |
|
"learning_rate": 0.00003082174647182673, |
|
"loss": 0.2987, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 1.1720337867736816, |
|
"learning_rate": 0.000030717516833086654, |
|
"loss": 0.284, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9668058156967163, |
|
"learning_rate": 0.000030613287194346585, |
|
"loss": 0.2961, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.0331960916519165, |
|
"learning_rate": 0.000030509057555606513, |
|
"loss": 0.3126, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.011672854423523, |
|
"learning_rate": 0.00003040482791686644, |
|
"loss": 0.2909, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.7572203874588013, |
|
"learning_rate": 0.00003030059827812637, |
|
"loss": 0.3003, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.2314035892486572, |
|
"learning_rate": 0.000030196368639386297, |
|
"loss": 0.2967, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.7441611886024475, |
|
"learning_rate": 0.000030092139000646225, |
|
"loss": 0.3081, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.1014299392700195, |
|
"learning_rate": 0.000029987909361906152, |
|
"loss": 0.3076, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8321337699890137, |
|
"learning_rate": 0.00002988367972316608, |
|
"loss": 0.3, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.0691934823989868, |
|
"learning_rate": 0.000029779450084426008, |
|
"loss": 0.3113, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8844496607780457, |
|
"learning_rate": 0.000029675220445685936, |
|
"loss": 0.3135, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9948704242706299, |
|
"learning_rate": 0.000029570990806945864, |
|
"loss": 0.2905, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.7587433457374573, |
|
"learning_rate": 0.000029466761168205792, |
|
"loss": 0.3187, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.8162456750869751, |
|
"learning_rate": 0.00002936253152946572, |
|
"loss": 0.3101, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.5907912850379944, |
|
"learning_rate": 0.000029258301890725648, |
|
"loss": 0.2926, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9535139799118042, |
|
"learning_rate": 0.000029154072251985576, |
|
"loss": 0.3237, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.9831104874610901, |
|
"learning_rate": 0.000029049842613245504, |
|
"loss": 0.2922, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8171008229255676, |
|
"learning_rate": 0.00002894561297450543, |
|
"loss": 0.3127, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.9013519287109375, |
|
"learning_rate": 0.00002884138333576536, |
|
"loss": 0.2974, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.0881197452545166, |
|
"learning_rate": 0.000028737153697025287, |
|
"loss": 0.3033, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1.0165661573410034, |
|
"learning_rate": 0.000028632924058285215, |
|
"loss": 0.2969, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.9244837760925293, |
|
"learning_rate": 0.000028528694419545143, |
|
"loss": 0.3074, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8180138468742371, |
|
"learning_rate": 0.00002842446478080507, |
|
"loss": 0.3022, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8913204073905945, |
|
"learning_rate": 0.000028320235142065, |
|
"loss": 0.3041, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.944990336894989, |
|
"learning_rate": 0.000028216005503324927, |
|
"loss": 0.3034, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.6920465230941772, |
|
"learning_rate": 0.000028111775864584855, |
|
"loss": 0.287, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.9530170559883118, |
|
"learning_rate": 0.000028007546225844783, |
|
"loss": 0.2924, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 1.0395070314407349, |
|
"learning_rate": 0.00002790331658710471, |
|
"loss": 0.2976, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 1.0013209581375122, |
|
"learning_rate": 0.00002779908694836464, |
|
"loss": 0.2765, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6239937543869019, |
|
"learning_rate": 0.000027694857309624566, |
|
"loss": 0.3001, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6583734154701233, |
|
"learning_rate": 0.000027590627670884494, |
|
"loss": 0.2984, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.9175625443458557, |
|
"learning_rate": 0.000027486398032144422, |
|
"loss": 0.305, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.192954182624817, |
|
"learning_rate": 0.00002738216839340435, |
|
"loss": 0.2913, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.9170815944671631, |
|
"learning_rate": 0.000027277938754664278, |
|
"loss": 0.2886, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.7114281058311462, |
|
"learning_rate": 0.000027173709115924206, |
|
"loss": 0.2872, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.1340280771255493, |
|
"learning_rate": 0.000027069479477184134, |
|
"loss": 0.2927, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 1.0684434175491333, |
|
"learning_rate": 0.00002696524983844406, |
|
"loss": 0.2762, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.8208048343658447, |
|
"learning_rate": 0.00002686102019970399, |
|
"loss": 0.2933, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.9516797065734863, |
|
"learning_rate": 0.000026756790560963917, |
|
"loss": 0.2809, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.7305499315261841, |
|
"learning_rate": 0.000026652560922223845, |
|
"loss": 0.2968, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 0.7523550391197205, |
|
"learning_rate": 0.000026548331283483773, |
|
"loss": 0.289, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 1.0975223779678345, |
|
"learning_rate": 0.0000264441016447437, |
|
"loss": 0.2961, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.906672477722168, |
|
"learning_rate": 0.00002633987200600363, |
|
"loss": 0.2904, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 0.8232882022857666, |
|
"learning_rate": 0.000026235642367263557, |
|
"loss": 0.2876, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 1.0251737833023071, |
|
"learning_rate": 0.000026131412728523485, |
|
"loss": 0.2842, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 1.1119544506072998, |
|
"learning_rate": 0.000026027183089783413, |
|
"loss": 0.2778, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.1404234170913696, |
|
"learning_rate": 0.00002592295345104334, |
|
"loss": 0.2865, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.0083096027374268, |
|
"learning_rate": 0.00002581872381230327, |
|
"loss": 0.2923, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1.1562743186950684, |
|
"learning_rate": 0.000025714494173563196, |
|
"loss": 0.3009, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 0.7771180272102356, |
|
"learning_rate": 0.000025610264534823124, |
|
"loss": 0.2938, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8902438282966614, |
|
"learning_rate": 0.000025506034896083052, |
|
"loss": 0.3102, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.9011629819869995, |
|
"learning_rate": 0.00002540180525734298, |
|
"loss": 0.2885, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.8738746643066406, |
|
"learning_rate": 0.000025297575618602908, |
|
"loss": 0.2739, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.5818448662757874, |
|
"learning_rate": 0.000025193345979862836, |
|
"loss": 0.281, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 0.927828311920166, |
|
"learning_rate": 0.000025089116341122764, |
|
"loss": 0.3039, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.8303672671318054, |
|
"learning_rate": 0.00002498488670238269, |
|
"loss": 0.2921, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 1.7229905128479004, |
|
"learning_rate": 0.000024880657063642616, |
|
"loss": 0.3003, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.898224949836731, |
|
"learning_rate": 0.000024776427424902544, |
|
"loss": 0.3034, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.7359880805015564, |
|
"learning_rate": 0.000024672197786162472, |
|
"loss": 0.2918, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 0.9466255903244019, |
|
"learning_rate": 0.0000245679681474224, |
|
"loss": 0.2982, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 0.9300728440284729, |
|
"learning_rate": 0.00002446373850868233, |
|
"loss": 0.3049, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.2316802740097046, |
|
"learning_rate": 0.00002435950886994226, |
|
"loss": 0.3088, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.1127874851226807, |
|
"learning_rate": 0.000024255279231202187, |
|
"loss": 0.2844, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.0291988849639893, |
|
"learning_rate": 0.000024151049592462115, |
|
"loss": 0.2844, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 1.0049411058425903, |
|
"learning_rate": 0.000024046819953722043, |
|
"loss": 0.2902, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 7.540700912475586, |
|
"learning_rate": 0.00002394259031498197, |
|
"loss": 0.2966, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.7970338463783264, |
|
"learning_rate": 0.0000238383606762419, |
|
"loss": 0.2714, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.7852484583854675, |
|
"learning_rate": 0.000023734131037501827, |
|
"loss": 0.3103, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 1.1321452856063843, |
|
"learning_rate": 0.000023629901398761754, |
|
"loss": 0.2939, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"grad_norm": 0.8240513801574707, |
|
"learning_rate": 0.000023525671760021682, |
|
"loss": 0.2933, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.8263924717903137, |
|
"learning_rate": 0.00002342144212128161, |
|
"loss": 0.2764, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.48156994581222534, |
|
"learning_rate": 0.000023317212482541538, |
|
"loss": 0.2785, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.7034128904342651, |
|
"learning_rate": 0.000023212982843801466, |
|
"loss": 0.2884, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 0.7832284569740295, |
|
"learning_rate": 0.000023108753205061394, |
|
"loss": 0.2933, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1.0077266693115234, |
|
"learning_rate": 0.000023004523566321322, |
|
"loss": 0.2929, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.7659132480621338, |
|
"learning_rate": 0.00002290029392758125, |
|
"loss": 0.277, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.7871528267860413, |
|
"learning_rate": 0.000022796064288841178, |
|
"loss": 0.2995, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.9168059825897217, |
|
"learning_rate": 0.000022691834650101106, |
|
"loss": 0.3038, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.6936691403388977, |
|
"learning_rate": 0.000022587605011361033, |
|
"loss": 0.2838, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 0.9125056266784668, |
|
"learning_rate": 0.00002248337537262096, |
|
"loss": 0.2832, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.8433107733726501, |
|
"learning_rate": 0.00002237914573388089, |
|
"loss": 0.2963, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.6140872836112976, |
|
"learning_rate": 0.000022274916095140817, |
|
"loss": 0.256, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.9464013576507568, |
|
"learning_rate": 0.000022170686456400745, |
|
"loss": 0.279, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 0.6698806881904602, |
|
"learning_rate": 0.000022066456817660673, |
|
"loss": 0.2664, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 1.0536932945251465, |
|
"learning_rate": 0.0000219622271789206, |
|
"loss": 0.2768, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.9017496705055237, |
|
"learning_rate": 0.00002185799754018053, |
|
"loss": 0.2723, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.9188879728317261, |
|
"learning_rate": 0.000021753767901440457, |
|
"loss": 0.2925, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.716852068901062, |
|
"learning_rate": 0.000021649538262700385, |
|
"loss": 0.2789, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 0.8829745650291443, |
|
"learning_rate": 0.000021545308623960312, |
|
"loss": 0.2613, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8111447095870972, |
|
"learning_rate": 0.00002144107898522024, |
|
"loss": 0.2847, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.069851279258728, |
|
"learning_rate": 0.000021336849346480168, |
|
"loss": 0.2865, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 1.1249984502792358, |
|
"learning_rate": 0.000021232619707740096, |
|
"loss": 0.285, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8425958156585693, |
|
"learning_rate": 0.000021128390069000024, |
|
"loss": 0.2822, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 0.8334409594535828, |
|
"learning_rate": 0.000021024160430259952, |
|
"loss": 0.2915, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.8733804225921631, |
|
"learning_rate": 0.00002091993079151988, |
|
"loss": 0.2784, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 1.1405348777770996, |
|
"learning_rate": 0.000020815701152779808, |
|
"loss": 0.2632, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.930985689163208, |
|
"learning_rate": 0.000020711471514039732, |
|
"loss": 0.2804, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.7500423789024353, |
|
"learning_rate": 0.00002060724187529966, |
|
"loss": 0.271, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 0.830214262008667, |
|
"learning_rate": 0.000020503012236559588, |
|
"loss": 0.2802, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.1190361976623535, |
|
"learning_rate": 0.000020398782597819516, |
|
"loss": 0.2978, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.938105583190918, |
|
"learning_rate": 0.000020294552959079444, |
|
"loss": 0.273, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1.2504827976226807, |
|
"learning_rate": 0.000020190323320339372, |
|
"loss": 0.2802, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.6229212284088135, |
|
"learning_rate": 0.0000200860936815993, |
|
"loss": 0.2751, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 0.917674720287323, |
|
"learning_rate": 0.000019981864042859228, |
|
"loss": 0.2668, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.7844977378845215, |
|
"learning_rate": 0.000019877634404119155, |
|
"loss": 0.2772, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.7926586866378784, |
|
"learning_rate": 0.000019773404765379083, |
|
"loss": 0.2851, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.9423878192901611, |
|
"learning_rate": 0.00001966917512663901, |
|
"loss": 0.2973, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 1.1130839586257935, |
|
"learning_rate": 0.00001956494548789894, |
|
"loss": 0.2858, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 0.8832184076309204, |
|
"learning_rate": 0.000019460715849158867, |
|
"loss": 0.2714, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7974615693092346, |
|
"learning_rate": 0.000019356486210418795, |
|
"loss": 0.2672, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 1.0496609210968018, |
|
"learning_rate": 0.000019252256571678723, |
|
"loss": 0.2967, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.7052696943283081, |
|
"learning_rate": 0.00001914802693293865, |
|
"loss": 0.2682, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"grad_norm": 0.8177437782287598, |
|
"learning_rate": 0.00001904379729419858, |
|
"loss": 0.2772, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.1367732286453247, |
|
"learning_rate": 0.000018939567655458507, |
|
"loss": 0.2751, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.4945389032363892, |
|
"learning_rate": 0.000018835338016718434, |
|
"loss": 0.2845, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 1.035019040107727, |
|
"learning_rate": 0.000018731108377978362, |
|
"loss": 0.2881, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.4365576207637787, |
|
"learning_rate": 0.00001862687873923829, |
|
"loss": 0.2805, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 0.6886979341506958, |
|
"learning_rate": 0.000018522649100498218, |
|
"loss": 0.2755, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.7902783751487732, |
|
"learning_rate": 0.000018418419461758146, |
|
"loss": 0.2746, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.1486902236938477, |
|
"learning_rate": 0.000018314189823018074, |
|
"loss": 0.2734, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8375297784805298, |
|
"learning_rate": 0.000018209960184278002, |
|
"loss": 0.2672, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.8426914811134338, |
|
"learning_rate": 0.00001810573054553793, |
|
"loss": 0.266, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 1.0207959413528442, |
|
"learning_rate": 0.000018001500906797858, |
|
"loss": 0.2879, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8999843597412109, |
|
"learning_rate": 0.000017897271268057786, |
|
"loss": 0.2617, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.7014563083648682, |
|
"learning_rate": 0.000017793041629317713, |
|
"loss": 0.2757, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.8926547765731812, |
|
"learning_rate": 0.00001768881199057764, |
|
"loss": 0.2805, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.9726804494857788, |
|
"learning_rate": 0.00001758458235183757, |
|
"loss": 0.2824, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 0.949054479598999, |
|
"learning_rate": 0.000017480352713097497, |
|
"loss": 0.2578, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.0712580680847168, |
|
"learning_rate": 0.000017376123074357425, |
|
"loss": 0.259, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8288059234619141, |
|
"learning_rate": 0.000017271893435617353, |
|
"loss": 0.2788, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.9404420852661133, |
|
"learning_rate": 0.00001716766379687728, |
|
"loss": 0.2766, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 0.8590795993804932, |
|
"learning_rate": 0.00001706343415813721, |
|
"loss": 0.2768, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 1.1562901735305786, |
|
"learning_rate": 0.000016959204519397137, |
|
"loss": 0.2876, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8512832522392273, |
|
"learning_rate": 0.000016854974880657065, |
|
"loss": 0.2599, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.8616644740104675, |
|
"learning_rate": 0.000016750745241916992, |
|
"loss": 0.2725, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.4701212644577026, |
|
"learning_rate": 0.00001664651560317692, |
|
"loss": 0.2592, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 1.0344711542129517, |
|
"learning_rate": 0.000016542285964436848, |
|
"loss": 0.2782, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"grad_norm": 0.9090347290039062, |
|
"learning_rate": 0.000016438056325696776, |
|
"loss": 0.2683, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.9814984798431396, |
|
"learning_rate": 0.000016333826686956704, |
|
"loss": 0.2692, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8351295590400696, |
|
"learning_rate": 0.000016229597048216632, |
|
"loss": 0.2707, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8099172115325928, |
|
"learning_rate": 0.00001612536740947656, |
|
"loss": 0.2737, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 0.8135939240455627, |
|
"learning_rate": 0.000016021137770736488, |
|
"loss": 0.2641, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 1.0211602449417114, |
|
"learning_rate": 0.000015916908131996416, |
|
"loss": 0.2719, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.499255657196045, |
|
"learning_rate": 0.000015812678493256344, |
|
"loss": 0.2774, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 1.062286376953125, |
|
"learning_rate": 0.00001570844885451627, |
|
"loss": 0.2631, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.7480090260505676, |
|
"learning_rate": 0.0000156042192157762, |
|
"loss": 0.2697, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"grad_norm": 0.8319742679595947, |
|
"learning_rate": 0.000015499989577036127, |
|
"loss": 0.286, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.1275886297225952, |
|
"learning_rate": 0.000015395759938296052, |
|
"loss": 0.2855, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.9062207937240601, |
|
"learning_rate": 0.00001529153029955598, |
|
"loss": 0.268, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8947410583496094, |
|
"learning_rate": 0.00001518730066081591, |
|
"loss": 0.2679, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 1.0549769401550293, |
|
"learning_rate": 0.000015083071022075837, |
|
"loss": 0.2691, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 0.8998539447784424, |
|
"learning_rate": 0.000014978841383335765, |
|
"loss": 0.2688, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.6037629246711731, |
|
"learning_rate": 0.000014874611744595693, |
|
"loss": 0.267, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 0.92879718542099, |
|
"learning_rate": 0.000014770382105855621, |
|
"loss": 0.2667, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.562403678894043, |
|
"learning_rate": 0.000014666152467115549, |
|
"loss": 0.2812, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.0794057846069336, |
|
"learning_rate": 0.000014561922828375477, |
|
"loss": 0.2691, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1.1239275932312012, |
|
"learning_rate": 0.000014457693189635405, |
|
"loss": 0.2626, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.2695655822753906, |
|
"learning_rate": 0.000014353463550895332, |
|
"loss": 0.2679, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.9684453010559082, |
|
"learning_rate": 0.00001424923391215526, |
|
"loss": 0.2689, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8536735773086548, |
|
"learning_rate": 0.000014145004273415188, |
|
"loss": 0.2505, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 0.8420373797416687, |
|
"learning_rate": 0.000014040774634675116, |
|
"loss": 0.2808, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 1.7913522720336914, |
|
"learning_rate": 0.000013936544995935044, |
|
"loss": 0.2713, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.9944103360176086, |
|
"learning_rate": 0.000013832315357194972, |
|
"loss": 0.2767, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.647555947303772, |
|
"learning_rate": 0.0000137280857184549, |
|
"loss": 0.2637, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.6362162828445435, |
|
"learning_rate": 0.000013623856079714828, |
|
"loss": 0.2626, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8359599113464355, |
|
"learning_rate": 0.000013519626440974756, |
|
"loss": 0.2819, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"grad_norm": 0.8736098408699036, |
|
"learning_rate": 0.000013415396802234684, |
|
"loss": 0.2699, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8288611173629761, |
|
"learning_rate": 0.000013311167163494611, |
|
"loss": 0.2621, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.7499838471412659, |
|
"learning_rate": 0.00001320693752475454, |
|
"loss": 0.2692, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 1.0514994859695435, |
|
"learning_rate": 0.000013102707886014467, |
|
"loss": 0.2712, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.8135703206062317, |
|
"learning_rate": 0.000012998478247274395, |
|
"loss": 0.267, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 0.780437707901001, |
|
"learning_rate": 0.000012894248608534323, |
|
"loss": 0.2578, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 1.0182126760482788, |
|
"learning_rate": 0.000012790018969794251, |
|
"loss": 0.2848, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.9728681445121765, |
|
"learning_rate": 0.000012685789331054179, |
|
"loss": 0.2772, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.7952257394790649, |
|
"learning_rate": 0.000012581559692314107, |
|
"loss": 0.261, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 0.7164922952651978, |
|
"learning_rate": 0.000012477330053574035, |
|
"loss": 0.2403, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.6388373970985413, |
|
"learning_rate": 0.000012373100414833963, |
|
"loss": 0.2705, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1.4427376985549927, |
|
"learning_rate": 0.00001226887077609389, |
|
"loss": 0.277, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.9559366106987, |
|
"learning_rate": 0.000012164641137353818, |
|
"loss": 0.2784, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.8561313152313232, |
|
"learning_rate": 0.000012060411498613746, |
|
"loss": 0.272, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 0.9756980538368225, |
|
"learning_rate": 0.000011956181859873674, |
|
"loss": 0.2644, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.350045919418335, |
|
"learning_rate": 0.000011851952221133602, |
|
"loss": 0.2726, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.2612251043319702, |
|
"learning_rate": 0.00001174772258239353, |
|
"loss": 0.2696, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.9203113913536072, |
|
"learning_rate": 0.000011643492943653458, |
|
"loss": 0.2721, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 1.0679327249526978, |
|
"learning_rate": 0.000011539263304913386, |
|
"loss": 0.2738, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 0.7020294070243835, |
|
"learning_rate": 0.000011435033666173314, |
|
"loss": 0.2806, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.9969758987426758, |
|
"learning_rate": 0.000011330804027433242, |
|
"loss": 0.2556, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.7139135003089905, |
|
"learning_rate": 0.00001122657438869317, |
|
"loss": 0.2739, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 1.3120732307434082, |
|
"learning_rate": 0.000011122344749953097, |
|
"loss": 0.2774, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 0.9226138591766357, |
|
"learning_rate": 0.000011018115111213025, |
|
"loss": 0.2414, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 1.024018406867981, |
|
"learning_rate": 0.000010913885472472953, |
|
"loss": 0.2453, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.7024143934249878, |
|
"learning_rate": 0.000010809655833732881, |
|
"loss": 0.2681, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.008748173713684, |
|
"learning_rate": 0.000010705426194992809, |
|
"loss": 0.2656, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.8661462068557739, |
|
"learning_rate": 0.000010601196556252737, |
|
"loss": 0.2679, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 0.9956786632537842, |
|
"learning_rate": 0.000010496966917512665, |
|
"loss": 0.2622, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 1.3166755437850952, |
|
"learning_rate": 0.000010392737278772593, |
|
"loss": 0.2595, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 1.0491749048233032, |
|
"learning_rate": 0.00001028850764003252, |
|
"loss": 0.2682, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 1.0107011795043945, |
|
"learning_rate": 0.000010184278001292448, |
|
"loss": 0.2544, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 1.0187588930130005, |
|
"learning_rate": 0.000010080048362552376, |
|
"loss": 0.2611, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.8718214631080627, |
|
"learning_rate": 0.000009975818723812304, |
|
"loss": 0.2567, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 0.7450688481330872, |
|
"learning_rate": 0.000009871589085072232, |
|
"loss": 0.2642, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 1.0107316970825195, |
|
"learning_rate": 0.00000976735944633216, |
|
"loss": 0.2613, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8210121393203735, |
|
"learning_rate": 0.000009663129807592088, |
|
"loss": 0.2768, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8526451587677002, |
|
"learning_rate": 0.000009558900168852016, |
|
"loss": 0.2473, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.7720713019371033, |
|
"learning_rate": 0.000009454670530111944, |
|
"loss": 0.2715, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 0.8657224774360657, |
|
"learning_rate": 0.000009350440891371872, |
|
"loss": 0.2612, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9144985675811768, |
|
"learning_rate": 0.0000092462112526318, |
|
"loss": 0.2648, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1.0658880472183228, |
|
"learning_rate": 0.000009141981613891727, |
|
"loss": 0.2661, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9904933571815491, |
|
"learning_rate": 0.000009037751975151655, |
|
"loss": 0.2513, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 0.9688336253166199, |
|
"learning_rate": 0.000008933522336411583, |
|
"loss": 0.2622, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.6780712604522705, |
|
"learning_rate": 0.000008829292697671511, |
|
"loss": 0.2665, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.9881765246391296, |
|
"learning_rate": 0.000008725063058931437, |
|
"loss": 0.2486, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.9511051774024963, |
|
"learning_rate": 0.000008620833420191365, |
|
"loss": 0.2655, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 0.8047460317611694, |
|
"learning_rate": 0.000008516603781451293, |
|
"loss": 0.262, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 1.0902456045150757, |
|
"learning_rate": 0.000008412374142711221, |
|
"loss": 0.2681, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9015824794769287, |
|
"learning_rate": 0.000008308144503971149, |
|
"loss": 0.2689, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9453705549240112, |
|
"learning_rate": 0.000008203914865231077, |
|
"loss": 0.2693, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.8557701706886292, |
|
"learning_rate": 0.000008099685226491005, |
|
"loss": 0.2753, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 1.022003173828125, |
|
"learning_rate": 0.000007995455587750933, |
|
"loss": 0.2646, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 0.9399078488349915, |
|
"learning_rate": 0.00000789122594901086, |
|
"loss": 0.2671, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.1249728202819824, |
|
"learning_rate": 0.000007786996310270788, |
|
"loss": 0.2655, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.2603721618652344, |
|
"learning_rate": 0.000007682766671530716, |
|
"loss": 0.262, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.6507100462913513, |
|
"learning_rate": 0.000007578537032790644, |
|
"loss": 0.2648, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 1.0216801166534424, |
|
"learning_rate": 0.000007474307394050572, |
|
"loss": 0.2727, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 0.8479890823364258, |
|
"learning_rate": 0.0000073700777553105, |
|
"loss": 0.2644, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.953072726726532, |
|
"learning_rate": 0.000007265848116570428, |
|
"loss": 0.2601, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.7130771279335022, |
|
"learning_rate": 0.000007161618477830357, |
|
"loss": 0.2634, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.1879602670669556, |
|
"learning_rate": 0.000007057388839090285, |
|
"loss": 0.2588, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 0.9556657075881958, |
|
"learning_rate": 0.0000069531592003502125, |
|
"loss": 0.2673, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 1.1229112148284912, |
|
"learning_rate": 0.0000068489295616101404, |
|
"loss": 0.264, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.2128034830093384, |
|
"learning_rate": 0.000006744699922870068, |
|
"loss": 0.2626, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.9225343465805054, |
|
"learning_rate": 0.000006640470284129996, |
|
"loss": 0.2713, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1.0495810508728027, |
|
"learning_rate": 0.000006536240645389924, |
|
"loss": 0.2739, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.5960795879364014, |
|
"learning_rate": 0.000006432011006649852, |
|
"loss": 0.2629, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 0.653361439704895, |
|
"learning_rate": 0.00000632778136790978, |
|
"loss": 0.2585, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8684765100479126, |
|
"learning_rate": 0.000006223551729169707, |
|
"loss": 0.2426, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.5730326771736145, |
|
"learning_rate": 0.000006119322090429635, |
|
"loss": 0.2537, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.8172651529312134, |
|
"learning_rate": 0.000006015092451689563, |
|
"loss": 0.2553, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"grad_norm": 0.9099143147468567, |
|
"learning_rate": 0.000005910862812949491, |
|
"loss": 0.2569, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.9645257592201233, |
|
"learning_rate": 0.000005806633174209419, |
|
"loss": 0.2689, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 0.6569193005561829, |
|
"learning_rate": 0.0000057024035354693465, |
|
"loss": 0.2662, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.2637701034545898, |
|
"learning_rate": 0.000005598173896729274, |
|
"loss": 0.2874, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.0651448965072632, |
|
"learning_rate": 0.000005493944257989202, |
|
"loss": 0.255, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 1.1091278791427612, |
|
"learning_rate": 0.000005389714619249129, |
|
"loss": 0.283, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.9137535095214844, |
|
"learning_rate": 0.000005285484980509057, |
|
"loss": 0.2674, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.8751276731491089, |
|
"learning_rate": 0.000005181255341768985, |
|
"loss": 0.2638, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.9242710471153259, |
|
"learning_rate": 0.000005077025703028913, |
|
"loss": 0.2582, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.7860730886459351, |
|
"learning_rate": 0.000004972796064288841, |
|
"loss": 0.2557, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 0.9240506291389465, |
|
"learning_rate": 0.000004868566425548769, |
|
"loss": 0.2591, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.0865675210952759, |
|
"learning_rate": 0.000004764336786808697, |
|
"loss": 0.2511, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 1.0534336566925049, |
|
"learning_rate": 0.0000046601071480686255, |
|
"loss": 0.2651, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.9301556348800659, |
|
"learning_rate": 0.000004555877509328553, |
|
"loss": 0.2414, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.9242545366287231, |
|
"learning_rate": 0.000004451647870588481, |
|
"loss": 0.2636, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 0.7292113304138184, |
|
"learning_rate": 0.000004347418231848409, |
|
"loss": 0.2672, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1.2003198862075806, |
|
"learning_rate": 0.000004243188593108337, |
|
"loss": 0.2661, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.9352708458900452, |
|
"learning_rate": 0.000004138958954368265, |
|
"loss": 0.275, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.7767049074172974, |
|
"learning_rate": 0.000004034729315628192, |
|
"loss": 0.2709, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.8384830355644226, |
|
"learning_rate": 0.00000393049967688812, |
|
"loss": 0.2655, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 0.9187309145927429, |
|
"learning_rate": 0.000003826270038148048, |
|
"loss": 0.2609, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8865881562232971, |
|
"learning_rate": 0.0000037220403994079758, |
|
"loss": 0.2677, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.0104628801345825, |
|
"learning_rate": 0.0000036178107606679037, |
|
"loss": 0.2649, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 0.8638792634010315, |
|
"learning_rate": 0.0000035135811219278316, |
|
"loss": 0.2694, |
|
"step": 45100 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.0379241704940796, |
|
"learning_rate": 0.0000034093514831877595, |
|
"loss": 0.2489, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 1.123955488204956, |
|
"learning_rate": 0.0000033051218444476874, |
|
"loss": 0.2576, |
|
"step": 45300 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.1420966386795044, |
|
"learning_rate": 0.0000032008922057076153, |
|
"loss": 0.2706, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.0519020557403564, |
|
"learning_rate": 0.0000030966625669675427, |
|
"loss": 0.2616, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 1.040588140487671, |
|
"learning_rate": 0.0000029924329282274706, |
|
"loss": 0.2546, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.8306300640106201, |
|
"learning_rate": 0.0000028882032894873985, |
|
"loss": 0.2683, |
|
"step": 45700 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 0.6718655228614807, |
|
"learning_rate": 0.0000027839736507473264, |
|
"loss": 0.2483, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8920142650604248, |
|
"learning_rate": 0.0000026797440120072548, |
|
"loss": 0.2612, |
|
"step": 45900 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 1.09241783618927, |
|
"learning_rate": 0.0000025755143732671827, |
|
"loss": 0.2483, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.9406448602676392, |
|
"learning_rate": 0.0000024712847345271106, |
|
"loss": 0.2639, |
|
"step": 46100 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.9145969748497009, |
|
"learning_rate": 0.000002367055095787038, |
|
"loss": 0.2493, |
|
"step": 46200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.722114086151123, |
|
"learning_rate": 0.000002262825457046966, |
|
"loss": 0.266, |
|
"step": 46300 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.7077659964561462, |
|
"learning_rate": 0.000002158595818306894, |
|
"loss": 0.2712, |
|
"step": 46400 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.7417329549789429, |
|
"learning_rate": 0.0000020543661795668217, |
|
"loss": 0.2435, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.8229231238365173, |
|
"learning_rate": 0.0000019501365408267496, |
|
"loss": 0.2647, |
|
"step": 46600 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.49395614862442017, |
|
"learning_rate": 0.0000018459069020866775, |
|
"loss": 0.2508, |
|
"step": 46700 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 1.0020008087158203, |
|
"learning_rate": 0.0000017416772633466052, |
|
"loss": 0.2505, |
|
"step": 46800 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8525140881538391, |
|
"learning_rate": 0.0000016374476246065331, |
|
"loss": 0.2518, |
|
"step": 46900 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.8383646011352539, |
|
"learning_rate": 0.000001533217985866461, |
|
"loss": 0.2768, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.586855947971344, |
|
"learning_rate": 0.000001428988347126389, |
|
"loss": 0.2379, |
|
"step": 47100 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"grad_norm": 0.7666946053504944, |
|
"learning_rate": 0.0000013247587083863168, |
|
"loss": 0.2501, |
|
"step": 47200 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.0712182521820068, |
|
"learning_rate": 0.0000012205290696462445, |
|
"loss": 0.2539, |
|
"step": 47300 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.9935363531112671, |
|
"learning_rate": 0.0000011162994309061726, |
|
"loss": 0.2603, |
|
"step": 47400 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.8433282971382141, |
|
"learning_rate": 0.0000010120697921661005, |
|
"loss": 0.261, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 0.942387044429779, |
|
"learning_rate": 9.078401534260282e-7, |
|
"loss": 0.2702, |
|
"step": 47600 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1.199331521987915, |
|
"learning_rate": 8.036105146859561e-7, |
|
"loss": 0.2452, |
|
"step": 47700 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.9567368626594543, |
|
"learning_rate": 6.993808759458839e-7, |
|
"loss": 0.2691, |
|
"step": 47800 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 1.0403625965118408, |
|
"learning_rate": 5.951512372058119e-7, |
|
"loss": 0.2702, |
|
"step": 47900 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"grad_norm": 0.9573351740837097, |
|
"learning_rate": 4.909215984657397e-7, |
|
"loss": 0.2539, |
|
"step": 48000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 48471, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 16000, |
|
"total_flos": 104021020901376000, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|