|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.324324324324325, |
|
"eval_steps": 500, |
|
"global_step": 120, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.036036036036036036, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.2548, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.07207207207207207, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 1.7083, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.10810810810810811, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 1.9566, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.14414414414414414, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.2875, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.18018018018018017, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.0179, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.21621621621621623, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.1465, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.25225225225225223, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.0991, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.2882882882882883, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0, |
|
"loss": 2.0996, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.32432432432432434, |
|
"grad_norm": 3.127271890640259, |
|
"learning_rate": 2e-05, |
|
"loss": 2.0748, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.36036036036036034, |
|
"grad_norm": 2.1929919719696045, |
|
"learning_rate": 4e-05, |
|
"loss": 1.7543, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.3963963963963964, |
|
"grad_norm": 3.9285449981689453, |
|
"learning_rate": 6e-05, |
|
"loss": 2.1182, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.43243243243243246, |
|
"grad_norm": 3.9492204189300537, |
|
"learning_rate": 8e-05, |
|
"loss": 2.1398, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.46846846846846846, |
|
"grad_norm": 4.697811603546143, |
|
"learning_rate": 0.0001, |
|
"loss": 1.944, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.5045045045045045, |
|
"grad_norm": 2.257394313812256, |
|
"learning_rate": 0.00012, |
|
"loss": 1.6118, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.5405405405405406, |
|
"grad_norm": 1.470400333404541, |
|
"learning_rate": 0.00014, |
|
"loss": 1.6489, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.5765765765765766, |
|
"grad_norm": 0.9114196300506592, |
|
"learning_rate": 0.00016, |
|
"loss": 1.3067, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.6126126126126126, |
|
"grad_norm": 1.9045084714889526, |
|
"learning_rate": 0.00018, |
|
"loss": 1.4384, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.6486486486486487, |
|
"grad_norm": 1.669812560081482, |
|
"learning_rate": 0.0002, |
|
"loss": 1.3614, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.6846846846846847, |
|
"grad_norm": 0.982727587223053, |
|
"learning_rate": 0.00019818181818181821, |
|
"loss": 1.29, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.7207207207207207, |
|
"grad_norm": 0.5169259309768677, |
|
"learning_rate": 0.00019636363636363636, |
|
"loss": 1.2386, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.7567567567567568, |
|
"grad_norm": 0.3028675317764282, |
|
"learning_rate": 0.00019454545454545457, |
|
"loss": 1.1707, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.7927927927927928, |
|
"grad_norm": 0.35564807057380676, |
|
"learning_rate": 0.00019272727272727274, |
|
"loss": 0.9072, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.8288288288288288, |
|
"grad_norm": 0.30732882022857666, |
|
"learning_rate": 0.00019090909090909092, |
|
"loss": 0.9289, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.8648648648648649, |
|
"grad_norm": 0.31778138875961304, |
|
"learning_rate": 0.0001890909090909091, |
|
"loss": 0.9955, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.9009009009009009, |
|
"grad_norm": 0.3151562809944153, |
|
"learning_rate": 0.00018727272727272728, |
|
"loss": 0.8705, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.9369369369369369, |
|
"grad_norm": 0.3682255744934082, |
|
"learning_rate": 0.00018545454545454545, |
|
"loss": 1.197, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.972972972972973, |
|
"grad_norm": 0.25122687220573425, |
|
"learning_rate": 0.00018363636363636366, |
|
"loss": 1.0997, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 1.009009009009009, |
|
"grad_norm": 0.23811328411102295, |
|
"learning_rate": 0.00018181818181818183, |
|
"loss": 1.0265, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.045045045045045, |
|
"grad_norm": 0.23290874063968658, |
|
"learning_rate": 0.00018, |
|
"loss": 0.7358, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.0810810810810811, |
|
"grad_norm": 0.21789032220840454, |
|
"learning_rate": 0.0001781818181818182, |
|
"loss": 0.7165, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.117117117117117, |
|
"grad_norm": 0.3115369975566864, |
|
"learning_rate": 0.00017636363636363637, |
|
"loss": 0.7812, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 1.1531531531531531, |
|
"grad_norm": 0.21534988284111023, |
|
"learning_rate": 0.00017454545454545454, |
|
"loss": 0.706, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.1891891891891893, |
|
"grad_norm": 0.21201848983764648, |
|
"learning_rate": 0.00017272727272727275, |
|
"loss": 0.6389, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.2252252252252251, |
|
"grad_norm": 0.16613227128982544, |
|
"learning_rate": 0.0001709090909090909, |
|
"loss": 0.7041, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.2612612612612613, |
|
"grad_norm": 0.26245370507240295, |
|
"learning_rate": 0.0001690909090909091, |
|
"loss": 0.6188, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 1.2972972972972974, |
|
"grad_norm": 0.17487148940563202, |
|
"learning_rate": 0.00016727272727272728, |
|
"loss": 0.7382, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 0.22627073526382446, |
|
"learning_rate": 0.00016545454545454545, |
|
"loss": 0.9381, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 1.3693693693693694, |
|
"grad_norm": 0.19883334636688232, |
|
"learning_rate": 0.00016363636363636366, |
|
"loss": 1.0042, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.4054054054054055, |
|
"grad_norm": 0.18648912012577057, |
|
"learning_rate": 0.00016181818181818184, |
|
"loss": 0.7297, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 1.4414414414414414, |
|
"grad_norm": 0.2486204355955124, |
|
"learning_rate": 0.00016, |
|
"loss": 1.0105, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.4774774774774775, |
|
"grad_norm": 0.19586724042892456, |
|
"learning_rate": 0.0001581818181818182, |
|
"loss": 0.8904, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 1.5135135135135136, |
|
"grad_norm": 0.1361636519432068, |
|
"learning_rate": 0.00015636363636363637, |
|
"loss": 0.7655, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.5495495495495497, |
|
"grad_norm": 0.1422913819551468, |
|
"learning_rate": 0.00015454545454545454, |
|
"loss": 0.884, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 1.5855855855855856, |
|
"grad_norm": 0.12737137079238892, |
|
"learning_rate": 0.00015272727272727275, |
|
"loss": 0.4881, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.6216216216216215, |
|
"grad_norm": 0.14031246304512024, |
|
"learning_rate": 0.0001509090909090909, |
|
"loss": 0.4856, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.6576576576576576, |
|
"grad_norm": 0.152836412191391, |
|
"learning_rate": 0.0001490909090909091, |
|
"loss": 0.5526, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.6936936936936937, |
|
"grad_norm": 0.16750769317150116, |
|
"learning_rate": 0.00014727272727272728, |
|
"loss": 0.6862, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 1.7297297297297298, |
|
"grad_norm": 0.15049736201763153, |
|
"learning_rate": 0.00014545454545454546, |
|
"loss": 0.6051, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.7657657657657657, |
|
"grad_norm": 0.1748289316892624, |
|
"learning_rate": 0.00014363636363636363, |
|
"loss": 0.6521, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 1.8018018018018018, |
|
"grad_norm": 0.12289358675479889, |
|
"learning_rate": 0.00014181818181818184, |
|
"loss": 0.5631, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.8378378378378377, |
|
"grad_norm": 0.14275340735912323, |
|
"learning_rate": 0.00014, |
|
"loss": 0.3934, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 1.8738738738738738, |
|
"grad_norm": 0.14242391288280487, |
|
"learning_rate": 0.0001381818181818182, |
|
"loss": 0.4372, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.90990990990991, |
|
"grad_norm": 0.1477702260017395, |
|
"learning_rate": 0.00013636363636363637, |
|
"loss": 0.7059, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 1.945945945945946, |
|
"grad_norm": 0.14974962174892426, |
|
"learning_rate": 0.00013454545454545455, |
|
"loss": 0.7635, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.981981981981982, |
|
"grad_norm": 0.14689475297927856, |
|
"learning_rate": 0.00013272727272727275, |
|
"loss": 0.7075, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 2.018018018018018, |
|
"grad_norm": 0.1843816339969635, |
|
"learning_rate": 0.00013090909090909093, |
|
"loss": 0.6481, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 2.054054054054054, |
|
"grad_norm": 0.14116476476192474, |
|
"learning_rate": 0.0001290909090909091, |
|
"loss": 0.3951, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 2.09009009009009, |
|
"grad_norm": 0.15613681077957153, |
|
"learning_rate": 0.00012727272727272728, |
|
"loss": 0.4265, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 2.126126126126126, |
|
"grad_norm": 0.1918690949678421, |
|
"learning_rate": 0.00012545454545454546, |
|
"loss": 0.9761, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 2.1621621621621623, |
|
"grad_norm": 0.14524200558662415, |
|
"learning_rate": 0.00012363636363636364, |
|
"loss": 0.4906, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 2.1981981981981984, |
|
"grad_norm": 0.16868562996387482, |
|
"learning_rate": 0.00012181818181818183, |
|
"loss": 0.5676, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 2.234234234234234, |
|
"grad_norm": 0.14121738076210022, |
|
"learning_rate": 0.00012, |
|
"loss": 0.5162, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 2.27027027027027, |
|
"grad_norm": 0.180411696434021, |
|
"learning_rate": 0.0001181818181818182, |
|
"loss": 0.7837, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 2.3063063063063063, |
|
"grad_norm": 0.23065026104450226, |
|
"learning_rate": 0.00011636363636363636, |
|
"loss": 0.4959, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 2.3423423423423424, |
|
"grad_norm": 0.1973959058523178, |
|
"learning_rate": 0.00011454545454545456, |
|
"loss": 0.4391, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 2.3783783783783785, |
|
"grad_norm": 0.17862387001514435, |
|
"learning_rate": 0.00011272727272727272, |
|
"loss": 0.4177, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 2.4144144144144146, |
|
"grad_norm": 0.19349342584609985, |
|
"learning_rate": 0.00011090909090909092, |
|
"loss": 0.7445, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 2.4504504504504503, |
|
"grad_norm": 0.1336154341697693, |
|
"learning_rate": 0.00010909090909090909, |
|
"loss": 0.2191, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 2.4864864864864864, |
|
"grad_norm": 0.21199998259544373, |
|
"learning_rate": 0.00010727272727272728, |
|
"loss": 0.4845, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 2.5225225225225225, |
|
"grad_norm": 0.18532893061637878, |
|
"learning_rate": 0.00010545454545454545, |
|
"loss": 0.7252, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.5585585585585586, |
|
"grad_norm": 0.14763948321342468, |
|
"learning_rate": 0.00010363636363636364, |
|
"loss": 0.4816, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 2.5945945945945947, |
|
"grad_norm": 0.19213160872459412, |
|
"learning_rate": 0.00010181818181818181, |
|
"loss": 0.4775, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.6306306306306304, |
|
"grad_norm": 0.19077633321285248, |
|
"learning_rate": 0.0001, |
|
"loss": 0.451, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 0.15648439526557922, |
|
"learning_rate": 9.818181818181818e-05, |
|
"loss": 0.2877, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.7027027027027026, |
|
"grad_norm": 0.15609943866729736, |
|
"learning_rate": 9.636363636363637e-05, |
|
"loss": 0.2944, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 2.7387387387387387, |
|
"grad_norm": 0.3236052989959717, |
|
"learning_rate": 9.454545454545455e-05, |
|
"loss": 0.62, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.774774774774775, |
|
"grad_norm": 0.222801074385643, |
|
"learning_rate": 9.272727272727273e-05, |
|
"loss": 0.6515, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 2.810810810810811, |
|
"grad_norm": 0.2324117124080658, |
|
"learning_rate": 9.090909090909092e-05, |
|
"loss": 0.5766, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.846846846846847, |
|
"grad_norm": 0.20293325185775757, |
|
"learning_rate": 8.90909090909091e-05, |
|
"loss": 0.4531, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 2.8828828828828827, |
|
"grad_norm": 0.20165590941905975, |
|
"learning_rate": 8.727272727272727e-05, |
|
"loss": 0.5847, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.918918918918919, |
|
"grad_norm": 0.3000154197216034, |
|
"learning_rate": 8.545454545454545e-05, |
|
"loss": 0.4334, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 2.954954954954955, |
|
"grad_norm": 0.20399592816829681, |
|
"learning_rate": 8.363636363636364e-05, |
|
"loss": 0.2509, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.990990990990991, |
|
"grad_norm": 0.2015918642282486, |
|
"learning_rate": 8.181818181818183e-05, |
|
"loss": 0.3644, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 3.027027027027027, |
|
"grad_norm": 0.16283659636974335, |
|
"learning_rate": 8e-05, |
|
"loss": 0.3619, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 3.063063063063063, |
|
"grad_norm": 0.23384951055049896, |
|
"learning_rate": 7.818181818181818e-05, |
|
"loss": 0.4679, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 3.099099099099099, |
|
"grad_norm": 0.3092138171195984, |
|
"learning_rate": 7.636363636363637e-05, |
|
"loss": 0.504, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 3.135135135135135, |
|
"grad_norm": 0.2725450396537781, |
|
"learning_rate": 7.454545454545455e-05, |
|
"loss": 0.3763, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 3.171171171171171, |
|
"grad_norm": 0.2366735339164734, |
|
"learning_rate": 7.272727272727273e-05, |
|
"loss": 0.3513, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 3.2072072072072073, |
|
"grad_norm": 0.2573714554309845, |
|
"learning_rate": 7.090909090909092e-05, |
|
"loss": 0.3202, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 3.2432432432432434, |
|
"grad_norm": 0.39939162135124207, |
|
"learning_rate": 6.90909090909091e-05, |
|
"loss": 0.4374, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 3.279279279279279, |
|
"grad_norm": 0.3089916408061981, |
|
"learning_rate": 6.727272727272727e-05, |
|
"loss": 0.4418, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 3.315315315315315, |
|
"grad_norm": 0.26053813099861145, |
|
"learning_rate": 6.545454545454546e-05, |
|
"loss": 0.3301, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 3.3513513513513513, |
|
"grad_norm": 0.25036731362342834, |
|
"learning_rate": 6.363636363636364e-05, |
|
"loss": 0.3525, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 3.3873873873873874, |
|
"grad_norm": 0.4556732475757599, |
|
"learning_rate": 6.181818181818182e-05, |
|
"loss": 0.3305, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 3.4234234234234235, |
|
"grad_norm": 0.26551949977874756, |
|
"learning_rate": 6e-05, |
|
"loss": 0.2769, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 3.4594594594594597, |
|
"grad_norm": 0.2707691192626953, |
|
"learning_rate": 5.818181818181818e-05, |
|
"loss": 0.2242, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 3.4954954954954953, |
|
"grad_norm": 0.2323060780763626, |
|
"learning_rate": 5.636363636363636e-05, |
|
"loss": 0.4295, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 3.5315315315315314, |
|
"grad_norm": 0.26477906107902527, |
|
"learning_rate": 5.4545454545454546e-05, |
|
"loss": 0.3623, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 3.5675675675675675, |
|
"grad_norm": 0.2650379538536072, |
|
"learning_rate": 5.272727272727272e-05, |
|
"loss": 0.2648, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 3.6036036036036037, |
|
"grad_norm": 0.33467838168144226, |
|
"learning_rate": 5.090909090909091e-05, |
|
"loss": 0.3563, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.6396396396396398, |
|
"grad_norm": 0.2872948944568634, |
|
"learning_rate": 4.909090909090909e-05, |
|
"loss": 0.4959, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 3.6756756756756754, |
|
"grad_norm": 0.29369983077049255, |
|
"learning_rate": 4.7272727272727275e-05, |
|
"loss": 0.2455, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 3.7117117117117115, |
|
"grad_norm": 0.3010379374027252, |
|
"learning_rate": 4.545454545454546e-05, |
|
"loss": 0.4271, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 3.7477477477477477, |
|
"grad_norm": 0.31059783697128296, |
|
"learning_rate": 4.3636363636363636e-05, |
|
"loss": 0.4152, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 3.7837837837837838, |
|
"grad_norm": 0.2542209327220917, |
|
"learning_rate": 4.181818181818182e-05, |
|
"loss": 0.382, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 3.81981981981982, |
|
"grad_norm": 0.2900529205799103, |
|
"learning_rate": 4e-05, |
|
"loss": 0.2371, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 3.855855855855856, |
|
"grad_norm": 0.25713253021240234, |
|
"learning_rate": 3.818181818181819e-05, |
|
"loss": 0.2636, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 3.891891891891892, |
|
"grad_norm": 0.2913016378879547, |
|
"learning_rate": 3.6363636363636364e-05, |
|
"loss": 0.4829, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 3.9279279279279278, |
|
"grad_norm": 0.25630098581314087, |
|
"learning_rate": 3.454545454545455e-05, |
|
"loss": 0.2281, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 3.963963963963964, |
|
"grad_norm": 0.29647207260131836, |
|
"learning_rate": 3.272727272727273e-05, |
|
"loss": 0.3197, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.3527073860168457, |
|
"learning_rate": 3.090909090909091e-05, |
|
"loss": 0.3282, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 4.036036036036036, |
|
"grad_norm": 0.482024610042572, |
|
"learning_rate": 2.909090909090909e-05, |
|
"loss": 0.255, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 4.072072072072072, |
|
"grad_norm": 0.22163565456867218, |
|
"learning_rate": 2.7272727272727273e-05, |
|
"loss": 0.1771, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 4.108108108108108, |
|
"grad_norm": 0.25971901416778564, |
|
"learning_rate": 2.5454545454545454e-05, |
|
"loss": 0.2633, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 4.1441441441441444, |
|
"grad_norm": 0.30024221539497375, |
|
"learning_rate": 2.3636363636363637e-05, |
|
"loss": 0.2706, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 4.18018018018018, |
|
"grad_norm": 0.26607680320739746, |
|
"learning_rate": 2.1818181818181818e-05, |
|
"loss": 0.2513, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 4.216216216216216, |
|
"grad_norm": 0.2651140093803406, |
|
"learning_rate": 2e-05, |
|
"loss": 0.2306, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 4.252252252252252, |
|
"grad_norm": 0.29040372371673584, |
|
"learning_rate": 1.8181818181818182e-05, |
|
"loss": 0.348, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 4.288288288288288, |
|
"grad_norm": 0.3690294623374939, |
|
"learning_rate": 1.6363636363636366e-05, |
|
"loss": 0.1826, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 4.324324324324325, |
|
"grad_norm": 0.22308969497680664, |
|
"learning_rate": 1.4545454545454545e-05, |
|
"loss": 0.1656, |
|
"step": 120 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 120, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.1955401487160934e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|