|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.13333333333333333, |
|
"global_step": 800, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 4.0598, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 3.9538, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.4e-05, |
|
"loss": 4.0434, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 4.0516, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4e-05, |
|
"loss": 3.998, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.8e-05, |
|
"loss": 4.0225, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 3.904, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 3.792, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 7.2e-05, |
|
"loss": 3.5984, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8e-05, |
|
"loss": 3.4579, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 3.6368, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.6e-05, |
|
"loss": 3.4327, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 3.3701, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 3.2466, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00012, |
|
"loss": 3.2952, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 3.1374, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 3.139, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.000144, |
|
"loss": 3.0374, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.000152, |
|
"loss": 3.0814, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00016, |
|
"loss": 3.0121, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.000168, |
|
"loss": 3.0238, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 2.84, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 2.8892, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.000192, |
|
"loss": 2.8234, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.0002, |
|
"loss": 2.8687, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00020800000000000001, |
|
"loss": 2.7765, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00021600000000000002, |
|
"loss": 2.7827, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00022400000000000002, |
|
"loss": 2.8096, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.000232, |
|
"loss": 2.7986, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00024, |
|
"loss": 2.7335, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000248, |
|
"loss": 2.7041, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00025600000000000004, |
|
"loss": 2.6543, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000264, |
|
"loss": 2.6889, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00027200000000000005, |
|
"loss": 2.7073, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00028, |
|
"loss": 2.7532, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000288, |
|
"loss": 2.5479, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000296, |
|
"loss": 2.5846, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000304, |
|
"loss": 2.5425, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00031200000000000005, |
|
"loss": 2.6203, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00032, |
|
"loss": 2.578, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000328, |
|
"loss": 2.6235, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000336, |
|
"loss": 2.5181, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000344, |
|
"loss": 2.5592, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00035200000000000005, |
|
"loss": 2.4978, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00036, |
|
"loss": 2.5523, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00036800000000000005, |
|
"loss": 2.5996, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000376, |
|
"loss": 2.5497, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000384, |
|
"loss": 2.5293, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000392, |
|
"loss": 2.417, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0004, |
|
"loss": 2.4453, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00040800000000000005, |
|
"loss": 2.4302, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00041600000000000003, |
|
"loss": 2.3976, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00042400000000000006, |
|
"loss": 2.5218, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00043200000000000004, |
|
"loss": 2.4465, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00044000000000000007, |
|
"loss": 2.4865, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00044800000000000005, |
|
"loss": 2.5155, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00045599999999999997, |
|
"loss": 2.502, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000464, |
|
"loss": 2.4115, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000472, |
|
"loss": 2.4561, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00048, |
|
"loss": 2.4581, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000488, |
|
"loss": 2.4264, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000496, |
|
"loss": 2.4441, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000504, |
|
"loss": 2.4781, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0005120000000000001, |
|
"loss": 2.3868, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0005200000000000001, |
|
"loss": 2.4542, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000528, |
|
"loss": 2.4437, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000536, |
|
"loss": 2.4007, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0005440000000000001, |
|
"loss": 2.4116, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000552, |
|
"loss": 2.4022, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00056, |
|
"loss": 2.3727, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000568, |
|
"loss": 2.38, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000576, |
|
"loss": 2.3938, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000584, |
|
"loss": 2.4948, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000592, |
|
"loss": 2.3133, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0006000000000000001, |
|
"loss": 2.4094, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000608, |
|
"loss": 2.3753, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000616, |
|
"loss": 2.4388, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0006240000000000001, |
|
"loss": 2.4028, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0006320000000000001, |
|
"loss": 2.3912, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00064, |
|
"loss": 2.3202, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000648, |
|
"loss": 2.404, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000656, |
|
"loss": 2.3694, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000664, |
|
"loss": 2.3869, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000672, |
|
"loss": 2.3976, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00068, |
|
"loss": 2.3407, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000688, |
|
"loss": 2.35, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000696, |
|
"loss": 2.3244, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0007040000000000001, |
|
"loss": 2.2987, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0007120000000000001, |
|
"loss": 2.403, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00072, |
|
"loss": 2.4228, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000728, |
|
"loss": 2.3729, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007360000000000001, |
|
"loss": 2.2986, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007440000000000001, |
|
"loss": 2.3624, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000752, |
|
"loss": 2.3393, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00076, |
|
"loss": 2.3151, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000768, |
|
"loss": 2.3513, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000776, |
|
"loss": 2.2885, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000784, |
|
"loss": 2.3285, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007920000000000001, |
|
"loss": 2.3709, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0008, |
|
"loss": 2.4014, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000799864406779661, |
|
"loss": 2.3282, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007997288135593221, |
|
"loss": 2.3529, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007995932203389831, |
|
"loss": 2.3502, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007994576271186441, |
|
"loss": 2.2684, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.000799322033898305, |
|
"loss": 2.3121, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007991864406779661, |
|
"loss": 2.2489, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007990508474576271, |
|
"loss": 2.2469, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007989152542372882, |
|
"loss": 2.3522, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007987796610169492, |
|
"loss": 2.2518, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007986440677966103, |
|
"loss": 2.3545, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007985084745762711, |
|
"loss": 2.3159, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007983728813559322, |
|
"loss": 2.3242, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007982372881355932, |
|
"loss": 2.3071, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007981016949152543, |
|
"loss": 2.3003, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007979661016949153, |
|
"loss": 2.2448, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007978305084745764, |
|
"loss": 2.3383, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007976949152542374, |
|
"loss": 2.2184, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007975593220338983, |
|
"loss": 2.3312, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007974237288135593, |
|
"loss": 2.3349, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007972881355932204, |
|
"loss": 2.3324, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007971525423728814, |
|
"loss": 2.3189, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007970169491525424, |
|
"loss": 2.2545, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007968813559322035, |
|
"loss": 2.3608, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007967457627118644, |
|
"loss": 2.2787, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007966101694915254, |
|
"loss": 2.3173, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007964745762711864, |
|
"loss": 2.2544, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007963389830508475, |
|
"loss": 2.3787, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007962033898305085, |
|
"loss": 2.2431, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007960677966101696, |
|
"loss": 2.2985, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007959322033898305, |
|
"loss": 2.255, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007957966101694916, |
|
"loss": 2.2984, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007956610169491526, |
|
"loss": 2.2976, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007955254237288136, |
|
"loss": 2.2816, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007953898305084746, |
|
"loss": 2.2662, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007952542372881357, |
|
"loss": 2.2066, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007951186440677966, |
|
"loss": 2.2509, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007949830508474577, |
|
"loss": 2.2874, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007948474576271187, |
|
"loss": 2.2304, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007947118644067797, |
|
"loss": 2.2438, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007945762711864407, |
|
"loss": 2.293, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007944406779661018, |
|
"loss": 2.2882, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007943050847457627, |
|
"loss": 2.2687, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007941694915254237, |
|
"loss": 2.2579, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007940338983050848, |
|
"loss": 2.2091, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007938983050847458, |
|
"loss": 2.3201, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007937627118644069, |
|
"loss": 2.1964, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007936271186440678, |
|
"loss": 2.2587, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007934915254237288, |
|
"loss": 2.2987, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0007933559322033898, |
|
"loss": 2.1963, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007932203389830509, |
|
"loss": 2.2045, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007930847457627119, |
|
"loss": 2.3174, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000792949152542373, |
|
"loss": 2.3348, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000792813559322034, |
|
"loss": 2.2241, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007926779661016949, |
|
"loss": 2.2472, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007925423728813559, |
|
"loss": 2.2352, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000792406779661017, |
|
"loss": 2.2351, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000792271186440678, |
|
"loss": 2.2205, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007921355932203391, |
|
"loss": 2.1175, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007920000000000001, |
|
"loss": 2.1801, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000791864406779661, |
|
"loss": 2.2301, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.000791728813559322, |
|
"loss": 2.2605, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007915932203389831, |
|
"loss": 2.2462, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007914576271186441, |
|
"loss": 2.2029, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007913220338983051, |
|
"loss": 2.2551, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007911864406779662, |
|
"loss": 2.2825, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007910508474576271, |
|
"loss": 2.2582, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007909152542372882, |
|
"loss": 2.2453, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007907796610169492, |
|
"loss": 2.2599, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007906440677966102, |
|
"loss": 2.2336, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007905084745762712, |
|
"loss": 2.1311, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007903728813559323, |
|
"loss": 2.2047, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007902372881355932, |
|
"loss": 2.2847, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007901016949152543, |
|
"loss": 2.1759, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007899661016949153, |
|
"loss": 2.1877, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007898305084745763, |
|
"loss": 2.2365, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007896949152542373, |
|
"loss": 2.1855, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007895593220338984, |
|
"loss": 2.2359, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007894237288135593, |
|
"loss": 2.2366, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007892881355932204, |
|
"loss": 2.2493, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007891525423728814, |
|
"loss": 2.2278, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007890169491525425, |
|
"loss": 2.2498, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007888813559322034, |
|
"loss": 2.2337, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007887457627118645, |
|
"loss": 2.2675, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007886101694915254, |
|
"loss": 2.2142, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007884745762711864, |
|
"loss": 2.2686, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007883389830508475, |
|
"loss": 2.1442, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007882033898305085, |
|
"loss": 2.2155, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007880677966101696, |
|
"loss": 2.2874, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007879322033898306, |
|
"loss": 2.1709, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007877966101694915, |
|
"loss": 2.2595, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007876610169491525, |
|
"loss": 2.2017, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007875254237288136, |
|
"loss": 2.1778, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007873898305084746, |
|
"loss": 2.2272, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007872542372881357, |
|
"loss": 2.2734, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007871186440677967, |
|
"loss": 2.1794, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007869830508474576, |
|
"loss": 2.2155, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007868474576271186, |
|
"loss": 2.2478, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007867118644067797, |
|
"loss": 2.1717, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007865762711864407, |
|
"loss": 2.2421, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007864406779661018, |
|
"loss": 2.2584, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007863050847457628, |
|
"loss": 2.2263, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007861694915254239, |
|
"loss": 2.1951, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007860338983050848, |
|
"loss": 2.1717, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007858983050847457, |
|
"loss": 2.1705, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007857627118644068, |
|
"loss": 2.199, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007856271186440678, |
|
"loss": 2.3078, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007854915254237289, |
|
"loss": 2.1449, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007853559322033898, |
|
"loss": 2.1788, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0007852203389830509, |
|
"loss": 2.1998, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007850847457627119, |
|
"loss": 2.2799, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007849491525423729, |
|
"loss": 2.2501, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007848135593220339, |
|
"loss": 2.2081, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000784677966101695, |
|
"loss": 2.1697, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007845423728813559, |
|
"loss": 2.1713, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000784406779661017, |
|
"loss": 2.2226, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000784271186440678, |
|
"loss": 2.2084, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007841355932203391, |
|
"loss": 2.1771, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000784, |
|
"loss": 2.2013, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007838644067796611, |
|
"loss": 2.248, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.000783728813559322, |
|
"loss": 2.2238, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007835932203389831, |
|
"loss": 2.1268, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007834576271186441, |
|
"loss": 2.161, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007833220338983052, |
|
"loss": 2.1049, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007831864406779662, |
|
"loss": 2.1929, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007830508474576272, |
|
"loss": 2.2978, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007829152542372881, |
|
"loss": 2.175, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007827796610169491, |
|
"loss": 2.2298, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007826440677966102, |
|
"loss": 2.1707, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007825084745762712, |
|
"loss": 2.1759, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007823728813559323, |
|
"loss": 2.182, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007822372881355933, |
|
"loss": 2.219, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007821016949152542, |
|
"loss": 2.2042, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007819661016949152, |
|
"loss": 2.1706, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007818305084745763, |
|
"loss": 2.171, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007816949152542373, |
|
"loss": 2.0928, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007815593220338984, |
|
"loss": 2.2218, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007814237288135594, |
|
"loss": 2.1285, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007812881355932205, |
|
"loss": 2.2702, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007811525423728813, |
|
"loss": 2.2371, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007810169491525424, |
|
"loss": 2.1952, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007808813559322034, |
|
"loss": 2.228, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007807457627118645, |
|
"loss": 2.0863, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007806101694915255, |
|
"loss": 2.1102, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007804745762711866, |
|
"loss": 2.0992, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007803389830508475, |
|
"loss": 2.231, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007802033898305085, |
|
"loss": 2.1459, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007800677966101695, |
|
"loss": 2.2173, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007799322033898305, |
|
"loss": 2.1938, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007797966101694916, |
|
"loss": 2.1128, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007796610169491525, |
|
"loss": 2.1401, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007795254237288136, |
|
"loss": 2.1503, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007793898305084746, |
|
"loss": 2.0771, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007792542372881356, |
|
"loss": 2.0993, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007791186440677966, |
|
"loss": 2.24, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007789830508474577, |
|
"loss": 2.154, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007788474576271186, |
|
"loss": 2.1774, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007787118644067797, |
|
"loss": 2.0888, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007785762711864407, |
|
"loss": 2.216, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007784406779661018, |
|
"loss": 2.1042, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007783050847457628, |
|
"loss": 2.2299, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007781694915254238, |
|
"loss": 2.201, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007780338983050847, |
|
"loss": 2.1315, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007778983050847458, |
|
"loss": 2.1977, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007777627118644068, |
|
"loss": 2.2053, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007776271186440679, |
|
"loss": 2.1809, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007774915254237289, |
|
"loss": 2.2412, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007773559322033899, |
|
"loss": 2.2178, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007772203389830508, |
|
"loss": 2.2615, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007770847457627118, |
|
"loss": 2.128, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0007769491525423729, |
|
"loss": 2.2243, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007768135593220339, |
|
"loss": 2.1416, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000776677966101695, |
|
"loss": 2.1921, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000776542372881356, |
|
"loss": 2.1748, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007764067796610171, |
|
"loss": 2.1006, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007762711864406779, |
|
"loss": 2.1855, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000776135593220339, |
|
"loss": 2.0681, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000776, |
|
"loss": 2.2588, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007758644067796611, |
|
"loss": 2.1486, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007757288135593221, |
|
"loss": 2.206, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007755932203389832, |
|
"loss": 2.0873, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007754576271186441, |
|
"loss": 2.149, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007753220338983051, |
|
"loss": 2.1717, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007751864406779661, |
|
"loss": 2.1944, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007750508474576272, |
|
"loss": 2.1323, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007749152542372882, |
|
"loss": 2.2031, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007747796610169493, |
|
"loss": 2.1541, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007746440677966102, |
|
"loss": 2.1012, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007745084745762712, |
|
"loss": 2.1111, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007743728813559322, |
|
"loss": 2.1989, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007742372881355932, |
|
"loss": 2.1702, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007741016949152543, |
|
"loss": 2.095, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007739661016949152, |
|
"loss": 2.2087, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007738305084745763, |
|
"loss": 2.1621, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007736949152542373, |
|
"loss": 2.1102, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007735593220338984, |
|
"loss": 2.1092, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007734237288135594, |
|
"loss": 2.0995, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007732881355932204, |
|
"loss": 2.0985, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007731525423728813, |
|
"loss": 2.1749, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007730169491525424, |
|
"loss": 2.1833, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007728813559322034, |
|
"loss": 2.1927, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007727457627118645, |
|
"loss": 2.0672, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007726101694915255, |
|
"loss": 2.2456, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007724745762711865, |
|
"loss": 2.1152, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007723389830508474, |
|
"loss": 2.1347, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007722033898305085, |
|
"loss": 2.113, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007720677966101695, |
|
"loss": 2.1097, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007719322033898306, |
|
"loss": 2.1361, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007717966101694916, |
|
"loss": 2.1015, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007716610169491526, |
|
"loss": 2.1516, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007715254237288135, |
|
"loss": 2.1099, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007713898305084745, |
|
"loss": 2.1754, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007712542372881356, |
|
"loss": 2.1101, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007711186440677966, |
|
"loss": 2.1652, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007709830508474577, |
|
"loss": 2.0852, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007708474576271187, |
|
"loss": 2.0926, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007707118644067798, |
|
"loss": 2.1592, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007705762711864407, |
|
"loss": 2.2111, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007704406779661017, |
|
"loss": 2.08, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007703050847457627, |
|
"loss": 2.1532, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007701694915254238, |
|
"loss": 2.1141, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007700338983050848, |
|
"loss": 2.0578, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007698983050847459, |
|
"loss": 2.1405, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007697627118644068, |
|
"loss": 2.0921, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007696271186440678, |
|
"loss": 2.1161, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007694915254237288, |
|
"loss": 2.118, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007693559322033899, |
|
"loss": 2.054, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007692203389830509, |
|
"loss": 2.0837, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.000769084745762712, |
|
"loss": 2.1788, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0007689491525423729, |
|
"loss": 2.1688, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007688135593220339, |
|
"loss": 2.1397, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000768677966101695, |
|
"loss": 2.0973, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007685423728813559, |
|
"loss": 2.1558, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000768406779661017, |
|
"loss": 2.0784, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000768271186440678, |
|
"loss": 2.0901, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000768135593220339, |
|
"loss": 2.083, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000768, |
|
"loss": 2.0883, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007678644067796611, |
|
"loss": 2.1707, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007677288135593221, |
|
"loss": 2.1467, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007675932203389831, |
|
"loss": 2.0778, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000767457627118644, |
|
"loss": 2.158, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007673220338983051, |
|
"loss": 2.1849, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007671864406779661, |
|
"loss": 2.132, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007670508474576272, |
|
"loss": 2.1017, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007669152542372882, |
|
"loss": 2.1146, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007667796610169493, |
|
"loss": 2.0731, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007666440677966102, |
|
"loss": 2.1342, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007665084745762712, |
|
"loss": 2.1694, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007663728813559322, |
|
"loss": 2.0779, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007662372881355933, |
|
"loss": 2.151, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007661016949152543, |
|
"loss": 2.121, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007659661016949153, |
|
"loss": 2.0767, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007658305084745764, |
|
"loss": 2.1442, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007656949152542373, |
|
"loss": 2.1636, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007655593220338983, |
|
"loss": 2.168, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007654237288135593, |
|
"loss": 2.1254, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007652881355932204, |
|
"loss": 2.1253, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007651525423728814, |
|
"loss": 2.0905, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007650169491525425, |
|
"loss": 2.0981, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007648813559322034, |
|
"loss": 2.1296, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007647457627118644, |
|
"loss": 2.1614, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007646101694915254, |
|
"loss": 2.0396, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007644745762711865, |
|
"loss": 2.1202, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007643389830508475, |
|
"loss": 2.083, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007642033898305086, |
|
"loss": 2.1423, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007640677966101695, |
|
"loss": 2.2048, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007639322033898306, |
|
"loss": 2.1045, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007637966101694915, |
|
"loss": 2.0007, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007636610169491526, |
|
"loss": 2.1807, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007635254237288136, |
|
"loss": 2.1266, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007633898305084747, |
|
"loss": 2.1566, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007632542372881356, |
|
"loss": 1.9788, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007631186440677966, |
|
"loss": 2.1346, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007629830508474577, |
|
"loss": 2.1011, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007628474576271187, |
|
"loss": 2.1045, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007627118644067797, |
|
"loss": 2.0789, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007625762711864407, |
|
"loss": 2.1627, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007624406779661017, |
|
"loss": 2.1348, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007623050847457627, |
|
"loss": 2.0705, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007621694915254238, |
|
"loss": 2.1322, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007620338983050848, |
|
"loss": 2.0468, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007618983050847458, |
|
"loss": 2.0707, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007617627118644068, |
|
"loss": 2.1316, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007616271186440678, |
|
"loss": 2.115, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007614915254237288, |
|
"loss": 2.1173, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007613559322033899, |
|
"loss": 2.1069, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007612203389830509, |
|
"loss": 2.1525, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000761084745762712, |
|
"loss": 2.1556, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.000760949152542373, |
|
"loss": 2.1647, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0007608135593220339, |
|
"loss": 2.0778, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007606779661016949, |
|
"loss": 2.1567, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007605423728813559, |
|
"loss": 2.0568, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000760406779661017, |
|
"loss": 2.0702, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000760271186440678, |
|
"loss": 2.1035, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007601355932203391, |
|
"loss": 2.1456, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00076, |
|
"loss": 2.0825, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000759864406779661, |
|
"loss": 2.1152, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000759728813559322, |
|
"loss": 2.0893, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007595932203389831, |
|
"loss": 2.1312, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007594576271186441, |
|
"loss": 2.1767, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007593220338983052, |
|
"loss": 2.1467, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007591864406779661, |
|
"loss": 2.0759, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007590508474576272, |
|
"loss": 2.0779, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007589152542372881, |
|
"loss": 2.048, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007587796610169492, |
|
"loss": 2.1718, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007586440677966102, |
|
"loss": 2.1782, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007585084745762713, |
|
"loss": 2.0557, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007583728813559322, |
|
"loss": 2.091, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007582372881355933, |
|
"loss": 2.1526, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007581016949152543, |
|
"loss": 2.1478, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007579661016949153, |
|
"loss": 2.0608, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007578305084745763, |
|
"loss": 2.1078, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007576949152542373, |
|
"loss": 2.1001, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007575593220338983, |
|
"loss": 2.1838, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007574237288135593, |
|
"loss": 2.0487, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007572881355932204, |
|
"loss": 2.0995, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007571525423728814, |
|
"loss": 2.0434, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007570169491525424, |
|
"loss": 2.0978, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007568813559322034, |
|
"loss": 2.1123, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007567457627118644, |
|
"loss": 2.1078, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007566101694915254, |
|
"loss": 2.0992, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007564745762711865, |
|
"loss": 2.1488, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007563389830508475, |
|
"loss": 2.0831, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007562033898305086, |
|
"loss": 2.061, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007560677966101696, |
|
"loss": 2.0328, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007559322033898305, |
|
"loss": 2.0889, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007557966101694915, |
|
"loss": 2.023, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007556610169491526, |
|
"loss": 2.0881, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007555254237288136, |
|
"loss": 2.0711, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007553898305084747, |
|
"loss": 2.0338, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007552542372881357, |
|
"loss": 2.0204, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007551186440677966, |
|
"loss": 2.1019, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007549830508474576, |
|
"loss": 2.1264, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007548474576271186, |
|
"loss": 2.069, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007547118644067797, |
|
"loss": 2.0708, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007545762711864407, |
|
"loss": 2.0963, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007544406779661018, |
|
"loss": 2.0615, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007543050847457627, |
|
"loss": 2.0938, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007541694915254237, |
|
"loss": 2.1352, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007540338983050847, |
|
"loss": 2.0963, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007538983050847458, |
|
"loss": 2.0754, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007537627118644068, |
|
"loss": 2.1233, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007536271186440679, |
|
"loss": 2.1221, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007534915254237288, |
|
"loss": 1.9535, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007533559322033899, |
|
"loss": 2.1664, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007532203389830509, |
|
"loss": 2.0706, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007530847457627119, |
|
"loss": 2.0818, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007529491525423729, |
|
"loss": 2.1169, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000752813559322034, |
|
"loss": 2.0383, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0007526779661016949, |
|
"loss": 2.0979, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.000752542372881356, |
|
"loss": 2.1208, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000752406779661017, |
|
"loss": 2.0399, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000752271186440678, |
|
"loss": 1.978, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000752135593220339, |
|
"loss": 2.0162, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000752, |
|
"loss": 2.162, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000751864406779661, |
|
"loss": 2.0827, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000751728813559322, |
|
"loss": 2.1191, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007515932203389831, |
|
"loss": 2.0478, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007514576271186441, |
|
"loss": 2.083, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007513220338983052, |
|
"loss": 2.0972, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007511864406779661, |
|
"loss": 2.0498, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007510508474576271, |
|
"loss": 2.0427, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007509152542372881, |
|
"loss": 2.0778, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007507796610169492, |
|
"loss": 2.035, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007506440677966102, |
|
"loss": 2.0465, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007505084745762713, |
|
"loss": 2.0468, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007503728813559323, |
|
"loss": 2.027, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007502372881355932, |
|
"loss": 2.0677, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007501016949152542, |
|
"loss": 2.1495, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007499661016949153, |
|
"loss": 2.0571, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007498305084745763, |
|
"loss": 2.1215, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007496949152542374, |
|
"loss": 2.0422, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007495593220338984, |
|
"loss": 2.1475, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007494237288135595, |
|
"loss": 2.0874, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007492881355932203, |
|
"loss": 2.0303, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007491525423728813, |
|
"loss": 2.0438, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007490169491525424, |
|
"loss": 2.0452, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007488813559322034, |
|
"loss": 2.0535, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007487457627118645, |
|
"loss": 2.1007, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007486101694915254, |
|
"loss": 1.9706, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007484745762711865, |
|
"loss": 2.1412, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007483389830508475, |
|
"loss": 2.0067, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007482033898305085, |
|
"loss": 2.0692, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007480677966101695, |
|
"loss": 2.0327, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007479322033898306, |
|
"loss": 2.0774, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007477966101694915, |
|
"loss": 2.0983, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007476610169491526, |
|
"loss": 2.14, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007475254237288136, |
|
"loss": 2.1073, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007473898305084746, |
|
"loss": 2.0778, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007472542372881356, |
|
"loss": 1.9671, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007471186440677967, |
|
"loss": 2.0635, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007469830508474576, |
|
"loss": 2.0576, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007468474576271187, |
|
"loss": 2.101, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007467118644067797, |
|
"loss": 2.1317, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007465762711864408, |
|
"loss": 2.0516, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007464406779661017, |
|
"loss": 2.1134, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007463050847457627, |
|
"loss": 2.0108, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007461694915254237, |
|
"loss": 2.0717, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007460338983050847, |
|
"loss": 2.062, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007458983050847458, |
|
"loss": 2.136, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007457627118644068, |
|
"loss": 1.9764, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007456271186440679, |
|
"loss": 1.9965, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007454915254237289, |
|
"loss": 1.9598, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007453559322033898, |
|
"loss": 2.0202, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007452203389830508, |
|
"loss": 2.0236, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007450847457627119, |
|
"loss": 2.0403, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0007449491525423729, |
|
"loss": 2.0985, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000744813559322034, |
|
"loss": 2.0491, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000744677966101695, |
|
"loss": 2.0475, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.000744542372881356, |
|
"loss": 2.0317, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007444067796610169, |
|
"loss": 2.0934, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000744271186440678, |
|
"loss": 2.0595, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000744135593220339, |
|
"loss": 2.0955, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007440000000000001, |
|
"loss": 2.0815, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007438644067796611, |
|
"loss": 1.9808, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007437288135593222, |
|
"loss": 2.1262, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007435932203389831, |
|
"loss": 2.0687, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000743457627118644, |
|
"loss": 1.9892, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007433220338983051, |
|
"loss": 2.0759, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007431864406779661, |
|
"loss": 2.1157, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007430508474576272, |
|
"loss": 2.0402, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007429152542372881, |
|
"loss": 2.1073, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007427796610169492, |
|
"loss": 2.0317, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007426440677966102, |
|
"loss": 1.9934, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007425084745762712, |
|
"loss": 1.9867, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007423728813559322, |
|
"loss": 2.0976, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007422372881355933, |
|
"loss": 1.9917, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007421016949152542, |
|
"loss": 2.0608, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007419661016949153, |
|
"loss": 1.9874, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007418305084745763, |
|
"loss": 2.0441, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007416949152542374, |
|
"loss": 2.0061, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007415593220338983, |
|
"loss": 2.0045, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007414237288135594, |
|
"loss": 2.0507, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007412881355932203, |
|
"loss": 2.0163, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007411525423728814, |
|
"loss": 2.1097, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007410169491525424, |
|
"loss": 2.0441, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007408813559322035, |
|
"loss": 2.0815, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007407457627118645, |
|
"loss": 2.122, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007406101694915255, |
|
"loss": 2.1065, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007404745762711864, |
|
"loss": 1.9847, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007403389830508474, |
|
"loss": 2.0269, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007402033898305085, |
|
"loss": 2.0521, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007400677966101695, |
|
"loss": 1.9796, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007399322033898306, |
|
"loss": 2.0629, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007397966101694916, |
|
"loss": 2.0144, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007396610169491525, |
|
"loss": 1.9843, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007395254237288135, |
|
"loss": 2.0864, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007393898305084746, |
|
"loss": 2.0842, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007392542372881356, |
|
"loss": 1.9326, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007391186440677967, |
|
"loss": 2.0396, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007389830508474577, |
|
"loss": 2.0398, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007388474576271188, |
|
"loss": 2.0183, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007387118644067797, |
|
"loss": 2.0774, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007385762711864407, |
|
"loss": 2.0977, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007384406779661017, |
|
"loss": 2.0751, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007383050847457628, |
|
"loss": 2.0928, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007381694915254238, |
|
"loss": 2.0599, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007380338983050849, |
|
"loss": 2.1053, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007378983050847458, |
|
"loss": 2.0203, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007377627118644068, |
|
"loss": 1.9643, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007376271186440678, |
|
"loss": 2.1094, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007374915254237288, |
|
"loss": 2.0148, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007373559322033899, |
|
"loss": 2.0172, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007372203389830509, |
|
"loss": 1.9883, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007370847457627119, |
|
"loss": 2.0124, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007369491525423729, |
|
"loss": 2.0097, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007368135593220339, |
|
"loss": 2.023, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0007366779661016949, |
|
"loss": 2.079, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000736542372881356, |
|
"loss": 2.1124, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.000736406779661017, |
|
"loss": 2.0094, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000736271186440678, |
|
"loss": 2.1219, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000736135593220339, |
|
"loss": 1.9903, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007360000000000001, |
|
"loss": 1.9831, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007358644067796611, |
|
"loss": 2.0807, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007357288135593221, |
|
"loss": 2.062, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000735593220338983, |
|
"loss": 2.1067, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007354576271186441, |
|
"loss": 2.0922, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007353220338983051, |
|
"loss": 2.0415, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007351864406779661, |
|
"loss": 1.9444, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007350508474576272, |
|
"loss": 2.048, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007349152542372882, |
|
"loss": 1.9961, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007347796610169492, |
|
"loss": 2.083, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007346440677966101, |
|
"loss": 2.0359, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007345084745762712, |
|
"loss": 2.0901, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007343728813559322, |
|
"loss": 2.0923, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007342372881355933, |
|
"loss": 1.9558, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007341016949152543, |
|
"loss": 2.0329, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007339661016949154, |
|
"loss": 2.0818, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007338305084745762, |
|
"loss": 2.0637, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007336949152542373, |
|
"loss": 2.0038, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007335593220338983, |
|
"loss": 2.0047, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007334237288135594, |
|
"loss": 2.108, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007332881355932204, |
|
"loss": 1.9784, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007331525423728815, |
|
"loss": 2.032, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007330169491525424, |
|
"loss": 2.032, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007328813559322034, |
|
"loss": 2.0375, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007327457627118644, |
|
"loss": 1.9258, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007326101694915255, |
|
"loss": 2.0336, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007324745762711865, |
|
"loss": 1.9972, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007323389830508475, |
|
"loss": 2.0291, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007322033898305085, |
|
"loss": 2.0544, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007320677966101695, |
|
"loss": 2.1134, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007319322033898305, |
|
"loss": 1.9614, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007317966101694915, |
|
"loss": 2.0024, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007316610169491526, |
|
"loss": 2.0332, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007315254237288136, |
|
"loss": 2.1373, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007313898305084746, |
|
"loss": 2.077, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007312542372881356, |
|
"loss": 2.017, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007311186440677967, |
|
"loss": 2.0286, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007309830508474577, |
|
"loss": 1.9955, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007308474576271187, |
|
"loss": 2.0605, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007307118644067797, |
|
"loss": 2.0716, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007305762711864407, |
|
"loss": 1.9667, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007304406779661017, |
|
"loss": 2.037, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007303050847457628, |
|
"loss": 1.9846, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007301694915254238, |
|
"loss": 2.0542, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007300338983050848, |
|
"loss": 1.9881, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007298983050847458, |
|
"loss": 2.0391, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007297627118644068, |
|
"loss": 2.0122, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007296271186440678, |
|
"loss": 2.0181, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007294915254237288, |
|
"loss": 2.007, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007293559322033899, |
|
"loss": 1.9961, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007292203389830509, |
|
"loss": 2.07, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007290847457627119, |
|
"loss": 2.0669, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007289491525423728, |
|
"loss": 2.0185, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007288135593220339, |
|
"loss": 2.0647, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007286779661016949, |
|
"loss": 2.0569, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000728542372881356, |
|
"loss": 2.0817, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000728406779661017, |
|
"loss": 2.0427, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0007282711864406781, |
|
"loss": 2.0117, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.000728135593220339, |
|
"loss": 1.9909, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000728, |
|
"loss": 2.0554, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000727864406779661, |
|
"loss": 2.0608, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007277288135593221, |
|
"loss": 2.1312, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007275932203389831, |
|
"loss": 2.0998, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007274576271186442, |
|
"loss": 2.0376, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007273220338983051, |
|
"loss": 2.0206, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007271864406779661, |
|
"loss": 1.944, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007270508474576271, |
|
"loss": 2.0059, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007269152542372882, |
|
"loss": 1.9915, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007267796610169492, |
|
"loss": 2.0549, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007266440677966102, |
|
"loss": 1.9902, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007265084745762712, |
|
"loss": 2.0937, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007263728813559322, |
|
"loss": 2.0679, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007262372881355933, |
|
"loss": 2.0352, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007261016949152542, |
|
"loss": 2.0968, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007259661016949153, |
|
"loss": 2.0103, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007258305084745763, |
|
"loss": 2.006, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007256949152542373, |
|
"loss": 2.0409, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007255593220338983, |
|
"loss": 2.0289, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007254237288135594, |
|
"loss": 2.0262, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007252881355932204, |
|
"loss": 1.9559, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007251525423728814, |
|
"loss": 2.0752, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007250169491525424, |
|
"loss": 1.9741, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007248813559322034, |
|
"loss": 2.011, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007247457627118644, |
|
"loss": 1.9709, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007246101694915255, |
|
"loss": 2.0412, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007244745762711865, |
|
"loss": 2.0846, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007243389830508476, |
|
"loss": 2.0141, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007242033898305085, |
|
"loss": 2.0181, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007240677966101695, |
|
"loss": 2.0751, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007239322033898305, |
|
"loss": 2.03, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007237966101694915, |
|
"loss": 1.9691, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007236610169491526, |
|
"loss": 2.0183, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007235254237288136, |
|
"loss": 1.9751, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007233898305084747, |
|
"loss": 2.057, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007232542372881356, |
|
"loss": 1.9894, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007231186440677966, |
|
"loss": 2.0016, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007229830508474576, |
|
"loss": 2.0016, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007228474576271187, |
|
"loss": 2.068, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007227118644067797, |
|
"loss": 2.0453, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007225762711864408, |
|
"loss": 1.9701, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007224406779661017, |
|
"loss": 2.014, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007223050847457627, |
|
"loss": 1.9256, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007221694915254237, |
|
"loss": 1.9928, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007220338983050848, |
|
"loss": 2.0385, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007218983050847458, |
|
"loss": 2.0701, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007217627118644069, |
|
"loss": 2.028, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007216271186440678, |
|
"loss": 1.9851, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007214915254237289, |
|
"loss": 2.0014, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007213559322033899, |
|
"loss": 2.0437, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007212203389830509, |
|
"loss": 1.9366, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007210847457627119, |
|
"loss": 1.9544, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007209491525423729, |
|
"loss": 2.0587, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007208135593220339, |
|
"loss": 2.0092, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0007206779661016949, |
|
"loss": 2.0476, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000720542372881356, |
|
"loss": 2.0304, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000720406779661017, |
|
"loss": 1.999, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000720271186440678, |
|
"loss": 1.9694, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.000720135593220339, |
|
"loss": 2.0553, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00072, |
|
"loss": 1.993, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000719864406779661, |
|
"loss": 2.0117, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007197288135593221, |
|
"loss": 1.9563, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007195932203389831, |
|
"loss": 2.0712, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007194576271186441, |
|
"loss": 2.0268, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007193220338983051, |
|
"loss": 1.9227, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007191864406779661, |
|
"loss": 1.9944, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007190508474576271, |
|
"loss": 1.954, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007189152542372882, |
|
"loss": 2.0044, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007187796610169492, |
|
"loss": 1.9591, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007186440677966103, |
|
"loss": 1.9501, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007185084745762713, |
|
"loss": 2.0118, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007183728813559322, |
|
"loss": 2.0294, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007182372881355932, |
|
"loss": 2.0701, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007181016949152542, |
|
"loss": 1.9157, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007179661016949153, |
|
"loss": 2.0225, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007178305084745763, |
|
"loss": 1.98, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007176949152542374, |
|
"loss": 1.9733, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007175593220338983, |
|
"loss": 2.0685, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007174237288135593, |
|
"loss": 1.9963, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007172881355932203, |
|
"loss": 2.034, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007171525423728814, |
|
"loss": 1.977, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007170169491525424, |
|
"loss": 2.0206, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007168813559322035, |
|
"loss": 1.9977, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007167457627118644, |
|
"loss": 2.0333, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007166101694915255, |
|
"loss": 2.0027, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007164745762711864, |
|
"loss": 2.0202, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007163389830508475, |
|
"loss": 2.0055, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007162033898305085, |
|
"loss": 2.024, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007160677966101696, |
|
"loss": 1.9303, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007159322033898305, |
|
"loss": 2.0139, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007157966101694916, |
|
"loss": 1.9983, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007156610169491526, |
|
"loss": 1.9705, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007155254237288136, |
|
"loss": 2.0189, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007153898305084746, |
|
"loss": 2.0578, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007152542372881356, |
|
"loss": 2.0432, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007151186440677966, |
|
"loss": 1.9693, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007149830508474576, |
|
"loss": 2.0364, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007148474576271187, |
|
"loss": 2.0152, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007147118644067797, |
|
"loss": 2.0469, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007145762711864407, |
|
"loss": 2.0795, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007144406779661017, |
|
"loss": 2.0025, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007143050847457627, |
|
"loss": 1.9858, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007141694915254237, |
|
"loss": 2.0756, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007140338983050848, |
|
"loss": 1.9671, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007138983050847458, |
|
"loss": 2.085, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007137627118644069, |
|
"loss": 1.9999, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007136271186440679, |
|
"loss": 1.9725, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007134915254237288, |
|
"loss": 2.0105, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007133559322033898, |
|
"loss": 2.0743, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007132203389830509, |
|
"loss": 1.9675, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007130847457627119, |
|
"loss": 1.9963, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000712949152542373, |
|
"loss": 1.9855, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000712813559322034, |
|
"loss": 2.0843, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000712677966101695, |
|
"loss": 1.9715, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007125423728813559, |
|
"loss": 2.0824, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007124067796610169, |
|
"loss": 1.9904, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000712271186440678, |
|
"loss": 1.9993, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000712135593220339, |
|
"loss": 1.9947, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0007120000000000001, |
|
"loss": 1.975, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.000711864406779661, |
|
"loss": 1.9464, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000711728813559322, |
|
"loss": 1.9511, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000711593220338983, |
|
"loss": 2.0268, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007114576271186441, |
|
"loss": 2.1, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007113220338983051, |
|
"loss": 2.0567, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007111864406779662, |
|
"loss": 2.0099, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007110508474576271, |
|
"loss": 2.0488, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007109152542372882, |
|
"loss": 1.9789, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007107796610169492, |
|
"loss": 2.015, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007106440677966102, |
|
"loss": 1.9582, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007105084745762712, |
|
"loss": 1.9835, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007103728813559323, |
|
"loss": 1.9657, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007102372881355932, |
|
"loss": 1.8867, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007101016949152543, |
|
"loss": 1.9759, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007099661016949153, |
|
"loss": 2.028, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007098305084745763, |
|
"loss": 1.9511, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007096949152542373, |
|
"loss": 2.0709, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007095593220338983, |
|
"loss": 1.9723, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007094237288135593, |
|
"loss": 1.985, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007092881355932203, |
|
"loss": 1.9668, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007091525423728814, |
|
"loss": 1.9951, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007090169491525424, |
|
"loss": 2.0183, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007088813559322035, |
|
"loss": 2.0393, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007087457627118644, |
|
"loss": 1.8901, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007086101694915254, |
|
"loss": 1.9539, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007084745762711864, |
|
"loss": 2.0013, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007083389830508475, |
|
"loss": 1.9558, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007082033898305085, |
|
"loss": 1.9213, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007080677966101696, |
|
"loss": 2.0308, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007079322033898306, |
|
"loss": 2.0112, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007077966101694915, |
|
"loss": 1.9911, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007076610169491525, |
|
"loss": 2.0662, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007075254237288136, |
|
"loss": 2.0265, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007073898305084746, |
|
"loss": 1.9888, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007072542372881357, |
|
"loss": 2.0748, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007071186440677967, |
|
"loss": 1.9128, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007069830508474577, |
|
"loss": 2.0175, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007068474576271186, |
|
"loss": 1.9524, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007067118644067796, |
|
"loss": 2.0121, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007065762711864407, |
|
"loss": 1.9964, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007064406779661017, |
|
"loss": 1.8582, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007063050847457628, |
|
"loss": 2.0165, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007061694915254238, |
|
"loss": 1.9998, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007060338983050848, |
|
"loss": 2.0749, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007058983050847458, |
|
"loss": 2.0067, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007057627118644068, |
|
"loss": 1.9916, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007056271186440678, |
|
"loss": 1.9975, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007054915254237289, |
|
"loss": 2.0394, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007053559322033899, |
|
"loss": 1.9753, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007052203389830509, |
|
"loss": 1.9977, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0007050847457627119, |
|
"loss": 2.021, |
|
"step": 800 |
|
} |
|
], |
|
"max_steps": 6000, |
|
"num_train_epochs": 9223372036854775807, |
|
"total_flos": 4.12828419529728e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|