|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.994671403197158, |
|
"global_step": 843, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.4117647058823532e-05, |
|
"loss": 5.6404, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.8235294117647063e-05, |
|
"loss": 4.5746, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.235294117647059e-05, |
|
"loss": 3.3627, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.6470588235294126e-05, |
|
"loss": 2.7997, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.058823529411765e-05, |
|
"loss": 2.4203, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 7.999971068621779e-05, |
|
"loss": 2.217, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 7.999537106318664e-05, |
|
"loss": 2.2324, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 7.998582444493811e-05, |
|
"loss": 2.089, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 7.997107207434592e-05, |
|
"loss": 2.0586, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 7.995111587202043e-05, |
|
"loss": 1.9011, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 7.992595843605852e-05, |
|
"loss": 1.8797, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 7.989560304170546e-05, |
|
"loss": 1.8855, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 7.98600536409284e-05, |
|
"loss": 1.7122, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 7.981931486190192e-05, |
|
"loss": 1.7285, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 7.97733920084055e-05, |
|
"loss": 1.6935, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.972229105913295e-05, |
|
"loss": 1.5963, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 7.966601866691413e-05, |
|
"loss": 1.6079, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 7.96045821578488e-05, |
|
"loss": 1.6835, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 7.953798953035274e-05, |
|
"loss": 1.5928, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.946624945411658e-05, |
|
"loss": 1.6176, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.938937126897702e-05, |
|
"loss": 1.5979, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.930736498370085e-05, |
|
"loss": 1.5759, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.922024127468196e-05, |
|
"loss": 1.633, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.912801148455134e-05, |
|
"loss": 1.5562, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.903068762070047e-05, |
|
"loss": 1.5349, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.892828235371795e-05, |
|
"loss": 1.5454, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.882080901574001e-05, |
|
"loss": 1.5811, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.870828159871478e-05, |
|
"loss": 1.5442, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.859071475258065e-05, |
|
"loss": 1.5513, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.8468123783359e-05, |
|
"loss": 1.5349, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.83405246511616e-05, |
|
"loss": 1.5035, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.820793396811258e-05, |
|
"loss": 1.5159, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.80703689961859e-05, |
|
"loss": 1.4197, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.792784764495791e-05, |
|
"loss": 1.4168, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.778038846927566e-05, |
|
"loss": 1.4997, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.762801066684133e-05, |
|
"loss": 1.4985, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 7.747073407571291e-05, |
|
"loss": 1.2951, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 7.730857917172132e-05, |
|
"loss": 1.3738, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 7.714156706580487e-05, |
|
"loss": 1.385, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 7.696971950126068e-05, |
|
"loss": 1.4614, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 7.679305885091398e-05, |
|
"loss": 1.3669, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 7.661160811420536e-05, |
|
"loss": 1.3745, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 7.642539091419654e-05, |
|
"loss": 1.3872, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 7.623443149449479e-05, |
|
"loss": 1.3118, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 7.603875471609677e-05, |
|
"loss": 1.3973, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.583838605415177e-05, |
|
"loss": 1.4051, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.563335159464515e-05, |
|
"loss": 1.4145, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.542367803100223e-05, |
|
"loss": 1.342, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.520939266061304e-05, |
|
"loss": 1.3072, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.499052338127845e-05, |
|
"loss": 1.3066, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.476709868757825e-05, |
|
"loss": 1.3681, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.453914766716133e-05, |
|
"loss": 1.3416, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.430669999695885e-05, |
|
"loss": 1.4098, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.406978593932055e-05, |
|
"loss": 1.4484, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.382843633807488e-05, |
|
"loss": 1.2463, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.358268261451353e-05, |
|
"loss": 1.3308, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.33325567633006e-05, |
|
"loss": 1.2935, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.307809134830729e-05, |
|
"loss": 1.2542, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.281931949837233e-05, |
|
"loss": 1.2856, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.255627490298902e-05, |
|
"loss": 1.3299, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.228899180791914e-05, |
|
"loss": 1.2866, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.201750501073452e-05, |
|
"loss": 1.2212, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.174184985628679e-05, |
|
"loss": 1.2903, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.146206223210571e-05, |
|
"loss": 1.2567, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.117817856372709e-05, |
|
"loss": 1.2662, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.089023580995057e-05, |
|
"loss": 1.2442, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 7.059827145802777e-05, |
|
"loss": 1.2454, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 7.03023235187821e-05, |
|
"loss": 1.3152, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 7.000243052165989e-05, |
|
"loss": 1.2658, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.969863150971439e-05, |
|
"loss": 1.1871, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.939096603452269e-05, |
|
"loss": 1.2346, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.907947415103656e-05, |
|
"loss": 1.1971, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.87641964123676e-05, |
|
"loss": 1.2636, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.844517386450775e-05, |
|
"loss": 1.2181, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.812244804098544e-05, |
|
"loss": 1.2436, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.779606095745832e-05, |
|
"loss": 1.249, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.746605510624337e-05, |
|
"loss": 1.2698, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.713247345078465e-05, |
|
"loss": 1.1873, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.679535942006004e-05, |
|
"loss": 1.1497, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 6.645475690292716e-05, |
|
"loss": 1.2872, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 6.611071024240946e-05, |
|
"loss": 1.2413, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 6.576326422992326e-05, |
|
"loss": 1.213, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 6.541246409944634e-05, |
|
"loss": 1.2931, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 6.505835552162887e-05, |
|
"loss": 1.2878, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 6.470098459784768e-05, |
|
"loss": 1.2502, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 6.434039785420417e-05, |
|
"loss": 1.2416, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 6.397664223546721e-05, |
|
"loss": 1.2468, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 6.360976509896133e-05, |
|
"loss": 1.2207, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 6.323981420840124e-05, |
|
"loss": 1.2038, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 6.286683772767365e-05, |
|
"loss": 1.3596, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.249088421456656e-05, |
|
"loss": 1.1464, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 6.211200261444774e-05, |
|
"loss": 1.2218, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 6.173024225389242e-05, |
|
"loss": 1.1165, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_gen_len": 28.29241516966068, |
|
"eval_loss": 1.1089943647384644, |
|
"eval_rouge1": 57.9239, |
|
"eval_rouge2": 31.9259, |
|
"eval_rougeL": 53.8737, |
|
"eval_rougeLsum": 54.9811, |
|
"eval_runtime": 90.9427, |
|
"eval_samples_per_second": 11.018, |
|
"eval_steps_per_second": 2.76, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 6.134565283426149e-05, |
|
"loss": 1.1982, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 6.0958284425230914e-05, |
|
"loss": 1.117, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 6.0568187458273134e-05, |
|
"loss": 1.1124, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 6.017541272009135e-05, |
|
"loss": 1.0754, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 5.9780011346007745e-05, |
|
"loss": 1.046, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 5.9382034813306014e-05, |
|
"loss": 1.1768, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 5.898153493452965e-05, |
|
"loss": 1.1407, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 5.857856385073643e-05, |
|
"loss": 1.1402, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 5.817317402471016e-05, |
|
"loss": 1.0946, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 5.776541823413062e-05, |
|
"loss": 1.0745, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 5.735534956470233e-05, |
|
"loss": 1.1032, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 5.6943021403243415e-05, |
|
"loss": 1.0543, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 5.652848743073513e-05, |
|
"loss": 1.0755, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 5.611180161533318e-05, |
|
"loss": 1.0945, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 5.569301820534154e-05, |
|
"loss": 1.158, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 5.5272191722149965e-05, |
|
"loss": 1.1357, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 5.484937695313573e-05, |
|
"loss": 1.1591, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 5.4424628944531024e-05, |
|
"loss": 1.0965, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 5.399800299425631e-05, |
|
"loss": 1.0746, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.356955464472121e-05, |
|
"loss": 1.1068, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.3139339675593446e-05, |
|
"loss": 1.0997, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 5.2707414096536793e-05, |
|
"loss": 1.1239, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 5.227383413991925e-05, |
|
"loss": 1.1499, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 5.1838656253492175e-05, |
|
"loss": 1.0583, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 5.1401937093041305e-05, |
|
"loss": 0.9659, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 5.096373351501074e-05, |
|
"loss": 1.0781, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 5.052410256910085e-05, |
|
"loss": 1.0969, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 5.008310149084097e-05, |
|
"loss": 1.1756, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 4.964078769413782e-05, |
|
"loss": 1.0718, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 4.919721876380094e-05, |
|
"loss": 1.1295, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 4.8752452448045626e-05, |
|
"loss": 1.1278, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 4.830654665097469e-05, |
|
"loss": 1.1014, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 4.785955942504001e-05, |
|
"loss": 1.0543, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.741154896348458e-05, |
|
"loss": 1.126, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 4.696257359276641e-05, |
|
"loss": 1.0462, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.651269176496496e-05, |
|
"loss": 1.1578, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.606196205017133e-05, |
|
"loss": 1.1495, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 4.561044312886294e-05, |
|
"loss": 1.0439, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.5158193784263975e-05, |
|
"loss": 1.0773, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 4.4705272894692425e-05, |
|
"loss": 1.1168, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 4.425173942589462e-05, |
|
"loss": 1.0929, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 4.379765242336859e-05, |
|
"loss": 1.1132, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.3343071004676905e-05, |
|
"loss": 0.9778, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 4.288805435175011e-05, |
|
"loss": 1.0954, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.2432661703181875e-05, |
|
"loss": 1.0817, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 4.197695234651674e-05, |
|
"loss": 1.1092, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 4.152098561053148e-05, |
|
"loss": 1.1176, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.1064820857511e-05, |
|
"loss": 1.0565, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.0608517475520084e-05, |
|
"loss": 1.1288, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.015213487067163e-05, |
|
"loss": 1.1744, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.96957324593925e-05, |
|
"loss": 1.1086, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 3.92393696606882e-05, |
|
"loss": 1.0737, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 3.878310588840703e-05, |
|
"loss": 1.1223, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.832700054350511e-05, |
|
"loss": 1.0578, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 3.787111300631287e-05, |
|
"loss": 1.0141, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.7415502628804334e-05, |
|
"loss": 1.0913, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.6960228726870186e-05, |
|
"loss": 1.1208, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.650535057259536e-05, |
|
"loss": 1.073, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.6050927386542365e-05, |
|
"loss": 1.1406, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.5597018330041476e-05, |
|
"loss": 1.0705, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.514368249748849e-05, |
|
"loss": 1.0571, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.469097890865113e-05, |
|
"loss": 1.0793, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.4238966500985405e-05, |
|
"loss": 1.1336, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.378770412196249e-05, |
|
"loss": 1.0964, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.333725052140732e-05, |
|
"loss": 1.0749, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.2887664343850045e-05, |
|
"loss": 1.0811, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 3.243900412089106e-05, |
|
"loss": 1.082, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.199132826358075e-05, |
|
"loss": 1.0476, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.154469505481503e-05, |
|
"loss": 1.1107, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.109916264174743e-05, |
|
"loss": 1.1545, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.065478902821894e-05, |
|
"loss": 1.0738, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.0211632067206533e-05, |
|
"loss": 1.0883, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.976974945329125e-05, |
|
"loss": 1.0172, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.932919871514689e-05, |
|
"loss": 1.076, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.8890037208050486e-05, |
|
"loss": 1.1079, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.845232210641517e-05, |
|
"loss": 1.0263, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.801611039634654e-05, |
|
"loss": 1.0274, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.7581458868223796e-05, |
|
"loss": 1.0736, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.714842410930612e-05, |
|
"loss": 1.0472, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.6717062496365574e-05, |
|
"loss": 1.0982, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.6287430188347456e-05, |
|
"loss": 1.0779, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.5859583119058967e-05, |
|
"loss": 1.0563, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.5433576989887115e-05, |
|
"loss": 1.0575, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.5009467262547072e-05, |
|
"loss": 1.0763, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.458730915186153e-05, |
|
"loss": 1.0557, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.4167157618572312e-05, |
|
"loss": 0.9993, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.374906736218506e-05, |
|
"loss": 1.0631, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.333309281384784e-05, |
|
"loss": 0.9966, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.2919288129264837e-05, |
|
"loss": 1.0742, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.250770718164579e-05, |
|
"loss": 1.0572, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.209840355469223e-05, |
|
"loss": 1.0628, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.1691430535621432e-05, |
|
"loss": 1.0401, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.1286841108228987e-05, |
|
"loss": 0.9666, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.0884687945990744e-05, |
|
"loss": 1.0763, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_gen_len": 29.696606786427147, |
|
"eval_loss": 1.0267277956008911, |
|
"eval_rouge1": 59.9605, |
|
"eval_rouge2": 34.0298, |
|
"eval_rougeL": 55.7523, |
|
"eval_rougeLsum": 56.8021, |
|
"eval_runtime": 96.5427, |
|
"eval_samples_per_second": 10.379, |
|
"eval_steps_per_second": 2.6, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.048502340520538e-05, |
|
"loss": 1.0732, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.008789951817809e-05, |
|
"loss": 0.9671, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.9693367986446415e-05, |
|
"loss": 1.0325, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.930148017404936e-05, |
|
"loss": 0.9656, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.8912287100840275e-05, |
|
"loss": 0.9455, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.852583943584449e-05, |
|
"loss": 0.993, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.8142187490662827e-05, |
|
"loss": 0.9598, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.7761381212921508e-05, |
|
"loss": 0.978, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.7383470179769416e-05, |
|
"loss": 1.0325, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.700850359142373e-05, |
|
"loss": 0.9173, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.663653026476452e-05, |
|
"loss": 1.0184, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.6267598626979264e-05, |
|
"loss": 0.9831, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.5901756709258133e-05, |
|
"loss": 0.9981, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.553905214054082e-05, |
|
"loss": 0.9645, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.5179532141315685e-05, |
|
"loss": 0.9968, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.4823243517472187e-05, |
|
"loss": 0.9537, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.4470232654207208e-05, |
|
"loss": 0.9923, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.4120545509986103e-05, |
|
"loss": 0.9497, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.3774227610559461e-05, |
|
"loss": 0.9502, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.3431324043036104e-05, |
|
"loss": 0.9703, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.3091879450013086e-05, |
|
"loss": 0.9592, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2755938023763821e-05, |
|
"loss": 0.9748, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2423543500484607e-05, |
|
"loss": 0.9978, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.2094739154600616e-05, |
|
"loss": 1.0247, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.1769567793132048e-05, |
|
"loss": 0.9923, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.1448071750121042e-05, |
|
"loss": 0.9638, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.113029288112022e-05, |
|
"loss": 0.9403, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.0816272557743525e-05, |
|
"loss": 1.0162, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.050605166228006e-05, |
|
"loss": 0.9988, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.0199670582371573e-05, |
|
"loss": 1.004, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 9.897169205754461e-06, |
|
"loss": 0.9909, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 9.598586915066766e-06, |
|
"loss": 0.977, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 9.303962582720887e-06, |
|
"loss": 1.001, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 9.013334565842929e-06, |
|
"loss": 0.9925, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 8.72674070127881e-06, |
|
"loss": 0.9803, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 8.44421830066843e-06, |
|
"loss": 0.9129, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 8.165804145588008e-06, |
|
"loss": 1.0281, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.891534482761463e-06, |
|
"loss": 0.9433, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.621445019341522e-06, |
|
"loss": 0.9536, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.355570918260979e-06, |
|
"loss": 1.0091, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.0939467936548045e-06, |
|
"loss": 1.0031, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 6.836606706353786e-06, |
|
"loss": 0.9501, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 6.583584159450116e-06, |
|
"loss": 1.0059, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 6.334912093935605e-06, |
|
"loss": 1.0037, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 6.090622884413164e-06, |
|
"loss": 0.9347, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 5.85074833488191e-06, |
|
"loss": 0.9206, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 5.6153196745965774e-06, |
|
"loss": 0.9739, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 5.384367554001864e-06, |
|
"loss": 0.9265, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 5.157922040741996e-06, |
|
"loss": 0.9867, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.936012615746193e-06, |
|
"loss": 0.9579, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.71866816939063e-06, |
|
"loss": 0.9567, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.505916997737143e-06, |
|
"loss": 0.9833, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.29778679884937e-06, |
|
"loss": 0.9967, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.094304669186761e-06, |
|
"loss": 0.9324, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.895497100076892e-06, |
|
"loss": 0.9994, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.701389974266536e-06, |
|
"loss": 0.9551, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.512008562552005e-06, |
|
"loss": 1.001, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.3273775204891678e-06, |
|
"loss": 1.0111, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.1475208851834815e-06, |
|
"loss": 1.0089, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.972462072160682e-06, |
|
"loss": 0.9709, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.8022238723182635e-06, |
|
"loss": 0.9676, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.6368284489583396e-06, |
|
"loss": 0.9643, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.4762973349022355e-06, |
|
"loss": 0.9753, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.320651429687093e-06, |
|
"loss": 0.8655, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.169910996844968e-06, |
|
"loss": 1.0049, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.0240956612647487e-06, |
|
"loss": 1.0138, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.8832244066371696e-06, |
|
"loss": 1.0226, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.7473155729833234e-06, |
|
"loss": 0.9316, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.616386854266989e-06, |
|
"loss": 0.9972, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.4904552960910512e-06, |
|
"loss": 0.9322, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.3695372934783157e-06, |
|
"loss": 0.9403, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.2536485887370754e-06, |
|
"loss": 0.998, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.142804269411606e-06, |
|
"loss": 1.0496, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.0370187663178854e-06, |
|
"loss": 0.964, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 9.363058516649138e-07, |
|
"loss": 0.9786, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 8.406786372616405e-07, |
|
"loss": 0.9675, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.501495728100061e-07, |
|
"loss": 0.9009, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 6.647304442840652e-07, |
|
"loss": 0.9913, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 5.844323723955736e-07, |
|
"loss": 1.0268, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 5.092658111462179e-07, |
|
"loss": 0.9936, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 4.392405464665883e-07, |
|
"loss": 0.9537, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.743656949421315e-07, |
|
"loss": 1.0263, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.14649702626304e-07, |
|
"loss": 1.0091, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 2.6010034394095e-07, |
|
"loss": 0.9997, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.1072472066414606e-07, |
|
"loss": 1.0062, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.6652926100565237e-07, |
|
"loss": 0.9276, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.2751971876999504e-07, |
|
"loss": 0.9591, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 9.37011726073811e-08, |
|
"loss": 0.949, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 6.507802535253405e-08, |
|
"loss": 1.0096, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 4.165400345145898e-08, |
|
"loss": 0.9327, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.3432156476310698e-08, |
|
"loss": 1.0365, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.0414856728373501e-08, |
|
"loss": 1.0014, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.603798929206036e-09, |
|
"loss": 0.9712, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 0.0, |
|
"loss": 0.9595, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"eval_gen_len": 29.714570858283434, |
|
"eval_loss": 1.013558268547058, |
|
"eval_rouge1": 59.9513, |
|
"eval_rouge2": 33.9118, |
|
"eval_rougeL": 55.7815, |
|
"eval_rougeLsum": 56.9064, |
|
"eval_runtime": 94.9567, |
|
"eval_samples_per_second": 10.552, |
|
"eval_steps_per_second": 2.643, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"step": 843, |
|
"total_flos": 3.289406245752013e+16, |
|
"train_loss": 1.2042206371664859, |
|
"train_runtime": 2861.2173, |
|
"train_samples_per_second": 18.888, |
|
"train_steps_per_second": 0.295 |
|
} |
|
], |
|
"max_steps": 843, |
|
"num_train_epochs": 3, |
|
"total_flos": 3.289406245752013e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|