|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"global_step": 768801, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.9967481832099594e-05, |
|
"loss": 6.7212, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.993496366419919e-05, |
|
"loss": 6.2046, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.9902445496298784e-05, |
|
"loss": 6.1099, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.986992732839838e-05, |
|
"loss": 6.0605, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.9837409160497974e-05, |
|
"loss": 6.0447, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.980489099259757e-05, |
|
"loss": 6.0184, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.9772372824697164e-05, |
|
"loss": 5.9684, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.9739854656796756e-05, |
|
"loss": 5.9596, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.970733648889635e-05, |
|
"loss": 5.9604, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.967481832099594e-05, |
|
"loss": 5.9346, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.964230015309554e-05, |
|
"loss": 5.9298, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.960978198519513e-05, |
|
"loss": 5.9189, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.957726381729472e-05, |
|
"loss": 5.8765, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.954474564939432e-05, |
|
"loss": 5.8812, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.951222748149391e-05, |
|
"loss": 5.8845, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.947970931359351e-05, |
|
"loss": 5.867, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.94471911456931e-05, |
|
"loss": 5.8744, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.94146729777927e-05, |
|
"loss": 5.8506, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.938215480989229e-05, |
|
"loss": 5.8343, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.934963664199188e-05, |
|
"loss": 5.829, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.931711847409148e-05, |
|
"loss": 5.8303, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.928460030619107e-05, |
|
"loss": 5.8147, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.9252082138290664e-05, |
|
"loss": 5.8124, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9219563970390256e-05, |
|
"loss": 5.8069, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9187045802489855e-05, |
|
"loss": 5.8026, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9154527634589446e-05, |
|
"loss": 5.8412, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.912200946668904e-05, |
|
"loss": 5.8021, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.9089491298788636e-05, |
|
"loss": 5.7984, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.905697313088823e-05, |
|
"loss": 5.7936, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.9024454962987826e-05, |
|
"loss": 5.7868, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.899193679508742e-05, |
|
"loss": 5.7637, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.895941862718701e-05, |
|
"loss": 5.7895, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.892690045928661e-05, |
|
"loss": 5.7957, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.88943822913862e-05, |
|
"loss": 5.7764, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.88618641234858e-05, |
|
"loss": 5.7723, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.882934595558539e-05, |
|
"loss": 5.763, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.879682778768498e-05, |
|
"loss": 5.7695, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.876430961978457e-05, |
|
"loss": 5.768, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.8731791451884165e-05, |
|
"loss": 5.7776, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.869927328398376e-05, |
|
"loss": 5.7606, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.8666755116083355e-05, |
|
"loss": 5.7308, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.863423694818295e-05, |
|
"loss": 5.7412, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.8601718780282545e-05, |
|
"loss": 5.7627, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.8569200612382136e-05, |
|
"loss": 5.7449, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.8536682444481735e-05, |
|
"loss": 5.7572, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.8504164276581326e-05, |
|
"loss": 5.7505, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.8471646108680925e-05, |
|
"loss": 5.7188, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.8439127940780517e-05, |
|
"loss": 5.7193, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.8406609772880115e-05, |
|
"loss": 5.7303, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.8374091604979707e-05, |
|
"loss": 5.7279, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.83415734370793e-05, |
|
"loss": 5.716, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.83090552691789e-05, |
|
"loss": 5.7237, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.827653710127849e-05, |
|
"loss": 5.7276, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.824401893337808e-05, |
|
"loss": 5.7271, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.821150076547767e-05, |
|
"loss": 5.7384, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.817898259757727e-05, |
|
"loss": 5.6949, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.814646442967686e-05, |
|
"loss": 5.7068, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.811394626177645e-05, |
|
"loss": 5.7147, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.808142809387605e-05, |
|
"loss": 5.7132, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.804890992597564e-05, |
|
"loss": 5.7256, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.801639175807524e-05, |
|
"loss": 5.7195, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.7983873590174833e-05, |
|
"loss": 5.7022, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.7951355422274425e-05, |
|
"loss": 5.6978, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.7918837254374024e-05, |
|
"loss": 5.674, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.7886319086473615e-05, |
|
"loss": 5.7044, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.7853800918573214e-05, |
|
"loss": 5.7076, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.7821282750672805e-05, |
|
"loss": 5.7146, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.77887645827724e-05, |
|
"loss": 5.6924, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.775624641487199e-05, |
|
"loss": 5.6972, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.772372824697158e-05, |
|
"loss": 5.6929, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.769121007907118e-05, |
|
"loss": 5.6553, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.765869191117077e-05, |
|
"loss": 5.6857, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.762617374327037e-05, |
|
"loss": 5.7021, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.759365557536996e-05, |
|
"loss": 5.695, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.756113740746956e-05, |
|
"loss": 5.7066, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.752861923956915e-05, |
|
"loss": 5.6919, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.749610107166874e-05, |
|
"loss": 5.6823, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.746358290376834e-05, |
|
"loss": 5.6699, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.743106473586793e-05, |
|
"loss": 5.6837, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.739854656796753e-05, |
|
"loss": 5.6773, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.736602840006712e-05, |
|
"loss": 5.6796, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.7333510232166714e-05, |
|
"loss": 5.6663, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.7300992064266305e-05, |
|
"loss": 5.6548, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.72684738963659e-05, |
|
"loss": 5.6853, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.7235955728465495e-05, |
|
"loss": 5.691, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.720343756056509e-05, |
|
"loss": 5.657, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.7170919392664686e-05, |
|
"loss": 5.6815, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.713840122476428e-05, |
|
"loss": 5.6602, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.710588305686387e-05, |
|
"loss": 5.6691, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.707336488896347e-05, |
|
"loss": 5.6767, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.704084672106306e-05, |
|
"loss": 5.6499, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.700832855316266e-05, |
|
"loss": 5.669, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.697581038526225e-05, |
|
"loss": 5.6641, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.694329221736185e-05, |
|
"loss": 5.6509, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.691077404946144e-05, |
|
"loss": 5.6501, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.687825588156103e-05, |
|
"loss": 5.6611, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.684573771366062e-05, |
|
"loss": 5.6626, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.681321954576022e-05, |
|
"loss": 5.6591, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.678070137785981e-05, |
|
"loss": 5.6519, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6748183209959404e-05, |
|
"loss": 5.6552, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6715665042058996e-05, |
|
"loss": 5.6542, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6683146874158594e-05, |
|
"loss": 5.6521, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6650628706258186e-05, |
|
"loss": 5.6631, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6618110538357784e-05, |
|
"loss": 5.6296, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.6585592370457376e-05, |
|
"loss": 5.6541, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.6553074202556974e-05, |
|
"loss": 5.6373, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.6520556034656566e-05, |
|
"loss": 5.6358, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.648803786675616e-05, |
|
"loss": 5.6654, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.6455519698855756e-05, |
|
"loss": 5.6481, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.642300153095535e-05, |
|
"loss": 5.6545, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.6390483363054946e-05, |
|
"loss": 5.6379, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.635796519515454e-05, |
|
"loss": 5.6537, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.632544702725413e-05, |
|
"loss": 5.6611, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.629292885935372e-05, |
|
"loss": 5.6639, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.626041069145331e-05, |
|
"loss": 5.6579, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.622789252355291e-05, |
|
"loss": 5.6714, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.61953743556525e-05, |
|
"loss": 5.6345, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.61628561877521e-05, |
|
"loss": 5.6457, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.613033801985169e-05, |
|
"loss": 5.6613, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.6097819851951284e-05, |
|
"loss": 5.6517, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.606530168405088e-05, |
|
"loss": 5.6328, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.6032783516150474e-05, |
|
"loss": 5.6311, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.600026534825007e-05, |
|
"loss": 5.6228, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.5967747180349665e-05, |
|
"loss": 5.6283, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.593522901244926e-05, |
|
"loss": 5.656, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.5902710844548855e-05, |
|
"loss": 5.6389, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.5870192676648446e-05, |
|
"loss": 5.6361, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.583767450874804e-05, |
|
"loss": 5.6305, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.580515634084763e-05, |
|
"loss": 5.6367, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.577263817294723e-05, |
|
"loss": 5.6393, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.574012000504682e-05, |
|
"loss": 5.6073, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.570760183714641e-05, |
|
"loss": 5.6139, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.567508366924601e-05, |
|
"loss": 5.619, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.56425655013456e-05, |
|
"loss": 5.6307, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.56100473334452e-05, |
|
"loss": 5.634, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.557752916554479e-05, |
|
"loss": 5.6234, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.554501099764439e-05, |
|
"loss": 5.6219, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.551249282974398e-05, |
|
"loss": 5.6326, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.547997466184357e-05, |
|
"loss": 5.6293, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.544745649394317e-05, |
|
"loss": 5.6153, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.541493832604276e-05, |
|
"loss": 5.6193, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.5382420158142355e-05, |
|
"loss": 5.62, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.5349901990241946e-05, |
|
"loss": 5.6276, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.5317383822341545e-05, |
|
"loss": 5.6065, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.5284865654441136e-05, |
|
"loss": 5.6008, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.525234748654073e-05, |
|
"loss": 5.624, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.5219829318640327e-05, |
|
"loss": 5.6111, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.518731115073992e-05, |
|
"loss": 5.6309, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.5154792982839517e-05, |
|
"loss": 5.6356, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.512227481493911e-05, |
|
"loss": 5.6045, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.50897566470387e-05, |
|
"loss": 5.6079, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.50572384791383e-05, |
|
"loss": 5.609, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.502472031123789e-05, |
|
"loss": 5.6262, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.499220214333749e-05, |
|
"loss": 5.5966, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.495968397543708e-05, |
|
"loss": 5.598, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.492716580753668e-05, |
|
"loss": 5.618, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.489464763963627e-05, |
|
"loss": 5.6125, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.486212947173586e-05, |
|
"loss": 5.6182, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.482961130383545e-05, |
|
"loss": 5.5995, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.4797093135935045e-05, |
|
"loss": 5.6079, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.4764574968034643e-05, |
|
"loss": 5.6138, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.4732056800134235e-05, |
|
"loss": 5.6076, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.4699538632233834e-05, |
|
"loss": 5.598, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.4667020464333425e-05, |
|
"loss": 5.6038, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.463450229643302e-05, |
|
"loss": 5.5977, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.4601984128532615e-05, |
|
"loss": 5.621, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.456946596063221e-05, |
|
"loss": 5.6076, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.4536947792731805e-05, |
|
"loss": 5.6156, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.45044296248314e-05, |
|
"loss": 5.6229, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.447191145693099e-05, |
|
"loss": 5.5821, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.443939328903059e-05, |
|
"loss": 5.5937, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.440687512113018e-05, |
|
"loss": 5.5981, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.437435695322977e-05, |
|
"loss": 5.6175, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.434183878532936e-05, |
|
"loss": 5.5978, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.430932061742896e-05, |
|
"loss": 5.5959, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.427680244952855e-05, |
|
"loss": 5.5779, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.4244284281628144e-05, |
|
"loss": 5.6002, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.421176611372774e-05, |
|
"loss": 5.6177, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.4179247945827334e-05, |
|
"loss": 5.6038, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.414672977792693e-05, |
|
"loss": 5.6033, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.4114211610026524e-05, |
|
"loss": 5.6066, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.408169344212612e-05, |
|
"loss": 5.605, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.4049175274225714e-05, |
|
"loss": 5.613, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.4016657106325305e-05, |
|
"loss": 5.5924, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.3984138938424904e-05, |
|
"loss": 5.5983, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.3951620770524496e-05, |
|
"loss": 5.5947, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.391910260262409e-05, |
|
"loss": 5.5848, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.388658443472368e-05, |
|
"loss": 5.5853, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.385406626682327e-05, |
|
"loss": 5.5949, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.382154809892287e-05, |
|
"loss": 5.5795, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.378902993102246e-05, |
|
"loss": 5.5979, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.375651176312206e-05, |
|
"loss": 5.5873, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.372399359522165e-05, |
|
"loss": 5.6173, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.369147542732125e-05, |
|
"loss": 5.6014, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.365895725942084e-05, |
|
"loss": 5.597, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.362643909152043e-05, |
|
"loss": 5.5682, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.359392092362003e-05, |
|
"loss": 5.5819, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.356140275571962e-05, |
|
"loss": 5.5784, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.352888458781922e-05, |
|
"loss": 5.5703, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.349636641991881e-05, |
|
"loss": 5.5705, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.3463848252018404e-05, |
|
"loss": 5.5742, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.3431330084118e-05, |
|
"loss": 5.6006, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.3398811916217594e-05, |
|
"loss": 5.5678, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.3366293748317186e-05, |
|
"loss": 5.5967, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.333377558041678e-05, |
|
"loss": 5.5938, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.3301257412516376e-05, |
|
"loss": 5.5844, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.326873924461597e-05, |
|
"loss": 5.5882, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.323622107671556e-05, |
|
"loss": 5.5708, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.320370290881516e-05, |
|
"loss": 5.5687, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.317118474091475e-05, |
|
"loss": 5.5928, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.313866657301435e-05, |
|
"loss": 5.5926, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 4.310614840511394e-05, |
|
"loss": 5.577, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.307363023721354e-05, |
|
"loss": 5.5642, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.304111206931313e-05, |
|
"loss": 5.5927, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.300859390141272e-05, |
|
"loss": 5.5954, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.297607573351232e-05, |
|
"loss": 5.5789, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.294355756561191e-05, |
|
"loss": 5.5541, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.29110393977115e-05, |
|
"loss": 5.5594, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.2878521229811094e-05, |
|
"loss": 5.5868, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.2846003061910686e-05, |
|
"loss": 5.5632, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.2813484894010284e-05, |
|
"loss": 5.5895, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.2780966726109876e-05, |
|
"loss": 5.594, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.2748448558209475e-05, |
|
"loss": 5.5825, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.2715930390309066e-05, |
|
"loss": 5.588, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.2683412222408665e-05, |
|
"loss": 5.5711, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.2650894054508256e-05, |
|
"loss": 5.5995, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.261837588660785e-05, |
|
"loss": 5.571, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 4.2585857718707446e-05, |
|
"loss": 5.5743, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.255333955080704e-05, |
|
"loss": 5.5766, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.2520821382906636e-05, |
|
"loss": 5.5705, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.248830321500623e-05, |
|
"loss": 5.5684, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.245578504710582e-05, |
|
"loss": 5.5593, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 4.242326687920541e-05, |
|
"loss": 5.5969, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.2390748711305e-05, |
|
"loss": 5.5724, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.23582305434046e-05, |
|
"loss": 5.5601, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.232571237550419e-05, |
|
"loss": 5.572, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.229319420760379e-05, |
|
"loss": 5.5733, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 4.226067603970338e-05, |
|
"loss": 5.5782, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.2228157871802975e-05, |
|
"loss": 5.5429, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.219563970390257e-05, |
|
"loss": 5.5873, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.2163121536002165e-05, |
|
"loss": 5.5707, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.213060336810176e-05, |
|
"loss": 5.5741, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 4.2098085200201355e-05, |
|
"loss": 5.5847, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.206556703230095e-05, |
|
"loss": 5.5845, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.2033048864400545e-05, |
|
"loss": 5.5769, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.2000530696500137e-05, |
|
"loss": 5.5777, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.1968012528599735e-05, |
|
"loss": 5.5624, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.1935494360699327e-05, |
|
"loss": 5.5753, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.190297619279892e-05, |
|
"loss": 5.572, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.187045802489851e-05, |
|
"loss": 5.5825, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.183793985699811e-05, |
|
"loss": 5.5775, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.18054216890977e-05, |
|
"loss": 5.5575, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 4.177290352119729e-05, |
|
"loss": 5.5643, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.174038535329689e-05, |
|
"loss": 5.5624, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.170786718539648e-05, |
|
"loss": 5.57, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.167534901749608e-05, |
|
"loss": 5.581, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.164283084959567e-05, |
|
"loss": 5.5746, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 4.161031268169526e-05, |
|
"loss": 5.5536, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.157779451379486e-05, |
|
"loss": 5.5706, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.1545276345894453e-05, |
|
"loss": 5.5287, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.151275817799405e-05, |
|
"loss": 5.5688, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.1480240010093644e-05, |
|
"loss": 5.5577, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.1447721842193235e-05, |
|
"loss": 5.5561, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.141520367429283e-05, |
|
"loss": 5.5381, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.138268550639242e-05, |
|
"loss": 5.5509, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.135016733849202e-05, |
|
"loss": 5.5589, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.131764917059161e-05, |
|
"loss": 5.5433, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.128513100269121e-05, |
|
"loss": 5.5809, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.12526128347908e-05, |
|
"loss": 5.5478, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.122009466689039e-05, |
|
"loss": 5.5663, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.118757649898999e-05, |
|
"loss": 5.529, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.115505833108958e-05, |
|
"loss": 5.5764, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.112254016318918e-05, |
|
"loss": 5.5561, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.109002199528877e-05, |
|
"loss": 5.5781, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.105750382738837e-05, |
|
"loss": 5.572, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.102498565948796e-05, |
|
"loss": 5.5631, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.099246749158755e-05, |
|
"loss": 5.5649, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.0959949323687144e-05, |
|
"loss": 5.5553, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.0927431155786735e-05, |
|
"loss": 5.5324, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.0894912987886334e-05, |
|
"loss": 5.5562, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.0862394819985925e-05, |
|
"loss": 5.5345, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.0829876652085524e-05, |
|
"loss": 5.5513, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.0797358484185115e-05, |
|
"loss": 5.5401, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.076484031628471e-05, |
|
"loss": 5.5566, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0732322148384306e-05, |
|
"loss": 5.538, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.06998039804839e-05, |
|
"loss": 5.5495, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0667285812583496e-05, |
|
"loss": 5.5551, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.063476764468309e-05, |
|
"loss": 5.5617, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.060224947678268e-05, |
|
"loss": 5.5654, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.056973130888228e-05, |
|
"loss": 5.5485, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.053721314098187e-05, |
|
"loss": 5.5328, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.050469497308146e-05, |
|
"loss": 5.5312, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.047217680518106e-05, |
|
"loss": 5.5306, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.043965863728065e-05, |
|
"loss": 5.5662, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.040714046938024e-05, |
|
"loss": 5.531, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0374622301479834e-05, |
|
"loss": 5.5477, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.034210413357943e-05, |
|
"loss": 5.5436, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.0309585965679024e-05, |
|
"loss": 5.5617, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.027706779777862e-05, |
|
"loss": 5.5636, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.0244549629878214e-05, |
|
"loss": 5.5282, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.021203146197781e-05, |
|
"loss": 5.55, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.0179513294077404e-05, |
|
"loss": 5.5378, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.0146995126176996e-05, |
|
"loss": 5.5463, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.0114476958276594e-05, |
|
"loss": 5.5292, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.0081958790376186e-05, |
|
"loss": 5.5282, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.0049440622475784e-05, |
|
"loss": 5.5575, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.0016922454575376e-05, |
|
"loss": 5.5299, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.998440428667497e-05, |
|
"loss": 5.55, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.995188611877456e-05, |
|
"loss": 5.5286, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.991936795087415e-05, |
|
"loss": 5.5405, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.988684978297375e-05, |
|
"loss": 5.5244, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.985433161507334e-05, |
|
"loss": 5.5208, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.982181344717294e-05, |
|
"loss": 5.5472, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.978929527927253e-05, |
|
"loss": 5.5582, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.975677711137212e-05, |
|
"loss": 5.5539, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.972425894347172e-05, |
|
"loss": 5.5463, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.969174077557131e-05, |
|
"loss": 5.5498, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.965922260767091e-05, |
|
"loss": 5.5299, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.96267044397705e-05, |
|
"loss": 5.5384, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.95941862718701e-05, |
|
"loss": 5.5288, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.956166810396969e-05, |
|
"loss": 5.5213, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.9529149936069285e-05, |
|
"loss": 5.5386, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.9496631768168876e-05, |
|
"loss": 5.5413, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.946411360026847e-05, |
|
"loss": 5.5387, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.9431595432368066e-05, |
|
"loss": 5.5342, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.939907726446766e-05, |
|
"loss": 5.5485, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.936655909656725e-05, |
|
"loss": 5.532, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.933404092866685e-05, |
|
"loss": 5.518, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.930152276076644e-05, |
|
"loss": 5.5319, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.926900459286604e-05, |
|
"loss": 5.5335, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.923648642496563e-05, |
|
"loss": 5.5488, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.920396825706523e-05, |
|
"loss": 5.5252, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.917145008916482e-05, |
|
"loss": 5.5141, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.913893192126441e-05, |
|
"loss": 5.5417, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.910641375336401e-05, |
|
"loss": 5.5355, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.90738955854636e-05, |
|
"loss": 5.5415, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.904137741756319e-05, |
|
"loss": 5.534, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.9008859249662785e-05, |
|
"loss": 5.5257, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.897634108176238e-05, |
|
"loss": 5.5221, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.8943822913861975e-05, |
|
"loss": 5.5364, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.8911304745961566e-05, |
|
"loss": 5.5296, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.8878786578061165e-05, |
|
"loss": 5.5197, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.8846268410160756e-05, |
|
"loss": 5.5615, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.8813750242260355e-05, |
|
"loss": 5.5354, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.8781232074359947e-05, |
|
"loss": 5.5192, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.874871390645954e-05, |
|
"loss": 5.5497, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.8716195738559137e-05, |
|
"loss": 5.5419, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.868367757065873e-05, |
|
"loss": 5.5366, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.865115940275833e-05, |
|
"loss": 5.5031, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.861864123485792e-05, |
|
"loss": 5.5425, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.858612306695752e-05, |
|
"loss": 5.5195, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.855360489905711e-05, |
|
"loss": 5.5235, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.85210867311567e-05, |
|
"loss": 5.5299, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.848856856325629e-05, |
|
"loss": 5.5233, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.845605039535588e-05, |
|
"loss": 5.5386, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.842353222745548e-05, |
|
"loss": 5.5233, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.839101405955507e-05, |
|
"loss": 5.5128, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.8358495891654665e-05, |
|
"loss": 5.5286, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.8325977723754263e-05, |
|
"loss": 5.5022, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.8293459555853855e-05, |
|
"loss": 5.523, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.8260941387953454e-05, |
|
"loss": 5.5229, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.8228423220053045e-05, |
|
"loss": 5.5352, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.8195905052152644e-05, |
|
"loss": 5.5312, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.8163386884252235e-05, |
|
"loss": 5.5209, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.813086871635183e-05, |
|
"loss": 5.5273, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.8098350548451425e-05, |
|
"loss": 5.5092, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.806583238055102e-05, |
|
"loss": 5.5154, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.803331421265061e-05, |
|
"loss": 5.5246, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.80007960447502e-05, |
|
"loss": 5.5372, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.79682778768498e-05, |
|
"loss": 5.5258, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.793575970894939e-05, |
|
"loss": 5.5236, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.790324154104898e-05, |
|
"loss": 5.4988, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.787072337314858e-05, |
|
"loss": 5.5234, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.783820520524817e-05, |
|
"loss": 5.5285, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.780568703734777e-05, |
|
"loss": 5.5015, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.777316886944736e-05, |
|
"loss": 5.5222, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.7740650701546954e-05, |
|
"loss": 5.5113, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.770813253364655e-05, |
|
"loss": 5.524, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.7675614365746144e-05, |
|
"loss": 5.5357, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.764309619784574e-05, |
|
"loss": 5.5221, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.7610578029945334e-05, |
|
"loss": 5.5226, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.7578059862044925e-05, |
|
"loss": 5.5204, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.754554169414452e-05, |
|
"loss": 5.5201, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.751302352624411e-05, |
|
"loss": 5.521, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.748050535834371e-05, |
|
"loss": 5.5334, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.74479871904433e-05, |
|
"loss": 5.4979, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.74154690225429e-05, |
|
"loss": 5.5098, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.738295085464249e-05, |
|
"loss": 5.5429, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.735043268674209e-05, |
|
"loss": 5.506, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.731791451884168e-05, |
|
"loss": 5.5116, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.728539635094127e-05, |
|
"loss": 5.5075, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.725287818304087e-05, |
|
"loss": 5.5159, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.722036001514046e-05, |
|
"loss": 5.5192, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.718784184724006e-05, |
|
"loss": 5.5109, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.715532367933965e-05, |
|
"loss": 5.496, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.712280551143924e-05, |
|
"loss": 5.5329, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.709028734353884e-05, |
|
"loss": 5.4942, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.705776917563843e-05, |
|
"loss": 5.5314, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.7025251007738024e-05, |
|
"loss": 5.5117, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.6992732839837616e-05, |
|
"loss": 5.5068, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.6960214671937214e-05, |
|
"loss": 5.5034, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.6927696504036806e-05, |
|
"loss": 5.5165, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.68951783361364e-05, |
|
"loss": 5.5228, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.6862660168235996e-05, |
|
"loss": 5.522, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.683014200033559e-05, |
|
"loss": 5.5127, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.6797623832435186e-05, |
|
"loss": 5.498, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.676510566453478e-05, |
|
"loss": 5.5175, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.6732587496634376e-05, |
|
"loss": 5.4934, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.670006932873397e-05, |
|
"loss": 5.5252, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.666755116083356e-05, |
|
"loss": 5.4944, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.663503299293316e-05, |
|
"loss": 5.4928, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.660251482503275e-05, |
|
"loss": 5.4842, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.656999665713234e-05, |
|
"loss": 5.4963, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.653747848923193e-05, |
|
"loss": 5.4839, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.6504960321331524e-05, |
|
"loss": 5.5047, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.647244215343112e-05, |
|
"loss": 5.5055, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.6439923985530714e-05, |
|
"loss": 5.467, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.640740581763031e-05, |
|
"loss": 5.4914, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6374887649729904e-05, |
|
"loss": 5.4803, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.63423694818295e-05, |
|
"loss": 5.4874, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6309851313929095e-05, |
|
"loss": 5.4186, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.6277333146028686e-05, |
|
"loss": 5.3267, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.6244814978128285e-05, |
|
"loss": 5.2962, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.6212296810227876e-05, |
|
"loss": 5.2441, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.6179778642327475e-05, |
|
"loss": 5.1665, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.6147260474427066e-05, |
|
"loss": 5.1096, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.611474230652666e-05, |
|
"loss": 5.0509, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.608222413862625e-05, |
|
"loss": 5.0168, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.604970597072584e-05, |
|
"loss": 4.9811, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.601718780282544e-05, |
|
"loss": 4.9535, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.598466963492503e-05, |
|
"loss": 4.9113, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.595215146702463e-05, |
|
"loss": 4.8715, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.591963329912422e-05, |
|
"loss": 4.8401, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.588711513122381e-05, |
|
"loss": 4.8103, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.585459696332341e-05, |
|
"loss": 4.7615, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.5822078795423e-05, |
|
"loss": 4.7477, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.57895606275226e-05, |
|
"loss": 4.6929, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.575704245962219e-05, |
|
"loss": 4.6674, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.572452429172179e-05, |
|
"loss": 4.6552, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.569200612382138e-05, |
|
"loss": 4.5982, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.5659487955920975e-05, |
|
"loss": 4.4937, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.562696978802057e-05, |
|
"loss": 4.3654, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.5594451620120165e-05, |
|
"loss": 4.249, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.5561933452219757e-05, |
|
"loss": 4.122, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.552941528431935e-05, |
|
"loss": 3.9966, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.549689711641894e-05, |
|
"loss": 3.8664, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.546437894851854e-05, |
|
"loss": 3.7561, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.543186078061813e-05, |
|
"loss": 3.6266, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.539934261271773e-05, |
|
"loss": 3.5324, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.536682444481732e-05, |
|
"loss": 3.439, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.533430627691692e-05, |
|
"loss": 3.3637, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.530178810901651e-05, |
|
"loss": 3.2801, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.52692699411161e-05, |
|
"loss": 3.1894, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.52367517732157e-05, |
|
"loss": 3.1034, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.520423360531529e-05, |
|
"loss": 3.0252, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.517171543741489e-05, |
|
"loss": 2.9289, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.513919726951448e-05, |
|
"loss": 2.8581, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.5106679101614073e-05, |
|
"loss": 2.8208, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.5074160933713665e-05, |
|
"loss": 2.7686, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.504164276581326e-05, |
|
"loss": 2.7256, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.5009124597912855e-05, |
|
"loss": 2.6784, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.497660643001245e-05, |
|
"loss": 2.6355, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.4944088262112045e-05, |
|
"loss": 2.6078, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.491157009421164e-05, |
|
"loss": 2.575, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.487905192631123e-05, |
|
"loss": 2.5266, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.484653375841083e-05, |
|
"loss": 2.5291, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.481401559051042e-05, |
|
"loss": 2.4816, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.478149742261002e-05, |
|
"loss": 2.4666, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.474897925470961e-05, |
|
"loss": 2.4444, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.471646108680921e-05, |
|
"loss": 2.4253, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.46839429189088e-05, |
|
"loss": 2.403, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.465142475100839e-05, |
|
"loss": 2.3755, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.461890658310798e-05, |
|
"loss": 2.35, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.4586388415207574e-05, |
|
"loss": 2.3292, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.455387024730717e-05, |
|
"loss": 2.3076, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.4521352079406764e-05, |
|
"loss": 2.2888, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.448883391150636e-05, |
|
"loss": 2.267, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.4456315743605954e-05, |
|
"loss": 2.2615, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.4423797575705545e-05, |
|
"loss": 2.2433, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.4391279407805144e-05, |
|
"loss": 2.2348, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.4358761239904735e-05, |
|
"loss": 2.2083, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.4326243072004334e-05, |
|
"loss": 2.2022, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.4293724904103926e-05, |
|
"loss": 2.2007, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 3.426120673620352e-05, |
|
"loss": 2.1813, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.4228688568303116e-05, |
|
"loss": 2.157, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.419617040040271e-05, |
|
"loss": 2.1462, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.41636522325023e-05, |
|
"loss": 2.1346, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.41311340646019e-05, |
|
"loss": 2.1271, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.409861589670149e-05, |
|
"loss": 2.1027, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.406609772880108e-05, |
|
"loss": 2.1023, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.403357956090067e-05, |
|
"loss": 2.0849, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.400106139300027e-05, |
|
"loss": 2.0843, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.396854322509986e-05, |
|
"loss": 2.0666, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.393602505719946e-05, |
|
"loss": 2.0578, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.390350688929905e-05, |
|
"loss": 2.0446, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.3870988721398644e-05, |
|
"loss": 2.0486, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.383847055349824e-05, |
|
"loss": 2.0247, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.3805952385597834e-05, |
|
"loss": 2.0286, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.377343421769743e-05, |
|
"loss": 2.0144, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.3740916049797024e-05, |
|
"loss": 2.0024, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.370839788189662e-05, |
|
"loss": 1.9928, |
|
"step": 250500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.3675879713996214e-05, |
|
"loss": 1.9791, |
|
"step": 251000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.3643361546095806e-05, |
|
"loss": 1.9735, |
|
"step": 251500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.36108433781954e-05, |
|
"loss": 1.9699, |
|
"step": 252000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.357832521029499e-05, |
|
"loss": 1.9683, |
|
"step": 252500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.354580704239459e-05, |
|
"loss": 1.9456, |
|
"step": 253000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.351328887449418e-05, |
|
"loss": 1.9365, |
|
"step": 253500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.348077070659378e-05, |
|
"loss": 1.9347, |
|
"step": 254000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.344825253869337e-05, |
|
"loss": 1.9325, |
|
"step": 254500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.341573437079296e-05, |
|
"loss": 1.9314, |
|
"step": 255000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.338321620289256e-05, |
|
"loss": 1.9199, |
|
"step": 255500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.335069803499215e-05, |
|
"loss": 1.905, |
|
"step": 256000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.331817986709175e-05, |
|
"loss": 1.9004, |
|
"step": 256500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.328566169919134e-05, |
|
"loss": 1.8919, |
|
"step": 257000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.325314353129093e-05, |
|
"loss": 1.8801, |
|
"step": 257500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.322062536339053e-05, |
|
"loss": 1.8772, |
|
"step": 258000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.318810719549012e-05, |
|
"loss": 1.8774, |
|
"step": 258500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.3155589027589714e-05, |
|
"loss": 1.8752, |
|
"step": 259000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.3123070859689306e-05, |
|
"loss": 1.852, |
|
"step": 259500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.3090552691788905e-05, |
|
"loss": 1.8492, |
|
"step": 260000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.3058034523888496e-05, |
|
"loss": 1.8562, |
|
"step": 260500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.302551635598809e-05, |
|
"loss": 1.843, |
|
"step": 261000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.2992998188087686e-05, |
|
"loss": 1.8321, |
|
"step": 261500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.296048002018728e-05, |
|
"loss": 1.8233, |
|
"step": 262000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 3.2927961852286876e-05, |
|
"loss": 1.8265, |
|
"step": 262500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.289544368438647e-05, |
|
"loss": 1.818, |
|
"step": 263000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.2862925516486066e-05, |
|
"loss": 1.8186, |
|
"step": 263500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.283040734858566e-05, |
|
"loss": 1.7974, |
|
"step": 264000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.279788918068525e-05, |
|
"loss": 1.7867, |
|
"step": 264500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.276537101278485e-05, |
|
"loss": 1.7938, |
|
"step": 265000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.273285284488444e-05, |
|
"loss": 1.7877, |
|
"step": 265500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.270033467698403e-05, |
|
"loss": 1.7923, |
|
"step": 266000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.266781650908362e-05, |
|
"loss": 1.7794, |
|
"step": 266500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.263529834118322e-05, |
|
"loss": 1.7809, |
|
"step": 267000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.260278017328281e-05, |
|
"loss": 1.7654, |
|
"step": 267500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.2570262005382405e-05, |
|
"loss": 1.7569, |
|
"step": 268000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.2537743837482e-05, |
|
"loss": 1.7653, |
|
"step": 268500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.2505225669581595e-05, |
|
"loss": 1.7554, |
|
"step": 269000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.247270750168119e-05, |
|
"loss": 1.7461, |
|
"step": 269500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 3.2440189333780785e-05, |
|
"loss": 1.7386, |
|
"step": 270000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.2407671165880376e-05, |
|
"loss": 1.738, |
|
"step": 270500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.2375152997979975e-05, |
|
"loss": 1.7343, |
|
"step": 271000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.2342634830079567e-05, |
|
"loss": 1.7334, |
|
"step": 271500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.2310116662179165e-05, |
|
"loss": 1.7163, |
|
"step": 272000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.2277598494278757e-05, |
|
"loss": 1.7176, |
|
"step": 272500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.2245080326378355e-05, |
|
"loss": 1.7203, |
|
"step": 273000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.221256215847795e-05, |
|
"loss": 1.7172, |
|
"step": 273500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.218004399057754e-05, |
|
"loss": 1.703, |
|
"step": 274000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.214752582267713e-05, |
|
"loss": 1.6993, |
|
"step": 274500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.211500765477672e-05, |
|
"loss": 1.6958, |
|
"step": 275000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.208248948687632e-05, |
|
"loss": 1.6965, |
|
"step": 275500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.204997131897591e-05, |
|
"loss": 1.6939, |
|
"step": 276000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.20174531510755e-05, |
|
"loss": 1.6792, |
|
"step": 276500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.19849349831751e-05, |
|
"loss": 1.6956, |
|
"step": 277000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.195241681527469e-05, |
|
"loss": 1.6888, |
|
"step": 277500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.191989864737429e-05, |
|
"loss": 1.6803, |
|
"step": 278000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.1887380479473883e-05, |
|
"loss": 1.6753, |
|
"step": 278500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.185486231157348e-05, |
|
"loss": 1.6731, |
|
"step": 279000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.1822344143673074e-05, |
|
"loss": 1.6584, |
|
"step": 279500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.1789825975772665e-05, |
|
"loss": 1.6632, |
|
"step": 280000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.1757307807872264e-05, |
|
"loss": 1.6526, |
|
"step": 280500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.1724789639971855e-05, |
|
"loss": 1.6596, |
|
"step": 281000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.169227147207145e-05, |
|
"loss": 1.6466, |
|
"step": 281500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.165975330417104e-05, |
|
"loss": 1.6503, |
|
"step": 282000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.162723513627063e-05, |
|
"loss": 1.6462, |
|
"step": 282500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.159471696837023e-05, |
|
"loss": 1.6459, |
|
"step": 283000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.156219880046982e-05, |
|
"loss": 1.6391, |
|
"step": 283500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.152968063256942e-05, |
|
"loss": 1.6301, |
|
"step": 284000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.149716246466901e-05, |
|
"loss": 1.627, |
|
"step": 284500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.146464429676861e-05, |
|
"loss": 1.6331, |
|
"step": 285000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.14321261288682e-05, |
|
"loss": 1.6254, |
|
"step": 285500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.139960796096779e-05, |
|
"loss": 1.6201, |
|
"step": 286000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.136708979306739e-05, |
|
"loss": 1.6213, |
|
"step": 286500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.133457162516698e-05, |
|
"loss": 1.6158, |
|
"step": 287000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.130205345726658e-05, |
|
"loss": 1.6142, |
|
"step": 287500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.126953528936617e-05, |
|
"loss": 1.6048, |
|
"step": 288000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.1237017121465764e-05, |
|
"loss": 1.6069, |
|
"step": 288500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.1204498953565355e-05, |
|
"loss": 1.601, |
|
"step": 289000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.117198078566495e-05, |
|
"loss": 1.6069, |
|
"step": 289500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.1139462617764545e-05, |
|
"loss": 1.5992, |
|
"step": 290000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.110694444986414e-05, |
|
"loss": 1.6011, |
|
"step": 290500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.1074426281963736e-05, |
|
"loss": 1.5995, |
|
"step": 291000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.104190811406333e-05, |
|
"loss": 1.5854, |
|
"step": 291500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.100938994616292e-05, |
|
"loss": 1.5855, |
|
"step": 292000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.097687177826252e-05, |
|
"loss": 1.58, |
|
"step": 292500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.094435361036211e-05, |
|
"loss": 1.5733, |
|
"step": 293000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.091183544246171e-05, |
|
"loss": 1.5734, |
|
"step": 293500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.08793172745613e-05, |
|
"loss": 1.5749, |
|
"step": 294000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.08467991066609e-05, |
|
"loss": 1.5793, |
|
"step": 294500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.081428093876049e-05, |
|
"loss": 1.57, |
|
"step": 295000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 3.078176277086008e-05, |
|
"loss": 1.559, |
|
"step": 295500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.074924460295968e-05, |
|
"loss": 1.5605, |
|
"step": 296000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.071672643505927e-05, |
|
"loss": 1.5528, |
|
"step": 296500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.068420826715886e-05, |
|
"loss": 1.5572, |
|
"step": 297000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.0651690099258454e-05, |
|
"loss": 1.5568, |
|
"step": 297500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.061917193135805e-05, |
|
"loss": 1.556, |
|
"step": 298000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.0586653763457644e-05, |
|
"loss": 1.5599, |
|
"step": 298500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0554135595557236e-05, |
|
"loss": 1.5531, |
|
"step": 299000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0521617427656834e-05, |
|
"loss": 1.5557, |
|
"step": 299500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0489099259756426e-05, |
|
"loss": 1.5438, |
|
"step": 300000 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0456581091856024e-05, |
|
"loss": 1.5526, |
|
"step": 300500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.0424062923955616e-05, |
|
"loss": 1.5352, |
|
"step": 301000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.0391544756055208e-05, |
|
"loss": 1.5359, |
|
"step": 301500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.0359026588154803e-05, |
|
"loss": 1.5354, |
|
"step": 302000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.0326508420254394e-05, |
|
"loss": 1.5295, |
|
"step": 302500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.0293990252353993e-05, |
|
"loss": 1.5296, |
|
"step": 303000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.0261472084453584e-05, |
|
"loss": 1.5247, |
|
"step": 303500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0228953916553183e-05, |
|
"loss": 1.5268, |
|
"step": 304000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0196435748652774e-05, |
|
"loss": 1.5214, |
|
"step": 304500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0163917580752366e-05, |
|
"loss": 1.5246, |
|
"step": 305000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0131399412851964e-05, |
|
"loss": 1.5274, |
|
"step": 305500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0098881244951556e-05, |
|
"loss": 1.5078, |
|
"step": 306000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.006636307705115e-05, |
|
"loss": 1.5175, |
|
"step": 306500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.0033844909150743e-05, |
|
"loss": 1.5166, |
|
"step": 307000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.000132674125034e-05, |
|
"loss": 1.5126, |
|
"step": 307500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 2.9968808573349933e-05, |
|
"loss": 1.5188, |
|
"step": 308000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 2.9936290405449524e-05, |
|
"loss": 1.5015, |
|
"step": 308500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.9903772237549123e-05, |
|
"loss": 1.505, |
|
"step": 309000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.9871254069648715e-05, |
|
"loss": 1.4986, |
|
"step": 309500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.983873590174831e-05, |
|
"loss": 1.4994, |
|
"step": 310000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.98062177338479e-05, |
|
"loss": 1.4982, |
|
"step": 310500 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.9773699565947493e-05, |
|
"loss": 1.4874, |
|
"step": 311000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 2.974118139804709e-05, |
|
"loss": 1.4915, |
|
"step": 311500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 2.9708663230146683e-05, |
|
"loss": 1.4996, |
|
"step": 312000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 2.967614506224628e-05, |
|
"loss": 1.4948, |
|
"step": 312500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 2.9643626894345873e-05, |
|
"loss": 1.4842, |
|
"step": 313000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 2.9611108726445468e-05, |
|
"loss": 1.4798, |
|
"step": 313500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.957859055854506e-05, |
|
"loss": 1.491, |
|
"step": 314000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.954607239064465e-05, |
|
"loss": 1.4821, |
|
"step": 314500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.951355422274425e-05, |
|
"loss": 1.4841, |
|
"step": 315000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.948103605484384e-05, |
|
"loss": 1.4746, |
|
"step": 315500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.944851788694344e-05, |
|
"loss": 1.4767, |
|
"step": 316000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.941599971904303e-05, |
|
"loss": 1.4775, |
|
"step": 316500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.9383481551142623e-05, |
|
"loss": 1.4777, |
|
"step": 317000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.9350963383242218e-05, |
|
"loss": 1.4718, |
|
"step": 317500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.931844521534181e-05, |
|
"loss": 1.4654, |
|
"step": 318000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.9285927047441408e-05, |
|
"loss": 1.4667, |
|
"step": 318500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.9253408879541e-05, |
|
"loss": 1.4718, |
|
"step": 319000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.9220890711640598e-05, |
|
"loss": 1.4644, |
|
"step": 319500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.918837254374019e-05, |
|
"loss": 1.4597, |
|
"step": 320000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.915585437583978e-05, |
|
"loss": 1.4532, |
|
"step": 320500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.9123336207939377e-05, |
|
"loss": 1.4635, |
|
"step": 321000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.9090818040038968e-05, |
|
"loss": 1.4501, |
|
"step": 321500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.9058299872138567e-05, |
|
"loss": 1.4521, |
|
"step": 322000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.9025781704238158e-05, |
|
"loss": 1.4588, |
|
"step": 322500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.8993263536337757e-05, |
|
"loss": 1.4562, |
|
"step": 323000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.896074536843735e-05, |
|
"loss": 1.4555, |
|
"step": 323500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.892822720053694e-05, |
|
"loss": 1.4483, |
|
"step": 324000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.8895709032636535e-05, |
|
"loss": 1.4548, |
|
"step": 324500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.8863190864736127e-05, |
|
"loss": 1.4491, |
|
"step": 325000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.8830672696835725e-05, |
|
"loss": 1.4398, |
|
"step": 325500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.8798154528935317e-05, |
|
"loss": 1.4409, |
|
"step": 326000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.876563636103491e-05, |
|
"loss": 1.4472, |
|
"step": 326500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.8733118193134507e-05, |
|
"loss": 1.4316, |
|
"step": 327000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.87006000252341e-05, |
|
"loss": 1.4255, |
|
"step": 327500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.8668081857333693e-05, |
|
"loss": 1.4308, |
|
"step": 328000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.863556368943329e-05, |
|
"loss": 1.4355, |
|
"step": 328500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 2.8603045521532884e-05, |
|
"loss": 1.4294, |
|
"step": 329000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.8570527353632475e-05, |
|
"loss": 1.4267, |
|
"step": 329500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.8538009185732067e-05, |
|
"loss": 1.4333, |
|
"step": 330000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.8505491017831665e-05, |
|
"loss": 1.4254, |
|
"step": 330500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.8472972849931257e-05, |
|
"loss": 1.4281, |
|
"step": 331000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.8440454682030855e-05, |
|
"loss": 1.426, |
|
"step": 331500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8407936514130447e-05, |
|
"loss": 1.4294, |
|
"step": 332000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8375418346230042e-05, |
|
"loss": 1.4208, |
|
"step": 332500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8342900178329634e-05, |
|
"loss": 1.421, |
|
"step": 333000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8310382010429225e-05, |
|
"loss": 1.4226, |
|
"step": 333500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 2.8277863842528824e-05, |
|
"loss": 1.417, |
|
"step": 334000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 2.8245345674628415e-05, |
|
"loss": 1.4166, |
|
"step": 334500 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 2.8212827506728014e-05, |
|
"loss": 1.4207, |
|
"step": 335000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 2.8180309338827605e-05, |
|
"loss": 1.408, |
|
"step": 335500 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 2.8147791170927197e-05, |
|
"loss": 1.412, |
|
"step": 336000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 2.8115273003026792e-05, |
|
"loss": 1.412, |
|
"step": 336500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.8082754835126384e-05, |
|
"loss": 1.4137, |
|
"step": 337000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.8050236667225982e-05, |
|
"loss": 1.4032, |
|
"step": 337500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.8017718499325574e-05, |
|
"loss": 1.4022, |
|
"step": 338000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.7985200331425172e-05, |
|
"loss": 1.4038, |
|
"step": 338500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.7952682163524764e-05, |
|
"loss": 1.3994, |
|
"step": 339000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.7920163995624355e-05, |
|
"loss": 1.3996, |
|
"step": 339500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 2.788764582772395e-05, |
|
"loss": 1.3937, |
|
"step": 340000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 2.7855127659823542e-05, |
|
"loss": 1.3994, |
|
"step": 340500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 2.782260949192314e-05, |
|
"loss": 1.4008, |
|
"step": 341000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 2.7790091324022732e-05, |
|
"loss": 1.399, |
|
"step": 341500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 2.775757315612233e-05, |
|
"loss": 1.398, |
|
"step": 342000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.7725054988221922e-05, |
|
"loss": 1.3928, |
|
"step": 342500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.7692536820321514e-05, |
|
"loss": 1.3855, |
|
"step": 343000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.766001865242111e-05, |
|
"loss": 1.394, |
|
"step": 343500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.76275004845207e-05, |
|
"loss": 1.3863, |
|
"step": 344000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 2.75949823166203e-05, |
|
"loss": 1.3915, |
|
"step": 344500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.756246414871989e-05, |
|
"loss": 1.392, |
|
"step": 345000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.7529945980819482e-05, |
|
"loss": 1.389, |
|
"step": 345500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.749742781291908e-05, |
|
"loss": 1.3836, |
|
"step": 346000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.7464909645018672e-05, |
|
"loss": 1.3852, |
|
"step": 346500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.7432391477118267e-05, |
|
"loss": 1.3736, |
|
"step": 347000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.739987330921786e-05, |
|
"loss": 1.3818, |
|
"step": 347500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.7367355141317457e-05, |
|
"loss": 1.3816, |
|
"step": 348000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.733483697341705e-05, |
|
"loss": 1.3766, |
|
"step": 348500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.730231880551664e-05, |
|
"loss": 1.3835, |
|
"step": 349000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.726980063761624e-05, |
|
"loss": 1.3843, |
|
"step": 349500 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 2.723728246971583e-05, |
|
"loss": 1.3723, |
|
"step": 350000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 2.7204764301815426e-05, |
|
"loss": 1.3802, |
|
"step": 350500 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 2.7172246133915018e-05, |
|
"loss": 1.3747, |
|
"step": 351000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 2.7139727966014616e-05, |
|
"loss": 1.3662, |
|
"step": 351500 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 2.7107209798114208e-05, |
|
"loss": 1.3712, |
|
"step": 352000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.70746916302138e-05, |
|
"loss": 1.3805, |
|
"step": 352500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.7042173462313398e-05, |
|
"loss": 1.3647, |
|
"step": 353000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.700965529441299e-05, |
|
"loss": 1.3731, |
|
"step": 353500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.6977137126512588e-05, |
|
"loss": 1.3663, |
|
"step": 354000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 2.694461895861218e-05, |
|
"loss": 1.3559, |
|
"step": 354500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.691210079071177e-05, |
|
"loss": 1.3642, |
|
"step": 355000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.6879582622811366e-05, |
|
"loss": 1.3679, |
|
"step": 355500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.6847064454910958e-05, |
|
"loss": 1.3576, |
|
"step": 356000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.6814546287010556e-05, |
|
"loss": 1.3536, |
|
"step": 356500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.6782028119110148e-05, |
|
"loss": 1.3635, |
|
"step": 357000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6749509951209746e-05, |
|
"loss": 1.3622, |
|
"step": 357500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6716991783309338e-05, |
|
"loss": 1.3594, |
|
"step": 358000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.668447361540893e-05, |
|
"loss": 1.3559, |
|
"step": 358500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6651955447508525e-05, |
|
"loss": 1.351, |
|
"step": 359000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6619437279608116e-05, |
|
"loss": 1.3582, |
|
"step": 359500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.6586919111707715e-05, |
|
"loss": 1.3567, |
|
"step": 360000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 2.6554400943807306e-05, |
|
"loss": 1.3417, |
|
"step": 360500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 2.6521882775906898e-05, |
|
"loss": 1.3467, |
|
"step": 361000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 2.6489364608006496e-05, |
|
"loss": 1.3623, |
|
"step": 361500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 2.6456846440106088e-05, |
|
"loss": 1.3495, |
|
"step": 362000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 2.6424328272205683e-05, |
|
"loss": 1.3514, |
|
"step": 362500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.6391810104305275e-05, |
|
"loss": 1.342, |
|
"step": 363000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.6359291936404873e-05, |
|
"loss": 1.3363, |
|
"step": 363500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.6326773768504465e-05, |
|
"loss": 1.3467, |
|
"step": 364000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.6294255600604056e-05, |
|
"loss": 1.345, |
|
"step": 364500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 2.6261737432703655e-05, |
|
"loss": 1.3486, |
|
"step": 365000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.6229219264803246e-05, |
|
"loss": 1.3505, |
|
"step": 365500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.619670109690284e-05, |
|
"loss": 1.3397, |
|
"step": 366000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.6164182929002433e-05, |
|
"loss": 1.3381, |
|
"step": 366500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.613166476110203e-05, |
|
"loss": 1.3367, |
|
"step": 367000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.6099146593201623e-05, |
|
"loss": 1.3375, |
|
"step": 367500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.6066628425301215e-05, |
|
"loss": 1.3405, |
|
"step": 368000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.6034110257400813e-05, |
|
"loss": 1.3384, |
|
"step": 368500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.6001592089500405e-05, |
|
"loss": 1.3357, |
|
"step": 369000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.59690739216e-05, |
|
"loss": 1.3349, |
|
"step": 369500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 2.593655575369959e-05, |
|
"loss": 1.3386, |
|
"step": 370000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 2.5904037585799183e-05, |
|
"loss": 1.3269, |
|
"step": 370500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 2.587151941789878e-05, |
|
"loss": 1.3368, |
|
"step": 371000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 2.5839001249998373e-05, |
|
"loss": 1.3378, |
|
"step": 371500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 2.580648308209797e-05, |
|
"loss": 1.3308, |
|
"step": 372000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 2.5773964914197563e-05, |
|
"loss": 1.3222, |
|
"step": 372500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.574144674629716e-05, |
|
"loss": 1.3313, |
|
"step": 373000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.570892857839675e-05, |
|
"loss": 1.3274, |
|
"step": 373500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.567641041049634e-05, |
|
"loss": 1.3282, |
|
"step": 374000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.564389224259594e-05, |
|
"loss": 1.3198, |
|
"step": 374500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.561137407469553e-05, |
|
"loss": 1.319, |
|
"step": 375000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 2.557885590679513e-05, |
|
"loss": 1.3279, |
|
"step": 375500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 2.5546337738894722e-05, |
|
"loss": 1.3206, |
|
"step": 376000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 2.5513819570994317e-05, |
|
"loss": 1.3237, |
|
"step": 376500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 2.5481301403093912e-05, |
|
"loss": 1.317, |
|
"step": 377000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 2.5448783235193503e-05, |
|
"loss": 1.3168, |
|
"step": 377500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.54162650672931e-05, |
|
"loss": 1.3219, |
|
"step": 378000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.538374689939269e-05, |
|
"loss": 1.316, |
|
"step": 378500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.535122873149229e-05, |
|
"loss": 1.3099, |
|
"step": 379000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.531871056359188e-05, |
|
"loss": 1.3184, |
|
"step": 379500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.5286192395691472e-05, |
|
"loss": 1.32, |
|
"step": 380000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 2.525367422779107e-05, |
|
"loss": 1.3118, |
|
"step": 380500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.5221156059890662e-05, |
|
"loss": 1.301, |
|
"step": 381000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.5188637891990257e-05, |
|
"loss": 1.3119, |
|
"step": 381500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.515611972408985e-05, |
|
"loss": 1.3115, |
|
"step": 382000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.5123601556189447e-05, |
|
"loss": 1.3155, |
|
"step": 382500 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 2.509108338828904e-05, |
|
"loss": 1.318, |
|
"step": 383000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.505856522038863e-05, |
|
"loss": 1.3038, |
|
"step": 383500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.502604705248823e-05, |
|
"loss": 1.3007, |
|
"step": 384000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.499352888458782e-05, |
|
"loss": 1.3096, |
|
"step": 384500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.4961010716687412e-05, |
|
"loss": 1.2961, |
|
"step": 385000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.4928492548787007e-05, |
|
"loss": 1.3118, |
|
"step": 385500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.4895974380886602e-05, |
|
"loss": 1.3076, |
|
"step": 386000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.4863456212986197e-05, |
|
"loss": 1.3118, |
|
"step": 386500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.4830938045085792e-05, |
|
"loss": 1.2965, |
|
"step": 387000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.4798419877185387e-05, |
|
"loss": 1.302, |
|
"step": 387500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.476590170928498e-05, |
|
"loss": 1.3042, |
|
"step": 388000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 2.473338354138457e-05, |
|
"loss": 1.3032, |
|
"step": 388500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 2.4700865373484165e-05, |
|
"loss": 1.3031, |
|
"step": 389000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 2.466834720558376e-05, |
|
"loss": 1.2966, |
|
"step": 389500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 2.4635829037683356e-05, |
|
"loss": 1.3008, |
|
"step": 390000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 2.460331086978295e-05, |
|
"loss": 1.2896, |
|
"step": 390500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.4570792701882542e-05, |
|
"loss": 1.2999, |
|
"step": 391000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.4538274533982137e-05, |
|
"loss": 1.3017, |
|
"step": 391500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.4505756366081732e-05, |
|
"loss": 1.2969, |
|
"step": 392000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.4473238198181324e-05, |
|
"loss": 1.297, |
|
"step": 392500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.444072003028092e-05, |
|
"loss": 1.2959, |
|
"step": 393000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 2.4408201862380514e-05, |
|
"loss": 1.3034, |
|
"step": 393500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 2.4375683694480106e-05, |
|
"loss": 1.285, |
|
"step": 394000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 2.43431655265797e-05, |
|
"loss": 1.2913, |
|
"step": 394500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 2.4310647358679296e-05, |
|
"loss": 1.2806, |
|
"step": 395000 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 2.427812919077889e-05, |
|
"loss": 1.2842, |
|
"step": 395500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.4245611022878482e-05, |
|
"loss": 1.2776, |
|
"step": 396000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.4213092854978077e-05, |
|
"loss": 1.2867, |
|
"step": 396500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.4180574687077672e-05, |
|
"loss": 1.2906, |
|
"step": 397000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.4148056519177264e-05, |
|
"loss": 1.2907, |
|
"step": 397500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 2.411553835127686e-05, |
|
"loss": 1.2786, |
|
"step": 398000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.4083020183376454e-05, |
|
"loss": 1.2943, |
|
"step": 398500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.405050201547605e-05, |
|
"loss": 1.2758, |
|
"step": 399000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.401798384757564e-05, |
|
"loss": 1.2839, |
|
"step": 399500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.3985465679675236e-05, |
|
"loss": 1.2819, |
|
"step": 400000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.3952947511774828e-05, |
|
"loss": 1.2791, |
|
"step": 400500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 2.3920429343874423e-05, |
|
"loss": 1.2753, |
|
"step": 401000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3887911175974018e-05, |
|
"loss": 1.2791, |
|
"step": 401500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3855393008073613e-05, |
|
"loss": 1.2691, |
|
"step": 402000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3822874840173208e-05, |
|
"loss": 1.278, |
|
"step": 402500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3790356672272803e-05, |
|
"loss": 1.2769, |
|
"step": 403000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.3757838504372394e-05, |
|
"loss": 1.282, |
|
"step": 403500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.3725320336471986e-05, |
|
"loss": 1.272, |
|
"step": 404000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.369280216857158e-05, |
|
"loss": 1.2694, |
|
"step": 404500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.3660284000671176e-05, |
|
"loss": 1.2681, |
|
"step": 405000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.362776583277077e-05, |
|
"loss": 1.2774, |
|
"step": 405500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 2.3595247664870366e-05, |
|
"loss": 1.2685, |
|
"step": 406000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3562729496969958e-05, |
|
"loss": 1.2703, |
|
"step": 406500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3530211329069553e-05, |
|
"loss": 1.277, |
|
"step": 407000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3497693161169144e-05, |
|
"loss": 1.2666, |
|
"step": 407500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.346517499326874e-05, |
|
"loss": 1.2656, |
|
"step": 408000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3432656825368335e-05, |
|
"loss": 1.2666, |
|
"step": 408500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 2.340013865746793e-05, |
|
"loss": 1.2639, |
|
"step": 409000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 2.3367620489567525e-05, |
|
"loss": 1.2686, |
|
"step": 409500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 2.3335102321667116e-05, |
|
"loss": 1.2689, |
|
"step": 410000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 2.330258415376671e-05, |
|
"loss": 1.2643, |
|
"step": 410500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 2.3270065985866303e-05, |
|
"loss": 1.2647, |
|
"step": 411000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.3237547817965898e-05, |
|
"loss": 1.2546, |
|
"step": 411500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.3205029650065493e-05, |
|
"loss": 1.2731, |
|
"step": 412000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.3172511482165088e-05, |
|
"loss": 1.2624, |
|
"step": 412500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.313999331426468e-05, |
|
"loss": 1.2578, |
|
"step": 413000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.3107475146364275e-05, |
|
"loss": 1.2609, |
|
"step": 413500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.307495697846387e-05, |
|
"loss": 1.2538, |
|
"step": 414000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.3042438810563465e-05, |
|
"loss": 1.258, |
|
"step": 414500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.3009920642663056e-05, |
|
"loss": 1.2518, |
|
"step": 415000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.297740247476265e-05, |
|
"loss": 1.2589, |
|
"step": 415500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 2.2944884306862243e-05, |
|
"loss": 1.2574, |
|
"step": 416000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2912366138961838e-05, |
|
"loss": 1.2568, |
|
"step": 416500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2879847971061433e-05, |
|
"loss": 1.2668, |
|
"step": 417000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2847329803161028e-05, |
|
"loss": 1.2583, |
|
"step": 417500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2814811635260623e-05, |
|
"loss": 1.2515, |
|
"step": 418000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 2.2782293467360215e-05, |
|
"loss": 1.2536, |
|
"step": 418500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.2749775299459806e-05, |
|
"loss": 1.2568, |
|
"step": 419000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.27172571315594e-05, |
|
"loss": 1.2572, |
|
"step": 419500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.2684738963658997e-05, |
|
"loss": 1.2556, |
|
"step": 420000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.265222079575859e-05, |
|
"loss": 1.2529, |
|
"step": 420500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.2619702627858187e-05, |
|
"loss": 1.2514, |
|
"step": 421000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.258718445995778e-05, |
|
"loss": 1.2515, |
|
"step": 421500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.2554666292057373e-05, |
|
"loss": 1.2482, |
|
"step": 422000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.2522148124156965e-05, |
|
"loss": 1.25, |
|
"step": 422500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.248962995625656e-05, |
|
"loss": 1.2518, |
|
"step": 423000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.2457111788356155e-05, |
|
"loss": 1.2528, |
|
"step": 423500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 2.242459362045575e-05, |
|
"loss": 1.2491, |
|
"step": 424000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.2392075452555345e-05, |
|
"loss": 1.2438, |
|
"step": 424500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.235955728465494e-05, |
|
"loss": 1.2541, |
|
"step": 425000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.2327039116754532e-05, |
|
"loss": 1.2482, |
|
"step": 425500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.2294520948854127e-05, |
|
"loss": 1.26, |
|
"step": 426000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.226200278095372e-05, |
|
"loss": 1.2491, |
|
"step": 426500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.2229484613053313e-05, |
|
"loss": 1.2455, |
|
"step": 427000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.219696644515291e-05, |
|
"loss": 1.2415, |
|
"step": 427500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.2164448277252504e-05, |
|
"loss": 1.2429, |
|
"step": 428000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.2131930109352095e-05, |
|
"loss": 1.2357, |
|
"step": 428500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 2.209941194145169e-05, |
|
"loss": 1.2391, |
|
"step": 429000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.2066893773551285e-05, |
|
"loss": 1.234, |
|
"step": 429500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.2034375605650877e-05, |
|
"loss": 1.2468, |
|
"step": 430000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.2001857437750472e-05, |
|
"loss": 1.2346, |
|
"step": 430500 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.1969339269850067e-05, |
|
"loss": 1.2501, |
|
"step": 431000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.1936821101949662e-05, |
|
"loss": 1.2325, |
|
"step": 431500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.1904302934049254e-05, |
|
"loss": 1.2407, |
|
"step": 432000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.187178476614885e-05, |
|
"loss": 1.2381, |
|
"step": 432500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.1839266598248444e-05, |
|
"loss": 1.239, |
|
"step": 433000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.1806748430348035e-05, |
|
"loss": 1.2338, |
|
"step": 433500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 2.177423026244763e-05, |
|
"loss": 1.2336, |
|
"step": 434000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.1741712094547225e-05, |
|
"loss": 1.2403, |
|
"step": 434500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.1709193926646817e-05, |
|
"loss": 1.2359, |
|
"step": 435000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.1676675758746412e-05, |
|
"loss": 1.2263, |
|
"step": 435500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.1644157590846007e-05, |
|
"loss": 1.2363, |
|
"step": 436000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.1611639422945602e-05, |
|
"loss": 1.2334, |
|
"step": 436500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.1579121255045194e-05, |
|
"loss": 1.2317, |
|
"step": 437000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.154660308714479e-05, |
|
"loss": 1.2421, |
|
"step": 437500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.151408491924438e-05, |
|
"loss": 1.2329, |
|
"step": 438000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.1481566751343975e-05, |
|
"loss": 1.2237, |
|
"step": 438500 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.144904858344357e-05, |
|
"loss": 1.2316, |
|
"step": 439000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.1416530415543166e-05, |
|
"loss": 1.2239, |
|
"step": 439500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.138401224764276e-05, |
|
"loss": 1.2344, |
|
"step": 440000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.1351494079742356e-05, |
|
"loss": 1.2243, |
|
"step": 440500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.1318975911841947e-05, |
|
"loss": 1.2329, |
|
"step": 441000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.128645774394154e-05, |
|
"loss": 1.2228, |
|
"step": 441500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.1253939576041134e-05, |
|
"loss": 1.2296, |
|
"step": 442000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.122142140814073e-05, |
|
"loss": 1.22, |
|
"step": 442500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.1188903240240324e-05, |
|
"loss": 1.2317, |
|
"step": 443000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.115638507233992e-05, |
|
"loss": 1.2299, |
|
"step": 443500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.1123866904439514e-05, |
|
"loss": 1.2361, |
|
"step": 444000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.1091348736539106e-05, |
|
"loss": 1.2282, |
|
"step": 444500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.1058830568638697e-05, |
|
"loss": 1.2263, |
|
"step": 445000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.1026312400738292e-05, |
|
"loss": 1.218, |
|
"step": 445500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.0993794232837887e-05, |
|
"loss": 1.228, |
|
"step": 446000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.0961276064937482e-05, |
|
"loss": 1.2248, |
|
"step": 446500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.0928757897037077e-05, |
|
"loss": 1.2275, |
|
"step": 447000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.089623972913667e-05, |
|
"loss": 1.2223, |
|
"step": 447500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.0863721561236264e-05, |
|
"loss": 1.2224, |
|
"step": 448000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.0831203393335856e-05, |
|
"loss": 1.2217, |
|
"step": 448500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.079868522543545e-05, |
|
"loss": 1.2208, |
|
"step": 449000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.0766167057535046e-05, |
|
"loss": 1.217, |
|
"step": 449500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.073364888963464e-05, |
|
"loss": 1.214, |
|
"step": 450000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.0701130721734233e-05, |
|
"loss": 1.2142, |
|
"step": 450500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.0668612553833828e-05, |
|
"loss": 1.2072, |
|
"step": 451000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.0636094385933423e-05, |
|
"loss": 1.2121, |
|
"step": 451500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.0603576218033018e-05, |
|
"loss": 1.2185, |
|
"step": 452000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.057105805013261e-05, |
|
"loss": 1.2129, |
|
"step": 452500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.0538539882232204e-05, |
|
"loss": 1.2091, |
|
"step": 453000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.05060217143318e-05, |
|
"loss": 1.2087, |
|
"step": 453500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.047350354643139e-05, |
|
"loss": 1.2247, |
|
"step": 454000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.0440985378530986e-05, |
|
"loss": 1.2131, |
|
"step": 454500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.040846721063058e-05, |
|
"loss": 1.2172, |
|
"step": 455000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.0375949042730176e-05, |
|
"loss": 1.2044, |
|
"step": 455500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.0343430874829768e-05, |
|
"loss": 1.2159, |
|
"step": 456000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.0310912706929363e-05, |
|
"loss": 1.2087, |
|
"step": 456500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.0278394539028954e-05, |
|
"loss": 1.2067, |
|
"step": 457000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.024587637112855e-05, |
|
"loss": 1.1992, |
|
"step": 457500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.0213358203228144e-05, |
|
"loss": 1.202, |
|
"step": 458000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.018084003532774e-05, |
|
"loss": 1.2078, |
|
"step": 458500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.0148321867427335e-05, |
|
"loss": 1.2087, |
|
"step": 459000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.0115803699526926e-05, |
|
"loss": 1.2119, |
|
"step": 459500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0083285531626518e-05, |
|
"loss": 1.2056, |
|
"step": 460000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0050767363726113e-05, |
|
"loss": 1.1988, |
|
"step": 460500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0018249195825708e-05, |
|
"loss": 1.2137, |
|
"step": 461000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.9985731027925303e-05, |
|
"loss": 1.2125, |
|
"step": 461500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.9953212860024898e-05, |
|
"loss": 1.2043, |
|
"step": 462000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.9920694692124493e-05, |
|
"loss": 1.2074, |
|
"step": 462500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.9888176524224085e-05, |
|
"loss": 1.2036, |
|
"step": 463000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.985565835632368e-05, |
|
"loss": 1.2005, |
|
"step": 463500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.982314018842327e-05, |
|
"loss": 1.2057, |
|
"step": 464000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.9790622020522866e-05, |
|
"loss": 1.2043, |
|
"step": 464500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.975810385262246e-05, |
|
"loss": 1.2036, |
|
"step": 465000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9725585684722056e-05, |
|
"loss": 1.2022, |
|
"step": 465500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.969306751682165e-05, |
|
"loss": 1.2003, |
|
"step": 466000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9660549348921243e-05, |
|
"loss": 1.1992, |
|
"step": 466500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9628031181020838e-05, |
|
"loss": 1.2011, |
|
"step": 467000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.959551301312043e-05, |
|
"loss": 1.1997, |
|
"step": 467500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.9562994845220025e-05, |
|
"loss": 1.2033, |
|
"step": 468000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.953047667731962e-05, |
|
"loss": 1.199, |
|
"step": 468500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.9497958509419215e-05, |
|
"loss": 1.1972, |
|
"step": 469000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.9465440341518807e-05, |
|
"loss": 1.1979, |
|
"step": 469500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.94329221736184e-05, |
|
"loss": 1.2027, |
|
"step": 470000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.9400404005717997e-05, |
|
"loss": 1.1959, |
|
"step": 470500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.9367885837817588e-05, |
|
"loss": 1.1962, |
|
"step": 471000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.9335367669917183e-05, |
|
"loss": 1.1978, |
|
"step": 471500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.930284950201678e-05, |
|
"loss": 1.2043, |
|
"step": 472000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.927033133411637e-05, |
|
"loss": 1.1901, |
|
"step": 472500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.9237813166215965e-05, |
|
"loss": 1.2003, |
|
"step": 473000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.920529499831556e-05, |
|
"loss": 1.1963, |
|
"step": 473500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.9172776830415155e-05, |
|
"loss": 1.197, |
|
"step": 474000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.914025866251475e-05, |
|
"loss": 1.1969, |
|
"step": 474500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.9107740494614342e-05, |
|
"loss": 1.1896, |
|
"step": 475000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.9075222326713933e-05, |
|
"loss": 1.1928, |
|
"step": 475500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.904270415881353e-05, |
|
"loss": 1.1887, |
|
"step": 476000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.9010185990913123e-05, |
|
"loss": 1.1895, |
|
"step": 476500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.897766782301272e-05, |
|
"loss": 1.1926, |
|
"step": 477000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.8945149655112314e-05, |
|
"loss": 1.1861, |
|
"step": 477500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.891263148721191e-05, |
|
"loss": 1.2007, |
|
"step": 478000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.88801133193115e-05, |
|
"loss": 1.1893, |
|
"step": 478500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8847595151411092e-05, |
|
"loss": 1.1894, |
|
"step": 479000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8815076983510687e-05, |
|
"loss": 1.1954, |
|
"step": 479500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8782558815610282e-05, |
|
"loss": 1.186, |
|
"step": 480000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8750040647709877e-05, |
|
"loss": 1.1876, |
|
"step": 480500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8717522479809472e-05, |
|
"loss": 1.19, |
|
"step": 481000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8685004311909067e-05, |
|
"loss": 1.1834, |
|
"step": 481500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.865248614400866e-05, |
|
"loss": 1.1836, |
|
"step": 482000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.861996797610825e-05, |
|
"loss": 1.1895, |
|
"step": 482500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8587449808207845e-05, |
|
"loss": 1.1894, |
|
"step": 483000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.855493164030744e-05, |
|
"loss": 1.1848, |
|
"step": 483500 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.8522413472407035e-05, |
|
"loss": 1.1855, |
|
"step": 484000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.848989530450663e-05, |
|
"loss": 1.1856, |
|
"step": 484500 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.8457377136606222e-05, |
|
"loss": 1.1802, |
|
"step": 485000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.8424858968705817e-05, |
|
"loss": 1.1805, |
|
"step": 485500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8392340800805412e-05, |
|
"loss": 1.1837, |
|
"step": 486000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8359822632905004e-05, |
|
"loss": 1.1772, |
|
"step": 486500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.83273044650046e-05, |
|
"loss": 1.1849, |
|
"step": 487000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8294786297104194e-05, |
|
"loss": 1.1802, |
|
"step": 487500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.826226812920379e-05, |
|
"loss": 1.1749, |
|
"step": 488000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.822974996130338e-05, |
|
"loss": 1.1786, |
|
"step": 488500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.8197231793402976e-05, |
|
"loss": 1.182, |
|
"step": 489000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.816471362550257e-05, |
|
"loss": 1.1811, |
|
"step": 489500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.8132195457602162e-05, |
|
"loss": 1.1782, |
|
"step": 490000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.8099677289701757e-05, |
|
"loss": 1.1785, |
|
"step": 490500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.8067159121801352e-05, |
|
"loss": 1.1802, |
|
"step": 491000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.8034640953900944e-05, |
|
"loss": 1.1783, |
|
"step": 491500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.800212278600054e-05, |
|
"loss": 1.1862, |
|
"step": 492000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.7969604618100134e-05, |
|
"loss": 1.1749, |
|
"step": 492500 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.793708645019973e-05, |
|
"loss": 1.1816, |
|
"step": 493000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.790456828229932e-05, |
|
"loss": 1.175, |
|
"step": 493500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.7872050114398916e-05, |
|
"loss": 1.178, |
|
"step": 494000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.7839531946498507e-05, |
|
"loss": 1.1739, |
|
"step": 494500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.7807013778598102e-05, |
|
"loss": 1.1811, |
|
"step": 495000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.7774495610697697e-05, |
|
"loss": 1.1812, |
|
"step": 495500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.7741977442797292e-05, |
|
"loss": 1.1772, |
|
"step": 496000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.7709459274896887e-05, |
|
"loss": 1.1761, |
|
"step": 496500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.767694110699648e-05, |
|
"loss": 1.1731, |
|
"step": 497000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.764442293909607e-05, |
|
"loss": 1.1655, |
|
"step": 497500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.7611904771195666e-05, |
|
"loss": 1.1715, |
|
"step": 498000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.757938660329526e-05, |
|
"loss": 1.1777, |
|
"step": 498500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.7546868435394856e-05, |
|
"loss": 1.172, |
|
"step": 499000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.751435026749445e-05, |
|
"loss": 1.1716, |
|
"step": 499500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.7481832099594046e-05, |
|
"loss": 1.1733, |
|
"step": 500000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.744931393169364e-05, |
|
"loss": 1.1745, |
|
"step": 500500 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.7416795763793233e-05, |
|
"loss": 1.1722, |
|
"step": 501000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.7384277595892824e-05, |
|
"loss": 1.1628, |
|
"step": 501500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.735175942799242e-05, |
|
"loss": 1.1659, |
|
"step": 502000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.7319241260092014e-05, |
|
"loss": 1.1742, |
|
"step": 502500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.728672309219161e-05, |
|
"loss": 1.1674, |
|
"step": 503000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.7254204924291204e-05, |
|
"loss": 1.1704, |
|
"step": 503500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.7221686756390796e-05, |
|
"loss": 1.164, |
|
"step": 504000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.718916858849039e-05, |
|
"loss": 1.1759, |
|
"step": 504500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.7156650420589983e-05, |
|
"loss": 1.1658, |
|
"step": 505000 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.7124132252689578e-05, |
|
"loss": 1.164, |
|
"step": 505500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.7091614084789173e-05, |
|
"loss": 1.1667, |
|
"step": 506000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.7059095916888768e-05, |
|
"loss": 1.1729, |
|
"step": 506500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.702657774898836e-05, |
|
"loss": 1.1668, |
|
"step": 507000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.6994059581087954e-05, |
|
"loss": 1.1709, |
|
"step": 507500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.696154141318755e-05, |
|
"loss": 1.1647, |
|
"step": 508000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.692902324528714e-05, |
|
"loss": 1.1599, |
|
"step": 508500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6896505077386736e-05, |
|
"loss": 1.1652, |
|
"step": 509000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.686398690948633e-05, |
|
"loss": 1.1587, |
|
"step": 509500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6831468741585926e-05, |
|
"loss": 1.1691, |
|
"step": 510000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6798950573685518e-05, |
|
"loss": 1.1643, |
|
"step": 510500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6766432405785113e-05, |
|
"loss": 1.1606, |
|
"step": 511000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.6733914237884708e-05, |
|
"loss": 1.1571, |
|
"step": 511500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.6701396069984303e-05, |
|
"loss": 1.1628, |
|
"step": 512000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.6668877902083895e-05, |
|
"loss": 1.1619, |
|
"step": 512500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.663635973418349e-05, |
|
"loss": 1.1597, |
|
"step": 513000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.660384156628308e-05, |
|
"loss": 1.1605, |
|
"step": 513500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.6571323398382676e-05, |
|
"loss": 1.1593, |
|
"step": 514000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.653880523048227e-05, |
|
"loss": 1.167, |
|
"step": 514500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.6506287062581866e-05, |
|
"loss": 1.1579, |
|
"step": 515000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.647376889468146e-05, |
|
"loss": 1.1611, |
|
"step": 515500 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 1.6441250726781053e-05, |
|
"loss": 1.1498, |
|
"step": 516000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.6408732558880645e-05, |
|
"loss": 1.1566, |
|
"step": 516500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.637621439098024e-05, |
|
"loss": 1.1596, |
|
"step": 517000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.6343696223079835e-05, |
|
"loss": 1.1514, |
|
"step": 517500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.631117805517943e-05, |
|
"loss": 1.1478, |
|
"step": 518000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.6278659887279025e-05, |
|
"loss": 1.1547, |
|
"step": 518500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.624614171937862e-05, |
|
"loss": 1.1621, |
|
"step": 519000 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.621362355147821e-05, |
|
"loss": 1.1557, |
|
"step": 519500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.6181105383577803e-05, |
|
"loss": 1.1481, |
|
"step": 520000 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.6148587215677398e-05, |
|
"loss": 1.1573, |
|
"step": 520500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.6116069047776993e-05, |
|
"loss": 1.1651, |
|
"step": 521000 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.608355087987659e-05, |
|
"loss": 1.1524, |
|
"step": 521500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.6051032711976183e-05, |
|
"loss": 1.151, |
|
"step": 522000 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.601851454407578e-05, |
|
"loss": 1.1594, |
|
"step": 522500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.598599637617537e-05, |
|
"loss": 1.1506, |
|
"step": 523000 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.5953478208274965e-05, |
|
"loss": 1.1606, |
|
"step": 523500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.5920960040374557e-05, |
|
"loss": 1.1546, |
|
"step": 524000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5888441872474152e-05, |
|
"loss": 1.1559, |
|
"step": 524500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5855923704573747e-05, |
|
"loss": 1.1504, |
|
"step": 525000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5823405536673342e-05, |
|
"loss": 1.1538, |
|
"step": 525500 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.5790887368772933e-05, |
|
"loss": 1.1498, |
|
"step": 526000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.575836920087253e-05, |
|
"loss": 1.1576, |
|
"step": 526500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5725851032972124e-05, |
|
"loss": 1.1574, |
|
"step": 527000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5693332865071715e-05, |
|
"loss": 1.1434, |
|
"step": 527500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.566081469717131e-05, |
|
"loss": 1.1451, |
|
"step": 528000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5628296529270905e-05, |
|
"loss": 1.1555, |
|
"step": 528500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5595778361370497e-05, |
|
"loss": 1.1537, |
|
"step": 529000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5563260193470092e-05, |
|
"loss": 1.153, |
|
"step": 529500 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5530742025569687e-05, |
|
"loss": 1.1526, |
|
"step": 530000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5498223857669282e-05, |
|
"loss": 1.1589, |
|
"step": 530500 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.5465705689768874e-05, |
|
"loss": 1.1471, |
|
"step": 531000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.543318752186847e-05, |
|
"loss": 1.1536, |
|
"step": 531500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.540066935396806e-05, |
|
"loss": 1.1498, |
|
"step": 532000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.5368151186067655e-05, |
|
"loss": 1.1515, |
|
"step": 532500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.533563301816725e-05, |
|
"loss": 1.1455, |
|
"step": 533000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.5303114850266845e-05, |
|
"loss": 1.1417, |
|
"step": 533500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.527059668236644e-05, |
|
"loss": 1.1472, |
|
"step": 534000 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5238078514466034e-05, |
|
"loss": 1.1419, |
|
"step": 534500 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5205560346565629e-05, |
|
"loss": 1.1446, |
|
"step": 535000 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.517304217866522e-05, |
|
"loss": 1.1467, |
|
"step": 535500 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5140524010764814e-05, |
|
"loss": 1.1511, |
|
"step": 536000 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.5108005842864409e-05, |
|
"loss": 1.1476, |
|
"step": 536500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.5075487674964004e-05, |
|
"loss": 1.1456, |
|
"step": 537000 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.5042969507063597e-05, |
|
"loss": 1.146, |
|
"step": 537500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.5010451339163192e-05, |
|
"loss": 1.1491, |
|
"step": 538000 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.4977933171262784e-05, |
|
"loss": 1.1476, |
|
"step": 538500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.4945415003362379e-05, |
|
"loss": 1.1413, |
|
"step": 539000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4912896835461972e-05, |
|
"loss": 1.1439, |
|
"step": 539500 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4880378667561567e-05, |
|
"loss": 1.1463, |
|
"step": 540000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4847860499661162e-05, |
|
"loss": 1.1464, |
|
"step": 540500 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4815342331760756e-05, |
|
"loss": 1.1368, |
|
"step": 541000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4782824163860349e-05, |
|
"loss": 1.1439, |
|
"step": 541500 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.4750305995959942e-05, |
|
"loss": 1.1426, |
|
"step": 542000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.4717787828059537e-05, |
|
"loss": 1.1357, |
|
"step": 542500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.4685269660159132e-05, |
|
"loss": 1.1489, |
|
"step": 543000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.4652751492258726e-05, |
|
"loss": 1.1383, |
|
"step": 543500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.462023332435832e-05, |
|
"loss": 1.1401, |
|
"step": 544000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.4587715156457916e-05, |
|
"loss": 1.1387, |
|
"step": 544500 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.4555196988557507e-05, |
|
"loss": 1.1428, |
|
"step": 545000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.45226788206571e-05, |
|
"loss": 1.1404, |
|
"step": 545500 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.4490160652756696e-05, |
|
"loss": 1.1366, |
|
"step": 546000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.445764248485629e-05, |
|
"loss": 1.1338, |
|
"step": 546500 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.4425124316955884e-05, |
|
"loss": 1.1323, |
|
"step": 547000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.439260614905548e-05, |
|
"loss": 1.1413, |
|
"step": 547500 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.4360087981155071e-05, |
|
"loss": 1.1357, |
|
"step": 548000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.4327569813254666e-05, |
|
"loss": 1.1352, |
|
"step": 548500 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.429505164535426e-05, |
|
"loss": 1.1406, |
|
"step": 549000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.4262533477453854e-05, |
|
"loss": 1.1359, |
|
"step": 549500 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.423001530955345e-05, |
|
"loss": 1.1319, |
|
"step": 550000 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.4197497141653043e-05, |
|
"loss": 1.1335, |
|
"step": 550500 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.4164978973752634e-05, |
|
"loss": 1.1293, |
|
"step": 551000 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.413246080585223e-05, |
|
"loss": 1.1322, |
|
"step": 551500 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.4099942637951824e-05, |
|
"loss": 1.135, |
|
"step": 552000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.4067424470051418e-05, |
|
"loss": 1.1328, |
|
"step": 552500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.4034906302151013e-05, |
|
"loss": 1.1359, |
|
"step": 553000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.4002388134250608e-05, |
|
"loss": 1.1387, |
|
"step": 553500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.39698699663502e-05, |
|
"loss": 1.1368, |
|
"step": 554000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.3937351798449794e-05, |
|
"loss": 1.1342, |
|
"step": 554500 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3904833630549388e-05, |
|
"loss": 1.1309, |
|
"step": 555000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3872315462648983e-05, |
|
"loss": 1.1427, |
|
"step": 555500 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3839797294748578e-05, |
|
"loss": 1.1333, |
|
"step": 556000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3807279126848171e-05, |
|
"loss": 1.1328, |
|
"step": 556500 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.3774760958947766e-05, |
|
"loss": 1.1394, |
|
"step": 557000 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.3742242791047358e-05, |
|
"loss": 1.1368, |
|
"step": 557500 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.3709724623146953e-05, |
|
"loss": 1.1344, |
|
"step": 558000 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.3677206455246546e-05, |
|
"loss": 1.1375, |
|
"step": 558500 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.3644688287346141e-05, |
|
"loss": 1.1308, |
|
"step": 559000 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.3612170119445736e-05, |
|
"loss": 1.1339, |
|
"step": 559500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.357965195154533e-05, |
|
"loss": 1.1252, |
|
"step": 560000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.3547133783644921e-05, |
|
"loss": 1.1296, |
|
"step": 560500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.3514615615744516e-05, |
|
"loss": 1.1386, |
|
"step": 561000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.3482097447844111e-05, |
|
"loss": 1.1379, |
|
"step": 561500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.3449579279943705e-05, |
|
"loss": 1.1262, |
|
"step": 562000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.34170611120433e-05, |
|
"loss": 1.1237, |
|
"step": 562500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.3384542944142895e-05, |
|
"loss": 1.1263, |
|
"step": 563000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.3352024776242486e-05, |
|
"loss": 1.1255, |
|
"step": 563500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.331950660834208e-05, |
|
"loss": 1.1362, |
|
"step": 564000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.3286988440441675e-05, |
|
"loss": 1.1202, |
|
"step": 564500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.325447027254127e-05, |
|
"loss": 1.1261, |
|
"step": 565000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.3221952104640865e-05, |
|
"loss": 1.128, |
|
"step": 565500 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.3189433936740458e-05, |
|
"loss": 1.1336, |
|
"step": 566000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.3156915768840053e-05, |
|
"loss": 1.1263, |
|
"step": 566500 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.3124397600939645e-05, |
|
"loss": 1.1338, |
|
"step": 567000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.309187943303924e-05, |
|
"loss": 1.1248, |
|
"step": 567500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.3059361265138833e-05, |
|
"loss": 1.1261, |
|
"step": 568000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.3026843097238428e-05, |
|
"loss": 1.1328, |
|
"step": 568500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.2994324929338023e-05, |
|
"loss": 1.1213, |
|
"step": 569000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.2961806761437617e-05, |
|
"loss": 1.1279, |
|
"step": 569500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.2929288593537208e-05, |
|
"loss": 1.1243, |
|
"step": 570000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2896770425636803e-05, |
|
"loss": 1.1201, |
|
"step": 570500 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2864252257736398e-05, |
|
"loss": 1.1245, |
|
"step": 571000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2831734089835992e-05, |
|
"loss": 1.1251, |
|
"step": 571500 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2799215921935587e-05, |
|
"loss": 1.1285, |
|
"step": 572000 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.2766697754035182e-05, |
|
"loss": 1.1276, |
|
"step": 572500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2734179586134773e-05, |
|
"loss": 1.1199, |
|
"step": 573000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2701661418234367e-05, |
|
"loss": 1.1215, |
|
"step": 573500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2669143250333962e-05, |
|
"loss": 1.1203, |
|
"step": 574000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.2636625082433557e-05, |
|
"loss": 1.1168, |
|
"step": 574500 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.260410691453315e-05, |
|
"loss": 1.1223, |
|
"step": 575000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.2571588746632745e-05, |
|
"loss": 1.117, |
|
"step": 575500 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.2539070578732337e-05, |
|
"loss": 1.1251, |
|
"step": 576000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.2506552410831932e-05, |
|
"loss": 1.1169, |
|
"step": 576500 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.2474034242931527e-05, |
|
"loss": 1.1183, |
|
"step": 577000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.244151607503112e-05, |
|
"loss": 1.1251, |
|
"step": 577500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.2408997907130715e-05, |
|
"loss": 1.1188, |
|
"step": 578000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.2376479739230309e-05, |
|
"loss": 1.1198, |
|
"step": 578500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.2343961571329902e-05, |
|
"loss": 1.1304, |
|
"step": 579000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.2311443403429497e-05, |
|
"loss": 1.121, |
|
"step": 579500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.227892523552909e-05, |
|
"loss": 1.1215, |
|
"step": 580000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.2246407067628685e-05, |
|
"loss": 1.1213, |
|
"step": 580500 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.2213888899728279e-05, |
|
"loss": 1.1229, |
|
"step": 581000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.2181370731827872e-05, |
|
"loss": 1.1167, |
|
"step": 581500 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.2148852563927467e-05, |
|
"loss": 1.1308, |
|
"step": 582000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.211633439602706e-05, |
|
"loss": 1.1189, |
|
"step": 582500 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.2083816228126654e-05, |
|
"loss": 1.1161, |
|
"step": 583000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.2051298060226249e-05, |
|
"loss": 1.1188, |
|
"step": 583500 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.2018779892325844e-05, |
|
"loss": 1.1157, |
|
"step": 584000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.1986261724425437e-05, |
|
"loss": 1.1145, |
|
"step": 584500 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.195374355652503e-05, |
|
"loss": 1.1179, |
|
"step": 585000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.1921225388624625e-05, |
|
"loss": 1.1155, |
|
"step": 585500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.188870722072422e-05, |
|
"loss": 1.1277, |
|
"step": 586000 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.1856189052823812e-05, |
|
"loss": 1.1162, |
|
"step": 586500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.1823670884923407e-05, |
|
"loss": 1.122, |
|
"step": 587000 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.1791152717023002e-05, |
|
"loss": 1.1188, |
|
"step": 587500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.1758634549122596e-05, |
|
"loss": 1.1187, |
|
"step": 588000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1726116381222189e-05, |
|
"loss": 1.108, |
|
"step": 588500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1693598213321784e-05, |
|
"loss": 1.1199, |
|
"step": 589000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1661080045421377e-05, |
|
"loss": 1.1125, |
|
"step": 589500 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1628561877520972e-05, |
|
"loss": 1.117, |
|
"step": 590000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1596043709620566e-05, |
|
"loss": 1.1159, |
|
"step": 590500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.1563525541720159e-05, |
|
"loss": 1.1172, |
|
"step": 591000 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.1531007373819754e-05, |
|
"loss": 1.1136, |
|
"step": 591500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.1498489205919347e-05, |
|
"loss": 1.1146, |
|
"step": 592000 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.146597103801894e-05, |
|
"loss": 1.111, |
|
"step": 592500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.1433452870118536e-05, |
|
"loss": 1.1129, |
|
"step": 593000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.140093470221813e-05, |
|
"loss": 1.1069, |
|
"step": 593500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.1368416534317722e-05, |
|
"loss": 1.1168, |
|
"step": 594000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.1335898366417317e-05, |
|
"loss": 1.1157, |
|
"step": 594500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.1303380198516912e-05, |
|
"loss": 1.1111, |
|
"step": 595000 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.1270862030616506e-05, |
|
"loss": 1.1084, |
|
"step": 595500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.1238343862716099e-05, |
|
"loss": 1.1132, |
|
"step": 596000 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.1205825694815694e-05, |
|
"loss": 1.1139, |
|
"step": 596500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.117330752691529e-05, |
|
"loss": 1.11, |
|
"step": 597000 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.1140789359014883e-05, |
|
"loss": 1.1162, |
|
"step": 597500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.1108271191114476e-05, |
|
"loss": 1.11, |
|
"step": 598000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1075753023214071e-05, |
|
"loss": 1.1106, |
|
"step": 598500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1043234855313664e-05, |
|
"loss": 1.1054, |
|
"step": 599000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1010716687413258e-05, |
|
"loss": 1.1091, |
|
"step": 599500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.0978198519512853e-05, |
|
"loss": 1.1067, |
|
"step": 600000 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.0945680351612446e-05, |
|
"loss": 1.1033, |
|
"step": 600500 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.0913162183712041e-05, |
|
"loss": 1.1133, |
|
"step": 601000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.0880644015811634e-05, |
|
"loss": 1.1101, |
|
"step": 601500 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.0848125847911228e-05, |
|
"loss": 1.1105, |
|
"step": 602000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.0815607680010823e-05, |
|
"loss": 1.1083, |
|
"step": 602500 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.0783089512110418e-05, |
|
"loss": 1.1077, |
|
"step": 603000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.075057134421001e-05, |
|
"loss": 1.1051, |
|
"step": 603500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0718053176309604e-05, |
|
"loss": 1.1064, |
|
"step": 604000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.06855350084092e-05, |
|
"loss": 1.1054, |
|
"step": 604500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0653016840508793e-05, |
|
"loss": 1.1107, |
|
"step": 605000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0620498672608386e-05, |
|
"loss": 1.1031, |
|
"step": 605500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0587980504707981e-05, |
|
"loss": 1.1097, |
|
"step": 606000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0555462336807576e-05, |
|
"loss": 1.1074, |
|
"step": 606500 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0522944168907168e-05, |
|
"loss": 1.1144, |
|
"step": 607000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0490426001006763e-05, |
|
"loss": 1.1106, |
|
"step": 607500 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0457907833106358e-05, |
|
"loss": 1.1091, |
|
"step": 608000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0425389665205951e-05, |
|
"loss": 1.1057, |
|
"step": 608500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0392871497305545e-05, |
|
"loss": 1.1066, |
|
"step": 609000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.036035332940514e-05, |
|
"loss": 1.108, |
|
"step": 609500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0327835161504733e-05, |
|
"loss": 1.1104, |
|
"step": 610000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0295316993604328e-05, |
|
"loss": 1.1158, |
|
"step": 610500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0262798825703921e-05, |
|
"loss": 1.109, |
|
"step": 611000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0230280657803515e-05, |
|
"loss": 1.1011, |
|
"step": 611500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.019776248990311e-05, |
|
"loss": 1.0989, |
|
"step": 612000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0165244322002703e-05, |
|
"loss": 1.109, |
|
"step": 612500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0132726154102296e-05, |
|
"loss": 1.0999, |
|
"step": 613000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0100207986201891e-05, |
|
"loss": 1.1091, |
|
"step": 613500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.0067689818301486e-05, |
|
"loss": 1.1033, |
|
"step": 614000 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.003517165040108e-05, |
|
"loss": 1.1087, |
|
"step": 614500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.0002653482500673e-05, |
|
"loss": 1.0964, |
|
"step": 615000 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.970135314600268e-06, |
|
"loss": 1.0946, |
|
"step": 615500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.937617146699861e-06, |
|
"loss": 1.0994, |
|
"step": 616000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.905098978799455e-06, |
|
"loss": 1.1078, |
|
"step": 616500 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.87258081089905e-06, |
|
"loss": 1.101, |
|
"step": 617000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.840062642998645e-06, |
|
"loss": 1.1085, |
|
"step": 617500 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.807544475098238e-06, |
|
"loss": 1.1083, |
|
"step": 618000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.775026307197832e-06, |
|
"loss": 1.1069, |
|
"step": 618500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.742508139297427e-06, |
|
"loss": 1.1041, |
|
"step": 619000 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.70998997139702e-06, |
|
"loss": 1.1056, |
|
"step": 619500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.677471803496615e-06, |
|
"loss": 1.1013, |
|
"step": 620000 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.644953635596208e-06, |
|
"loss": 1.1007, |
|
"step": 620500 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.612435467695802e-06, |
|
"loss": 1.1092, |
|
"step": 621000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.579917299795397e-06, |
|
"loss": 1.104, |
|
"step": 621500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.54739913189499e-06, |
|
"loss": 1.098, |
|
"step": 622000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.514880963994583e-06, |
|
"loss": 1.0999, |
|
"step": 622500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.482362796094178e-06, |
|
"loss": 1.0956, |
|
"step": 623000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.449844628193773e-06, |
|
"loss": 1.1052, |
|
"step": 623500 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.417326460293365e-06, |
|
"loss": 1.1057, |
|
"step": 624000 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 9.38480829239296e-06, |
|
"loss": 1.0962, |
|
"step": 624500 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 9.352290124492555e-06, |
|
"loss": 1.103, |
|
"step": 625000 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 9.319771956592148e-06, |
|
"loss": 1.1092, |
|
"step": 625500 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 9.287253788691742e-06, |
|
"loss": 1.0881, |
|
"step": 626000 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 9.254735620791337e-06, |
|
"loss": 1.0999, |
|
"step": 626500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.22221745289093e-06, |
|
"loss": 1.1054, |
|
"step": 627000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.189699284990525e-06, |
|
"loss": 1.0941, |
|
"step": 627500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.157181117090119e-06, |
|
"loss": 1.1027, |
|
"step": 628000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.124662949189712e-06, |
|
"loss": 1.0986, |
|
"step": 628500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 9.092144781289307e-06, |
|
"loss": 1.0928, |
|
"step": 629000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 9.0596266133889e-06, |
|
"loss": 1.0906, |
|
"step": 629500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 9.027108445488495e-06, |
|
"loss": 1.0925, |
|
"step": 630000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.994590277588089e-06, |
|
"loss": 1.0926, |
|
"step": 630500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.962072109687684e-06, |
|
"loss": 1.0927, |
|
"step": 631000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.929553941787277e-06, |
|
"loss": 1.1047, |
|
"step": 631500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.89703577388687e-06, |
|
"loss": 1.0932, |
|
"step": 632000 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.864517605986465e-06, |
|
"loss": 1.099, |
|
"step": 632500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.83199943808606e-06, |
|
"loss": 1.0946, |
|
"step": 633000 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.799481270185652e-06, |
|
"loss": 1.1017, |
|
"step": 633500 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 8.766963102285247e-06, |
|
"loss": 1.1008, |
|
"step": 634000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.734444934384842e-06, |
|
"loss": 1.0916, |
|
"step": 634500 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.701926766484435e-06, |
|
"loss": 1.0949, |
|
"step": 635000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.669408598584029e-06, |
|
"loss": 1.0958, |
|
"step": 635500 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.636890430683624e-06, |
|
"loss": 1.1001, |
|
"step": 636000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 8.604372262783217e-06, |
|
"loss": 1.0931, |
|
"step": 636500 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.57185409488281e-06, |
|
"loss": 1.0975, |
|
"step": 637000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.539335926982406e-06, |
|
"loss": 1.1003, |
|
"step": 637500 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.506817759081999e-06, |
|
"loss": 1.0929, |
|
"step": 638000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.474299591181594e-06, |
|
"loss": 1.0987, |
|
"step": 638500 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 8.441781423281187e-06, |
|
"loss": 1.0915, |
|
"step": 639000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.40926325538078e-06, |
|
"loss": 1.0907, |
|
"step": 639500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.376745087480376e-06, |
|
"loss": 1.0959, |
|
"step": 640000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.34422691957997e-06, |
|
"loss": 1.094, |
|
"step": 640500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.311708751679564e-06, |
|
"loss": 1.0932, |
|
"step": 641000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.279190583779157e-06, |
|
"loss": 1.0887, |
|
"step": 641500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.246672415878752e-06, |
|
"loss": 1.0996, |
|
"step": 642000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.214154247978346e-06, |
|
"loss": 1.0919, |
|
"step": 642500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.181636080077939e-06, |
|
"loss": 1.0899, |
|
"step": 643000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.149117912177534e-06, |
|
"loss": 1.0899, |
|
"step": 643500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.116599744277129e-06, |
|
"loss": 1.0939, |
|
"step": 644000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 8.084081576376722e-06, |
|
"loss": 1.0915, |
|
"step": 644500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 8.051563408476316e-06, |
|
"loss": 1.0901, |
|
"step": 645000 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 8.01904524057591e-06, |
|
"loss": 1.0844, |
|
"step": 645500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.986527072675504e-06, |
|
"loss": 1.0808, |
|
"step": 646000 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.954008904775098e-06, |
|
"loss": 1.0843, |
|
"step": 646500 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.921490736874693e-06, |
|
"loss": 1.0931, |
|
"step": 647000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.888972568974286e-06, |
|
"loss": 1.0874, |
|
"step": 647500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.856454401073881e-06, |
|
"loss": 1.0893, |
|
"step": 648000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.823936233173474e-06, |
|
"loss": 1.0879, |
|
"step": 648500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.791418065273068e-06, |
|
"loss": 1.1044, |
|
"step": 649000 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.758899897372663e-06, |
|
"loss": 1.0898, |
|
"step": 649500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.726381729472256e-06, |
|
"loss": 1.0876, |
|
"step": 650000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.69386356157185e-06, |
|
"loss": 1.0968, |
|
"step": 650500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.661345393671444e-06, |
|
"loss": 1.087, |
|
"step": 651000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.6288272257710385e-06, |
|
"loss": 1.0867, |
|
"step": 651500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.5963090578706336e-06, |
|
"loss": 1.0834, |
|
"step": 652000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.563790889970226e-06, |
|
"loss": 1.0882, |
|
"step": 652500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.531272722069821e-06, |
|
"loss": 1.0883, |
|
"step": 653000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.498754554169415e-06, |
|
"loss": 1.0811, |
|
"step": 653500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.466236386269009e-06, |
|
"loss": 1.0852, |
|
"step": 654000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.433718218368603e-06, |
|
"loss": 1.0924, |
|
"step": 654500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.401200050468198e-06, |
|
"loss": 1.0938, |
|
"step": 655000 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.36868188256779e-06, |
|
"loss": 1.0888, |
|
"step": 655500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.336163714667385e-06, |
|
"loss": 1.0905, |
|
"step": 656000 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.3036455467669795e-06, |
|
"loss": 1.0882, |
|
"step": 656500 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.271127378866573e-06, |
|
"loss": 1.0871, |
|
"step": 657000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.238609210966167e-06, |
|
"loss": 1.0855, |
|
"step": 657500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.206091043065761e-06, |
|
"loss": 1.0888, |
|
"step": 658000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.173572875165355e-06, |
|
"loss": 1.0819, |
|
"step": 658500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.141054707264949e-06, |
|
"loss": 1.089, |
|
"step": 659000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.108536539364544e-06, |
|
"loss": 1.0899, |
|
"step": 659500 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.076018371464137e-06, |
|
"loss": 1.0841, |
|
"step": 660000 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.043500203563731e-06, |
|
"loss": 1.0852, |
|
"step": 660500 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.0109820356633255e-06, |
|
"loss": 1.0779, |
|
"step": 661000 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 6.978463867762919e-06, |
|
"loss": 1.084, |
|
"step": 661500 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 6.945945699862513e-06, |
|
"loss": 1.0835, |
|
"step": 662000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.913427531962108e-06, |
|
"loss": 1.0893, |
|
"step": 662500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.880909364061702e-06, |
|
"loss": 1.0897, |
|
"step": 663000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.848391196161296e-06, |
|
"loss": 1.0888, |
|
"step": 663500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.81587302826089e-06, |
|
"loss": 1.0907, |
|
"step": 664000 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.783354860360484e-06, |
|
"loss": 1.0827, |
|
"step": 664500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 6.750836692460077e-06, |
|
"loss": 1.0874, |
|
"step": 665000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 6.718318524559672e-06, |
|
"loss": 1.0859, |
|
"step": 665500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 6.6858003566592665e-06, |
|
"loss": 1.0768, |
|
"step": 666000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 6.65328218875886e-06, |
|
"loss": 1.084, |
|
"step": 666500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 6.620764020858454e-06, |
|
"loss": 1.0814, |
|
"step": 667000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 6.588245852958048e-06, |
|
"loss": 1.0858, |
|
"step": 667500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.5557276850576416e-06, |
|
"loss": 1.0886, |
|
"step": 668000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.523209517157236e-06, |
|
"loss": 1.077, |
|
"step": 668500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.490691349256831e-06, |
|
"loss": 1.0803, |
|
"step": 669000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.458173181356423e-06, |
|
"loss": 1.0757, |
|
"step": 669500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 6.425655013456018e-06, |
|
"loss": 1.0778, |
|
"step": 670000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.3931368455556125e-06, |
|
"loss": 1.0805, |
|
"step": 670500 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.360618677655206e-06, |
|
"loss": 1.0889, |
|
"step": 671000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.3281005097548e-06, |
|
"loss": 1.0801, |
|
"step": 671500 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.295582341854395e-06, |
|
"loss": 1.0801, |
|
"step": 672000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.2630641739539876e-06, |
|
"loss": 1.0843, |
|
"step": 672500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.230546006053583e-06, |
|
"loss": 1.0856, |
|
"step": 673000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.198027838153177e-06, |
|
"loss": 1.0809, |
|
"step": 673500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.16550967025277e-06, |
|
"loss": 1.0775, |
|
"step": 674000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.132991502352364e-06, |
|
"loss": 1.0819, |
|
"step": 674500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 6.1004733344519585e-06, |
|
"loss": 1.0753, |
|
"step": 675000 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 6.067955166551553e-06, |
|
"loss": 1.0901, |
|
"step": 675500 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 6.035436998651146e-06, |
|
"loss": 1.0796, |
|
"step": 676000 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 6.002918830750741e-06, |
|
"loss": 1.0775, |
|
"step": 676500 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 5.970400662850334e-06, |
|
"loss": 1.0789, |
|
"step": 677000 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 5.937882494949929e-06, |
|
"loss": 1.0898, |
|
"step": 677500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.905364327049523e-06, |
|
"loss": 1.0753, |
|
"step": 678000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.872846159149117e-06, |
|
"loss": 1.0732, |
|
"step": 678500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.840327991248711e-06, |
|
"loss": 1.0776, |
|
"step": 679000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.807809823348305e-06, |
|
"loss": 1.0794, |
|
"step": 679500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 5.775291655447899e-06, |
|
"loss": 1.0765, |
|
"step": 680000 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.742773487547493e-06, |
|
"loss": 1.0737, |
|
"step": 680500 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.710255319647087e-06, |
|
"loss": 1.0817, |
|
"step": 681000 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.677737151746681e-06, |
|
"loss": 1.0755, |
|
"step": 681500 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.645218983846275e-06, |
|
"loss": 1.0853, |
|
"step": 682000 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 5.612700815945869e-06, |
|
"loss": 1.0847, |
|
"step": 682500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.580182648045464e-06, |
|
"loss": 1.086, |
|
"step": 683000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.547664480145057e-06, |
|
"loss": 1.0752, |
|
"step": 683500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.515146312244651e-06, |
|
"loss": 1.0851, |
|
"step": 684000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.4826281443442455e-06, |
|
"loss": 1.0743, |
|
"step": 684500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.45010997644384e-06, |
|
"loss": 1.0807, |
|
"step": 685000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 5.417591808543433e-06, |
|
"loss": 1.0754, |
|
"step": 685500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.385073640643028e-06, |
|
"loss": 1.0702, |
|
"step": 686000 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.352555472742621e-06, |
|
"loss": 1.0796, |
|
"step": 686500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.3200373048422156e-06, |
|
"loss": 1.0742, |
|
"step": 687000 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.28751913694181e-06, |
|
"loss": 1.0724, |
|
"step": 687500 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.255000969041404e-06, |
|
"loss": 1.0806, |
|
"step": 688000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.222482801140998e-06, |
|
"loss": 1.0784, |
|
"step": 688500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.1899646332405914e-06, |
|
"loss": 1.0779, |
|
"step": 689000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.157446465340186e-06, |
|
"loss": 1.0747, |
|
"step": 689500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.12492829743978e-06, |
|
"loss": 1.0713, |
|
"step": 690000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.092410129539374e-06, |
|
"loss": 1.0813, |
|
"step": 690500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.059891961638967e-06, |
|
"loss": 1.0787, |
|
"step": 691000 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.027373793738562e-06, |
|
"loss": 1.0704, |
|
"step": 691500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.994855625838156e-06, |
|
"loss": 1.0789, |
|
"step": 692000 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.96233745793775e-06, |
|
"loss": 1.0808, |
|
"step": 692500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.929819290037344e-06, |
|
"loss": 1.077, |
|
"step": 693000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.897301122136938e-06, |
|
"loss": 1.0727, |
|
"step": 693500 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.8647829542365325e-06, |
|
"loss": 1.0761, |
|
"step": 694000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.832264786336127e-06, |
|
"loss": 1.072, |
|
"step": 694500 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.79974661843572e-06, |
|
"loss": 1.077, |
|
"step": 695000 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.767228450535314e-06, |
|
"loss": 1.0719, |
|
"step": 695500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.734710282634908e-06, |
|
"loss": 1.0732, |
|
"step": 696000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.7021921147345025e-06, |
|
"loss": 1.0697, |
|
"step": 696500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.669673946834097e-06, |
|
"loss": 1.0658, |
|
"step": 697000 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.63715577893369e-06, |
|
"loss": 1.0835, |
|
"step": 697500 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.604637611033284e-06, |
|
"loss": 1.0741, |
|
"step": 698000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.5721194431328784e-06, |
|
"loss": 1.0666, |
|
"step": 698500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.539601275232473e-06, |
|
"loss": 1.0721, |
|
"step": 699000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.507083107332067e-06, |
|
"loss": 1.0756, |
|
"step": 699500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.474564939431661e-06, |
|
"loss": 1.0732, |
|
"step": 700000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.442046771531254e-06, |
|
"loss": 1.0761, |
|
"step": 700500 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.409528603630849e-06, |
|
"loss": 1.0742, |
|
"step": 701000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.377010435730443e-06, |
|
"loss": 1.0779, |
|
"step": 701500 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.344492267830037e-06, |
|
"loss": 1.072, |
|
"step": 702000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.311974099929631e-06, |
|
"loss": 1.0762, |
|
"step": 702500 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.279455932029225e-06, |
|
"loss": 1.0707, |
|
"step": 703000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.246937764128819e-06, |
|
"loss": 1.0772, |
|
"step": 703500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.214419596228413e-06, |
|
"loss": 1.0619, |
|
"step": 704000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.181901428328007e-06, |
|
"loss": 1.0868, |
|
"step": 704500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.149383260427601e-06, |
|
"loss": 1.0695, |
|
"step": 705000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.116865092527195e-06, |
|
"loss": 1.0613, |
|
"step": 705500 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.084346924626789e-06, |
|
"loss": 1.0673, |
|
"step": 706000 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.051828756726384e-06, |
|
"loss": 1.0672, |
|
"step": 706500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.019310588825977e-06, |
|
"loss": 1.067, |
|
"step": 707000 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.986792420925571e-06, |
|
"loss": 1.0703, |
|
"step": 707500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.9542742530251654e-06, |
|
"loss": 1.0661, |
|
"step": 708000 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.92175608512476e-06, |
|
"loss": 1.0697, |
|
"step": 708500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.889237917224353e-06, |
|
"loss": 1.072, |
|
"step": 709000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.856719749323948e-06, |
|
"loss": 1.0708, |
|
"step": 709500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.824201581423541e-06, |
|
"loss": 1.0715, |
|
"step": 710000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.791683413523135e-06, |
|
"loss": 1.0731, |
|
"step": 710500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.7591652456227297e-06, |
|
"loss": 1.0731, |
|
"step": 711000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.7266470777223235e-06, |
|
"loss": 1.0659, |
|
"step": 711500 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.694128909821918e-06, |
|
"loss": 1.0744, |
|
"step": 712000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.661610741921512e-06, |
|
"loss": 1.0691, |
|
"step": 712500 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.6290925740211056e-06, |
|
"loss": 1.078, |
|
"step": 713000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.5965744061207e-06, |
|
"loss": 1.0713, |
|
"step": 713500 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.564056238220294e-06, |
|
"loss": 1.0709, |
|
"step": 714000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.5315380703198877e-06, |
|
"loss": 1.0647, |
|
"step": 714500 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.499019902419482e-06, |
|
"loss": 1.0637, |
|
"step": 715000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.466501734519076e-06, |
|
"loss": 1.0696, |
|
"step": 715500 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.43398356661867e-06, |
|
"loss": 1.0796, |
|
"step": 716000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.401465398718264e-06, |
|
"loss": 1.0665, |
|
"step": 716500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.368947230817858e-06, |
|
"loss": 1.0692, |
|
"step": 717000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.3364290629174524e-06, |
|
"loss": 1.0693, |
|
"step": 717500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.303910895017046e-06, |
|
"loss": 1.0715, |
|
"step": 718000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.27139272711664e-06, |
|
"loss": 1.0652, |
|
"step": 718500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.2388745592162346e-06, |
|
"loss": 1.0846, |
|
"step": 719000 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.2063563913158283e-06, |
|
"loss": 1.0657, |
|
"step": 719500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.173838223415422e-06, |
|
"loss": 1.0693, |
|
"step": 720000 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.1413200555150167e-06, |
|
"loss": 1.0736, |
|
"step": 720500 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.1088018876146104e-06, |
|
"loss": 1.0636, |
|
"step": 721000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.0762837197142046e-06, |
|
"loss": 1.069, |
|
"step": 721500 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.043765551813799e-06, |
|
"loss": 1.072, |
|
"step": 722000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.0112473839133926e-06, |
|
"loss": 1.0663, |
|
"step": 722500 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.9787292160129868e-06, |
|
"loss": 1.072, |
|
"step": 723000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.9462110481125805e-06, |
|
"loss": 1.0611, |
|
"step": 723500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.9136928802121747e-06, |
|
"loss": 1.0741, |
|
"step": 724000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.8811747123117685e-06, |
|
"loss": 1.0712, |
|
"step": 724500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.8486565444113627e-06, |
|
"loss": 1.07, |
|
"step": 725000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.816138376510957e-06, |
|
"loss": 1.0655, |
|
"step": 725500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.7836202086105506e-06, |
|
"loss": 1.0663, |
|
"step": 726000 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 2.751102040710145e-06, |
|
"loss": 1.0648, |
|
"step": 726500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.718583872809739e-06, |
|
"loss": 1.068, |
|
"step": 727000 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.686065704909333e-06, |
|
"loss": 1.0661, |
|
"step": 727500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.653547537008927e-06, |
|
"loss": 1.0593, |
|
"step": 728000 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.621029369108521e-06, |
|
"loss": 1.0652, |
|
"step": 728500 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.5885112012081153e-06, |
|
"loss": 1.0744, |
|
"step": 729000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.5559930333077095e-06, |
|
"loss": 1.0704, |
|
"step": 729500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.5234748654073033e-06, |
|
"loss": 1.0706, |
|
"step": 730000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.490956697506897e-06, |
|
"loss": 1.0681, |
|
"step": 730500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.458438529606491e-06, |
|
"loss": 1.078, |
|
"step": 731000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.425920361706085e-06, |
|
"loss": 1.0632, |
|
"step": 731500 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.393402193805679e-06, |
|
"loss": 1.0671, |
|
"step": 732000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.3608840259052733e-06, |
|
"loss": 1.0681, |
|
"step": 732500 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.3283658580048675e-06, |
|
"loss": 1.0653, |
|
"step": 733000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.2958476901044613e-06, |
|
"loss": 1.0697, |
|
"step": 733500 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.2633295222040555e-06, |
|
"loss": 1.0691, |
|
"step": 734000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.2308113543036497e-06, |
|
"loss": 1.0584, |
|
"step": 734500 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.198293186403244e-06, |
|
"loss": 1.0656, |
|
"step": 735000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.1657750185028376e-06, |
|
"loss": 1.0707, |
|
"step": 735500 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.133256850602432e-06, |
|
"loss": 1.062, |
|
"step": 736000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.100738682702026e-06, |
|
"loss": 1.0601, |
|
"step": 736500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.0682205148016197e-06, |
|
"loss": 1.0657, |
|
"step": 737000 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.035702346901214e-06, |
|
"loss": 1.0618, |
|
"step": 737500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.0031841790008077e-06, |
|
"loss": 1.0607, |
|
"step": 738000 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.970666011100402e-06, |
|
"loss": 1.0701, |
|
"step": 738500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.9381478431999956e-06, |
|
"loss": 1.0678, |
|
"step": 739000 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.90562967529959e-06, |
|
"loss": 1.071, |
|
"step": 739500 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.873111507399184e-06, |
|
"loss": 1.0649, |
|
"step": 740000 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.8405933394987782e-06, |
|
"loss": 1.0644, |
|
"step": 740500 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.808075171598372e-06, |
|
"loss": 1.0663, |
|
"step": 741000 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.7755570036979661e-06, |
|
"loss": 1.0682, |
|
"step": 741500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.7430388357975603e-06, |
|
"loss": 1.0641, |
|
"step": 742000 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.710520667897154e-06, |
|
"loss": 1.0597, |
|
"step": 742500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6780024999967483e-06, |
|
"loss": 1.0585, |
|
"step": 743000 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6454843320963425e-06, |
|
"loss": 1.0633, |
|
"step": 743500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.6129661641959364e-06, |
|
"loss": 1.0659, |
|
"step": 744000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.5804479962955302e-06, |
|
"loss": 1.0535, |
|
"step": 744500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.5479298283951244e-06, |
|
"loss": 1.0733, |
|
"step": 745000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.5154116604947186e-06, |
|
"loss": 1.0663, |
|
"step": 745500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.4828934925943125e-06, |
|
"loss": 1.0696, |
|
"step": 746000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.4503753246939065e-06, |
|
"loss": 1.0688, |
|
"step": 746500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.4178571567935007e-06, |
|
"loss": 1.0593, |
|
"step": 747000 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.3853389888930947e-06, |
|
"loss": 1.0683, |
|
"step": 747500 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.3528208209926887e-06, |
|
"loss": 1.0631, |
|
"step": 748000 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.3203026530922826e-06, |
|
"loss": 1.0726, |
|
"step": 748500 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.2877844851918768e-06, |
|
"loss": 1.0675, |
|
"step": 749000 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.2552663172914708e-06, |
|
"loss": 1.0617, |
|
"step": 749500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.222748149391065e-06, |
|
"loss": 1.0617, |
|
"step": 750000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.190229981490659e-06, |
|
"loss": 1.0654, |
|
"step": 750500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.157711813590253e-06, |
|
"loss": 1.0681, |
|
"step": 751000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.1251936456898469e-06, |
|
"loss": 1.0669, |
|
"step": 751500 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.0926754777894409e-06, |
|
"loss": 1.0634, |
|
"step": 752000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.060157309889035e-06, |
|
"loss": 1.0596, |
|
"step": 752500 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.027639141988629e-06, |
|
"loss": 1.0609, |
|
"step": 753000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 9.951209740882232e-07, |
|
"loss": 1.0641, |
|
"step": 753500 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 9.626028061878172e-07, |
|
"loss": 1.0555, |
|
"step": 754000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 9.300846382874113e-07, |
|
"loss": 1.0574, |
|
"step": 754500 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.975664703870052e-07, |
|
"loss": 1.0654, |
|
"step": 755000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.650483024865994e-07, |
|
"loss": 1.0592, |
|
"step": 755500 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.325301345861933e-07, |
|
"loss": 1.057, |
|
"step": 756000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 8.000119666857873e-07, |
|
"loss": 1.0593, |
|
"step": 756500 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 7.674937987853815e-07, |
|
"loss": 1.0655, |
|
"step": 757000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.349756308849755e-07, |
|
"loss": 1.071, |
|
"step": 757500 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 7.024574629845695e-07, |
|
"loss": 1.0546, |
|
"step": 758000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 6.699392950841635e-07, |
|
"loss": 1.0573, |
|
"step": 758500 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 6.374211271837576e-07, |
|
"loss": 1.0652, |
|
"step": 759000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 6.049029592833516e-07, |
|
"loss": 1.065, |
|
"step": 759500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.723847913829456e-07, |
|
"loss": 1.071, |
|
"step": 760000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.398666234825397e-07, |
|
"loss": 1.0628, |
|
"step": 760500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.073484555821338e-07, |
|
"loss": 1.0618, |
|
"step": 761000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 4.748302876817278e-07, |
|
"loss": 1.0604, |
|
"step": 761500 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 4.423121197813219e-07, |
|
"loss": 1.0612, |
|
"step": 762000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 4.097939518809159e-07, |
|
"loss": 1.0581, |
|
"step": 762500 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.7727578398050994e-07, |
|
"loss": 1.0603, |
|
"step": 763000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.4475761608010396e-07, |
|
"loss": 1.0605, |
|
"step": 763500 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.1223944817969804e-07, |
|
"loss": 1.0618, |
|
"step": 764000 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.7972128027929207e-07, |
|
"loss": 1.064, |
|
"step": 764500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.472031123788861e-07, |
|
"loss": 1.0619, |
|
"step": 765000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.1468494447848013e-07, |
|
"loss": 1.0642, |
|
"step": 765500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.8216677657807418e-07, |
|
"loss": 1.0553, |
|
"step": 766000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.4964860867766823e-07, |
|
"loss": 1.0646, |
|
"step": 766500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.1713044077726226e-07, |
|
"loss": 1.0582, |
|
"step": 767000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 8.46122728768563e-08, |
|
"loss": 1.057, |
|
"step": 767500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 5.209410497645035e-08, |
|
"loss": 1.0583, |
|
"step": 768000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.9575937076044387e-08, |
|
"loss": 1.0635, |
|
"step": 768500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 768801, |
|
"total_flos": 6.47787050209493e+18, |
|
"train_loss": 2.538804254183028, |
|
"train_runtime": 274559.0377, |
|
"train_samples_per_second": 89.604, |
|
"train_steps_per_second": 2.8 |
|
} |
|
], |
|
"max_steps": 768801, |
|
"num_train_epochs": 3, |
|
"total_flos": 6.47787050209493e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|