|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 30.0, |
|
"global_step": 486870, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 3.6192, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 3.0017, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 3e-06, |
|
"loss": 2.7552, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.632, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 5e-06, |
|
"loss": 2.5416, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 6e-06, |
|
"loss": 2.468, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7e-06, |
|
"loss": 2.404, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.3388, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9e-06, |
|
"loss": 2.2985, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1e-05, |
|
"loss": 2.2616, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 2.2156, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.1814, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 2.1473, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.4e-05, |
|
"loss": 2.1205, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 2.1008, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 2.0718, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.7e-05, |
|
"loss": 2.0496, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8e-05, |
|
"loss": 2.0264, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.9e-05, |
|
"loss": 2.0085, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2e-05, |
|
"loss": 1.9873, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.9979029924298028e-05, |
|
"loss": 1.9619, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.9958059848596054e-05, |
|
"loss": 1.952, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9937089772894084e-05, |
|
"loss": 1.9409, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.991611969719211e-05, |
|
"loss": 1.9217, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.9895149621490136e-05, |
|
"loss": 1.9073, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9874179545788163e-05, |
|
"loss": 1.8983, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.985320947008619e-05, |
|
"loss": 1.8814, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.9832239394384215e-05, |
|
"loss": 1.8657, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.981126931868224e-05, |
|
"loss": 1.8473, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.9790299242980267e-05, |
|
"loss": 1.8407, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.9769329167278294e-05, |
|
"loss": 1.8356, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.974835909157632e-05, |
|
"loss": 1.8291, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.972738901587435e-05, |
|
"loss": 1.8095, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.9706418940172376e-05, |
|
"loss": 1.8137, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.9685448864470402e-05, |
|
"loss": 1.8022, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.966447878876843e-05, |
|
"loss": 1.7933, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.9643508713066455e-05, |
|
"loss": 1.7749, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.9622538637364484e-05, |
|
"loss": 1.7731, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.960156856166251e-05, |
|
"loss": 1.7717, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.9580598485960537e-05, |
|
"loss": 1.769, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.9559628410258563e-05, |
|
"loss": 1.7564, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.953865833455659e-05, |
|
"loss": 1.7541, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.9517688258854615e-05, |
|
"loss": 1.7503, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.9496718183152642e-05, |
|
"loss": 1.7405, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.9475748107450668e-05, |
|
"loss": 1.7392, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.9454778031748694e-05, |
|
"loss": 1.7311, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.943380795604672e-05, |
|
"loss": 1.728, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.941283788034475e-05, |
|
"loss": 1.7181, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.9391867804642776e-05, |
|
"loss": 1.7145, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.9370897728940803e-05, |
|
"loss": 1.7176, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.934992765323883e-05, |
|
"loss": 1.7116, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.9328957577536855e-05, |
|
"loss": 1.7063, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.9307987501834885e-05, |
|
"loss": 1.6982, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.928701742613291e-05, |
|
"loss": 1.6941, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.9266047350430937e-05, |
|
"loss": 1.6868, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.9245077274728963e-05, |
|
"loss": 1.6887, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.922410719902699e-05, |
|
"loss": 1.6855, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.9203137123325016e-05, |
|
"loss": 1.6802, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9182167047623042e-05, |
|
"loss": 1.6739, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.916119697192107e-05, |
|
"loss": 1.674, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.9140226896219095e-05, |
|
"loss": 1.667, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.9119256820517124e-05, |
|
"loss": 1.6611, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.909828674481515e-05, |
|
"loss": 1.6597, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.9077316669113177e-05, |
|
"loss": 1.6639, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.9056346593411203e-05, |
|
"loss": 1.651, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.903537651770923e-05, |
|
"loss": 1.6532, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.901440644200726e-05, |
|
"loss": 1.6374, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.8993436366305285e-05, |
|
"loss": 1.6437, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.897246629060331e-05, |
|
"loss": 1.6372, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.8951496214901338e-05, |
|
"loss": 1.6427, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.8930526139199364e-05, |
|
"loss": 1.6327, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.890955606349739e-05, |
|
"loss": 1.6289, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.8888585987795416e-05, |
|
"loss": 1.6294, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.8867615912093443e-05, |
|
"loss": 1.6255, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.884664583639147e-05, |
|
"loss": 1.6289, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.88256757606895e-05, |
|
"loss": 1.6217, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.8804705684987525e-05, |
|
"loss": 1.6137, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.878373560928555e-05, |
|
"loss": 1.6145, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.8762765533583577e-05, |
|
"loss": 1.6193, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.8741795457881604e-05, |
|
"loss": 1.6082, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.872082538217963e-05, |
|
"loss": 1.6114, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.869985530647766e-05, |
|
"loss": 1.6123, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.8678885230775686e-05, |
|
"loss": 1.605, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 1.8657915155073712e-05, |
|
"loss": 1.6008, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.8636945079371738e-05, |
|
"loss": 1.5934, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.8615975003669764e-05, |
|
"loss": 1.6007, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.8595004927967794e-05, |
|
"loss": 1.5928, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.857403485226582e-05, |
|
"loss": 1.5988, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.8553064776563847e-05, |
|
"loss": 1.5943, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.8532094700861873e-05, |
|
"loss": 1.5842, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.85111246251599e-05, |
|
"loss": 1.5901, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.8490154549457925e-05, |
|
"loss": 1.5841, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.846918447375595e-05, |
|
"loss": 1.5784, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.8448214398053978e-05, |
|
"loss": 1.58, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.8427244322352004e-05, |
|
"loss": 1.5736, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.840627424665003e-05, |
|
"loss": 1.5697, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.838530417094806e-05, |
|
"loss": 1.5739, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.8364334095246086e-05, |
|
"loss": 1.5688, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.8343364019544112e-05, |
|
"loss": 1.5644, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 1.832239394384214e-05, |
|
"loss": 1.562, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.8301423868140165e-05, |
|
"loss": 1.5709, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.8280453792438195e-05, |
|
"loss": 1.5617, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 1.825948371673622e-05, |
|
"loss": 1.5603, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.8238513641034247e-05, |
|
"loss": 1.5625, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.8217543565332273e-05, |
|
"loss": 1.5583, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 1.81965734896303e-05, |
|
"loss": 1.5597, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.8175603413928326e-05, |
|
"loss": 1.553, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 1.8154633338226352e-05, |
|
"loss": 1.5525, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.8133663262524378e-05, |
|
"loss": 1.5521, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.8112693186822404e-05, |
|
"loss": 1.5501, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.809172311112043e-05, |
|
"loss": 1.55, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.807075303541846e-05, |
|
"loss": 1.5504, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.8049782959716487e-05, |
|
"loss": 1.5451, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.8028812884014513e-05, |
|
"loss": 1.5443, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.800784280831254e-05, |
|
"loss": 1.5412, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.7986872732610565e-05, |
|
"loss": 1.5398, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.7965902656908595e-05, |
|
"loss": 1.5425, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.794493258120662e-05, |
|
"loss": 1.538, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.7923962505504647e-05, |
|
"loss": 1.5406, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.7902992429802674e-05, |
|
"loss": 1.5365, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.78820223541007e-05, |
|
"loss": 1.5344, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.7861052278398726e-05, |
|
"loss": 1.5332, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.7840082202696752e-05, |
|
"loss": 1.5311, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.781911212699478e-05, |
|
"loss": 1.5316, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.7798142051292805e-05, |
|
"loss": 1.5241, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.777717197559083e-05, |
|
"loss": 1.523, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.775620189988886e-05, |
|
"loss": 1.5247, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 1.7735231824186887e-05, |
|
"loss": 1.5246, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 1.7714261748484913e-05, |
|
"loss": 1.5215, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 1.769329167278294e-05, |
|
"loss": 1.5256, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.7672321597080966e-05, |
|
"loss": 1.5244, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 1.7651351521378995e-05, |
|
"loss": 1.5104, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.7630381445677022e-05, |
|
"loss": 1.5083, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.7609411369975048e-05, |
|
"loss": 1.5076, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 1.7588441294273074e-05, |
|
"loss": 1.5132, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 1.75674712185711e-05, |
|
"loss": 1.5163, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 1.7546501142869127e-05, |
|
"loss": 1.5133, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.7525531067167153e-05, |
|
"loss": 1.5099, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 1.750456099146518e-05, |
|
"loss": 1.5054, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.7483590915763205e-05, |
|
"loss": 1.5081, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.7462620840061235e-05, |
|
"loss": 1.5063, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 1.744165076435926e-05, |
|
"loss": 1.5016, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 1.7420680688657288e-05, |
|
"loss": 1.5061, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 1.7399710612955314e-05, |
|
"loss": 1.5001, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 1.737874053725334e-05, |
|
"loss": 1.5002, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.735777046155137e-05, |
|
"loss": 1.5027, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.7336800385849396e-05, |
|
"loss": 1.4996, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.7315830310147422e-05, |
|
"loss": 1.4975, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 1.729486023444545e-05, |
|
"loss": 1.5008, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.7273890158743475e-05, |
|
"loss": 1.4927, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 1.72529200830415e-05, |
|
"loss": 1.4955, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.7231950007339527e-05, |
|
"loss": 1.4892, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 1.7210979931637553e-05, |
|
"loss": 1.4918, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 1.719000985593558e-05, |
|
"loss": 1.4886, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 1.7169039780233606e-05, |
|
"loss": 1.4964, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 1.7148069704531636e-05, |
|
"loss": 1.4863, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 1.7127099628829662e-05, |
|
"loss": 1.4895, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.7106129553127688e-05, |
|
"loss": 1.4893, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 1.7085159477425714e-05, |
|
"loss": 1.4906, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 1.706418940172374e-05, |
|
"loss": 1.4911, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 1.704321932602177e-05, |
|
"loss": 1.4903, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.7022249250319796e-05, |
|
"loss": 1.4839, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 1.7001279174617823e-05, |
|
"loss": 1.4805, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 1.698030909891585e-05, |
|
"loss": 1.4767, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 1.6959339023213875e-05, |
|
"loss": 1.4758, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 1.69383689475119e-05, |
|
"loss": 1.4798, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 1.6917398871809928e-05, |
|
"loss": 1.4724, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 1.6896428796107954e-05, |
|
"loss": 1.4736, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 1.687545872040598e-05, |
|
"loss": 1.4691, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 1.6854488644704006e-05, |
|
"loss": 1.4756, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 1.6833518569002036e-05, |
|
"loss": 1.4813, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 1.6812548493300062e-05, |
|
"loss": 1.4753, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 1.679157841759809e-05, |
|
"loss": 1.4701, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 1.6770608341896115e-05, |
|
"loss": 1.4697, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 1.674963826619414e-05, |
|
"loss": 1.4728, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 1.672866819049217e-05, |
|
"loss": 1.4701, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 1.6707698114790197e-05, |
|
"loss": 1.4675, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 1.6686728039088223e-05, |
|
"loss": 1.4667, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 1.666575796338625e-05, |
|
"loss": 1.4627, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 1.6644787887684276e-05, |
|
"loss": 1.4633, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 1.6623817811982302e-05, |
|
"loss": 1.4629, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 1.6602847736280328e-05, |
|
"loss": 1.4638, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 1.6581877660578354e-05, |
|
"loss": 1.4673, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 1.656090758487638e-05, |
|
"loss": 1.461, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 1.6539937509174407e-05, |
|
"loss": 1.4613, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 1.6518967433472437e-05, |
|
"loss": 1.4629, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 1.6497997357770463e-05, |
|
"loss": 1.4579, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 1.647702728206849e-05, |
|
"loss": 1.4546, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.6456057206366515e-05, |
|
"loss": 1.4541, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 1.643508713066454e-05, |
|
"loss": 1.4586, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.641411705496257e-05, |
|
"loss": 1.4567, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 1.6393146979260597e-05, |
|
"loss": 1.4485, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 1.6372176903558624e-05, |
|
"loss": 1.4503, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 1.635120682785665e-05, |
|
"loss": 1.4566, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 1.6330236752154676e-05, |
|
"loss": 1.4526, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 1.6309266676452706e-05, |
|
"loss": 1.4417, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 1.6288296600750732e-05, |
|
"loss": 1.4384, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 1.6267326525048758e-05, |
|
"loss": 1.4511, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 1.6246356449346784e-05, |
|
"loss": 1.4437, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 1.622538637364481e-05, |
|
"loss": 1.4464, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 1.6204416297942837e-05, |
|
"loss": 1.4437, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 1.6183446222240863e-05, |
|
"loss": 1.4424, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 1.616247614653889e-05, |
|
"loss": 1.4423, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 1.6141506070836916e-05, |
|
"loss": 1.4483, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 1.6120535995134942e-05, |
|
"loss": 1.4487, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 1.609956591943297e-05, |
|
"loss": 1.4403, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.6078595843730998e-05, |
|
"loss": 1.4417, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 1.6057625768029024e-05, |
|
"loss": 1.4403, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.603665569232705e-05, |
|
"loss": 1.4426, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 1.6015685616625077e-05, |
|
"loss": 1.441, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 1.5994715540923106e-05, |
|
"loss": 1.4416, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.5973745465221132e-05, |
|
"loss": 1.4478, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 1.595277538951916e-05, |
|
"loss": 1.4347, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 1.5931805313817185e-05, |
|
"loss": 1.4347, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 1.591083523811521e-05, |
|
"loss": 1.4346, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.5889865162413237e-05, |
|
"loss": 1.4344, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 1.5868895086711264e-05, |
|
"loss": 1.4405, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 1.584792501100929e-05, |
|
"loss": 1.439, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 1.5826954935307316e-05, |
|
"loss": 1.4293, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.5805984859605346e-05, |
|
"loss": 1.4289, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.5785014783903372e-05, |
|
"loss": 1.4316, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 1.57640447082014e-05, |
|
"loss": 1.4329, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 1.5743074632499425e-05, |
|
"loss": 1.4318, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 1.572210455679745e-05, |
|
"loss": 1.435, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 1.570113448109548e-05, |
|
"loss": 1.4323, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 1.5680164405393507e-05, |
|
"loss": 1.4334, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 1.5659194329691533e-05, |
|
"loss": 1.4303, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.563822425398956e-05, |
|
"loss": 1.4241, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 1.5617254178287585e-05, |
|
"loss": 1.4261, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 1.559628410258561e-05, |
|
"loss": 1.424, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 1.5575314026883638e-05, |
|
"loss": 1.4229, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 1.5554343951181664e-05, |
|
"loss": 1.4258, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 1.553337387547969e-05, |
|
"loss": 1.4203, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 1.5512403799777717e-05, |
|
"loss": 1.4193, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 1.5491433724075746e-05, |
|
"loss": 1.4302, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 1.5470463648373773e-05, |
|
"loss": 1.4206, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 1.54494935726718e-05, |
|
"loss": 1.4181, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 1.5428523496969825e-05, |
|
"loss": 1.4179, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 1.540755342126785e-05, |
|
"loss": 1.4207, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 1.538658334556588e-05, |
|
"loss": 1.4117, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 1.5365613269863907e-05, |
|
"loss": 1.4153, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 1.5344643194161933e-05, |
|
"loss": 1.419, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 1.532367311845996e-05, |
|
"loss": 1.4232, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 1.5302703042757986e-05, |
|
"loss": 1.421, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 1.5281732967056012e-05, |
|
"loss": 1.4236, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 1.526076289135404e-05, |
|
"loss": 1.4154, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 1.5239792815652066e-05, |
|
"loss": 1.4103, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 1.5218822739950093e-05, |
|
"loss": 1.4092, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 1.5197852664248119e-05, |
|
"loss": 1.4167, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 1.5176882588546147e-05, |
|
"loss": 1.4183, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 1.5155912512844173e-05, |
|
"loss": 1.4099, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 1.51349424371422e-05, |
|
"loss": 1.413, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 1.5113972361440226e-05, |
|
"loss": 1.411, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 1.5093002285738252e-05, |
|
"loss": 1.4109, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 1.507203221003628e-05, |
|
"loss": 1.414, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 1.5051062134334306e-05, |
|
"loss": 1.4149, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 1.5030092058632332e-05, |
|
"loss": 1.4143, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 1.5009121982930358e-05, |
|
"loss": 1.4164, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 1.4988151907228385e-05, |
|
"loss": 1.4118, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 1.4967181831526414e-05, |
|
"loss": 1.4053, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 1.494621175582444e-05, |
|
"loss": 1.4026, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 1.4925241680122467e-05, |
|
"loss": 1.4048, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 1.4904271604420493e-05, |
|
"loss": 1.4014, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 1.488330152871852e-05, |
|
"loss": 1.4036, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 1.4862331453016547e-05, |
|
"loss": 1.4054, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 1.4841361377314574e-05, |
|
"loss": 1.4043, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 1.48203913016126e-05, |
|
"loss": 1.4035, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 1.4799421225910626e-05, |
|
"loss": 1.4045, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 1.4778451150208652e-05, |
|
"loss": 1.4024, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 1.475748107450668e-05, |
|
"loss": 1.3982, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 1.4736510998804706e-05, |
|
"loss": 1.3996, |
|
"step": 135500 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 1.4715540923102733e-05, |
|
"loss": 1.397, |
|
"step": 136000 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 1.469457084740076e-05, |
|
"loss": 1.4058, |
|
"step": 136500 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 1.4673600771698787e-05, |
|
"loss": 1.3984, |
|
"step": 137000 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 1.4652630695996815e-05, |
|
"loss": 1.4037, |
|
"step": 137500 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 1.4631660620294841e-05, |
|
"loss": 1.4021, |
|
"step": 138000 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 1.4610690544592867e-05, |
|
"loss": 1.3966, |
|
"step": 138500 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 1.4589720468890894e-05, |
|
"loss": 1.397, |
|
"step": 139000 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 1.456875039318892e-05, |
|
"loss": 1.3994, |
|
"step": 139500 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 1.4547780317486948e-05, |
|
"loss": 1.3992, |
|
"step": 140000 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 1.4526810241784974e-05, |
|
"loss": 1.399, |
|
"step": 140500 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 1.4505840166083e-05, |
|
"loss": 1.4036, |
|
"step": 141000 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 1.4484870090381026e-05, |
|
"loss": 1.3949, |
|
"step": 141500 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 1.4463900014679053e-05, |
|
"loss": 1.3994, |
|
"step": 142000 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 1.4442929938977082e-05, |
|
"loss": 1.3941, |
|
"step": 142500 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 1.4421959863275109e-05, |
|
"loss": 1.3963, |
|
"step": 143000 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 1.4400989787573135e-05, |
|
"loss": 1.3996, |
|
"step": 143500 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 1.4380019711871161e-05, |
|
"loss": 1.3914, |
|
"step": 144000 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 1.4359049636169187e-05, |
|
"loss": 1.3923, |
|
"step": 144500 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 1.4338079560467215e-05, |
|
"loss": 1.3958, |
|
"step": 145000 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 1.4317109484765242e-05, |
|
"loss": 1.3906, |
|
"step": 145500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 1.4296139409063268e-05, |
|
"loss": 1.3867, |
|
"step": 146000 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 1.4275169333361294e-05, |
|
"loss": 1.3821, |
|
"step": 146500 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.425419925765932e-05, |
|
"loss": 1.3917, |
|
"step": 147000 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 1.4233229181957348e-05, |
|
"loss": 1.3972, |
|
"step": 147500 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 1.4212259106255374e-05, |
|
"loss": 1.3848, |
|
"step": 148000 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 1.41912890305534e-05, |
|
"loss": 1.386, |
|
"step": 148500 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 1.4170318954851427e-05, |
|
"loss": 1.3855, |
|
"step": 149000 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 1.4149348879149457e-05, |
|
"loss": 1.3876, |
|
"step": 149500 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 1.4128378803447483e-05, |
|
"loss": 1.3836, |
|
"step": 150000 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 1.4107408727745509e-05, |
|
"loss": 1.3841, |
|
"step": 150500 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 1.4086438652043535e-05, |
|
"loss": 1.3828, |
|
"step": 151000 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 1.4065468576341562e-05, |
|
"loss": 1.3914, |
|
"step": 151500 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 1.404449850063959e-05, |
|
"loss": 1.3813, |
|
"step": 152000 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 1.4023528424937616e-05, |
|
"loss": 1.3803, |
|
"step": 152500 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 1.4002558349235642e-05, |
|
"loss": 1.3852, |
|
"step": 153000 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 1.3981588273533668e-05, |
|
"loss": 1.3791, |
|
"step": 153500 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 1.3960618197831695e-05, |
|
"loss": 1.3798, |
|
"step": 154000 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 1.3939648122129722e-05, |
|
"loss": 1.384, |
|
"step": 154500 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 1.3918678046427749e-05, |
|
"loss": 1.3834, |
|
"step": 155000 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 1.3897707970725775e-05, |
|
"loss": 1.3852, |
|
"step": 155500 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 1.3876737895023801e-05, |
|
"loss": 1.3827, |
|
"step": 156000 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.3855767819321827e-05, |
|
"loss": 1.3846, |
|
"step": 156500 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 1.3834797743619857e-05, |
|
"loss": 1.3849, |
|
"step": 157000 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 1.3813827667917883e-05, |
|
"loss": 1.3816, |
|
"step": 157500 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 1.379285759221591e-05, |
|
"loss": 1.3777, |
|
"step": 158000 |
|
}, |
|
{ |
|
"epoch": 9.77, |
|
"learning_rate": 1.3771887516513936e-05, |
|
"loss": 1.3773, |
|
"step": 158500 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 1.3750917440811962e-05, |
|
"loss": 1.379, |
|
"step": 159000 |
|
}, |
|
{ |
|
"epoch": 9.83, |
|
"learning_rate": 1.372994736510999e-05, |
|
"loss": 1.3778, |
|
"step": 159500 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 1.3708977289408016e-05, |
|
"loss": 1.3772, |
|
"step": 160000 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 1.3688007213706042e-05, |
|
"loss": 1.3793, |
|
"step": 160500 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 1.3667037138004069e-05, |
|
"loss": 1.3763, |
|
"step": 161000 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 1.3646067062302095e-05, |
|
"loss": 1.3781, |
|
"step": 161500 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 1.3625096986600123e-05, |
|
"loss": 1.3821, |
|
"step": 162000 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 1.360412691089815e-05, |
|
"loss": 1.3729, |
|
"step": 162500 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 1.3583156835196175e-05, |
|
"loss": 1.3756, |
|
"step": 163000 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 1.3562186759494202e-05, |
|
"loss": 1.3719, |
|
"step": 163500 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 1.354121668379223e-05, |
|
"loss": 1.3704, |
|
"step": 164000 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 1.3520246608090258e-05, |
|
"loss": 1.3729, |
|
"step": 164500 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 1.3499276532388284e-05, |
|
"loss": 1.3721, |
|
"step": 165000 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 1.347830645668631e-05, |
|
"loss": 1.3708, |
|
"step": 165500 |
|
}, |
|
{ |
|
"epoch": 10.23, |
|
"learning_rate": 1.3457336380984336e-05, |
|
"loss": 1.369, |
|
"step": 166000 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 1.3436366305282363e-05, |
|
"loss": 1.3703, |
|
"step": 166500 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 1.341539622958039e-05, |
|
"loss": 1.3715, |
|
"step": 167000 |
|
}, |
|
{ |
|
"epoch": 10.32, |
|
"learning_rate": 1.3394426153878417e-05, |
|
"loss": 1.3714, |
|
"step": 167500 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 1.3373456078176443e-05, |
|
"loss": 1.3711, |
|
"step": 168000 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 1.335248600247447e-05, |
|
"loss": 1.3759, |
|
"step": 168500 |
|
}, |
|
{ |
|
"epoch": 10.41, |
|
"learning_rate": 1.3331515926772495e-05, |
|
"loss": 1.3701, |
|
"step": 169000 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 1.3310545851070525e-05, |
|
"loss": 1.3694, |
|
"step": 169500 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 1.3289575775368551e-05, |
|
"loss": 1.3724, |
|
"step": 170000 |
|
}, |
|
{ |
|
"epoch": 10.51, |
|
"learning_rate": 1.3268605699666578e-05, |
|
"loss": 1.3705, |
|
"step": 170500 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 1.3247635623964604e-05, |
|
"loss": 1.3749, |
|
"step": 171000 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 1.322666554826263e-05, |
|
"loss": 1.3747, |
|
"step": 171500 |
|
}, |
|
{ |
|
"epoch": 10.6, |
|
"learning_rate": 1.3205695472560658e-05, |
|
"loss": 1.3645, |
|
"step": 172000 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 1.3184725396858684e-05, |
|
"loss": 1.3676, |
|
"step": 172500 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 1.316375532115671e-05, |
|
"loss": 1.3634, |
|
"step": 173000 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 1.3142785245454737e-05, |
|
"loss": 1.3682, |
|
"step": 173500 |
|
}, |
|
{ |
|
"epoch": 10.72, |
|
"learning_rate": 1.3121815169752763e-05, |
|
"loss": 1.366, |
|
"step": 174000 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 1.3100845094050791e-05, |
|
"loss": 1.3656, |
|
"step": 174500 |
|
}, |
|
{ |
|
"epoch": 10.78, |
|
"learning_rate": 1.3079875018348817e-05, |
|
"loss": 1.36, |
|
"step": 175000 |
|
}, |
|
{ |
|
"epoch": 10.81, |
|
"learning_rate": 1.3058904942646843e-05, |
|
"loss": 1.3637, |
|
"step": 175500 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"learning_rate": 1.303793486694487e-05, |
|
"loss": 1.3677, |
|
"step": 176000 |
|
}, |
|
{ |
|
"epoch": 10.88, |
|
"learning_rate": 1.3016964791242896e-05, |
|
"loss": 1.3574, |
|
"step": 176500 |
|
}, |
|
{ |
|
"epoch": 10.91, |
|
"learning_rate": 1.2995994715540926e-05, |
|
"loss": 1.3666, |
|
"step": 177000 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 1.2975024639838952e-05, |
|
"loss": 1.3712, |
|
"step": 177500 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 1.2954054564136978e-05, |
|
"loss": 1.3607, |
|
"step": 178000 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 1.2933084488435004e-05, |
|
"loss": 1.3613, |
|
"step": 178500 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 1.291211441273303e-05, |
|
"loss": 1.3596, |
|
"step": 179000 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 1.2891144337031058e-05, |
|
"loss": 1.3634, |
|
"step": 179500 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 1.2870174261329085e-05, |
|
"loss": 1.365, |
|
"step": 180000 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 1.2849204185627111e-05, |
|
"loss": 1.3575, |
|
"step": 180500 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 1.2828234109925137e-05, |
|
"loss": 1.3589, |
|
"step": 181000 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 1.2807264034223163e-05, |
|
"loss": 1.3601, |
|
"step": 181500 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 1.2786293958521191e-05, |
|
"loss": 1.3621, |
|
"step": 182000 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 1.2765323882819218e-05, |
|
"loss": 1.3605, |
|
"step": 182500 |
|
}, |
|
{ |
|
"epoch": 11.28, |
|
"learning_rate": 1.2744353807117244e-05, |
|
"loss": 1.3596, |
|
"step": 183000 |
|
}, |
|
{ |
|
"epoch": 11.31, |
|
"learning_rate": 1.272338373141527e-05, |
|
"loss": 1.3597, |
|
"step": 183500 |
|
}, |
|
{ |
|
"epoch": 11.34, |
|
"learning_rate": 1.2702413655713296e-05, |
|
"loss": 1.3599, |
|
"step": 184000 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 1.2681443580011326e-05, |
|
"loss": 1.3595, |
|
"step": 184500 |
|
}, |
|
{ |
|
"epoch": 11.4, |
|
"learning_rate": 1.2660473504309352e-05, |
|
"loss": 1.3613, |
|
"step": 185000 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"learning_rate": 1.2639503428607379e-05, |
|
"loss": 1.3594, |
|
"step": 185500 |
|
}, |
|
{ |
|
"epoch": 11.46, |
|
"learning_rate": 1.2618533352905405e-05, |
|
"loss": 1.358, |
|
"step": 186000 |
|
}, |
|
{ |
|
"epoch": 11.49, |
|
"learning_rate": 1.2597563277203431e-05, |
|
"loss": 1.3542, |
|
"step": 186500 |
|
}, |
|
{ |
|
"epoch": 11.52, |
|
"learning_rate": 1.2576593201501459e-05, |
|
"loss": 1.3655, |
|
"step": 187000 |
|
}, |
|
{ |
|
"epoch": 11.55, |
|
"learning_rate": 1.2555623125799485e-05, |
|
"loss": 1.3545, |
|
"step": 187500 |
|
}, |
|
{ |
|
"epoch": 11.58, |
|
"learning_rate": 1.2534653050097511e-05, |
|
"loss": 1.3558, |
|
"step": 188000 |
|
}, |
|
{ |
|
"epoch": 11.62, |
|
"learning_rate": 1.2513682974395538e-05, |
|
"loss": 1.3526, |
|
"step": 188500 |
|
}, |
|
{ |
|
"epoch": 11.65, |
|
"learning_rate": 1.2492712898693566e-05, |
|
"loss": 1.3565, |
|
"step": 189000 |
|
}, |
|
{ |
|
"epoch": 11.68, |
|
"learning_rate": 1.2471742822991592e-05, |
|
"loss": 1.3515, |
|
"step": 189500 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 1.2450772747289618e-05, |
|
"loss": 1.3506, |
|
"step": 190000 |
|
}, |
|
{ |
|
"epoch": 11.74, |
|
"learning_rate": 1.2429802671587644e-05, |
|
"loss": 1.3578, |
|
"step": 190500 |
|
}, |
|
{ |
|
"epoch": 11.77, |
|
"learning_rate": 1.240883259588567e-05, |
|
"loss": 1.3566, |
|
"step": 191000 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 1.23878625201837e-05, |
|
"loss": 1.3566, |
|
"step": 191500 |
|
}, |
|
{ |
|
"epoch": 11.83, |
|
"learning_rate": 1.2366892444481727e-05, |
|
"loss": 1.3562, |
|
"step": 192000 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 1.2345922368779753e-05, |
|
"loss": 1.3509, |
|
"step": 192500 |
|
}, |
|
{ |
|
"epoch": 11.89, |
|
"learning_rate": 1.2324952293077779e-05, |
|
"loss": 1.3492, |
|
"step": 193000 |
|
}, |
|
{ |
|
"epoch": 11.92, |
|
"learning_rate": 1.2303982217375805e-05, |
|
"loss": 1.3487, |
|
"step": 193500 |
|
}, |
|
{ |
|
"epoch": 11.95, |
|
"learning_rate": 1.2283012141673833e-05, |
|
"loss": 1.3525, |
|
"step": 194000 |
|
}, |
|
{ |
|
"epoch": 11.98, |
|
"learning_rate": 1.226204206597186e-05, |
|
"loss": 1.3531, |
|
"step": 194500 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 1.2241071990269886e-05, |
|
"loss": 1.3475, |
|
"step": 195000 |
|
}, |
|
{ |
|
"epoch": 12.05, |
|
"learning_rate": 1.2220101914567912e-05, |
|
"loss": 1.3472, |
|
"step": 195500 |
|
}, |
|
{ |
|
"epoch": 12.08, |
|
"learning_rate": 1.2199131838865938e-05, |
|
"loss": 1.3536, |
|
"step": 196000 |
|
}, |
|
{ |
|
"epoch": 12.11, |
|
"learning_rate": 1.2178161763163968e-05, |
|
"loss": 1.3525, |
|
"step": 196500 |
|
}, |
|
{ |
|
"epoch": 12.14, |
|
"learning_rate": 1.2157191687461994e-05, |
|
"loss": 1.3526, |
|
"step": 197000 |
|
}, |
|
{ |
|
"epoch": 12.17, |
|
"learning_rate": 1.213622161176002e-05, |
|
"loss": 1.3515, |
|
"step": 197500 |
|
}, |
|
{ |
|
"epoch": 12.2, |
|
"learning_rate": 1.2115251536058047e-05, |
|
"loss": 1.3457, |
|
"step": 198000 |
|
}, |
|
{ |
|
"epoch": 12.23, |
|
"learning_rate": 1.2094281460356073e-05, |
|
"loss": 1.3538, |
|
"step": 198500 |
|
}, |
|
{ |
|
"epoch": 12.26, |
|
"learning_rate": 1.20733113846541e-05, |
|
"loss": 1.3506, |
|
"step": 199000 |
|
}, |
|
{ |
|
"epoch": 12.29, |
|
"learning_rate": 1.2052341308952127e-05, |
|
"loss": 1.3438, |
|
"step": 199500 |
|
}, |
|
{ |
|
"epoch": 12.32, |
|
"learning_rate": 1.2031371233250153e-05, |
|
"loss": 1.3471, |
|
"step": 200000 |
|
}, |
|
{ |
|
"epoch": 12.35, |
|
"learning_rate": 1.201040115754818e-05, |
|
"loss": 1.352, |
|
"step": 200500 |
|
}, |
|
{ |
|
"epoch": 12.39, |
|
"learning_rate": 1.1989431081846206e-05, |
|
"loss": 1.3458, |
|
"step": 201000 |
|
}, |
|
{ |
|
"epoch": 12.42, |
|
"learning_rate": 1.1968461006144234e-05, |
|
"loss": 1.3539, |
|
"step": 201500 |
|
}, |
|
{ |
|
"epoch": 12.45, |
|
"learning_rate": 1.194749093044226e-05, |
|
"loss": 1.3478, |
|
"step": 202000 |
|
}, |
|
{ |
|
"epoch": 12.48, |
|
"learning_rate": 1.1926520854740286e-05, |
|
"loss": 1.347, |
|
"step": 202500 |
|
}, |
|
{ |
|
"epoch": 12.51, |
|
"learning_rate": 1.1905550779038312e-05, |
|
"loss": 1.3503, |
|
"step": 203000 |
|
}, |
|
{ |
|
"epoch": 12.54, |
|
"learning_rate": 1.1884580703336339e-05, |
|
"loss": 1.3459, |
|
"step": 203500 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 1.1863610627634368e-05, |
|
"loss": 1.3464, |
|
"step": 204000 |
|
}, |
|
{ |
|
"epoch": 12.6, |
|
"learning_rate": 1.1842640551932395e-05, |
|
"loss": 1.3463, |
|
"step": 204500 |
|
}, |
|
{ |
|
"epoch": 12.63, |
|
"learning_rate": 1.182167047623042e-05, |
|
"loss": 1.3512, |
|
"step": 205000 |
|
}, |
|
{ |
|
"epoch": 12.66, |
|
"learning_rate": 1.1800700400528447e-05, |
|
"loss": 1.344, |
|
"step": 205500 |
|
}, |
|
{ |
|
"epoch": 12.69, |
|
"learning_rate": 1.1779730324826473e-05, |
|
"loss": 1.3481, |
|
"step": 206000 |
|
}, |
|
{ |
|
"epoch": 12.72, |
|
"learning_rate": 1.1758760249124501e-05, |
|
"loss": 1.3392, |
|
"step": 206500 |
|
}, |
|
{ |
|
"epoch": 12.75, |
|
"learning_rate": 1.1737790173422527e-05, |
|
"loss": 1.3333, |
|
"step": 207000 |
|
}, |
|
{ |
|
"epoch": 12.79, |
|
"learning_rate": 1.1716820097720554e-05, |
|
"loss": 1.3428, |
|
"step": 207500 |
|
}, |
|
{ |
|
"epoch": 12.82, |
|
"learning_rate": 1.169585002201858e-05, |
|
"loss": 1.3449, |
|
"step": 208000 |
|
}, |
|
{ |
|
"epoch": 12.85, |
|
"learning_rate": 1.1674879946316606e-05, |
|
"loss": 1.3458, |
|
"step": 208500 |
|
}, |
|
{ |
|
"epoch": 12.88, |
|
"learning_rate": 1.1653909870614634e-05, |
|
"loss": 1.3378, |
|
"step": 209000 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 1.163293979491266e-05, |
|
"loss": 1.3392, |
|
"step": 209500 |
|
}, |
|
{ |
|
"epoch": 12.94, |
|
"learning_rate": 1.1611969719210687e-05, |
|
"loss": 1.3436, |
|
"step": 210000 |
|
}, |
|
{ |
|
"epoch": 12.97, |
|
"learning_rate": 1.1590999643508713e-05, |
|
"loss": 1.3411, |
|
"step": 210500 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 1.1570029567806739e-05, |
|
"loss": 1.3425, |
|
"step": 211000 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 1.1549059492104769e-05, |
|
"loss": 1.3406, |
|
"step": 211500 |
|
}, |
|
{ |
|
"epoch": 13.06, |
|
"learning_rate": 1.1528089416402795e-05, |
|
"loss": 1.3393, |
|
"step": 212000 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 1.1507119340700821e-05, |
|
"loss": 1.3422, |
|
"step": 212500 |
|
}, |
|
{ |
|
"epoch": 13.12, |
|
"learning_rate": 1.1486149264998848e-05, |
|
"loss": 1.3359, |
|
"step": 213000 |
|
}, |
|
{ |
|
"epoch": 13.16, |
|
"learning_rate": 1.1465179189296874e-05, |
|
"loss": 1.335, |
|
"step": 213500 |
|
}, |
|
{ |
|
"epoch": 13.19, |
|
"learning_rate": 1.1444209113594902e-05, |
|
"loss": 1.3396, |
|
"step": 214000 |
|
}, |
|
{ |
|
"epoch": 13.22, |
|
"learning_rate": 1.1423239037892928e-05, |
|
"loss": 1.3373, |
|
"step": 214500 |
|
}, |
|
{ |
|
"epoch": 13.25, |
|
"learning_rate": 1.1402268962190954e-05, |
|
"loss": 1.3377, |
|
"step": 215000 |
|
}, |
|
{ |
|
"epoch": 13.28, |
|
"learning_rate": 1.138129888648898e-05, |
|
"loss": 1.3372, |
|
"step": 215500 |
|
}, |
|
{ |
|
"epoch": 13.31, |
|
"learning_rate": 1.1360328810787007e-05, |
|
"loss": 1.3281, |
|
"step": 216000 |
|
}, |
|
{ |
|
"epoch": 13.34, |
|
"learning_rate": 1.1339358735085035e-05, |
|
"loss": 1.3378, |
|
"step": 216500 |
|
}, |
|
{ |
|
"epoch": 13.37, |
|
"learning_rate": 1.1318388659383061e-05, |
|
"loss": 1.3313, |
|
"step": 217000 |
|
}, |
|
{ |
|
"epoch": 13.4, |
|
"learning_rate": 1.1297418583681087e-05, |
|
"loss": 1.3396, |
|
"step": 217500 |
|
}, |
|
{ |
|
"epoch": 13.43, |
|
"learning_rate": 1.1276448507979113e-05, |
|
"loss": 1.3375, |
|
"step": 218000 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 1.125547843227714e-05, |
|
"loss": 1.3345, |
|
"step": 218500 |
|
}, |
|
{ |
|
"epoch": 13.49, |
|
"learning_rate": 1.123450835657517e-05, |
|
"loss": 1.336, |
|
"step": 219000 |
|
}, |
|
{ |
|
"epoch": 13.53, |
|
"learning_rate": 1.1213538280873195e-05, |
|
"loss": 1.336, |
|
"step": 219500 |
|
}, |
|
{ |
|
"epoch": 13.56, |
|
"learning_rate": 1.1192568205171222e-05, |
|
"loss": 1.3383, |
|
"step": 220000 |
|
}, |
|
{ |
|
"epoch": 13.59, |
|
"learning_rate": 1.1171598129469248e-05, |
|
"loss": 1.3372, |
|
"step": 220500 |
|
}, |
|
{ |
|
"epoch": 13.62, |
|
"learning_rate": 1.1150628053767274e-05, |
|
"loss": 1.33, |
|
"step": 221000 |
|
}, |
|
{ |
|
"epoch": 13.65, |
|
"learning_rate": 1.1129657978065302e-05, |
|
"loss": 1.3366, |
|
"step": 221500 |
|
}, |
|
{ |
|
"epoch": 13.68, |
|
"learning_rate": 1.1108687902363328e-05, |
|
"loss": 1.3339, |
|
"step": 222000 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 1.1087717826661355e-05, |
|
"loss": 1.3392, |
|
"step": 222500 |
|
}, |
|
{ |
|
"epoch": 13.74, |
|
"learning_rate": 1.1066747750959381e-05, |
|
"loss": 1.3307, |
|
"step": 223000 |
|
}, |
|
{ |
|
"epoch": 13.77, |
|
"learning_rate": 1.1045777675257407e-05, |
|
"loss": 1.3412, |
|
"step": 223500 |
|
}, |
|
{ |
|
"epoch": 13.8, |
|
"learning_rate": 1.1024807599555437e-05, |
|
"loss": 1.339, |
|
"step": 224000 |
|
}, |
|
{ |
|
"epoch": 13.83, |
|
"learning_rate": 1.1003837523853463e-05, |
|
"loss": 1.336, |
|
"step": 224500 |
|
}, |
|
{ |
|
"epoch": 13.86, |
|
"learning_rate": 1.098286744815149e-05, |
|
"loss": 1.3337, |
|
"step": 225000 |
|
}, |
|
{ |
|
"epoch": 13.89, |
|
"learning_rate": 1.0961897372449516e-05, |
|
"loss": 1.3322, |
|
"step": 225500 |
|
}, |
|
{ |
|
"epoch": 13.93, |
|
"learning_rate": 1.0940927296747542e-05, |
|
"loss": 1.3313, |
|
"step": 226000 |
|
}, |
|
{ |
|
"epoch": 13.96, |
|
"learning_rate": 1.091995722104557e-05, |
|
"loss": 1.3289, |
|
"step": 226500 |
|
}, |
|
{ |
|
"epoch": 13.99, |
|
"learning_rate": 1.0898987145343596e-05, |
|
"loss": 1.3371, |
|
"step": 227000 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 1.0878017069641622e-05, |
|
"loss": 1.3318, |
|
"step": 227500 |
|
}, |
|
{ |
|
"epoch": 14.05, |
|
"learning_rate": 1.0857046993939648e-05, |
|
"loss": 1.3302, |
|
"step": 228000 |
|
}, |
|
{ |
|
"epoch": 14.08, |
|
"learning_rate": 1.0836076918237676e-05, |
|
"loss": 1.3276, |
|
"step": 228500 |
|
}, |
|
{ |
|
"epoch": 14.11, |
|
"learning_rate": 1.0815106842535703e-05, |
|
"loss": 1.3294, |
|
"step": 229000 |
|
}, |
|
{ |
|
"epoch": 14.14, |
|
"learning_rate": 1.0794136766833729e-05, |
|
"loss": 1.3297, |
|
"step": 229500 |
|
}, |
|
{ |
|
"epoch": 14.17, |
|
"learning_rate": 1.0773166691131755e-05, |
|
"loss": 1.3238, |
|
"step": 230000 |
|
}, |
|
{ |
|
"epoch": 14.2, |
|
"learning_rate": 1.0752196615429781e-05, |
|
"loss": 1.3286, |
|
"step": 230500 |
|
}, |
|
{ |
|
"epoch": 14.23, |
|
"learning_rate": 1.0731226539727811e-05, |
|
"loss": 1.3285, |
|
"step": 231000 |
|
}, |
|
{ |
|
"epoch": 14.26, |
|
"learning_rate": 1.0710256464025837e-05, |
|
"loss": 1.3294, |
|
"step": 231500 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 1.0689286388323864e-05, |
|
"loss": 1.3311, |
|
"step": 232000 |
|
}, |
|
{ |
|
"epoch": 14.33, |
|
"learning_rate": 1.066831631262189e-05, |
|
"loss": 1.3265, |
|
"step": 232500 |
|
}, |
|
{ |
|
"epoch": 14.36, |
|
"learning_rate": 1.0647346236919916e-05, |
|
"loss": 1.3249, |
|
"step": 233000 |
|
}, |
|
{ |
|
"epoch": 14.39, |
|
"learning_rate": 1.0626376161217944e-05, |
|
"loss": 1.3237, |
|
"step": 233500 |
|
}, |
|
{ |
|
"epoch": 14.42, |
|
"learning_rate": 1.060540608551597e-05, |
|
"loss": 1.3308, |
|
"step": 234000 |
|
}, |
|
{ |
|
"epoch": 14.45, |
|
"learning_rate": 1.0584436009813996e-05, |
|
"loss": 1.3268, |
|
"step": 234500 |
|
}, |
|
{ |
|
"epoch": 14.48, |
|
"learning_rate": 1.0563465934112023e-05, |
|
"loss": 1.3246, |
|
"step": 235000 |
|
}, |
|
{ |
|
"epoch": 14.51, |
|
"learning_rate": 1.0542495858410049e-05, |
|
"loss": 1.3308, |
|
"step": 235500 |
|
}, |
|
{ |
|
"epoch": 14.54, |
|
"learning_rate": 1.0521525782708077e-05, |
|
"loss": 1.3237, |
|
"step": 236000 |
|
}, |
|
{ |
|
"epoch": 14.57, |
|
"learning_rate": 1.0500555707006103e-05, |
|
"loss": 1.33, |
|
"step": 236500 |
|
}, |
|
{ |
|
"epoch": 14.6, |
|
"learning_rate": 1.047958563130413e-05, |
|
"loss": 1.3233, |
|
"step": 237000 |
|
}, |
|
{ |
|
"epoch": 14.63, |
|
"learning_rate": 1.0458615555602156e-05, |
|
"loss": 1.3297, |
|
"step": 237500 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"learning_rate": 1.0437645479900182e-05, |
|
"loss": 1.3294, |
|
"step": 238000 |
|
}, |
|
{ |
|
"epoch": 14.7, |
|
"learning_rate": 1.0416675404198212e-05, |
|
"loss": 1.3229, |
|
"step": 238500 |
|
}, |
|
{ |
|
"epoch": 14.73, |
|
"learning_rate": 1.0395705328496238e-05, |
|
"loss": 1.3252, |
|
"step": 239000 |
|
}, |
|
{ |
|
"epoch": 14.76, |
|
"learning_rate": 1.0374735252794264e-05, |
|
"loss": 1.3222, |
|
"step": 239500 |
|
}, |
|
{ |
|
"epoch": 14.79, |
|
"learning_rate": 1.035376517709229e-05, |
|
"loss": 1.3243, |
|
"step": 240000 |
|
}, |
|
{ |
|
"epoch": 14.82, |
|
"learning_rate": 1.0332795101390316e-05, |
|
"loss": 1.3217, |
|
"step": 240500 |
|
}, |
|
{ |
|
"epoch": 14.85, |
|
"learning_rate": 1.0311825025688344e-05, |
|
"loss": 1.3261, |
|
"step": 241000 |
|
}, |
|
{ |
|
"epoch": 14.88, |
|
"learning_rate": 1.029085494998637e-05, |
|
"loss": 1.3253, |
|
"step": 241500 |
|
}, |
|
{ |
|
"epoch": 14.91, |
|
"learning_rate": 1.0269884874284397e-05, |
|
"loss": 1.3211, |
|
"step": 242000 |
|
}, |
|
{ |
|
"epoch": 14.94, |
|
"learning_rate": 1.0248914798582423e-05, |
|
"loss": 1.325, |
|
"step": 242500 |
|
}, |
|
{ |
|
"epoch": 14.97, |
|
"learning_rate": 1.022794472288045e-05, |
|
"loss": 1.331, |
|
"step": 243000 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 1.0206974647178477e-05, |
|
"loss": 1.3201, |
|
"step": 243500 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 1.0186004571476504e-05, |
|
"loss": 1.3229, |
|
"step": 244000 |
|
}, |
|
{ |
|
"epoch": 15.07, |
|
"learning_rate": 1.016503449577453e-05, |
|
"loss": 1.3253, |
|
"step": 244500 |
|
}, |
|
{ |
|
"epoch": 15.1, |
|
"learning_rate": 1.0144064420072556e-05, |
|
"loss": 1.3235, |
|
"step": 245000 |
|
}, |
|
{ |
|
"epoch": 15.13, |
|
"learning_rate": 1.0123094344370582e-05, |
|
"loss": 1.3269, |
|
"step": 245500 |
|
}, |
|
{ |
|
"epoch": 15.16, |
|
"learning_rate": 1.0102124268668612e-05, |
|
"loss": 1.3211, |
|
"step": 246000 |
|
}, |
|
{ |
|
"epoch": 15.19, |
|
"learning_rate": 1.0081154192966638e-05, |
|
"loss": 1.3222, |
|
"step": 246500 |
|
}, |
|
{ |
|
"epoch": 15.22, |
|
"learning_rate": 1.0060184117264664e-05, |
|
"loss": 1.3162, |
|
"step": 247000 |
|
}, |
|
{ |
|
"epoch": 15.25, |
|
"learning_rate": 1.003921404156269e-05, |
|
"loss": 1.3243, |
|
"step": 247500 |
|
}, |
|
{ |
|
"epoch": 15.28, |
|
"learning_rate": 1.0018243965860717e-05, |
|
"loss": 1.3196, |
|
"step": 248000 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"learning_rate": 9.997273890158743e-06, |
|
"loss": 1.3228, |
|
"step": 248500 |
|
}, |
|
{ |
|
"epoch": 15.34, |
|
"learning_rate": 9.976303814456771e-06, |
|
"loss": 1.3166, |
|
"step": 249000 |
|
}, |
|
{ |
|
"epoch": 15.37, |
|
"learning_rate": 9.955333738754797e-06, |
|
"loss": 1.3183, |
|
"step": 249500 |
|
}, |
|
{ |
|
"epoch": 15.4, |
|
"learning_rate": 9.934363663052824e-06, |
|
"loss": 1.3199, |
|
"step": 250000 |
|
}, |
|
{ |
|
"epoch": 15.44, |
|
"learning_rate": 9.913393587350852e-06, |
|
"loss": 1.316, |
|
"step": 250500 |
|
}, |
|
{ |
|
"epoch": 15.47, |
|
"learning_rate": 9.892423511648878e-06, |
|
"loss": 1.3208, |
|
"step": 251000 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"learning_rate": 9.871453435946906e-06, |
|
"loss": 1.3207, |
|
"step": 251500 |
|
}, |
|
{ |
|
"epoch": 15.53, |
|
"learning_rate": 9.850483360244932e-06, |
|
"loss": 1.3233, |
|
"step": 252000 |
|
}, |
|
{ |
|
"epoch": 15.56, |
|
"learning_rate": 9.829513284542958e-06, |
|
"loss": 1.3193, |
|
"step": 252500 |
|
}, |
|
{ |
|
"epoch": 15.59, |
|
"learning_rate": 9.808543208840985e-06, |
|
"loss": 1.3207, |
|
"step": 253000 |
|
}, |
|
{ |
|
"epoch": 15.62, |
|
"learning_rate": 9.78757313313901e-06, |
|
"loss": 1.3271, |
|
"step": 253500 |
|
}, |
|
{ |
|
"epoch": 15.65, |
|
"learning_rate": 9.766603057437039e-06, |
|
"loss": 1.3233, |
|
"step": 254000 |
|
}, |
|
{ |
|
"epoch": 15.68, |
|
"learning_rate": 9.745632981735065e-06, |
|
"loss": 1.316, |
|
"step": 254500 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"learning_rate": 9.724662906033091e-06, |
|
"loss": 1.3188, |
|
"step": 255000 |
|
}, |
|
{ |
|
"epoch": 15.74, |
|
"learning_rate": 9.703692830331119e-06, |
|
"loss": 1.3183, |
|
"step": 255500 |
|
}, |
|
{ |
|
"epoch": 15.77, |
|
"learning_rate": 9.682722754629145e-06, |
|
"loss": 1.3248, |
|
"step": 256000 |
|
}, |
|
{ |
|
"epoch": 15.81, |
|
"learning_rate": 9.661752678927172e-06, |
|
"loss": 1.3113, |
|
"step": 256500 |
|
}, |
|
{ |
|
"epoch": 15.84, |
|
"learning_rate": 9.640782603225198e-06, |
|
"loss": 1.3195, |
|
"step": 257000 |
|
}, |
|
{ |
|
"epoch": 15.87, |
|
"learning_rate": 9.619812527523226e-06, |
|
"loss": 1.3179, |
|
"step": 257500 |
|
}, |
|
{ |
|
"epoch": 15.9, |
|
"learning_rate": 9.598842451821252e-06, |
|
"loss": 1.3163, |
|
"step": 258000 |
|
}, |
|
{ |
|
"epoch": 15.93, |
|
"learning_rate": 9.577872376119278e-06, |
|
"loss": 1.3124, |
|
"step": 258500 |
|
}, |
|
{ |
|
"epoch": 15.96, |
|
"learning_rate": 9.556902300417306e-06, |
|
"loss": 1.3143, |
|
"step": 259000 |
|
}, |
|
{ |
|
"epoch": 15.99, |
|
"learning_rate": 9.535932224715332e-06, |
|
"loss": 1.31, |
|
"step": 259500 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 9.514962149013359e-06, |
|
"loss": 1.3128, |
|
"step": 260000 |
|
}, |
|
{ |
|
"epoch": 16.05, |
|
"learning_rate": 9.493992073311385e-06, |
|
"loss": 1.3154, |
|
"step": 260500 |
|
}, |
|
{ |
|
"epoch": 16.08, |
|
"learning_rate": 9.473021997609411e-06, |
|
"loss": 1.3104, |
|
"step": 261000 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 9.45205192190744e-06, |
|
"loss": 1.3185, |
|
"step": 261500 |
|
}, |
|
{ |
|
"epoch": 16.14, |
|
"learning_rate": 9.431081846205465e-06, |
|
"loss": 1.3144, |
|
"step": 262000 |
|
}, |
|
{ |
|
"epoch": 16.17, |
|
"learning_rate": 9.410111770503493e-06, |
|
"loss": 1.318, |
|
"step": 262500 |
|
}, |
|
{ |
|
"epoch": 16.21, |
|
"learning_rate": 9.38914169480152e-06, |
|
"loss": 1.3094, |
|
"step": 263000 |
|
}, |
|
{ |
|
"epoch": 16.24, |
|
"learning_rate": 9.368171619099546e-06, |
|
"loss": 1.3149, |
|
"step": 263500 |
|
}, |
|
{ |
|
"epoch": 16.27, |
|
"learning_rate": 9.347201543397572e-06, |
|
"loss": 1.3144, |
|
"step": 264000 |
|
}, |
|
{ |
|
"epoch": 16.3, |
|
"learning_rate": 9.326231467695598e-06, |
|
"loss": 1.3162, |
|
"step": 264500 |
|
}, |
|
{ |
|
"epoch": 16.33, |
|
"learning_rate": 9.305261391993626e-06, |
|
"loss": 1.3158, |
|
"step": 265000 |
|
}, |
|
{ |
|
"epoch": 16.36, |
|
"learning_rate": 9.284291316291653e-06, |
|
"loss": 1.3096, |
|
"step": 265500 |
|
}, |
|
{ |
|
"epoch": 16.39, |
|
"learning_rate": 9.263321240589679e-06, |
|
"loss": 1.3145, |
|
"step": 266000 |
|
}, |
|
{ |
|
"epoch": 16.42, |
|
"learning_rate": 9.242351164887707e-06, |
|
"loss": 1.3126, |
|
"step": 266500 |
|
}, |
|
{ |
|
"epoch": 16.45, |
|
"learning_rate": 9.221381089185733e-06, |
|
"loss": 1.3086, |
|
"step": 267000 |
|
}, |
|
{ |
|
"epoch": 16.48, |
|
"learning_rate": 9.20041101348376e-06, |
|
"loss": 1.3075, |
|
"step": 267500 |
|
}, |
|
{ |
|
"epoch": 16.51, |
|
"learning_rate": 9.179440937781785e-06, |
|
"loss": 1.3131, |
|
"step": 268000 |
|
}, |
|
{ |
|
"epoch": 16.54, |
|
"learning_rate": 9.158470862079812e-06, |
|
"loss": 1.312, |
|
"step": 268500 |
|
}, |
|
{ |
|
"epoch": 16.58, |
|
"learning_rate": 9.13750078637784e-06, |
|
"loss": 1.317, |
|
"step": 269000 |
|
}, |
|
{ |
|
"epoch": 16.61, |
|
"learning_rate": 9.116530710675866e-06, |
|
"loss": 1.3111, |
|
"step": 269500 |
|
}, |
|
{ |
|
"epoch": 16.64, |
|
"learning_rate": 9.095560634973894e-06, |
|
"loss": 1.3057, |
|
"step": 270000 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"learning_rate": 9.07459055927192e-06, |
|
"loss": 1.311, |
|
"step": 270500 |
|
}, |
|
{ |
|
"epoch": 16.7, |
|
"learning_rate": 9.053620483569946e-06, |
|
"loss": 1.3112, |
|
"step": 271000 |
|
}, |
|
{ |
|
"epoch": 16.73, |
|
"learning_rate": 9.032650407867973e-06, |
|
"loss": 1.3145, |
|
"step": 271500 |
|
}, |
|
{ |
|
"epoch": 16.76, |
|
"learning_rate": 9.011680332165999e-06, |
|
"loss": 1.3092, |
|
"step": 272000 |
|
}, |
|
{ |
|
"epoch": 16.79, |
|
"learning_rate": 8.990710256464027e-06, |
|
"loss": 1.3128, |
|
"step": 272500 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 8.969740180762053e-06, |
|
"loss": 1.3098, |
|
"step": 273000 |
|
}, |
|
{ |
|
"epoch": 16.85, |
|
"learning_rate": 8.94877010506008e-06, |
|
"loss": 1.3087, |
|
"step": 273500 |
|
}, |
|
{ |
|
"epoch": 16.88, |
|
"learning_rate": 8.927800029358107e-06, |
|
"loss": 1.309, |
|
"step": 274000 |
|
}, |
|
{ |
|
"epoch": 16.91, |
|
"learning_rate": 8.906829953656133e-06, |
|
"loss": 1.3106, |
|
"step": 274500 |
|
}, |
|
{ |
|
"epoch": 16.94, |
|
"learning_rate": 8.88585987795416e-06, |
|
"loss": 1.3093, |
|
"step": 275000 |
|
}, |
|
{ |
|
"epoch": 16.98, |
|
"learning_rate": 8.864889802252186e-06, |
|
"loss": 1.3075, |
|
"step": 275500 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 8.843919726550212e-06, |
|
"loss": 1.3093, |
|
"step": 276000 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 8.82294965084824e-06, |
|
"loss": 1.3107, |
|
"step": 276500 |
|
}, |
|
{ |
|
"epoch": 17.07, |
|
"learning_rate": 8.801979575146266e-06, |
|
"loss": 1.3104, |
|
"step": 277000 |
|
}, |
|
{ |
|
"epoch": 17.1, |
|
"learning_rate": 8.781009499444294e-06, |
|
"loss": 1.3117, |
|
"step": 277500 |
|
}, |
|
{ |
|
"epoch": 17.13, |
|
"learning_rate": 8.76003942374232e-06, |
|
"loss": 1.3042, |
|
"step": 278000 |
|
}, |
|
{ |
|
"epoch": 17.16, |
|
"learning_rate": 8.739069348040349e-06, |
|
"loss": 1.3026, |
|
"step": 278500 |
|
}, |
|
{ |
|
"epoch": 17.19, |
|
"learning_rate": 8.718099272338375e-06, |
|
"loss": 1.3066, |
|
"step": 279000 |
|
}, |
|
{ |
|
"epoch": 17.22, |
|
"learning_rate": 8.697129196636401e-06, |
|
"loss": 1.2996, |
|
"step": 279500 |
|
}, |
|
{ |
|
"epoch": 17.25, |
|
"learning_rate": 8.676159120934427e-06, |
|
"loss": 1.3066, |
|
"step": 280000 |
|
}, |
|
{ |
|
"epoch": 17.28, |
|
"learning_rate": 8.655189045232453e-06, |
|
"loss": 1.3032, |
|
"step": 280500 |
|
}, |
|
{ |
|
"epoch": 17.31, |
|
"learning_rate": 8.634218969530481e-06, |
|
"loss": 1.3055, |
|
"step": 281000 |
|
}, |
|
{ |
|
"epoch": 17.35, |
|
"learning_rate": 8.613248893828508e-06, |
|
"loss": 1.3061, |
|
"step": 281500 |
|
}, |
|
{ |
|
"epoch": 17.38, |
|
"learning_rate": 8.592278818126534e-06, |
|
"loss": 1.3043, |
|
"step": 282000 |
|
}, |
|
{ |
|
"epoch": 17.41, |
|
"learning_rate": 8.571308742424562e-06, |
|
"loss": 1.3089, |
|
"step": 282500 |
|
}, |
|
{ |
|
"epoch": 17.44, |
|
"learning_rate": 8.550338666722588e-06, |
|
"loss": 1.3078, |
|
"step": 283000 |
|
}, |
|
{ |
|
"epoch": 17.47, |
|
"learning_rate": 8.529368591020614e-06, |
|
"loss": 1.3042, |
|
"step": 283500 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 8.50839851531864e-06, |
|
"loss": 1.3045, |
|
"step": 284000 |
|
}, |
|
{ |
|
"epoch": 17.53, |
|
"learning_rate": 8.487428439616667e-06, |
|
"loss": 1.3074, |
|
"step": 284500 |
|
}, |
|
{ |
|
"epoch": 17.56, |
|
"learning_rate": 8.466458363914695e-06, |
|
"loss": 1.3, |
|
"step": 285000 |
|
}, |
|
{ |
|
"epoch": 17.59, |
|
"learning_rate": 8.445488288212721e-06, |
|
"loss": 1.3044, |
|
"step": 285500 |
|
}, |
|
{ |
|
"epoch": 17.62, |
|
"learning_rate": 8.424518212510749e-06, |
|
"loss": 1.3092, |
|
"step": 286000 |
|
}, |
|
{ |
|
"epoch": 17.65, |
|
"learning_rate": 8.403548136808775e-06, |
|
"loss": 1.2999, |
|
"step": 286500 |
|
}, |
|
{ |
|
"epoch": 17.68, |
|
"learning_rate": 8.382578061106801e-06, |
|
"loss": 1.3109, |
|
"step": 287000 |
|
}, |
|
{ |
|
"epoch": 17.72, |
|
"learning_rate": 8.361607985404828e-06, |
|
"loss": 1.3071, |
|
"step": 287500 |
|
}, |
|
{ |
|
"epoch": 17.75, |
|
"learning_rate": 8.340637909702854e-06, |
|
"loss": 1.3054, |
|
"step": 288000 |
|
}, |
|
{ |
|
"epoch": 17.78, |
|
"learning_rate": 8.319667834000882e-06, |
|
"loss": 1.3018, |
|
"step": 288500 |
|
}, |
|
{ |
|
"epoch": 17.81, |
|
"learning_rate": 8.298697758298908e-06, |
|
"loss": 1.3061, |
|
"step": 289000 |
|
}, |
|
{ |
|
"epoch": 17.84, |
|
"learning_rate": 8.277727682596934e-06, |
|
"loss": 1.3032, |
|
"step": 289500 |
|
}, |
|
{ |
|
"epoch": 17.87, |
|
"learning_rate": 8.256757606894962e-06, |
|
"loss": 1.3088, |
|
"step": 290000 |
|
}, |
|
{ |
|
"epoch": 17.9, |
|
"learning_rate": 8.235787531192989e-06, |
|
"loss": 1.3006, |
|
"step": 290500 |
|
}, |
|
{ |
|
"epoch": 17.93, |
|
"learning_rate": 8.214817455491015e-06, |
|
"loss": 1.3041, |
|
"step": 291000 |
|
}, |
|
{ |
|
"epoch": 17.96, |
|
"learning_rate": 8.193847379789041e-06, |
|
"loss": 1.3017, |
|
"step": 291500 |
|
}, |
|
{ |
|
"epoch": 17.99, |
|
"learning_rate": 8.172877304087067e-06, |
|
"loss": 1.3003, |
|
"step": 292000 |
|
}, |
|
{ |
|
"epoch": 18.02, |
|
"learning_rate": 8.151907228385095e-06, |
|
"loss": 1.3033, |
|
"step": 292500 |
|
}, |
|
{ |
|
"epoch": 18.05, |
|
"learning_rate": 8.130937152683122e-06, |
|
"loss": 1.2964, |
|
"step": 293000 |
|
}, |
|
{ |
|
"epoch": 18.08, |
|
"learning_rate": 8.10996707698115e-06, |
|
"loss": 1.299, |
|
"step": 293500 |
|
}, |
|
{ |
|
"epoch": 18.12, |
|
"learning_rate": 8.088997001279176e-06, |
|
"loss": 1.3047, |
|
"step": 294000 |
|
}, |
|
{ |
|
"epoch": 18.15, |
|
"learning_rate": 8.068026925577202e-06, |
|
"loss": 1.3035, |
|
"step": 294500 |
|
}, |
|
{ |
|
"epoch": 18.18, |
|
"learning_rate": 8.047056849875228e-06, |
|
"loss": 1.3021, |
|
"step": 295000 |
|
}, |
|
{ |
|
"epoch": 18.21, |
|
"learning_rate": 8.026086774173254e-06, |
|
"loss": 1.3094, |
|
"step": 295500 |
|
}, |
|
{ |
|
"epoch": 18.24, |
|
"learning_rate": 8.005116698471282e-06, |
|
"loss": 1.2999, |
|
"step": 296000 |
|
}, |
|
{ |
|
"epoch": 18.27, |
|
"learning_rate": 7.984146622769309e-06, |
|
"loss": 1.2984, |
|
"step": 296500 |
|
}, |
|
{ |
|
"epoch": 18.3, |
|
"learning_rate": 7.963176547067337e-06, |
|
"loss": 1.2982, |
|
"step": 297000 |
|
}, |
|
{ |
|
"epoch": 18.33, |
|
"learning_rate": 7.942206471365363e-06, |
|
"loss": 1.3021, |
|
"step": 297500 |
|
}, |
|
{ |
|
"epoch": 18.36, |
|
"learning_rate": 7.921236395663389e-06, |
|
"loss": 1.2956, |
|
"step": 298000 |
|
}, |
|
{ |
|
"epoch": 18.39, |
|
"learning_rate": 7.900266319961415e-06, |
|
"loss": 1.2943, |
|
"step": 298500 |
|
}, |
|
{ |
|
"epoch": 18.42, |
|
"learning_rate": 7.879296244259442e-06, |
|
"loss": 1.3039, |
|
"step": 299000 |
|
}, |
|
{ |
|
"epoch": 18.45, |
|
"learning_rate": 7.85832616855747e-06, |
|
"loss": 1.2958, |
|
"step": 299500 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"learning_rate": 7.837356092855496e-06, |
|
"loss": 1.2928, |
|
"step": 300000 |
|
}, |
|
{ |
|
"epoch": 18.52, |
|
"learning_rate": 7.816386017153522e-06, |
|
"loss": 1.2986, |
|
"step": 300500 |
|
}, |
|
{ |
|
"epoch": 18.55, |
|
"learning_rate": 7.79541594145155e-06, |
|
"loss": 1.3018, |
|
"step": 301000 |
|
}, |
|
{ |
|
"epoch": 18.58, |
|
"learning_rate": 7.774445865749576e-06, |
|
"loss": 1.3009, |
|
"step": 301500 |
|
}, |
|
{ |
|
"epoch": 18.61, |
|
"learning_rate": 7.753475790047602e-06, |
|
"loss": 1.3003, |
|
"step": 302000 |
|
}, |
|
{ |
|
"epoch": 18.64, |
|
"learning_rate": 7.732505714345629e-06, |
|
"loss": 1.2967, |
|
"step": 302500 |
|
}, |
|
{ |
|
"epoch": 18.67, |
|
"learning_rate": 7.711535638643655e-06, |
|
"loss": 1.2978, |
|
"step": 303000 |
|
}, |
|
{ |
|
"epoch": 18.7, |
|
"learning_rate": 7.690565562941683e-06, |
|
"loss": 1.2977, |
|
"step": 303500 |
|
}, |
|
{ |
|
"epoch": 18.73, |
|
"learning_rate": 7.669595487239709e-06, |
|
"loss": 1.2957, |
|
"step": 304000 |
|
}, |
|
{ |
|
"epoch": 18.76, |
|
"learning_rate": 7.648625411537737e-06, |
|
"loss": 1.2994, |
|
"step": 304500 |
|
}, |
|
{ |
|
"epoch": 18.79, |
|
"learning_rate": 7.627655335835763e-06, |
|
"loss": 1.2978, |
|
"step": 305000 |
|
}, |
|
{ |
|
"epoch": 18.82, |
|
"learning_rate": 7.6066852601337895e-06, |
|
"loss": 1.2934, |
|
"step": 305500 |
|
}, |
|
{ |
|
"epoch": 18.86, |
|
"learning_rate": 7.585715184431817e-06, |
|
"loss": 1.2969, |
|
"step": 306000 |
|
}, |
|
{ |
|
"epoch": 18.89, |
|
"learning_rate": 7.564745108729843e-06, |
|
"loss": 1.2997, |
|
"step": 306500 |
|
}, |
|
{ |
|
"epoch": 18.92, |
|
"learning_rate": 7.54377503302787e-06, |
|
"loss": 1.2984, |
|
"step": 307000 |
|
}, |
|
{ |
|
"epoch": 18.95, |
|
"learning_rate": 7.522804957325896e-06, |
|
"loss": 1.2957, |
|
"step": 307500 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"learning_rate": 7.5018348816239225e-06, |
|
"loss": 1.304, |
|
"step": 308000 |
|
}, |
|
{ |
|
"epoch": 19.01, |
|
"learning_rate": 7.48086480592195e-06, |
|
"loss": 1.2993, |
|
"step": 308500 |
|
}, |
|
{ |
|
"epoch": 19.04, |
|
"learning_rate": 7.459894730219977e-06, |
|
"loss": 1.2971, |
|
"step": 309000 |
|
}, |
|
{ |
|
"epoch": 19.07, |
|
"learning_rate": 7.438924654518004e-06, |
|
"loss": 1.2952, |
|
"step": 309500 |
|
}, |
|
{ |
|
"epoch": 19.1, |
|
"learning_rate": 7.41795457881603e-06, |
|
"loss": 1.2903, |
|
"step": 310000 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"learning_rate": 7.396984503114056e-06, |
|
"loss": 1.2917, |
|
"step": 310500 |
|
}, |
|
{ |
|
"epoch": 19.16, |
|
"learning_rate": 7.376014427412083e-06, |
|
"loss": 1.2967, |
|
"step": 311000 |
|
}, |
|
{ |
|
"epoch": 19.19, |
|
"learning_rate": 7.35504435171011e-06, |
|
"loss": 1.2945, |
|
"step": 311500 |
|
}, |
|
{ |
|
"epoch": 19.22, |
|
"learning_rate": 7.3340742760081375e-06, |
|
"loss": 1.2947, |
|
"step": 312000 |
|
}, |
|
{ |
|
"epoch": 19.26, |
|
"learning_rate": 7.313104200306164e-06, |
|
"loss": 1.2929, |
|
"step": 312500 |
|
}, |
|
{ |
|
"epoch": 19.29, |
|
"learning_rate": 7.29213412460419e-06, |
|
"loss": 1.2942, |
|
"step": 313000 |
|
}, |
|
{ |
|
"epoch": 19.32, |
|
"learning_rate": 7.271164048902217e-06, |
|
"loss": 1.2931, |
|
"step": 313500 |
|
}, |
|
{ |
|
"epoch": 19.35, |
|
"learning_rate": 7.250193973200243e-06, |
|
"loss": 1.2932, |
|
"step": 314000 |
|
}, |
|
{ |
|
"epoch": 19.38, |
|
"learning_rate": 7.229223897498271e-06, |
|
"loss": 1.2944, |
|
"step": 314500 |
|
}, |
|
{ |
|
"epoch": 19.41, |
|
"learning_rate": 7.2082538217962975e-06, |
|
"loss": 1.2957, |
|
"step": 315000 |
|
}, |
|
{ |
|
"epoch": 19.44, |
|
"learning_rate": 7.187283746094324e-06, |
|
"loss": 1.2971, |
|
"step": 315500 |
|
}, |
|
{ |
|
"epoch": 19.47, |
|
"learning_rate": 7.166313670392351e-06, |
|
"loss": 1.2964, |
|
"step": 316000 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"learning_rate": 7.145343594690377e-06, |
|
"loss": 1.2943, |
|
"step": 316500 |
|
}, |
|
{ |
|
"epoch": 19.53, |
|
"learning_rate": 7.124373518988404e-06, |
|
"loss": 1.2935, |
|
"step": 317000 |
|
}, |
|
{ |
|
"epoch": 19.56, |
|
"learning_rate": 7.1034034432864305e-06, |
|
"loss": 1.2963, |
|
"step": 317500 |
|
}, |
|
{ |
|
"epoch": 19.59, |
|
"learning_rate": 7.082433367584458e-06, |
|
"loss": 1.2899, |
|
"step": 318000 |
|
}, |
|
{ |
|
"epoch": 19.63, |
|
"learning_rate": 7.061463291882485e-06, |
|
"loss": 1.2935, |
|
"step": 318500 |
|
}, |
|
{ |
|
"epoch": 19.66, |
|
"learning_rate": 7.040493216180511e-06, |
|
"loss": 1.2964, |
|
"step": 319000 |
|
}, |
|
{ |
|
"epoch": 19.69, |
|
"learning_rate": 7.019523140478538e-06, |
|
"loss": 1.2892, |
|
"step": 319500 |
|
}, |
|
{ |
|
"epoch": 19.72, |
|
"learning_rate": 6.998553064776564e-06, |
|
"loss": 1.2942, |
|
"step": 320000 |
|
}, |
|
{ |
|
"epoch": 19.75, |
|
"learning_rate": 6.977582989074591e-06, |
|
"loss": 1.2943, |
|
"step": 320500 |
|
}, |
|
{ |
|
"epoch": 19.78, |
|
"learning_rate": 6.956612913372618e-06, |
|
"loss": 1.2904, |
|
"step": 321000 |
|
}, |
|
{ |
|
"epoch": 19.81, |
|
"learning_rate": 6.935642837670644e-06, |
|
"loss": 1.294, |
|
"step": 321500 |
|
}, |
|
{ |
|
"epoch": 19.84, |
|
"learning_rate": 6.914672761968672e-06, |
|
"loss": 1.2958, |
|
"step": 322000 |
|
}, |
|
{ |
|
"epoch": 19.87, |
|
"learning_rate": 6.893702686266698e-06, |
|
"loss": 1.2941, |
|
"step": 322500 |
|
}, |
|
{ |
|
"epoch": 19.9, |
|
"learning_rate": 6.872732610564725e-06, |
|
"loss": 1.2935, |
|
"step": 323000 |
|
}, |
|
{ |
|
"epoch": 19.93, |
|
"learning_rate": 6.851762534862751e-06, |
|
"loss": 1.2917, |
|
"step": 323500 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"learning_rate": 6.830792459160778e-06, |
|
"loss": 1.2899, |
|
"step": 324000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 6.809822383458805e-06, |
|
"loss": 1.2946, |
|
"step": 324500 |
|
}, |
|
{ |
|
"epoch": 20.03, |
|
"learning_rate": 6.788852307756831e-06, |
|
"loss": 1.293, |
|
"step": 325000 |
|
}, |
|
{ |
|
"epoch": 20.06, |
|
"learning_rate": 6.767882232054859e-06, |
|
"loss": 1.2877, |
|
"step": 325500 |
|
}, |
|
{ |
|
"epoch": 20.09, |
|
"learning_rate": 6.746912156352885e-06, |
|
"loss": 1.2873, |
|
"step": 326000 |
|
}, |
|
{ |
|
"epoch": 20.12, |
|
"learning_rate": 6.725942080650911e-06, |
|
"loss": 1.2932, |
|
"step": 326500 |
|
}, |
|
{ |
|
"epoch": 20.15, |
|
"learning_rate": 6.7049720049489385e-06, |
|
"loss": 1.2928, |
|
"step": 327000 |
|
}, |
|
{ |
|
"epoch": 20.18, |
|
"learning_rate": 6.684001929246965e-06, |
|
"loss": 1.2875, |
|
"step": 327500 |
|
}, |
|
{ |
|
"epoch": 20.21, |
|
"learning_rate": 6.663031853544993e-06, |
|
"loss": 1.2935, |
|
"step": 328000 |
|
}, |
|
{ |
|
"epoch": 20.24, |
|
"learning_rate": 6.642061777843019e-06, |
|
"loss": 1.2839, |
|
"step": 328500 |
|
}, |
|
{ |
|
"epoch": 20.27, |
|
"learning_rate": 6.621091702141045e-06, |
|
"loss": 1.2932, |
|
"step": 329000 |
|
}, |
|
{ |
|
"epoch": 20.3, |
|
"learning_rate": 6.600121626439072e-06, |
|
"loss": 1.2851, |
|
"step": 329500 |
|
}, |
|
{ |
|
"epoch": 20.33, |
|
"learning_rate": 6.5791515507370985e-06, |
|
"loss": 1.293, |
|
"step": 330000 |
|
}, |
|
{ |
|
"epoch": 20.36, |
|
"learning_rate": 6.558181475035126e-06, |
|
"loss": 1.2895, |
|
"step": 330500 |
|
}, |
|
{ |
|
"epoch": 20.4, |
|
"learning_rate": 6.537211399333152e-06, |
|
"loss": 1.2879, |
|
"step": 331000 |
|
}, |
|
{ |
|
"epoch": 20.43, |
|
"learning_rate": 6.516241323631178e-06, |
|
"loss": 1.2901, |
|
"step": 331500 |
|
}, |
|
{ |
|
"epoch": 20.46, |
|
"learning_rate": 6.495271247929206e-06, |
|
"loss": 1.289, |
|
"step": 332000 |
|
}, |
|
{ |
|
"epoch": 20.49, |
|
"learning_rate": 6.474301172227232e-06, |
|
"loss": 1.2876, |
|
"step": 332500 |
|
}, |
|
{ |
|
"epoch": 20.52, |
|
"learning_rate": 6.453331096525259e-06, |
|
"loss": 1.288, |
|
"step": 333000 |
|
}, |
|
{ |
|
"epoch": 20.55, |
|
"learning_rate": 6.432361020823286e-06, |
|
"loss": 1.2881, |
|
"step": 333500 |
|
}, |
|
{ |
|
"epoch": 20.58, |
|
"learning_rate": 6.411390945121312e-06, |
|
"loss": 1.2891, |
|
"step": 334000 |
|
}, |
|
{ |
|
"epoch": 20.61, |
|
"learning_rate": 6.390420869419339e-06, |
|
"loss": 1.2864, |
|
"step": 334500 |
|
}, |
|
{ |
|
"epoch": 20.64, |
|
"learning_rate": 6.369450793717365e-06, |
|
"loss": 1.2895, |
|
"step": 335000 |
|
}, |
|
{ |
|
"epoch": 20.67, |
|
"learning_rate": 6.348480718015393e-06, |
|
"loss": 1.2865, |
|
"step": 335500 |
|
}, |
|
{ |
|
"epoch": 20.7, |
|
"learning_rate": 6.327510642313419e-06, |
|
"loss": 1.2809, |
|
"step": 336000 |
|
}, |
|
{ |
|
"epoch": 20.73, |
|
"learning_rate": 6.3065405666114465e-06, |
|
"loss": 1.2884, |
|
"step": 336500 |
|
}, |
|
{ |
|
"epoch": 20.77, |
|
"learning_rate": 6.285570490909473e-06, |
|
"loss": 1.2866, |
|
"step": 337000 |
|
}, |
|
{ |
|
"epoch": 20.8, |
|
"learning_rate": 6.264600415207499e-06, |
|
"loss": 1.2957, |
|
"step": 337500 |
|
}, |
|
{ |
|
"epoch": 20.83, |
|
"learning_rate": 6.243630339505526e-06, |
|
"loss": 1.2829, |
|
"step": 338000 |
|
}, |
|
{ |
|
"epoch": 20.86, |
|
"learning_rate": 6.222660263803552e-06, |
|
"loss": 1.2866, |
|
"step": 338500 |
|
}, |
|
{ |
|
"epoch": 20.89, |
|
"learning_rate": 6.20169018810158e-06, |
|
"loss": 1.2896, |
|
"step": 339000 |
|
}, |
|
{ |
|
"epoch": 20.92, |
|
"learning_rate": 6.1807201123996065e-06, |
|
"loss": 1.2863, |
|
"step": 339500 |
|
}, |
|
{ |
|
"epoch": 20.95, |
|
"learning_rate": 6.159750036697633e-06, |
|
"loss": 1.2854, |
|
"step": 340000 |
|
}, |
|
{ |
|
"epoch": 20.98, |
|
"learning_rate": 6.13877996099566e-06, |
|
"loss": 1.2834, |
|
"step": 340500 |
|
}, |
|
{ |
|
"epoch": 21.01, |
|
"learning_rate": 6.117809885293686e-06, |
|
"loss": 1.2887, |
|
"step": 341000 |
|
}, |
|
{ |
|
"epoch": 21.04, |
|
"learning_rate": 6.096839809591714e-06, |
|
"loss": 1.2845, |
|
"step": 341500 |
|
}, |
|
{ |
|
"epoch": 21.07, |
|
"learning_rate": 6.07586973388974e-06, |
|
"loss": 1.2899, |
|
"step": 342000 |
|
}, |
|
{ |
|
"epoch": 21.1, |
|
"learning_rate": 6.0548996581877665e-06, |
|
"loss": 1.2808, |
|
"step": 342500 |
|
}, |
|
{ |
|
"epoch": 21.14, |
|
"learning_rate": 6.033929582485794e-06, |
|
"loss": 1.2853, |
|
"step": 343000 |
|
}, |
|
{ |
|
"epoch": 21.17, |
|
"learning_rate": 6.01295950678382e-06, |
|
"loss": 1.2858, |
|
"step": 343500 |
|
}, |
|
{ |
|
"epoch": 21.2, |
|
"learning_rate": 5.991989431081847e-06, |
|
"loss": 1.288, |
|
"step": 344000 |
|
}, |
|
{ |
|
"epoch": 21.23, |
|
"learning_rate": 5.971019355379873e-06, |
|
"loss": 1.281, |
|
"step": 344500 |
|
}, |
|
{ |
|
"epoch": 21.26, |
|
"learning_rate": 5.9500492796778995e-06, |
|
"loss": 1.2854, |
|
"step": 345000 |
|
}, |
|
{ |
|
"epoch": 21.29, |
|
"learning_rate": 5.929079203975927e-06, |
|
"loss": 1.2904, |
|
"step": 345500 |
|
}, |
|
{ |
|
"epoch": 21.32, |
|
"learning_rate": 5.908109128273954e-06, |
|
"loss": 1.2812, |
|
"step": 346000 |
|
}, |
|
{ |
|
"epoch": 21.35, |
|
"learning_rate": 5.887139052571981e-06, |
|
"loss": 1.2815, |
|
"step": 346500 |
|
}, |
|
{ |
|
"epoch": 21.38, |
|
"learning_rate": 5.866168976870007e-06, |
|
"loss": 1.2881, |
|
"step": 347000 |
|
}, |
|
{ |
|
"epoch": 21.41, |
|
"learning_rate": 5.845198901168033e-06, |
|
"loss": 1.2815, |
|
"step": 347500 |
|
}, |
|
{ |
|
"epoch": 21.44, |
|
"learning_rate": 5.82422882546606e-06, |
|
"loss": 1.2813, |
|
"step": 348000 |
|
}, |
|
{ |
|
"epoch": 21.47, |
|
"learning_rate": 5.8032587497640866e-06, |
|
"loss": 1.2855, |
|
"step": 348500 |
|
}, |
|
{ |
|
"epoch": 21.5, |
|
"learning_rate": 5.7822886740621145e-06, |
|
"loss": 1.2856, |
|
"step": 349000 |
|
}, |
|
{ |
|
"epoch": 21.54, |
|
"learning_rate": 5.761318598360141e-06, |
|
"loss": 1.2873, |
|
"step": 349500 |
|
}, |
|
{ |
|
"epoch": 21.57, |
|
"learning_rate": 5.740348522658167e-06, |
|
"loss": 1.288, |
|
"step": 350000 |
|
}, |
|
{ |
|
"epoch": 21.6, |
|
"learning_rate": 5.719378446956194e-06, |
|
"loss": 1.2841, |
|
"step": 350500 |
|
}, |
|
{ |
|
"epoch": 21.63, |
|
"learning_rate": 5.69840837125422e-06, |
|
"loss": 1.2827, |
|
"step": 351000 |
|
}, |
|
{ |
|
"epoch": 21.66, |
|
"learning_rate": 5.6774382955522474e-06, |
|
"loss": 1.2872, |
|
"step": 351500 |
|
}, |
|
{ |
|
"epoch": 21.69, |
|
"learning_rate": 5.656468219850274e-06, |
|
"loss": 1.2843, |
|
"step": 352000 |
|
}, |
|
{ |
|
"epoch": 21.72, |
|
"learning_rate": 5.6354981441483e-06, |
|
"loss": 1.2841, |
|
"step": 352500 |
|
}, |
|
{ |
|
"epoch": 21.75, |
|
"learning_rate": 5.614528068446328e-06, |
|
"loss": 1.2853, |
|
"step": 353000 |
|
}, |
|
{ |
|
"epoch": 21.78, |
|
"learning_rate": 5.593557992744354e-06, |
|
"loss": 1.2827, |
|
"step": 353500 |
|
}, |
|
{ |
|
"epoch": 21.81, |
|
"learning_rate": 5.572587917042381e-06, |
|
"loss": 1.2842, |
|
"step": 354000 |
|
}, |
|
{ |
|
"epoch": 21.84, |
|
"learning_rate": 5.5516178413404075e-06, |
|
"loss": 1.283, |
|
"step": 354500 |
|
}, |
|
{ |
|
"epoch": 21.87, |
|
"learning_rate": 5.530647765638434e-06, |
|
"loss": 1.2859, |
|
"step": 355000 |
|
}, |
|
{ |
|
"epoch": 21.91, |
|
"learning_rate": 5.509677689936462e-06, |
|
"loss": 1.279, |
|
"step": 355500 |
|
}, |
|
{ |
|
"epoch": 21.94, |
|
"learning_rate": 5.488707614234488e-06, |
|
"loss": 1.2767, |
|
"step": 356000 |
|
}, |
|
{ |
|
"epoch": 21.97, |
|
"learning_rate": 5.467737538532515e-06, |
|
"loss": 1.2789, |
|
"step": 356500 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"learning_rate": 5.446767462830541e-06, |
|
"loss": 1.2863, |
|
"step": 357000 |
|
}, |
|
{ |
|
"epoch": 22.03, |
|
"learning_rate": 5.425797387128568e-06, |
|
"loss": 1.2829, |
|
"step": 357500 |
|
}, |
|
{ |
|
"epoch": 22.06, |
|
"learning_rate": 5.4048273114265946e-06, |
|
"loss": 1.2766, |
|
"step": 358000 |
|
}, |
|
{ |
|
"epoch": 22.09, |
|
"learning_rate": 5.383857235724621e-06, |
|
"loss": 1.2841, |
|
"step": 358500 |
|
}, |
|
{ |
|
"epoch": 22.12, |
|
"learning_rate": 5.362887160022649e-06, |
|
"loss": 1.2837, |
|
"step": 359000 |
|
}, |
|
{ |
|
"epoch": 22.15, |
|
"learning_rate": 5.341917084320675e-06, |
|
"loss": 1.2837, |
|
"step": 359500 |
|
}, |
|
{ |
|
"epoch": 22.18, |
|
"learning_rate": 5.320947008618702e-06, |
|
"loss": 1.2829, |
|
"step": 360000 |
|
}, |
|
{ |
|
"epoch": 22.21, |
|
"learning_rate": 5.299976932916728e-06, |
|
"loss": 1.2827, |
|
"step": 360500 |
|
}, |
|
{ |
|
"epoch": 22.24, |
|
"learning_rate": 5.279006857214755e-06, |
|
"loss": 1.286, |
|
"step": 361000 |
|
}, |
|
{ |
|
"epoch": 22.27, |
|
"learning_rate": 5.258036781512782e-06, |
|
"loss": 1.2832, |
|
"step": 361500 |
|
}, |
|
{ |
|
"epoch": 22.31, |
|
"learning_rate": 5.237066705810808e-06, |
|
"loss": 1.2812, |
|
"step": 362000 |
|
}, |
|
{ |
|
"epoch": 22.34, |
|
"learning_rate": 5.216096630108836e-06, |
|
"loss": 1.2797, |
|
"step": 362500 |
|
}, |
|
{ |
|
"epoch": 22.37, |
|
"learning_rate": 5.195126554406862e-06, |
|
"loss": 1.2752, |
|
"step": 363000 |
|
}, |
|
{ |
|
"epoch": 22.4, |
|
"learning_rate": 5.174156478704888e-06, |
|
"loss": 1.2824, |
|
"step": 363500 |
|
}, |
|
{ |
|
"epoch": 22.43, |
|
"learning_rate": 5.1531864030029155e-06, |
|
"loss": 1.2782, |
|
"step": 364000 |
|
}, |
|
{ |
|
"epoch": 22.46, |
|
"learning_rate": 5.132216327300942e-06, |
|
"loss": 1.2891, |
|
"step": 364500 |
|
}, |
|
{ |
|
"epoch": 22.49, |
|
"learning_rate": 5.111246251598969e-06, |
|
"loss": 1.2871, |
|
"step": 365000 |
|
}, |
|
{ |
|
"epoch": 22.52, |
|
"learning_rate": 5.090276175896995e-06, |
|
"loss": 1.2844, |
|
"step": 365500 |
|
}, |
|
{ |
|
"epoch": 22.55, |
|
"learning_rate": 5.069306100195021e-06, |
|
"loss": 1.2804, |
|
"step": 366000 |
|
}, |
|
{ |
|
"epoch": 22.58, |
|
"learning_rate": 5.048336024493049e-06, |
|
"loss": 1.2774, |
|
"step": 366500 |
|
}, |
|
{ |
|
"epoch": 22.61, |
|
"learning_rate": 5.0273659487910755e-06, |
|
"loss": 1.2817, |
|
"step": 367000 |
|
}, |
|
{ |
|
"epoch": 22.64, |
|
"learning_rate": 5.0063958730891026e-06, |
|
"loss": 1.2802, |
|
"step": 367500 |
|
}, |
|
{ |
|
"epoch": 22.68, |
|
"learning_rate": 4.985425797387129e-06, |
|
"loss": 1.2757, |
|
"step": 368000 |
|
}, |
|
{ |
|
"epoch": 22.71, |
|
"learning_rate": 4.964455721685156e-06, |
|
"loss": 1.2831, |
|
"step": 368500 |
|
}, |
|
{ |
|
"epoch": 22.74, |
|
"learning_rate": 4.943485645983183e-06, |
|
"loss": 1.2814, |
|
"step": 369000 |
|
}, |
|
{ |
|
"epoch": 22.77, |
|
"learning_rate": 4.922515570281209e-06, |
|
"loss": 1.2796, |
|
"step": 369500 |
|
}, |
|
{ |
|
"epoch": 22.8, |
|
"learning_rate": 4.9015454945792355e-06, |
|
"loss": 1.2783, |
|
"step": 370000 |
|
}, |
|
{ |
|
"epoch": 22.83, |
|
"learning_rate": 4.880575418877263e-06, |
|
"loss": 1.2782, |
|
"step": 370500 |
|
}, |
|
{ |
|
"epoch": 22.86, |
|
"learning_rate": 4.85960534317529e-06, |
|
"loss": 1.28, |
|
"step": 371000 |
|
}, |
|
{ |
|
"epoch": 22.89, |
|
"learning_rate": 4.838635267473316e-06, |
|
"loss": 1.2813, |
|
"step": 371500 |
|
}, |
|
{ |
|
"epoch": 22.92, |
|
"learning_rate": 4.817665191771343e-06, |
|
"loss": 1.2784, |
|
"step": 372000 |
|
}, |
|
{ |
|
"epoch": 22.95, |
|
"learning_rate": 4.796695116069369e-06, |
|
"loss": 1.2782, |
|
"step": 372500 |
|
}, |
|
{ |
|
"epoch": 22.98, |
|
"learning_rate": 4.775725040367396e-06, |
|
"loss": 1.2735, |
|
"step": 373000 |
|
}, |
|
{ |
|
"epoch": 23.01, |
|
"learning_rate": 4.754754964665423e-06, |
|
"loss": 1.2787, |
|
"step": 373500 |
|
}, |
|
{ |
|
"epoch": 23.05, |
|
"learning_rate": 4.73378488896345e-06, |
|
"loss": 1.2787, |
|
"step": 374000 |
|
}, |
|
{ |
|
"epoch": 23.08, |
|
"learning_rate": 4.712814813261477e-06, |
|
"loss": 1.2744, |
|
"step": 374500 |
|
}, |
|
{ |
|
"epoch": 23.11, |
|
"learning_rate": 4.691844737559503e-06, |
|
"loss": 1.2754, |
|
"step": 375000 |
|
}, |
|
{ |
|
"epoch": 23.14, |
|
"learning_rate": 4.670874661857529e-06, |
|
"loss": 1.2789, |
|
"step": 375500 |
|
}, |
|
{ |
|
"epoch": 23.17, |
|
"learning_rate": 4.649904586155556e-06, |
|
"loss": 1.2795, |
|
"step": 376000 |
|
}, |
|
{ |
|
"epoch": 23.2, |
|
"learning_rate": 4.6289345104535835e-06, |
|
"loss": 1.2757, |
|
"step": 376500 |
|
}, |
|
{ |
|
"epoch": 23.23, |
|
"learning_rate": 4.60796443475161e-06, |
|
"loss": 1.2772, |
|
"step": 377000 |
|
}, |
|
{ |
|
"epoch": 23.26, |
|
"learning_rate": 4.586994359049636e-06, |
|
"loss": 1.274, |
|
"step": 377500 |
|
}, |
|
{ |
|
"epoch": 23.29, |
|
"learning_rate": 4.566024283347663e-06, |
|
"loss": 1.2771, |
|
"step": 378000 |
|
}, |
|
{ |
|
"epoch": 23.32, |
|
"learning_rate": 4.54505420764569e-06, |
|
"loss": 1.277, |
|
"step": 378500 |
|
}, |
|
{ |
|
"epoch": 23.35, |
|
"learning_rate": 4.524084131943716e-06, |
|
"loss": 1.2804, |
|
"step": 379000 |
|
}, |
|
{ |
|
"epoch": 23.38, |
|
"learning_rate": 4.5031140562417435e-06, |
|
"loss": 1.2783, |
|
"step": 379500 |
|
}, |
|
{ |
|
"epoch": 23.41, |
|
"learning_rate": 4.48214398053977e-06, |
|
"loss": 1.2762, |
|
"step": 380000 |
|
}, |
|
{ |
|
"epoch": 23.45, |
|
"learning_rate": 4.461173904837797e-06, |
|
"loss": 1.2779, |
|
"step": 380500 |
|
}, |
|
{ |
|
"epoch": 23.48, |
|
"learning_rate": 4.440203829135823e-06, |
|
"loss": 1.273, |
|
"step": 381000 |
|
}, |
|
{ |
|
"epoch": 23.51, |
|
"learning_rate": 4.41923375343385e-06, |
|
"loss": 1.2811, |
|
"step": 381500 |
|
}, |
|
{ |
|
"epoch": 23.54, |
|
"learning_rate": 4.398263677731877e-06, |
|
"loss": 1.2747, |
|
"step": 382000 |
|
}, |
|
{ |
|
"epoch": 23.57, |
|
"learning_rate": 4.3772936020299035e-06, |
|
"loss": 1.2711, |
|
"step": 382500 |
|
}, |
|
{ |
|
"epoch": 23.6, |
|
"learning_rate": 4.356323526327931e-06, |
|
"loss": 1.2784, |
|
"step": 383000 |
|
}, |
|
{ |
|
"epoch": 23.63, |
|
"learning_rate": 4.335353450625957e-06, |
|
"loss": 1.2776, |
|
"step": 383500 |
|
}, |
|
{ |
|
"epoch": 23.66, |
|
"learning_rate": 4.314383374923984e-06, |
|
"loss": 1.2702, |
|
"step": 384000 |
|
}, |
|
{ |
|
"epoch": 23.69, |
|
"learning_rate": 4.293413299222011e-06, |
|
"loss": 1.2728, |
|
"step": 384500 |
|
}, |
|
{ |
|
"epoch": 23.72, |
|
"learning_rate": 4.272443223520037e-06, |
|
"loss": 1.279, |
|
"step": 385000 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 4.2514731478180635e-06, |
|
"loss": 1.2786, |
|
"step": 385500 |
|
}, |
|
{ |
|
"epoch": 23.78, |
|
"learning_rate": 4.230503072116091e-06, |
|
"loss": 1.2733, |
|
"step": 386000 |
|
}, |
|
{ |
|
"epoch": 23.82, |
|
"learning_rate": 4.209532996414118e-06, |
|
"loss": 1.2778, |
|
"step": 386500 |
|
}, |
|
{ |
|
"epoch": 23.85, |
|
"learning_rate": 4.188562920712144e-06, |
|
"loss": 1.2726, |
|
"step": 387000 |
|
}, |
|
{ |
|
"epoch": 23.88, |
|
"learning_rate": 4.167592845010171e-06, |
|
"loss": 1.2729, |
|
"step": 387500 |
|
}, |
|
{ |
|
"epoch": 23.91, |
|
"learning_rate": 4.146622769308197e-06, |
|
"loss": 1.2781, |
|
"step": 388000 |
|
}, |
|
{ |
|
"epoch": 23.94, |
|
"learning_rate": 4.125652693606224e-06, |
|
"loss": 1.2811, |
|
"step": 388500 |
|
}, |
|
{ |
|
"epoch": 23.97, |
|
"learning_rate": 4.104682617904251e-06, |
|
"loss": 1.2833, |
|
"step": 389000 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"learning_rate": 4.083712542202278e-06, |
|
"loss": 1.2734, |
|
"step": 389500 |
|
}, |
|
{ |
|
"epoch": 24.03, |
|
"learning_rate": 4.062742466500305e-06, |
|
"loss": 1.2759, |
|
"step": 390000 |
|
}, |
|
{ |
|
"epoch": 24.06, |
|
"learning_rate": 4.041772390798331e-06, |
|
"loss": 1.2716, |
|
"step": 390500 |
|
}, |
|
{ |
|
"epoch": 24.09, |
|
"learning_rate": 4.020802315096357e-06, |
|
"loss": 1.2753, |
|
"step": 391000 |
|
}, |
|
{ |
|
"epoch": 24.12, |
|
"learning_rate": 3.9998322393943844e-06, |
|
"loss": 1.2755, |
|
"step": 391500 |
|
}, |
|
{ |
|
"epoch": 24.15, |
|
"learning_rate": 3.9788621636924115e-06, |
|
"loss": 1.2715, |
|
"step": 392000 |
|
}, |
|
{ |
|
"epoch": 24.19, |
|
"learning_rate": 3.957892087990438e-06, |
|
"loss": 1.2727, |
|
"step": 392500 |
|
}, |
|
{ |
|
"epoch": 24.22, |
|
"learning_rate": 3.936922012288465e-06, |
|
"loss": 1.2693, |
|
"step": 393000 |
|
}, |
|
{ |
|
"epoch": 24.25, |
|
"learning_rate": 3.915951936586491e-06, |
|
"loss": 1.2767, |
|
"step": 393500 |
|
}, |
|
{ |
|
"epoch": 24.28, |
|
"learning_rate": 3.894981860884518e-06, |
|
"loss": 1.271, |
|
"step": 394000 |
|
}, |
|
{ |
|
"epoch": 24.31, |
|
"learning_rate": 3.8740117851825445e-06, |
|
"loss": 1.2744, |
|
"step": 394500 |
|
}, |
|
{ |
|
"epoch": 24.34, |
|
"learning_rate": 3.8530417094805716e-06, |
|
"loss": 1.2777, |
|
"step": 395000 |
|
}, |
|
{ |
|
"epoch": 24.37, |
|
"learning_rate": 3.832071633778599e-06, |
|
"loss": 1.2758, |
|
"step": 395500 |
|
}, |
|
{ |
|
"epoch": 24.4, |
|
"learning_rate": 3.811101558076625e-06, |
|
"loss": 1.2717, |
|
"step": 396000 |
|
}, |
|
{ |
|
"epoch": 24.43, |
|
"learning_rate": 3.7901314823746516e-06, |
|
"loss": 1.2728, |
|
"step": 396500 |
|
}, |
|
{ |
|
"epoch": 24.46, |
|
"learning_rate": 3.7691614066726782e-06, |
|
"loss": 1.2732, |
|
"step": 397000 |
|
}, |
|
{ |
|
"epoch": 24.49, |
|
"learning_rate": 3.7481913309707053e-06, |
|
"loss": 1.2742, |
|
"step": 397500 |
|
}, |
|
{ |
|
"epoch": 24.52, |
|
"learning_rate": 3.727221255268732e-06, |
|
"loss": 1.2731, |
|
"step": 398000 |
|
}, |
|
{ |
|
"epoch": 24.55, |
|
"learning_rate": 3.7062511795667582e-06, |
|
"loss": 1.2732, |
|
"step": 398500 |
|
}, |
|
{ |
|
"epoch": 24.59, |
|
"learning_rate": 3.685281103864785e-06, |
|
"loss": 1.2652, |
|
"step": 399000 |
|
}, |
|
{ |
|
"epoch": 24.62, |
|
"learning_rate": 3.664311028162812e-06, |
|
"loss": 1.2754, |
|
"step": 399500 |
|
}, |
|
{ |
|
"epoch": 24.65, |
|
"learning_rate": 3.6433409524608387e-06, |
|
"loss": 1.2714, |
|
"step": 400000 |
|
}, |
|
{ |
|
"epoch": 24.68, |
|
"learning_rate": 3.6223708767588658e-06, |
|
"loss": 1.2773, |
|
"step": 400500 |
|
}, |
|
{ |
|
"epoch": 24.71, |
|
"learning_rate": 3.6014008010568924e-06, |
|
"loss": 1.2697, |
|
"step": 401000 |
|
}, |
|
{ |
|
"epoch": 24.74, |
|
"learning_rate": 3.5804307253549187e-06, |
|
"loss": 1.2769, |
|
"step": 401500 |
|
}, |
|
{ |
|
"epoch": 24.77, |
|
"learning_rate": 3.5594606496529454e-06, |
|
"loss": 1.2767, |
|
"step": 402000 |
|
}, |
|
{ |
|
"epoch": 24.8, |
|
"learning_rate": 3.5384905739509724e-06, |
|
"loss": 1.2758, |
|
"step": 402500 |
|
}, |
|
{ |
|
"epoch": 24.83, |
|
"learning_rate": 3.517520498248999e-06, |
|
"loss": 1.2686, |
|
"step": 403000 |
|
}, |
|
{ |
|
"epoch": 24.86, |
|
"learning_rate": 3.496550422547026e-06, |
|
"loss": 1.2733, |
|
"step": 403500 |
|
}, |
|
{ |
|
"epoch": 24.89, |
|
"learning_rate": 3.475580346845052e-06, |
|
"loss": 1.2718, |
|
"step": 404000 |
|
}, |
|
{ |
|
"epoch": 24.92, |
|
"learning_rate": 3.454610271143079e-06, |
|
"loss": 1.2678, |
|
"step": 404500 |
|
}, |
|
{ |
|
"epoch": 24.96, |
|
"learning_rate": 3.433640195441106e-06, |
|
"loss": 1.2721, |
|
"step": 405000 |
|
}, |
|
{ |
|
"epoch": 24.99, |
|
"learning_rate": 3.4126701197391325e-06, |
|
"loss": 1.2799, |
|
"step": 405500 |
|
}, |
|
{ |
|
"epoch": 25.02, |
|
"learning_rate": 3.3917000440371596e-06, |
|
"loss": 1.2721, |
|
"step": 406000 |
|
}, |
|
{ |
|
"epoch": 25.05, |
|
"learning_rate": 3.370729968335186e-06, |
|
"loss": 1.268, |
|
"step": 406500 |
|
}, |
|
{ |
|
"epoch": 25.08, |
|
"learning_rate": 3.3497598926332125e-06, |
|
"loss": 1.2737, |
|
"step": 407000 |
|
}, |
|
{ |
|
"epoch": 25.11, |
|
"learning_rate": 3.3287898169312396e-06, |
|
"loss": 1.2692, |
|
"step": 407500 |
|
}, |
|
{ |
|
"epoch": 25.14, |
|
"learning_rate": 3.3078197412292662e-06, |
|
"loss": 1.2697, |
|
"step": 408000 |
|
}, |
|
{ |
|
"epoch": 25.17, |
|
"learning_rate": 3.286849665527293e-06, |
|
"loss": 1.271, |
|
"step": 408500 |
|
}, |
|
{ |
|
"epoch": 25.2, |
|
"learning_rate": 3.265879589825319e-06, |
|
"loss": 1.2711, |
|
"step": 409000 |
|
}, |
|
{ |
|
"epoch": 25.23, |
|
"learning_rate": 3.2449095141233463e-06, |
|
"loss": 1.2736, |
|
"step": 409500 |
|
}, |
|
{ |
|
"epoch": 25.26, |
|
"learning_rate": 3.223939438421373e-06, |
|
"loss": 1.2667, |
|
"step": 410000 |
|
}, |
|
{ |
|
"epoch": 25.29, |
|
"learning_rate": 3.2029693627193996e-06, |
|
"loss": 1.2706, |
|
"step": 410500 |
|
}, |
|
{ |
|
"epoch": 25.33, |
|
"learning_rate": 3.1819992870174267e-06, |
|
"loss": 1.276, |
|
"step": 411000 |
|
}, |
|
{ |
|
"epoch": 25.36, |
|
"learning_rate": 3.1610292113154534e-06, |
|
"loss": 1.2682, |
|
"step": 411500 |
|
}, |
|
{ |
|
"epoch": 25.39, |
|
"learning_rate": 3.1400591356134796e-06, |
|
"loss": 1.2695, |
|
"step": 412000 |
|
}, |
|
{ |
|
"epoch": 25.42, |
|
"learning_rate": 3.1190890599115063e-06, |
|
"loss": 1.2733, |
|
"step": 412500 |
|
}, |
|
{ |
|
"epoch": 25.45, |
|
"learning_rate": 3.0981189842095334e-06, |
|
"loss": 1.273, |
|
"step": 413000 |
|
}, |
|
{ |
|
"epoch": 25.48, |
|
"learning_rate": 3.07714890850756e-06, |
|
"loss": 1.2699, |
|
"step": 413500 |
|
}, |
|
{ |
|
"epoch": 25.51, |
|
"learning_rate": 3.0561788328055867e-06, |
|
"loss": 1.2692, |
|
"step": 414000 |
|
}, |
|
{ |
|
"epoch": 25.54, |
|
"learning_rate": 3.0352087571036134e-06, |
|
"loss": 1.2688, |
|
"step": 414500 |
|
}, |
|
{ |
|
"epoch": 25.57, |
|
"learning_rate": 3.01423868140164e-06, |
|
"loss": 1.27, |
|
"step": 415000 |
|
}, |
|
{ |
|
"epoch": 25.6, |
|
"learning_rate": 2.9932686056996667e-06, |
|
"loss": 1.2701, |
|
"step": 415500 |
|
}, |
|
{ |
|
"epoch": 25.63, |
|
"learning_rate": 2.972298529997694e-06, |
|
"loss": 1.2712, |
|
"step": 416000 |
|
}, |
|
{ |
|
"epoch": 25.66, |
|
"learning_rate": 2.9513284542957205e-06, |
|
"loss": 1.272, |
|
"step": 416500 |
|
}, |
|
{ |
|
"epoch": 25.69, |
|
"learning_rate": 2.9303583785937467e-06, |
|
"loss": 1.264, |
|
"step": 417000 |
|
}, |
|
{ |
|
"epoch": 25.73, |
|
"learning_rate": 2.9093883028917734e-06, |
|
"loss": 1.2692, |
|
"step": 417500 |
|
}, |
|
{ |
|
"epoch": 25.76, |
|
"learning_rate": 2.8884182271898005e-06, |
|
"loss": 1.2689, |
|
"step": 418000 |
|
}, |
|
{ |
|
"epoch": 25.79, |
|
"learning_rate": 2.867448151487827e-06, |
|
"loss": 1.2705, |
|
"step": 418500 |
|
}, |
|
{ |
|
"epoch": 25.82, |
|
"learning_rate": 2.846478075785854e-06, |
|
"loss": 1.2724, |
|
"step": 419000 |
|
}, |
|
{ |
|
"epoch": 25.85, |
|
"learning_rate": 2.82550800008388e-06, |
|
"loss": 1.2646, |
|
"step": 419500 |
|
}, |
|
{ |
|
"epoch": 25.88, |
|
"learning_rate": 2.804537924381907e-06, |
|
"loss": 1.2666, |
|
"step": 420000 |
|
}, |
|
{ |
|
"epoch": 25.91, |
|
"learning_rate": 2.783567848679934e-06, |
|
"loss": 1.2604, |
|
"step": 420500 |
|
}, |
|
{ |
|
"epoch": 25.94, |
|
"learning_rate": 2.7625977729779605e-06, |
|
"loss": 1.2696, |
|
"step": 421000 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"learning_rate": 2.7416276972759876e-06, |
|
"loss": 1.2694, |
|
"step": 421500 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"learning_rate": 2.7206576215740143e-06, |
|
"loss": 1.268, |
|
"step": 422000 |
|
}, |
|
{ |
|
"epoch": 26.03, |
|
"learning_rate": 2.6996875458720405e-06, |
|
"loss": 1.2649, |
|
"step": 422500 |
|
}, |
|
{ |
|
"epoch": 26.06, |
|
"learning_rate": 2.6787174701700676e-06, |
|
"loss": 1.2719, |
|
"step": 423000 |
|
}, |
|
{ |
|
"epoch": 26.1, |
|
"learning_rate": 2.6577473944680943e-06, |
|
"loss": 1.2722, |
|
"step": 423500 |
|
}, |
|
{ |
|
"epoch": 26.13, |
|
"learning_rate": 2.636777318766121e-06, |
|
"loss": 1.2684, |
|
"step": 424000 |
|
}, |
|
{ |
|
"epoch": 26.16, |
|
"learning_rate": 2.615807243064148e-06, |
|
"loss": 1.2734, |
|
"step": 424500 |
|
}, |
|
{ |
|
"epoch": 26.19, |
|
"learning_rate": 2.5948371673621743e-06, |
|
"loss": 1.2674, |
|
"step": 425000 |
|
}, |
|
{ |
|
"epoch": 26.22, |
|
"learning_rate": 2.573867091660201e-06, |
|
"loss": 1.2631, |
|
"step": 425500 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 2.5528970159582276e-06, |
|
"loss": 1.2697, |
|
"step": 426000 |
|
}, |
|
{ |
|
"epoch": 26.28, |
|
"learning_rate": 2.5319269402562547e-06, |
|
"loss": 1.2682, |
|
"step": 426500 |
|
}, |
|
{ |
|
"epoch": 26.31, |
|
"learning_rate": 2.5109568645542814e-06, |
|
"loss": 1.2729, |
|
"step": 427000 |
|
}, |
|
{ |
|
"epoch": 26.34, |
|
"learning_rate": 2.489986788852308e-06, |
|
"loss": 1.2661, |
|
"step": 427500 |
|
}, |
|
{ |
|
"epoch": 26.37, |
|
"learning_rate": 2.4690167131503347e-06, |
|
"loss": 1.2705, |
|
"step": 428000 |
|
}, |
|
{ |
|
"epoch": 26.4, |
|
"learning_rate": 2.4480466374483614e-06, |
|
"loss": 1.2699, |
|
"step": 428500 |
|
}, |
|
{ |
|
"epoch": 26.43, |
|
"learning_rate": 2.427076561746388e-06, |
|
"loss": 1.2705, |
|
"step": 429000 |
|
}, |
|
{ |
|
"epoch": 26.46, |
|
"learning_rate": 2.4061064860444148e-06, |
|
"loss": 1.2708, |
|
"step": 429500 |
|
}, |
|
{ |
|
"epoch": 26.5, |
|
"learning_rate": 2.3851364103424414e-06, |
|
"loss": 1.2694, |
|
"step": 430000 |
|
}, |
|
{ |
|
"epoch": 26.53, |
|
"learning_rate": 2.3641663346404685e-06, |
|
"loss": 1.2632, |
|
"step": 430500 |
|
}, |
|
{ |
|
"epoch": 26.56, |
|
"learning_rate": 2.3431962589384948e-06, |
|
"loss": 1.2693, |
|
"step": 431000 |
|
}, |
|
{ |
|
"epoch": 26.59, |
|
"learning_rate": 2.322226183236522e-06, |
|
"loss": 1.2687, |
|
"step": 431500 |
|
}, |
|
{ |
|
"epoch": 26.62, |
|
"learning_rate": 2.301256107534548e-06, |
|
"loss": 1.2635, |
|
"step": 432000 |
|
}, |
|
{ |
|
"epoch": 26.65, |
|
"learning_rate": 2.280286031832575e-06, |
|
"loss": 1.2701, |
|
"step": 432500 |
|
}, |
|
{ |
|
"epoch": 26.68, |
|
"learning_rate": 2.259315956130602e-06, |
|
"loss": 1.27, |
|
"step": 433000 |
|
}, |
|
{ |
|
"epoch": 26.71, |
|
"learning_rate": 2.2383458804286285e-06, |
|
"loss": 1.2635, |
|
"step": 433500 |
|
}, |
|
{ |
|
"epoch": 26.74, |
|
"learning_rate": 2.2173758047266552e-06, |
|
"loss": 1.2722, |
|
"step": 434000 |
|
}, |
|
{ |
|
"epoch": 26.77, |
|
"learning_rate": 2.196405729024682e-06, |
|
"loss": 1.2657, |
|
"step": 434500 |
|
}, |
|
{ |
|
"epoch": 26.8, |
|
"learning_rate": 2.1754356533227086e-06, |
|
"loss": 1.2601, |
|
"step": 435000 |
|
}, |
|
{ |
|
"epoch": 26.83, |
|
"learning_rate": 2.1544655776207356e-06, |
|
"loss": 1.2641, |
|
"step": 435500 |
|
}, |
|
{ |
|
"epoch": 26.87, |
|
"learning_rate": 2.133495501918762e-06, |
|
"loss": 1.267, |
|
"step": 436000 |
|
}, |
|
{ |
|
"epoch": 26.9, |
|
"learning_rate": 2.112525426216789e-06, |
|
"loss": 1.2722, |
|
"step": 436500 |
|
}, |
|
{ |
|
"epoch": 26.93, |
|
"learning_rate": 2.0915553505148157e-06, |
|
"loss": 1.2638, |
|
"step": 437000 |
|
}, |
|
{ |
|
"epoch": 26.96, |
|
"learning_rate": 2.0705852748128423e-06, |
|
"loss": 1.269, |
|
"step": 437500 |
|
}, |
|
{ |
|
"epoch": 26.99, |
|
"learning_rate": 2.049615199110869e-06, |
|
"loss": 1.2711, |
|
"step": 438000 |
|
}, |
|
{ |
|
"epoch": 27.02, |
|
"learning_rate": 2.0286451234088957e-06, |
|
"loss": 1.266, |
|
"step": 438500 |
|
}, |
|
{ |
|
"epoch": 27.05, |
|
"learning_rate": 2.0076750477069223e-06, |
|
"loss": 1.2658, |
|
"step": 439000 |
|
}, |
|
{ |
|
"epoch": 27.08, |
|
"learning_rate": 1.986704972004949e-06, |
|
"loss": 1.2638, |
|
"step": 439500 |
|
}, |
|
{ |
|
"epoch": 27.11, |
|
"learning_rate": 1.9657348963029757e-06, |
|
"loss": 1.2662, |
|
"step": 440000 |
|
}, |
|
{ |
|
"epoch": 27.14, |
|
"learning_rate": 1.9447648206010023e-06, |
|
"loss": 1.2596, |
|
"step": 440500 |
|
}, |
|
{ |
|
"epoch": 27.17, |
|
"learning_rate": 1.9237947448990294e-06, |
|
"loss": 1.268, |
|
"step": 441000 |
|
}, |
|
{ |
|
"epoch": 27.2, |
|
"learning_rate": 1.902824669197056e-06, |
|
"loss": 1.264, |
|
"step": 441500 |
|
}, |
|
{ |
|
"epoch": 27.24, |
|
"learning_rate": 1.8818545934950828e-06, |
|
"loss": 1.2689, |
|
"step": 442000 |
|
}, |
|
{ |
|
"epoch": 27.27, |
|
"learning_rate": 1.8608845177931095e-06, |
|
"loss": 1.2614, |
|
"step": 442500 |
|
}, |
|
{ |
|
"epoch": 27.3, |
|
"learning_rate": 1.8399144420911361e-06, |
|
"loss": 1.2648, |
|
"step": 443000 |
|
}, |
|
{ |
|
"epoch": 27.33, |
|
"learning_rate": 1.8189443663891628e-06, |
|
"loss": 1.2694, |
|
"step": 443500 |
|
}, |
|
{ |
|
"epoch": 27.36, |
|
"learning_rate": 1.7979742906871895e-06, |
|
"loss": 1.2676, |
|
"step": 444000 |
|
}, |
|
{ |
|
"epoch": 27.39, |
|
"learning_rate": 1.7770042149852163e-06, |
|
"loss": 1.2672, |
|
"step": 444500 |
|
}, |
|
{ |
|
"epoch": 27.42, |
|
"learning_rate": 1.756034139283243e-06, |
|
"loss": 1.2616, |
|
"step": 445000 |
|
}, |
|
{ |
|
"epoch": 27.45, |
|
"learning_rate": 1.7350640635812697e-06, |
|
"loss": 1.2684, |
|
"step": 445500 |
|
}, |
|
{ |
|
"epoch": 27.48, |
|
"learning_rate": 1.7140939878792964e-06, |
|
"loss": 1.27, |
|
"step": 446000 |
|
}, |
|
{ |
|
"epoch": 27.51, |
|
"learning_rate": 1.693123912177323e-06, |
|
"loss": 1.2657, |
|
"step": 446500 |
|
}, |
|
{ |
|
"epoch": 27.54, |
|
"learning_rate": 1.6721538364753497e-06, |
|
"loss": 1.2603, |
|
"step": 447000 |
|
}, |
|
{ |
|
"epoch": 27.57, |
|
"learning_rate": 1.6511837607733766e-06, |
|
"loss": 1.2679, |
|
"step": 447500 |
|
}, |
|
{ |
|
"epoch": 27.6, |
|
"learning_rate": 1.6302136850714032e-06, |
|
"loss": 1.2692, |
|
"step": 448000 |
|
}, |
|
{ |
|
"epoch": 27.64, |
|
"learning_rate": 1.60924360936943e-06, |
|
"loss": 1.2659, |
|
"step": 448500 |
|
}, |
|
{ |
|
"epoch": 27.67, |
|
"learning_rate": 1.5882735336674568e-06, |
|
"loss": 1.2624, |
|
"step": 449000 |
|
}, |
|
{ |
|
"epoch": 27.7, |
|
"learning_rate": 1.5673034579654833e-06, |
|
"loss": 1.2632, |
|
"step": 449500 |
|
}, |
|
{ |
|
"epoch": 27.73, |
|
"learning_rate": 1.5463333822635101e-06, |
|
"loss": 1.264, |
|
"step": 450000 |
|
}, |
|
{ |
|
"epoch": 27.76, |
|
"learning_rate": 1.5253633065615366e-06, |
|
"loss": 1.2651, |
|
"step": 450500 |
|
}, |
|
{ |
|
"epoch": 27.79, |
|
"learning_rate": 1.5043932308595635e-06, |
|
"loss": 1.262, |
|
"step": 451000 |
|
}, |
|
{ |
|
"epoch": 27.82, |
|
"learning_rate": 1.4834231551575904e-06, |
|
"loss": 1.2618, |
|
"step": 451500 |
|
}, |
|
{ |
|
"epoch": 27.85, |
|
"learning_rate": 1.4624530794556168e-06, |
|
"loss": 1.2625, |
|
"step": 452000 |
|
}, |
|
{ |
|
"epoch": 27.88, |
|
"learning_rate": 1.4414830037536437e-06, |
|
"loss": 1.2667, |
|
"step": 452500 |
|
}, |
|
{ |
|
"epoch": 27.91, |
|
"learning_rate": 1.4205129280516706e-06, |
|
"loss": 1.2566, |
|
"step": 453000 |
|
}, |
|
{ |
|
"epoch": 27.94, |
|
"learning_rate": 1.399542852349697e-06, |
|
"loss": 1.2645, |
|
"step": 453500 |
|
}, |
|
{ |
|
"epoch": 27.97, |
|
"learning_rate": 1.378572776647724e-06, |
|
"loss": 1.2661, |
|
"step": 454000 |
|
}, |
|
{ |
|
"epoch": 28.01, |
|
"learning_rate": 1.3576027009457504e-06, |
|
"loss": 1.272, |
|
"step": 454500 |
|
}, |
|
{ |
|
"epoch": 28.04, |
|
"learning_rate": 1.3366326252437773e-06, |
|
"loss": 1.2631, |
|
"step": 455000 |
|
}, |
|
{ |
|
"epoch": 28.07, |
|
"learning_rate": 1.3156625495418041e-06, |
|
"loss": 1.2666, |
|
"step": 455500 |
|
}, |
|
{ |
|
"epoch": 28.1, |
|
"learning_rate": 1.2946924738398306e-06, |
|
"loss": 1.2583, |
|
"step": 456000 |
|
}, |
|
{ |
|
"epoch": 28.13, |
|
"learning_rate": 1.2737223981378575e-06, |
|
"loss": 1.2597, |
|
"step": 456500 |
|
}, |
|
{ |
|
"epoch": 28.16, |
|
"learning_rate": 1.252752322435884e-06, |
|
"loss": 1.2678, |
|
"step": 457000 |
|
}, |
|
{ |
|
"epoch": 28.19, |
|
"learning_rate": 1.2317822467339108e-06, |
|
"loss": 1.265, |
|
"step": 457500 |
|
}, |
|
{ |
|
"epoch": 28.22, |
|
"learning_rate": 1.2108121710319375e-06, |
|
"loss": 1.2601, |
|
"step": 458000 |
|
}, |
|
{ |
|
"epoch": 28.25, |
|
"learning_rate": 1.1898420953299644e-06, |
|
"loss": 1.2625, |
|
"step": 458500 |
|
}, |
|
{ |
|
"epoch": 28.28, |
|
"learning_rate": 1.168872019627991e-06, |
|
"loss": 1.2695, |
|
"step": 459000 |
|
}, |
|
{ |
|
"epoch": 28.31, |
|
"learning_rate": 1.1479019439260177e-06, |
|
"loss": 1.2688, |
|
"step": 459500 |
|
}, |
|
{ |
|
"epoch": 28.34, |
|
"learning_rate": 1.1269318682240444e-06, |
|
"loss": 1.2643, |
|
"step": 460000 |
|
}, |
|
{ |
|
"epoch": 28.38, |
|
"learning_rate": 1.105961792522071e-06, |
|
"loss": 1.2694, |
|
"step": 460500 |
|
}, |
|
{ |
|
"epoch": 28.41, |
|
"learning_rate": 1.0849917168200977e-06, |
|
"loss": 1.264, |
|
"step": 461000 |
|
}, |
|
{ |
|
"epoch": 28.44, |
|
"learning_rate": 1.0640216411181244e-06, |
|
"loss": 1.266, |
|
"step": 461500 |
|
}, |
|
{ |
|
"epoch": 28.47, |
|
"learning_rate": 1.0430515654161513e-06, |
|
"loss": 1.268, |
|
"step": 462000 |
|
}, |
|
{ |
|
"epoch": 28.5, |
|
"learning_rate": 1.022081489714178e-06, |
|
"loss": 1.2654, |
|
"step": 462500 |
|
}, |
|
{ |
|
"epoch": 28.53, |
|
"learning_rate": 1.0011114140122046e-06, |
|
"loss": 1.2625, |
|
"step": 463000 |
|
}, |
|
{ |
|
"epoch": 28.56, |
|
"learning_rate": 9.801413383102313e-07, |
|
"loss": 1.2651, |
|
"step": 463500 |
|
}, |
|
{ |
|
"epoch": 28.59, |
|
"learning_rate": 9.59171262608258e-07, |
|
"loss": 1.2644, |
|
"step": 464000 |
|
}, |
|
{ |
|
"epoch": 28.62, |
|
"learning_rate": 9.382011869062848e-07, |
|
"loss": 1.2602, |
|
"step": 464500 |
|
}, |
|
{ |
|
"epoch": 28.65, |
|
"learning_rate": 9.172311112043115e-07, |
|
"loss": 1.2597, |
|
"step": 465000 |
|
}, |
|
{ |
|
"epoch": 28.68, |
|
"learning_rate": 8.962610355023382e-07, |
|
"loss": 1.2638, |
|
"step": 465500 |
|
}, |
|
{ |
|
"epoch": 28.71, |
|
"learning_rate": 8.75290959800365e-07, |
|
"loss": 1.2593, |
|
"step": 466000 |
|
}, |
|
{ |
|
"epoch": 28.74, |
|
"learning_rate": 8.543208840983917e-07, |
|
"loss": 1.2641, |
|
"step": 466500 |
|
}, |
|
{ |
|
"epoch": 28.78, |
|
"learning_rate": 8.333508083964184e-07, |
|
"loss": 1.2642, |
|
"step": 467000 |
|
}, |
|
{ |
|
"epoch": 28.81, |
|
"learning_rate": 8.123807326944451e-07, |
|
"loss": 1.2659, |
|
"step": 467500 |
|
}, |
|
{ |
|
"epoch": 28.84, |
|
"learning_rate": 7.914106569924717e-07, |
|
"loss": 1.2644, |
|
"step": 468000 |
|
}, |
|
{ |
|
"epoch": 28.87, |
|
"learning_rate": 7.704405812904986e-07, |
|
"loss": 1.2615, |
|
"step": 468500 |
|
}, |
|
{ |
|
"epoch": 28.9, |
|
"learning_rate": 7.494705055885253e-07, |
|
"loss": 1.2662, |
|
"step": 469000 |
|
}, |
|
{ |
|
"epoch": 28.93, |
|
"learning_rate": 7.28500429886552e-07, |
|
"loss": 1.2582, |
|
"step": 469500 |
|
}, |
|
{ |
|
"epoch": 28.96, |
|
"learning_rate": 7.075303541845786e-07, |
|
"loss": 1.262, |
|
"step": 470000 |
|
}, |
|
{ |
|
"epoch": 28.99, |
|
"learning_rate": 6.865602784826053e-07, |
|
"loss": 1.2566, |
|
"step": 470500 |
|
}, |
|
{ |
|
"epoch": 29.02, |
|
"learning_rate": 6.655902027806321e-07, |
|
"loss": 1.2596, |
|
"step": 471000 |
|
}, |
|
{ |
|
"epoch": 29.05, |
|
"learning_rate": 6.446201270786589e-07, |
|
"loss": 1.2651, |
|
"step": 471500 |
|
}, |
|
{ |
|
"epoch": 29.08, |
|
"learning_rate": 6.236500513766855e-07, |
|
"loss": 1.2588, |
|
"step": 472000 |
|
}, |
|
{ |
|
"epoch": 29.11, |
|
"learning_rate": 6.026799756747122e-07, |
|
"loss": 1.2596, |
|
"step": 472500 |
|
}, |
|
{ |
|
"epoch": 29.15, |
|
"learning_rate": 5.817098999727389e-07, |
|
"loss": 1.2645, |
|
"step": 473000 |
|
}, |
|
{ |
|
"epoch": 29.18, |
|
"learning_rate": 5.607398242707656e-07, |
|
"loss": 1.2631, |
|
"step": 473500 |
|
}, |
|
{ |
|
"epoch": 29.21, |
|
"learning_rate": 5.397697485687923e-07, |
|
"loss": 1.2614, |
|
"step": 474000 |
|
}, |
|
{ |
|
"epoch": 29.24, |
|
"learning_rate": 5.187996728668191e-07, |
|
"loss": 1.2659, |
|
"step": 474500 |
|
}, |
|
{ |
|
"epoch": 29.27, |
|
"learning_rate": 4.978295971648458e-07, |
|
"loss": 1.262, |
|
"step": 475000 |
|
}, |
|
{ |
|
"epoch": 29.3, |
|
"learning_rate": 4.768595214628725e-07, |
|
"loss": 1.2602, |
|
"step": 475500 |
|
}, |
|
{ |
|
"epoch": 29.33, |
|
"learning_rate": 4.558894457608992e-07, |
|
"loss": 1.2637, |
|
"step": 476000 |
|
}, |
|
{ |
|
"epoch": 29.36, |
|
"learning_rate": 4.34919370058926e-07, |
|
"loss": 1.2657, |
|
"step": 476500 |
|
}, |
|
{ |
|
"epoch": 29.39, |
|
"learning_rate": 4.1394929435695266e-07, |
|
"loss": 1.256, |
|
"step": 477000 |
|
}, |
|
{ |
|
"epoch": 29.42, |
|
"learning_rate": 3.929792186549794e-07, |
|
"loss": 1.2673, |
|
"step": 477500 |
|
}, |
|
{ |
|
"epoch": 29.45, |
|
"learning_rate": 3.720091429530061e-07, |
|
"loss": 1.2656, |
|
"step": 478000 |
|
}, |
|
{ |
|
"epoch": 29.48, |
|
"learning_rate": 3.5103906725103283e-07, |
|
"loss": 1.266, |
|
"step": 478500 |
|
}, |
|
{ |
|
"epoch": 29.52, |
|
"learning_rate": 3.300689915490595e-07, |
|
"loss": 1.2594, |
|
"step": 479000 |
|
}, |
|
{ |
|
"epoch": 29.55, |
|
"learning_rate": 3.090989158470862e-07, |
|
"loss": 1.2628, |
|
"step": 479500 |
|
}, |
|
{ |
|
"epoch": 29.58, |
|
"learning_rate": 2.8812884014511294e-07, |
|
"loss": 1.2628, |
|
"step": 480000 |
|
}, |
|
{ |
|
"epoch": 29.61, |
|
"learning_rate": 2.6715876444313967e-07, |
|
"loss": 1.2603, |
|
"step": 480500 |
|
}, |
|
{ |
|
"epoch": 29.64, |
|
"learning_rate": 2.461886887411664e-07, |
|
"loss": 1.2595, |
|
"step": 481000 |
|
}, |
|
{ |
|
"epoch": 29.67, |
|
"learning_rate": 2.252186130391931e-07, |
|
"loss": 1.2576, |
|
"step": 481500 |
|
}, |
|
{ |
|
"epoch": 29.7, |
|
"learning_rate": 2.042485373372198e-07, |
|
"loss": 1.2598, |
|
"step": 482000 |
|
}, |
|
{ |
|
"epoch": 29.73, |
|
"learning_rate": 1.832784616352465e-07, |
|
"loss": 1.264, |
|
"step": 482500 |
|
}, |
|
{ |
|
"epoch": 29.76, |
|
"learning_rate": 1.6230838593327323e-07, |
|
"loss": 1.2564, |
|
"step": 483000 |
|
}, |
|
{ |
|
"epoch": 29.79, |
|
"learning_rate": 1.4133831023129995e-07, |
|
"loss": 1.2626, |
|
"step": 483500 |
|
}, |
|
{ |
|
"epoch": 29.82, |
|
"learning_rate": 1.2036823452932668e-07, |
|
"loss": 1.262, |
|
"step": 484000 |
|
}, |
|
{ |
|
"epoch": 29.85, |
|
"learning_rate": 9.939815882735339e-08, |
|
"loss": 1.2612, |
|
"step": 484500 |
|
}, |
|
{ |
|
"epoch": 29.88, |
|
"learning_rate": 7.84280831253801e-08, |
|
"loss": 1.2666, |
|
"step": 485000 |
|
}, |
|
{ |
|
"epoch": 29.92, |
|
"learning_rate": 5.7458007423406806e-08, |
|
"loss": 1.2569, |
|
"step": 485500 |
|
}, |
|
{ |
|
"epoch": 29.95, |
|
"learning_rate": 3.6487931721433516e-08, |
|
"loss": 1.2639, |
|
"step": 486000 |
|
}, |
|
{ |
|
"epoch": 29.98, |
|
"learning_rate": 1.5517856019460232e-08, |
|
"loss": 1.2638, |
|
"step": 486500 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"step": 486870, |
|
"total_flos": 8.200903785279652e+18, |
|
"train_loss": 1.3885125489469685, |
|
"train_runtime": 402878.8865, |
|
"train_samples_per_second": 77.338, |
|
"train_steps_per_second": 1.208 |
|
} |
|
], |
|
"max_steps": 486870, |
|
"num_train_epochs": 30, |
|
"total_flos": 8.200903785279652e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|