End of training
Browse files
all_results.json
CHANGED
@@ -1,15 +1,15 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_cer": 0.
|
4 |
-
"eval_loss": 0.
|
5 |
-
"eval_runtime":
|
6 |
"eval_samples": 1647,
|
7 |
-
"eval_samples_per_second":
|
8 |
-
"eval_steps_per_second": 2.
|
9 |
-
"eval_wer": 0.
|
10 |
-
"train_loss": 0.
|
11 |
-
"train_runtime":
|
12 |
"train_samples": 3478,
|
13 |
-
"train_samples_per_second":
|
14 |
-
"train_steps_per_second":
|
15 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 96.33,
|
3 |
+
"eval_cer": 0.13390005847164124,
|
4 |
+
"eval_loss": 0.7359636425971985,
|
5 |
+
"eval_runtime": 83.055,
|
6 |
"eval_samples": 1647,
|
7 |
+
"eval_samples_per_second": 19.83,
|
8 |
+
"eval_steps_per_second": 2.48,
|
9 |
+
"eval_wer": 0.4754366254723726,
|
10 |
+
"train_loss": 0.0,
|
11 |
+
"train_runtime": 113.9955,
|
12 |
"train_samples": 3478,
|
13 |
+
"train_samples_per_second": 152.55,
|
14 |
+
"train_steps_per_second": 4.781
|
15 |
}
|
eval_results.json
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_cer": 0.
|
4 |
-
"eval_loss": 0.
|
5 |
-
"eval_runtime":
|
6 |
"eval_samples": 1647,
|
7 |
-
"eval_samples_per_second":
|
8 |
-
"eval_steps_per_second": 2.
|
9 |
-
"eval_wer": 0.
|
10 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 96.33,
|
3 |
+
"eval_cer": 0.13390005847164124,
|
4 |
+
"eval_loss": 0.7359636425971985,
|
5 |
+
"eval_runtime": 83.055,
|
6 |
"eval_samples": 1647,
|
7 |
+
"eval_samples_per_second": 19.83,
|
8 |
+
"eval_steps_per_second": 2.48,
|
9 |
+
"eval_wer": 0.4754366254723726
|
10 |
}
|
runs/Feb02_08-46-53_job-98a7eb1d-c6a6-4e4b-936b-3898557b5e73/events.out.tfevents.1643792324.job-98a7eb1d-c6a6-4e4b-936b-3898557b5e73.544082.2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08a917ccd708b36361a9ce5fa79328ee4416e8a8f0e92f2d35ec8d196cb7b111
|
3 |
+
size 405
|
train_results.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"train_loss": 0.
|
4 |
-
"train_runtime":
|
5 |
"train_samples": 3478,
|
6 |
-
"train_samples_per_second":
|
7 |
-
"train_steps_per_second":
|
8 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 96.33,
|
3 |
+
"train_loss": 0.0,
|
4 |
+
"train_runtime": 113.9955,
|
5 |
"train_samples": 3478,
|
6 |
+
"train_samples_per_second": 152.55,
|
7 |
+
"train_steps_per_second": 4.781
|
8 |
}
|
trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -63218,2418 +63218,18 @@
|
|
63218 |
"step": 10500
|
63219 |
},
|
63220 |
{
|
63221 |
-
"epoch": 96.
|
63222 |
-
"
|
63223 |
-
"
|
63224 |
-
"
|
63225 |
-
|
63226 |
-
|
63227 |
-
"
|
63228 |
-
"learning_rate": 1.932692307692308e-05,
|
63229 |
-
"loss": 0.1755,
|
63230 |
-
"step": 10502
|
63231 |
-
},
|
63232 |
-
{
|
63233 |
-
"epoch": 96.36,
|
63234 |
-
"learning_rate": 1.9278846153846155e-05,
|
63235 |
-
"loss": 0.0978,
|
63236 |
-
"step": 10503
|
63237 |
-
},
|
63238 |
-
{
|
63239 |
-
"epoch": 96.37,
|
63240 |
-
"learning_rate": 1.923076923076923e-05,
|
63241 |
-
"loss": 0.1605,
|
63242 |
-
"step": 10504
|
63243 |
-
},
|
63244 |
-
{
|
63245 |
-
"epoch": 96.38,
|
63246 |
-
"learning_rate": 1.9182692307692307e-05,
|
63247 |
-
"loss": 0.1321,
|
63248 |
-
"step": 10505
|
63249 |
-
},
|
63250 |
-
{
|
63251 |
-
"epoch": 96.39,
|
63252 |
-
"learning_rate": 1.9134615384615387e-05,
|
63253 |
-
"loss": 0.1023,
|
63254 |
-
"step": 10506
|
63255 |
-
},
|
63256 |
-
{
|
63257 |
-
"epoch": 96.39,
|
63258 |
-
"learning_rate": 1.908653846153846e-05,
|
63259 |
-
"loss": 0.0784,
|
63260 |
-
"step": 10507
|
63261 |
-
},
|
63262 |
-
{
|
63263 |
-
"epoch": 96.4,
|
63264 |
-
"learning_rate": 1.903846153846154e-05,
|
63265 |
-
"loss": 0.069,
|
63266 |
-
"step": 10508
|
63267 |
-
},
|
63268 |
-
{
|
63269 |
-
"epoch": 96.41,
|
63270 |
-
"learning_rate": 1.8990384615384615e-05,
|
63271 |
-
"loss": 0.0694,
|
63272 |
-
"step": 10509
|
63273 |
-
},
|
63274 |
-
{
|
63275 |
-
"epoch": 96.42,
|
63276 |
-
"learning_rate": 1.8942307692307694e-05,
|
63277 |
-
"loss": 0.1372,
|
63278 |
-
"step": 10510
|
63279 |
-
},
|
63280 |
-
{
|
63281 |
-
"epoch": 96.43,
|
63282 |
-
"learning_rate": 1.889423076923077e-05,
|
63283 |
-
"loss": 0.0912,
|
63284 |
-
"step": 10511
|
63285 |
-
},
|
63286 |
-
{
|
63287 |
-
"epoch": 96.44,
|
63288 |
-
"learning_rate": 1.8846153846153846e-05,
|
63289 |
-
"loss": 0.1011,
|
63290 |
-
"step": 10512
|
63291 |
-
},
|
63292 |
-
{
|
63293 |
-
"epoch": 96.45,
|
63294 |
-
"learning_rate": 1.8798076923076926e-05,
|
63295 |
-
"loss": 0.1284,
|
63296 |
-
"step": 10513
|
63297 |
-
},
|
63298 |
-
{
|
63299 |
-
"epoch": 96.46,
|
63300 |
-
"learning_rate": 1.875e-05,
|
63301 |
-
"loss": 0.1563,
|
63302 |
-
"step": 10514
|
63303 |
-
},
|
63304 |
-
{
|
63305 |
-
"epoch": 96.47,
|
63306 |
-
"learning_rate": 1.8701923076923078e-05,
|
63307 |
-
"loss": 0.0666,
|
63308 |
-
"step": 10515
|
63309 |
-
},
|
63310 |
-
{
|
63311 |
-
"epoch": 96.48,
|
63312 |
-
"learning_rate": 1.8653846153846154e-05,
|
63313 |
-
"loss": 0.1044,
|
63314 |
-
"step": 10516
|
63315 |
-
},
|
63316 |
-
{
|
63317 |
-
"epoch": 96.49,
|
63318 |
-
"learning_rate": 1.8605769230769233e-05,
|
63319 |
-
"loss": 0.0935,
|
63320 |
-
"step": 10517
|
63321 |
-
},
|
63322 |
-
{
|
63323 |
-
"epoch": 96.5,
|
63324 |
-
"learning_rate": 1.8557692307692306e-05,
|
63325 |
-
"loss": 0.112,
|
63326 |
-
"step": 10518
|
63327 |
-
},
|
63328 |
-
{
|
63329 |
-
"epoch": 96.5,
|
63330 |
-
"learning_rate": 1.8509615384615385e-05,
|
63331 |
-
"loss": 0.126,
|
63332 |
-
"step": 10519
|
63333 |
-
},
|
63334 |
-
{
|
63335 |
-
"epoch": 96.51,
|
63336 |
-
"learning_rate": 1.8461538461538465e-05,
|
63337 |
-
"loss": 0.1627,
|
63338 |
-
"step": 10520
|
63339 |
-
},
|
63340 |
-
{
|
63341 |
-
"epoch": 96.52,
|
63342 |
-
"learning_rate": 1.8413461538461537e-05,
|
63343 |
-
"loss": 0.1363,
|
63344 |
-
"step": 10521
|
63345 |
-
},
|
63346 |
-
{
|
63347 |
-
"epoch": 96.53,
|
63348 |
-
"learning_rate": 1.8365384615384617e-05,
|
63349 |
-
"loss": 0.1524,
|
63350 |
-
"step": 10522
|
63351 |
-
},
|
63352 |
-
{
|
63353 |
-
"epoch": 96.54,
|
63354 |
-
"learning_rate": 1.8317307692307693e-05,
|
63355 |
-
"loss": 0.1185,
|
63356 |
-
"step": 10523
|
63357 |
-
},
|
63358 |
-
{
|
63359 |
-
"epoch": 96.55,
|
63360 |
-
"learning_rate": 1.8269230769230772e-05,
|
63361 |
-
"loss": 0.111,
|
63362 |
-
"step": 10524
|
63363 |
-
},
|
63364 |
-
{
|
63365 |
-
"epoch": 96.56,
|
63366 |
-
"learning_rate": 1.8221153846153845e-05,
|
63367 |
-
"loss": 0.1093,
|
63368 |
-
"step": 10525
|
63369 |
-
},
|
63370 |
-
{
|
63371 |
-
"epoch": 96.57,
|
63372 |
-
"learning_rate": 1.8173076923076924e-05,
|
63373 |
-
"loss": 0.058,
|
63374 |
-
"step": 10526
|
63375 |
-
},
|
63376 |
-
{
|
63377 |
-
"epoch": 96.58,
|
63378 |
-
"learning_rate": 1.8125e-05,
|
63379 |
-
"loss": 0.0667,
|
63380 |
-
"step": 10527
|
63381 |
-
},
|
63382 |
-
{
|
63383 |
-
"epoch": 96.59,
|
63384 |
-
"learning_rate": 1.8076923076923076e-05,
|
63385 |
-
"loss": 0.1126,
|
63386 |
-
"step": 10528
|
63387 |
-
},
|
63388 |
-
{
|
63389 |
-
"epoch": 96.6,
|
63390 |
-
"learning_rate": 1.8028846153846152e-05,
|
63391 |
-
"loss": 0.0928,
|
63392 |
-
"step": 10529
|
63393 |
-
},
|
63394 |
-
{
|
63395 |
-
"epoch": 96.61,
|
63396 |
-
"learning_rate": 1.798076923076923e-05,
|
63397 |
-
"loss": 0.0966,
|
63398 |
-
"step": 10530
|
63399 |
-
},
|
63400 |
-
{
|
63401 |
-
"epoch": 96.61,
|
63402 |
-
"learning_rate": 1.793269230769231e-05,
|
63403 |
-
"loss": 0.0637,
|
63404 |
-
"step": 10531
|
63405 |
-
},
|
63406 |
-
{
|
63407 |
-
"epoch": 96.62,
|
63408 |
-
"learning_rate": 1.7884615384615384e-05,
|
63409 |
-
"loss": 0.0786,
|
63410 |
-
"step": 10532
|
63411 |
-
},
|
63412 |
-
{
|
63413 |
-
"epoch": 96.63,
|
63414 |
-
"learning_rate": 1.7836538461538463e-05,
|
63415 |
-
"loss": 0.1502,
|
63416 |
-
"step": 10533
|
63417 |
-
},
|
63418 |
-
{
|
63419 |
-
"epoch": 96.64,
|
63420 |
-
"learning_rate": 1.778846153846154e-05,
|
63421 |
-
"loss": 0.0982,
|
63422 |
-
"step": 10534
|
63423 |
-
},
|
63424 |
-
{
|
63425 |
-
"epoch": 96.65,
|
63426 |
-
"learning_rate": 1.7740384615384615e-05,
|
63427 |
-
"loss": 0.1326,
|
63428 |
-
"step": 10535
|
63429 |
-
},
|
63430 |
-
{
|
63431 |
-
"epoch": 96.66,
|
63432 |
-
"learning_rate": 1.769230769230769e-05,
|
63433 |
-
"loss": 0.0662,
|
63434 |
-
"step": 10536
|
63435 |
-
},
|
63436 |
-
{
|
63437 |
-
"epoch": 96.67,
|
63438 |
-
"learning_rate": 1.764423076923077e-05,
|
63439 |
-
"loss": 0.1025,
|
63440 |
-
"step": 10537
|
63441 |
-
},
|
63442 |
-
{
|
63443 |
-
"epoch": 96.68,
|
63444 |
-
"learning_rate": 1.7596153846153846e-05,
|
63445 |
-
"loss": 0.1385,
|
63446 |
-
"step": 10538
|
63447 |
-
},
|
63448 |
-
{
|
63449 |
-
"epoch": 96.69,
|
63450 |
-
"learning_rate": 1.7548076923076922e-05,
|
63451 |
-
"loss": 0.0851,
|
63452 |
-
"step": 10539
|
63453 |
-
},
|
63454 |
-
{
|
63455 |
-
"epoch": 96.7,
|
63456 |
-
"learning_rate": 1.7500000000000002e-05,
|
63457 |
-
"loss": 0.0788,
|
63458 |
-
"step": 10540
|
63459 |
-
},
|
63460 |
-
{
|
63461 |
-
"epoch": 96.71,
|
63462 |
-
"learning_rate": 1.7451923076923078e-05,
|
63463 |
-
"loss": 0.0495,
|
63464 |
-
"step": 10541
|
63465 |
-
},
|
63466 |
-
{
|
63467 |
-
"epoch": 96.72,
|
63468 |
-
"learning_rate": 1.7403846153846154e-05,
|
63469 |
-
"loss": 0.0613,
|
63470 |
-
"step": 10542
|
63471 |
-
},
|
63472 |
-
{
|
63473 |
-
"epoch": 96.72,
|
63474 |
-
"learning_rate": 1.735576923076923e-05,
|
63475 |
-
"loss": 0.083,
|
63476 |
-
"step": 10543
|
63477 |
-
},
|
63478 |
-
{
|
63479 |
-
"epoch": 96.73,
|
63480 |
-
"learning_rate": 1.730769230769231e-05,
|
63481 |
-
"loss": 0.1314,
|
63482 |
-
"step": 10544
|
63483 |
-
},
|
63484 |
-
{
|
63485 |
-
"epoch": 96.74,
|
63486 |
-
"learning_rate": 1.7259615384615385e-05,
|
63487 |
-
"loss": 0.1278,
|
63488 |
-
"step": 10545
|
63489 |
-
},
|
63490 |
-
{
|
63491 |
-
"epoch": 96.75,
|
63492 |
-
"learning_rate": 1.721153846153846e-05,
|
63493 |
-
"loss": 0.1405,
|
63494 |
-
"step": 10546
|
63495 |
-
},
|
63496 |
-
{
|
63497 |
-
"epoch": 96.76,
|
63498 |
-
"learning_rate": 1.7163461538461537e-05,
|
63499 |
-
"loss": 0.0983,
|
63500 |
-
"step": 10547
|
63501 |
-
},
|
63502 |
-
{
|
63503 |
-
"epoch": 96.77,
|
63504 |
-
"learning_rate": 1.7115384615384617e-05,
|
63505 |
-
"loss": 0.1169,
|
63506 |
-
"step": 10548
|
63507 |
-
},
|
63508 |
-
{
|
63509 |
-
"epoch": 96.78,
|
63510 |
-
"learning_rate": 1.706730769230769e-05,
|
63511 |
-
"loss": 0.0914,
|
63512 |
-
"step": 10549
|
63513 |
-
},
|
63514 |
-
{
|
63515 |
-
"epoch": 96.79,
|
63516 |
-
"learning_rate": 1.701923076923077e-05,
|
63517 |
-
"loss": 0.0971,
|
63518 |
-
"step": 10550
|
63519 |
-
},
|
63520 |
-
{
|
63521 |
-
"epoch": 96.8,
|
63522 |
-
"learning_rate": 1.6971153846153848e-05,
|
63523 |
-
"loss": 0.0951,
|
63524 |
-
"step": 10551
|
63525 |
-
},
|
63526 |
-
{
|
63527 |
-
"epoch": 96.81,
|
63528 |
-
"learning_rate": 1.6923076923076924e-05,
|
63529 |
-
"loss": 0.0813,
|
63530 |
-
"step": 10552
|
63531 |
-
},
|
63532 |
-
{
|
63533 |
-
"epoch": 96.82,
|
63534 |
-
"learning_rate": 1.6875e-05,
|
63535 |
-
"loss": 0.0793,
|
63536 |
-
"step": 10553
|
63537 |
-
},
|
63538 |
-
{
|
63539 |
-
"epoch": 96.83,
|
63540 |
-
"learning_rate": 1.6826923076923076e-05,
|
63541 |
-
"loss": 0.0885,
|
63542 |
-
"step": 10554
|
63543 |
-
},
|
63544 |
-
{
|
63545 |
-
"epoch": 96.83,
|
63546 |
-
"learning_rate": 1.6778846153846156e-05,
|
63547 |
-
"loss": 0.073,
|
63548 |
-
"step": 10555
|
63549 |
-
},
|
63550 |
-
{
|
63551 |
-
"epoch": 96.84,
|
63552 |
-
"learning_rate": 1.673076923076923e-05,
|
63553 |
-
"loss": 0.0892,
|
63554 |
-
"step": 10556
|
63555 |
-
},
|
63556 |
-
{
|
63557 |
-
"epoch": 96.85,
|
63558 |
-
"learning_rate": 1.6682692307692308e-05,
|
63559 |
-
"loss": 0.0787,
|
63560 |
-
"step": 10557
|
63561 |
-
},
|
63562 |
-
{
|
63563 |
-
"epoch": 96.86,
|
63564 |
-
"learning_rate": 1.6634615384615384e-05,
|
63565 |
-
"loss": 0.0782,
|
63566 |
-
"step": 10558
|
63567 |
-
},
|
63568 |
-
{
|
63569 |
-
"epoch": 96.87,
|
63570 |
-
"learning_rate": 1.6586538461538463e-05,
|
63571 |
-
"loss": 0.0622,
|
63572 |
-
"step": 10559
|
63573 |
-
},
|
63574 |
-
{
|
63575 |
-
"epoch": 96.88,
|
63576 |
-
"learning_rate": 1.653846153846154e-05,
|
63577 |
-
"loss": 0.0956,
|
63578 |
-
"step": 10560
|
63579 |
-
},
|
63580 |
-
{
|
63581 |
-
"epoch": 96.89,
|
63582 |
-
"learning_rate": 1.6490384615384615e-05,
|
63583 |
-
"loss": 0.0741,
|
63584 |
-
"step": 10561
|
63585 |
-
},
|
63586 |
-
{
|
63587 |
-
"epoch": 96.9,
|
63588 |
-
"learning_rate": 1.6442307692307695e-05,
|
63589 |
-
"loss": 0.1073,
|
63590 |
-
"step": 10562
|
63591 |
-
},
|
63592 |
-
{
|
63593 |
-
"epoch": 96.91,
|
63594 |
-
"learning_rate": 1.639423076923077e-05,
|
63595 |
-
"loss": 0.0747,
|
63596 |
-
"step": 10563
|
63597 |
-
},
|
63598 |
-
{
|
63599 |
-
"epoch": 96.92,
|
63600 |
-
"learning_rate": 1.6346153846153847e-05,
|
63601 |
-
"loss": 0.1143,
|
63602 |
-
"step": 10564
|
63603 |
-
},
|
63604 |
-
{
|
63605 |
-
"epoch": 96.93,
|
63606 |
-
"learning_rate": 1.6298076923076923e-05,
|
63607 |
-
"loss": 0.0906,
|
63608 |
-
"step": 10565
|
63609 |
-
},
|
63610 |
-
{
|
63611 |
-
"epoch": 96.94,
|
63612 |
-
"learning_rate": 1.6250000000000002e-05,
|
63613 |
-
"loss": 0.1303,
|
63614 |
-
"step": 10566
|
63615 |
-
},
|
63616 |
-
{
|
63617 |
-
"epoch": 96.94,
|
63618 |
-
"learning_rate": 1.6201923076923075e-05,
|
63619 |
-
"loss": 0.1188,
|
63620 |
-
"step": 10567
|
63621 |
-
},
|
63622 |
-
{
|
63623 |
-
"epoch": 96.95,
|
63624 |
-
"learning_rate": 1.6153846153846154e-05,
|
63625 |
-
"loss": 0.0773,
|
63626 |
-
"step": 10568
|
63627 |
-
},
|
63628 |
-
{
|
63629 |
-
"epoch": 96.96,
|
63630 |
-
"learning_rate": 1.610576923076923e-05,
|
63631 |
-
"loss": 0.1376,
|
63632 |
-
"step": 10569
|
63633 |
-
},
|
63634 |
-
{
|
63635 |
-
"epoch": 96.97,
|
63636 |
-
"learning_rate": 1.605769230769231e-05,
|
63637 |
-
"loss": 0.091,
|
63638 |
-
"step": 10570
|
63639 |
-
},
|
63640 |
-
{
|
63641 |
-
"epoch": 96.98,
|
63642 |
-
"learning_rate": 1.6009615384615385e-05,
|
63643 |
-
"loss": 0.0843,
|
63644 |
-
"step": 10571
|
63645 |
-
},
|
63646 |
-
{
|
63647 |
-
"epoch": 96.99,
|
63648 |
-
"learning_rate": 1.596153846153846e-05,
|
63649 |
-
"loss": 0.1066,
|
63650 |
-
"step": 10572
|
63651 |
-
},
|
63652 |
-
{
|
63653 |
-
"epoch": 97.0,
|
63654 |
-
"learning_rate": 1.591346153846154e-05,
|
63655 |
-
"loss": 0.0846,
|
63656 |
-
"step": 10573
|
63657 |
-
},
|
63658 |
-
{
|
63659 |
-
"epoch": 97.01,
|
63660 |
-
"learning_rate": 1.5865384615384613e-05,
|
63661 |
-
"loss": 0.1501,
|
63662 |
-
"step": 10574
|
63663 |
-
},
|
63664 |
-
{
|
63665 |
-
"epoch": 97.02,
|
63666 |
-
"learning_rate": 1.5817307692307693e-05,
|
63667 |
-
"loss": 0.1045,
|
63668 |
-
"step": 10575
|
63669 |
-
},
|
63670 |
-
{
|
63671 |
-
"epoch": 97.03,
|
63672 |
-
"learning_rate": 1.576923076923077e-05,
|
63673 |
-
"loss": 0.1179,
|
63674 |
-
"step": 10576
|
63675 |
-
},
|
63676 |
-
{
|
63677 |
-
"epoch": 97.04,
|
63678 |
-
"learning_rate": 1.5721153846153848e-05,
|
63679 |
-
"loss": 0.1085,
|
63680 |
-
"step": 10577
|
63681 |
-
},
|
63682 |
-
{
|
63683 |
-
"epoch": 97.05,
|
63684 |
-
"learning_rate": 1.567307692307692e-05,
|
63685 |
-
"loss": 0.1578,
|
63686 |
-
"step": 10578
|
63687 |
-
},
|
63688 |
-
{
|
63689 |
-
"epoch": 97.06,
|
63690 |
-
"learning_rate": 1.5625e-05,
|
63691 |
-
"loss": 0.0982,
|
63692 |
-
"step": 10579
|
63693 |
-
},
|
63694 |
-
{
|
63695 |
-
"epoch": 97.06,
|
63696 |
-
"learning_rate": 1.5576923076923076e-05,
|
63697 |
-
"loss": 0.0764,
|
63698 |
-
"step": 10580
|
63699 |
-
},
|
63700 |
-
{
|
63701 |
-
"epoch": 97.07,
|
63702 |
-
"learning_rate": 1.5528846153846152e-05,
|
63703 |
-
"loss": 0.1044,
|
63704 |
-
"step": 10581
|
63705 |
-
},
|
63706 |
-
{
|
63707 |
-
"epoch": 97.08,
|
63708 |
-
"learning_rate": 1.5480769230769232e-05,
|
63709 |
-
"loss": 0.1084,
|
63710 |
-
"step": 10582
|
63711 |
-
},
|
63712 |
-
{
|
63713 |
-
"epoch": 97.09,
|
63714 |
-
"learning_rate": 1.5432692307692308e-05,
|
63715 |
-
"loss": 0.1081,
|
63716 |
-
"step": 10583
|
63717 |
-
},
|
63718 |
-
{
|
63719 |
-
"epoch": 97.1,
|
63720 |
-
"learning_rate": 1.5384615384615387e-05,
|
63721 |
-
"loss": 0.0863,
|
63722 |
-
"step": 10584
|
63723 |
-
},
|
63724 |
-
{
|
63725 |
-
"epoch": 97.11,
|
63726 |
-
"learning_rate": 1.5336538461538463e-05,
|
63727 |
-
"loss": 0.1192,
|
63728 |
-
"step": 10585
|
63729 |
-
},
|
63730 |
-
{
|
63731 |
-
"epoch": 97.12,
|
63732 |
-
"learning_rate": 1.528846153846154e-05,
|
63733 |
-
"loss": 0.125,
|
63734 |
-
"step": 10586
|
63735 |
-
},
|
63736 |
-
{
|
63737 |
-
"epoch": 97.13,
|
63738 |
-
"learning_rate": 1.5240384615384615e-05,
|
63739 |
-
"loss": 0.0623,
|
63740 |
-
"step": 10587
|
63741 |
-
},
|
63742 |
-
{
|
63743 |
-
"epoch": 97.14,
|
63744 |
-
"learning_rate": 1.5192307692307693e-05,
|
63745 |
-
"loss": 0.0727,
|
63746 |
-
"step": 10588
|
63747 |
-
},
|
63748 |
-
{
|
63749 |
-
"epoch": 97.15,
|
63750 |
-
"learning_rate": 1.5144230769230769e-05,
|
63751 |
-
"loss": 0.0703,
|
63752 |
-
"step": 10589
|
63753 |
-
},
|
63754 |
-
{
|
63755 |
-
"epoch": 97.16,
|
63756 |
-
"learning_rate": 1.5096153846153847e-05,
|
63757 |
-
"loss": 0.113,
|
63758 |
-
"step": 10590
|
63759 |
-
},
|
63760 |
-
{
|
63761 |
-
"epoch": 97.17,
|
63762 |
-
"learning_rate": 1.5048076923076923e-05,
|
63763 |
-
"loss": 0.0971,
|
63764 |
-
"step": 10591
|
63765 |
-
},
|
63766 |
-
{
|
63767 |
-
"epoch": 97.17,
|
63768 |
-
"learning_rate": 1.5e-05,
|
63769 |
-
"loss": 0.0691,
|
63770 |
-
"step": 10592
|
63771 |
-
},
|
63772 |
-
{
|
63773 |
-
"epoch": 97.18,
|
63774 |
-
"learning_rate": 1.4951923076923076e-05,
|
63775 |
-
"loss": 0.1618,
|
63776 |
-
"step": 10593
|
63777 |
-
},
|
63778 |
-
{
|
63779 |
-
"epoch": 97.19,
|
63780 |
-
"learning_rate": 1.4903846153846156e-05,
|
63781 |
-
"loss": 0.086,
|
63782 |
-
"step": 10594
|
63783 |
-
},
|
63784 |
-
{
|
63785 |
-
"epoch": 97.2,
|
63786 |
-
"learning_rate": 1.4855769230769232e-05,
|
63787 |
-
"loss": 0.1307,
|
63788 |
-
"step": 10595
|
63789 |
-
},
|
63790 |
-
{
|
63791 |
-
"epoch": 97.21,
|
63792 |
-
"learning_rate": 1.4807692307692308e-05,
|
63793 |
-
"loss": 0.0762,
|
63794 |
-
"step": 10596
|
63795 |
-
},
|
63796 |
-
{
|
63797 |
-
"epoch": 97.22,
|
63798 |
-
"learning_rate": 1.4759615384615386e-05,
|
63799 |
-
"loss": 0.1148,
|
63800 |
-
"step": 10597
|
63801 |
-
},
|
63802 |
-
{
|
63803 |
-
"epoch": 97.23,
|
63804 |
-
"learning_rate": 1.4711538461538462e-05,
|
63805 |
-
"loss": 0.0723,
|
63806 |
-
"step": 10598
|
63807 |
-
},
|
63808 |
-
{
|
63809 |
-
"epoch": 97.24,
|
63810 |
-
"learning_rate": 1.466346153846154e-05,
|
63811 |
-
"loss": 0.0957,
|
63812 |
-
"step": 10599
|
63813 |
-
},
|
63814 |
-
{
|
63815 |
-
"epoch": 97.25,
|
63816 |
-
"learning_rate": 1.4615384615384615e-05,
|
63817 |
-
"loss": 0.085,
|
63818 |
-
"step": 10600
|
63819 |
-
},
|
63820 |
-
{
|
63821 |
-
"epoch": 97.26,
|
63822 |
-
"learning_rate": 1.4567307692307693e-05,
|
63823 |
-
"loss": 0.1458,
|
63824 |
-
"step": 10601
|
63825 |
-
},
|
63826 |
-
{
|
63827 |
-
"epoch": 97.27,
|
63828 |
-
"learning_rate": 1.4519230769230769e-05,
|
63829 |
-
"loss": 0.1268,
|
63830 |
-
"step": 10602
|
63831 |
-
},
|
63832 |
-
{
|
63833 |
-
"epoch": 97.28,
|
63834 |
-
"learning_rate": 1.4471153846153845e-05,
|
63835 |
-
"loss": 0.1339,
|
63836 |
-
"step": 10603
|
63837 |
-
},
|
63838 |
-
{
|
63839 |
-
"epoch": 97.28,
|
63840 |
-
"learning_rate": 1.4423076923076924e-05,
|
63841 |
-
"loss": 0.1093,
|
63842 |
-
"step": 10604
|
63843 |
-
},
|
63844 |
-
{
|
63845 |
-
"epoch": 97.29,
|
63846 |
-
"learning_rate": 1.4375e-05,
|
63847 |
-
"loss": 0.1123,
|
63848 |
-
"step": 10605
|
63849 |
-
},
|
63850 |
-
{
|
63851 |
-
"epoch": 97.3,
|
63852 |
-
"learning_rate": 1.4326923076923078e-05,
|
63853 |
-
"loss": 0.1232,
|
63854 |
-
"step": 10606
|
63855 |
-
},
|
63856 |
-
{
|
63857 |
-
"epoch": 97.31,
|
63858 |
-
"learning_rate": 1.4278846153846154e-05,
|
63859 |
-
"loss": 0.0946,
|
63860 |
-
"step": 10607
|
63861 |
-
},
|
63862 |
-
{
|
63863 |
-
"epoch": 97.32,
|
63864 |
-
"learning_rate": 1.4230769230769232e-05,
|
63865 |
-
"loss": 0.0789,
|
63866 |
-
"step": 10608
|
63867 |
-
},
|
63868 |
-
{
|
63869 |
-
"epoch": 97.33,
|
63870 |
-
"learning_rate": 1.4182692307692308e-05,
|
63871 |
-
"loss": 0.1177,
|
63872 |
-
"step": 10609
|
63873 |
-
},
|
63874 |
-
{
|
63875 |
-
"epoch": 97.34,
|
63876 |
-
"learning_rate": 1.4134615384615384e-05,
|
63877 |
-
"loss": 0.0587,
|
63878 |
-
"step": 10610
|
63879 |
-
},
|
63880 |
-
{
|
63881 |
-
"epoch": 97.35,
|
63882 |
-
"learning_rate": 1.4086538461538462e-05,
|
63883 |
-
"loss": 0.088,
|
63884 |
-
"step": 10611
|
63885 |
-
},
|
63886 |
-
{
|
63887 |
-
"epoch": 97.36,
|
63888 |
-
"learning_rate": 1.4038461538461538e-05,
|
63889 |
-
"loss": 0.0568,
|
63890 |
-
"step": 10612
|
63891 |
-
},
|
63892 |
-
{
|
63893 |
-
"epoch": 97.37,
|
63894 |
-
"learning_rate": 1.3990384615384615e-05,
|
63895 |
-
"loss": 0.096,
|
63896 |
-
"step": 10613
|
63897 |
-
},
|
63898 |
-
{
|
63899 |
-
"epoch": 97.38,
|
63900 |
-
"learning_rate": 1.3942307692307693e-05,
|
63901 |
-
"loss": 0.1626,
|
63902 |
-
"step": 10614
|
63903 |
-
},
|
63904 |
-
{
|
63905 |
-
"epoch": 97.39,
|
63906 |
-
"learning_rate": 1.389423076923077e-05,
|
63907 |
-
"loss": 0.0847,
|
63908 |
-
"step": 10615
|
63909 |
-
},
|
63910 |
-
{
|
63911 |
-
"epoch": 97.39,
|
63912 |
-
"learning_rate": 1.3846153846153847e-05,
|
63913 |
-
"loss": 0.1289,
|
63914 |
-
"step": 10616
|
63915 |
-
},
|
63916 |
-
{
|
63917 |
-
"epoch": 97.4,
|
63918 |
-
"learning_rate": 1.3798076923076923e-05,
|
63919 |
-
"loss": 0.0813,
|
63920 |
-
"step": 10617
|
63921 |
-
},
|
63922 |
-
{
|
63923 |
-
"epoch": 97.41,
|
63924 |
-
"learning_rate": 1.375e-05,
|
63925 |
-
"loss": 0.0931,
|
63926 |
-
"step": 10618
|
63927 |
-
},
|
63928 |
-
{
|
63929 |
-
"epoch": 97.42,
|
63930 |
-
"learning_rate": 1.3701923076923077e-05,
|
63931 |
-
"loss": 0.1229,
|
63932 |
-
"step": 10619
|
63933 |
-
},
|
63934 |
-
{
|
63935 |
-
"epoch": 97.43,
|
63936 |
-
"learning_rate": 1.3653846153846154e-05,
|
63937 |
-
"loss": 0.1042,
|
63938 |
-
"step": 10620
|
63939 |
-
},
|
63940 |
-
{
|
63941 |
-
"epoch": 97.44,
|
63942 |
-
"learning_rate": 1.360576923076923e-05,
|
63943 |
-
"loss": 0.1809,
|
63944 |
-
"step": 10621
|
63945 |
-
},
|
63946 |
-
{
|
63947 |
-
"epoch": 97.45,
|
63948 |
-
"learning_rate": 1.3557692307692308e-05,
|
63949 |
-
"loss": 0.163,
|
63950 |
-
"step": 10622
|
63951 |
-
},
|
63952 |
-
{
|
63953 |
-
"epoch": 97.46,
|
63954 |
-
"learning_rate": 1.3509615384615384e-05,
|
63955 |
-
"loss": 0.0389,
|
63956 |
-
"step": 10623
|
63957 |
-
},
|
63958 |
-
{
|
63959 |
-
"epoch": 97.47,
|
63960 |
-
"learning_rate": 1.3461538461538463e-05,
|
63961 |
-
"loss": 0.1089,
|
63962 |
-
"step": 10624
|
63963 |
-
},
|
63964 |
-
{
|
63965 |
-
"epoch": 97.48,
|
63966 |
-
"learning_rate": 1.341346153846154e-05,
|
63967 |
-
"loss": 0.0626,
|
63968 |
-
"step": 10625
|
63969 |
-
},
|
63970 |
-
{
|
63971 |
-
"epoch": 97.49,
|
63972 |
-
"learning_rate": 1.3365384615384615e-05,
|
63973 |
-
"loss": 0.0596,
|
63974 |
-
"step": 10626
|
63975 |
-
},
|
63976 |
-
{
|
63977 |
-
"epoch": 97.5,
|
63978 |
-
"learning_rate": 1.3317307692307693e-05,
|
63979 |
-
"loss": 0.0524,
|
63980 |
-
"step": 10627
|
63981 |
-
},
|
63982 |
-
{
|
63983 |
-
"epoch": 97.5,
|
63984 |
-
"learning_rate": 1.3269230769230769e-05,
|
63985 |
-
"loss": 0.1355,
|
63986 |
-
"step": 10628
|
63987 |
-
},
|
63988 |
-
{
|
63989 |
-
"epoch": 97.51,
|
63990 |
-
"learning_rate": 1.3221153846153847e-05,
|
63991 |
-
"loss": 0.1323,
|
63992 |
-
"step": 10629
|
63993 |
-
},
|
63994 |
-
{
|
63995 |
-
"epoch": 97.52,
|
63996 |
-
"learning_rate": 1.3173076923076923e-05,
|
63997 |
-
"loss": 0.1384,
|
63998 |
-
"step": 10630
|
63999 |
-
},
|
64000 |
-
{
|
64001 |
-
"epoch": 97.53,
|
64002 |
-
"learning_rate": 1.3125e-05,
|
64003 |
-
"loss": 0.0982,
|
64004 |
-
"step": 10631
|
64005 |
-
},
|
64006 |
-
{
|
64007 |
-
"epoch": 97.54,
|
64008 |
-
"learning_rate": 1.3076923076923077e-05,
|
64009 |
-
"loss": 0.0849,
|
64010 |
-
"step": 10632
|
64011 |
-
},
|
64012 |
-
{
|
64013 |
-
"epoch": 97.55,
|
64014 |
-
"learning_rate": 1.3028846153846153e-05,
|
64015 |
-
"loss": 0.0658,
|
64016 |
-
"step": 10633
|
64017 |
-
},
|
64018 |
-
{
|
64019 |
-
"epoch": 97.56,
|
64020 |
-
"learning_rate": 1.2980769230769232e-05,
|
64021 |
-
"loss": 0.0823,
|
64022 |
-
"step": 10634
|
64023 |
-
},
|
64024 |
-
{
|
64025 |
-
"epoch": 97.57,
|
64026 |
-
"learning_rate": 1.2932692307692308e-05,
|
64027 |
-
"loss": 0.1178,
|
64028 |
-
"step": 10635
|
64029 |
-
},
|
64030 |
-
{
|
64031 |
-
"epoch": 97.58,
|
64032 |
-
"learning_rate": 1.2884615384615386e-05,
|
64033 |
-
"loss": 0.0725,
|
64034 |
-
"step": 10636
|
64035 |
-
},
|
64036 |
-
{
|
64037 |
-
"epoch": 97.59,
|
64038 |
-
"learning_rate": 1.2836538461538462e-05,
|
64039 |
-
"loss": 0.0602,
|
64040 |
-
"step": 10637
|
64041 |
-
},
|
64042 |
-
{
|
64043 |
-
"epoch": 97.6,
|
64044 |
-
"learning_rate": 1.278846153846154e-05,
|
64045 |
-
"loss": 0.0946,
|
64046 |
-
"step": 10638
|
64047 |
-
},
|
64048 |
-
{
|
64049 |
-
"epoch": 97.61,
|
64050 |
-
"learning_rate": 1.2740384615384615e-05,
|
64051 |
-
"loss": 0.0841,
|
64052 |
-
"step": 10639
|
64053 |
-
},
|
64054 |
-
{
|
64055 |
-
"epoch": 97.61,
|
64056 |
-
"learning_rate": 1.2692307692307691e-05,
|
64057 |
-
"loss": 0.0951,
|
64058 |
-
"step": 10640
|
64059 |
-
},
|
64060 |
-
{
|
64061 |
-
"epoch": 97.62,
|
64062 |
-
"learning_rate": 1.264423076923077e-05,
|
64063 |
-
"loss": 0.121,
|
64064 |
-
"step": 10641
|
64065 |
-
},
|
64066 |
-
{
|
64067 |
-
"epoch": 97.63,
|
64068 |
-
"learning_rate": 1.2596153846153845e-05,
|
64069 |
-
"loss": 0.0961,
|
64070 |
-
"step": 10642
|
64071 |
-
},
|
64072 |
-
{
|
64073 |
-
"epoch": 97.64,
|
64074 |
-
"learning_rate": 1.2548076923076923e-05,
|
64075 |
-
"loss": 0.0875,
|
64076 |
-
"step": 10643
|
64077 |
-
},
|
64078 |
-
{
|
64079 |
-
"epoch": 97.65,
|
64080 |
-
"learning_rate": 1.25e-05,
|
64081 |
-
"loss": 0.0905,
|
64082 |
-
"step": 10644
|
64083 |
-
},
|
64084 |
-
{
|
64085 |
-
"epoch": 97.66,
|
64086 |
-
"learning_rate": 1.2451923076923078e-05,
|
64087 |
-
"loss": 0.087,
|
64088 |
-
"step": 10645
|
64089 |
-
},
|
64090 |
-
{
|
64091 |
-
"epoch": 97.67,
|
64092 |
-
"learning_rate": 1.2403846153846154e-05,
|
64093 |
-
"loss": 0.0773,
|
64094 |
-
"step": 10646
|
64095 |
-
},
|
64096 |
-
{
|
64097 |
-
"epoch": 97.68,
|
64098 |
-
"learning_rate": 1.2355769230769232e-05,
|
64099 |
-
"loss": 0.0643,
|
64100 |
-
"step": 10647
|
64101 |
-
},
|
64102 |
-
{
|
64103 |
-
"epoch": 97.69,
|
64104 |
-
"learning_rate": 1.2307692307692308e-05,
|
64105 |
-
"loss": 0.0863,
|
64106 |
-
"step": 10648
|
64107 |
-
},
|
64108 |
-
{
|
64109 |
-
"epoch": 97.7,
|
64110 |
-
"learning_rate": 1.2259615384615384e-05,
|
64111 |
-
"loss": 0.0755,
|
64112 |
-
"step": 10649
|
64113 |
-
},
|
64114 |
-
{
|
64115 |
-
"epoch": 97.71,
|
64116 |
-
"learning_rate": 1.2211538461538462e-05,
|
64117 |
-
"loss": 0.094,
|
64118 |
-
"step": 10650
|
64119 |
-
},
|
64120 |
-
{
|
64121 |
-
"epoch": 97.72,
|
64122 |
-
"learning_rate": 1.2163461538461538e-05,
|
64123 |
-
"loss": 0.0653,
|
64124 |
-
"step": 10651
|
64125 |
-
},
|
64126 |
-
{
|
64127 |
-
"epoch": 97.72,
|
64128 |
-
"learning_rate": 1.2115384615384615e-05,
|
64129 |
-
"loss": 0.0869,
|
64130 |
-
"step": 10652
|
64131 |
-
},
|
64132 |
-
{
|
64133 |
-
"epoch": 97.73,
|
64134 |
-
"learning_rate": 1.2067307692307693e-05,
|
64135 |
-
"loss": 0.0762,
|
64136 |
-
"step": 10653
|
64137 |
-
},
|
64138 |
-
{
|
64139 |
-
"epoch": 97.74,
|
64140 |
-
"learning_rate": 1.2019230769230771e-05,
|
64141 |
-
"loss": 0.0212,
|
64142 |
-
"step": 10654
|
64143 |
-
},
|
64144 |
-
{
|
64145 |
-
"epoch": 97.75,
|
64146 |
-
"learning_rate": 1.1971153846153847e-05,
|
64147 |
-
"loss": 0.1531,
|
64148 |
-
"step": 10655
|
64149 |
-
},
|
64150 |
-
{
|
64151 |
-
"epoch": 97.76,
|
64152 |
-
"learning_rate": 1.1923076923076923e-05,
|
64153 |
-
"loss": 0.1183,
|
64154 |
-
"step": 10656
|
64155 |
-
},
|
64156 |
-
{
|
64157 |
-
"epoch": 97.77,
|
64158 |
-
"learning_rate": 1.1875e-05,
|
64159 |
-
"loss": 0.0916,
|
64160 |
-
"step": 10657
|
64161 |
-
},
|
64162 |
-
{
|
64163 |
-
"epoch": 97.78,
|
64164 |
-
"learning_rate": 1.1826923076923077e-05,
|
64165 |
-
"loss": 0.086,
|
64166 |
-
"step": 10658
|
64167 |
-
},
|
64168 |
-
{
|
64169 |
-
"epoch": 97.79,
|
64170 |
-
"learning_rate": 1.1778846153846154e-05,
|
64171 |
-
"loss": 0.1233,
|
64172 |
-
"step": 10659
|
64173 |
-
},
|
64174 |
-
{
|
64175 |
-
"epoch": 97.8,
|
64176 |
-
"learning_rate": 1.173076923076923e-05,
|
64177 |
-
"loss": 0.0986,
|
64178 |
-
"step": 10660
|
64179 |
-
},
|
64180 |
-
{
|
64181 |
-
"epoch": 97.81,
|
64182 |
-
"learning_rate": 1.1682692307692308e-05,
|
64183 |
-
"loss": 0.1258,
|
64184 |
-
"step": 10661
|
64185 |
-
},
|
64186 |
-
{
|
64187 |
-
"epoch": 97.82,
|
64188 |
-
"learning_rate": 1.1634615384615384e-05,
|
64189 |
-
"loss": 0.1512,
|
64190 |
-
"step": 10662
|
64191 |
-
},
|
64192 |
-
{
|
64193 |
-
"epoch": 97.83,
|
64194 |
-
"learning_rate": 1.1586538461538462e-05,
|
64195 |
-
"loss": 0.0741,
|
64196 |
-
"step": 10663
|
64197 |
-
},
|
64198 |
-
{
|
64199 |
-
"epoch": 97.83,
|
64200 |
-
"learning_rate": 1.153846153846154e-05,
|
64201 |
-
"loss": 0.1224,
|
64202 |
-
"step": 10664
|
64203 |
-
},
|
64204 |
-
{
|
64205 |
-
"epoch": 97.84,
|
64206 |
-
"learning_rate": 1.1490384615384616e-05,
|
64207 |
-
"loss": 0.1029,
|
64208 |
-
"step": 10665
|
64209 |
-
},
|
64210 |
-
{
|
64211 |
-
"epoch": 97.85,
|
64212 |
-
"learning_rate": 1.1442307692307693e-05,
|
64213 |
-
"loss": 0.1199,
|
64214 |
-
"step": 10666
|
64215 |
-
},
|
64216 |
-
{
|
64217 |
-
"epoch": 97.86,
|
64218 |
-
"learning_rate": 1.139423076923077e-05,
|
64219 |
-
"loss": 0.0765,
|
64220 |
-
"step": 10667
|
64221 |
-
},
|
64222 |
-
{
|
64223 |
-
"epoch": 97.87,
|
64224 |
-
"learning_rate": 1.1346153846153847e-05,
|
64225 |
-
"loss": 0.1132,
|
64226 |
-
"step": 10668
|
64227 |
-
},
|
64228 |
-
{
|
64229 |
-
"epoch": 97.88,
|
64230 |
-
"learning_rate": 1.1298076923076923e-05,
|
64231 |
-
"loss": 0.1571,
|
64232 |
-
"step": 10669
|
64233 |
-
},
|
64234 |
-
{
|
64235 |
-
"epoch": 97.89,
|
64236 |
-
"learning_rate": 1.1249999999999999e-05,
|
64237 |
-
"loss": 0.0853,
|
64238 |
-
"step": 10670
|
64239 |
-
},
|
64240 |
-
{
|
64241 |
-
"epoch": 97.9,
|
64242 |
-
"learning_rate": 1.1201923076923077e-05,
|
64243 |
-
"loss": 0.1177,
|
64244 |
-
"step": 10671
|
64245 |
-
},
|
64246 |
-
{
|
64247 |
-
"epoch": 97.91,
|
64248 |
-
"learning_rate": 1.1153846153846153e-05,
|
64249 |
-
"loss": 0.1029,
|
64250 |
-
"step": 10672
|
64251 |
-
},
|
64252 |
-
{
|
64253 |
-
"epoch": 97.92,
|
64254 |
-
"learning_rate": 1.1105769230769232e-05,
|
64255 |
-
"loss": 0.0737,
|
64256 |
-
"step": 10673
|
64257 |
-
},
|
64258 |
-
{
|
64259 |
-
"epoch": 97.93,
|
64260 |
-
"learning_rate": 1.1057692307692308e-05,
|
64261 |
-
"loss": 0.071,
|
64262 |
-
"step": 10674
|
64263 |
-
},
|
64264 |
-
{
|
64265 |
-
"epoch": 97.94,
|
64266 |
-
"learning_rate": 1.1009615384615386e-05,
|
64267 |
-
"loss": 0.0793,
|
64268 |
-
"step": 10675
|
64269 |
-
},
|
64270 |
-
{
|
64271 |
-
"epoch": 97.94,
|
64272 |
-
"learning_rate": 1.0961538461538462e-05,
|
64273 |
-
"loss": 0.0541,
|
64274 |
-
"step": 10676
|
64275 |
-
},
|
64276 |
-
{
|
64277 |
-
"epoch": 97.95,
|
64278 |
-
"learning_rate": 1.091346153846154e-05,
|
64279 |
-
"loss": 0.0715,
|
64280 |
-
"step": 10677
|
64281 |
-
},
|
64282 |
-
{
|
64283 |
-
"epoch": 97.96,
|
64284 |
-
"learning_rate": 1.0865384615384616e-05,
|
64285 |
-
"loss": 0.0896,
|
64286 |
-
"step": 10678
|
64287 |
-
},
|
64288 |
-
{
|
64289 |
-
"epoch": 97.97,
|
64290 |
-
"learning_rate": 1.0817307692307692e-05,
|
64291 |
-
"loss": 0.1121,
|
64292 |
-
"step": 10679
|
64293 |
-
},
|
64294 |
-
{
|
64295 |
-
"epoch": 97.98,
|
64296 |
-
"learning_rate": 1.076923076923077e-05,
|
64297 |
-
"loss": 0.0806,
|
64298 |
-
"step": 10680
|
64299 |
-
},
|
64300 |
-
{
|
64301 |
-
"epoch": 97.99,
|
64302 |
-
"learning_rate": 1.0721153846153845e-05,
|
64303 |
-
"loss": 0.039,
|
64304 |
-
"step": 10681
|
64305 |
-
},
|
64306 |
-
{
|
64307 |
-
"epoch": 98.0,
|
64308 |
-
"learning_rate": 1.0673076923076923e-05,
|
64309 |
-
"loss": 0.0613,
|
64310 |
-
"step": 10682
|
64311 |
-
},
|
64312 |
-
{
|
64313 |
-
"epoch": 98.01,
|
64314 |
-
"learning_rate": 1.0625e-05,
|
64315 |
-
"loss": 0.1759,
|
64316 |
-
"step": 10683
|
64317 |
-
},
|
64318 |
-
{
|
64319 |
-
"epoch": 98.02,
|
64320 |
-
"learning_rate": 1.0576923076923078e-05,
|
64321 |
-
"loss": 0.1243,
|
64322 |
-
"step": 10684
|
64323 |
-
},
|
64324 |
-
{
|
64325 |
-
"epoch": 98.03,
|
64326 |
-
"learning_rate": 1.0528846153846154e-05,
|
64327 |
-
"loss": 0.1195,
|
64328 |
-
"step": 10685
|
64329 |
-
},
|
64330 |
-
{
|
64331 |
-
"epoch": 98.04,
|
64332 |
-
"learning_rate": 1.048076923076923e-05,
|
64333 |
-
"loss": 0.0845,
|
64334 |
-
"step": 10686
|
64335 |
-
},
|
64336 |
-
{
|
64337 |
-
"epoch": 98.05,
|
64338 |
-
"learning_rate": 1.0432692307692308e-05,
|
64339 |
-
"loss": 0.1433,
|
64340 |
-
"step": 10687
|
64341 |
-
},
|
64342 |
-
{
|
64343 |
-
"epoch": 98.06,
|
64344 |
-
"learning_rate": 1.0384615384615384e-05,
|
64345 |
-
"loss": 0.1045,
|
64346 |
-
"step": 10688
|
64347 |
-
},
|
64348 |
-
{
|
64349 |
-
"epoch": 98.06,
|
64350 |
-
"learning_rate": 1.0336538461538462e-05,
|
64351 |
-
"loss": 0.1342,
|
64352 |
-
"step": 10689
|
64353 |
-
},
|
64354 |
-
{
|
64355 |
-
"epoch": 98.07,
|
64356 |
-
"learning_rate": 1.0288461538461538e-05,
|
64357 |
-
"loss": 0.1045,
|
64358 |
-
"step": 10690
|
64359 |
-
},
|
64360 |
-
{
|
64361 |
-
"epoch": 98.08,
|
64362 |
-
"learning_rate": 1.0240384615384616e-05,
|
64363 |
-
"loss": 0.1304,
|
64364 |
-
"step": 10691
|
64365 |
-
},
|
64366 |
-
{
|
64367 |
-
"epoch": 98.09,
|
64368 |
-
"learning_rate": 1.0192307692307692e-05,
|
64369 |
-
"loss": 0.1098,
|
64370 |
-
"step": 10692
|
64371 |
-
},
|
64372 |
-
{
|
64373 |
-
"epoch": 98.1,
|
64374 |
-
"learning_rate": 1.014423076923077e-05,
|
64375 |
-
"loss": 0.1077,
|
64376 |
-
"step": 10693
|
64377 |
-
},
|
64378 |
-
{
|
64379 |
-
"epoch": 98.11,
|
64380 |
-
"learning_rate": 1.0096153846153847e-05,
|
64381 |
-
"loss": 0.1068,
|
64382 |
-
"step": 10694
|
64383 |
-
},
|
64384 |
-
{
|
64385 |
-
"epoch": 98.12,
|
64386 |
-
"learning_rate": 1.0048076923076923e-05,
|
64387 |
-
"loss": 0.1021,
|
64388 |
-
"step": 10695
|
64389 |
-
},
|
64390 |
-
{
|
64391 |
-
"epoch": 98.13,
|
64392 |
-
"learning_rate": 1e-05,
|
64393 |
-
"loss": 0.0923,
|
64394 |
-
"step": 10696
|
64395 |
-
},
|
64396 |
-
{
|
64397 |
-
"epoch": 98.14,
|
64398 |
-
"learning_rate": 9.951923076923077e-06,
|
64399 |
-
"loss": 0.1061,
|
64400 |
-
"step": 10697
|
64401 |
-
},
|
64402 |
-
{
|
64403 |
-
"epoch": 98.15,
|
64404 |
-
"learning_rate": 9.903846153846155e-06,
|
64405 |
-
"loss": 0.1195,
|
64406 |
-
"step": 10698
|
64407 |
-
},
|
64408 |
-
{
|
64409 |
-
"epoch": 98.16,
|
64410 |
-
"learning_rate": 9.85576923076923e-06,
|
64411 |
-
"loss": 0.0543,
|
64412 |
-
"step": 10699
|
64413 |
-
},
|
64414 |
-
{
|
64415 |
-
"epoch": 98.17,
|
64416 |
-
"learning_rate": 9.807692307692307e-06,
|
64417 |
-
"loss": 0.1391,
|
64418 |
-
"step": 10700
|
64419 |
-
},
|
64420 |
-
{
|
64421 |
-
"epoch": 98.17,
|
64422 |
-
"learning_rate": 9.759615384615384e-06,
|
64423 |
-
"loss": 0.0912,
|
64424 |
-
"step": 10701
|
64425 |
-
},
|
64426 |
-
{
|
64427 |
-
"epoch": 98.18,
|
64428 |
-
"learning_rate": 9.71153846153846e-06,
|
64429 |
-
"loss": 0.0989,
|
64430 |
-
"step": 10702
|
64431 |
-
},
|
64432 |
-
{
|
64433 |
-
"epoch": 98.19,
|
64434 |
-
"learning_rate": 9.66346153846154e-06,
|
64435 |
-
"loss": 0.1448,
|
64436 |
-
"step": 10703
|
64437 |
-
},
|
64438 |
-
{
|
64439 |
-
"epoch": 98.2,
|
64440 |
-
"learning_rate": 9.615384615384616e-06,
|
64441 |
-
"loss": 0.0309,
|
64442 |
-
"step": 10704
|
64443 |
-
},
|
64444 |
-
{
|
64445 |
-
"epoch": 98.21,
|
64446 |
-
"learning_rate": 9.567307692307693e-06,
|
64447 |
-
"loss": 0.0515,
|
64448 |
-
"step": 10705
|
64449 |
-
},
|
64450 |
-
{
|
64451 |
-
"epoch": 98.22,
|
64452 |
-
"learning_rate": 9.51923076923077e-06,
|
64453 |
-
"loss": 0.0897,
|
64454 |
-
"step": 10706
|
64455 |
-
},
|
64456 |
-
{
|
64457 |
-
"epoch": 98.23,
|
64458 |
-
"learning_rate": 9.471153846153847e-06,
|
64459 |
-
"loss": 0.0825,
|
64460 |
-
"step": 10707
|
64461 |
-
},
|
64462 |
-
{
|
64463 |
-
"epoch": 98.24,
|
64464 |
-
"learning_rate": 9.423076923076923e-06,
|
64465 |
-
"loss": 0.07,
|
64466 |
-
"step": 10708
|
64467 |
-
},
|
64468 |
-
{
|
64469 |
-
"epoch": 98.25,
|
64470 |
-
"learning_rate": 9.375e-06,
|
64471 |
-
"loss": 0.0725,
|
64472 |
-
"step": 10709
|
64473 |
-
},
|
64474 |
-
{
|
64475 |
-
"epoch": 98.26,
|
64476 |
-
"learning_rate": 9.326923076923077e-06,
|
64477 |
-
"loss": 0.1557,
|
64478 |
-
"step": 10710
|
64479 |
-
},
|
64480 |
-
{
|
64481 |
-
"epoch": 98.27,
|
64482 |
-
"learning_rate": 9.278846153846153e-06,
|
64483 |
-
"loss": 0.1375,
|
64484 |
-
"step": 10711
|
64485 |
-
},
|
64486 |
-
{
|
64487 |
-
"epoch": 98.28,
|
64488 |
-
"learning_rate": 9.230769230769232e-06,
|
64489 |
-
"loss": 0.1014,
|
64490 |
-
"step": 10712
|
64491 |
-
},
|
64492 |
-
{
|
64493 |
-
"epoch": 98.28,
|
64494 |
-
"learning_rate": 9.182692307692308e-06,
|
64495 |
-
"loss": 0.1071,
|
64496 |
-
"step": 10713
|
64497 |
-
},
|
64498 |
-
{
|
64499 |
-
"epoch": 98.29,
|
64500 |
-
"learning_rate": 9.134615384615386e-06,
|
64501 |
-
"loss": 0.1428,
|
64502 |
-
"step": 10714
|
64503 |
-
},
|
64504 |
-
{
|
64505 |
-
"epoch": 98.3,
|
64506 |
-
"learning_rate": 9.086538461538462e-06,
|
64507 |
-
"loss": 0.1182,
|
64508 |
-
"step": 10715
|
64509 |
-
},
|
64510 |
-
{
|
64511 |
-
"epoch": 98.31,
|
64512 |
-
"learning_rate": 9.038461538461538e-06,
|
64513 |
-
"loss": 0.1034,
|
64514 |
-
"step": 10716
|
64515 |
-
},
|
64516 |
-
{
|
64517 |
-
"epoch": 98.32,
|
64518 |
-
"learning_rate": 8.990384615384616e-06,
|
64519 |
-
"loss": 0.1138,
|
64520 |
-
"step": 10717
|
64521 |
-
},
|
64522 |
-
{
|
64523 |
-
"epoch": 98.33,
|
64524 |
-
"learning_rate": 8.942307692307692e-06,
|
64525 |
-
"loss": 0.1355,
|
64526 |
-
"step": 10718
|
64527 |
-
},
|
64528 |
-
{
|
64529 |
-
"epoch": 98.34,
|
64530 |
-
"learning_rate": 8.89423076923077e-06,
|
64531 |
-
"loss": 0.0971,
|
64532 |
-
"step": 10719
|
64533 |
-
},
|
64534 |
-
{
|
64535 |
-
"epoch": 98.35,
|
64536 |
-
"learning_rate": 8.846153846153846e-06,
|
64537 |
-
"loss": 0.1325,
|
64538 |
-
"step": 10720
|
64539 |
-
},
|
64540 |
-
{
|
64541 |
-
"epoch": 98.36,
|
64542 |
-
"learning_rate": 8.798076923076923e-06,
|
64543 |
-
"loss": 0.1429,
|
64544 |
-
"step": 10721
|
64545 |
-
},
|
64546 |
-
{
|
64547 |
-
"epoch": 98.37,
|
64548 |
-
"learning_rate": 8.750000000000001e-06,
|
64549 |
-
"loss": 0.1403,
|
64550 |
-
"step": 10722
|
64551 |
-
},
|
64552 |
-
{
|
64553 |
-
"epoch": 98.38,
|
64554 |
-
"learning_rate": 8.701923076923077e-06,
|
64555 |
-
"loss": 0.1079,
|
64556 |
-
"step": 10723
|
64557 |
-
},
|
64558 |
-
{
|
64559 |
-
"epoch": 98.39,
|
64560 |
-
"learning_rate": 8.653846153846155e-06,
|
64561 |
-
"loss": 0.1083,
|
64562 |
-
"step": 10724
|
64563 |
-
},
|
64564 |
-
{
|
64565 |
-
"epoch": 98.39,
|
64566 |
-
"learning_rate": 8.60576923076923e-06,
|
64567 |
-
"loss": 0.1192,
|
64568 |
-
"step": 10725
|
64569 |
-
},
|
64570 |
-
{
|
64571 |
-
"epoch": 98.4,
|
64572 |
-
"learning_rate": 8.557692307692308e-06,
|
64573 |
-
"loss": 0.1056,
|
64574 |
-
"step": 10726
|
64575 |
-
},
|
64576 |
-
{
|
64577 |
-
"epoch": 98.41,
|
64578 |
-
"learning_rate": 8.509615384615384e-06,
|
64579 |
-
"loss": 0.1136,
|
64580 |
-
"step": 10727
|
64581 |
-
},
|
64582 |
-
{
|
64583 |
-
"epoch": 98.42,
|
64584 |
-
"learning_rate": 8.461538461538462e-06,
|
64585 |
-
"loss": 0.088,
|
64586 |
-
"step": 10728
|
64587 |
-
},
|
64588 |
-
{
|
64589 |
-
"epoch": 98.43,
|
64590 |
-
"learning_rate": 8.413461538461538e-06,
|
64591 |
-
"loss": 0.0983,
|
64592 |
-
"step": 10729
|
64593 |
-
},
|
64594 |
-
{
|
64595 |
-
"epoch": 98.44,
|
64596 |
-
"learning_rate": 8.365384615384616e-06,
|
64597 |
-
"loss": 0.1053,
|
64598 |
-
"step": 10730
|
64599 |
-
},
|
64600 |
-
{
|
64601 |
-
"epoch": 98.45,
|
64602 |
-
"learning_rate": 8.317307692307692e-06,
|
64603 |
-
"loss": 0.0869,
|
64604 |
-
"step": 10731
|
64605 |
-
},
|
64606 |
-
{
|
64607 |
-
"epoch": 98.46,
|
64608 |
-
"learning_rate": 8.26923076923077e-06,
|
64609 |
-
"loss": 0.071,
|
64610 |
-
"step": 10732
|
64611 |
-
},
|
64612 |
-
{
|
64613 |
-
"epoch": 98.47,
|
64614 |
-
"learning_rate": 8.221153846153847e-06,
|
64615 |
-
"loss": 0.1736,
|
64616 |
-
"step": 10733
|
64617 |
-
},
|
64618 |
-
{
|
64619 |
-
"epoch": 98.48,
|
64620 |
-
"learning_rate": 8.173076923076923e-06,
|
64621 |
-
"loss": 0.0977,
|
64622 |
-
"step": 10734
|
64623 |
-
},
|
64624 |
-
{
|
64625 |
-
"epoch": 98.49,
|
64626 |
-
"learning_rate": 8.125000000000001e-06,
|
64627 |
-
"loss": 0.0914,
|
64628 |
-
"step": 10735
|
64629 |
-
},
|
64630 |
-
{
|
64631 |
-
"epoch": 98.5,
|
64632 |
-
"learning_rate": 8.076923076923077e-06,
|
64633 |
-
"loss": 0.0306,
|
64634 |
-
"step": 10736
|
64635 |
-
},
|
64636 |
-
{
|
64637 |
-
"epoch": 98.5,
|
64638 |
-
"learning_rate": 8.028846153846155e-06,
|
64639 |
-
"loss": 0.1421,
|
64640 |
-
"step": 10737
|
64641 |
-
},
|
64642 |
-
{
|
64643 |
-
"epoch": 98.51,
|
64644 |
-
"learning_rate": 7.98076923076923e-06,
|
64645 |
-
"loss": 0.1401,
|
64646 |
-
"step": 10738
|
64647 |
-
},
|
64648 |
-
{
|
64649 |
-
"epoch": 98.52,
|
64650 |
-
"learning_rate": 7.932692307692307e-06,
|
64651 |
-
"loss": 0.1104,
|
64652 |
-
"step": 10739
|
64653 |
-
},
|
64654 |
-
{
|
64655 |
-
"epoch": 98.53,
|
64656 |
-
"learning_rate": 7.884615384615384e-06,
|
64657 |
-
"loss": 0.1197,
|
64658 |
-
"step": 10740
|
64659 |
-
},
|
64660 |
-
{
|
64661 |
-
"epoch": 98.54,
|
64662 |
-
"learning_rate": 7.83653846153846e-06,
|
64663 |
-
"loss": 0.133,
|
64664 |
-
"step": 10741
|
64665 |
-
},
|
64666 |
-
{
|
64667 |
-
"epoch": 98.55,
|
64668 |
-
"learning_rate": 7.788461538461538e-06,
|
64669 |
-
"loss": 0.1172,
|
64670 |
-
"step": 10742
|
64671 |
-
},
|
64672 |
-
{
|
64673 |
-
"epoch": 98.56,
|
64674 |
-
"learning_rate": 7.740384615384616e-06,
|
64675 |
-
"loss": 0.1163,
|
64676 |
-
"step": 10743
|
64677 |
-
},
|
64678 |
-
{
|
64679 |
-
"epoch": 98.57,
|
64680 |
-
"learning_rate": 7.692307692307694e-06,
|
64681 |
-
"loss": 0.1027,
|
64682 |
-
"step": 10744
|
64683 |
-
},
|
64684 |
-
{
|
64685 |
-
"epoch": 98.58,
|
64686 |
-
"learning_rate": 7.64423076923077e-06,
|
64687 |
-
"loss": 0.1016,
|
64688 |
-
"step": 10745
|
64689 |
-
},
|
64690 |
-
{
|
64691 |
-
"epoch": 98.59,
|
64692 |
-
"learning_rate": 7.5961538461538465e-06,
|
64693 |
-
"loss": 0.123,
|
64694 |
-
"step": 10746
|
64695 |
-
},
|
64696 |
-
{
|
64697 |
-
"epoch": 98.6,
|
64698 |
-
"learning_rate": 7.548076923076923e-06,
|
64699 |
-
"loss": 0.135,
|
64700 |
-
"step": 10747
|
64701 |
-
},
|
64702 |
-
{
|
64703 |
-
"epoch": 98.61,
|
64704 |
-
"learning_rate": 7.5e-06,
|
64705 |
-
"loss": 0.1509,
|
64706 |
-
"step": 10748
|
64707 |
-
},
|
64708 |
-
{
|
64709 |
-
"epoch": 98.61,
|
64710 |
-
"learning_rate": 7.451923076923078e-06,
|
64711 |
-
"loss": 0.0623,
|
64712 |
-
"step": 10749
|
64713 |
-
},
|
64714 |
-
{
|
64715 |
-
"epoch": 98.62,
|
64716 |
-
"learning_rate": 7.403846153846154e-06,
|
64717 |
-
"loss": 0.0884,
|
64718 |
-
"step": 10750
|
64719 |
-
},
|
64720 |
-
{
|
64721 |
-
"epoch": 98.63,
|
64722 |
-
"learning_rate": 7.355769230769231e-06,
|
64723 |
-
"loss": 0.1031,
|
64724 |
-
"step": 10751
|
64725 |
-
},
|
64726 |
-
{
|
64727 |
-
"epoch": 98.64,
|
64728 |
-
"learning_rate": 7.307692307692308e-06,
|
64729 |
-
"loss": 0.0448,
|
64730 |
-
"step": 10752
|
64731 |
-
},
|
64732 |
-
{
|
64733 |
-
"epoch": 98.65,
|
64734 |
-
"learning_rate": 7.2596153846153845e-06,
|
64735 |
-
"loss": 0.0493,
|
64736 |
-
"step": 10753
|
64737 |
-
},
|
64738 |
-
{
|
64739 |
-
"epoch": 98.66,
|
64740 |
-
"learning_rate": 7.211538461538462e-06,
|
64741 |
-
"loss": 0.0913,
|
64742 |
-
"step": 10754
|
64743 |
-
},
|
64744 |
-
{
|
64745 |
-
"epoch": 98.67,
|
64746 |
-
"learning_rate": 7.163461538461539e-06,
|
64747 |
-
"loss": 0.0801,
|
64748 |
-
"step": 10755
|
64749 |
-
},
|
64750 |
-
{
|
64751 |
-
"epoch": 98.68,
|
64752 |
-
"learning_rate": 7.115384615384616e-06,
|
64753 |
-
"loss": 0.1115,
|
64754 |
-
"step": 10756
|
64755 |
-
},
|
64756 |
-
{
|
64757 |
-
"epoch": 98.69,
|
64758 |
-
"learning_rate": 7.067307692307692e-06,
|
64759 |
-
"loss": 0.0649,
|
64760 |
-
"step": 10757
|
64761 |
-
},
|
64762 |
-
{
|
64763 |
-
"epoch": 98.7,
|
64764 |
-
"learning_rate": 7.019230769230769e-06,
|
64765 |
-
"loss": 0.1154,
|
64766 |
-
"step": 10758
|
64767 |
-
},
|
64768 |
-
{
|
64769 |
-
"epoch": 98.71,
|
64770 |
-
"learning_rate": 6.9711538461538465e-06,
|
64771 |
-
"loss": 0.0921,
|
64772 |
-
"step": 10759
|
64773 |
-
},
|
64774 |
-
{
|
64775 |
-
"epoch": 98.72,
|
64776 |
-
"learning_rate": 6.923076923076923e-06,
|
64777 |
-
"loss": 0.1203,
|
64778 |
-
"step": 10760
|
64779 |
-
},
|
64780 |
-
{
|
64781 |
-
"epoch": 98.72,
|
64782 |
-
"learning_rate": 6.875e-06,
|
64783 |
-
"loss": 0.07,
|
64784 |
-
"step": 10761
|
64785 |
-
},
|
64786 |
-
{
|
64787 |
-
"epoch": 98.73,
|
64788 |
-
"learning_rate": 6.826923076923077e-06,
|
64789 |
-
"loss": 0.0516,
|
64790 |
-
"step": 10762
|
64791 |
-
},
|
64792 |
-
{
|
64793 |
-
"epoch": 98.74,
|
64794 |
-
"learning_rate": 6.778846153846154e-06,
|
64795 |
-
"loss": 0.0659,
|
64796 |
-
"step": 10763
|
64797 |
-
},
|
64798 |
-
{
|
64799 |
-
"epoch": 98.75,
|
64800 |
-
"learning_rate": 6.730769230769232e-06,
|
64801 |
-
"loss": 0.1547,
|
64802 |
-
"step": 10764
|
64803 |
-
},
|
64804 |
-
{
|
64805 |
-
"epoch": 98.76,
|
64806 |
-
"learning_rate": 6.682692307692308e-06,
|
64807 |
-
"loss": 0.0945,
|
64808 |
-
"step": 10765
|
64809 |
-
},
|
64810 |
-
{
|
64811 |
-
"epoch": 98.77,
|
64812 |
-
"learning_rate": 6.6346153846153846e-06,
|
64813 |
-
"loss": 0.1401,
|
64814 |
-
"step": 10766
|
64815 |
-
},
|
64816 |
-
{
|
64817 |
-
"epoch": 98.78,
|
64818 |
-
"learning_rate": 6.586538461538461e-06,
|
64819 |
-
"loss": 0.1126,
|
64820 |
-
"step": 10767
|
64821 |
-
},
|
64822 |
-
{
|
64823 |
-
"epoch": 98.79,
|
64824 |
-
"learning_rate": 6.538461538461538e-06,
|
64825 |
-
"loss": 0.1139,
|
64826 |
-
"step": 10768
|
64827 |
-
},
|
64828 |
-
{
|
64829 |
-
"epoch": 98.8,
|
64830 |
-
"learning_rate": 6.490384615384616e-06,
|
64831 |
-
"loss": 0.1408,
|
64832 |
-
"step": 10769
|
64833 |
-
},
|
64834 |
-
{
|
64835 |
-
"epoch": 98.81,
|
64836 |
-
"learning_rate": 6.442307692307693e-06,
|
64837 |
-
"loss": 0.112,
|
64838 |
-
"step": 10770
|
64839 |
-
},
|
64840 |
-
{
|
64841 |
-
"epoch": 98.82,
|
64842 |
-
"learning_rate": 6.39423076923077e-06,
|
64843 |
-
"loss": 0.0731,
|
64844 |
-
"step": 10771
|
64845 |
-
},
|
64846 |
-
{
|
64847 |
-
"epoch": 98.83,
|
64848 |
-
"learning_rate": 6.346153846153846e-06,
|
64849 |
-
"loss": 0.1398,
|
64850 |
-
"step": 10772
|
64851 |
-
},
|
64852 |
-
{
|
64853 |
-
"epoch": 98.83,
|
64854 |
-
"learning_rate": 6.298076923076923e-06,
|
64855 |
-
"loss": 0.0956,
|
64856 |
-
"step": 10773
|
64857 |
-
},
|
64858 |
-
{
|
64859 |
-
"epoch": 98.84,
|
64860 |
-
"learning_rate": 6.25e-06,
|
64861 |
-
"loss": 0.1047,
|
64862 |
-
"step": 10774
|
64863 |
-
},
|
64864 |
-
{
|
64865 |
-
"epoch": 98.85,
|
64866 |
-
"learning_rate": 6.201923076923077e-06,
|
64867 |
-
"loss": 0.0632,
|
64868 |
-
"step": 10775
|
64869 |
-
},
|
64870 |
-
{
|
64871 |
-
"epoch": 98.86,
|
64872 |
-
"learning_rate": 6.153846153846154e-06,
|
64873 |
-
"loss": 0.0971,
|
64874 |
-
"step": 10776
|
64875 |
-
},
|
64876 |
-
{
|
64877 |
-
"epoch": 98.87,
|
64878 |
-
"learning_rate": 6.105769230769231e-06,
|
64879 |
-
"loss": 0.1108,
|
64880 |
-
"step": 10777
|
64881 |
-
},
|
64882 |
-
{
|
64883 |
-
"epoch": 98.88,
|
64884 |
-
"learning_rate": 6.057692307692308e-06,
|
64885 |
-
"loss": 0.154,
|
64886 |
-
"step": 10778
|
64887 |
-
},
|
64888 |
-
{
|
64889 |
-
"epoch": 98.89,
|
64890 |
-
"learning_rate": 6.0096153846153855e-06,
|
64891 |
-
"loss": 0.0705,
|
64892 |
-
"step": 10779
|
64893 |
-
},
|
64894 |
-
{
|
64895 |
-
"epoch": 98.9,
|
64896 |
-
"learning_rate": 5.9615384615384615e-06,
|
64897 |
-
"loss": 0.1322,
|
64898 |
-
"step": 10780
|
64899 |
-
},
|
64900 |
-
{
|
64901 |
-
"epoch": 98.91,
|
64902 |
-
"learning_rate": 5.913461538461538e-06,
|
64903 |
-
"loss": 0.0784,
|
64904 |
-
"step": 10781
|
64905 |
-
},
|
64906 |
-
{
|
64907 |
-
"epoch": 98.92,
|
64908 |
-
"learning_rate": 5.865384615384615e-06,
|
64909 |
-
"loss": 0.0775,
|
64910 |
-
"step": 10782
|
64911 |
-
},
|
64912 |
-
{
|
64913 |
-
"epoch": 98.93,
|
64914 |
-
"learning_rate": 5.817307692307692e-06,
|
64915 |
-
"loss": 0.1013,
|
64916 |
-
"step": 10783
|
64917 |
-
},
|
64918 |
-
{
|
64919 |
-
"epoch": 98.94,
|
64920 |
-
"learning_rate": 5.76923076923077e-06,
|
64921 |
-
"loss": 0.0877,
|
64922 |
-
"step": 10784
|
64923 |
-
},
|
64924 |
-
{
|
64925 |
-
"epoch": 98.94,
|
64926 |
-
"learning_rate": 5.721153846153847e-06,
|
64927 |
-
"loss": 0.1856,
|
64928 |
-
"step": 10785
|
64929 |
-
},
|
64930 |
-
{
|
64931 |
-
"epoch": 98.95,
|
64932 |
-
"learning_rate": 5.6730769230769235e-06,
|
64933 |
-
"loss": 0.0554,
|
64934 |
-
"step": 10786
|
64935 |
-
},
|
64936 |
-
{
|
64937 |
-
"epoch": 98.96,
|
64938 |
-
"learning_rate": 5.6249999999999995e-06,
|
64939 |
-
"loss": 0.0956,
|
64940 |
-
"step": 10787
|
64941 |
-
},
|
64942 |
-
{
|
64943 |
-
"epoch": 98.97,
|
64944 |
-
"learning_rate": 5.576923076923076e-06,
|
64945 |
-
"loss": 0.0672,
|
64946 |
-
"step": 10788
|
64947 |
-
},
|
64948 |
-
{
|
64949 |
-
"epoch": 98.98,
|
64950 |
-
"learning_rate": 5.528846153846154e-06,
|
64951 |
-
"loss": 0.1047,
|
64952 |
-
"step": 10789
|
64953 |
-
},
|
64954 |
-
{
|
64955 |
-
"epoch": 98.99,
|
64956 |
-
"learning_rate": 5.480769230769231e-06,
|
64957 |
-
"loss": 0.0257,
|
64958 |
-
"step": 10790
|
64959 |
-
},
|
64960 |
-
{
|
64961 |
-
"epoch": 99.0,
|
64962 |
-
"learning_rate": 5.432692307692308e-06,
|
64963 |
-
"loss": 0.1263,
|
64964 |
-
"step": 10791
|
64965 |
-
},
|
64966 |
-
{
|
64967 |
-
"epoch": 99.01,
|
64968 |
-
"learning_rate": 5.384615384615385e-06,
|
64969 |
-
"loss": 0.1653,
|
64970 |
-
"step": 10792
|
64971 |
-
},
|
64972 |
-
{
|
64973 |
-
"epoch": 99.02,
|
64974 |
-
"learning_rate": 5.3365384615384615e-06,
|
64975 |
-
"loss": 0.1337,
|
64976 |
-
"step": 10793
|
64977 |
-
},
|
64978 |
-
{
|
64979 |
-
"epoch": 99.03,
|
64980 |
-
"learning_rate": 5.288461538461539e-06,
|
64981 |
-
"loss": 0.1166,
|
64982 |
-
"step": 10794
|
64983 |
-
},
|
64984 |
-
{
|
64985 |
-
"epoch": 99.04,
|
64986 |
-
"learning_rate": 5.240384615384615e-06,
|
64987 |
-
"loss": 0.1185,
|
64988 |
-
"step": 10795
|
64989 |
-
},
|
64990 |
-
{
|
64991 |
-
"epoch": 99.05,
|
64992 |
-
"learning_rate": 5.192307692307692e-06,
|
64993 |
-
"loss": 0.1101,
|
64994 |
-
"step": 10796
|
64995 |
-
},
|
64996 |
-
{
|
64997 |
-
"epoch": 99.06,
|
64998 |
-
"learning_rate": 5.144230769230769e-06,
|
64999 |
-
"loss": 0.1459,
|
65000 |
-
"step": 10797
|
65001 |
-
},
|
65002 |
-
{
|
65003 |
-
"epoch": 99.06,
|
65004 |
-
"learning_rate": 5.096153846153846e-06,
|
65005 |
-
"loss": 0.1221,
|
65006 |
-
"step": 10798
|
65007 |
-
},
|
65008 |
-
{
|
65009 |
-
"epoch": 99.07,
|
65010 |
-
"learning_rate": 5.0480769230769235e-06,
|
65011 |
-
"loss": 0.106,
|
65012 |
-
"step": 10799
|
65013 |
-
},
|
65014 |
-
{
|
65015 |
-
"epoch": 99.08,
|
65016 |
-
"learning_rate": 5e-06,
|
65017 |
-
"loss": 0.1366,
|
65018 |
-
"step": 10800
|
65019 |
-
},
|
65020 |
-
{
|
65021 |
-
"epoch": 99.09,
|
65022 |
-
"learning_rate": 4.951923076923077e-06,
|
65023 |
-
"loss": 0.0836,
|
65024 |
-
"step": 10801
|
65025 |
-
},
|
65026 |
-
{
|
65027 |
-
"epoch": 99.1,
|
65028 |
-
"learning_rate": 4.903846153846153e-06,
|
65029 |
-
"loss": 0.1291,
|
65030 |
-
"step": 10802
|
65031 |
-
},
|
65032 |
-
{
|
65033 |
-
"epoch": 99.11,
|
65034 |
-
"learning_rate": 4.85576923076923e-06,
|
65035 |
-
"loss": 0.1328,
|
65036 |
-
"step": 10803
|
65037 |
-
},
|
65038 |
-
{
|
65039 |
-
"epoch": 99.12,
|
65040 |
-
"learning_rate": 4.807692307692308e-06,
|
65041 |
-
"loss": 0.0745,
|
65042 |
-
"step": 10804
|
65043 |
-
},
|
65044 |
-
{
|
65045 |
-
"epoch": 99.13,
|
65046 |
-
"learning_rate": 4.759615384615385e-06,
|
65047 |
-
"loss": 0.1366,
|
65048 |
-
"step": 10805
|
65049 |
-
},
|
65050 |
-
{
|
65051 |
-
"epoch": 99.14,
|
65052 |
-
"learning_rate": 4.711538461538462e-06,
|
65053 |
-
"loss": 0.0914,
|
65054 |
-
"step": 10806
|
65055 |
-
},
|
65056 |
-
{
|
65057 |
-
"epoch": 99.15,
|
65058 |
-
"learning_rate": 4.6634615384615384e-06,
|
65059 |
-
"loss": 0.1165,
|
65060 |
-
"step": 10807
|
65061 |
-
},
|
65062 |
-
{
|
65063 |
-
"epoch": 99.16,
|
65064 |
-
"learning_rate": 4.615384615384616e-06,
|
65065 |
-
"loss": 0.0281,
|
65066 |
-
"step": 10808
|
65067 |
-
},
|
65068 |
-
{
|
65069 |
-
"epoch": 99.17,
|
65070 |
-
"learning_rate": 4.567307692307693e-06,
|
65071 |
-
"loss": 0.1348,
|
65072 |
-
"step": 10809
|
65073 |
-
},
|
65074 |
-
{
|
65075 |
-
"epoch": 99.17,
|
65076 |
-
"learning_rate": 4.519230769230769e-06,
|
65077 |
-
"loss": 0.0689,
|
65078 |
-
"step": 10810
|
65079 |
-
},
|
65080 |
-
{
|
65081 |
-
"epoch": 99.18,
|
65082 |
-
"learning_rate": 4.471153846153846e-06,
|
65083 |
-
"loss": 0.1174,
|
65084 |
-
"step": 10811
|
65085 |
-
},
|
65086 |
-
{
|
65087 |
-
"epoch": 99.19,
|
65088 |
-
"learning_rate": 4.423076923076923e-06,
|
65089 |
-
"loss": 0.079,
|
65090 |
-
"step": 10812
|
65091 |
-
},
|
65092 |
-
{
|
65093 |
-
"epoch": 99.2,
|
65094 |
-
"learning_rate": 4.3750000000000005e-06,
|
65095 |
-
"loss": 0.1269,
|
65096 |
-
"step": 10813
|
65097 |
-
},
|
65098 |
-
{
|
65099 |
-
"epoch": 99.21,
|
65100 |
-
"learning_rate": 4.326923076923077e-06,
|
65101 |
-
"loss": 0.0556,
|
65102 |
-
"step": 10814
|
65103 |
-
},
|
65104 |
-
{
|
65105 |
-
"epoch": 99.22,
|
65106 |
-
"learning_rate": 4.278846153846154e-06,
|
65107 |
-
"loss": 0.0863,
|
65108 |
-
"step": 10815
|
65109 |
-
},
|
65110 |
-
{
|
65111 |
-
"epoch": 99.23,
|
65112 |
-
"learning_rate": 4.230769230769231e-06,
|
65113 |
-
"loss": 0.0922,
|
65114 |
-
"step": 10816
|
65115 |
-
},
|
65116 |
-
{
|
65117 |
-
"epoch": 99.24,
|
65118 |
-
"learning_rate": 4.182692307692308e-06,
|
65119 |
-
"loss": 0.0968,
|
65120 |
-
"step": 10817
|
65121 |
-
},
|
65122 |
-
{
|
65123 |
-
"epoch": 99.25,
|
65124 |
-
"learning_rate": 4.134615384615385e-06,
|
65125 |
-
"loss": 0.077,
|
65126 |
-
"step": 10818
|
65127 |
-
},
|
65128 |
-
{
|
65129 |
-
"epoch": 99.26,
|
65130 |
-
"learning_rate": 4.086538461538462e-06,
|
65131 |
-
"loss": 0.1543,
|
65132 |
-
"step": 10819
|
65133 |
-
},
|
65134 |
-
{
|
65135 |
-
"epoch": 99.27,
|
65136 |
-
"learning_rate": 4.0384615384615385e-06,
|
65137 |
-
"loss": 0.1635,
|
65138 |
-
"step": 10820
|
65139 |
-
},
|
65140 |
-
{
|
65141 |
-
"epoch": 99.28,
|
65142 |
-
"learning_rate": 3.990384615384615e-06,
|
65143 |
-
"loss": 0.0929,
|
65144 |
-
"step": 10821
|
65145 |
-
},
|
65146 |
-
{
|
65147 |
-
"epoch": 99.28,
|
65148 |
-
"learning_rate": 3.942307692307692e-06,
|
65149 |
-
"loss": 0.1135,
|
65150 |
-
"step": 10822
|
65151 |
-
},
|
65152 |
-
{
|
65153 |
-
"epoch": 99.29,
|
65154 |
-
"learning_rate": 3.894230769230769e-06,
|
65155 |
-
"loss": 0.1106,
|
65156 |
-
"step": 10823
|
65157 |
-
},
|
65158 |
-
{
|
65159 |
-
"epoch": 99.3,
|
65160 |
-
"learning_rate": 3.846153846153847e-06,
|
65161 |
-
"loss": 0.1107,
|
65162 |
-
"step": 10824
|
65163 |
-
},
|
65164 |
-
{
|
65165 |
-
"epoch": 99.31,
|
65166 |
-
"learning_rate": 3.7980769230769232e-06,
|
65167 |
-
"loss": 0.1289,
|
65168 |
-
"step": 10825
|
65169 |
-
},
|
65170 |
-
{
|
65171 |
-
"epoch": 99.32,
|
65172 |
-
"learning_rate": 3.75e-06,
|
65173 |
-
"loss": 0.0919,
|
65174 |
-
"step": 10826
|
65175 |
-
},
|
65176 |
-
{
|
65177 |
-
"epoch": 99.33,
|
65178 |
-
"learning_rate": 3.701923076923077e-06,
|
65179 |
-
"loss": 0.1318,
|
65180 |
-
"step": 10827
|
65181 |
-
},
|
65182 |
-
{
|
65183 |
-
"epoch": 99.34,
|
65184 |
-
"learning_rate": 3.653846153846154e-06,
|
65185 |
-
"loss": 0.0655,
|
65186 |
-
"step": 10828
|
65187 |
-
},
|
65188 |
-
{
|
65189 |
-
"epoch": 99.35,
|
65190 |
-
"learning_rate": 3.605769230769231e-06,
|
65191 |
-
"loss": 0.0838,
|
65192 |
-
"step": 10829
|
65193 |
-
},
|
65194 |
-
{
|
65195 |
-
"epoch": 99.36,
|
65196 |
-
"learning_rate": 3.557692307692308e-06,
|
65197 |
-
"loss": 0.0583,
|
65198 |
-
"step": 10830
|
65199 |
-
},
|
65200 |
-
{
|
65201 |
-
"epoch": 99.37,
|
65202 |
-
"learning_rate": 3.5096153846153844e-06,
|
65203 |
-
"loss": 0.0635,
|
65204 |
-
"step": 10831
|
65205 |
-
},
|
65206 |
-
{
|
65207 |
-
"epoch": 99.38,
|
65208 |
-
"learning_rate": 3.4615384615384617e-06,
|
65209 |
-
"loss": 0.1053,
|
65210 |
-
"step": 10832
|
65211 |
-
},
|
65212 |
-
{
|
65213 |
-
"epoch": 99.39,
|
65214 |
-
"learning_rate": 3.4134615384615386e-06,
|
65215 |
-
"loss": 0.1396,
|
65216 |
-
"step": 10833
|
65217 |
-
},
|
65218 |
-
{
|
65219 |
-
"epoch": 99.39,
|
65220 |
-
"learning_rate": 3.365384615384616e-06,
|
65221 |
-
"loss": 0.0984,
|
65222 |
-
"step": 10834
|
65223 |
-
},
|
65224 |
-
{
|
65225 |
-
"epoch": 99.4,
|
65226 |
-
"learning_rate": 3.3173076923076923e-06,
|
65227 |
-
"loss": 0.0749,
|
65228 |
-
"step": 10835
|
65229 |
-
},
|
65230 |
-
{
|
65231 |
-
"epoch": 99.41,
|
65232 |
-
"learning_rate": 3.269230769230769e-06,
|
65233 |
-
"loss": 0.1174,
|
65234 |
-
"step": 10836
|
65235 |
-
},
|
65236 |
-
{
|
65237 |
-
"epoch": 99.42,
|
65238 |
-
"learning_rate": 3.2211538461538464e-06,
|
65239 |
-
"loss": 0.0928,
|
65240 |
-
"step": 10837
|
65241 |
-
},
|
65242 |
-
{
|
65243 |
-
"epoch": 99.43,
|
65244 |
-
"learning_rate": 3.173076923076923e-06,
|
65245 |
-
"loss": 0.0549,
|
65246 |
-
"step": 10838
|
65247 |
-
},
|
65248 |
-
{
|
65249 |
-
"epoch": 99.44,
|
65250 |
-
"learning_rate": 3.125e-06,
|
65251 |
-
"loss": 0.0872,
|
65252 |
-
"step": 10839
|
65253 |
-
},
|
65254 |
-
{
|
65255 |
-
"epoch": 99.45,
|
65256 |
-
"learning_rate": 3.076923076923077e-06,
|
65257 |
-
"loss": 0.0704,
|
65258 |
-
"step": 10840
|
65259 |
-
},
|
65260 |
-
{
|
65261 |
-
"epoch": 99.46,
|
65262 |
-
"learning_rate": 3.028846153846154e-06,
|
65263 |
-
"loss": 0.0398,
|
65264 |
-
"step": 10841
|
65265 |
-
},
|
65266 |
-
{
|
65267 |
-
"epoch": 99.47,
|
65268 |
-
"learning_rate": 2.9807692307692307e-06,
|
65269 |
-
"loss": 0.094,
|
65270 |
-
"step": 10842
|
65271 |
-
},
|
65272 |
-
{
|
65273 |
-
"epoch": 99.48,
|
65274 |
-
"learning_rate": 2.9326923076923076e-06,
|
65275 |
-
"loss": 0.0902,
|
65276 |
-
"step": 10843
|
65277 |
-
},
|
65278 |
-
{
|
65279 |
-
"epoch": 99.49,
|
65280 |
-
"learning_rate": 2.884615384615385e-06,
|
65281 |
-
"loss": 0.0994,
|
65282 |
-
"step": 10844
|
65283 |
-
},
|
65284 |
-
{
|
65285 |
-
"epoch": 99.5,
|
65286 |
-
"learning_rate": 2.8365384615384617e-06,
|
65287 |
-
"loss": 0.0513,
|
65288 |
-
"step": 10845
|
65289 |
-
},
|
65290 |
-
{
|
65291 |
-
"epoch": 99.5,
|
65292 |
-
"learning_rate": 2.788461538461538e-06,
|
65293 |
-
"loss": 0.2055,
|
65294 |
-
"step": 10846
|
65295 |
-
},
|
65296 |
-
{
|
65297 |
-
"epoch": 99.51,
|
65298 |
-
"learning_rate": 2.7403846153846155e-06,
|
65299 |
-
"loss": 0.1316,
|
65300 |
-
"step": 10847
|
65301 |
-
},
|
65302 |
-
{
|
65303 |
-
"epoch": 99.52,
|
65304 |
-
"learning_rate": 2.6923076923076923e-06,
|
65305 |
-
"loss": 0.1219,
|
65306 |
-
"step": 10848
|
65307 |
-
},
|
65308 |
-
{
|
65309 |
-
"epoch": 99.53,
|
65310 |
-
"learning_rate": 2.6442307692307696e-06,
|
65311 |
-
"loss": 0.1132,
|
65312 |
-
"step": 10849
|
65313 |
-
},
|
65314 |
-
{
|
65315 |
-
"epoch": 99.54,
|
65316 |
-
"learning_rate": 2.596153846153846e-06,
|
65317 |
-
"loss": 0.1072,
|
65318 |
-
"step": 10850
|
65319 |
-
},
|
65320 |
-
{
|
65321 |
-
"epoch": 99.55,
|
65322 |
-
"learning_rate": 2.548076923076923e-06,
|
65323 |
-
"loss": 0.0983,
|
65324 |
-
"step": 10851
|
65325 |
-
},
|
65326 |
-
{
|
65327 |
-
"epoch": 99.56,
|
65328 |
-
"learning_rate": 2.5e-06,
|
65329 |
-
"loss": 0.1457,
|
65330 |
-
"step": 10852
|
65331 |
-
},
|
65332 |
-
{
|
65333 |
-
"epoch": 99.57,
|
65334 |
-
"learning_rate": 2.4519230769230766e-06,
|
65335 |
-
"loss": 0.0745,
|
65336 |
-
"step": 10853
|
65337 |
-
},
|
65338 |
-
{
|
65339 |
-
"epoch": 99.58,
|
65340 |
-
"learning_rate": 2.403846153846154e-06,
|
65341 |
-
"loss": 0.0876,
|
65342 |
-
"step": 10854
|
65343 |
-
},
|
65344 |
-
{
|
65345 |
-
"epoch": 99.59,
|
65346 |
-
"learning_rate": 2.355769230769231e-06,
|
65347 |
-
"loss": 0.1684,
|
65348 |
-
"step": 10855
|
65349 |
-
},
|
65350 |
-
{
|
65351 |
-
"epoch": 99.6,
|
65352 |
-
"learning_rate": 2.307692307692308e-06,
|
65353 |
-
"loss": 0.0859,
|
65354 |
-
"step": 10856
|
65355 |
-
},
|
65356 |
-
{
|
65357 |
-
"epoch": 99.61,
|
65358 |
-
"learning_rate": 2.2596153846153845e-06,
|
65359 |
-
"loss": 0.1476,
|
65360 |
-
"step": 10857
|
65361 |
-
},
|
65362 |
-
{
|
65363 |
-
"epoch": 99.61,
|
65364 |
-
"learning_rate": 2.2115384615384614e-06,
|
65365 |
-
"loss": 0.0793,
|
65366 |
-
"step": 10858
|
65367 |
-
},
|
65368 |
-
{
|
65369 |
-
"epoch": 99.62,
|
65370 |
-
"learning_rate": 2.1634615384615387e-06,
|
65371 |
-
"loss": 0.1262,
|
65372 |
-
"step": 10859
|
65373 |
-
},
|
65374 |
-
{
|
65375 |
-
"epoch": 99.63,
|
65376 |
-
"learning_rate": 2.1153846153846155e-06,
|
65377 |
-
"loss": 0.1145,
|
65378 |
-
"step": 10860
|
65379 |
-
},
|
65380 |
-
{
|
65381 |
-
"epoch": 99.64,
|
65382 |
-
"learning_rate": 2.0673076923076924e-06,
|
65383 |
-
"loss": 0.0778,
|
65384 |
-
"step": 10861
|
65385 |
-
},
|
65386 |
-
{
|
65387 |
-
"epoch": 99.65,
|
65388 |
-
"learning_rate": 2.0192307692307692e-06,
|
65389 |
-
"loss": 0.067,
|
65390 |
-
"step": 10862
|
65391 |
-
},
|
65392 |
-
{
|
65393 |
-
"epoch": 99.66,
|
65394 |
-
"learning_rate": 1.971153846153846e-06,
|
65395 |
-
"loss": 0.0686,
|
65396 |
-
"step": 10863
|
65397 |
-
},
|
65398 |
-
{
|
65399 |
-
"epoch": 99.67,
|
65400 |
-
"learning_rate": 1.9230769230769234e-06,
|
65401 |
-
"loss": 0.0865,
|
65402 |
-
"step": 10864
|
65403 |
-
},
|
65404 |
-
{
|
65405 |
-
"epoch": 99.68,
|
65406 |
-
"learning_rate": 1.875e-06,
|
65407 |
-
"loss": 0.1209,
|
65408 |
-
"step": 10865
|
65409 |
-
},
|
65410 |
-
{
|
65411 |
-
"epoch": 99.69,
|
65412 |
-
"learning_rate": 1.826923076923077e-06,
|
65413 |
-
"loss": 0.0729,
|
65414 |
-
"step": 10866
|
65415 |
-
},
|
65416 |
-
{
|
65417 |
-
"epoch": 99.7,
|
65418 |
-
"learning_rate": 1.778846153846154e-06,
|
65419 |
-
"loss": 0.1472,
|
65420 |
-
"step": 10867
|
65421 |
-
},
|
65422 |
-
{
|
65423 |
-
"epoch": 99.71,
|
65424 |
-
"learning_rate": 1.7307692307692308e-06,
|
65425 |
-
"loss": 0.0516,
|
65426 |
-
"step": 10868
|
65427 |
-
},
|
65428 |
-
{
|
65429 |
-
"epoch": 99.72,
|
65430 |
-
"learning_rate": 1.682692307692308e-06,
|
65431 |
-
"loss": 0.0901,
|
65432 |
-
"step": 10869
|
65433 |
-
},
|
65434 |
-
{
|
65435 |
-
"epoch": 99.72,
|
65436 |
-
"learning_rate": 1.6346153846153846e-06,
|
65437 |
-
"loss": 0.0505,
|
65438 |
-
"step": 10870
|
65439 |
-
},
|
65440 |
-
{
|
65441 |
-
"epoch": 99.73,
|
65442 |
-
"learning_rate": 1.5865384615384614e-06,
|
65443 |
-
"loss": 0.0629,
|
65444 |
-
"step": 10871
|
65445 |
-
},
|
65446 |
-
{
|
65447 |
-
"epoch": 99.74,
|
65448 |
-
"learning_rate": 1.5384615384615385e-06,
|
65449 |
-
"loss": 0.2219,
|
65450 |
-
"step": 10872
|
65451 |
-
},
|
65452 |
-
{
|
65453 |
-
"epoch": 99.75,
|
65454 |
-
"learning_rate": 1.4903846153846154e-06,
|
65455 |
-
"loss": 0.1274,
|
65456 |
-
"step": 10873
|
65457 |
-
},
|
65458 |
-
{
|
65459 |
-
"epoch": 99.76,
|
65460 |
-
"learning_rate": 1.4423076923076924e-06,
|
65461 |
-
"loss": 0.1312,
|
65462 |
-
"step": 10874
|
65463 |
-
},
|
65464 |
-
{
|
65465 |
-
"epoch": 99.77,
|
65466 |
-
"learning_rate": 1.394230769230769e-06,
|
65467 |
-
"loss": 0.1167,
|
65468 |
-
"step": 10875
|
65469 |
-
},
|
65470 |
-
{
|
65471 |
-
"epoch": 99.78,
|
65472 |
-
"learning_rate": 1.3461538461538462e-06,
|
65473 |
-
"loss": 0.0919,
|
65474 |
-
"step": 10876
|
65475 |
-
},
|
65476 |
-
{
|
65477 |
-
"epoch": 99.79,
|
65478 |
-
"learning_rate": 1.298076923076923e-06,
|
65479 |
-
"loss": 0.17,
|
65480 |
-
"step": 10877
|
65481 |
-
},
|
65482 |
-
{
|
65483 |
-
"epoch": 99.8,
|
65484 |
-
"learning_rate": 1.25e-06,
|
65485 |
-
"loss": 0.1601,
|
65486 |
-
"step": 10878
|
65487 |
-
},
|
65488 |
-
{
|
65489 |
-
"epoch": 99.81,
|
65490 |
-
"learning_rate": 1.201923076923077e-06,
|
65491 |
-
"loss": 0.0664,
|
65492 |
-
"step": 10879
|
65493 |
-
},
|
65494 |
-
{
|
65495 |
-
"epoch": 99.82,
|
65496 |
-
"learning_rate": 1.153846153846154e-06,
|
65497 |
-
"loss": 0.08,
|
65498 |
-
"step": 10880
|
65499 |
-
},
|
65500 |
-
{
|
65501 |
-
"epoch": 99.83,
|
65502 |
-
"learning_rate": 1.1057692307692307e-06,
|
65503 |
-
"loss": 0.1484,
|
65504 |
-
"step": 10881
|
65505 |
-
},
|
65506 |
-
{
|
65507 |
-
"epoch": 99.83,
|
65508 |
-
"learning_rate": 1.0576923076923078e-06,
|
65509 |
-
"loss": 0.1291,
|
65510 |
-
"step": 10882
|
65511 |
-
},
|
65512 |
-
{
|
65513 |
-
"epoch": 99.84,
|
65514 |
-
"learning_rate": 1.0096153846153846e-06,
|
65515 |
-
"loss": 0.1387,
|
65516 |
-
"step": 10883
|
65517 |
-
},
|
65518 |
-
{
|
65519 |
-
"epoch": 99.85,
|
65520 |
-
"learning_rate": 9.615384615384617e-07,
|
65521 |
-
"loss": 0.1047,
|
65522 |
-
"step": 10884
|
65523 |
-
},
|
65524 |
-
{
|
65525 |
-
"epoch": 99.86,
|
65526 |
-
"learning_rate": 9.134615384615385e-07,
|
65527 |
-
"loss": 0.1096,
|
65528 |
-
"step": 10885
|
65529 |
-
},
|
65530 |
-
{
|
65531 |
-
"epoch": 99.87,
|
65532 |
-
"learning_rate": 8.653846153846154e-07,
|
65533 |
-
"loss": 0.1193,
|
65534 |
-
"step": 10886
|
65535 |
-
},
|
65536 |
-
{
|
65537 |
-
"epoch": 99.88,
|
65538 |
-
"learning_rate": 8.173076923076923e-07,
|
65539 |
-
"loss": 0.1043,
|
65540 |
-
"step": 10887
|
65541 |
-
},
|
65542 |
-
{
|
65543 |
-
"epoch": 99.89,
|
65544 |
-
"learning_rate": 7.692307692307693e-07,
|
65545 |
-
"loss": 0.178,
|
65546 |
-
"step": 10888
|
65547 |
-
},
|
65548 |
-
{
|
65549 |
-
"epoch": 99.9,
|
65550 |
-
"learning_rate": 7.211538461538462e-07,
|
65551 |
-
"loss": 0.0839,
|
65552 |
-
"step": 10889
|
65553 |
-
},
|
65554 |
-
{
|
65555 |
-
"epoch": 99.91,
|
65556 |
-
"learning_rate": 6.730769230769231e-07,
|
65557 |
-
"loss": 0.0829,
|
65558 |
-
"step": 10890
|
65559 |
-
},
|
65560 |
-
{
|
65561 |
-
"epoch": 99.92,
|
65562 |
-
"learning_rate": 6.25e-07,
|
65563 |
-
"loss": 0.1001,
|
65564 |
-
"step": 10891
|
65565 |
-
},
|
65566 |
-
{
|
65567 |
-
"epoch": 99.93,
|
65568 |
-
"learning_rate": 5.76923076923077e-07,
|
65569 |
-
"loss": 0.0748,
|
65570 |
-
"step": 10892
|
65571 |
-
},
|
65572 |
-
{
|
65573 |
-
"epoch": 99.94,
|
65574 |
-
"learning_rate": 5.288461538461539e-07,
|
65575 |
-
"loss": 0.0612,
|
65576 |
-
"step": 10893
|
65577 |
-
},
|
65578 |
-
{
|
65579 |
-
"epoch": 99.94,
|
65580 |
-
"learning_rate": 4.807692307692308e-07,
|
65581 |
-
"loss": 0.1034,
|
65582 |
-
"step": 10894
|
65583 |
-
},
|
65584 |
-
{
|
65585 |
-
"epoch": 99.95,
|
65586 |
-
"learning_rate": 4.326923076923077e-07,
|
65587 |
-
"loss": 0.0369,
|
65588 |
-
"step": 10895
|
65589 |
-
},
|
65590 |
-
{
|
65591 |
-
"epoch": 99.96,
|
65592 |
-
"learning_rate": 3.8461538461538463e-07,
|
65593 |
-
"loss": 0.1068,
|
65594 |
-
"step": 10896
|
65595 |
-
},
|
65596 |
-
{
|
65597 |
-
"epoch": 99.97,
|
65598 |
-
"learning_rate": 3.3653846153846154e-07,
|
65599 |
-
"loss": 0.0879,
|
65600 |
-
"step": 10897
|
65601 |
-
},
|
65602 |
-
{
|
65603 |
-
"epoch": 99.98,
|
65604 |
-
"learning_rate": 2.884615384615385e-07,
|
65605 |
-
"loss": 0.1357,
|
65606 |
-
"step": 10898
|
65607 |
-
},
|
65608 |
-
{
|
65609 |
-
"epoch": 99.99,
|
65610 |
-
"learning_rate": 2.403846153846154e-07,
|
65611 |
-
"loss": 0.0615,
|
65612 |
-
"step": 10899
|
65613 |
-
},
|
65614 |
-
{
|
65615 |
-
"epoch": 100.0,
|
65616 |
-
"learning_rate": 1.9230769230769231e-07,
|
65617 |
-
"loss": 0.0967,
|
65618 |
-
"step": 10900
|
65619 |
-
},
|
65620 |
-
{
|
65621 |
-
"epoch": 100.0,
|
65622 |
-
"step": 10900,
|
65623 |
-
"total_flos": 1.3360714276118677e+20,
|
65624 |
-
"train_loss": 0.44451899920906357,
|
65625 |
-
"train_runtime": 28212.0393,
|
65626 |
-
"train_samples_per_second": 12.328,
|
65627 |
-
"train_steps_per_second": 0.386
|
65628 |
}
|
65629 |
],
|
65630 |
-
"max_steps":
|
65631 |
-
"num_train_epochs":
|
65632 |
-
"total_flos": 1.
|
65633 |
"trial_name": null,
|
65634 |
"trial_params": null
|
65635 |
}
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 96.3302752293578,
|
5 |
+
"global_step": 10500,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
63218 |
"step": 10500
|
63219 |
},
|
63220 |
{
|
63221 |
+
"epoch": 96.33,
|
63222 |
+
"step": 10500,
|
63223 |
+
"total_flos": 1.2873670788396168e+20,
|
63224 |
+
"train_loss": 0.0,
|
63225 |
+
"train_runtime": 113.9955,
|
63226 |
+
"train_samples_per_second": 152.55,
|
63227 |
+
"train_steps_per_second": 4.781
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
63228 |
}
|
63229 |
],
|
63230 |
+
"max_steps": 545,
|
63231 |
+
"num_train_epochs": 5,
|
63232 |
+
"total_flos": 1.2873670788396168e+20,
|
63233 |
"trial_name": null,
|
63234 |
"trial_params": null
|
63235 |
}
|