|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 7.0, |
|
"eval_steps": 500, |
|
"global_step": 399, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 0.422, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 0.5154, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 5e-06, |
|
"loss": 0.3768, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.356, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 8.333333333333334e-06, |
|
"loss": 0.2775, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1e-05, |
|
"loss": 0.1761, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.1666666666666668e-05, |
|
"loss": 0.218, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 0.2161, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.2634, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 0.2294, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8333333333333333e-05, |
|
"loss": 0.1694, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 2e-05, |
|
"loss": 0.1751, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9999670507574945e-05, |
|
"loss": 0.1785, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.999868205201284e-05, |
|
"loss": 0.1852, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9997034698451396e-05, |
|
"loss": 0.1508, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9994728555448723e-05, |
|
"loss": 0.1505, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.999176377497616e-05, |
|
"loss": 0.1586, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.998814055240823e-05, |
|
"loss": 0.1438, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9983859126509827e-05, |
|
"loss": 0.1479, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.9978919779420427e-05, |
|
"loss": 0.1612, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9973322836635517e-05, |
|
"loss": 0.1547, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.996706866698515e-05, |
|
"loss": 0.1871, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9960157682609634e-05, |
|
"loss": 0.136, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.995259033893236e-05, |
|
"loss": 0.1541, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.994436713462982e-05, |
|
"loss": 0.1466, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9935488611598716e-05, |
|
"loss": 0.1409, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9925955354920265e-05, |
|
"loss": 0.1455, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.9915767992821642e-05, |
|
"loss": 0.1687, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.9904927196634572e-05, |
|
"loss": 0.1488, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.9893433680751105e-05, |
|
"loss": 0.1545, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.9881288202576518e-05, |
|
"loss": 0.1533, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.986849156247943e-05, |
|
"loss": 0.158, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.985504460373903e-05, |
|
"loss": 0.1545, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.9840948212489527e-05, |
|
"loss": 0.1597, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.9826203317661758e-05, |
|
"loss": 0.1556, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.9810810890921943e-05, |
|
"loss": 0.1581, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.979477194660769e-05, |
|
"loss": 0.1625, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.977808754166113e-05, |
|
"loss": 0.1562, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.9760758775559275e-05, |
|
"loss": 0.1503, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.9742786790241548e-05, |
|
"loss": 0.1549, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.9724172770034566e-05, |
|
"loss": 0.1736, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.9704917941574053e-05, |
|
"loss": 0.1537, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.9685023573724036e-05, |
|
"loss": 0.1483, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.9664490977493223e-05, |
|
"loss": 0.1675, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.9643321505948588e-05, |
|
"loss": 0.1436, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.962151655412624e-05, |
|
"loss": 0.1536, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.9599077558939468e-05, |
|
"loss": 0.146, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.957600599908406e-05, |
|
"loss": 0.1559, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.9552303394940862e-05, |
|
"loss": 0.1508, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.952797130847557e-05, |
|
"loss": 0.1468, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.9503011343135828e-05, |
|
"loss": 0.1562, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.947742514374553e-05, |
|
"loss": 0.174, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.9451214396396453e-05, |
|
"loss": 0.1482, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.9424380828337146e-05, |
|
"loss": 0.1531, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.9396926207859085e-05, |
|
"loss": 0.1744, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.9368852344180168e-05, |
|
"loss": 0.1439, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.9340161087325483e-05, |
|
"loss": 0.1617, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.9310854328005383e-05, |
|
"loss": 0.1442, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.9280933997490912e-05, |
|
"loss": 0.1461, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.9250402067486523e-05, |
|
"loss": 0.1375, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9219260550000144e-05, |
|
"loss": 0.1385, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.91875114972106e-05, |
|
"loss": 0.1347, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.9155157001332374e-05, |
|
"loss": 0.1273, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.9122199194477723e-05, |
|
"loss": 0.1539, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.9088640248516185e-05, |
|
"loss": 0.1375, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.905448237493147e-05, |
|
"loss": 0.1694, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.9019727824675686e-05, |
|
"loss": 0.1813, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.8984378888021045e-05, |
|
"loss": 0.1409, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.894843789440892e-05, |
|
"loss": 0.1333, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.8911907212296343e-05, |
|
"loss": 0.1657, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.8874789248999915e-05, |
|
"loss": 0.1627, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.8837086450537195e-05, |
|
"loss": 0.14, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.879880130146547e-05, |
|
"loss": 0.1388, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.8759936324718068e-05, |
|
"loss": 0.1353, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.872049408143808e-05, |
|
"loss": 0.1495, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.8680477170809573e-05, |
|
"loss": 0.1539, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.8639888229886342e-05, |
|
"loss": 0.1406, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.8598729933418102e-05, |
|
"loss": 0.1413, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.855700499367423e-05, |
|
"loss": 0.1384, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.851471616026505e-05, |
|
"loss": 0.1638, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.8471866219960604e-05, |
|
"loss": 0.1602, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.8428457996507053e-05, |
|
"loss": 0.1592, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.8384494350440553e-05, |
|
"loss": 0.1455, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.833997817889878e-05, |
|
"loss": 0.1575, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.8294912415429995e-05, |
|
"loss": 0.1343, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.8249300029799735e-05, |
|
"loss": 0.1668, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.820314402779511e-05, |
|
"loss": 0.1696, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.815644745102673e-05, |
|
"loss": 0.1436, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.810921337672826e-05, |
|
"loss": 0.1473, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.806144491755363e-05, |
|
"loss": 0.145, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.8013145221371937e-05, |
|
"loss": 0.136, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.7964317471059982e-05, |
|
"loss": 0.142, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.7914964884292543e-05, |
|
"loss": 0.1548, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.7865090713330313e-05, |
|
"loss": 0.1357, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.7814698244805605e-05, |
|
"loss": 0.1433, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.7763790799505746e-05, |
|
"loss": 0.1339, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.771237173215426e-05, |
|
"loss": 0.1215, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.766044443118978e-05, |
|
"loss": 0.1359, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.760801231854278e-05, |
|
"loss": 0.136, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.7555078849410044e-05, |
|
"loss": 0.1477, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.7501647512026996e-05, |
|
"loss": 0.137, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.744772182743782e-05, |
|
"loss": 0.1322, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.7393305349263433e-05, |
|
"loss": 0.1319, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.733840166346731e-05, |
|
"loss": 0.139, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.728301438811916e-05, |
|
"loss": 0.1355, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.7227147173156525e-05, |
|
"loss": 0.1202, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.7170803700144227e-05, |
|
"loss": 0.1146, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.711398768203178e-05, |
|
"loss": 0.1274, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.7056702862908704e-05, |
|
"loss": 0.1246, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.6998953017757787e-05, |
|
"loss": 0.1471, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.6940741952206342e-05, |
|
"loss": 0.1569, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.6882073502275394e-05, |
|
"loss": 0.1115, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.682295153412691e-05, |
|
"loss": 0.1191, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.676337994380903e-05, |
|
"loss": 0.1409, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.67033626569993e-05, |
|
"loss": 0.1032, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.6642903628745995e-05, |
|
"loss": 0.1077, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 1.658200684320748e-05, |
|
"loss": 0.1096, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 1.652067631338967e-05, |
|
"loss": 0.1263, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.6458916080881566e-05, |
|
"loss": 0.1081, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 1.6396730215588913e-05, |
|
"loss": 0.1276, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.6334122815466035e-05, |
|
"loss": 0.1161, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.6271098006245742e-05, |
|
"loss": 0.1141, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.6207659941167485e-05, |
|
"loss": 0.121, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.6143812800703646e-05, |
|
"loss": 0.093, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.6079560792284046e-05, |
|
"loss": 0.131, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.6014908150018703e-05, |
|
"loss": 0.0955, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.5949859134418798e-05, |
|
"loss": 0.1054, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.5884418032115907e-05, |
|
"loss": 0.1262, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.581858915557953e-05, |
|
"loss": 0.1168, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.57523768428329e-05, |
|
"loss": 0.1003, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.5685785457167114e-05, |
|
"loss": 0.111, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.5618819386853607e-05, |
|
"loss": 0.1134, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.5551483044854954e-05, |
|
"loss": 0.1193, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.5483780868534087e-05, |
|
"loss": 0.1217, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.541571731936185e-05, |
|
"loss": 0.1028, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.534729688262302e-05, |
|
"loss": 0.115, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 1.527852406712072e-05, |
|
"loss": 0.1047, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.5209403404879305e-05, |
|
"loss": 0.1238, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.5139939450845699e-05, |
|
"loss": 0.1031, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.5070136782589236e-05, |
|
"loss": 0.0983, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.1298, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.4929533724985712e-05, |
|
"loss": 0.1263, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.485874260116714e-05, |
|
"loss": 0.1331, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.4787631293572094e-05, |
|
"loss": 0.1134, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.4716204488328007e-05, |
|
"loss": 0.125, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.4644466892353142e-05, |
|
"loss": 0.1047, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.4572423233046386e-05, |
|
"loss": 0.0993, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.4500078257975748e-05, |
|
"loss": 0.1071, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.4427436734565475e-05, |
|
"loss": 0.1047, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 1.4354503449781914e-05, |
|
"loss": 0.1008, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 1.428128320981804e-05, |
|
"loss": 0.1173, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.4207780839776735e-05, |
|
"loss": 0.1268, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.4134001183352833e-05, |
|
"loss": 0.119, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.4059949102513914e-05, |
|
"loss": 0.1138, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.3985629477179915e-05, |
|
"loss": 0.125, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.391104720490156e-05, |
|
"loss": 0.1024, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.3836207200537597e-05, |
|
"loss": 0.0989, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.376111439593093e-05, |
|
"loss": 0.1017, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.368577373958362e-05, |
|
"loss": 0.1113, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.3610190196330777e-05, |
|
"loss": 0.1196, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.3534368747013396e-05, |
|
"loss": 0.1314, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.3458314388150115e-05, |
|
"loss": 0.1279, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.3382032131607967e-05, |
|
"loss": 0.1216, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.3305527004272087e-05, |
|
"loss": 0.1212, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.3228804047714462e-05, |
|
"loss": 0.1388, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.31518683178617e-05, |
|
"loss": 0.1193, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.3074724884661833e-05, |
|
"loss": 0.113, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.2997378831750242e-05, |
|
"loss": 0.1115, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.2919835256114639e-05, |
|
"loss": 0.1232, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.2842099267759176e-05, |
|
"loss": 0.118, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.2764175989367717e-05, |
|
"loss": 0.1046, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.2686070555966255e-05, |
|
"loss": 0.1047, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 1.2607788114584522e-05, |
|
"loss": 0.1192, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.2529333823916807e-05, |
|
"loss": 0.1008, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.2450712853982014e-05, |
|
"loss": 0.0812, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.2371930385782944e-05, |
|
"loss": 0.0961, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 1.2292991610964902e-05, |
|
"loss": 0.0839, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.2213901731473555e-05, |
|
"loss": 0.0851, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.2134665959212138e-05, |
|
"loss": 0.0851, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.2055289515698008e-05, |
|
"loss": 0.0709, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.1975777631718533e-05, |
|
"loss": 0.0732, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.189613554698641e-05, |
|
"loss": 0.1016, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.1816368509794365e-05, |
|
"loss": 0.0832, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.1736481776669307e-05, |
|
"loss": 0.073, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.1656480612025914e-05, |
|
"loss": 0.0744, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.1576370287819737e-05, |
|
"loss": 0.0743, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.149615608319978e-05, |
|
"loss": 0.0573, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.14158432841606e-05, |
|
"loss": 0.0829, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 1.133543718319398e-05, |
|
"loss": 0.0564, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 1.1254943078940161e-05, |
|
"loss": 0.1077, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.1174366275838664e-05, |
|
"loss": 0.0831, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.1093712083778748e-05, |
|
"loss": 0.079, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.1012985817749465e-05, |
|
"loss": 0.0762, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.093219279748946e-05, |
|
"loss": 0.0594, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.0851338347136358e-05, |
|
"loss": 0.0734, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.077042779487595e-05, |
|
"loss": 0.0868, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.068946647259105e-05, |
|
"loss": 0.0958, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.060845971551014e-05, |
|
"loss": 0.0836, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.052741286185579e-05, |
|
"loss": 0.1155, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.0446331252492864e-05, |
|
"loss": 0.0799, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.0365220230576592e-05, |
|
"loss": 0.0935, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.0284085141200424e-05, |
|
"loss": 0.0853, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.0202931331043839e-05, |
|
"loss": 0.0703, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 1.0121764148019977e-05, |
|
"loss": 0.0951, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.004058894092323e-05, |
|
"loss": 0.0804, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 9.95941105907677e-06, |
|
"loss": 0.0963, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 9.878235851980027e-06, |
|
"loss": 0.0955, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 9.797068668956163e-06, |
|
"loss": 0.0862, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 9.715914858799576e-06, |
|
"loss": 0.0737, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 9.634779769423412e-06, |
|
"loss": 0.1326, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 9.553668747507139e-06, |
|
"loss": 0.079, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 9.472587138144215e-06, |
|
"loss": 0.0948, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 9.391540284489862e-06, |
|
"loss": 0.0698, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 9.310533527408953e-06, |
|
"loss": 0.0785, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 9.229572205124052e-06, |
|
"loss": 0.078, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 9.148661652863644e-06, |
|
"loss": 0.0861, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 9.067807202510543e-06, |
|
"loss": 0.0847, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 8.987014182250538e-06, |
|
"loss": 0.073, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 8.906287916221259e-06, |
|
"loss": 0.0897, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 8.825633724161336e-06, |
|
"loss": 0.0642, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 8.74505692105984e-06, |
|
"loss": 0.0862, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 8.664562816806022e-06, |
|
"loss": 0.0767, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 8.584156715839402e-06, |
|
"loss": 0.0593, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 8.503843916800221e-06, |
|
"loss": 0.0947, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 8.423629712180265e-06, |
|
"loss": 0.0717, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 8.343519387974091e-06, |
|
"loss": 0.0909, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 8.263518223330698e-06, |
|
"loss": 0.0657, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.183631490205636e-06, |
|
"loss": 0.0593, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.103864453013593e-06, |
|
"loss": 0.0517, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 8.024222368281469e-06, |
|
"loss": 0.0422, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 7.944710484301995e-06, |
|
"loss": 0.0432, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 7.865334040787866e-06, |
|
"loss": 0.0531, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 7.786098268526447e-06, |
|
"loss": 0.043, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 7.707008389035102e-06, |
|
"loss": 0.0303, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 7.628069614217058e-06, |
|
"loss": 0.0582, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 7.54928714601799e-06, |
|
"loss": 0.0372, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 7.470666176083193e-06, |
|
"loss": 0.0364, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 7.3922118854154815e-06, |
|
"loss": 0.0444, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 7.31392944403375e-06, |
|
"loss": 0.0577, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 7.235824010632284e-06, |
|
"loss": 0.0252, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 7.157900732240828e-06, |
|
"loss": 0.0406, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 7.080164743885364e-06, |
|
"loss": 0.0426, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 7.002621168249759e-06, |
|
"loss": 0.0229, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 6.925275115338168e-06, |
|
"loss": 0.0299, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 6.848131682138304e-06, |
|
"loss": 0.0474, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 6.771195952285541e-06, |
|
"loss": 0.0421, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 6.694472995727914e-06, |
|
"loss": 0.031, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 6.617967868392036e-06, |
|
"loss": 0.0392, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 6.5416856118498874e-06, |
|
"loss": 0.0357, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 6.4656312529866086e-06, |
|
"loss": 0.0436, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 6.389809803669226e-06, |
|
"loss": 0.0355, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 6.314226260416383e-06, |
|
"loss": 0.0492, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 6.238885604069075e-06, |
|
"loss": 0.016, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 6.163792799462404e-06, |
|
"loss": 0.0327, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 6.088952795098442e-06, |
|
"loss": 0.0511, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 6.014370522820085e-06, |
|
"loss": 0.0609, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 5.9400508974860895e-06, |
|
"loss": 0.0286, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 5.8659988166471715e-06, |
|
"loss": 0.0436, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 5.792219160223268e-06, |
|
"loss": 0.0556, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 5.7187167901819665e-06, |
|
"loss": 0.0761, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 5.645496550218089e-06, |
|
"loss": 0.046, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 5.572563265434528e-06, |
|
"loss": 0.0292, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 5.499921742024258e-06, |
|
"loss": 0.0283, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 5.427576766953615e-06, |
|
"loss": 0.0351, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 5.3555331076468585e-06, |
|
"loss": 0.0391, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 5.283795511671994e-06, |
|
"loss": 0.0524, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 5.212368706427913e-06, |
|
"loss": 0.0389, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 5.141257398832863e-06, |
|
"loss": 0.0379, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 5.0704662750142875e-06, |
|
"loss": 0.045, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 5.000000000000003e-06, |
|
"loss": 0.0348, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 4.929863217410768e-06, |
|
"loss": 0.0548, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 4.860060549154302e-06, |
|
"loss": 0.0467, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 4.790596595120699e-06, |
|
"loss": 0.0423, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 4.721475932879283e-06, |
|
"loss": 0.0434, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 4.652703117376987e-06, |
|
"loss": 0.0383, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 4.584282680638155e-06, |
|
"loss": 0.0431, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 4.51621913146592e-06, |
|
"loss": 0.0346, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 4.448516955145048e-06, |
|
"loss": 0.0323, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 4.381180613146396e-06, |
|
"loss": 0.032, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 4.314214542832889e-06, |
|
"loss": 0.051, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 4.2476231571671025e-06, |
|
"loss": 0.05, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 4.181410844420473e-06, |
|
"loss": 0.0302, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 4.1155819678840935e-06, |
|
"loss": 0.0471, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 4.050140865581205e-06, |
|
"loss": 0.0379, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 3.9850918499812976e-06, |
|
"loss": 0.0414, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 3.920439207715955e-06, |
|
"loss": 0.032, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 3.8561871992963585e-06, |
|
"loss": 0.0191, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 3.7923400588325156e-06, |
|
"loss": 0.0167, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 3.7289019937542603e-06, |
|
"loss": 0.0157, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 5.09, |
|
"learning_rate": 3.665877184533968e-06, |
|
"loss": 0.0247, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 3.6032697844110896e-06, |
|
"loss": 0.0128, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 3.5410839191184386e-06, |
|
"loss": 0.0147, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 3.4793236866103296e-06, |
|
"loss": 0.019, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 3.4179931567925216e-06, |
|
"loss": 0.0067, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 3.3570963712540083e-06, |
|
"loss": 0.0141, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 3.2966373430007047e-06, |
|
"loss": 0.0192, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 3.236620056190972e-06, |
|
"loss": 0.0114, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 3.1770484658730896e-06, |
|
"loss": 0.0092, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 3.1179264977246106e-06, |
|
"loss": 0.0263, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 3.0592580477936606e-06, |
|
"loss": 0.0119, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 3.001046982242216e-06, |
|
"loss": 0.0164, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 2.9432971370913e-06, |
|
"loss": 0.0086, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 2.8860123179682244e-06, |
|
"loss": 0.0173, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 2.8291962998557754e-06, |
|
"loss": 0.0255, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 2.772852826843476e-06, |
|
"loss": 0.0195, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 2.7169856118808414e-06, |
|
"loss": 0.0115, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 2.6615983365326926e-06, |
|
"loss": 0.0176, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 2.60669465073657e-06, |
|
"loss": 0.0088, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 2.5522781725621814e-06, |
|
"loss": 0.0313, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 2.498352487973007e-06, |
|
"loss": 0.0221, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 2.444921150589957e-06, |
|
"loss": 0.0201, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 2.3919876814572197e-06, |
|
"loss": 0.018, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 2.339555568810221e-06, |
|
"loss": 0.0125, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 2.287628267845744e-06, |
|
"loss": 0.0087, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 2.2362092004942583e-06, |
|
"loss": 0.0098, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 2.185301755194399e-06, |
|
"loss": 0.0135, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 2.1349092866696906e-06, |
|
"loss": 0.005, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 2.08503511570746e-06, |
|
"loss": 0.0176, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 2.0356825289400185e-06, |
|
"loss": 0.0114, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 1.986854778628067e-06, |
|
"loss": 0.0074, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 1.9385550824463727e-06, |
|
"loss": 0.0171, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 1.890786623271743e-06, |
|
"loss": 0.0166, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 1.843552548973272e-06, |
|
"loss": 0.0227, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 1.7968559722048906e-06, |
|
"loss": 0.0134, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 1.7506999702002682e-06, |
|
"loss": 0.0135, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 1.7050875845700066e-06, |
|
"loss": 0.0126, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.660021821101222e-06, |
|
"loss": 0.0065, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 1.6155056495594467e-06, |
|
"loss": 0.0061, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 1.5715420034929475e-06, |
|
"loss": 0.0069, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 1.528133780039397e-06, |
|
"loss": 0.0384, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 1.4852838397349545e-06, |
|
"loss": 0.0142, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.4429950063257713e-06, |
|
"loss": 0.0043, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 1.401270066581899e-06, |
|
"loss": 0.0065, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.360111770113659e-06, |
|
"loss": 0.0038, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.3195228291904271e-06, |
|
"loss": 0.0084, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 1.279505918561923e-06, |
|
"loss": 0.0068, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 1.240063675281934e-06, |
|
"loss": 0.0193, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 1.2011986985345313e-06, |
|
"loss": 0.0093, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 1.1629135494628097e-06, |
|
"loss": 0.0127, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.1252107510000843e-06, |
|
"loss": 0.0112, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 1.0880927877036608e-06, |
|
"loss": 0.0128, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 1.051562105591082e-06, |
|
"loss": 0.0144, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 1.0156211119789582e-06, |
|
"loss": 0.0027, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 9.802721753243193e-07, |
|
"loss": 0.003, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 9.455176250685338e-07, |
|
"loss": 0.0044, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 9.113597514838135e-07, |
|
"loss": 0.0108, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 8.778008055222809e-07, |
|
"loss": 0.0014, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 8.448429986676298e-07, |
|
"loss": 0.0029, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 8.124885027894014e-07, |
|
"loss": 0.0044, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 7.807394499998578e-07, |
|
"loss": 0.0037, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 7.495979325134806e-07, |
|
"loss": 0.004, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 7.19066002509089e-07, |
|
"loss": 0.0028, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 6.891456719946188e-07, |
|
"loss": 0.004, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 6.598389126745209e-07, |
|
"loss": 0.0025, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 6.311476558198337e-07, |
|
"loss": 0.0106, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 6.030737921409169e-07, |
|
"loss": 0.0023, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 5.756191716628556e-07, |
|
"loss": 0.0025, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 5.487856036035488e-07, |
|
"loss": 0.0025, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 5.225748562544741e-07, |
|
"loss": 0.0021, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 4.969886568641757e-07, |
|
"loss": 0.0016, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 4.7202869152443096e-07, |
|
"loss": 0.0103, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 4.4769660505914136e-07, |
|
"loss": 0.0015, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 4.2399400091594154e-07, |
|
"loss": 0.0015, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 4.0092244106053393e-07, |
|
"loss": 0.0049, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 3.7848344587376297e-07, |
|
"loss": 0.0031, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 3.566784940514145e-07, |
|
"loss": 0.0028, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 3.355090225067792e-07, |
|
"loss": 0.0023, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 3.1497642627596247e-07, |
|
"loss": 0.0022, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 2.9508205842594727e-07, |
|
"loss": 0.0014, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 2.758272299654341e-07, |
|
"loss": 0.014, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 2.572132097584523e-07, |
|
"loss": 0.0018, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 2.392412244407294e-07, |
|
"loss": 0.0013, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 2.2191245833886988e-07, |
|
"loss": 0.0055, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 2.052280533923101e-07, |
|
"loss": 0.0022, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 1.8918910907805733e-07, |
|
"loss": 0.0039, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.737966823382442e-07, |
|
"loss": 0.0012, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 1.5905178751047135e-07, |
|
"loss": 0.0032, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 1.4495539626097289e-07, |
|
"loss": 0.0026, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.3150843752057442e-07, |
|
"loss": 0.0017, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 1.1871179742348416e-07, |
|
"loss": 0.0059, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 1.0656631924889749e-07, |
|
"loss": 0.0023, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 9.507280336542912e-08, |
|
"loss": 0.0049, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 8.423200717835978e-08, |
|
"loss": 0.0133, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 7.404464507973608e-08, |
|
"loss": 0.0018, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 6.451138840128601e-08, |
|
"loss": 0.0014, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 5.5632865370183196e-08, |
|
"loss": 0.0033, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 4.740966106764222e-08, |
|
"loss": 0.0023, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 3.984231739036948e-08, |
|
"loss": 0.0029, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 3.293133301485063e-08, |
|
"loss": 0.0071, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 2.667716336448356e-08, |
|
"loss": 0.0015, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 2.1080220579573485e-08, |
|
"loss": 0.0034, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 1.614087349017246e-08, |
|
"loss": 0.0016, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 1.1859447591769934e-08, |
|
"loss": 0.0037, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 8.236225023844357e-09, |
|
"loss": 0.0051, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 5.271444551276928e-09, |
|
"loss": 0.0018, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 2.9653015486064143e-09, |
|
"loss": 0.0055, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 1.3179479871638923e-09, |
|
"loss": 0.002, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 3.2949242505408987e-10, |
|
"loss": 0.003, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.002, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"step": 399, |
|
"total_flos": 9464738217984.0, |
|
"train_loss": 0.08329384368762635, |
|
"train_runtime": 745.7277, |
|
"train_samples_per_second": 34.074, |
|
"train_steps_per_second": 0.535 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 399, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 7, |
|
"save_steps": 800, |
|
"total_flos": 9464738217984.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|