|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"global_step": 396, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 1.0912, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.1973, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5e-06, |
|
"loss": 1.8214, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 2.1457, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.333333333333334e-06, |
|
"loss": 2.2347, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1e-05, |
|
"loss": 1.4613, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.1666666666666668e-05, |
|
"loss": 1.9206, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 1.7732, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 1.8977, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.1657, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8333333333333333e-05, |
|
"loss": 0.9736, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 2e-05, |
|
"loss": 1.2143, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9999665339174013e-05, |
|
"loss": 0.9434, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9998661379095622e-05, |
|
"loss": 0.9365, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9996988186962044e-05, |
|
"loss": 1.0491, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9994645874763657e-05, |
|
"loss": 1.3624, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9991634599276488e-05, |
|
"loss": 0.9719, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9987954562051724e-05, |
|
"loss": 1.0088, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9983606009402224e-05, |
|
"loss": 0.8494, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9978589232386036e-05, |
|
"loss": 1.4835, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9972904566786903e-05, |
|
"loss": 0.4601, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9966552393091804e-05, |
|
"loss": 1.2571, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.995953313646548e-05, |
|
"loss": 0.6248, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.995184726672197e-05, |
|
"loss": 0.4497, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9943495298293182e-05, |
|
"loss": 0.7671, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9934477790194445e-05, |
|
"loss": 0.8551, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9924795345987103e-05, |
|
"loss": 0.9292, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9914448613738107e-05, |
|
"loss": 0.9321, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.990343828597665e-05, |
|
"loss": 0.9564, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.989176509964781e-05, |
|
"loss": 0.7888, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9879429836063228e-05, |
|
"loss": 0.3541, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.9866433320848793e-05, |
|
"loss": 0.7242, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9852776423889414e-05, |
|
"loss": 0.9379, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9838460059270775e-05, |
|
"loss": 1.5263, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.982348518521816e-05, |
|
"loss": 1.241, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9807852804032306e-05, |
|
"loss": 0.585, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.979156396202234e-05, |
|
"loss": 0.4718, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.977461974943572e-05, |
|
"loss": 0.8916, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9757021300385288e-05, |
|
"loss": 0.7767, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9738769792773338e-05, |
|
"loss": 0.7323, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.9719866448212794e-05, |
|
"loss": 0.8039, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.9700312531945444e-05, |
|
"loss": 0.9728, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9680109352757228e-05, |
|
"loss": 1.4139, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9659258262890683e-05, |
|
"loss": 0.5891, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.96377606579544e-05, |
|
"loss": 0.8547, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.9615617976829622e-05, |
|
"loss": 0.6453, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.959283170157394e-05, |
|
"loss": 0.9798, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.956940335732209e-05, |
|
"loss": 0.8966, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9545334512183886e-05, |
|
"loss": 0.5067, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9520626777139243e-05, |
|
"loss": 1.1425, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.949528180593037e-05, |
|
"loss": 1.3068, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.946930129495106e-05, |
|
"loss": 0.7687, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.944268698313317e-05, |
|
"loss": 0.9368, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.941544065183021e-05, |
|
"loss": 0.6697, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.9387564124698133e-05, |
|
"loss": 0.6488, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.935905926757326e-05, |
|
"loss": 0.9918, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.932992798834739e-05, |
|
"loss": 0.7319, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.930017223684012e-05, |
|
"loss": 0.6621, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.926979400466833e-05, |
|
"loss": 0.6273, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.9238795325112867e-05, |
|
"loss": 0.8412, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9207178272982477e-05, |
|
"loss": 0.7979, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9174944964474914e-05, |
|
"loss": 1.0536, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.914209755703531e-05, |
|
"loss": 0.3799, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.910863824921176e-05, |
|
"loss": 1.4296, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.9074569280508188e-05, |
|
"loss": 0.637, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.9039892931234434e-05, |
|
"loss": 1.3104, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.900461152235364e-05, |
|
"loss": 0.7303, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8968727415326885e-05, |
|
"loss": 0.3171, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8932243011955154e-05, |
|
"loss": 0.3529, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.8895160754218562e-05, |
|
"loss": 1.5304, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.885748312411291e-05, |
|
"loss": 1.1709, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.881921264348355e-05, |
|
"loss": 1.4026, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8780351873856627e-05, |
|
"loss": 0.5411, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.874090341626759e-05, |
|
"loss": 0.3518, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.8700869911087115e-05, |
|
"loss": 1.2801, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.866025403784439e-05, |
|
"loss": 1.1898, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.8619058515047746e-05, |
|
"loss": 1.1509, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8577286100002723e-05, |
|
"loss": 1.0352, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8534939588627506e-05, |
|
"loss": 0.8594, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.849202181526579e-05, |
|
"loss": 0.4596, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8448535652497073e-05, |
|
"loss": 0.7898, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.840448401094438e-05, |
|
"loss": 0.5138, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.835986983907947e-05, |
|
"loss": 1.443, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8314696123025456e-05, |
|
"loss": 0.321, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8268965886356965e-05, |
|
"loss": 0.4334, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.822268218989775e-05, |
|
"loss": 1.2314, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.817584813151584e-05, |
|
"loss": 0.667, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8128466845916156e-05, |
|
"loss": 0.399, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8080541504430732e-05, |
|
"loss": 1.1662, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.803207531480645e-05, |
|
"loss": 0.7352, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.7983071520990316e-05, |
|
"loss": 0.7251, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.7933533402912354e-05, |
|
"loss": 0.5131, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.7883464276266064e-05, |
|
"loss": 0.9221, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.7832867492286506e-05, |
|
"loss": 0.6758, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.778174643752598e-05, |
|
"loss": 0.6372, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.773010453362737e-05, |
|
"loss": 0.4088, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7677945237095122e-05, |
|
"loss": 0.8338, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.7625272039063884e-05, |
|
"loss": 0.5329, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.7572088465064847e-05, |
|
"loss": 0.7356, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.7518398074789776e-05, |
|
"loss": 0.5893, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.746420446185274e-05, |
|
"loss": 1.0192, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.7409511253549592e-05, |
|
"loss": 0.6132, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.735432211061519e-05, |
|
"loss": 0.5129, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.7298640726978357e-05, |
|
"loss": 1.3157, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.7242470829514674e-05, |
|
"loss": 1.1499, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.718581617779698e-05, |
|
"loss": 0.6059, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.71286805638438e-05, |
|
"loss": 0.7768, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.7071067811865477e-05, |
|
"loss": 0.3944, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.7012981778008245e-05, |
|
"loss": 0.4616, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.6954426350096118e-05, |
|
"loss": 0.392, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.689540544737067e-05, |
|
"loss": 0.4423, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.6835923020228714e-05, |
|
"loss": 0.5843, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.6775983049957888e-05, |
|
"loss": 0.4968, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.6715589548470187e-05, |
|
"loss": 0.6635, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.6654746558033424e-05, |
|
"loss": 0.5345, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.659345815100069e-05, |
|
"loss": 0.3274, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.6531728429537766e-05, |
|
"loss": 0.758, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.6469561525348576e-05, |
|
"loss": 0.6526, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.640696159939861e-05, |
|
"loss": 0.5578, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.6343932841636455e-05, |
|
"loss": 0.9087, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6280479470713343e-05, |
|
"loss": 0.3794, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6216605733700776e-05, |
|
"loss": 0.6467, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.615231590580627e-05, |
|
"loss": 0.5239, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.608761429008721e-05, |
|
"loss": 0.4212, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.6022505217162824e-05, |
|
"loss": 0.3344, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5956993044924334e-05, |
|
"loss": 1.239, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.5891082158243283e-05, |
|
"loss": 0.8274, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.5824776968678024e-05, |
|
"loss": 0.5102, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.5758081914178457e-05, |
|
"loss": 0.8099, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.5691001458788984e-05, |
|
"loss": 0.9807, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.5623540092349735e-05, |
|
"loss": 0.6831, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.5555702330196024e-05, |
|
"loss": 0.3847, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.548749271285616e-05, |
|
"loss": 0.9832, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.5418915805747518e-05, |
|
"loss": 0.4993, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.5349976198870974e-05, |
|
"loss": 0.2899, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.528067850650368e-05, |
|
"loss": 0.52, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.5211027366890237e-05, |
|
"loss": 0.6157, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.5141027441932217e-05, |
|
"loss": 0.2785, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.5070683416876172e-05, |
|
"loss": 0.4745, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.2615, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.4928981922297842e-05, |
|
"loss": 0.5145, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.4857633937163402e-05, |
|
"loss": 0.347, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.4785960820071812e-05, |
|
"loss": 0.2387, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.4713967368259981e-05, |
|
"loss": 0.7635, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.4641658400405516e-05, |
|
"loss": 0.3601, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4569038756304209e-05, |
|
"loss": 0.3369, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4496113296546068e-05, |
|
"loss": 0.3824, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4422886902190014e-05, |
|
"loss": 0.287, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.4349364474437172e-05, |
|
"loss": 0.9414, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.4275550934302822e-05, |
|
"loss": 0.4918, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.4201451222287026e-05, |
|
"loss": 0.5476, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.4127070298043949e-05, |
|
"loss": 0.2683, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.4052413140049898e-05, |
|
"loss": 0.2498, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.3977484745270112e-05, |
|
"loss": 0.4185, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.3902290128824282e-05, |
|
"loss": 0.3995, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.3826834323650899e-05, |
|
"loss": 0.3908, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.375112238017038e-05, |
|
"loss": 0.2061, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.3675159365947038e-05, |
|
"loss": 0.3934, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.3598950365349884e-05, |
|
"loss": 0.3242, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.3522500479212337e-05, |
|
"loss": 0.4355, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.3445814824490806e-05, |
|
"loss": 0.4813, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.3368898533922202e-05, |
|
"loss": 0.4271, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.329175675568039e-05, |
|
"loss": 0.2628, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.3214394653031616e-05, |
|
"loss": 0.2714, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.3136817403988918e-05, |
|
"loss": 0.3173, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.3059030200965536e-05, |
|
"loss": 0.2928, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.2981038250427402e-05, |
|
"loss": 0.3756, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.2902846772544625e-05, |
|
"loss": 0.25, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.2824461000842126e-05, |
|
"loss": 0.389, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.2745886181849325e-05, |
|
"loss": 0.3511, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.2667127574748985e-05, |
|
"loss": 0.6542, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.2588190451025209e-05, |
|
"loss": 0.1331, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.2509080094110604e-05, |
|
"loss": 0.4515, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.242980179903264e-05, |
|
"loss": 0.4946, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.2350360872059268e-05, |
|
"loss": 0.453, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.2270762630343734e-05, |
|
"loss": 0.3207, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.2191012401568698e-05, |
|
"loss": 0.2817, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.2111115523589651e-05, |
|
"loss": 0.5249, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.203107734407763e-05, |
|
"loss": 0.616, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1950903220161286e-05, |
|
"loss": 0.3736, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1870598518068321e-05, |
|
"loss": 0.3653, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1790168612766331e-05, |
|
"loss": 0.3652, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1709618887603013e-05, |
|
"loss": 0.4013, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.162895473394589e-05, |
|
"loss": 0.3294, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1548181550821411e-05, |
|
"loss": 0.6704, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.1467304744553618e-05, |
|
"loss": 0.3813, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.1386329728402269e-05, |
|
"loss": 0.3289, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.130526192220052e-05, |
|
"loss": 0.492, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.1224106751992164e-05, |
|
"loss": 0.6376, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.1142869649668467e-05, |
|
"loss": 0.4514, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.1061556052604579e-05, |
|
"loss": 0.1564, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.098017140329561e-05, |
|
"loss": 0.2852, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.089872114899235e-05, |
|
"loss": 0.3373, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0817210741336684e-05, |
|
"loss": 0.2011, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0735645635996676e-05, |
|
"loss": 0.6372, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0654031292301432e-05, |
|
"loss": 0.1721, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.057237317287569e-05, |
|
"loss": 0.314, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0490676743274181e-05, |
|
"loss": 0.3923, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0408947471615836e-05, |
|
"loss": 0.4527, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0327190828217763e-05, |
|
"loss": 0.2541, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0245412285229124e-05, |
|
"loss": 0.1693, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0163617316264869e-05, |
|
"loss": 0.7895, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0081811396039374e-05, |
|
"loss": 0.6897, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1e-05, |
|
"loss": 0.3858, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.918188603960632e-06, |
|
"loss": 0.495, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.836382683735133e-06, |
|
"loss": 0.2469, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.75458771477088e-06, |
|
"loss": 0.5765, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.67280917178224e-06, |
|
"loss": 0.4859, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.591052528384168e-06, |
|
"loss": 0.2056, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.50932325672582e-06, |
|
"loss": 0.5528, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.427626827124316e-06, |
|
"loss": 0.2854, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.34596870769857e-06, |
|
"loss": 0.2586, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.264354364003327e-06, |
|
"loss": 0.2958, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.182789258663321e-06, |
|
"loss": 0.2754, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.101278851007651e-06, |
|
"loss": 0.1672, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.019828596704394e-06, |
|
"loss": 0.474, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.938443947395428e-06, |
|
"loss": 0.7002, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.857130350331535e-06, |
|
"loss": 0.487, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.77589324800784e-06, |
|
"loss": 0.5058, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.694738077799487e-06, |
|
"loss": 0.3104, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.613670271597733e-06, |
|
"loss": 0.2636, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.532695255446384e-06, |
|
"loss": 0.2949, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.451818449178592e-06, |
|
"loss": 0.4606, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.371045266054114e-06, |
|
"loss": 0.2796, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.290381112396989e-06, |
|
"loss": 0.3996, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.209831387233675e-06, |
|
"loss": 0.7789, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.12940148193168e-06, |
|
"loss": 0.3954, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.04909677983872e-06, |
|
"loss": 0.293, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 7.968922655922375e-06, |
|
"loss": 0.2245, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 7.888884476410348e-06, |
|
"loss": 0.3602, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.808987598431303e-06, |
|
"loss": 0.1273, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 7.72923736965627e-06, |
|
"loss": 0.2048, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.649639127940734e-06, |
|
"loss": 0.4198, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.570198200967363e-06, |
|
"loss": 0.536, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 7.490919905889403e-06, |
|
"loss": 0.6785, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.411809548974792e-06, |
|
"loss": 0.5334, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.332872425251017e-06, |
|
"loss": 0.6093, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.25411381815068e-06, |
|
"loss": 0.2398, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.175538999157876e-06, |
|
"loss": 0.2231, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.097153227455379e-06, |
|
"loss": 0.2658, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.018961749572604e-06, |
|
"loss": 0.3377, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 6.940969799034465e-06, |
|
"loss": 0.2729, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.8631825960110866e-06, |
|
"loss": 0.2559, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.785605346968387e-06, |
|
"loss": 0.5784, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.708243244319611e-06, |
|
"loss": 0.2154, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.631101466077801e-06, |
|
"loss": 0.3358, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.554185175509196e-06, |
|
"loss": 0.8953, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.4774995207876654e-06, |
|
"loss": 0.3955, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.401049634650119e-06, |
|
"loss": 0.554, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.3248406340529665e-06, |
|
"loss": 0.2883, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.2488776198296195e-06, |
|
"loss": 0.21, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.173165676349103e-06, |
|
"loss": 0.4709, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.097709871175723e-06, |
|
"loss": 0.4529, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.02251525472989e-06, |
|
"loss": 0.7958, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.947586859950103e-06, |
|
"loss": 0.5224, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.872929701956054e-06, |
|
"loss": 0.4974, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.798548777712977e-06, |
|
"loss": 0.3196, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.724449065697182e-06, |
|
"loss": 0.2064, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.650635525562829e-06, |
|
"loss": 0.6573, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.5771130978099896e-06, |
|
"loss": 0.308, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.503886703453933e-06, |
|
"loss": 0.4254, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.430961243695794e-06, |
|
"loss": 0.2067, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.358341599594483e-06, |
|
"loss": 0.2164, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.286032631740023e-06, |
|
"loss": 0.71, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.214039179928194e-06, |
|
"loss": 0.4526, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.142366062836599e-06, |
|
"loss": 0.2104, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.071018077702161e-06, |
|
"loss": 0.2086, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.000000000000003e-06, |
|
"loss": 0.1537, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.9293165831238295e-06, |
|
"loss": 0.0942, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.858972558067784e-06, |
|
"loss": 0.1073, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.788972633109768e-06, |
|
"loss": 0.2291, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.7193214934963204e-06, |
|
"loss": 0.0834, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.65002380112903e-06, |
|
"loss": 0.1469, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.581084194252486e-06, |
|
"loss": 0.2772, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.5125072871438426e-06, |
|
"loss": 0.1829, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.444297669803981e-06, |
|
"loss": 0.1034, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.3764599076502705e-06, |
|
"loss": 0.1104, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.308998541211016e-06, |
|
"loss": 0.1409, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.241918085821547e-06, |
|
"loss": 0.1215, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.17522303132198e-06, |
|
"loss": 0.1462, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.108917841756717e-06, |
|
"loss": 0.2358, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.043006955075667e-06, |
|
"loss": 0.2254, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 3.977494782837182e-06, |
|
"loss": 0.1637, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.912385709912794e-06, |
|
"loss": 0.1373, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.847684094193733e-06, |
|
"loss": 0.2059, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.7833942662992286e-06, |
|
"loss": 0.1392, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.7195205292866588e-06, |
|
"loss": 0.1367, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.6560671583635467e-06, |
|
"loss": 0.1172, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.593038400601395e-06, |
|
"loss": 0.0897, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.5304384746514273e-06, |
|
"loss": 0.2328, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.468271570462235e-06, |
|
"loss": 0.2492, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.4065418489993118e-06, |
|
"loss": 0.1344, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.3452534419665793e-06, |
|
"loss": 0.1469, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.284410451529816e-06, |
|
"loss": 0.1339, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.224016950042114e-06, |
|
"loss": 0.0994, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.1640769797712865e-06, |
|
"loss": 0.2619, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.1045945526293307e-06, |
|
"loss": 0.0797, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.0455736499038847e-06, |
|
"loss": 0.181, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.9870182219917564e-06, |
|
"loss": 0.2399, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.9289321881345257e-06, |
|
"loss": 0.1349, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.8713194361562036e-06, |
|
"loss": 0.1067, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.8141838222030195e-06, |
|
"loss": 0.1468, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.7575291704853325e-06, |
|
"loss": 0.1159, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.7013592730216464e-06, |
|
"loss": 0.1167, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.6456778893848144e-06, |
|
"loss": 0.2314, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.5904887464504115e-06, |
|
"loss": 0.2353, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.5357955381472622e-06, |
|
"loss": 0.1086, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.4816019252102274e-06, |
|
"loss": 0.352, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.4279115349351546e-06, |
|
"loss": 0.2142, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.3747279609361197e-06, |
|
"loss": 0.1142, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.3220547629048796e-06, |
|
"loss": 0.1481, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.26989546637263e-06, |
|
"loss": 0.0982, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2182535624740233e-06, |
|
"loss": 0.149, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.1671325077134963e-06, |
|
"loss": 0.1391, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.116535723733938e-06, |
|
"loss": 0.1366, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.0664665970876496e-06, |
|
"loss": 0.1159, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.0169284790096856e-06, |
|
"loss": 0.161, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 1.967924685193552e-06, |
|
"loss": 0.1869, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.919458495569271e-06, |
|
"loss": 0.2142, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.8715331540838488e-06, |
|
"loss": 0.4326, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.8241518684841642e-06, |
|
"loss": 0.1299, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.7773178101022514e-06, |
|
"loss": 0.1008, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.7310341136430386e-06, |
|
"loss": 0.1049, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.6853038769745466e-06, |
|
"loss": 0.205, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.6401301609205333e-06, |
|
"loss": 0.2219, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 1.5955159890556182e-06, |
|
"loss": 0.1405, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.551464347502929e-06, |
|
"loss": 0.1053, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.5079781847342122e-06, |
|
"loss": 0.1873, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.4650604113724953e-06, |
|
"loss": 0.26, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.4227138999972801e-06, |
|
"loss": 0.1335, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.3809414849522585e-06, |
|
"loss": 0.1168, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.339745962155613e-06, |
|
"loss": 0.1731, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.2991300889128867e-06, |
|
"loss": 0.2434, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.2590965837324132e-06, |
|
"loss": 0.1344, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.2196481261433735e-06, |
|
"loss": 0.1206, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.1807873565164507e-06, |
|
"loss": 0.1283, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.1425168758870963e-06, |
|
"loss": 0.1008, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.1048392457814406e-06, |
|
"loss": 0.1143, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.0677569880448479e-06, |
|
"loss": 0.1137, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.0312725846731174e-06, |
|
"loss": 0.137, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 9.953884776463652e-07, |
|
"loss": 0.0847, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 9.601070687655667e-07, |
|
"loss": 0.1254, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 9.254307194918144e-07, |
|
"loss": 0.1035, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 8.91361750788241e-07, |
|
"loss": 0.126, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 8.579024429646932e-07, |
|
"loss": 0.1385, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 8.250550355250875e-07, |
|
"loss": 0.3491, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.928217270175231e-07, |
|
"loss": 0.1928, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.612046748871327e-07, |
|
"loss": 0.121, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.30205995331672e-07, |
|
"loss": 0.0909, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 6.998277631598793e-07, |
|
"loss": 0.1568, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 6.700720116526116e-07, |
|
"loss": 0.1316, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.409407324267448e-07, |
|
"loss": 0.3285, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.124358753018688e-07, |
|
"loss": 0.1659, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.845593481697931e-07, |
|
"loss": 0.1013, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.57313016866835e-07, |
|
"loss": 0.2537, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.306987050489442e-07, |
|
"loss": 0.1518, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.047181940696333e-07, |
|
"loss": 0.0987, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.793732228607573e-07, |
|
"loss": 0.2286, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.54665487816115e-07, |
|
"loss": 0.1637, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.305966426779118e-07, |
|
"loss": 0.1226, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.0716829842606385e-07, |
|
"loss": 0.1352, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.8438202317037987e-07, |
|
"loss": 0.1597, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.6223934204560165e-07, |
|
"loss": 0.2184, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.4074173710931804e-07, |
|
"loss": 0.174, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.198906472427732e-07, |
|
"loss": 0.1711, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.996874680545603e-07, |
|
"loss": 0.2215, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.8013355178720613e-07, |
|
"loss": 0.1262, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.612302072266637e-07, |
|
"loss": 0.1392, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.4297869961471544e-07, |
|
"loss": 0.1611, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.2538025056428216e-07, |
|
"loss": 0.1515, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 2.0843603797766287e-07, |
|
"loss": 0.1478, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.921471959676957e-07, |
|
"loss": 0.1685, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.7651481478184297e-07, |
|
"loss": 0.1206, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.615399407292251e-07, |
|
"loss": 0.0952, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.472235761105878e-07, |
|
"loss": 0.1378, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.3356667915121025e-07, |
|
"loss": 0.1138, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.2057016393677623e-07, |
|
"loss": 0.0912, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.0823490035218986e-07, |
|
"loss": 0.1429, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 9.656171402335213e-08, |
|
"loss": 0.1181, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.555138626189619e-08, |
|
"loss": 0.1142, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.520465401290033e-08, |
|
"loss": 0.2712, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 6.552220980555635e-08, |
|
"loss": 0.1408, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 5.650470170681876e-08, |
|
"loss": 0.1736, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.815273327803183e-08, |
|
"loss": 0.1527, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.0466863534522895e-08, |
|
"loss": 0.1103, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.3447606908196815e-08, |
|
"loss": 0.1522, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.7095433213097933e-08, |
|
"loss": 0.1824, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.1410767613965212e-08, |
|
"loss": 0.1105, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.639399059777502e-08, |
|
"loss": 0.0768, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.2045437948275952e-08, |
|
"loss": 0.0796, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 8.365400723512329e-09, |
|
"loss": 0.0868, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 5.354125236343155e-09, |
|
"loss": 0.1575, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.0118130379575005e-09, |
|
"loss": 0.1257, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.3386209043819708e-09, |
|
"loss": 0.1137, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.346608259890971e-10, |
|
"loss": 0.1299, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.1945, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 396, |
|
"total_flos": 4.02857154245034e+17, |
|
"train_loss": 0.47617627023672215, |
|
"train_runtime": 8141.7482, |
|
"train_samples_per_second": 0.776, |
|
"train_steps_per_second": 0.049 |
|
} |
|
], |
|
"max_steps": 396, |
|
"num_train_epochs": 3, |
|
"total_flos": 4.02857154245034e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|