{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9998900252941825, "eval_steps": 500, "global_step": 9092, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00021994941163532388, "grad_norm": 0.7960259914398193, "learning_rate": 8e-05, "loss": 2.2745, "step": 1 }, { "epoch": 0.00043989882327064776, "grad_norm": 0.3960127830505371, "learning_rate": 8e-05, "loss": 1.9155, "step": 2 }, { "epoch": 0.0006598482349059716, "grad_norm": 0.3869185745716095, "learning_rate": 8e-05, "loss": 1.8754, "step": 3 }, { "epoch": 0.0008797976465412955, "grad_norm": 0.33234134316444397, "learning_rate": 8e-05, "loss": 1.9126, "step": 4 }, { "epoch": 0.0010997470581766194, "grad_norm": 0.3670472502708435, "learning_rate": 8e-05, "loss": 1.9912, "step": 5 }, { "epoch": 0.0013196964698119432, "grad_norm": 0.32942938804626465, "learning_rate": 8e-05, "loss": 1.8746, "step": 6 }, { "epoch": 0.001539645881447267, "grad_norm": 0.3588086664676666, "learning_rate": 8e-05, "loss": 1.9545, "step": 7 }, { "epoch": 0.001759595293082591, "grad_norm": 0.33002492785453796, "learning_rate": 8e-05, "loss": 1.8781, "step": 8 }, { "epoch": 0.001979544704717915, "grad_norm": 0.3024381101131439, "learning_rate": 8e-05, "loss": 1.8859, "step": 9 }, { "epoch": 0.002199494116353239, "grad_norm": 0.3224199712276459, "learning_rate": 8e-05, "loss": 1.829, "step": 10 }, { "epoch": 0.0024194435279885624, "grad_norm": 0.31481102108955383, "learning_rate": 8e-05, "loss": 1.817, "step": 11 }, { "epoch": 0.0026393929396238865, "grad_norm": 0.3078259825706482, "learning_rate": 8e-05, "loss": 1.9631, "step": 12 }, { "epoch": 0.0028593423512592105, "grad_norm": 0.3141743540763855, "learning_rate": 8e-05, "loss": 1.8879, "step": 13 }, { "epoch": 0.003079291762894534, "grad_norm": 0.29574745893478394, "learning_rate": 8e-05, "loss": 1.799, "step": 14 }, { "epoch": 0.003299241174529858, "grad_norm": 0.3095031976699829, "learning_rate": 8e-05, "loss": 1.8741, "step": 15 }, { "epoch": 0.003519190586165182, "grad_norm": 0.28804337978363037, "learning_rate": 8e-05, "loss": 1.9931, "step": 16 }, { "epoch": 0.0037391399978005057, "grad_norm": 0.25137585401535034, "learning_rate": 8e-05, "loss": 1.6762, "step": 17 }, { "epoch": 0.00395908940943583, "grad_norm": 0.28565698862075806, "learning_rate": 8e-05, "loss": 1.8489, "step": 18 }, { "epoch": 0.004179038821071153, "grad_norm": 0.2877500355243683, "learning_rate": 8e-05, "loss": 1.8871, "step": 19 }, { "epoch": 0.004398988232706478, "grad_norm": 0.28803154826164246, "learning_rate": 8e-05, "loss": 1.6956, "step": 20 }, { "epoch": 0.004618937644341801, "grad_norm": 0.32161301374435425, "learning_rate": 8e-05, "loss": 1.8243, "step": 21 }, { "epoch": 0.004838887055977125, "grad_norm": 0.2959391176700592, "learning_rate": 8e-05, "loss": 1.8991, "step": 22 }, { "epoch": 0.005058836467612449, "grad_norm": 0.3021189868450165, "learning_rate": 8e-05, "loss": 1.9975, "step": 23 }, { "epoch": 0.005278785879247773, "grad_norm": 0.2793104946613312, "learning_rate": 8e-05, "loss": 1.8792, "step": 24 }, { "epoch": 0.0054987352908830965, "grad_norm": 0.2658381760120392, "learning_rate": 8e-05, "loss": 1.6467, "step": 25 }, { "epoch": 0.005718684702518421, "grad_norm": 0.2793010175228119, "learning_rate": 8e-05, "loss": 1.7479, "step": 26 }, { "epoch": 0.0059386341141537445, "grad_norm": 0.2800044119358063, "learning_rate": 8e-05, "loss": 1.7885, "step": 27 }, { "epoch": 0.006158583525789068, "grad_norm": 0.2864585220813751, "learning_rate": 8e-05, "loss": 1.9257, "step": 28 }, { "epoch": 0.006378532937424393, "grad_norm": 0.301496684551239, "learning_rate": 8e-05, "loss": 1.8586, "step": 29 }, { "epoch": 0.006598482349059716, "grad_norm": 0.2858293354511261, "learning_rate": 8e-05, "loss": 1.8541, "step": 30 }, { "epoch": 0.00681843176069504, "grad_norm": 0.31271278858184814, "learning_rate": 8e-05, "loss": 1.8774, "step": 31 }, { "epoch": 0.007038381172330364, "grad_norm": 0.30428266525268555, "learning_rate": 8e-05, "loss": 1.8464, "step": 32 }, { "epoch": 0.007258330583965688, "grad_norm": 0.26637139916419983, "learning_rate": 8e-05, "loss": 1.7896, "step": 33 }, { "epoch": 0.007478279995601011, "grad_norm": 0.2802716791629791, "learning_rate": 8e-05, "loss": 1.9534, "step": 34 }, { "epoch": 0.007698229407236336, "grad_norm": 0.35580113530158997, "learning_rate": 8e-05, "loss": 1.8236, "step": 35 }, { "epoch": 0.00791817881887166, "grad_norm": 0.2794848382472992, "learning_rate": 8e-05, "loss": 1.8089, "step": 36 }, { "epoch": 0.008138128230506983, "grad_norm": 0.27942711114883423, "learning_rate": 8e-05, "loss": 1.7725, "step": 37 }, { "epoch": 0.008358077642142307, "grad_norm": 0.2882610857486725, "learning_rate": 8e-05, "loss": 1.8279, "step": 38 }, { "epoch": 0.008578027053777632, "grad_norm": 0.29375842213630676, "learning_rate": 8e-05, "loss": 2.0123, "step": 39 }, { "epoch": 0.008797976465412955, "grad_norm": 0.26120567321777344, "learning_rate": 8e-05, "loss": 1.6731, "step": 40 }, { "epoch": 0.009017925877048279, "grad_norm": 0.25272971391677856, "learning_rate": 8e-05, "loss": 1.5723, "step": 41 }, { "epoch": 0.009237875288683603, "grad_norm": 0.30548569560050964, "learning_rate": 8e-05, "loss": 1.9269, "step": 42 }, { "epoch": 0.009457824700318926, "grad_norm": 0.2714739441871643, "learning_rate": 8e-05, "loss": 1.6715, "step": 43 }, { "epoch": 0.00967777411195425, "grad_norm": 0.3086313009262085, "learning_rate": 8e-05, "loss": 1.8903, "step": 44 }, { "epoch": 0.009897723523589575, "grad_norm": 0.28676554560661316, "learning_rate": 8e-05, "loss": 1.8257, "step": 45 }, { "epoch": 0.010117672935224899, "grad_norm": 0.2898331880569458, "learning_rate": 8e-05, "loss": 1.822, "step": 46 }, { "epoch": 0.010337622346860222, "grad_norm": 0.2887754440307617, "learning_rate": 8e-05, "loss": 1.7629, "step": 47 }, { "epoch": 0.010557571758495546, "grad_norm": 0.28026437759399414, "learning_rate": 8e-05, "loss": 1.8874, "step": 48 }, { "epoch": 0.01077752117013087, "grad_norm": 0.29256439208984375, "learning_rate": 8e-05, "loss": 1.9169, "step": 49 }, { "epoch": 0.010997470581766193, "grad_norm": 0.29388460516929626, "learning_rate": 8e-05, "loss": 1.8341, "step": 50 }, { "epoch": 0.011217419993401518, "grad_norm": 0.29456326365470886, "learning_rate": 8e-05, "loss": 1.7088, "step": 51 }, { "epoch": 0.011437369405036842, "grad_norm": 0.2810533046722412, "learning_rate": 8e-05, "loss": 1.8564, "step": 52 }, { "epoch": 0.011657318816672166, "grad_norm": 0.3049224019050598, "learning_rate": 8e-05, "loss": 2.0114, "step": 53 }, { "epoch": 0.011877268228307489, "grad_norm": 0.347817987203598, "learning_rate": 8e-05, "loss": 1.6936, "step": 54 }, { "epoch": 0.012097217639942813, "grad_norm": 0.28999242186546326, "learning_rate": 8e-05, "loss": 1.5852, "step": 55 }, { "epoch": 0.012317167051578136, "grad_norm": 0.32856103777885437, "learning_rate": 8e-05, "loss": 1.8249, "step": 56 }, { "epoch": 0.012537116463213462, "grad_norm": 0.3450610339641571, "learning_rate": 8e-05, "loss": 1.9309, "step": 57 }, { "epoch": 0.012757065874848785, "grad_norm": 0.27445971965789795, "learning_rate": 8e-05, "loss": 1.8153, "step": 58 }, { "epoch": 0.012977015286484109, "grad_norm": 0.28595077991485596, "learning_rate": 8e-05, "loss": 1.8061, "step": 59 }, { "epoch": 0.013196964698119432, "grad_norm": 0.2909082770347595, "learning_rate": 8e-05, "loss": 1.7793, "step": 60 }, { "epoch": 0.013416914109754756, "grad_norm": 0.28822049498558044, "learning_rate": 8e-05, "loss": 1.7218, "step": 61 }, { "epoch": 0.01363686352139008, "grad_norm": 0.29159948229789734, "learning_rate": 8e-05, "loss": 1.7985, "step": 62 }, { "epoch": 0.013856812933025405, "grad_norm": 0.29802417755126953, "learning_rate": 8e-05, "loss": 1.8903, "step": 63 }, { "epoch": 0.014076762344660728, "grad_norm": 0.29128944873809814, "learning_rate": 8e-05, "loss": 1.7928, "step": 64 }, { "epoch": 0.014296711756296052, "grad_norm": 0.3093227446079254, "learning_rate": 8e-05, "loss": 1.8409, "step": 65 }, { "epoch": 0.014516661167931376, "grad_norm": 0.2688956558704376, "learning_rate": 8e-05, "loss": 1.6136, "step": 66 }, { "epoch": 0.0147366105795667, "grad_norm": 0.316579133272171, "learning_rate": 8e-05, "loss": 1.8153, "step": 67 }, { "epoch": 0.014956559991202023, "grad_norm": 0.30234795808792114, "learning_rate": 8e-05, "loss": 1.7311, "step": 68 }, { "epoch": 0.015176509402837348, "grad_norm": 0.2790556848049164, "learning_rate": 8e-05, "loss": 1.7714, "step": 69 }, { "epoch": 0.015396458814472672, "grad_norm": 0.29012972116470337, "learning_rate": 8e-05, "loss": 1.7528, "step": 70 }, { "epoch": 0.015616408226107995, "grad_norm": 0.28507527709007263, "learning_rate": 8e-05, "loss": 1.6452, "step": 71 }, { "epoch": 0.01583635763774332, "grad_norm": 0.28862133622169495, "learning_rate": 8e-05, "loss": 1.7473, "step": 72 }, { "epoch": 0.016056307049378642, "grad_norm": 0.2726048231124878, "learning_rate": 8e-05, "loss": 1.7519, "step": 73 }, { "epoch": 0.016276256461013966, "grad_norm": 0.26808786392211914, "learning_rate": 8e-05, "loss": 1.6332, "step": 74 }, { "epoch": 0.01649620587264929, "grad_norm": 0.32144519686698914, "learning_rate": 8e-05, "loss": 1.7115, "step": 75 }, { "epoch": 0.016716155284284613, "grad_norm": 0.26930421590805054, "learning_rate": 8e-05, "loss": 1.7854, "step": 76 }, { "epoch": 0.016936104695919937, "grad_norm": 0.29462486505508423, "learning_rate": 8e-05, "loss": 1.6919, "step": 77 }, { "epoch": 0.017156054107555264, "grad_norm": 0.2780003249645233, "learning_rate": 8e-05, "loss": 1.6355, "step": 78 }, { "epoch": 0.017376003519190587, "grad_norm": 0.29219016432762146, "learning_rate": 8e-05, "loss": 1.883, "step": 79 }, { "epoch": 0.01759595293082591, "grad_norm": 0.2893241047859192, "learning_rate": 8e-05, "loss": 1.8548, "step": 80 }, { "epoch": 0.017815902342461234, "grad_norm": 0.283512145280838, "learning_rate": 8e-05, "loss": 1.79, "step": 81 }, { "epoch": 0.018035851754096558, "grad_norm": 0.2679024040699005, "learning_rate": 8e-05, "loss": 1.5866, "step": 82 }, { "epoch": 0.01825580116573188, "grad_norm": 0.2892123758792877, "learning_rate": 8e-05, "loss": 1.9033, "step": 83 }, { "epoch": 0.018475750577367205, "grad_norm": 0.2680201530456543, "learning_rate": 8e-05, "loss": 1.8557, "step": 84 }, { "epoch": 0.01869569998900253, "grad_norm": 0.30922645330429077, "learning_rate": 8e-05, "loss": 1.8885, "step": 85 }, { "epoch": 0.018915649400637852, "grad_norm": 0.2735271751880646, "learning_rate": 8e-05, "loss": 1.8765, "step": 86 }, { "epoch": 0.019135598812273176, "grad_norm": 0.28639712929725647, "learning_rate": 8e-05, "loss": 1.9429, "step": 87 }, { "epoch": 0.0193555482239085, "grad_norm": 0.28437235951423645, "learning_rate": 8e-05, "loss": 1.8405, "step": 88 }, { "epoch": 0.019575497635543827, "grad_norm": 0.276517778635025, "learning_rate": 8e-05, "loss": 1.7496, "step": 89 }, { "epoch": 0.01979544704717915, "grad_norm": 0.273404598236084, "learning_rate": 8e-05, "loss": 1.704, "step": 90 }, { "epoch": 0.020015396458814474, "grad_norm": 0.2707740366458893, "learning_rate": 8e-05, "loss": 1.8274, "step": 91 }, { "epoch": 0.020235345870449797, "grad_norm": 0.26880595088005066, "learning_rate": 8e-05, "loss": 1.7695, "step": 92 }, { "epoch": 0.02045529528208512, "grad_norm": 0.28712528944015503, "learning_rate": 8e-05, "loss": 1.9436, "step": 93 }, { "epoch": 0.020675244693720445, "grad_norm": 0.26633599400520325, "learning_rate": 8e-05, "loss": 1.7877, "step": 94 }, { "epoch": 0.020895194105355768, "grad_norm": 0.2843431532382965, "learning_rate": 8e-05, "loss": 1.8389, "step": 95 }, { "epoch": 0.02111514351699109, "grad_norm": 0.2597465515136719, "learning_rate": 8e-05, "loss": 1.7047, "step": 96 }, { "epoch": 0.021335092928626415, "grad_norm": 0.2804902493953705, "learning_rate": 8e-05, "loss": 1.9375, "step": 97 }, { "epoch": 0.02155504234026174, "grad_norm": 0.2825285792350769, "learning_rate": 8e-05, "loss": 1.8348, "step": 98 }, { "epoch": 0.021774991751897062, "grad_norm": 0.26459112763404846, "learning_rate": 8e-05, "loss": 1.7416, "step": 99 }, { "epoch": 0.021994941163532386, "grad_norm": 0.28523096442222595, "learning_rate": 8e-05, "loss": 1.9202, "step": 100 }, { "epoch": 0.022214890575167713, "grad_norm": 0.2679818570613861, "learning_rate": 8e-05, "loss": 1.6741, "step": 101 }, { "epoch": 0.022434839986803037, "grad_norm": 0.2798464894294739, "learning_rate": 8e-05, "loss": 1.6622, "step": 102 }, { "epoch": 0.02265478939843836, "grad_norm": 0.2826269567012787, "learning_rate": 8e-05, "loss": 1.7577, "step": 103 }, { "epoch": 0.022874738810073684, "grad_norm": 0.3859495222568512, "learning_rate": 8e-05, "loss": 1.9705, "step": 104 }, { "epoch": 0.023094688221709007, "grad_norm": 0.2766650319099426, "learning_rate": 8e-05, "loss": 1.7706, "step": 105 }, { "epoch": 0.02331463763334433, "grad_norm": 0.2804067134857178, "learning_rate": 8e-05, "loss": 1.8007, "step": 106 }, { "epoch": 0.023534587044979655, "grad_norm": 0.27818629145622253, "learning_rate": 8e-05, "loss": 1.7913, "step": 107 }, { "epoch": 0.023754536456614978, "grad_norm": 0.2697458267211914, "learning_rate": 8e-05, "loss": 1.8458, "step": 108 }, { "epoch": 0.023974485868250302, "grad_norm": 0.28805410861968994, "learning_rate": 8e-05, "loss": 1.7543, "step": 109 }, { "epoch": 0.024194435279885625, "grad_norm": 0.28452396392822266, "learning_rate": 8e-05, "loss": 1.8499, "step": 110 }, { "epoch": 0.02441438469152095, "grad_norm": 0.2837978005409241, "learning_rate": 8e-05, "loss": 1.797, "step": 111 }, { "epoch": 0.024634334103156273, "grad_norm": 0.2965853810310364, "learning_rate": 8e-05, "loss": 1.7988, "step": 112 }, { "epoch": 0.0248542835147916, "grad_norm": 0.28529393672943115, "learning_rate": 8e-05, "loss": 1.7886, "step": 113 }, { "epoch": 0.025074232926426923, "grad_norm": 0.285199910402298, "learning_rate": 8e-05, "loss": 1.9112, "step": 114 }, { "epoch": 0.025294182338062247, "grad_norm": 0.286316454410553, "learning_rate": 8e-05, "loss": 1.6735, "step": 115 }, { "epoch": 0.02551413174969757, "grad_norm": 0.2648874819278717, "learning_rate": 8e-05, "loss": 1.5333, "step": 116 }, { "epoch": 0.025734081161332894, "grad_norm": 0.2834017276763916, "learning_rate": 8e-05, "loss": 1.7524, "step": 117 }, { "epoch": 0.025954030572968217, "grad_norm": 0.27846938371658325, "learning_rate": 8e-05, "loss": 1.8448, "step": 118 }, { "epoch": 0.02617397998460354, "grad_norm": 0.3278025984764099, "learning_rate": 8e-05, "loss": 1.9158, "step": 119 }, { "epoch": 0.026393929396238865, "grad_norm": 0.30259498953819275, "learning_rate": 8e-05, "loss": 1.7897, "step": 120 }, { "epoch": 0.026613878807874188, "grad_norm": 0.27566099166870117, "learning_rate": 8e-05, "loss": 1.682, "step": 121 }, { "epoch": 0.026833828219509512, "grad_norm": 0.2959173321723938, "learning_rate": 8e-05, "loss": 1.9032, "step": 122 }, { "epoch": 0.027053777631144835, "grad_norm": 0.29449525475502014, "learning_rate": 8e-05, "loss": 1.6174, "step": 123 }, { "epoch": 0.02727372704278016, "grad_norm": 0.3012568950653076, "learning_rate": 8e-05, "loss": 1.6817, "step": 124 }, { "epoch": 0.027493676454415486, "grad_norm": 0.29086676239967346, "learning_rate": 8e-05, "loss": 1.833, "step": 125 }, { "epoch": 0.02771362586605081, "grad_norm": 0.2756067216396332, "learning_rate": 8e-05, "loss": 1.7807, "step": 126 }, { "epoch": 0.027933575277686133, "grad_norm": 0.3420695662498474, "learning_rate": 8e-05, "loss": 1.8652, "step": 127 }, { "epoch": 0.028153524689321457, "grad_norm": 0.2899749279022217, "learning_rate": 8e-05, "loss": 1.7199, "step": 128 }, { "epoch": 0.02837347410095678, "grad_norm": 0.274718701839447, "learning_rate": 8e-05, "loss": 1.7322, "step": 129 }, { "epoch": 0.028593423512592104, "grad_norm": 0.3784034848213196, "learning_rate": 8e-05, "loss": 1.8917, "step": 130 }, { "epoch": 0.028813372924227428, "grad_norm": 0.2814437448978424, "learning_rate": 8e-05, "loss": 1.726, "step": 131 }, { "epoch": 0.02903332233586275, "grad_norm": 0.287701815366745, "learning_rate": 8e-05, "loss": 1.8166, "step": 132 }, { "epoch": 0.029253271747498075, "grad_norm": 0.28487101197242737, "learning_rate": 8e-05, "loss": 1.7183, "step": 133 }, { "epoch": 0.0294732211591334, "grad_norm": 0.27141597867012024, "learning_rate": 8e-05, "loss": 1.7436, "step": 134 }, { "epoch": 0.029693170570768722, "grad_norm": 0.2708652913570404, "learning_rate": 8e-05, "loss": 1.8116, "step": 135 }, { "epoch": 0.029913119982404045, "grad_norm": 0.2789991796016693, "learning_rate": 8e-05, "loss": 1.7942, "step": 136 }, { "epoch": 0.030133069394039372, "grad_norm": 0.3053725063800812, "learning_rate": 8e-05, "loss": 1.8508, "step": 137 }, { "epoch": 0.030353018805674696, "grad_norm": 0.30432772636413574, "learning_rate": 8e-05, "loss": 1.8129, "step": 138 }, { "epoch": 0.03057296821731002, "grad_norm": 0.2873070240020752, "learning_rate": 8e-05, "loss": 1.8713, "step": 139 }, { "epoch": 0.030792917628945343, "grad_norm": 0.2777135968208313, "learning_rate": 8e-05, "loss": 1.7065, "step": 140 }, { "epoch": 0.031012867040580667, "grad_norm": 0.29774004220962524, "learning_rate": 8e-05, "loss": 1.6471, "step": 141 }, { "epoch": 0.03123281645221599, "grad_norm": 0.2803782522678375, "learning_rate": 8e-05, "loss": 1.6992, "step": 142 }, { "epoch": 0.03145276586385132, "grad_norm": 0.2777007818222046, "learning_rate": 8e-05, "loss": 1.8398, "step": 143 }, { "epoch": 0.03167271527548664, "grad_norm": 0.26938894391059875, "learning_rate": 8e-05, "loss": 1.6082, "step": 144 }, { "epoch": 0.031892664687121965, "grad_norm": 0.2934747338294983, "learning_rate": 8e-05, "loss": 1.6929, "step": 145 }, { "epoch": 0.032112614098757285, "grad_norm": 0.2687772214412689, "learning_rate": 8e-05, "loss": 1.6472, "step": 146 }, { "epoch": 0.03233256351039261, "grad_norm": 0.2758256793022156, "learning_rate": 8e-05, "loss": 1.7128, "step": 147 }, { "epoch": 0.03255251292202793, "grad_norm": 0.26065707206726074, "learning_rate": 8e-05, "loss": 1.7108, "step": 148 }, { "epoch": 0.03277246233366326, "grad_norm": 0.31668898463249207, "learning_rate": 8e-05, "loss": 1.9365, "step": 149 }, { "epoch": 0.03299241174529858, "grad_norm": 0.2915947437286377, "learning_rate": 8e-05, "loss": 1.855, "step": 150 }, { "epoch": 0.033212361156933906, "grad_norm": 0.2741534113883972, "learning_rate": 8e-05, "loss": 1.735, "step": 151 }, { "epoch": 0.033432310568569226, "grad_norm": 0.300800085067749, "learning_rate": 8e-05, "loss": 1.7161, "step": 152 }, { "epoch": 0.03365225998020455, "grad_norm": 0.26691076159477234, "learning_rate": 8e-05, "loss": 1.665, "step": 153 }, { "epoch": 0.03387220939183987, "grad_norm": 0.2605098485946655, "learning_rate": 8e-05, "loss": 1.7288, "step": 154 }, { "epoch": 0.0340921588034752, "grad_norm": 0.2728619873523712, "learning_rate": 8e-05, "loss": 1.7237, "step": 155 }, { "epoch": 0.03431210821511053, "grad_norm": 0.29627877473831177, "learning_rate": 8e-05, "loss": 1.8024, "step": 156 }, { "epoch": 0.03453205762674585, "grad_norm": 0.27106964588165283, "learning_rate": 8e-05, "loss": 1.8166, "step": 157 }, { "epoch": 0.034752007038381175, "grad_norm": 0.26806893944740295, "learning_rate": 8e-05, "loss": 1.7061, "step": 158 }, { "epoch": 0.034971956450016495, "grad_norm": 0.2509767413139343, "learning_rate": 8e-05, "loss": 1.6897, "step": 159 }, { "epoch": 0.03519190586165182, "grad_norm": 0.34342750906944275, "learning_rate": 8e-05, "loss": 1.7151, "step": 160 }, { "epoch": 0.03541185527328714, "grad_norm": 0.27948594093322754, "learning_rate": 8e-05, "loss": 1.6574, "step": 161 }, { "epoch": 0.03563180468492247, "grad_norm": 0.28651687502861023, "learning_rate": 8e-05, "loss": 1.839, "step": 162 }, { "epoch": 0.03585175409655779, "grad_norm": 0.2787701189517975, "learning_rate": 8e-05, "loss": 1.8146, "step": 163 }, { "epoch": 0.036071703508193116, "grad_norm": 0.2596721351146698, "learning_rate": 8e-05, "loss": 1.6088, "step": 164 }, { "epoch": 0.036291652919828436, "grad_norm": 0.2630285322666168, "learning_rate": 8e-05, "loss": 1.6941, "step": 165 }, { "epoch": 0.03651160233146376, "grad_norm": 0.30072465538978577, "learning_rate": 8e-05, "loss": 1.8684, "step": 166 }, { "epoch": 0.03673155174309909, "grad_norm": 0.2789234519004822, "learning_rate": 8e-05, "loss": 1.9136, "step": 167 }, { "epoch": 0.03695150115473441, "grad_norm": 0.25597283244132996, "learning_rate": 8e-05, "loss": 1.669, "step": 168 }, { "epoch": 0.03717145056636974, "grad_norm": 0.30354219675064087, "learning_rate": 8e-05, "loss": 1.7845, "step": 169 }, { "epoch": 0.03739139997800506, "grad_norm": 0.26998043060302734, "learning_rate": 8e-05, "loss": 1.6626, "step": 170 }, { "epoch": 0.037611349389640385, "grad_norm": 0.27418825030326843, "learning_rate": 8e-05, "loss": 1.6444, "step": 171 }, { "epoch": 0.037831298801275705, "grad_norm": 0.2858507037162781, "learning_rate": 8e-05, "loss": 1.8584, "step": 172 }, { "epoch": 0.03805124821291103, "grad_norm": 0.26513633131980896, "learning_rate": 8e-05, "loss": 1.7107, "step": 173 }, { "epoch": 0.03827119762454635, "grad_norm": 0.3162567913532257, "learning_rate": 8e-05, "loss": 1.7153, "step": 174 }, { "epoch": 0.03849114703618168, "grad_norm": 0.28961601853370667, "learning_rate": 8e-05, "loss": 1.8455, "step": 175 }, { "epoch": 0.038711096447817, "grad_norm": 0.29676249623298645, "learning_rate": 8e-05, "loss": 1.9303, "step": 176 }, { "epoch": 0.038931045859452326, "grad_norm": 0.2863664925098419, "learning_rate": 8e-05, "loss": 1.6975, "step": 177 }, { "epoch": 0.03915099527108765, "grad_norm": 0.2715422213077545, "learning_rate": 8e-05, "loss": 1.5472, "step": 178 }, { "epoch": 0.03937094468272297, "grad_norm": 0.2740415036678314, "learning_rate": 8e-05, "loss": 1.7113, "step": 179 }, { "epoch": 0.0395908940943583, "grad_norm": 0.29612302780151367, "learning_rate": 8e-05, "loss": 1.8689, "step": 180 }, { "epoch": 0.03981084350599362, "grad_norm": 0.26745903491973877, "learning_rate": 8e-05, "loss": 1.6076, "step": 181 }, { "epoch": 0.04003079291762895, "grad_norm": 0.296695739030838, "learning_rate": 8e-05, "loss": 1.846, "step": 182 }, { "epoch": 0.04025074232926427, "grad_norm": 0.27626705169677734, "learning_rate": 8e-05, "loss": 1.8103, "step": 183 }, { "epoch": 0.040470691740899595, "grad_norm": 0.2597677409648895, "learning_rate": 8e-05, "loss": 1.6432, "step": 184 }, { "epoch": 0.040690641152534915, "grad_norm": 0.2738899290561676, "learning_rate": 8e-05, "loss": 1.8351, "step": 185 }, { "epoch": 0.04091059056417024, "grad_norm": 0.2683742344379425, "learning_rate": 8e-05, "loss": 1.6453, "step": 186 }, { "epoch": 0.04113053997580556, "grad_norm": 0.28722816705703735, "learning_rate": 8e-05, "loss": 1.7685, "step": 187 }, { "epoch": 0.04135048938744089, "grad_norm": 0.2851015627384186, "learning_rate": 8e-05, "loss": 1.8464, "step": 188 }, { "epoch": 0.04157043879907621, "grad_norm": 0.2630920112133026, "learning_rate": 8e-05, "loss": 1.7176, "step": 189 }, { "epoch": 0.041790388210711536, "grad_norm": 0.2678779661655426, "learning_rate": 8e-05, "loss": 1.671, "step": 190 }, { "epoch": 0.04201033762234686, "grad_norm": 0.27810946106910706, "learning_rate": 8e-05, "loss": 1.6467, "step": 191 }, { "epoch": 0.04223028703398218, "grad_norm": 0.2831014394760132, "learning_rate": 8e-05, "loss": 1.8784, "step": 192 }, { "epoch": 0.04245023644561751, "grad_norm": 0.2643384635448456, "learning_rate": 8e-05, "loss": 1.6239, "step": 193 }, { "epoch": 0.04267018585725283, "grad_norm": 0.27143070101737976, "learning_rate": 8e-05, "loss": 1.8012, "step": 194 }, { "epoch": 0.04289013526888816, "grad_norm": 0.28524088859558105, "learning_rate": 8e-05, "loss": 1.7534, "step": 195 }, { "epoch": 0.04311008468052348, "grad_norm": 0.27226153016090393, "learning_rate": 8e-05, "loss": 1.847, "step": 196 }, { "epoch": 0.043330034092158805, "grad_norm": 0.27042534947395325, "learning_rate": 8e-05, "loss": 1.698, "step": 197 }, { "epoch": 0.043549983503794125, "grad_norm": 0.2673223912715912, "learning_rate": 8e-05, "loss": 1.7825, "step": 198 }, { "epoch": 0.04376993291542945, "grad_norm": 0.26485180854797363, "learning_rate": 8e-05, "loss": 1.7755, "step": 199 }, { "epoch": 0.04398988232706477, "grad_norm": 0.26945164799690247, "learning_rate": 8e-05, "loss": 1.8612, "step": 200 }, { "epoch": 0.0442098317387001, "grad_norm": 0.30337756872177124, "learning_rate": 8e-05, "loss": 1.8556, "step": 201 }, { "epoch": 0.044429781150335426, "grad_norm": 0.26593855023384094, "learning_rate": 8e-05, "loss": 1.7633, "step": 202 }, { "epoch": 0.044649730561970746, "grad_norm": 0.26703208684921265, "learning_rate": 8e-05, "loss": 1.7787, "step": 203 }, { "epoch": 0.04486967997360607, "grad_norm": 0.2799319922924042, "learning_rate": 8e-05, "loss": 1.8946, "step": 204 }, { "epoch": 0.04508962938524139, "grad_norm": 0.261406809091568, "learning_rate": 8e-05, "loss": 1.714, "step": 205 }, { "epoch": 0.04530957879687672, "grad_norm": 0.30923140048980713, "learning_rate": 8e-05, "loss": 1.9953, "step": 206 }, { "epoch": 0.04552952820851204, "grad_norm": 0.28189903497695923, "learning_rate": 8e-05, "loss": 1.8068, "step": 207 }, { "epoch": 0.04574947762014737, "grad_norm": 0.28659504652023315, "learning_rate": 8e-05, "loss": 1.7961, "step": 208 }, { "epoch": 0.04596942703178269, "grad_norm": 0.27828094363212585, "learning_rate": 8e-05, "loss": 1.6398, "step": 209 }, { "epoch": 0.046189376443418015, "grad_norm": 0.2826248109340668, "learning_rate": 8e-05, "loss": 1.8442, "step": 210 }, { "epoch": 0.046409325855053335, "grad_norm": 0.2596709430217743, "learning_rate": 8e-05, "loss": 1.7269, "step": 211 }, { "epoch": 0.04662927526668866, "grad_norm": 0.26883357763290405, "learning_rate": 8e-05, "loss": 1.7396, "step": 212 }, { "epoch": 0.04684922467832398, "grad_norm": 0.2834852933883667, "learning_rate": 8e-05, "loss": 1.6992, "step": 213 }, { "epoch": 0.04706917408995931, "grad_norm": 0.30232125520706177, "learning_rate": 8e-05, "loss": 1.8216, "step": 214 }, { "epoch": 0.047289123501594636, "grad_norm": 0.2887151539325714, "learning_rate": 8e-05, "loss": 1.5633, "step": 215 }, { "epoch": 0.047509072913229956, "grad_norm": 0.27171874046325684, "learning_rate": 8e-05, "loss": 1.8272, "step": 216 }, { "epoch": 0.04772902232486528, "grad_norm": 0.35441088676452637, "learning_rate": 8e-05, "loss": 1.8308, "step": 217 }, { "epoch": 0.047948971736500604, "grad_norm": 0.28351160883903503, "learning_rate": 8e-05, "loss": 1.8697, "step": 218 }, { "epoch": 0.04816892114813593, "grad_norm": 0.26361364126205444, "learning_rate": 8e-05, "loss": 1.7044, "step": 219 }, { "epoch": 0.04838887055977125, "grad_norm": 0.2720041871070862, "learning_rate": 8e-05, "loss": 1.7718, "step": 220 }, { "epoch": 0.04860881997140658, "grad_norm": 0.28131023049354553, "learning_rate": 8e-05, "loss": 1.8066, "step": 221 }, { "epoch": 0.0488287693830419, "grad_norm": 0.2640543580055237, "learning_rate": 8e-05, "loss": 1.69, "step": 222 }, { "epoch": 0.049048718794677225, "grad_norm": 0.26101046800613403, "learning_rate": 8e-05, "loss": 1.6372, "step": 223 }, { "epoch": 0.049268668206312545, "grad_norm": 0.3021651804447174, "learning_rate": 8e-05, "loss": 1.8528, "step": 224 }, { "epoch": 0.04948861761794787, "grad_norm": 0.2655261158943176, "learning_rate": 8e-05, "loss": 1.7406, "step": 225 }, { "epoch": 0.0497085670295832, "grad_norm": 0.2873914837837219, "learning_rate": 8e-05, "loss": 1.7643, "step": 226 }, { "epoch": 0.04992851644121852, "grad_norm": 0.31813880801200867, "learning_rate": 8e-05, "loss": 1.8645, "step": 227 }, { "epoch": 0.050148465852853846, "grad_norm": 0.2996014654636383, "learning_rate": 8e-05, "loss": 1.6685, "step": 228 }, { "epoch": 0.050368415264489166, "grad_norm": 0.2837509512901306, "learning_rate": 8e-05, "loss": 1.9227, "step": 229 }, { "epoch": 0.05058836467612449, "grad_norm": 0.29532885551452637, "learning_rate": 8e-05, "loss": 1.9073, "step": 230 }, { "epoch": 0.050808314087759814, "grad_norm": 0.285295307636261, "learning_rate": 8e-05, "loss": 1.8248, "step": 231 }, { "epoch": 0.05102826349939514, "grad_norm": 0.26331770420074463, "learning_rate": 8e-05, "loss": 1.7146, "step": 232 }, { "epoch": 0.05124821291103046, "grad_norm": 0.24956567585468292, "learning_rate": 8e-05, "loss": 1.5574, "step": 233 }, { "epoch": 0.05146816232266579, "grad_norm": 0.27515965700149536, "learning_rate": 8e-05, "loss": 1.7854, "step": 234 }, { "epoch": 0.05168811173430111, "grad_norm": 0.28268730640411377, "learning_rate": 8e-05, "loss": 1.8294, "step": 235 }, { "epoch": 0.051908061145936435, "grad_norm": 0.25420427322387695, "learning_rate": 8e-05, "loss": 1.6735, "step": 236 }, { "epoch": 0.052128010557571755, "grad_norm": 0.2869463860988617, "learning_rate": 8e-05, "loss": 1.808, "step": 237 }, { "epoch": 0.05234795996920708, "grad_norm": 0.2574792206287384, "learning_rate": 8e-05, "loss": 1.7563, "step": 238 }, { "epoch": 0.05256790938084241, "grad_norm": 0.26652273535728455, "learning_rate": 8e-05, "loss": 1.743, "step": 239 }, { "epoch": 0.05278785879247773, "grad_norm": 0.2956235408782959, "learning_rate": 8e-05, "loss": 1.9169, "step": 240 }, { "epoch": 0.053007808204113056, "grad_norm": 0.274142861366272, "learning_rate": 8e-05, "loss": 1.8321, "step": 241 }, { "epoch": 0.053227757615748376, "grad_norm": 0.27525436878204346, "learning_rate": 8e-05, "loss": 1.8206, "step": 242 }, { "epoch": 0.053447707027383703, "grad_norm": 0.26323091983795166, "learning_rate": 8e-05, "loss": 1.7574, "step": 243 }, { "epoch": 0.053667656439019024, "grad_norm": 0.28554126620292664, "learning_rate": 8e-05, "loss": 1.9293, "step": 244 }, { "epoch": 0.05388760585065435, "grad_norm": 0.2651476562023163, "learning_rate": 8e-05, "loss": 1.808, "step": 245 }, { "epoch": 0.05410755526228967, "grad_norm": 0.27941837906837463, "learning_rate": 8e-05, "loss": 1.7838, "step": 246 }, { "epoch": 0.054327504673925, "grad_norm": 0.26575711369514465, "learning_rate": 8e-05, "loss": 1.6117, "step": 247 }, { "epoch": 0.05454745408556032, "grad_norm": 0.2620556354522705, "learning_rate": 8e-05, "loss": 1.7703, "step": 248 }, { "epoch": 0.054767403497195645, "grad_norm": 0.2782936990261078, "learning_rate": 8e-05, "loss": 1.753, "step": 249 }, { "epoch": 0.05498735290883097, "grad_norm": 0.28347843885421753, "learning_rate": 8e-05, "loss": 1.8365, "step": 250 }, { "epoch": 0.05520730232046629, "grad_norm": 0.2740314304828644, "learning_rate": 8e-05, "loss": 1.7448, "step": 251 }, { "epoch": 0.05542725173210162, "grad_norm": 0.2779199779033661, "learning_rate": 8e-05, "loss": 1.8025, "step": 252 }, { "epoch": 0.05564720114373694, "grad_norm": 0.27700838446617126, "learning_rate": 8e-05, "loss": 1.6368, "step": 253 }, { "epoch": 0.055867150555372266, "grad_norm": 0.2753797173500061, "learning_rate": 8e-05, "loss": 1.7058, "step": 254 }, { "epoch": 0.056087099967007586, "grad_norm": 0.2677604556083679, "learning_rate": 8e-05, "loss": 1.772, "step": 255 }, { "epoch": 0.056307049378642914, "grad_norm": 0.291358083486557, "learning_rate": 8e-05, "loss": 1.7229, "step": 256 }, { "epoch": 0.056526998790278234, "grad_norm": 0.2605611979961395, "learning_rate": 8e-05, "loss": 1.6654, "step": 257 }, { "epoch": 0.05674694820191356, "grad_norm": 0.2726796865463257, "learning_rate": 8e-05, "loss": 1.8524, "step": 258 }, { "epoch": 0.05696689761354888, "grad_norm": 0.2769307494163513, "learning_rate": 8e-05, "loss": 1.913, "step": 259 }, { "epoch": 0.05718684702518421, "grad_norm": 0.27163514494895935, "learning_rate": 8e-05, "loss": 1.7076, "step": 260 }, { "epoch": 0.057406796436819535, "grad_norm": 0.27037522196769714, "learning_rate": 8e-05, "loss": 1.7461, "step": 261 }, { "epoch": 0.057626745848454855, "grad_norm": 0.2570153772830963, "learning_rate": 8e-05, "loss": 1.6714, "step": 262 }, { "epoch": 0.05784669526009018, "grad_norm": 0.2802227735519409, "learning_rate": 8e-05, "loss": 1.6782, "step": 263 }, { "epoch": 0.0580666446717255, "grad_norm": 0.293969064950943, "learning_rate": 8e-05, "loss": 1.6253, "step": 264 }, { "epoch": 0.05828659408336083, "grad_norm": 0.28199446201324463, "learning_rate": 8e-05, "loss": 1.791, "step": 265 }, { "epoch": 0.05850654349499615, "grad_norm": 0.3037835657596588, "learning_rate": 8e-05, "loss": 1.8553, "step": 266 }, { "epoch": 0.058726492906631476, "grad_norm": 0.2814860939979553, "learning_rate": 8e-05, "loss": 1.7237, "step": 267 }, { "epoch": 0.0589464423182668, "grad_norm": 0.29769864678382874, "learning_rate": 8e-05, "loss": 1.8635, "step": 268 }, { "epoch": 0.059166391729902124, "grad_norm": 0.26650169491767883, "learning_rate": 8e-05, "loss": 1.8173, "step": 269 }, { "epoch": 0.059386341141537444, "grad_norm": 0.29682958126068115, "learning_rate": 8e-05, "loss": 1.6548, "step": 270 }, { "epoch": 0.05960629055317277, "grad_norm": 0.2702498137950897, "learning_rate": 8e-05, "loss": 1.6022, "step": 271 }, { "epoch": 0.05982623996480809, "grad_norm": 0.2940424680709839, "learning_rate": 8e-05, "loss": 1.7955, "step": 272 }, { "epoch": 0.06004618937644342, "grad_norm": 0.2655317485332489, "learning_rate": 8e-05, "loss": 1.786, "step": 273 }, { "epoch": 0.060266138788078745, "grad_norm": 0.28093400597572327, "learning_rate": 8e-05, "loss": 1.9798, "step": 274 }, { "epoch": 0.060486088199714065, "grad_norm": 0.2635514736175537, "learning_rate": 8e-05, "loss": 1.6737, "step": 275 }, { "epoch": 0.06070603761134939, "grad_norm": 0.2648226320743561, "learning_rate": 8e-05, "loss": 1.8771, "step": 276 }, { "epoch": 0.06092598702298471, "grad_norm": 0.2934603691101074, "learning_rate": 8e-05, "loss": 1.4751, "step": 277 }, { "epoch": 0.06114593643462004, "grad_norm": 0.26369500160217285, "learning_rate": 8e-05, "loss": 1.7832, "step": 278 }, { "epoch": 0.06136588584625536, "grad_norm": 0.26159989833831787, "learning_rate": 8e-05, "loss": 1.7276, "step": 279 }, { "epoch": 0.061585835257890686, "grad_norm": 0.2826705873012543, "learning_rate": 8e-05, "loss": 1.8767, "step": 280 }, { "epoch": 0.06180578466952601, "grad_norm": 0.2911459505558014, "learning_rate": 8e-05, "loss": 1.7795, "step": 281 }, { "epoch": 0.062025734081161334, "grad_norm": 0.27846869826316833, "learning_rate": 8e-05, "loss": 1.838, "step": 282 }, { "epoch": 0.062245683492796654, "grad_norm": 0.33195585012435913, "learning_rate": 8e-05, "loss": 1.8576, "step": 283 }, { "epoch": 0.06246563290443198, "grad_norm": 0.26306337118148804, "learning_rate": 8e-05, "loss": 1.7202, "step": 284 }, { "epoch": 0.0626855823160673, "grad_norm": 0.2703022360801697, "learning_rate": 8e-05, "loss": 1.6962, "step": 285 }, { "epoch": 0.06290553172770263, "grad_norm": 0.2754605710506439, "learning_rate": 8e-05, "loss": 1.6468, "step": 286 }, { "epoch": 0.06312548113933796, "grad_norm": 0.2995694577693939, "learning_rate": 8e-05, "loss": 1.9298, "step": 287 }, { "epoch": 0.06334543055097328, "grad_norm": 0.27501800656318665, "learning_rate": 8e-05, "loss": 1.8152, "step": 288 }, { "epoch": 0.0635653799626086, "grad_norm": 0.2668202519416809, "learning_rate": 8e-05, "loss": 1.8809, "step": 289 }, { "epoch": 0.06378532937424393, "grad_norm": 0.26209571957588196, "learning_rate": 8e-05, "loss": 1.4927, "step": 290 }, { "epoch": 0.06400527878587925, "grad_norm": 0.35276591777801514, "learning_rate": 8e-05, "loss": 1.9654, "step": 291 }, { "epoch": 0.06422522819751457, "grad_norm": 0.26070040464401245, "learning_rate": 8e-05, "loss": 1.7332, "step": 292 }, { "epoch": 0.06444517760914989, "grad_norm": 0.26518604159355164, "learning_rate": 8e-05, "loss": 1.6867, "step": 293 }, { "epoch": 0.06466512702078522, "grad_norm": 0.28992095589637756, "learning_rate": 8e-05, "loss": 1.7498, "step": 294 }, { "epoch": 0.06488507643242054, "grad_norm": 0.27465108036994934, "learning_rate": 8e-05, "loss": 1.6095, "step": 295 }, { "epoch": 0.06510502584405586, "grad_norm": 0.2841359078884125, "learning_rate": 8e-05, "loss": 1.6869, "step": 296 }, { "epoch": 0.0653249752556912, "grad_norm": 0.28873759508132935, "learning_rate": 8e-05, "loss": 1.7954, "step": 297 }, { "epoch": 0.06554492466732652, "grad_norm": 0.2542605698108673, "learning_rate": 8e-05, "loss": 1.6075, "step": 298 }, { "epoch": 0.06576487407896184, "grad_norm": 0.270823210477829, "learning_rate": 8e-05, "loss": 1.7238, "step": 299 }, { "epoch": 0.06598482349059716, "grad_norm": 0.2610267102718353, "learning_rate": 8e-05, "loss": 1.697, "step": 300 }, { "epoch": 0.06620477290223249, "grad_norm": 0.28088685870170593, "learning_rate": 8e-05, "loss": 1.6806, "step": 301 }, { "epoch": 0.06642472231386781, "grad_norm": 0.2656930088996887, "learning_rate": 8e-05, "loss": 1.8744, "step": 302 }, { "epoch": 0.06664467172550313, "grad_norm": 0.2721637189388275, "learning_rate": 8e-05, "loss": 1.6903, "step": 303 }, { "epoch": 0.06686462113713845, "grad_norm": 0.2612883746623993, "learning_rate": 8e-05, "loss": 1.7444, "step": 304 }, { "epoch": 0.06708457054877379, "grad_norm": 0.2533530592918396, "learning_rate": 8e-05, "loss": 1.6427, "step": 305 }, { "epoch": 0.0673045199604091, "grad_norm": 0.27200043201446533, "learning_rate": 8e-05, "loss": 1.769, "step": 306 }, { "epoch": 0.06752446937204443, "grad_norm": 0.2626403272151947, "learning_rate": 8e-05, "loss": 1.64, "step": 307 }, { "epoch": 0.06774441878367975, "grad_norm": 0.3720408082008362, "learning_rate": 8e-05, "loss": 1.9055, "step": 308 }, { "epoch": 0.06796436819531508, "grad_norm": 0.2745527923107147, "learning_rate": 8e-05, "loss": 1.7844, "step": 309 }, { "epoch": 0.0681843176069504, "grad_norm": 0.2568323612213135, "learning_rate": 8e-05, "loss": 1.6728, "step": 310 }, { "epoch": 0.06840426701858572, "grad_norm": 0.2704140543937683, "learning_rate": 8e-05, "loss": 1.7685, "step": 311 }, { "epoch": 0.06862421643022105, "grad_norm": 0.27828502655029297, "learning_rate": 8e-05, "loss": 1.7957, "step": 312 }, { "epoch": 0.06884416584185638, "grad_norm": 0.2951858341693878, "learning_rate": 8e-05, "loss": 1.7709, "step": 313 }, { "epoch": 0.0690641152534917, "grad_norm": 0.2756475806236267, "learning_rate": 8e-05, "loss": 1.6348, "step": 314 }, { "epoch": 0.06928406466512702, "grad_norm": 0.2913607954978943, "learning_rate": 8e-05, "loss": 1.7888, "step": 315 }, { "epoch": 0.06950401407676235, "grad_norm": 0.2798636853694916, "learning_rate": 8e-05, "loss": 1.7806, "step": 316 }, { "epoch": 0.06972396348839767, "grad_norm": 0.27596554160118103, "learning_rate": 8e-05, "loss": 1.7458, "step": 317 }, { "epoch": 0.06994391290003299, "grad_norm": 0.26655322313308716, "learning_rate": 8e-05, "loss": 1.5985, "step": 318 }, { "epoch": 0.07016386231166831, "grad_norm": 0.2731332778930664, "learning_rate": 8e-05, "loss": 1.5995, "step": 319 }, { "epoch": 0.07038381172330364, "grad_norm": 0.2769210934638977, "learning_rate": 8e-05, "loss": 1.6748, "step": 320 }, { "epoch": 0.07060376113493896, "grad_norm": 0.290889173746109, "learning_rate": 8e-05, "loss": 1.9427, "step": 321 }, { "epoch": 0.07082371054657428, "grad_norm": 0.2911258339881897, "learning_rate": 8e-05, "loss": 1.7723, "step": 322 }, { "epoch": 0.07104365995820962, "grad_norm": 0.301992267370224, "learning_rate": 8e-05, "loss": 1.7772, "step": 323 }, { "epoch": 0.07126360936984494, "grad_norm": 0.3023516535758972, "learning_rate": 8e-05, "loss": 1.8363, "step": 324 }, { "epoch": 0.07148355878148026, "grad_norm": 0.3058542013168335, "learning_rate": 8e-05, "loss": 1.8762, "step": 325 }, { "epoch": 0.07170350819311558, "grad_norm": 0.3215092718601227, "learning_rate": 8e-05, "loss": 1.7265, "step": 326 }, { "epoch": 0.07192345760475091, "grad_norm": 0.2762998342514038, "learning_rate": 8e-05, "loss": 1.6361, "step": 327 }, { "epoch": 0.07214340701638623, "grad_norm": 0.258635014295578, "learning_rate": 8e-05, "loss": 1.7031, "step": 328 }, { "epoch": 0.07236335642802155, "grad_norm": 0.27160710096359253, "learning_rate": 8e-05, "loss": 1.6759, "step": 329 }, { "epoch": 0.07258330583965687, "grad_norm": 0.31089314818382263, "learning_rate": 8e-05, "loss": 1.8141, "step": 330 }, { "epoch": 0.0728032552512922, "grad_norm": 0.3026575744152069, "learning_rate": 8e-05, "loss": 1.9513, "step": 331 }, { "epoch": 0.07302320466292753, "grad_norm": 0.2692122161388397, "learning_rate": 8e-05, "loss": 1.8277, "step": 332 }, { "epoch": 0.07324315407456285, "grad_norm": 0.27460286021232605, "learning_rate": 8e-05, "loss": 1.6426, "step": 333 }, { "epoch": 0.07346310348619818, "grad_norm": 0.2557325065135956, "learning_rate": 8e-05, "loss": 1.6418, "step": 334 }, { "epoch": 0.0736830528978335, "grad_norm": 0.28074318170547485, "learning_rate": 8e-05, "loss": 1.79, "step": 335 }, { "epoch": 0.07390300230946882, "grad_norm": 0.28538671135902405, "learning_rate": 8e-05, "loss": 1.7363, "step": 336 }, { "epoch": 0.07412295172110414, "grad_norm": 0.27379995584487915, "learning_rate": 8e-05, "loss": 1.7881, "step": 337 }, { "epoch": 0.07434290113273948, "grad_norm": 0.2628316283226013, "learning_rate": 8e-05, "loss": 1.745, "step": 338 }, { "epoch": 0.0745628505443748, "grad_norm": 0.2573058009147644, "learning_rate": 8e-05, "loss": 1.7997, "step": 339 }, { "epoch": 0.07478279995601012, "grad_norm": 0.31905651092529297, "learning_rate": 8e-05, "loss": 1.8125, "step": 340 }, { "epoch": 0.07500274936764544, "grad_norm": 0.2501446604728699, "learning_rate": 8e-05, "loss": 1.557, "step": 341 }, { "epoch": 0.07522269877928077, "grad_norm": 0.26969289779663086, "learning_rate": 8e-05, "loss": 1.7819, "step": 342 }, { "epoch": 0.07544264819091609, "grad_norm": 0.28457415103912354, "learning_rate": 8e-05, "loss": 1.7682, "step": 343 }, { "epoch": 0.07566259760255141, "grad_norm": 0.27833452820777893, "learning_rate": 8e-05, "loss": 1.8436, "step": 344 }, { "epoch": 0.07588254701418674, "grad_norm": 0.2574867010116577, "learning_rate": 8e-05, "loss": 1.7196, "step": 345 }, { "epoch": 0.07610249642582206, "grad_norm": 0.30035245418548584, "learning_rate": 8e-05, "loss": 1.7159, "step": 346 }, { "epoch": 0.07632244583745738, "grad_norm": 0.284169465303421, "learning_rate": 8e-05, "loss": 1.7238, "step": 347 }, { "epoch": 0.0765423952490927, "grad_norm": 0.257168173789978, "learning_rate": 8e-05, "loss": 1.8531, "step": 348 }, { "epoch": 0.07676234466072804, "grad_norm": 0.2611413300037384, "learning_rate": 8e-05, "loss": 1.7753, "step": 349 }, { "epoch": 0.07698229407236336, "grad_norm": 0.26592132449150085, "learning_rate": 8e-05, "loss": 1.7557, "step": 350 }, { "epoch": 0.07720224348399868, "grad_norm": 0.27427396178245544, "learning_rate": 8e-05, "loss": 1.8699, "step": 351 }, { "epoch": 0.077422192895634, "grad_norm": 0.27014485001564026, "learning_rate": 8e-05, "loss": 1.816, "step": 352 }, { "epoch": 0.07764214230726933, "grad_norm": 0.27720019221305847, "learning_rate": 8e-05, "loss": 1.9601, "step": 353 }, { "epoch": 0.07786209171890465, "grad_norm": 0.3222314417362213, "learning_rate": 8e-05, "loss": 1.6726, "step": 354 }, { "epoch": 0.07808204113053997, "grad_norm": 0.2675410211086273, "learning_rate": 8e-05, "loss": 1.7113, "step": 355 }, { "epoch": 0.0783019905421753, "grad_norm": 0.2902251183986664, "learning_rate": 8e-05, "loss": 1.7734, "step": 356 }, { "epoch": 0.07852193995381063, "grad_norm": 0.2985514998435974, "learning_rate": 8e-05, "loss": 1.9182, "step": 357 }, { "epoch": 0.07874188936544595, "grad_norm": 0.30351343750953674, "learning_rate": 8e-05, "loss": 1.7795, "step": 358 }, { "epoch": 0.07896183877708127, "grad_norm": 0.2885829210281372, "learning_rate": 8e-05, "loss": 1.8054, "step": 359 }, { "epoch": 0.0791817881887166, "grad_norm": 0.273366242647171, "learning_rate": 8e-05, "loss": 1.7903, "step": 360 }, { "epoch": 0.07940173760035192, "grad_norm": 0.2959200441837311, "learning_rate": 8e-05, "loss": 1.9163, "step": 361 }, { "epoch": 0.07962168701198724, "grad_norm": 0.2587856948375702, "learning_rate": 8e-05, "loss": 1.5969, "step": 362 }, { "epoch": 0.07984163642362256, "grad_norm": 0.27777665853500366, "learning_rate": 8e-05, "loss": 1.8769, "step": 363 }, { "epoch": 0.0800615858352579, "grad_norm": 0.2635156512260437, "learning_rate": 8e-05, "loss": 1.8236, "step": 364 }, { "epoch": 0.08028153524689322, "grad_norm": 0.26534774899482727, "learning_rate": 8e-05, "loss": 1.6824, "step": 365 }, { "epoch": 0.08050148465852854, "grad_norm": 0.26372772455215454, "learning_rate": 8e-05, "loss": 1.517, "step": 366 }, { "epoch": 0.08072143407016386, "grad_norm": 0.2707895338535309, "learning_rate": 8e-05, "loss": 1.6757, "step": 367 }, { "epoch": 0.08094138348179919, "grad_norm": 0.2712070345878601, "learning_rate": 8e-05, "loss": 1.7261, "step": 368 }, { "epoch": 0.08116133289343451, "grad_norm": 0.2870525121688843, "learning_rate": 8e-05, "loss": 1.6337, "step": 369 }, { "epoch": 0.08138128230506983, "grad_norm": 0.30548396706581116, "learning_rate": 8e-05, "loss": 1.8733, "step": 370 }, { "epoch": 0.08160123171670516, "grad_norm": 0.2853962182998657, "learning_rate": 8e-05, "loss": 1.7938, "step": 371 }, { "epoch": 0.08182118112834048, "grad_norm": 0.2716579735279083, "learning_rate": 8e-05, "loss": 1.6733, "step": 372 }, { "epoch": 0.0820411305399758, "grad_norm": 0.3110131025314331, "learning_rate": 8e-05, "loss": 1.8554, "step": 373 }, { "epoch": 0.08226107995161112, "grad_norm": 0.28003835678100586, "learning_rate": 8e-05, "loss": 1.8032, "step": 374 }, { "epoch": 0.08248102936324646, "grad_norm": 0.28504347801208496, "learning_rate": 8e-05, "loss": 1.942, "step": 375 }, { "epoch": 0.08270097877488178, "grad_norm": 0.2593232989311218, "learning_rate": 8e-05, "loss": 1.4993, "step": 376 }, { "epoch": 0.0829209281865171, "grad_norm": 0.35680094361305237, "learning_rate": 8e-05, "loss": 1.8997, "step": 377 }, { "epoch": 0.08314087759815242, "grad_norm": 0.2747777998447418, "learning_rate": 8e-05, "loss": 1.7364, "step": 378 }, { "epoch": 0.08336082700978775, "grad_norm": 0.26816287636756897, "learning_rate": 8e-05, "loss": 1.7011, "step": 379 }, { "epoch": 0.08358077642142307, "grad_norm": 0.31877851486206055, "learning_rate": 8e-05, "loss": 1.6131, "step": 380 }, { "epoch": 0.08380072583305839, "grad_norm": 0.2845601737499237, "learning_rate": 8e-05, "loss": 1.6544, "step": 381 }, { "epoch": 0.08402067524469373, "grad_norm": 0.27758803963661194, "learning_rate": 8e-05, "loss": 1.8891, "step": 382 }, { "epoch": 0.08424062465632905, "grad_norm": 0.2832657992839813, "learning_rate": 8e-05, "loss": 1.7505, "step": 383 }, { "epoch": 0.08446057406796437, "grad_norm": 0.2901705801486969, "learning_rate": 8e-05, "loss": 1.7501, "step": 384 }, { "epoch": 0.08468052347959969, "grad_norm": 0.31189531087875366, "learning_rate": 8e-05, "loss": 1.8132, "step": 385 }, { "epoch": 0.08490047289123502, "grad_norm": 0.27582603693008423, "learning_rate": 8e-05, "loss": 1.7693, "step": 386 }, { "epoch": 0.08512042230287034, "grad_norm": 0.3030100464820862, "learning_rate": 8e-05, "loss": 1.7327, "step": 387 }, { "epoch": 0.08534037171450566, "grad_norm": 0.26879045367240906, "learning_rate": 8e-05, "loss": 1.6614, "step": 388 }, { "epoch": 0.08556032112614098, "grad_norm": 0.29507508873939514, "learning_rate": 8e-05, "loss": 1.9483, "step": 389 }, { "epoch": 0.08578027053777632, "grad_norm": 0.27386122941970825, "learning_rate": 8e-05, "loss": 1.8974, "step": 390 }, { "epoch": 0.08600021994941164, "grad_norm": 0.27103161811828613, "learning_rate": 8e-05, "loss": 1.7579, "step": 391 }, { "epoch": 0.08622016936104696, "grad_norm": 0.3045141100883484, "learning_rate": 8e-05, "loss": 1.8175, "step": 392 }, { "epoch": 0.08644011877268229, "grad_norm": 0.29032695293426514, "learning_rate": 8e-05, "loss": 1.7493, "step": 393 }, { "epoch": 0.08666006818431761, "grad_norm": 0.27853158116340637, "learning_rate": 8e-05, "loss": 1.7297, "step": 394 }, { "epoch": 0.08688001759595293, "grad_norm": 0.3007650375366211, "learning_rate": 8e-05, "loss": 1.6736, "step": 395 }, { "epoch": 0.08709996700758825, "grad_norm": 0.28009670972824097, "learning_rate": 8e-05, "loss": 1.9539, "step": 396 }, { "epoch": 0.08731991641922358, "grad_norm": 0.2512955665588379, "learning_rate": 8e-05, "loss": 1.6362, "step": 397 }, { "epoch": 0.0875398658308589, "grad_norm": 0.297489732503891, "learning_rate": 8e-05, "loss": 1.9097, "step": 398 }, { "epoch": 0.08775981524249422, "grad_norm": 0.2735532522201538, "learning_rate": 8e-05, "loss": 1.8348, "step": 399 }, { "epoch": 0.08797976465412954, "grad_norm": 0.2559053897857666, "learning_rate": 8e-05, "loss": 1.685, "step": 400 }, { "epoch": 0.08819971406576488, "grad_norm": 0.27982097864151, "learning_rate": 8e-05, "loss": 1.6801, "step": 401 }, { "epoch": 0.0884196634774002, "grad_norm": 0.26066988706588745, "learning_rate": 8e-05, "loss": 1.7732, "step": 402 }, { "epoch": 0.08863961288903552, "grad_norm": 0.26763463020324707, "learning_rate": 8e-05, "loss": 1.7214, "step": 403 }, { "epoch": 0.08885956230067085, "grad_norm": 0.2795925736427307, "learning_rate": 8e-05, "loss": 1.8387, "step": 404 }, { "epoch": 0.08907951171230617, "grad_norm": 0.266305148601532, "learning_rate": 8e-05, "loss": 1.6515, "step": 405 }, { "epoch": 0.08929946112394149, "grad_norm": 0.27049583196640015, "learning_rate": 8e-05, "loss": 1.7824, "step": 406 }, { "epoch": 0.08951941053557681, "grad_norm": 0.2959458529949188, "learning_rate": 8e-05, "loss": 1.8766, "step": 407 }, { "epoch": 0.08973935994721215, "grad_norm": 0.28563347458839417, "learning_rate": 8e-05, "loss": 1.8618, "step": 408 }, { "epoch": 0.08995930935884747, "grad_norm": 0.2840110659599304, "learning_rate": 8e-05, "loss": 1.6834, "step": 409 }, { "epoch": 0.09017925877048279, "grad_norm": 0.25303247570991516, "learning_rate": 8e-05, "loss": 1.6477, "step": 410 }, { "epoch": 0.09039920818211811, "grad_norm": 0.27236899733543396, "learning_rate": 8e-05, "loss": 1.7004, "step": 411 }, { "epoch": 0.09061915759375344, "grad_norm": 0.2795659899711609, "learning_rate": 8e-05, "loss": 1.7492, "step": 412 }, { "epoch": 0.09083910700538876, "grad_norm": 0.26019132137298584, "learning_rate": 8e-05, "loss": 1.691, "step": 413 }, { "epoch": 0.09105905641702408, "grad_norm": 0.26624274253845215, "learning_rate": 8e-05, "loss": 1.7001, "step": 414 }, { "epoch": 0.09127900582865942, "grad_norm": 0.2661585509777069, "learning_rate": 8e-05, "loss": 1.6762, "step": 415 }, { "epoch": 0.09149895524029474, "grad_norm": 0.2719002068042755, "learning_rate": 8e-05, "loss": 1.6915, "step": 416 }, { "epoch": 0.09171890465193006, "grad_norm": 0.24670244753360748, "learning_rate": 8e-05, "loss": 1.5598, "step": 417 }, { "epoch": 0.09193885406356538, "grad_norm": 0.2550405263900757, "learning_rate": 8e-05, "loss": 1.4817, "step": 418 }, { "epoch": 0.09215880347520071, "grad_norm": 0.26272761821746826, "learning_rate": 8e-05, "loss": 1.7016, "step": 419 }, { "epoch": 0.09237875288683603, "grad_norm": 0.2673632502555847, "learning_rate": 8e-05, "loss": 1.7626, "step": 420 }, { "epoch": 0.09259870229847135, "grad_norm": 0.25949448347091675, "learning_rate": 8e-05, "loss": 1.6273, "step": 421 }, { "epoch": 0.09281865171010667, "grad_norm": 0.27953028678894043, "learning_rate": 8e-05, "loss": 1.8843, "step": 422 }, { "epoch": 0.093038601121742, "grad_norm": 0.2534630298614502, "learning_rate": 8e-05, "loss": 1.7305, "step": 423 }, { "epoch": 0.09325855053337732, "grad_norm": 0.2573072910308838, "learning_rate": 8e-05, "loss": 1.6397, "step": 424 }, { "epoch": 0.09347849994501264, "grad_norm": 0.2604135572910309, "learning_rate": 8e-05, "loss": 1.6696, "step": 425 }, { "epoch": 0.09369844935664796, "grad_norm": 0.25805628299713135, "learning_rate": 8e-05, "loss": 1.6441, "step": 426 }, { "epoch": 0.0939183987682833, "grad_norm": 0.2935563027858734, "learning_rate": 8e-05, "loss": 1.6475, "step": 427 }, { "epoch": 0.09413834817991862, "grad_norm": 0.25222933292388916, "learning_rate": 8e-05, "loss": 1.727, "step": 428 }, { "epoch": 0.09435829759155394, "grad_norm": 0.2593076527118683, "learning_rate": 8e-05, "loss": 1.7066, "step": 429 }, { "epoch": 0.09457824700318927, "grad_norm": 0.25259336829185486, "learning_rate": 8e-05, "loss": 1.6821, "step": 430 }, { "epoch": 0.09479819641482459, "grad_norm": 0.2512541115283966, "learning_rate": 8e-05, "loss": 1.5923, "step": 431 }, { "epoch": 0.09501814582645991, "grad_norm": 0.2711183726787567, "learning_rate": 8e-05, "loss": 1.755, "step": 432 }, { "epoch": 0.09523809523809523, "grad_norm": 0.2782961130142212, "learning_rate": 8e-05, "loss": 1.8914, "step": 433 }, { "epoch": 0.09545804464973057, "grad_norm": 0.25964146852493286, "learning_rate": 8e-05, "loss": 1.6588, "step": 434 }, { "epoch": 0.09567799406136589, "grad_norm": 0.27077510952949524, "learning_rate": 8e-05, "loss": 1.753, "step": 435 }, { "epoch": 0.09589794347300121, "grad_norm": 0.2923937141895294, "learning_rate": 8e-05, "loss": 1.8218, "step": 436 }, { "epoch": 0.09611789288463653, "grad_norm": 0.2513190805912018, "learning_rate": 8e-05, "loss": 1.6232, "step": 437 }, { "epoch": 0.09633784229627186, "grad_norm": 0.28531181812286377, "learning_rate": 8e-05, "loss": 1.7199, "step": 438 }, { "epoch": 0.09655779170790718, "grad_norm": 0.302020400762558, "learning_rate": 8e-05, "loss": 1.8359, "step": 439 }, { "epoch": 0.0967777411195425, "grad_norm": 0.28001338243484497, "learning_rate": 8e-05, "loss": 1.8434, "step": 440 }, { "epoch": 0.09699769053117784, "grad_norm": 0.2990663945674896, "learning_rate": 8e-05, "loss": 1.6995, "step": 441 }, { "epoch": 0.09721763994281316, "grad_norm": 0.266197144985199, "learning_rate": 8e-05, "loss": 1.6195, "step": 442 }, { "epoch": 0.09743758935444848, "grad_norm": 0.28108519315719604, "learning_rate": 8e-05, "loss": 1.8108, "step": 443 }, { "epoch": 0.0976575387660838, "grad_norm": 0.26744788885116577, "learning_rate": 8e-05, "loss": 1.6497, "step": 444 }, { "epoch": 0.09787748817771913, "grad_norm": 0.28030574321746826, "learning_rate": 8e-05, "loss": 1.8143, "step": 445 }, { "epoch": 0.09809743758935445, "grad_norm": 0.27872079610824585, "learning_rate": 8e-05, "loss": 1.6319, "step": 446 }, { "epoch": 0.09831738700098977, "grad_norm": 0.2816067039966583, "learning_rate": 8e-05, "loss": 1.8385, "step": 447 }, { "epoch": 0.09853733641262509, "grad_norm": 0.25677627325057983, "learning_rate": 8e-05, "loss": 1.6885, "step": 448 }, { "epoch": 0.09875728582426042, "grad_norm": 0.276569128036499, "learning_rate": 8e-05, "loss": 1.7652, "step": 449 }, { "epoch": 0.09897723523589574, "grad_norm": 0.2765633463859558, "learning_rate": 8e-05, "loss": 1.7763, "step": 450 }, { "epoch": 0.09919718464753106, "grad_norm": 0.27050015330314636, "learning_rate": 8e-05, "loss": 1.6459, "step": 451 }, { "epoch": 0.0994171340591664, "grad_norm": 0.2552846372127533, "learning_rate": 8e-05, "loss": 1.6877, "step": 452 }, { "epoch": 0.09963708347080172, "grad_norm": 0.2653469741344452, "learning_rate": 8e-05, "loss": 1.6536, "step": 453 }, { "epoch": 0.09985703288243704, "grad_norm": 0.28801941871643066, "learning_rate": 8e-05, "loss": 1.7643, "step": 454 }, { "epoch": 0.10007698229407236, "grad_norm": 0.2930269241333008, "learning_rate": 8e-05, "loss": 1.7766, "step": 455 }, { "epoch": 0.10029693170570769, "grad_norm": 0.2718334496021271, "learning_rate": 8e-05, "loss": 1.7347, "step": 456 }, { "epoch": 0.10051688111734301, "grad_norm": 0.2807629704475403, "learning_rate": 8e-05, "loss": 1.7245, "step": 457 }, { "epoch": 0.10073683052897833, "grad_norm": 0.2801489531993866, "learning_rate": 8e-05, "loss": 1.7854, "step": 458 }, { "epoch": 0.10095677994061365, "grad_norm": 0.2616996765136719, "learning_rate": 8e-05, "loss": 1.6179, "step": 459 }, { "epoch": 0.10117672935224899, "grad_norm": 0.2626480758190155, "learning_rate": 8e-05, "loss": 1.7475, "step": 460 }, { "epoch": 0.10139667876388431, "grad_norm": 0.27338841557502747, "learning_rate": 8e-05, "loss": 1.8972, "step": 461 }, { "epoch": 0.10161662817551963, "grad_norm": 0.2695038616657257, "learning_rate": 8e-05, "loss": 1.7279, "step": 462 }, { "epoch": 0.10183657758715496, "grad_norm": 0.25614050030708313, "learning_rate": 8e-05, "loss": 1.6, "step": 463 }, { "epoch": 0.10205652699879028, "grad_norm": 0.2722180187702179, "learning_rate": 8e-05, "loss": 1.9241, "step": 464 }, { "epoch": 0.1022764764104256, "grad_norm": 0.2580203115940094, "learning_rate": 8e-05, "loss": 1.693, "step": 465 }, { "epoch": 0.10249642582206092, "grad_norm": 0.2848857641220093, "learning_rate": 8e-05, "loss": 1.9072, "step": 466 }, { "epoch": 0.10271637523369626, "grad_norm": 0.2783052325248718, "learning_rate": 8e-05, "loss": 1.9102, "step": 467 }, { "epoch": 0.10293632464533158, "grad_norm": 0.279695987701416, "learning_rate": 8e-05, "loss": 1.7491, "step": 468 }, { "epoch": 0.1031562740569669, "grad_norm": 0.2493034154176712, "learning_rate": 8e-05, "loss": 1.6789, "step": 469 }, { "epoch": 0.10337622346860222, "grad_norm": 0.2751196622848511, "learning_rate": 8e-05, "loss": 1.8132, "step": 470 }, { "epoch": 0.10359617288023755, "grad_norm": 0.2739677131175995, "learning_rate": 8e-05, "loss": 1.7945, "step": 471 }, { "epoch": 0.10381612229187287, "grad_norm": 0.30357351899147034, "learning_rate": 8e-05, "loss": 1.9113, "step": 472 }, { "epoch": 0.10403607170350819, "grad_norm": 0.2646970748901367, "learning_rate": 8e-05, "loss": 1.811, "step": 473 }, { "epoch": 0.10425602111514351, "grad_norm": 0.2626940608024597, "learning_rate": 8e-05, "loss": 1.6911, "step": 474 }, { "epoch": 0.10447597052677884, "grad_norm": 0.2613508701324463, "learning_rate": 8e-05, "loss": 1.7209, "step": 475 }, { "epoch": 0.10469591993841416, "grad_norm": 0.2609264552593231, "learning_rate": 8e-05, "loss": 1.6303, "step": 476 }, { "epoch": 0.10491586935004948, "grad_norm": 0.2549975514411926, "learning_rate": 8e-05, "loss": 1.7769, "step": 477 }, { "epoch": 0.10513581876168482, "grad_norm": 0.2742570638656616, "learning_rate": 8e-05, "loss": 1.8101, "step": 478 }, { "epoch": 0.10535576817332014, "grad_norm": 0.267070472240448, "learning_rate": 8e-05, "loss": 1.787, "step": 479 }, { "epoch": 0.10557571758495546, "grad_norm": 0.2735085189342499, "learning_rate": 8e-05, "loss": 1.8112, "step": 480 }, { "epoch": 0.10579566699659078, "grad_norm": 0.260111540555954, "learning_rate": 8e-05, "loss": 1.6926, "step": 481 }, { "epoch": 0.10601561640822611, "grad_norm": 0.26309284567832947, "learning_rate": 8e-05, "loss": 1.778, "step": 482 }, { "epoch": 0.10623556581986143, "grad_norm": 0.2658458948135376, "learning_rate": 8e-05, "loss": 1.7179, "step": 483 }, { "epoch": 0.10645551523149675, "grad_norm": 0.27498647570610046, "learning_rate": 8e-05, "loss": 1.6689, "step": 484 }, { "epoch": 0.10667546464313207, "grad_norm": 0.2658367156982422, "learning_rate": 8e-05, "loss": 1.6786, "step": 485 }, { "epoch": 0.10689541405476741, "grad_norm": 0.26023292541503906, "learning_rate": 8e-05, "loss": 1.6995, "step": 486 }, { "epoch": 0.10711536346640273, "grad_norm": 0.25749459862709045, "learning_rate": 8e-05, "loss": 1.6614, "step": 487 }, { "epoch": 0.10733531287803805, "grad_norm": 0.26305267214775085, "learning_rate": 8e-05, "loss": 1.6838, "step": 488 }, { "epoch": 0.10755526228967338, "grad_norm": 0.25277695059776306, "learning_rate": 8e-05, "loss": 1.6975, "step": 489 }, { "epoch": 0.1077752117013087, "grad_norm": 0.2584420144557953, "learning_rate": 8e-05, "loss": 1.7434, "step": 490 }, { "epoch": 0.10799516111294402, "grad_norm": 0.28107360005378723, "learning_rate": 8e-05, "loss": 1.8037, "step": 491 }, { "epoch": 0.10821511052457934, "grad_norm": 0.553341269493103, "learning_rate": 8e-05, "loss": 1.8896, "step": 492 }, { "epoch": 0.10843505993621468, "grad_norm": 0.2718677222728729, "learning_rate": 8e-05, "loss": 1.6646, "step": 493 }, { "epoch": 0.10865500934785, "grad_norm": 0.27301734685897827, "learning_rate": 8e-05, "loss": 1.6663, "step": 494 }, { "epoch": 0.10887495875948532, "grad_norm": 0.26952439546585083, "learning_rate": 8e-05, "loss": 1.7228, "step": 495 }, { "epoch": 0.10909490817112064, "grad_norm": 0.3017599582672119, "learning_rate": 8e-05, "loss": 1.7936, "step": 496 }, { "epoch": 0.10931485758275597, "grad_norm": 0.2676602303981781, "learning_rate": 8e-05, "loss": 1.7861, "step": 497 }, { "epoch": 0.10953480699439129, "grad_norm": 0.27192267775535583, "learning_rate": 8e-05, "loss": 1.8032, "step": 498 }, { "epoch": 0.10975475640602661, "grad_norm": 0.2807183861732483, "learning_rate": 8e-05, "loss": 1.6331, "step": 499 }, { "epoch": 0.10997470581766194, "grad_norm": 0.2652963399887085, "learning_rate": 8e-05, "loss": 1.6231, "step": 500 }, { "epoch": 0.11019465522929726, "grad_norm": 0.26010751724243164, "learning_rate": 8e-05, "loss": 1.729, "step": 501 }, { "epoch": 0.11041460464093258, "grad_norm": 0.29573148488998413, "learning_rate": 8e-05, "loss": 1.8082, "step": 502 }, { "epoch": 0.1106345540525679, "grad_norm": 0.28008025884628296, "learning_rate": 8e-05, "loss": 1.6829, "step": 503 }, { "epoch": 0.11085450346420324, "grad_norm": 0.3029135763645172, "learning_rate": 8e-05, "loss": 1.7699, "step": 504 }, { "epoch": 0.11107445287583856, "grad_norm": 0.2821674346923828, "learning_rate": 8e-05, "loss": 1.7337, "step": 505 }, { "epoch": 0.11129440228747388, "grad_norm": 0.274880975484848, "learning_rate": 8e-05, "loss": 1.7973, "step": 506 }, { "epoch": 0.1115143516991092, "grad_norm": 0.28885796666145325, "learning_rate": 8e-05, "loss": 1.7756, "step": 507 }, { "epoch": 0.11173430111074453, "grad_norm": 0.2744079530239105, "learning_rate": 8e-05, "loss": 1.7991, "step": 508 }, { "epoch": 0.11195425052237985, "grad_norm": 0.2645000219345093, "learning_rate": 8e-05, "loss": 1.6566, "step": 509 }, { "epoch": 0.11217419993401517, "grad_norm": 0.2640466094017029, "learning_rate": 8e-05, "loss": 1.6649, "step": 510 }, { "epoch": 0.11239414934565051, "grad_norm": 0.2965867817401886, "learning_rate": 8e-05, "loss": 1.7733, "step": 511 }, { "epoch": 0.11261409875728583, "grad_norm": 0.2533203661441803, "learning_rate": 8e-05, "loss": 1.7194, "step": 512 }, { "epoch": 0.11283404816892115, "grad_norm": 0.261994868516922, "learning_rate": 8e-05, "loss": 1.7387, "step": 513 }, { "epoch": 0.11305399758055647, "grad_norm": 0.2868165969848633, "learning_rate": 8e-05, "loss": 1.7444, "step": 514 }, { "epoch": 0.1132739469921918, "grad_norm": 0.2836281657218933, "learning_rate": 8e-05, "loss": 1.6507, "step": 515 }, { "epoch": 0.11349389640382712, "grad_norm": 0.28675276041030884, "learning_rate": 8e-05, "loss": 1.7054, "step": 516 }, { "epoch": 0.11371384581546244, "grad_norm": 0.2745465040206909, "learning_rate": 8e-05, "loss": 1.77, "step": 517 }, { "epoch": 0.11393379522709776, "grad_norm": 0.27250972390174866, "learning_rate": 8e-05, "loss": 1.9102, "step": 518 }, { "epoch": 0.1141537446387331, "grad_norm": 0.2781262695789337, "learning_rate": 8e-05, "loss": 1.8126, "step": 519 }, { "epoch": 0.11437369405036842, "grad_norm": 0.2691183388233185, "learning_rate": 8e-05, "loss": 1.5978, "step": 520 }, { "epoch": 0.11459364346200374, "grad_norm": 0.29496780037879944, "learning_rate": 8e-05, "loss": 1.9214, "step": 521 }, { "epoch": 0.11481359287363907, "grad_norm": 0.27725401520729065, "learning_rate": 8e-05, "loss": 1.8722, "step": 522 }, { "epoch": 0.11503354228527439, "grad_norm": 0.28819364309310913, "learning_rate": 8e-05, "loss": 1.6739, "step": 523 }, { "epoch": 0.11525349169690971, "grad_norm": 0.278857946395874, "learning_rate": 8e-05, "loss": 1.8137, "step": 524 }, { "epoch": 0.11547344110854503, "grad_norm": 0.26911258697509766, "learning_rate": 8e-05, "loss": 1.7123, "step": 525 }, { "epoch": 0.11569339052018036, "grad_norm": 0.2656850814819336, "learning_rate": 8e-05, "loss": 1.8124, "step": 526 }, { "epoch": 0.11591333993181568, "grad_norm": 0.26521819829940796, "learning_rate": 8e-05, "loss": 1.8188, "step": 527 }, { "epoch": 0.116133289343451, "grad_norm": 0.2821720540523529, "learning_rate": 8e-05, "loss": 1.7607, "step": 528 }, { "epoch": 0.11635323875508632, "grad_norm": 0.294612854719162, "learning_rate": 8e-05, "loss": 1.8142, "step": 529 }, { "epoch": 0.11657318816672166, "grad_norm": 0.29858094453811646, "learning_rate": 8e-05, "loss": 1.8795, "step": 530 }, { "epoch": 0.11679313757835698, "grad_norm": 0.2726878821849823, "learning_rate": 8e-05, "loss": 1.7988, "step": 531 }, { "epoch": 0.1170130869899923, "grad_norm": 0.2651258111000061, "learning_rate": 8e-05, "loss": 1.8106, "step": 532 }, { "epoch": 0.11723303640162762, "grad_norm": 0.2681291997432709, "learning_rate": 8e-05, "loss": 1.6692, "step": 533 }, { "epoch": 0.11745298581326295, "grad_norm": 0.2641060948371887, "learning_rate": 8e-05, "loss": 1.6479, "step": 534 }, { "epoch": 0.11767293522489827, "grad_norm": 0.2850191593170166, "learning_rate": 8e-05, "loss": 1.7337, "step": 535 }, { "epoch": 0.1178928846365336, "grad_norm": 0.2718667685985565, "learning_rate": 8e-05, "loss": 1.7069, "step": 536 }, { "epoch": 0.11811283404816893, "grad_norm": 0.27950581908226013, "learning_rate": 8e-05, "loss": 1.83, "step": 537 }, { "epoch": 0.11833278345980425, "grad_norm": 0.26720213890075684, "learning_rate": 8e-05, "loss": 1.6787, "step": 538 }, { "epoch": 0.11855273287143957, "grad_norm": 0.25440508127212524, "learning_rate": 8e-05, "loss": 1.5966, "step": 539 }, { "epoch": 0.11877268228307489, "grad_norm": 0.2716729938983917, "learning_rate": 8e-05, "loss": 1.793, "step": 540 }, { "epoch": 0.11899263169471022, "grad_norm": 0.26204821467399597, "learning_rate": 8e-05, "loss": 1.5882, "step": 541 }, { "epoch": 0.11921258110634554, "grad_norm": 0.2756775915622711, "learning_rate": 8e-05, "loss": 1.7529, "step": 542 }, { "epoch": 0.11943253051798086, "grad_norm": 0.27235740423202515, "learning_rate": 8e-05, "loss": 1.7607, "step": 543 }, { "epoch": 0.11965247992961618, "grad_norm": 0.27712538838386536, "learning_rate": 8e-05, "loss": 1.7504, "step": 544 }, { "epoch": 0.11987242934125152, "grad_norm": 0.27800193428993225, "learning_rate": 8e-05, "loss": 1.7421, "step": 545 }, { "epoch": 0.12009237875288684, "grad_norm": 0.27911701798439026, "learning_rate": 8e-05, "loss": 1.6683, "step": 546 }, { "epoch": 0.12031232816452216, "grad_norm": 0.27643364667892456, "learning_rate": 8e-05, "loss": 1.6393, "step": 547 }, { "epoch": 0.12053227757615749, "grad_norm": 0.25785166025161743, "learning_rate": 8e-05, "loss": 1.641, "step": 548 }, { "epoch": 0.12075222698779281, "grad_norm": 0.2791956067085266, "learning_rate": 8e-05, "loss": 1.7585, "step": 549 }, { "epoch": 0.12097217639942813, "grad_norm": 0.28245967626571655, "learning_rate": 8e-05, "loss": 1.8716, "step": 550 }, { "epoch": 0.12119212581106345, "grad_norm": 0.27160346508026123, "learning_rate": 8e-05, "loss": 1.7023, "step": 551 }, { "epoch": 0.12141207522269878, "grad_norm": 0.2670506536960602, "learning_rate": 8e-05, "loss": 1.5844, "step": 552 }, { "epoch": 0.1216320246343341, "grad_norm": 0.2762441337108612, "learning_rate": 8e-05, "loss": 1.7286, "step": 553 }, { "epoch": 0.12185197404596942, "grad_norm": 0.29608720541000366, "learning_rate": 8e-05, "loss": 1.7875, "step": 554 }, { "epoch": 0.12207192345760474, "grad_norm": 0.2847777307033539, "learning_rate": 8e-05, "loss": 1.7388, "step": 555 }, { "epoch": 0.12229187286924008, "grad_norm": 0.2769443988800049, "learning_rate": 8e-05, "loss": 1.8129, "step": 556 }, { "epoch": 0.1225118222808754, "grad_norm": 0.27490487694740295, "learning_rate": 8e-05, "loss": 1.678, "step": 557 }, { "epoch": 0.12273177169251072, "grad_norm": 0.2851822078227997, "learning_rate": 8e-05, "loss": 1.8268, "step": 558 }, { "epoch": 0.12295172110414605, "grad_norm": 0.31336653232574463, "learning_rate": 8e-05, "loss": 1.8247, "step": 559 }, { "epoch": 0.12317167051578137, "grad_norm": 0.26455923914909363, "learning_rate": 8e-05, "loss": 1.5548, "step": 560 }, { "epoch": 0.12339161992741669, "grad_norm": 0.2750054597854614, "learning_rate": 8e-05, "loss": 1.7912, "step": 561 }, { "epoch": 0.12361156933905201, "grad_norm": 0.28016433119773865, "learning_rate": 8e-05, "loss": 1.7367, "step": 562 }, { "epoch": 0.12383151875068735, "grad_norm": 0.30594533681869507, "learning_rate": 8e-05, "loss": 1.7959, "step": 563 }, { "epoch": 0.12405146816232267, "grad_norm": 0.2753421664237976, "learning_rate": 8e-05, "loss": 1.6714, "step": 564 }, { "epoch": 0.12427141757395799, "grad_norm": 0.3309609889984131, "learning_rate": 8e-05, "loss": 1.7632, "step": 565 }, { "epoch": 0.12449136698559331, "grad_norm": 0.3116569221019745, "learning_rate": 8e-05, "loss": 1.8312, "step": 566 }, { "epoch": 0.12471131639722864, "grad_norm": 0.27756184339523315, "learning_rate": 8e-05, "loss": 1.6622, "step": 567 }, { "epoch": 0.12493126580886396, "grad_norm": 0.2740349769592285, "learning_rate": 8e-05, "loss": 1.7015, "step": 568 }, { "epoch": 0.1251512152204993, "grad_norm": 0.2696126401424408, "learning_rate": 8e-05, "loss": 1.6063, "step": 569 }, { "epoch": 0.1253711646321346, "grad_norm": 0.29191461205482483, "learning_rate": 8e-05, "loss": 1.8429, "step": 570 }, { "epoch": 0.12559111404376994, "grad_norm": 0.2984013855457306, "learning_rate": 8e-05, "loss": 1.8194, "step": 571 }, { "epoch": 0.12581106345540527, "grad_norm": 0.27315613627433777, "learning_rate": 8e-05, "loss": 1.7027, "step": 572 }, { "epoch": 0.12603101286704058, "grad_norm": 0.28547149896621704, "learning_rate": 8e-05, "loss": 1.694, "step": 573 }, { "epoch": 0.1262509622786759, "grad_norm": 0.26458805799484253, "learning_rate": 8e-05, "loss": 1.7978, "step": 574 }, { "epoch": 0.12647091169031122, "grad_norm": 0.29676830768585205, "learning_rate": 8e-05, "loss": 1.8295, "step": 575 }, { "epoch": 0.12669086110194655, "grad_norm": 0.28077611327171326, "learning_rate": 8e-05, "loss": 1.7711, "step": 576 }, { "epoch": 0.12691081051358188, "grad_norm": 0.256736159324646, "learning_rate": 8e-05, "loss": 1.5371, "step": 577 }, { "epoch": 0.1271307599252172, "grad_norm": 0.2888578474521637, "learning_rate": 8e-05, "loss": 1.7532, "step": 578 }, { "epoch": 0.12735070933685252, "grad_norm": 0.29349133372306824, "learning_rate": 8e-05, "loss": 1.856, "step": 579 }, { "epoch": 0.12757065874848786, "grad_norm": 0.2626110911369324, "learning_rate": 8e-05, "loss": 1.5482, "step": 580 }, { "epoch": 0.12779060816012316, "grad_norm": 0.2715248167514801, "learning_rate": 8e-05, "loss": 1.7003, "step": 581 }, { "epoch": 0.1280105575717585, "grad_norm": 0.2800534963607788, "learning_rate": 8e-05, "loss": 1.7065, "step": 582 }, { "epoch": 0.12823050698339383, "grad_norm": 0.3190186619758606, "learning_rate": 8e-05, "loss": 1.8099, "step": 583 }, { "epoch": 0.12845045639502914, "grad_norm": 0.2689470648765564, "learning_rate": 8e-05, "loss": 1.7824, "step": 584 }, { "epoch": 0.12867040580666447, "grad_norm": 0.2715473473072052, "learning_rate": 8e-05, "loss": 1.7721, "step": 585 }, { "epoch": 0.12889035521829978, "grad_norm": 0.27956798672676086, "learning_rate": 8e-05, "loss": 1.7888, "step": 586 }, { "epoch": 0.1291103046299351, "grad_norm": 0.2842330038547516, "learning_rate": 8e-05, "loss": 1.7131, "step": 587 }, { "epoch": 0.12933025404157045, "grad_norm": 0.2888692021369934, "learning_rate": 8e-05, "loss": 1.7509, "step": 588 }, { "epoch": 0.12955020345320575, "grad_norm": 0.27673423290252686, "learning_rate": 8e-05, "loss": 1.7235, "step": 589 }, { "epoch": 0.1297701528648411, "grad_norm": 0.26007330417633057, "learning_rate": 8e-05, "loss": 1.7157, "step": 590 }, { "epoch": 0.12999010227647642, "grad_norm": 0.27521616220474243, "learning_rate": 8e-05, "loss": 1.7519, "step": 591 }, { "epoch": 0.13021005168811173, "grad_norm": 0.2753496766090393, "learning_rate": 8e-05, "loss": 1.6956, "step": 592 }, { "epoch": 0.13043000109974706, "grad_norm": 0.25559505820274353, "learning_rate": 8e-05, "loss": 1.522, "step": 593 }, { "epoch": 0.1306499505113824, "grad_norm": 0.26815375685691833, "learning_rate": 8e-05, "loss": 1.7658, "step": 594 }, { "epoch": 0.1308698999230177, "grad_norm": 0.26870042085647583, "learning_rate": 8e-05, "loss": 1.779, "step": 595 }, { "epoch": 0.13108984933465304, "grad_norm": 0.27346327900886536, "learning_rate": 8e-05, "loss": 1.7397, "step": 596 }, { "epoch": 0.13130979874628834, "grad_norm": 0.26674172282218933, "learning_rate": 8e-05, "loss": 1.906, "step": 597 }, { "epoch": 0.13152974815792368, "grad_norm": 0.266916960477829, "learning_rate": 8e-05, "loss": 1.6896, "step": 598 }, { "epoch": 0.131749697569559, "grad_norm": 0.2620035707950592, "learning_rate": 8e-05, "loss": 1.8032, "step": 599 }, { "epoch": 0.13196964698119432, "grad_norm": 0.2721168100833893, "learning_rate": 8e-05, "loss": 1.7992, "step": 600 }, { "epoch": 0.13218959639282965, "grad_norm": 0.2902929186820984, "learning_rate": 8e-05, "loss": 1.8392, "step": 601 }, { "epoch": 0.13240954580446498, "grad_norm": 0.267459899187088, "learning_rate": 8e-05, "loss": 1.8469, "step": 602 }, { "epoch": 0.1326294952161003, "grad_norm": 0.25643131136894226, "learning_rate": 8e-05, "loss": 1.5562, "step": 603 }, { "epoch": 0.13284944462773562, "grad_norm": 0.2919185757637024, "learning_rate": 8e-05, "loss": 1.7108, "step": 604 }, { "epoch": 0.13306939403937096, "grad_norm": 0.2631925046443939, "learning_rate": 8e-05, "loss": 1.4344, "step": 605 }, { "epoch": 0.13328934345100626, "grad_norm": 0.2710738182067871, "learning_rate": 8e-05, "loss": 1.6774, "step": 606 }, { "epoch": 0.1335092928626416, "grad_norm": 0.2641798257827759, "learning_rate": 8e-05, "loss": 1.8032, "step": 607 }, { "epoch": 0.1337292422742769, "grad_norm": 0.2571311891078949, "learning_rate": 8e-05, "loss": 1.63, "step": 608 }, { "epoch": 0.13394919168591224, "grad_norm": 0.24528057873249054, "learning_rate": 8e-05, "loss": 1.4576, "step": 609 }, { "epoch": 0.13416914109754757, "grad_norm": 0.270641028881073, "learning_rate": 8e-05, "loss": 1.7896, "step": 610 }, { "epoch": 0.13438909050918288, "grad_norm": 0.2723008990287781, "learning_rate": 8e-05, "loss": 1.7894, "step": 611 }, { "epoch": 0.1346090399208182, "grad_norm": 0.26487669348716736, "learning_rate": 8e-05, "loss": 1.7646, "step": 612 }, { "epoch": 0.13482898933245355, "grad_norm": 0.26771143078804016, "learning_rate": 8e-05, "loss": 1.8015, "step": 613 }, { "epoch": 0.13504893874408885, "grad_norm": 0.2585919499397278, "learning_rate": 8e-05, "loss": 1.6487, "step": 614 }, { "epoch": 0.1352688881557242, "grad_norm": 0.28161996603012085, "learning_rate": 8e-05, "loss": 1.7813, "step": 615 }, { "epoch": 0.1354888375673595, "grad_norm": 0.25246456265449524, "learning_rate": 8e-05, "loss": 1.5549, "step": 616 }, { "epoch": 0.13570878697899483, "grad_norm": 0.2803630530834198, "learning_rate": 8e-05, "loss": 1.7545, "step": 617 }, { "epoch": 0.13592873639063016, "grad_norm": 0.2587769031524658, "learning_rate": 8e-05, "loss": 1.6755, "step": 618 }, { "epoch": 0.13614868580226547, "grad_norm": 0.2890148162841797, "learning_rate": 8e-05, "loss": 1.9753, "step": 619 }, { "epoch": 0.1363686352139008, "grad_norm": 0.2924948036670685, "learning_rate": 8e-05, "loss": 1.7611, "step": 620 }, { "epoch": 0.13658858462553614, "grad_norm": 0.2594594359397888, "learning_rate": 8e-05, "loss": 1.6945, "step": 621 }, { "epoch": 0.13680853403717144, "grad_norm": 0.2853068709373474, "learning_rate": 8e-05, "loss": 1.8637, "step": 622 }, { "epoch": 0.13702848344880678, "grad_norm": 0.2696111798286438, "learning_rate": 8e-05, "loss": 1.7777, "step": 623 }, { "epoch": 0.1372484328604421, "grad_norm": 0.3137861490249634, "learning_rate": 8e-05, "loss": 1.8799, "step": 624 }, { "epoch": 0.13746838227207742, "grad_norm": 0.25645750761032104, "learning_rate": 8e-05, "loss": 1.5023, "step": 625 }, { "epoch": 0.13768833168371275, "grad_norm": 0.29853489995002747, "learning_rate": 8e-05, "loss": 1.9131, "step": 626 }, { "epoch": 0.13790828109534806, "grad_norm": 0.2653225362300873, "learning_rate": 8e-05, "loss": 1.6835, "step": 627 }, { "epoch": 0.1381282305069834, "grad_norm": 0.26686328649520874, "learning_rate": 8e-05, "loss": 1.7667, "step": 628 }, { "epoch": 0.13834817991861872, "grad_norm": 0.26114073395729065, "learning_rate": 8e-05, "loss": 1.6925, "step": 629 }, { "epoch": 0.13856812933025403, "grad_norm": 0.2520682215690613, "learning_rate": 8e-05, "loss": 1.6065, "step": 630 }, { "epoch": 0.13878807874188936, "grad_norm": 0.2676456868648529, "learning_rate": 8e-05, "loss": 1.7353, "step": 631 }, { "epoch": 0.1390080281535247, "grad_norm": 0.2525452673435211, "learning_rate": 8e-05, "loss": 1.5993, "step": 632 }, { "epoch": 0.13922797756516, "grad_norm": 0.25620371103286743, "learning_rate": 8e-05, "loss": 1.7188, "step": 633 }, { "epoch": 0.13944792697679534, "grad_norm": 0.4071904420852661, "learning_rate": 8e-05, "loss": 1.9348, "step": 634 }, { "epoch": 0.13966787638843067, "grad_norm": 0.2656376361846924, "learning_rate": 8e-05, "loss": 1.7833, "step": 635 }, { "epoch": 0.13988782580006598, "grad_norm": 0.25558993220329285, "learning_rate": 8e-05, "loss": 1.715, "step": 636 }, { "epoch": 0.1401077752117013, "grad_norm": 0.28318601846694946, "learning_rate": 8e-05, "loss": 1.8012, "step": 637 }, { "epoch": 0.14032772462333662, "grad_norm": 0.2558564245700836, "learning_rate": 8e-05, "loss": 1.5802, "step": 638 }, { "epoch": 0.14054767403497195, "grad_norm": 0.26874974370002747, "learning_rate": 8e-05, "loss": 1.7884, "step": 639 }, { "epoch": 0.1407676234466073, "grad_norm": 0.2960795760154724, "learning_rate": 8e-05, "loss": 1.7884, "step": 640 }, { "epoch": 0.1409875728582426, "grad_norm": 0.3098964989185333, "learning_rate": 8e-05, "loss": 1.8844, "step": 641 }, { "epoch": 0.14120752226987793, "grad_norm": 0.2819165885448456, "learning_rate": 8e-05, "loss": 1.7111, "step": 642 }, { "epoch": 0.14142747168151326, "grad_norm": 0.26352617144584656, "learning_rate": 8e-05, "loss": 1.7337, "step": 643 }, { "epoch": 0.14164742109314857, "grad_norm": 0.2622654139995575, "learning_rate": 8e-05, "loss": 1.7284, "step": 644 }, { "epoch": 0.1418673705047839, "grad_norm": 0.2793010473251343, "learning_rate": 8e-05, "loss": 1.8534, "step": 645 }, { "epoch": 0.14208731991641924, "grad_norm": 0.27972397208213806, "learning_rate": 8e-05, "loss": 1.7658, "step": 646 }, { "epoch": 0.14230726932805454, "grad_norm": 0.25940972566604614, "learning_rate": 8e-05, "loss": 1.6676, "step": 647 }, { "epoch": 0.14252721873968988, "grad_norm": 0.29578897356987, "learning_rate": 8e-05, "loss": 1.8002, "step": 648 }, { "epoch": 0.14274716815132518, "grad_norm": 0.2577681541442871, "learning_rate": 8e-05, "loss": 1.6154, "step": 649 }, { "epoch": 0.14296711756296052, "grad_norm": 0.2615002989768982, "learning_rate": 8e-05, "loss": 1.7539, "step": 650 }, { "epoch": 0.14318706697459585, "grad_norm": 0.26044437289237976, "learning_rate": 8e-05, "loss": 1.5284, "step": 651 }, { "epoch": 0.14340701638623116, "grad_norm": 0.28386443853378296, "learning_rate": 8e-05, "loss": 1.7188, "step": 652 }, { "epoch": 0.1436269657978665, "grad_norm": 0.2579086124897003, "learning_rate": 8e-05, "loss": 1.6758, "step": 653 }, { "epoch": 0.14384691520950182, "grad_norm": 0.263192743062973, "learning_rate": 8e-05, "loss": 1.7013, "step": 654 }, { "epoch": 0.14406686462113713, "grad_norm": 0.26551106572151184, "learning_rate": 8e-05, "loss": 1.7314, "step": 655 }, { "epoch": 0.14428681403277246, "grad_norm": 0.26143091917037964, "learning_rate": 8e-05, "loss": 1.7041, "step": 656 }, { "epoch": 0.1445067634444078, "grad_norm": 0.26432663202285767, "learning_rate": 8e-05, "loss": 1.601, "step": 657 }, { "epoch": 0.1447267128560431, "grad_norm": 0.2831920087337494, "learning_rate": 8e-05, "loss": 1.8573, "step": 658 }, { "epoch": 0.14494666226767844, "grad_norm": 0.3045855462551117, "learning_rate": 8e-05, "loss": 1.7853, "step": 659 }, { "epoch": 0.14516661167931375, "grad_norm": 0.28249257802963257, "learning_rate": 8e-05, "loss": 1.7525, "step": 660 }, { "epoch": 0.14538656109094908, "grad_norm": 0.27501189708709717, "learning_rate": 8e-05, "loss": 1.6939, "step": 661 }, { "epoch": 0.1456065105025844, "grad_norm": 0.28419750928878784, "learning_rate": 8e-05, "loss": 1.837, "step": 662 }, { "epoch": 0.14582645991421972, "grad_norm": 0.28872454166412354, "learning_rate": 8e-05, "loss": 1.623, "step": 663 }, { "epoch": 0.14604640932585505, "grad_norm": 0.2926316559314728, "learning_rate": 8e-05, "loss": 1.7438, "step": 664 }, { "epoch": 0.1462663587374904, "grad_norm": 0.2716543972492218, "learning_rate": 8e-05, "loss": 1.8925, "step": 665 }, { "epoch": 0.1464863081491257, "grad_norm": 0.2707289159297943, "learning_rate": 8e-05, "loss": 1.8218, "step": 666 }, { "epoch": 0.14670625756076103, "grad_norm": 0.2609579265117645, "learning_rate": 8e-05, "loss": 1.4612, "step": 667 }, { "epoch": 0.14692620697239636, "grad_norm": 0.2958548367023468, "learning_rate": 8e-05, "loss": 1.6191, "step": 668 }, { "epoch": 0.14714615638403167, "grad_norm": 0.2585492730140686, "learning_rate": 8e-05, "loss": 1.7161, "step": 669 }, { "epoch": 0.147366105795667, "grad_norm": 0.2637808322906494, "learning_rate": 8e-05, "loss": 1.6534, "step": 670 }, { "epoch": 0.1475860552073023, "grad_norm": 0.2885671854019165, "learning_rate": 8e-05, "loss": 1.7663, "step": 671 }, { "epoch": 0.14780600461893764, "grad_norm": 0.27028244733810425, "learning_rate": 8e-05, "loss": 1.7718, "step": 672 }, { "epoch": 0.14802595403057298, "grad_norm": 0.27723586559295654, "learning_rate": 8e-05, "loss": 1.7762, "step": 673 }, { "epoch": 0.14824590344220828, "grad_norm": 0.26336848735809326, "learning_rate": 8e-05, "loss": 1.6114, "step": 674 }, { "epoch": 0.14846585285384362, "grad_norm": 0.26031750440597534, "learning_rate": 8e-05, "loss": 1.7259, "step": 675 }, { "epoch": 0.14868580226547895, "grad_norm": 0.30176040530204773, "learning_rate": 8e-05, "loss": 1.7007, "step": 676 }, { "epoch": 0.14890575167711426, "grad_norm": 0.25952771306037903, "learning_rate": 8e-05, "loss": 1.6573, "step": 677 }, { "epoch": 0.1491257010887496, "grad_norm": 0.2727009356021881, "learning_rate": 8e-05, "loss": 1.7725, "step": 678 }, { "epoch": 0.14934565050038492, "grad_norm": 0.26398420333862305, "learning_rate": 8e-05, "loss": 1.7245, "step": 679 }, { "epoch": 0.14956559991202023, "grad_norm": 0.273967981338501, "learning_rate": 8e-05, "loss": 1.7231, "step": 680 }, { "epoch": 0.14978554932365556, "grad_norm": 0.27241724729537964, "learning_rate": 8e-05, "loss": 1.6896, "step": 681 }, { "epoch": 0.15000549873529087, "grad_norm": 0.26996085047721863, "learning_rate": 8e-05, "loss": 1.6767, "step": 682 }, { "epoch": 0.1502254481469262, "grad_norm": 0.27165672183036804, "learning_rate": 8e-05, "loss": 1.7747, "step": 683 }, { "epoch": 0.15044539755856154, "grad_norm": 0.26840028166770935, "learning_rate": 8e-05, "loss": 1.7616, "step": 684 }, { "epoch": 0.15066534697019685, "grad_norm": 0.27101555466651917, "learning_rate": 8e-05, "loss": 1.622, "step": 685 }, { "epoch": 0.15088529638183218, "grad_norm": 0.2691043019294739, "learning_rate": 8e-05, "loss": 1.7514, "step": 686 }, { "epoch": 0.1511052457934675, "grad_norm": 0.2926357090473175, "learning_rate": 8e-05, "loss": 1.6953, "step": 687 }, { "epoch": 0.15132519520510282, "grad_norm": 0.2730226516723633, "learning_rate": 8e-05, "loss": 1.6286, "step": 688 }, { "epoch": 0.15154514461673815, "grad_norm": 0.2618841826915741, "learning_rate": 8e-05, "loss": 1.7194, "step": 689 }, { "epoch": 0.1517650940283735, "grad_norm": 0.2584119737148285, "learning_rate": 8e-05, "loss": 1.6032, "step": 690 }, { "epoch": 0.1519850434400088, "grad_norm": 0.26063093543052673, "learning_rate": 8e-05, "loss": 1.63, "step": 691 }, { "epoch": 0.15220499285164413, "grad_norm": 0.267938494682312, "learning_rate": 8e-05, "loss": 1.7087, "step": 692 }, { "epoch": 0.15242494226327943, "grad_norm": 0.2709169089794159, "learning_rate": 8e-05, "loss": 1.6663, "step": 693 }, { "epoch": 0.15264489167491477, "grad_norm": 0.3015836775302887, "learning_rate": 8e-05, "loss": 1.6797, "step": 694 }, { "epoch": 0.1528648410865501, "grad_norm": 0.27824944257736206, "learning_rate": 8e-05, "loss": 1.7972, "step": 695 }, { "epoch": 0.1530847904981854, "grad_norm": 0.31089073419570923, "learning_rate": 8e-05, "loss": 1.7352, "step": 696 }, { "epoch": 0.15330473990982074, "grad_norm": 0.2804546654224396, "learning_rate": 8e-05, "loss": 1.6898, "step": 697 }, { "epoch": 0.15352468932145608, "grad_norm": 0.2804514765739441, "learning_rate": 8e-05, "loss": 1.8409, "step": 698 }, { "epoch": 0.15374463873309138, "grad_norm": 0.31666815280914307, "learning_rate": 8e-05, "loss": 1.6569, "step": 699 }, { "epoch": 0.15396458814472672, "grad_norm": 0.2846215069293976, "learning_rate": 8e-05, "loss": 1.8081, "step": 700 }, { "epoch": 0.15418453755636205, "grad_norm": 0.2656068801879883, "learning_rate": 8e-05, "loss": 1.5747, "step": 701 }, { "epoch": 0.15440448696799736, "grad_norm": 0.2633317708969116, "learning_rate": 8e-05, "loss": 1.6027, "step": 702 }, { "epoch": 0.1546244363796327, "grad_norm": 0.2669740319252014, "learning_rate": 8e-05, "loss": 1.6964, "step": 703 }, { "epoch": 0.154844385791268, "grad_norm": 0.2878497540950775, "learning_rate": 8e-05, "loss": 1.677, "step": 704 }, { "epoch": 0.15506433520290333, "grad_norm": 0.2624325156211853, "learning_rate": 8e-05, "loss": 1.6247, "step": 705 }, { "epoch": 0.15528428461453866, "grad_norm": 0.2894291579723358, "learning_rate": 8e-05, "loss": 1.7271, "step": 706 }, { "epoch": 0.15550423402617397, "grad_norm": 0.2924456298351288, "learning_rate": 8e-05, "loss": 1.7475, "step": 707 }, { "epoch": 0.1557241834378093, "grad_norm": 0.2519112229347229, "learning_rate": 8e-05, "loss": 1.6306, "step": 708 }, { "epoch": 0.15594413284944464, "grad_norm": 0.2831405699253082, "learning_rate": 8e-05, "loss": 1.7571, "step": 709 }, { "epoch": 0.15616408226107995, "grad_norm": 0.2804257273674011, "learning_rate": 8e-05, "loss": 1.6721, "step": 710 }, { "epoch": 0.15638403167271528, "grad_norm": 0.27130362391471863, "learning_rate": 8e-05, "loss": 1.7451, "step": 711 }, { "epoch": 0.1566039810843506, "grad_norm": 0.27843937277793884, "learning_rate": 8e-05, "loss": 1.7187, "step": 712 }, { "epoch": 0.15682393049598592, "grad_norm": 0.26205387711524963, "learning_rate": 8e-05, "loss": 1.7667, "step": 713 }, { "epoch": 0.15704387990762125, "grad_norm": 0.25978967547416687, "learning_rate": 8e-05, "loss": 1.6595, "step": 714 }, { "epoch": 0.15726382931925656, "grad_norm": 0.26331478357315063, "learning_rate": 8e-05, "loss": 1.8067, "step": 715 }, { "epoch": 0.1574837787308919, "grad_norm": 0.26023924350738525, "learning_rate": 8e-05, "loss": 1.8533, "step": 716 }, { "epoch": 0.15770372814252723, "grad_norm": 0.27147844433784485, "learning_rate": 8e-05, "loss": 1.6309, "step": 717 }, { "epoch": 0.15792367755416253, "grad_norm": 0.286035418510437, "learning_rate": 8e-05, "loss": 1.72, "step": 718 }, { "epoch": 0.15814362696579787, "grad_norm": 0.3167229890823364, "learning_rate": 8e-05, "loss": 1.9007, "step": 719 }, { "epoch": 0.1583635763774332, "grad_norm": 0.283975750207901, "learning_rate": 8e-05, "loss": 1.6662, "step": 720 }, { "epoch": 0.1585835257890685, "grad_norm": 0.2812137007713318, "learning_rate": 8e-05, "loss": 1.7651, "step": 721 }, { "epoch": 0.15880347520070384, "grad_norm": 0.2737642526626587, "learning_rate": 8e-05, "loss": 1.7679, "step": 722 }, { "epoch": 0.15902342461233915, "grad_norm": 0.30812978744506836, "learning_rate": 8e-05, "loss": 1.8408, "step": 723 }, { "epoch": 0.15924337402397448, "grad_norm": 0.27026352286338806, "learning_rate": 8e-05, "loss": 1.7362, "step": 724 }, { "epoch": 0.15946332343560982, "grad_norm": 0.2788861393928528, "learning_rate": 8e-05, "loss": 1.8371, "step": 725 }, { "epoch": 0.15968327284724512, "grad_norm": 0.2623996138572693, "learning_rate": 8e-05, "loss": 1.5855, "step": 726 }, { "epoch": 0.15990322225888046, "grad_norm": 0.2764820158481598, "learning_rate": 8e-05, "loss": 1.8185, "step": 727 }, { "epoch": 0.1601231716705158, "grad_norm": 0.27394816279411316, "learning_rate": 8e-05, "loss": 1.641, "step": 728 }, { "epoch": 0.1603431210821511, "grad_norm": 0.2726307511329651, "learning_rate": 8e-05, "loss": 1.6128, "step": 729 }, { "epoch": 0.16056307049378643, "grad_norm": 0.28221258521080017, "learning_rate": 8e-05, "loss": 1.8413, "step": 730 }, { "epoch": 0.16078301990542176, "grad_norm": 0.2649543881416321, "learning_rate": 8e-05, "loss": 1.5707, "step": 731 }, { "epoch": 0.16100296931705707, "grad_norm": 0.2659435570240021, "learning_rate": 8e-05, "loss": 1.6761, "step": 732 }, { "epoch": 0.1612229187286924, "grad_norm": 0.3131570518016815, "learning_rate": 8e-05, "loss": 1.9439, "step": 733 }, { "epoch": 0.1614428681403277, "grad_norm": 0.263069748878479, "learning_rate": 8e-05, "loss": 1.7069, "step": 734 }, { "epoch": 0.16166281755196305, "grad_norm": 0.2708505392074585, "learning_rate": 8e-05, "loss": 1.8031, "step": 735 }, { "epoch": 0.16188276696359838, "grad_norm": 0.26446613669395447, "learning_rate": 8e-05, "loss": 1.6419, "step": 736 }, { "epoch": 0.16210271637523369, "grad_norm": 0.27720367908477783, "learning_rate": 8e-05, "loss": 1.8291, "step": 737 }, { "epoch": 0.16232266578686902, "grad_norm": 0.25950226187705994, "learning_rate": 8e-05, "loss": 1.7498, "step": 738 }, { "epoch": 0.16254261519850435, "grad_norm": 0.25445327162742615, "learning_rate": 8e-05, "loss": 1.6804, "step": 739 }, { "epoch": 0.16276256461013966, "grad_norm": 0.2868766784667969, "learning_rate": 8e-05, "loss": 1.8058, "step": 740 }, { "epoch": 0.162982514021775, "grad_norm": 0.2775559425354004, "learning_rate": 8e-05, "loss": 1.7971, "step": 741 }, { "epoch": 0.16320246343341033, "grad_norm": 0.2822381556034088, "learning_rate": 8e-05, "loss": 1.7294, "step": 742 }, { "epoch": 0.16342241284504563, "grad_norm": 0.26617857813835144, "learning_rate": 8e-05, "loss": 1.8011, "step": 743 }, { "epoch": 0.16364236225668097, "grad_norm": 0.25615090131759644, "learning_rate": 8e-05, "loss": 1.6328, "step": 744 }, { "epoch": 0.16386231166831627, "grad_norm": 0.25831338763237, "learning_rate": 8e-05, "loss": 1.6174, "step": 745 }, { "epoch": 0.1640822610799516, "grad_norm": 0.2707291543483734, "learning_rate": 8e-05, "loss": 1.8217, "step": 746 }, { "epoch": 0.16430221049158694, "grad_norm": 0.3028862774372101, "learning_rate": 8e-05, "loss": 1.5852, "step": 747 }, { "epoch": 0.16452215990322225, "grad_norm": 0.26598575711250305, "learning_rate": 8e-05, "loss": 1.7213, "step": 748 }, { "epoch": 0.16474210931485758, "grad_norm": 0.27408871054649353, "learning_rate": 8e-05, "loss": 1.7109, "step": 749 }, { "epoch": 0.16496205872649292, "grad_norm": 0.27065837383270264, "learning_rate": 8e-05, "loss": 1.6696, "step": 750 }, { "epoch": 0.16518200813812822, "grad_norm": 0.2721879184246063, "learning_rate": 8e-05, "loss": 1.7055, "step": 751 }, { "epoch": 0.16540195754976356, "grad_norm": 0.29569125175476074, "learning_rate": 8e-05, "loss": 1.5921, "step": 752 }, { "epoch": 0.1656219069613989, "grad_norm": 0.28580978512763977, "learning_rate": 8e-05, "loss": 1.7518, "step": 753 }, { "epoch": 0.1658418563730342, "grad_norm": 0.2869469225406647, "learning_rate": 8e-05, "loss": 1.8164, "step": 754 }, { "epoch": 0.16606180578466953, "grad_norm": 0.2796071171760559, "learning_rate": 8e-05, "loss": 1.8325, "step": 755 }, { "epoch": 0.16628175519630484, "grad_norm": 0.27365031838417053, "learning_rate": 8e-05, "loss": 1.7287, "step": 756 }, { "epoch": 0.16650170460794017, "grad_norm": 0.2524491846561432, "learning_rate": 8e-05, "loss": 1.5379, "step": 757 }, { "epoch": 0.1667216540195755, "grad_norm": 0.259860634803772, "learning_rate": 8e-05, "loss": 1.5204, "step": 758 }, { "epoch": 0.1669416034312108, "grad_norm": 0.2714100182056427, "learning_rate": 8e-05, "loss": 1.7245, "step": 759 }, { "epoch": 0.16716155284284615, "grad_norm": 0.2729417383670807, "learning_rate": 8e-05, "loss": 1.6889, "step": 760 }, { "epoch": 0.16738150225448148, "grad_norm": 0.2753896415233612, "learning_rate": 8e-05, "loss": 1.7345, "step": 761 }, { "epoch": 0.16760145166611679, "grad_norm": 0.2830727994441986, "learning_rate": 8e-05, "loss": 1.6884, "step": 762 }, { "epoch": 0.16782140107775212, "grad_norm": 0.27818116545677185, "learning_rate": 8e-05, "loss": 1.7819, "step": 763 }, { "epoch": 0.16804135048938745, "grad_norm": 0.2601570785045624, "learning_rate": 8e-05, "loss": 1.6323, "step": 764 }, { "epoch": 0.16826129990102276, "grad_norm": 0.2638706564903259, "learning_rate": 8e-05, "loss": 1.5957, "step": 765 }, { "epoch": 0.1684812493126581, "grad_norm": 0.2798631489276886, "learning_rate": 8e-05, "loss": 1.7946, "step": 766 }, { "epoch": 0.1687011987242934, "grad_norm": 0.2975100874900818, "learning_rate": 8e-05, "loss": 1.871, "step": 767 }, { "epoch": 0.16892114813592873, "grad_norm": 0.28308364748954773, "learning_rate": 8e-05, "loss": 1.7184, "step": 768 }, { "epoch": 0.16914109754756407, "grad_norm": 0.2594911456108093, "learning_rate": 8e-05, "loss": 1.5867, "step": 769 }, { "epoch": 0.16936104695919937, "grad_norm": 0.27594470977783203, "learning_rate": 8e-05, "loss": 1.7722, "step": 770 }, { "epoch": 0.1695809963708347, "grad_norm": 0.2783298194408417, "learning_rate": 8e-05, "loss": 1.7891, "step": 771 }, { "epoch": 0.16980094578247004, "grad_norm": 0.2863733172416687, "learning_rate": 8e-05, "loss": 1.6274, "step": 772 }, { "epoch": 0.17002089519410535, "grad_norm": 0.27953147888183594, "learning_rate": 8e-05, "loss": 1.7287, "step": 773 }, { "epoch": 0.17024084460574068, "grad_norm": 0.2736772894859314, "learning_rate": 8e-05, "loss": 1.6802, "step": 774 }, { "epoch": 0.17046079401737602, "grad_norm": 0.27663713693618774, "learning_rate": 8e-05, "loss": 1.6607, "step": 775 }, { "epoch": 0.17068074342901132, "grad_norm": 0.3064086437225342, "learning_rate": 8e-05, "loss": 1.8244, "step": 776 }, { "epoch": 0.17090069284064666, "grad_norm": 0.29848581552505493, "learning_rate": 8e-05, "loss": 1.7702, "step": 777 }, { "epoch": 0.17112064225228196, "grad_norm": 0.3101220726966858, "learning_rate": 8e-05, "loss": 1.7714, "step": 778 }, { "epoch": 0.1713405916639173, "grad_norm": 0.2754581868648529, "learning_rate": 8e-05, "loss": 1.6367, "step": 779 }, { "epoch": 0.17156054107555263, "grad_norm": 0.2706362307071686, "learning_rate": 8e-05, "loss": 1.6236, "step": 780 }, { "epoch": 0.17178049048718794, "grad_norm": 0.29135438799858093, "learning_rate": 8e-05, "loss": 1.8478, "step": 781 }, { "epoch": 0.17200043989882327, "grad_norm": 0.2751868963241577, "learning_rate": 8e-05, "loss": 1.751, "step": 782 }, { "epoch": 0.1722203893104586, "grad_norm": 0.2871004045009613, "learning_rate": 8e-05, "loss": 1.6793, "step": 783 }, { "epoch": 0.1724403387220939, "grad_norm": 0.31024861335754395, "learning_rate": 8e-05, "loss": 1.7419, "step": 784 }, { "epoch": 0.17266028813372924, "grad_norm": 0.2917722165584564, "learning_rate": 8e-05, "loss": 1.8913, "step": 785 }, { "epoch": 0.17288023754536458, "grad_norm": 0.25443291664123535, "learning_rate": 8e-05, "loss": 1.6991, "step": 786 }, { "epoch": 0.17310018695699989, "grad_norm": 0.2827921211719513, "learning_rate": 8e-05, "loss": 1.8408, "step": 787 }, { "epoch": 0.17332013636863522, "grad_norm": 0.26190435886383057, "learning_rate": 8e-05, "loss": 1.6841, "step": 788 }, { "epoch": 0.17354008578027053, "grad_norm": 0.31557098031044006, "learning_rate": 8e-05, "loss": 1.8838, "step": 789 }, { "epoch": 0.17376003519190586, "grad_norm": 0.27622002363204956, "learning_rate": 8e-05, "loss": 1.6558, "step": 790 }, { "epoch": 0.1739799846035412, "grad_norm": 0.3161294758319855, "learning_rate": 8e-05, "loss": 1.5771, "step": 791 }, { "epoch": 0.1741999340151765, "grad_norm": 0.3014603555202484, "learning_rate": 8e-05, "loss": 1.7742, "step": 792 }, { "epoch": 0.17441988342681183, "grad_norm": 0.24996457993984222, "learning_rate": 8e-05, "loss": 1.5667, "step": 793 }, { "epoch": 0.17463983283844717, "grad_norm": 0.29180648922920227, "learning_rate": 8e-05, "loss": 1.7703, "step": 794 }, { "epoch": 0.17485978225008247, "grad_norm": 0.26707547903060913, "learning_rate": 8e-05, "loss": 1.7964, "step": 795 }, { "epoch": 0.1750797316617178, "grad_norm": 0.24924349784851074, "learning_rate": 8e-05, "loss": 1.6619, "step": 796 }, { "epoch": 0.17529968107335314, "grad_norm": 0.29872292280197144, "learning_rate": 8e-05, "loss": 1.8561, "step": 797 }, { "epoch": 0.17551963048498845, "grad_norm": 0.2770175337791443, "learning_rate": 8e-05, "loss": 1.6352, "step": 798 }, { "epoch": 0.17573957989662378, "grad_norm": 0.26890453696250916, "learning_rate": 8e-05, "loss": 1.885, "step": 799 }, { "epoch": 0.1759595293082591, "grad_norm": 0.2830483317375183, "learning_rate": 8e-05, "loss": 1.6029, "step": 800 }, { "epoch": 0.17617947871989442, "grad_norm": 0.27421921491622925, "learning_rate": 8e-05, "loss": 1.6845, "step": 801 }, { "epoch": 0.17639942813152976, "grad_norm": 0.29273220896720886, "learning_rate": 8e-05, "loss": 1.8135, "step": 802 }, { "epoch": 0.17661937754316506, "grad_norm": 0.2675575315952301, "learning_rate": 8e-05, "loss": 1.571, "step": 803 }, { "epoch": 0.1768393269548004, "grad_norm": 0.2821138799190521, "learning_rate": 8e-05, "loss": 1.9244, "step": 804 }, { "epoch": 0.17705927636643573, "grad_norm": 0.28082311153411865, "learning_rate": 8e-05, "loss": 1.7395, "step": 805 }, { "epoch": 0.17727922577807104, "grad_norm": 0.27897313237190247, "learning_rate": 8e-05, "loss": 1.6347, "step": 806 }, { "epoch": 0.17749917518970637, "grad_norm": 0.27358707785606384, "learning_rate": 8e-05, "loss": 1.7643, "step": 807 }, { "epoch": 0.1777191246013417, "grad_norm": 0.284059077501297, "learning_rate": 8e-05, "loss": 1.5789, "step": 808 }, { "epoch": 0.177939074012977, "grad_norm": 0.26125824451446533, "learning_rate": 8e-05, "loss": 1.7029, "step": 809 }, { "epoch": 0.17815902342461234, "grad_norm": 0.26438888907432556, "learning_rate": 8e-05, "loss": 1.6424, "step": 810 }, { "epoch": 0.17837897283624765, "grad_norm": 0.2746163010597229, "learning_rate": 8e-05, "loss": 1.7992, "step": 811 }, { "epoch": 0.17859892224788299, "grad_norm": 0.27717527747154236, "learning_rate": 8e-05, "loss": 1.7603, "step": 812 }, { "epoch": 0.17881887165951832, "grad_norm": 0.28336596488952637, "learning_rate": 8e-05, "loss": 1.7133, "step": 813 }, { "epoch": 0.17903882107115363, "grad_norm": 0.2701306939125061, "learning_rate": 8e-05, "loss": 1.7724, "step": 814 }, { "epoch": 0.17925877048278896, "grad_norm": 0.2807336449623108, "learning_rate": 8e-05, "loss": 1.7832, "step": 815 }, { "epoch": 0.1794787198944243, "grad_norm": 0.2847912907600403, "learning_rate": 8e-05, "loss": 1.7042, "step": 816 }, { "epoch": 0.1796986693060596, "grad_norm": 0.2836345434188843, "learning_rate": 8e-05, "loss": 1.8506, "step": 817 }, { "epoch": 0.17991861871769493, "grad_norm": 0.30620551109313965, "learning_rate": 8e-05, "loss": 1.7695, "step": 818 }, { "epoch": 0.18013856812933027, "grad_norm": 0.2698993980884552, "learning_rate": 8e-05, "loss": 1.6388, "step": 819 }, { "epoch": 0.18035851754096557, "grad_norm": 0.2937266528606415, "learning_rate": 8e-05, "loss": 1.8648, "step": 820 }, { "epoch": 0.1805784669526009, "grad_norm": 0.2661988139152527, "learning_rate": 8e-05, "loss": 1.7563, "step": 821 }, { "epoch": 0.18079841636423621, "grad_norm": 0.2944018840789795, "learning_rate": 8e-05, "loss": 1.882, "step": 822 }, { "epoch": 0.18101836577587155, "grad_norm": 0.2774435579776764, "learning_rate": 8e-05, "loss": 1.8117, "step": 823 }, { "epoch": 0.18123831518750688, "grad_norm": 0.27865204215049744, "learning_rate": 8e-05, "loss": 1.8815, "step": 824 }, { "epoch": 0.1814582645991422, "grad_norm": 0.26444011926651, "learning_rate": 8e-05, "loss": 1.5844, "step": 825 }, { "epoch": 0.18167821401077752, "grad_norm": 0.27044716477394104, "learning_rate": 8e-05, "loss": 1.7403, "step": 826 }, { "epoch": 0.18189816342241286, "grad_norm": 0.28727805614471436, "learning_rate": 8e-05, "loss": 1.8556, "step": 827 }, { "epoch": 0.18211811283404816, "grad_norm": 0.26131972670555115, "learning_rate": 8e-05, "loss": 1.7727, "step": 828 }, { "epoch": 0.1823380622456835, "grad_norm": 0.269638329744339, "learning_rate": 8e-05, "loss": 1.6795, "step": 829 }, { "epoch": 0.18255801165731883, "grad_norm": 0.2671653628349304, "learning_rate": 8e-05, "loss": 1.6811, "step": 830 }, { "epoch": 0.18277796106895414, "grad_norm": 0.2659014165401459, "learning_rate": 8e-05, "loss": 1.7166, "step": 831 }, { "epoch": 0.18299791048058947, "grad_norm": 0.2719801962375641, "learning_rate": 8e-05, "loss": 1.6938, "step": 832 }, { "epoch": 0.18321785989222478, "grad_norm": 0.3272366225719452, "learning_rate": 8e-05, "loss": 1.8213, "step": 833 }, { "epoch": 0.1834378093038601, "grad_norm": 0.2635113000869751, "learning_rate": 8e-05, "loss": 1.6291, "step": 834 }, { "epoch": 0.18365775871549544, "grad_norm": 0.29401281476020813, "learning_rate": 8e-05, "loss": 1.8234, "step": 835 }, { "epoch": 0.18387770812713075, "grad_norm": 0.29188451170921326, "learning_rate": 8e-05, "loss": 1.7359, "step": 836 }, { "epoch": 0.18409765753876609, "grad_norm": 0.2688080072402954, "learning_rate": 8e-05, "loss": 1.7088, "step": 837 }, { "epoch": 0.18431760695040142, "grad_norm": 0.27907344698905945, "learning_rate": 8e-05, "loss": 1.6762, "step": 838 }, { "epoch": 0.18453755636203673, "grad_norm": 0.2875908315181732, "learning_rate": 8e-05, "loss": 1.7612, "step": 839 }, { "epoch": 0.18475750577367206, "grad_norm": 0.2683177888393402, "learning_rate": 8e-05, "loss": 1.5965, "step": 840 }, { "epoch": 0.18497745518530737, "grad_norm": 0.29948660731315613, "learning_rate": 8e-05, "loss": 1.7358, "step": 841 }, { "epoch": 0.1851974045969427, "grad_norm": 0.28153204917907715, "learning_rate": 8e-05, "loss": 1.8089, "step": 842 }, { "epoch": 0.18541735400857803, "grad_norm": 0.29185283184051514, "learning_rate": 8e-05, "loss": 1.8357, "step": 843 }, { "epoch": 0.18563730342021334, "grad_norm": 0.27565860748291016, "learning_rate": 8e-05, "loss": 1.9565, "step": 844 }, { "epoch": 0.18585725283184867, "grad_norm": 0.2811479866504669, "learning_rate": 8e-05, "loss": 1.8493, "step": 845 }, { "epoch": 0.186077202243484, "grad_norm": 0.271893173456192, "learning_rate": 8e-05, "loss": 1.7622, "step": 846 }, { "epoch": 0.18629715165511931, "grad_norm": 0.26383113861083984, "learning_rate": 8e-05, "loss": 1.7392, "step": 847 }, { "epoch": 0.18651710106675465, "grad_norm": 0.2863881289958954, "learning_rate": 8e-05, "loss": 1.7367, "step": 848 }, { "epoch": 0.18673705047838998, "grad_norm": 0.28036433458328247, "learning_rate": 8e-05, "loss": 1.6587, "step": 849 }, { "epoch": 0.1869569998900253, "grad_norm": 0.2938581705093384, "learning_rate": 8e-05, "loss": 1.7411, "step": 850 }, { "epoch": 0.18717694930166062, "grad_norm": 0.27487799525260925, "learning_rate": 8e-05, "loss": 1.8054, "step": 851 }, { "epoch": 0.18739689871329593, "grad_norm": 0.2693670690059662, "learning_rate": 8e-05, "loss": 1.7361, "step": 852 }, { "epoch": 0.18761684812493126, "grad_norm": 0.2999705970287323, "learning_rate": 8e-05, "loss": 1.909, "step": 853 }, { "epoch": 0.1878367975365666, "grad_norm": 0.28235265612602234, "learning_rate": 8e-05, "loss": 1.8611, "step": 854 }, { "epoch": 0.1880567469482019, "grad_norm": 0.28417298197746277, "learning_rate": 8e-05, "loss": 1.683, "step": 855 }, { "epoch": 0.18827669635983724, "grad_norm": 0.2697356045246124, "learning_rate": 8e-05, "loss": 1.7138, "step": 856 }, { "epoch": 0.18849664577147257, "grad_norm": 0.26900357007980347, "learning_rate": 8e-05, "loss": 1.5579, "step": 857 }, { "epoch": 0.18871659518310788, "grad_norm": 0.259941041469574, "learning_rate": 8e-05, "loss": 1.7106, "step": 858 }, { "epoch": 0.1889365445947432, "grad_norm": 0.26958781480789185, "learning_rate": 8e-05, "loss": 1.6454, "step": 859 }, { "epoch": 0.18915649400637854, "grad_norm": 0.26425305008888245, "learning_rate": 8e-05, "loss": 1.6408, "step": 860 }, { "epoch": 0.18937644341801385, "grad_norm": 0.26996907591819763, "learning_rate": 8e-05, "loss": 1.6949, "step": 861 }, { "epoch": 0.18959639282964919, "grad_norm": 0.25882837176322937, "learning_rate": 8e-05, "loss": 1.6142, "step": 862 }, { "epoch": 0.1898163422412845, "grad_norm": 0.28000783920288086, "learning_rate": 8e-05, "loss": 1.8007, "step": 863 }, { "epoch": 0.19003629165291983, "grad_norm": 0.2744222581386566, "learning_rate": 8e-05, "loss": 1.6604, "step": 864 }, { "epoch": 0.19025624106455516, "grad_norm": 0.2791576683521271, "learning_rate": 8e-05, "loss": 1.7061, "step": 865 }, { "epoch": 0.19047619047619047, "grad_norm": 0.27878084778785706, "learning_rate": 8e-05, "loss": 1.8604, "step": 866 }, { "epoch": 0.1906961398878258, "grad_norm": 0.3818608820438385, "learning_rate": 8e-05, "loss": 1.8616, "step": 867 }, { "epoch": 0.19091608929946113, "grad_norm": 0.27952665090560913, "learning_rate": 8e-05, "loss": 1.7616, "step": 868 }, { "epoch": 0.19113603871109644, "grad_norm": 0.2711832523345947, "learning_rate": 8e-05, "loss": 1.7974, "step": 869 }, { "epoch": 0.19135598812273177, "grad_norm": 0.2572176456451416, "learning_rate": 8e-05, "loss": 1.584, "step": 870 }, { "epoch": 0.1915759375343671, "grad_norm": 0.2847760319709778, "learning_rate": 8e-05, "loss": 1.8598, "step": 871 }, { "epoch": 0.19179588694600241, "grad_norm": 0.29798731207847595, "learning_rate": 8e-05, "loss": 1.6689, "step": 872 }, { "epoch": 0.19201583635763775, "grad_norm": 0.2674097716808319, "learning_rate": 8e-05, "loss": 1.6694, "step": 873 }, { "epoch": 0.19223578576927305, "grad_norm": 0.27707335352897644, "learning_rate": 8e-05, "loss": 1.7251, "step": 874 }, { "epoch": 0.1924557351809084, "grad_norm": 0.2801666259765625, "learning_rate": 8e-05, "loss": 1.7251, "step": 875 }, { "epoch": 0.19267568459254372, "grad_norm": 0.2656191885471344, "learning_rate": 8e-05, "loss": 1.6232, "step": 876 }, { "epoch": 0.19289563400417903, "grad_norm": 0.2588733732700348, "learning_rate": 8e-05, "loss": 1.7595, "step": 877 }, { "epoch": 0.19311558341581436, "grad_norm": 0.2999958097934723, "learning_rate": 8e-05, "loss": 1.9095, "step": 878 }, { "epoch": 0.1933355328274497, "grad_norm": 0.27143120765686035, "learning_rate": 8e-05, "loss": 1.6698, "step": 879 }, { "epoch": 0.193555482239085, "grad_norm": 0.29155731201171875, "learning_rate": 8e-05, "loss": 1.6437, "step": 880 }, { "epoch": 0.19377543165072034, "grad_norm": 0.26307716965675354, "learning_rate": 8e-05, "loss": 1.6161, "step": 881 }, { "epoch": 0.19399538106235567, "grad_norm": 0.27041196823120117, "learning_rate": 8e-05, "loss": 1.5374, "step": 882 }, { "epoch": 0.19421533047399098, "grad_norm": 0.2752692699432373, "learning_rate": 8e-05, "loss": 1.6543, "step": 883 }, { "epoch": 0.1944352798856263, "grad_norm": 0.2883388102054596, "learning_rate": 8e-05, "loss": 1.7503, "step": 884 }, { "epoch": 0.19465522929726162, "grad_norm": 0.27332282066345215, "learning_rate": 8e-05, "loss": 1.8456, "step": 885 }, { "epoch": 0.19487517870889695, "grad_norm": 0.26226627826690674, "learning_rate": 8e-05, "loss": 1.2577, "step": 886 }, { "epoch": 0.19509512812053229, "grad_norm": 0.2709749639034271, "learning_rate": 8e-05, "loss": 1.7854, "step": 887 }, { "epoch": 0.1953150775321676, "grad_norm": 0.28380879759788513, "learning_rate": 8e-05, "loss": 1.7151, "step": 888 }, { "epoch": 0.19553502694380293, "grad_norm": 0.2702254354953766, "learning_rate": 8e-05, "loss": 1.6779, "step": 889 }, { "epoch": 0.19575497635543826, "grad_norm": 0.2620486617088318, "learning_rate": 8e-05, "loss": 1.7166, "step": 890 }, { "epoch": 0.19597492576707357, "grad_norm": 0.27195873856544495, "learning_rate": 8e-05, "loss": 1.7263, "step": 891 }, { "epoch": 0.1961948751787089, "grad_norm": 0.2719867527484894, "learning_rate": 8e-05, "loss": 1.6982, "step": 892 }, { "epoch": 0.19641482459034423, "grad_norm": 0.27889111638069153, "learning_rate": 8e-05, "loss": 1.7726, "step": 893 }, { "epoch": 0.19663477400197954, "grad_norm": 0.2745397686958313, "learning_rate": 8e-05, "loss": 1.7737, "step": 894 }, { "epoch": 0.19685472341361487, "grad_norm": 0.2698670029640198, "learning_rate": 8e-05, "loss": 1.7149, "step": 895 }, { "epoch": 0.19707467282525018, "grad_norm": 0.27113667130470276, "learning_rate": 8e-05, "loss": 1.7887, "step": 896 }, { "epoch": 0.19729462223688551, "grad_norm": 0.2772979140281677, "learning_rate": 8e-05, "loss": 1.8163, "step": 897 }, { "epoch": 0.19751457164852085, "grad_norm": 0.2757657766342163, "learning_rate": 8e-05, "loss": 1.636, "step": 898 }, { "epoch": 0.19773452106015615, "grad_norm": 0.26945242285728455, "learning_rate": 8e-05, "loss": 1.7639, "step": 899 }, { "epoch": 0.1979544704717915, "grad_norm": 0.27328991889953613, "learning_rate": 8e-05, "loss": 1.8421, "step": 900 }, { "epoch": 0.19817441988342682, "grad_norm": 0.2721468210220337, "learning_rate": 8e-05, "loss": 1.6926, "step": 901 }, { "epoch": 0.19839436929506213, "grad_norm": 0.2633766233921051, "learning_rate": 8e-05, "loss": 1.7629, "step": 902 }, { "epoch": 0.19861431870669746, "grad_norm": 0.26183879375457764, "learning_rate": 8e-05, "loss": 1.7149, "step": 903 }, { "epoch": 0.1988342681183328, "grad_norm": 0.2837960422039032, "learning_rate": 8e-05, "loss": 1.7923, "step": 904 }, { "epoch": 0.1990542175299681, "grad_norm": 0.30745571851730347, "learning_rate": 8e-05, "loss": 1.8315, "step": 905 }, { "epoch": 0.19927416694160344, "grad_norm": 0.2734341323375702, "learning_rate": 8e-05, "loss": 1.7424, "step": 906 }, { "epoch": 0.19949411635323874, "grad_norm": 0.2613460123538971, "learning_rate": 8e-05, "loss": 1.6445, "step": 907 }, { "epoch": 0.19971406576487408, "grad_norm": 0.27867522835731506, "learning_rate": 8e-05, "loss": 1.7075, "step": 908 }, { "epoch": 0.1999340151765094, "grad_norm": 0.269789457321167, "learning_rate": 8e-05, "loss": 1.6801, "step": 909 }, { "epoch": 0.20015396458814472, "grad_norm": 0.2684427797794342, "learning_rate": 8e-05, "loss": 1.6862, "step": 910 }, { "epoch": 0.20037391399978005, "grad_norm": 0.2929883897304535, "learning_rate": 8e-05, "loss": 1.7972, "step": 911 }, { "epoch": 0.20059386341141539, "grad_norm": 0.2757764756679535, "learning_rate": 8e-05, "loss": 1.6198, "step": 912 }, { "epoch": 0.2008138128230507, "grad_norm": 0.28071129322052, "learning_rate": 8e-05, "loss": 1.8268, "step": 913 }, { "epoch": 0.20103376223468603, "grad_norm": 0.2964448928833008, "learning_rate": 8e-05, "loss": 1.7806, "step": 914 }, { "epoch": 0.20125371164632136, "grad_norm": 0.2682490050792694, "learning_rate": 8e-05, "loss": 1.7016, "step": 915 }, { "epoch": 0.20147366105795667, "grad_norm": 0.2838338613510132, "learning_rate": 8e-05, "loss": 1.7402, "step": 916 }, { "epoch": 0.201693610469592, "grad_norm": 0.27621790766716003, "learning_rate": 8e-05, "loss": 1.7442, "step": 917 }, { "epoch": 0.2019135598812273, "grad_norm": 0.29265734553337097, "learning_rate": 8e-05, "loss": 1.6924, "step": 918 }, { "epoch": 0.20213350929286264, "grad_norm": 0.27404630184173584, "learning_rate": 8e-05, "loss": 1.7312, "step": 919 }, { "epoch": 0.20235345870449797, "grad_norm": 0.2742730975151062, "learning_rate": 8e-05, "loss": 1.8645, "step": 920 }, { "epoch": 0.20257340811613328, "grad_norm": 0.28536343574523926, "learning_rate": 8e-05, "loss": 1.717, "step": 921 }, { "epoch": 0.20279335752776861, "grad_norm": 0.28739288449287415, "learning_rate": 8e-05, "loss": 1.8356, "step": 922 }, { "epoch": 0.20301330693940395, "grad_norm": 0.2650564908981323, "learning_rate": 8e-05, "loss": 1.717, "step": 923 }, { "epoch": 0.20323325635103925, "grad_norm": 0.28638410568237305, "learning_rate": 8e-05, "loss": 1.8822, "step": 924 }, { "epoch": 0.2034532057626746, "grad_norm": 0.25474488735198975, "learning_rate": 8e-05, "loss": 1.5533, "step": 925 }, { "epoch": 0.20367315517430992, "grad_norm": 0.2719588279724121, "learning_rate": 8e-05, "loss": 1.7887, "step": 926 }, { "epoch": 0.20389310458594523, "grad_norm": 0.2572193741798401, "learning_rate": 8e-05, "loss": 1.5735, "step": 927 }, { "epoch": 0.20411305399758056, "grad_norm": 0.2975933253765106, "learning_rate": 8e-05, "loss": 1.7875, "step": 928 }, { "epoch": 0.20433300340921587, "grad_norm": 0.2562117874622345, "learning_rate": 8e-05, "loss": 1.5977, "step": 929 }, { "epoch": 0.2045529528208512, "grad_norm": 0.2524821162223816, "learning_rate": 8e-05, "loss": 1.6356, "step": 930 }, { "epoch": 0.20477290223248654, "grad_norm": 0.2621130347251892, "learning_rate": 8e-05, "loss": 1.6082, "step": 931 }, { "epoch": 0.20499285164412184, "grad_norm": 0.27930355072021484, "learning_rate": 8e-05, "loss": 1.7618, "step": 932 }, { "epoch": 0.20521280105575718, "grad_norm": 0.29147934913635254, "learning_rate": 8e-05, "loss": 1.8223, "step": 933 }, { "epoch": 0.2054327504673925, "grad_norm": 0.2584928870201111, "learning_rate": 8e-05, "loss": 1.6916, "step": 934 }, { "epoch": 0.20565269987902782, "grad_norm": 0.27299705147743225, "learning_rate": 8e-05, "loss": 1.5535, "step": 935 }, { "epoch": 0.20587264929066315, "grad_norm": 0.2682443857192993, "learning_rate": 8e-05, "loss": 1.7119, "step": 936 }, { "epoch": 0.20609259870229849, "grad_norm": 0.29716598987579346, "learning_rate": 8e-05, "loss": 2.0561, "step": 937 }, { "epoch": 0.2063125481139338, "grad_norm": 0.27801281213760376, "learning_rate": 8e-05, "loss": 1.7605, "step": 938 }, { "epoch": 0.20653249752556913, "grad_norm": 0.26767662167549133, "learning_rate": 8e-05, "loss": 1.6462, "step": 939 }, { "epoch": 0.20675244693720443, "grad_norm": 0.27354639768600464, "learning_rate": 8e-05, "loss": 1.7241, "step": 940 }, { "epoch": 0.20697239634883977, "grad_norm": 0.2684631049633026, "learning_rate": 8e-05, "loss": 1.8066, "step": 941 }, { "epoch": 0.2071923457604751, "grad_norm": 0.27846816182136536, "learning_rate": 8e-05, "loss": 1.7553, "step": 942 }, { "epoch": 0.2074122951721104, "grad_norm": 0.2820284366607666, "learning_rate": 8e-05, "loss": 1.8302, "step": 943 }, { "epoch": 0.20763224458374574, "grad_norm": 0.28080835938453674, "learning_rate": 8e-05, "loss": 1.7544, "step": 944 }, { "epoch": 0.20785219399538107, "grad_norm": 0.28095102310180664, "learning_rate": 8e-05, "loss": 1.7271, "step": 945 }, { "epoch": 0.20807214340701638, "grad_norm": 0.27856144309043884, "learning_rate": 8e-05, "loss": 1.8441, "step": 946 }, { "epoch": 0.20829209281865171, "grad_norm": 0.27816230058670044, "learning_rate": 8e-05, "loss": 1.981, "step": 947 }, { "epoch": 0.20851204223028702, "grad_norm": 0.2954215705394745, "learning_rate": 8e-05, "loss": 1.8001, "step": 948 }, { "epoch": 0.20873199164192235, "grad_norm": 0.24413350224494934, "learning_rate": 8e-05, "loss": 1.5436, "step": 949 }, { "epoch": 0.2089519410535577, "grad_norm": 0.2849874198436737, "learning_rate": 8e-05, "loss": 1.7027, "step": 950 }, { "epoch": 0.209171890465193, "grad_norm": 0.2710252106189728, "learning_rate": 8e-05, "loss": 1.7222, "step": 951 }, { "epoch": 0.20939183987682833, "grad_norm": 0.2557348608970642, "learning_rate": 8e-05, "loss": 1.6469, "step": 952 }, { "epoch": 0.20961178928846366, "grad_norm": 0.2688618004322052, "learning_rate": 8e-05, "loss": 1.6471, "step": 953 }, { "epoch": 0.20983173870009897, "grad_norm": 0.28641626238822937, "learning_rate": 8e-05, "loss": 1.8796, "step": 954 }, { "epoch": 0.2100516881117343, "grad_norm": 0.2582222521305084, "learning_rate": 8e-05, "loss": 1.5961, "step": 955 }, { "epoch": 0.21027163752336964, "grad_norm": 0.2615504562854767, "learning_rate": 8e-05, "loss": 1.6668, "step": 956 }, { "epoch": 0.21049158693500494, "grad_norm": 0.2669670879840851, "learning_rate": 8e-05, "loss": 1.663, "step": 957 }, { "epoch": 0.21071153634664028, "grad_norm": 0.2649092972278595, "learning_rate": 8e-05, "loss": 1.5377, "step": 958 }, { "epoch": 0.21093148575827558, "grad_norm": 0.2936461865901947, "learning_rate": 8e-05, "loss": 1.5659, "step": 959 }, { "epoch": 0.21115143516991092, "grad_norm": 0.2878846824169159, "learning_rate": 8e-05, "loss": 1.8567, "step": 960 }, { "epoch": 0.21137138458154625, "grad_norm": 0.2928799092769623, "learning_rate": 8e-05, "loss": 1.8423, "step": 961 }, { "epoch": 0.21159133399318156, "grad_norm": 0.2641200125217438, "learning_rate": 8e-05, "loss": 1.6403, "step": 962 }, { "epoch": 0.2118112834048169, "grad_norm": 0.26553985476493835, "learning_rate": 8e-05, "loss": 1.7436, "step": 963 }, { "epoch": 0.21203123281645223, "grad_norm": 0.25616276264190674, "learning_rate": 8e-05, "loss": 1.5959, "step": 964 }, { "epoch": 0.21225118222808753, "grad_norm": 0.29729175567626953, "learning_rate": 8e-05, "loss": 1.7164, "step": 965 }, { "epoch": 0.21247113163972287, "grad_norm": 0.2739759683609009, "learning_rate": 8e-05, "loss": 1.7797, "step": 966 }, { "epoch": 0.2126910810513582, "grad_norm": 0.2686353921890259, "learning_rate": 8e-05, "loss": 1.6974, "step": 967 }, { "epoch": 0.2129110304629935, "grad_norm": 0.261820912361145, "learning_rate": 8e-05, "loss": 1.6864, "step": 968 }, { "epoch": 0.21313097987462884, "grad_norm": 0.26877105236053467, "learning_rate": 8e-05, "loss": 1.6164, "step": 969 }, { "epoch": 0.21335092928626415, "grad_norm": 0.2555043399333954, "learning_rate": 8e-05, "loss": 1.6898, "step": 970 }, { "epoch": 0.21357087869789948, "grad_norm": 0.28584909439086914, "learning_rate": 8e-05, "loss": 1.9125, "step": 971 }, { "epoch": 0.21379082810953481, "grad_norm": 0.2830945551395416, "learning_rate": 8e-05, "loss": 1.6416, "step": 972 }, { "epoch": 0.21401077752117012, "grad_norm": 0.27979904413223267, "learning_rate": 8e-05, "loss": 1.8355, "step": 973 }, { "epoch": 0.21423072693280545, "grad_norm": 0.2672286033630371, "learning_rate": 8e-05, "loss": 1.685, "step": 974 }, { "epoch": 0.2144506763444408, "grad_norm": 0.26699069142341614, "learning_rate": 8e-05, "loss": 1.5951, "step": 975 }, { "epoch": 0.2146706257560761, "grad_norm": 0.2720418870449066, "learning_rate": 8e-05, "loss": 1.7558, "step": 976 }, { "epoch": 0.21489057516771143, "grad_norm": 0.26792463660240173, "learning_rate": 8e-05, "loss": 1.7407, "step": 977 }, { "epoch": 0.21511052457934676, "grad_norm": 0.2763652503490448, "learning_rate": 8e-05, "loss": 1.7525, "step": 978 }, { "epoch": 0.21533047399098207, "grad_norm": 0.2952554225921631, "learning_rate": 8e-05, "loss": 1.6535, "step": 979 }, { "epoch": 0.2155504234026174, "grad_norm": 0.24981874227523804, "learning_rate": 8e-05, "loss": 1.6055, "step": 980 }, { "epoch": 0.2157703728142527, "grad_norm": 0.29071807861328125, "learning_rate": 8e-05, "loss": 1.7461, "step": 981 }, { "epoch": 0.21599032222588804, "grad_norm": 0.26875782012939453, "learning_rate": 8e-05, "loss": 1.5809, "step": 982 }, { "epoch": 0.21621027163752338, "grad_norm": 0.2519072890281677, "learning_rate": 8e-05, "loss": 1.7001, "step": 983 }, { "epoch": 0.21643022104915868, "grad_norm": 0.2748781144618988, "learning_rate": 8e-05, "loss": 1.8367, "step": 984 }, { "epoch": 0.21665017046079402, "grad_norm": 0.274047315120697, "learning_rate": 8e-05, "loss": 1.7698, "step": 985 }, { "epoch": 0.21687011987242935, "grad_norm": 0.2614712119102478, "learning_rate": 8e-05, "loss": 1.5411, "step": 986 }, { "epoch": 0.21709006928406466, "grad_norm": 0.2714536190032959, "learning_rate": 8e-05, "loss": 1.6058, "step": 987 }, { "epoch": 0.2173100186957, "grad_norm": 0.28763729333877563, "learning_rate": 8e-05, "loss": 1.6711, "step": 988 }, { "epoch": 0.21752996810733533, "grad_norm": 0.26780402660369873, "learning_rate": 8e-05, "loss": 1.549, "step": 989 }, { "epoch": 0.21774991751897063, "grad_norm": 0.28782159090042114, "learning_rate": 8e-05, "loss": 1.8305, "step": 990 }, { "epoch": 0.21796986693060597, "grad_norm": 0.2859013080596924, "learning_rate": 8e-05, "loss": 1.7794, "step": 991 }, { "epoch": 0.21818981634224127, "grad_norm": 0.2893369197845459, "learning_rate": 8e-05, "loss": 1.6284, "step": 992 }, { "epoch": 0.2184097657538766, "grad_norm": 0.2809627652168274, "learning_rate": 8e-05, "loss": 1.6401, "step": 993 }, { "epoch": 0.21862971516551194, "grad_norm": 0.2700895667076111, "learning_rate": 8e-05, "loss": 1.7153, "step": 994 }, { "epoch": 0.21884966457714725, "grad_norm": 0.26506903767585754, "learning_rate": 8e-05, "loss": 1.4688, "step": 995 }, { "epoch": 0.21906961398878258, "grad_norm": 0.28202024102211, "learning_rate": 8e-05, "loss": 1.7009, "step": 996 }, { "epoch": 0.21928956340041791, "grad_norm": 0.2625409960746765, "learning_rate": 8e-05, "loss": 1.6491, "step": 997 }, { "epoch": 0.21950951281205322, "grad_norm": 0.29967787861824036, "learning_rate": 8e-05, "loss": 1.7231, "step": 998 }, { "epoch": 0.21972946222368855, "grad_norm": 0.2992357909679413, "learning_rate": 8e-05, "loss": 1.7028, "step": 999 }, { "epoch": 0.2199494116353239, "grad_norm": 0.28712475299835205, "learning_rate": 8e-05, "loss": 1.763, "step": 1000 }, { "epoch": 0.2201693610469592, "grad_norm": 0.26186901330947876, "learning_rate": 8e-05, "loss": 1.5695, "step": 1001 }, { "epoch": 0.22038931045859453, "grad_norm": 0.2897952198982239, "learning_rate": 8e-05, "loss": 1.6303, "step": 1002 }, { "epoch": 0.22060925987022983, "grad_norm": 0.2761494815349579, "learning_rate": 8e-05, "loss": 1.7448, "step": 1003 }, { "epoch": 0.22082920928186517, "grad_norm": 0.2604154944419861, "learning_rate": 8e-05, "loss": 1.53, "step": 1004 }, { "epoch": 0.2210491586935005, "grad_norm": 0.2897418737411499, "learning_rate": 8e-05, "loss": 1.7639, "step": 1005 }, { "epoch": 0.2212691081051358, "grad_norm": 0.28289687633514404, "learning_rate": 8e-05, "loss": 1.7202, "step": 1006 }, { "epoch": 0.22148905751677114, "grad_norm": 0.26917099952697754, "learning_rate": 8e-05, "loss": 1.7183, "step": 1007 }, { "epoch": 0.22170900692840648, "grad_norm": 0.26708024740219116, "learning_rate": 8e-05, "loss": 1.636, "step": 1008 }, { "epoch": 0.22192895634004178, "grad_norm": 0.2759459316730499, "learning_rate": 8e-05, "loss": 1.6537, "step": 1009 }, { "epoch": 0.22214890575167712, "grad_norm": 0.3040393590927124, "learning_rate": 8e-05, "loss": 1.7849, "step": 1010 }, { "epoch": 0.22236885516331245, "grad_norm": 0.2729750871658325, "learning_rate": 8e-05, "loss": 1.8199, "step": 1011 }, { "epoch": 0.22258880457494776, "grad_norm": 0.28002965450286865, "learning_rate": 8e-05, "loss": 1.8286, "step": 1012 }, { "epoch": 0.2228087539865831, "grad_norm": 0.27389100193977356, "learning_rate": 8e-05, "loss": 1.6472, "step": 1013 }, { "epoch": 0.2230287033982184, "grad_norm": 0.2610195279121399, "learning_rate": 8e-05, "loss": 1.6096, "step": 1014 }, { "epoch": 0.22324865280985373, "grad_norm": 0.2683162987232208, "learning_rate": 8e-05, "loss": 1.6477, "step": 1015 }, { "epoch": 0.22346860222148907, "grad_norm": 0.26524773240089417, "learning_rate": 8e-05, "loss": 1.6224, "step": 1016 }, { "epoch": 0.22368855163312437, "grad_norm": 0.26295366883277893, "learning_rate": 8e-05, "loss": 1.7316, "step": 1017 }, { "epoch": 0.2239085010447597, "grad_norm": 0.2837565243244171, "learning_rate": 8e-05, "loss": 1.9452, "step": 1018 }, { "epoch": 0.22412845045639504, "grad_norm": 0.28365132212638855, "learning_rate": 8e-05, "loss": 1.6577, "step": 1019 }, { "epoch": 0.22434839986803035, "grad_norm": 0.2736522853374481, "learning_rate": 8e-05, "loss": 1.6644, "step": 1020 }, { "epoch": 0.22456834927966568, "grad_norm": 0.2878374755382538, "learning_rate": 8e-05, "loss": 1.5844, "step": 1021 }, { "epoch": 0.22478829869130101, "grad_norm": 0.28223422169685364, "learning_rate": 8e-05, "loss": 1.881, "step": 1022 }, { "epoch": 0.22500824810293632, "grad_norm": 0.26408734917640686, "learning_rate": 8e-05, "loss": 1.6201, "step": 1023 }, { "epoch": 0.22522819751457165, "grad_norm": 0.28506824374198914, "learning_rate": 8e-05, "loss": 1.8146, "step": 1024 }, { "epoch": 0.22544814692620696, "grad_norm": 0.2808188796043396, "learning_rate": 8e-05, "loss": 1.8394, "step": 1025 }, { "epoch": 0.2256680963378423, "grad_norm": 0.2950645387172699, "learning_rate": 8e-05, "loss": 1.7993, "step": 1026 }, { "epoch": 0.22588804574947763, "grad_norm": 0.27935850620269775, "learning_rate": 8e-05, "loss": 1.6506, "step": 1027 }, { "epoch": 0.22610799516111293, "grad_norm": 0.2576957643032074, "learning_rate": 8e-05, "loss": 1.6987, "step": 1028 }, { "epoch": 0.22632794457274827, "grad_norm": 0.2719384729862213, "learning_rate": 8e-05, "loss": 1.6407, "step": 1029 }, { "epoch": 0.2265478939843836, "grad_norm": 0.25457167625427246, "learning_rate": 8e-05, "loss": 1.6877, "step": 1030 }, { "epoch": 0.2267678433960189, "grad_norm": 0.2758035659790039, "learning_rate": 8e-05, "loss": 1.6739, "step": 1031 }, { "epoch": 0.22698779280765424, "grad_norm": 0.27135321497917175, "learning_rate": 8e-05, "loss": 1.7124, "step": 1032 }, { "epoch": 0.22720774221928958, "grad_norm": 0.2675740420818329, "learning_rate": 8e-05, "loss": 1.7857, "step": 1033 }, { "epoch": 0.22742769163092488, "grad_norm": 0.28627943992614746, "learning_rate": 8e-05, "loss": 1.7012, "step": 1034 }, { "epoch": 0.22764764104256022, "grad_norm": 0.2710109353065491, "learning_rate": 8e-05, "loss": 1.6463, "step": 1035 }, { "epoch": 0.22786759045419552, "grad_norm": 0.27190473675727844, "learning_rate": 8e-05, "loss": 1.7288, "step": 1036 }, { "epoch": 0.22808753986583086, "grad_norm": 0.2503564953804016, "learning_rate": 8e-05, "loss": 1.566, "step": 1037 }, { "epoch": 0.2283074892774662, "grad_norm": 0.26503992080688477, "learning_rate": 8e-05, "loss": 1.7034, "step": 1038 }, { "epoch": 0.2285274386891015, "grad_norm": 0.29445260763168335, "learning_rate": 8e-05, "loss": 1.6739, "step": 1039 }, { "epoch": 0.22874738810073683, "grad_norm": 0.25705471634864807, "learning_rate": 8e-05, "loss": 1.6503, "step": 1040 }, { "epoch": 0.22896733751237217, "grad_norm": 0.27109014987945557, "learning_rate": 8e-05, "loss": 1.8045, "step": 1041 }, { "epoch": 0.22918728692400747, "grad_norm": 0.2972055673599243, "learning_rate": 8e-05, "loss": 1.6439, "step": 1042 }, { "epoch": 0.2294072363356428, "grad_norm": 0.27126485109329224, "learning_rate": 8e-05, "loss": 1.672, "step": 1043 }, { "epoch": 0.22962718574727814, "grad_norm": 0.2731145918369293, "learning_rate": 8e-05, "loss": 1.7795, "step": 1044 }, { "epoch": 0.22984713515891345, "grad_norm": 0.2768365442752838, "learning_rate": 8e-05, "loss": 1.7145, "step": 1045 }, { "epoch": 0.23006708457054878, "grad_norm": 0.2606940269470215, "learning_rate": 8e-05, "loss": 1.6169, "step": 1046 }, { "epoch": 0.2302870339821841, "grad_norm": 0.2898729741573334, "learning_rate": 8e-05, "loss": 1.7315, "step": 1047 }, { "epoch": 0.23050698339381942, "grad_norm": 0.2772413194179535, "learning_rate": 8e-05, "loss": 1.8632, "step": 1048 }, { "epoch": 0.23072693280545475, "grad_norm": 0.25808605551719666, "learning_rate": 8e-05, "loss": 1.6626, "step": 1049 }, { "epoch": 0.23094688221709006, "grad_norm": 0.2727161645889282, "learning_rate": 8e-05, "loss": 1.7848, "step": 1050 }, { "epoch": 0.2311668316287254, "grad_norm": 0.25677087903022766, "learning_rate": 8e-05, "loss": 1.6168, "step": 1051 }, { "epoch": 0.23138678104036073, "grad_norm": 0.2761050760746002, "learning_rate": 8e-05, "loss": 1.8615, "step": 1052 }, { "epoch": 0.23160673045199603, "grad_norm": 0.2862778604030609, "learning_rate": 8e-05, "loss": 1.8728, "step": 1053 }, { "epoch": 0.23182667986363137, "grad_norm": 0.27526941895484924, "learning_rate": 8e-05, "loss": 1.7627, "step": 1054 }, { "epoch": 0.23204662927526667, "grad_norm": 0.2932235896587372, "learning_rate": 8e-05, "loss": 1.8539, "step": 1055 }, { "epoch": 0.232266578686902, "grad_norm": 0.2770839035511017, "learning_rate": 8e-05, "loss": 1.7393, "step": 1056 }, { "epoch": 0.23248652809853734, "grad_norm": 0.2741580307483673, "learning_rate": 8e-05, "loss": 1.6076, "step": 1057 }, { "epoch": 0.23270647751017265, "grad_norm": 0.2788783311843872, "learning_rate": 8e-05, "loss": 1.7615, "step": 1058 }, { "epoch": 0.23292642692180798, "grad_norm": 0.28565406799316406, "learning_rate": 8e-05, "loss": 1.6266, "step": 1059 }, { "epoch": 0.23314637633344332, "grad_norm": 0.26543545722961426, "learning_rate": 8e-05, "loss": 1.7192, "step": 1060 }, { "epoch": 0.23336632574507862, "grad_norm": 0.2770478129386902, "learning_rate": 8e-05, "loss": 1.8056, "step": 1061 }, { "epoch": 0.23358627515671396, "grad_norm": 0.27805015444755554, "learning_rate": 8e-05, "loss": 1.6735, "step": 1062 }, { "epoch": 0.2338062245683493, "grad_norm": 0.309862345457077, "learning_rate": 8e-05, "loss": 1.7235, "step": 1063 }, { "epoch": 0.2340261739799846, "grad_norm": 0.27140697836875916, "learning_rate": 8e-05, "loss": 1.6883, "step": 1064 }, { "epoch": 0.23424612339161993, "grad_norm": 0.3052090108394623, "learning_rate": 8e-05, "loss": 1.8792, "step": 1065 }, { "epoch": 0.23446607280325524, "grad_norm": 0.2995065450668335, "learning_rate": 8e-05, "loss": 1.6632, "step": 1066 }, { "epoch": 0.23468602221489057, "grad_norm": 0.2782532870769501, "learning_rate": 8e-05, "loss": 1.7395, "step": 1067 }, { "epoch": 0.2349059716265259, "grad_norm": 0.28436902165412903, "learning_rate": 8e-05, "loss": 1.8416, "step": 1068 }, { "epoch": 0.2351259210381612, "grad_norm": 0.2740377187728882, "learning_rate": 8e-05, "loss": 1.9026, "step": 1069 }, { "epoch": 0.23534587044979655, "grad_norm": 0.2978285849094391, "learning_rate": 8e-05, "loss": 1.7277, "step": 1070 }, { "epoch": 0.23556581986143188, "grad_norm": 0.27265986800193787, "learning_rate": 8e-05, "loss": 1.7376, "step": 1071 }, { "epoch": 0.2357857692730672, "grad_norm": 0.24915599822998047, "learning_rate": 8e-05, "loss": 1.6151, "step": 1072 }, { "epoch": 0.23600571868470252, "grad_norm": 0.28203171491622925, "learning_rate": 8e-05, "loss": 1.7713, "step": 1073 }, { "epoch": 0.23622566809633785, "grad_norm": 0.278793066740036, "learning_rate": 8e-05, "loss": 1.6717, "step": 1074 }, { "epoch": 0.23644561750797316, "grad_norm": 0.2760609984397888, "learning_rate": 8e-05, "loss": 1.5866, "step": 1075 }, { "epoch": 0.2366655669196085, "grad_norm": 0.2726036012172699, "learning_rate": 8e-05, "loss": 1.6774, "step": 1076 }, { "epoch": 0.2368855163312438, "grad_norm": 0.27443891763687134, "learning_rate": 8e-05, "loss": 1.7615, "step": 1077 }, { "epoch": 0.23710546574287913, "grad_norm": 0.2818880081176758, "learning_rate": 8e-05, "loss": 1.7433, "step": 1078 }, { "epoch": 0.23732541515451447, "grad_norm": 0.2646252512931824, "learning_rate": 8e-05, "loss": 1.5498, "step": 1079 }, { "epoch": 0.23754536456614977, "grad_norm": 0.2964784502983093, "learning_rate": 8e-05, "loss": 1.6162, "step": 1080 }, { "epoch": 0.2377653139777851, "grad_norm": 0.3044411242008209, "learning_rate": 8e-05, "loss": 1.7395, "step": 1081 }, { "epoch": 0.23798526338942044, "grad_norm": 0.28679221868515015, "learning_rate": 8e-05, "loss": 1.8126, "step": 1082 }, { "epoch": 0.23820521280105575, "grad_norm": 0.26326417922973633, "learning_rate": 8e-05, "loss": 1.6451, "step": 1083 }, { "epoch": 0.23842516221269108, "grad_norm": 0.28527480363845825, "learning_rate": 8e-05, "loss": 1.8442, "step": 1084 }, { "epoch": 0.23864511162432642, "grad_norm": 0.28897759318351746, "learning_rate": 8e-05, "loss": 1.8224, "step": 1085 }, { "epoch": 0.23886506103596172, "grad_norm": 0.2955721616744995, "learning_rate": 8e-05, "loss": 1.7304, "step": 1086 }, { "epoch": 0.23908501044759706, "grad_norm": 0.26267075538635254, "learning_rate": 8e-05, "loss": 1.67, "step": 1087 }, { "epoch": 0.23930495985923236, "grad_norm": 0.27105912566185, "learning_rate": 8e-05, "loss": 1.7461, "step": 1088 }, { "epoch": 0.2395249092708677, "grad_norm": 0.26483941078186035, "learning_rate": 8e-05, "loss": 1.6215, "step": 1089 }, { "epoch": 0.23974485868250303, "grad_norm": 0.2804373800754547, "learning_rate": 8e-05, "loss": 1.6618, "step": 1090 }, { "epoch": 0.23996480809413834, "grad_norm": 0.26146185398101807, "learning_rate": 8e-05, "loss": 1.6641, "step": 1091 }, { "epoch": 0.24018475750577367, "grad_norm": 0.2839837372303009, "learning_rate": 8e-05, "loss": 1.5898, "step": 1092 }, { "epoch": 0.240404706917409, "grad_norm": 0.26833322644233704, "learning_rate": 8e-05, "loss": 1.8341, "step": 1093 }, { "epoch": 0.2406246563290443, "grad_norm": 0.2779574394226074, "learning_rate": 8e-05, "loss": 1.7142, "step": 1094 }, { "epoch": 0.24084460574067965, "grad_norm": 0.2821759879589081, "learning_rate": 8e-05, "loss": 1.7261, "step": 1095 }, { "epoch": 0.24106455515231498, "grad_norm": 0.2849150002002716, "learning_rate": 8e-05, "loss": 1.6834, "step": 1096 }, { "epoch": 0.24128450456395029, "grad_norm": 0.277148574590683, "learning_rate": 8e-05, "loss": 1.5617, "step": 1097 }, { "epoch": 0.24150445397558562, "grad_norm": 0.28307756781578064, "learning_rate": 8e-05, "loss": 1.8104, "step": 1098 }, { "epoch": 0.24172440338722093, "grad_norm": 0.28540289402008057, "learning_rate": 8e-05, "loss": 1.7331, "step": 1099 }, { "epoch": 0.24194435279885626, "grad_norm": 0.277544766664505, "learning_rate": 8e-05, "loss": 1.762, "step": 1100 }, { "epoch": 0.2421643022104916, "grad_norm": 0.259435772895813, "learning_rate": 8e-05, "loss": 1.5474, "step": 1101 }, { "epoch": 0.2423842516221269, "grad_norm": 0.2759372591972351, "learning_rate": 8e-05, "loss": 1.6535, "step": 1102 }, { "epoch": 0.24260420103376223, "grad_norm": 0.27163347601890564, "learning_rate": 8e-05, "loss": 1.6035, "step": 1103 }, { "epoch": 0.24282415044539757, "grad_norm": 0.26722922921180725, "learning_rate": 8e-05, "loss": 1.7607, "step": 1104 }, { "epoch": 0.24304409985703287, "grad_norm": 0.2925039529800415, "learning_rate": 8e-05, "loss": 1.6441, "step": 1105 }, { "epoch": 0.2432640492686682, "grad_norm": 0.271672785282135, "learning_rate": 8e-05, "loss": 1.658, "step": 1106 }, { "epoch": 0.24348399868030354, "grad_norm": 0.2827896773815155, "learning_rate": 8e-05, "loss": 1.6258, "step": 1107 }, { "epoch": 0.24370394809193885, "grad_norm": 0.2732497751712799, "learning_rate": 8e-05, "loss": 1.6379, "step": 1108 }, { "epoch": 0.24392389750357418, "grad_norm": 0.28081193566322327, "learning_rate": 8e-05, "loss": 1.7901, "step": 1109 }, { "epoch": 0.2441438469152095, "grad_norm": 0.2799675762653351, "learning_rate": 8e-05, "loss": 1.8323, "step": 1110 }, { "epoch": 0.24436379632684482, "grad_norm": 0.2677648961544037, "learning_rate": 8e-05, "loss": 1.7372, "step": 1111 }, { "epoch": 0.24458374573848016, "grad_norm": 0.2644648551940918, "learning_rate": 8e-05, "loss": 1.6594, "step": 1112 }, { "epoch": 0.24480369515011546, "grad_norm": 0.2704750895500183, "learning_rate": 8e-05, "loss": 1.706, "step": 1113 }, { "epoch": 0.2450236445617508, "grad_norm": 0.2762587368488312, "learning_rate": 8e-05, "loss": 1.7445, "step": 1114 }, { "epoch": 0.24524359397338613, "grad_norm": 0.2578018307685852, "learning_rate": 8e-05, "loss": 1.5707, "step": 1115 }, { "epoch": 0.24546354338502144, "grad_norm": 0.2892129719257355, "learning_rate": 8e-05, "loss": 1.805, "step": 1116 }, { "epoch": 0.24568349279665677, "grad_norm": 0.2868081033229828, "learning_rate": 8e-05, "loss": 1.7756, "step": 1117 }, { "epoch": 0.2459034422082921, "grad_norm": 0.2820534110069275, "learning_rate": 8e-05, "loss": 1.7826, "step": 1118 }, { "epoch": 0.2461233916199274, "grad_norm": 0.2824958264827728, "learning_rate": 8e-05, "loss": 1.6752, "step": 1119 }, { "epoch": 0.24634334103156275, "grad_norm": 0.2782610356807709, "learning_rate": 8e-05, "loss": 1.7536, "step": 1120 }, { "epoch": 0.24656329044319805, "grad_norm": 0.27147912979125977, "learning_rate": 8e-05, "loss": 1.6783, "step": 1121 }, { "epoch": 0.24678323985483339, "grad_norm": 0.2740795612335205, "learning_rate": 8e-05, "loss": 1.7702, "step": 1122 }, { "epoch": 0.24700318926646872, "grad_norm": 0.2922619879245758, "learning_rate": 8e-05, "loss": 1.8204, "step": 1123 }, { "epoch": 0.24722313867810403, "grad_norm": 0.2872619926929474, "learning_rate": 8e-05, "loss": 1.714, "step": 1124 }, { "epoch": 0.24744308808973936, "grad_norm": 0.27333369851112366, "learning_rate": 8e-05, "loss": 1.6575, "step": 1125 }, { "epoch": 0.2476630375013747, "grad_norm": 0.28192320466041565, "learning_rate": 8e-05, "loss": 1.7221, "step": 1126 }, { "epoch": 0.24788298691301, "grad_norm": 0.26607248187065125, "learning_rate": 8e-05, "loss": 1.7262, "step": 1127 }, { "epoch": 0.24810293632464533, "grad_norm": 0.279690682888031, "learning_rate": 8e-05, "loss": 1.7004, "step": 1128 }, { "epoch": 0.24832288573628067, "grad_norm": 0.27289190888404846, "learning_rate": 8e-05, "loss": 1.6916, "step": 1129 }, { "epoch": 0.24854283514791597, "grad_norm": 0.27388349175453186, "learning_rate": 8e-05, "loss": 1.6656, "step": 1130 }, { "epoch": 0.2487627845595513, "grad_norm": 0.2912501096725464, "learning_rate": 8e-05, "loss": 1.8086, "step": 1131 }, { "epoch": 0.24898273397118661, "grad_norm": 0.2999799847602844, "learning_rate": 8e-05, "loss": 1.7659, "step": 1132 }, { "epoch": 0.24920268338282195, "grad_norm": 0.262207955121994, "learning_rate": 8e-05, "loss": 1.6581, "step": 1133 }, { "epoch": 0.24942263279445728, "grad_norm": 0.2571624517440796, "learning_rate": 8e-05, "loss": 1.6509, "step": 1134 }, { "epoch": 0.2496425822060926, "grad_norm": 0.26213690638542175, "learning_rate": 8e-05, "loss": 1.6044, "step": 1135 }, { "epoch": 0.24986253161772792, "grad_norm": 0.2870398461818695, "learning_rate": 8e-05, "loss": 1.6678, "step": 1136 }, { "epoch": 0.25008248102936326, "grad_norm": 0.2672583758831024, "learning_rate": 8e-05, "loss": 1.6563, "step": 1137 }, { "epoch": 0.2503024304409986, "grad_norm": 0.29864680767059326, "learning_rate": 8e-05, "loss": 1.858, "step": 1138 }, { "epoch": 0.25052237985263387, "grad_norm": 0.3096907436847687, "learning_rate": 8e-05, "loss": 1.7731, "step": 1139 }, { "epoch": 0.2507423292642692, "grad_norm": 0.2668014466762543, "learning_rate": 8e-05, "loss": 1.6173, "step": 1140 }, { "epoch": 0.25096227867590454, "grad_norm": 0.275074303150177, "learning_rate": 8e-05, "loss": 1.704, "step": 1141 }, { "epoch": 0.25118222808753987, "grad_norm": 0.29657119512557983, "learning_rate": 8e-05, "loss": 1.9789, "step": 1142 }, { "epoch": 0.2514021774991752, "grad_norm": 0.26117807626724243, "learning_rate": 8e-05, "loss": 1.6815, "step": 1143 }, { "epoch": 0.25162212691081054, "grad_norm": 0.2738019824028015, "learning_rate": 8e-05, "loss": 1.7031, "step": 1144 }, { "epoch": 0.2518420763224458, "grad_norm": 0.27922967076301575, "learning_rate": 8e-05, "loss": 1.7914, "step": 1145 }, { "epoch": 0.25206202573408115, "grad_norm": 0.2876172661781311, "learning_rate": 8e-05, "loss": 1.721, "step": 1146 }, { "epoch": 0.2522819751457165, "grad_norm": 0.28017961978912354, "learning_rate": 8e-05, "loss": 1.6731, "step": 1147 }, { "epoch": 0.2525019245573518, "grad_norm": 0.2898389399051666, "learning_rate": 8e-05, "loss": 1.8749, "step": 1148 }, { "epoch": 0.25272187396898715, "grad_norm": 0.2742408812046051, "learning_rate": 8e-05, "loss": 1.6811, "step": 1149 }, { "epoch": 0.25294182338062243, "grad_norm": 0.2806207835674286, "learning_rate": 8e-05, "loss": 1.7082, "step": 1150 }, { "epoch": 0.25316177279225777, "grad_norm": 0.27871328592300415, "learning_rate": 8e-05, "loss": 1.7142, "step": 1151 }, { "epoch": 0.2533817222038931, "grad_norm": 0.2792799472808838, "learning_rate": 8e-05, "loss": 1.5703, "step": 1152 }, { "epoch": 0.25360167161552843, "grad_norm": 0.27358901500701904, "learning_rate": 8e-05, "loss": 1.7576, "step": 1153 }, { "epoch": 0.25382162102716377, "grad_norm": 0.26983192563056946, "learning_rate": 8e-05, "loss": 1.6646, "step": 1154 }, { "epoch": 0.2540415704387991, "grad_norm": 0.2711959183216095, "learning_rate": 8e-05, "loss": 1.7698, "step": 1155 }, { "epoch": 0.2542615198504344, "grad_norm": 0.28412333130836487, "learning_rate": 8e-05, "loss": 1.7446, "step": 1156 }, { "epoch": 0.2544814692620697, "grad_norm": 0.2698575258255005, "learning_rate": 8e-05, "loss": 1.6861, "step": 1157 }, { "epoch": 0.25470141867370505, "grad_norm": 0.2806732952594757, "learning_rate": 8e-05, "loss": 1.7308, "step": 1158 }, { "epoch": 0.2549213680853404, "grad_norm": 0.2715948522090912, "learning_rate": 8e-05, "loss": 1.852, "step": 1159 }, { "epoch": 0.2551413174969757, "grad_norm": 0.33048170804977417, "learning_rate": 8e-05, "loss": 1.881, "step": 1160 }, { "epoch": 0.255361266908611, "grad_norm": 0.27907994389533997, "learning_rate": 8e-05, "loss": 1.6501, "step": 1161 }, { "epoch": 0.25558121632024633, "grad_norm": 0.2747988998889923, "learning_rate": 8e-05, "loss": 1.7265, "step": 1162 }, { "epoch": 0.25580116573188166, "grad_norm": 0.28321677446365356, "learning_rate": 8e-05, "loss": 1.8602, "step": 1163 }, { "epoch": 0.256021115143517, "grad_norm": 0.2695465683937073, "learning_rate": 8e-05, "loss": 1.6091, "step": 1164 }, { "epoch": 0.25624106455515233, "grad_norm": 0.272135466337204, "learning_rate": 8e-05, "loss": 1.6236, "step": 1165 }, { "epoch": 0.25646101396678767, "grad_norm": 0.2715020775794983, "learning_rate": 8e-05, "loss": 1.674, "step": 1166 }, { "epoch": 0.25668096337842294, "grad_norm": 0.2879820764064789, "learning_rate": 8e-05, "loss": 1.8393, "step": 1167 }, { "epoch": 0.2569009127900583, "grad_norm": 0.2616657018661499, "learning_rate": 8e-05, "loss": 1.6391, "step": 1168 }, { "epoch": 0.2571208622016936, "grad_norm": 0.2558441460132599, "learning_rate": 8e-05, "loss": 1.606, "step": 1169 }, { "epoch": 0.25734081161332895, "grad_norm": 0.26944512128829956, "learning_rate": 8e-05, "loss": 1.7288, "step": 1170 }, { "epoch": 0.2575607610249643, "grad_norm": 0.26958367228507996, "learning_rate": 8e-05, "loss": 1.6233, "step": 1171 }, { "epoch": 0.25778071043659956, "grad_norm": 0.29003527760505676, "learning_rate": 8e-05, "loss": 1.777, "step": 1172 }, { "epoch": 0.2580006598482349, "grad_norm": 0.2677457630634308, "learning_rate": 8e-05, "loss": 1.6835, "step": 1173 }, { "epoch": 0.2582206092598702, "grad_norm": 0.28062689304351807, "learning_rate": 8e-05, "loss": 1.726, "step": 1174 }, { "epoch": 0.25844055867150556, "grad_norm": 0.26764920353889465, "learning_rate": 8e-05, "loss": 1.6575, "step": 1175 }, { "epoch": 0.2586605080831409, "grad_norm": 0.28183332085609436, "learning_rate": 8e-05, "loss": 1.784, "step": 1176 }, { "epoch": 0.25888045749477623, "grad_norm": 0.25718390941619873, "learning_rate": 8e-05, "loss": 1.6317, "step": 1177 }, { "epoch": 0.2591004069064115, "grad_norm": 0.25523149967193604, "learning_rate": 8e-05, "loss": 1.5634, "step": 1178 }, { "epoch": 0.25932035631804684, "grad_norm": 0.2539874315261841, "learning_rate": 8e-05, "loss": 1.5878, "step": 1179 }, { "epoch": 0.2595403057296822, "grad_norm": 0.2868393659591675, "learning_rate": 8e-05, "loss": 1.7301, "step": 1180 }, { "epoch": 0.2597602551413175, "grad_norm": 0.27819645404815674, "learning_rate": 8e-05, "loss": 1.6895, "step": 1181 }, { "epoch": 0.25998020455295284, "grad_norm": 0.27499255537986755, "learning_rate": 8e-05, "loss": 1.7462, "step": 1182 }, { "epoch": 0.2602001539645881, "grad_norm": 0.2858695685863495, "learning_rate": 8e-05, "loss": 1.8199, "step": 1183 }, { "epoch": 0.26042010337622346, "grad_norm": 0.2646760642528534, "learning_rate": 8e-05, "loss": 1.6597, "step": 1184 }, { "epoch": 0.2606400527878588, "grad_norm": 0.2831268310546875, "learning_rate": 8e-05, "loss": 1.8383, "step": 1185 }, { "epoch": 0.2608600021994941, "grad_norm": 0.2593746483325958, "learning_rate": 8e-05, "loss": 1.6115, "step": 1186 }, { "epoch": 0.26107995161112946, "grad_norm": 0.26519641280174255, "learning_rate": 8e-05, "loss": 1.5959, "step": 1187 }, { "epoch": 0.2612999010227648, "grad_norm": 0.2733252942562103, "learning_rate": 8e-05, "loss": 1.6318, "step": 1188 }, { "epoch": 0.26151985043440007, "grad_norm": 0.27299511432647705, "learning_rate": 8e-05, "loss": 1.7313, "step": 1189 }, { "epoch": 0.2617397998460354, "grad_norm": 0.2684955894947052, "learning_rate": 8e-05, "loss": 1.5826, "step": 1190 }, { "epoch": 0.26195974925767074, "grad_norm": 0.2747553586959839, "learning_rate": 8e-05, "loss": 1.7008, "step": 1191 }, { "epoch": 0.26217969866930607, "grad_norm": 0.26033639907836914, "learning_rate": 8e-05, "loss": 1.5571, "step": 1192 }, { "epoch": 0.2623996480809414, "grad_norm": 0.2640804350376129, "learning_rate": 8e-05, "loss": 1.5317, "step": 1193 }, { "epoch": 0.2626195974925767, "grad_norm": 0.27063700556755066, "learning_rate": 8e-05, "loss": 1.5501, "step": 1194 }, { "epoch": 0.262839546904212, "grad_norm": 0.2677111029624939, "learning_rate": 8e-05, "loss": 1.5894, "step": 1195 }, { "epoch": 0.26305949631584735, "grad_norm": 0.28144168853759766, "learning_rate": 8e-05, "loss": 1.7496, "step": 1196 }, { "epoch": 0.2632794457274827, "grad_norm": 0.2602388858795166, "learning_rate": 8e-05, "loss": 1.571, "step": 1197 }, { "epoch": 0.263499395139118, "grad_norm": 0.2941505014896393, "learning_rate": 8e-05, "loss": 1.6692, "step": 1198 }, { "epoch": 0.26371934455075335, "grad_norm": 0.264433354139328, "learning_rate": 8e-05, "loss": 1.6922, "step": 1199 }, { "epoch": 0.26393929396238863, "grad_norm": 0.25587090849876404, "learning_rate": 8e-05, "loss": 1.5599, "step": 1200 }, { "epoch": 0.26415924337402397, "grad_norm": 0.3012869358062744, "learning_rate": 8e-05, "loss": 1.9195, "step": 1201 }, { "epoch": 0.2643791927856593, "grad_norm": 0.2762719392776489, "learning_rate": 8e-05, "loss": 1.898, "step": 1202 }, { "epoch": 0.26459914219729463, "grad_norm": 0.2701188325881958, "learning_rate": 8e-05, "loss": 1.7312, "step": 1203 }, { "epoch": 0.26481909160892997, "grad_norm": 0.29665982723236084, "learning_rate": 8e-05, "loss": 1.8089, "step": 1204 }, { "epoch": 0.26503904102056525, "grad_norm": 0.26700517535209656, "learning_rate": 8e-05, "loss": 1.8401, "step": 1205 }, { "epoch": 0.2652589904322006, "grad_norm": 0.2828493118286133, "learning_rate": 8e-05, "loss": 1.8622, "step": 1206 }, { "epoch": 0.2654789398438359, "grad_norm": 0.2746271789073944, "learning_rate": 8e-05, "loss": 1.6521, "step": 1207 }, { "epoch": 0.26569888925547125, "grad_norm": 0.2882270812988281, "learning_rate": 8e-05, "loss": 1.7168, "step": 1208 }, { "epoch": 0.2659188386671066, "grad_norm": 0.29784512519836426, "learning_rate": 8e-05, "loss": 1.6968, "step": 1209 }, { "epoch": 0.2661387880787419, "grad_norm": 0.2807427942752838, "learning_rate": 8e-05, "loss": 1.6004, "step": 1210 }, { "epoch": 0.2663587374903772, "grad_norm": 0.2956424951553345, "learning_rate": 8e-05, "loss": 1.8325, "step": 1211 }, { "epoch": 0.26657868690201253, "grad_norm": 0.2647739350795746, "learning_rate": 8e-05, "loss": 1.6391, "step": 1212 }, { "epoch": 0.26679863631364786, "grad_norm": 0.2955171465873718, "learning_rate": 8e-05, "loss": 1.7893, "step": 1213 }, { "epoch": 0.2670185857252832, "grad_norm": 0.27241894602775574, "learning_rate": 8e-05, "loss": 1.781, "step": 1214 }, { "epoch": 0.26723853513691853, "grad_norm": 0.2841251492500305, "learning_rate": 8e-05, "loss": 1.8612, "step": 1215 }, { "epoch": 0.2674584845485538, "grad_norm": 0.327891081571579, "learning_rate": 8e-05, "loss": 1.844, "step": 1216 }, { "epoch": 0.26767843396018914, "grad_norm": 0.26434099674224854, "learning_rate": 8e-05, "loss": 1.6325, "step": 1217 }, { "epoch": 0.2678983833718245, "grad_norm": 0.2868417799472809, "learning_rate": 8e-05, "loss": 1.7087, "step": 1218 }, { "epoch": 0.2681183327834598, "grad_norm": 0.27408069372177124, "learning_rate": 8e-05, "loss": 1.6006, "step": 1219 }, { "epoch": 0.26833828219509515, "grad_norm": 0.2697390019893646, "learning_rate": 8e-05, "loss": 1.6833, "step": 1220 }, { "epoch": 0.2685582316067304, "grad_norm": 0.27598559856414795, "learning_rate": 8e-05, "loss": 1.7192, "step": 1221 }, { "epoch": 0.26877818101836576, "grad_norm": 0.26871007680892944, "learning_rate": 8e-05, "loss": 1.6301, "step": 1222 }, { "epoch": 0.2689981304300011, "grad_norm": 0.2739337980747223, "learning_rate": 8e-05, "loss": 1.6828, "step": 1223 }, { "epoch": 0.2692180798416364, "grad_norm": 0.286530464887619, "learning_rate": 8e-05, "loss": 1.6484, "step": 1224 }, { "epoch": 0.26943802925327176, "grad_norm": 0.27509886026382446, "learning_rate": 8e-05, "loss": 1.6647, "step": 1225 }, { "epoch": 0.2696579786649071, "grad_norm": 0.2916969358921051, "learning_rate": 8e-05, "loss": 1.7908, "step": 1226 }, { "epoch": 0.2698779280765424, "grad_norm": 0.26566174626350403, "learning_rate": 8e-05, "loss": 1.6075, "step": 1227 }, { "epoch": 0.2700978774881777, "grad_norm": 0.27648022770881653, "learning_rate": 8e-05, "loss": 1.7536, "step": 1228 }, { "epoch": 0.27031782689981304, "grad_norm": 0.27313023805618286, "learning_rate": 8e-05, "loss": 1.6978, "step": 1229 }, { "epoch": 0.2705377763114484, "grad_norm": 0.2755061388015747, "learning_rate": 8e-05, "loss": 1.7196, "step": 1230 }, { "epoch": 0.2707577257230837, "grad_norm": 0.25907769799232483, "learning_rate": 8e-05, "loss": 1.5518, "step": 1231 }, { "epoch": 0.270977675134719, "grad_norm": 0.26485681533813477, "learning_rate": 8e-05, "loss": 1.5053, "step": 1232 }, { "epoch": 0.2711976245463543, "grad_norm": 0.27980178594589233, "learning_rate": 8e-05, "loss": 1.7824, "step": 1233 }, { "epoch": 0.27141757395798966, "grad_norm": 0.2750954329967499, "learning_rate": 8e-05, "loss": 1.6973, "step": 1234 }, { "epoch": 0.271637523369625, "grad_norm": 0.27367594838142395, "learning_rate": 8e-05, "loss": 1.6691, "step": 1235 }, { "epoch": 0.2718574727812603, "grad_norm": 0.27089521288871765, "learning_rate": 8e-05, "loss": 1.7532, "step": 1236 }, { "epoch": 0.27207742219289566, "grad_norm": 0.30656641721725464, "learning_rate": 8e-05, "loss": 1.8411, "step": 1237 }, { "epoch": 0.27229737160453094, "grad_norm": 0.25732672214508057, "learning_rate": 8e-05, "loss": 1.5599, "step": 1238 }, { "epoch": 0.27251732101616627, "grad_norm": 0.2643807828426361, "learning_rate": 8e-05, "loss": 1.6654, "step": 1239 }, { "epoch": 0.2727372704278016, "grad_norm": 0.2703326344490051, "learning_rate": 8e-05, "loss": 1.594, "step": 1240 }, { "epoch": 0.27295721983943694, "grad_norm": 0.27907243371009827, "learning_rate": 8e-05, "loss": 1.7531, "step": 1241 }, { "epoch": 0.27317716925107227, "grad_norm": 0.2482902854681015, "learning_rate": 8e-05, "loss": 1.3586, "step": 1242 }, { "epoch": 0.27339711866270755, "grad_norm": 0.2879469394683838, "learning_rate": 8e-05, "loss": 1.76, "step": 1243 }, { "epoch": 0.2736170680743429, "grad_norm": 0.26334571838378906, "learning_rate": 8e-05, "loss": 1.536, "step": 1244 }, { "epoch": 0.2738370174859782, "grad_norm": 0.27328065037727356, "learning_rate": 8e-05, "loss": 1.7199, "step": 1245 }, { "epoch": 0.27405696689761355, "grad_norm": 0.27392926812171936, "learning_rate": 8e-05, "loss": 1.7731, "step": 1246 }, { "epoch": 0.2742769163092489, "grad_norm": 0.29755476117134094, "learning_rate": 8e-05, "loss": 1.7184, "step": 1247 }, { "epoch": 0.2744968657208842, "grad_norm": 0.29554107785224915, "learning_rate": 8e-05, "loss": 1.7442, "step": 1248 }, { "epoch": 0.2747168151325195, "grad_norm": 0.2562367618083954, "learning_rate": 8e-05, "loss": 1.63, "step": 1249 }, { "epoch": 0.27493676454415483, "grad_norm": 0.27746453881263733, "learning_rate": 8e-05, "loss": 1.7396, "step": 1250 }, { "epoch": 0.27515671395579017, "grad_norm": 0.2747843265533447, "learning_rate": 8e-05, "loss": 1.6628, "step": 1251 }, { "epoch": 0.2753766633674255, "grad_norm": 0.2650463581085205, "learning_rate": 8e-05, "loss": 1.6409, "step": 1252 }, { "epoch": 0.27559661277906083, "grad_norm": 0.30537328124046326, "learning_rate": 8e-05, "loss": 1.4927, "step": 1253 }, { "epoch": 0.2758165621906961, "grad_norm": 0.26015424728393555, "learning_rate": 8e-05, "loss": 1.718, "step": 1254 }, { "epoch": 0.27603651160233145, "grad_norm": 0.2512992322444916, "learning_rate": 8e-05, "loss": 1.4757, "step": 1255 }, { "epoch": 0.2762564610139668, "grad_norm": 0.28478461503982544, "learning_rate": 8e-05, "loss": 1.9081, "step": 1256 }, { "epoch": 0.2764764104256021, "grad_norm": 0.28490516543388367, "learning_rate": 8e-05, "loss": 1.8495, "step": 1257 }, { "epoch": 0.27669635983723745, "grad_norm": 0.2758481204509735, "learning_rate": 8e-05, "loss": 1.767, "step": 1258 }, { "epoch": 0.2769163092488728, "grad_norm": 0.28743213415145874, "learning_rate": 8e-05, "loss": 1.6548, "step": 1259 }, { "epoch": 0.27713625866050806, "grad_norm": 0.2738385796546936, "learning_rate": 8e-05, "loss": 1.5616, "step": 1260 }, { "epoch": 0.2773562080721434, "grad_norm": 0.27758583426475525, "learning_rate": 8e-05, "loss": 1.7793, "step": 1261 }, { "epoch": 0.27757615748377873, "grad_norm": 0.2830480635166168, "learning_rate": 8e-05, "loss": 1.8048, "step": 1262 }, { "epoch": 0.27779610689541406, "grad_norm": 0.296036034822464, "learning_rate": 8e-05, "loss": 1.7844, "step": 1263 }, { "epoch": 0.2780160563070494, "grad_norm": 0.28651297092437744, "learning_rate": 8e-05, "loss": 1.7239, "step": 1264 }, { "epoch": 0.2782360057186847, "grad_norm": 0.2826116979122162, "learning_rate": 8e-05, "loss": 1.8415, "step": 1265 }, { "epoch": 0.27845595513032, "grad_norm": 0.27445724606513977, "learning_rate": 8e-05, "loss": 1.6738, "step": 1266 }, { "epoch": 0.27867590454195534, "grad_norm": 0.28153640031814575, "learning_rate": 8e-05, "loss": 1.6519, "step": 1267 }, { "epoch": 0.2788958539535907, "grad_norm": 0.27389946579933167, "learning_rate": 8e-05, "loss": 1.681, "step": 1268 }, { "epoch": 0.279115803365226, "grad_norm": 0.2639203667640686, "learning_rate": 8e-05, "loss": 1.6398, "step": 1269 }, { "epoch": 0.27933575277686135, "grad_norm": 0.2787509560585022, "learning_rate": 8e-05, "loss": 1.7199, "step": 1270 }, { "epoch": 0.2795557021884966, "grad_norm": 0.28468430042266846, "learning_rate": 8e-05, "loss": 1.8668, "step": 1271 }, { "epoch": 0.27977565160013196, "grad_norm": 0.2907005250453949, "learning_rate": 8e-05, "loss": 1.9328, "step": 1272 }, { "epoch": 0.2799956010117673, "grad_norm": 0.2607463300228119, "learning_rate": 8e-05, "loss": 1.5958, "step": 1273 }, { "epoch": 0.2802155504234026, "grad_norm": 0.2695181965827942, "learning_rate": 8e-05, "loss": 1.6708, "step": 1274 }, { "epoch": 0.28043549983503796, "grad_norm": 0.28671538829803467, "learning_rate": 8e-05, "loss": 1.7736, "step": 1275 }, { "epoch": 0.28065544924667324, "grad_norm": 0.3246489465236664, "learning_rate": 8e-05, "loss": 1.8145, "step": 1276 }, { "epoch": 0.2808753986583086, "grad_norm": 0.2879314720630646, "learning_rate": 8e-05, "loss": 1.782, "step": 1277 }, { "epoch": 0.2810953480699439, "grad_norm": 0.27141574025154114, "learning_rate": 8e-05, "loss": 1.8069, "step": 1278 }, { "epoch": 0.28131529748157924, "grad_norm": 0.2893892228603363, "learning_rate": 8e-05, "loss": 1.7893, "step": 1279 }, { "epoch": 0.2815352468932146, "grad_norm": 0.2985538840293884, "learning_rate": 8e-05, "loss": 1.7804, "step": 1280 }, { "epoch": 0.2817551963048499, "grad_norm": 0.2664276957511902, "learning_rate": 8e-05, "loss": 1.6785, "step": 1281 }, { "epoch": 0.2819751457164852, "grad_norm": 0.3002198040485382, "learning_rate": 8e-05, "loss": 1.6109, "step": 1282 }, { "epoch": 0.2821950951281205, "grad_norm": 0.27687907218933105, "learning_rate": 8e-05, "loss": 1.6322, "step": 1283 }, { "epoch": 0.28241504453975586, "grad_norm": 0.28822144865989685, "learning_rate": 8e-05, "loss": 1.6785, "step": 1284 }, { "epoch": 0.2826349939513912, "grad_norm": 0.2801685333251953, "learning_rate": 8e-05, "loss": 1.69, "step": 1285 }, { "epoch": 0.2828549433630265, "grad_norm": 0.27876734733581543, "learning_rate": 8e-05, "loss": 1.6442, "step": 1286 }, { "epoch": 0.2830748927746618, "grad_norm": 0.2990095317363739, "learning_rate": 8e-05, "loss": 1.7439, "step": 1287 }, { "epoch": 0.28329484218629714, "grad_norm": 0.2710682451725006, "learning_rate": 8e-05, "loss": 1.6908, "step": 1288 }, { "epoch": 0.28351479159793247, "grad_norm": 0.2922731935977936, "learning_rate": 8e-05, "loss": 1.8361, "step": 1289 }, { "epoch": 0.2837347410095678, "grad_norm": 0.2638223171234131, "learning_rate": 8e-05, "loss": 1.6233, "step": 1290 }, { "epoch": 0.28395469042120314, "grad_norm": 0.27564552426338196, "learning_rate": 8e-05, "loss": 1.7624, "step": 1291 }, { "epoch": 0.28417463983283847, "grad_norm": 0.28238940238952637, "learning_rate": 8e-05, "loss": 1.8649, "step": 1292 }, { "epoch": 0.28439458924447375, "grad_norm": 0.27798035740852356, "learning_rate": 8e-05, "loss": 1.7877, "step": 1293 }, { "epoch": 0.2846145386561091, "grad_norm": 0.29618534445762634, "learning_rate": 8e-05, "loss": 1.816, "step": 1294 }, { "epoch": 0.2848344880677444, "grad_norm": 0.27669045329093933, "learning_rate": 8e-05, "loss": 1.7014, "step": 1295 }, { "epoch": 0.28505443747937975, "grad_norm": 0.27973508834838867, "learning_rate": 8e-05, "loss": 1.7491, "step": 1296 }, { "epoch": 0.2852743868910151, "grad_norm": 0.28833356499671936, "learning_rate": 8e-05, "loss": 1.6948, "step": 1297 }, { "epoch": 0.28549433630265036, "grad_norm": 0.2751030921936035, "learning_rate": 8e-05, "loss": 1.6846, "step": 1298 }, { "epoch": 0.2857142857142857, "grad_norm": 0.2766781449317932, "learning_rate": 8e-05, "loss": 1.5442, "step": 1299 }, { "epoch": 0.28593423512592103, "grad_norm": 0.29664894938468933, "learning_rate": 8e-05, "loss": 1.6884, "step": 1300 }, { "epoch": 0.28615418453755637, "grad_norm": 0.2771795392036438, "learning_rate": 8e-05, "loss": 1.6479, "step": 1301 }, { "epoch": 0.2863741339491917, "grad_norm": 0.2623322904109955, "learning_rate": 8e-05, "loss": 1.5803, "step": 1302 }, { "epoch": 0.28659408336082703, "grad_norm": 0.2821153998374939, "learning_rate": 8e-05, "loss": 1.7758, "step": 1303 }, { "epoch": 0.2868140327724623, "grad_norm": 0.29058384895324707, "learning_rate": 8e-05, "loss": 1.7244, "step": 1304 }, { "epoch": 0.28703398218409765, "grad_norm": 0.2811940312385559, "learning_rate": 8e-05, "loss": 1.6708, "step": 1305 }, { "epoch": 0.287253931595733, "grad_norm": 0.2773367762565613, "learning_rate": 8e-05, "loss": 1.7857, "step": 1306 }, { "epoch": 0.2874738810073683, "grad_norm": 0.2689999043941498, "learning_rate": 8e-05, "loss": 1.7432, "step": 1307 }, { "epoch": 0.28769383041900365, "grad_norm": 0.26896870136260986, "learning_rate": 8e-05, "loss": 1.6389, "step": 1308 }, { "epoch": 0.2879137798306389, "grad_norm": 0.2981964349746704, "learning_rate": 8e-05, "loss": 1.8771, "step": 1309 }, { "epoch": 0.28813372924227426, "grad_norm": 0.2872856855392456, "learning_rate": 8e-05, "loss": 1.785, "step": 1310 }, { "epoch": 0.2883536786539096, "grad_norm": 0.3186649680137634, "learning_rate": 8e-05, "loss": 1.9051, "step": 1311 }, { "epoch": 0.28857362806554493, "grad_norm": 0.2802119255065918, "learning_rate": 8e-05, "loss": 1.6532, "step": 1312 }, { "epoch": 0.28879357747718026, "grad_norm": 0.2864134907722473, "learning_rate": 8e-05, "loss": 1.7373, "step": 1313 }, { "epoch": 0.2890135268888156, "grad_norm": 0.2739737331867218, "learning_rate": 8e-05, "loss": 1.5365, "step": 1314 }, { "epoch": 0.2892334763004509, "grad_norm": 0.2707555294036865, "learning_rate": 8e-05, "loss": 1.6516, "step": 1315 }, { "epoch": 0.2894534257120862, "grad_norm": 0.2895212173461914, "learning_rate": 8e-05, "loss": 1.5634, "step": 1316 }, { "epoch": 0.28967337512372154, "grad_norm": 0.26424047350883484, "learning_rate": 8e-05, "loss": 1.6543, "step": 1317 }, { "epoch": 0.2898933245353569, "grad_norm": 0.26237159967422485, "learning_rate": 8e-05, "loss": 1.6814, "step": 1318 }, { "epoch": 0.2901132739469922, "grad_norm": 0.27964159846305847, "learning_rate": 8e-05, "loss": 1.7505, "step": 1319 }, { "epoch": 0.2903332233586275, "grad_norm": 0.27128270268440247, "learning_rate": 8e-05, "loss": 1.7229, "step": 1320 }, { "epoch": 0.2905531727702628, "grad_norm": 0.3012688159942627, "learning_rate": 8e-05, "loss": 1.6851, "step": 1321 }, { "epoch": 0.29077312218189816, "grad_norm": 0.2725695073604584, "learning_rate": 8e-05, "loss": 1.6552, "step": 1322 }, { "epoch": 0.2909930715935335, "grad_norm": 0.2855455279350281, "learning_rate": 8e-05, "loss": 1.7779, "step": 1323 }, { "epoch": 0.2912130210051688, "grad_norm": 0.2906174659729004, "learning_rate": 8e-05, "loss": 1.8209, "step": 1324 }, { "epoch": 0.29143297041680416, "grad_norm": 0.26015472412109375, "learning_rate": 8e-05, "loss": 1.5403, "step": 1325 }, { "epoch": 0.29165291982843944, "grad_norm": 0.29065820574760437, "learning_rate": 8e-05, "loss": 1.8499, "step": 1326 }, { "epoch": 0.2918728692400748, "grad_norm": 0.28715917468070984, "learning_rate": 8e-05, "loss": 1.854, "step": 1327 }, { "epoch": 0.2920928186517101, "grad_norm": 0.26932859420776367, "learning_rate": 8e-05, "loss": 1.6254, "step": 1328 }, { "epoch": 0.29231276806334544, "grad_norm": 0.2757404148578644, "learning_rate": 8e-05, "loss": 1.6077, "step": 1329 }, { "epoch": 0.2925327174749808, "grad_norm": 0.26532551646232605, "learning_rate": 8e-05, "loss": 1.6551, "step": 1330 }, { "epoch": 0.29275266688661605, "grad_norm": 0.2754289209842682, "learning_rate": 8e-05, "loss": 1.7276, "step": 1331 }, { "epoch": 0.2929726162982514, "grad_norm": 0.290568470954895, "learning_rate": 8e-05, "loss": 1.7622, "step": 1332 }, { "epoch": 0.2931925657098867, "grad_norm": 0.3045903742313385, "learning_rate": 8e-05, "loss": 1.7937, "step": 1333 }, { "epoch": 0.29341251512152206, "grad_norm": 0.2594483196735382, "learning_rate": 8e-05, "loss": 1.6163, "step": 1334 }, { "epoch": 0.2936324645331574, "grad_norm": 0.3054102957248688, "learning_rate": 8e-05, "loss": 1.7767, "step": 1335 }, { "epoch": 0.2938524139447927, "grad_norm": 0.27347666025161743, "learning_rate": 8e-05, "loss": 1.682, "step": 1336 }, { "epoch": 0.294072363356428, "grad_norm": 0.2639494836330414, "learning_rate": 8e-05, "loss": 1.4616, "step": 1337 }, { "epoch": 0.29429231276806334, "grad_norm": 0.2842942178249359, "learning_rate": 8e-05, "loss": 1.7625, "step": 1338 }, { "epoch": 0.29451226217969867, "grad_norm": 0.2895960509777069, "learning_rate": 8e-05, "loss": 1.7127, "step": 1339 }, { "epoch": 0.294732211591334, "grad_norm": 0.2836678624153137, "learning_rate": 8e-05, "loss": 1.7765, "step": 1340 }, { "epoch": 0.29495216100296934, "grad_norm": 0.26315444707870483, "learning_rate": 8e-05, "loss": 1.6592, "step": 1341 }, { "epoch": 0.2951721104146046, "grad_norm": 0.2601313591003418, "learning_rate": 8e-05, "loss": 1.5803, "step": 1342 }, { "epoch": 0.29539205982623995, "grad_norm": 0.28084784746170044, "learning_rate": 8e-05, "loss": 1.6172, "step": 1343 }, { "epoch": 0.2956120092378753, "grad_norm": 0.27707698941230774, "learning_rate": 8e-05, "loss": 1.6774, "step": 1344 }, { "epoch": 0.2958319586495106, "grad_norm": 0.28750407695770264, "learning_rate": 8e-05, "loss": 1.7775, "step": 1345 }, { "epoch": 0.29605190806114595, "grad_norm": 0.27315664291381836, "learning_rate": 8e-05, "loss": 1.6578, "step": 1346 }, { "epoch": 0.2962718574727813, "grad_norm": 0.26131486892700195, "learning_rate": 8e-05, "loss": 1.6429, "step": 1347 }, { "epoch": 0.29649180688441656, "grad_norm": 0.27198976278305054, "learning_rate": 8e-05, "loss": 1.6594, "step": 1348 }, { "epoch": 0.2967117562960519, "grad_norm": 0.2785218060016632, "learning_rate": 8e-05, "loss": 1.6959, "step": 1349 }, { "epoch": 0.29693170570768723, "grad_norm": 0.26987215876579285, "learning_rate": 8e-05, "loss": 1.6561, "step": 1350 }, { "epoch": 0.29715165511932257, "grad_norm": 0.2634013295173645, "learning_rate": 8e-05, "loss": 1.6817, "step": 1351 }, { "epoch": 0.2973716045309579, "grad_norm": 0.2584557831287384, "learning_rate": 8e-05, "loss": 1.5104, "step": 1352 }, { "epoch": 0.2975915539425932, "grad_norm": 0.28787991404533386, "learning_rate": 8e-05, "loss": 1.8217, "step": 1353 }, { "epoch": 0.2978115033542285, "grad_norm": 0.5047094225883484, "learning_rate": 8e-05, "loss": 1.7733, "step": 1354 }, { "epoch": 0.29803145276586385, "grad_norm": 0.26776471734046936, "learning_rate": 8e-05, "loss": 1.6961, "step": 1355 }, { "epoch": 0.2982514021774992, "grad_norm": 0.30351778864860535, "learning_rate": 8e-05, "loss": 1.7104, "step": 1356 }, { "epoch": 0.2984713515891345, "grad_norm": 0.27889010310173035, "learning_rate": 8e-05, "loss": 1.7276, "step": 1357 }, { "epoch": 0.29869130100076985, "grad_norm": 0.2656184136867523, "learning_rate": 8e-05, "loss": 1.7438, "step": 1358 }, { "epoch": 0.2989112504124051, "grad_norm": 0.27338340878486633, "learning_rate": 8e-05, "loss": 1.7526, "step": 1359 }, { "epoch": 0.29913119982404046, "grad_norm": 0.3266398310661316, "learning_rate": 8e-05, "loss": 1.8091, "step": 1360 }, { "epoch": 0.2993511492356758, "grad_norm": 0.309469997882843, "learning_rate": 8e-05, "loss": 2.0485, "step": 1361 }, { "epoch": 0.29957109864731113, "grad_norm": 0.2768929600715637, "learning_rate": 8e-05, "loss": 1.7977, "step": 1362 }, { "epoch": 0.29979104805894646, "grad_norm": 0.27685433626174927, "learning_rate": 8e-05, "loss": 1.5712, "step": 1363 }, { "epoch": 0.30001099747058174, "grad_norm": 0.26404622197151184, "learning_rate": 8e-05, "loss": 1.6639, "step": 1364 }, { "epoch": 0.3002309468822171, "grad_norm": 0.2719237208366394, "learning_rate": 8e-05, "loss": 1.788, "step": 1365 }, { "epoch": 0.3004508962938524, "grad_norm": 0.27983394265174866, "learning_rate": 8e-05, "loss": 1.7361, "step": 1366 }, { "epoch": 0.30067084570548774, "grad_norm": 0.2673875689506531, "learning_rate": 8e-05, "loss": 1.6288, "step": 1367 }, { "epoch": 0.3008907951171231, "grad_norm": 0.2850426435470581, "learning_rate": 8e-05, "loss": 1.8328, "step": 1368 }, { "epoch": 0.3011107445287584, "grad_norm": 0.2577967345714569, "learning_rate": 8e-05, "loss": 1.6267, "step": 1369 }, { "epoch": 0.3013306939403937, "grad_norm": 0.276094913482666, "learning_rate": 8e-05, "loss": 1.7673, "step": 1370 }, { "epoch": 0.301550643352029, "grad_norm": 0.2834344208240509, "learning_rate": 8e-05, "loss": 1.6692, "step": 1371 }, { "epoch": 0.30177059276366436, "grad_norm": 0.2617560029029846, "learning_rate": 8e-05, "loss": 1.7734, "step": 1372 }, { "epoch": 0.3019905421752997, "grad_norm": 0.27122870087623596, "learning_rate": 8e-05, "loss": 1.6988, "step": 1373 }, { "epoch": 0.302210491586935, "grad_norm": 0.26526594161987305, "learning_rate": 8e-05, "loss": 1.7459, "step": 1374 }, { "epoch": 0.3024304409985703, "grad_norm": 0.2893051207065582, "learning_rate": 8e-05, "loss": 1.8214, "step": 1375 }, { "epoch": 0.30265039041020564, "grad_norm": 0.2735356092453003, "learning_rate": 8e-05, "loss": 1.8437, "step": 1376 }, { "epoch": 0.302870339821841, "grad_norm": 0.2743459939956665, "learning_rate": 8e-05, "loss": 1.8365, "step": 1377 }, { "epoch": 0.3030902892334763, "grad_norm": 0.28047019243240356, "learning_rate": 8e-05, "loss": 1.6143, "step": 1378 }, { "epoch": 0.30331023864511164, "grad_norm": 0.268197238445282, "learning_rate": 8e-05, "loss": 1.591, "step": 1379 }, { "epoch": 0.303530188056747, "grad_norm": 0.2890843451023102, "learning_rate": 8e-05, "loss": 1.7757, "step": 1380 }, { "epoch": 0.30375013746838225, "grad_norm": 0.2765072286128998, "learning_rate": 8e-05, "loss": 1.6363, "step": 1381 }, { "epoch": 0.3039700868800176, "grad_norm": 0.290147602558136, "learning_rate": 8e-05, "loss": 1.7615, "step": 1382 }, { "epoch": 0.3041900362916529, "grad_norm": 0.2721220850944519, "learning_rate": 8e-05, "loss": 1.7101, "step": 1383 }, { "epoch": 0.30440998570328826, "grad_norm": 0.27125662565231323, "learning_rate": 8e-05, "loss": 1.7291, "step": 1384 }, { "epoch": 0.3046299351149236, "grad_norm": 0.2594304084777832, "learning_rate": 8e-05, "loss": 1.6754, "step": 1385 }, { "epoch": 0.30484988452655887, "grad_norm": 0.28582707047462463, "learning_rate": 8e-05, "loss": 1.6808, "step": 1386 }, { "epoch": 0.3050698339381942, "grad_norm": 0.2853895425796509, "learning_rate": 8e-05, "loss": 1.779, "step": 1387 }, { "epoch": 0.30528978334982954, "grad_norm": 0.2580530345439911, "learning_rate": 8e-05, "loss": 1.6316, "step": 1388 }, { "epoch": 0.30550973276146487, "grad_norm": 0.2793220281600952, "learning_rate": 8e-05, "loss": 1.7326, "step": 1389 }, { "epoch": 0.3057296821731002, "grad_norm": 0.2672085165977478, "learning_rate": 8e-05, "loss": 1.6544, "step": 1390 }, { "epoch": 0.30594963158473554, "grad_norm": 0.27718111872673035, "learning_rate": 8e-05, "loss": 1.6307, "step": 1391 }, { "epoch": 0.3061695809963708, "grad_norm": 0.29295554757118225, "learning_rate": 8e-05, "loss": 1.502, "step": 1392 }, { "epoch": 0.30638953040800615, "grad_norm": 0.2840512990951538, "learning_rate": 8e-05, "loss": 1.6326, "step": 1393 }, { "epoch": 0.3066094798196415, "grad_norm": 0.2897029519081116, "learning_rate": 8e-05, "loss": 1.7543, "step": 1394 }, { "epoch": 0.3068294292312768, "grad_norm": 0.28060710430145264, "learning_rate": 8e-05, "loss": 1.7227, "step": 1395 }, { "epoch": 0.30704937864291215, "grad_norm": 0.27874305844306946, "learning_rate": 8e-05, "loss": 1.6639, "step": 1396 }, { "epoch": 0.30726932805454743, "grad_norm": 0.2679193615913391, "learning_rate": 8e-05, "loss": 1.7226, "step": 1397 }, { "epoch": 0.30748927746618276, "grad_norm": 0.2769779562950134, "learning_rate": 8e-05, "loss": 1.6384, "step": 1398 }, { "epoch": 0.3077092268778181, "grad_norm": 0.26620879769325256, "learning_rate": 8e-05, "loss": 1.7134, "step": 1399 }, { "epoch": 0.30792917628945343, "grad_norm": 0.277423620223999, "learning_rate": 8e-05, "loss": 1.7376, "step": 1400 }, { "epoch": 0.30814912570108877, "grad_norm": 0.2629416882991791, "learning_rate": 8e-05, "loss": 1.598, "step": 1401 }, { "epoch": 0.3083690751127241, "grad_norm": 0.2844812572002411, "learning_rate": 8e-05, "loss": 1.7067, "step": 1402 }, { "epoch": 0.3085890245243594, "grad_norm": 0.2731526494026184, "learning_rate": 8e-05, "loss": 1.8571, "step": 1403 }, { "epoch": 0.3088089739359947, "grad_norm": 0.287438303232193, "learning_rate": 8e-05, "loss": 1.7612, "step": 1404 }, { "epoch": 0.30902892334763005, "grad_norm": 0.266718327999115, "learning_rate": 8e-05, "loss": 1.6106, "step": 1405 }, { "epoch": 0.3092488727592654, "grad_norm": 0.28080686926841736, "learning_rate": 8e-05, "loss": 1.8281, "step": 1406 }, { "epoch": 0.3094688221709007, "grad_norm": 0.27558308839797974, "learning_rate": 8e-05, "loss": 1.8677, "step": 1407 }, { "epoch": 0.309688771582536, "grad_norm": 0.2798183262348175, "learning_rate": 8e-05, "loss": 1.7867, "step": 1408 }, { "epoch": 0.3099087209941713, "grad_norm": 0.25823187828063965, "learning_rate": 8e-05, "loss": 1.6743, "step": 1409 }, { "epoch": 0.31012867040580666, "grad_norm": 0.27356335520744324, "learning_rate": 8e-05, "loss": 1.7039, "step": 1410 }, { "epoch": 0.310348619817442, "grad_norm": 0.2842661440372467, "learning_rate": 8e-05, "loss": 1.7046, "step": 1411 }, { "epoch": 0.31056856922907733, "grad_norm": 0.2561197876930237, "learning_rate": 8e-05, "loss": 1.4887, "step": 1412 }, { "epoch": 0.31078851864071266, "grad_norm": 0.2851184904575348, "learning_rate": 8e-05, "loss": 1.7074, "step": 1413 }, { "epoch": 0.31100846805234794, "grad_norm": 0.2655506432056427, "learning_rate": 8e-05, "loss": 1.6049, "step": 1414 }, { "epoch": 0.3112284174639833, "grad_norm": 0.26412099599838257, "learning_rate": 8e-05, "loss": 1.6052, "step": 1415 }, { "epoch": 0.3114483668756186, "grad_norm": 0.3026227056980133, "learning_rate": 8e-05, "loss": 1.7085, "step": 1416 }, { "epoch": 0.31166831628725394, "grad_norm": 0.28821703791618347, "learning_rate": 8e-05, "loss": 1.7573, "step": 1417 }, { "epoch": 0.3118882656988893, "grad_norm": 0.26806455850601196, "learning_rate": 8e-05, "loss": 1.7136, "step": 1418 }, { "epoch": 0.31210821511052456, "grad_norm": 0.28336799144744873, "learning_rate": 8e-05, "loss": 1.8445, "step": 1419 }, { "epoch": 0.3123281645221599, "grad_norm": 0.2772139012813568, "learning_rate": 8e-05, "loss": 1.692, "step": 1420 }, { "epoch": 0.3125481139337952, "grad_norm": 0.2815256714820862, "learning_rate": 8e-05, "loss": 1.77, "step": 1421 }, { "epoch": 0.31276806334543056, "grad_norm": 0.4029920697212219, "learning_rate": 8e-05, "loss": 1.8103, "step": 1422 }, { "epoch": 0.3129880127570659, "grad_norm": 0.2677610218524933, "learning_rate": 8e-05, "loss": 1.5898, "step": 1423 }, { "epoch": 0.3132079621687012, "grad_norm": 0.2605397701263428, "learning_rate": 8e-05, "loss": 1.5735, "step": 1424 }, { "epoch": 0.3134279115803365, "grad_norm": 0.2831586003303528, "learning_rate": 8e-05, "loss": 1.6641, "step": 1425 }, { "epoch": 0.31364786099197184, "grad_norm": 0.2746485471725464, "learning_rate": 8e-05, "loss": 1.618, "step": 1426 }, { "epoch": 0.3138678104036072, "grad_norm": 0.283342182636261, "learning_rate": 8e-05, "loss": 1.6963, "step": 1427 }, { "epoch": 0.3140877598152425, "grad_norm": 0.27635300159454346, "learning_rate": 8e-05, "loss": 1.6911, "step": 1428 }, { "epoch": 0.31430770922687784, "grad_norm": 0.2719132900238037, "learning_rate": 8e-05, "loss": 1.7063, "step": 1429 }, { "epoch": 0.3145276586385131, "grad_norm": 0.27162256836891174, "learning_rate": 8e-05, "loss": 1.6397, "step": 1430 }, { "epoch": 0.31474760805014845, "grad_norm": 0.2934938073158264, "learning_rate": 8e-05, "loss": 1.7555, "step": 1431 }, { "epoch": 0.3149675574617838, "grad_norm": 0.3060123920440674, "learning_rate": 8e-05, "loss": 1.642, "step": 1432 }, { "epoch": 0.3151875068734191, "grad_norm": 0.280846506357193, "learning_rate": 8e-05, "loss": 1.6805, "step": 1433 }, { "epoch": 0.31540745628505445, "grad_norm": 0.2768997550010681, "learning_rate": 8e-05, "loss": 1.7359, "step": 1434 }, { "epoch": 0.3156274056966898, "grad_norm": 0.29172810912132263, "learning_rate": 8e-05, "loss": 1.821, "step": 1435 }, { "epoch": 0.31584735510832507, "grad_norm": 0.30742648243904114, "learning_rate": 8e-05, "loss": 1.8198, "step": 1436 }, { "epoch": 0.3160673045199604, "grad_norm": 0.2889997065067291, "learning_rate": 8e-05, "loss": 1.6733, "step": 1437 }, { "epoch": 0.31628725393159574, "grad_norm": 0.2859675884246826, "learning_rate": 8e-05, "loss": 1.7655, "step": 1438 }, { "epoch": 0.31650720334323107, "grad_norm": 0.2926831543445587, "learning_rate": 8e-05, "loss": 1.7871, "step": 1439 }, { "epoch": 0.3167271527548664, "grad_norm": 0.28924524784088135, "learning_rate": 8e-05, "loss": 1.665, "step": 1440 }, { "epoch": 0.3169471021665017, "grad_norm": 0.2940097749233246, "learning_rate": 8e-05, "loss": 1.8364, "step": 1441 }, { "epoch": 0.317167051578137, "grad_norm": 0.2923974096775055, "learning_rate": 8e-05, "loss": 1.8071, "step": 1442 }, { "epoch": 0.31738700098977235, "grad_norm": 0.28991878032684326, "learning_rate": 8e-05, "loss": 1.7445, "step": 1443 }, { "epoch": 0.3176069504014077, "grad_norm": 0.283600777387619, "learning_rate": 8e-05, "loss": 1.8043, "step": 1444 }, { "epoch": 0.317826899813043, "grad_norm": 0.3082323372364044, "learning_rate": 8e-05, "loss": 1.7858, "step": 1445 }, { "epoch": 0.3180468492246783, "grad_norm": 0.28433462977409363, "learning_rate": 8e-05, "loss": 1.6911, "step": 1446 }, { "epoch": 0.31826679863631363, "grad_norm": 0.27776578068733215, "learning_rate": 8e-05, "loss": 1.7212, "step": 1447 }, { "epoch": 0.31848674804794896, "grad_norm": 0.29395151138305664, "learning_rate": 8e-05, "loss": 1.7221, "step": 1448 }, { "epoch": 0.3187066974595843, "grad_norm": 0.27507245540618896, "learning_rate": 8e-05, "loss": 1.7358, "step": 1449 }, { "epoch": 0.31892664687121963, "grad_norm": 0.25614190101623535, "learning_rate": 8e-05, "loss": 1.5138, "step": 1450 }, { "epoch": 0.31914659628285497, "grad_norm": 0.2908024489879608, "learning_rate": 8e-05, "loss": 1.756, "step": 1451 }, { "epoch": 0.31936654569449024, "grad_norm": 0.2729463577270508, "learning_rate": 8e-05, "loss": 1.5542, "step": 1452 }, { "epoch": 0.3195864951061256, "grad_norm": 0.27094194293022156, "learning_rate": 8e-05, "loss": 1.5917, "step": 1453 }, { "epoch": 0.3198064445177609, "grad_norm": 0.28125494718551636, "learning_rate": 8e-05, "loss": 1.6584, "step": 1454 }, { "epoch": 0.32002639392939625, "grad_norm": 0.29033198952674866, "learning_rate": 8e-05, "loss": 1.7332, "step": 1455 }, { "epoch": 0.3202463433410316, "grad_norm": 0.26570284366607666, "learning_rate": 8e-05, "loss": 1.6159, "step": 1456 }, { "epoch": 0.32046629275266686, "grad_norm": 0.307412713766098, "learning_rate": 8e-05, "loss": 1.7351, "step": 1457 }, { "epoch": 0.3206862421643022, "grad_norm": 0.29387474060058594, "learning_rate": 8e-05, "loss": 1.9386, "step": 1458 }, { "epoch": 0.3209061915759375, "grad_norm": 0.26545315980911255, "learning_rate": 8e-05, "loss": 1.6343, "step": 1459 }, { "epoch": 0.32112614098757286, "grad_norm": 0.279238224029541, "learning_rate": 8e-05, "loss": 1.6245, "step": 1460 }, { "epoch": 0.3213460903992082, "grad_norm": 0.2766862213611603, "learning_rate": 8e-05, "loss": 1.7135, "step": 1461 }, { "epoch": 0.32156603981084353, "grad_norm": 0.2705351412296295, "learning_rate": 8e-05, "loss": 1.6526, "step": 1462 }, { "epoch": 0.3217859892224788, "grad_norm": 0.27870967984199524, "learning_rate": 8e-05, "loss": 1.6512, "step": 1463 }, { "epoch": 0.32200593863411414, "grad_norm": 0.284407377243042, "learning_rate": 8e-05, "loss": 1.755, "step": 1464 }, { "epoch": 0.3222258880457495, "grad_norm": 0.2897641062736511, "learning_rate": 8e-05, "loss": 1.8383, "step": 1465 }, { "epoch": 0.3224458374573848, "grad_norm": 0.2667568624019623, "learning_rate": 8e-05, "loss": 1.6989, "step": 1466 }, { "epoch": 0.32266578686902014, "grad_norm": 0.26580294966697693, "learning_rate": 8e-05, "loss": 1.5895, "step": 1467 }, { "epoch": 0.3228857362806554, "grad_norm": 0.26188549399375916, "learning_rate": 8e-05, "loss": 1.5799, "step": 1468 }, { "epoch": 0.32310568569229076, "grad_norm": 0.27703747153282166, "learning_rate": 8e-05, "loss": 1.8306, "step": 1469 }, { "epoch": 0.3233256351039261, "grad_norm": 0.27643802762031555, "learning_rate": 8e-05, "loss": 1.6864, "step": 1470 }, { "epoch": 0.3235455845155614, "grad_norm": 0.27216553688049316, "learning_rate": 8e-05, "loss": 1.6006, "step": 1471 }, { "epoch": 0.32376553392719676, "grad_norm": 0.2984940707683563, "learning_rate": 8e-05, "loss": 1.7548, "step": 1472 }, { "epoch": 0.3239854833388321, "grad_norm": 0.30579298734664917, "learning_rate": 8e-05, "loss": 1.8307, "step": 1473 }, { "epoch": 0.32420543275046737, "grad_norm": 0.27524709701538086, "learning_rate": 8e-05, "loss": 1.6134, "step": 1474 }, { "epoch": 0.3244253821621027, "grad_norm": 0.2788650393486023, "learning_rate": 8e-05, "loss": 1.8194, "step": 1475 }, { "epoch": 0.32464533157373804, "grad_norm": 0.28263744711875916, "learning_rate": 8e-05, "loss": 1.7633, "step": 1476 }, { "epoch": 0.3248652809853734, "grad_norm": 0.30234408378601074, "learning_rate": 8e-05, "loss": 1.6057, "step": 1477 }, { "epoch": 0.3250852303970087, "grad_norm": 0.2820134162902832, "learning_rate": 8e-05, "loss": 1.6913, "step": 1478 }, { "epoch": 0.325305179808644, "grad_norm": 0.28929245471954346, "learning_rate": 8e-05, "loss": 1.9538, "step": 1479 }, { "epoch": 0.3255251292202793, "grad_norm": 0.26399463415145874, "learning_rate": 8e-05, "loss": 1.7309, "step": 1480 }, { "epoch": 0.32574507863191465, "grad_norm": 0.2722630202770233, "learning_rate": 8e-05, "loss": 1.5595, "step": 1481 }, { "epoch": 0.32596502804355, "grad_norm": 0.2759261727333069, "learning_rate": 8e-05, "loss": 1.6272, "step": 1482 }, { "epoch": 0.3261849774551853, "grad_norm": 0.28047022223472595, "learning_rate": 8e-05, "loss": 1.6933, "step": 1483 }, { "epoch": 0.32640492686682065, "grad_norm": 0.2835995554924011, "learning_rate": 8e-05, "loss": 1.7165, "step": 1484 }, { "epoch": 0.32662487627845593, "grad_norm": 0.28965097665786743, "learning_rate": 8e-05, "loss": 1.6399, "step": 1485 }, { "epoch": 0.32684482569009127, "grad_norm": 0.2729817032814026, "learning_rate": 8e-05, "loss": 1.7397, "step": 1486 }, { "epoch": 0.3270647751017266, "grad_norm": 0.26809874176979065, "learning_rate": 8e-05, "loss": 1.6954, "step": 1487 }, { "epoch": 0.32728472451336194, "grad_norm": 0.29766684770584106, "learning_rate": 8e-05, "loss": 1.6947, "step": 1488 }, { "epoch": 0.32750467392499727, "grad_norm": 0.27032670378685, "learning_rate": 8e-05, "loss": 1.8036, "step": 1489 }, { "epoch": 0.32772462333663255, "grad_norm": 0.2694716453552246, "learning_rate": 8e-05, "loss": 1.6856, "step": 1490 }, { "epoch": 0.3279445727482679, "grad_norm": 0.27968841791152954, "learning_rate": 8e-05, "loss": 1.7466, "step": 1491 }, { "epoch": 0.3281645221599032, "grad_norm": 0.2956348955631256, "learning_rate": 8e-05, "loss": 1.833, "step": 1492 }, { "epoch": 0.32838447157153855, "grad_norm": 0.27069491147994995, "learning_rate": 8e-05, "loss": 1.715, "step": 1493 }, { "epoch": 0.3286044209831739, "grad_norm": 0.26747795939445496, "learning_rate": 8e-05, "loss": 1.6663, "step": 1494 }, { "epoch": 0.3288243703948092, "grad_norm": 0.2619915008544922, "learning_rate": 8e-05, "loss": 1.6503, "step": 1495 }, { "epoch": 0.3290443198064445, "grad_norm": 0.2720276117324829, "learning_rate": 8e-05, "loss": 1.7174, "step": 1496 }, { "epoch": 0.32926426921807983, "grad_norm": 0.26874253153800964, "learning_rate": 8e-05, "loss": 1.6503, "step": 1497 }, { "epoch": 0.32948421862971516, "grad_norm": 0.28397336602211, "learning_rate": 8e-05, "loss": 1.67, "step": 1498 }, { "epoch": 0.3297041680413505, "grad_norm": 0.2544403076171875, "learning_rate": 8e-05, "loss": 1.5153, "step": 1499 }, { "epoch": 0.32992411745298583, "grad_norm": 0.2819180488586426, "learning_rate": 8e-05, "loss": 1.6704, "step": 1500 }, { "epoch": 0.3301440668646211, "grad_norm": 0.28150951862335205, "learning_rate": 8e-05, "loss": 1.8451, "step": 1501 }, { "epoch": 0.33036401627625644, "grad_norm": 0.27396339178085327, "learning_rate": 8e-05, "loss": 1.7631, "step": 1502 }, { "epoch": 0.3305839656878918, "grad_norm": 0.2954351007938385, "learning_rate": 8e-05, "loss": 1.8101, "step": 1503 }, { "epoch": 0.3308039150995271, "grad_norm": 0.27129319310188293, "learning_rate": 8e-05, "loss": 1.6484, "step": 1504 }, { "epoch": 0.33102386451116245, "grad_norm": 0.27612754702568054, "learning_rate": 8e-05, "loss": 1.6178, "step": 1505 }, { "epoch": 0.3312438139227978, "grad_norm": 0.26097655296325684, "learning_rate": 8e-05, "loss": 1.5781, "step": 1506 }, { "epoch": 0.33146376333443306, "grad_norm": 0.2704753577709198, "learning_rate": 8e-05, "loss": 1.6919, "step": 1507 }, { "epoch": 0.3316837127460684, "grad_norm": 0.26866593956947327, "learning_rate": 8e-05, "loss": 1.6795, "step": 1508 }, { "epoch": 0.3319036621577037, "grad_norm": 0.31797948479652405, "learning_rate": 8e-05, "loss": 1.7511, "step": 1509 }, { "epoch": 0.33212361156933906, "grad_norm": 0.29456841945648193, "learning_rate": 8e-05, "loss": 1.7041, "step": 1510 }, { "epoch": 0.3323435609809744, "grad_norm": 0.28345033526420593, "learning_rate": 8e-05, "loss": 1.7499, "step": 1511 }, { "epoch": 0.3325635103926097, "grad_norm": 0.28679129481315613, "learning_rate": 8e-05, "loss": 1.8304, "step": 1512 }, { "epoch": 0.332783459804245, "grad_norm": 0.2799399793148041, "learning_rate": 8e-05, "loss": 1.6461, "step": 1513 }, { "epoch": 0.33300340921588034, "grad_norm": 0.3234422206878662, "learning_rate": 8e-05, "loss": 1.5622, "step": 1514 }, { "epoch": 0.3332233586275157, "grad_norm": 0.27786344289779663, "learning_rate": 8e-05, "loss": 1.6718, "step": 1515 }, { "epoch": 0.333443308039151, "grad_norm": 0.27040839195251465, "learning_rate": 8e-05, "loss": 1.7428, "step": 1516 }, { "epoch": 0.33366325745078634, "grad_norm": 0.2837252616882324, "learning_rate": 8e-05, "loss": 1.6929, "step": 1517 }, { "epoch": 0.3338832068624216, "grad_norm": 0.27352792024612427, "learning_rate": 8e-05, "loss": 1.7804, "step": 1518 }, { "epoch": 0.33410315627405696, "grad_norm": 0.27237218618392944, "learning_rate": 8e-05, "loss": 1.7652, "step": 1519 }, { "epoch": 0.3343231056856923, "grad_norm": 0.3166270852088928, "learning_rate": 8e-05, "loss": 1.6363, "step": 1520 }, { "epoch": 0.3345430550973276, "grad_norm": 0.2650817930698395, "learning_rate": 8e-05, "loss": 1.6954, "step": 1521 }, { "epoch": 0.33476300450896296, "grad_norm": 0.2907481789588928, "learning_rate": 8e-05, "loss": 1.809, "step": 1522 }, { "epoch": 0.33498295392059824, "grad_norm": 0.2754502296447754, "learning_rate": 8e-05, "loss": 1.8143, "step": 1523 }, { "epoch": 0.33520290333223357, "grad_norm": 0.2890012264251709, "learning_rate": 8e-05, "loss": 1.6603, "step": 1524 }, { "epoch": 0.3354228527438689, "grad_norm": 0.271720826625824, "learning_rate": 8e-05, "loss": 1.7186, "step": 1525 }, { "epoch": 0.33564280215550424, "grad_norm": 0.2845331132411957, "learning_rate": 8e-05, "loss": 1.7739, "step": 1526 }, { "epoch": 0.3358627515671396, "grad_norm": 0.2787776291370392, "learning_rate": 8e-05, "loss": 1.6146, "step": 1527 }, { "epoch": 0.3360827009787749, "grad_norm": 0.2612919211387634, "learning_rate": 8e-05, "loss": 1.5575, "step": 1528 }, { "epoch": 0.3363026503904102, "grad_norm": 0.279220849275589, "learning_rate": 8e-05, "loss": 1.7661, "step": 1529 }, { "epoch": 0.3365225998020455, "grad_norm": 0.2812168300151825, "learning_rate": 8e-05, "loss": 1.7011, "step": 1530 }, { "epoch": 0.33674254921368085, "grad_norm": 0.28216826915740967, "learning_rate": 8e-05, "loss": 1.7856, "step": 1531 }, { "epoch": 0.3369624986253162, "grad_norm": 0.279895156621933, "learning_rate": 8e-05, "loss": 1.6793, "step": 1532 }, { "epoch": 0.3371824480369515, "grad_norm": 0.2694056034088135, "learning_rate": 8e-05, "loss": 1.6289, "step": 1533 }, { "epoch": 0.3374023974485868, "grad_norm": 0.2692592740058899, "learning_rate": 8e-05, "loss": 1.5595, "step": 1534 }, { "epoch": 0.33762234686022213, "grad_norm": 0.32149383425712585, "learning_rate": 8e-05, "loss": 1.6667, "step": 1535 }, { "epoch": 0.33784229627185747, "grad_norm": 0.28884437680244446, "learning_rate": 8e-05, "loss": 1.7836, "step": 1536 }, { "epoch": 0.3380622456834928, "grad_norm": 0.276017963886261, "learning_rate": 8e-05, "loss": 1.712, "step": 1537 }, { "epoch": 0.33828219509512814, "grad_norm": 0.26901450753211975, "learning_rate": 8e-05, "loss": 1.6442, "step": 1538 }, { "epoch": 0.33850214450676347, "grad_norm": 0.29827412962913513, "learning_rate": 8e-05, "loss": 1.7619, "step": 1539 }, { "epoch": 0.33872209391839875, "grad_norm": 0.2763231098651886, "learning_rate": 8e-05, "loss": 1.6344, "step": 1540 }, { "epoch": 0.3389420433300341, "grad_norm": 0.26493677496910095, "learning_rate": 8e-05, "loss": 1.6964, "step": 1541 }, { "epoch": 0.3391619927416694, "grad_norm": 0.2956371605396271, "learning_rate": 8e-05, "loss": 1.7328, "step": 1542 }, { "epoch": 0.33938194215330475, "grad_norm": 0.2845339775085449, "learning_rate": 8e-05, "loss": 1.6477, "step": 1543 }, { "epoch": 0.3396018915649401, "grad_norm": 0.29501214623451233, "learning_rate": 8e-05, "loss": 1.7951, "step": 1544 }, { "epoch": 0.33982184097657536, "grad_norm": 0.2859644591808319, "learning_rate": 8e-05, "loss": 1.6607, "step": 1545 }, { "epoch": 0.3400417903882107, "grad_norm": 0.2733168303966522, "learning_rate": 8e-05, "loss": 1.6397, "step": 1546 }, { "epoch": 0.34026173979984603, "grad_norm": 0.2580598294734955, "learning_rate": 8e-05, "loss": 1.5692, "step": 1547 }, { "epoch": 0.34048168921148136, "grad_norm": 0.3042803406715393, "learning_rate": 8e-05, "loss": 1.7063, "step": 1548 }, { "epoch": 0.3407016386231167, "grad_norm": 0.2833859324455261, "learning_rate": 8e-05, "loss": 1.7531, "step": 1549 }, { "epoch": 0.34092158803475203, "grad_norm": 0.259620726108551, "learning_rate": 8e-05, "loss": 1.6179, "step": 1550 }, { "epoch": 0.3411415374463873, "grad_norm": 0.268355131149292, "learning_rate": 8e-05, "loss": 1.6009, "step": 1551 }, { "epoch": 0.34136148685802264, "grad_norm": 0.2858780324459076, "learning_rate": 8e-05, "loss": 1.7033, "step": 1552 }, { "epoch": 0.341581436269658, "grad_norm": 0.2777354121208191, "learning_rate": 8e-05, "loss": 1.7615, "step": 1553 }, { "epoch": 0.3418013856812933, "grad_norm": 0.27899524569511414, "learning_rate": 8e-05, "loss": 1.6684, "step": 1554 }, { "epoch": 0.34202133509292865, "grad_norm": 0.3156200349330902, "learning_rate": 8e-05, "loss": 1.5658, "step": 1555 }, { "epoch": 0.3422412845045639, "grad_norm": 0.27549582719802856, "learning_rate": 8e-05, "loss": 1.692, "step": 1556 }, { "epoch": 0.34246123391619926, "grad_norm": 0.27770310640335083, "learning_rate": 8e-05, "loss": 1.5891, "step": 1557 }, { "epoch": 0.3426811833278346, "grad_norm": 0.28138646483421326, "learning_rate": 8e-05, "loss": 1.5949, "step": 1558 }, { "epoch": 0.3429011327394699, "grad_norm": 0.2790684998035431, "learning_rate": 8e-05, "loss": 1.6371, "step": 1559 }, { "epoch": 0.34312108215110526, "grad_norm": 0.303230345249176, "learning_rate": 8e-05, "loss": 1.7416, "step": 1560 }, { "epoch": 0.3433410315627406, "grad_norm": 0.26891767978668213, "learning_rate": 8e-05, "loss": 1.8044, "step": 1561 }, { "epoch": 0.3435609809743759, "grad_norm": 0.2734631896018982, "learning_rate": 8e-05, "loss": 1.7171, "step": 1562 }, { "epoch": 0.3437809303860112, "grad_norm": 0.29556018114089966, "learning_rate": 8e-05, "loss": 1.9085, "step": 1563 }, { "epoch": 0.34400087979764654, "grad_norm": 0.26478004455566406, "learning_rate": 8e-05, "loss": 1.6153, "step": 1564 }, { "epoch": 0.3442208292092819, "grad_norm": 0.27655404806137085, "learning_rate": 8e-05, "loss": 1.7384, "step": 1565 }, { "epoch": 0.3444407786209172, "grad_norm": 0.2902698218822479, "learning_rate": 8e-05, "loss": 1.6589, "step": 1566 }, { "epoch": 0.3446607280325525, "grad_norm": 0.2857147455215454, "learning_rate": 8e-05, "loss": 1.6598, "step": 1567 }, { "epoch": 0.3448806774441878, "grad_norm": 0.28339943289756775, "learning_rate": 8e-05, "loss": 1.7356, "step": 1568 }, { "epoch": 0.34510062685582316, "grad_norm": 0.29340776801109314, "learning_rate": 8e-05, "loss": 1.8316, "step": 1569 }, { "epoch": 0.3453205762674585, "grad_norm": 0.26669397950172424, "learning_rate": 8e-05, "loss": 1.6803, "step": 1570 }, { "epoch": 0.3455405256790938, "grad_norm": 0.28508248925209045, "learning_rate": 8e-05, "loss": 1.7702, "step": 1571 }, { "epoch": 0.34576047509072916, "grad_norm": 0.25610047578811646, "learning_rate": 8e-05, "loss": 1.6343, "step": 1572 }, { "epoch": 0.34598042450236444, "grad_norm": 0.2758273482322693, "learning_rate": 8e-05, "loss": 1.7875, "step": 1573 }, { "epoch": 0.34620037391399977, "grad_norm": 0.2674688398838043, "learning_rate": 8e-05, "loss": 1.6804, "step": 1574 }, { "epoch": 0.3464203233256351, "grad_norm": 0.2796163558959961, "learning_rate": 8e-05, "loss": 1.6135, "step": 1575 }, { "epoch": 0.34664027273727044, "grad_norm": 0.26260775327682495, "learning_rate": 8e-05, "loss": 1.6752, "step": 1576 }, { "epoch": 0.3468602221489058, "grad_norm": 0.2897137403488159, "learning_rate": 8e-05, "loss": 1.6743, "step": 1577 }, { "epoch": 0.34708017156054105, "grad_norm": 0.27681732177734375, "learning_rate": 8e-05, "loss": 1.6436, "step": 1578 }, { "epoch": 0.3473001209721764, "grad_norm": 0.2694265842437744, "learning_rate": 8e-05, "loss": 1.6343, "step": 1579 }, { "epoch": 0.3475200703838117, "grad_norm": 0.28179508447647095, "learning_rate": 8e-05, "loss": 1.676, "step": 1580 }, { "epoch": 0.34774001979544705, "grad_norm": 0.29600057005882263, "learning_rate": 8e-05, "loss": 1.786, "step": 1581 }, { "epoch": 0.3479599692070824, "grad_norm": 0.28932616114616394, "learning_rate": 8e-05, "loss": 1.7151, "step": 1582 }, { "epoch": 0.3481799186187177, "grad_norm": 0.2912417948246002, "learning_rate": 8e-05, "loss": 1.7788, "step": 1583 }, { "epoch": 0.348399868030353, "grad_norm": 0.2844431698322296, "learning_rate": 8e-05, "loss": 1.5585, "step": 1584 }, { "epoch": 0.34861981744198833, "grad_norm": 0.2916630804538727, "learning_rate": 8e-05, "loss": 1.7484, "step": 1585 }, { "epoch": 0.34883976685362367, "grad_norm": 0.2785089612007141, "learning_rate": 8e-05, "loss": 1.61, "step": 1586 }, { "epoch": 0.349059716265259, "grad_norm": 0.2777422368526459, "learning_rate": 8e-05, "loss": 1.7183, "step": 1587 }, { "epoch": 0.34927966567689434, "grad_norm": 0.28772565722465515, "learning_rate": 8e-05, "loss": 1.7161, "step": 1588 }, { "epoch": 0.3494996150885296, "grad_norm": 0.28452831506729126, "learning_rate": 8e-05, "loss": 1.8004, "step": 1589 }, { "epoch": 0.34971956450016495, "grad_norm": 0.2837449014186859, "learning_rate": 8e-05, "loss": 1.7992, "step": 1590 }, { "epoch": 0.3499395139118003, "grad_norm": 0.2874920666217804, "learning_rate": 8e-05, "loss": 1.6408, "step": 1591 }, { "epoch": 0.3501594633234356, "grad_norm": 0.26615065336227417, "learning_rate": 8e-05, "loss": 1.658, "step": 1592 }, { "epoch": 0.35037941273507095, "grad_norm": 0.27493569254875183, "learning_rate": 8e-05, "loss": 1.6843, "step": 1593 }, { "epoch": 0.3505993621467063, "grad_norm": 0.291886568069458, "learning_rate": 8e-05, "loss": 1.7683, "step": 1594 }, { "epoch": 0.35081931155834156, "grad_norm": 0.2868814468383789, "learning_rate": 8e-05, "loss": 1.7825, "step": 1595 }, { "epoch": 0.3510392609699769, "grad_norm": 0.30988067388534546, "learning_rate": 8e-05, "loss": 1.7511, "step": 1596 }, { "epoch": 0.35125921038161223, "grad_norm": 0.2746553122997284, "learning_rate": 8e-05, "loss": 1.6298, "step": 1597 }, { "epoch": 0.35147915979324756, "grad_norm": 0.3013536036014557, "learning_rate": 8e-05, "loss": 1.7883, "step": 1598 }, { "epoch": 0.3516991092048829, "grad_norm": 0.2906748056411743, "learning_rate": 8e-05, "loss": 1.5819, "step": 1599 }, { "epoch": 0.3519190586165182, "grad_norm": 0.28082364797592163, "learning_rate": 8e-05, "loss": 1.7525, "step": 1600 }, { "epoch": 0.3521390080281535, "grad_norm": 0.28713324666023254, "learning_rate": 8e-05, "loss": 1.79, "step": 1601 }, { "epoch": 0.35235895743978884, "grad_norm": 0.2819896638393402, "learning_rate": 8e-05, "loss": 1.6514, "step": 1602 }, { "epoch": 0.3525789068514242, "grad_norm": 0.27669310569763184, "learning_rate": 8e-05, "loss": 1.5888, "step": 1603 }, { "epoch": 0.3527988562630595, "grad_norm": 0.2873641848564148, "learning_rate": 8e-05, "loss": 1.8206, "step": 1604 }, { "epoch": 0.35301880567469485, "grad_norm": 0.28426647186279297, "learning_rate": 8e-05, "loss": 1.7736, "step": 1605 }, { "epoch": 0.3532387550863301, "grad_norm": 0.2733590602874756, "learning_rate": 8e-05, "loss": 1.6653, "step": 1606 }, { "epoch": 0.35345870449796546, "grad_norm": 0.26751479506492615, "learning_rate": 8e-05, "loss": 1.5841, "step": 1607 }, { "epoch": 0.3536786539096008, "grad_norm": 0.2767663598060608, "learning_rate": 8e-05, "loss": 1.6859, "step": 1608 }, { "epoch": 0.3538986033212361, "grad_norm": 0.28359255194664, "learning_rate": 8e-05, "loss": 1.8799, "step": 1609 }, { "epoch": 0.35411855273287146, "grad_norm": 0.27551594376564026, "learning_rate": 8e-05, "loss": 1.6429, "step": 1610 }, { "epoch": 0.35433850214450674, "grad_norm": 0.26260972023010254, "learning_rate": 8e-05, "loss": 1.6068, "step": 1611 }, { "epoch": 0.3545584515561421, "grad_norm": 0.2778937518596649, "learning_rate": 8e-05, "loss": 1.8057, "step": 1612 }, { "epoch": 0.3547784009677774, "grad_norm": 0.27607765793800354, "learning_rate": 8e-05, "loss": 1.7439, "step": 1613 }, { "epoch": 0.35499835037941274, "grad_norm": 0.2628287076950073, "learning_rate": 8e-05, "loss": 1.6916, "step": 1614 }, { "epoch": 0.3552182997910481, "grad_norm": 0.2767592966556549, "learning_rate": 8e-05, "loss": 1.7185, "step": 1615 }, { "epoch": 0.3554382492026834, "grad_norm": 0.2666943669319153, "learning_rate": 8e-05, "loss": 1.7351, "step": 1616 }, { "epoch": 0.3556581986143187, "grad_norm": 0.28780093789100647, "learning_rate": 8e-05, "loss": 1.757, "step": 1617 }, { "epoch": 0.355878148025954, "grad_norm": 0.30761584639549255, "learning_rate": 8e-05, "loss": 1.8096, "step": 1618 }, { "epoch": 0.35609809743758936, "grad_norm": 0.2926090359687805, "learning_rate": 8e-05, "loss": 1.8609, "step": 1619 }, { "epoch": 0.3563180468492247, "grad_norm": 0.27546852827072144, "learning_rate": 8e-05, "loss": 1.6422, "step": 1620 }, { "epoch": 0.35653799626086, "grad_norm": 0.28559309244155884, "learning_rate": 8e-05, "loss": 1.8225, "step": 1621 }, { "epoch": 0.3567579456724953, "grad_norm": 0.2804494798183441, "learning_rate": 8e-05, "loss": 1.9108, "step": 1622 }, { "epoch": 0.35697789508413064, "grad_norm": 0.2643645703792572, "learning_rate": 8e-05, "loss": 1.5462, "step": 1623 }, { "epoch": 0.35719784449576597, "grad_norm": 0.2888531982898712, "learning_rate": 8e-05, "loss": 1.701, "step": 1624 }, { "epoch": 0.3574177939074013, "grad_norm": 0.28601035475730896, "learning_rate": 8e-05, "loss": 1.628, "step": 1625 }, { "epoch": 0.35763774331903664, "grad_norm": 0.2877524197101593, "learning_rate": 8e-05, "loss": 1.8403, "step": 1626 }, { "epoch": 0.357857692730672, "grad_norm": 0.2658945918083191, "learning_rate": 8e-05, "loss": 1.4552, "step": 1627 }, { "epoch": 0.35807764214230725, "grad_norm": 0.2911885976791382, "learning_rate": 8e-05, "loss": 1.7753, "step": 1628 }, { "epoch": 0.3582975915539426, "grad_norm": 0.29072439670562744, "learning_rate": 8e-05, "loss": 1.7229, "step": 1629 }, { "epoch": 0.3585175409655779, "grad_norm": 0.29961150884628296, "learning_rate": 8e-05, "loss": 1.7694, "step": 1630 }, { "epoch": 0.35873749037721325, "grad_norm": 0.2760653793811798, "learning_rate": 8e-05, "loss": 1.719, "step": 1631 }, { "epoch": 0.3589574397888486, "grad_norm": 0.2739832103252411, "learning_rate": 8e-05, "loss": 1.7367, "step": 1632 }, { "epoch": 0.35917738920048387, "grad_norm": 0.2669771611690521, "learning_rate": 8e-05, "loss": 1.5306, "step": 1633 }, { "epoch": 0.3593973386121192, "grad_norm": 0.2744583189487457, "learning_rate": 8e-05, "loss": 1.5713, "step": 1634 }, { "epoch": 0.35961728802375453, "grad_norm": 0.2943086326122284, "learning_rate": 8e-05, "loss": 1.6569, "step": 1635 }, { "epoch": 0.35983723743538987, "grad_norm": 0.2873243987560272, "learning_rate": 8e-05, "loss": 1.6864, "step": 1636 }, { "epoch": 0.3600571868470252, "grad_norm": 0.27217867970466614, "learning_rate": 8e-05, "loss": 1.7519, "step": 1637 }, { "epoch": 0.36027713625866054, "grad_norm": 0.28656938672065735, "learning_rate": 8e-05, "loss": 1.7892, "step": 1638 }, { "epoch": 0.3604970856702958, "grad_norm": 0.2876884937286377, "learning_rate": 8e-05, "loss": 1.6709, "step": 1639 }, { "epoch": 0.36071703508193115, "grad_norm": 0.2873481512069702, "learning_rate": 8e-05, "loss": 1.8336, "step": 1640 }, { "epoch": 0.3609369844935665, "grad_norm": 0.28285419940948486, "learning_rate": 8e-05, "loss": 1.5887, "step": 1641 }, { "epoch": 0.3611569339052018, "grad_norm": 0.2624582052230835, "learning_rate": 8e-05, "loss": 1.6248, "step": 1642 }, { "epoch": 0.36137688331683715, "grad_norm": 0.2794424891471863, "learning_rate": 8e-05, "loss": 1.7191, "step": 1643 }, { "epoch": 0.36159683272847243, "grad_norm": 0.2890479862689972, "learning_rate": 8e-05, "loss": 1.905, "step": 1644 }, { "epoch": 0.36181678214010776, "grad_norm": 0.28444570302963257, "learning_rate": 8e-05, "loss": 1.6948, "step": 1645 }, { "epoch": 0.3620367315517431, "grad_norm": 0.27037203311920166, "learning_rate": 8e-05, "loss": 1.6245, "step": 1646 }, { "epoch": 0.36225668096337843, "grad_norm": 0.2864437699317932, "learning_rate": 8e-05, "loss": 1.688, "step": 1647 }, { "epoch": 0.36247663037501376, "grad_norm": 0.27912065386772156, "learning_rate": 8e-05, "loss": 1.6056, "step": 1648 }, { "epoch": 0.3626965797866491, "grad_norm": 0.26467230916023254, "learning_rate": 8e-05, "loss": 1.5786, "step": 1649 }, { "epoch": 0.3629165291982844, "grad_norm": 0.2793690264225006, "learning_rate": 8e-05, "loss": 1.6003, "step": 1650 }, { "epoch": 0.3631364786099197, "grad_norm": 0.288629949092865, "learning_rate": 8e-05, "loss": 1.6752, "step": 1651 }, { "epoch": 0.36335642802155504, "grad_norm": 0.283195823431015, "learning_rate": 8e-05, "loss": 1.6854, "step": 1652 }, { "epoch": 0.3635763774331904, "grad_norm": 0.2929665446281433, "learning_rate": 8e-05, "loss": 1.7191, "step": 1653 }, { "epoch": 0.3637963268448257, "grad_norm": 0.28676289319992065, "learning_rate": 8e-05, "loss": 1.6959, "step": 1654 }, { "epoch": 0.364016276256461, "grad_norm": 0.264635294675827, "learning_rate": 8e-05, "loss": 1.6232, "step": 1655 }, { "epoch": 0.3642362256680963, "grad_norm": 0.2763380706310272, "learning_rate": 8e-05, "loss": 1.7631, "step": 1656 }, { "epoch": 0.36445617507973166, "grad_norm": 0.2624233365058899, "learning_rate": 8e-05, "loss": 1.6635, "step": 1657 }, { "epoch": 0.364676124491367, "grad_norm": 0.2564058303833008, "learning_rate": 8e-05, "loss": 1.4745, "step": 1658 }, { "epoch": 0.3648960739030023, "grad_norm": 0.2966236174106598, "learning_rate": 8e-05, "loss": 1.6892, "step": 1659 }, { "epoch": 0.36511602331463766, "grad_norm": 0.30588555335998535, "learning_rate": 8e-05, "loss": 1.6884, "step": 1660 }, { "epoch": 0.36533597272627294, "grad_norm": 0.2692076861858368, "learning_rate": 8e-05, "loss": 1.7158, "step": 1661 }, { "epoch": 0.3655559221379083, "grad_norm": 0.29388558864593506, "learning_rate": 8e-05, "loss": 1.7133, "step": 1662 }, { "epoch": 0.3657758715495436, "grad_norm": 0.28685635328292847, "learning_rate": 8e-05, "loss": 1.7444, "step": 1663 }, { "epoch": 0.36599582096117894, "grad_norm": 0.2885795831680298, "learning_rate": 8e-05, "loss": 1.7537, "step": 1664 }, { "epoch": 0.3662157703728143, "grad_norm": 0.3066631853580475, "learning_rate": 8e-05, "loss": 1.7843, "step": 1665 }, { "epoch": 0.36643571978444955, "grad_norm": 0.31112298369407654, "learning_rate": 8e-05, "loss": 1.4934, "step": 1666 }, { "epoch": 0.3666556691960849, "grad_norm": 0.2751656472682953, "learning_rate": 8e-05, "loss": 1.7463, "step": 1667 }, { "epoch": 0.3668756186077202, "grad_norm": 0.2834889590740204, "learning_rate": 8e-05, "loss": 1.7757, "step": 1668 }, { "epoch": 0.36709556801935556, "grad_norm": 0.2778145968914032, "learning_rate": 8e-05, "loss": 1.7423, "step": 1669 }, { "epoch": 0.3673155174309909, "grad_norm": 0.32161521911621094, "learning_rate": 8e-05, "loss": 1.8312, "step": 1670 }, { "epoch": 0.36753546684262617, "grad_norm": 0.27995115518569946, "learning_rate": 8e-05, "loss": 1.7694, "step": 1671 }, { "epoch": 0.3677554162542615, "grad_norm": 0.27701541781425476, "learning_rate": 8e-05, "loss": 1.8054, "step": 1672 }, { "epoch": 0.36797536566589684, "grad_norm": 0.2757355570793152, "learning_rate": 8e-05, "loss": 1.7036, "step": 1673 }, { "epoch": 0.36819531507753217, "grad_norm": 0.27305907011032104, "learning_rate": 8e-05, "loss": 1.6627, "step": 1674 }, { "epoch": 0.3684152644891675, "grad_norm": 0.299679696559906, "learning_rate": 8e-05, "loss": 1.7552, "step": 1675 }, { "epoch": 0.36863521390080284, "grad_norm": 0.2728777825832367, "learning_rate": 8e-05, "loss": 1.7649, "step": 1676 }, { "epoch": 0.3688551633124381, "grad_norm": 0.26330089569091797, "learning_rate": 8e-05, "loss": 1.5887, "step": 1677 }, { "epoch": 0.36907511272407345, "grad_norm": 0.2850317060947418, "learning_rate": 8e-05, "loss": 1.6255, "step": 1678 }, { "epoch": 0.3692950621357088, "grad_norm": 0.2784862220287323, "learning_rate": 8e-05, "loss": 1.7123, "step": 1679 }, { "epoch": 0.3695150115473441, "grad_norm": 0.284298300743103, "learning_rate": 8e-05, "loss": 1.5809, "step": 1680 }, { "epoch": 0.36973496095897945, "grad_norm": 0.2725334167480469, "learning_rate": 8e-05, "loss": 1.7037, "step": 1681 }, { "epoch": 0.36995491037061473, "grad_norm": 0.2760758399963379, "learning_rate": 8e-05, "loss": 1.6827, "step": 1682 }, { "epoch": 0.37017485978225007, "grad_norm": 0.2661541700363159, "learning_rate": 8e-05, "loss": 1.7042, "step": 1683 }, { "epoch": 0.3703948091938854, "grad_norm": 0.27737516164779663, "learning_rate": 8e-05, "loss": 1.7689, "step": 1684 }, { "epoch": 0.37061475860552073, "grad_norm": 0.2607424259185791, "learning_rate": 8e-05, "loss": 1.6356, "step": 1685 }, { "epoch": 0.37083470801715607, "grad_norm": 0.2802969217300415, "learning_rate": 8e-05, "loss": 1.7004, "step": 1686 }, { "epoch": 0.3710546574287914, "grad_norm": 0.2660817801952362, "learning_rate": 8e-05, "loss": 1.5539, "step": 1687 }, { "epoch": 0.3712746068404267, "grad_norm": 0.27867192029953003, "learning_rate": 8e-05, "loss": 1.6531, "step": 1688 }, { "epoch": 0.371494556252062, "grad_norm": 0.27857083082199097, "learning_rate": 8e-05, "loss": 1.8023, "step": 1689 }, { "epoch": 0.37171450566369735, "grad_norm": 0.2689161002635956, "learning_rate": 8e-05, "loss": 1.7601, "step": 1690 }, { "epoch": 0.3719344550753327, "grad_norm": 0.297826886177063, "learning_rate": 8e-05, "loss": 1.7627, "step": 1691 }, { "epoch": 0.372154404486968, "grad_norm": 0.2592705190181732, "learning_rate": 8e-05, "loss": 1.7132, "step": 1692 }, { "epoch": 0.3723743538986033, "grad_norm": 0.28288522362709045, "learning_rate": 8e-05, "loss": 1.7604, "step": 1693 }, { "epoch": 0.37259430331023863, "grad_norm": 0.30823859572410583, "learning_rate": 8e-05, "loss": 1.8563, "step": 1694 }, { "epoch": 0.37281425272187396, "grad_norm": 0.27835527062416077, "learning_rate": 8e-05, "loss": 1.6816, "step": 1695 }, { "epoch": 0.3730342021335093, "grad_norm": 0.2626672089099884, "learning_rate": 8e-05, "loss": 1.6185, "step": 1696 }, { "epoch": 0.37325415154514463, "grad_norm": 0.2489227056503296, "learning_rate": 8e-05, "loss": 1.6119, "step": 1697 }, { "epoch": 0.37347410095677996, "grad_norm": 0.28637897968292236, "learning_rate": 8e-05, "loss": 1.6695, "step": 1698 }, { "epoch": 0.37369405036841524, "grad_norm": 0.27077022194862366, "learning_rate": 8e-05, "loss": 1.6095, "step": 1699 }, { "epoch": 0.3739139997800506, "grad_norm": 0.32049357891082764, "learning_rate": 8e-05, "loss": 1.875, "step": 1700 }, { "epoch": 0.3741339491916859, "grad_norm": 0.2890382707118988, "learning_rate": 8e-05, "loss": 1.7129, "step": 1701 }, { "epoch": 0.37435389860332124, "grad_norm": 0.2785224914550781, "learning_rate": 8e-05, "loss": 1.7162, "step": 1702 }, { "epoch": 0.3745738480149566, "grad_norm": 0.2685299217700958, "learning_rate": 8e-05, "loss": 1.7358, "step": 1703 }, { "epoch": 0.37479379742659186, "grad_norm": 0.2840120494365692, "learning_rate": 8e-05, "loss": 1.9123, "step": 1704 }, { "epoch": 0.3750137468382272, "grad_norm": 0.27426856756210327, "learning_rate": 8e-05, "loss": 1.7144, "step": 1705 }, { "epoch": 0.3752336962498625, "grad_norm": 0.2707318663597107, "learning_rate": 8e-05, "loss": 1.6961, "step": 1706 }, { "epoch": 0.37545364566149786, "grad_norm": 0.3059745728969574, "learning_rate": 8e-05, "loss": 1.7491, "step": 1707 }, { "epoch": 0.3756735950731332, "grad_norm": 0.27109962701797485, "learning_rate": 8e-05, "loss": 1.6515, "step": 1708 }, { "epoch": 0.3758935444847685, "grad_norm": 0.26874709129333496, "learning_rate": 8e-05, "loss": 1.7119, "step": 1709 }, { "epoch": 0.3761134938964038, "grad_norm": 0.27959340810775757, "learning_rate": 8e-05, "loss": 1.5449, "step": 1710 }, { "epoch": 0.37633344330803914, "grad_norm": 0.284386545419693, "learning_rate": 8e-05, "loss": 1.8336, "step": 1711 }, { "epoch": 0.3765533927196745, "grad_norm": 0.27861231565475464, "learning_rate": 8e-05, "loss": 1.7547, "step": 1712 }, { "epoch": 0.3767733421313098, "grad_norm": 0.26845625042915344, "learning_rate": 8e-05, "loss": 1.6838, "step": 1713 }, { "epoch": 0.37699329154294514, "grad_norm": 0.31240981817245483, "learning_rate": 8e-05, "loss": 1.7489, "step": 1714 }, { "epoch": 0.3772132409545804, "grad_norm": 0.2878013253211975, "learning_rate": 8e-05, "loss": 1.7533, "step": 1715 }, { "epoch": 0.37743319036621575, "grad_norm": 0.27676892280578613, "learning_rate": 8e-05, "loss": 1.6218, "step": 1716 }, { "epoch": 0.3776531397778511, "grad_norm": 0.2782065272331238, "learning_rate": 8e-05, "loss": 1.6311, "step": 1717 }, { "epoch": 0.3778730891894864, "grad_norm": 0.2829797863960266, "learning_rate": 8e-05, "loss": 1.5863, "step": 1718 }, { "epoch": 0.37809303860112176, "grad_norm": 0.2851261794567108, "learning_rate": 8e-05, "loss": 1.8365, "step": 1719 }, { "epoch": 0.3783129880127571, "grad_norm": 0.2844488322734833, "learning_rate": 8e-05, "loss": 1.7765, "step": 1720 }, { "epoch": 0.37853293742439237, "grad_norm": 0.2976120412349701, "learning_rate": 8e-05, "loss": 1.7334, "step": 1721 }, { "epoch": 0.3787528868360277, "grad_norm": 0.27947840094566345, "learning_rate": 8e-05, "loss": 1.6641, "step": 1722 }, { "epoch": 0.37897283624766304, "grad_norm": 0.2986278831958771, "learning_rate": 8e-05, "loss": 1.8201, "step": 1723 }, { "epoch": 0.37919278565929837, "grad_norm": 0.26200374960899353, "learning_rate": 8e-05, "loss": 1.5835, "step": 1724 }, { "epoch": 0.3794127350709337, "grad_norm": 0.2846388816833496, "learning_rate": 8e-05, "loss": 1.7863, "step": 1725 }, { "epoch": 0.379632684482569, "grad_norm": 0.2809320390224457, "learning_rate": 8e-05, "loss": 1.6667, "step": 1726 }, { "epoch": 0.3798526338942043, "grad_norm": 0.28523099422454834, "learning_rate": 8e-05, "loss": 1.6647, "step": 1727 }, { "epoch": 0.38007258330583965, "grad_norm": 0.2719436287879944, "learning_rate": 8e-05, "loss": 1.6, "step": 1728 }, { "epoch": 0.380292532717475, "grad_norm": 0.2762429118156433, "learning_rate": 8e-05, "loss": 1.6888, "step": 1729 }, { "epoch": 0.3805124821291103, "grad_norm": 0.30161863565444946, "learning_rate": 8e-05, "loss": 1.6659, "step": 1730 }, { "epoch": 0.38073243154074565, "grad_norm": 0.27962687611579895, "learning_rate": 8e-05, "loss": 1.629, "step": 1731 }, { "epoch": 0.38095238095238093, "grad_norm": 0.27580323815345764, "learning_rate": 8e-05, "loss": 1.689, "step": 1732 }, { "epoch": 0.38117233036401627, "grad_norm": 0.2676113545894623, "learning_rate": 8e-05, "loss": 1.7195, "step": 1733 }, { "epoch": 0.3813922797756516, "grad_norm": 0.27840152382850647, "learning_rate": 8e-05, "loss": 1.6433, "step": 1734 }, { "epoch": 0.38161222918728693, "grad_norm": 0.27100005745887756, "learning_rate": 8e-05, "loss": 1.6517, "step": 1735 }, { "epoch": 0.38183217859892227, "grad_norm": 0.2874828577041626, "learning_rate": 8e-05, "loss": 1.9139, "step": 1736 }, { "epoch": 0.38205212801055755, "grad_norm": 0.2685931324958801, "learning_rate": 8e-05, "loss": 1.7373, "step": 1737 }, { "epoch": 0.3822720774221929, "grad_norm": 0.2895548641681671, "learning_rate": 8e-05, "loss": 1.7828, "step": 1738 }, { "epoch": 0.3824920268338282, "grad_norm": 0.29109206795692444, "learning_rate": 8e-05, "loss": 1.6347, "step": 1739 }, { "epoch": 0.38271197624546355, "grad_norm": 0.2804923951625824, "learning_rate": 8e-05, "loss": 1.5978, "step": 1740 }, { "epoch": 0.3829319256570989, "grad_norm": 0.2829732894897461, "learning_rate": 8e-05, "loss": 1.6271, "step": 1741 }, { "epoch": 0.3831518750687342, "grad_norm": 0.28979840874671936, "learning_rate": 8e-05, "loss": 1.7244, "step": 1742 }, { "epoch": 0.3833718244803695, "grad_norm": 0.30159792304039, "learning_rate": 8e-05, "loss": 1.8074, "step": 1743 }, { "epoch": 0.38359177389200483, "grad_norm": 0.28228580951690674, "learning_rate": 8e-05, "loss": 1.6669, "step": 1744 }, { "epoch": 0.38381172330364016, "grad_norm": 0.27950945496559143, "learning_rate": 8e-05, "loss": 1.6583, "step": 1745 }, { "epoch": 0.3840316727152755, "grad_norm": 0.2708896994590759, "learning_rate": 8e-05, "loss": 1.5793, "step": 1746 }, { "epoch": 0.38425162212691083, "grad_norm": 0.27368029952049255, "learning_rate": 8e-05, "loss": 1.6371, "step": 1747 }, { "epoch": 0.3844715715385461, "grad_norm": 0.27621379494667053, "learning_rate": 8e-05, "loss": 1.5737, "step": 1748 }, { "epoch": 0.38469152095018144, "grad_norm": 0.27143922448158264, "learning_rate": 8e-05, "loss": 1.7289, "step": 1749 }, { "epoch": 0.3849114703618168, "grad_norm": 0.28887274861335754, "learning_rate": 8e-05, "loss": 1.7262, "step": 1750 }, { "epoch": 0.3851314197734521, "grad_norm": 0.26516541838645935, "learning_rate": 8e-05, "loss": 1.6358, "step": 1751 }, { "epoch": 0.38535136918508744, "grad_norm": 0.31475701928138733, "learning_rate": 8e-05, "loss": 1.8599, "step": 1752 }, { "epoch": 0.3855713185967228, "grad_norm": 0.27711552381515503, "learning_rate": 8e-05, "loss": 1.632, "step": 1753 }, { "epoch": 0.38579126800835806, "grad_norm": 0.27542901039123535, "learning_rate": 8e-05, "loss": 1.6741, "step": 1754 }, { "epoch": 0.3860112174199934, "grad_norm": 0.2941054701805115, "learning_rate": 8e-05, "loss": 1.6396, "step": 1755 }, { "epoch": 0.3862311668316287, "grad_norm": 0.27836698293685913, "learning_rate": 8e-05, "loss": 1.5689, "step": 1756 }, { "epoch": 0.38645111624326406, "grad_norm": 0.29147645831108093, "learning_rate": 8e-05, "loss": 1.7523, "step": 1757 }, { "epoch": 0.3866710656548994, "grad_norm": 0.30084285140037537, "learning_rate": 8e-05, "loss": 1.6932, "step": 1758 }, { "epoch": 0.38689101506653467, "grad_norm": 0.2850727140903473, "learning_rate": 8e-05, "loss": 1.7276, "step": 1759 }, { "epoch": 0.38711096447817, "grad_norm": 0.27011391520500183, "learning_rate": 8e-05, "loss": 1.6635, "step": 1760 }, { "epoch": 0.38733091388980534, "grad_norm": 0.28682348132133484, "learning_rate": 8e-05, "loss": 1.7642, "step": 1761 }, { "epoch": 0.3875508633014407, "grad_norm": 0.27676117420196533, "learning_rate": 8e-05, "loss": 1.7406, "step": 1762 }, { "epoch": 0.387770812713076, "grad_norm": 0.2654523551464081, "learning_rate": 8e-05, "loss": 1.6484, "step": 1763 }, { "epoch": 0.38799076212471134, "grad_norm": 0.28026026487350464, "learning_rate": 8e-05, "loss": 1.6714, "step": 1764 }, { "epoch": 0.3882107115363466, "grad_norm": 0.3003789782524109, "learning_rate": 8e-05, "loss": 1.8121, "step": 1765 }, { "epoch": 0.38843066094798195, "grad_norm": 0.35523107647895813, "learning_rate": 8e-05, "loss": 1.9299, "step": 1766 }, { "epoch": 0.3886506103596173, "grad_norm": 0.26844245195388794, "learning_rate": 8e-05, "loss": 1.6358, "step": 1767 }, { "epoch": 0.3888705597712526, "grad_norm": 0.27308356761932373, "learning_rate": 8e-05, "loss": 1.6104, "step": 1768 }, { "epoch": 0.38909050918288796, "grad_norm": 0.2775373160839081, "learning_rate": 8e-05, "loss": 1.5679, "step": 1769 }, { "epoch": 0.38931045859452323, "grad_norm": 0.29753705859184265, "learning_rate": 8e-05, "loss": 1.7678, "step": 1770 }, { "epoch": 0.38953040800615857, "grad_norm": 0.2798722982406616, "learning_rate": 8e-05, "loss": 1.7034, "step": 1771 }, { "epoch": 0.3897503574177939, "grad_norm": 0.2842818796634674, "learning_rate": 8e-05, "loss": 1.727, "step": 1772 }, { "epoch": 0.38997030682942924, "grad_norm": 0.27555832266807556, "learning_rate": 8e-05, "loss": 1.7438, "step": 1773 }, { "epoch": 0.39019025624106457, "grad_norm": 0.2824547588825226, "learning_rate": 8e-05, "loss": 1.5733, "step": 1774 }, { "epoch": 0.3904102056526999, "grad_norm": 0.2658035159111023, "learning_rate": 8e-05, "loss": 1.5997, "step": 1775 }, { "epoch": 0.3906301550643352, "grad_norm": 0.27601394057273865, "learning_rate": 8e-05, "loss": 1.7025, "step": 1776 }, { "epoch": 0.3908501044759705, "grad_norm": 0.2990022897720337, "learning_rate": 8e-05, "loss": 1.8278, "step": 1777 }, { "epoch": 0.39107005388760585, "grad_norm": 0.29378873109817505, "learning_rate": 8e-05, "loss": 1.7549, "step": 1778 }, { "epoch": 0.3912900032992412, "grad_norm": 0.29202136397361755, "learning_rate": 8e-05, "loss": 1.6602, "step": 1779 }, { "epoch": 0.3915099527108765, "grad_norm": 0.28191903233528137, "learning_rate": 8e-05, "loss": 1.6618, "step": 1780 }, { "epoch": 0.3917299021225118, "grad_norm": 0.26916682720184326, "learning_rate": 8e-05, "loss": 1.6523, "step": 1781 }, { "epoch": 0.39194985153414713, "grad_norm": 0.2886850833892822, "learning_rate": 8e-05, "loss": 1.5752, "step": 1782 }, { "epoch": 0.39216980094578247, "grad_norm": 0.2749246656894684, "learning_rate": 8e-05, "loss": 1.6731, "step": 1783 }, { "epoch": 0.3923897503574178, "grad_norm": 0.28945374488830566, "learning_rate": 8e-05, "loss": 1.7404, "step": 1784 }, { "epoch": 0.39260969976905313, "grad_norm": 0.27297836542129517, "learning_rate": 8e-05, "loss": 1.7176, "step": 1785 }, { "epoch": 0.39282964918068847, "grad_norm": 0.2738782465457916, "learning_rate": 8e-05, "loss": 1.6844, "step": 1786 }, { "epoch": 0.39304959859232375, "grad_norm": 0.2897050082683563, "learning_rate": 8e-05, "loss": 1.6522, "step": 1787 }, { "epoch": 0.3932695480039591, "grad_norm": 0.31031668186187744, "learning_rate": 8e-05, "loss": 1.8216, "step": 1788 }, { "epoch": 0.3934894974155944, "grad_norm": 0.2869516909122467, "learning_rate": 8e-05, "loss": 1.7598, "step": 1789 }, { "epoch": 0.39370944682722975, "grad_norm": 0.3080596625804901, "learning_rate": 8e-05, "loss": 1.7645, "step": 1790 }, { "epoch": 0.3939293962388651, "grad_norm": 0.27992716431617737, "learning_rate": 8e-05, "loss": 1.8373, "step": 1791 }, { "epoch": 0.39414934565050036, "grad_norm": 0.2761777341365814, "learning_rate": 8e-05, "loss": 1.6767, "step": 1792 }, { "epoch": 0.3943692950621357, "grad_norm": 0.30193084478378296, "learning_rate": 8e-05, "loss": 1.7394, "step": 1793 }, { "epoch": 0.39458924447377103, "grad_norm": 0.29375529289245605, "learning_rate": 8e-05, "loss": 1.761, "step": 1794 }, { "epoch": 0.39480919388540636, "grad_norm": 0.32190364599227905, "learning_rate": 8e-05, "loss": 1.7045, "step": 1795 }, { "epoch": 0.3950291432970417, "grad_norm": 0.27505311369895935, "learning_rate": 8e-05, "loss": 1.6678, "step": 1796 }, { "epoch": 0.39524909270867703, "grad_norm": 0.28678107261657715, "learning_rate": 8e-05, "loss": 1.7155, "step": 1797 }, { "epoch": 0.3954690421203123, "grad_norm": 0.28372088074684143, "learning_rate": 8e-05, "loss": 1.8833, "step": 1798 }, { "epoch": 0.39568899153194764, "grad_norm": 0.27803388237953186, "learning_rate": 8e-05, "loss": 1.7125, "step": 1799 }, { "epoch": 0.395908940943583, "grad_norm": 0.278728187084198, "learning_rate": 8e-05, "loss": 1.7019, "step": 1800 }, { "epoch": 0.3961288903552183, "grad_norm": 0.29563671350479126, "learning_rate": 8e-05, "loss": 1.6588, "step": 1801 }, { "epoch": 0.39634883976685364, "grad_norm": 0.35585105419158936, "learning_rate": 8e-05, "loss": 1.8864, "step": 1802 }, { "epoch": 0.3965687891784889, "grad_norm": 0.27399691939353943, "learning_rate": 8e-05, "loss": 1.6222, "step": 1803 }, { "epoch": 0.39678873859012426, "grad_norm": 0.2557234764099121, "learning_rate": 8e-05, "loss": 1.4835, "step": 1804 }, { "epoch": 0.3970086880017596, "grad_norm": 0.2929818332195282, "learning_rate": 8e-05, "loss": 1.7737, "step": 1805 }, { "epoch": 0.3972286374133949, "grad_norm": 0.279729425907135, "learning_rate": 8e-05, "loss": 1.6073, "step": 1806 }, { "epoch": 0.39744858682503026, "grad_norm": 0.2622847259044647, "learning_rate": 8e-05, "loss": 1.5368, "step": 1807 }, { "epoch": 0.3976685362366656, "grad_norm": 0.2704038619995117, "learning_rate": 8e-05, "loss": 1.5809, "step": 1808 }, { "epoch": 0.39788848564830087, "grad_norm": 0.2785516679286957, "learning_rate": 8e-05, "loss": 1.6416, "step": 1809 }, { "epoch": 0.3981084350599362, "grad_norm": 0.29611852765083313, "learning_rate": 8e-05, "loss": 1.7186, "step": 1810 }, { "epoch": 0.39832838447157154, "grad_norm": 0.28127896785736084, "learning_rate": 8e-05, "loss": 1.6248, "step": 1811 }, { "epoch": 0.3985483338832069, "grad_norm": 0.2746615707874298, "learning_rate": 8e-05, "loss": 1.7029, "step": 1812 }, { "epoch": 0.3987682832948422, "grad_norm": 0.2650880515575409, "learning_rate": 8e-05, "loss": 1.7828, "step": 1813 }, { "epoch": 0.3989882327064775, "grad_norm": 0.28278401494026184, "learning_rate": 8e-05, "loss": 1.6383, "step": 1814 }, { "epoch": 0.3992081821181128, "grad_norm": 0.2749755382537842, "learning_rate": 8e-05, "loss": 1.7551, "step": 1815 }, { "epoch": 0.39942813152974815, "grad_norm": 0.26788878440856934, "learning_rate": 8e-05, "loss": 1.5206, "step": 1816 }, { "epoch": 0.3996480809413835, "grad_norm": 0.28166842460632324, "learning_rate": 8e-05, "loss": 1.7134, "step": 1817 }, { "epoch": 0.3998680303530188, "grad_norm": 0.2781674563884735, "learning_rate": 8e-05, "loss": 1.7864, "step": 1818 }, { "epoch": 0.40008797976465416, "grad_norm": 0.2810186445713043, "learning_rate": 8e-05, "loss": 1.7017, "step": 1819 }, { "epoch": 0.40030792917628943, "grad_norm": 0.2872167229652405, "learning_rate": 8e-05, "loss": 1.563, "step": 1820 }, { "epoch": 0.40052787858792477, "grad_norm": 0.2864447832107544, "learning_rate": 8e-05, "loss": 1.8261, "step": 1821 }, { "epoch": 0.4007478279995601, "grad_norm": 0.2633639872074127, "learning_rate": 8e-05, "loss": 1.5864, "step": 1822 }, { "epoch": 0.40096777741119544, "grad_norm": 0.3556326925754547, "learning_rate": 8e-05, "loss": 1.5781, "step": 1823 }, { "epoch": 0.40118772682283077, "grad_norm": 0.2832813560962677, "learning_rate": 8e-05, "loss": 1.7145, "step": 1824 }, { "epoch": 0.40140767623446605, "grad_norm": 0.2862699627876282, "learning_rate": 8e-05, "loss": 1.7753, "step": 1825 }, { "epoch": 0.4016276256461014, "grad_norm": 0.3274460732936859, "learning_rate": 8e-05, "loss": 1.7979, "step": 1826 }, { "epoch": 0.4018475750577367, "grad_norm": 0.277118444442749, "learning_rate": 8e-05, "loss": 1.6089, "step": 1827 }, { "epoch": 0.40206752446937205, "grad_norm": 0.278337687253952, "learning_rate": 8e-05, "loss": 1.5788, "step": 1828 }, { "epoch": 0.4022874738810074, "grad_norm": 0.28072914481163025, "learning_rate": 8e-05, "loss": 1.6735, "step": 1829 }, { "epoch": 0.4025074232926427, "grad_norm": 0.2815505564212799, "learning_rate": 8e-05, "loss": 1.7593, "step": 1830 }, { "epoch": 0.402727372704278, "grad_norm": 0.2957006096839905, "learning_rate": 8e-05, "loss": 1.865, "step": 1831 }, { "epoch": 0.40294732211591333, "grad_norm": 0.3079582452774048, "learning_rate": 8e-05, "loss": 1.7421, "step": 1832 }, { "epoch": 0.40316727152754867, "grad_norm": 0.2924387454986572, "learning_rate": 8e-05, "loss": 1.7462, "step": 1833 }, { "epoch": 0.403387220939184, "grad_norm": 0.28879454731941223, "learning_rate": 8e-05, "loss": 1.7433, "step": 1834 }, { "epoch": 0.40360717035081933, "grad_norm": 0.27446237206459045, "learning_rate": 8e-05, "loss": 1.5869, "step": 1835 }, { "epoch": 0.4038271197624546, "grad_norm": 0.3164878487586975, "learning_rate": 8e-05, "loss": 1.7505, "step": 1836 }, { "epoch": 0.40404706917408995, "grad_norm": 0.25979530811309814, "learning_rate": 8e-05, "loss": 1.6001, "step": 1837 }, { "epoch": 0.4042670185857253, "grad_norm": 0.30625709891319275, "learning_rate": 8e-05, "loss": 1.7907, "step": 1838 }, { "epoch": 0.4044869679973606, "grad_norm": 0.27351540327072144, "learning_rate": 8e-05, "loss": 1.5835, "step": 1839 }, { "epoch": 0.40470691740899595, "grad_norm": 0.302372545003891, "learning_rate": 8e-05, "loss": 1.7821, "step": 1840 }, { "epoch": 0.4049268668206313, "grad_norm": 0.2910183370113373, "learning_rate": 8e-05, "loss": 1.7993, "step": 1841 }, { "epoch": 0.40514681623226656, "grad_norm": 0.2934883236885071, "learning_rate": 8e-05, "loss": 1.5928, "step": 1842 }, { "epoch": 0.4053667656439019, "grad_norm": 0.2586327791213989, "learning_rate": 8e-05, "loss": 1.5714, "step": 1843 }, { "epoch": 0.40558671505553723, "grad_norm": 0.27952027320861816, "learning_rate": 8e-05, "loss": 1.8168, "step": 1844 }, { "epoch": 0.40580666446717256, "grad_norm": 0.2987437844276428, "learning_rate": 8e-05, "loss": 1.555, "step": 1845 }, { "epoch": 0.4060266138788079, "grad_norm": 0.29165002703666687, "learning_rate": 8e-05, "loss": 1.6865, "step": 1846 }, { "epoch": 0.4062465632904432, "grad_norm": 0.2825503945350647, "learning_rate": 8e-05, "loss": 1.8814, "step": 1847 }, { "epoch": 0.4064665127020785, "grad_norm": 0.27995482087135315, "learning_rate": 8e-05, "loss": 1.6897, "step": 1848 }, { "epoch": 0.40668646211371384, "grad_norm": 0.2735064923763275, "learning_rate": 8e-05, "loss": 1.7279, "step": 1849 }, { "epoch": 0.4069064115253492, "grad_norm": 0.2850511074066162, "learning_rate": 8e-05, "loss": 1.6459, "step": 1850 }, { "epoch": 0.4071263609369845, "grad_norm": 0.3000599145889282, "learning_rate": 8e-05, "loss": 1.7301, "step": 1851 }, { "epoch": 0.40734631034861984, "grad_norm": 0.2768002152442932, "learning_rate": 8e-05, "loss": 1.5748, "step": 1852 }, { "epoch": 0.4075662597602551, "grad_norm": 0.26737141609191895, "learning_rate": 8e-05, "loss": 1.5895, "step": 1853 }, { "epoch": 0.40778620917189046, "grad_norm": 0.26408424973487854, "learning_rate": 8e-05, "loss": 1.5611, "step": 1854 }, { "epoch": 0.4080061585835258, "grad_norm": 0.2646276354789734, "learning_rate": 8e-05, "loss": 1.5865, "step": 1855 }, { "epoch": 0.4082261079951611, "grad_norm": 0.27871212363243103, "learning_rate": 8e-05, "loss": 1.8202, "step": 1856 }, { "epoch": 0.40844605740679646, "grad_norm": 0.3234533965587616, "learning_rate": 8e-05, "loss": 1.8213, "step": 1857 }, { "epoch": 0.40866600681843174, "grad_norm": 0.2705099284648895, "learning_rate": 8e-05, "loss": 1.6637, "step": 1858 }, { "epoch": 0.40888595623006707, "grad_norm": 0.28647711873054504, "learning_rate": 8e-05, "loss": 1.7396, "step": 1859 }, { "epoch": 0.4091059056417024, "grad_norm": 0.2812083959579468, "learning_rate": 8e-05, "loss": 1.5663, "step": 1860 }, { "epoch": 0.40932585505333774, "grad_norm": 0.2818193733692169, "learning_rate": 8e-05, "loss": 1.6073, "step": 1861 }, { "epoch": 0.4095458044649731, "grad_norm": 0.29906994104385376, "learning_rate": 8e-05, "loss": 1.7061, "step": 1862 }, { "epoch": 0.4097657538766084, "grad_norm": 0.27941465377807617, "learning_rate": 8e-05, "loss": 1.7282, "step": 1863 }, { "epoch": 0.4099857032882437, "grad_norm": 0.27629899978637695, "learning_rate": 8e-05, "loss": 1.6879, "step": 1864 }, { "epoch": 0.410205652699879, "grad_norm": 0.2792319059371948, "learning_rate": 8e-05, "loss": 1.7196, "step": 1865 }, { "epoch": 0.41042560211151435, "grad_norm": 0.2763090431690216, "learning_rate": 8e-05, "loss": 1.673, "step": 1866 }, { "epoch": 0.4106455515231497, "grad_norm": 0.2930999994277954, "learning_rate": 8e-05, "loss": 1.6919, "step": 1867 }, { "epoch": 0.410865500934785, "grad_norm": 0.2748461365699768, "learning_rate": 8e-05, "loss": 1.7553, "step": 1868 }, { "epoch": 0.4110854503464203, "grad_norm": 0.2742187976837158, "learning_rate": 8e-05, "loss": 1.6786, "step": 1869 }, { "epoch": 0.41130539975805563, "grad_norm": 0.3050731420516968, "learning_rate": 8e-05, "loss": 1.4902, "step": 1870 }, { "epoch": 0.41152534916969097, "grad_norm": 0.29456627368927, "learning_rate": 8e-05, "loss": 1.758, "step": 1871 }, { "epoch": 0.4117452985813263, "grad_norm": 0.2844219505786896, "learning_rate": 8e-05, "loss": 1.6206, "step": 1872 }, { "epoch": 0.41196524799296164, "grad_norm": 0.28889915347099304, "learning_rate": 8e-05, "loss": 1.6907, "step": 1873 }, { "epoch": 0.41218519740459697, "grad_norm": 0.27245181798934937, "learning_rate": 8e-05, "loss": 1.6749, "step": 1874 }, { "epoch": 0.41240514681623225, "grad_norm": 0.2927252948284149, "learning_rate": 8e-05, "loss": 1.6382, "step": 1875 }, { "epoch": 0.4126250962278676, "grad_norm": 0.27153030037879944, "learning_rate": 8e-05, "loss": 1.6011, "step": 1876 }, { "epoch": 0.4128450456395029, "grad_norm": 0.2807110846042633, "learning_rate": 8e-05, "loss": 1.7126, "step": 1877 }, { "epoch": 0.41306499505113825, "grad_norm": 0.27375784516334534, "learning_rate": 8e-05, "loss": 1.7443, "step": 1878 }, { "epoch": 0.4132849444627736, "grad_norm": 0.27330929040908813, "learning_rate": 8e-05, "loss": 1.6305, "step": 1879 }, { "epoch": 0.41350489387440886, "grad_norm": 0.27126336097717285, "learning_rate": 8e-05, "loss": 1.6688, "step": 1880 }, { "epoch": 0.4137248432860442, "grad_norm": 0.2768147885799408, "learning_rate": 8e-05, "loss": 1.7274, "step": 1881 }, { "epoch": 0.41394479269767953, "grad_norm": 0.2686031460762024, "learning_rate": 8e-05, "loss": 1.6445, "step": 1882 }, { "epoch": 0.41416474210931487, "grad_norm": 0.27737778425216675, "learning_rate": 8e-05, "loss": 1.5226, "step": 1883 }, { "epoch": 0.4143846915209502, "grad_norm": 0.2761901319026947, "learning_rate": 8e-05, "loss": 1.6884, "step": 1884 }, { "epoch": 0.41460464093258553, "grad_norm": 0.28609856963157654, "learning_rate": 8e-05, "loss": 1.7719, "step": 1885 }, { "epoch": 0.4148245903442208, "grad_norm": 0.2904943525791168, "learning_rate": 8e-05, "loss": 1.6979, "step": 1886 }, { "epoch": 0.41504453975585615, "grad_norm": 0.3016435503959656, "learning_rate": 8e-05, "loss": 1.7912, "step": 1887 }, { "epoch": 0.4152644891674915, "grad_norm": 0.27562782168388367, "learning_rate": 8e-05, "loss": 1.5822, "step": 1888 }, { "epoch": 0.4154844385791268, "grad_norm": 0.2841348648071289, "learning_rate": 8e-05, "loss": 1.7524, "step": 1889 }, { "epoch": 0.41570438799076215, "grad_norm": 0.26393935084342957, "learning_rate": 8e-05, "loss": 1.6219, "step": 1890 }, { "epoch": 0.4159243374023974, "grad_norm": 0.2792678773403168, "learning_rate": 8e-05, "loss": 1.7243, "step": 1891 }, { "epoch": 0.41614428681403276, "grad_norm": 0.291425496339798, "learning_rate": 8e-05, "loss": 1.7499, "step": 1892 }, { "epoch": 0.4163642362256681, "grad_norm": 0.2737634778022766, "learning_rate": 8e-05, "loss": 1.6565, "step": 1893 }, { "epoch": 0.41658418563730343, "grad_norm": 0.26807767152786255, "learning_rate": 8e-05, "loss": 1.6149, "step": 1894 }, { "epoch": 0.41680413504893876, "grad_norm": 0.28826507925987244, "learning_rate": 8e-05, "loss": 1.6857, "step": 1895 }, { "epoch": 0.41702408446057404, "grad_norm": 0.27604466676712036, "learning_rate": 8e-05, "loss": 1.7689, "step": 1896 }, { "epoch": 0.4172440338722094, "grad_norm": 0.27355703711509705, "learning_rate": 8e-05, "loss": 1.7254, "step": 1897 }, { "epoch": 0.4174639832838447, "grad_norm": 0.26692044734954834, "learning_rate": 8e-05, "loss": 1.6372, "step": 1898 }, { "epoch": 0.41768393269548004, "grad_norm": 0.27527916431427, "learning_rate": 8e-05, "loss": 1.6913, "step": 1899 }, { "epoch": 0.4179038821071154, "grad_norm": 0.26881837844848633, "learning_rate": 8e-05, "loss": 1.6663, "step": 1900 }, { "epoch": 0.4181238315187507, "grad_norm": 0.27977946400642395, "learning_rate": 8e-05, "loss": 1.8117, "step": 1901 }, { "epoch": 0.418343780930386, "grad_norm": 0.2958911955356598, "learning_rate": 8e-05, "loss": 1.6603, "step": 1902 }, { "epoch": 0.4185637303420213, "grad_norm": 0.2845151424407959, "learning_rate": 8e-05, "loss": 1.7517, "step": 1903 }, { "epoch": 0.41878367975365666, "grad_norm": 0.2804581820964813, "learning_rate": 8e-05, "loss": 1.765, "step": 1904 }, { "epoch": 0.419003629165292, "grad_norm": 0.29568520188331604, "learning_rate": 8e-05, "loss": 1.6995, "step": 1905 }, { "epoch": 0.4192235785769273, "grad_norm": 0.303100049495697, "learning_rate": 8e-05, "loss": 1.77, "step": 1906 }, { "epoch": 0.4194435279885626, "grad_norm": 0.26847636699676514, "learning_rate": 8e-05, "loss": 1.6964, "step": 1907 }, { "epoch": 0.41966347740019794, "grad_norm": 0.2791590094566345, "learning_rate": 8e-05, "loss": 1.5912, "step": 1908 }, { "epoch": 0.41988342681183327, "grad_norm": 0.2687268555164337, "learning_rate": 8e-05, "loss": 1.6163, "step": 1909 }, { "epoch": 0.4201033762234686, "grad_norm": 0.29087433218955994, "learning_rate": 8e-05, "loss": 1.6588, "step": 1910 }, { "epoch": 0.42032332563510394, "grad_norm": 0.29639971256256104, "learning_rate": 8e-05, "loss": 1.6911, "step": 1911 }, { "epoch": 0.4205432750467393, "grad_norm": 0.27669841051101685, "learning_rate": 8e-05, "loss": 1.7078, "step": 1912 }, { "epoch": 0.42076322445837455, "grad_norm": 0.2851327955722809, "learning_rate": 8e-05, "loss": 1.7217, "step": 1913 }, { "epoch": 0.4209831738700099, "grad_norm": 0.27069011330604553, "learning_rate": 8e-05, "loss": 1.7076, "step": 1914 }, { "epoch": 0.4212031232816452, "grad_norm": 0.26195240020751953, "learning_rate": 8e-05, "loss": 1.6647, "step": 1915 }, { "epoch": 0.42142307269328055, "grad_norm": 0.3046209216117859, "learning_rate": 8e-05, "loss": 1.5303, "step": 1916 }, { "epoch": 0.4216430221049159, "grad_norm": 0.29437899589538574, "learning_rate": 8e-05, "loss": 1.6589, "step": 1917 }, { "epoch": 0.42186297151655117, "grad_norm": 0.2954728603363037, "learning_rate": 8e-05, "loss": 1.777, "step": 1918 }, { "epoch": 0.4220829209281865, "grad_norm": 0.2612738609313965, "learning_rate": 8e-05, "loss": 1.5668, "step": 1919 }, { "epoch": 0.42230287033982183, "grad_norm": 0.3015122413635254, "learning_rate": 8e-05, "loss": 1.7861, "step": 1920 }, { "epoch": 0.42252281975145717, "grad_norm": 0.3785838484764099, "learning_rate": 8e-05, "loss": 1.8979, "step": 1921 }, { "epoch": 0.4227427691630925, "grad_norm": 0.2849038541316986, "learning_rate": 8e-05, "loss": 1.6891, "step": 1922 }, { "epoch": 0.42296271857472784, "grad_norm": 0.278728723526001, "learning_rate": 8e-05, "loss": 1.7891, "step": 1923 }, { "epoch": 0.4231826679863631, "grad_norm": 0.27032172679901123, "learning_rate": 8e-05, "loss": 1.6963, "step": 1924 }, { "epoch": 0.42340261739799845, "grad_norm": 0.2731832265853882, "learning_rate": 8e-05, "loss": 1.632, "step": 1925 }, { "epoch": 0.4236225668096338, "grad_norm": 0.30378425121307373, "learning_rate": 8e-05, "loss": 1.7823, "step": 1926 }, { "epoch": 0.4238425162212691, "grad_norm": 0.27693971991539, "learning_rate": 8e-05, "loss": 1.483, "step": 1927 }, { "epoch": 0.42406246563290445, "grad_norm": 0.2719477415084839, "learning_rate": 8e-05, "loss": 1.6708, "step": 1928 }, { "epoch": 0.42428241504453973, "grad_norm": 0.26625335216522217, "learning_rate": 8e-05, "loss": 1.4946, "step": 1929 }, { "epoch": 0.42450236445617506, "grad_norm": 0.2843473553657532, "learning_rate": 8e-05, "loss": 1.722, "step": 1930 }, { "epoch": 0.4247223138678104, "grad_norm": 0.3453083336353302, "learning_rate": 8e-05, "loss": 1.7238, "step": 1931 }, { "epoch": 0.42494226327944573, "grad_norm": 0.25626078248023987, "learning_rate": 8e-05, "loss": 1.4706, "step": 1932 }, { "epoch": 0.42516221269108107, "grad_norm": 0.2908123731613159, "learning_rate": 8e-05, "loss": 1.7105, "step": 1933 }, { "epoch": 0.4253821621027164, "grad_norm": 0.33517104387283325, "learning_rate": 8e-05, "loss": 1.8023, "step": 1934 }, { "epoch": 0.4256021115143517, "grad_norm": 0.28047069907188416, "learning_rate": 8e-05, "loss": 1.6266, "step": 1935 }, { "epoch": 0.425822060925987, "grad_norm": 0.2778942584991455, "learning_rate": 8e-05, "loss": 1.6866, "step": 1936 }, { "epoch": 0.42604201033762235, "grad_norm": 0.3038877248764038, "learning_rate": 8e-05, "loss": 1.6075, "step": 1937 }, { "epoch": 0.4262619597492577, "grad_norm": 0.2814297378063202, "learning_rate": 8e-05, "loss": 1.5939, "step": 1938 }, { "epoch": 0.426481909160893, "grad_norm": 0.27854403853416443, "learning_rate": 8e-05, "loss": 1.5943, "step": 1939 }, { "epoch": 0.4267018585725283, "grad_norm": 0.2924019694328308, "learning_rate": 8e-05, "loss": 1.8193, "step": 1940 }, { "epoch": 0.4269218079841636, "grad_norm": 0.2862766683101654, "learning_rate": 8e-05, "loss": 1.6065, "step": 1941 }, { "epoch": 0.42714175739579896, "grad_norm": 0.2696346342563629, "learning_rate": 8e-05, "loss": 1.5343, "step": 1942 }, { "epoch": 0.4273617068074343, "grad_norm": 0.2578338384628296, "learning_rate": 8e-05, "loss": 1.5786, "step": 1943 }, { "epoch": 0.42758165621906963, "grad_norm": 0.28594937920570374, "learning_rate": 8e-05, "loss": 1.725, "step": 1944 }, { "epoch": 0.42780160563070496, "grad_norm": 0.2808282971382141, "learning_rate": 8e-05, "loss": 1.7951, "step": 1945 }, { "epoch": 0.42802155504234024, "grad_norm": 0.32533401250839233, "learning_rate": 8e-05, "loss": 1.9645, "step": 1946 }, { "epoch": 0.4282415044539756, "grad_norm": 0.2737642228603363, "learning_rate": 8e-05, "loss": 1.6243, "step": 1947 }, { "epoch": 0.4284614538656109, "grad_norm": 0.2885657250881195, "learning_rate": 8e-05, "loss": 1.7338, "step": 1948 }, { "epoch": 0.42868140327724624, "grad_norm": 0.2788100242614746, "learning_rate": 8e-05, "loss": 1.76, "step": 1949 }, { "epoch": 0.4289013526888816, "grad_norm": 0.2899073362350464, "learning_rate": 8e-05, "loss": 1.7739, "step": 1950 }, { "epoch": 0.42912130210051685, "grad_norm": 0.2874782681465149, "learning_rate": 8e-05, "loss": 1.8283, "step": 1951 }, { "epoch": 0.4293412515121522, "grad_norm": 0.2757413685321808, "learning_rate": 8e-05, "loss": 1.641, "step": 1952 }, { "epoch": 0.4295612009237875, "grad_norm": 0.2811121940612793, "learning_rate": 8e-05, "loss": 1.7231, "step": 1953 }, { "epoch": 0.42978115033542286, "grad_norm": 0.3400493860244751, "learning_rate": 8e-05, "loss": 1.8431, "step": 1954 }, { "epoch": 0.4300010997470582, "grad_norm": 0.29006627202033997, "learning_rate": 8e-05, "loss": 1.7438, "step": 1955 }, { "epoch": 0.4302210491586935, "grad_norm": 0.30233392119407654, "learning_rate": 8e-05, "loss": 1.6603, "step": 1956 }, { "epoch": 0.4304409985703288, "grad_norm": 0.2921263873577118, "learning_rate": 8e-05, "loss": 1.6604, "step": 1957 }, { "epoch": 0.43066094798196414, "grad_norm": 0.27695250511169434, "learning_rate": 8e-05, "loss": 1.7536, "step": 1958 }, { "epoch": 0.43088089739359947, "grad_norm": 0.2827337980270386, "learning_rate": 8e-05, "loss": 1.6324, "step": 1959 }, { "epoch": 0.4311008468052348, "grad_norm": 0.27993375062942505, "learning_rate": 8e-05, "loss": 1.7168, "step": 1960 }, { "epoch": 0.43132079621687014, "grad_norm": 0.2801220417022705, "learning_rate": 8e-05, "loss": 1.705, "step": 1961 }, { "epoch": 0.4315407456285054, "grad_norm": 0.27520567178726196, "learning_rate": 8e-05, "loss": 1.664, "step": 1962 }, { "epoch": 0.43176069504014075, "grad_norm": 0.26910632848739624, "learning_rate": 8e-05, "loss": 1.3616, "step": 1963 }, { "epoch": 0.4319806444517761, "grad_norm": 0.27770352363586426, "learning_rate": 8e-05, "loss": 1.6689, "step": 1964 }, { "epoch": 0.4322005938634114, "grad_norm": 0.27606719732284546, "learning_rate": 8e-05, "loss": 1.6644, "step": 1965 }, { "epoch": 0.43242054327504675, "grad_norm": 0.27787330746650696, "learning_rate": 8e-05, "loss": 1.8854, "step": 1966 }, { "epoch": 0.4326404926866821, "grad_norm": 0.26479870080947876, "learning_rate": 8e-05, "loss": 1.5904, "step": 1967 }, { "epoch": 0.43286044209831737, "grad_norm": 0.27598053216934204, "learning_rate": 8e-05, "loss": 1.5666, "step": 1968 }, { "epoch": 0.4330803915099527, "grad_norm": 0.27461937069892883, "learning_rate": 8e-05, "loss": 1.5487, "step": 1969 }, { "epoch": 0.43330034092158803, "grad_norm": 0.2928270399570465, "learning_rate": 8e-05, "loss": 1.8173, "step": 1970 }, { "epoch": 0.43352029033322337, "grad_norm": 0.30754199624061584, "learning_rate": 8e-05, "loss": 1.6762, "step": 1971 }, { "epoch": 0.4337402397448587, "grad_norm": 0.2676936089992523, "learning_rate": 8e-05, "loss": 1.7314, "step": 1972 }, { "epoch": 0.433960189156494, "grad_norm": 0.2919710576534271, "learning_rate": 8e-05, "loss": 1.8586, "step": 1973 }, { "epoch": 0.4341801385681293, "grad_norm": 0.28165963292121887, "learning_rate": 8e-05, "loss": 1.7943, "step": 1974 }, { "epoch": 0.43440008797976465, "grad_norm": 0.2700537443161011, "learning_rate": 8e-05, "loss": 1.613, "step": 1975 }, { "epoch": 0.4346200373914, "grad_norm": 0.26830658316612244, "learning_rate": 8e-05, "loss": 1.5854, "step": 1976 }, { "epoch": 0.4348399868030353, "grad_norm": 0.28799256682395935, "learning_rate": 8e-05, "loss": 1.8246, "step": 1977 }, { "epoch": 0.43505993621467065, "grad_norm": 0.27226150035858154, "learning_rate": 8e-05, "loss": 1.6252, "step": 1978 }, { "epoch": 0.43527988562630593, "grad_norm": 0.2646162807941437, "learning_rate": 8e-05, "loss": 1.5699, "step": 1979 }, { "epoch": 0.43549983503794126, "grad_norm": 0.27331140637397766, "learning_rate": 8e-05, "loss": 1.6893, "step": 1980 }, { "epoch": 0.4357197844495766, "grad_norm": 0.26996269822120667, "learning_rate": 8e-05, "loss": 1.6004, "step": 1981 }, { "epoch": 0.43593973386121193, "grad_norm": 0.29484307765960693, "learning_rate": 8e-05, "loss": 1.551, "step": 1982 }, { "epoch": 0.43615968327284727, "grad_norm": 0.28224268555641174, "learning_rate": 8e-05, "loss": 1.6898, "step": 1983 }, { "epoch": 0.43637963268448254, "grad_norm": 0.26172178983688354, "learning_rate": 8e-05, "loss": 1.4375, "step": 1984 }, { "epoch": 0.4365995820961179, "grad_norm": 0.2603735029697418, "learning_rate": 8e-05, "loss": 1.4528, "step": 1985 }, { "epoch": 0.4368195315077532, "grad_norm": 0.30643707513809204, "learning_rate": 8e-05, "loss": 1.688, "step": 1986 }, { "epoch": 0.43703948091938855, "grad_norm": 0.2951216995716095, "learning_rate": 8e-05, "loss": 1.7769, "step": 1987 }, { "epoch": 0.4372594303310239, "grad_norm": 0.2939329445362091, "learning_rate": 8e-05, "loss": 1.8161, "step": 1988 }, { "epoch": 0.4374793797426592, "grad_norm": 0.27539846301078796, "learning_rate": 8e-05, "loss": 1.6019, "step": 1989 }, { "epoch": 0.4376993291542945, "grad_norm": 0.2770693898200989, "learning_rate": 8e-05, "loss": 1.5972, "step": 1990 }, { "epoch": 0.4379192785659298, "grad_norm": 0.2832552492618561, "learning_rate": 8e-05, "loss": 1.7467, "step": 1991 }, { "epoch": 0.43813922797756516, "grad_norm": 0.2983148992061615, "learning_rate": 8e-05, "loss": 1.7181, "step": 1992 }, { "epoch": 0.4383591773892005, "grad_norm": 0.2829340994358063, "learning_rate": 8e-05, "loss": 1.5984, "step": 1993 }, { "epoch": 0.43857912680083583, "grad_norm": 0.2857687473297119, "learning_rate": 8e-05, "loss": 1.6471, "step": 1994 }, { "epoch": 0.4387990762124711, "grad_norm": 0.2669824957847595, "learning_rate": 8e-05, "loss": 1.6215, "step": 1995 }, { "epoch": 0.43901902562410644, "grad_norm": 0.28832894563674927, "learning_rate": 8e-05, "loss": 1.6884, "step": 1996 }, { "epoch": 0.4392389750357418, "grad_norm": 0.2919970154762268, "learning_rate": 8e-05, "loss": 1.7462, "step": 1997 }, { "epoch": 0.4394589244473771, "grad_norm": 0.2998509109020233, "learning_rate": 8e-05, "loss": 1.7219, "step": 1998 }, { "epoch": 0.43967887385901244, "grad_norm": 0.2780647575855255, "learning_rate": 8e-05, "loss": 1.8219, "step": 1999 }, { "epoch": 0.4398988232706478, "grad_norm": 0.2833268940448761, "learning_rate": 8e-05, "loss": 1.6873, "step": 2000 }, { "epoch": 0.44011877268228305, "grad_norm": 0.2802470624446869, "learning_rate": 8e-05, "loss": 1.7249, "step": 2001 }, { "epoch": 0.4403387220939184, "grad_norm": 0.2767699658870697, "learning_rate": 8e-05, "loss": 1.6789, "step": 2002 }, { "epoch": 0.4405586715055537, "grad_norm": 0.28534451127052307, "learning_rate": 8e-05, "loss": 1.6269, "step": 2003 }, { "epoch": 0.44077862091718906, "grad_norm": 0.28716540336608887, "learning_rate": 8e-05, "loss": 1.8097, "step": 2004 }, { "epoch": 0.4409985703288244, "grad_norm": 0.27516409754753113, "learning_rate": 8e-05, "loss": 1.837, "step": 2005 }, { "epoch": 0.44121851974045967, "grad_norm": 0.2910866439342499, "learning_rate": 8e-05, "loss": 1.61, "step": 2006 }, { "epoch": 0.441438469152095, "grad_norm": 0.2851128578186035, "learning_rate": 8e-05, "loss": 1.6939, "step": 2007 }, { "epoch": 0.44165841856373034, "grad_norm": 0.2709331214427948, "learning_rate": 8e-05, "loss": 1.6352, "step": 2008 }, { "epoch": 0.44187836797536567, "grad_norm": 0.28077712655067444, "learning_rate": 8e-05, "loss": 1.6119, "step": 2009 }, { "epoch": 0.442098317387001, "grad_norm": 0.2804681956768036, "learning_rate": 8e-05, "loss": 1.706, "step": 2010 }, { "epoch": 0.44231826679863634, "grad_norm": 0.28305575251579285, "learning_rate": 8e-05, "loss": 1.7501, "step": 2011 }, { "epoch": 0.4425382162102716, "grad_norm": 0.30372944474220276, "learning_rate": 8e-05, "loss": 1.69, "step": 2012 }, { "epoch": 0.44275816562190695, "grad_norm": 0.2695739269256592, "learning_rate": 8e-05, "loss": 1.4976, "step": 2013 }, { "epoch": 0.4429781150335423, "grad_norm": 0.27175822854042053, "learning_rate": 8e-05, "loss": 1.6941, "step": 2014 }, { "epoch": 0.4431980644451776, "grad_norm": 0.2786177396774292, "learning_rate": 8e-05, "loss": 1.719, "step": 2015 }, { "epoch": 0.44341801385681295, "grad_norm": 0.26625001430511475, "learning_rate": 8e-05, "loss": 1.6197, "step": 2016 }, { "epoch": 0.44363796326844823, "grad_norm": 0.29516807198524475, "learning_rate": 8e-05, "loss": 1.7305, "step": 2017 }, { "epoch": 0.44385791268008357, "grad_norm": 0.39562076330184937, "learning_rate": 8e-05, "loss": 1.783, "step": 2018 }, { "epoch": 0.4440778620917189, "grad_norm": 0.27659425139427185, "learning_rate": 8e-05, "loss": 1.6948, "step": 2019 }, { "epoch": 0.44429781150335423, "grad_norm": 0.2787366211414337, "learning_rate": 8e-05, "loss": 1.6237, "step": 2020 }, { "epoch": 0.44451776091498957, "grad_norm": 0.27939459681510925, "learning_rate": 8e-05, "loss": 1.7628, "step": 2021 }, { "epoch": 0.4447377103266249, "grad_norm": 0.27395081520080566, "learning_rate": 8e-05, "loss": 1.5619, "step": 2022 }, { "epoch": 0.4449576597382602, "grad_norm": 0.28255096077919006, "learning_rate": 8e-05, "loss": 1.7556, "step": 2023 }, { "epoch": 0.4451776091498955, "grad_norm": 0.2922489643096924, "learning_rate": 8e-05, "loss": 1.624, "step": 2024 }, { "epoch": 0.44539755856153085, "grad_norm": 0.4039583206176758, "learning_rate": 8e-05, "loss": 1.507, "step": 2025 }, { "epoch": 0.4456175079731662, "grad_norm": 0.28025928139686584, "learning_rate": 8e-05, "loss": 1.8057, "step": 2026 }, { "epoch": 0.4458374573848015, "grad_norm": 0.2777588963508606, "learning_rate": 8e-05, "loss": 1.6104, "step": 2027 }, { "epoch": 0.4460574067964368, "grad_norm": 0.2689501941204071, "learning_rate": 8e-05, "loss": 1.5339, "step": 2028 }, { "epoch": 0.44627735620807213, "grad_norm": 0.28021785616874695, "learning_rate": 8e-05, "loss": 1.7819, "step": 2029 }, { "epoch": 0.44649730561970746, "grad_norm": 0.26980918645858765, "learning_rate": 8e-05, "loss": 1.5995, "step": 2030 }, { "epoch": 0.4467172550313428, "grad_norm": 0.293047696352005, "learning_rate": 8e-05, "loss": 1.8571, "step": 2031 }, { "epoch": 0.44693720444297813, "grad_norm": 0.2841939330101013, "learning_rate": 8e-05, "loss": 1.5794, "step": 2032 }, { "epoch": 0.44715715385461346, "grad_norm": 0.2845712900161743, "learning_rate": 8e-05, "loss": 1.6635, "step": 2033 }, { "epoch": 0.44737710326624874, "grad_norm": 0.27919885516166687, "learning_rate": 8e-05, "loss": 1.7094, "step": 2034 }, { "epoch": 0.4475970526778841, "grad_norm": 0.30076712369918823, "learning_rate": 8e-05, "loss": 1.6989, "step": 2035 }, { "epoch": 0.4478170020895194, "grad_norm": 0.2666080892086029, "learning_rate": 8e-05, "loss": 1.689, "step": 2036 }, { "epoch": 0.44803695150115475, "grad_norm": 0.29133087396621704, "learning_rate": 8e-05, "loss": 1.8559, "step": 2037 }, { "epoch": 0.4482569009127901, "grad_norm": 0.28610843420028687, "learning_rate": 8e-05, "loss": 1.6556, "step": 2038 }, { "epoch": 0.44847685032442536, "grad_norm": 0.26724278926849365, "learning_rate": 8e-05, "loss": 1.6104, "step": 2039 }, { "epoch": 0.4486967997360607, "grad_norm": 0.2724173665046692, "learning_rate": 8e-05, "loss": 1.645, "step": 2040 }, { "epoch": 0.448916749147696, "grad_norm": 0.26570823788642883, "learning_rate": 8e-05, "loss": 1.4611, "step": 2041 }, { "epoch": 0.44913669855933136, "grad_norm": 0.29498788714408875, "learning_rate": 8e-05, "loss": 1.8787, "step": 2042 }, { "epoch": 0.4493566479709667, "grad_norm": 0.28211459517478943, "learning_rate": 8e-05, "loss": 1.6062, "step": 2043 }, { "epoch": 0.44957659738260203, "grad_norm": 0.3148192763328552, "learning_rate": 8e-05, "loss": 1.8734, "step": 2044 }, { "epoch": 0.4497965467942373, "grad_norm": 0.27721115946769714, "learning_rate": 8e-05, "loss": 1.7274, "step": 2045 }, { "epoch": 0.45001649620587264, "grad_norm": 0.29178541898727417, "learning_rate": 8e-05, "loss": 1.7568, "step": 2046 }, { "epoch": 0.450236445617508, "grad_norm": 0.27845948934555054, "learning_rate": 8e-05, "loss": 1.6615, "step": 2047 }, { "epoch": 0.4504563950291433, "grad_norm": 0.2741856873035431, "learning_rate": 8e-05, "loss": 1.5667, "step": 2048 }, { "epoch": 0.45067634444077864, "grad_norm": 0.28572753071784973, "learning_rate": 8e-05, "loss": 1.7115, "step": 2049 }, { "epoch": 0.4508962938524139, "grad_norm": 0.2769505977630615, "learning_rate": 8e-05, "loss": 1.6297, "step": 2050 }, { "epoch": 0.45111624326404925, "grad_norm": 0.27633893489837646, "learning_rate": 8e-05, "loss": 1.6455, "step": 2051 }, { "epoch": 0.4513361926756846, "grad_norm": 0.28455862402915955, "learning_rate": 8e-05, "loss": 1.7496, "step": 2052 }, { "epoch": 0.4515561420873199, "grad_norm": 0.2920532524585724, "learning_rate": 8e-05, "loss": 1.7671, "step": 2053 }, { "epoch": 0.45177609149895526, "grad_norm": 0.27528077363967896, "learning_rate": 8e-05, "loss": 1.6535, "step": 2054 }, { "epoch": 0.4519960409105906, "grad_norm": 0.2949519157409668, "learning_rate": 8e-05, "loss": 1.6796, "step": 2055 }, { "epoch": 0.45221599032222587, "grad_norm": 0.2740989923477173, "learning_rate": 8e-05, "loss": 1.4341, "step": 2056 }, { "epoch": 0.4524359397338612, "grad_norm": 0.31732696294784546, "learning_rate": 8e-05, "loss": 1.717, "step": 2057 }, { "epoch": 0.45265588914549654, "grad_norm": 0.2747776210308075, "learning_rate": 8e-05, "loss": 1.7162, "step": 2058 }, { "epoch": 0.45287583855713187, "grad_norm": 0.3037000596523285, "learning_rate": 8e-05, "loss": 1.6399, "step": 2059 }, { "epoch": 0.4530957879687672, "grad_norm": 0.29499107599258423, "learning_rate": 8e-05, "loss": 1.6306, "step": 2060 }, { "epoch": 0.4533157373804025, "grad_norm": 0.2890235483646393, "learning_rate": 8e-05, "loss": 1.8048, "step": 2061 }, { "epoch": 0.4535356867920378, "grad_norm": 0.28515708446502686, "learning_rate": 8e-05, "loss": 1.843, "step": 2062 }, { "epoch": 0.45375563620367315, "grad_norm": 0.28525930643081665, "learning_rate": 8e-05, "loss": 1.6568, "step": 2063 }, { "epoch": 0.4539755856153085, "grad_norm": 0.27117711305618286, "learning_rate": 8e-05, "loss": 1.6605, "step": 2064 }, { "epoch": 0.4541955350269438, "grad_norm": 0.2771861255168915, "learning_rate": 8e-05, "loss": 1.7553, "step": 2065 }, { "epoch": 0.45441548443857915, "grad_norm": 0.29263827204704285, "learning_rate": 8e-05, "loss": 1.7919, "step": 2066 }, { "epoch": 0.45463543385021443, "grad_norm": 0.2785603702068329, "learning_rate": 8e-05, "loss": 1.6686, "step": 2067 }, { "epoch": 0.45485538326184977, "grad_norm": 0.2752209007740021, "learning_rate": 8e-05, "loss": 1.5984, "step": 2068 }, { "epoch": 0.4550753326734851, "grad_norm": 0.279784232378006, "learning_rate": 8e-05, "loss": 1.6903, "step": 2069 }, { "epoch": 0.45529528208512043, "grad_norm": 0.2957722246646881, "learning_rate": 8e-05, "loss": 1.8176, "step": 2070 }, { "epoch": 0.45551523149675577, "grad_norm": 0.2798726260662079, "learning_rate": 8e-05, "loss": 1.6637, "step": 2071 }, { "epoch": 0.45573518090839105, "grad_norm": 0.26509538292884827, "learning_rate": 8e-05, "loss": 1.5034, "step": 2072 }, { "epoch": 0.4559551303200264, "grad_norm": 0.2984442114830017, "learning_rate": 8e-05, "loss": 1.8225, "step": 2073 }, { "epoch": 0.4561750797316617, "grad_norm": 0.28242239356040955, "learning_rate": 8e-05, "loss": 1.6707, "step": 2074 }, { "epoch": 0.45639502914329705, "grad_norm": 0.2722650468349457, "learning_rate": 8e-05, "loss": 1.5723, "step": 2075 }, { "epoch": 0.4566149785549324, "grad_norm": 0.25942909717559814, "learning_rate": 8e-05, "loss": 1.5417, "step": 2076 }, { "epoch": 0.4568349279665677, "grad_norm": 0.2782632112503052, "learning_rate": 8e-05, "loss": 1.5857, "step": 2077 }, { "epoch": 0.457054877378203, "grad_norm": 0.28298354148864746, "learning_rate": 8e-05, "loss": 1.6124, "step": 2078 }, { "epoch": 0.45727482678983833, "grad_norm": 0.2920227348804474, "learning_rate": 8e-05, "loss": 1.6247, "step": 2079 }, { "epoch": 0.45749477620147366, "grad_norm": 0.30804532766342163, "learning_rate": 8e-05, "loss": 1.8689, "step": 2080 }, { "epoch": 0.457714725613109, "grad_norm": 0.2759280204772949, "learning_rate": 8e-05, "loss": 1.6366, "step": 2081 }, { "epoch": 0.45793467502474433, "grad_norm": 0.27967414259910583, "learning_rate": 8e-05, "loss": 1.6391, "step": 2082 }, { "epoch": 0.4581546244363796, "grad_norm": 0.30624908208847046, "learning_rate": 8e-05, "loss": 1.8188, "step": 2083 }, { "epoch": 0.45837457384801494, "grad_norm": 0.2747632563114166, "learning_rate": 8e-05, "loss": 1.6394, "step": 2084 }, { "epoch": 0.4585945232596503, "grad_norm": 0.29921606183052063, "learning_rate": 8e-05, "loss": 1.6264, "step": 2085 }, { "epoch": 0.4588144726712856, "grad_norm": 0.27374643087387085, "learning_rate": 8e-05, "loss": 1.5955, "step": 2086 }, { "epoch": 0.45903442208292095, "grad_norm": 0.2804218530654907, "learning_rate": 8e-05, "loss": 1.6936, "step": 2087 }, { "epoch": 0.4592543714945563, "grad_norm": 0.288095623254776, "learning_rate": 8e-05, "loss": 1.658, "step": 2088 }, { "epoch": 0.45947432090619156, "grad_norm": 0.2622469961643219, "learning_rate": 8e-05, "loss": 1.5468, "step": 2089 }, { "epoch": 0.4596942703178269, "grad_norm": 0.298968106508255, "learning_rate": 8e-05, "loss": 1.8415, "step": 2090 }, { "epoch": 0.4599142197294622, "grad_norm": 0.27127620577812195, "learning_rate": 8e-05, "loss": 1.5154, "step": 2091 }, { "epoch": 0.46013416914109756, "grad_norm": 0.3025810122489929, "learning_rate": 8e-05, "loss": 1.7251, "step": 2092 }, { "epoch": 0.4603541185527329, "grad_norm": 0.2805241346359253, "learning_rate": 8e-05, "loss": 1.786, "step": 2093 }, { "epoch": 0.4605740679643682, "grad_norm": 0.28292620182037354, "learning_rate": 8e-05, "loss": 1.6205, "step": 2094 }, { "epoch": 0.4607940173760035, "grad_norm": 0.27890294790267944, "learning_rate": 8e-05, "loss": 1.6182, "step": 2095 }, { "epoch": 0.46101396678763884, "grad_norm": 0.2704887390136719, "learning_rate": 8e-05, "loss": 1.643, "step": 2096 }, { "epoch": 0.4612339161992742, "grad_norm": 0.27034714818000793, "learning_rate": 8e-05, "loss": 1.6918, "step": 2097 }, { "epoch": 0.4614538656109095, "grad_norm": 0.2763729691505432, "learning_rate": 8e-05, "loss": 1.6218, "step": 2098 }, { "epoch": 0.46167381502254484, "grad_norm": 0.28457143902778625, "learning_rate": 8e-05, "loss": 1.5602, "step": 2099 }, { "epoch": 0.4618937644341801, "grad_norm": 0.3102862536907196, "learning_rate": 8e-05, "loss": 1.6946, "step": 2100 }, { "epoch": 0.46211371384581545, "grad_norm": 0.2817099690437317, "learning_rate": 8e-05, "loss": 1.69, "step": 2101 }, { "epoch": 0.4623336632574508, "grad_norm": 0.2816404104232788, "learning_rate": 8e-05, "loss": 1.6651, "step": 2102 }, { "epoch": 0.4625536126690861, "grad_norm": 0.2756252884864807, "learning_rate": 8e-05, "loss": 1.6935, "step": 2103 }, { "epoch": 0.46277356208072146, "grad_norm": 0.28443071246147156, "learning_rate": 8e-05, "loss": 1.8278, "step": 2104 }, { "epoch": 0.46299351149235674, "grad_norm": 0.2955114543437958, "learning_rate": 8e-05, "loss": 1.7654, "step": 2105 }, { "epoch": 0.46321346090399207, "grad_norm": 0.30527764558792114, "learning_rate": 8e-05, "loss": 1.921, "step": 2106 }, { "epoch": 0.4634334103156274, "grad_norm": 0.28985050320625305, "learning_rate": 8e-05, "loss": 1.7637, "step": 2107 }, { "epoch": 0.46365335972726274, "grad_norm": 0.2904118299484253, "learning_rate": 8e-05, "loss": 1.7645, "step": 2108 }, { "epoch": 0.46387330913889807, "grad_norm": 0.3137964606285095, "learning_rate": 8e-05, "loss": 1.8301, "step": 2109 }, { "epoch": 0.46409325855053335, "grad_norm": 0.2634297013282776, "learning_rate": 8e-05, "loss": 1.5916, "step": 2110 }, { "epoch": 0.4643132079621687, "grad_norm": 0.26435586810112, "learning_rate": 8e-05, "loss": 1.5805, "step": 2111 }, { "epoch": 0.464533157373804, "grad_norm": 0.2845149040222168, "learning_rate": 8e-05, "loss": 1.6894, "step": 2112 }, { "epoch": 0.46475310678543935, "grad_norm": 0.3034592568874359, "learning_rate": 8e-05, "loss": 1.7122, "step": 2113 }, { "epoch": 0.4649730561970747, "grad_norm": 0.2862027585506439, "learning_rate": 8e-05, "loss": 1.5886, "step": 2114 }, { "epoch": 0.46519300560871, "grad_norm": 0.2709789574146271, "learning_rate": 8e-05, "loss": 1.6112, "step": 2115 }, { "epoch": 0.4654129550203453, "grad_norm": 0.3048953115940094, "learning_rate": 8e-05, "loss": 1.718, "step": 2116 }, { "epoch": 0.46563290443198063, "grad_norm": 0.295149564743042, "learning_rate": 8e-05, "loss": 1.7339, "step": 2117 }, { "epoch": 0.46585285384361597, "grad_norm": 0.27533626556396484, "learning_rate": 8e-05, "loss": 1.6143, "step": 2118 }, { "epoch": 0.4660728032552513, "grad_norm": 0.30383235216140747, "learning_rate": 8e-05, "loss": 1.7154, "step": 2119 }, { "epoch": 0.46629275266688663, "grad_norm": 0.2834450304508209, "learning_rate": 8e-05, "loss": 1.6892, "step": 2120 }, { "epoch": 0.4665127020785219, "grad_norm": 0.27407264709472656, "learning_rate": 8e-05, "loss": 1.7138, "step": 2121 }, { "epoch": 0.46673265149015725, "grad_norm": 0.28417688608169556, "learning_rate": 8e-05, "loss": 1.7416, "step": 2122 }, { "epoch": 0.4669526009017926, "grad_norm": 0.2915797233581543, "learning_rate": 8e-05, "loss": 1.7104, "step": 2123 }, { "epoch": 0.4671725503134279, "grad_norm": 0.29155269265174866, "learning_rate": 8e-05, "loss": 1.7374, "step": 2124 }, { "epoch": 0.46739249972506325, "grad_norm": 0.27683204412460327, "learning_rate": 8e-05, "loss": 1.5669, "step": 2125 }, { "epoch": 0.4676124491366986, "grad_norm": 0.2835148870944977, "learning_rate": 8e-05, "loss": 1.7426, "step": 2126 }, { "epoch": 0.46783239854833386, "grad_norm": 0.27906641364097595, "learning_rate": 8e-05, "loss": 1.6297, "step": 2127 }, { "epoch": 0.4680523479599692, "grad_norm": 0.28407955169677734, "learning_rate": 8e-05, "loss": 1.7351, "step": 2128 }, { "epoch": 0.46827229737160453, "grad_norm": 0.2793600261211395, "learning_rate": 8e-05, "loss": 1.5786, "step": 2129 }, { "epoch": 0.46849224678323986, "grad_norm": 0.2806802988052368, "learning_rate": 8e-05, "loss": 1.7466, "step": 2130 }, { "epoch": 0.4687121961948752, "grad_norm": 0.30251967906951904, "learning_rate": 8e-05, "loss": 1.6703, "step": 2131 }, { "epoch": 0.4689321456065105, "grad_norm": 0.275473415851593, "learning_rate": 8e-05, "loss": 1.5881, "step": 2132 }, { "epoch": 0.4691520950181458, "grad_norm": 0.28925517201423645, "learning_rate": 8e-05, "loss": 1.7369, "step": 2133 }, { "epoch": 0.46937204442978114, "grad_norm": 0.26768866181373596, "learning_rate": 8e-05, "loss": 1.5469, "step": 2134 }, { "epoch": 0.4695919938414165, "grad_norm": 0.272969514131546, "learning_rate": 8e-05, "loss": 1.6235, "step": 2135 }, { "epoch": 0.4698119432530518, "grad_norm": 0.39006346464157104, "learning_rate": 8e-05, "loss": 1.6943, "step": 2136 }, { "epoch": 0.47003189266468715, "grad_norm": 0.2898694574832916, "learning_rate": 8e-05, "loss": 1.7196, "step": 2137 }, { "epoch": 0.4702518420763224, "grad_norm": 0.28824204206466675, "learning_rate": 8e-05, "loss": 1.6807, "step": 2138 }, { "epoch": 0.47047179148795776, "grad_norm": 0.3024749755859375, "learning_rate": 8e-05, "loss": 1.7435, "step": 2139 }, { "epoch": 0.4706917408995931, "grad_norm": 0.2894933521747589, "learning_rate": 8e-05, "loss": 1.8047, "step": 2140 }, { "epoch": 0.4709116903112284, "grad_norm": 0.2900967299938202, "learning_rate": 8e-05, "loss": 1.6536, "step": 2141 }, { "epoch": 0.47113163972286376, "grad_norm": 0.2727701961994171, "learning_rate": 8e-05, "loss": 1.6187, "step": 2142 }, { "epoch": 0.47135158913449904, "grad_norm": 0.2630798816680908, "learning_rate": 8e-05, "loss": 1.5558, "step": 2143 }, { "epoch": 0.4715715385461344, "grad_norm": 0.2755641043186188, "learning_rate": 8e-05, "loss": 1.6247, "step": 2144 }, { "epoch": 0.4717914879577697, "grad_norm": 0.26855289936065674, "learning_rate": 8e-05, "loss": 1.6908, "step": 2145 }, { "epoch": 0.47201143736940504, "grad_norm": 0.26333558559417725, "learning_rate": 8e-05, "loss": 1.6467, "step": 2146 }, { "epoch": 0.4722313867810404, "grad_norm": 0.2696126103401184, "learning_rate": 8e-05, "loss": 1.646, "step": 2147 }, { "epoch": 0.4724513361926757, "grad_norm": 0.2838461101055145, "learning_rate": 8e-05, "loss": 1.7269, "step": 2148 }, { "epoch": 0.472671285604311, "grad_norm": 0.27359622716903687, "learning_rate": 8e-05, "loss": 1.6578, "step": 2149 }, { "epoch": 0.4728912350159463, "grad_norm": 0.28489992022514343, "learning_rate": 8e-05, "loss": 1.7451, "step": 2150 }, { "epoch": 0.47311118442758165, "grad_norm": 0.30069005489349365, "learning_rate": 8e-05, "loss": 1.7789, "step": 2151 }, { "epoch": 0.473331133839217, "grad_norm": 0.2787550091743469, "learning_rate": 8e-05, "loss": 1.8532, "step": 2152 }, { "epoch": 0.4735510832508523, "grad_norm": 0.28521937131881714, "learning_rate": 8e-05, "loss": 1.6202, "step": 2153 }, { "epoch": 0.4737710326624876, "grad_norm": 0.27512073516845703, "learning_rate": 8e-05, "loss": 1.6982, "step": 2154 }, { "epoch": 0.47399098207412294, "grad_norm": 0.28500398993492126, "learning_rate": 8e-05, "loss": 1.7262, "step": 2155 }, { "epoch": 0.47421093148575827, "grad_norm": 0.2889910340309143, "learning_rate": 8e-05, "loss": 1.7001, "step": 2156 }, { "epoch": 0.4744308808973936, "grad_norm": 0.2868637144565582, "learning_rate": 8e-05, "loss": 1.6569, "step": 2157 }, { "epoch": 0.47465083030902894, "grad_norm": 0.27974042296409607, "learning_rate": 8e-05, "loss": 1.7308, "step": 2158 }, { "epoch": 0.47487077972066427, "grad_norm": 0.2812412977218628, "learning_rate": 8e-05, "loss": 1.6586, "step": 2159 }, { "epoch": 0.47509072913229955, "grad_norm": 0.27973487973213196, "learning_rate": 8e-05, "loss": 1.7688, "step": 2160 }, { "epoch": 0.4753106785439349, "grad_norm": 0.2852223515510559, "learning_rate": 8e-05, "loss": 1.7773, "step": 2161 }, { "epoch": 0.4755306279555702, "grad_norm": 0.2702232301235199, "learning_rate": 8e-05, "loss": 1.5842, "step": 2162 }, { "epoch": 0.47575057736720555, "grad_norm": 0.26885986328125, "learning_rate": 8e-05, "loss": 1.6179, "step": 2163 }, { "epoch": 0.4759705267788409, "grad_norm": 0.26561740040779114, "learning_rate": 8e-05, "loss": 1.6105, "step": 2164 }, { "epoch": 0.47619047619047616, "grad_norm": 0.3136744499206543, "learning_rate": 8e-05, "loss": 1.7263, "step": 2165 }, { "epoch": 0.4764104256021115, "grad_norm": 0.29980844259262085, "learning_rate": 8e-05, "loss": 1.6428, "step": 2166 }, { "epoch": 0.47663037501374683, "grad_norm": 0.2782588303089142, "learning_rate": 8e-05, "loss": 1.5905, "step": 2167 }, { "epoch": 0.47685032442538217, "grad_norm": 0.3134911358356476, "learning_rate": 8e-05, "loss": 1.7437, "step": 2168 }, { "epoch": 0.4770702738370175, "grad_norm": 0.28305792808532715, "learning_rate": 8e-05, "loss": 1.6187, "step": 2169 }, { "epoch": 0.47729022324865283, "grad_norm": 0.2741806209087372, "learning_rate": 8e-05, "loss": 1.7017, "step": 2170 }, { "epoch": 0.4775101726602881, "grad_norm": 0.2861132323741913, "learning_rate": 8e-05, "loss": 1.7167, "step": 2171 }, { "epoch": 0.47773012207192345, "grad_norm": 0.2796178162097931, "learning_rate": 8e-05, "loss": 1.7747, "step": 2172 }, { "epoch": 0.4779500714835588, "grad_norm": 0.3019583821296692, "learning_rate": 8e-05, "loss": 1.9244, "step": 2173 }, { "epoch": 0.4781700208951941, "grad_norm": 0.2874825894832611, "learning_rate": 8e-05, "loss": 1.676, "step": 2174 }, { "epoch": 0.47838997030682945, "grad_norm": 0.2864963412284851, "learning_rate": 8e-05, "loss": 1.8289, "step": 2175 }, { "epoch": 0.4786099197184647, "grad_norm": 0.3347536623477936, "learning_rate": 8e-05, "loss": 1.8204, "step": 2176 }, { "epoch": 0.47882986913010006, "grad_norm": 0.2859993577003479, "learning_rate": 8e-05, "loss": 1.5472, "step": 2177 }, { "epoch": 0.4790498185417354, "grad_norm": 0.2972160875797272, "learning_rate": 8e-05, "loss": 1.6577, "step": 2178 }, { "epoch": 0.47926976795337073, "grad_norm": 0.26402032375335693, "learning_rate": 8e-05, "loss": 1.4936, "step": 2179 }, { "epoch": 0.47948971736500606, "grad_norm": 0.28069886565208435, "learning_rate": 8e-05, "loss": 1.8158, "step": 2180 }, { "epoch": 0.4797096667766414, "grad_norm": 0.2630525529384613, "learning_rate": 8e-05, "loss": 1.5814, "step": 2181 }, { "epoch": 0.4799296161882767, "grad_norm": 0.2999349534511566, "learning_rate": 8e-05, "loss": 1.7879, "step": 2182 }, { "epoch": 0.480149565599912, "grad_norm": 0.28722846508026123, "learning_rate": 8e-05, "loss": 1.7143, "step": 2183 }, { "epoch": 0.48036951501154734, "grad_norm": 0.2746049165725708, "learning_rate": 8e-05, "loss": 1.5197, "step": 2184 }, { "epoch": 0.4805894644231827, "grad_norm": 0.2719694972038269, "learning_rate": 8e-05, "loss": 1.6642, "step": 2185 }, { "epoch": 0.480809413834818, "grad_norm": 0.286636084318161, "learning_rate": 8e-05, "loss": 1.6885, "step": 2186 }, { "epoch": 0.4810293632464533, "grad_norm": 0.3076469600200653, "learning_rate": 8e-05, "loss": 1.7681, "step": 2187 }, { "epoch": 0.4812493126580886, "grad_norm": 0.3075680434703827, "learning_rate": 8e-05, "loss": 1.8007, "step": 2188 }, { "epoch": 0.48146926206972396, "grad_norm": 0.2793465852737427, "learning_rate": 8e-05, "loss": 1.6552, "step": 2189 }, { "epoch": 0.4816892114813593, "grad_norm": 0.2886781692504883, "learning_rate": 8e-05, "loss": 1.7827, "step": 2190 }, { "epoch": 0.4819091608929946, "grad_norm": 0.27432283759117126, "learning_rate": 8e-05, "loss": 1.7724, "step": 2191 }, { "epoch": 0.48212911030462996, "grad_norm": 0.26780393719673157, "learning_rate": 8e-05, "loss": 1.4613, "step": 2192 }, { "epoch": 0.48234905971626524, "grad_norm": 0.27178165316581726, "learning_rate": 8e-05, "loss": 1.6439, "step": 2193 }, { "epoch": 0.48256900912790057, "grad_norm": 0.27651992440223694, "learning_rate": 8e-05, "loss": 1.6672, "step": 2194 }, { "epoch": 0.4827889585395359, "grad_norm": 0.26919931173324585, "learning_rate": 8e-05, "loss": 1.594, "step": 2195 }, { "epoch": 0.48300890795117124, "grad_norm": 0.267678439617157, "learning_rate": 8e-05, "loss": 1.5852, "step": 2196 }, { "epoch": 0.4832288573628066, "grad_norm": 0.2895921766757965, "learning_rate": 8e-05, "loss": 1.6456, "step": 2197 }, { "epoch": 0.48344880677444185, "grad_norm": 0.27052855491638184, "learning_rate": 8e-05, "loss": 1.5393, "step": 2198 }, { "epoch": 0.4836687561860772, "grad_norm": 0.2793048024177551, "learning_rate": 8e-05, "loss": 1.7006, "step": 2199 }, { "epoch": 0.4838887055977125, "grad_norm": 0.2838841676712036, "learning_rate": 8e-05, "loss": 1.6547, "step": 2200 }, { "epoch": 0.48410865500934785, "grad_norm": 0.29814717173576355, "learning_rate": 8e-05, "loss": 1.7392, "step": 2201 }, { "epoch": 0.4843286044209832, "grad_norm": 0.28823426365852356, "learning_rate": 8e-05, "loss": 1.6434, "step": 2202 }, { "epoch": 0.4845485538326185, "grad_norm": 0.2645476460456848, "learning_rate": 8e-05, "loss": 1.5663, "step": 2203 }, { "epoch": 0.4847685032442538, "grad_norm": 0.3046651780605316, "learning_rate": 8e-05, "loss": 1.5941, "step": 2204 }, { "epoch": 0.48498845265588914, "grad_norm": 0.28958046436309814, "learning_rate": 8e-05, "loss": 1.7348, "step": 2205 }, { "epoch": 0.48520840206752447, "grad_norm": 0.25703537464141846, "learning_rate": 8e-05, "loss": 1.5712, "step": 2206 }, { "epoch": 0.4854283514791598, "grad_norm": 0.2969980537891388, "learning_rate": 8e-05, "loss": 1.7966, "step": 2207 }, { "epoch": 0.48564830089079514, "grad_norm": 0.27638381719589233, "learning_rate": 8e-05, "loss": 1.6478, "step": 2208 }, { "epoch": 0.4858682503024304, "grad_norm": 0.283682644367218, "learning_rate": 8e-05, "loss": 1.6524, "step": 2209 }, { "epoch": 0.48608819971406575, "grad_norm": 0.2837259769439697, "learning_rate": 8e-05, "loss": 1.7302, "step": 2210 }, { "epoch": 0.4863081491257011, "grad_norm": 0.3042410612106323, "learning_rate": 8e-05, "loss": 1.5969, "step": 2211 }, { "epoch": 0.4865280985373364, "grad_norm": 0.2819627821445465, "learning_rate": 8e-05, "loss": 1.5621, "step": 2212 }, { "epoch": 0.48674804794897175, "grad_norm": 0.28049173951148987, "learning_rate": 8e-05, "loss": 1.6831, "step": 2213 }, { "epoch": 0.4869679973606071, "grad_norm": 0.29762500524520874, "learning_rate": 8e-05, "loss": 1.7883, "step": 2214 }, { "epoch": 0.48718794677224236, "grad_norm": 0.3022189140319824, "learning_rate": 8e-05, "loss": 1.5744, "step": 2215 }, { "epoch": 0.4874078961838777, "grad_norm": 0.28249025344848633, "learning_rate": 8e-05, "loss": 1.6514, "step": 2216 }, { "epoch": 0.48762784559551303, "grad_norm": 0.3398612141609192, "learning_rate": 8e-05, "loss": 1.4215, "step": 2217 }, { "epoch": 0.48784779500714837, "grad_norm": 0.28481197357177734, "learning_rate": 8e-05, "loss": 1.7065, "step": 2218 }, { "epoch": 0.4880677444187837, "grad_norm": 0.3076950013637543, "learning_rate": 8e-05, "loss": 1.7665, "step": 2219 }, { "epoch": 0.488287693830419, "grad_norm": 0.28533896803855896, "learning_rate": 8e-05, "loss": 1.6156, "step": 2220 }, { "epoch": 0.4885076432420543, "grad_norm": 0.2940129041671753, "learning_rate": 8e-05, "loss": 1.6957, "step": 2221 }, { "epoch": 0.48872759265368965, "grad_norm": 0.30342915654182434, "learning_rate": 8e-05, "loss": 1.6014, "step": 2222 }, { "epoch": 0.488947542065325, "grad_norm": 0.2851787507534027, "learning_rate": 8e-05, "loss": 1.8816, "step": 2223 }, { "epoch": 0.4891674914769603, "grad_norm": 0.3813328146934509, "learning_rate": 8e-05, "loss": 1.6608, "step": 2224 }, { "epoch": 0.48938744088859565, "grad_norm": 0.329373300075531, "learning_rate": 8e-05, "loss": 1.7455, "step": 2225 }, { "epoch": 0.4896073903002309, "grad_norm": 0.2751460373401642, "learning_rate": 8e-05, "loss": 1.6854, "step": 2226 }, { "epoch": 0.48982733971186626, "grad_norm": 0.283033549785614, "learning_rate": 8e-05, "loss": 1.7018, "step": 2227 }, { "epoch": 0.4900472891235016, "grad_norm": 0.2869894504547119, "learning_rate": 8e-05, "loss": 1.7062, "step": 2228 }, { "epoch": 0.49026723853513693, "grad_norm": 0.2895123064517975, "learning_rate": 8e-05, "loss": 1.625, "step": 2229 }, { "epoch": 0.49048718794677226, "grad_norm": 0.28288763761520386, "learning_rate": 8e-05, "loss": 1.6901, "step": 2230 }, { "epoch": 0.49070713735840754, "grad_norm": 0.28054291009902954, "learning_rate": 8e-05, "loss": 1.7782, "step": 2231 }, { "epoch": 0.4909270867700429, "grad_norm": 0.2727196216583252, "learning_rate": 8e-05, "loss": 1.531, "step": 2232 }, { "epoch": 0.4911470361816782, "grad_norm": 0.28192493319511414, "learning_rate": 8e-05, "loss": 1.7144, "step": 2233 }, { "epoch": 0.49136698559331354, "grad_norm": 0.28133469820022583, "learning_rate": 8e-05, "loss": 1.6448, "step": 2234 }, { "epoch": 0.4915869350049489, "grad_norm": 0.3175356686115265, "learning_rate": 8e-05, "loss": 1.648, "step": 2235 }, { "epoch": 0.4918068844165842, "grad_norm": 0.2990395128726959, "learning_rate": 8e-05, "loss": 1.6871, "step": 2236 }, { "epoch": 0.4920268338282195, "grad_norm": 0.3121372163295746, "learning_rate": 8e-05, "loss": 1.7671, "step": 2237 }, { "epoch": 0.4922467832398548, "grad_norm": 0.2772499918937683, "learning_rate": 8e-05, "loss": 1.674, "step": 2238 }, { "epoch": 0.49246673265149016, "grad_norm": 0.2990477383136749, "learning_rate": 8e-05, "loss": 1.8202, "step": 2239 }, { "epoch": 0.4926866820631255, "grad_norm": 0.28028279542922974, "learning_rate": 8e-05, "loss": 1.7255, "step": 2240 }, { "epoch": 0.4929066314747608, "grad_norm": 0.28883370757102966, "learning_rate": 8e-05, "loss": 1.6102, "step": 2241 }, { "epoch": 0.4931265808863961, "grad_norm": 0.2775261700153351, "learning_rate": 8e-05, "loss": 1.8572, "step": 2242 }, { "epoch": 0.49334653029803144, "grad_norm": 0.2821192443370819, "learning_rate": 8e-05, "loss": 1.8068, "step": 2243 }, { "epoch": 0.49356647970966677, "grad_norm": 0.29555544257164, "learning_rate": 8e-05, "loss": 1.6875, "step": 2244 }, { "epoch": 0.4937864291213021, "grad_norm": 0.28827783465385437, "learning_rate": 8e-05, "loss": 1.7123, "step": 2245 }, { "epoch": 0.49400637853293744, "grad_norm": 0.27672290802001953, "learning_rate": 8e-05, "loss": 1.4739, "step": 2246 }, { "epoch": 0.4942263279445728, "grad_norm": 0.27185946702957153, "learning_rate": 8e-05, "loss": 1.7748, "step": 2247 }, { "epoch": 0.49444627735620805, "grad_norm": 0.2972213327884674, "learning_rate": 8e-05, "loss": 1.7122, "step": 2248 }, { "epoch": 0.4946662267678434, "grad_norm": 0.30817538499832153, "learning_rate": 8e-05, "loss": 1.926, "step": 2249 }, { "epoch": 0.4948861761794787, "grad_norm": 0.2821509838104248, "learning_rate": 8e-05, "loss": 1.7608, "step": 2250 }, { "epoch": 0.49510612559111405, "grad_norm": 0.29807963967323303, "learning_rate": 8e-05, "loss": 1.8233, "step": 2251 }, { "epoch": 0.4953260750027494, "grad_norm": 0.29549431800842285, "learning_rate": 8e-05, "loss": 1.7963, "step": 2252 }, { "epoch": 0.49554602441438467, "grad_norm": 0.3025868535041809, "learning_rate": 8e-05, "loss": 1.7597, "step": 2253 }, { "epoch": 0.49576597382602, "grad_norm": 0.30950862169265747, "learning_rate": 8e-05, "loss": 1.8901, "step": 2254 }, { "epoch": 0.49598592323765534, "grad_norm": 0.3357299864292145, "learning_rate": 8e-05, "loss": 1.7692, "step": 2255 }, { "epoch": 0.49620587264929067, "grad_norm": 0.2873973548412323, "learning_rate": 8e-05, "loss": 1.6934, "step": 2256 }, { "epoch": 0.496425822060926, "grad_norm": 0.2997465431690216, "learning_rate": 8e-05, "loss": 1.5939, "step": 2257 }, { "epoch": 0.49664577147256134, "grad_norm": 0.269217312335968, "learning_rate": 8e-05, "loss": 1.5687, "step": 2258 }, { "epoch": 0.4968657208841966, "grad_norm": 0.27386826276779175, "learning_rate": 8e-05, "loss": 1.6858, "step": 2259 }, { "epoch": 0.49708567029583195, "grad_norm": 0.2911466658115387, "learning_rate": 8e-05, "loss": 1.701, "step": 2260 }, { "epoch": 0.4973056197074673, "grad_norm": 0.2837962508201599, "learning_rate": 8e-05, "loss": 1.6886, "step": 2261 }, { "epoch": 0.4975255691191026, "grad_norm": 0.3071229159832001, "learning_rate": 8e-05, "loss": 1.73, "step": 2262 }, { "epoch": 0.49774551853073795, "grad_norm": 0.303252249956131, "learning_rate": 8e-05, "loss": 1.7693, "step": 2263 }, { "epoch": 0.49796546794237323, "grad_norm": 0.2802221179008484, "learning_rate": 8e-05, "loss": 1.6394, "step": 2264 }, { "epoch": 0.49818541735400856, "grad_norm": 0.28856000304222107, "learning_rate": 8e-05, "loss": 1.6035, "step": 2265 }, { "epoch": 0.4984053667656439, "grad_norm": 0.28943875432014465, "learning_rate": 8e-05, "loss": 1.7989, "step": 2266 }, { "epoch": 0.49862531617727923, "grad_norm": 0.26969149708747864, "learning_rate": 8e-05, "loss": 1.6607, "step": 2267 }, { "epoch": 0.49884526558891457, "grad_norm": 0.311819851398468, "learning_rate": 8e-05, "loss": 1.7912, "step": 2268 }, { "epoch": 0.4990652150005499, "grad_norm": 0.296274334192276, "learning_rate": 8e-05, "loss": 1.6953, "step": 2269 }, { "epoch": 0.4992851644121852, "grad_norm": 0.26551195979118347, "learning_rate": 8e-05, "loss": 1.3549, "step": 2270 }, { "epoch": 0.4995051138238205, "grad_norm": 0.28540030121803284, "learning_rate": 8e-05, "loss": 1.7142, "step": 2271 }, { "epoch": 0.49972506323545585, "grad_norm": 0.2834233045578003, "learning_rate": 8e-05, "loss": 1.7293, "step": 2272 }, { "epoch": 0.4999450126470912, "grad_norm": 0.34650975465774536, "learning_rate": 8e-05, "loss": 1.7652, "step": 2273 }, { "epoch": 0.5001649620587265, "grad_norm": 0.2988453507423401, "learning_rate": 8e-05, "loss": 1.8895, "step": 2274 }, { "epoch": 0.5003849114703618, "grad_norm": 0.2912983000278473, "learning_rate": 8e-05, "loss": 1.6431, "step": 2275 }, { "epoch": 0.5006048608819972, "grad_norm": 0.28406545519828796, "learning_rate": 8e-05, "loss": 1.7805, "step": 2276 }, { "epoch": 0.5008248102936325, "grad_norm": 0.2748315632343292, "learning_rate": 8e-05, "loss": 1.5514, "step": 2277 }, { "epoch": 0.5010447597052677, "grad_norm": 0.3016912341117859, "learning_rate": 8e-05, "loss": 1.6271, "step": 2278 }, { "epoch": 0.5012647091169031, "grad_norm": 0.3006996512413025, "learning_rate": 8e-05, "loss": 1.7195, "step": 2279 }, { "epoch": 0.5014846585285384, "grad_norm": 0.27598950266838074, "learning_rate": 8e-05, "loss": 1.4975, "step": 2280 }, { "epoch": 0.5017046079401738, "grad_norm": 0.2810399830341339, "learning_rate": 8e-05, "loss": 1.6498, "step": 2281 }, { "epoch": 0.5019245573518091, "grad_norm": 0.3018679916858673, "learning_rate": 8e-05, "loss": 1.6711, "step": 2282 }, { "epoch": 0.5021445067634444, "grad_norm": 0.2889658510684967, "learning_rate": 8e-05, "loss": 1.7318, "step": 2283 }, { "epoch": 0.5023644561750797, "grad_norm": 0.28475597500801086, "learning_rate": 8e-05, "loss": 1.7573, "step": 2284 }, { "epoch": 0.502584405586715, "grad_norm": 0.3001987338066101, "learning_rate": 8e-05, "loss": 1.8463, "step": 2285 }, { "epoch": 0.5028043549983504, "grad_norm": 0.273773193359375, "learning_rate": 8e-05, "loss": 1.6465, "step": 2286 }, { "epoch": 0.5030243044099857, "grad_norm": 0.34727615118026733, "learning_rate": 8e-05, "loss": 1.9431, "step": 2287 }, { "epoch": 0.5032442538216211, "grad_norm": 0.2818615138530731, "learning_rate": 8e-05, "loss": 1.6113, "step": 2288 }, { "epoch": 0.5034642032332564, "grad_norm": 0.26619333028793335, "learning_rate": 8e-05, "loss": 1.5167, "step": 2289 }, { "epoch": 0.5036841526448916, "grad_norm": 0.2748722434043884, "learning_rate": 8e-05, "loss": 1.6384, "step": 2290 }, { "epoch": 0.503904102056527, "grad_norm": 0.29873546957969666, "learning_rate": 8e-05, "loss": 1.7309, "step": 2291 }, { "epoch": 0.5041240514681623, "grad_norm": 0.28378361463546753, "learning_rate": 8e-05, "loss": 1.6788, "step": 2292 }, { "epoch": 0.5043440008797977, "grad_norm": 0.28786107897758484, "learning_rate": 8e-05, "loss": 1.7281, "step": 2293 }, { "epoch": 0.504563950291433, "grad_norm": 0.2831546366214752, "learning_rate": 8e-05, "loss": 1.7644, "step": 2294 }, { "epoch": 0.5047838997030683, "grad_norm": 0.28964316844940186, "learning_rate": 8e-05, "loss": 1.6409, "step": 2295 }, { "epoch": 0.5050038491147036, "grad_norm": 0.2778918743133545, "learning_rate": 8e-05, "loss": 1.6078, "step": 2296 }, { "epoch": 0.5052237985263389, "grad_norm": 0.2749491035938263, "learning_rate": 8e-05, "loss": 1.668, "step": 2297 }, { "epoch": 0.5054437479379743, "grad_norm": 0.2856389880180359, "learning_rate": 8e-05, "loss": 1.7052, "step": 2298 }, { "epoch": 0.5056636973496096, "grad_norm": 0.28082379698753357, "learning_rate": 8e-05, "loss": 1.6627, "step": 2299 }, { "epoch": 0.5058836467612449, "grad_norm": 0.27894240617752075, "learning_rate": 8e-05, "loss": 1.6408, "step": 2300 }, { "epoch": 0.5061035961728803, "grad_norm": 0.31712663173675537, "learning_rate": 8e-05, "loss": 1.6148, "step": 2301 }, { "epoch": 0.5063235455845155, "grad_norm": 0.28600454330444336, "learning_rate": 8e-05, "loss": 1.7695, "step": 2302 }, { "epoch": 0.5065434949961509, "grad_norm": 0.3285694718360901, "learning_rate": 8e-05, "loss": 1.7882, "step": 2303 }, { "epoch": 0.5067634444077862, "grad_norm": 0.27823877334594727, "learning_rate": 8e-05, "loss": 1.4648, "step": 2304 }, { "epoch": 0.5069833938194215, "grad_norm": 0.3157597482204437, "learning_rate": 8e-05, "loss": 1.8441, "step": 2305 }, { "epoch": 0.5072033432310569, "grad_norm": 0.2913108170032501, "learning_rate": 8e-05, "loss": 1.7584, "step": 2306 }, { "epoch": 0.5074232926426921, "grad_norm": 0.28753626346588135, "learning_rate": 8e-05, "loss": 1.7045, "step": 2307 }, { "epoch": 0.5076432420543275, "grad_norm": 0.2981377840042114, "learning_rate": 8e-05, "loss": 1.54, "step": 2308 }, { "epoch": 0.5078631914659628, "grad_norm": 0.2911180853843689, "learning_rate": 8e-05, "loss": 1.8266, "step": 2309 }, { "epoch": 0.5080831408775982, "grad_norm": 0.2862488031387329, "learning_rate": 8e-05, "loss": 1.7503, "step": 2310 }, { "epoch": 0.5083030902892335, "grad_norm": 0.3015568256378174, "learning_rate": 8e-05, "loss": 1.8209, "step": 2311 }, { "epoch": 0.5085230397008688, "grad_norm": 0.27230823040008545, "learning_rate": 8e-05, "loss": 1.7977, "step": 2312 }, { "epoch": 0.5087429891125042, "grad_norm": 0.27871981263160706, "learning_rate": 8e-05, "loss": 1.5971, "step": 2313 }, { "epoch": 0.5089629385241394, "grad_norm": 0.2884436249732971, "learning_rate": 8e-05, "loss": 1.7726, "step": 2314 }, { "epoch": 0.5091828879357748, "grad_norm": 0.2778714895248413, "learning_rate": 8e-05, "loss": 1.7104, "step": 2315 }, { "epoch": 0.5094028373474101, "grad_norm": 0.30202385783195496, "learning_rate": 8e-05, "loss": 1.658, "step": 2316 }, { "epoch": 0.5096227867590454, "grad_norm": 0.2785106599330902, "learning_rate": 8e-05, "loss": 1.717, "step": 2317 }, { "epoch": 0.5098427361706808, "grad_norm": 0.28846096992492676, "learning_rate": 8e-05, "loss": 1.7456, "step": 2318 }, { "epoch": 0.510062685582316, "grad_norm": 0.27753984928131104, "learning_rate": 8e-05, "loss": 1.6569, "step": 2319 }, { "epoch": 0.5102826349939514, "grad_norm": 0.2834784686565399, "learning_rate": 8e-05, "loss": 1.6348, "step": 2320 }, { "epoch": 0.5105025844055867, "grad_norm": 0.27789169549942017, "learning_rate": 8e-05, "loss": 1.5586, "step": 2321 }, { "epoch": 0.510722533817222, "grad_norm": 0.28466710448265076, "learning_rate": 8e-05, "loss": 1.6793, "step": 2322 }, { "epoch": 0.5109424832288574, "grad_norm": 0.2759189009666443, "learning_rate": 8e-05, "loss": 1.5311, "step": 2323 }, { "epoch": 0.5111624326404927, "grad_norm": 0.2931334674358368, "learning_rate": 8e-05, "loss": 1.7258, "step": 2324 }, { "epoch": 0.511382382052128, "grad_norm": 0.2740546464920044, "learning_rate": 8e-05, "loss": 1.558, "step": 2325 }, { "epoch": 0.5116023314637633, "grad_norm": 0.29584407806396484, "learning_rate": 8e-05, "loss": 1.7777, "step": 2326 }, { "epoch": 0.5118222808753986, "grad_norm": 0.2948019504547119, "learning_rate": 8e-05, "loss": 1.7722, "step": 2327 }, { "epoch": 0.512042230287034, "grad_norm": 0.27409225702285767, "learning_rate": 8e-05, "loss": 1.6524, "step": 2328 }, { "epoch": 0.5122621796986693, "grad_norm": 0.26278048753738403, "learning_rate": 8e-05, "loss": 1.5945, "step": 2329 }, { "epoch": 0.5124821291103047, "grad_norm": 0.29483261704444885, "learning_rate": 8e-05, "loss": 1.8132, "step": 2330 }, { "epoch": 0.5127020785219399, "grad_norm": 0.27037349343299866, "learning_rate": 8e-05, "loss": 1.5657, "step": 2331 }, { "epoch": 0.5129220279335753, "grad_norm": 0.2826361060142517, "learning_rate": 8e-05, "loss": 1.6955, "step": 2332 }, { "epoch": 0.5131419773452106, "grad_norm": 0.2957696318626404, "learning_rate": 8e-05, "loss": 1.814, "step": 2333 }, { "epoch": 0.5133619267568459, "grad_norm": 0.2752826511859894, "learning_rate": 8e-05, "loss": 1.636, "step": 2334 }, { "epoch": 0.5135818761684813, "grad_norm": 0.28523313999176025, "learning_rate": 8e-05, "loss": 1.5444, "step": 2335 }, { "epoch": 0.5138018255801166, "grad_norm": 0.286304235458374, "learning_rate": 8e-05, "loss": 1.4665, "step": 2336 }, { "epoch": 0.514021774991752, "grad_norm": 0.28738734126091003, "learning_rate": 8e-05, "loss": 1.6802, "step": 2337 }, { "epoch": 0.5142417244033872, "grad_norm": 0.2669237554073334, "learning_rate": 8e-05, "loss": 1.6011, "step": 2338 }, { "epoch": 0.5144616738150225, "grad_norm": 0.274325430393219, "learning_rate": 8e-05, "loss": 1.6248, "step": 2339 }, { "epoch": 0.5146816232266579, "grad_norm": 0.2798522710800171, "learning_rate": 8e-05, "loss": 1.7636, "step": 2340 }, { "epoch": 0.5149015726382932, "grad_norm": 0.27266305685043335, "learning_rate": 8e-05, "loss": 1.6434, "step": 2341 }, { "epoch": 0.5151215220499286, "grad_norm": 0.2740791440010071, "learning_rate": 8e-05, "loss": 1.7084, "step": 2342 }, { "epoch": 0.5153414714615638, "grad_norm": 0.28098320960998535, "learning_rate": 8e-05, "loss": 1.6356, "step": 2343 }, { "epoch": 0.5155614208731991, "grad_norm": 0.2760515809059143, "learning_rate": 8e-05, "loss": 1.661, "step": 2344 }, { "epoch": 0.5157813702848345, "grad_norm": 0.27894532680511475, "learning_rate": 8e-05, "loss": 1.6794, "step": 2345 }, { "epoch": 0.5160013196964698, "grad_norm": 0.2972679138183594, "learning_rate": 8e-05, "loss": 1.6943, "step": 2346 }, { "epoch": 0.5162212691081052, "grad_norm": 0.3100125789642334, "learning_rate": 8e-05, "loss": 1.7214, "step": 2347 }, { "epoch": 0.5164412185197405, "grad_norm": 0.2743578255176544, "learning_rate": 8e-05, "loss": 1.7021, "step": 2348 }, { "epoch": 0.5166611679313757, "grad_norm": 0.29266777634620667, "learning_rate": 8e-05, "loss": 1.7585, "step": 2349 }, { "epoch": 0.5168811173430111, "grad_norm": 0.2791600227355957, "learning_rate": 8e-05, "loss": 1.7012, "step": 2350 }, { "epoch": 0.5171010667546464, "grad_norm": 0.28535401821136475, "learning_rate": 8e-05, "loss": 1.6695, "step": 2351 }, { "epoch": 0.5173210161662818, "grad_norm": 0.2860865592956543, "learning_rate": 8e-05, "loss": 1.7419, "step": 2352 }, { "epoch": 0.5175409655779171, "grad_norm": 0.27693790197372437, "learning_rate": 8e-05, "loss": 1.5459, "step": 2353 }, { "epoch": 0.5177609149895525, "grad_norm": 0.2858433723449707, "learning_rate": 8e-05, "loss": 1.799, "step": 2354 }, { "epoch": 0.5179808644011877, "grad_norm": 0.30761632323265076, "learning_rate": 8e-05, "loss": 1.5971, "step": 2355 }, { "epoch": 0.518200813812823, "grad_norm": 0.2943046987056732, "learning_rate": 8e-05, "loss": 1.7399, "step": 2356 }, { "epoch": 0.5184207632244584, "grad_norm": 0.2748922109603882, "learning_rate": 8e-05, "loss": 1.7202, "step": 2357 }, { "epoch": 0.5186407126360937, "grad_norm": 0.2818071246147156, "learning_rate": 8e-05, "loss": 1.5918, "step": 2358 }, { "epoch": 0.5188606620477291, "grad_norm": 0.28235137462615967, "learning_rate": 8e-05, "loss": 1.5728, "step": 2359 }, { "epoch": 0.5190806114593643, "grad_norm": 0.27833959460258484, "learning_rate": 8e-05, "loss": 1.5966, "step": 2360 }, { "epoch": 0.5193005608709996, "grad_norm": 0.2731468677520752, "learning_rate": 8e-05, "loss": 1.6518, "step": 2361 }, { "epoch": 0.519520510282635, "grad_norm": 0.2777821719646454, "learning_rate": 8e-05, "loss": 1.6885, "step": 2362 }, { "epoch": 0.5197404596942703, "grad_norm": 0.2685951590538025, "learning_rate": 8e-05, "loss": 1.6315, "step": 2363 }, { "epoch": 0.5199604091059057, "grad_norm": 0.3087875545024872, "learning_rate": 8e-05, "loss": 1.7525, "step": 2364 }, { "epoch": 0.520180358517541, "grad_norm": 0.2693195044994354, "learning_rate": 8e-05, "loss": 1.5993, "step": 2365 }, { "epoch": 0.5204003079291762, "grad_norm": 0.2832968235015869, "learning_rate": 8e-05, "loss": 1.5798, "step": 2366 }, { "epoch": 0.5206202573408116, "grad_norm": 0.2873738706111908, "learning_rate": 8e-05, "loss": 1.7373, "step": 2367 }, { "epoch": 0.5208402067524469, "grad_norm": 0.2888682782649994, "learning_rate": 8e-05, "loss": 1.6995, "step": 2368 }, { "epoch": 0.5210601561640823, "grad_norm": 0.2788809537887573, "learning_rate": 8e-05, "loss": 1.5928, "step": 2369 }, { "epoch": 0.5212801055757176, "grad_norm": 0.28021880984306335, "learning_rate": 8e-05, "loss": 1.6692, "step": 2370 }, { "epoch": 0.5215000549873529, "grad_norm": 0.3041718006134033, "learning_rate": 8e-05, "loss": 1.7378, "step": 2371 }, { "epoch": 0.5217200043989882, "grad_norm": 0.2774748206138611, "learning_rate": 8e-05, "loss": 1.5802, "step": 2372 }, { "epoch": 0.5219399538106235, "grad_norm": 0.2876451015472412, "learning_rate": 8e-05, "loss": 1.8057, "step": 2373 }, { "epoch": 0.5221599032222589, "grad_norm": 0.2740166485309601, "learning_rate": 8e-05, "loss": 1.6694, "step": 2374 }, { "epoch": 0.5223798526338942, "grad_norm": 0.288555771112442, "learning_rate": 8e-05, "loss": 1.792, "step": 2375 }, { "epoch": 0.5225998020455296, "grad_norm": 0.2777664065361023, "learning_rate": 8e-05, "loss": 1.6781, "step": 2376 }, { "epoch": 0.5228197514571649, "grad_norm": 0.27108079195022583, "learning_rate": 8e-05, "loss": 1.4881, "step": 2377 }, { "epoch": 0.5230397008688001, "grad_norm": 0.2909669578075409, "learning_rate": 8e-05, "loss": 1.7174, "step": 2378 }, { "epoch": 0.5232596502804355, "grad_norm": 0.2978494167327881, "learning_rate": 8e-05, "loss": 1.8641, "step": 2379 }, { "epoch": 0.5234795996920708, "grad_norm": 0.2649437487125397, "learning_rate": 8e-05, "loss": 1.473, "step": 2380 }, { "epoch": 0.5236995491037062, "grad_norm": 0.28939372301101685, "learning_rate": 8e-05, "loss": 1.5567, "step": 2381 }, { "epoch": 0.5239194985153415, "grad_norm": 0.2740820646286011, "learning_rate": 8e-05, "loss": 1.6321, "step": 2382 }, { "epoch": 0.5241394479269768, "grad_norm": 0.28426122665405273, "learning_rate": 8e-05, "loss": 1.5952, "step": 2383 }, { "epoch": 0.5243593973386121, "grad_norm": 0.28176257014274597, "learning_rate": 8e-05, "loss": 1.6231, "step": 2384 }, { "epoch": 0.5245793467502474, "grad_norm": 0.29681360721588135, "learning_rate": 8e-05, "loss": 1.8203, "step": 2385 }, { "epoch": 0.5247992961618828, "grad_norm": 0.272658109664917, "learning_rate": 8e-05, "loss": 1.6942, "step": 2386 }, { "epoch": 0.5250192455735181, "grad_norm": 0.27786141633987427, "learning_rate": 8e-05, "loss": 1.6081, "step": 2387 }, { "epoch": 0.5252391949851534, "grad_norm": 0.2938309609889984, "learning_rate": 8e-05, "loss": 1.7454, "step": 2388 }, { "epoch": 0.5254591443967888, "grad_norm": 0.2710343599319458, "learning_rate": 8e-05, "loss": 1.6391, "step": 2389 }, { "epoch": 0.525679093808424, "grad_norm": 0.2757870554924011, "learning_rate": 8e-05, "loss": 1.6526, "step": 2390 }, { "epoch": 0.5258990432200594, "grad_norm": 0.2581859827041626, "learning_rate": 8e-05, "loss": 1.4344, "step": 2391 }, { "epoch": 0.5261189926316947, "grad_norm": 0.2732166647911072, "learning_rate": 8e-05, "loss": 1.5987, "step": 2392 }, { "epoch": 0.52633894204333, "grad_norm": 0.2859753370285034, "learning_rate": 8e-05, "loss": 1.6654, "step": 2393 }, { "epoch": 0.5265588914549654, "grad_norm": 0.2680748701095581, "learning_rate": 8e-05, "loss": 1.5764, "step": 2394 }, { "epoch": 0.5267788408666007, "grad_norm": 0.2866816818714142, "learning_rate": 8e-05, "loss": 1.7725, "step": 2395 }, { "epoch": 0.526998790278236, "grad_norm": 0.27792397141456604, "learning_rate": 8e-05, "loss": 1.5932, "step": 2396 }, { "epoch": 0.5272187396898713, "grad_norm": 0.26985982060432434, "learning_rate": 8e-05, "loss": 1.5955, "step": 2397 }, { "epoch": 0.5274386891015067, "grad_norm": 0.28183454275131226, "learning_rate": 8e-05, "loss": 1.553, "step": 2398 }, { "epoch": 0.527658638513142, "grad_norm": 0.29282763600349426, "learning_rate": 8e-05, "loss": 1.7858, "step": 2399 }, { "epoch": 0.5278785879247773, "grad_norm": 0.30619367957115173, "learning_rate": 8e-05, "loss": 1.718, "step": 2400 }, { "epoch": 0.5280985373364127, "grad_norm": 0.26707130670547485, "learning_rate": 8e-05, "loss": 1.5899, "step": 2401 }, { "epoch": 0.5283184867480479, "grad_norm": 0.3182383179664612, "learning_rate": 8e-05, "loss": 1.7268, "step": 2402 }, { "epoch": 0.5285384361596833, "grad_norm": 0.3178313374519348, "learning_rate": 8e-05, "loss": 1.7282, "step": 2403 }, { "epoch": 0.5287583855713186, "grad_norm": 0.26504799723625183, "learning_rate": 8e-05, "loss": 1.6046, "step": 2404 }, { "epoch": 0.5289783349829539, "grad_norm": 0.2749512195587158, "learning_rate": 8e-05, "loss": 1.6251, "step": 2405 }, { "epoch": 0.5291982843945893, "grad_norm": 0.27312803268432617, "learning_rate": 8e-05, "loss": 1.746, "step": 2406 }, { "epoch": 0.5294182338062245, "grad_norm": 0.26339027285575867, "learning_rate": 8e-05, "loss": 1.5214, "step": 2407 }, { "epoch": 0.5296381832178599, "grad_norm": 0.28254935145378113, "learning_rate": 8e-05, "loss": 1.629, "step": 2408 }, { "epoch": 0.5298581326294952, "grad_norm": 0.2761283218860626, "learning_rate": 8e-05, "loss": 1.7202, "step": 2409 }, { "epoch": 0.5300780820411305, "grad_norm": 0.27570095658302307, "learning_rate": 8e-05, "loss": 1.7042, "step": 2410 }, { "epoch": 0.5302980314527659, "grad_norm": 0.2886349856853485, "learning_rate": 8e-05, "loss": 1.7923, "step": 2411 }, { "epoch": 0.5305179808644012, "grad_norm": 0.29611504077911377, "learning_rate": 8e-05, "loss": 1.7219, "step": 2412 }, { "epoch": 0.5307379302760366, "grad_norm": 0.28122174739837646, "learning_rate": 8e-05, "loss": 1.6888, "step": 2413 }, { "epoch": 0.5309578796876718, "grad_norm": 0.2690391540527344, "learning_rate": 8e-05, "loss": 1.2745, "step": 2414 }, { "epoch": 0.5311778290993071, "grad_norm": 0.2676471173763275, "learning_rate": 8e-05, "loss": 1.6422, "step": 2415 }, { "epoch": 0.5313977785109425, "grad_norm": 0.2947712540626526, "learning_rate": 8e-05, "loss": 1.654, "step": 2416 }, { "epoch": 0.5316177279225778, "grad_norm": 0.27766644954681396, "learning_rate": 8e-05, "loss": 1.6208, "step": 2417 }, { "epoch": 0.5318376773342132, "grad_norm": 0.28579944372177124, "learning_rate": 8e-05, "loss": 1.8171, "step": 2418 }, { "epoch": 0.5320576267458484, "grad_norm": 0.2734217345714569, "learning_rate": 8e-05, "loss": 1.658, "step": 2419 }, { "epoch": 0.5322775761574838, "grad_norm": 0.28343021869659424, "learning_rate": 8e-05, "loss": 1.6291, "step": 2420 }, { "epoch": 0.5324975255691191, "grad_norm": 0.2881801128387451, "learning_rate": 8e-05, "loss": 1.5212, "step": 2421 }, { "epoch": 0.5327174749807544, "grad_norm": 0.27267688512802124, "learning_rate": 8e-05, "loss": 1.6599, "step": 2422 }, { "epoch": 0.5329374243923898, "grad_norm": 0.29100489616394043, "learning_rate": 8e-05, "loss": 1.6636, "step": 2423 }, { "epoch": 0.5331573738040251, "grad_norm": 0.301812082529068, "learning_rate": 8e-05, "loss": 1.9097, "step": 2424 }, { "epoch": 0.5333773232156604, "grad_norm": 0.2864093482494354, "learning_rate": 8e-05, "loss": 1.6535, "step": 2425 }, { "epoch": 0.5335972726272957, "grad_norm": 0.28721320629119873, "learning_rate": 8e-05, "loss": 1.7307, "step": 2426 }, { "epoch": 0.533817222038931, "grad_norm": 0.3100323975086212, "learning_rate": 8e-05, "loss": 1.7155, "step": 2427 }, { "epoch": 0.5340371714505664, "grad_norm": 0.2595236301422119, "learning_rate": 8e-05, "loss": 1.4525, "step": 2428 }, { "epoch": 0.5342571208622017, "grad_norm": 0.27981269359588623, "learning_rate": 8e-05, "loss": 1.6821, "step": 2429 }, { "epoch": 0.5344770702738371, "grad_norm": 0.28523892164230347, "learning_rate": 8e-05, "loss": 1.6213, "step": 2430 }, { "epoch": 0.5346970196854723, "grad_norm": 0.2951820194721222, "learning_rate": 8e-05, "loss": 1.7798, "step": 2431 }, { "epoch": 0.5349169690971076, "grad_norm": 0.27744752168655396, "learning_rate": 8e-05, "loss": 1.664, "step": 2432 }, { "epoch": 0.535136918508743, "grad_norm": 0.2700327932834625, "learning_rate": 8e-05, "loss": 1.6476, "step": 2433 }, { "epoch": 0.5353568679203783, "grad_norm": 0.3043116331100464, "learning_rate": 8e-05, "loss": 1.8377, "step": 2434 }, { "epoch": 0.5355768173320137, "grad_norm": 0.2886519730091095, "learning_rate": 8e-05, "loss": 1.7098, "step": 2435 }, { "epoch": 0.535796766743649, "grad_norm": 0.28121626377105713, "learning_rate": 8e-05, "loss": 1.5902, "step": 2436 }, { "epoch": 0.5360167161552842, "grad_norm": 0.28657859563827515, "learning_rate": 8e-05, "loss": 1.6769, "step": 2437 }, { "epoch": 0.5362366655669196, "grad_norm": 0.3111754059791565, "learning_rate": 8e-05, "loss": 1.8352, "step": 2438 }, { "epoch": 0.5364566149785549, "grad_norm": 0.27172762155532837, "learning_rate": 8e-05, "loss": 1.5897, "step": 2439 }, { "epoch": 0.5366765643901903, "grad_norm": 0.28469017148017883, "learning_rate": 8e-05, "loss": 1.6976, "step": 2440 }, { "epoch": 0.5368965138018256, "grad_norm": 0.29801180958747864, "learning_rate": 8e-05, "loss": 1.7811, "step": 2441 }, { "epoch": 0.5371164632134608, "grad_norm": 0.2860267758369446, "learning_rate": 8e-05, "loss": 1.7026, "step": 2442 }, { "epoch": 0.5373364126250962, "grad_norm": 0.3069910705089569, "learning_rate": 8e-05, "loss": 1.823, "step": 2443 }, { "epoch": 0.5375563620367315, "grad_norm": 0.29847028851509094, "learning_rate": 8e-05, "loss": 1.6314, "step": 2444 }, { "epoch": 0.5377763114483669, "grad_norm": 0.2970685660839081, "learning_rate": 8e-05, "loss": 1.8591, "step": 2445 }, { "epoch": 0.5379962608600022, "grad_norm": 0.28767916560173035, "learning_rate": 8e-05, "loss": 1.6244, "step": 2446 }, { "epoch": 0.5382162102716376, "grad_norm": 0.2858954966068268, "learning_rate": 8e-05, "loss": 1.6002, "step": 2447 }, { "epoch": 0.5384361596832729, "grad_norm": 0.25083082914352417, "learning_rate": 8e-05, "loss": 1.4772, "step": 2448 }, { "epoch": 0.5386561090949081, "grad_norm": 0.28772327303886414, "learning_rate": 8e-05, "loss": 1.7885, "step": 2449 }, { "epoch": 0.5388760585065435, "grad_norm": 0.300503671169281, "learning_rate": 8e-05, "loss": 1.8074, "step": 2450 }, { "epoch": 0.5390960079181788, "grad_norm": 0.29243797063827515, "learning_rate": 8e-05, "loss": 1.7033, "step": 2451 }, { "epoch": 0.5393159573298142, "grad_norm": 0.28921830654144287, "learning_rate": 8e-05, "loss": 1.734, "step": 2452 }, { "epoch": 0.5395359067414495, "grad_norm": 0.2754501700401306, "learning_rate": 8e-05, "loss": 1.577, "step": 2453 }, { "epoch": 0.5397558561530847, "grad_norm": 0.26824522018432617, "learning_rate": 8e-05, "loss": 1.6434, "step": 2454 }, { "epoch": 0.5399758055647201, "grad_norm": 0.26851388812065125, "learning_rate": 8e-05, "loss": 1.6706, "step": 2455 }, { "epoch": 0.5401957549763554, "grad_norm": 0.2697846293449402, "learning_rate": 8e-05, "loss": 1.6052, "step": 2456 }, { "epoch": 0.5404157043879908, "grad_norm": 0.27774059772491455, "learning_rate": 8e-05, "loss": 1.663, "step": 2457 }, { "epoch": 0.5406356537996261, "grad_norm": 0.2799103558063507, "learning_rate": 8e-05, "loss": 1.9412, "step": 2458 }, { "epoch": 0.5408556032112614, "grad_norm": 0.2874007523059845, "learning_rate": 8e-05, "loss": 1.6366, "step": 2459 }, { "epoch": 0.5410755526228967, "grad_norm": 0.29054176807403564, "learning_rate": 8e-05, "loss": 1.7546, "step": 2460 }, { "epoch": 0.541295502034532, "grad_norm": 0.29359421133995056, "learning_rate": 8e-05, "loss": 1.8274, "step": 2461 }, { "epoch": 0.5415154514461674, "grad_norm": 0.29589033126831055, "learning_rate": 8e-05, "loss": 1.8619, "step": 2462 }, { "epoch": 0.5417354008578027, "grad_norm": 0.2997150421142578, "learning_rate": 8e-05, "loss": 1.7685, "step": 2463 }, { "epoch": 0.541955350269438, "grad_norm": 0.2759319543838501, "learning_rate": 8e-05, "loss": 1.5652, "step": 2464 }, { "epoch": 0.5421752996810734, "grad_norm": 0.27741214632987976, "learning_rate": 8e-05, "loss": 1.6247, "step": 2465 }, { "epoch": 0.5423952490927086, "grad_norm": 0.29365673661231995, "learning_rate": 8e-05, "loss": 1.7768, "step": 2466 }, { "epoch": 0.542615198504344, "grad_norm": 0.2897026836872101, "learning_rate": 8e-05, "loss": 1.7435, "step": 2467 }, { "epoch": 0.5428351479159793, "grad_norm": 0.2963312566280365, "learning_rate": 8e-05, "loss": 1.7792, "step": 2468 }, { "epoch": 0.5430550973276147, "grad_norm": 0.3142043948173523, "learning_rate": 8e-05, "loss": 1.8348, "step": 2469 }, { "epoch": 0.54327504673925, "grad_norm": 0.28869184851646423, "learning_rate": 8e-05, "loss": 1.6916, "step": 2470 }, { "epoch": 0.5434949961508853, "grad_norm": 0.27220281958580017, "learning_rate": 8e-05, "loss": 1.5963, "step": 2471 }, { "epoch": 0.5437149455625206, "grad_norm": 0.3002524971961975, "learning_rate": 8e-05, "loss": 1.7516, "step": 2472 }, { "epoch": 0.5439348949741559, "grad_norm": 0.27016308903694153, "learning_rate": 8e-05, "loss": 1.5655, "step": 2473 }, { "epoch": 0.5441548443857913, "grad_norm": 0.2886146903038025, "learning_rate": 8e-05, "loss": 1.6563, "step": 2474 }, { "epoch": 0.5443747937974266, "grad_norm": 0.2743261158466339, "learning_rate": 8e-05, "loss": 1.7916, "step": 2475 }, { "epoch": 0.5445947432090619, "grad_norm": 0.27933475375175476, "learning_rate": 8e-05, "loss": 1.7397, "step": 2476 }, { "epoch": 0.5448146926206973, "grad_norm": 0.2805885672569275, "learning_rate": 8e-05, "loss": 1.6608, "step": 2477 }, { "epoch": 0.5450346420323325, "grad_norm": 0.26985716819763184, "learning_rate": 8e-05, "loss": 1.6551, "step": 2478 }, { "epoch": 0.5452545914439679, "grad_norm": 0.2778765857219696, "learning_rate": 8e-05, "loss": 1.6792, "step": 2479 }, { "epoch": 0.5454745408556032, "grad_norm": 0.27623313665390015, "learning_rate": 8e-05, "loss": 1.6867, "step": 2480 }, { "epoch": 0.5456944902672385, "grad_norm": 0.27185389399528503, "learning_rate": 8e-05, "loss": 1.6184, "step": 2481 }, { "epoch": 0.5459144396788739, "grad_norm": 0.29302138090133667, "learning_rate": 8e-05, "loss": 1.7075, "step": 2482 }, { "epoch": 0.5461343890905092, "grad_norm": 0.26639193296432495, "learning_rate": 8e-05, "loss": 1.6138, "step": 2483 }, { "epoch": 0.5463543385021445, "grad_norm": 0.28048211336135864, "learning_rate": 8e-05, "loss": 1.6672, "step": 2484 }, { "epoch": 0.5465742879137798, "grad_norm": 0.2844570577144623, "learning_rate": 8e-05, "loss": 1.6888, "step": 2485 }, { "epoch": 0.5467942373254151, "grad_norm": 0.2801128923892975, "learning_rate": 8e-05, "loss": 1.8506, "step": 2486 }, { "epoch": 0.5470141867370505, "grad_norm": 0.2718241810798645, "learning_rate": 8e-05, "loss": 1.6105, "step": 2487 }, { "epoch": 0.5472341361486858, "grad_norm": 0.28759825229644775, "learning_rate": 8e-05, "loss": 1.7449, "step": 2488 }, { "epoch": 0.5474540855603212, "grad_norm": 0.29218876361846924, "learning_rate": 8e-05, "loss": 1.8732, "step": 2489 }, { "epoch": 0.5476740349719564, "grad_norm": 0.29760751128196716, "learning_rate": 8e-05, "loss": 1.7804, "step": 2490 }, { "epoch": 0.5478939843835918, "grad_norm": 0.28636956214904785, "learning_rate": 8e-05, "loss": 1.6994, "step": 2491 }, { "epoch": 0.5481139337952271, "grad_norm": 0.2892046570777893, "learning_rate": 8e-05, "loss": 1.759, "step": 2492 }, { "epoch": 0.5483338832068624, "grad_norm": 0.280556857585907, "learning_rate": 8e-05, "loss": 1.8084, "step": 2493 }, { "epoch": 0.5485538326184978, "grad_norm": 0.2733471691608429, "learning_rate": 8e-05, "loss": 1.7293, "step": 2494 }, { "epoch": 0.548773782030133, "grad_norm": 0.2813643515110016, "learning_rate": 8e-05, "loss": 1.6178, "step": 2495 }, { "epoch": 0.5489937314417684, "grad_norm": 0.27255943417549133, "learning_rate": 8e-05, "loss": 1.6827, "step": 2496 }, { "epoch": 0.5492136808534037, "grad_norm": 0.2690375745296478, "learning_rate": 8e-05, "loss": 1.684, "step": 2497 }, { "epoch": 0.549433630265039, "grad_norm": 0.30036401748657227, "learning_rate": 8e-05, "loss": 1.8676, "step": 2498 }, { "epoch": 0.5496535796766744, "grad_norm": 0.27924251556396484, "learning_rate": 8e-05, "loss": 1.6619, "step": 2499 }, { "epoch": 0.5498735290883097, "grad_norm": 0.2792947590351105, "learning_rate": 8e-05, "loss": 1.6121, "step": 2500 }, { "epoch": 0.550093478499945, "grad_norm": 0.27976930141448975, "learning_rate": 8e-05, "loss": 1.5815, "step": 2501 }, { "epoch": 0.5503134279115803, "grad_norm": 0.28429850935935974, "learning_rate": 8e-05, "loss": 1.6713, "step": 2502 }, { "epoch": 0.5505333773232156, "grad_norm": 0.2669944763183594, "learning_rate": 8e-05, "loss": 1.5065, "step": 2503 }, { "epoch": 0.550753326734851, "grad_norm": 0.2846994400024414, "learning_rate": 8e-05, "loss": 1.7238, "step": 2504 }, { "epoch": 0.5509732761464863, "grad_norm": 0.27598071098327637, "learning_rate": 8e-05, "loss": 1.5364, "step": 2505 }, { "epoch": 0.5511932255581217, "grad_norm": 0.27275460958480835, "learning_rate": 8e-05, "loss": 1.6171, "step": 2506 }, { "epoch": 0.551413174969757, "grad_norm": 0.2846895456314087, "learning_rate": 8e-05, "loss": 1.7082, "step": 2507 }, { "epoch": 0.5516331243813922, "grad_norm": 0.3010547161102295, "learning_rate": 8e-05, "loss": 1.7946, "step": 2508 }, { "epoch": 0.5518530737930276, "grad_norm": 0.28405773639678955, "learning_rate": 8e-05, "loss": 1.6063, "step": 2509 }, { "epoch": 0.5520730232046629, "grad_norm": 0.2855536639690399, "learning_rate": 8e-05, "loss": 1.6138, "step": 2510 }, { "epoch": 0.5522929726162983, "grad_norm": 0.2949456572532654, "learning_rate": 8e-05, "loss": 1.7734, "step": 2511 }, { "epoch": 0.5525129220279336, "grad_norm": 0.31665512919425964, "learning_rate": 8e-05, "loss": 1.7163, "step": 2512 }, { "epoch": 0.552732871439569, "grad_norm": 0.2881389260292053, "learning_rate": 8e-05, "loss": 1.5617, "step": 2513 }, { "epoch": 0.5529528208512042, "grad_norm": 0.26758721470832825, "learning_rate": 8e-05, "loss": 1.6714, "step": 2514 }, { "epoch": 0.5531727702628395, "grad_norm": 0.29549580812454224, "learning_rate": 8e-05, "loss": 1.6516, "step": 2515 }, { "epoch": 0.5533927196744749, "grad_norm": 0.2811340391635895, "learning_rate": 8e-05, "loss": 1.6026, "step": 2516 }, { "epoch": 0.5536126690861102, "grad_norm": 0.2837204039096832, "learning_rate": 8e-05, "loss": 1.8413, "step": 2517 }, { "epoch": 0.5538326184977456, "grad_norm": 0.276216983795166, "learning_rate": 8e-05, "loss": 1.6671, "step": 2518 }, { "epoch": 0.5540525679093808, "grad_norm": 0.2781767249107361, "learning_rate": 8e-05, "loss": 1.5886, "step": 2519 }, { "epoch": 0.5542725173210161, "grad_norm": 0.2830861210823059, "learning_rate": 8e-05, "loss": 1.6097, "step": 2520 }, { "epoch": 0.5544924667326515, "grad_norm": 0.2746805250644684, "learning_rate": 8e-05, "loss": 1.7153, "step": 2521 }, { "epoch": 0.5547124161442868, "grad_norm": 0.2781994640827179, "learning_rate": 8e-05, "loss": 1.6597, "step": 2522 }, { "epoch": 0.5549323655559222, "grad_norm": 0.2919979393482208, "learning_rate": 8e-05, "loss": 1.715, "step": 2523 }, { "epoch": 0.5551523149675575, "grad_norm": 0.27563661336898804, "learning_rate": 8e-05, "loss": 1.6129, "step": 2524 }, { "epoch": 0.5553722643791927, "grad_norm": 0.3070942163467407, "learning_rate": 8e-05, "loss": 1.875, "step": 2525 }, { "epoch": 0.5555922137908281, "grad_norm": 0.278039813041687, "learning_rate": 8e-05, "loss": 1.6894, "step": 2526 }, { "epoch": 0.5558121632024634, "grad_norm": 0.2709571421146393, "learning_rate": 8e-05, "loss": 1.5268, "step": 2527 }, { "epoch": 0.5560321126140988, "grad_norm": 0.27659523487091064, "learning_rate": 8e-05, "loss": 1.5817, "step": 2528 }, { "epoch": 0.5562520620257341, "grad_norm": 0.33376970887184143, "learning_rate": 8e-05, "loss": 1.6275, "step": 2529 }, { "epoch": 0.5564720114373694, "grad_norm": 0.28663134574890137, "learning_rate": 8e-05, "loss": 1.6497, "step": 2530 }, { "epoch": 0.5566919608490047, "grad_norm": 0.27400556206703186, "learning_rate": 8e-05, "loss": 1.6768, "step": 2531 }, { "epoch": 0.55691191026064, "grad_norm": 0.3359694182872772, "learning_rate": 8e-05, "loss": 1.7628, "step": 2532 }, { "epoch": 0.5571318596722754, "grad_norm": 0.3009445071220398, "learning_rate": 8e-05, "loss": 1.6368, "step": 2533 }, { "epoch": 0.5573518090839107, "grad_norm": 0.2951606512069702, "learning_rate": 8e-05, "loss": 1.6468, "step": 2534 }, { "epoch": 0.5575717584955461, "grad_norm": 0.298835426568985, "learning_rate": 8e-05, "loss": 1.679, "step": 2535 }, { "epoch": 0.5577917079071814, "grad_norm": 0.29196399450302124, "learning_rate": 8e-05, "loss": 1.6865, "step": 2536 }, { "epoch": 0.5580116573188166, "grad_norm": 0.3057127296924591, "learning_rate": 8e-05, "loss": 1.7979, "step": 2537 }, { "epoch": 0.558231606730452, "grad_norm": 0.3170565664768219, "learning_rate": 8e-05, "loss": 1.7815, "step": 2538 }, { "epoch": 0.5584515561420873, "grad_norm": 0.28287273645401, "learning_rate": 8e-05, "loss": 1.7151, "step": 2539 }, { "epoch": 0.5586715055537227, "grad_norm": 0.30313780903816223, "learning_rate": 8e-05, "loss": 1.6501, "step": 2540 }, { "epoch": 0.558891454965358, "grad_norm": 0.28195586800575256, "learning_rate": 8e-05, "loss": 1.7281, "step": 2541 }, { "epoch": 0.5591114043769932, "grad_norm": 0.2734014391899109, "learning_rate": 8e-05, "loss": 1.6504, "step": 2542 }, { "epoch": 0.5593313537886286, "grad_norm": 0.28178513050079346, "learning_rate": 8e-05, "loss": 1.5946, "step": 2543 }, { "epoch": 0.5595513032002639, "grad_norm": 0.2800062894821167, "learning_rate": 8e-05, "loss": 1.7498, "step": 2544 }, { "epoch": 0.5597712526118993, "grad_norm": 0.28368762135505676, "learning_rate": 8e-05, "loss": 1.6732, "step": 2545 }, { "epoch": 0.5599912020235346, "grad_norm": 0.3069396913051605, "learning_rate": 8e-05, "loss": 1.7647, "step": 2546 }, { "epoch": 0.5602111514351699, "grad_norm": 0.27336394786834717, "learning_rate": 8e-05, "loss": 1.4557, "step": 2547 }, { "epoch": 0.5604311008468053, "grad_norm": 0.28363245725631714, "learning_rate": 8e-05, "loss": 1.7301, "step": 2548 }, { "epoch": 0.5606510502584405, "grad_norm": 0.3097067177295685, "learning_rate": 8e-05, "loss": 1.7322, "step": 2549 }, { "epoch": 0.5608709996700759, "grad_norm": 0.28125154972076416, "learning_rate": 8e-05, "loss": 1.5551, "step": 2550 }, { "epoch": 0.5610909490817112, "grad_norm": 0.3111821413040161, "learning_rate": 8e-05, "loss": 1.8965, "step": 2551 }, { "epoch": 0.5613108984933465, "grad_norm": 0.2920529842376709, "learning_rate": 8e-05, "loss": 1.6835, "step": 2552 }, { "epoch": 0.5615308479049819, "grad_norm": 0.27278631925582886, "learning_rate": 8e-05, "loss": 1.7175, "step": 2553 }, { "epoch": 0.5617507973166171, "grad_norm": 0.2742355763912201, "learning_rate": 8e-05, "loss": 1.6745, "step": 2554 }, { "epoch": 0.5619707467282525, "grad_norm": 0.2675003707408905, "learning_rate": 8e-05, "loss": 1.5771, "step": 2555 }, { "epoch": 0.5621906961398878, "grad_norm": 0.2805350422859192, "learning_rate": 8e-05, "loss": 1.7503, "step": 2556 }, { "epoch": 0.5624106455515232, "grad_norm": 0.27205830812454224, "learning_rate": 8e-05, "loss": 1.6591, "step": 2557 }, { "epoch": 0.5626305949631585, "grad_norm": 0.26984983682632446, "learning_rate": 8e-05, "loss": 1.6351, "step": 2558 }, { "epoch": 0.5628505443747938, "grad_norm": 0.3067481517791748, "learning_rate": 8e-05, "loss": 1.5304, "step": 2559 }, { "epoch": 0.5630704937864291, "grad_norm": 0.28945624828338623, "learning_rate": 8e-05, "loss": 1.7099, "step": 2560 }, { "epoch": 0.5632904431980644, "grad_norm": 0.269144743680954, "learning_rate": 8e-05, "loss": 1.557, "step": 2561 }, { "epoch": 0.5635103926096998, "grad_norm": 0.329520583152771, "learning_rate": 8e-05, "loss": 1.7867, "step": 2562 }, { "epoch": 0.5637303420213351, "grad_norm": 0.35944700241088867, "learning_rate": 8e-05, "loss": 1.8146, "step": 2563 }, { "epoch": 0.5639502914329704, "grad_norm": 0.30693116784095764, "learning_rate": 8e-05, "loss": 1.7709, "step": 2564 }, { "epoch": 0.5641702408446058, "grad_norm": 0.31814631819725037, "learning_rate": 8e-05, "loss": 1.8679, "step": 2565 }, { "epoch": 0.564390190256241, "grad_norm": 0.2988479435443878, "learning_rate": 8e-05, "loss": 1.7675, "step": 2566 }, { "epoch": 0.5646101396678764, "grad_norm": 0.2955850064754486, "learning_rate": 8e-05, "loss": 1.7082, "step": 2567 }, { "epoch": 0.5648300890795117, "grad_norm": 0.27773404121398926, "learning_rate": 8e-05, "loss": 1.5378, "step": 2568 }, { "epoch": 0.565050038491147, "grad_norm": 0.2847524583339691, "learning_rate": 8e-05, "loss": 1.792, "step": 2569 }, { "epoch": 0.5652699879027824, "grad_norm": 0.29024967551231384, "learning_rate": 8e-05, "loss": 1.6185, "step": 2570 }, { "epoch": 0.5654899373144177, "grad_norm": 0.27534323930740356, "learning_rate": 8e-05, "loss": 1.7044, "step": 2571 }, { "epoch": 0.565709886726053, "grad_norm": 0.28059902787208557, "learning_rate": 8e-05, "loss": 1.7199, "step": 2572 }, { "epoch": 0.5659298361376883, "grad_norm": 0.29140958189964294, "learning_rate": 8e-05, "loss": 1.6534, "step": 2573 }, { "epoch": 0.5661497855493236, "grad_norm": 0.303821861743927, "learning_rate": 8e-05, "loss": 1.7163, "step": 2574 }, { "epoch": 0.566369734960959, "grad_norm": 0.3073093295097351, "learning_rate": 8e-05, "loss": 1.7885, "step": 2575 }, { "epoch": 0.5665896843725943, "grad_norm": 0.2976214289665222, "learning_rate": 8e-05, "loss": 1.7059, "step": 2576 }, { "epoch": 0.5668096337842297, "grad_norm": 0.3081284761428833, "learning_rate": 8e-05, "loss": 1.5529, "step": 2577 }, { "epoch": 0.5670295831958649, "grad_norm": 0.2893354594707489, "learning_rate": 8e-05, "loss": 1.8564, "step": 2578 }, { "epoch": 0.5672495326075003, "grad_norm": 0.2904176115989685, "learning_rate": 8e-05, "loss": 1.6903, "step": 2579 }, { "epoch": 0.5674694820191356, "grad_norm": 0.2907819449901581, "learning_rate": 8e-05, "loss": 1.5663, "step": 2580 }, { "epoch": 0.5676894314307709, "grad_norm": 0.27938172221183777, "learning_rate": 8e-05, "loss": 1.6031, "step": 2581 }, { "epoch": 0.5679093808424063, "grad_norm": 0.28864786028862, "learning_rate": 8e-05, "loss": 1.666, "step": 2582 }, { "epoch": 0.5681293302540416, "grad_norm": 0.29587891697883606, "learning_rate": 8e-05, "loss": 1.7545, "step": 2583 }, { "epoch": 0.5683492796656769, "grad_norm": 0.26541203260421753, "learning_rate": 8e-05, "loss": 1.6059, "step": 2584 }, { "epoch": 0.5685692290773122, "grad_norm": 0.2819576561450958, "learning_rate": 8e-05, "loss": 1.7227, "step": 2585 }, { "epoch": 0.5687891784889475, "grad_norm": 0.2920463979244232, "learning_rate": 8e-05, "loss": 1.7553, "step": 2586 }, { "epoch": 0.5690091279005829, "grad_norm": 0.29490089416503906, "learning_rate": 8e-05, "loss": 1.7117, "step": 2587 }, { "epoch": 0.5692290773122182, "grad_norm": 0.29847970604896545, "learning_rate": 8e-05, "loss": 1.8931, "step": 2588 }, { "epoch": 0.5694490267238536, "grad_norm": 0.28575995564460754, "learning_rate": 8e-05, "loss": 1.595, "step": 2589 }, { "epoch": 0.5696689761354888, "grad_norm": 0.28053271770477295, "learning_rate": 8e-05, "loss": 1.6089, "step": 2590 }, { "epoch": 0.5698889255471241, "grad_norm": 0.27538979053497314, "learning_rate": 8e-05, "loss": 1.7808, "step": 2591 }, { "epoch": 0.5701088749587595, "grad_norm": 0.2819748520851135, "learning_rate": 8e-05, "loss": 1.7355, "step": 2592 }, { "epoch": 0.5703288243703948, "grad_norm": 0.3023085594177246, "learning_rate": 8e-05, "loss": 1.759, "step": 2593 }, { "epoch": 0.5705487737820302, "grad_norm": 0.28369995951652527, "learning_rate": 8e-05, "loss": 1.7796, "step": 2594 }, { "epoch": 0.5707687231936655, "grad_norm": 0.27062156796455383, "learning_rate": 8e-05, "loss": 1.6206, "step": 2595 }, { "epoch": 0.5709886726053007, "grad_norm": 0.2928752303123474, "learning_rate": 8e-05, "loss": 1.7578, "step": 2596 }, { "epoch": 0.5712086220169361, "grad_norm": 0.28366369009017944, "learning_rate": 8e-05, "loss": 1.6367, "step": 2597 }, { "epoch": 0.5714285714285714, "grad_norm": 0.2794798016548157, "learning_rate": 8e-05, "loss": 1.7006, "step": 2598 }, { "epoch": 0.5716485208402068, "grad_norm": 0.278814435005188, "learning_rate": 8e-05, "loss": 1.6883, "step": 2599 }, { "epoch": 0.5718684702518421, "grad_norm": 0.28789058327674866, "learning_rate": 8e-05, "loss": 1.76, "step": 2600 }, { "epoch": 0.5720884196634775, "grad_norm": 0.289120614528656, "learning_rate": 8e-05, "loss": 1.5449, "step": 2601 }, { "epoch": 0.5723083690751127, "grad_norm": 0.27491265535354614, "learning_rate": 8e-05, "loss": 1.6287, "step": 2602 }, { "epoch": 0.572528318486748, "grad_norm": 0.2837536931037903, "learning_rate": 8e-05, "loss": 1.6618, "step": 2603 }, { "epoch": 0.5727482678983834, "grad_norm": 0.27386194467544556, "learning_rate": 8e-05, "loss": 1.5815, "step": 2604 }, { "epoch": 0.5729682173100187, "grad_norm": 0.2818918228149414, "learning_rate": 8e-05, "loss": 1.511, "step": 2605 }, { "epoch": 0.5731881667216541, "grad_norm": 0.29329514503479004, "learning_rate": 8e-05, "loss": 1.5494, "step": 2606 }, { "epoch": 0.5734081161332893, "grad_norm": 0.29942408204078674, "learning_rate": 8e-05, "loss": 1.8049, "step": 2607 }, { "epoch": 0.5736280655449246, "grad_norm": 0.30527159571647644, "learning_rate": 8e-05, "loss": 1.8735, "step": 2608 }, { "epoch": 0.57384801495656, "grad_norm": 0.2842453122138977, "learning_rate": 8e-05, "loss": 1.7124, "step": 2609 }, { "epoch": 0.5740679643681953, "grad_norm": 0.3305295407772064, "learning_rate": 8e-05, "loss": 1.6555, "step": 2610 }, { "epoch": 0.5742879137798307, "grad_norm": 0.28134140372276306, "learning_rate": 8e-05, "loss": 1.639, "step": 2611 }, { "epoch": 0.574507863191466, "grad_norm": 0.2862444818019867, "learning_rate": 8e-05, "loss": 1.7949, "step": 2612 }, { "epoch": 0.5747278126031012, "grad_norm": 0.3089071214199066, "learning_rate": 8e-05, "loss": 1.7782, "step": 2613 }, { "epoch": 0.5749477620147366, "grad_norm": 0.3113284111022949, "learning_rate": 8e-05, "loss": 1.5588, "step": 2614 }, { "epoch": 0.5751677114263719, "grad_norm": 0.2865052819252014, "learning_rate": 8e-05, "loss": 1.6919, "step": 2615 }, { "epoch": 0.5753876608380073, "grad_norm": 0.26997220516204834, "learning_rate": 8e-05, "loss": 1.6853, "step": 2616 }, { "epoch": 0.5756076102496426, "grad_norm": 0.3056239187717438, "learning_rate": 8e-05, "loss": 1.7971, "step": 2617 }, { "epoch": 0.5758275596612779, "grad_norm": 0.3041035234928131, "learning_rate": 8e-05, "loss": 1.7889, "step": 2618 }, { "epoch": 0.5760475090729132, "grad_norm": 0.2829764187335968, "learning_rate": 8e-05, "loss": 1.6371, "step": 2619 }, { "epoch": 0.5762674584845485, "grad_norm": 0.29050111770629883, "learning_rate": 8e-05, "loss": 1.8909, "step": 2620 }, { "epoch": 0.5764874078961839, "grad_norm": 0.29888811707496643, "learning_rate": 8e-05, "loss": 1.6974, "step": 2621 }, { "epoch": 0.5767073573078192, "grad_norm": 0.3193587362766266, "learning_rate": 8e-05, "loss": 1.7083, "step": 2622 }, { "epoch": 0.5769273067194546, "grad_norm": 0.2855699360370636, "learning_rate": 8e-05, "loss": 1.6106, "step": 2623 }, { "epoch": 0.5771472561310899, "grad_norm": 0.29608815908432007, "learning_rate": 8e-05, "loss": 1.7813, "step": 2624 }, { "epoch": 0.5773672055427251, "grad_norm": 0.2846873700618744, "learning_rate": 8e-05, "loss": 1.6164, "step": 2625 }, { "epoch": 0.5775871549543605, "grad_norm": 0.3074873983860016, "learning_rate": 8e-05, "loss": 1.7278, "step": 2626 }, { "epoch": 0.5778071043659958, "grad_norm": 0.3016159236431122, "learning_rate": 8e-05, "loss": 1.6286, "step": 2627 }, { "epoch": 0.5780270537776312, "grad_norm": 0.28926798701286316, "learning_rate": 8e-05, "loss": 1.746, "step": 2628 }, { "epoch": 0.5782470031892665, "grad_norm": 0.3222711682319641, "learning_rate": 8e-05, "loss": 1.8108, "step": 2629 }, { "epoch": 0.5784669526009018, "grad_norm": 0.30052945017814636, "learning_rate": 8e-05, "loss": 1.7374, "step": 2630 }, { "epoch": 0.5786869020125371, "grad_norm": 0.2880706787109375, "learning_rate": 8e-05, "loss": 1.627, "step": 2631 }, { "epoch": 0.5789068514241724, "grad_norm": 0.30028629302978516, "learning_rate": 8e-05, "loss": 1.8345, "step": 2632 }, { "epoch": 0.5791268008358078, "grad_norm": 0.3164263665676117, "learning_rate": 8e-05, "loss": 1.9713, "step": 2633 }, { "epoch": 0.5793467502474431, "grad_norm": 0.294114887714386, "learning_rate": 8e-05, "loss": 1.6083, "step": 2634 }, { "epoch": 0.5795666996590784, "grad_norm": 0.31809002161026, "learning_rate": 8e-05, "loss": 1.794, "step": 2635 }, { "epoch": 0.5797866490707138, "grad_norm": 0.3005049526691437, "learning_rate": 8e-05, "loss": 1.469, "step": 2636 }, { "epoch": 0.580006598482349, "grad_norm": 0.2874310314655304, "learning_rate": 8e-05, "loss": 1.7345, "step": 2637 }, { "epoch": 0.5802265478939844, "grad_norm": 0.295523077249527, "learning_rate": 8e-05, "loss": 1.6995, "step": 2638 }, { "epoch": 0.5804464973056197, "grad_norm": 0.29120928049087524, "learning_rate": 8e-05, "loss": 1.7736, "step": 2639 }, { "epoch": 0.580666446717255, "grad_norm": 0.2916790246963501, "learning_rate": 8e-05, "loss": 1.74, "step": 2640 }, { "epoch": 0.5808863961288904, "grad_norm": 0.285230427980423, "learning_rate": 8e-05, "loss": 1.7685, "step": 2641 }, { "epoch": 0.5811063455405256, "grad_norm": 0.2743189334869385, "learning_rate": 8e-05, "loss": 1.6751, "step": 2642 }, { "epoch": 0.581326294952161, "grad_norm": 0.2997332811355591, "learning_rate": 8e-05, "loss": 1.5959, "step": 2643 }, { "epoch": 0.5815462443637963, "grad_norm": 0.28394201397895813, "learning_rate": 8e-05, "loss": 1.6288, "step": 2644 }, { "epoch": 0.5817661937754317, "grad_norm": 0.2787470519542694, "learning_rate": 8e-05, "loss": 1.7496, "step": 2645 }, { "epoch": 0.581986143187067, "grad_norm": 0.2853599488735199, "learning_rate": 8e-05, "loss": 1.6439, "step": 2646 }, { "epoch": 0.5822060925987023, "grad_norm": 0.2939299941062927, "learning_rate": 8e-05, "loss": 1.7293, "step": 2647 }, { "epoch": 0.5824260420103377, "grad_norm": 0.27831408381462097, "learning_rate": 8e-05, "loss": 1.6748, "step": 2648 }, { "epoch": 0.5826459914219729, "grad_norm": 0.296762615442276, "learning_rate": 8e-05, "loss": 1.6735, "step": 2649 }, { "epoch": 0.5828659408336083, "grad_norm": 0.27961719036102295, "learning_rate": 8e-05, "loss": 1.6837, "step": 2650 }, { "epoch": 0.5830858902452436, "grad_norm": 0.27915704250335693, "learning_rate": 8e-05, "loss": 1.6745, "step": 2651 }, { "epoch": 0.5833058396568789, "grad_norm": 0.273799329996109, "learning_rate": 8e-05, "loss": 1.5609, "step": 2652 }, { "epoch": 0.5835257890685143, "grad_norm": 0.287383109331131, "learning_rate": 8e-05, "loss": 1.7569, "step": 2653 }, { "epoch": 0.5837457384801495, "grad_norm": 0.27745500206947327, "learning_rate": 8e-05, "loss": 1.6093, "step": 2654 }, { "epoch": 0.5839656878917849, "grad_norm": 0.2954557240009308, "learning_rate": 8e-05, "loss": 1.788, "step": 2655 }, { "epoch": 0.5841856373034202, "grad_norm": 0.28464850783348083, "learning_rate": 8e-05, "loss": 1.7079, "step": 2656 }, { "epoch": 0.5844055867150555, "grad_norm": 0.27475497126579285, "learning_rate": 8e-05, "loss": 1.6137, "step": 2657 }, { "epoch": 0.5846255361266909, "grad_norm": 0.27928462624549866, "learning_rate": 8e-05, "loss": 1.5776, "step": 2658 }, { "epoch": 0.5848454855383262, "grad_norm": 0.2889251708984375, "learning_rate": 8e-05, "loss": 1.7871, "step": 2659 }, { "epoch": 0.5850654349499615, "grad_norm": 0.29489466547966003, "learning_rate": 8e-05, "loss": 1.7299, "step": 2660 }, { "epoch": 0.5852853843615968, "grad_norm": 0.27761825919151306, "learning_rate": 8e-05, "loss": 1.6772, "step": 2661 }, { "epoch": 0.5855053337732321, "grad_norm": 0.2886674702167511, "learning_rate": 8e-05, "loss": 1.6718, "step": 2662 }, { "epoch": 0.5857252831848675, "grad_norm": 0.2736080586910248, "learning_rate": 8e-05, "loss": 1.5834, "step": 2663 }, { "epoch": 0.5859452325965028, "grad_norm": 0.29493847489356995, "learning_rate": 8e-05, "loss": 1.7123, "step": 2664 }, { "epoch": 0.5861651820081382, "grad_norm": 0.2919282615184784, "learning_rate": 8e-05, "loss": 1.7192, "step": 2665 }, { "epoch": 0.5863851314197734, "grad_norm": 0.2883647680282593, "learning_rate": 8e-05, "loss": 1.6271, "step": 2666 }, { "epoch": 0.5866050808314087, "grad_norm": 0.2852446734905243, "learning_rate": 8e-05, "loss": 1.7295, "step": 2667 }, { "epoch": 0.5868250302430441, "grad_norm": 0.3113778531551361, "learning_rate": 8e-05, "loss": 1.6605, "step": 2668 }, { "epoch": 0.5870449796546794, "grad_norm": 0.2629379630088806, "learning_rate": 8e-05, "loss": 1.3457, "step": 2669 }, { "epoch": 0.5872649290663148, "grad_norm": 0.28648287057876587, "learning_rate": 8e-05, "loss": 1.7137, "step": 2670 }, { "epoch": 0.5874848784779501, "grad_norm": 0.30140426754951477, "learning_rate": 8e-05, "loss": 1.7612, "step": 2671 }, { "epoch": 0.5877048278895854, "grad_norm": 0.29059261083602905, "learning_rate": 8e-05, "loss": 1.8526, "step": 2672 }, { "epoch": 0.5879247773012207, "grad_norm": 0.2913878560066223, "learning_rate": 8e-05, "loss": 1.7214, "step": 2673 }, { "epoch": 0.588144726712856, "grad_norm": 0.3046487271785736, "learning_rate": 8e-05, "loss": 1.8342, "step": 2674 }, { "epoch": 0.5883646761244914, "grad_norm": 0.2699670195579529, "learning_rate": 8e-05, "loss": 1.6057, "step": 2675 }, { "epoch": 0.5885846255361267, "grad_norm": 0.2722747027873993, "learning_rate": 8e-05, "loss": 1.5067, "step": 2676 }, { "epoch": 0.5888045749477621, "grad_norm": 0.27758973836898804, "learning_rate": 8e-05, "loss": 1.613, "step": 2677 }, { "epoch": 0.5890245243593973, "grad_norm": 0.30234992504119873, "learning_rate": 8e-05, "loss": 1.7266, "step": 2678 }, { "epoch": 0.5892444737710326, "grad_norm": 0.3146234452724457, "learning_rate": 8e-05, "loss": 1.6588, "step": 2679 }, { "epoch": 0.589464423182668, "grad_norm": 0.2867683470249176, "learning_rate": 8e-05, "loss": 1.6726, "step": 2680 }, { "epoch": 0.5896843725943033, "grad_norm": 0.28295040130615234, "learning_rate": 8e-05, "loss": 1.7338, "step": 2681 }, { "epoch": 0.5899043220059387, "grad_norm": 0.28655725717544556, "learning_rate": 8e-05, "loss": 1.6791, "step": 2682 }, { "epoch": 0.590124271417574, "grad_norm": 0.2897862493991852, "learning_rate": 8e-05, "loss": 1.7127, "step": 2683 }, { "epoch": 0.5903442208292092, "grad_norm": 0.278427392244339, "learning_rate": 8e-05, "loss": 1.7166, "step": 2684 }, { "epoch": 0.5905641702408446, "grad_norm": 0.28383758664131165, "learning_rate": 8e-05, "loss": 1.8498, "step": 2685 }, { "epoch": 0.5907841196524799, "grad_norm": 0.2690020501613617, "learning_rate": 8e-05, "loss": 1.5887, "step": 2686 }, { "epoch": 0.5910040690641153, "grad_norm": 0.2910546362400055, "learning_rate": 8e-05, "loss": 1.7525, "step": 2687 }, { "epoch": 0.5912240184757506, "grad_norm": 0.2932651937007904, "learning_rate": 8e-05, "loss": 1.6113, "step": 2688 }, { "epoch": 0.5914439678873858, "grad_norm": 0.275622695684433, "learning_rate": 8e-05, "loss": 1.6322, "step": 2689 }, { "epoch": 0.5916639172990212, "grad_norm": 0.2838039696216583, "learning_rate": 8e-05, "loss": 1.8021, "step": 2690 }, { "epoch": 0.5918838667106565, "grad_norm": 0.290005087852478, "learning_rate": 8e-05, "loss": 1.6378, "step": 2691 }, { "epoch": 0.5921038161222919, "grad_norm": 0.2730334401130676, "learning_rate": 8e-05, "loss": 1.6665, "step": 2692 }, { "epoch": 0.5923237655339272, "grad_norm": 0.27828192710876465, "learning_rate": 8e-05, "loss": 1.6905, "step": 2693 }, { "epoch": 0.5925437149455626, "grad_norm": 0.26481491327285767, "learning_rate": 8e-05, "loss": 1.5056, "step": 2694 }, { "epoch": 0.5927636643571979, "grad_norm": 0.2684583365917206, "learning_rate": 8e-05, "loss": 1.4181, "step": 2695 }, { "epoch": 0.5929836137688331, "grad_norm": 0.2848527431488037, "learning_rate": 8e-05, "loss": 1.6243, "step": 2696 }, { "epoch": 0.5932035631804685, "grad_norm": 0.2943567931652069, "learning_rate": 8e-05, "loss": 1.7296, "step": 2697 }, { "epoch": 0.5934235125921038, "grad_norm": 0.2790435552597046, "learning_rate": 8e-05, "loss": 1.611, "step": 2698 }, { "epoch": 0.5936434620037392, "grad_norm": 0.3020678460597992, "learning_rate": 8e-05, "loss": 1.8619, "step": 2699 }, { "epoch": 0.5938634114153745, "grad_norm": 0.2809624969959259, "learning_rate": 8e-05, "loss": 1.7253, "step": 2700 }, { "epoch": 0.5940833608270097, "grad_norm": 0.2655926048755646, "learning_rate": 8e-05, "loss": 1.5514, "step": 2701 }, { "epoch": 0.5943033102386451, "grad_norm": 0.28663522005081177, "learning_rate": 8e-05, "loss": 1.6708, "step": 2702 }, { "epoch": 0.5945232596502804, "grad_norm": 0.28419819474220276, "learning_rate": 8e-05, "loss": 1.6551, "step": 2703 }, { "epoch": 0.5947432090619158, "grad_norm": 0.29084041714668274, "learning_rate": 8e-05, "loss": 1.7509, "step": 2704 }, { "epoch": 0.5949631584735511, "grad_norm": 0.27892929315567017, "learning_rate": 8e-05, "loss": 1.6507, "step": 2705 }, { "epoch": 0.5951831078851864, "grad_norm": 0.29692748188972473, "learning_rate": 8e-05, "loss": 1.8667, "step": 2706 }, { "epoch": 0.5954030572968217, "grad_norm": 0.2867085933685303, "learning_rate": 8e-05, "loss": 1.6157, "step": 2707 }, { "epoch": 0.595623006708457, "grad_norm": 0.29867735505104065, "learning_rate": 8e-05, "loss": 1.7813, "step": 2708 }, { "epoch": 0.5958429561200924, "grad_norm": 0.28061944246292114, "learning_rate": 8e-05, "loss": 1.6157, "step": 2709 }, { "epoch": 0.5960629055317277, "grad_norm": 0.2807196080684662, "learning_rate": 8e-05, "loss": 1.5745, "step": 2710 }, { "epoch": 0.596282854943363, "grad_norm": 0.2854728698730469, "learning_rate": 8e-05, "loss": 1.78, "step": 2711 }, { "epoch": 0.5965028043549984, "grad_norm": 0.2980540990829468, "learning_rate": 8e-05, "loss": 1.8421, "step": 2712 }, { "epoch": 0.5967227537666336, "grad_norm": 0.2892910838127136, "learning_rate": 8e-05, "loss": 1.6555, "step": 2713 }, { "epoch": 0.596942703178269, "grad_norm": 0.2773078680038452, "learning_rate": 8e-05, "loss": 1.609, "step": 2714 }, { "epoch": 0.5971626525899043, "grad_norm": 0.29283806681632996, "learning_rate": 8e-05, "loss": 1.7398, "step": 2715 }, { "epoch": 0.5973826020015397, "grad_norm": 0.2872734069824219, "learning_rate": 8e-05, "loss": 1.6864, "step": 2716 }, { "epoch": 0.597602551413175, "grad_norm": 0.26770031452178955, "learning_rate": 8e-05, "loss": 1.5877, "step": 2717 }, { "epoch": 0.5978225008248103, "grad_norm": 0.2958748936653137, "learning_rate": 8e-05, "loss": 1.6565, "step": 2718 }, { "epoch": 0.5980424502364456, "grad_norm": 0.30203044414520264, "learning_rate": 8e-05, "loss": 1.7878, "step": 2719 }, { "epoch": 0.5982623996480809, "grad_norm": 0.29320842027664185, "learning_rate": 8e-05, "loss": 1.5706, "step": 2720 }, { "epoch": 0.5984823490597163, "grad_norm": 0.29835638403892517, "learning_rate": 8e-05, "loss": 1.7271, "step": 2721 }, { "epoch": 0.5987022984713516, "grad_norm": 0.36251741647720337, "learning_rate": 8e-05, "loss": 1.5832, "step": 2722 }, { "epoch": 0.5989222478829869, "grad_norm": 0.28875645995140076, "learning_rate": 8e-05, "loss": 1.8148, "step": 2723 }, { "epoch": 0.5991421972946223, "grad_norm": 0.27607399225234985, "learning_rate": 8e-05, "loss": 1.6024, "step": 2724 }, { "epoch": 0.5993621467062575, "grad_norm": 0.290351539850235, "learning_rate": 8e-05, "loss": 1.7287, "step": 2725 }, { "epoch": 0.5995820961178929, "grad_norm": 0.28432413935661316, "learning_rate": 8e-05, "loss": 1.8603, "step": 2726 }, { "epoch": 0.5998020455295282, "grad_norm": 0.2780609130859375, "learning_rate": 8e-05, "loss": 1.6839, "step": 2727 }, { "epoch": 0.6000219949411635, "grad_norm": 0.31952062249183655, "learning_rate": 8e-05, "loss": 1.585, "step": 2728 }, { "epoch": 0.6002419443527989, "grad_norm": 0.2631243169307709, "learning_rate": 8e-05, "loss": 1.6074, "step": 2729 }, { "epoch": 0.6004618937644342, "grad_norm": 0.28518691658973694, "learning_rate": 8e-05, "loss": 1.6944, "step": 2730 }, { "epoch": 0.6006818431760695, "grad_norm": 0.29021504521369934, "learning_rate": 8e-05, "loss": 1.5919, "step": 2731 }, { "epoch": 0.6009017925877048, "grad_norm": 0.2772546410560608, "learning_rate": 8e-05, "loss": 1.6372, "step": 2732 }, { "epoch": 0.6011217419993401, "grad_norm": 0.27938538789749146, "learning_rate": 8e-05, "loss": 1.7311, "step": 2733 }, { "epoch": 0.6013416914109755, "grad_norm": 0.2936658561229706, "learning_rate": 8e-05, "loss": 1.695, "step": 2734 }, { "epoch": 0.6015616408226108, "grad_norm": 0.2893039286136627, "learning_rate": 8e-05, "loss": 1.6837, "step": 2735 }, { "epoch": 0.6017815902342462, "grad_norm": 0.28634974360466003, "learning_rate": 8e-05, "loss": 1.6155, "step": 2736 }, { "epoch": 0.6020015396458814, "grad_norm": 0.2868409752845764, "learning_rate": 8e-05, "loss": 1.5901, "step": 2737 }, { "epoch": 0.6022214890575168, "grad_norm": 0.28888818621635437, "learning_rate": 8e-05, "loss": 1.6951, "step": 2738 }, { "epoch": 0.6024414384691521, "grad_norm": 0.2881872355937958, "learning_rate": 8e-05, "loss": 1.7648, "step": 2739 }, { "epoch": 0.6026613878807874, "grad_norm": 0.29601114988327026, "learning_rate": 8e-05, "loss": 1.6499, "step": 2740 }, { "epoch": 0.6028813372924228, "grad_norm": 0.28861135244369507, "learning_rate": 8e-05, "loss": 1.6896, "step": 2741 }, { "epoch": 0.603101286704058, "grad_norm": 0.30852892994880676, "learning_rate": 8e-05, "loss": 1.8539, "step": 2742 }, { "epoch": 0.6033212361156934, "grad_norm": 0.2659029960632324, "learning_rate": 8e-05, "loss": 1.5581, "step": 2743 }, { "epoch": 0.6035411855273287, "grad_norm": 0.2938629686832428, "learning_rate": 8e-05, "loss": 1.6893, "step": 2744 }, { "epoch": 0.603761134938964, "grad_norm": 0.3215024769306183, "learning_rate": 8e-05, "loss": 1.809, "step": 2745 }, { "epoch": 0.6039810843505994, "grad_norm": 0.3320122957229614, "learning_rate": 8e-05, "loss": 1.7137, "step": 2746 }, { "epoch": 0.6042010337622347, "grad_norm": 0.2901141047477722, "learning_rate": 8e-05, "loss": 1.7675, "step": 2747 }, { "epoch": 0.60442098317387, "grad_norm": 0.2905901074409485, "learning_rate": 8e-05, "loss": 1.7454, "step": 2748 }, { "epoch": 0.6046409325855053, "grad_norm": 0.2820628583431244, "learning_rate": 8e-05, "loss": 1.7143, "step": 2749 }, { "epoch": 0.6048608819971406, "grad_norm": 0.29754188656806946, "learning_rate": 8e-05, "loss": 1.6957, "step": 2750 }, { "epoch": 0.605080831408776, "grad_norm": 0.28644484281539917, "learning_rate": 8e-05, "loss": 1.864, "step": 2751 }, { "epoch": 0.6053007808204113, "grad_norm": 0.2816253900527954, "learning_rate": 8e-05, "loss": 1.6956, "step": 2752 }, { "epoch": 0.6055207302320467, "grad_norm": 0.27785420417785645, "learning_rate": 8e-05, "loss": 1.6618, "step": 2753 }, { "epoch": 0.605740679643682, "grad_norm": 0.2993432283401489, "learning_rate": 8e-05, "loss": 1.6782, "step": 2754 }, { "epoch": 0.6059606290553172, "grad_norm": 0.2837073802947998, "learning_rate": 8e-05, "loss": 1.6096, "step": 2755 }, { "epoch": 0.6061805784669526, "grad_norm": 0.2930501401424408, "learning_rate": 8e-05, "loss": 1.7036, "step": 2756 }, { "epoch": 0.6064005278785879, "grad_norm": 0.2830953299999237, "learning_rate": 8e-05, "loss": 1.7393, "step": 2757 }, { "epoch": 0.6066204772902233, "grad_norm": 0.3069010078907013, "learning_rate": 8e-05, "loss": 1.5636, "step": 2758 }, { "epoch": 0.6068404267018586, "grad_norm": 0.2761766314506531, "learning_rate": 8e-05, "loss": 1.7233, "step": 2759 }, { "epoch": 0.607060376113494, "grad_norm": 0.28254058957099915, "learning_rate": 8e-05, "loss": 1.7132, "step": 2760 }, { "epoch": 0.6072803255251292, "grad_norm": 0.27911651134490967, "learning_rate": 8e-05, "loss": 1.7782, "step": 2761 }, { "epoch": 0.6075002749367645, "grad_norm": 0.2875358462333679, "learning_rate": 8e-05, "loss": 1.6974, "step": 2762 }, { "epoch": 0.6077202243483999, "grad_norm": 0.28940457105636597, "learning_rate": 8e-05, "loss": 1.6274, "step": 2763 }, { "epoch": 0.6079401737600352, "grad_norm": 0.27163782715797424, "learning_rate": 8e-05, "loss": 1.6507, "step": 2764 }, { "epoch": 0.6081601231716706, "grad_norm": 0.2914412021636963, "learning_rate": 8e-05, "loss": 1.6826, "step": 2765 }, { "epoch": 0.6083800725833058, "grad_norm": 0.31414681673049927, "learning_rate": 8e-05, "loss": 1.8466, "step": 2766 }, { "epoch": 0.6086000219949411, "grad_norm": 0.3015105426311493, "learning_rate": 8e-05, "loss": 1.599, "step": 2767 }, { "epoch": 0.6088199714065765, "grad_norm": 0.27743127942085266, "learning_rate": 8e-05, "loss": 1.5278, "step": 2768 }, { "epoch": 0.6090399208182118, "grad_norm": 0.2868049442768097, "learning_rate": 8e-05, "loss": 1.7008, "step": 2769 }, { "epoch": 0.6092598702298472, "grad_norm": 0.2832272946834564, "learning_rate": 8e-05, "loss": 1.6368, "step": 2770 }, { "epoch": 0.6094798196414825, "grad_norm": 0.28054770827293396, "learning_rate": 8e-05, "loss": 1.6475, "step": 2771 }, { "epoch": 0.6096997690531177, "grad_norm": 0.28185421228408813, "learning_rate": 8e-05, "loss": 1.7731, "step": 2772 }, { "epoch": 0.6099197184647531, "grad_norm": 0.2819845676422119, "learning_rate": 8e-05, "loss": 1.6824, "step": 2773 }, { "epoch": 0.6101396678763884, "grad_norm": 0.2764539420604706, "learning_rate": 8e-05, "loss": 1.7001, "step": 2774 }, { "epoch": 0.6103596172880238, "grad_norm": 0.30475977063179016, "learning_rate": 8e-05, "loss": 1.8297, "step": 2775 }, { "epoch": 0.6105795666996591, "grad_norm": 0.2848237454891205, "learning_rate": 8e-05, "loss": 1.7453, "step": 2776 }, { "epoch": 0.6107995161112943, "grad_norm": 0.28268033266067505, "learning_rate": 8e-05, "loss": 1.5241, "step": 2777 }, { "epoch": 0.6110194655229297, "grad_norm": 0.27673062682151794, "learning_rate": 8e-05, "loss": 1.5273, "step": 2778 }, { "epoch": 0.611239414934565, "grad_norm": 0.28202882409095764, "learning_rate": 8e-05, "loss": 1.5769, "step": 2779 }, { "epoch": 0.6114593643462004, "grad_norm": 0.28480303287506104, "learning_rate": 8e-05, "loss": 1.7456, "step": 2780 }, { "epoch": 0.6116793137578357, "grad_norm": 0.3028055727481842, "learning_rate": 8e-05, "loss": 1.6652, "step": 2781 }, { "epoch": 0.6118992631694711, "grad_norm": 0.28677237033843994, "learning_rate": 8e-05, "loss": 1.6877, "step": 2782 }, { "epoch": 0.6121192125811064, "grad_norm": 0.3057413399219513, "learning_rate": 8e-05, "loss": 1.9811, "step": 2783 }, { "epoch": 0.6123391619927416, "grad_norm": 0.2802276313304901, "learning_rate": 8e-05, "loss": 1.5965, "step": 2784 }, { "epoch": 0.612559111404377, "grad_norm": 0.27934229373931885, "learning_rate": 8e-05, "loss": 1.5959, "step": 2785 }, { "epoch": 0.6127790608160123, "grad_norm": 0.2864493429660797, "learning_rate": 8e-05, "loss": 1.6289, "step": 2786 }, { "epoch": 0.6129990102276477, "grad_norm": 0.26668915152549744, "learning_rate": 8e-05, "loss": 1.4584, "step": 2787 }, { "epoch": 0.613218959639283, "grad_norm": 0.28092291951179504, "learning_rate": 8e-05, "loss": 1.6641, "step": 2788 }, { "epoch": 0.6134389090509182, "grad_norm": 0.2933676242828369, "learning_rate": 8e-05, "loss": 1.9453, "step": 2789 }, { "epoch": 0.6136588584625536, "grad_norm": 0.31618431210517883, "learning_rate": 8e-05, "loss": 1.594, "step": 2790 }, { "epoch": 0.6138788078741889, "grad_norm": 0.28090760111808777, "learning_rate": 8e-05, "loss": 1.7531, "step": 2791 }, { "epoch": 0.6140987572858243, "grad_norm": 0.3137405216693878, "learning_rate": 8e-05, "loss": 1.7243, "step": 2792 }, { "epoch": 0.6143187066974596, "grad_norm": 0.2949986755847931, "learning_rate": 8e-05, "loss": 1.7581, "step": 2793 }, { "epoch": 0.6145386561090949, "grad_norm": 0.28396037220954895, "learning_rate": 8e-05, "loss": 1.6995, "step": 2794 }, { "epoch": 0.6147586055207303, "grad_norm": 0.26976051926612854, "learning_rate": 8e-05, "loss": 1.654, "step": 2795 }, { "epoch": 0.6149785549323655, "grad_norm": 0.27323633432388306, "learning_rate": 8e-05, "loss": 1.6944, "step": 2796 }, { "epoch": 0.6151985043440009, "grad_norm": 0.29849350452423096, "learning_rate": 8e-05, "loss": 1.6127, "step": 2797 }, { "epoch": 0.6154184537556362, "grad_norm": 0.28575918078422546, "learning_rate": 8e-05, "loss": 1.7579, "step": 2798 }, { "epoch": 0.6156384031672715, "grad_norm": 0.26723456382751465, "learning_rate": 8e-05, "loss": 1.5461, "step": 2799 }, { "epoch": 0.6158583525789069, "grad_norm": 0.29076528549194336, "learning_rate": 8e-05, "loss": 1.7001, "step": 2800 }, { "epoch": 0.6160783019905421, "grad_norm": 0.27913492918014526, "learning_rate": 8e-05, "loss": 1.3878, "step": 2801 }, { "epoch": 0.6162982514021775, "grad_norm": 0.2841816246509552, "learning_rate": 8e-05, "loss": 1.6871, "step": 2802 }, { "epoch": 0.6165182008138128, "grad_norm": 0.26845458149909973, "learning_rate": 8e-05, "loss": 1.5518, "step": 2803 }, { "epoch": 0.6167381502254482, "grad_norm": 0.30308809876441956, "learning_rate": 8e-05, "loss": 1.7735, "step": 2804 }, { "epoch": 0.6169580996370835, "grad_norm": 0.2812938690185547, "learning_rate": 8e-05, "loss": 1.7043, "step": 2805 }, { "epoch": 0.6171780490487188, "grad_norm": 0.27101054787635803, "learning_rate": 8e-05, "loss": 1.6156, "step": 2806 }, { "epoch": 0.6173979984603541, "grad_norm": 0.2900649607181549, "learning_rate": 8e-05, "loss": 1.7119, "step": 2807 }, { "epoch": 0.6176179478719894, "grad_norm": 0.3011523187160492, "learning_rate": 8e-05, "loss": 1.7085, "step": 2808 }, { "epoch": 0.6178378972836248, "grad_norm": 0.2845047116279602, "learning_rate": 8e-05, "loss": 1.6691, "step": 2809 }, { "epoch": 0.6180578466952601, "grad_norm": 0.31060662865638733, "learning_rate": 8e-05, "loss": 1.831, "step": 2810 }, { "epoch": 0.6182777961068954, "grad_norm": 0.27987706661224365, "learning_rate": 8e-05, "loss": 1.4881, "step": 2811 }, { "epoch": 0.6184977455185308, "grad_norm": 0.3197080194950104, "learning_rate": 8e-05, "loss": 1.7112, "step": 2812 }, { "epoch": 0.618717694930166, "grad_norm": 0.31402066349983215, "learning_rate": 8e-05, "loss": 1.6535, "step": 2813 }, { "epoch": 0.6189376443418014, "grad_norm": 0.303529292345047, "learning_rate": 8e-05, "loss": 1.6563, "step": 2814 }, { "epoch": 0.6191575937534367, "grad_norm": 0.26674556732177734, "learning_rate": 8e-05, "loss": 1.5202, "step": 2815 }, { "epoch": 0.619377543165072, "grad_norm": 0.30466997623443604, "learning_rate": 8e-05, "loss": 1.6014, "step": 2816 }, { "epoch": 0.6195974925767074, "grad_norm": 0.2991195619106293, "learning_rate": 8e-05, "loss": 1.765, "step": 2817 }, { "epoch": 0.6198174419883427, "grad_norm": 0.30000337958335876, "learning_rate": 8e-05, "loss": 1.7794, "step": 2818 }, { "epoch": 0.620037391399978, "grad_norm": 0.29237842559814453, "learning_rate": 8e-05, "loss": 1.8753, "step": 2819 }, { "epoch": 0.6202573408116133, "grad_norm": 0.2896344065666199, "learning_rate": 8e-05, "loss": 1.6137, "step": 2820 }, { "epoch": 0.6204772902232486, "grad_norm": 0.34269601106643677, "learning_rate": 8e-05, "loss": 1.7693, "step": 2821 }, { "epoch": 0.620697239634884, "grad_norm": 0.30044153332710266, "learning_rate": 8e-05, "loss": 1.7286, "step": 2822 }, { "epoch": 0.6209171890465193, "grad_norm": 0.2616185247898102, "learning_rate": 8e-05, "loss": 1.5373, "step": 2823 }, { "epoch": 0.6211371384581547, "grad_norm": 0.3217238485813141, "learning_rate": 8e-05, "loss": 1.7681, "step": 2824 }, { "epoch": 0.6213570878697899, "grad_norm": 0.284446120262146, "learning_rate": 8e-05, "loss": 1.5597, "step": 2825 }, { "epoch": 0.6215770372814253, "grad_norm": 0.28698036074638367, "learning_rate": 8e-05, "loss": 1.6844, "step": 2826 }, { "epoch": 0.6217969866930606, "grad_norm": 0.2828524708747864, "learning_rate": 8e-05, "loss": 1.7142, "step": 2827 }, { "epoch": 0.6220169361046959, "grad_norm": 0.3004125952720642, "learning_rate": 8e-05, "loss": 1.6812, "step": 2828 }, { "epoch": 0.6222368855163313, "grad_norm": 0.30438825488090515, "learning_rate": 8e-05, "loss": 1.6696, "step": 2829 }, { "epoch": 0.6224568349279666, "grad_norm": 0.2654431164264679, "learning_rate": 8e-05, "loss": 1.5656, "step": 2830 }, { "epoch": 0.6226767843396019, "grad_norm": 0.28561410307884216, "learning_rate": 8e-05, "loss": 1.5924, "step": 2831 }, { "epoch": 0.6228967337512372, "grad_norm": 0.29075953364372253, "learning_rate": 8e-05, "loss": 1.6026, "step": 2832 }, { "epoch": 0.6231166831628725, "grad_norm": 0.3002355098724365, "learning_rate": 8e-05, "loss": 1.7783, "step": 2833 }, { "epoch": 0.6233366325745079, "grad_norm": 0.2757151424884796, "learning_rate": 8e-05, "loss": 1.6404, "step": 2834 }, { "epoch": 0.6235565819861432, "grad_norm": 0.28108781576156616, "learning_rate": 8e-05, "loss": 1.6727, "step": 2835 }, { "epoch": 0.6237765313977786, "grad_norm": 0.31818297505378723, "learning_rate": 8e-05, "loss": 1.8859, "step": 2836 }, { "epoch": 0.6239964808094138, "grad_norm": 0.30283015966415405, "learning_rate": 8e-05, "loss": 1.6967, "step": 2837 }, { "epoch": 0.6242164302210491, "grad_norm": 0.28991082310676575, "learning_rate": 8e-05, "loss": 1.7564, "step": 2838 }, { "epoch": 0.6244363796326845, "grad_norm": 0.27985113859176636, "learning_rate": 8e-05, "loss": 1.645, "step": 2839 }, { "epoch": 0.6246563290443198, "grad_norm": 0.28599318861961365, "learning_rate": 8e-05, "loss": 1.7069, "step": 2840 }, { "epoch": 0.6248762784559552, "grad_norm": 0.291962593793869, "learning_rate": 8e-05, "loss": 1.7303, "step": 2841 }, { "epoch": 0.6250962278675904, "grad_norm": 0.2977605164051056, "learning_rate": 8e-05, "loss": 1.6251, "step": 2842 }, { "epoch": 0.6253161772792257, "grad_norm": 0.280979186296463, "learning_rate": 8e-05, "loss": 1.6132, "step": 2843 }, { "epoch": 0.6255361266908611, "grad_norm": 0.30565154552459717, "learning_rate": 8e-05, "loss": 1.8351, "step": 2844 }, { "epoch": 0.6257560761024964, "grad_norm": 0.2870398759841919, "learning_rate": 8e-05, "loss": 1.7169, "step": 2845 }, { "epoch": 0.6259760255141318, "grad_norm": 0.2740568518638611, "learning_rate": 8e-05, "loss": 1.6666, "step": 2846 }, { "epoch": 0.6261959749257671, "grad_norm": 0.27255693078041077, "learning_rate": 8e-05, "loss": 1.6112, "step": 2847 }, { "epoch": 0.6264159243374025, "grad_norm": 0.2785317003726959, "learning_rate": 8e-05, "loss": 1.6532, "step": 2848 }, { "epoch": 0.6266358737490377, "grad_norm": 0.2979902923107147, "learning_rate": 8e-05, "loss": 1.7981, "step": 2849 }, { "epoch": 0.626855823160673, "grad_norm": 0.29625701904296875, "learning_rate": 8e-05, "loss": 1.7212, "step": 2850 }, { "epoch": 0.6270757725723084, "grad_norm": 0.2768239676952362, "learning_rate": 8e-05, "loss": 1.6358, "step": 2851 }, { "epoch": 0.6272957219839437, "grad_norm": 0.2931036055088043, "learning_rate": 8e-05, "loss": 1.7728, "step": 2852 }, { "epoch": 0.6275156713955791, "grad_norm": 0.2883271872997284, "learning_rate": 8e-05, "loss": 1.6403, "step": 2853 }, { "epoch": 0.6277356208072143, "grad_norm": 0.31137219071388245, "learning_rate": 8e-05, "loss": 1.6454, "step": 2854 }, { "epoch": 0.6279555702188496, "grad_norm": 0.3026840388774872, "learning_rate": 8e-05, "loss": 1.6534, "step": 2855 }, { "epoch": 0.628175519630485, "grad_norm": 0.2950657606124878, "learning_rate": 8e-05, "loss": 1.7298, "step": 2856 }, { "epoch": 0.6283954690421203, "grad_norm": 0.29347553849220276, "learning_rate": 8e-05, "loss": 1.58, "step": 2857 }, { "epoch": 0.6286154184537557, "grad_norm": 0.28075262904167175, "learning_rate": 8e-05, "loss": 1.6262, "step": 2858 }, { "epoch": 0.628835367865391, "grad_norm": 0.26556506752967834, "learning_rate": 8e-05, "loss": 1.5666, "step": 2859 }, { "epoch": 0.6290553172770262, "grad_norm": 0.28918468952178955, "learning_rate": 8e-05, "loss": 1.6918, "step": 2860 }, { "epoch": 0.6292752666886616, "grad_norm": 0.2816839814186096, "learning_rate": 8e-05, "loss": 1.6387, "step": 2861 }, { "epoch": 0.6294952161002969, "grad_norm": 0.2819633185863495, "learning_rate": 8e-05, "loss": 1.6945, "step": 2862 }, { "epoch": 0.6297151655119323, "grad_norm": 0.2847195863723755, "learning_rate": 8e-05, "loss": 1.744, "step": 2863 }, { "epoch": 0.6299351149235676, "grad_norm": 0.2706061899662018, "learning_rate": 8e-05, "loss": 1.6246, "step": 2864 }, { "epoch": 0.6301550643352029, "grad_norm": 0.281125545501709, "learning_rate": 8e-05, "loss": 1.49, "step": 2865 }, { "epoch": 0.6303750137468382, "grad_norm": 0.2861780822277069, "learning_rate": 8e-05, "loss": 1.7375, "step": 2866 }, { "epoch": 0.6305949631584735, "grad_norm": 0.2654918134212494, "learning_rate": 8e-05, "loss": 1.5997, "step": 2867 }, { "epoch": 0.6308149125701089, "grad_norm": 0.29169219732284546, "learning_rate": 8e-05, "loss": 1.8921, "step": 2868 }, { "epoch": 0.6310348619817442, "grad_norm": 0.2858426570892334, "learning_rate": 8e-05, "loss": 1.8006, "step": 2869 }, { "epoch": 0.6312548113933796, "grad_norm": 0.2712969183921814, "learning_rate": 8e-05, "loss": 1.5761, "step": 2870 }, { "epoch": 0.6314747608050149, "grad_norm": 0.28961536288261414, "learning_rate": 8e-05, "loss": 1.618, "step": 2871 }, { "epoch": 0.6316947102166501, "grad_norm": 0.2879860997200012, "learning_rate": 8e-05, "loss": 1.6116, "step": 2872 }, { "epoch": 0.6319146596282855, "grad_norm": 0.3009500801563263, "learning_rate": 8e-05, "loss": 1.8099, "step": 2873 }, { "epoch": 0.6321346090399208, "grad_norm": 0.3012961149215698, "learning_rate": 8e-05, "loss": 1.8123, "step": 2874 }, { "epoch": 0.6323545584515562, "grad_norm": 0.27382341027259827, "learning_rate": 8e-05, "loss": 1.5486, "step": 2875 }, { "epoch": 0.6325745078631915, "grad_norm": 0.45538756251335144, "learning_rate": 8e-05, "loss": 1.7382, "step": 2876 }, { "epoch": 0.6327944572748267, "grad_norm": 0.27454543113708496, "learning_rate": 8e-05, "loss": 1.6488, "step": 2877 }, { "epoch": 0.6330144066864621, "grad_norm": 0.28111204504966736, "learning_rate": 8e-05, "loss": 1.7262, "step": 2878 }, { "epoch": 0.6332343560980974, "grad_norm": 0.2855817675590515, "learning_rate": 8e-05, "loss": 1.6841, "step": 2879 }, { "epoch": 0.6334543055097328, "grad_norm": 0.3017145097255707, "learning_rate": 8e-05, "loss": 1.759, "step": 2880 }, { "epoch": 0.6336742549213681, "grad_norm": 0.27578651905059814, "learning_rate": 8e-05, "loss": 1.5341, "step": 2881 }, { "epoch": 0.6338942043330034, "grad_norm": 0.28522011637687683, "learning_rate": 8e-05, "loss": 1.5042, "step": 2882 }, { "epoch": 0.6341141537446388, "grad_norm": 0.28013676404953003, "learning_rate": 8e-05, "loss": 1.591, "step": 2883 }, { "epoch": 0.634334103156274, "grad_norm": 0.30440640449523926, "learning_rate": 8e-05, "loss": 1.6698, "step": 2884 }, { "epoch": 0.6345540525679094, "grad_norm": 0.28555527329444885, "learning_rate": 8e-05, "loss": 1.6199, "step": 2885 }, { "epoch": 0.6347740019795447, "grad_norm": 0.31451916694641113, "learning_rate": 8e-05, "loss": 1.6717, "step": 2886 }, { "epoch": 0.63499395139118, "grad_norm": 0.3116842806339264, "learning_rate": 8e-05, "loss": 1.7043, "step": 2887 }, { "epoch": 0.6352139008028154, "grad_norm": 0.30441299080848694, "learning_rate": 8e-05, "loss": 1.6953, "step": 2888 }, { "epoch": 0.6354338502144506, "grad_norm": 0.2890806496143341, "learning_rate": 8e-05, "loss": 1.7363, "step": 2889 }, { "epoch": 0.635653799626086, "grad_norm": 0.2715187072753906, "learning_rate": 8e-05, "loss": 1.5412, "step": 2890 }, { "epoch": 0.6358737490377213, "grad_norm": 0.32213905453681946, "learning_rate": 8e-05, "loss": 1.7735, "step": 2891 }, { "epoch": 0.6360936984493566, "grad_norm": 0.28850191831588745, "learning_rate": 8e-05, "loss": 1.5156, "step": 2892 }, { "epoch": 0.636313647860992, "grad_norm": 0.2934744358062744, "learning_rate": 8e-05, "loss": 1.6822, "step": 2893 }, { "epoch": 0.6365335972726273, "grad_norm": 0.29068851470947266, "learning_rate": 8e-05, "loss": 1.7723, "step": 2894 }, { "epoch": 0.6367535466842627, "grad_norm": 0.28490251302719116, "learning_rate": 8e-05, "loss": 1.7666, "step": 2895 }, { "epoch": 0.6369734960958979, "grad_norm": 0.28677448630332947, "learning_rate": 8e-05, "loss": 1.6792, "step": 2896 }, { "epoch": 0.6371934455075333, "grad_norm": 0.29424387216567993, "learning_rate": 8e-05, "loss": 1.6224, "step": 2897 }, { "epoch": 0.6374133949191686, "grad_norm": 0.2872456908226013, "learning_rate": 8e-05, "loss": 1.6677, "step": 2898 }, { "epoch": 0.6376333443308039, "grad_norm": 0.26886799931526184, "learning_rate": 8e-05, "loss": 1.5086, "step": 2899 }, { "epoch": 0.6378532937424393, "grad_norm": 0.2737233638763428, "learning_rate": 8e-05, "loss": 1.6832, "step": 2900 }, { "epoch": 0.6380732431540745, "grad_norm": 0.2912994623184204, "learning_rate": 8e-05, "loss": 1.5945, "step": 2901 }, { "epoch": 0.6382931925657099, "grad_norm": 0.2800372540950775, "learning_rate": 8e-05, "loss": 1.5661, "step": 2902 }, { "epoch": 0.6385131419773452, "grad_norm": 0.3248150050640106, "learning_rate": 8e-05, "loss": 1.6098, "step": 2903 }, { "epoch": 0.6387330913889805, "grad_norm": 0.27953609824180603, "learning_rate": 8e-05, "loss": 1.7255, "step": 2904 }, { "epoch": 0.6389530408006159, "grad_norm": 0.276395708322525, "learning_rate": 8e-05, "loss": 1.6782, "step": 2905 }, { "epoch": 0.6391729902122512, "grad_norm": 0.2754693031311035, "learning_rate": 8e-05, "loss": 1.5135, "step": 2906 }, { "epoch": 0.6393929396238865, "grad_norm": 0.2756873369216919, "learning_rate": 8e-05, "loss": 1.6208, "step": 2907 }, { "epoch": 0.6396128890355218, "grad_norm": 0.3032161593437195, "learning_rate": 8e-05, "loss": 1.7675, "step": 2908 }, { "epoch": 0.6398328384471571, "grad_norm": 0.2915925085544586, "learning_rate": 8e-05, "loss": 1.6899, "step": 2909 }, { "epoch": 0.6400527878587925, "grad_norm": 0.28415006399154663, "learning_rate": 8e-05, "loss": 1.7218, "step": 2910 }, { "epoch": 0.6402727372704278, "grad_norm": 0.3023785352706909, "learning_rate": 8e-05, "loss": 1.7959, "step": 2911 }, { "epoch": 0.6404926866820632, "grad_norm": 0.2656283974647522, "learning_rate": 8e-05, "loss": 1.6287, "step": 2912 }, { "epoch": 0.6407126360936984, "grad_norm": 0.2835081219673157, "learning_rate": 8e-05, "loss": 1.6597, "step": 2913 }, { "epoch": 0.6409325855053337, "grad_norm": 0.2756771743297577, "learning_rate": 8e-05, "loss": 1.625, "step": 2914 }, { "epoch": 0.6411525349169691, "grad_norm": 0.283149778842926, "learning_rate": 8e-05, "loss": 1.662, "step": 2915 }, { "epoch": 0.6413724843286044, "grad_norm": 0.28902921080589294, "learning_rate": 8e-05, "loss": 1.4862, "step": 2916 }, { "epoch": 0.6415924337402398, "grad_norm": 0.28932076692581177, "learning_rate": 8e-05, "loss": 1.6109, "step": 2917 }, { "epoch": 0.641812383151875, "grad_norm": 0.30964934825897217, "learning_rate": 8e-05, "loss": 1.7021, "step": 2918 }, { "epoch": 0.6420323325635104, "grad_norm": 0.28258854150772095, "learning_rate": 8e-05, "loss": 1.6536, "step": 2919 }, { "epoch": 0.6422522819751457, "grad_norm": 0.2939313054084778, "learning_rate": 8e-05, "loss": 1.6299, "step": 2920 }, { "epoch": 0.642472231386781, "grad_norm": 0.31722190976142883, "learning_rate": 8e-05, "loss": 1.6789, "step": 2921 }, { "epoch": 0.6426921807984164, "grad_norm": 0.29024428129196167, "learning_rate": 8e-05, "loss": 1.7429, "step": 2922 }, { "epoch": 0.6429121302100517, "grad_norm": 0.2716485559940338, "learning_rate": 8e-05, "loss": 1.619, "step": 2923 }, { "epoch": 0.6431320796216871, "grad_norm": 0.2986311614513397, "learning_rate": 8e-05, "loss": 1.7574, "step": 2924 }, { "epoch": 0.6433520290333223, "grad_norm": 0.29542550444602966, "learning_rate": 8e-05, "loss": 1.7322, "step": 2925 }, { "epoch": 0.6435719784449576, "grad_norm": 0.27215078473091125, "learning_rate": 8e-05, "loss": 1.5544, "step": 2926 }, { "epoch": 0.643791927856593, "grad_norm": 0.29105404019355774, "learning_rate": 8e-05, "loss": 1.7499, "step": 2927 }, { "epoch": 0.6440118772682283, "grad_norm": 0.2990782558917999, "learning_rate": 8e-05, "loss": 1.7987, "step": 2928 }, { "epoch": 0.6442318266798637, "grad_norm": 0.27296003699302673, "learning_rate": 8e-05, "loss": 1.6271, "step": 2929 }, { "epoch": 0.644451776091499, "grad_norm": 0.27194517850875854, "learning_rate": 8e-05, "loss": 1.5178, "step": 2930 }, { "epoch": 0.6446717255031342, "grad_norm": 0.2720150053501129, "learning_rate": 8e-05, "loss": 1.5665, "step": 2931 }, { "epoch": 0.6448916749147696, "grad_norm": 0.2825513780117035, "learning_rate": 8e-05, "loss": 1.6276, "step": 2932 }, { "epoch": 0.6451116243264049, "grad_norm": 0.2869420349597931, "learning_rate": 8e-05, "loss": 1.5306, "step": 2933 }, { "epoch": 0.6453315737380403, "grad_norm": 0.2829979360103607, "learning_rate": 8e-05, "loss": 1.7132, "step": 2934 }, { "epoch": 0.6455515231496756, "grad_norm": 0.28047260642051697, "learning_rate": 8e-05, "loss": 1.5932, "step": 2935 }, { "epoch": 0.6457714725613108, "grad_norm": 0.287765234708786, "learning_rate": 8e-05, "loss": 1.6029, "step": 2936 }, { "epoch": 0.6459914219729462, "grad_norm": 0.2858487665653229, "learning_rate": 8e-05, "loss": 1.5959, "step": 2937 }, { "epoch": 0.6462113713845815, "grad_norm": 0.29041311144828796, "learning_rate": 8e-05, "loss": 1.6348, "step": 2938 }, { "epoch": 0.6464313207962169, "grad_norm": 0.2873425781726837, "learning_rate": 8e-05, "loss": 1.6514, "step": 2939 }, { "epoch": 0.6466512702078522, "grad_norm": 0.2767978310585022, "learning_rate": 8e-05, "loss": 1.5471, "step": 2940 }, { "epoch": 0.6468712196194876, "grad_norm": 0.27061501145362854, "learning_rate": 8e-05, "loss": 1.5439, "step": 2941 }, { "epoch": 0.6470911690311228, "grad_norm": 0.2724677324295044, "learning_rate": 8e-05, "loss": 1.6808, "step": 2942 }, { "epoch": 0.6473111184427581, "grad_norm": 0.2804121971130371, "learning_rate": 8e-05, "loss": 1.7923, "step": 2943 }, { "epoch": 0.6475310678543935, "grad_norm": 0.2881599962711334, "learning_rate": 8e-05, "loss": 1.745, "step": 2944 }, { "epoch": 0.6477510172660288, "grad_norm": 0.3064921200275421, "learning_rate": 8e-05, "loss": 1.7387, "step": 2945 }, { "epoch": 0.6479709666776642, "grad_norm": 0.2994825541973114, "learning_rate": 8e-05, "loss": 1.7143, "step": 2946 }, { "epoch": 0.6481909160892995, "grad_norm": 0.29315468668937683, "learning_rate": 8e-05, "loss": 1.7712, "step": 2947 }, { "epoch": 0.6484108655009347, "grad_norm": 0.2923111915588379, "learning_rate": 8e-05, "loss": 1.6861, "step": 2948 }, { "epoch": 0.6486308149125701, "grad_norm": 0.2861957848072052, "learning_rate": 8e-05, "loss": 1.5951, "step": 2949 }, { "epoch": 0.6488507643242054, "grad_norm": 0.2978787422180176, "learning_rate": 8e-05, "loss": 1.6617, "step": 2950 }, { "epoch": 0.6490707137358408, "grad_norm": 0.28596314787864685, "learning_rate": 8e-05, "loss": 1.7623, "step": 2951 }, { "epoch": 0.6492906631474761, "grad_norm": 0.29754844307899475, "learning_rate": 8e-05, "loss": 1.6497, "step": 2952 }, { "epoch": 0.6495106125591114, "grad_norm": 0.2931132912635803, "learning_rate": 8e-05, "loss": 1.5972, "step": 2953 }, { "epoch": 0.6497305619707467, "grad_norm": 0.2667228877544403, "learning_rate": 8e-05, "loss": 1.5598, "step": 2954 }, { "epoch": 0.649950511382382, "grad_norm": 0.2866271436214447, "learning_rate": 8e-05, "loss": 1.5908, "step": 2955 }, { "epoch": 0.6501704607940174, "grad_norm": 0.28429698944091797, "learning_rate": 8e-05, "loss": 1.6582, "step": 2956 }, { "epoch": 0.6503904102056527, "grad_norm": 0.27636975049972534, "learning_rate": 8e-05, "loss": 1.6778, "step": 2957 }, { "epoch": 0.650610359617288, "grad_norm": 0.3005516529083252, "learning_rate": 8e-05, "loss": 1.7191, "step": 2958 }, { "epoch": 0.6508303090289234, "grad_norm": 0.28478094935417175, "learning_rate": 8e-05, "loss": 1.5651, "step": 2959 }, { "epoch": 0.6510502584405586, "grad_norm": 0.2879832088947296, "learning_rate": 8e-05, "loss": 1.7055, "step": 2960 }, { "epoch": 0.651270207852194, "grad_norm": 0.2899249196052551, "learning_rate": 8e-05, "loss": 1.6682, "step": 2961 }, { "epoch": 0.6514901572638293, "grad_norm": 0.26806798577308655, "learning_rate": 8e-05, "loss": 1.6962, "step": 2962 }, { "epoch": 0.6517101066754647, "grad_norm": 0.2929481565952301, "learning_rate": 8e-05, "loss": 1.6388, "step": 2963 }, { "epoch": 0.6519300560871, "grad_norm": 0.2920469641685486, "learning_rate": 8e-05, "loss": 1.8584, "step": 2964 }, { "epoch": 0.6521500054987353, "grad_norm": 0.285696417093277, "learning_rate": 8e-05, "loss": 1.6491, "step": 2965 }, { "epoch": 0.6523699549103706, "grad_norm": 0.2991807162761688, "learning_rate": 8e-05, "loss": 1.6565, "step": 2966 }, { "epoch": 0.6525899043220059, "grad_norm": 0.27987217903137207, "learning_rate": 8e-05, "loss": 1.6274, "step": 2967 }, { "epoch": 0.6528098537336413, "grad_norm": 0.2810576856136322, "learning_rate": 8e-05, "loss": 1.5294, "step": 2968 }, { "epoch": 0.6530298031452766, "grad_norm": 0.2755715847015381, "learning_rate": 8e-05, "loss": 1.6319, "step": 2969 }, { "epoch": 0.6532497525569119, "grad_norm": 0.3041331171989441, "learning_rate": 8e-05, "loss": 1.5308, "step": 2970 }, { "epoch": 0.6534697019685473, "grad_norm": 0.2858032286167145, "learning_rate": 8e-05, "loss": 1.6557, "step": 2971 }, { "epoch": 0.6536896513801825, "grad_norm": 0.3001968562602997, "learning_rate": 8e-05, "loss": 1.6841, "step": 2972 }, { "epoch": 0.6539096007918179, "grad_norm": 0.29567384719848633, "learning_rate": 8e-05, "loss": 1.6329, "step": 2973 }, { "epoch": 0.6541295502034532, "grad_norm": 0.29874905943870544, "learning_rate": 8e-05, "loss": 1.7178, "step": 2974 }, { "epoch": 0.6543494996150885, "grad_norm": 0.34721627831459045, "learning_rate": 8e-05, "loss": 1.8192, "step": 2975 }, { "epoch": 0.6545694490267239, "grad_norm": 0.2965874671936035, "learning_rate": 8e-05, "loss": 1.8156, "step": 2976 }, { "epoch": 0.6547893984383591, "grad_norm": 0.2710880637168884, "learning_rate": 8e-05, "loss": 1.5101, "step": 2977 }, { "epoch": 0.6550093478499945, "grad_norm": 0.2852049171924591, "learning_rate": 8e-05, "loss": 1.5835, "step": 2978 }, { "epoch": 0.6552292972616298, "grad_norm": 0.2898092567920685, "learning_rate": 8e-05, "loss": 1.599, "step": 2979 }, { "epoch": 0.6554492466732651, "grad_norm": 0.2880117893218994, "learning_rate": 8e-05, "loss": 1.5904, "step": 2980 }, { "epoch": 0.6556691960849005, "grad_norm": 0.2850951850414276, "learning_rate": 8e-05, "loss": 1.5551, "step": 2981 }, { "epoch": 0.6558891454965358, "grad_norm": 0.292883038520813, "learning_rate": 8e-05, "loss": 1.5587, "step": 2982 }, { "epoch": 0.6561090949081712, "grad_norm": 0.3050660490989685, "learning_rate": 8e-05, "loss": 1.7904, "step": 2983 }, { "epoch": 0.6563290443198064, "grad_norm": 0.31059566140174866, "learning_rate": 8e-05, "loss": 1.7716, "step": 2984 }, { "epoch": 0.6565489937314418, "grad_norm": 0.33118265867233276, "learning_rate": 8e-05, "loss": 1.7878, "step": 2985 }, { "epoch": 0.6567689431430771, "grad_norm": 0.27835318446159363, "learning_rate": 8e-05, "loss": 1.7069, "step": 2986 }, { "epoch": 0.6569888925547124, "grad_norm": 0.33706921339035034, "learning_rate": 8e-05, "loss": 1.7135, "step": 2987 }, { "epoch": 0.6572088419663478, "grad_norm": 0.3115323781967163, "learning_rate": 8e-05, "loss": 1.7396, "step": 2988 }, { "epoch": 0.657428791377983, "grad_norm": 0.2827862501144409, "learning_rate": 8e-05, "loss": 1.6889, "step": 2989 }, { "epoch": 0.6576487407896184, "grad_norm": 0.29057440161705017, "learning_rate": 8e-05, "loss": 1.7208, "step": 2990 }, { "epoch": 0.6578686902012537, "grad_norm": 0.2977316081523895, "learning_rate": 8e-05, "loss": 1.7192, "step": 2991 }, { "epoch": 0.658088639612889, "grad_norm": 0.296475887298584, "learning_rate": 8e-05, "loss": 1.6815, "step": 2992 }, { "epoch": 0.6583085890245244, "grad_norm": 0.29579752683639526, "learning_rate": 8e-05, "loss": 1.781, "step": 2993 }, { "epoch": 0.6585285384361597, "grad_norm": 0.2853552997112274, "learning_rate": 8e-05, "loss": 1.69, "step": 2994 }, { "epoch": 0.658748487847795, "grad_norm": 0.2831558883190155, "learning_rate": 8e-05, "loss": 1.7118, "step": 2995 }, { "epoch": 0.6589684372594303, "grad_norm": 0.30975469946861267, "learning_rate": 8e-05, "loss": 1.7153, "step": 2996 }, { "epoch": 0.6591883866710656, "grad_norm": 0.287047803401947, "learning_rate": 8e-05, "loss": 1.7241, "step": 2997 }, { "epoch": 0.659408336082701, "grad_norm": 0.2812976837158203, "learning_rate": 8e-05, "loss": 1.6536, "step": 2998 }, { "epoch": 0.6596282854943363, "grad_norm": 0.2794138491153717, "learning_rate": 8e-05, "loss": 1.6176, "step": 2999 }, { "epoch": 0.6598482349059717, "grad_norm": 0.2949649691581726, "learning_rate": 8e-05, "loss": 1.6499, "step": 3000 }, { "epoch": 0.6600681843176069, "grad_norm": 0.2770403027534485, "learning_rate": 8e-05, "loss": 1.6235, "step": 3001 }, { "epoch": 0.6602881337292422, "grad_norm": 0.30226513743400574, "learning_rate": 8e-05, "loss": 1.7714, "step": 3002 }, { "epoch": 0.6605080831408776, "grad_norm": 0.27673766016960144, "learning_rate": 8e-05, "loss": 1.6372, "step": 3003 }, { "epoch": 0.6607280325525129, "grad_norm": 0.2805183231830597, "learning_rate": 8e-05, "loss": 1.5495, "step": 3004 }, { "epoch": 0.6609479819641483, "grad_norm": 0.33441999554634094, "learning_rate": 8e-05, "loss": 1.7972, "step": 3005 }, { "epoch": 0.6611679313757836, "grad_norm": 0.29396557807922363, "learning_rate": 8e-05, "loss": 1.7203, "step": 3006 }, { "epoch": 0.661387880787419, "grad_norm": 0.27457180619239807, "learning_rate": 8e-05, "loss": 1.5044, "step": 3007 }, { "epoch": 0.6616078301990542, "grad_norm": 0.2927014231681824, "learning_rate": 8e-05, "loss": 1.6993, "step": 3008 }, { "epoch": 0.6618277796106895, "grad_norm": 0.2836621403694153, "learning_rate": 8e-05, "loss": 1.6748, "step": 3009 }, { "epoch": 0.6620477290223249, "grad_norm": 0.28402644395828247, "learning_rate": 8e-05, "loss": 1.6562, "step": 3010 }, { "epoch": 0.6622676784339602, "grad_norm": 0.2613022029399872, "learning_rate": 8e-05, "loss": 1.5386, "step": 3011 }, { "epoch": 0.6624876278455956, "grad_norm": 0.2824957072734833, "learning_rate": 8e-05, "loss": 1.6881, "step": 3012 }, { "epoch": 0.6627075772572308, "grad_norm": 0.301944375038147, "learning_rate": 8e-05, "loss": 1.6021, "step": 3013 }, { "epoch": 0.6629275266688661, "grad_norm": 0.2858961820602417, "learning_rate": 8e-05, "loss": 1.7544, "step": 3014 }, { "epoch": 0.6631474760805015, "grad_norm": 0.3019895851612091, "learning_rate": 8e-05, "loss": 1.611, "step": 3015 }, { "epoch": 0.6633674254921368, "grad_norm": 0.30784347653388977, "learning_rate": 8e-05, "loss": 1.5494, "step": 3016 }, { "epoch": 0.6635873749037722, "grad_norm": 0.2946648597717285, "learning_rate": 8e-05, "loss": 1.6664, "step": 3017 }, { "epoch": 0.6638073243154075, "grad_norm": 0.29524630308151245, "learning_rate": 8e-05, "loss": 1.6755, "step": 3018 }, { "epoch": 0.6640272737270427, "grad_norm": 0.2957313656806946, "learning_rate": 8e-05, "loss": 1.7904, "step": 3019 }, { "epoch": 0.6642472231386781, "grad_norm": 0.2753024697303772, "learning_rate": 8e-05, "loss": 1.5431, "step": 3020 }, { "epoch": 0.6644671725503134, "grad_norm": 0.27610090374946594, "learning_rate": 8e-05, "loss": 1.5585, "step": 3021 }, { "epoch": 0.6646871219619488, "grad_norm": 0.33159691095352173, "learning_rate": 8e-05, "loss": 1.6951, "step": 3022 }, { "epoch": 0.6649070713735841, "grad_norm": 0.3014865219593048, "learning_rate": 8e-05, "loss": 1.8178, "step": 3023 }, { "epoch": 0.6651270207852193, "grad_norm": 0.29317575693130493, "learning_rate": 8e-05, "loss": 1.699, "step": 3024 }, { "epoch": 0.6653469701968547, "grad_norm": 0.28347086906433105, "learning_rate": 8e-05, "loss": 1.658, "step": 3025 }, { "epoch": 0.66556691960849, "grad_norm": 0.27827227115631104, "learning_rate": 8e-05, "loss": 1.6284, "step": 3026 }, { "epoch": 0.6657868690201254, "grad_norm": 0.2806004285812378, "learning_rate": 8e-05, "loss": 1.6494, "step": 3027 }, { "epoch": 0.6660068184317607, "grad_norm": 0.27863019704818726, "learning_rate": 8e-05, "loss": 1.5617, "step": 3028 }, { "epoch": 0.6662267678433961, "grad_norm": 0.3057478368282318, "learning_rate": 8e-05, "loss": 1.7866, "step": 3029 }, { "epoch": 0.6664467172550314, "grad_norm": 0.29229068756103516, "learning_rate": 8e-05, "loss": 1.5757, "step": 3030 }, { "epoch": 0.6666666666666666, "grad_norm": 0.35514572262763977, "learning_rate": 8e-05, "loss": 1.8164, "step": 3031 }, { "epoch": 0.666886616078302, "grad_norm": 0.2878423035144806, "learning_rate": 8e-05, "loss": 1.6587, "step": 3032 }, { "epoch": 0.6671065654899373, "grad_norm": 0.34859347343444824, "learning_rate": 8e-05, "loss": 1.8458, "step": 3033 }, { "epoch": 0.6673265149015727, "grad_norm": 0.2848225235939026, "learning_rate": 8e-05, "loss": 1.6178, "step": 3034 }, { "epoch": 0.667546464313208, "grad_norm": 0.29574286937713623, "learning_rate": 8e-05, "loss": 1.5993, "step": 3035 }, { "epoch": 0.6677664137248432, "grad_norm": 0.2836141586303711, "learning_rate": 8e-05, "loss": 1.696, "step": 3036 }, { "epoch": 0.6679863631364786, "grad_norm": 0.32046374678611755, "learning_rate": 8e-05, "loss": 1.8144, "step": 3037 }, { "epoch": 0.6682063125481139, "grad_norm": 0.2817988991737366, "learning_rate": 8e-05, "loss": 1.675, "step": 3038 }, { "epoch": 0.6684262619597493, "grad_norm": 0.3105819523334503, "learning_rate": 8e-05, "loss": 1.756, "step": 3039 }, { "epoch": 0.6686462113713846, "grad_norm": 0.28836145997047424, "learning_rate": 8e-05, "loss": 1.7165, "step": 3040 }, { "epoch": 0.6688661607830199, "grad_norm": 0.35061419010162354, "learning_rate": 8e-05, "loss": 1.6084, "step": 3041 }, { "epoch": 0.6690861101946552, "grad_norm": 0.2887910008430481, "learning_rate": 8e-05, "loss": 1.7425, "step": 3042 }, { "epoch": 0.6693060596062905, "grad_norm": 0.34143608808517456, "learning_rate": 8e-05, "loss": 1.8527, "step": 3043 }, { "epoch": 0.6695260090179259, "grad_norm": 0.2963760197162628, "learning_rate": 8e-05, "loss": 1.7126, "step": 3044 }, { "epoch": 0.6697459584295612, "grad_norm": 0.27499714493751526, "learning_rate": 8e-05, "loss": 1.5678, "step": 3045 }, { "epoch": 0.6699659078411965, "grad_norm": 0.2994243800640106, "learning_rate": 8e-05, "loss": 1.7, "step": 3046 }, { "epoch": 0.6701858572528319, "grad_norm": 0.28100645542144775, "learning_rate": 8e-05, "loss": 1.5831, "step": 3047 }, { "epoch": 0.6704058066644671, "grad_norm": 0.3079579472541809, "learning_rate": 8e-05, "loss": 1.5761, "step": 3048 }, { "epoch": 0.6706257560761025, "grad_norm": 0.30679425597190857, "learning_rate": 8e-05, "loss": 1.8592, "step": 3049 }, { "epoch": 0.6708457054877378, "grad_norm": 0.2816236913204193, "learning_rate": 8e-05, "loss": 1.61, "step": 3050 }, { "epoch": 0.6710656548993732, "grad_norm": 0.29027682542800903, "learning_rate": 8e-05, "loss": 1.7003, "step": 3051 }, { "epoch": 0.6712856043110085, "grad_norm": 0.29806220531463623, "learning_rate": 8e-05, "loss": 1.7301, "step": 3052 }, { "epoch": 0.6715055537226438, "grad_norm": 0.3061826527118683, "learning_rate": 8e-05, "loss": 1.7959, "step": 3053 }, { "epoch": 0.6717255031342791, "grad_norm": 0.2916790544986725, "learning_rate": 8e-05, "loss": 1.6444, "step": 3054 }, { "epoch": 0.6719454525459144, "grad_norm": 0.29990294575691223, "learning_rate": 8e-05, "loss": 1.6154, "step": 3055 }, { "epoch": 0.6721654019575498, "grad_norm": 0.2841048240661621, "learning_rate": 8e-05, "loss": 1.604, "step": 3056 }, { "epoch": 0.6723853513691851, "grad_norm": 0.2929285764694214, "learning_rate": 8e-05, "loss": 1.7562, "step": 3057 }, { "epoch": 0.6726053007808204, "grad_norm": 0.3035488724708557, "learning_rate": 8e-05, "loss": 1.7122, "step": 3058 }, { "epoch": 0.6728252501924558, "grad_norm": 0.2829075753688812, "learning_rate": 8e-05, "loss": 1.5672, "step": 3059 }, { "epoch": 0.673045199604091, "grad_norm": 0.2959376871585846, "learning_rate": 8e-05, "loss": 1.6747, "step": 3060 }, { "epoch": 0.6732651490157264, "grad_norm": 0.2833213806152344, "learning_rate": 8e-05, "loss": 1.6152, "step": 3061 }, { "epoch": 0.6734850984273617, "grad_norm": 0.30541667342185974, "learning_rate": 8e-05, "loss": 1.8179, "step": 3062 }, { "epoch": 0.673705047838997, "grad_norm": 0.2929080128669739, "learning_rate": 8e-05, "loss": 1.7548, "step": 3063 }, { "epoch": 0.6739249972506324, "grad_norm": 0.28685638308525085, "learning_rate": 8e-05, "loss": 1.6509, "step": 3064 }, { "epoch": 0.6741449466622677, "grad_norm": 0.2862042486667633, "learning_rate": 8e-05, "loss": 1.6994, "step": 3065 }, { "epoch": 0.674364896073903, "grad_norm": 0.298554927110672, "learning_rate": 8e-05, "loss": 1.7516, "step": 3066 }, { "epoch": 0.6745848454855383, "grad_norm": 0.27519044280052185, "learning_rate": 8e-05, "loss": 1.6233, "step": 3067 }, { "epoch": 0.6748047948971736, "grad_norm": 0.2909212112426758, "learning_rate": 8e-05, "loss": 1.6986, "step": 3068 }, { "epoch": 0.675024744308809, "grad_norm": 0.3514604866504669, "learning_rate": 8e-05, "loss": 1.5948, "step": 3069 }, { "epoch": 0.6752446937204443, "grad_norm": 0.2866019308567047, "learning_rate": 8e-05, "loss": 1.6945, "step": 3070 }, { "epoch": 0.6754646431320797, "grad_norm": 0.28309282660484314, "learning_rate": 8e-05, "loss": 1.5978, "step": 3071 }, { "epoch": 0.6756845925437149, "grad_norm": 0.29281336069107056, "learning_rate": 8e-05, "loss": 1.7229, "step": 3072 }, { "epoch": 0.6759045419553503, "grad_norm": 0.2977350056171417, "learning_rate": 8e-05, "loss": 1.6669, "step": 3073 }, { "epoch": 0.6761244913669856, "grad_norm": 0.2664392590522766, "learning_rate": 8e-05, "loss": 1.5452, "step": 3074 }, { "epoch": 0.6763444407786209, "grad_norm": 0.27810007333755493, "learning_rate": 8e-05, "loss": 1.5437, "step": 3075 }, { "epoch": 0.6765643901902563, "grad_norm": 0.28598442673683167, "learning_rate": 8e-05, "loss": 1.6154, "step": 3076 }, { "epoch": 0.6767843396018915, "grad_norm": 0.2768900692462921, "learning_rate": 8e-05, "loss": 1.7086, "step": 3077 }, { "epoch": 0.6770042890135269, "grad_norm": 0.2813318073749542, "learning_rate": 8e-05, "loss": 1.638, "step": 3078 }, { "epoch": 0.6772242384251622, "grad_norm": 0.2905293107032776, "learning_rate": 8e-05, "loss": 1.8223, "step": 3079 }, { "epoch": 0.6774441878367975, "grad_norm": 0.3056020140647888, "learning_rate": 8e-05, "loss": 1.6754, "step": 3080 }, { "epoch": 0.6776641372484329, "grad_norm": 0.29252004623413086, "learning_rate": 8e-05, "loss": 1.764, "step": 3081 }, { "epoch": 0.6778840866600682, "grad_norm": 0.29702186584472656, "learning_rate": 8e-05, "loss": 1.707, "step": 3082 }, { "epoch": 0.6781040360717036, "grad_norm": 0.286920964717865, "learning_rate": 8e-05, "loss": 1.6433, "step": 3083 }, { "epoch": 0.6783239854833388, "grad_norm": 0.28408947587013245, "learning_rate": 8e-05, "loss": 1.5688, "step": 3084 }, { "epoch": 0.6785439348949741, "grad_norm": 0.30250096321105957, "learning_rate": 8e-05, "loss": 1.6018, "step": 3085 }, { "epoch": 0.6787638843066095, "grad_norm": 0.28146034479141235, "learning_rate": 8e-05, "loss": 1.5733, "step": 3086 }, { "epoch": 0.6789838337182448, "grad_norm": 0.27325439453125, "learning_rate": 8e-05, "loss": 1.5913, "step": 3087 }, { "epoch": 0.6792037831298802, "grad_norm": 0.2941316068172455, "learning_rate": 8e-05, "loss": 1.8549, "step": 3088 }, { "epoch": 0.6794237325415154, "grad_norm": 0.29817789793014526, "learning_rate": 8e-05, "loss": 1.8493, "step": 3089 }, { "epoch": 0.6796436819531507, "grad_norm": 0.29364877939224243, "learning_rate": 8e-05, "loss": 1.6039, "step": 3090 }, { "epoch": 0.6798636313647861, "grad_norm": 0.3067622482776642, "learning_rate": 8e-05, "loss": 1.7219, "step": 3091 }, { "epoch": 0.6800835807764214, "grad_norm": 0.2989698648452759, "learning_rate": 8e-05, "loss": 1.6856, "step": 3092 }, { "epoch": 0.6803035301880568, "grad_norm": 0.27398690581321716, "learning_rate": 8e-05, "loss": 1.6872, "step": 3093 }, { "epoch": 0.6805234795996921, "grad_norm": 0.27870994806289673, "learning_rate": 8e-05, "loss": 1.5519, "step": 3094 }, { "epoch": 0.6807434290113274, "grad_norm": 0.27137231826782227, "learning_rate": 8e-05, "loss": 1.703, "step": 3095 }, { "epoch": 0.6809633784229627, "grad_norm": 0.30240193009376526, "learning_rate": 8e-05, "loss": 1.7728, "step": 3096 }, { "epoch": 0.681183327834598, "grad_norm": 0.29213348031044006, "learning_rate": 8e-05, "loss": 1.7236, "step": 3097 }, { "epoch": 0.6814032772462334, "grad_norm": 0.2953360974788666, "learning_rate": 8e-05, "loss": 1.6609, "step": 3098 }, { "epoch": 0.6816232266578687, "grad_norm": 0.29802900552749634, "learning_rate": 8e-05, "loss": 1.6969, "step": 3099 }, { "epoch": 0.6818431760695041, "grad_norm": 0.2967725396156311, "learning_rate": 8e-05, "loss": 1.8439, "step": 3100 }, { "epoch": 0.6820631254811393, "grad_norm": 0.2971365451812744, "learning_rate": 8e-05, "loss": 1.7695, "step": 3101 }, { "epoch": 0.6822830748927746, "grad_norm": 0.2880017161369324, "learning_rate": 8e-05, "loss": 1.6916, "step": 3102 }, { "epoch": 0.68250302430441, "grad_norm": 0.2691252529621124, "learning_rate": 8e-05, "loss": 1.5439, "step": 3103 }, { "epoch": 0.6827229737160453, "grad_norm": 0.2928919494152069, "learning_rate": 8e-05, "loss": 1.6325, "step": 3104 }, { "epoch": 0.6829429231276807, "grad_norm": 0.28234854340553284, "learning_rate": 8e-05, "loss": 1.6778, "step": 3105 }, { "epoch": 0.683162872539316, "grad_norm": 0.2900925278663635, "learning_rate": 8e-05, "loss": 1.7047, "step": 3106 }, { "epoch": 0.6833828219509512, "grad_norm": 0.30338430404663086, "learning_rate": 8e-05, "loss": 1.7422, "step": 3107 }, { "epoch": 0.6836027713625866, "grad_norm": 0.2576045095920563, "learning_rate": 8e-05, "loss": 1.2735, "step": 3108 }, { "epoch": 0.6838227207742219, "grad_norm": 0.2839035987854004, "learning_rate": 8e-05, "loss": 1.6443, "step": 3109 }, { "epoch": 0.6840426701858573, "grad_norm": 0.29431045055389404, "learning_rate": 8e-05, "loss": 1.7928, "step": 3110 }, { "epoch": 0.6842626195974926, "grad_norm": 0.301283597946167, "learning_rate": 8e-05, "loss": 1.7339, "step": 3111 }, { "epoch": 0.6844825690091279, "grad_norm": 0.27778249979019165, "learning_rate": 8e-05, "loss": 1.6481, "step": 3112 }, { "epoch": 0.6847025184207632, "grad_norm": 0.317439466714859, "learning_rate": 8e-05, "loss": 1.7339, "step": 3113 }, { "epoch": 0.6849224678323985, "grad_norm": 0.26562100648880005, "learning_rate": 8e-05, "loss": 1.5077, "step": 3114 }, { "epoch": 0.6851424172440339, "grad_norm": 0.27444925904273987, "learning_rate": 8e-05, "loss": 1.6178, "step": 3115 }, { "epoch": 0.6853623666556692, "grad_norm": 0.302569180727005, "learning_rate": 8e-05, "loss": 1.7989, "step": 3116 }, { "epoch": 0.6855823160673045, "grad_norm": 0.27267923951148987, "learning_rate": 8e-05, "loss": 1.5604, "step": 3117 }, { "epoch": 0.6858022654789399, "grad_norm": 0.295436292886734, "learning_rate": 8e-05, "loss": 1.6847, "step": 3118 }, { "epoch": 0.6860222148905751, "grad_norm": 0.29056796431541443, "learning_rate": 8e-05, "loss": 1.5835, "step": 3119 }, { "epoch": 0.6862421643022105, "grad_norm": 0.29942628741264343, "learning_rate": 8e-05, "loss": 1.692, "step": 3120 }, { "epoch": 0.6864621137138458, "grad_norm": 0.29042521119117737, "learning_rate": 8e-05, "loss": 1.7639, "step": 3121 }, { "epoch": 0.6866820631254812, "grad_norm": 0.2988620102405548, "learning_rate": 8e-05, "loss": 1.7096, "step": 3122 }, { "epoch": 0.6869020125371165, "grad_norm": 0.30227869749069214, "learning_rate": 8e-05, "loss": 1.7103, "step": 3123 }, { "epoch": 0.6871219619487517, "grad_norm": 0.2963883578777313, "learning_rate": 8e-05, "loss": 1.6252, "step": 3124 }, { "epoch": 0.6873419113603871, "grad_norm": 0.267903596162796, "learning_rate": 8e-05, "loss": 1.5459, "step": 3125 }, { "epoch": 0.6875618607720224, "grad_norm": 0.34292858839035034, "learning_rate": 8e-05, "loss": 1.7557, "step": 3126 }, { "epoch": 0.6877818101836578, "grad_norm": 0.31191956996917725, "learning_rate": 8e-05, "loss": 1.8102, "step": 3127 }, { "epoch": 0.6880017595952931, "grad_norm": 0.2759200632572174, "learning_rate": 8e-05, "loss": 1.4965, "step": 3128 }, { "epoch": 0.6882217090069284, "grad_norm": 0.279722660779953, "learning_rate": 8e-05, "loss": 1.6268, "step": 3129 }, { "epoch": 0.6884416584185638, "grad_norm": 0.296171635389328, "learning_rate": 8e-05, "loss": 1.8015, "step": 3130 }, { "epoch": 0.688661607830199, "grad_norm": 0.2832602560520172, "learning_rate": 8e-05, "loss": 1.6554, "step": 3131 }, { "epoch": 0.6888815572418344, "grad_norm": 0.28423500061035156, "learning_rate": 8e-05, "loss": 1.5898, "step": 3132 }, { "epoch": 0.6891015066534697, "grad_norm": 0.27690091729164124, "learning_rate": 8e-05, "loss": 1.5216, "step": 3133 }, { "epoch": 0.689321456065105, "grad_norm": 0.36139482259750366, "learning_rate": 8e-05, "loss": 1.7153, "step": 3134 }, { "epoch": 0.6895414054767404, "grad_norm": 0.3002482056617737, "learning_rate": 8e-05, "loss": 1.8588, "step": 3135 }, { "epoch": 0.6897613548883756, "grad_norm": 0.29645755887031555, "learning_rate": 8e-05, "loss": 1.6894, "step": 3136 }, { "epoch": 0.689981304300011, "grad_norm": 0.27819156646728516, "learning_rate": 8e-05, "loss": 1.5448, "step": 3137 }, { "epoch": 0.6902012537116463, "grad_norm": 0.2958146035671234, "learning_rate": 8e-05, "loss": 1.7416, "step": 3138 }, { "epoch": 0.6904212031232816, "grad_norm": 0.30198779702186584, "learning_rate": 8e-05, "loss": 1.7765, "step": 3139 }, { "epoch": 0.690641152534917, "grad_norm": 0.2883966565132141, "learning_rate": 8e-05, "loss": 1.6467, "step": 3140 }, { "epoch": 0.6908611019465523, "grad_norm": 0.2731442451477051, "learning_rate": 8e-05, "loss": 1.5186, "step": 3141 }, { "epoch": 0.6910810513581876, "grad_norm": 0.2791924774646759, "learning_rate": 8e-05, "loss": 1.7721, "step": 3142 }, { "epoch": 0.6913010007698229, "grad_norm": 0.2815117835998535, "learning_rate": 8e-05, "loss": 1.5362, "step": 3143 }, { "epoch": 0.6915209501814583, "grad_norm": 0.2724231481552124, "learning_rate": 8e-05, "loss": 1.7015, "step": 3144 }, { "epoch": 0.6917408995930936, "grad_norm": 0.3006933629512787, "learning_rate": 8e-05, "loss": 1.7298, "step": 3145 }, { "epoch": 0.6919608490047289, "grad_norm": 0.3087989389896393, "learning_rate": 8e-05, "loss": 1.6349, "step": 3146 }, { "epoch": 0.6921807984163643, "grad_norm": 0.2806054949760437, "learning_rate": 8e-05, "loss": 1.6347, "step": 3147 }, { "epoch": 0.6924007478279995, "grad_norm": 0.28338587284088135, "learning_rate": 8e-05, "loss": 1.7277, "step": 3148 }, { "epoch": 0.6926206972396349, "grad_norm": 0.27475735545158386, "learning_rate": 8e-05, "loss": 1.6705, "step": 3149 }, { "epoch": 0.6928406466512702, "grad_norm": 0.3012952506542206, "learning_rate": 8e-05, "loss": 1.8124, "step": 3150 }, { "epoch": 0.6930605960629055, "grad_norm": 0.28966784477233887, "learning_rate": 8e-05, "loss": 1.6413, "step": 3151 }, { "epoch": 0.6932805454745409, "grad_norm": 0.27356377243995667, "learning_rate": 8e-05, "loss": 1.5894, "step": 3152 }, { "epoch": 0.6935004948861762, "grad_norm": 0.3299154043197632, "learning_rate": 8e-05, "loss": 1.6931, "step": 3153 }, { "epoch": 0.6937204442978115, "grad_norm": 0.2960734963417053, "learning_rate": 8e-05, "loss": 1.7089, "step": 3154 }, { "epoch": 0.6939403937094468, "grad_norm": 0.2994215190410614, "learning_rate": 8e-05, "loss": 1.6326, "step": 3155 }, { "epoch": 0.6941603431210821, "grad_norm": 0.2847454249858856, "learning_rate": 8e-05, "loss": 1.7982, "step": 3156 }, { "epoch": 0.6943802925327175, "grad_norm": 0.28896382451057434, "learning_rate": 8e-05, "loss": 1.6695, "step": 3157 }, { "epoch": 0.6946002419443528, "grad_norm": 0.2861023545265198, "learning_rate": 8e-05, "loss": 1.7209, "step": 3158 }, { "epoch": 0.6948201913559882, "grad_norm": 0.26986655592918396, "learning_rate": 8e-05, "loss": 1.535, "step": 3159 }, { "epoch": 0.6950401407676234, "grad_norm": 0.3060385286808014, "learning_rate": 8e-05, "loss": 1.6242, "step": 3160 }, { "epoch": 0.6952600901792587, "grad_norm": 0.2831737995147705, "learning_rate": 8e-05, "loss": 1.6341, "step": 3161 }, { "epoch": 0.6954800395908941, "grad_norm": 0.30474355816841125, "learning_rate": 8e-05, "loss": 1.7048, "step": 3162 }, { "epoch": 0.6956999890025294, "grad_norm": 0.2686191499233246, "learning_rate": 8e-05, "loss": 1.5553, "step": 3163 }, { "epoch": 0.6959199384141648, "grad_norm": 0.2899799048900604, "learning_rate": 8e-05, "loss": 1.5918, "step": 3164 }, { "epoch": 0.6961398878258, "grad_norm": 0.2757590115070343, "learning_rate": 8e-05, "loss": 1.6582, "step": 3165 }, { "epoch": 0.6963598372374354, "grad_norm": 0.2934139668941498, "learning_rate": 8e-05, "loss": 1.6196, "step": 3166 }, { "epoch": 0.6965797866490707, "grad_norm": 0.2803792953491211, "learning_rate": 8e-05, "loss": 1.4932, "step": 3167 }, { "epoch": 0.696799736060706, "grad_norm": 0.27724161744117737, "learning_rate": 8e-05, "loss": 1.6452, "step": 3168 }, { "epoch": 0.6970196854723414, "grad_norm": 0.31233644485473633, "learning_rate": 8e-05, "loss": 1.7572, "step": 3169 }, { "epoch": 0.6972396348839767, "grad_norm": 0.2862301766872406, "learning_rate": 8e-05, "loss": 1.663, "step": 3170 }, { "epoch": 0.6974595842956121, "grad_norm": 0.284385085105896, "learning_rate": 8e-05, "loss": 1.7268, "step": 3171 }, { "epoch": 0.6976795337072473, "grad_norm": 0.2832756042480469, "learning_rate": 8e-05, "loss": 1.6006, "step": 3172 }, { "epoch": 0.6978994831188826, "grad_norm": 0.2860611081123352, "learning_rate": 8e-05, "loss": 1.6932, "step": 3173 }, { "epoch": 0.698119432530518, "grad_norm": 0.2913680970668793, "learning_rate": 8e-05, "loss": 1.6705, "step": 3174 }, { "epoch": 0.6983393819421533, "grad_norm": 0.2927596867084503, "learning_rate": 8e-05, "loss": 1.7578, "step": 3175 }, { "epoch": 0.6985593313537887, "grad_norm": 0.2793924808502197, "learning_rate": 8e-05, "loss": 1.6293, "step": 3176 }, { "epoch": 0.698779280765424, "grad_norm": 0.28402742743492126, "learning_rate": 8e-05, "loss": 1.6703, "step": 3177 }, { "epoch": 0.6989992301770592, "grad_norm": 0.2879323661327362, "learning_rate": 8e-05, "loss": 1.5645, "step": 3178 }, { "epoch": 0.6992191795886946, "grad_norm": 0.28840363025665283, "learning_rate": 8e-05, "loss": 1.6608, "step": 3179 }, { "epoch": 0.6994391290003299, "grad_norm": 0.2971773147583008, "learning_rate": 8e-05, "loss": 1.7994, "step": 3180 }, { "epoch": 0.6996590784119653, "grad_norm": 0.2872611880302429, "learning_rate": 8e-05, "loss": 1.6931, "step": 3181 }, { "epoch": 0.6998790278236006, "grad_norm": 0.2892748713493347, "learning_rate": 8e-05, "loss": 1.6808, "step": 3182 }, { "epoch": 0.7000989772352358, "grad_norm": 0.3782177269458771, "learning_rate": 8e-05, "loss": 1.7109, "step": 3183 }, { "epoch": 0.7003189266468712, "grad_norm": 0.2824467420578003, "learning_rate": 8e-05, "loss": 1.6967, "step": 3184 }, { "epoch": 0.7005388760585065, "grad_norm": 0.3341313898563385, "learning_rate": 8e-05, "loss": 1.8551, "step": 3185 }, { "epoch": 0.7007588254701419, "grad_norm": 0.2792004346847534, "learning_rate": 8e-05, "loss": 1.6054, "step": 3186 }, { "epoch": 0.7009787748817772, "grad_norm": 0.2860063910484314, "learning_rate": 8e-05, "loss": 1.7671, "step": 3187 }, { "epoch": 0.7011987242934126, "grad_norm": 0.2876088321208954, "learning_rate": 8e-05, "loss": 1.6843, "step": 3188 }, { "epoch": 0.7014186737050478, "grad_norm": 0.28482869267463684, "learning_rate": 8e-05, "loss": 1.42, "step": 3189 }, { "epoch": 0.7016386231166831, "grad_norm": 0.30407440662384033, "learning_rate": 8e-05, "loss": 1.7995, "step": 3190 }, { "epoch": 0.7018585725283185, "grad_norm": 0.28979554772377014, "learning_rate": 8e-05, "loss": 1.6972, "step": 3191 }, { "epoch": 0.7020785219399538, "grad_norm": 0.2807050049304962, "learning_rate": 8e-05, "loss": 1.6641, "step": 3192 }, { "epoch": 0.7022984713515892, "grad_norm": 0.3025875985622406, "learning_rate": 8e-05, "loss": 1.5833, "step": 3193 }, { "epoch": 0.7025184207632245, "grad_norm": 0.3045402467250824, "learning_rate": 8e-05, "loss": 1.5986, "step": 3194 }, { "epoch": 0.7027383701748597, "grad_norm": 0.2938578128814697, "learning_rate": 8e-05, "loss": 1.5732, "step": 3195 }, { "epoch": 0.7029583195864951, "grad_norm": 0.270671546459198, "learning_rate": 8e-05, "loss": 1.4949, "step": 3196 }, { "epoch": 0.7031782689981304, "grad_norm": 0.2783917784690857, "learning_rate": 8e-05, "loss": 1.6004, "step": 3197 }, { "epoch": 0.7033982184097658, "grad_norm": 0.26526427268981934, "learning_rate": 8e-05, "loss": 1.5312, "step": 3198 }, { "epoch": 0.7036181678214011, "grad_norm": 0.2944982647895813, "learning_rate": 8e-05, "loss": 1.7079, "step": 3199 }, { "epoch": 0.7038381172330364, "grad_norm": 0.2870388627052307, "learning_rate": 8e-05, "loss": 1.5144, "step": 3200 }, { "epoch": 0.7040580666446717, "grad_norm": 0.28370967507362366, "learning_rate": 8e-05, "loss": 1.5572, "step": 3201 }, { "epoch": 0.704278016056307, "grad_norm": 0.2962104380130768, "learning_rate": 8e-05, "loss": 1.768, "step": 3202 }, { "epoch": 0.7044979654679424, "grad_norm": 0.30153971910476685, "learning_rate": 8e-05, "loss": 1.5896, "step": 3203 }, { "epoch": 0.7047179148795777, "grad_norm": 0.27244117856025696, "learning_rate": 8e-05, "loss": 1.6106, "step": 3204 }, { "epoch": 0.704937864291213, "grad_norm": 0.27167847752571106, "learning_rate": 8e-05, "loss": 1.6106, "step": 3205 }, { "epoch": 0.7051578137028484, "grad_norm": 0.28725895285606384, "learning_rate": 8e-05, "loss": 1.6029, "step": 3206 }, { "epoch": 0.7053777631144836, "grad_norm": 0.2840006947517395, "learning_rate": 8e-05, "loss": 1.5254, "step": 3207 }, { "epoch": 0.705597712526119, "grad_norm": 0.30838871002197266, "learning_rate": 8e-05, "loss": 1.711, "step": 3208 }, { "epoch": 0.7058176619377543, "grad_norm": 0.28863799571990967, "learning_rate": 8e-05, "loss": 1.6534, "step": 3209 }, { "epoch": 0.7060376113493897, "grad_norm": 0.27136942744255066, "learning_rate": 8e-05, "loss": 1.6272, "step": 3210 }, { "epoch": 0.706257560761025, "grad_norm": 0.29609814286231995, "learning_rate": 8e-05, "loss": 1.7805, "step": 3211 }, { "epoch": 0.7064775101726603, "grad_norm": 0.2786168158054352, "learning_rate": 8e-05, "loss": 1.5998, "step": 3212 }, { "epoch": 0.7066974595842956, "grad_norm": 0.2873217761516571, "learning_rate": 8e-05, "loss": 1.8394, "step": 3213 }, { "epoch": 0.7069174089959309, "grad_norm": 0.28696301579475403, "learning_rate": 8e-05, "loss": 1.6836, "step": 3214 }, { "epoch": 0.7071373584075663, "grad_norm": 0.2748275101184845, "learning_rate": 8e-05, "loss": 1.632, "step": 3215 }, { "epoch": 0.7073573078192016, "grad_norm": 0.30858665704727173, "learning_rate": 8e-05, "loss": 1.6296, "step": 3216 }, { "epoch": 0.7075772572308369, "grad_norm": 0.28793784976005554, "learning_rate": 8e-05, "loss": 1.7079, "step": 3217 }, { "epoch": 0.7077972066424723, "grad_norm": 0.3041824996471405, "learning_rate": 8e-05, "loss": 1.756, "step": 3218 }, { "epoch": 0.7080171560541075, "grad_norm": 0.27776944637298584, "learning_rate": 8e-05, "loss": 1.5994, "step": 3219 }, { "epoch": 0.7082371054657429, "grad_norm": 0.291361927986145, "learning_rate": 8e-05, "loss": 1.6744, "step": 3220 }, { "epoch": 0.7084570548773782, "grad_norm": 0.2865651249885559, "learning_rate": 8e-05, "loss": 1.7511, "step": 3221 }, { "epoch": 0.7086770042890135, "grad_norm": 0.28433775901794434, "learning_rate": 8e-05, "loss": 1.7272, "step": 3222 }, { "epoch": 0.7088969537006489, "grad_norm": 0.2853262722492218, "learning_rate": 8e-05, "loss": 1.6281, "step": 3223 }, { "epoch": 0.7091169031122841, "grad_norm": 0.2622532844543457, "learning_rate": 8e-05, "loss": 1.6179, "step": 3224 }, { "epoch": 0.7093368525239195, "grad_norm": 0.32692235708236694, "learning_rate": 8e-05, "loss": 1.784, "step": 3225 }, { "epoch": 0.7095568019355548, "grad_norm": 0.3035888671875, "learning_rate": 8e-05, "loss": 1.6944, "step": 3226 }, { "epoch": 0.7097767513471901, "grad_norm": 0.28113415837287903, "learning_rate": 8e-05, "loss": 1.5492, "step": 3227 }, { "epoch": 0.7099967007588255, "grad_norm": 0.282117635011673, "learning_rate": 8e-05, "loss": 1.5972, "step": 3228 }, { "epoch": 0.7102166501704608, "grad_norm": 0.28914523124694824, "learning_rate": 8e-05, "loss": 1.6345, "step": 3229 }, { "epoch": 0.7104365995820962, "grad_norm": 0.2931251525878906, "learning_rate": 8e-05, "loss": 1.7021, "step": 3230 }, { "epoch": 0.7106565489937314, "grad_norm": 0.29080888628959656, "learning_rate": 8e-05, "loss": 1.6678, "step": 3231 }, { "epoch": 0.7108764984053668, "grad_norm": 0.29373523592948914, "learning_rate": 8e-05, "loss": 1.6674, "step": 3232 }, { "epoch": 0.7110964478170021, "grad_norm": 0.2820954918861389, "learning_rate": 8e-05, "loss": 1.6162, "step": 3233 }, { "epoch": 0.7113163972286374, "grad_norm": 0.2817211449146271, "learning_rate": 8e-05, "loss": 1.572, "step": 3234 }, { "epoch": 0.7115363466402728, "grad_norm": 0.28479206562042236, "learning_rate": 8e-05, "loss": 1.4741, "step": 3235 }, { "epoch": 0.711756296051908, "grad_norm": 0.2774266004562378, "learning_rate": 8e-05, "loss": 1.5465, "step": 3236 }, { "epoch": 0.7119762454635434, "grad_norm": 0.2834816873073578, "learning_rate": 8e-05, "loss": 1.6505, "step": 3237 }, { "epoch": 0.7121961948751787, "grad_norm": 0.28653544187545776, "learning_rate": 8e-05, "loss": 1.6307, "step": 3238 }, { "epoch": 0.712416144286814, "grad_norm": 0.28194478154182434, "learning_rate": 8e-05, "loss": 1.7033, "step": 3239 }, { "epoch": 0.7126360936984494, "grad_norm": 0.2983120083808899, "learning_rate": 8e-05, "loss": 1.7034, "step": 3240 }, { "epoch": 0.7128560431100847, "grad_norm": 0.3055756390094757, "learning_rate": 8e-05, "loss": 1.6579, "step": 3241 }, { "epoch": 0.71307599252172, "grad_norm": 0.3022952377796173, "learning_rate": 8e-05, "loss": 1.7379, "step": 3242 }, { "epoch": 0.7132959419333553, "grad_norm": 0.28210487961769104, "learning_rate": 8e-05, "loss": 1.519, "step": 3243 }, { "epoch": 0.7135158913449906, "grad_norm": 0.28924715518951416, "learning_rate": 8e-05, "loss": 1.6344, "step": 3244 }, { "epoch": 0.713735840756626, "grad_norm": 0.285334974527359, "learning_rate": 8e-05, "loss": 1.7545, "step": 3245 }, { "epoch": 0.7139557901682613, "grad_norm": 0.29755088686943054, "learning_rate": 8e-05, "loss": 1.7103, "step": 3246 }, { "epoch": 0.7141757395798967, "grad_norm": 0.30941107869148254, "learning_rate": 8e-05, "loss": 1.6855, "step": 3247 }, { "epoch": 0.7143956889915319, "grad_norm": 0.29600104689598083, "learning_rate": 8e-05, "loss": 1.8132, "step": 3248 }, { "epoch": 0.7146156384031672, "grad_norm": 0.290912926197052, "learning_rate": 8e-05, "loss": 1.7253, "step": 3249 }, { "epoch": 0.7148355878148026, "grad_norm": 0.2880486845970154, "learning_rate": 8e-05, "loss": 1.7233, "step": 3250 }, { "epoch": 0.7150555372264379, "grad_norm": 0.27967116236686707, "learning_rate": 8e-05, "loss": 1.4805, "step": 3251 }, { "epoch": 0.7152754866380733, "grad_norm": 0.2888146638870239, "learning_rate": 8e-05, "loss": 1.7523, "step": 3252 }, { "epoch": 0.7154954360497086, "grad_norm": 0.2760499119758606, "learning_rate": 8e-05, "loss": 1.6595, "step": 3253 }, { "epoch": 0.715715385461344, "grad_norm": 0.29169565439224243, "learning_rate": 8e-05, "loss": 1.6981, "step": 3254 }, { "epoch": 0.7159353348729792, "grad_norm": 0.2904864251613617, "learning_rate": 8e-05, "loss": 1.708, "step": 3255 }, { "epoch": 0.7161552842846145, "grad_norm": 0.2833339273929596, "learning_rate": 8e-05, "loss": 1.5798, "step": 3256 }, { "epoch": 0.7163752336962499, "grad_norm": 0.3214869499206543, "learning_rate": 8e-05, "loss": 1.823, "step": 3257 }, { "epoch": 0.7165951831078852, "grad_norm": 0.3174017071723938, "learning_rate": 8e-05, "loss": 1.7109, "step": 3258 }, { "epoch": 0.7168151325195206, "grad_norm": 0.2913649082183838, "learning_rate": 8e-05, "loss": 1.5767, "step": 3259 }, { "epoch": 0.7170350819311558, "grad_norm": 0.2901788651943207, "learning_rate": 8e-05, "loss": 1.6476, "step": 3260 }, { "epoch": 0.7172550313427911, "grad_norm": 0.2945404052734375, "learning_rate": 8e-05, "loss": 1.6527, "step": 3261 }, { "epoch": 0.7174749807544265, "grad_norm": 0.29439011216163635, "learning_rate": 8e-05, "loss": 1.7253, "step": 3262 }, { "epoch": 0.7176949301660618, "grad_norm": 0.2744368016719818, "learning_rate": 8e-05, "loss": 1.5482, "step": 3263 }, { "epoch": 0.7179148795776972, "grad_norm": 0.3027225434780121, "learning_rate": 8e-05, "loss": 1.646, "step": 3264 }, { "epoch": 0.7181348289893325, "grad_norm": 0.271297425031662, "learning_rate": 8e-05, "loss": 1.4867, "step": 3265 }, { "epoch": 0.7183547784009677, "grad_norm": 0.32890835404396057, "learning_rate": 8e-05, "loss": 1.8444, "step": 3266 }, { "epoch": 0.7185747278126031, "grad_norm": 0.2833615839481354, "learning_rate": 8e-05, "loss": 1.7631, "step": 3267 }, { "epoch": 0.7187946772242384, "grad_norm": 0.2767006456851959, "learning_rate": 8e-05, "loss": 1.7149, "step": 3268 }, { "epoch": 0.7190146266358738, "grad_norm": 0.2924560308456421, "learning_rate": 8e-05, "loss": 1.6564, "step": 3269 }, { "epoch": 0.7192345760475091, "grad_norm": 0.28660503029823303, "learning_rate": 8e-05, "loss": 1.5485, "step": 3270 }, { "epoch": 0.7194545254591443, "grad_norm": 0.28452107310295105, "learning_rate": 8e-05, "loss": 1.4871, "step": 3271 }, { "epoch": 0.7196744748707797, "grad_norm": 0.2906745374202728, "learning_rate": 8e-05, "loss": 1.7993, "step": 3272 }, { "epoch": 0.719894424282415, "grad_norm": 0.28533515334129333, "learning_rate": 8e-05, "loss": 1.6419, "step": 3273 }, { "epoch": 0.7201143736940504, "grad_norm": 0.282487690448761, "learning_rate": 8e-05, "loss": 1.6759, "step": 3274 }, { "epoch": 0.7203343231056857, "grad_norm": 0.3005145490169525, "learning_rate": 8e-05, "loss": 1.7566, "step": 3275 }, { "epoch": 0.7205542725173211, "grad_norm": 0.2659122347831726, "learning_rate": 8e-05, "loss": 1.443, "step": 3276 }, { "epoch": 0.7207742219289563, "grad_norm": 0.30296915769577026, "learning_rate": 8e-05, "loss": 1.6464, "step": 3277 }, { "epoch": 0.7209941713405916, "grad_norm": 0.2925507128238678, "learning_rate": 8e-05, "loss": 1.5286, "step": 3278 }, { "epoch": 0.721214120752227, "grad_norm": 0.27778342366218567, "learning_rate": 8e-05, "loss": 1.5417, "step": 3279 }, { "epoch": 0.7214340701638623, "grad_norm": 0.3245250880718231, "learning_rate": 8e-05, "loss": 1.9343, "step": 3280 }, { "epoch": 0.7216540195754977, "grad_norm": 0.2841079831123352, "learning_rate": 8e-05, "loss": 1.6095, "step": 3281 }, { "epoch": 0.721873968987133, "grad_norm": 0.2818967401981354, "learning_rate": 8e-05, "loss": 1.775, "step": 3282 }, { "epoch": 0.7220939183987682, "grad_norm": 0.30750346183776855, "learning_rate": 8e-05, "loss": 1.837, "step": 3283 }, { "epoch": 0.7223138678104036, "grad_norm": 0.2764554023742676, "learning_rate": 8e-05, "loss": 1.5843, "step": 3284 }, { "epoch": 0.7225338172220389, "grad_norm": 0.2906947433948517, "learning_rate": 8e-05, "loss": 1.748, "step": 3285 }, { "epoch": 0.7227537666336743, "grad_norm": 0.30081701278686523, "learning_rate": 8e-05, "loss": 1.6306, "step": 3286 }, { "epoch": 0.7229737160453096, "grad_norm": 0.2862705886363983, "learning_rate": 8e-05, "loss": 1.7076, "step": 3287 }, { "epoch": 0.7231936654569449, "grad_norm": 0.29947561025619507, "learning_rate": 8e-05, "loss": 1.6987, "step": 3288 }, { "epoch": 0.7234136148685802, "grad_norm": 0.2880048453807831, "learning_rate": 8e-05, "loss": 1.6921, "step": 3289 }, { "epoch": 0.7236335642802155, "grad_norm": 0.2952638268470764, "learning_rate": 8e-05, "loss": 1.533, "step": 3290 }, { "epoch": 0.7238535136918509, "grad_norm": 0.29756438732147217, "learning_rate": 8e-05, "loss": 1.6473, "step": 3291 }, { "epoch": 0.7240734631034862, "grad_norm": 0.29125919938087463, "learning_rate": 8e-05, "loss": 1.6892, "step": 3292 }, { "epoch": 0.7242934125151215, "grad_norm": 0.26599353551864624, "learning_rate": 8e-05, "loss": 1.4273, "step": 3293 }, { "epoch": 0.7245133619267569, "grad_norm": 0.2843374013900757, "learning_rate": 8e-05, "loss": 1.7237, "step": 3294 }, { "epoch": 0.7247333113383921, "grad_norm": 0.28228116035461426, "learning_rate": 8e-05, "loss": 1.7063, "step": 3295 }, { "epoch": 0.7249532607500275, "grad_norm": 0.2917938828468323, "learning_rate": 8e-05, "loss": 1.7018, "step": 3296 }, { "epoch": 0.7251732101616628, "grad_norm": 0.28480616211891174, "learning_rate": 8e-05, "loss": 1.5869, "step": 3297 }, { "epoch": 0.7253931595732982, "grad_norm": 0.3087390661239624, "learning_rate": 8e-05, "loss": 1.8078, "step": 3298 }, { "epoch": 0.7256131089849335, "grad_norm": 0.2808952331542969, "learning_rate": 8e-05, "loss": 1.7724, "step": 3299 }, { "epoch": 0.7258330583965688, "grad_norm": 0.30944177508354187, "learning_rate": 8e-05, "loss": 1.7574, "step": 3300 }, { "epoch": 0.7260530078082041, "grad_norm": 0.2862800657749176, "learning_rate": 8e-05, "loss": 1.5506, "step": 3301 }, { "epoch": 0.7262729572198394, "grad_norm": 0.29587674140930176, "learning_rate": 8e-05, "loss": 1.6555, "step": 3302 }, { "epoch": 0.7264929066314748, "grad_norm": 0.279253214597702, "learning_rate": 8e-05, "loss": 1.5783, "step": 3303 }, { "epoch": 0.7267128560431101, "grad_norm": 0.2954663634300232, "learning_rate": 8e-05, "loss": 1.898, "step": 3304 }, { "epoch": 0.7269328054547454, "grad_norm": 0.27713724970817566, "learning_rate": 8e-05, "loss": 1.6111, "step": 3305 }, { "epoch": 0.7271527548663808, "grad_norm": 0.2909323275089264, "learning_rate": 8e-05, "loss": 1.7006, "step": 3306 }, { "epoch": 0.727372704278016, "grad_norm": 0.28384488821029663, "learning_rate": 8e-05, "loss": 1.5633, "step": 3307 }, { "epoch": 0.7275926536896514, "grad_norm": 0.30551770329475403, "learning_rate": 8e-05, "loss": 1.7208, "step": 3308 }, { "epoch": 0.7278126031012867, "grad_norm": 0.29794755578041077, "learning_rate": 8e-05, "loss": 1.7839, "step": 3309 }, { "epoch": 0.728032552512922, "grad_norm": 0.30165839195251465, "learning_rate": 8e-05, "loss": 1.7569, "step": 3310 }, { "epoch": 0.7282525019245574, "grad_norm": 0.2731498181819916, "learning_rate": 8e-05, "loss": 1.4881, "step": 3311 }, { "epoch": 0.7284724513361927, "grad_norm": 0.29713213443756104, "learning_rate": 8e-05, "loss": 1.7023, "step": 3312 }, { "epoch": 0.728692400747828, "grad_norm": 0.30032122135162354, "learning_rate": 8e-05, "loss": 1.7717, "step": 3313 }, { "epoch": 0.7289123501594633, "grad_norm": 0.2873935401439667, "learning_rate": 8e-05, "loss": 1.8062, "step": 3314 }, { "epoch": 0.7291322995710986, "grad_norm": 0.2813476026058197, "learning_rate": 8e-05, "loss": 1.6503, "step": 3315 }, { "epoch": 0.729352248982734, "grad_norm": 0.28851833939552307, "learning_rate": 8e-05, "loss": 1.5839, "step": 3316 }, { "epoch": 0.7295721983943693, "grad_norm": 0.29246899485588074, "learning_rate": 8e-05, "loss": 1.7905, "step": 3317 }, { "epoch": 0.7297921478060047, "grad_norm": 0.2823907434940338, "learning_rate": 8e-05, "loss": 1.695, "step": 3318 }, { "epoch": 0.7300120972176399, "grad_norm": 0.2800486981868744, "learning_rate": 8e-05, "loss": 1.7131, "step": 3319 }, { "epoch": 0.7302320466292753, "grad_norm": 0.28211459517478943, "learning_rate": 8e-05, "loss": 1.5183, "step": 3320 }, { "epoch": 0.7304519960409106, "grad_norm": 0.28503546118736267, "learning_rate": 8e-05, "loss": 1.4416, "step": 3321 }, { "epoch": 0.7306719454525459, "grad_norm": 0.2907847464084625, "learning_rate": 8e-05, "loss": 1.6119, "step": 3322 }, { "epoch": 0.7308918948641813, "grad_norm": 0.2849101722240448, "learning_rate": 8e-05, "loss": 1.5296, "step": 3323 }, { "epoch": 0.7311118442758165, "grad_norm": 0.2884346842765808, "learning_rate": 8e-05, "loss": 1.6141, "step": 3324 }, { "epoch": 0.7313317936874519, "grad_norm": 0.30081263184547424, "learning_rate": 8e-05, "loss": 1.7284, "step": 3325 }, { "epoch": 0.7315517430990872, "grad_norm": 0.2910812497138977, "learning_rate": 8e-05, "loss": 1.5262, "step": 3326 }, { "epoch": 0.7317716925107225, "grad_norm": 0.2986319661140442, "learning_rate": 8e-05, "loss": 1.6879, "step": 3327 }, { "epoch": 0.7319916419223579, "grad_norm": 0.2961769700050354, "learning_rate": 8e-05, "loss": 1.7522, "step": 3328 }, { "epoch": 0.7322115913339932, "grad_norm": 0.2825765907764435, "learning_rate": 8e-05, "loss": 1.7106, "step": 3329 }, { "epoch": 0.7324315407456286, "grad_norm": 0.2914576530456543, "learning_rate": 8e-05, "loss": 1.6491, "step": 3330 }, { "epoch": 0.7326514901572638, "grad_norm": 0.2786273658275604, "learning_rate": 8e-05, "loss": 1.7187, "step": 3331 }, { "epoch": 0.7328714395688991, "grad_norm": 0.27736160159111023, "learning_rate": 8e-05, "loss": 1.5219, "step": 3332 }, { "epoch": 0.7330913889805345, "grad_norm": 0.289495587348938, "learning_rate": 8e-05, "loss": 1.7318, "step": 3333 }, { "epoch": 0.7333113383921698, "grad_norm": 0.31175169348716736, "learning_rate": 8e-05, "loss": 1.8213, "step": 3334 }, { "epoch": 0.7335312878038052, "grad_norm": 0.29931631684303284, "learning_rate": 8e-05, "loss": 1.8126, "step": 3335 }, { "epoch": 0.7337512372154404, "grad_norm": 0.2750184237957001, "learning_rate": 8e-05, "loss": 1.5831, "step": 3336 }, { "epoch": 0.7339711866270757, "grad_norm": 0.30192333459854126, "learning_rate": 8e-05, "loss": 1.7635, "step": 3337 }, { "epoch": 0.7341911360387111, "grad_norm": 0.28221380710601807, "learning_rate": 8e-05, "loss": 1.5748, "step": 3338 }, { "epoch": 0.7344110854503464, "grad_norm": 0.2903328239917755, "learning_rate": 8e-05, "loss": 1.6313, "step": 3339 }, { "epoch": 0.7346310348619818, "grad_norm": 0.2983735501766205, "learning_rate": 8e-05, "loss": 1.6551, "step": 3340 }, { "epoch": 0.7348509842736171, "grad_norm": 0.29272475838661194, "learning_rate": 8e-05, "loss": 1.7012, "step": 3341 }, { "epoch": 0.7350709336852523, "grad_norm": 0.3180513083934784, "learning_rate": 8e-05, "loss": 1.7281, "step": 3342 }, { "epoch": 0.7352908830968877, "grad_norm": 0.3015908896923065, "learning_rate": 8e-05, "loss": 1.7734, "step": 3343 }, { "epoch": 0.735510832508523, "grad_norm": 0.303092360496521, "learning_rate": 8e-05, "loss": 1.708, "step": 3344 }, { "epoch": 0.7357307819201584, "grad_norm": 0.2860123813152313, "learning_rate": 8e-05, "loss": 1.7253, "step": 3345 }, { "epoch": 0.7359507313317937, "grad_norm": 0.2848500907421112, "learning_rate": 8e-05, "loss": 1.6338, "step": 3346 }, { "epoch": 0.7361706807434291, "grad_norm": 0.2954801321029663, "learning_rate": 8e-05, "loss": 1.7751, "step": 3347 }, { "epoch": 0.7363906301550643, "grad_norm": 0.36568450927734375, "learning_rate": 8e-05, "loss": 1.9713, "step": 3348 }, { "epoch": 0.7366105795666996, "grad_norm": 0.2932252585887909, "learning_rate": 8e-05, "loss": 1.6413, "step": 3349 }, { "epoch": 0.736830528978335, "grad_norm": 0.2913023829460144, "learning_rate": 8e-05, "loss": 1.6776, "step": 3350 }, { "epoch": 0.7370504783899703, "grad_norm": 0.30034127831459045, "learning_rate": 8e-05, "loss": 1.8318, "step": 3351 }, { "epoch": 0.7372704278016057, "grad_norm": 0.28994688391685486, "learning_rate": 8e-05, "loss": 1.6315, "step": 3352 }, { "epoch": 0.737490377213241, "grad_norm": 0.30393558740615845, "learning_rate": 8e-05, "loss": 1.7876, "step": 3353 }, { "epoch": 0.7377103266248762, "grad_norm": 0.3077211380004883, "learning_rate": 8e-05, "loss": 1.7739, "step": 3354 }, { "epoch": 0.7379302760365116, "grad_norm": 0.2963906526565552, "learning_rate": 8e-05, "loss": 1.7134, "step": 3355 }, { "epoch": 0.7381502254481469, "grad_norm": 0.30240267515182495, "learning_rate": 8e-05, "loss": 1.7058, "step": 3356 }, { "epoch": 0.7383701748597823, "grad_norm": 0.2841428518295288, "learning_rate": 8e-05, "loss": 1.5885, "step": 3357 }, { "epoch": 0.7385901242714176, "grad_norm": 0.2864632308483124, "learning_rate": 8e-05, "loss": 1.4941, "step": 3358 }, { "epoch": 0.7388100736830528, "grad_norm": 0.3023029565811157, "learning_rate": 8e-05, "loss": 1.7757, "step": 3359 }, { "epoch": 0.7390300230946882, "grad_norm": 0.29607269167900085, "learning_rate": 8e-05, "loss": 1.7288, "step": 3360 }, { "epoch": 0.7392499725063235, "grad_norm": 0.28938642144203186, "learning_rate": 8e-05, "loss": 1.6216, "step": 3361 }, { "epoch": 0.7394699219179589, "grad_norm": 0.2934838831424713, "learning_rate": 8e-05, "loss": 1.7762, "step": 3362 }, { "epoch": 0.7396898713295942, "grad_norm": 0.3048613369464874, "learning_rate": 8e-05, "loss": 1.7296, "step": 3363 }, { "epoch": 0.7399098207412295, "grad_norm": 0.2915562689304352, "learning_rate": 8e-05, "loss": 1.6688, "step": 3364 }, { "epoch": 0.7401297701528649, "grad_norm": 0.2817826271057129, "learning_rate": 8e-05, "loss": 1.7981, "step": 3365 }, { "epoch": 0.7403497195645001, "grad_norm": 0.31884127855300903, "learning_rate": 8e-05, "loss": 1.7536, "step": 3366 }, { "epoch": 0.7405696689761355, "grad_norm": 0.2934412956237793, "learning_rate": 8e-05, "loss": 1.6895, "step": 3367 }, { "epoch": 0.7407896183877708, "grad_norm": 0.279256671667099, "learning_rate": 8e-05, "loss": 1.681, "step": 3368 }, { "epoch": 0.7410095677994062, "grad_norm": 0.2802670896053314, "learning_rate": 8e-05, "loss": 1.6775, "step": 3369 }, { "epoch": 0.7412295172110415, "grad_norm": 0.28715866804122925, "learning_rate": 8e-05, "loss": 1.7328, "step": 3370 }, { "epoch": 0.7414494666226767, "grad_norm": 0.28366830945014954, "learning_rate": 8e-05, "loss": 1.6503, "step": 3371 }, { "epoch": 0.7416694160343121, "grad_norm": 0.27768293023109436, "learning_rate": 8e-05, "loss": 1.6495, "step": 3372 }, { "epoch": 0.7418893654459474, "grad_norm": 0.27918893098831177, "learning_rate": 8e-05, "loss": 1.5964, "step": 3373 }, { "epoch": 0.7421093148575828, "grad_norm": 0.31968846917152405, "learning_rate": 8e-05, "loss": 1.7707, "step": 3374 }, { "epoch": 0.7423292642692181, "grad_norm": 0.2908543348312378, "learning_rate": 8e-05, "loss": 1.7626, "step": 3375 }, { "epoch": 0.7425492136808534, "grad_norm": 0.2790231704711914, "learning_rate": 8e-05, "loss": 1.571, "step": 3376 }, { "epoch": 0.7427691630924887, "grad_norm": 0.2957472503185272, "learning_rate": 8e-05, "loss": 1.6063, "step": 3377 }, { "epoch": 0.742989112504124, "grad_norm": 0.2858772277832031, "learning_rate": 8e-05, "loss": 1.5614, "step": 3378 }, { "epoch": 0.7432090619157594, "grad_norm": 0.2928299605846405, "learning_rate": 8e-05, "loss": 1.6377, "step": 3379 }, { "epoch": 0.7434290113273947, "grad_norm": 0.28481486439704895, "learning_rate": 8e-05, "loss": 1.5916, "step": 3380 }, { "epoch": 0.74364896073903, "grad_norm": 0.30001428723335266, "learning_rate": 8e-05, "loss": 1.6408, "step": 3381 }, { "epoch": 0.7438689101506654, "grad_norm": 0.31076952815055847, "learning_rate": 8e-05, "loss": 1.7663, "step": 3382 }, { "epoch": 0.7440888595623006, "grad_norm": 0.2954659163951874, "learning_rate": 8e-05, "loss": 1.8063, "step": 3383 }, { "epoch": 0.744308808973936, "grad_norm": 0.2676610052585602, "learning_rate": 8e-05, "loss": 1.6039, "step": 3384 }, { "epoch": 0.7445287583855713, "grad_norm": 0.28663912415504456, "learning_rate": 8e-05, "loss": 1.486, "step": 3385 }, { "epoch": 0.7447487077972066, "grad_norm": 0.28007110953330994, "learning_rate": 8e-05, "loss": 1.5365, "step": 3386 }, { "epoch": 0.744968657208842, "grad_norm": 0.28811848163604736, "learning_rate": 8e-05, "loss": 1.6558, "step": 3387 }, { "epoch": 0.7451886066204773, "grad_norm": 0.313472181558609, "learning_rate": 8e-05, "loss": 1.7308, "step": 3388 }, { "epoch": 0.7454085560321126, "grad_norm": 0.28725314140319824, "learning_rate": 8e-05, "loss": 1.7269, "step": 3389 }, { "epoch": 0.7456285054437479, "grad_norm": 0.3030818998813629, "learning_rate": 8e-05, "loss": 1.6803, "step": 3390 }, { "epoch": 0.7458484548553833, "grad_norm": 0.34181302785873413, "learning_rate": 8e-05, "loss": 1.8658, "step": 3391 }, { "epoch": 0.7460684042670186, "grad_norm": 0.29316088557243347, "learning_rate": 8e-05, "loss": 1.6458, "step": 3392 }, { "epoch": 0.7462883536786539, "grad_norm": 0.3064456582069397, "learning_rate": 8e-05, "loss": 1.8798, "step": 3393 }, { "epoch": 0.7465083030902893, "grad_norm": 0.28297048807144165, "learning_rate": 8e-05, "loss": 1.6796, "step": 3394 }, { "epoch": 0.7467282525019245, "grad_norm": 0.2969253659248352, "learning_rate": 8e-05, "loss": 1.6369, "step": 3395 }, { "epoch": 0.7469482019135599, "grad_norm": 0.3293825089931488, "learning_rate": 8e-05, "loss": 1.9732, "step": 3396 }, { "epoch": 0.7471681513251952, "grad_norm": 0.2745886445045471, "learning_rate": 8e-05, "loss": 1.537, "step": 3397 }, { "epoch": 0.7473881007368305, "grad_norm": 0.28334876894950867, "learning_rate": 8e-05, "loss": 1.582, "step": 3398 }, { "epoch": 0.7476080501484659, "grad_norm": 0.29402390122413635, "learning_rate": 8e-05, "loss": 1.8205, "step": 3399 }, { "epoch": 0.7478279995601012, "grad_norm": 0.3036934733390808, "learning_rate": 8e-05, "loss": 1.6388, "step": 3400 }, { "epoch": 0.7480479489717365, "grad_norm": 0.2951960563659668, "learning_rate": 8e-05, "loss": 1.8011, "step": 3401 }, { "epoch": 0.7482678983833718, "grad_norm": 0.29489466547966003, "learning_rate": 8e-05, "loss": 1.8069, "step": 3402 }, { "epoch": 0.7484878477950071, "grad_norm": 0.2927097976207733, "learning_rate": 8e-05, "loss": 1.7201, "step": 3403 }, { "epoch": 0.7487077972066425, "grad_norm": 0.277259886264801, "learning_rate": 8e-05, "loss": 1.6, "step": 3404 }, { "epoch": 0.7489277466182778, "grad_norm": 0.2830856144428253, "learning_rate": 8e-05, "loss": 1.6689, "step": 3405 }, { "epoch": 0.7491476960299132, "grad_norm": 0.3067016899585724, "learning_rate": 8e-05, "loss": 1.8059, "step": 3406 }, { "epoch": 0.7493676454415484, "grad_norm": 0.2836841940879822, "learning_rate": 8e-05, "loss": 1.6509, "step": 3407 }, { "epoch": 0.7495875948531837, "grad_norm": 0.2890060842037201, "learning_rate": 8e-05, "loss": 1.7764, "step": 3408 }, { "epoch": 0.7498075442648191, "grad_norm": 0.29502347111701965, "learning_rate": 8e-05, "loss": 1.6978, "step": 3409 }, { "epoch": 0.7500274936764544, "grad_norm": 0.3138415217399597, "learning_rate": 8e-05, "loss": 1.6973, "step": 3410 }, { "epoch": 0.7502474430880898, "grad_norm": 0.2842028737068176, "learning_rate": 8e-05, "loss": 1.5305, "step": 3411 }, { "epoch": 0.750467392499725, "grad_norm": 0.3058598041534424, "learning_rate": 8e-05, "loss": 1.7835, "step": 3412 }, { "epoch": 0.7506873419113604, "grad_norm": 0.2859271466732025, "learning_rate": 8e-05, "loss": 1.6406, "step": 3413 }, { "epoch": 0.7509072913229957, "grad_norm": 0.2928830683231354, "learning_rate": 8e-05, "loss": 1.5819, "step": 3414 }, { "epoch": 0.751127240734631, "grad_norm": 0.3138205409049988, "learning_rate": 8e-05, "loss": 1.5155, "step": 3415 }, { "epoch": 0.7513471901462664, "grad_norm": 0.3159571886062622, "learning_rate": 8e-05, "loss": 1.8749, "step": 3416 }, { "epoch": 0.7515671395579017, "grad_norm": 0.2998178005218506, "learning_rate": 8e-05, "loss": 1.6651, "step": 3417 }, { "epoch": 0.751787088969537, "grad_norm": 0.30106493830680847, "learning_rate": 8e-05, "loss": 1.7626, "step": 3418 }, { "epoch": 0.7520070383811723, "grad_norm": 0.2966890037059784, "learning_rate": 8e-05, "loss": 1.6505, "step": 3419 }, { "epoch": 0.7522269877928076, "grad_norm": 0.28747057914733887, "learning_rate": 8e-05, "loss": 1.5537, "step": 3420 }, { "epoch": 0.752446937204443, "grad_norm": 0.3139256536960602, "learning_rate": 8e-05, "loss": 1.65, "step": 3421 }, { "epoch": 0.7526668866160783, "grad_norm": 0.2905026972293854, "learning_rate": 8e-05, "loss": 1.7456, "step": 3422 }, { "epoch": 0.7528868360277137, "grad_norm": 0.29663148522377014, "learning_rate": 8e-05, "loss": 1.6986, "step": 3423 }, { "epoch": 0.753106785439349, "grad_norm": 0.2942657470703125, "learning_rate": 8e-05, "loss": 1.7396, "step": 3424 }, { "epoch": 0.7533267348509842, "grad_norm": 0.2846404016017914, "learning_rate": 8e-05, "loss": 1.5673, "step": 3425 }, { "epoch": 0.7535466842626196, "grad_norm": 0.3121284544467926, "learning_rate": 8e-05, "loss": 1.7852, "step": 3426 }, { "epoch": 0.7537666336742549, "grad_norm": 0.3095207214355469, "learning_rate": 8e-05, "loss": 1.7747, "step": 3427 }, { "epoch": 0.7539865830858903, "grad_norm": 0.28364792466163635, "learning_rate": 8e-05, "loss": 1.6513, "step": 3428 }, { "epoch": 0.7542065324975256, "grad_norm": 0.2978142201900482, "learning_rate": 8e-05, "loss": 1.7227, "step": 3429 }, { "epoch": 0.7544264819091608, "grad_norm": 0.2860453724861145, "learning_rate": 8e-05, "loss": 1.5699, "step": 3430 }, { "epoch": 0.7546464313207962, "grad_norm": 0.30176058411598206, "learning_rate": 8e-05, "loss": 1.683, "step": 3431 }, { "epoch": 0.7548663807324315, "grad_norm": 0.2823827266693115, "learning_rate": 8e-05, "loss": 1.4892, "step": 3432 }, { "epoch": 0.7550863301440669, "grad_norm": 0.2879229784011841, "learning_rate": 8e-05, "loss": 1.5821, "step": 3433 }, { "epoch": 0.7553062795557022, "grad_norm": 0.29188740253448486, "learning_rate": 8e-05, "loss": 1.5946, "step": 3434 }, { "epoch": 0.7555262289673376, "grad_norm": 0.29899924993515015, "learning_rate": 8e-05, "loss": 1.7849, "step": 3435 }, { "epoch": 0.7557461783789728, "grad_norm": 0.29541707038879395, "learning_rate": 8e-05, "loss": 1.6732, "step": 3436 }, { "epoch": 0.7559661277906081, "grad_norm": 0.2920849919319153, "learning_rate": 8e-05, "loss": 1.5479, "step": 3437 }, { "epoch": 0.7561860772022435, "grad_norm": 0.3060014545917511, "learning_rate": 8e-05, "loss": 1.7094, "step": 3438 }, { "epoch": 0.7564060266138788, "grad_norm": 0.28043729066848755, "learning_rate": 8e-05, "loss": 1.4717, "step": 3439 }, { "epoch": 0.7566259760255142, "grad_norm": 0.29130029678344727, "learning_rate": 8e-05, "loss": 1.6355, "step": 3440 }, { "epoch": 0.7568459254371495, "grad_norm": 0.2969004213809967, "learning_rate": 8e-05, "loss": 1.6808, "step": 3441 }, { "epoch": 0.7570658748487847, "grad_norm": 0.3089723289012909, "learning_rate": 8e-05, "loss": 1.7368, "step": 3442 }, { "epoch": 0.7572858242604201, "grad_norm": 0.29811516404151917, "learning_rate": 8e-05, "loss": 1.6317, "step": 3443 }, { "epoch": 0.7575057736720554, "grad_norm": 0.2753433883190155, "learning_rate": 8e-05, "loss": 1.598, "step": 3444 }, { "epoch": 0.7577257230836908, "grad_norm": 0.2914634943008423, "learning_rate": 8e-05, "loss": 1.7117, "step": 3445 }, { "epoch": 0.7579456724953261, "grad_norm": 0.3008229732513428, "learning_rate": 8e-05, "loss": 1.7158, "step": 3446 }, { "epoch": 0.7581656219069614, "grad_norm": 0.297671377658844, "learning_rate": 8e-05, "loss": 1.793, "step": 3447 }, { "epoch": 0.7583855713185967, "grad_norm": 0.27893882989883423, "learning_rate": 8e-05, "loss": 1.6701, "step": 3448 }, { "epoch": 0.758605520730232, "grad_norm": 0.2874872386455536, "learning_rate": 8e-05, "loss": 1.5559, "step": 3449 }, { "epoch": 0.7588254701418674, "grad_norm": 0.30345162749290466, "learning_rate": 8e-05, "loss": 1.6925, "step": 3450 }, { "epoch": 0.7590454195535027, "grad_norm": 0.3018907308578491, "learning_rate": 8e-05, "loss": 1.6325, "step": 3451 }, { "epoch": 0.759265368965138, "grad_norm": 0.30021166801452637, "learning_rate": 8e-05, "loss": 1.6077, "step": 3452 }, { "epoch": 0.7594853183767734, "grad_norm": 0.2934636175632477, "learning_rate": 8e-05, "loss": 1.681, "step": 3453 }, { "epoch": 0.7597052677884086, "grad_norm": 0.30079886317253113, "learning_rate": 8e-05, "loss": 1.7378, "step": 3454 }, { "epoch": 0.759925217200044, "grad_norm": 0.29279422760009766, "learning_rate": 8e-05, "loss": 1.5918, "step": 3455 }, { "epoch": 0.7601451666116793, "grad_norm": 0.2980215847492218, "learning_rate": 8e-05, "loss": 1.7888, "step": 3456 }, { "epoch": 0.7603651160233147, "grad_norm": 0.27554240822792053, "learning_rate": 8e-05, "loss": 1.5351, "step": 3457 }, { "epoch": 0.76058506543495, "grad_norm": 0.2831701338291168, "learning_rate": 8e-05, "loss": 1.7558, "step": 3458 }, { "epoch": 0.7608050148465852, "grad_norm": 0.29401731491088867, "learning_rate": 8e-05, "loss": 1.69, "step": 3459 }, { "epoch": 0.7610249642582206, "grad_norm": 0.29017552733421326, "learning_rate": 8e-05, "loss": 1.6145, "step": 3460 }, { "epoch": 0.7612449136698559, "grad_norm": 0.2881031036376953, "learning_rate": 8e-05, "loss": 1.6464, "step": 3461 }, { "epoch": 0.7614648630814913, "grad_norm": 0.3012145757675171, "learning_rate": 8e-05, "loss": 1.6904, "step": 3462 }, { "epoch": 0.7616848124931266, "grad_norm": 0.29406794905662537, "learning_rate": 8e-05, "loss": 1.6375, "step": 3463 }, { "epoch": 0.7619047619047619, "grad_norm": 0.28518983721733093, "learning_rate": 8e-05, "loss": 1.5478, "step": 3464 }, { "epoch": 0.7621247113163973, "grad_norm": 0.3052254021167755, "learning_rate": 8e-05, "loss": 1.8076, "step": 3465 }, { "epoch": 0.7623446607280325, "grad_norm": 0.2873866856098175, "learning_rate": 8e-05, "loss": 1.7547, "step": 3466 }, { "epoch": 0.7625646101396679, "grad_norm": 0.28812938928604126, "learning_rate": 8e-05, "loss": 1.6502, "step": 3467 }, { "epoch": 0.7627845595513032, "grad_norm": 0.28548526763916016, "learning_rate": 8e-05, "loss": 1.7744, "step": 3468 }, { "epoch": 0.7630045089629385, "grad_norm": 0.30266743898391724, "learning_rate": 8e-05, "loss": 1.7142, "step": 3469 }, { "epoch": 0.7632244583745739, "grad_norm": 0.29223671555519104, "learning_rate": 8e-05, "loss": 1.5969, "step": 3470 }, { "epoch": 0.7634444077862091, "grad_norm": 0.2857435643672943, "learning_rate": 8e-05, "loss": 1.6398, "step": 3471 }, { "epoch": 0.7636643571978445, "grad_norm": 0.28153562545776367, "learning_rate": 8e-05, "loss": 1.5967, "step": 3472 }, { "epoch": 0.7638843066094798, "grad_norm": 0.28425905108451843, "learning_rate": 8e-05, "loss": 1.5901, "step": 3473 }, { "epoch": 0.7641042560211151, "grad_norm": 0.3140738308429718, "learning_rate": 8e-05, "loss": 1.772, "step": 3474 }, { "epoch": 0.7643242054327505, "grad_norm": 0.2958965301513672, "learning_rate": 8e-05, "loss": 1.8743, "step": 3475 }, { "epoch": 0.7645441548443858, "grad_norm": 0.3059069514274597, "learning_rate": 8e-05, "loss": 1.6789, "step": 3476 }, { "epoch": 0.7647641042560211, "grad_norm": 0.3015092611312866, "learning_rate": 8e-05, "loss": 1.7492, "step": 3477 }, { "epoch": 0.7649840536676564, "grad_norm": 0.2840091586112976, "learning_rate": 8e-05, "loss": 1.6442, "step": 3478 }, { "epoch": 0.7652040030792918, "grad_norm": 0.2928544282913208, "learning_rate": 8e-05, "loss": 1.5534, "step": 3479 }, { "epoch": 0.7654239524909271, "grad_norm": 0.2872399389743805, "learning_rate": 8e-05, "loss": 1.7049, "step": 3480 }, { "epoch": 0.7656439019025624, "grad_norm": 0.29002323746681213, "learning_rate": 8e-05, "loss": 1.6036, "step": 3481 }, { "epoch": 0.7658638513141978, "grad_norm": 0.28830286860466003, "learning_rate": 8e-05, "loss": 1.6805, "step": 3482 }, { "epoch": 0.766083800725833, "grad_norm": 0.2776120901107788, "learning_rate": 8e-05, "loss": 1.609, "step": 3483 }, { "epoch": 0.7663037501374684, "grad_norm": 0.2959437072277069, "learning_rate": 8e-05, "loss": 1.6657, "step": 3484 }, { "epoch": 0.7665236995491037, "grad_norm": 0.29850873351097107, "learning_rate": 8e-05, "loss": 1.4705, "step": 3485 }, { "epoch": 0.766743648960739, "grad_norm": 0.2813439667224884, "learning_rate": 8e-05, "loss": 1.5541, "step": 3486 }, { "epoch": 0.7669635983723744, "grad_norm": 0.3008776009082794, "learning_rate": 8e-05, "loss": 1.7506, "step": 3487 }, { "epoch": 0.7671835477840097, "grad_norm": 0.28104785084724426, "learning_rate": 8e-05, "loss": 1.5551, "step": 3488 }, { "epoch": 0.767403497195645, "grad_norm": 0.28244099020957947, "learning_rate": 8e-05, "loss": 1.644, "step": 3489 }, { "epoch": 0.7676234466072803, "grad_norm": 0.2998720407485962, "learning_rate": 8e-05, "loss": 1.7757, "step": 3490 }, { "epoch": 0.7678433960189156, "grad_norm": 0.28794363141059875, "learning_rate": 8e-05, "loss": 1.5814, "step": 3491 }, { "epoch": 0.768063345430551, "grad_norm": 0.2936362326145172, "learning_rate": 8e-05, "loss": 1.602, "step": 3492 }, { "epoch": 0.7682832948421863, "grad_norm": 0.282718300819397, "learning_rate": 8e-05, "loss": 1.5837, "step": 3493 }, { "epoch": 0.7685032442538217, "grad_norm": 0.2850697934627533, "learning_rate": 8e-05, "loss": 1.708, "step": 3494 }, { "epoch": 0.7687231936654569, "grad_norm": 0.3146105110645294, "learning_rate": 8e-05, "loss": 1.7864, "step": 3495 }, { "epoch": 0.7689431430770922, "grad_norm": 0.295669823884964, "learning_rate": 8e-05, "loss": 1.7563, "step": 3496 }, { "epoch": 0.7691630924887276, "grad_norm": 0.2910419702529907, "learning_rate": 8e-05, "loss": 1.7159, "step": 3497 }, { "epoch": 0.7693830419003629, "grad_norm": 0.2934516668319702, "learning_rate": 8e-05, "loss": 1.7183, "step": 3498 }, { "epoch": 0.7696029913119983, "grad_norm": 0.2876817584037781, "learning_rate": 8e-05, "loss": 1.652, "step": 3499 }, { "epoch": 0.7698229407236336, "grad_norm": 0.2993420660495758, "learning_rate": 8e-05, "loss": 1.8512, "step": 3500 }, { "epoch": 0.7700428901352689, "grad_norm": 0.28241509199142456, "learning_rate": 8e-05, "loss": 1.6526, "step": 3501 }, { "epoch": 0.7702628395469042, "grad_norm": 0.2991597056388855, "learning_rate": 8e-05, "loss": 1.656, "step": 3502 }, { "epoch": 0.7704827889585395, "grad_norm": 0.26715072989463806, "learning_rate": 8e-05, "loss": 1.4417, "step": 3503 }, { "epoch": 0.7707027383701749, "grad_norm": 0.28751102089881897, "learning_rate": 8e-05, "loss": 1.6471, "step": 3504 }, { "epoch": 0.7709226877818102, "grad_norm": 0.297297865152359, "learning_rate": 8e-05, "loss": 1.6943, "step": 3505 }, { "epoch": 0.7711426371934456, "grad_norm": 0.2984234690666199, "learning_rate": 8e-05, "loss": 1.6245, "step": 3506 }, { "epoch": 0.7713625866050808, "grad_norm": 0.29237473011016846, "learning_rate": 8e-05, "loss": 1.6624, "step": 3507 }, { "epoch": 0.7715825360167161, "grad_norm": 0.2652399241924286, "learning_rate": 8e-05, "loss": 1.3897, "step": 3508 }, { "epoch": 0.7718024854283515, "grad_norm": 0.28701311349868774, "learning_rate": 8e-05, "loss": 1.5524, "step": 3509 }, { "epoch": 0.7720224348399868, "grad_norm": 0.2850416600704193, "learning_rate": 8e-05, "loss": 1.6508, "step": 3510 }, { "epoch": 0.7722423842516222, "grad_norm": 0.30162402987480164, "learning_rate": 8e-05, "loss": 1.7517, "step": 3511 }, { "epoch": 0.7724623336632574, "grad_norm": 0.29350200295448303, "learning_rate": 8e-05, "loss": 1.7633, "step": 3512 }, { "epoch": 0.7726822830748927, "grad_norm": 0.28733333945274353, "learning_rate": 8e-05, "loss": 1.5516, "step": 3513 }, { "epoch": 0.7729022324865281, "grad_norm": 0.32462671399116516, "learning_rate": 8e-05, "loss": 1.7854, "step": 3514 }, { "epoch": 0.7731221818981634, "grad_norm": 0.3113280236721039, "learning_rate": 8e-05, "loss": 1.8449, "step": 3515 }, { "epoch": 0.7733421313097988, "grad_norm": 0.2827896773815155, "learning_rate": 8e-05, "loss": 1.7656, "step": 3516 }, { "epoch": 0.7735620807214341, "grad_norm": 0.2928074598312378, "learning_rate": 8e-05, "loss": 1.6439, "step": 3517 }, { "epoch": 0.7737820301330693, "grad_norm": 0.2791458070278168, "learning_rate": 8e-05, "loss": 1.618, "step": 3518 }, { "epoch": 0.7740019795447047, "grad_norm": 0.3217780590057373, "learning_rate": 8e-05, "loss": 1.7436, "step": 3519 }, { "epoch": 0.77422192895634, "grad_norm": 0.28555285930633545, "learning_rate": 8e-05, "loss": 1.6563, "step": 3520 }, { "epoch": 0.7744418783679754, "grad_norm": 0.30171915888786316, "learning_rate": 8e-05, "loss": 1.5573, "step": 3521 }, { "epoch": 0.7746618277796107, "grad_norm": 0.29558464884757996, "learning_rate": 8e-05, "loss": 1.7783, "step": 3522 }, { "epoch": 0.7748817771912461, "grad_norm": 0.28546035289764404, "learning_rate": 8e-05, "loss": 1.5793, "step": 3523 }, { "epoch": 0.7751017266028813, "grad_norm": 0.30701640248298645, "learning_rate": 8e-05, "loss": 1.8204, "step": 3524 }, { "epoch": 0.7753216760145166, "grad_norm": 0.2884214222431183, "learning_rate": 8e-05, "loss": 1.6179, "step": 3525 }, { "epoch": 0.775541625426152, "grad_norm": 0.3010881245136261, "learning_rate": 8e-05, "loss": 1.7521, "step": 3526 }, { "epoch": 0.7757615748377873, "grad_norm": 0.2960352301597595, "learning_rate": 8e-05, "loss": 1.6982, "step": 3527 }, { "epoch": 0.7759815242494227, "grad_norm": 0.3347279131412506, "learning_rate": 8e-05, "loss": 1.8114, "step": 3528 }, { "epoch": 0.776201473661058, "grad_norm": 0.298384428024292, "learning_rate": 8e-05, "loss": 1.676, "step": 3529 }, { "epoch": 0.7764214230726932, "grad_norm": 0.30956321954727173, "learning_rate": 8e-05, "loss": 1.7505, "step": 3530 }, { "epoch": 0.7766413724843286, "grad_norm": 0.29873716831207275, "learning_rate": 8e-05, "loss": 1.6829, "step": 3531 }, { "epoch": 0.7768613218959639, "grad_norm": 0.2934054136276245, "learning_rate": 8e-05, "loss": 1.6946, "step": 3532 }, { "epoch": 0.7770812713075993, "grad_norm": 0.3284706175327301, "learning_rate": 8e-05, "loss": 1.881, "step": 3533 }, { "epoch": 0.7773012207192346, "grad_norm": 0.29005488753318787, "learning_rate": 8e-05, "loss": 1.6785, "step": 3534 }, { "epoch": 0.7775211701308699, "grad_norm": 0.2805337607860565, "learning_rate": 8e-05, "loss": 1.6025, "step": 3535 }, { "epoch": 0.7777411195425052, "grad_norm": 0.308052122592926, "learning_rate": 8e-05, "loss": 1.8216, "step": 3536 }, { "epoch": 0.7779610689541405, "grad_norm": 0.2906242907047272, "learning_rate": 8e-05, "loss": 1.7335, "step": 3537 }, { "epoch": 0.7781810183657759, "grad_norm": 0.28683236241340637, "learning_rate": 8e-05, "loss": 1.6598, "step": 3538 }, { "epoch": 0.7784009677774112, "grad_norm": 0.3011239469051361, "learning_rate": 8e-05, "loss": 1.5983, "step": 3539 }, { "epoch": 0.7786209171890465, "grad_norm": 0.2807599604129791, "learning_rate": 8e-05, "loss": 1.5252, "step": 3540 }, { "epoch": 0.7788408666006819, "grad_norm": 0.3378777503967285, "learning_rate": 8e-05, "loss": 1.6314, "step": 3541 }, { "epoch": 0.7790608160123171, "grad_norm": 0.28623417019844055, "learning_rate": 8e-05, "loss": 1.7219, "step": 3542 }, { "epoch": 0.7792807654239525, "grad_norm": 0.2920507788658142, "learning_rate": 8e-05, "loss": 1.5725, "step": 3543 }, { "epoch": 0.7795007148355878, "grad_norm": 0.3120638132095337, "learning_rate": 8e-05, "loss": 1.637, "step": 3544 }, { "epoch": 0.7797206642472232, "grad_norm": 0.29561811685562134, "learning_rate": 8e-05, "loss": 1.6626, "step": 3545 }, { "epoch": 0.7799406136588585, "grad_norm": 0.299492746591568, "learning_rate": 8e-05, "loss": 1.7206, "step": 3546 }, { "epoch": 0.7801605630704938, "grad_norm": 0.3011503517627716, "learning_rate": 8e-05, "loss": 1.695, "step": 3547 }, { "epoch": 0.7803805124821291, "grad_norm": 0.3122527003288269, "learning_rate": 8e-05, "loss": 1.8032, "step": 3548 }, { "epoch": 0.7806004618937644, "grad_norm": 0.28565728664398193, "learning_rate": 8e-05, "loss": 1.6263, "step": 3549 }, { "epoch": 0.7808204113053998, "grad_norm": 0.28905171155929565, "learning_rate": 8e-05, "loss": 1.7008, "step": 3550 }, { "epoch": 0.7810403607170351, "grad_norm": 0.2950139343738556, "learning_rate": 8e-05, "loss": 1.6454, "step": 3551 }, { "epoch": 0.7812603101286704, "grad_norm": 0.29627618193626404, "learning_rate": 8e-05, "loss": 1.6792, "step": 3552 }, { "epoch": 0.7814802595403058, "grad_norm": 0.28478100895881653, "learning_rate": 8e-05, "loss": 1.65, "step": 3553 }, { "epoch": 0.781700208951941, "grad_norm": 0.2886268198490143, "learning_rate": 8e-05, "loss": 1.655, "step": 3554 }, { "epoch": 0.7819201583635764, "grad_norm": 0.30898573994636536, "learning_rate": 8e-05, "loss": 1.6732, "step": 3555 }, { "epoch": 0.7821401077752117, "grad_norm": 0.2944103479385376, "learning_rate": 8e-05, "loss": 1.6625, "step": 3556 }, { "epoch": 0.782360057186847, "grad_norm": 0.2840811610221863, "learning_rate": 8e-05, "loss": 1.7059, "step": 3557 }, { "epoch": 0.7825800065984824, "grad_norm": 0.31668660044670105, "learning_rate": 8e-05, "loss": 1.7343, "step": 3558 }, { "epoch": 0.7827999560101176, "grad_norm": 0.2928016483783722, "learning_rate": 8e-05, "loss": 1.5946, "step": 3559 }, { "epoch": 0.783019905421753, "grad_norm": 0.30465641617774963, "learning_rate": 8e-05, "loss": 1.8133, "step": 3560 }, { "epoch": 0.7832398548333883, "grad_norm": 0.3166663646697998, "learning_rate": 8e-05, "loss": 1.7594, "step": 3561 }, { "epoch": 0.7834598042450236, "grad_norm": 0.2896936535835266, "learning_rate": 8e-05, "loss": 1.7585, "step": 3562 }, { "epoch": 0.783679753656659, "grad_norm": 0.313091903924942, "learning_rate": 8e-05, "loss": 1.8129, "step": 3563 }, { "epoch": 0.7838997030682943, "grad_norm": 0.306034117937088, "learning_rate": 8e-05, "loss": 1.6265, "step": 3564 }, { "epoch": 0.7841196524799297, "grad_norm": 0.2841499149799347, "learning_rate": 8e-05, "loss": 1.7271, "step": 3565 }, { "epoch": 0.7843396018915649, "grad_norm": 0.29752469062805176, "learning_rate": 8e-05, "loss": 1.5481, "step": 3566 }, { "epoch": 0.7845595513032002, "grad_norm": 0.29329928755760193, "learning_rate": 8e-05, "loss": 1.7561, "step": 3567 }, { "epoch": 0.7847795007148356, "grad_norm": 0.290488600730896, "learning_rate": 8e-05, "loss": 1.7908, "step": 3568 }, { "epoch": 0.7849994501264709, "grad_norm": 0.2861136198043823, "learning_rate": 8e-05, "loss": 1.6646, "step": 3569 }, { "epoch": 0.7852193995381063, "grad_norm": 0.28423643112182617, "learning_rate": 8e-05, "loss": 1.5909, "step": 3570 }, { "epoch": 0.7854393489497415, "grad_norm": 0.3002358675003052, "learning_rate": 8e-05, "loss": 1.6773, "step": 3571 }, { "epoch": 0.7856592983613769, "grad_norm": 0.2944103479385376, "learning_rate": 8e-05, "loss": 1.5379, "step": 3572 }, { "epoch": 0.7858792477730122, "grad_norm": 0.28983041644096375, "learning_rate": 8e-05, "loss": 1.5886, "step": 3573 }, { "epoch": 0.7860991971846475, "grad_norm": 0.297059565782547, "learning_rate": 8e-05, "loss": 1.6874, "step": 3574 }, { "epoch": 0.7863191465962829, "grad_norm": 0.2881814241409302, "learning_rate": 8e-05, "loss": 1.6546, "step": 3575 }, { "epoch": 0.7865390960079182, "grad_norm": 0.30335813760757446, "learning_rate": 8e-05, "loss": 1.6688, "step": 3576 }, { "epoch": 0.7867590454195535, "grad_norm": 0.2669346332550049, "learning_rate": 8e-05, "loss": 1.414, "step": 3577 }, { "epoch": 0.7869789948311888, "grad_norm": 0.29250532388687134, "learning_rate": 8e-05, "loss": 1.6435, "step": 3578 }, { "epoch": 0.7871989442428241, "grad_norm": 0.3067176043987274, "learning_rate": 8e-05, "loss": 1.7231, "step": 3579 }, { "epoch": 0.7874188936544595, "grad_norm": 0.2805663049221039, "learning_rate": 8e-05, "loss": 1.5512, "step": 3580 }, { "epoch": 0.7876388430660948, "grad_norm": 0.2998615503311157, "learning_rate": 8e-05, "loss": 1.6355, "step": 3581 }, { "epoch": 0.7878587924777302, "grad_norm": 0.28142982721328735, "learning_rate": 8e-05, "loss": 1.605, "step": 3582 }, { "epoch": 0.7880787418893654, "grad_norm": 0.30738621950149536, "learning_rate": 8e-05, "loss": 1.6741, "step": 3583 }, { "epoch": 0.7882986913010007, "grad_norm": 0.3007453382015228, "learning_rate": 8e-05, "loss": 1.8258, "step": 3584 }, { "epoch": 0.7885186407126361, "grad_norm": 0.29813793301582336, "learning_rate": 8e-05, "loss": 1.6551, "step": 3585 }, { "epoch": 0.7887385901242714, "grad_norm": 0.29949578642845154, "learning_rate": 8e-05, "loss": 1.6033, "step": 3586 }, { "epoch": 0.7889585395359068, "grad_norm": 0.2934853136539459, "learning_rate": 8e-05, "loss": 1.5856, "step": 3587 }, { "epoch": 0.7891784889475421, "grad_norm": 0.3007175326347351, "learning_rate": 8e-05, "loss": 1.5335, "step": 3588 }, { "epoch": 0.7893984383591773, "grad_norm": 0.2819620370864868, "learning_rate": 8e-05, "loss": 1.6833, "step": 3589 }, { "epoch": 0.7896183877708127, "grad_norm": 0.27381354570388794, "learning_rate": 8e-05, "loss": 1.5009, "step": 3590 }, { "epoch": 0.789838337182448, "grad_norm": 0.2849341928958893, "learning_rate": 8e-05, "loss": 1.7086, "step": 3591 }, { "epoch": 0.7900582865940834, "grad_norm": 0.3121960759162903, "learning_rate": 8e-05, "loss": 1.8541, "step": 3592 }, { "epoch": 0.7902782360057187, "grad_norm": 0.28709515929222107, "learning_rate": 8e-05, "loss": 1.6198, "step": 3593 }, { "epoch": 0.7904981854173541, "grad_norm": 0.29474472999572754, "learning_rate": 8e-05, "loss": 1.6322, "step": 3594 }, { "epoch": 0.7907181348289893, "grad_norm": 0.2754751443862915, "learning_rate": 8e-05, "loss": 1.5421, "step": 3595 }, { "epoch": 0.7909380842406246, "grad_norm": 0.3164171576499939, "learning_rate": 8e-05, "loss": 1.8512, "step": 3596 }, { "epoch": 0.79115803365226, "grad_norm": 0.2999406158924103, "learning_rate": 8e-05, "loss": 1.6801, "step": 3597 }, { "epoch": 0.7913779830638953, "grad_norm": 0.2944709360599518, "learning_rate": 8e-05, "loss": 1.6419, "step": 3598 }, { "epoch": 0.7915979324755307, "grad_norm": 0.28523892164230347, "learning_rate": 8e-05, "loss": 1.4772, "step": 3599 }, { "epoch": 0.791817881887166, "grad_norm": 0.3085593581199646, "learning_rate": 8e-05, "loss": 1.643, "step": 3600 }, { "epoch": 0.7920378312988012, "grad_norm": 0.2771039605140686, "learning_rate": 8e-05, "loss": 1.701, "step": 3601 }, { "epoch": 0.7922577807104366, "grad_norm": 0.2826726734638214, "learning_rate": 8e-05, "loss": 1.5694, "step": 3602 }, { "epoch": 0.7924777301220719, "grad_norm": 0.28886765241622925, "learning_rate": 8e-05, "loss": 1.6117, "step": 3603 }, { "epoch": 0.7926976795337073, "grad_norm": 0.30170392990112305, "learning_rate": 8e-05, "loss": 1.7294, "step": 3604 }, { "epoch": 0.7929176289453426, "grad_norm": 0.2870231866836548, "learning_rate": 8e-05, "loss": 1.5244, "step": 3605 }, { "epoch": 0.7931375783569778, "grad_norm": 0.28652969002723694, "learning_rate": 8e-05, "loss": 1.5694, "step": 3606 }, { "epoch": 0.7933575277686132, "grad_norm": 0.2963985204696655, "learning_rate": 8e-05, "loss": 1.7553, "step": 3607 }, { "epoch": 0.7935774771802485, "grad_norm": 0.302688866853714, "learning_rate": 8e-05, "loss": 1.689, "step": 3608 }, { "epoch": 0.7937974265918839, "grad_norm": 0.32751408219337463, "learning_rate": 8e-05, "loss": 1.8021, "step": 3609 }, { "epoch": 0.7940173760035192, "grad_norm": 0.3023033142089844, "learning_rate": 8e-05, "loss": 1.7521, "step": 3610 }, { "epoch": 0.7942373254151545, "grad_norm": 0.2758845090866089, "learning_rate": 8e-05, "loss": 1.5365, "step": 3611 }, { "epoch": 0.7944572748267898, "grad_norm": 0.2906113564968109, "learning_rate": 8e-05, "loss": 1.7001, "step": 3612 }, { "epoch": 0.7946772242384251, "grad_norm": 0.32935893535614014, "learning_rate": 8e-05, "loss": 1.8016, "step": 3613 }, { "epoch": 0.7948971736500605, "grad_norm": 0.29360225796699524, "learning_rate": 8e-05, "loss": 1.5381, "step": 3614 }, { "epoch": 0.7951171230616958, "grad_norm": 0.3403272330760956, "learning_rate": 8e-05, "loss": 1.6639, "step": 3615 }, { "epoch": 0.7953370724733312, "grad_norm": 0.27274513244628906, "learning_rate": 8e-05, "loss": 1.5926, "step": 3616 }, { "epoch": 0.7955570218849665, "grad_norm": 0.2820468544960022, "learning_rate": 8e-05, "loss": 1.5791, "step": 3617 }, { "epoch": 0.7957769712966017, "grad_norm": 0.3011898398399353, "learning_rate": 8e-05, "loss": 1.7195, "step": 3618 }, { "epoch": 0.7959969207082371, "grad_norm": 0.2950305938720703, "learning_rate": 8e-05, "loss": 1.6335, "step": 3619 }, { "epoch": 0.7962168701198724, "grad_norm": 0.27936869859695435, "learning_rate": 8e-05, "loss": 1.5635, "step": 3620 }, { "epoch": 0.7964368195315078, "grad_norm": 0.3040826916694641, "learning_rate": 8e-05, "loss": 1.672, "step": 3621 }, { "epoch": 0.7966567689431431, "grad_norm": 0.3088817894458771, "learning_rate": 8e-05, "loss": 1.7677, "step": 3622 }, { "epoch": 0.7968767183547784, "grad_norm": 0.28857895731925964, "learning_rate": 8e-05, "loss": 1.6923, "step": 3623 }, { "epoch": 0.7970966677664137, "grad_norm": 0.3066072165966034, "learning_rate": 8e-05, "loss": 1.5091, "step": 3624 }, { "epoch": 0.797316617178049, "grad_norm": 0.3056313991546631, "learning_rate": 8e-05, "loss": 1.8183, "step": 3625 }, { "epoch": 0.7975365665896844, "grad_norm": 0.27834954857826233, "learning_rate": 8e-05, "loss": 1.6914, "step": 3626 }, { "epoch": 0.7977565160013197, "grad_norm": 0.2966832220554352, "learning_rate": 8e-05, "loss": 1.6422, "step": 3627 }, { "epoch": 0.797976465412955, "grad_norm": 0.31216248869895935, "learning_rate": 8e-05, "loss": 1.7629, "step": 3628 }, { "epoch": 0.7981964148245904, "grad_norm": 0.3048402965068817, "learning_rate": 8e-05, "loss": 1.7893, "step": 3629 }, { "epoch": 0.7984163642362256, "grad_norm": 0.3240011930465698, "learning_rate": 8e-05, "loss": 1.7933, "step": 3630 }, { "epoch": 0.798636313647861, "grad_norm": 0.3020577132701874, "learning_rate": 8e-05, "loss": 1.6685, "step": 3631 }, { "epoch": 0.7988562630594963, "grad_norm": 0.3189142644405365, "learning_rate": 8e-05, "loss": 1.7346, "step": 3632 }, { "epoch": 0.7990762124711316, "grad_norm": 0.2925734519958496, "learning_rate": 8e-05, "loss": 1.6465, "step": 3633 }, { "epoch": 0.799296161882767, "grad_norm": 0.32868942618370056, "learning_rate": 8e-05, "loss": 1.6621, "step": 3634 }, { "epoch": 0.7995161112944023, "grad_norm": 0.28033214807510376, "learning_rate": 8e-05, "loss": 1.6006, "step": 3635 }, { "epoch": 0.7997360607060376, "grad_norm": 0.28859859704971313, "learning_rate": 8e-05, "loss": 1.7644, "step": 3636 }, { "epoch": 0.7999560101176729, "grad_norm": 0.2980223000049591, "learning_rate": 8e-05, "loss": 1.699, "step": 3637 }, { "epoch": 0.8001759595293083, "grad_norm": 0.288064181804657, "learning_rate": 8e-05, "loss": 1.6435, "step": 3638 }, { "epoch": 0.8003959089409436, "grad_norm": 0.29023095965385437, "learning_rate": 8e-05, "loss": 1.8842, "step": 3639 }, { "epoch": 0.8006158583525789, "grad_norm": 0.2843625247478485, "learning_rate": 8e-05, "loss": 1.6155, "step": 3640 }, { "epoch": 0.8008358077642143, "grad_norm": 0.292680948972702, "learning_rate": 8e-05, "loss": 1.7985, "step": 3641 }, { "epoch": 0.8010557571758495, "grad_norm": 0.2843727171421051, "learning_rate": 8e-05, "loss": 1.8262, "step": 3642 }, { "epoch": 0.8012757065874849, "grad_norm": 0.3109283745288849, "learning_rate": 8e-05, "loss": 1.719, "step": 3643 }, { "epoch": 0.8014956559991202, "grad_norm": 0.2911458909511566, "learning_rate": 8e-05, "loss": 1.6543, "step": 3644 }, { "epoch": 0.8017156054107555, "grad_norm": 0.29368388652801514, "learning_rate": 8e-05, "loss": 1.6606, "step": 3645 }, { "epoch": 0.8019355548223909, "grad_norm": 0.27842018008232117, "learning_rate": 8e-05, "loss": 1.6225, "step": 3646 }, { "epoch": 0.8021555042340262, "grad_norm": 0.2872051000595093, "learning_rate": 8e-05, "loss": 1.5608, "step": 3647 }, { "epoch": 0.8023754536456615, "grad_norm": 0.2749423086643219, "learning_rate": 8e-05, "loss": 1.579, "step": 3648 }, { "epoch": 0.8025954030572968, "grad_norm": 0.28996387124061584, "learning_rate": 8e-05, "loss": 1.6742, "step": 3649 }, { "epoch": 0.8028153524689321, "grad_norm": 0.3018381595611572, "learning_rate": 8e-05, "loss": 1.7604, "step": 3650 }, { "epoch": 0.8030353018805675, "grad_norm": 0.29183822870254517, "learning_rate": 8e-05, "loss": 1.7273, "step": 3651 }, { "epoch": 0.8032552512922028, "grad_norm": 0.2697616517543793, "learning_rate": 8e-05, "loss": 1.4865, "step": 3652 }, { "epoch": 0.8034752007038382, "grad_norm": 0.29433682560920715, "learning_rate": 8e-05, "loss": 1.7453, "step": 3653 }, { "epoch": 0.8036951501154734, "grad_norm": 0.3100878596305847, "learning_rate": 8e-05, "loss": 1.6805, "step": 3654 }, { "epoch": 0.8039150995271087, "grad_norm": 0.3025181293487549, "learning_rate": 8e-05, "loss": 1.6999, "step": 3655 }, { "epoch": 0.8041350489387441, "grad_norm": 0.29453882575035095, "learning_rate": 8e-05, "loss": 1.5694, "step": 3656 }, { "epoch": 0.8043549983503794, "grad_norm": 0.30523207783699036, "learning_rate": 8e-05, "loss": 1.6819, "step": 3657 }, { "epoch": 0.8045749477620148, "grad_norm": 0.28335437178611755, "learning_rate": 8e-05, "loss": 1.7137, "step": 3658 }, { "epoch": 0.80479489717365, "grad_norm": 0.30658838152885437, "learning_rate": 8e-05, "loss": 1.6486, "step": 3659 }, { "epoch": 0.8050148465852854, "grad_norm": 0.30316153168678284, "learning_rate": 8e-05, "loss": 1.7012, "step": 3660 }, { "epoch": 0.8052347959969207, "grad_norm": 0.2958810329437256, "learning_rate": 8e-05, "loss": 1.5989, "step": 3661 }, { "epoch": 0.805454745408556, "grad_norm": 0.2858383059501648, "learning_rate": 8e-05, "loss": 1.6466, "step": 3662 }, { "epoch": 0.8056746948201914, "grad_norm": 0.3087503910064697, "learning_rate": 8e-05, "loss": 1.7894, "step": 3663 }, { "epoch": 0.8058946442318267, "grad_norm": 0.29160773754119873, "learning_rate": 8e-05, "loss": 1.5268, "step": 3664 }, { "epoch": 0.806114593643462, "grad_norm": 0.3402816653251648, "learning_rate": 8e-05, "loss": 1.728, "step": 3665 }, { "epoch": 0.8063345430550973, "grad_norm": 0.2856731712818146, "learning_rate": 8e-05, "loss": 1.5064, "step": 3666 }, { "epoch": 0.8065544924667326, "grad_norm": 0.29663363099098206, "learning_rate": 8e-05, "loss": 1.7799, "step": 3667 }, { "epoch": 0.806774441878368, "grad_norm": 0.2893684208393097, "learning_rate": 8e-05, "loss": 1.6961, "step": 3668 }, { "epoch": 0.8069943912900033, "grad_norm": 0.2832998037338257, "learning_rate": 8e-05, "loss": 1.514, "step": 3669 }, { "epoch": 0.8072143407016387, "grad_norm": 0.3098543584346771, "learning_rate": 8e-05, "loss": 1.7395, "step": 3670 }, { "epoch": 0.807434290113274, "grad_norm": 0.30097678303718567, "learning_rate": 8e-05, "loss": 1.6402, "step": 3671 }, { "epoch": 0.8076542395249092, "grad_norm": 0.3011092245578766, "learning_rate": 8e-05, "loss": 1.6554, "step": 3672 }, { "epoch": 0.8078741889365446, "grad_norm": 0.3125768303871155, "learning_rate": 8e-05, "loss": 1.6503, "step": 3673 }, { "epoch": 0.8080941383481799, "grad_norm": 0.3440975546836853, "learning_rate": 8e-05, "loss": 1.6666, "step": 3674 }, { "epoch": 0.8083140877598153, "grad_norm": 0.3339664340019226, "learning_rate": 8e-05, "loss": 1.7333, "step": 3675 }, { "epoch": 0.8085340371714506, "grad_norm": 0.3038443922996521, "learning_rate": 8e-05, "loss": 1.6352, "step": 3676 }, { "epoch": 0.8087539865830858, "grad_norm": 0.2839770019054413, "learning_rate": 8e-05, "loss": 1.6714, "step": 3677 }, { "epoch": 0.8089739359947212, "grad_norm": 0.28624406456947327, "learning_rate": 8e-05, "loss": 1.5826, "step": 3678 }, { "epoch": 0.8091938854063565, "grad_norm": 0.28209447860717773, "learning_rate": 8e-05, "loss": 1.5905, "step": 3679 }, { "epoch": 0.8094138348179919, "grad_norm": 0.3045743703842163, "learning_rate": 8e-05, "loss": 1.5201, "step": 3680 }, { "epoch": 0.8096337842296272, "grad_norm": 0.2729533910751343, "learning_rate": 8e-05, "loss": 1.5097, "step": 3681 }, { "epoch": 0.8098537336412626, "grad_norm": 0.3302726447582245, "learning_rate": 8e-05, "loss": 1.9036, "step": 3682 }, { "epoch": 0.8100736830528978, "grad_norm": 0.3054613471031189, "learning_rate": 8e-05, "loss": 1.6474, "step": 3683 }, { "epoch": 0.8102936324645331, "grad_norm": 0.306883305311203, "learning_rate": 8e-05, "loss": 1.8867, "step": 3684 }, { "epoch": 0.8105135818761685, "grad_norm": 0.2889968156814575, "learning_rate": 8e-05, "loss": 1.7218, "step": 3685 }, { "epoch": 0.8107335312878038, "grad_norm": 0.2770795226097107, "learning_rate": 8e-05, "loss": 1.5795, "step": 3686 }, { "epoch": 0.8109534806994392, "grad_norm": 0.29530447721481323, "learning_rate": 8e-05, "loss": 1.681, "step": 3687 }, { "epoch": 0.8111734301110745, "grad_norm": 0.28968799114227295, "learning_rate": 8e-05, "loss": 1.6409, "step": 3688 }, { "epoch": 0.8113933795227097, "grad_norm": 0.2868937849998474, "learning_rate": 8e-05, "loss": 1.6418, "step": 3689 }, { "epoch": 0.8116133289343451, "grad_norm": 0.29399046301841736, "learning_rate": 8e-05, "loss": 1.6872, "step": 3690 }, { "epoch": 0.8118332783459804, "grad_norm": 0.2806103229522705, "learning_rate": 8e-05, "loss": 1.5778, "step": 3691 }, { "epoch": 0.8120532277576158, "grad_norm": 0.3034190535545349, "learning_rate": 8e-05, "loss": 1.605, "step": 3692 }, { "epoch": 0.8122731771692511, "grad_norm": 0.29207345843315125, "learning_rate": 8e-05, "loss": 1.7604, "step": 3693 }, { "epoch": 0.8124931265808863, "grad_norm": 0.28544682264328003, "learning_rate": 8e-05, "loss": 1.5085, "step": 3694 }, { "epoch": 0.8127130759925217, "grad_norm": 0.30136585235595703, "learning_rate": 8e-05, "loss": 1.7208, "step": 3695 }, { "epoch": 0.812933025404157, "grad_norm": 0.29252931475639343, "learning_rate": 8e-05, "loss": 1.5529, "step": 3696 }, { "epoch": 0.8131529748157924, "grad_norm": 0.3112747371196747, "learning_rate": 8e-05, "loss": 1.6788, "step": 3697 }, { "epoch": 0.8133729242274277, "grad_norm": 0.28653424978256226, "learning_rate": 8e-05, "loss": 1.6755, "step": 3698 }, { "epoch": 0.813592873639063, "grad_norm": 0.2881905436515808, "learning_rate": 8e-05, "loss": 1.598, "step": 3699 }, { "epoch": 0.8138128230506984, "grad_norm": 0.3594631850719452, "learning_rate": 8e-05, "loss": 1.7263, "step": 3700 }, { "epoch": 0.8140327724623336, "grad_norm": 0.31845301389694214, "learning_rate": 8e-05, "loss": 1.7378, "step": 3701 }, { "epoch": 0.814252721873969, "grad_norm": 0.29628971219062805, "learning_rate": 8e-05, "loss": 1.6484, "step": 3702 }, { "epoch": 0.8144726712856043, "grad_norm": 0.2896782159805298, "learning_rate": 8e-05, "loss": 1.6006, "step": 3703 }, { "epoch": 0.8146926206972397, "grad_norm": 0.32776281237602234, "learning_rate": 8e-05, "loss": 1.8014, "step": 3704 }, { "epoch": 0.814912570108875, "grad_norm": 0.29571112990379333, "learning_rate": 8e-05, "loss": 1.7807, "step": 3705 }, { "epoch": 0.8151325195205102, "grad_norm": 0.2930830717086792, "learning_rate": 8e-05, "loss": 1.6407, "step": 3706 }, { "epoch": 0.8153524689321456, "grad_norm": 0.30788877606391907, "learning_rate": 8e-05, "loss": 1.8575, "step": 3707 }, { "epoch": 0.8155724183437809, "grad_norm": 0.2772351801395416, "learning_rate": 8e-05, "loss": 1.7, "step": 3708 }, { "epoch": 0.8157923677554163, "grad_norm": 0.2957149147987366, "learning_rate": 8e-05, "loss": 1.6791, "step": 3709 }, { "epoch": 0.8160123171670516, "grad_norm": 0.29258304834365845, "learning_rate": 8e-05, "loss": 1.6694, "step": 3710 }, { "epoch": 0.8162322665786869, "grad_norm": 0.3325083255767822, "learning_rate": 8e-05, "loss": 1.649, "step": 3711 }, { "epoch": 0.8164522159903222, "grad_norm": 0.2934175431728363, "learning_rate": 8e-05, "loss": 1.5983, "step": 3712 }, { "epoch": 0.8166721654019575, "grad_norm": 0.2800202965736389, "learning_rate": 8e-05, "loss": 1.6778, "step": 3713 }, { "epoch": 0.8168921148135929, "grad_norm": 0.28423386812210083, "learning_rate": 8e-05, "loss": 1.4981, "step": 3714 }, { "epoch": 0.8171120642252282, "grad_norm": 0.3425268530845642, "learning_rate": 8e-05, "loss": 1.7839, "step": 3715 }, { "epoch": 0.8173320136368635, "grad_norm": 0.3047974705696106, "learning_rate": 8e-05, "loss": 1.6118, "step": 3716 }, { "epoch": 0.8175519630484989, "grad_norm": 0.2974521815776825, "learning_rate": 8e-05, "loss": 1.6564, "step": 3717 }, { "epoch": 0.8177719124601341, "grad_norm": 0.27390167117118835, "learning_rate": 8e-05, "loss": 1.5317, "step": 3718 }, { "epoch": 0.8179918618717695, "grad_norm": 0.3012712597846985, "learning_rate": 8e-05, "loss": 1.5497, "step": 3719 }, { "epoch": 0.8182118112834048, "grad_norm": 0.2798818349838257, "learning_rate": 8e-05, "loss": 1.6036, "step": 3720 }, { "epoch": 0.8184317606950401, "grad_norm": 0.2981497049331665, "learning_rate": 8e-05, "loss": 1.7168, "step": 3721 }, { "epoch": 0.8186517101066755, "grad_norm": 0.30722394585609436, "learning_rate": 8e-05, "loss": 1.5669, "step": 3722 }, { "epoch": 0.8188716595183108, "grad_norm": 0.317390114068985, "learning_rate": 8e-05, "loss": 1.7005, "step": 3723 }, { "epoch": 0.8190916089299461, "grad_norm": 0.306482195854187, "learning_rate": 8e-05, "loss": 1.7386, "step": 3724 }, { "epoch": 0.8193115583415814, "grad_norm": 0.2971460521221161, "learning_rate": 8e-05, "loss": 1.6786, "step": 3725 }, { "epoch": 0.8195315077532168, "grad_norm": 0.30549904704093933, "learning_rate": 8e-05, "loss": 1.7212, "step": 3726 }, { "epoch": 0.8197514571648521, "grad_norm": 0.2948817312717438, "learning_rate": 8e-05, "loss": 1.5523, "step": 3727 }, { "epoch": 0.8199714065764874, "grad_norm": 0.29448559880256653, "learning_rate": 8e-05, "loss": 1.6445, "step": 3728 }, { "epoch": 0.8201913559881228, "grad_norm": 0.31158149242401123, "learning_rate": 8e-05, "loss": 1.7997, "step": 3729 }, { "epoch": 0.820411305399758, "grad_norm": 0.29270637035369873, "learning_rate": 8e-05, "loss": 1.6418, "step": 3730 }, { "epoch": 0.8206312548113934, "grad_norm": 0.29338178038597107, "learning_rate": 8e-05, "loss": 1.6839, "step": 3731 }, { "epoch": 0.8208512042230287, "grad_norm": 0.2772715389728546, "learning_rate": 8e-05, "loss": 1.6326, "step": 3732 }, { "epoch": 0.821071153634664, "grad_norm": 0.2976597547531128, "learning_rate": 8e-05, "loss": 1.6613, "step": 3733 }, { "epoch": 0.8212911030462994, "grad_norm": 0.3194999396800995, "learning_rate": 8e-05, "loss": 1.665, "step": 3734 }, { "epoch": 0.8215110524579347, "grad_norm": 0.28494882583618164, "learning_rate": 8e-05, "loss": 1.5707, "step": 3735 }, { "epoch": 0.82173100186957, "grad_norm": 0.29339084029197693, "learning_rate": 8e-05, "loss": 1.6276, "step": 3736 }, { "epoch": 0.8219509512812053, "grad_norm": 0.2872979938983917, "learning_rate": 8e-05, "loss": 1.5653, "step": 3737 }, { "epoch": 0.8221709006928406, "grad_norm": 0.2818096876144409, "learning_rate": 8e-05, "loss": 1.5631, "step": 3738 }, { "epoch": 0.822390850104476, "grad_norm": 0.2971639633178711, "learning_rate": 8e-05, "loss": 1.6698, "step": 3739 }, { "epoch": 0.8226107995161113, "grad_norm": 0.27923282980918884, "learning_rate": 8e-05, "loss": 1.6127, "step": 3740 }, { "epoch": 0.8228307489277467, "grad_norm": 0.2899787425994873, "learning_rate": 8e-05, "loss": 1.6323, "step": 3741 }, { "epoch": 0.8230506983393819, "grad_norm": 0.2971678376197815, "learning_rate": 8e-05, "loss": 1.6495, "step": 3742 }, { "epoch": 0.8232706477510172, "grad_norm": 0.3197123408317566, "learning_rate": 8e-05, "loss": 1.6807, "step": 3743 }, { "epoch": 0.8234905971626526, "grad_norm": 0.2743538022041321, "learning_rate": 8e-05, "loss": 1.659, "step": 3744 }, { "epoch": 0.8237105465742879, "grad_norm": 0.29337090253829956, "learning_rate": 8e-05, "loss": 1.7992, "step": 3745 }, { "epoch": 0.8239304959859233, "grad_norm": 0.295331209897995, "learning_rate": 8e-05, "loss": 1.5459, "step": 3746 }, { "epoch": 0.8241504453975586, "grad_norm": 0.2844454050064087, "learning_rate": 8e-05, "loss": 1.7079, "step": 3747 }, { "epoch": 0.8243703948091939, "grad_norm": 0.32962703704833984, "learning_rate": 8e-05, "loss": 1.8193, "step": 3748 }, { "epoch": 0.8245903442208292, "grad_norm": 0.2843455374240875, "learning_rate": 8e-05, "loss": 1.6964, "step": 3749 }, { "epoch": 0.8248102936324645, "grad_norm": 0.28151237964630127, "learning_rate": 8e-05, "loss": 1.7152, "step": 3750 }, { "epoch": 0.8250302430440999, "grad_norm": 0.30013763904571533, "learning_rate": 8e-05, "loss": 1.7262, "step": 3751 }, { "epoch": 0.8252501924557352, "grad_norm": 0.2816784083843231, "learning_rate": 8e-05, "loss": 1.7431, "step": 3752 }, { "epoch": 0.8254701418673706, "grad_norm": 0.32148584723472595, "learning_rate": 8e-05, "loss": 1.7909, "step": 3753 }, { "epoch": 0.8256900912790058, "grad_norm": 0.2902558147907257, "learning_rate": 8e-05, "loss": 1.5782, "step": 3754 }, { "epoch": 0.8259100406906411, "grad_norm": 0.3017018735408783, "learning_rate": 8e-05, "loss": 1.7332, "step": 3755 }, { "epoch": 0.8261299901022765, "grad_norm": 0.29928895831108093, "learning_rate": 8e-05, "loss": 1.6972, "step": 3756 }, { "epoch": 0.8263499395139118, "grad_norm": 0.2922762632369995, "learning_rate": 8e-05, "loss": 1.5663, "step": 3757 }, { "epoch": 0.8265698889255472, "grad_norm": 0.2903006672859192, "learning_rate": 8e-05, "loss": 1.6253, "step": 3758 }, { "epoch": 0.8267898383371824, "grad_norm": 0.2981746196746826, "learning_rate": 8e-05, "loss": 1.69, "step": 3759 }, { "epoch": 0.8270097877488177, "grad_norm": 0.2905546724796295, "learning_rate": 8e-05, "loss": 1.5513, "step": 3760 }, { "epoch": 0.8272297371604531, "grad_norm": 0.30684736371040344, "learning_rate": 8e-05, "loss": 1.7188, "step": 3761 }, { "epoch": 0.8274496865720884, "grad_norm": 0.29893192648887634, "learning_rate": 8e-05, "loss": 1.7759, "step": 3762 }, { "epoch": 0.8276696359837238, "grad_norm": 0.2762022316455841, "learning_rate": 8e-05, "loss": 1.5835, "step": 3763 }, { "epoch": 0.8278895853953591, "grad_norm": 0.3005400002002716, "learning_rate": 8e-05, "loss": 1.7313, "step": 3764 }, { "epoch": 0.8281095348069943, "grad_norm": 0.33047348260879517, "learning_rate": 8e-05, "loss": 1.7504, "step": 3765 }, { "epoch": 0.8283294842186297, "grad_norm": 0.29011067748069763, "learning_rate": 8e-05, "loss": 1.677, "step": 3766 }, { "epoch": 0.828549433630265, "grad_norm": 0.28470665216445923, "learning_rate": 8e-05, "loss": 1.5375, "step": 3767 }, { "epoch": 0.8287693830419004, "grad_norm": 0.47327253222465515, "learning_rate": 8e-05, "loss": 1.9681, "step": 3768 }, { "epoch": 0.8289893324535357, "grad_norm": 0.3004605174064636, "learning_rate": 8e-05, "loss": 1.6489, "step": 3769 }, { "epoch": 0.8292092818651711, "grad_norm": 0.29962432384490967, "learning_rate": 8e-05, "loss": 1.5485, "step": 3770 }, { "epoch": 0.8294292312768063, "grad_norm": 0.3226446509361267, "learning_rate": 8e-05, "loss": 1.5853, "step": 3771 }, { "epoch": 0.8296491806884416, "grad_norm": 0.2956733703613281, "learning_rate": 8e-05, "loss": 1.6783, "step": 3772 }, { "epoch": 0.829869130100077, "grad_norm": 0.30481475591659546, "learning_rate": 8e-05, "loss": 1.7441, "step": 3773 }, { "epoch": 0.8300890795117123, "grad_norm": 0.29181525111198425, "learning_rate": 8e-05, "loss": 1.7026, "step": 3774 }, { "epoch": 0.8303090289233477, "grad_norm": 0.28580138087272644, "learning_rate": 8e-05, "loss": 1.6131, "step": 3775 }, { "epoch": 0.830528978334983, "grad_norm": 0.28029665350914, "learning_rate": 8e-05, "loss": 1.618, "step": 3776 }, { "epoch": 0.8307489277466182, "grad_norm": 0.31312623620033264, "learning_rate": 8e-05, "loss": 1.629, "step": 3777 }, { "epoch": 0.8309688771582536, "grad_norm": 0.31540897488594055, "learning_rate": 8e-05, "loss": 1.8215, "step": 3778 }, { "epoch": 0.8311888265698889, "grad_norm": 0.3031502664089203, "learning_rate": 8e-05, "loss": 1.6722, "step": 3779 }, { "epoch": 0.8314087759815243, "grad_norm": 0.31536054611206055, "learning_rate": 8e-05, "loss": 1.753, "step": 3780 }, { "epoch": 0.8316287253931596, "grad_norm": 0.31534263491630554, "learning_rate": 8e-05, "loss": 1.6841, "step": 3781 }, { "epoch": 0.8318486748047949, "grad_norm": 0.29810741543769836, "learning_rate": 8e-05, "loss": 1.8143, "step": 3782 }, { "epoch": 0.8320686242164302, "grad_norm": 0.2857937812805176, "learning_rate": 8e-05, "loss": 1.6004, "step": 3783 }, { "epoch": 0.8322885736280655, "grad_norm": 0.2927904427051544, "learning_rate": 8e-05, "loss": 1.6385, "step": 3784 }, { "epoch": 0.8325085230397009, "grad_norm": 0.29359960556030273, "learning_rate": 8e-05, "loss": 1.762, "step": 3785 }, { "epoch": 0.8327284724513362, "grad_norm": 0.2871841788291931, "learning_rate": 8e-05, "loss": 1.6141, "step": 3786 }, { "epoch": 0.8329484218629715, "grad_norm": 0.32198578119277954, "learning_rate": 8e-05, "loss": 1.8785, "step": 3787 }, { "epoch": 0.8331683712746069, "grad_norm": 0.2883181571960449, "learning_rate": 8e-05, "loss": 1.6136, "step": 3788 }, { "epoch": 0.8333883206862421, "grad_norm": 0.2934539020061493, "learning_rate": 8e-05, "loss": 1.6878, "step": 3789 }, { "epoch": 0.8336082700978775, "grad_norm": 0.29860734939575195, "learning_rate": 8e-05, "loss": 1.5412, "step": 3790 }, { "epoch": 0.8338282195095128, "grad_norm": 0.32730698585510254, "learning_rate": 8e-05, "loss": 1.8215, "step": 3791 }, { "epoch": 0.8340481689211481, "grad_norm": 0.2875627875328064, "learning_rate": 8e-05, "loss": 1.571, "step": 3792 }, { "epoch": 0.8342681183327835, "grad_norm": 0.33305448293685913, "learning_rate": 8e-05, "loss": 1.8454, "step": 3793 }, { "epoch": 0.8344880677444187, "grad_norm": 0.2842869162559509, "learning_rate": 8e-05, "loss": 1.6616, "step": 3794 }, { "epoch": 0.8347080171560541, "grad_norm": 0.30781999230384827, "learning_rate": 8e-05, "loss": 1.622, "step": 3795 }, { "epoch": 0.8349279665676894, "grad_norm": 0.2892124056816101, "learning_rate": 8e-05, "loss": 1.5731, "step": 3796 }, { "epoch": 0.8351479159793248, "grad_norm": 0.2950188219547272, "learning_rate": 8e-05, "loss": 1.6374, "step": 3797 }, { "epoch": 0.8353678653909601, "grad_norm": 0.283393919467926, "learning_rate": 8e-05, "loss": 1.5603, "step": 3798 }, { "epoch": 0.8355878148025954, "grad_norm": 0.28449615836143494, "learning_rate": 8e-05, "loss": 1.5694, "step": 3799 }, { "epoch": 0.8358077642142308, "grad_norm": 0.27379027009010315, "learning_rate": 8e-05, "loss": 1.4331, "step": 3800 }, { "epoch": 0.836027713625866, "grad_norm": 0.31976261734962463, "learning_rate": 8e-05, "loss": 1.7362, "step": 3801 }, { "epoch": 0.8362476630375014, "grad_norm": 0.27755287289619446, "learning_rate": 8e-05, "loss": 1.577, "step": 3802 }, { "epoch": 0.8364676124491367, "grad_norm": 0.2974432110786438, "learning_rate": 8e-05, "loss": 1.7651, "step": 3803 }, { "epoch": 0.836687561860772, "grad_norm": 0.320019006729126, "learning_rate": 8e-05, "loss": 1.8053, "step": 3804 }, { "epoch": 0.8369075112724074, "grad_norm": 0.29677972197532654, "learning_rate": 8e-05, "loss": 1.7302, "step": 3805 }, { "epoch": 0.8371274606840426, "grad_norm": 0.28835824131965637, "learning_rate": 8e-05, "loss": 1.6395, "step": 3806 }, { "epoch": 0.837347410095678, "grad_norm": 0.29761141538619995, "learning_rate": 8e-05, "loss": 1.5725, "step": 3807 }, { "epoch": 0.8375673595073133, "grad_norm": 0.30719003081321716, "learning_rate": 8e-05, "loss": 1.6291, "step": 3808 }, { "epoch": 0.8377873089189486, "grad_norm": 0.3065665662288666, "learning_rate": 8e-05, "loss": 1.6901, "step": 3809 }, { "epoch": 0.838007258330584, "grad_norm": 0.3014419376850128, "learning_rate": 8e-05, "loss": 1.7154, "step": 3810 }, { "epoch": 0.8382272077422193, "grad_norm": 0.29619207978248596, "learning_rate": 8e-05, "loss": 1.6578, "step": 3811 }, { "epoch": 0.8384471571538546, "grad_norm": 0.2940784692764282, "learning_rate": 8e-05, "loss": 1.7499, "step": 3812 }, { "epoch": 0.8386671065654899, "grad_norm": 0.29198816418647766, "learning_rate": 8e-05, "loss": 1.6565, "step": 3813 }, { "epoch": 0.8388870559771252, "grad_norm": 0.28611230850219727, "learning_rate": 8e-05, "loss": 1.7194, "step": 3814 }, { "epoch": 0.8391070053887606, "grad_norm": 0.2848166227340698, "learning_rate": 8e-05, "loss": 1.6492, "step": 3815 }, { "epoch": 0.8393269548003959, "grad_norm": 0.31025707721710205, "learning_rate": 8e-05, "loss": 1.7402, "step": 3816 }, { "epoch": 0.8395469042120313, "grad_norm": 0.2825357913970947, "learning_rate": 8e-05, "loss": 1.6927, "step": 3817 }, { "epoch": 0.8397668536236665, "grad_norm": 0.27436748147010803, "learning_rate": 8e-05, "loss": 1.551, "step": 3818 }, { "epoch": 0.8399868030353019, "grad_norm": 0.26923510432243347, "learning_rate": 8e-05, "loss": 1.5776, "step": 3819 }, { "epoch": 0.8402067524469372, "grad_norm": 0.29636648297309875, "learning_rate": 8e-05, "loss": 1.5779, "step": 3820 }, { "epoch": 0.8404267018585725, "grad_norm": 0.2913194000720978, "learning_rate": 8e-05, "loss": 1.6921, "step": 3821 }, { "epoch": 0.8406466512702079, "grad_norm": 0.28964900970458984, "learning_rate": 8e-05, "loss": 1.6866, "step": 3822 }, { "epoch": 0.8408666006818432, "grad_norm": 0.2947314381599426, "learning_rate": 8e-05, "loss": 1.7541, "step": 3823 }, { "epoch": 0.8410865500934785, "grad_norm": 0.2964346706867218, "learning_rate": 8e-05, "loss": 1.7498, "step": 3824 }, { "epoch": 0.8413064995051138, "grad_norm": 0.2943279445171356, "learning_rate": 8e-05, "loss": 1.6294, "step": 3825 }, { "epoch": 0.8415264489167491, "grad_norm": 0.305585116147995, "learning_rate": 8e-05, "loss": 1.6573, "step": 3826 }, { "epoch": 0.8417463983283845, "grad_norm": 0.2842620313167572, "learning_rate": 8e-05, "loss": 1.6747, "step": 3827 }, { "epoch": 0.8419663477400198, "grad_norm": 0.29704785346984863, "learning_rate": 8e-05, "loss": 1.7986, "step": 3828 }, { "epoch": 0.8421862971516552, "grad_norm": 0.28527146577835083, "learning_rate": 8e-05, "loss": 1.722, "step": 3829 }, { "epoch": 0.8424062465632904, "grad_norm": 0.2996181547641754, "learning_rate": 8e-05, "loss": 1.8015, "step": 3830 }, { "epoch": 0.8426261959749257, "grad_norm": 0.3034017086029053, "learning_rate": 8e-05, "loss": 1.7774, "step": 3831 }, { "epoch": 0.8428461453865611, "grad_norm": 0.2914513647556305, "learning_rate": 8e-05, "loss": 1.672, "step": 3832 }, { "epoch": 0.8430660947981964, "grad_norm": 0.2868506610393524, "learning_rate": 8e-05, "loss": 1.6906, "step": 3833 }, { "epoch": 0.8432860442098318, "grad_norm": 0.2967677712440491, "learning_rate": 8e-05, "loss": 1.4859, "step": 3834 }, { "epoch": 0.843505993621467, "grad_norm": 0.2904664874076843, "learning_rate": 8e-05, "loss": 1.7559, "step": 3835 }, { "epoch": 0.8437259430331023, "grad_norm": 0.2900542616844177, "learning_rate": 8e-05, "loss": 1.6315, "step": 3836 }, { "epoch": 0.8439458924447377, "grad_norm": 0.28981515765190125, "learning_rate": 8e-05, "loss": 1.6254, "step": 3837 }, { "epoch": 0.844165841856373, "grad_norm": 0.29468080401420593, "learning_rate": 8e-05, "loss": 1.6417, "step": 3838 }, { "epoch": 0.8443857912680084, "grad_norm": 0.3301861584186554, "learning_rate": 8e-05, "loss": 1.7982, "step": 3839 }, { "epoch": 0.8446057406796437, "grad_norm": 0.28386393189430237, "learning_rate": 8e-05, "loss": 1.7077, "step": 3840 }, { "epoch": 0.8448256900912791, "grad_norm": 0.27421340346336365, "learning_rate": 8e-05, "loss": 1.5492, "step": 3841 }, { "epoch": 0.8450456395029143, "grad_norm": 0.30234119296073914, "learning_rate": 8e-05, "loss": 1.6162, "step": 3842 }, { "epoch": 0.8452655889145496, "grad_norm": 0.3157059848308563, "learning_rate": 8e-05, "loss": 1.7228, "step": 3843 }, { "epoch": 0.845485538326185, "grad_norm": 0.2899254858493805, "learning_rate": 8e-05, "loss": 1.451, "step": 3844 }, { "epoch": 0.8457054877378203, "grad_norm": 0.3020762801170349, "learning_rate": 8e-05, "loss": 1.8273, "step": 3845 }, { "epoch": 0.8459254371494557, "grad_norm": 0.3134917616844177, "learning_rate": 8e-05, "loss": 1.8578, "step": 3846 }, { "epoch": 0.846145386561091, "grad_norm": 0.29482507705688477, "learning_rate": 8e-05, "loss": 1.6039, "step": 3847 }, { "epoch": 0.8463653359727262, "grad_norm": 0.3106936514377594, "learning_rate": 8e-05, "loss": 1.7019, "step": 3848 }, { "epoch": 0.8465852853843616, "grad_norm": 0.2869921028614044, "learning_rate": 8e-05, "loss": 1.6515, "step": 3849 }, { "epoch": 0.8468052347959969, "grad_norm": 0.32278892397880554, "learning_rate": 8e-05, "loss": 1.955, "step": 3850 }, { "epoch": 0.8470251842076323, "grad_norm": 0.28249430656433105, "learning_rate": 8e-05, "loss": 1.575, "step": 3851 }, { "epoch": 0.8472451336192676, "grad_norm": 0.3081803321838379, "learning_rate": 8e-05, "loss": 1.5634, "step": 3852 }, { "epoch": 0.8474650830309028, "grad_norm": 0.3033173978328705, "learning_rate": 8e-05, "loss": 1.6182, "step": 3853 }, { "epoch": 0.8476850324425382, "grad_norm": 0.3237481415271759, "learning_rate": 8e-05, "loss": 1.5605, "step": 3854 }, { "epoch": 0.8479049818541735, "grad_norm": 0.3091139495372772, "learning_rate": 8e-05, "loss": 1.7551, "step": 3855 }, { "epoch": 0.8481249312658089, "grad_norm": 0.2904627025127411, "learning_rate": 8e-05, "loss": 1.6982, "step": 3856 }, { "epoch": 0.8483448806774442, "grad_norm": 0.3025551438331604, "learning_rate": 8e-05, "loss": 1.6156, "step": 3857 }, { "epoch": 0.8485648300890795, "grad_norm": 0.28311145305633545, "learning_rate": 8e-05, "loss": 1.6919, "step": 3858 }, { "epoch": 0.8487847795007148, "grad_norm": 0.2805883288383484, "learning_rate": 8e-05, "loss": 1.5491, "step": 3859 }, { "epoch": 0.8490047289123501, "grad_norm": 0.31418681144714355, "learning_rate": 8e-05, "loss": 1.7864, "step": 3860 }, { "epoch": 0.8492246783239855, "grad_norm": 0.2807353734970093, "learning_rate": 8e-05, "loss": 1.4794, "step": 3861 }, { "epoch": 0.8494446277356208, "grad_norm": 0.29171043634414673, "learning_rate": 8e-05, "loss": 1.7223, "step": 3862 }, { "epoch": 0.8496645771472562, "grad_norm": 0.30100691318511963, "learning_rate": 8e-05, "loss": 1.8106, "step": 3863 }, { "epoch": 0.8498845265588915, "grad_norm": 0.29176101088523865, "learning_rate": 8e-05, "loss": 1.6441, "step": 3864 }, { "epoch": 0.8501044759705267, "grad_norm": 0.27088651061058044, "learning_rate": 8e-05, "loss": 1.4872, "step": 3865 }, { "epoch": 0.8503244253821621, "grad_norm": 0.28077685832977295, "learning_rate": 8e-05, "loss": 1.6243, "step": 3866 }, { "epoch": 0.8505443747937974, "grad_norm": 0.30027422308921814, "learning_rate": 8e-05, "loss": 1.7244, "step": 3867 }, { "epoch": 0.8507643242054328, "grad_norm": 0.29100343585014343, "learning_rate": 8e-05, "loss": 1.6324, "step": 3868 }, { "epoch": 0.8509842736170681, "grad_norm": 0.29306820034980774, "learning_rate": 8e-05, "loss": 1.7432, "step": 3869 }, { "epoch": 0.8512042230287034, "grad_norm": 0.29503849148750305, "learning_rate": 8e-05, "loss": 1.5348, "step": 3870 }, { "epoch": 0.8514241724403387, "grad_norm": 0.28499239683151245, "learning_rate": 8e-05, "loss": 1.6653, "step": 3871 }, { "epoch": 0.851644121851974, "grad_norm": 0.3126235008239746, "learning_rate": 8e-05, "loss": 1.6023, "step": 3872 }, { "epoch": 0.8518640712636094, "grad_norm": 0.2873520255088806, "learning_rate": 8e-05, "loss": 1.6659, "step": 3873 }, { "epoch": 0.8520840206752447, "grad_norm": 0.28033289313316345, "learning_rate": 8e-05, "loss": 1.6133, "step": 3874 }, { "epoch": 0.85230397008688, "grad_norm": 0.28731170296669006, "learning_rate": 8e-05, "loss": 1.6344, "step": 3875 }, { "epoch": 0.8525239194985154, "grad_norm": 0.28616607189178467, "learning_rate": 8e-05, "loss": 1.5874, "step": 3876 }, { "epoch": 0.8527438689101506, "grad_norm": 0.29681506752967834, "learning_rate": 8e-05, "loss": 1.6628, "step": 3877 }, { "epoch": 0.852963818321786, "grad_norm": 0.3163807690143585, "learning_rate": 8e-05, "loss": 1.7927, "step": 3878 }, { "epoch": 0.8531837677334213, "grad_norm": 0.28302204608917236, "learning_rate": 8e-05, "loss": 1.6406, "step": 3879 }, { "epoch": 0.8534037171450566, "grad_norm": 0.29161083698272705, "learning_rate": 8e-05, "loss": 1.7386, "step": 3880 }, { "epoch": 0.853623666556692, "grad_norm": 0.29691028594970703, "learning_rate": 8e-05, "loss": 1.6582, "step": 3881 }, { "epoch": 0.8538436159683273, "grad_norm": 0.28576239943504333, "learning_rate": 8e-05, "loss": 1.6866, "step": 3882 }, { "epoch": 0.8540635653799626, "grad_norm": 0.2867792844772339, "learning_rate": 8e-05, "loss": 1.4684, "step": 3883 }, { "epoch": 0.8542835147915979, "grad_norm": 0.2788017690181732, "learning_rate": 8e-05, "loss": 1.6202, "step": 3884 }, { "epoch": 0.8545034642032333, "grad_norm": 0.2767166197299957, "learning_rate": 8e-05, "loss": 1.5813, "step": 3885 }, { "epoch": 0.8547234136148686, "grad_norm": 0.29414990544319153, "learning_rate": 8e-05, "loss": 1.6617, "step": 3886 }, { "epoch": 0.8549433630265039, "grad_norm": 0.2955497205257416, "learning_rate": 8e-05, "loss": 1.6456, "step": 3887 }, { "epoch": 0.8551633124381393, "grad_norm": 0.31308290362358093, "learning_rate": 8e-05, "loss": 1.6532, "step": 3888 }, { "epoch": 0.8553832618497745, "grad_norm": 0.301305890083313, "learning_rate": 8e-05, "loss": 1.6877, "step": 3889 }, { "epoch": 0.8556032112614099, "grad_norm": 0.290628582239151, "learning_rate": 8e-05, "loss": 1.7098, "step": 3890 }, { "epoch": 0.8558231606730452, "grad_norm": 0.30904853343963623, "learning_rate": 8e-05, "loss": 1.6352, "step": 3891 }, { "epoch": 0.8560431100846805, "grad_norm": 0.29854530096054077, "learning_rate": 8e-05, "loss": 1.6971, "step": 3892 }, { "epoch": 0.8562630594963159, "grad_norm": 0.2870941460132599, "learning_rate": 8e-05, "loss": 1.6655, "step": 3893 }, { "epoch": 0.8564830089079511, "grad_norm": 0.29070189595222473, "learning_rate": 8e-05, "loss": 1.5808, "step": 3894 }, { "epoch": 0.8567029583195865, "grad_norm": 0.3046536445617676, "learning_rate": 8e-05, "loss": 1.7834, "step": 3895 }, { "epoch": 0.8569229077312218, "grad_norm": 0.2883049249649048, "learning_rate": 8e-05, "loss": 1.5898, "step": 3896 }, { "epoch": 0.8571428571428571, "grad_norm": 0.29613447189331055, "learning_rate": 8e-05, "loss": 1.7619, "step": 3897 }, { "epoch": 0.8573628065544925, "grad_norm": 0.27996566891670227, "learning_rate": 8e-05, "loss": 1.5863, "step": 3898 }, { "epoch": 0.8575827559661278, "grad_norm": 0.3026728332042694, "learning_rate": 8e-05, "loss": 1.7488, "step": 3899 }, { "epoch": 0.8578027053777632, "grad_norm": 0.28238317370414734, "learning_rate": 8e-05, "loss": 1.6908, "step": 3900 }, { "epoch": 0.8580226547893984, "grad_norm": 0.271651953458786, "learning_rate": 8e-05, "loss": 1.5228, "step": 3901 }, { "epoch": 0.8582426042010337, "grad_norm": 0.2936142385005951, "learning_rate": 8e-05, "loss": 1.7332, "step": 3902 }, { "epoch": 0.8584625536126691, "grad_norm": 0.29982343316078186, "learning_rate": 8e-05, "loss": 1.6792, "step": 3903 }, { "epoch": 0.8586825030243044, "grad_norm": 0.30300748348236084, "learning_rate": 8e-05, "loss": 1.7689, "step": 3904 }, { "epoch": 0.8589024524359398, "grad_norm": 0.3021508455276489, "learning_rate": 8e-05, "loss": 1.7082, "step": 3905 }, { "epoch": 0.859122401847575, "grad_norm": 0.2989715039730072, "learning_rate": 8e-05, "loss": 1.6786, "step": 3906 }, { "epoch": 0.8593423512592104, "grad_norm": 0.2968369424343109, "learning_rate": 8e-05, "loss": 1.7742, "step": 3907 }, { "epoch": 0.8595623006708457, "grad_norm": 0.2943005859851837, "learning_rate": 8e-05, "loss": 1.6512, "step": 3908 }, { "epoch": 0.859782250082481, "grad_norm": 0.2851683497428894, "learning_rate": 8e-05, "loss": 1.6554, "step": 3909 }, { "epoch": 0.8600021994941164, "grad_norm": 0.295810729265213, "learning_rate": 8e-05, "loss": 1.7791, "step": 3910 }, { "epoch": 0.8602221489057517, "grad_norm": 0.28922411799430847, "learning_rate": 8e-05, "loss": 1.7462, "step": 3911 }, { "epoch": 0.860442098317387, "grad_norm": 0.2724452316761017, "learning_rate": 8e-05, "loss": 1.3677, "step": 3912 }, { "epoch": 0.8606620477290223, "grad_norm": 0.28607258200645447, "learning_rate": 8e-05, "loss": 1.7326, "step": 3913 }, { "epoch": 0.8608819971406576, "grad_norm": 0.26861658692359924, "learning_rate": 8e-05, "loss": 1.4499, "step": 3914 }, { "epoch": 0.861101946552293, "grad_norm": 0.2905106246471405, "learning_rate": 8e-05, "loss": 1.6292, "step": 3915 }, { "epoch": 0.8613218959639283, "grad_norm": 0.28987225890159607, "learning_rate": 8e-05, "loss": 1.5838, "step": 3916 }, { "epoch": 0.8615418453755637, "grad_norm": 0.3006553053855896, "learning_rate": 8e-05, "loss": 1.5658, "step": 3917 }, { "epoch": 0.8617617947871989, "grad_norm": 0.2888564467430115, "learning_rate": 8e-05, "loss": 1.6067, "step": 3918 }, { "epoch": 0.8619817441988342, "grad_norm": 0.3040676414966583, "learning_rate": 8e-05, "loss": 1.8178, "step": 3919 }, { "epoch": 0.8622016936104696, "grad_norm": 0.29286056756973267, "learning_rate": 8e-05, "loss": 1.6235, "step": 3920 }, { "epoch": 0.8624216430221049, "grad_norm": 0.33138954639434814, "learning_rate": 8e-05, "loss": 1.6806, "step": 3921 }, { "epoch": 0.8626415924337403, "grad_norm": 0.2946946620941162, "learning_rate": 8e-05, "loss": 1.631, "step": 3922 }, { "epoch": 0.8628615418453756, "grad_norm": 0.3030078411102295, "learning_rate": 8e-05, "loss": 1.7346, "step": 3923 }, { "epoch": 0.8630814912570108, "grad_norm": 0.2995181381702423, "learning_rate": 8e-05, "loss": 1.6159, "step": 3924 }, { "epoch": 0.8633014406686462, "grad_norm": 0.31188464164733887, "learning_rate": 8e-05, "loss": 1.7165, "step": 3925 }, { "epoch": 0.8635213900802815, "grad_norm": 0.2989533245563507, "learning_rate": 8e-05, "loss": 1.7557, "step": 3926 }, { "epoch": 0.8637413394919169, "grad_norm": 0.2965574264526367, "learning_rate": 8e-05, "loss": 1.678, "step": 3927 }, { "epoch": 0.8639612889035522, "grad_norm": 0.28041690587997437, "learning_rate": 8e-05, "loss": 1.553, "step": 3928 }, { "epoch": 0.8641812383151876, "grad_norm": 0.2907460033893585, "learning_rate": 8e-05, "loss": 1.6468, "step": 3929 }, { "epoch": 0.8644011877268228, "grad_norm": 0.3051550090312958, "learning_rate": 8e-05, "loss": 1.7234, "step": 3930 }, { "epoch": 0.8646211371384581, "grad_norm": 0.2832759618759155, "learning_rate": 8e-05, "loss": 1.5426, "step": 3931 }, { "epoch": 0.8648410865500935, "grad_norm": 0.2915174961090088, "learning_rate": 8e-05, "loss": 1.649, "step": 3932 }, { "epoch": 0.8650610359617288, "grad_norm": 0.3146245777606964, "learning_rate": 8e-05, "loss": 1.7486, "step": 3933 }, { "epoch": 0.8652809853733642, "grad_norm": 0.2949977517127991, "learning_rate": 8e-05, "loss": 1.6221, "step": 3934 }, { "epoch": 0.8655009347849995, "grad_norm": 0.34237998723983765, "learning_rate": 8e-05, "loss": 1.6641, "step": 3935 }, { "epoch": 0.8657208841966347, "grad_norm": 0.3072294592857361, "learning_rate": 8e-05, "loss": 1.6981, "step": 3936 }, { "epoch": 0.8659408336082701, "grad_norm": 0.2796785235404968, "learning_rate": 8e-05, "loss": 1.5687, "step": 3937 }, { "epoch": 0.8661607830199054, "grad_norm": 0.27884992957115173, "learning_rate": 8e-05, "loss": 1.4457, "step": 3938 }, { "epoch": 0.8663807324315408, "grad_norm": 0.30189958214759827, "learning_rate": 8e-05, "loss": 1.7328, "step": 3939 }, { "epoch": 0.8666006818431761, "grad_norm": 0.3078850209712982, "learning_rate": 8e-05, "loss": 1.8462, "step": 3940 }, { "epoch": 0.8668206312548113, "grad_norm": 0.27258560061454773, "learning_rate": 8e-05, "loss": 1.501, "step": 3941 }, { "epoch": 0.8670405806664467, "grad_norm": 0.31592684984207153, "learning_rate": 8e-05, "loss": 1.573, "step": 3942 }, { "epoch": 0.867260530078082, "grad_norm": 0.2867029309272766, "learning_rate": 8e-05, "loss": 1.6152, "step": 3943 }, { "epoch": 0.8674804794897174, "grad_norm": 0.31841394305229187, "learning_rate": 8e-05, "loss": 1.6078, "step": 3944 }, { "epoch": 0.8677004289013527, "grad_norm": 0.3039727509021759, "learning_rate": 8e-05, "loss": 1.6775, "step": 3945 }, { "epoch": 0.867920378312988, "grad_norm": 0.3054428994655609, "learning_rate": 8e-05, "loss": 1.8358, "step": 3946 }, { "epoch": 0.8681403277246234, "grad_norm": 0.2971208691596985, "learning_rate": 8e-05, "loss": 1.6011, "step": 3947 }, { "epoch": 0.8683602771362586, "grad_norm": 0.30911338329315186, "learning_rate": 8e-05, "loss": 1.7928, "step": 3948 }, { "epoch": 0.868580226547894, "grad_norm": 0.3277186453342438, "learning_rate": 8e-05, "loss": 1.5783, "step": 3949 }, { "epoch": 0.8688001759595293, "grad_norm": 0.29121798276901245, "learning_rate": 8e-05, "loss": 1.6368, "step": 3950 }, { "epoch": 0.8690201253711647, "grad_norm": 0.30979645252227783, "learning_rate": 8e-05, "loss": 1.6603, "step": 3951 }, { "epoch": 0.8692400747828, "grad_norm": 0.29589441418647766, "learning_rate": 8e-05, "loss": 1.7536, "step": 3952 }, { "epoch": 0.8694600241944352, "grad_norm": 0.3051590323448181, "learning_rate": 8e-05, "loss": 1.6639, "step": 3953 }, { "epoch": 0.8696799736060706, "grad_norm": 0.30917656421661377, "learning_rate": 8e-05, "loss": 1.8418, "step": 3954 }, { "epoch": 0.8698999230177059, "grad_norm": 0.27990633249282837, "learning_rate": 8e-05, "loss": 1.7211, "step": 3955 }, { "epoch": 0.8701198724293413, "grad_norm": 0.28773045539855957, "learning_rate": 8e-05, "loss": 1.5785, "step": 3956 }, { "epoch": 0.8703398218409766, "grad_norm": 0.2980699837207794, "learning_rate": 8e-05, "loss": 1.5905, "step": 3957 }, { "epoch": 0.8705597712526119, "grad_norm": 0.30660519003868103, "learning_rate": 8e-05, "loss": 1.6561, "step": 3958 }, { "epoch": 0.8707797206642472, "grad_norm": 0.2980074882507324, "learning_rate": 8e-05, "loss": 1.5939, "step": 3959 }, { "epoch": 0.8709996700758825, "grad_norm": 0.30249714851379395, "learning_rate": 8e-05, "loss": 1.7815, "step": 3960 }, { "epoch": 0.8712196194875179, "grad_norm": 0.3185308277606964, "learning_rate": 8e-05, "loss": 1.6994, "step": 3961 }, { "epoch": 0.8714395688991532, "grad_norm": 0.29345861077308655, "learning_rate": 8e-05, "loss": 1.6363, "step": 3962 }, { "epoch": 0.8716595183107885, "grad_norm": 0.29790154099464417, "learning_rate": 8e-05, "loss": 1.6433, "step": 3963 }, { "epoch": 0.8718794677224239, "grad_norm": 0.30221304297447205, "learning_rate": 8e-05, "loss": 1.6036, "step": 3964 }, { "epoch": 0.8720994171340591, "grad_norm": 0.3027768135070801, "learning_rate": 8e-05, "loss": 1.7376, "step": 3965 }, { "epoch": 0.8723193665456945, "grad_norm": 0.32072335481643677, "learning_rate": 8e-05, "loss": 1.7161, "step": 3966 }, { "epoch": 0.8725393159573298, "grad_norm": 0.3282933831214905, "learning_rate": 8e-05, "loss": 1.8066, "step": 3967 }, { "epoch": 0.8727592653689651, "grad_norm": 0.2951596677303314, "learning_rate": 8e-05, "loss": 1.5761, "step": 3968 }, { "epoch": 0.8729792147806005, "grad_norm": 0.3005920946598053, "learning_rate": 8e-05, "loss": 1.6086, "step": 3969 }, { "epoch": 0.8731991641922358, "grad_norm": 0.2846725881099701, "learning_rate": 8e-05, "loss": 1.6208, "step": 3970 }, { "epoch": 0.8734191136038711, "grad_norm": 0.3003213107585907, "learning_rate": 8e-05, "loss": 1.6404, "step": 3971 }, { "epoch": 0.8736390630155064, "grad_norm": 0.2906554639339447, "learning_rate": 8e-05, "loss": 1.5947, "step": 3972 }, { "epoch": 0.8738590124271418, "grad_norm": 0.2852460741996765, "learning_rate": 8e-05, "loss": 1.6473, "step": 3973 }, { "epoch": 0.8740789618387771, "grad_norm": 0.2833881378173828, "learning_rate": 8e-05, "loss": 1.6245, "step": 3974 }, { "epoch": 0.8742989112504124, "grad_norm": 0.3263465464115143, "learning_rate": 8e-05, "loss": 1.6252, "step": 3975 }, { "epoch": 0.8745188606620478, "grad_norm": 0.3010908365249634, "learning_rate": 8e-05, "loss": 1.6072, "step": 3976 }, { "epoch": 0.874738810073683, "grad_norm": 0.29418328404426575, "learning_rate": 8e-05, "loss": 1.7066, "step": 3977 }, { "epoch": 0.8749587594853184, "grad_norm": 0.31706956028938293, "learning_rate": 8e-05, "loss": 1.6545, "step": 3978 }, { "epoch": 0.8751787088969537, "grad_norm": 0.31312480568885803, "learning_rate": 8e-05, "loss": 1.6229, "step": 3979 }, { "epoch": 0.875398658308589, "grad_norm": 0.27910348773002625, "learning_rate": 8e-05, "loss": 1.7239, "step": 3980 }, { "epoch": 0.8756186077202244, "grad_norm": 0.33857640624046326, "learning_rate": 8e-05, "loss": 1.7275, "step": 3981 }, { "epoch": 0.8758385571318597, "grad_norm": 0.2809359133243561, "learning_rate": 8e-05, "loss": 1.482, "step": 3982 }, { "epoch": 0.876058506543495, "grad_norm": 0.3103906214237213, "learning_rate": 8e-05, "loss": 1.7277, "step": 3983 }, { "epoch": 0.8762784559551303, "grad_norm": 0.2988683581352234, "learning_rate": 8e-05, "loss": 1.7167, "step": 3984 }, { "epoch": 0.8764984053667656, "grad_norm": 0.3110254406929016, "learning_rate": 8e-05, "loss": 1.6631, "step": 3985 }, { "epoch": 0.876718354778401, "grad_norm": 0.2963966429233551, "learning_rate": 8e-05, "loss": 1.7406, "step": 3986 }, { "epoch": 0.8769383041900363, "grad_norm": 0.3057374954223633, "learning_rate": 8e-05, "loss": 1.5424, "step": 3987 }, { "epoch": 0.8771582536016717, "grad_norm": 0.32236918807029724, "learning_rate": 8e-05, "loss": 1.6419, "step": 3988 }, { "epoch": 0.8773782030133069, "grad_norm": 0.35666584968566895, "learning_rate": 8e-05, "loss": 1.7104, "step": 3989 }, { "epoch": 0.8775981524249422, "grad_norm": 0.2982276380062103, "learning_rate": 8e-05, "loss": 1.5707, "step": 3990 }, { "epoch": 0.8778181018365776, "grad_norm": 0.28403279185295105, "learning_rate": 8e-05, "loss": 1.592, "step": 3991 }, { "epoch": 0.8780380512482129, "grad_norm": 0.3072356581687927, "learning_rate": 8e-05, "loss": 1.7964, "step": 3992 }, { "epoch": 0.8782580006598483, "grad_norm": 0.29571470618247986, "learning_rate": 8e-05, "loss": 1.7442, "step": 3993 }, { "epoch": 0.8784779500714835, "grad_norm": 0.30304571986198425, "learning_rate": 8e-05, "loss": 1.7008, "step": 3994 }, { "epoch": 0.8786978994831188, "grad_norm": 0.2830870747566223, "learning_rate": 8e-05, "loss": 1.6545, "step": 3995 }, { "epoch": 0.8789178488947542, "grad_norm": 0.3138338029384613, "learning_rate": 8e-05, "loss": 1.649, "step": 3996 }, { "epoch": 0.8791377983063895, "grad_norm": 0.28111809492111206, "learning_rate": 8e-05, "loss": 1.6918, "step": 3997 }, { "epoch": 0.8793577477180249, "grad_norm": 0.30100250244140625, "learning_rate": 8e-05, "loss": 1.7132, "step": 3998 }, { "epoch": 0.8795776971296602, "grad_norm": 0.3129050135612488, "learning_rate": 8e-05, "loss": 1.7433, "step": 3999 }, { "epoch": 0.8797976465412956, "grad_norm": 0.2925567030906677, "learning_rate": 8e-05, "loss": 1.7867, "step": 4000 }, { "epoch": 0.8800175959529308, "grad_norm": 0.28086230158805847, "learning_rate": 8e-05, "loss": 1.5538, "step": 4001 }, { "epoch": 0.8802375453645661, "grad_norm": 0.29163795709609985, "learning_rate": 8e-05, "loss": 1.6125, "step": 4002 }, { "epoch": 0.8804574947762015, "grad_norm": 0.278808057308197, "learning_rate": 8e-05, "loss": 1.4841, "step": 4003 }, { "epoch": 0.8806774441878368, "grad_norm": 0.3127332031726837, "learning_rate": 8e-05, "loss": 1.821, "step": 4004 }, { "epoch": 0.8808973935994722, "grad_norm": 0.3092391788959503, "learning_rate": 8e-05, "loss": 1.4779, "step": 4005 }, { "epoch": 0.8811173430111074, "grad_norm": 0.30426523089408875, "learning_rate": 8e-05, "loss": 1.7699, "step": 4006 }, { "epoch": 0.8813372924227427, "grad_norm": 0.28605568408966064, "learning_rate": 8e-05, "loss": 1.7, "step": 4007 }, { "epoch": 0.8815572418343781, "grad_norm": 0.2959664762020111, "learning_rate": 8e-05, "loss": 1.733, "step": 4008 }, { "epoch": 0.8817771912460134, "grad_norm": 0.33198082447052, "learning_rate": 8e-05, "loss": 1.7399, "step": 4009 }, { "epoch": 0.8819971406576488, "grad_norm": 0.3039609491825104, "learning_rate": 8e-05, "loss": 1.7289, "step": 4010 }, { "epoch": 0.8822170900692841, "grad_norm": 0.3058362603187561, "learning_rate": 8e-05, "loss": 1.7897, "step": 4011 }, { "epoch": 0.8824370394809193, "grad_norm": 0.3194586932659149, "learning_rate": 8e-05, "loss": 1.806, "step": 4012 }, { "epoch": 0.8826569888925547, "grad_norm": 0.31681060791015625, "learning_rate": 8e-05, "loss": 1.7131, "step": 4013 }, { "epoch": 0.88287693830419, "grad_norm": 0.2874566912651062, "learning_rate": 8e-05, "loss": 1.5243, "step": 4014 }, { "epoch": 0.8830968877158254, "grad_norm": 0.3297685980796814, "learning_rate": 8e-05, "loss": 1.9256, "step": 4015 }, { "epoch": 0.8833168371274607, "grad_norm": 0.30325692892074585, "learning_rate": 8e-05, "loss": 1.7809, "step": 4016 }, { "epoch": 0.883536786539096, "grad_norm": 0.38110217452049255, "learning_rate": 8e-05, "loss": 1.7246, "step": 4017 }, { "epoch": 0.8837567359507313, "grad_norm": 0.3073844313621521, "learning_rate": 8e-05, "loss": 1.6379, "step": 4018 }, { "epoch": 0.8839766853623666, "grad_norm": 0.3144959509372711, "learning_rate": 8e-05, "loss": 1.7369, "step": 4019 }, { "epoch": 0.884196634774002, "grad_norm": 0.29416364431381226, "learning_rate": 8e-05, "loss": 1.6181, "step": 4020 }, { "epoch": 0.8844165841856373, "grad_norm": 0.31616780161857605, "learning_rate": 8e-05, "loss": 1.6965, "step": 4021 }, { "epoch": 0.8846365335972727, "grad_norm": 0.3114292621612549, "learning_rate": 8e-05, "loss": 1.7151, "step": 4022 }, { "epoch": 0.884856483008908, "grad_norm": 0.2938895523548126, "learning_rate": 8e-05, "loss": 1.6919, "step": 4023 }, { "epoch": 0.8850764324205432, "grad_norm": 0.3129356801509857, "learning_rate": 8e-05, "loss": 1.8082, "step": 4024 }, { "epoch": 0.8852963818321786, "grad_norm": 0.2698216140270233, "learning_rate": 8e-05, "loss": 1.5458, "step": 4025 }, { "epoch": 0.8855163312438139, "grad_norm": 0.3061481714248657, "learning_rate": 8e-05, "loss": 1.7474, "step": 4026 }, { "epoch": 0.8857362806554493, "grad_norm": 0.3081664741039276, "learning_rate": 8e-05, "loss": 1.7645, "step": 4027 }, { "epoch": 0.8859562300670846, "grad_norm": 0.29356658458709717, "learning_rate": 8e-05, "loss": 1.7544, "step": 4028 }, { "epoch": 0.8861761794787198, "grad_norm": 0.287942498922348, "learning_rate": 8e-05, "loss": 1.5793, "step": 4029 }, { "epoch": 0.8863961288903552, "grad_norm": 0.2899230718612671, "learning_rate": 8e-05, "loss": 1.7263, "step": 4030 }, { "epoch": 0.8866160783019905, "grad_norm": 0.28524360060691833, "learning_rate": 8e-05, "loss": 1.6558, "step": 4031 }, { "epoch": 0.8868360277136259, "grad_norm": 0.34471920132637024, "learning_rate": 8e-05, "loss": 1.7586, "step": 4032 }, { "epoch": 0.8870559771252612, "grad_norm": 0.2993306815624237, "learning_rate": 8e-05, "loss": 1.8712, "step": 4033 }, { "epoch": 0.8872759265368965, "grad_norm": 0.30809563398361206, "learning_rate": 8e-05, "loss": 1.4949, "step": 4034 }, { "epoch": 0.8874958759485319, "grad_norm": 0.3870789110660553, "learning_rate": 8e-05, "loss": 1.7098, "step": 4035 }, { "epoch": 0.8877158253601671, "grad_norm": 0.27530455589294434, "learning_rate": 8e-05, "loss": 1.5475, "step": 4036 }, { "epoch": 0.8879357747718025, "grad_norm": 0.3203597366809845, "learning_rate": 8e-05, "loss": 1.5524, "step": 4037 }, { "epoch": 0.8881557241834378, "grad_norm": 0.2910875678062439, "learning_rate": 8e-05, "loss": 1.6297, "step": 4038 }, { "epoch": 0.8883756735950731, "grad_norm": 0.362775057554245, "learning_rate": 8e-05, "loss": 2.0182, "step": 4039 }, { "epoch": 0.8885956230067085, "grad_norm": 0.31645676493644714, "learning_rate": 8e-05, "loss": 1.7838, "step": 4040 }, { "epoch": 0.8888155724183437, "grad_norm": 0.3013925552368164, "learning_rate": 8e-05, "loss": 1.7254, "step": 4041 }, { "epoch": 0.8890355218299791, "grad_norm": 0.3068082928657532, "learning_rate": 8e-05, "loss": 1.7238, "step": 4042 }, { "epoch": 0.8892554712416144, "grad_norm": 0.29745063185691833, "learning_rate": 8e-05, "loss": 1.8604, "step": 4043 }, { "epoch": 0.8894754206532498, "grad_norm": 0.30625858902931213, "learning_rate": 8e-05, "loss": 1.8517, "step": 4044 }, { "epoch": 0.8896953700648851, "grad_norm": 0.2904675006866455, "learning_rate": 8e-05, "loss": 1.5672, "step": 4045 }, { "epoch": 0.8899153194765204, "grad_norm": 0.28286129236221313, "learning_rate": 8e-05, "loss": 1.6136, "step": 4046 }, { "epoch": 0.8901352688881558, "grad_norm": 0.2881507873535156, "learning_rate": 8e-05, "loss": 1.6415, "step": 4047 }, { "epoch": 0.890355218299791, "grad_norm": 0.2884863018989563, "learning_rate": 8e-05, "loss": 1.7578, "step": 4048 }, { "epoch": 0.8905751677114264, "grad_norm": 0.2926565706729889, "learning_rate": 8e-05, "loss": 1.6658, "step": 4049 }, { "epoch": 0.8907951171230617, "grad_norm": 0.3014330565929413, "learning_rate": 8e-05, "loss": 1.6047, "step": 4050 }, { "epoch": 0.891015066534697, "grad_norm": 0.3148139417171478, "learning_rate": 8e-05, "loss": 1.4756, "step": 4051 }, { "epoch": 0.8912350159463324, "grad_norm": 0.31396764516830444, "learning_rate": 8e-05, "loss": 1.8031, "step": 4052 }, { "epoch": 0.8914549653579676, "grad_norm": 0.2879650890827179, "learning_rate": 8e-05, "loss": 1.7601, "step": 4053 }, { "epoch": 0.891674914769603, "grad_norm": 0.2978127896785736, "learning_rate": 8e-05, "loss": 1.5957, "step": 4054 }, { "epoch": 0.8918948641812383, "grad_norm": 0.28600072860717773, "learning_rate": 8e-05, "loss": 1.629, "step": 4055 }, { "epoch": 0.8921148135928736, "grad_norm": 0.30629444122314453, "learning_rate": 8e-05, "loss": 1.7707, "step": 4056 }, { "epoch": 0.892334763004509, "grad_norm": 0.2934803366661072, "learning_rate": 8e-05, "loss": 1.669, "step": 4057 }, { "epoch": 0.8925547124161443, "grad_norm": 0.2665422558784485, "learning_rate": 8e-05, "loss": 1.4797, "step": 4058 }, { "epoch": 0.8927746618277796, "grad_norm": 0.2825581431388855, "learning_rate": 8e-05, "loss": 1.662, "step": 4059 }, { "epoch": 0.8929946112394149, "grad_norm": 0.2984130382537842, "learning_rate": 8e-05, "loss": 1.7903, "step": 4060 }, { "epoch": 0.8932145606510502, "grad_norm": 0.2899414300918579, "learning_rate": 8e-05, "loss": 1.6148, "step": 4061 }, { "epoch": 0.8934345100626856, "grad_norm": 0.29609423875808716, "learning_rate": 8e-05, "loss": 1.7082, "step": 4062 }, { "epoch": 0.8936544594743209, "grad_norm": 0.2987944185733795, "learning_rate": 8e-05, "loss": 1.7379, "step": 4063 }, { "epoch": 0.8938744088859563, "grad_norm": 0.3441619873046875, "learning_rate": 8e-05, "loss": 1.5096, "step": 4064 }, { "epoch": 0.8940943582975915, "grad_norm": 0.27636975049972534, "learning_rate": 8e-05, "loss": 1.4661, "step": 4065 }, { "epoch": 0.8943143077092269, "grad_norm": 0.2951783835887909, "learning_rate": 8e-05, "loss": 1.801, "step": 4066 }, { "epoch": 0.8945342571208622, "grad_norm": 0.2693262994289398, "learning_rate": 8e-05, "loss": 1.5644, "step": 4067 }, { "epoch": 0.8947542065324975, "grad_norm": 0.283433198928833, "learning_rate": 8e-05, "loss": 1.6118, "step": 4068 }, { "epoch": 0.8949741559441329, "grad_norm": 0.32869231700897217, "learning_rate": 8e-05, "loss": 1.7622, "step": 4069 }, { "epoch": 0.8951941053557682, "grad_norm": 0.29133233428001404, "learning_rate": 8e-05, "loss": 1.6443, "step": 4070 }, { "epoch": 0.8954140547674035, "grad_norm": 0.3477209210395813, "learning_rate": 8e-05, "loss": 1.9147, "step": 4071 }, { "epoch": 0.8956340041790388, "grad_norm": 0.30423182249069214, "learning_rate": 8e-05, "loss": 1.6524, "step": 4072 }, { "epoch": 0.8958539535906741, "grad_norm": 0.3046300411224365, "learning_rate": 8e-05, "loss": 1.5708, "step": 4073 }, { "epoch": 0.8960739030023095, "grad_norm": 0.31017544865608215, "learning_rate": 8e-05, "loss": 1.8136, "step": 4074 }, { "epoch": 0.8962938524139448, "grad_norm": 0.2947177588939667, "learning_rate": 8e-05, "loss": 1.6837, "step": 4075 }, { "epoch": 0.8965138018255802, "grad_norm": 0.3107057213783264, "learning_rate": 8e-05, "loss": 1.6765, "step": 4076 }, { "epoch": 0.8967337512372154, "grad_norm": 0.3040451109409332, "learning_rate": 8e-05, "loss": 1.6128, "step": 4077 }, { "epoch": 0.8969537006488507, "grad_norm": 0.31244924664497375, "learning_rate": 8e-05, "loss": 1.6391, "step": 4078 }, { "epoch": 0.8971736500604861, "grad_norm": 0.2981853485107422, "learning_rate": 8e-05, "loss": 1.6423, "step": 4079 }, { "epoch": 0.8973935994721214, "grad_norm": 0.3044477701187134, "learning_rate": 8e-05, "loss": 1.6362, "step": 4080 }, { "epoch": 0.8976135488837568, "grad_norm": 0.30110621452331543, "learning_rate": 8e-05, "loss": 1.6289, "step": 4081 }, { "epoch": 0.897833498295392, "grad_norm": 0.2834107577800751, "learning_rate": 8e-05, "loss": 1.593, "step": 4082 }, { "epoch": 0.8980534477070273, "grad_norm": 0.34450557827949524, "learning_rate": 8e-05, "loss": 1.7081, "step": 4083 }, { "epoch": 0.8982733971186627, "grad_norm": 0.27985870838165283, "learning_rate": 8e-05, "loss": 1.6028, "step": 4084 }, { "epoch": 0.898493346530298, "grad_norm": 0.31170904636383057, "learning_rate": 8e-05, "loss": 1.67, "step": 4085 }, { "epoch": 0.8987132959419334, "grad_norm": 0.30392032861709595, "learning_rate": 8e-05, "loss": 1.6134, "step": 4086 }, { "epoch": 0.8989332453535687, "grad_norm": 0.30208954215049744, "learning_rate": 8e-05, "loss": 1.7703, "step": 4087 }, { "epoch": 0.8991531947652041, "grad_norm": 0.2879715859889984, "learning_rate": 8e-05, "loss": 1.6479, "step": 4088 }, { "epoch": 0.8993731441768393, "grad_norm": 0.3037298619747162, "learning_rate": 8e-05, "loss": 1.5704, "step": 4089 }, { "epoch": 0.8995930935884746, "grad_norm": 0.3177938163280487, "learning_rate": 8e-05, "loss": 1.6422, "step": 4090 }, { "epoch": 0.89981304300011, "grad_norm": 0.28877830505371094, "learning_rate": 8e-05, "loss": 1.5779, "step": 4091 }, { "epoch": 0.9000329924117453, "grad_norm": 0.2998355031013489, "learning_rate": 8e-05, "loss": 1.6163, "step": 4092 }, { "epoch": 0.9002529418233807, "grad_norm": 0.28364381194114685, "learning_rate": 8e-05, "loss": 1.6757, "step": 4093 }, { "epoch": 0.900472891235016, "grad_norm": 0.31235194206237793, "learning_rate": 8e-05, "loss": 1.7467, "step": 4094 }, { "epoch": 0.9006928406466512, "grad_norm": 0.2888253927230835, "learning_rate": 8e-05, "loss": 1.6793, "step": 4095 }, { "epoch": 0.9009127900582866, "grad_norm": 0.3084707260131836, "learning_rate": 8e-05, "loss": 1.7183, "step": 4096 }, { "epoch": 0.9011327394699219, "grad_norm": 0.2737601697444916, "learning_rate": 8e-05, "loss": 1.4579, "step": 4097 }, { "epoch": 0.9013526888815573, "grad_norm": 0.2910381257534027, "learning_rate": 8e-05, "loss": 1.6289, "step": 4098 }, { "epoch": 0.9015726382931926, "grad_norm": 0.3083101809024811, "learning_rate": 8e-05, "loss": 1.8007, "step": 4099 }, { "epoch": 0.9017925877048278, "grad_norm": 0.2874706983566284, "learning_rate": 8e-05, "loss": 1.7087, "step": 4100 }, { "epoch": 0.9020125371164632, "grad_norm": 0.27825212478637695, "learning_rate": 8e-05, "loss": 1.5046, "step": 4101 }, { "epoch": 0.9022324865280985, "grad_norm": 0.30026623606681824, "learning_rate": 8e-05, "loss": 1.7011, "step": 4102 }, { "epoch": 0.9024524359397339, "grad_norm": 0.3117727041244507, "learning_rate": 8e-05, "loss": 1.8456, "step": 4103 }, { "epoch": 0.9026723853513692, "grad_norm": 0.2942452132701874, "learning_rate": 8e-05, "loss": 1.6479, "step": 4104 }, { "epoch": 0.9028923347630045, "grad_norm": 0.29466378688812256, "learning_rate": 8e-05, "loss": 1.6942, "step": 4105 }, { "epoch": 0.9031122841746398, "grad_norm": 0.2969743013381958, "learning_rate": 8e-05, "loss": 1.6088, "step": 4106 }, { "epoch": 0.9033322335862751, "grad_norm": 0.29155898094177246, "learning_rate": 8e-05, "loss": 1.7225, "step": 4107 }, { "epoch": 0.9035521829979105, "grad_norm": 0.27798759937286377, "learning_rate": 8e-05, "loss": 1.5386, "step": 4108 }, { "epoch": 0.9037721324095458, "grad_norm": 0.29623040556907654, "learning_rate": 8e-05, "loss": 1.7057, "step": 4109 }, { "epoch": 0.9039920818211812, "grad_norm": 0.28695133328437805, "learning_rate": 8e-05, "loss": 1.6513, "step": 4110 }, { "epoch": 0.9042120312328165, "grad_norm": 0.283668577671051, "learning_rate": 8e-05, "loss": 1.8331, "step": 4111 }, { "epoch": 0.9044319806444517, "grad_norm": 0.3282906115055084, "learning_rate": 8e-05, "loss": 1.7565, "step": 4112 }, { "epoch": 0.9046519300560871, "grad_norm": 0.2832440435886383, "learning_rate": 8e-05, "loss": 1.4466, "step": 4113 }, { "epoch": 0.9048718794677224, "grad_norm": 0.2924807369709015, "learning_rate": 8e-05, "loss": 1.6251, "step": 4114 }, { "epoch": 0.9050918288793578, "grad_norm": 0.30800655484199524, "learning_rate": 8e-05, "loss": 1.6581, "step": 4115 }, { "epoch": 0.9053117782909931, "grad_norm": 0.2765842080116272, "learning_rate": 8e-05, "loss": 1.6272, "step": 4116 }, { "epoch": 0.9055317277026284, "grad_norm": 0.29153522849082947, "learning_rate": 8e-05, "loss": 1.5286, "step": 4117 }, { "epoch": 0.9057516771142637, "grad_norm": 0.2986176013946533, "learning_rate": 8e-05, "loss": 1.5265, "step": 4118 }, { "epoch": 0.905971626525899, "grad_norm": 0.28605547547340393, "learning_rate": 8e-05, "loss": 1.6859, "step": 4119 }, { "epoch": 0.9061915759375344, "grad_norm": 0.28253173828125, "learning_rate": 8e-05, "loss": 1.6302, "step": 4120 }, { "epoch": 0.9064115253491697, "grad_norm": 0.2970622479915619, "learning_rate": 8e-05, "loss": 1.7511, "step": 4121 }, { "epoch": 0.906631474760805, "grad_norm": 0.2855907082557678, "learning_rate": 8e-05, "loss": 1.6366, "step": 4122 }, { "epoch": 0.9068514241724404, "grad_norm": 0.29107728600502014, "learning_rate": 8e-05, "loss": 1.7313, "step": 4123 }, { "epoch": 0.9070713735840756, "grad_norm": 0.2963238060474396, "learning_rate": 8e-05, "loss": 1.7331, "step": 4124 }, { "epoch": 0.907291322995711, "grad_norm": 0.27756524085998535, "learning_rate": 8e-05, "loss": 1.6023, "step": 4125 }, { "epoch": 0.9075112724073463, "grad_norm": 0.30239957571029663, "learning_rate": 8e-05, "loss": 1.7414, "step": 4126 }, { "epoch": 0.9077312218189816, "grad_norm": 0.28329378366470337, "learning_rate": 8e-05, "loss": 1.6604, "step": 4127 }, { "epoch": 0.907951171230617, "grad_norm": 0.2846319377422333, "learning_rate": 8e-05, "loss": 1.5306, "step": 4128 }, { "epoch": 0.9081711206422522, "grad_norm": 0.294981986284256, "learning_rate": 8e-05, "loss": 1.6668, "step": 4129 }, { "epoch": 0.9083910700538876, "grad_norm": 0.30630192160606384, "learning_rate": 8e-05, "loss": 1.789, "step": 4130 }, { "epoch": 0.9086110194655229, "grad_norm": 0.29193127155303955, "learning_rate": 8e-05, "loss": 1.7436, "step": 4131 }, { "epoch": 0.9088309688771583, "grad_norm": 0.28358370065689087, "learning_rate": 8e-05, "loss": 1.6301, "step": 4132 }, { "epoch": 0.9090509182887936, "grad_norm": 0.29850831627845764, "learning_rate": 8e-05, "loss": 1.7101, "step": 4133 }, { "epoch": 0.9092708677004289, "grad_norm": 0.3096507787704468, "learning_rate": 8e-05, "loss": 1.7025, "step": 4134 }, { "epoch": 0.9094908171120643, "grad_norm": 0.3215219974517822, "learning_rate": 8e-05, "loss": 1.5914, "step": 4135 }, { "epoch": 0.9097107665236995, "grad_norm": 0.3059280812740326, "learning_rate": 8e-05, "loss": 1.6822, "step": 4136 }, { "epoch": 0.9099307159353349, "grad_norm": 0.3068895936012268, "learning_rate": 8e-05, "loss": 1.63, "step": 4137 }, { "epoch": 0.9101506653469702, "grad_norm": 0.2992387115955353, "learning_rate": 8e-05, "loss": 1.7165, "step": 4138 }, { "epoch": 0.9103706147586055, "grad_norm": 0.29406991600990295, "learning_rate": 8e-05, "loss": 1.65, "step": 4139 }, { "epoch": 0.9105905641702409, "grad_norm": 0.2930237650871277, "learning_rate": 8e-05, "loss": 1.7292, "step": 4140 }, { "epoch": 0.9108105135818761, "grad_norm": 0.2610776722431183, "learning_rate": 8e-05, "loss": 1.4707, "step": 4141 }, { "epoch": 0.9110304629935115, "grad_norm": 0.2843222916126251, "learning_rate": 8e-05, "loss": 1.5374, "step": 4142 }, { "epoch": 0.9112504124051468, "grad_norm": 0.30256927013397217, "learning_rate": 8e-05, "loss": 1.7271, "step": 4143 }, { "epoch": 0.9114703618167821, "grad_norm": 0.29254963994026184, "learning_rate": 8e-05, "loss": 1.656, "step": 4144 }, { "epoch": 0.9116903112284175, "grad_norm": 0.2959951162338257, "learning_rate": 8e-05, "loss": 1.5848, "step": 4145 }, { "epoch": 0.9119102606400528, "grad_norm": 0.3133392632007599, "learning_rate": 8e-05, "loss": 1.6898, "step": 4146 }, { "epoch": 0.9121302100516882, "grad_norm": 0.31604188680648804, "learning_rate": 8e-05, "loss": 1.727, "step": 4147 }, { "epoch": 0.9123501594633234, "grad_norm": 0.28543493151664734, "learning_rate": 8e-05, "loss": 1.6329, "step": 4148 }, { "epoch": 0.9125701088749587, "grad_norm": 0.28241634368896484, "learning_rate": 8e-05, "loss": 1.5312, "step": 4149 }, { "epoch": 0.9127900582865941, "grad_norm": 0.2962937653064728, "learning_rate": 8e-05, "loss": 1.6362, "step": 4150 }, { "epoch": 0.9130100076982294, "grad_norm": 0.299084335565567, "learning_rate": 8e-05, "loss": 1.6091, "step": 4151 }, { "epoch": 0.9132299571098648, "grad_norm": 0.28923937678337097, "learning_rate": 8e-05, "loss": 1.6693, "step": 4152 }, { "epoch": 0.9134499065215, "grad_norm": 0.2867174446582794, "learning_rate": 8e-05, "loss": 1.5203, "step": 4153 }, { "epoch": 0.9136698559331354, "grad_norm": 0.28517308831214905, "learning_rate": 8e-05, "loss": 1.6159, "step": 4154 }, { "epoch": 0.9138898053447707, "grad_norm": 0.29169774055480957, "learning_rate": 8e-05, "loss": 1.707, "step": 4155 }, { "epoch": 0.914109754756406, "grad_norm": 0.28574511408805847, "learning_rate": 8e-05, "loss": 1.6871, "step": 4156 }, { "epoch": 0.9143297041680414, "grad_norm": 0.29836469888687134, "learning_rate": 8e-05, "loss": 1.6859, "step": 4157 }, { "epoch": 0.9145496535796767, "grad_norm": 0.32184654474258423, "learning_rate": 8e-05, "loss": 1.667, "step": 4158 }, { "epoch": 0.914769602991312, "grad_norm": 0.28499841690063477, "learning_rate": 8e-05, "loss": 1.6998, "step": 4159 }, { "epoch": 0.9149895524029473, "grad_norm": 0.28232523798942566, "learning_rate": 8e-05, "loss": 1.5786, "step": 4160 }, { "epoch": 0.9152095018145826, "grad_norm": 0.29549431800842285, "learning_rate": 8e-05, "loss": 1.6957, "step": 4161 }, { "epoch": 0.915429451226218, "grad_norm": 0.30055925250053406, "learning_rate": 8e-05, "loss": 1.6295, "step": 4162 }, { "epoch": 0.9156494006378533, "grad_norm": 0.30925673246383667, "learning_rate": 8e-05, "loss": 1.5685, "step": 4163 }, { "epoch": 0.9158693500494887, "grad_norm": 0.30817949771881104, "learning_rate": 8e-05, "loss": 1.732, "step": 4164 }, { "epoch": 0.9160892994611239, "grad_norm": 0.28672513365745544, "learning_rate": 8e-05, "loss": 1.6787, "step": 4165 }, { "epoch": 0.9163092488727592, "grad_norm": 0.30071011185646057, "learning_rate": 8e-05, "loss": 1.7844, "step": 4166 }, { "epoch": 0.9165291982843946, "grad_norm": 0.292388916015625, "learning_rate": 8e-05, "loss": 1.579, "step": 4167 }, { "epoch": 0.9167491476960299, "grad_norm": 0.31311121582984924, "learning_rate": 8e-05, "loss": 1.6822, "step": 4168 }, { "epoch": 0.9169690971076653, "grad_norm": 0.31428346037864685, "learning_rate": 8e-05, "loss": 1.7232, "step": 4169 }, { "epoch": 0.9171890465193006, "grad_norm": 0.30659395456314087, "learning_rate": 8e-05, "loss": 1.7332, "step": 4170 }, { "epoch": 0.9174089959309358, "grad_norm": 0.32651838660240173, "learning_rate": 8e-05, "loss": 1.7149, "step": 4171 }, { "epoch": 0.9176289453425712, "grad_norm": 0.3228572607040405, "learning_rate": 8e-05, "loss": 1.6848, "step": 4172 }, { "epoch": 0.9178488947542065, "grad_norm": 0.2901211380958557, "learning_rate": 8e-05, "loss": 1.7337, "step": 4173 }, { "epoch": 0.9180688441658419, "grad_norm": 0.2871779799461365, "learning_rate": 8e-05, "loss": 1.6134, "step": 4174 }, { "epoch": 0.9182887935774772, "grad_norm": 0.30921611189842224, "learning_rate": 8e-05, "loss": 1.7896, "step": 4175 }, { "epoch": 0.9185087429891126, "grad_norm": 0.2913828492164612, "learning_rate": 8e-05, "loss": 1.7384, "step": 4176 }, { "epoch": 0.9187286924007478, "grad_norm": 0.31601426005363464, "learning_rate": 8e-05, "loss": 1.6391, "step": 4177 }, { "epoch": 0.9189486418123831, "grad_norm": 0.2996780276298523, "learning_rate": 8e-05, "loss": 1.5581, "step": 4178 }, { "epoch": 0.9191685912240185, "grad_norm": 0.3079301416873932, "learning_rate": 8e-05, "loss": 1.5813, "step": 4179 }, { "epoch": 0.9193885406356538, "grad_norm": 0.28514519333839417, "learning_rate": 8e-05, "loss": 1.702, "step": 4180 }, { "epoch": 0.9196084900472892, "grad_norm": 0.29393357038497925, "learning_rate": 8e-05, "loss": 1.6957, "step": 4181 }, { "epoch": 0.9198284394589245, "grad_norm": 0.2881946563720703, "learning_rate": 8e-05, "loss": 1.6014, "step": 4182 }, { "epoch": 0.9200483888705597, "grad_norm": 0.2976199686527252, "learning_rate": 8e-05, "loss": 1.7531, "step": 4183 }, { "epoch": 0.9202683382821951, "grad_norm": 0.3015999495983124, "learning_rate": 8e-05, "loss": 1.6589, "step": 4184 }, { "epoch": 0.9204882876938304, "grad_norm": 0.2981117367744446, "learning_rate": 8e-05, "loss": 1.6802, "step": 4185 }, { "epoch": 0.9207082371054658, "grad_norm": 0.29663896560668945, "learning_rate": 8e-05, "loss": 1.7327, "step": 4186 }, { "epoch": 0.9209281865171011, "grad_norm": 0.28317975997924805, "learning_rate": 8e-05, "loss": 1.6129, "step": 4187 }, { "epoch": 0.9211481359287363, "grad_norm": 0.27301955223083496, "learning_rate": 8e-05, "loss": 1.3794, "step": 4188 }, { "epoch": 0.9213680853403717, "grad_norm": 0.28590822219848633, "learning_rate": 8e-05, "loss": 1.6962, "step": 4189 }, { "epoch": 0.921588034752007, "grad_norm": 0.29811057448387146, "learning_rate": 8e-05, "loss": 1.8319, "step": 4190 }, { "epoch": 0.9218079841636424, "grad_norm": 0.2986351549625397, "learning_rate": 8e-05, "loss": 1.6403, "step": 4191 }, { "epoch": 0.9220279335752777, "grad_norm": 0.29516395926475525, "learning_rate": 8e-05, "loss": 1.5754, "step": 4192 }, { "epoch": 0.922247882986913, "grad_norm": 0.29312869906425476, "learning_rate": 8e-05, "loss": 1.6533, "step": 4193 }, { "epoch": 0.9224678323985483, "grad_norm": 0.28571614623069763, "learning_rate": 8e-05, "loss": 1.6711, "step": 4194 }, { "epoch": 0.9226877818101836, "grad_norm": 0.28665637969970703, "learning_rate": 8e-05, "loss": 1.6586, "step": 4195 }, { "epoch": 0.922907731221819, "grad_norm": 0.3119889795780182, "learning_rate": 8e-05, "loss": 1.8049, "step": 4196 }, { "epoch": 0.9231276806334543, "grad_norm": 0.2987707853317261, "learning_rate": 8e-05, "loss": 1.8385, "step": 4197 }, { "epoch": 0.9233476300450897, "grad_norm": 0.2872779071331024, "learning_rate": 8e-05, "loss": 1.7374, "step": 4198 }, { "epoch": 0.923567579456725, "grad_norm": 0.30723926424980164, "learning_rate": 8e-05, "loss": 1.7982, "step": 4199 }, { "epoch": 0.9237875288683602, "grad_norm": 0.2957054376602173, "learning_rate": 8e-05, "loss": 1.712, "step": 4200 }, { "epoch": 0.9240074782799956, "grad_norm": 0.2941366136074066, "learning_rate": 8e-05, "loss": 1.6782, "step": 4201 }, { "epoch": 0.9242274276916309, "grad_norm": 0.3279740512371063, "learning_rate": 8e-05, "loss": 1.7321, "step": 4202 }, { "epoch": 0.9244473771032663, "grad_norm": 0.3030930459499359, "learning_rate": 8e-05, "loss": 1.6227, "step": 4203 }, { "epoch": 0.9246673265149016, "grad_norm": 0.29185232520103455, "learning_rate": 8e-05, "loss": 1.6529, "step": 4204 }, { "epoch": 0.9248872759265369, "grad_norm": 0.31919175386428833, "learning_rate": 8e-05, "loss": 1.6879, "step": 4205 }, { "epoch": 0.9251072253381722, "grad_norm": 0.2989010214805603, "learning_rate": 8e-05, "loss": 1.7958, "step": 4206 }, { "epoch": 0.9253271747498075, "grad_norm": 0.28767868876457214, "learning_rate": 8e-05, "loss": 1.6155, "step": 4207 }, { "epoch": 0.9255471241614429, "grad_norm": 0.30420902371406555, "learning_rate": 8e-05, "loss": 1.74, "step": 4208 }, { "epoch": 0.9257670735730782, "grad_norm": 0.28389647603034973, "learning_rate": 8e-05, "loss": 1.6953, "step": 4209 }, { "epoch": 0.9259870229847135, "grad_norm": 0.2950130105018616, "learning_rate": 8e-05, "loss": 1.8119, "step": 4210 }, { "epoch": 0.9262069723963489, "grad_norm": 0.28202083706855774, "learning_rate": 8e-05, "loss": 1.6032, "step": 4211 }, { "epoch": 0.9264269218079841, "grad_norm": 0.2873939871788025, "learning_rate": 8e-05, "loss": 1.6406, "step": 4212 }, { "epoch": 0.9266468712196195, "grad_norm": 0.29845505952835083, "learning_rate": 8e-05, "loss": 1.7685, "step": 4213 }, { "epoch": 0.9268668206312548, "grad_norm": 0.3224455714225769, "learning_rate": 8e-05, "loss": 1.8116, "step": 4214 }, { "epoch": 0.9270867700428901, "grad_norm": 0.2965445816516876, "learning_rate": 8e-05, "loss": 1.6149, "step": 4215 }, { "epoch": 0.9273067194545255, "grad_norm": 0.2944619357585907, "learning_rate": 8e-05, "loss": 1.5827, "step": 4216 }, { "epoch": 0.9275266688661608, "grad_norm": 0.29647067189216614, "learning_rate": 8e-05, "loss": 1.6907, "step": 4217 }, { "epoch": 0.9277466182777961, "grad_norm": 0.27989983558654785, "learning_rate": 8e-05, "loss": 1.5422, "step": 4218 }, { "epoch": 0.9279665676894314, "grad_norm": 0.30345508456230164, "learning_rate": 8e-05, "loss": 1.6439, "step": 4219 }, { "epoch": 0.9281865171010667, "grad_norm": 0.29629063606262207, "learning_rate": 8e-05, "loss": 1.5947, "step": 4220 }, { "epoch": 0.9284064665127021, "grad_norm": 0.29597344994544983, "learning_rate": 8e-05, "loss": 1.6203, "step": 4221 }, { "epoch": 0.9286264159243374, "grad_norm": 0.2957248389720917, "learning_rate": 8e-05, "loss": 1.5987, "step": 4222 }, { "epoch": 0.9288463653359728, "grad_norm": 0.2960350513458252, "learning_rate": 8e-05, "loss": 1.5776, "step": 4223 }, { "epoch": 0.929066314747608, "grad_norm": 0.29710423946380615, "learning_rate": 8e-05, "loss": 1.5518, "step": 4224 }, { "epoch": 0.9292862641592434, "grad_norm": 0.29832521080970764, "learning_rate": 8e-05, "loss": 1.6414, "step": 4225 }, { "epoch": 0.9295062135708787, "grad_norm": 0.26050087809562683, "learning_rate": 8e-05, "loss": 1.3936, "step": 4226 }, { "epoch": 0.929726162982514, "grad_norm": 0.3182241916656494, "learning_rate": 8e-05, "loss": 1.8863, "step": 4227 }, { "epoch": 0.9299461123941494, "grad_norm": 0.2917816936969757, "learning_rate": 8e-05, "loss": 1.7444, "step": 4228 }, { "epoch": 0.9301660618057846, "grad_norm": 0.30499497056007385, "learning_rate": 8e-05, "loss": 1.6962, "step": 4229 }, { "epoch": 0.93038601121742, "grad_norm": 0.30994856357574463, "learning_rate": 8e-05, "loss": 1.8838, "step": 4230 }, { "epoch": 0.9306059606290553, "grad_norm": 0.3050210177898407, "learning_rate": 8e-05, "loss": 1.6575, "step": 4231 }, { "epoch": 0.9308259100406906, "grad_norm": 0.2949892282485962, "learning_rate": 8e-05, "loss": 1.7391, "step": 4232 }, { "epoch": 0.931045859452326, "grad_norm": 0.2921696603298187, "learning_rate": 8e-05, "loss": 1.8789, "step": 4233 }, { "epoch": 0.9312658088639613, "grad_norm": 0.2918970584869385, "learning_rate": 8e-05, "loss": 1.5761, "step": 4234 }, { "epoch": 0.9314857582755967, "grad_norm": 0.2940721809864044, "learning_rate": 8e-05, "loss": 1.6322, "step": 4235 }, { "epoch": 0.9317057076872319, "grad_norm": 0.27402326464653015, "learning_rate": 8e-05, "loss": 1.5636, "step": 4236 }, { "epoch": 0.9319256570988672, "grad_norm": 0.34060537815093994, "learning_rate": 8e-05, "loss": 1.693, "step": 4237 }, { "epoch": 0.9321456065105026, "grad_norm": 0.2948283851146698, "learning_rate": 8e-05, "loss": 1.7477, "step": 4238 }, { "epoch": 0.9323655559221379, "grad_norm": 0.29711946845054626, "learning_rate": 8e-05, "loss": 1.654, "step": 4239 }, { "epoch": 0.9325855053337733, "grad_norm": 0.28953954577445984, "learning_rate": 8e-05, "loss": 1.6674, "step": 4240 }, { "epoch": 0.9328054547454085, "grad_norm": 0.30088186264038086, "learning_rate": 8e-05, "loss": 1.5417, "step": 4241 }, { "epoch": 0.9330254041570438, "grad_norm": 0.3125405013561249, "learning_rate": 8e-05, "loss": 1.7275, "step": 4242 }, { "epoch": 0.9332453535686792, "grad_norm": 0.2886843979358673, "learning_rate": 8e-05, "loss": 1.6494, "step": 4243 }, { "epoch": 0.9334653029803145, "grad_norm": 0.30982518196105957, "learning_rate": 8e-05, "loss": 1.8119, "step": 4244 }, { "epoch": 0.9336852523919499, "grad_norm": 0.27964144945144653, "learning_rate": 8e-05, "loss": 1.6068, "step": 4245 }, { "epoch": 0.9339052018035852, "grad_norm": 0.29574427008628845, "learning_rate": 8e-05, "loss": 1.7418, "step": 4246 }, { "epoch": 0.9341251512152206, "grad_norm": 0.3052821755409241, "learning_rate": 8e-05, "loss": 1.6943, "step": 4247 }, { "epoch": 0.9343451006268558, "grad_norm": 0.28099381923675537, "learning_rate": 8e-05, "loss": 1.5774, "step": 4248 }, { "epoch": 0.9345650500384911, "grad_norm": 0.2942703664302826, "learning_rate": 8e-05, "loss": 1.7297, "step": 4249 }, { "epoch": 0.9347849994501265, "grad_norm": 0.2937106490135193, "learning_rate": 8e-05, "loss": 1.6643, "step": 4250 }, { "epoch": 0.9350049488617618, "grad_norm": 0.30792105197906494, "learning_rate": 8e-05, "loss": 1.8709, "step": 4251 }, { "epoch": 0.9352248982733972, "grad_norm": 0.2843385338783264, "learning_rate": 8e-05, "loss": 1.6057, "step": 4252 }, { "epoch": 0.9354448476850324, "grad_norm": 0.30580613017082214, "learning_rate": 8e-05, "loss": 1.5744, "step": 4253 }, { "epoch": 0.9356647970966677, "grad_norm": 0.29129788279533386, "learning_rate": 8e-05, "loss": 1.6922, "step": 4254 }, { "epoch": 0.9358847465083031, "grad_norm": 0.31255224347114563, "learning_rate": 8e-05, "loss": 1.8804, "step": 4255 }, { "epoch": 0.9361046959199384, "grad_norm": 0.30788132548332214, "learning_rate": 8e-05, "loss": 1.7001, "step": 4256 }, { "epoch": 0.9363246453315738, "grad_norm": 0.2987869083881378, "learning_rate": 8e-05, "loss": 1.6763, "step": 4257 }, { "epoch": 0.9365445947432091, "grad_norm": 0.28915441036224365, "learning_rate": 8e-05, "loss": 1.5499, "step": 4258 }, { "epoch": 0.9367645441548443, "grad_norm": 0.3005789518356323, "learning_rate": 8e-05, "loss": 1.7255, "step": 4259 }, { "epoch": 0.9369844935664797, "grad_norm": 0.299762099981308, "learning_rate": 8e-05, "loss": 1.6768, "step": 4260 }, { "epoch": 0.937204442978115, "grad_norm": 0.3155803680419922, "learning_rate": 8e-05, "loss": 1.684, "step": 4261 }, { "epoch": 0.9374243923897504, "grad_norm": 0.2879777252674103, "learning_rate": 8e-05, "loss": 1.5999, "step": 4262 }, { "epoch": 0.9376443418013857, "grad_norm": 0.304275244474411, "learning_rate": 8e-05, "loss": 1.6666, "step": 4263 }, { "epoch": 0.937864291213021, "grad_norm": 0.30361175537109375, "learning_rate": 8e-05, "loss": 1.6459, "step": 4264 }, { "epoch": 0.9380842406246563, "grad_norm": 0.2961748242378235, "learning_rate": 8e-05, "loss": 1.6006, "step": 4265 }, { "epoch": 0.9383041900362916, "grad_norm": 0.30298399925231934, "learning_rate": 8e-05, "loss": 1.7134, "step": 4266 }, { "epoch": 0.938524139447927, "grad_norm": 0.2866550087928772, "learning_rate": 8e-05, "loss": 1.6643, "step": 4267 }, { "epoch": 0.9387440888595623, "grad_norm": 0.3346495032310486, "learning_rate": 8e-05, "loss": 1.6863, "step": 4268 }, { "epoch": 0.9389640382711977, "grad_norm": 0.2898212671279907, "learning_rate": 8e-05, "loss": 1.5883, "step": 4269 }, { "epoch": 0.939183987682833, "grad_norm": 0.29734665155410767, "learning_rate": 8e-05, "loss": 1.7226, "step": 4270 }, { "epoch": 0.9394039370944682, "grad_norm": 0.3220420181751251, "learning_rate": 8e-05, "loss": 1.663, "step": 4271 }, { "epoch": 0.9396238865061036, "grad_norm": 0.30133485794067383, "learning_rate": 8e-05, "loss": 1.4772, "step": 4272 }, { "epoch": 0.9398438359177389, "grad_norm": 0.28944581747055054, "learning_rate": 8e-05, "loss": 1.5403, "step": 4273 }, { "epoch": 0.9400637853293743, "grad_norm": 0.3234364092350006, "learning_rate": 8e-05, "loss": 1.8958, "step": 4274 }, { "epoch": 0.9402837347410096, "grad_norm": 0.3140423595905304, "learning_rate": 8e-05, "loss": 1.6434, "step": 4275 }, { "epoch": 0.9405036841526448, "grad_norm": 0.31199538707733154, "learning_rate": 8e-05, "loss": 1.7016, "step": 4276 }, { "epoch": 0.9407236335642802, "grad_norm": 0.31812262535095215, "learning_rate": 8e-05, "loss": 1.6401, "step": 4277 }, { "epoch": 0.9409435829759155, "grad_norm": 0.30784544348716736, "learning_rate": 8e-05, "loss": 1.745, "step": 4278 }, { "epoch": 0.9411635323875509, "grad_norm": 0.2914595901966095, "learning_rate": 8e-05, "loss": 1.6515, "step": 4279 }, { "epoch": 0.9413834817991862, "grad_norm": 0.3110421597957611, "learning_rate": 8e-05, "loss": 1.5437, "step": 4280 }, { "epoch": 0.9416034312108215, "grad_norm": 0.32427704334259033, "learning_rate": 8e-05, "loss": 1.7521, "step": 4281 }, { "epoch": 0.9418233806224569, "grad_norm": 0.3069719672203064, "learning_rate": 8e-05, "loss": 1.7473, "step": 4282 }, { "epoch": 0.9420433300340921, "grad_norm": 0.3008507192134857, "learning_rate": 8e-05, "loss": 1.6536, "step": 4283 }, { "epoch": 0.9422632794457275, "grad_norm": 0.28315114974975586, "learning_rate": 8e-05, "loss": 1.5788, "step": 4284 }, { "epoch": 0.9424832288573628, "grad_norm": 0.2987407445907593, "learning_rate": 8e-05, "loss": 1.7148, "step": 4285 }, { "epoch": 0.9427031782689981, "grad_norm": 0.2861090302467346, "learning_rate": 8e-05, "loss": 1.6858, "step": 4286 }, { "epoch": 0.9429231276806335, "grad_norm": 0.297574907541275, "learning_rate": 8e-05, "loss": 1.7461, "step": 4287 }, { "epoch": 0.9431430770922687, "grad_norm": 0.28463122248649597, "learning_rate": 8e-05, "loss": 1.6208, "step": 4288 }, { "epoch": 0.9433630265039041, "grad_norm": 0.3073120415210724, "learning_rate": 8e-05, "loss": 1.6601, "step": 4289 }, { "epoch": 0.9435829759155394, "grad_norm": 0.2836720049381256, "learning_rate": 8e-05, "loss": 1.525, "step": 4290 }, { "epoch": 0.9438029253271748, "grad_norm": 0.2875766456127167, "learning_rate": 8e-05, "loss": 1.6625, "step": 4291 }, { "epoch": 0.9440228747388101, "grad_norm": 0.3201231062412262, "learning_rate": 8e-05, "loss": 1.7481, "step": 4292 }, { "epoch": 0.9442428241504454, "grad_norm": 0.30221429467201233, "learning_rate": 8e-05, "loss": 1.6486, "step": 4293 }, { "epoch": 0.9444627735620807, "grad_norm": 0.3243577182292938, "learning_rate": 8e-05, "loss": 1.6943, "step": 4294 }, { "epoch": 0.944682722973716, "grad_norm": 0.2791222035884857, "learning_rate": 8e-05, "loss": 1.6036, "step": 4295 }, { "epoch": 0.9449026723853514, "grad_norm": 0.2943405210971832, "learning_rate": 8e-05, "loss": 1.6321, "step": 4296 }, { "epoch": 0.9451226217969867, "grad_norm": 0.303776353597641, "learning_rate": 8e-05, "loss": 1.6788, "step": 4297 }, { "epoch": 0.945342571208622, "grad_norm": 0.2904314398765564, "learning_rate": 8e-05, "loss": 1.5924, "step": 4298 }, { "epoch": 0.9455625206202574, "grad_norm": 0.29499560594558716, "learning_rate": 8e-05, "loss": 1.767, "step": 4299 }, { "epoch": 0.9457824700318926, "grad_norm": 0.2957536578178406, "learning_rate": 8e-05, "loss": 1.6784, "step": 4300 }, { "epoch": 0.946002419443528, "grad_norm": 0.3278692364692688, "learning_rate": 8e-05, "loss": 1.7183, "step": 4301 }, { "epoch": 0.9462223688551633, "grad_norm": 0.32448363304138184, "learning_rate": 8e-05, "loss": 1.7439, "step": 4302 }, { "epoch": 0.9464423182667986, "grad_norm": 0.29863062500953674, "learning_rate": 8e-05, "loss": 1.8053, "step": 4303 }, { "epoch": 0.946662267678434, "grad_norm": 0.29405081272125244, "learning_rate": 8e-05, "loss": 1.5501, "step": 4304 }, { "epoch": 0.9468822170900693, "grad_norm": 0.29146045446395874, "learning_rate": 8e-05, "loss": 1.6857, "step": 4305 }, { "epoch": 0.9471021665017046, "grad_norm": 0.31284937262535095, "learning_rate": 8e-05, "loss": 1.6328, "step": 4306 }, { "epoch": 0.9473221159133399, "grad_norm": 0.29520636796951294, "learning_rate": 8e-05, "loss": 1.6156, "step": 4307 }, { "epoch": 0.9475420653249752, "grad_norm": 0.3117775022983551, "learning_rate": 8e-05, "loss": 1.6738, "step": 4308 }, { "epoch": 0.9477620147366106, "grad_norm": 0.3293139338493347, "learning_rate": 8e-05, "loss": 1.6706, "step": 4309 }, { "epoch": 0.9479819641482459, "grad_norm": 0.30195489525794983, "learning_rate": 8e-05, "loss": 1.6393, "step": 4310 }, { "epoch": 0.9482019135598813, "grad_norm": 0.2867063581943512, "learning_rate": 8e-05, "loss": 1.6381, "step": 4311 }, { "epoch": 0.9484218629715165, "grad_norm": 0.31880491971969604, "learning_rate": 8e-05, "loss": 1.6891, "step": 4312 }, { "epoch": 0.9486418123831519, "grad_norm": 0.28975731134414673, "learning_rate": 8e-05, "loss": 1.6369, "step": 4313 }, { "epoch": 0.9488617617947872, "grad_norm": 0.30534425377845764, "learning_rate": 8e-05, "loss": 1.6479, "step": 4314 }, { "epoch": 0.9490817112064225, "grad_norm": 0.29459723830223083, "learning_rate": 8e-05, "loss": 1.6362, "step": 4315 }, { "epoch": 0.9493016606180579, "grad_norm": 0.31398990750312805, "learning_rate": 8e-05, "loss": 1.5502, "step": 4316 }, { "epoch": 0.9495216100296932, "grad_norm": 0.3026418089866638, "learning_rate": 8e-05, "loss": 1.8244, "step": 4317 }, { "epoch": 0.9497415594413285, "grad_norm": 0.29010850191116333, "learning_rate": 8e-05, "loss": 1.6547, "step": 4318 }, { "epoch": 0.9499615088529638, "grad_norm": 0.28357642889022827, "learning_rate": 8e-05, "loss": 1.562, "step": 4319 }, { "epoch": 0.9501814582645991, "grad_norm": 0.3127571940422058, "learning_rate": 8e-05, "loss": 1.5851, "step": 4320 }, { "epoch": 0.9504014076762345, "grad_norm": 0.31412258744239807, "learning_rate": 8e-05, "loss": 1.7569, "step": 4321 }, { "epoch": 0.9506213570878698, "grad_norm": 0.29025739431381226, "learning_rate": 8e-05, "loss": 1.6701, "step": 4322 }, { "epoch": 0.9508413064995052, "grad_norm": 0.36086446046829224, "learning_rate": 8e-05, "loss": 1.8504, "step": 4323 }, { "epoch": 0.9510612559111404, "grad_norm": 0.29903754591941833, "learning_rate": 8e-05, "loss": 1.8401, "step": 4324 }, { "epoch": 0.9512812053227757, "grad_norm": 0.31516164541244507, "learning_rate": 8e-05, "loss": 1.6723, "step": 4325 }, { "epoch": 0.9515011547344111, "grad_norm": 0.2996736168861389, "learning_rate": 8e-05, "loss": 1.8117, "step": 4326 }, { "epoch": 0.9517211041460464, "grad_norm": 0.2747475504875183, "learning_rate": 8e-05, "loss": 1.4716, "step": 4327 }, { "epoch": 0.9519410535576818, "grad_norm": 0.2930557429790497, "learning_rate": 8e-05, "loss": 1.7099, "step": 4328 }, { "epoch": 0.952161002969317, "grad_norm": 0.3209344446659088, "learning_rate": 8e-05, "loss": 1.7772, "step": 4329 }, { "epoch": 0.9523809523809523, "grad_norm": 0.2976103127002716, "learning_rate": 8e-05, "loss": 1.6168, "step": 4330 }, { "epoch": 0.9526009017925877, "grad_norm": 0.2832762598991394, "learning_rate": 8e-05, "loss": 1.5194, "step": 4331 }, { "epoch": 0.952820851204223, "grad_norm": 0.3471873998641968, "learning_rate": 8e-05, "loss": 1.6147, "step": 4332 }, { "epoch": 0.9530408006158584, "grad_norm": 0.323137491941452, "learning_rate": 8e-05, "loss": 1.6678, "step": 4333 }, { "epoch": 0.9532607500274937, "grad_norm": 0.3110518157482147, "learning_rate": 8e-05, "loss": 1.7733, "step": 4334 }, { "epoch": 0.953480699439129, "grad_norm": 0.30080094933509827, "learning_rate": 8e-05, "loss": 1.6189, "step": 4335 }, { "epoch": 0.9537006488507643, "grad_norm": 0.3108077645301819, "learning_rate": 8e-05, "loss": 1.6818, "step": 4336 }, { "epoch": 0.9539205982623996, "grad_norm": 0.32039645314216614, "learning_rate": 8e-05, "loss": 1.7102, "step": 4337 }, { "epoch": 0.954140547674035, "grad_norm": 0.29062017798423767, "learning_rate": 8e-05, "loss": 1.6722, "step": 4338 }, { "epoch": 0.9543604970856703, "grad_norm": 0.314155250787735, "learning_rate": 8e-05, "loss": 1.7383, "step": 4339 }, { "epoch": 0.9545804464973057, "grad_norm": 0.29008060693740845, "learning_rate": 8e-05, "loss": 1.5542, "step": 4340 }, { "epoch": 0.954800395908941, "grad_norm": 0.31939879059791565, "learning_rate": 8e-05, "loss": 1.7417, "step": 4341 }, { "epoch": 0.9550203453205762, "grad_norm": 0.32870885729789734, "learning_rate": 8e-05, "loss": 1.8094, "step": 4342 }, { "epoch": 0.9552402947322116, "grad_norm": 0.3048943877220154, "learning_rate": 8e-05, "loss": 1.5526, "step": 4343 }, { "epoch": 0.9554602441438469, "grad_norm": 0.3018760681152344, "learning_rate": 8e-05, "loss": 1.7414, "step": 4344 }, { "epoch": 0.9556801935554823, "grad_norm": 0.299770325422287, "learning_rate": 8e-05, "loss": 1.7656, "step": 4345 }, { "epoch": 0.9559001429671176, "grad_norm": 0.3222806751728058, "learning_rate": 8e-05, "loss": 1.6876, "step": 4346 }, { "epoch": 0.9561200923787528, "grad_norm": 0.2997550666332245, "learning_rate": 8e-05, "loss": 1.6424, "step": 4347 }, { "epoch": 0.9563400417903882, "grad_norm": 0.298745721578598, "learning_rate": 8e-05, "loss": 1.6877, "step": 4348 }, { "epoch": 0.9565599912020235, "grad_norm": 0.3091779947280884, "learning_rate": 8e-05, "loss": 1.6664, "step": 4349 }, { "epoch": 0.9567799406136589, "grad_norm": 0.29578372836112976, "learning_rate": 8e-05, "loss": 1.7281, "step": 4350 }, { "epoch": 0.9569998900252942, "grad_norm": 0.29080474376678467, "learning_rate": 8e-05, "loss": 1.7009, "step": 4351 }, { "epoch": 0.9572198394369295, "grad_norm": 0.30178776383399963, "learning_rate": 8e-05, "loss": 1.7002, "step": 4352 }, { "epoch": 0.9574397888485648, "grad_norm": 0.3175758719444275, "learning_rate": 8e-05, "loss": 1.8114, "step": 4353 }, { "epoch": 0.9576597382602001, "grad_norm": 0.2928202748298645, "learning_rate": 8e-05, "loss": 1.5807, "step": 4354 }, { "epoch": 0.9578796876718355, "grad_norm": 0.30198556184768677, "learning_rate": 8e-05, "loss": 1.6466, "step": 4355 }, { "epoch": 0.9580996370834708, "grad_norm": 0.30202385783195496, "learning_rate": 8e-05, "loss": 1.7117, "step": 4356 }, { "epoch": 0.9583195864951062, "grad_norm": 0.30362236499786377, "learning_rate": 8e-05, "loss": 1.7272, "step": 4357 }, { "epoch": 0.9585395359067415, "grad_norm": 0.29718053340911865, "learning_rate": 8e-05, "loss": 1.5618, "step": 4358 }, { "epoch": 0.9587594853183767, "grad_norm": 0.28676435351371765, "learning_rate": 8e-05, "loss": 1.5514, "step": 4359 }, { "epoch": 0.9589794347300121, "grad_norm": 0.35272395610809326, "learning_rate": 8e-05, "loss": 1.6748, "step": 4360 }, { "epoch": 0.9591993841416474, "grad_norm": 0.28872179985046387, "learning_rate": 8e-05, "loss": 1.6616, "step": 4361 }, { "epoch": 0.9594193335532828, "grad_norm": 0.3032558262348175, "learning_rate": 8e-05, "loss": 1.7008, "step": 4362 }, { "epoch": 0.9596392829649181, "grad_norm": 0.307427316904068, "learning_rate": 8e-05, "loss": 1.6507, "step": 4363 }, { "epoch": 0.9598592323765534, "grad_norm": 0.28583332896232605, "learning_rate": 8e-05, "loss": 1.578, "step": 4364 }, { "epoch": 0.9600791817881887, "grad_norm": 0.3125254809856415, "learning_rate": 8e-05, "loss": 1.6767, "step": 4365 }, { "epoch": 0.960299131199824, "grad_norm": 0.30561694502830505, "learning_rate": 8e-05, "loss": 1.797, "step": 4366 }, { "epoch": 0.9605190806114594, "grad_norm": 0.30248764157295227, "learning_rate": 8e-05, "loss": 1.5862, "step": 4367 }, { "epoch": 0.9607390300230947, "grad_norm": 0.3098229169845581, "learning_rate": 8e-05, "loss": 1.7568, "step": 4368 }, { "epoch": 0.96095897943473, "grad_norm": 0.30071017146110535, "learning_rate": 8e-05, "loss": 1.6039, "step": 4369 }, { "epoch": 0.9611789288463654, "grad_norm": 0.3008226454257965, "learning_rate": 8e-05, "loss": 1.7515, "step": 4370 }, { "epoch": 0.9613988782580006, "grad_norm": 0.3019086420536041, "learning_rate": 8e-05, "loss": 1.7186, "step": 4371 }, { "epoch": 0.961618827669636, "grad_norm": 0.2837902307510376, "learning_rate": 8e-05, "loss": 1.5705, "step": 4372 }, { "epoch": 0.9618387770812713, "grad_norm": 0.27051079273223877, "learning_rate": 8e-05, "loss": 1.5068, "step": 4373 }, { "epoch": 0.9620587264929066, "grad_norm": 0.3019634485244751, "learning_rate": 8e-05, "loss": 1.8051, "step": 4374 }, { "epoch": 0.962278675904542, "grad_norm": 0.29091939330101013, "learning_rate": 8e-05, "loss": 1.5171, "step": 4375 }, { "epoch": 0.9624986253161772, "grad_norm": 0.3006875514984131, "learning_rate": 8e-05, "loss": 1.619, "step": 4376 }, { "epoch": 0.9627185747278126, "grad_norm": 0.2964954078197479, "learning_rate": 8e-05, "loss": 1.5904, "step": 4377 }, { "epoch": 0.9629385241394479, "grad_norm": 0.2927916944026947, "learning_rate": 8e-05, "loss": 1.6793, "step": 4378 }, { "epoch": 0.9631584735510833, "grad_norm": 0.2932109236717224, "learning_rate": 8e-05, "loss": 1.655, "step": 4379 }, { "epoch": 0.9633784229627186, "grad_norm": 0.29698604345321655, "learning_rate": 8e-05, "loss": 1.6694, "step": 4380 }, { "epoch": 0.9635983723743539, "grad_norm": 0.28351879119873047, "learning_rate": 8e-05, "loss": 1.6355, "step": 4381 }, { "epoch": 0.9638183217859893, "grad_norm": 0.3210040330886841, "learning_rate": 8e-05, "loss": 1.7633, "step": 4382 }, { "epoch": 0.9640382711976245, "grad_norm": 0.2798837423324585, "learning_rate": 8e-05, "loss": 1.6254, "step": 4383 }, { "epoch": 0.9642582206092599, "grad_norm": 0.2931169867515564, "learning_rate": 8e-05, "loss": 1.6197, "step": 4384 }, { "epoch": 0.9644781700208952, "grad_norm": 0.2779824137687683, "learning_rate": 8e-05, "loss": 1.623, "step": 4385 }, { "epoch": 0.9646981194325305, "grad_norm": 0.3030729293823242, "learning_rate": 8e-05, "loss": 1.5768, "step": 4386 }, { "epoch": 0.9649180688441659, "grad_norm": 0.2920215427875519, "learning_rate": 8e-05, "loss": 1.5899, "step": 4387 }, { "epoch": 0.9651380182558011, "grad_norm": 0.30520427227020264, "learning_rate": 8e-05, "loss": 1.6882, "step": 4388 }, { "epoch": 0.9653579676674365, "grad_norm": 0.3132491409778595, "learning_rate": 8e-05, "loss": 1.8433, "step": 4389 }, { "epoch": 0.9655779170790718, "grad_norm": 0.305482417345047, "learning_rate": 8e-05, "loss": 1.7519, "step": 4390 }, { "epoch": 0.9657978664907071, "grad_norm": 0.3076193630695343, "learning_rate": 8e-05, "loss": 1.6502, "step": 4391 }, { "epoch": 0.9660178159023425, "grad_norm": 0.30150118470191956, "learning_rate": 8e-05, "loss": 1.6548, "step": 4392 }, { "epoch": 0.9662377653139778, "grad_norm": 0.3134932219982147, "learning_rate": 8e-05, "loss": 1.6598, "step": 4393 }, { "epoch": 0.9664577147256131, "grad_norm": 0.31480950117111206, "learning_rate": 8e-05, "loss": 1.654, "step": 4394 }, { "epoch": 0.9666776641372484, "grad_norm": 0.2868606150150299, "learning_rate": 8e-05, "loss": 1.7108, "step": 4395 }, { "epoch": 0.9668976135488837, "grad_norm": 0.2981705069541931, "learning_rate": 8e-05, "loss": 1.6678, "step": 4396 }, { "epoch": 0.9671175629605191, "grad_norm": 0.30914050340652466, "learning_rate": 8e-05, "loss": 1.7751, "step": 4397 }, { "epoch": 0.9673375123721544, "grad_norm": 0.2951439619064331, "learning_rate": 8e-05, "loss": 1.7323, "step": 4398 }, { "epoch": 0.9675574617837898, "grad_norm": 0.3048958480358124, "learning_rate": 8e-05, "loss": 1.6566, "step": 4399 }, { "epoch": 0.967777411195425, "grad_norm": 0.2965305745601654, "learning_rate": 8e-05, "loss": 1.7116, "step": 4400 }, { "epoch": 0.9679973606070604, "grad_norm": 0.2833409011363983, "learning_rate": 8e-05, "loss": 1.5868, "step": 4401 }, { "epoch": 0.9682173100186957, "grad_norm": 0.3302156925201416, "learning_rate": 8e-05, "loss": 1.7029, "step": 4402 }, { "epoch": 0.968437259430331, "grad_norm": 0.3424098789691925, "learning_rate": 8e-05, "loss": 1.7407, "step": 4403 }, { "epoch": 0.9686572088419664, "grad_norm": 0.3008859157562256, "learning_rate": 8e-05, "loss": 1.6784, "step": 4404 }, { "epoch": 0.9688771582536017, "grad_norm": 0.3263012170791626, "learning_rate": 8e-05, "loss": 1.759, "step": 4405 }, { "epoch": 0.969097107665237, "grad_norm": 0.2940591871738434, "learning_rate": 8e-05, "loss": 1.6771, "step": 4406 }, { "epoch": 0.9693170570768723, "grad_norm": 0.2776276767253876, "learning_rate": 8e-05, "loss": 1.5939, "step": 4407 }, { "epoch": 0.9695370064885076, "grad_norm": 0.3084886968135834, "learning_rate": 8e-05, "loss": 1.8447, "step": 4408 }, { "epoch": 0.969756955900143, "grad_norm": 0.30279961228370667, "learning_rate": 8e-05, "loss": 1.6919, "step": 4409 }, { "epoch": 0.9699769053117783, "grad_norm": 0.2889377772808075, "learning_rate": 8e-05, "loss": 1.6302, "step": 4410 }, { "epoch": 0.9701968547234137, "grad_norm": 0.3116249740123749, "learning_rate": 8e-05, "loss": 1.6945, "step": 4411 }, { "epoch": 0.9704168041350489, "grad_norm": 0.2997124493122101, "learning_rate": 8e-05, "loss": 1.5636, "step": 4412 }, { "epoch": 0.9706367535466842, "grad_norm": 0.29871103167533875, "learning_rate": 8e-05, "loss": 1.6513, "step": 4413 }, { "epoch": 0.9708567029583196, "grad_norm": 0.2757253348827362, "learning_rate": 8e-05, "loss": 1.5724, "step": 4414 }, { "epoch": 0.9710766523699549, "grad_norm": 0.29090118408203125, "learning_rate": 8e-05, "loss": 1.7293, "step": 4415 }, { "epoch": 0.9712966017815903, "grad_norm": 0.2890031337738037, "learning_rate": 8e-05, "loss": 1.6034, "step": 4416 }, { "epoch": 0.9715165511932256, "grad_norm": 0.2936271131038666, "learning_rate": 8e-05, "loss": 1.5668, "step": 4417 }, { "epoch": 0.9717365006048608, "grad_norm": 0.336303174495697, "learning_rate": 8e-05, "loss": 1.6313, "step": 4418 }, { "epoch": 0.9719564500164962, "grad_norm": 0.31786924600601196, "learning_rate": 8e-05, "loss": 1.7597, "step": 4419 }, { "epoch": 0.9721763994281315, "grad_norm": 0.3035045266151428, "learning_rate": 8e-05, "loss": 1.6743, "step": 4420 }, { "epoch": 0.9723963488397669, "grad_norm": 0.27620363235473633, "learning_rate": 8e-05, "loss": 1.4299, "step": 4421 }, { "epoch": 0.9726162982514022, "grad_norm": 0.2746250629425049, "learning_rate": 8e-05, "loss": 1.5365, "step": 4422 }, { "epoch": 0.9728362476630376, "grad_norm": 0.3262247145175934, "learning_rate": 8e-05, "loss": 1.7204, "step": 4423 }, { "epoch": 0.9730561970746728, "grad_norm": 0.31040772795677185, "learning_rate": 8e-05, "loss": 1.648, "step": 4424 }, { "epoch": 0.9732761464863081, "grad_norm": 0.2918784022331238, "learning_rate": 8e-05, "loss": 1.6991, "step": 4425 }, { "epoch": 0.9734960958979435, "grad_norm": 0.3132600784301758, "learning_rate": 8e-05, "loss": 1.682, "step": 4426 }, { "epoch": 0.9737160453095788, "grad_norm": 0.2929805517196655, "learning_rate": 8e-05, "loss": 1.6979, "step": 4427 }, { "epoch": 0.9739359947212142, "grad_norm": 0.30973419547080994, "learning_rate": 8e-05, "loss": 1.5817, "step": 4428 }, { "epoch": 0.9741559441328494, "grad_norm": 0.3428287208080292, "learning_rate": 8e-05, "loss": 1.8223, "step": 4429 }, { "epoch": 0.9743758935444847, "grad_norm": 0.29221585392951965, "learning_rate": 8e-05, "loss": 1.5699, "step": 4430 }, { "epoch": 0.9745958429561201, "grad_norm": 0.28617948293685913, "learning_rate": 8e-05, "loss": 1.5019, "step": 4431 }, { "epoch": 0.9748157923677554, "grad_norm": 0.31226715445518494, "learning_rate": 8e-05, "loss": 1.581, "step": 4432 }, { "epoch": 0.9750357417793908, "grad_norm": 0.3052622675895691, "learning_rate": 8e-05, "loss": 1.6488, "step": 4433 }, { "epoch": 0.9752556911910261, "grad_norm": 0.29430046677589417, "learning_rate": 8e-05, "loss": 1.5932, "step": 4434 }, { "epoch": 0.9754756406026613, "grad_norm": 0.2936963140964508, "learning_rate": 8e-05, "loss": 1.5609, "step": 4435 }, { "epoch": 0.9756955900142967, "grad_norm": 0.2764322757720947, "learning_rate": 8e-05, "loss": 1.6329, "step": 4436 }, { "epoch": 0.975915539425932, "grad_norm": 0.28910988569259644, "learning_rate": 8e-05, "loss": 1.59, "step": 4437 }, { "epoch": 0.9761354888375674, "grad_norm": 0.28807327151298523, "learning_rate": 8e-05, "loss": 1.6683, "step": 4438 }, { "epoch": 0.9763554382492027, "grad_norm": 0.29345518350601196, "learning_rate": 8e-05, "loss": 1.6477, "step": 4439 }, { "epoch": 0.976575387660838, "grad_norm": 0.28553545475006104, "learning_rate": 8e-05, "loss": 1.6333, "step": 4440 }, { "epoch": 0.9767953370724733, "grad_norm": 0.3055817484855652, "learning_rate": 8e-05, "loss": 1.7352, "step": 4441 }, { "epoch": 0.9770152864841086, "grad_norm": 0.26843395829200745, "learning_rate": 8e-05, "loss": 1.5031, "step": 4442 }, { "epoch": 0.977235235895744, "grad_norm": 0.29834216833114624, "learning_rate": 8e-05, "loss": 1.7358, "step": 4443 }, { "epoch": 0.9774551853073793, "grad_norm": 0.32550013065338135, "learning_rate": 8e-05, "loss": 1.5645, "step": 4444 }, { "epoch": 0.9776751347190146, "grad_norm": 0.29521921277046204, "learning_rate": 8e-05, "loss": 1.6498, "step": 4445 }, { "epoch": 0.97789508413065, "grad_norm": 0.29316624999046326, "learning_rate": 8e-05, "loss": 1.5499, "step": 4446 }, { "epoch": 0.9781150335422852, "grad_norm": 0.2895122468471527, "learning_rate": 8e-05, "loss": 1.6961, "step": 4447 }, { "epoch": 0.9783349829539206, "grad_norm": 0.3027707040309906, "learning_rate": 8e-05, "loss": 1.9138, "step": 4448 }, { "epoch": 0.9785549323655559, "grad_norm": 0.2765319049358368, "learning_rate": 8e-05, "loss": 1.6376, "step": 4449 }, { "epoch": 0.9787748817771913, "grad_norm": 0.28939464688301086, "learning_rate": 8e-05, "loss": 1.6433, "step": 4450 }, { "epoch": 0.9789948311888266, "grad_norm": 0.3009035289287567, "learning_rate": 8e-05, "loss": 1.6584, "step": 4451 }, { "epoch": 0.9792147806004619, "grad_norm": 0.30962783098220825, "learning_rate": 8e-05, "loss": 1.673, "step": 4452 }, { "epoch": 0.9794347300120972, "grad_norm": 0.27727317810058594, "learning_rate": 8e-05, "loss": 1.5244, "step": 4453 }, { "epoch": 0.9796546794237325, "grad_norm": 0.28220412135124207, "learning_rate": 8e-05, "loss": 1.4242, "step": 4454 }, { "epoch": 0.9798746288353679, "grad_norm": 0.29269176721572876, "learning_rate": 8e-05, "loss": 1.5952, "step": 4455 }, { "epoch": 0.9800945782470032, "grad_norm": 0.30379146337509155, "learning_rate": 8e-05, "loss": 1.6384, "step": 4456 }, { "epoch": 0.9803145276586385, "grad_norm": 0.3239672780036926, "learning_rate": 8e-05, "loss": 1.5808, "step": 4457 }, { "epoch": 0.9805344770702739, "grad_norm": 0.28674960136413574, "learning_rate": 8e-05, "loss": 1.4705, "step": 4458 }, { "epoch": 0.9807544264819091, "grad_norm": 0.2956444025039673, "learning_rate": 8e-05, "loss": 1.7746, "step": 4459 }, { "epoch": 0.9809743758935445, "grad_norm": 0.2885272800922394, "learning_rate": 8e-05, "loss": 1.5695, "step": 4460 }, { "epoch": 0.9811943253051798, "grad_norm": 0.29144173860549927, "learning_rate": 8e-05, "loss": 1.7118, "step": 4461 }, { "epoch": 0.9814142747168151, "grad_norm": 0.31238171458244324, "learning_rate": 8e-05, "loss": 1.7893, "step": 4462 }, { "epoch": 0.9816342241284505, "grad_norm": 0.2996987998485565, "learning_rate": 8e-05, "loss": 1.7206, "step": 4463 }, { "epoch": 0.9818541735400858, "grad_norm": 0.29437190294265747, "learning_rate": 8e-05, "loss": 1.6398, "step": 4464 }, { "epoch": 0.9820741229517211, "grad_norm": 0.28236857056617737, "learning_rate": 8e-05, "loss": 1.477, "step": 4465 }, { "epoch": 0.9822940723633564, "grad_norm": 0.290854811668396, "learning_rate": 8e-05, "loss": 1.6445, "step": 4466 }, { "epoch": 0.9825140217749917, "grad_norm": 0.28919798135757446, "learning_rate": 8e-05, "loss": 1.5369, "step": 4467 }, { "epoch": 0.9827339711866271, "grad_norm": 0.3495406210422516, "learning_rate": 8e-05, "loss": 1.8034, "step": 4468 }, { "epoch": 0.9829539205982624, "grad_norm": 0.2776128053665161, "learning_rate": 8e-05, "loss": 1.4966, "step": 4469 }, { "epoch": 0.9831738700098978, "grad_norm": 0.32572057843208313, "learning_rate": 8e-05, "loss": 1.6874, "step": 4470 }, { "epoch": 0.983393819421533, "grad_norm": 0.3179634213447571, "learning_rate": 8e-05, "loss": 1.5736, "step": 4471 }, { "epoch": 0.9836137688331684, "grad_norm": 0.30530181527137756, "learning_rate": 8e-05, "loss": 1.7474, "step": 4472 }, { "epoch": 0.9838337182448037, "grad_norm": 0.30747902393341064, "learning_rate": 8e-05, "loss": 1.7044, "step": 4473 }, { "epoch": 0.984053667656439, "grad_norm": 0.2948608100414276, "learning_rate": 8e-05, "loss": 1.6953, "step": 4474 }, { "epoch": 0.9842736170680744, "grad_norm": 0.32307469844818115, "learning_rate": 8e-05, "loss": 1.6535, "step": 4475 }, { "epoch": 0.9844935664797096, "grad_norm": 0.314513623714447, "learning_rate": 8e-05, "loss": 1.6058, "step": 4476 }, { "epoch": 0.984713515891345, "grad_norm": 0.30157166719436646, "learning_rate": 8e-05, "loss": 1.7344, "step": 4477 }, { "epoch": 0.9849334653029803, "grad_norm": 0.32073357701301575, "learning_rate": 8e-05, "loss": 1.5717, "step": 4478 }, { "epoch": 0.9851534147146156, "grad_norm": 0.29884982109069824, "learning_rate": 8e-05, "loss": 1.6779, "step": 4479 }, { "epoch": 0.985373364126251, "grad_norm": 0.3231385052204132, "learning_rate": 8e-05, "loss": 1.8388, "step": 4480 }, { "epoch": 0.9855933135378863, "grad_norm": 0.2887023091316223, "learning_rate": 8e-05, "loss": 1.6754, "step": 4481 }, { "epoch": 0.9858132629495217, "grad_norm": 0.2872041165828705, "learning_rate": 8e-05, "loss": 1.6078, "step": 4482 }, { "epoch": 0.9860332123611569, "grad_norm": 0.2921448051929474, "learning_rate": 8e-05, "loss": 1.5941, "step": 4483 }, { "epoch": 0.9862531617727922, "grad_norm": 0.3134671747684479, "learning_rate": 8e-05, "loss": 1.6015, "step": 4484 }, { "epoch": 0.9864731111844276, "grad_norm": 0.31237107515335083, "learning_rate": 8e-05, "loss": 1.6342, "step": 4485 }, { "epoch": 0.9866930605960629, "grad_norm": 0.28489091992378235, "learning_rate": 8e-05, "loss": 1.564, "step": 4486 }, { "epoch": 0.9869130100076983, "grad_norm": 0.300727903842926, "learning_rate": 8e-05, "loss": 1.7434, "step": 4487 }, { "epoch": 0.9871329594193335, "grad_norm": 0.28028106689453125, "learning_rate": 8e-05, "loss": 1.4857, "step": 4488 }, { "epoch": 0.9873529088309688, "grad_norm": 0.28882718086242676, "learning_rate": 8e-05, "loss": 1.6234, "step": 4489 }, { "epoch": 0.9875728582426042, "grad_norm": 0.3193677067756653, "learning_rate": 8e-05, "loss": 1.7581, "step": 4490 }, { "epoch": 0.9877928076542395, "grad_norm": 0.2990550398826599, "learning_rate": 8e-05, "loss": 1.55, "step": 4491 }, { "epoch": 0.9880127570658749, "grad_norm": 0.3127286434173584, "learning_rate": 8e-05, "loss": 1.8427, "step": 4492 }, { "epoch": 0.9882327064775102, "grad_norm": 0.3024497926235199, "learning_rate": 8e-05, "loss": 1.6602, "step": 4493 }, { "epoch": 0.9884526558891455, "grad_norm": 0.2969488799571991, "learning_rate": 8e-05, "loss": 1.7345, "step": 4494 }, { "epoch": 0.9886726053007808, "grad_norm": 0.2769564092159271, "learning_rate": 8e-05, "loss": 1.5995, "step": 4495 }, { "epoch": 0.9888925547124161, "grad_norm": 0.29575562477111816, "learning_rate": 8e-05, "loss": 1.6056, "step": 4496 }, { "epoch": 0.9891125041240515, "grad_norm": 0.3045266270637512, "learning_rate": 8e-05, "loss": 1.5789, "step": 4497 }, { "epoch": 0.9893324535356868, "grad_norm": 0.2926347553730011, "learning_rate": 8e-05, "loss": 1.5467, "step": 4498 }, { "epoch": 0.9895524029473222, "grad_norm": 0.2982698976993561, "learning_rate": 8e-05, "loss": 1.7507, "step": 4499 }, { "epoch": 0.9897723523589574, "grad_norm": 0.31570032238960266, "learning_rate": 8e-05, "loss": 1.6138, "step": 4500 }, { "epoch": 0.9899923017705927, "grad_norm": 0.29251834750175476, "learning_rate": 8e-05, "loss": 1.6789, "step": 4501 }, { "epoch": 0.9902122511822281, "grad_norm": 0.28332188725471497, "learning_rate": 8e-05, "loss": 1.517, "step": 4502 }, { "epoch": 0.9904322005938634, "grad_norm": 0.31072255969047546, "learning_rate": 8e-05, "loss": 1.8599, "step": 4503 }, { "epoch": 0.9906521500054988, "grad_norm": 0.3108634650707245, "learning_rate": 8e-05, "loss": 1.7261, "step": 4504 }, { "epoch": 0.9908720994171341, "grad_norm": 0.29612377285957336, "learning_rate": 8e-05, "loss": 1.5988, "step": 4505 }, { "epoch": 0.9910920488287693, "grad_norm": 0.29026466608047485, "learning_rate": 8e-05, "loss": 1.6283, "step": 4506 }, { "epoch": 0.9913119982404047, "grad_norm": 0.31100019812583923, "learning_rate": 8e-05, "loss": 1.7241, "step": 4507 }, { "epoch": 0.99153194765204, "grad_norm": 0.29672232270240784, "learning_rate": 8e-05, "loss": 1.49, "step": 4508 }, { "epoch": 0.9917518970636754, "grad_norm": 0.28797340393066406, "learning_rate": 8e-05, "loss": 1.6308, "step": 4509 }, { "epoch": 0.9919718464753107, "grad_norm": 0.3070496618747711, "learning_rate": 8e-05, "loss": 1.6575, "step": 4510 }, { "epoch": 0.992191795886946, "grad_norm": 0.2943253517150879, "learning_rate": 8e-05, "loss": 1.6108, "step": 4511 }, { "epoch": 0.9924117452985813, "grad_norm": 0.311239093542099, "learning_rate": 8e-05, "loss": 1.7555, "step": 4512 }, { "epoch": 0.9926316947102166, "grad_norm": 0.3025715947151184, "learning_rate": 8e-05, "loss": 1.6224, "step": 4513 }, { "epoch": 0.992851644121852, "grad_norm": 0.3275901675224304, "learning_rate": 8e-05, "loss": 1.7496, "step": 4514 }, { "epoch": 0.9930715935334873, "grad_norm": 0.2779105603694916, "learning_rate": 8e-05, "loss": 1.5484, "step": 4515 }, { "epoch": 0.9932915429451227, "grad_norm": 0.2887669503688812, "learning_rate": 8e-05, "loss": 1.6076, "step": 4516 }, { "epoch": 0.993511492356758, "grad_norm": 0.28616613149642944, "learning_rate": 8e-05, "loss": 1.6341, "step": 4517 }, { "epoch": 0.9937314417683932, "grad_norm": 0.2875531017780304, "learning_rate": 8e-05, "loss": 1.5669, "step": 4518 }, { "epoch": 0.9939513911800286, "grad_norm": 0.28946763277053833, "learning_rate": 8e-05, "loss": 1.634, "step": 4519 }, { "epoch": 0.9941713405916639, "grad_norm": 0.2973674535751343, "learning_rate": 8e-05, "loss": 1.8391, "step": 4520 }, { "epoch": 0.9943912900032993, "grad_norm": 0.3605339229106903, "learning_rate": 8e-05, "loss": 1.7529, "step": 4521 }, { "epoch": 0.9946112394149346, "grad_norm": 0.28719213604927063, "learning_rate": 8e-05, "loss": 1.6061, "step": 4522 }, { "epoch": 0.9948311888265698, "grad_norm": 0.29998138546943665, "learning_rate": 8e-05, "loss": 1.7375, "step": 4523 }, { "epoch": 0.9950511382382052, "grad_norm": 0.3062797486782074, "learning_rate": 8e-05, "loss": 1.625, "step": 4524 }, { "epoch": 0.9952710876498405, "grad_norm": 0.28792884945869446, "learning_rate": 8e-05, "loss": 1.6572, "step": 4525 }, { "epoch": 0.9954910370614759, "grad_norm": 0.33734482526779175, "learning_rate": 8e-05, "loss": 1.9469, "step": 4526 }, { "epoch": 0.9957109864731112, "grad_norm": 0.2947852313518524, "learning_rate": 8e-05, "loss": 1.6837, "step": 4527 }, { "epoch": 0.9959309358847465, "grad_norm": 0.30927664041519165, "learning_rate": 8e-05, "loss": 1.7504, "step": 4528 }, { "epoch": 0.9961508852963818, "grad_norm": 0.2859261631965637, "learning_rate": 8e-05, "loss": 1.7234, "step": 4529 }, { "epoch": 0.9963708347080171, "grad_norm": 0.29213234782218933, "learning_rate": 8e-05, "loss": 1.6802, "step": 4530 }, { "epoch": 0.9965907841196525, "grad_norm": 0.30377697944641113, "learning_rate": 8e-05, "loss": 1.7073, "step": 4531 }, { "epoch": 0.9968107335312878, "grad_norm": 0.32840946316719055, "learning_rate": 8e-05, "loss": 1.7767, "step": 4532 }, { "epoch": 0.9970306829429231, "grad_norm": 0.29963135719299316, "learning_rate": 8e-05, "loss": 1.6862, "step": 4533 }, { "epoch": 0.9972506323545585, "grad_norm": 0.28282636404037476, "learning_rate": 8e-05, "loss": 1.6298, "step": 4534 }, { "epoch": 0.9974705817661937, "grad_norm": 0.29182666540145874, "learning_rate": 8e-05, "loss": 1.6182, "step": 4535 }, { "epoch": 0.9976905311778291, "grad_norm": 0.290402352809906, "learning_rate": 8e-05, "loss": 1.7541, "step": 4536 }, { "epoch": 0.9979104805894644, "grad_norm": 0.2885053753852844, "learning_rate": 8e-05, "loss": 1.619, "step": 4537 }, { "epoch": 0.9981304300010998, "grad_norm": 0.30916741490364075, "learning_rate": 8e-05, "loss": 1.6159, "step": 4538 }, { "epoch": 0.9983503794127351, "grad_norm": 0.2906371057033539, "learning_rate": 8e-05, "loss": 1.5887, "step": 4539 }, { "epoch": 0.9985703288243704, "grad_norm": 0.2910885810852051, "learning_rate": 8e-05, "loss": 1.6174, "step": 4540 }, { "epoch": 0.9987902782360057, "grad_norm": 0.325016051530838, "learning_rate": 8e-05, "loss": 1.6684, "step": 4541 }, { "epoch": 0.999010227647641, "grad_norm": 0.29865211248397827, "learning_rate": 8e-05, "loss": 1.6947, "step": 4542 }, { "epoch": 0.9992301770592764, "grad_norm": 0.29670125246047974, "learning_rate": 8e-05, "loss": 1.678, "step": 4543 }, { "epoch": 0.9994501264709117, "grad_norm": 0.28062552213668823, "learning_rate": 8e-05, "loss": 1.5117, "step": 4544 }, { "epoch": 0.999670075882547, "grad_norm": 0.3313736319541931, "learning_rate": 8e-05, "loss": 1.7289, "step": 4545 }, { "epoch": 0.9998900252941824, "grad_norm": 0.3050709068775177, "learning_rate": 8e-05, "loss": 1.6675, "step": 4546 }, { "epoch": 1.0002199494116353, "grad_norm": 0.30780911445617676, "learning_rate": 8e-05, "loss": 1.7135, "step": 4547 }, { "epoch": 1.0004398988232706, "grad_norm": 0.3015933334827423, "learning_rate": 8e-05, "loss": 1.6974, "step": 4548 }, { "epoch": 1.000659848234906, "grad_norm": 0.33518365025520325, "learning_rate": 8e-05, "loss": 1.6258, "step": 4549 }, { "epoch": 1.0008797976465413, "grad_norm": 0.2950296700000763, "learning_rate": 8e-05, "loss": 1.592, "step": 4550 }, { "epoch": 1.0010997470581766, "grad_norm": 0.29944804310798645, "learning_rate": 8e-05, "loss": 1.4951, "step": 4551 }, { "epoch": 1.001319696469812, "grad_norm": 0.31224048137664795, "learning_rate": 8e-05, "loss": 1.5955, "step": 4552 }, { "epoch": 1.0015396458814472, "grad_norm": 0.3046097159385681, "learning_rate": 8e-05, "loss": 1.5619, "step": 4553 }, { "epoch": 1.0017595952930827, "grad_norm": 0.3094725012779236, "learning_rate": 8e-05, "loss": 1.5112, "step": 4554 }, { "epoch": 1.001979544704718, "grad_norm": 0.3166990578174591, "learning_rate": 8e-05, "loss": 1.5901, "step": 4555 }, { "epoch": 1.0021994941163532, "grad_norm": 0.31150323152542114, "learning_rate": 8e-05, "loss": 1.6286, "step": 4556 }, { "epoch": 1.0024194435279885, "grad_norm": 0.2905597388744354, "learning_rate": 8e-05, "loss": 1.5102, "step": 4557 }, { "epoch": 1.0026393929396238, "grad_norm": 0.3298214077949524, "learning_rate": 8e-05, "loss": 1.7204, "step": 4558 }, { "epoch": 1.0028593423512593, "grad_norm": 0.31622079014778137, "learning_rate": 8e-05, "loss": 1.6093, "step": 4559 }, { "epoch": 1.0030792917628946, "grad_norm": 0.3177032768726349, "learning_rate": 8e-05, "loss": 1.606, "step": 4560 }, { "epoch": 1.0032992411745298, "grad_norm": 0.3607257902622223, "learning_rate": 8e-05, "loss": 1.8143, "step": 4561 }, { "epoch": 1.0035191905861651, "grad_norm": 0.3000054359436035, "learning_rate": 8e-05, "loss": 1.577, "step": 4562 }, { "epoch": 1.0037391399978004, "grad_norm": 0.30288824439048767, "learning_rate": 8e-05, "loss": 1.5576, "step": 4563 }, { "epoch": 1.003959089409436, "grad_norm": 0.300060510635376, "learning_rate": 8e-05, "loss": 1.4984, "step": 4564 }, { "epoch": 1.0041790388210712, "grad_norm": 0.3101755976676941, "learning_rate": 8e-05, "loss": 1.6304, "step": 4565 }, { "epoch": 1.0043989882327065, "grad_norm": 0.31316637992858887, "learning_rate": 8e-05, "loss": 1.6561, "step": 4566 }, { "epoch": 1.0046189376443417, "grad_norm": 0.3024763762950897, "learning_rate": 8e-05, "loss": 1.53, "step": 4567 }, { "epoch": 1.004838887055977, "grad_norm": 0.31294289231300354, "learning_rate": 8e-05, "loss": 1.5028, "step": 4568 }, { "epoch": 1.0050588364676125, "grad_norm": 0.3366270363330841, "learning_rate": 8e-05, "loss": 1.6038, "step": 4569 }, { "epoch": 1.0052787858792478, "grad_norm": 0.3199102580547333, "learning_rate": 8e-05, "loss": 1.658, "step": 4570 }, { "epoch": 1.005498735290883, "grad_norm": 0.2984057664871216, "learning_rate": 8e-05, "loss": 1.4897, "step": 4571 }, { "epoch": 1.0057186847025184, "grad_norm": 0.3316095471382141, "learning_rate": 8e-05, "loss": 1.627, "step": 4572 }, { "epoch": 1.0059386341141539, "grad_norm": 0.3173426687717438, "learning_rate": 8e-05, "loss": 1.7036, "step": 4573 }, { "epoch": 1.0061585835257891, "grad_norm": 0.29876741766929626, "learning_rate": 8e-05, "loss": 1.5216, "step": 4574 }, { "epoch": 1.0063785329374244, "grad_norm": 0.29442480206489563, "learning_rate": 8e-05, "loss": 1.5123, "step": 4575 }, { "epoch": 1.0065984823490597, "grad_norm": 0.338815838098526, "learning_rate": 8e-05, "loss": 1.6351, "step": 4576 }, { "epoch": 1.006818431760695, "grad_norm": 0.3387576937675476, "learning_rate": 8e-05, "loss": 1.4414, "step": 4577 }, { "epoch": 1.0070383811723305, "grad_norm": 0.30262520909309387, "learning_rate": 8e-05, "loss": 1.5072, "step": 4578 }, { "epoch": 1.0072583305839657, "grad_norm": 0.3149118423461914, "learning_rate": 8e-05, "loss": 1.7121, "step": 4579 }, { "epoch": 1.007478279995601, "grad_norm": 0.32053691148757935, "learning_rate": 8e-05, "loss": 1.5372, "step": 4580 }, { "epoch": 1.0076982294072363, "grad_norm": 0.3288147449493408, "learning_rate": 8e-05, "loss": 1.4396, "step": 4581 }, { "epoch": 1.0079181788188716, "grad_norm": 0.3048933148384094, "learning_rate": 8e-05, "loss": 1.5685, "step": 4582 }, { "epoch": 1.008138128230507, "grad_norm": 0.30691149830818176, "learning_rate": 8e-05, "loss": 1.679, "step": 4583 }, { "epoch": 1.0083580776421424, "grad_norm": 0.31473320722579956, "learning_rate": 8e-05, "loss": 1.5019, "step": 4584 }, { "epoch": 1.0085780270537776, "grad_norm": 0.31538698077201843, "learning_rate": 8e-05, "loss": 1.5076, "step": 4585 }, { "epoch": 1.008797976465413, "grad_norm": 0.31171107292175293, "learning_rate": 8e-05, "loss": 1.3906, "step": 4586 }, { "epoch": 1.0090179258770482, "grad_norm": 0.30550599098205566, "learning_rate": 8e-05, "loss": 1.5685, "step": 4587 }, { "epoch": 1.0092378752886837, "grad_norm": 0.3264256417751312, "learning_rate": 8e-05, "loss": 1.6312, "step": 4588 }, { "epoch": 1.009457824700319, "grad_norm": 0.32684946060180664, "learning_rate": 8e-05, "loss": 1.6074, "step": 4589 }, { "epoch": 1.0096777741119543, "grad_norm": 0.3070155382156372, "learning_rate": 8e-05, "loss": 1.5128, "step": 4590 }, { "epoch": 1.0098977235235895, "grad_norm": 0.3298015594482422, "learning_rate": 8e-05, "loss": 1.6372, "step": 4591 }, { "epoch": 1.0101176729352248, "grad_norm": 0.30557525157928467, "learning_rate": 8e-05, "loss": 1.4791, "step": 4592 }, { "epoch": 1.0103376223468603, "grad_norm": 0.31129592657089233, "learning_rate": 8e-05, "loss": 1.6922, "step": 4593 }, { "epoch": 1.0105575717584956, "grad_norm": 0.3519895374774933, "learning_rate": 8e-05, "loss": 1.6122, "step": 4594 }, { "epoch": 1.0107775211701309, "grad_norm": 0.3283763527870178, "learning_rate": 8e-05, "loss": 1.5753, "step": 4595 }, { "epoch": 1.0109974705817661, "grad_norm": 0.2999582290649414, "learning_rate": 8e-05, "loss": 1.4436, "step": 4596 }, { "epoch": 1.0112174199934014, "grad_norm": 0.31484997272491455, "learning_rate": 8e-05, "loss": 1.6768, "step": 4597 }, { "epoch": 1.011437369405037, "grad_norm": 0.300536185503006, "learning_rate": 8e-05, "loss": 1.6283, "step": 4598 }, { "epoch": 1.0116573188166722, "grad_norm": 0.3382927477359772, "learning_rate": 8e-05, "loss": 1.6532, "step": 4599 }, { "epoch": 1.0118772682283075, "grad_norm": 0.33039289712905884, "learning_rate": 8e-05, "loss": 1.6642, "step": 4600 }, { "epoch": 1.0120972176399428, "grad_norm": 0.3481544554233551, "learning_rate": 8e-05, "loss": 1.7762, "step": 4601 }, { "epoch": 1.012317167051578, "grad_norm": 0.3230370581150055, "learning_rate": 8e-05, "loss": 1.7961, "step": 4602 }, { "epoch": 1.0125371164632135, "grad_norm": 0.3128516972064972, "learning_rate": 8e-05, "loss": 1.5489, "step": 4603 }, { "epoch": 1.0127570658748488, "grad_norm": 0.3378541171550751, "learning_rate": 8e-05, "loss": 1.5365, "step": 4604 }, { "epoch": 1.012977015286484, "grad_norm": 0.35683292150497437, "learning_rate": 8e-05, "loss": 1.6517, "step": 4605 }, { "epoch": 1.0131969646981194, "grad_norm": 0.33353936672210693, "learning_rate": 8e-05, "loss": 1.6275, "step": 4606 }, { "epoch": 1.0134169141097547, "grad_norm": 0.3142012059688568, "learning_rate": 8e-05, "loss": 1.6154, "step": 4607 }, { "epoch": 1.0136368635213902, "grad_norm": 0.32411718368530273, "learning_rate": 8e-05, "loss": 1.5863, "step": 4608 }, { "epoch": 1.0138568129330254, "grad_norm": 0.3335449993610382, "learning_rate": 8e-05, "loss": 1.5405, "step": 4609 }, { "epoch": 1.0140767623446607, "grad_norm": 0.32759708166122437, "learning_rate": 8e-05, "loss": 1.7108, "step": 4610 }, { "epoch": 1.014296711756296, "grad_norm": 0.2864172160625458, "learning_rate": 8e-05, "loss": 1.5193, "step": 4611 }, { "epoch": 1.0145166611679313, "grad_norm": 0.29074007272720337, "learning_rate": 8e-05, "loss": 1.4968, "step": 4612 }, { "epoch": 1.0147366105795668, "grad_norm": 0.32190489768981934, "learning_rate": 8e-05, "loss": 1.5788, "step": 4613 }, { "epoch": 1.014956559991202, "grad_norm": 0.2896258234977722, "learning_rate": 8e-05, "loss": 1.5404, "step": 4614 }, { "epoch": 1.0151765094028373, "grad_norm": 0.3059661090373993, "learning_rate": 8e-05, "loss": 1.5562, "step": 4615 }, { "epoch": 1.0153964588144726, "grad_norm": 0.30121976137161255, "learning_rate": 8e-05, "loss": 1.53, "step": 4616 }, { "epoch": 1.015616408226108, "grad_norm": 0.32223767042160034, "learning_rate": 8e-05, "loss": 1.6296, "step": 4617 }, { "epoch": 1.0158363576377434, "grad_norm": 0.320445716381073, "learning_rate": 8e-05, "loss": 1.3921, "step": 4618 }, { "epoch": 1.0160563070493787, "grad_norm": 0.30459362268447876, "learning_rate": 8e-05, "loss": 1.5742, "step": 4619 }, { "epoch": 1.016276256461014, "grad_norm": 0.3256053328514099, "learning_rate": 8e-05, "loss": 1.5106, "step": 4620 }, { "epoch": 1.0164962058726492, "grad_norm": 0.29514098167419434, "learning_rate": 8e-05, "loss": 1.6195, "step": 4621 }, { "epoch": 1.0167161552842847, "grad_norm": 0.30088508129119873, "learning_rate": 8e-05, "loss": 1.4971, "step": 4622 }, { "epoch": 1.01693610469592, "grad_norm": 0.34027406573295593, "learning_rate": 8e-05, "loss": 1.6916, "step": 4623 }, { "epoch": 1.0171560541075553, "grad_norm": 0.3266535997390747, "learning_rate": 8e-05, "loss": 1.556, "step": 4624 }, { "epoch": 1.0173760035191906, "grad_norm": 0.3183201253414154, "learning_rate": 8e-05, "loss": 1.6892, "step": 4625 }, { "epoch": 1.0175959529308258, "grad_norm": 0.28214922547340393, "learning_rate": 8e-05, "loss": 1.4613, "step": 4626 }, { "epoch": 1.0178159023424613, "grad_norm": 0.35053750872612, "learning_rate": 8e-05, "loss": 1.608, "step": 4627 }, { "epoch": 1.0180358517540966, "grad_norm": 0.2983100712299347, "learning_rate": 8e-05, "loss": 1.7106, "step": 4628 }, { "epoch": 1.0182558011657319, "grad_norm": 0.3033554255962372, "learning_rate": 8e-05, "loss": 1.4703, "step": 4629 }, { "epoch": 1.0184757505773672, "grad_norm": 0.36974287033081055, "learning_rate": 8e-05, "loss": 1.7206, "step": 4630 }, { "epoch": 1.0186956999890024, "grad_norm": 0.32487836480140686, "learning_rate": 8e-05, "loss": 1.6613, "step": 4631 }, { "epoch": 1.018915649400638, "grad_norm": 0.32653507590293884, "learning_rate": 8e-05, "loss": 1.6942, "step": 4632 }, { "epoch": 1.0191355988122732, "grad_norm": 0.3333089053630829, "learning_rate": 8e-05, "loss": 1.5744, "step": 4633 }, { "epoch": 1.0193555482239085, "grad_norm": 0.344862699508667, "learning_rate": 8e-05, "loss": 1.7667, "step": 4634 }, { "epoch": 1.0195754976355438, "grad_norm": 0.2952236831188202, "learning_rate": 8e-05, "loss": 1.5045, "step": 4635 }, { "epoch": 1.019795447047179, "grad_norm": 0.31951904296875, "learning_rate": 8e-05, "loss": 1.7083, "step": 4636 }, { "epoch": 1.0200153964588146, "grad_norm": 0.31778484582901, "learning_rate": 8e-05, "loss": 1.442, "step": 4637 }, { "epoch": 1.0202353458704498, "grad_norm": 0.30361348390579224, "learning_rate": 8e-05, "loss": 1.6425, "step": 4638 }, { "epoch": 1.0204552952820851, "grad_norm": 0.29916316270828247, "learning_rate": 8e-05, "loss": 1.5198, "step": 4639 }, { "epoch": 1.0206752446937204, "grad_norm": 0.293755441904068, "learning_rate": 8e-05, "loss": 1.6721, "step": 4640 }, { "epoch": 1.0208951941053557, "grad_norm": 0.3150803744792938, "learning_rate": 8e-05, "loss": 1.4918, "step": 4641 }, { "epoch": 1.0211151435169912, "grad_norm": 0.30666598677635193, "learning_rate": 8e-05, "loss": 1.6305, "step": 4642 }, { "epoch": 1.0213350929286265, "grad_norm": 0.2995915710926056, "learning_rate": 8e-05, "loss": 1.6421, "step": 4643 }, { "epoch": 1.0215550423402617, "grad_norm": 0.29210665822029114, "learning_rate": 8e-05, "loss": 1.4322, "step": 4644 }, { "epoch": 1.021774991751897, "grad_norm": 0.31763955950737, "learning_rate": 8e-05, "loss": 1.7082, "step": 4645 }, { "epoch": 1.0219949411635323, "grad_norm": 0.3176196217536926, "learning_rate": 8e-05, "loss": 1.7565, "step": 4646 }, { "epoch": 1.0222148905751678, "grad_norm": 0.3182832896709442, "learning_rate": 8e-05, "loss": 1.7318, "step": 4647 }, { "epoch": 1.022434839986803, "grad_norm": 0.30931442975997925, "learning_rate": 8e-05, "loss": 1.5531, "step": 4648 }, { "epoch": 1.0226547893984383, "grad_norm": 0.3212144076824188, "learning_rate": 8e-05, "loss": 1.536, "step": 4649 }, { "epoch": 1.0228747388100736, "grad_norm": 0.302397757768631, "learning_rate": 8e-05, "loss": 1.588, "step": 4650 }, { "epoch": 1.023094688221709, "grad_norm": 0.31494298577308655, "learning_rate": 8e-05, "loss": 1.6312, "step": 4651 }, { "epoch": 1.0233146376333444, "grad_norm": 0.322308748960495, "learning_rate": 8e-05, "loss": 1.739, "step": 4652 }, { "epoch": 1.0235345870449797, "grad_norm": 0.2970570921897888, "learning_rate": 8e-05, "loss": 1.5554, "step": 4653 }, { "epoch": 1.023754536456615, "grad_norm": 0.31364747881889343, "learning_rate": 8e-05, "loss": 1.6556, "step": 4654 }, { "epoch": 1.0239744858682502, "grad_norm": 0.31346267461776733, "learning_rate": 8e-05, "loss": 1.5616, "step": 4655 }, { "epoch": 1.0241944352798855, "grad_norm": 0.3113657236099243, "learning_rate": 8e-05, "loss": 1.517, "step": 4656 }, { "epoch": 1.024414384691521, "grad_norm": 0.3142276406288147, "learning_rate": 8e-05, "loss": 1.4922, "step": 4657 }, { "epoch": 1.0246343341031563, "grad_norm": 0.2907213866710663, "learning_rate": 8e-05, "loss": 1.5592, "step": 4658 }, { "epoch": 1.0248542835147916, "grad_norm": 0.30547502636909485, "learning_rate": 8e-05, "loss": 1.6472, "step": 4659 }, { "epoch": 1.0250742329264269, "grad_norm": 0.30475929379463196, "learning_rate": 8e-05, "loss": 1.5432, "step": 4660 }, { "epoch": 1.0252941823380624, "grad_norm": 0.321879506111145, "learning_rate": 8e-05, "loss": 1.5975, "step": 4661 }, { "epoch": 1.0255141317496976, "grad_norm": 0.2996005117893219, "learning_rate": 8e-05, "loss": 1.6203, "step": 4662 }, { "epoch": 1.025734081161333, "grad_norm": 0.3133268654346466, "learning_rate": 8e-05, "loss": 1.5739, "step": 4663 }, { "epoch": 1.0259540305729682, "grad_norm": 0.3288825750350952, "learning_rate": 8e-05, "loss": 1.5984, "step": 4664 }, { "epoch": 1.0261739799846035, "grad_norm": 0.31704506278038025, "learning_rate": 8e-05, "loss": 1.5251, "step": 4665 }, { "epoch": 1.026393929396239, "grad_norm": 0.3139953017234802, "learning_rate": 8e-05, "loss": 1.552, "step": 4666 }, { "epoch": 1.0266138788078742, "grad_norm": 0.31135937571525574, "learning_rate": 8e-05, "loss": 1.4934, "step": 4667 }, { "epoch": 1.0268338282195095, "grad_norm": 0.3169006109237671, "learning_rate": 8e-05, "loss": 1.6894, "step": 4668 }, { "epoch": 1.0270537776311448, "grad_norm": 0.31006866693496704, "learning_rate": 8e-05, "loss": 1.6107, "step": 4669 }, { "epoch": 1.02727372704278, "grad_norm": 0.3118205666542053, "learning_rate": 8e-05, "loss": 1.5491, "step": 4670 }, { "epoch": 1.0274936764544156, "grad_norm": 0.32314392924308777, "learning_rate": 8e-05, "loss": 1.5864, "step": 4671 }, { "epoch": 1.0277136258660509, "grad_norm": 0.3034246563911438, "learning_rate": 8e-05, "loss": 1.5226, "step": 4672 }, { "epoch": 1.0279335752776861, "grad_norm": 0.30772241950035095, "learning_rate": 8e-05, "loss": 1.7155, "step": 4673 }, { "epoch": 1.0281535246893214, "grad_norm": 0.3146454691886902, "learning_rate": 8e-05, "loss": 1.5578, "step": 4674 }, { "epoch": 1.0283734741009567, "grad_norm": 0.30904579162597656, "learning_rate": 8e-05, "loss": 1.6624, "step": 4675 }, { "epoch": 1.0285934235125922, "grad_norm": 0.3104855418205261, "learning_rate": 8e-05, "loss": 1.51, "step": 4676 }, { "epoch": 1.0288133729242275, "grad_norm": 0.3260266184806824, "learning_rate": 8e-05, "loss": 1.6716, "step": 4677 }, { "epoch": 1.0290333223358628, "grad_norm": 0.30246251821517944, "learning_rate": 8e-05, "loss": 1.5087, "step": 4678 }, { "epoch": 1.029253271747498, "grad_norm": 0.3298100531101227, "learning_rate": 8e-05, "loss": 1.665, "step": 4679 }, { "epoch": 1.0294732211591333, "grad_norm": 0.3139473795890808, "learning_rate": 8e-05, "loss": 1.6546, "step": 4680 }, { "epoch": 1.0296931705707688, "grad_norm": 0.29749056696891785, "learning_rate": 8e-05, "loss": 1.4827, "step": 4681 }, { "epoch": 1.029913119982404, "grad_norm": 0.2903030514717102, "learning_rate": 8e-05, "loss": 1.4577, "step": 4682 }, { "epoch": 1.0301330693940394, "grad_norm": 0.3296603262424469, "learning_rate": 8e-05, "loss": 1.6203, "step": 4683 }, { "epoch": 1.0303530188056746, "grad_norm": 0.3165712058544159, "learning_rate": 8e-05, "loss": 1.7285, "step": 4684 }, { "epoch": 1.03057296821731, "grad_norm": 0.3168621063232422, "learning_rate": 8e-05, "loss": 1.6063, "step": 4685 }, { "epoch": 1.0307929176289454, "grad_norm": 0.31027770042419434, "learning_rate": 8e-05, "loss": 1.6322, "step": 4686 }, { "epoch": 1.0310128670405807, "grad_norm": 0.3032212555408478, "learning_rate": 8e-05, "loss": 1.6127, "step": 4687 }, { "epoch": 1.031232816452216, "grad_norm": 0.3149271011352539, "learning_rate": 8e-05, "loss": 1.7228, "step": 4688 }, { "epoch": 1.0314527658638513, "grad_norm": 0.3136771619319916, "learning_rate": 8e-05, "loss": 1.6046, "step": 4689 }, { "epoch": 1.0316727152754865, "grad_norm": 0.29865020513534546, "learning_rate": 8e-05, "loss": 1.5209, "step": 4690 }, { "epoch": 1.031892664687122, "grad_norm": 0.31305360794067383, "learning_rate": 8e-05, "loss": 1.5171, "step": 4691 }, { "epoch": 1.0321126140987573, "grad_norm": 0.3163146674633026, "learning_rate": 8e-05, "loss": 1.6065, "step": 4692 }, { "epoch": 1.0323325635103926, "grad_norm": 0.332800954580307, "learning_rate": 8e-05, "loss": 1.663, "step": 4693 }, { "epoch": 1.0325525129220279, "grad_norm": 0.311379075050354, "learning_rate": 8e-05, "loss": 1.4946, "step": 4694 }, { "epoch": 1.0327724623336632, "grad_norm": 0.3351609408855438, "learning_rate": 8e-05, "loss": 1.6556, "step": 4695 }, { "epoch": 1.0329924117452987, "grad_norm": 0.33250653743743896, "learning_rate": 8e-05, "loss": 1.6326, "step": 4696 }, { "epoch": 1.033212361156934, "grad_norm": 0.3125477135181427, "learning_rate": 8e-05, "loss": 1.547, "step": 4697 }, { "epoch": 1.0334323105685692, "grad_norm": 0.3594234883785248, "learning_rate": 8e-05, "loss": 1.6578, "step": 4698 }, { "epoch": 1.0336522599802045, "grad_norm": 0.328444242477417, "learning_rate": 8e-05, "loss": 1.5171, "step": 4699 }, { "epoch": 1.0338722093918398, "grad_norm": 0.2991880178451538, "learning_rate": 8e-05, "loss": 1.5349, "step": 4700 }, { "epoch": 1.0340921588034753, "grad_norm": 0.3097067177295685, "learning_rate": 8e-05, "loss": 1.4742, "step": 4701 }, { "epoch": 1.0343121082151105, "grad_norm": 0.3114356994628906, "learning_rate": 8e-05, "loss": 1.627, "step": 4702 }, { "epoch": 1.0345320576267458, "grad_norm": 0.3157333433628082, "learning_rate": 8e-05, "loss": 1.5068, "step": 4703 }, { "epoch": 1.034752007038381, "grad_norm": 0.3133704364299774, "learning_rate": 8e-05, "loss": 1.6701, "step": 4704 }, { "epoch": 1.0349719564500166, "grad_norm": 0.3917182981967926, "learning_rate": 8e-05, "loss": 1.7513, "step": 4705 }, { "epoch": 1.0351919058616519, "grad_norm": 0.3267260789871216, "learning_rate": 8e-05, "loss": 1.6925, "step": 4706 }, { "epoch": 1.0354118552732872, "grad_norm": 0.3231751024723053, "learning_rate": 8e-05, "loss": 1.6328, "step": 4707 }, { "epoch": 1.0356318046849224, "grad_norm": 0.33699479699134827, "learning_rate": 8e-05, "loss": 1.5769, "step": 4708 }, { "epoch": 1.0358517540965577, "grad_norm": 0.31869399547576904, "learning_rate": 8e-05, "loss": 1.6545, "step": 4709 }, { "epoch": 1.0360717035081932, "grad_norm": 0.34719544649124146, "learning_rate": 8e-05, "loss": 1.5932, "step": 4710 }, { "epoch": 1.0362916529198285, "grad_norm": 0.3298405110836029, "learning_rate": 8e-05, "loss": 1.5435, "step": 4711 }, { "epoch": 1.0365116023314638, "grad_norm": 0.29324913024902344, "learning_rate": 8e-05, "loss": 1.4689, "step": 4712 }, { "epoch": 1.036731551743099, "grad_norm": 0.3122355341911316, "learning_rate": 8e-05, "loss": 1.4814, "step": 4713 }, { "epoch": 1.0369515011547343, "grad_norm": 0.2854824662208557, "learning_rate": 8e-05, "loss": 1.4768, "step": 4714 }, { "epoch": 1.0371714505663698, "grad_norm": 0.3275965452194214, "learning_rate": 8e-05, "loss": 1.7056, "step": 4715 }, { "epoch": 1.0373913999780051, "grad_norm": 0.3156280815601349, "learning_rate": 8e-05, "loss": 1.5638, "step": 4716 }, { "epoch": 1.0376113493896404, "grad_norm": 0.36162593960762024, "learning_rate": 8e-05, "loss": 1.8022, "step": 4717 }, { "epoch": 1.0378312988012757, "grad_norm": 0.3237744867801666, "learning_rate": 8e-05, "loss": 1.4659, "step": 4718 }, { "epoch": 1.038051248212911, "grad_norm": 0.3478328585624695, "learning_rate": 8e-05, "loss": 1.6539, "step": 4719 }, { "epoch": 1.0382711976245464, "grad_norm": 0.3335779905319214, "learning_rate": 8e-05, "loss": 1.6616, "step": 4720 }, { "epoch": 1.0384911470361817, "grad_norm": 0.3193053603172302, "learning_rate": 8e-05, "loss": 1.6958, "step": 4721 }, { "epoch": 1.038711096447817, "grad_norm": 0.3104461133480072, "learning_rate": 8e-05, "loss": 1.5925, "step": 4722 }, { "epoch": 1.0389310458594523, "grad_norm": 0.3187297582626343, "learning_rate": 8e-05, "loss": 1.7027, "step": 4723 }, { "epoch": 1.0391509952710876, "grad_norm": 0.3145037889480591, "learning_rate": 8e-05, "loss": 1.6404, "step": 4724 }, { "epoch": 1.039370944682723, "grad_norm": 0.3282637894153595, "learning_rate": 8e-05, "loss": 1.82, "step": 4725 }, { "epoch": 1.0395908940943583, "grad_norm": 0.31352725625038147, "learning_rate": 8e-05, "loss": 1.564, "step": 4726 }, { "epoch": 1.0398108435059936, "grad_norm": 0.34648942947387695, "learning_rate": 8e-05, "loss": 1.599, "step": 4727 }, { "epoch": 1.040030792917629, "grad_norm": 0.3547716438770294, "learning_rate": 8e-05, "loss": 1.5763, "step": 4728 }, { "epoch": 1.0402507423292642, "grad_norm": 0.3460184335708618, "learning_rate": 8e-05, "loss": 1.6263, "step": 4729 }, { "epoch": 1.0404706917408997, "grad_norm": 0.30453264713287354, "learning_rate": 8e-05, "loss": 1.6035, "step": 4730 }, { "epoch": 1.040690641152535, "grad_norm": 0.3143294155597687, "learning_rate": 8e-05, "loss": 1.6496, "step": 4731 }, { "epoch": 1.0409105905641702, "grad_norm": 0.33055970072746277, "learning_rate": 8e-05, "loss": 1.6192, "step": 4732 }, { "epoch": 1.0411305399758055, "grad_norm": 0.3228357434272766, "learning_rate": 8e-05, "loss": 1.4648, "step": 4733 }, { "epoch": 1.0413504893874408, "grad_norm": 0.320232629776001, "learning_rate": 8e-05, "loss": 1.5879, "step": 4734 }, { "epoch": 1.0415704387990763, "grad_norm": 0.32477840781211853, "learning_rate": 8e-05, "loss": 1.5708, "step": 4735 }, { "epoch": 1.0417903882107116, "grad_norm": 0.3234589397907257, "learning_rate": 8e-05, "loss": 1.4557, "step": 4736 }, { "epoch": 1.0420103376223468, "grad_norm": 0.337787002325058, "learning_rate": 8e-05, "loss": 1.6308, "step": 4737 }, { "epoch": 1.0422302870339821, "grad_norm": 0.32591602206230164, "learning_rate": 8e-05, "loss": 1.5828, "step": 4738 }, { "epoch": 1.0424502364456174, "grad_norm": 0.3175657093524933, "learning_rate": 8e-05, "loss": 1.6663, "step": 4739 }, { "epoch": 1.042670185857253, "grad_norm": 0.34104838967323303, "learning_rate": 8e-05, "loss": 1.7052, "step": 4740 }, { "epoch": 1.0428901352688882, "grad_norm": 0.3005687892436981, "learning_rate": 8e-05, "loss": 1.5028, "step": 4741 }, { "epoch": 1.0431100846805235, "grad_norm": 0.3174095153808594, "learning_rate": 8e-05, "loss": 1.5769, "step": 4742 }, { "epoch": 1.0433300340921587, "grad_norm": 0.30731111764907837, "learning_rate": 8e-05, "loss": 1.528, "step": 4743 }, { "epoch": 1.043549983503794, "grad_norm": 0.31795534491539, "learning_rate": 8e-05, "loss": 1.5162, "step": 4744 }, { "epoch": 1.0437699329154295, "grad_norm": 0.32395127415657043, "learning_rate": 8e-05, "loss": 1.6462, "step": 4745 }, { "epoch": 1.0439898823270648, "grad_norm": 0.31554362177848816, "learning_rate": 8e-05, "loss": 1.6257, "step": 4746 }, { "epoch": 1.0442098317387, "grad_norm": 0.3178892731666565, "learning_rate": 8e-05, "loss": 1.5241, "step": 4747 }, { "epoch": 1.0444297811503354, "grad_norm": 0.32683679461479187, "learning_rate": 8e-05, "loss": 1.4666, "step": 4748 }, { "epoch": 1.0446497305619706, "grad_norm": 0.32143253087997437, "learning_rate": 8e-05, "loss": 1.5521, "step": 4749 }, { "epoch": 1.0448696799736061, "grad_norm": 0.3408335745334625, "learning_rate": 8e-05, "loss": 1.5882, "step": 4750 }, { "epoch": 1.0450896293852414, "grad_norm": 0.32550984621047974, "learning_rate": 8e-05, "loss": 1.4965, "step": 4751 }, { "epoch": 1.0453095787968767, "grad_norm": 0.34926268458366394, "learning_rate": 8e-05, "loss": 1.7002, "step": 4752 }, { "epoch": 1.045529528208512, "grad_norm": 0.3265923857688904, "learning_rate": 8e-05, "loss": 1.7342, "step": 4753 }, { "epoch": 1.0457494776201475, "grad_norm": 0.3274342715740204, "learning_rate": 8e-05, "loss": 1.4837, "step": 4754 }, { "epoch": 1.0459694270317828, "grad_norm": 0.3310771584510803, "learning_rate": 8e-05, "loss": 1.5561, "step": 4755 }, { "epoch": 1.046189376443418, "grad_norm": 0.32436877489089966, "learning_rate": 8e-05, "loss": 1.5857, "step": 4756 }, { "epoch": 1.0464093258550533, "grad_norm": 0.3209306001663208, "learning_rate": 8e-05, "loss": 1.63, "step": 4757 }, { "epoch": 1.0466292752666886, "grad_norm": 0.32431018352508545, "learning_rate": 8e-05, "loss": 1.5799, "step": 4758 }, { "epoch": 1.046849224678324, "grad_norm": 0.3614148497581482, "learning_rate": 8e-05, "loss": 1.7825, "step": 4759 }, { "epoch": 1.0470691740899594, "grad_norm": 0.30742087960243225, "learning_rate": 8e-05, "loss": 1.5974, "step": 4760 }, { "epoch": 1.0472891235015946, "grad_norm": 0.32750600576400757, "learning_rate": 8e-05, "loss": 1.5656, "step": 4761 }, { "epoch": 1.04750907291323, "grad_norm": 0.3058852255344391, "learning_rate": 8e-05, "loss": 1.4428, "step": 4762 }, { "epoch": 1.0477290223248652, "grad_norm": 0.33409029245376587, "learning_rate": 8e-05, "loss": 1.7169, "step": 4763 }, { "epoch": 1.0479489717365007, "grad_norm": 0.3593020439147949, "learning_rate": 8e-05, "loss": 1.7302, "step": 4764 }, { "epoch": 1.048168921148136, "grad_norm": 0.32970818877220154, "learning_rate": 8e-05, "loss": 1.5177, "step": 4765 }, { "epoch": 1.0483888705597713, "grad_norm": 0.32168370485305786, "learning_rate": 8e-05, "loss": 1.5795, "step": 4766 }, { "epoch": 1.0486088199714065, "grad_norm": 0.3235938847064972, "learning_rate": 8e-05, "loss": 1.5497, "step": 4767 }, { "epoch": 1.0488287693830418, "grad_norm": 0.32496464252471924, "learning_rate": 8e-05, "loss": 1.4826, "step": 4768 }, { "epoch": 1.0490487187946773, "grad_norm": 0.3207201361656189, "learning_rate": 8e-05, "loss": 1.4221, "step": 4769 }, { "epoch": 1.0492686682063126, "grad_norm": 0.3103640377521515, "learning_rate": 8e-05, "loss": 1.5009, "step": 4770 }, { "epoch": 1.0494886176179479, "grad_norm": 0.3269423544406891, "learning_rate": 8e-05, "loss": 1.5583, "step": 4771 }, { "epoch": 1.0497085670295832, "grad_norm": 0.31951606273651123, "learning_rate": 8e-05, "loss": 1.5848, "step": 4772 }, { "epoch": 1.0499285164412184, "grad_norm": 0.3351285457611084, "learning_rate": 8e-05, "loss": 1.5673, "step": 4773 }, { "epoch": 1.050148465852854, "grad_norm": 0.3169887065887451, "learning_rate": 8e-05, "loss": 1.5521, "step": 4774 }, { "epoch": 1.0503684152644892, "grad_norm": 0.34145545959472656, "learning_rate": 8e-05, "loss": 1.6899, "step": 4775 }, { "epoch": 1.0505883646761245, "grad_norm": 0.3373747169971466, "learning_rate": 8e-05, "loss": 1.599, "step": 4776 }, { "epoch": 1.0508083140877598, "grad_norm": 0.3361704349517822, "learning_rate": 8e-05, "loss": 1.7051, "step": 4777 }, { "epoch": 1.051028263499395, "grad_norm": 0.3099362254142761, "learning_rate": 8e-05, "loss": 1.5401, "step": 4778 }, { "epoch": 1.0512482129110305, "grad_norm": 0.3178727328777313, "learning_rate": 8e-05, "loss": 1.6139, "step": 4779 }, { "epoch": 1.0514681623226658, "grad_norm": 0.32660022377967834, "learning_rate": 8e-05, "loss": 1.59, "step": 4780 }, { "epoch": 1.051688111734301, "grad_norm": 0.30725130438804626, "learning_rate": 8e-05, "loss": 1.5828, "step": 4781 }, { "epoch": 1.0519080611459364, "grad_norm": 0.34365415573120117, "learning_rate": 8e-05, "loss": 1.6672, "step": 4782 }, { "epoch": 1.0521280105575717, "grad_norm": 0.3104267418384552, "learning_rate": 8e-05, "loss": 1.2426, "step": 4783 }, { "epoch": 1.0523479599692072, "grad_norm": 0.32874515652656555, "learning_rate": 8e-05, "loss": 1.6987, "step": 4784 }, { "epoch": 1.0525679093808424, "grad_norm": 0.3329507112503052, "learning_rate": 8e-05, "loss": 1.6884, "step": 4785 }, { "epoch": 1.0527878587924777, "grad_norm": 0.31262141466140747, "learning_rate": 8e-05, "loss": 1.5659, "step": 4786 }, { "epoch": 1.053007808204113, "grad_norm": 0.3158217668533325, "learning_rate": 8e-05, "loss": 1.5119, "step": 4787 }, { "epoch": 1.0532277576157483, "grad_norm": 0.3056739866733551, "learning_rate": 8e-05, "loss": 1.5402, "step": 4788 }, { "epoch": 1.0534477070273838, "grad_norm": 0.3142102062702179, "learning_rate": 8e-05, "loss": 1.5725, "step": 4789 }, { "epoch": 1.053667656439019, "grad_norm": 0.31378012895584106, "learning_rate": 8e-05, "loss": 1.4328, "step": 4790 }, { "epoch": 1.0538876058506543, "grad_norm": 0.301178902387619, "learning_rate": 8e-05, "loss": 1.5987, "step": 4791 }, { "epoch": 1.0541075552622896, "grad_norm": 0.31317493319511414, "learning_rate": 8e-05, "loss": 1.4711, "step": 4792 }, { "epoch": 1.054327504673925, "grad_norm": 0.3427419066429138, "learning_rate": 8e-05, "loss": 1.5135, "step": 4793 }, { "epoch": 1.0545474540855604, "grad_norm": 0.3269393742084503, "learning_rate": 8e-05, "loss": 1.5583, "step": 4794 }, { "epoch": 1.0547674034971957, "grad_norm": 0.3082657754421234, "learning_rate": 8e-05, "loss": 1.6388, "step": 4795 }, { "epoch": 1.054987352908831, "grad_norm": 0.3138137757778168, "learning_rate": 8e-05, "loss": 1.4592, "step": 4796 }, { "epoch": 1.0552073023204662, "grad_norm": 0.31918948888778687, "learning_rate": 8e-05, "loss": 1.676, "step": 4797 }, { "epoch": 1.0554272517321017, "grad_norm": 0.3299674689769745, "learning_rate": 8e-05, "loss": 1.7152, "step": 4798 }, { "epoch": 1.055647201143737, "grad_norm": 0.2974132001399994, "learning_rate": 8e-05, "loss": 1.5336, "step": 4799 }, { "epoch": 1.0558671505553723, "grad_norm": 0.3264104723930359, "learning_rate": 8e-05, "loss": 1.6765, "step": 4800 }, { "epoch": 1.0560870999670076, "grad_norm": 0.30670997500419617, "learning_rate": 8e-05, "loss": 1.6277, "step": 4801 }, { "epoch": 1.0563070493786428, "grad_norm": 0.3084074556827545, "learning_rate": 8e-05, "loss": 1.6029, "step": 4802 }, { "epoch": 1.0565269987902783, "grad_norm": 0.3156238794326782, "learning_rate": 8e-05, "loss": 1.5551, "step": 4803 }, { "epoch": 1.0567469482019136, "grad_norm": 0.2925167381763458, "learning_rate": 8e-05, "loss": 1.3627, "step": 4804 }, { "epoch": 1.056966897613549, "grad_norm": 0.30956152081489563, "learning_rate": 8e-05, "loss": 1.5579, "step": 4805 }, { "epoch": 1.0571868470251842, "grad_norm": 0.3217785358428955, "learning_rate": 8e-05, "loss": 1.6563, "step": 4806 }, { "epoch": 1.0574067964368195, "grad_norm": 0.3405359089374542, "learning_rate": 8e-05, "loss": 1.6587, "step": 4807 }, { "epoch": 1.057626745848455, "grad_norm": 0.3022216856479645, "learning_rate": 8e-05, "loss": 1.5408, "step": 4808 }, { "epoch": 1.0578466952600902, "grad_norm": 0.31406083703041077, "learning_rate": 8e-05, "loss": 1.5555, "step": 4809 }, { "epoch": 1.0580666446717255, "grad_norm": 0.3455413281917572, "learning_rate": 8e-05, "loss": 1.6544, "step": 4810 }, { "epoch": 1.0582865940833608, "grad_norm": 0.31688791513442993, "learning_rate": 8e-05, "loss": 1.5264, "step": 4811 }, { "epoch": 1.058506543494996, "grad_norm": 0.34746018052101135, "learning_rate": 8e-05, "loss": 1.6542, "step": 4812 }, { "epoch": 1.0587264929066316, "grad_norm": 0.3529677391052246, "learning_rate": 8e-05, "loss": 1.721, "step": 4813 }, { "epoch": 1.0589464423182668, "grad_norm": 0.3068278431892395, "learning_rate": 8e-05, "loss": 1.5508, "step": 4814 }, { "epoch": 1.0591663917299021, "grad_norm": 0.3102533221244812, "learning_rate": 8e-05, "loss": 1.4624, "step": 4815 }, { "epoch": 1.0593863411415374, "grad_norm": 0.3388904631137848, "learning_rate": 8e-05, "loss": 1.6366, "step": 4816 }, { "epoch": 1.0596062905531727, "grad_norm": 0.32976335287094116, "learning_rate": 8e-05, "loss": 1.6446, "step": 4817 }, { "epoch": 1.0598262399648082, "grad_norm": 0.32190465927124023, "learning_rate": 8e-05, "loss": 1.5006, "step": 4818 }, { "epoch": 1.0600461893764435, "grad_norm": 0.3166998624801636, "learning_rate": 8e-05, "loss": 1.6239, "step": 4819 }, { "epoch": 1.0602661387880787, "grad_norm": 0.311631977558136, "learning_rate": 8e-05, "loss": 1.5813, "step": 4820 }, { "epoch": 1.060486088199714, "grad_norm": 0.31014204025268555, "learning_rate": 8e-05, "loss": 1.648, "step": 4821 }, { "epoch": 1.0607060376113493, "grad_norm": 0.31572502851486206, "learning_rate": 8e-05, "loss": 1.6727, "step": 4822 }, { "epoch": 1.0609259870229848, "grad_norm": 0.30379027128219604, "learning_rate": 8e-05, "loss": 1.4965, "step": 4823 }, { "epoch": 1.06114593643462, "grad_norm": 0.32408884167671204, "learning_rate": 8e-05, "loss": 1.6728, "step": 4824 }, { "epoch": 1.0613658858462554, "grad_norm": 0.3221605122089386, "learning_rate": 8e-05, "loss": 1.4892, "step": 4825 }, { "epoch": 1.0615858352578906, "grad_norm": 0.326666921377182, "learning_rate": 8e-05, "loss": 1.65, "step": 4826 }, { "epoch": 1.061805784669526, "grad_norm": 0.3198178708553314, "learning_rate": 8e-05, "loss": 1.6942, "step": 4827 }, { "epoch": 1.0620257340811614, "grad_norm": 0.35545653104782104, "learning_rate": 8e-05, "loss": 1.8287, "step": 4828 }, { "epoch": 1.0622456834927967, "grad_norm": 0.31938254833221436, "learning_rate": 8e-05, "loss": 1.5497, "step": 4829 }, { "epoch": 1.062465632904432, "grad_norm": 0.30391940474510193, "learning_rate": 8e-05, "loss": 1.6244, "step": 4830 }, { "epoch": 1.0626855823160672, "grad_norm": 0.30713897943496704, "learning_rate": 8e-05, "loss": 1.4544, "step": 4831 }, { "epoch": 1.0629055317277025, "grad_norm": 0.3189101815223694, "learning_rate": 8e-05, "loss": 1.5632, "step": 4832 }, { "epoch": 1.063125481139338, "grad_norm": 0.352477103471756, "learning_rate": 8e-05, "loss": 1.5738, "step": 4833 }, { "epoch": 1.0633454305509733, "grad_norm": 0.30898699164390564, "learning_rate": 8e-05, "loss": 1.5122, "step": 4834 }, { "epoch": 1.0635653799626086, "grad_norm": 0.36749008297920227, "learning_rate": 8e-05, "loss": 1.655, "step": 4835 }, { "epoch": 1.0637853293742439, "grad_norm": 0.33112525939941406, "learning_rate": 8e-05, "loss": 1.7262, "step": 4836 }, { "epoch": 1.0640052787858791, "grad_norm": 0.33594053983688354, "learning_rate": 8e-05, "loss": 1.5578, "step": 4837 }, { "epoch": 1.0642252281975146, "grad_norm": 0.3227829933166504, "learning_rate": 8e-05, "loss": 1.5597, "step": 4838 }, { "epoch": 1.06444517760915, "grad_norm": 0.32178738713264465, "learning_rate": 8e-05, "loss": 1.571, "step": 4839 }, { "epoch": 1.0646651270207852, "grad_norm": 0.35744771361351013, "learning_rate": 8e-05, "loss": 1.8214, "step": 4840 }, { "epoch": 1.0648850764324205, "grad_norm": 0.33018893003463745, "learning_rate": 8e-05, "loss": 1.599, "step": 4841 }, { "epoch": 1.0651050258440558, "grad_norm": 0.33655205368995667, "learning_rate": 8e-05, "loss": 1.58, "step": 4842 }, { "epoch": 1.0653249752556913, "grad_norm": 0.2989410161972046, "learning_rate": 8e-05, "loss": 1.5106, "step": 4843 }, { "epoch": 1.0655449246673265, "grad_norm": 0.3095400929450989, "learning_rate": 8e-05, "loss": 1.5495, "step": 4844 }, { "epoch": 1.0657648740789618, "grad_norm": 0.3345370590686798, "learning_rate": 8e-05, "loss": 1.5898, "step": 4845 }, { "epoch": 1.065984823490597, "grad_norm": 0.31667864322662354, "learning_rate": 8e-05, "loss": 1.6177, "step": 4846 }, { "epoch": 1.0662047729022326, "grad_norm": 0.3012166917324066, "learning_rate": 8e-05, "loss": 1.5757, "step": 4847 }, { "epoch": 1.0664247223138679, "grad_norm": 0.30444058775901794, "learning_rate": 8e-05, "loss": 1.4331, "step": 4848 }, { "epoch": 1.0666446717255031, "grad_norm": 0.31133100390434265, "learning_rate": 8e-05, "loss": 1.5776, "step": 4849 }, { "epoch": 1.0668646211371384, "grad_norm": 0.37968552112579346, "learning_rate": 8e-05, "loss": 1.6019, "step": 4850 }, { "epoch": 1.0670845705487737, "grad_norm": 0.3456275463104248, "learning_rate": 8e-05, "loss": 1.5473, "step": 4851 }, { "epoch": 1.0673045199604092, "grad_norm": 0.3273290991783142, "learning_rate": 8e-05, "loss": 1.5897, "step": 4852 }, { "epoch": 1.0675244693720445, "grad_norm": 0.3222695291042328, "learning_rate": 8e-05, "loss": 1.6031, "step": 4853 }, { "epoch": 1.0677444187836798, "grad_norm": 0.3495657742023468, "learning_rate": 8e-05, "loss": 1.4644, "step": 4854 }, { "epoch": 1.067964368195315, "grad_norm": 0.4430721402168274, "learning_rate": 8e-05, "loss": 1.7669, "step": 4855 }, { "epoch": 1.0681843176069503, "grad_norm": 0.3258362114429474, "learning_rate": 8e-05, "loss": 1.6228, "step": 4856 }, { "epoch": 1.0684042670185858, "grad_norm": 0.34792041778564453, "learning_rate": 8e-05, "loss": 1.6783, "step": 4857 }, { "epoch": 1.068624216430221, "grad_norm": 0.3170369863510132, "learning_rate": 8e-05, "loss": 1.5951, "step": 4858 }, { "epoch": 1.0688441658418564, "grad_norm": 0.3128320872783661, "learning_rate": 8e-05, "loss": 1.5104, "step": 4859 }, { "epoch": 1.0690641152534917, "grad_norm": 0.3237660825252533, "learning_rate": 8e-05, "loss": 1.5945, "step": 4860 }, { "epoch": 1.069284064665127, "grad_norm": 0.340315043926239, "learning_rate": 8e-05, "loss": 1.7501, "step": 4861 }, { "epoch": 1.0695040140767624, "grad_norm": 0.3219820261001587, "learning_rate": 8e-05, "loss": 1.4895, "step": 4862 }, { "epoch": 1.0697239634883977, "grad_norm": 0.3298594057559967, "learning_rate": 8e-05, "loss": 1.5992, "step": 4863 }, { "epoch": 1.069943912900033, "grad_norm": 0.3744049072265625, "learning_rate": 8e-05, "loss": 1.7039, "step": 4864 }, { "epoch": 1.0701638623116683, "grad_norm": 0.33990490436553955, "learning_rate": 8e-05, "loss": 1.5291, "step": 4865 }, { "epoch": 1.0703838117233035, "grad_norm": 0.3289349377155304, "learning_rate": 8e-05, "loss": 1.5791, "step": 4866 }, { "epoch": 1.070603761134939, "grad_norm": 0.39308100938796997, "learning_rate": 8e-05, "loss": 1.626, "step": 4867 }, { "epoch": 1.0708237105465743, "grad_norm": 0.3331829607486725, "learning_rate": 8e-05, "loss": 1.575, "step": 4868 }, { "epoch": 1.0710436599582096, "grad_norm": 0.3172522485256195, "learning_rate": 8e-05, "loss": 1.6756, "step": 4869 }, { "epoch": 1.0712636093698449, "grad_norm": 0.3198091685771942, "learning_rate": 8e-05, "loss": 1.563, "step": 4870 }, { "epoch": 1.0714835587814802, "grad_norm": 0.333723247051239, "learning_rate": 8e-05, "loss": 1.6656, "step": 4871 }, { "epoch": 1.0717035081931157, "grad_norm": 0.305131196975708, "learning_rate": 8e-05, "loss": 1.5726, "step": 4872 }, { "epoch": 1.071923457604751, "grad_norm": 0.30560487508773804, "learning_rate": 8e-05, "loss": 1.5096, "step": 4873 }, { "epoch": 1.0721434070163862, "grad_norm": 0.3272777795791626, "learning_rate": 8e-05, "loss": 1.6676, "step": 4874 }, { "epoch": 1.0723633564280215, "grad_norm": 0.30223652720451355, "learning_rate": 8e-05, "loss": 1.4938, "step": 4875 }, { "epoch": 1.0725833058396568, "grad_norm": 0.32536906003952026, "learning_rate": 8e-05, "loss": 1.3859, "step": 4876 }, { "epoch": 1.0728032552512923, "grad_norm": 0.328176885843277, "learning_rate": 8e-05, "loss": 1.6068, "step": 4877 }, { "epoch": 1.0730232046629276, "grad_norm": 0.31937548518180847, "learning_rate": 8e-05, "loss": 1.4856, "step": 4878 }, { "epoch": 1.0732431540745628, "grad_norm": 0.32114046812057495, "learning_rate": 8e-05, "loss": 1.6312, "step": 4879 }, { "epoch": 1.0734631034861981, "grad_norm": 0.309573769569397, "learning_rate": 8e-05, "loss": 1.5023, "step": 4880 }, { "epoch": 1.0736830528978336, "grad_norm": 0.32102400064468384, "learning_rate": 8e-05, "loss": 1.6974, "step": 4881 }, { "epoch": 1.073903002309469, "grad_norm": 0.34739330410957336, "learning_rate": 8e-05, "loss": 1.6681, "step": 4882 }, { "epoch": 1.0741229517211042, "grad_norm": 0.34210023283958435, "learning_rate": 8e-05, "loss": 1.7147, "step": 4883 }, { "epoch": 1.0743429011327394, "grad_norm": 0.319172739982605, "learning_rate": 8e-05, "loss": 1.5843, "step": 4884 }, { "epoch": 1.0745628505443747, "grad_norm": 0.329084187746048, "learning_rate": 8e-05, "loss": 1.6324, "step": 4885 }, { "epoch": 1.0747827999560102, "grad_norm": 0.32300254702568054, "learning_rate": 8e-05, "loss": 1.5752, "step": 4886 }, { "epoch": 1.0750027493676455, "grad_norm": 0.35015806555747986, "learning_rate": 8e-05, "loss": 1.811, "step": 4887 }, { "epoch": 1.0752226987792808, "grad_norm": 0.30605348944664, "learning_rate": 8e-05, "loss": 1.5134, "step": 4888 }, { "epoch": 1.075442648190916, "grad_norm": 0.31354203820228577, "learning_rate": 8e-05, "loss": 1.7043, "step": 4889 }, { "epoch": 1.0756625976025513, "grad_norm": 0.33387675881385803, "learning_rate": 8e-05, "loss": 1.582, "step": 4890 }, { "epoch": 1.0758825470141868, "grad_norm": 0.32679006457328796, "learning_rate": 8e-05, "loss": 1.5891, "step": 4891 }, { "epoch": 1.0761024964258221, "grad_norm": 0.3183629512786865, "learning_rate": 8e-05, "loss": 1.4813, "step": 4892 }, { "epoch": 1.0763224458374574, "grad_norm": 0.3158898949623108, "learning_rate": 8e-05, "loss": 1.5122, "step": 4893 }, { "epoch": 1.0765423952490927, "grad_norm": 0.305123507976532, "learning_rate": 8e-05, "loss": 1.5704, "step": 4894 }, { "epoch": 1.076762344660728, "grad_norm": 0.33120593428611755, "learning_rate": 8e-05, "loss": 1.5365, "step": 4895 }, { "epoch": 1.0769822940723635, "grad_norm": 0.33490070700645447, "learning_rate": 8e-05, "loss": 1.5856, "step": 4896 }, { "epoch": 1.0772022434839987, "grad_norm": 0.34703588485717773, "learning_rate": 8e-05, "loss": 1.694, "step": 4897 }, { "epoch": 1.077422192895634, "grad_norm": 0.31777331233024597, "learning_rate": 8e-05, "loss": 1.6613, "step": 4898 }, { "epoch": 1.0776421423072693, "grad_norm": 0.3249010443687439, "learning_rate": 8e-05, "loss": 1.3772, "step": 4899 }, { "epoch": 1.0778620917189046, "grad_norm": 0.32455241680145264, "learning_rate": 8e-05, "loss": 1.6902, "step": 4900 }, { "epoch": 1.07808204113054, "grad_norm": 0.33786752820014954, "learning_rate": 8e-05, "loss": 1.7336, "step": 4901 }, { "epoch": 1.0783019905421753, "grad_norm": 0.3316012918949127, "learning_rate": 8e-05, "loss": 1.7529, "step": 4902 }, { "epoch": 1.0785219399538106, "grad_norm": 0.31583714485168457, "learning_rate": 8e-05, "loss": 1.5726, "step": 4903 }, { "epoch": 1.078741889365446, "grad_norm": 0.3362698256969452, "learning_rate": 8e-05, "loss": 1.5493, "step": 4904 }, { "epoch": 1.0789618387770812, "grad_norm": 0.3300097584724426, "learning_rate": 8e-05, "loss": 1.621, "step": 4905 }, { "epoch": 1.0791817881887167, "grad_norm": 0.30957186222076416, "learning_rate": 8e-05, "loss": 1.6994, "step": 4906 }, { "epoch": 1.079401737600352, "grad_norm": 0.3522869348526001, "learning_rate": 8e-05, "loss": 1.6715, "step": 4907 }, { "epoch": 1.0796216870119872, "grad_norm": 0.3219485580921173, "learning_rate": 8e-05, "loss": 1.5669, "step": 4908 }, { "epoch": 1.0798416364236225, "grad_norm": 0.3225310444831848, "learning_rate": 8e-05, "loss": 1.6448, "step": 4909 }, { "epoch": 1.0800615858352578, "grad_norm": 0.31042027473449707, "learning_rate": 8e-05, "loss": 1.6721, "step": 4910 }, { "epoch": 1.0802815352468933, "grad_norm": 0.32192590832710266, "learning_rate": 8e-05, "loss": 1.5515, "step": 4911 }, { "epoch": 1.0805014846585286, "grad_norm": 0.31783169507980347, "learning_rate": 8e-05, "loss": 1.5808, "step": 4912 }, { "epoch": 1.0807214340701639, "grad_norm": 0.3307020366191864, "learning_rate": 8e-05, "loss": 1.6376, "step": 4913 }, { "epoch": 1.0809413834817991, "grad_norm": 0.312326043844223, "learning_rate": 8e-05, "loss": 1.5677, "step": 4914 }, { "epoch": 1.0811613328934344, "grad_norm": 0.3150627911090851, "learning_rate": 8e-05, "loss": 1.432, "step": 4915 }, { "epoch": 1.08138128230507, "grad_norm": 0.3163559138774872, "learning_rate": 8e-05, "loss": 1.4882, "step": 4916 }, { "epoch": 1.0816012317167052, "grad_norm": 0.30999115109443665, "learning_rate": 8e-05, "loss": 1.406, "step": 4917 }, { "epoch": 1.0818211811283405, "grad_norm": 0.31164002418518066, "learning_rate": 8e-05, "loss": 1.4772, "step": 4918 }, { "epoch": 1.0820411305399757, "grad_norm": 0.3401453197002411, "learning_rate": 8e-05, "loss": 1.3823, "step": 4919 }, { "epoch": 1.082261079951611, "grad_norm": 0.3215329349040985, "learning_rate": 8e-05, "loss": 1.5591, "step": 4920 }, { "epoch": 1.0824810293632465, "grad_norm": 0.30950450897216797, "learning_rate": 8e-05, "loss": 1.5672, "step": 4921 }, { "epoch": 1.0827009787748818, "grad_norm": 0.3296648859977722, "learning_rate": 8e-05, "loss": 1.5825, "step": 4922 }, { "epoch": 1.082920928186517, "grad_norm": 0.32215073704719543, "learning_rate": 8e-05, "loss": 1.5357, "step": 4923 }, { "epoch": 1.0831408775981524, "grad_norm": 0.3348964750766754, "learning_rate": 8e-05, "loss": 1.6334, "step": 4924 }, { "epoch": 1.0833608270097876, "grad_norm": 0.3218708336353302, "learning_rate": 8e-05, "loss": 1.6409, "step": 4925 }, { "epoch": 1.0835807764214231, "grad_norm": 0.3146817088127136, "learning_rate": 8e-05, "loss": 1.6956, "step": 4926 }, { "epoch": 1.0838007258330584, "grad_norm": 0.34461474418640137, "learning_rate": 8e-05, "loss": 1.5127, "step": 4927 }, { "epoch": 1.0840206752446937, "grad_norm": 0.31843042373657227, "learning_rate": 8e-05, "loss": 1.5658, "step": 4928 }, { "epoch": 1.084240624656329, "grad_norm": 0.3925442397594452, "learning_rate": 8e-05, "loss": 1.6608, "step": 4929 }, { "epoch": 1.0844605740679643, "grad_norm": 0.3256430923938751, "learning_rate": 8e-05, "loss": 1.6466, "step": 4930 }, { "epoch": 1.0846805234795998, "grad_norm": 0.3160317838191986, "learning_rate": 8e-05, "loss": 1.6675, "step": 4931 }, { "epoch": 1.084900472891235, "grad_norm": 0.3316245675086975, "learning_rate": 8e-05, "loss": 1.6027, "step": 4932 }, { "epoch": 1.0851204223028703, "grad_norm": 0.34334978461265564, "learning_rate": 8e-05, "loss": 1.7329, "step": 4933 }, { "epoch": 1.0853403717145056, "grad_norm": 0.3201568126678467, "learning_rate": 8e-05, "loss": 1.6491, "step": 4934 }, { "epoch": 1.0855603211261409, "grad_norm": 0.31593194603919983, "learning_rate": 8e-05, "loss": 1.5853, "step": 4935 }, { "epoch": 1.0857802705377764, "grad_norm": 0.3154931962490082, "learning_rate": 8e-05, "loss": 1.6237, "step": 4936 }, { "epoch": 1.0860002199494116, "grad_norm": 0.31061291694641113, "learning_rate": 8e-05, "loss": 1.6123, "step": 4937 }, { "epoch": 1.086220169361047, "grad_norm": 0.3058033585548401, "learning_rate": 8e-05, "loss": 1.6003, "step": 4938 }, { "epoch": 1.0864401187726822, "grad_norm": 0.3258708119392395, "learning_rate": 8e-05, "loss": 1.6418, "step": 4939 }, { "epoch": 1.0866600681843177, "grad_norm": 0.3100445866584778, "learning_rate": 8e-05, "loss": 1.6339, "step": 4940 }, { "epoch": 1.086880017595953, "grad_norm": 0.329999178647995, "learning_rate": 8e-05, "loss": 1.5911, "step": 4941 }, { "epoch": 1.0870999670075883, "grad_norm": 0.2916867733001709, "learning_rate": 8e-05, "loss": 1.4917, "step": 4942 }, { "epoch": 1.0873199164192235, "grad_norm": 0.31672465801239014, "learning_rate": 8e-05, "loss": 1.6361, "step": 4943 }, { "epoch": 1.0875398658308588, "grad_norm": 0.31611114740371704, "learning_rate": 8e-05, "loss": 1.4749, "step": 4944 }, { "epoch": 1.0877598152424943, "grad_norm": 0.32682448625564575, "learning_rate": 8e-05, "loss": 1.736, "step": 4945 }, { "epoch": 1.0879797646541296, "grad_norm": 0.31582289934158325, "learning_rate": 8e-05, "loss": 1.5794, "step": 4946 }, { "epoch": 1.0881997140657649, "grad_norm": 0.3109477758407593, "learning_rate": 8e-05, "loss": 1.5223, "step": 4947 }, { "epoch": 1.0884196634774002, "grad_norm": 0.32298681139945984, "learning_rate": 8e-05, "loss": 1.5584, "step": 4948 }, { "epoch": 1.0886396128890354, "grad_norm": 0.3679943084716797, "learning_rate": 8e-05, "loss": 1.5822, "step": 4949 }, { "epoch": 1.088859562300671, "grad_norm": 0.3428889811038971, "learning_rate": 8e-05, "loss": 1.6102, "step": 4950 }, { "epoch": 1.0890795117123062, "grad_norm": 0.32236382365226746, "learning_rate": 8e-05, "loss": 1.4775, "step": 4951 }, { "epoch": 1.0892994611239415, "grad_norm": 0.3363906443119049, "learning_rate": 8e-05, "loss": 1.7252, "step": 4952 }, { "epoch": 1.0895194105355768, "grad_norm": 0.33397215604782104, "learning_rate": 8e-05, "loss": 1.5865, "step": 4953 }, { "epoch": 1.089739359947212, "grad_norm": 0.34392181038856506, "learning_rate": 8e-05, "loss": 1.62, "step": 4954 }, { "epoch": 1.0899593093588476, "grad_norm": 0.342862069606781, "learning_rate": 8e-05, "loss": 1.5946, "step": 4955 }, { "epoch": 1.0901792587704828, "grad_norm": 0.30339252948760986, "learning_rate": 8e-05, "loss": 1.4822, "step": 4956 }, { "epoch": 1.090399208182118, "grad_norm": 0.3095758557319641, "learning_rate": 8e-05, "loss": 1.5285, "step": 4957 }, { "epoch": 1.0906191575937534, "grad_norm": 0.32964375615119934, "learning_rate": 8e-05, "loss": 1.5745, "step": 4958 }, { "epoch": 1.0908391070053887, "grad_norm": 0.3260382413864136, "learning_rate": 8e-05, "loss": 1.5889, "step": 4959 }, { "epoch": 1.0910590564170242, "grad_norm": 0.30548831820487976, "learning_rate": 8e-05, "loss": 1.5435, "step": 4960 }, { "epoch": 1.0912790058286594, "grad_norm": 0.30104225873947144, "learning_rate": 8e-05, "loss": 1.4736, "step": 4961 }, { "epoch": 1.0914989552402947, "grad_norm": 0.3617940843105316, "learning_rate": 8e-05, "loss": 1.7213, "step": 4962 }, { "epoch": 1.09171890465193, "grad_norm": 0.2966761291027069, "learning_rate": 8e-05, "loss": 1.477, "step": 4963 }, { "epoch": 1.0919388540635653, "grad_norm": 0.331989586353302, "learning_rate": 8e-05, "loss": 1.5351, "step": 4964 }, { "epoch": 1.0921588034752008, "grad_norm": 0.30379363894462585, "learning_rate": 8e-05, "loss": 1.3846, "step": 4965 }, { "epoch": 1.092378752886836, "grad_norm": 0.3260042667388916, "learning_rate": 8e-05, "loss": 1.6168, "step": 4966 }, { "epoch": 1.0925987022984713, "grad_norm": 0.3615175783634186, "learning_rate": 8e-05, "loss": 1.5874, "step": 4967 }, { "epoch": 1.0928186517101066, "grad_norm": 0.35878193378448486, "learning_rate": 8e-05, "loss": 1.6804, "step": 4968 }, { "epoch": 1.0930386011217421, "grad_norm": 0.32007548213005066, "learning_rate": 8e-05, "loss": 1.5191, "step": 4969 }, { "epoch": 1.0932585505333774, "grad_norm": 0.32241368293762207, "learning_rate": 8e-05, "loss": 1.4805, "step": 4970 }, { "epoch": 1.0934784999450127, "grad_norm": 0.30216753482818604, "learning_rate": 8e-05, "loss": 1.4739, "step": 4971 }, { "epoch": 1.093698449356648, "grad_norm": 0.33534616231918335, "learning_rate": 8e-05, "loss": 1.5657, "step": 4972 }, { "epoch": 1.0939183987682832, "grad_norm": 0.31524744629859924, "learning_rate": 8e-05, "loss": 1.4906, "step": 4973 }, { "epoch": 1.0941383481799187, "grad_norm": 0.29395657777786255, "learning_rate": 8e-05, "loss": 1.4084, "step": 4974 }, { "epoch": 1.094358297591554, "grad_norm": 0.3654805123806, "learning_rate": 8e-05, "loss": 1.6225, "step": 4975 }, { "epoch": 1.0945782470031893, "grad_norm": 0.32975268363952637, "learning_rate": 8e-05, "loss": 1.6546, "step": 4976 }, { "epoch": 1.0947981964148246, "grad_norm": 0.3227517008781433, "learning_rate": 8e-05, "loss": 1.6451, "step": 4977 }, { "epoch": 1.0950181458264598, "grad_norm": 0.3046882450580597, "learning_rate": 8e-05, "loss": 1.3408, "step": 4978 }, { "epoch": 1.0952380952380953, "grad_norm": 0.35542967915534973, "learning_rate": 8e-05, "loss": 1.5862, "step": 4979 }, { "epoch": 1.0954580446497306, "grad_norm": 0.3299451470375061, "learning_rate": 8e-05, "loss": 1.6062, "step": 4980 }, { "epoch": 1.095677994061366, "grad_norm": 0.34388840198516846, "learning_rate": 8e-05, "loss": 1.5033, "step": 4981 }, { "epoch": 1.0958979434730012, "grad_norm": 0.33758535981178284, "learning_rate": 8e-05, "loss": 1.6178, "step": 4982 }, { "epoch": 1.0961178928846365, "grad_norm": 0.3469134271144867, "learning_rate": 8e-05, "loss": 1.7144, "step": 4983 }, { "epoch": 1.096337842296272, "grad_norm": 0.3169137239456177, "learning_rate": 8e-05, "loss": 1.6077, "step": 4984 }, { "epoch": 1.0965577917079072, "grad_norm": 0.31431668996810913, "learning_rate": 8e-05, "loss": 1.6228, "step": 4985 }, { "epoch": 1.0967777411195425, "grad_norm": 0.32845214009284973, "learning_rate": 8e-05, "loss": 1.7445, "step": 4986 }, { "epoch": 1.0969976905311778, "grad_norm": 0.3124347925186157, "learning_rate": 8e-05, "loss": 1.5874, "step": 4987 }, { "epoch": 1.097217639942813, "grad_norm": 0.32342013716697693, "learning_rate": 8e-05, "loss": 1.6841, "step": 4988 }, { "epoch": 1.0974375893544486, "grad_norm": 0.3514409065246582, "learning_rate": 8e-05, "loss": 1.4936, "step": 4989 }, { "epoch": 1.0976575387660839, "grad_norm": 0.31863728165626526, "learning_rate": 8e-05, "loss": 1.5986, "step": 4990 }, { "epoch": 1.0978774881777191, "grad_norm": 0.3084011375904083, "learning_rate": 8e-05, "loss": 1.5223, "step": 4991 }, { "epoch": 1.0980974375893544, "grad_norm": 0.31163445115089417, "learning_rate": 8e-05, "loss": 1.3865, "step": 4992 }, { "epoch": 1.0983173870009897, "grad_norm": 0.35192030668258667, "learning_rate": 8e-05, "loss": 1.7001, "step": 4993 }, { "epoch": 1.0985373364126252, "grad_norm": 0.3234604597091675, "learning_rate": 8e-05, "loss": 1.6378, "step": 4994 }, { "epoch": 1.0987572858242605, "grad_norm": 0.3249831795692444, "learning_rate": 8e-05, "loss": 1.4746, "step": 4995 }, { "epoch": 1.0989772352358957, "grad_norm": 0.33398959040641785, "learning_rate": 8e-05, "loss": 1.5323, "step": 4996 }, { "epoch": 1.099197184647531, "grad_norm": 0.30378586053848267, "learning_rate": 8e-05, "loss": 1.4505, "step": 4997 }, { "epoch": 1.0994171340591663, "grad_norm": 0.31826141476631165, "learning_rate": 8e-05, "loss": 1.6144, "step": 4998 }, { "epoch": 1.0996370834708018, "grad_norm": 0.32833409309387207, "learning_rate": 8e-05, "loss": 1.6505, "step": 4999 }, { "epoch": 1.099857032882437, "grad_norm": 0.3189920485019684, "learning_rate": 8e-05, "loss": 1.6229, "step": 5000 }, { "epoch": 1.1000769822940724, "grad_norm": 0.3404131233692169, "learning_rate": 8e-05, "loss": 1.7067, "step": 5001 }, { "epoch": 1.1002969317057076, "grad_norm": 0.3161638081073761, "learning_rate": 8e-05, "loss": 1.4904, "step": 5002 }, { "epoch": 1.100516881117343, "grad_norm": 0.3301504850387573, "learning_rate": 8e-05, "loss": 1.5062, "step": 5003 }, { "epoch": 1.1007368305289784, "grad_norm": 0.3327349126338959, "learning_rate": 8e-05, "loss": 1.7645, "step": 5004 }, { "epoch": 1.1009567799406137, "grad_norm": 0.3130214512348175, "learning_rate": 8e-05, "loss": 1.5332, "step": 5005 }, { "epoch": 1.101176729352249, "grad_norm": 0.3195625841617584, "learning_rate": 8e-05, "loss": 1.5502, "step": 5006 }, { "epoch": 1.1013966787638843, "grad_norm": 0.3334658145904541, "learning_rate": 8e-05, "loss": 1.555, "step": 5007 }, { "epoch": 1.1016166281755195, "grad_norm": 0.32564017176628113, "learning_rate": 8e-05, "loss": 1.5327, "step": 5008 }, { "epoch": 1.101836577587155, "grad_norm": 0.33153125643730164, "learning_rate": 8e-05, "loss": 1.5682, "step": 5009 }, { "epoch": 1.1020565269987903, "grad_norm": 0.31199514865875244, "learning_rate": 8e-05, "loss": 1.3827, "step": 5010 }, { "epoch": 1.1022764764104256, "grad_norm": 0.3132420778274536, "learning_rate": 8e-05, "loss": 1.5693, "step": 5011 }, { "epoch": 1.1024964258220609, "grad_norm": 0.33908095955848694, "learning_rate": 8e-05, "loss": 1.5801, "step": 5012 }, { "epoch": 1.1027163752336961, "grad_norm": 0.32728588581085205, "learning_rate": 8e-05, "loss": 1.5984, "step": 5013 }, { "epoch": 1.1029363246453316, "grad_norm": 0.3394905924797058, "learning_rate": 8e-05, "loss": 1.5596, "step": 5014 }, { "epoch": 1.103156274056967, "grad_norm": 0.33040326833724976, "learning_rate": 8e-05, "loss": 1.5424, "step": 5015 }, { "epoch": 1.1033762234686022, "grad_norm": 0.3371153175830841, "learning_rate": 8e-05, "loss": 1.6111, "step": 5016 }, { "epoch": 1.1035961728802375, "grad_norm": 0.3386276662349701, "learning_rate": 8e-05, "loss": 1.6907, "step": 5017 }, { "epoch": 1.1038161222918728, "grad_norm": 0.3508915603160858, "learning_rate": 8e-05, "loss": 1.5415, "step": 5018 }, { "epoch": 1.1040360717035083, "grad_norm": 0.33191850781440735, "learning_rate": 8e-05, "loss": 1.4879, "step": 5019 }, { "epoch": 1.1042560211151435, "grad_norm": 0.3415803909301758, "learning_rate": 8e-05, "loss": 1.6555, "step": 5020 }, { "epoch": 1.1044759705267788, "grad_norm": 0.3358510434627533, "learning_rate": 8e-05, "loss": 1.6675, "step": 5021 }, { "epoch": 1.104695919938414, "grad_norm": 0.34978464245796204, "learning_rate": 8e-05, "loss": 1.5632, "step": 5022 }, { "epoch": 1.1049158693500494, "grad_norm": 0.33762186765670776, "learning_rate": 8e-05, "loss": 1.597, "step": 5023 }, { "epoch": 1.1051358187616849, "grad_norm": 0.32468852400779724, "learning_rate": 8e-05, "loss": 1.4187, "step": 5024 }, { "epoch": 1.1053557681733202, "grad_norm": 0.3331574499607086, "learning_rate": 8e-05, "loss": 1.6283, "step": 5025 }, { "epoch": 1.1055757175849554, "grad_norm": 0.31028181314468384, "learning_rate": 8e-05, "loss": 1.5533, "step": 5026 }, { "epoch": 1.1057956669965907, "grad_norm": 0.34552764892578125, "learning_rate": 8e-05, "loss": 1.5866, "step": 5027 }, { "epoch": 1.1060156164082262, "grad_norm": 0.34182167053222656, "learning_rate": 8e-05, "loss": 1.5581, "step": 5028 }, { "epoch": 1.1062355658198615, "grad_norm": 0.32179248332977295, "learning_rate": 8e-05, "loss": 1.7732, "step": 5029 }, { "epoch": 1.1064555152314968, "grad_norm": 0.30047982931137085, "learning_rate": 8e-05, "loss": 1.487, "step": 5030 }, { "epoch": 1.106675464643132, "grad_norm": 0.32344919443130493, "learning_rate": 8e-05, "loss": 1.6408, "step": 5031 }, { "epoch": 1.1068954140547673, "grad_norm": 0.33185386657714844, "learning_rate": 8e-05, "loss": 1.5058, "step": 5032 }, { "epoch": 1.1071153634664028, "grad_norm": 0.32186517119407654, "learning_rate": 8e-05, "loss": 1.5306, "step": 5033 }, { "epoch": 1.107335312878038, "grad_norm": 0.33195388317108154, "learning_rate": 8e-05, "loss": 1.5559, "step": 5034 }, { "epoch": 1.1075552622896734, "grad_norm": 0.3100522756576538, "learning_rate": 8e-05, "loss": 1.4017, "step": 5035 }, { "epoch": 1.1077752117013087, "grad_norm": 0.33637264370918274, "learning_rate": 8e-05, "loss": 1.6882, "step": 5036 }, { "epoch": 1.107995161112944, "grad_norm": 0.33185431361198425, "learning_rate": 8e-05, "loss": 1.5474, "step": 5037 }, { "epoch": 1.1082151105245794, "grad_norm": 0.3057461082935333, "learning_rate": 8e-05, "loss": 1.4498, "step": 5038 }, { "epoch": 1.1084350599362147, "grad_norm": 0.31757861375808716, "learning_rate": 8e-05, "loss": 1.651, "step": 5039 }, { "epoch": 1.10865500934785, "grad_norm": 0.31345975399017334, "learning_rate": 8e-05, "loss": 1.4606, "step": 5040 }, { "epoch": 1.1088749587594853, "grad_norm": 0.310784250497818, "learning_rate": 8e-05, "loss": 1.5412, "step": 5041 }, { "epoch": 1.1090949081711206, "grad_norm": 0.35223492980003357, "learning_rate": 8e-05, "loss": 1.7017, "step": 5042 }, { "epoch": 1.109314857582756, "grad_norm": 0.3677147328853607, "learning_rate": 8e-05, "loss": 1.6374, "step": 5043 }, { "epoch": 1.1095348069943913, "grad_norm": 0.3289753198623657, "learning_rate": 8e-05, "loss": 1.4801, "step": 5044 }, { "epoch": 1.1097547564060266, "grad_norm": 0.3232313394546509, "learning_rate": 8e-05, "loss": 1.5673, "step": 5045 }, { "epoch": 1.1099747058176619, "grad_norm": 0.3122326135635376, "learning_rate": 8e-05, "loss": 1.5411, "step": 5046 }, { "epoch": 1.1101946552292972, "grad_norm": 0.31872233748435974, "learning_rate": 8e-05, "loss": 1.3971, "step": 5047 }, { "epoch": 1.1104146046409327, "grad_norm": 0.33294031023979187, "learning_rate": 8e-05, "loss": 1.714, "step": 5048 }, { "epoch": 1.110634554052568, "grad_norm": 0.3526622951030731, "learning_rate": 8e-05, "loss": 1.6007, "step": 5049 }, { "epoch": 1.1108545034642032, "grad_norm": 0.3182721734046936, "learning_rate": 8e-05, "loss": 1.548, "step": 5050 }, { "epoch": 1.1110744528758385, "grad_norm": 0.3197391629219055, "learning_rate": 8e-05, "loss": 1.5852, "step": 5051 }, { "epoch": 1.1112944022874738, "grad_norm": 0.34332001209259033, "learning_rate": 8e-05, "loss": 1.609, "step": 5052 }, { "epoch": 1.1115143516991093, "grad_norm": 0.29913702607154846, "learning_rate": 8e-05, "loss": 1.4465, "step": 5053 }, { "epoch": 1.1117343011107446, "grad_norm": 0.32063722610473633, "learning_rate": 8e-05, "loss": 1.4696, "step": 5054 }, { "epoch": 1.1119542505223798, "grad_norm": 0.3325173556804657, "learning_rate": 8e-05, "loss": 1.5659, "step": 5055 }, { "epoch": 1.1121741999340151, "grad_norm": 0.33699125051498413, "learning_rate": 8e-05, "loss": 1.5844, "step": 5056 }, { "epoch": 1.1123941493456506, "grad_norm": 0.3417279124259949, "learning_rate": 8e-05, "loss": 1.5125, "step": 5057 }, { "epoch": 1.112614098757286, "grad_norm": 0.3325464725494385, "learning_rate": 8e-05, "loss": 1.677, "step": 5058 }, { "epoch": 1.1128340481689212, "grad_norm": 0.33324649930000305, "learning_rate": 8e-05, "loss": 1.5552, "step": 5059 }, { "epoch": 1.1130539975805565, "grad_norm": 0.3148370683193207, "learning_rate": 8e-05, "loss": 1.5716, "step": 5060 }, { "epoch": 1.1132739469921917, "grad_norm": 0.3522774577140808, "learning_rate": 8e-05, "loss": 1.6748, "step": 5061 }, { "epoch": 1.1134938964038272, "grad_norm": 0.4092399477958679, "learning_rate": 8e-05, "loss": 1.8287, "step": 5062 }, { "epoch": 1.1137138458154625, "grad_norm": 0.3320629894733429, "learning_rate": 8e-05, "loss": 1.6218, "step": 5063 }, { "epoch": 1.1139337952270978, "grad_norm": 0.32833537459373474, "learning_rate": 8e-05, "loss": 1.4546, "step": 5064 }, { "epoch": 1.114153744638733, "grad_norm": 0.3366737961769104, "learning_rate": 8e-05, "loss": 1.5454, "step": 5065 }, { "epoch": 1.1143736940503683, "grad_norm": 0.3287598788738251, "learning_rate": 8e-05, "loss": 1.5318, "step": 5066 }, { "epoch": 1.1145936434620038, "grad_norm": 0.3353724181652069, "learning_rate": 8e-05, "loss": 1.5376, "step": 5067 }, { "epoch": 1.1148135928736391, "grad_norm": 0.3277460038661957, "learning_rate": 8e-05, "loss": 1.5895, "step": 5068 }, { "epoch": 1.1150335422852744, "grad_norm": 0.3464389443397522, "learning_rate": 8e-05, "loss": 1.6818, "step": 5069 }, { "epoch": 1.1152534916969097, "grad_norm": 0.31511595845222473, "learning_rate": 8e-05, "loss": 1.5304, "step": 5070 }, { "epoch": 1.115473441108545, "grad_norm": 0.31136149168014526, "learning_rate": 8e-05, "loss": 1.6456, "step": 5071 }, { "epoch": 1.1156933905201805, "grad_norm": 0.31155848503112793, "learning_rate": 8e-05, "loss": 1.5911, "step": 5072 }, { "epoch": 1.1159133399318157, "grad_norm": 0.3226284980773926, "learning_rate": 8e-05, "loss": 1.5256, "step": 5073 }, { "epoch": 1.116133289343451, "grad_norm": 0.3271101415157318, "learning_rate": 8e-05, "loss": 1.6583, "step": 5074 }, { "epoch": 1.1163532387550863, "grad_norm": 0.3435283303260803, "learning_rate": 8e-05, "loss": 1.5363, "step": 5075 }, { "epoch": 1.1165731881667216, "grad_norm": 0.3330495059490204, "learning_rate": 8e-05, "loss": 1.7619, "step": 5076 }, { "epoch": 1.116793137578357, "grad_norm": 0.31265267729759216, "learning_rate": 8e-05, "loss": 1.5341, "step": 5077 }, { "epoch": 1.1170130869899924, "grad_norm": 0.37733444571495056, "learning_rate": 8e-05, "loss": 1.6618, "step": 5078 }, { "epoch": 1.1172330364016276, "grad_norm": 0.3527967929840088, "learning_rate": 8e-05, "loss": 1.7136, "step": 5079 }, { "epoch": 1.117452985813263, "grad_norm": 0.3261122703552246, "learning_rate": 8e-05, "loss": 1.594, "step": 5080 }, { "epoch": 1.1176729352248982, "grad_norm": 0.3716588318347931, "learning_rate": 8e-05, "loss": 1.6035, "step": 5081 }, { "epoch": 1.1178928846365337, "grad_norm": 0.3316706418991089, "learning_rate": 8e-05, "loss": 1.6709, "step": 5082 }, { "epoch": 1.118112834048169, "grad_norm": 0.3227297067642212, "learning_rate": 8e-05, "loss": 1.6054, "step": 5083 }, { "epoch": 1.1183327834598042, "grad_norm": 0.34253278374671936, "learning_rate": 8e-05, "loss": 1.5486, "step": 5084 }, { "epoch": 1.1185527328714395, "grad_norm": 0.3321642279624939, "learning_rate": 8e-05, "loss": 1.5896, "step": 5085 }, { "epoch": 1.1187726822830748, "grad_norm": 0.3106277883052826, "learning_rate": 8e-05, "loss": 1.5263, "step": 5086 }, { "epoch": 1.1189926316947103, "grad_norm": 0.31817442178726196, "learning_rate": 8e-05, "loss": 1.5057, "step": 5087 }, { "epoch": 1.1192125811063456, "grad_norm": 0.3145909607410431, "learning_rate": 8e-05, "loss": 1.4045, "step": 5088 }, { "epoch": 1.1194325305179809, "grad_norm": 0.32602986693382263, "learning_rate": 8e-05, "loss": 1.6293, "step": 5089 }, { "epoch": 1.1196524799296161, "grad_norm": 0.3545764684677124, "learning_rate": 8e-05, "loss": 1.5405, "step": 5090 }, { "epoch": 1.1198724293412514, "grad_norm": 0.32123392820358276, "learning_rate": 8e-05, "loss": 1.5323, "step": 5091 }, { "epoch": 1.120092378752887, "grad_norm": 0.306854248046875, "learning_rate": 8e-05, "loss": 1.42, "step": 5092 }, { "epoch": 1.1203123281645222, "grad_norm": 0.341370165348053, "learning_rate": 8e-05, "loss": 1.5608, "step": 5093 }, { "epoch": 1.1205322775761575, "grad_norm": 0.3305357098579407, "learning_rate": 8e-05, "loss": 1.6057, "step": 5094 }, { "epoch": 1.1207522269877928, "grad_norm": 0.33589455485343933, "learning_rate": 8e-05, "loss": 1.6233, "step": 5095 }, { "epoch": 1.120972176399428, "grad_norm": 0.3268560767173767, "learning_rate": 8e-05, "loss": 1.6387, "step": 5096 }, { "epoch": 1.1211921258110635, "grad_norm": 0.379149854183197, "learning_rate": 8e-05, "loss": 1.5716, "step": 5097 }, { "epoch": 1.1214120752226988, "grad_norm": 0.32057011127471924, "learning_rate": 8e-05, "loss": 1.4858, "step": 5098 }, { "epoch": 1.121632024634334, "grad_norm": 0.3142772614955902, "learning_rate": 8e-05, "loss": 1.5841, "step": 5099 }, { "epoch": 1.1218519740459694, "grad_norm": 0.3322499394416809, "learning_rate": 8e-05, "loss": 1.6728, "step": 5100 }, { "epoch": 1.1220719234576046, "grad_norm": 0.32901647686958313, "learning_rate": 8e-05, "loss": 1.5542, "step": 5101 }, { "epoch": 1.1222918728692401, "grad_norm": 0.31181979179382324, "learning_rate": 8e-05, "loss": 1.4743, "step": 5102 }, { "epoch": 1.1225118222808754, "grad_norm": 0.3399113416671753, "learning_rate": 8e-05, "loss": 1.5871, "step": 5103 }, { "epoch": 1.1227317716925107, "grad_norm": 0.33891019225120544, "learning_rate": 8e-05, "loss": 1.5407, "step": 5104 }, { "epoch": 1.122951721104146, "grad_norm": 0.33827653527259827, "learning_rate": 8e-05, "loss": 1.7274, "step": 5105 }, { "epoch": 1.1231716705157813, "grad_norm": 0.3077896535396576, "learning_rate": 8e-05, "loss": 1.5706, "step": 5106 }, { "epoch": 1.1233916199274168, "grad_norm": 0.3457901179790497, "learning_rate": 8e-05, "loss": 1.6504, "step": 5107 }, { "epoch": 1.123611569339052, "grad_norm": 0.3221600651741028, "learning_rate": 8e-05, "loss": 1.454, "step": 5108 }, { "epoch": 1.1238315187506873, "grad_norm": 0.320821613073349, "learning_rate": 8e-05, "loss": 1.5007, "step": 5109 }, { "epoch": 1.1240514681623226, "grad_norm": 0.33881670236587524, "learning_rate": 8e-05, "loss": 1.6184, "step": 5110 }, { "epoch": 1.1242714175739579, "grad_norm": 0.32681500911712646, "learning_rate": 8e-05, "loss": 1.594, "step": 5111 }, { "epoch": 1.1244913669855934, "grad_norm": 0.3243984282016754, "learning_rate": 8e-05, "loss": 1.5437, "step": 5112 }, { "epoch": 1.1247113163972287, "grad_norm": 0.3210890591144562, "learning_rate": 8e-05, "loss": 1.6374, "step": 5113 }, { "epoch": 1.124931265808864, "grad_norm": 0.3472272455692291, "learning_rate": 8e-05, "loss": 1.6653, "step": 5114 }, { "epoch": 1.1251512152204992, "grad_norm": 0.33331987261772156, "learning_rate": 8e-05, "loss": 1.5885, "step": 5115 }, { "epoch": 1.1253711646321345, "grad_norm": 0.3285711407661438, "learning_rate": 8e-05, "loss": 1.6174, "step": 5116 }, { "epoch": 1.12559111404377, "grad_norm": 0.32795506715774536, "learning_rate": 8e-05, "loss": 1.3565, "step": 5117 }, { "epoch": 1.1258110634554053, "grad_norm": 0.3626576066017151, "learning_rate": 8e-05, "loss": 1.6297, "step": 5118 }, { "epoch": 1.1260310128670405, "grad_norm": 0.3133982717990875, "learning_rate": 8e-05, "loss": 1.4358, "step": 5119 }, { "epoch": 1.1262509622786758, "grad_norm": 0.31942275166511536, "learning_rate": 8e-05, "loss": 1.5068, "step": 5120 }, { "epoch": 1.126470911690311, "grad_norm": 0.33955660462379456, "learning_rate": 8e-05, "loss": 1.6789, "step": 5121 }, { "epoch": 1.1266908611019466, "grad_norm": 0.37339428067207336, "learning_rate": 8e-05, "loss": 1.7943, "step": 5122 }, { "epoch": 1.1269108105135819, "grad_norm": 0.30857279896736145, "learning_rate": 8e-05, "loss": 1.492, "step": 5123 }, { "epoch": 1.1271307599252172, "grad_norm": 0.3245776295661926, "learning_rate": 8e-05, "loss": 1.5625, "step": 5124 }, { "epoch": 1.1273507093368524, "grad_norm": 0.31873783469200134, "learning_rate": 8e-05, "loss": 1.5676, "step": 5125 }, { "epoch": 1.127570658748488, "grad_norm": 0.3208547532558441, "learning_rate": 8e-05, "loss": 1.3796, "step": 5126 }, { "epoch": 1.1277906081601232, "grad_norm": 0.3517221510410309, "learning_rate": 8e-05, "loss": 1.4954, "step": 5127 }, { "epoch": 1.1280105575717585, "grad_norm": 0.319656103849411, "learning_rate": 8e-05, "loss": 1.5633, "step": 5128 }, { "epoch": 1.1282305069833938, "grad_norm": 0.3175421953201294, "learning_rate": 8e-05, "loss": 1.4826, "step": 5129 }, { "epoch": 1.128450456395029, "grad_norm": 0.3292050063610077, "learning_rate": 8e-05, "loss": 1.5922, "step": 5130 }, { "epoch": 1.1286704058066646, "grad_norm": 0.31769126653671265, "learning_rate": 8e-05, "loss": 1.4714, "step": 5131 }, { "epoch": 1.1288903552182998, "grad_norm": 0.33158057928085327, "learning_rate": 8e-05, "loss": 1.5359, "step": 5132 }, { "epoch": 1.1291103046299351, "grad_norm": 0.3481951057910919, "learning_rate": 8e-05, "loss": 1.7386, "step": 5133 }, { "epoch": 1.1293302540415704, "grad_norm": 0.33302003145217896, "learning_rate": 8e-05, "loss": 1.5563, "step": 5134 }, { "epoch": 1.1295502034532057, "grad_norm": 0.3279911279678345, "learning_rate": 8e-05, "loss": 1.5511, "step": 5135 }, { "epoch": 1.1297701528648412, "grad_norm": 0.32343900203704834, "learning_rate": 8e-05, "loss": 1.5811, "step": 5136 }, { "epoch": 1.1299901022764764, "grad_norm": 0.3172832727432251, "learning_rate": 8e-05, "loss": 1.6458, "step": 5137 }, { "epoch": 1.1302100516881117, "grad_norm": 0.3302123546600342, "learning_rate": 8e-05, "loss": 1.6981, "step": 5138 }, { "epoch": 1.130430001099747, "grad_norm": 0.3414337933063507, "learning_rate": 8e-05, "loss": 1.8442, "step": 5139 }, { "epoch": 1.1306499505113825, "grad_norm": 0.3398199677467346, "learning_rate": 8e-05, "loss": 1.6196, "step": 5140 }, { "epoch": 1.1308698999230178, "grad_norm": 0.3407689332962036, "learning_rate": 8e-05, "loss": 1.7702, "step": 5141 }, { "epoch": 1.131089849334653, "grad_norm": 0.33658891916275024, "learning_rate": 8e-05, "loss": 1.6651, "step": 5142 }, { "epoch": 1.1313097987462883, "grad_norm": 0.32450464367866516, "learning_rate": 8e-05, "loss": 1.5154, "step": 5143 }, { "epoch": 1.1315297481579236, "grad_norm": 0.32278215885162354, "learning_rate": 8e-05, "loss": 1.6561, "step": 5144 }, { "epoch": 1.1317496975695591, "grad_norm": 0.3105833828449249, "learning_rate": 8e-05, "loss": 1.5237, "step": 5145 }, { "epoch": 1.1319696469811944, "grad_norm": 0.3566106855869293, "learning_rate": 8e-05, "loss": 1.8513, "step": 5146 }, { "epoch": 1.1321895963928297, "grad_norm": 0.3124619722366333, "learning_rate": 8e-05, "loss": 1.5371, "step": 5147 }, { "epoch": 1.132409545804465, "grad_norm": 0.30793124437332153, "learning_rate": 8e-05, "loss": 1.484, "step": 5148 }, { "epoch": 1.1326294952161002, "grad_norm": 0.3573998212814331, "learning_rate": 8e-05, "loss": 1.4632, "step": 5149 }, { "epoch": 1.1328494446277357, "grad_norm": 0.3092050552368164, "learning_rate": 8e-05, "loss": 1.5631, "step": 5150 }, { "epoch": 1.133069394039371, "grad_norm": 0.31325066089630127, "learning_rate": 8e-05, "loss": 1.4787, "step": 5151 }, { "epoch": 1.1332893434510063, "grad_norm": 0.31855833530426025, "learning_rate": 8e-05, "loss": 1.3667, "step": 5152 }, { "epoch": 1.1335092928626416, "grad_norm": 0.3058852255344391, "learning_rate": 8e-05, "loss": 1.5633, "step": 5153 }, { "epoch": 1.1337292422742768, "grad_norm": 0.32730501890182495, "learning_rate": 8e-05, "loss": 1.5121, "step": 5154 }, { "epoch": 1.1339491916859123, "grad_norm": 0.3253394663333893, "learning_rate": 8e-05, "loss": 1.6508, "step": 5155 }, { "epoch": 1.1341691410975476, "grad_norm": 0.33001965284347534, "learning_rate": 8e-05, "loss": 1.6086, "step": 5156 }, { "epoch": 1.134389090509183, "grad_norm": 0.3528082072734833, "learning_rate": 8e-05, "loss": 1.4909, "step": 5157 }, { "epoch": 1.1346090399208182, "grad_norm": 0.3358846604824066, "learning_rate": 8e-05, "loss": 1.4018, "step": 5158 }, { "epoch": 1.1348289893324535, "grad_norm": 0.3301566243171692, "learning_rate": 8e-05, "loss": 1.4602, "step": 5159 }, { "epoch": 1.135048938744089, "grad_norm": 0.3090686500072479, "learning_rate": 8e-05, "loss": 1.3886, "step": 5160 }, { "epoch": 1.1352688881557242, "grad_norm": 0.3259727656841278, "learning_rate": 8e-05, "loss": 1.4529, "step": 5161 }, { "epoch": 1.1354888375673595, "grad_norm": 0.31200674176216125, "learning_rate": 8e-05, "loss": 1.4417, "step": 5162 }, { "epoch": 1.1357087869789948, "grad_norm": 0.3312123119831085, "learning_rate": 8e-05, "loss": 1.6306, "step": 5163 }, { "epoch": 1.13592873639063, "grad_norm": 0.34394145011901855, "learning_rate": 8e-05, "loss": 1.7628, "step": 5164 }, { "epoch": 1.1361486858022656, "grad_norm": 0.3211493194103241, "learning_rate": 8e-05, "loss": 1.5522, "step": 5165 }, { "epoch": 1.1363686352139009, "grad_norm": 0.3305715322494507, "learning_rate": 8e-05, "loss": 1.5752, "step": 5166 }, { "epoch": 1.1365885846255361, "grad_norm": 0.3293518126010895, "learning_rate": 8e-05, "loss": 1.5852, "step": 5167 }, { "epoch": 1.1368085340371714, "grad_norm": 0.3357377052307129, "learning_rate": 8e-05, "loss": 1.5334, "step": 5168 }, { "epoch": 1.1370284834488067, "grad_norm": 0.3383323848247528, "learning_rate": 8e-05, "loss": 1.5979, "step": 5169 }, { "epoch": 1.1372484328604422, "grad_norm": 0.3330705165863037, "learning_rate": 8e-05, "loss": 1.656, "step": 5170 }, { "epoch": 1.1374683822720775, "grad_norm": 0.3366387188434601, "learning_rate": 8e-05, "loss": 1.6028, "step": 5171 }, { "epoch": 1.1376883316837128, "grad_norm": 0.3154458701610565, "learning_rate": 8e-05, "loss": 1.5559, "step": 5172 }, { "epoch": 1.137908281095348, "grad_norm": 0.32994750142097473, "learning_rate": 8e-05, "loss": 1.5113, "step": 5173 }, { "epoch": 1.1381282305069833, "grad_norm": 0.32376259565353394, "learning_rate": 8e-05, "loss": 1.7197, "step": 5174 }, { "epoch": 1.1383481799186188, "grad_norm": 0.34894615411758423, "learning_rate": 8e-05, "loss": 1.4991, "step": 5175 }, { "epoch": 1.138568129330254, "grad_norm": 0.32551950216293335, "learning_rate": 8e-05, "loss": 1.7635, "step": 5176 }, { "epoch": 1.1387880787418894, "grad_norm": 0.32671916484832764, "learning_rate": 8e-05, "loss": 1.6442, "step": 5177 }, { "epoch": 1.1390080281535246, "grad_norm": 0.32519635558128357, "learning_rate": 8e-05, "loss": 1.5207, "step": 5178 }, { "epoch": 1.13922797756516, "grad_norm": 0.3291681408882141, "learning_rate": 8e-05, "loss": 1.666, "step": 5179 }, { "epoch": 1.1394479269767954, "grad_norm": 0.31787407398223877, "learning_rate": 8e-05, "loss": 1.6009, "step": 5180 }, { "epoch": 1.1396678763884307, "grad_norm": 0.3146039843559265, "learning_rate": 8e-05, "loss": 1.5795, "step": 5181 }, { "epoch": 1.139887825800066, "grad_norm": 0.32676151394844055, "learning_rate": 8e-05, "loss": 1.578, "step": 5182 }, { "epoch": 1.1401077752117013, "grad_norm": 0.3422422409057617, "learning_rate": 8e-05, "loss": 1.6054, "step": 5183 }, { "epoch": 1.1403277246233365, "grad_norm": 0.3256738781929016, "learning_rate": 8e-05, "loss": 1.6739, "step": 5184 }, { "epoch": 1.140547674034972, "grad_norm": 0.3405067026615143, "learning_rate": 8e-05, "loss": 1.6888, "step": 5185 }, { "epoch": 1.1407676234466073, "grad_norm": 0.3133193850517273, "learning_rate": 8e-05, "loss": 1.3727, "step": 5186 }, { "epoch": 1.1409875728582426, "grad_norm": 0.3357144594192505, "learning_rate": 8e-05, "loss": 1.5833, "step": 5187 }, { "epoch": 1.1412075222698779, "grad_norm": 0.3582302927970886, "learning_rate": 8e-05, "loss": 1.5144, "step": 5188 }, { "epoch": 1.1414274716815132, "grad_norm": 0.3194274604320526, "learning_rate": 8e-05, "loss": 1.6315, "step": 5189 }, { "epoch": 1.1416474210931487, "grad_norm": 0.3333747684955597, "learning_rate": 8e-05, "loss": 1.6568, "step": 5190 }, { "epoch": 1.141867370504784, "grad_norm": 0.3355323374271393, "learning_rate": 8e-05, "loss": 1.5113, "step": 5191 }, { "epoch": 1.1420873199164192, "grad_norm": 0.3134694993495941, "learning_rate": 8e-05, "loss": 1.5214, "step": 5192 }, { "epoch": 1.1423072693280545, "grad_norm": 0.3097131848335266, "learning_rate": 8e-05, "loss": 1.5307, "step": 5193 }, { "epoch": 1.1425272187396898, "grad_norm": 0.3294723629951477, "learning_rate": 8e-05, "loss": 1.6075, "step": 5194 }, { "epoch": 1.1427471681513253, "grad_norm": 0.34683695435523987, "learning_rate": 8e-05, "loss": 1.4961, "step": 5195 }, { "epoch": 1.1429671175629605, "grad_norm": 0.31840071082115173, "learning_rate": 8e-05, "loss": 1.5651, "step": 5196 }, { "epoch": 1.1431870669745958, "grad_norm": 0.321075439453125, "learning_rate": 8e-05, "loss": 1.6502, "step": 5197 }, { "epoch": 1.143407016386231, "grad_norm": 0.3229003846645355, "learning_rate": 8e-05, "loss": 1.5082, "step": 5198 }, { "epoch": 1.1436269657978664, "grad_norm": 0.33982300758361816, "learning_rate": 8e-05, "loss": 1.675, "step": 5199 }, { "epoch": 1.1438469152095019, "grad_norm": 0.3364129662513733, "learning_rate": 8e-05, "loss": 1.5487, "step": 5200 }, { "epoch": 1.1440668646211372, "grad_norm": 0.3294176459312439, "learning_rate": 8e-05, "loss": 1.6401, "step": 5201 }, { "epoch": 1.1442868140327724, "grad_norm": 0.3198351562023163, "learning_rate": 8e-05, "loss": 1.6079, "step": 5202 }, { "epoch": 1.1445067634444077, "grad_norm": 0.3304404020309448, "learning_rate": 8e-05, "loss": 1.6805, "step": 5203 }, { "epoch": 1.144726712856043, "grad_norm": 0.3445228934288025, "learning_rate": 8e-05, "loss": 1.4646, "step": 5204 }, { "epoch": 1.1449466622676785, "grad_norm": 0.32834136486053467, "learning_rate": 8e-05, "loss": 1.6929, "step": 5205 }, { "epoch": 1.1451666116793138, "grad_norm": 0.3230096995830536, "learning_rate": 8e-05, "loss": 1.6398, "step": 5206 }, { "epoch": 1.145386561090949, "grad_norm": 0.33999907970428467, "learning_rate": 8e-05, "loss": 1.5168, "step": 5207 }, { "epoch": 1.1456065105025843, "grad_norm": 0.3245522677898407, "learning_rate": 8e-05, "loss": 1.5785, "step": 5208 }, { "epoch": 1.1458264599142196, "grad_norm": 0.341769278049469, "learning_rate": 8e-05, "loss": 1.6476, "step": 5209 }, { "epoch": 1.146046409325855, "grad_norm": 0.352146714925766, "learning_rate": 8e-05, "loss": 1.6664, "step": 5210 }, { "epoch": 1.1462663587374904, "grad_norm": 0.3433312475681305, "learning_rate": 8e-05, "loss": 1.6613, "step": 5211 }, { "epoch": 1.1464863081491257, "grad_norm": 0.3751370906829834, "learning_rate": 8e-05, "loss": 1.6801, "step": 5212 }, { "epoch": 1.146706257560761, "grad_norm": 0.31403568387031555, "learning_rate": 8e-05, "loss": 1.4792, "step": 5213 }, { "epoch": 1.1469262069723964, "grad_norm": 0.3489457964897156, "learning_rate": 8e-05, "loss": 1.6575, "step": 5214 }, { "epoch": 1.1471461563840317, "grad_norm": 0.34716835618019104, "learning_rate": 8e-05, "loss": 1.725, "step": 5215 }, { "epoch": 1.147366105795667, "grad_norm": 0.32516106963157654, "learning_rate": 8e-05, "loss": 1.6056, "step": 5216 }, { "epoch": 1.1475860552073023, "grad_norm": 0.3346816897392273, "learning_rate": 8e-05, "loss": 1.6168, "step": 5217 }, { "epoch": 1.1478060046189376, "grad_norm": 0.3132788836956024, "learning_rate": 8e-05, "loss": 1.4822, "step": 5218 }, { "epoch": 1.148025954030573, "grad_norm": 0.34313905239105225, "learning_rate": 8e-05, "loss": 1.6065, "step": 5219 }, { "epoch": 1.1482459034422083, "grad_norm": 0.32414478063583374, "learning_rate": 8e-05, "loss": 1.4274, "step": 5220 }, { "epoch": 1.1484658528538436, "grad_norm": 0.3358854651451111, "learning_rate": 8e-05, "loss": 1.5507, "step": 5221 }, { "epoch": 1.148685802265479, "grad_norm": 0.32924723625183105, "learning_rate": 8e-05, "loss": 1.6701, "step": 5222 }, { "epoch": 1.1489057516771142, "grad_norm": 0.31682083010673523, "learning_rate": 8e-05, "loss": 1.4948, "step": 5223 }, { "epoch": 1.1491257010887497, "grad_norm": 0.3245946764945984, "learning_rate": 8e-05, "loss": 1.5186, "step": 5224 }, { "epoch": 1.149345650500385, "grad_norm": 0.33114445209503174, "learning_rate": 8e-05, "loss": 1.6288, "step": 5225 }, { "epoch": 1.1495655999120202, "grad_norm": 0.35070741176605225, "learning_rate": 8e-05, "loss": 1.549, "step": 5226 }, { "epoch": 1.1497855493236555, "grad_norm": 0.33421263098716736, "learning_rate": 8e-05, "loss": 1.4953, "step": 5227 }, { "epoch": 1.1500054987352908, "grad_norm": 0.31994763016700745, "learning_rate": 8e-05, "loss": 1.5215, "step": 5228 }, { "epoch": 1.1502254481469263, "grad_norm": 0.33780449628829956, "learning_rate": 8e-05, "loss": 1.7376, "step": 5229 }, { "epoch": 1.1504453975585616, "grad_norm": 0.3094518184661865, "learning_rate": 8e-05, "loss": 1.5061, "step": 5230 }, { "epoch": 1.1506653469701968, "grad_norm": 0.33406639099121094, "learning_rate": 8e-05, "loss": 1.6612, "step": 5231 }, { "epoch": 1.1508852963818321, "grad_norm": 0.3320581614971161, "learning_rate": 8e-05, "loss": 1.5844, "step": 5232 }, { "epoch": 1.1511052457934676, "grad_norm": 0.32815083861351013, "learning_rate": 8e-05, "loss": 1.7442, "step": 5233 }, { "epoch": 1.151325195205103, "grad_norm": 0.3417838513851166, "learning_rate": 8e-05, "loss": 1.6012, "step": 5234 }, { "epoch": 1.1515451446167382, "grad_norm": 0.34620028734207153, "learning_rate": 8e-05, "loss": 1.7045, "step": 5235 }, { "epoch": 1.1517650940283735, "grad_norm": 0.3147315979003906, "learning_rate": 8e-05, "loss": 1.5557, "step": 5236 }, { "epoch": 1.1519850434400087, "grad_norm": 0.3592619299888611, "learning_rate": 8e-05, "loss": 1.7826, "step": 5237 }, { "epoch": 1.1522049928516442, "grad_norm": 0.3434905409812927, "learning_rate": 8e-05, "loss": 1.5314, "step": 5238 }, { "epoch": 1.1524249422632795, "grad_norm": 0.34384992718696594, "learning_rate": 8e-05, "loss": 1.7704, "step": 5239 }, { "epoch": 1.1526448916749148, "grad_norm": 0.36034470796585083, "learning_rate": 8e-05, "loss": 1.5648, "step": 5240 }, { "epoch": 1.15286484108655, "grad_norm": 0.3212474584579468, "learning_rate": 8e-05, "loss": 1.5768, "step": 5241 }, { "epoch": 1.1530847904981854, "grad_norm": 0.34541085362434387, "learning_rate": 8e-05, "loss": 1.7168, "step": 5242 }, { "epoch": 1.1533047399098209, "grad_norm": 0.38996708393096924, "learning_rate": 8e-05, "loss": 1.7377, "step": 5243 }, { "epoch": 1.1535246893214561, "grad_norm": 0.34553247690200806, "learning_rate": 8e-05, "loss": 1.5117, "step": 5244 }, { "epoch": 1.1537446387330914, "grad_norm": 0.3314375579357147, "learning_rate": 8e-05, "loss": 1.5537, "step": 5245 }, { "epoch": 1.1539645881447267, "grad_norm": 0.3203626573085785, "learning_rate": 8e-05, "loss": 1.6466, "step": 5246 }, { "epoch": 1.154184537556362, "grad_norm": 0.31628188490867615, "learning_rate": 8e-05, "loss": 1.5232, "step": 5247 }, { "epoch": 1.1544044869679975, "grad_norm": 0.3399323523044586, "learning_rate": 8e-05, "loss": 1.7626, "step": 5248 }, { "epoch": 1.1546244363796327, "grad_norm": 0.3148668110370636, "learning_rate": 8e-05, "loss": 1.5592, "step": 5249 }, { "epoch": 1.154844385791268, "grad_norm": 0.3220260739326477, "learning_rate": 8e-05, "loss": 1.6002, "step": 5250 }, { "epoch": 1.1550643352029033, "grad_norm": 0.3275005519390106, "learning_rate": 8e-05, "loss": 1.4953, "step": 5251 }, { "epoch": 1.1552842846145386, "grad_norm": 0.3274150788784027, "learning_rate": 8e-05, "loss": 1.562, "step": 5252 }, { "epoch": 1.155504234026174, "grad_norm": 0.337016761302948, "learning_rate": 8e-05, "loss": 1.7178, "step": 5253 }, { "epoch": 1.1557241834378094, "grad_norm": 0.3254145681858063, "learning_rate": 8e-05, "loss": 1.6088, "step": 5254 }, { "epoch": 1.1559441328494446, "grad_norm": 0.3864974081516266, "learning_rate": 8e-05, "loss": 1.5511, "step": 5255 }, { "epoch": 1.15616408226108, "grad_norm": 0.36195075511932373, "learning_rate": 8e-05, "loss": 1.7118, "step": 5256 }, { "epoch": 1.1563840316727152, "grad_norm": 0.3282105326652527, "learning_rate": 8e-05, "loss": 1.6142, "step": 5257 }, { "epoch": 1.1566039810843507, "grad_norm": 0.34599289298057556, "learning_rate": 8e-05, "loss": 1.6525, "step": 5258 }, { "epoch": 1.156823930495986, "grad_norm": 0.3619854152202606, "learning_rate": 8e-05, "loss": 1.6138, "step": 5259 }, { "epoch": 1.1570438799076213, "grad_norm": 0.33085936307907104, "learning_rate": 8e-05, "loss": 1.5321, "step": 5260 }, { "epoch": 1.1572638293192565, "grad_norm": 0.3586635887622833, "learning_rate": 8e-05, "loss": 1.496, "step": 5261 }, { "epoch": 1.1574837787308918, "grad_norm": 0.33476370573043823, "learning_rate": 8e-05, "loss": 1.7385, "step": 5262 }, { "epoch": 1.1577037281425273, "grad_norm": 0.36151987314224243, "learning_rate": 8e-05, "loss": 1.4995, "step": 5263 }, { "epoch": 1.1579236775541626, "grad_norm": 0.34880131483078003, "learning_rate": 8e-05, "loss": 1.5992, "step": 5264 }, { "epoch": 1.1581436269657979, "grad_norm": 0.3399914801120758, "learning_rate": 8e-05, "loss": 1.5337, "step": 5265 }, { "epoch": 1.1583635763774331, "grad_norm": 0.3159194886684418, "learning_rate": 8e-05, "loss": 1.5235, "step": 5266 }, { "epoch": 1.1585835257890684, "grad_norm": 0.3623919188976288, "learning_rate": 8e-05, "loss": 1.6601, "step": 5267 }, { "epoch": 1.158803475200704, "grad_norm": 0.33179324865341187, "learning_rate": 8e-05, "loss": 1.5351, "step": 5268 }, { "epoch": 1.1590234246123392, "grad_norm": 0.32558172941207886, "learning_rate": 8e-05, "loss": 1.58, "step": 5269 }, { "epoch": 1.1592433740239745, "grad_norm": 0.3160456120967865, "learning_rate": 8e-05, "loss": 1.6792, "step": 5270 }, { "epoch": 1.1594633234356098, "grad_norm": 0.3259306848049164, "learning_rate": 8e-05, "loss": 1.5792, "step": 5271 }, { "epoch": 1.159683272847245, "grad_norm": 0.3459818661212921, "learning_rate": 8e-05, "loss": 1.7299, "step": 5272 }, { "epoch": 1.1599032222588805, "grad_norm": 0.3362138271331787, "learning_rate": 8e-05, "loss": 1.6218, "step": 5273 }, { "epoch": 1.1601231716705158, "grad_norm": 0.3218403160572052, "learning_rate": 8e-05, "loss": 1.525, "step": 5274 }, { "epoch": 1.160343121082151, "grad_norm": 0.32988420128822327, "learning_rate": 8e-05, "loss": 1.6058, "step": 5275 }, { "epoch": 1.1605630704937864, "grad_norm": 0.32046428322792053, "learning_rate": 8e-05, "loss": 1.4607, "step": 5276 }, { "epoch": 1.1607830199054217, "grad_norm": 0.31565141677856445, "learning_rate": 8e-05, "loss": 1.4754, "step": 5277 }, { "epoch": 1.1610029693170572, "grad_norm": 0.32645583152770996, "learning_rate": 8e-05, "loss": 1.6695, "step": 5278 }, { "epoch": 1.1612229187286924, "grad_norm": 0.329095721244812, "learning_rate": 8e-05, "loss": 1.5851, "step": 5279 }, { "epoch": 1.1614428681403277, "grad_norm": 0.35523852705955505, "learning_rate": 8e-05, "loss": 1.7029, "step": 5280 }, { "epoch": 1.161662817551963, "grad_norm": 0.32568076252937317, "learning_rate": 8e-05, "loss": 1.4652, "step": 5281 }, { "epoch": 1.1618827669635983, "grad_norm": 0.3269540071487427, "learning_rate": 8e-05, "loss": 1.6503, "step": 5282 }, { "epoch": 1.1621027163752338, "grad_norm": 0.3148273825645447, "learning_rate": 8e-05, "loss": 1.5795, "step": 5283 }, { "epoch": 1.162322665786869, "grad_norm": 0.32738733291625977, "learning_rate": 8e-05, "loss": 1.7003, "step": 5284 }, { "epoch": 1.1625426151985043, "grad_norm": 0.3145153522491455, "learning_rate": 8e-05, "loss": 1.4548, "step": 5285 }, { "epoch": 1.1627625646101396, "grad_norm": 0.32286927103996277, "learning_rate": 8e-05, "loss": 1.5347, "step": 5286 }, { "epoch": 1.1629825140217749, "grad_norm": 0.32443732023239136, "learning_rate": 8e-05, "loss": 1.6173, "step": 5287 }, { "epoch": 1.1632024634334104, "grad_norm": 0.3473280072212219, "learning_rate": 8e-05, "loss": 1.6949, "step": 5288 }, { "epoch": 1.1634224128450457, "grad_norm": 0.31731489300727844, "learning_rate": 8e-05, "loss": 1.5367, "step": 5289 }, { "epoch": 1.163642362256681, "grad_norm": 0.29897063970565796, "learning_rate": 8e-05, "loss": 1.434, "step": 5290 }, { "epoch": 1.1638623116683162, "grad_norm": 0.31201422214508057, "learning_rate": 8e-05, "loss": 1.4966, "step": 5291 }, { "epoch": 1.1640822610799515, "grad_norm": 0.32904836535453796, "learning_rate": 8e-05, "loss": 1.5503, "step": 5292 }, { "epoch": 1.164302210491587, "grad_norm": 0.3221074938774109, "learning_rate": 8e-05, "loss": 1.6269, "step": 5293 }, { "epoch": 1.1645221599032223, "grad_norm": 0.32835301756858826, "learning_rate": 8e-05, "loss": 1.6373, "step": 5294 }, { "epoch": 1.1647421093148576, "grad_norm": 0.34563249349594116, "learning_rate": 8e-05, "loss": 1.5665, "step": 5295 }, { "epoch": 1.1649620587264928, "grad_norm": 0.3434504270553589, "learning_rate": 8e-05, "loss": 1.58, "step": 5296 }, { "epoch": 1.1651820081381281, "grad_norm": 0.3113488554954529, "learning_rate": 8e-05, "loss": 1.5862, "step": 5297 }, { "epoch": 1.1654019575497636, "grad_norm": 0.3258254826068878, "learning_rate": 8e-05, "loss": 1.7167, "step": 5298 }, { "epoch": 1.165621906961399, "grad_norm": 0.3251175284385681, "learning_rate": 8e-05, "loss": 1.5661, "step": 5299 }, { "epoch": 1.1658418563730342, "grad_norm": 0.3212796449661255, "learning_rate": 8e-05, "loss": 1.4364, "step": 5300 }, { "epoch": 1.1660618057846694, "grad_norm": 0.3193243741989136, "learning_rate": 8e-05, "loss": 1.604, "step": 5301 }, { "epoch": 1.1662817551963047, "grad_norm": 0.33087998628616333, "learning_rate": 8e-05, "loss": 1.5147, "step": 5302 }, { "epoch": 1.1665017046079402, "grad_norm": 0.3636016547679901, "learning_rate": 8e-05, "loss": 1.7103, "step": 5303 }, { "epoch": 1.1667216540195755, "grad_norm": 0.31153616309165955, "learning_rate": 8e-05, "loss": 1.56, "step": 5304 }, { "epoch": 1.1669416034312108, "grad_norm": 0.329251229763031, "learning_rate": 8e-05, "loss": 1.4533, "step": 5305 }, { "epoch": 1.167161552842846, "grad_norm": 0.3564174175262451, "learning_rate": 8e-05, "loss": 1.6549, "step": 5306 }, { "epoch": 1.1673815022544816, "grad_norm": 0.31766775250434875, "learning_rate": 8e-05, "loss": 1.5366, "step": 5307 }, { "epoch": 1.1676014516661168, "grad_norm": 0.335202157497406, "learning_rate": 8e-05, "loss": 1.6693, "step": 5308 }, { "epoch": 1.1678214010777521, "grad_norm": 0.329578697681427, "learning_rate": 8e-05, "loss": 1.6142, "step": 5309 }, { "epoch": 1.1680413504893874, "grad_norm": 0.3394508957862854, "learning_rate": 8e-05, "loss": 1.6416, "step": 5310 }, { "epoch": 1.1682612999010227, "grad_norm": 0.3284442722797394, "learning_rate": 8e-05, "loss": 1.5215, "step": 5311 }, { "epoch": 1.1684812493126582, "grad_norm": 0.3243228793144226, "learning_rate": 8e-05, "loss": 1.5567, "step": 5312 }, { "epoch": 1.1687011987242935, "grad_norm": 0.3502015769481659, "learning_rate": 8e-05, "loss": 1.6602, "step": 5313 }, { "epoch": 1.1689211481359287, "grad_norm": 0.285980761051178, "learning_rate": 8e-05, "loss": 1.2477, "step": 5314 }, { "epoch": 1.169141097547564, "grad_norm": 0.38234028220176697, "learning_rate": 8e-05, "loss": 1.539, "step": 5315 }, { "epoch": 1.1693610469591993, "grad_norm": 0.30444514751434326, "learning_rate": 8e-05, "loss": 1.4346, "step": 5316 }, { "epoch": 1.1695809963708348, "grad_norm": 0.3514763116836548, "learning_rate": 8e-05, "loss": 1.7157, "step": 5317 }, { "epoch": 1.16980094578247, "grad_norm": 0.333297461271286, "learning_rate": 8e-05, "loss": 1.6805, "step": 5318 }, { "epoch": 1.1700208951941053, "grad_norm": 0.31654053926467896, "learning_rate": 8e-05, "loss": 1.6026, "step": 5319 }, { "epoch": 1.1702408446057406, "grad_norm": 0.31996950507164, "learning_rate": 8e-05, "loss": 1.5327, "step": 5320 }, { "epoch": 1.1704607940173761, "grad_norm": 0.34770169854164124, "learning_rate": 8e-05, "loss": 1.6016, "step": 5321 }, { "epoch": 1.1706807434290114, "grad_norm": 0.49952423572540283, "learning_rate": 8e-05, "loss": 1.6909, "step": 5322 }, { "epoch": 1.1709006928406467, "grad_norm": 0.374234139919281, "learning_rate": 8e-05, "loss": 1.7079, "step": 5323 }, { "epoch": 1.171120642252282, "grad_norm": 0.3237394094467163, "learning_rate": 8e-05, "loss": 1.5821, "step": 5324 }, { "epoch": 1.1713405916639172, "grad_norm": 0.37319716811180115, "learning_rate": 8e-05, "loss": 1.5109, "step": 5325 }, { "epoch": 1.1715605410755527, "grad_norm": 0.35475215315818787, "learning_rate": 8e-05, "loss": 1.7489, "step": 5326 }, { "epoch": 1.171780490487188, "grad_norm": 0.3292977511882782, "learning_rate": 8e-05, "loss": 1.5134, "step": 5327 }, { "epoch": 1.1720004398988233, "grad_norm": 0.32654982805252075, "learning_rate": 8e-05, "loss": 1.631, "step": 5328 }, { "epoch": 1.1722203893104586, "grad_norm": 0.33397769927978516, "learning_rate": 8e-05, "loss": 1.6963, "step": 5329 }, { "epoch": 1.1724403387220939, "grad_norm": 0.350326806306839, "learning_rate": 8e-05, "loss": 1.7515, "step": 5330 }, { "epoch": 1.1726602881337294, "grad_norm": 0.3254203796386719, "learning_rate": 8e-05, "loss": 1.5239, "step": 5331 }, { "epoch": 1.1728802375453646, "grad_norm": 0.33540964126586914, "learning_rate": 8e-05, "loss": 1.6718, "step": 5332 }, { "epoch": 1.173100186957, "grad_norm": 0.3301061987876892, "learning_rate": 8e-05, "loss": 1.7011, "step": 5333 }, { "epoch": 1.1733201363686352, "grad_norm": 0.3411114513874054, "learning_rate": 8e-05, "loss": 1.5857, "step": 5334 }, { "epoch": 1.1735400857802705, "grad_norm": 0.3515836000442505, "learning_rate": 8e-05, "loss": 1.6121, "step": 5335 }, { "epoch": 1.173760035191906, "grad_norm": 0.31727758049964905, "learning_rate": 8e-05, "loss": 1.3963, "step": 5336 }, { "epoch": 1.1739799846035412, "grad_norm": 0.3143068552017212, "learning_rate": 8e-05, "loss": 1.4553, "step": 5337 }, { "epoch": 1.1741999340151765, "grad_norm": 0.32039743661880493, "learning_rate": 8e-05, "loss": 1.532, "step": 5338 }, { "epoch": 1.1744198834268118, "grad_norm": 0.3140069842338562, "learning_rate": 8e-05, "loss": 1.6045, "step": 5339 }, { "epoch": 1.174639832838447, "grad_norm": 0.30989301204681396, "learning_rate": 8e-05, "loss": 1.5716, "step": 5340 }, { "epoch": 1.1748597822500826, "grad_norm": 0.31228935718536377, "learning_rate": 8e-05, "loss": 1.4734, "step": 5341 }, { "epoch": 1.1750797316617179, "grad_norm": 0.33272165060043335, "learning_rate": 8e-05, "loss": 1.5597, "step": 5342 }, { "epoch": 1.1752996810733531, "grad_norm": 0.3431159555912018, "learning_rate": 8e-05, "loss": 1.5546, "step": 5343 }, { "epoch": 1.1755196304849884, "grad_norm": 0.34407317638397217, "learning_rate": 8e-05, "loss": 1.6361, "step": 5344 }, { "epoch": 1.1757395798966237, "grad_norm": 0.3060881197452545, "learning_rate": 8e-05, "loss": 1.5371, "step": 5345 }, { "epoch": 1.1759595293082592, "grad_norm": 0.31113526225090027, "learning_rate": 8e-05, "loss": 1.3654, "step": 5346 }, { "epoch": 1.1761794787198945, "grad_norm": 0.3776507079601288, "learning_rate": 8e-05, "loss": 1.5653, "step": 5347 }, { "epoch": 1.1763994281315298, "grad_norm": 0.32404646277427673, "learning_rate": 8e-05, "loss": 1.6833, "step": 5348 }, { "epoch": 1.176619377543165, "grad_norm": 0.3355356454849243, "learning_rate": 8e-05, "loss": 1.5837, "step": 5349 }, { "epoch": 1.1768393269548003, "grad_norm": 0.3192623257637024, "learning_rate": 8e-05, "loss": 1.5325, "step": 5350 }, { "epoch": 1.1770592763664358, "grad_norm": 0.3288591206073761, "learning_rate": 8e-05, "loss": 1.4653, "step": 5351 }, { "epoch": 1.177279225778071, "grad_norm": 0.34229061007499695, "learning_rate": 8e-05, "loss": 1.6407, "step": 5352 }, { "epoch": 1.1774991751897064, "grad_norm": 0.3215787410736084, "learning_rate": 8e-05, "loss": 1.404, "step": 5353 }, { "epoch": 1.1777191246013416, "grad_norm": 0.34103938937187195, "learning_rate": 8e-05, "loss": 1.6437, "step": 5354 }, { "epoch": 1.177939074012977, "grad_norm": 0.32909271121025085, "learning_rate": 8e-05, "loss": 1.6261, "step": 5355 }, { "epoch": 1.1781590234246124, "grad_norm": 0.31601282954216003, "learning_rate": 8e-05, "loss": 1.5727, "step": 5356 }, { "epoch": 1.1783789728362477, "grad_norm": 0.3170057535171509, "learning_rate": 8e-05, "loss": 1.5005, "step": 5357 }, { "epoch": 1.178598922247883, "grad_norm": 0.32038405537605286, "learning_rate": 8e-05, "loss": 1.5457, "step": 5358 }, { "epoch": 1.1788188716595183, "grad_norm": 0.31853222846984863, "learning_rate": 8e-05, "loss": 1.5491, "step": 5359 }, { "epoch": 1.1790388210711535, "grad_norm": 0.34559717774391174, "learning_rate": 8e-05, "loss": 1.6177, "step": 5360 }, { "epoch": 1.179258770482789, "grad_norm": 0.33484289050102234, "learning_rate": 8e-05, "loss": 1.5433, "step": 5361 }, { "epoch": 1.1794787198944243, "grad_norm": 0.32582131028175354, "learning_rate": 8e-05, "loss": 1.5999, "step": 5362 }, { "epoch": 1.1796986693060596, "grad_norm": 0.3173767626285553, "learning_rate": 8e-05, "loss": 1.5223, "step": 5363 }, { "epoch": 1.1799186187176949, "grad_norm": 0.3221052587032318, "learning_rate": 8e-05, "loss": 1.6375, "step": 5364 }, { "epoch": 1.1801385681293302, "grad_norm": 0.33772122859954834, "learning_rate": 8e-05, "loss": 1.6406, "step": 5365 }, { "epoch": 1.1803585175409657, "grad_norm": 0.3308549225330353, "learning_rate": 8e-05, "loss": 1.6118, "step": 5366 }, { "epoch": 1.180578466952601, "grad_norm": 0.32646870613098145, "learning_rate": 8e-05, "loss": 1.5866, "step": 5367 }, { "epoch": 1.1807984163642362, "grad_norm": 0.3572963774204254, "learning_rate": 8e-05, "loss": 1.5774, "step": 5368 }, { "epoch": 1.1810183657758715, "grad_norm": 0.32707181572914124, "learning_rate": 8e-05, "loss": 1.5438, "step": 5369 }, { "epoch": 1.1812383151875068, "grad_norm": 0.3376196026802063, "learning_rate": 8e-05, "loss": 1.6258, "step": 5370 }, { "epoch": 1.1814582645991423, "grad_norm": 0.3346554934978485, "learning_rate": 8e-05, "loss": 1.4286, "step": 5371 }, { "epoch": 1.1816782140107776, "grad_norm": 0.32654672861099243, "learning_rate": 8e-05, "loss": 1.4564, "step": 5372 }, { "epoch": 1.1818981634224128, "grad_norm": 0.3526204824447632, "learning_rate": 8e-05, "loss": 1.6333, "step": 5373 }, { "epoch": 1.182118112834048, "grad_norm": 0.3578186333179474, "learning_rate": 8e-05, "loss": 1.5559, "step": 5374 }, { "epoch": 1.1823380622456834, "grad_norm": 0.3193061053752899, "learning_rate": 8e-05, "loss": 1.5236, "step": 5375 }, { "epoch": 1.1825580116573189, "grad_norm": 0.3291642665863037, "learning_rate": 8e-05, "loss": 1.6211, "step": 5376 }, { "epoch": 1.1827779610689542, "grad_norm": 0.3349018096923828, "learning_rate": 8e-05, "loss": 1.554, "step": 5377 }, { "epoch": 1.1829979104805894, "grad_norm": 0.32746621966362, "learning_rate": 8e-05, "loss": 1.5785, "step": 5378 }, { "epoch": 1.1832178598922247, "grad_norm": 0.34521356225013733, "learning_rate": 8e-05, "loss": 1.5574, "step": 5379 }, { "epoch": 1.18343780930386, "grad_norm": 0.350311279296875, "learning_rate": 8e-05, "loss": 1.6907, "step": 5380 }, { "epoch": 1.1836577587154955, "grad_norm": 0.36038681864738464, "learning_rate": 8e-05, "loss": 1.6784, "step": 5381 }, { "epoch": 1.1838777081271308, "grad_norm": 0.3376932740211487, "learning_rate": 8e-05, "loss": 1.6443, "step": 5382 }, { "epoch": 1.184097657538766, "grad_norm": 0.3266044855117798, "learning_rate": 8e-05, "loss": 1.5875, "step": 5383 }, { "epoch": 1.1843176069504013, "grad_norm": 0.3391602635383606, "learning_rate": 8e-05, "loss": 1.5126, "step": 5384 }, { "epoch": 1.1845375563620366, "grad_norm": 0.3197435140609741, "learning_rate": 8e-05, "loss": 1.3945, "step": 5385 }, { "epoch": 1.1847575057736721, "grad_norm": 0.3665485978126526, "learning_rate": 8e-05, "loss": 1.7937, "step": 5386 }, { "epoch": 1.1849774551853074, "grad_norm": 0.32803964614868164, "learning_rate": 8e-05, "loss": 1.616, "step": 5387 }, { "epoch": 1.1851974045969427, "grad_norm": 0.3301391005516052, "learning_rate": 8e-05, "loss": 1.627, "step": 5388 }, { "epoch": 1.185417354008578, "grad_norm": 0.33791351318359375, "learning_rate": 8e-05, "loss": 1.4732, "step": 5389 }, { "epoch": 1.1856373034202132, "grad_norm": 0.30893924832344055, "learning_rate": 8e-05, "loss": 1.5131, "step": 5390 }, { "epoch": 1.1858572528318487, "grad_norm": 0.3352710008621216, "learning_rate": 8e-05, "loss": 1.5509, "step": 5391 }, { "epoch": 1.186077202243484, "grad_norm": 0.31556859612464905, "learning_rate": 8e-05, "loss": 1.4581, "step": 5392 }, { "epoch": 1.1862971516551193, "grad_norm": 0.3348480463027954, "learning_rate": 8e-05, "loss": 1.6183, "step": 5393 }, { "epoch": 1.1865171010667546, "grad_norm": 0.3428530693054199, "learning_rate": 8e-05, "loss": 1.6325, "step": 5394 }, { "epoch": 1.18673705047839, "grad_norm": 0.3312858045101166, "learning_rate": 8e-05, "loss": 1.7025, "step": 5395 }, { "epoch": 1.1869569998900253, "grad_norm": 0.3252064287662506, "learning_rate": 8e-05, "loss": 1.5835, "step": 5396 }, { "epoch": 1.1871769493016606, "grad_norm": 0.3331434428691864, "learning_rate": 8e-05, "loss": 1.4693, "step": 5397 }, { "epoch": 1.187396898713296, "grad_norm": 0.37382298707962036, "learning_rate": 8e-05, "loss": 1.6329, "step": 5398 }, { "epoch": 1.1876168481249312, "grad_norm": 0.36379164457321167, "learning_rate": 8e-05, "loss": 1.7061, "step": 5399 }, { "epoch": 1.1878367975365667, "grad_norm": 0.33243992924690247, "learning_rate": 8e-05, "loss": 1.5528, "step": 5400 }, { "epoch": 1.188056746948202, "grad_norm": 0.32361292839050293, "learning_rate": 8e-05, "loss": 1.5996, "step": 5401 }, { "epoch": 1.1882766963598372, "grad_norm": 0.344622403383255, "learning_rate": 8e-05, "loss": 1.5965, "step": 5402 }, { "epoch": 1.1884966457714725, "grad_norm": 0.3345627784729004, "learning_rate": 8e-05, "loss": 1.5684, "step": 5403 }, { "epoch": 1.1887165951831078, "grad_norm": 0.34564343094825745, "learning_rate": 8e-05, "loss": 1.6639, "step": 5404 }, { "epoch": 1.1889365445947433, "grad_norm": 0.3181012272834778, "learning_rate": 8e-05, "loss": 1.4918, "step": 5405 }, { "epoch": 1.1891564940063786, "grad_norm": 0.3159765601158142, "learning_rate": 8e-05, "loss": 1.4248, "step": 5406 }, { "epoch": 1.1893764434180139, "grad_norm": 0.3497673571109772, "learning_rate": 8e-05, "loss": 1.5633, "step": 5407 }, { "epoch": 1.1895963928296491, "grad_norm": 0.323435515165329, "learning_rate": 8e-05, "loss": 1.5633, "step": 5408 }, { "epoch": 1.1898163422412844, "grad_norm": 0.33802157640457153, "learning_rate": 8e-05, "loss": 1.5716, "step": 5409 }, { "epoch": 1.19003629165292, "grad_norm": 0.35690879821777344, "learning_rate": 8e-05, "loss": 1.612, "step": 5410 }, { "epoch": 1.1902562410645552, "grad_norm": 0.33218786120414734, "learning_rate": 8e-05, "loss": 1.5855, "step": 5411 }, { "epoch": 1.1904761904761905, "grad_norm": 0.3270936608314514, "learning_rate": 8e-05, "loss": 1.5826, "step": 5412 }, { "epoch": 1.1906961398878257, "grad_norm": 0.374396950006485, "learning_rate": 8e-05, "loss": 1.5081, "step": 5413 }, { "epoch": 1.1909160892994612, "grad_norm": 0.3247188627719879, "learning_rate": 8e-05, "loss": 1.6128, "step": 5414 }, { "epoch": 1.1911360387110965, "grad_norm": 0.33673202991485596, "learning_rate": 8e-05, "loss": 1.6072, "step": 5415 }, { "epoch": 1.1913559881227318, "grad_norm": 0.36152219772338867, "learning_rate": 8e-05, "loss": 1.6127, "step": 5416 }, { "epoch": 1.191575937534367, "grad_norm": 0.3505416810512543, "learning_rate": 8e-05, "loss": 1.6225, "step": 5417 }, { "epoch": 1.1917958869460024, "grad_norm": 0.33250531554222107, "learning_rate": 8e-05, "loss": 1.661, "step": 5418 }, { "epoch": 1.1920158363576379, "grad_norm": 0.33932605385780334, "learning_rate": 8e-05, "loss": 1.6953, "step": 5419 }, { "epoch": 1.1922357857692731, "grad_norm": 0.3214292526245117, "learning_rate": 8e-05, "loss": 1.5422, "step": 5420 }, { "epoch": 1.1924557351809084, "grad_norm": 0.33164653182029724, "learning_rate": 8e-05, "loss": 1.5462, "step": 5421 }, { "epoch": 1.1926756845925437, "grad_norm": 0.3508250117301941, "learning_rate": 8e-05, "loss": 1.5378, "step": 5422 }, { "epoch": 1.192895634004179, "grad_norm": 0.3195161521434784, "learning_rate": 8e-05, "loss": 1.523, "step": 5423 }, { "epoch": 1.1931155834158145, "grad_norm": 0.3141813576221466, "learning_rate": 8e-05, "loss": 1.5201, "step": 5424 }, { "epoch": 1.1933355328274498, "grad_norm": 0.3240199089050293, "learning_rate": 8e-05, "loss": 1.5772, "step": 5425 }, { "epoch": 1.193555482239085, "grad_norm": 0.3305763900279999, "learning_rate": 8e-05, "loss": 1.6038, "step": 5426 }, { "epoch": 1.1937754316507203, "grad_norm": 0.3213599920272827, "learning_rate": 8e-05, "loss": 1.6062, "step": 5427 }, { "epoch": 1.1939953810623556, "grad_norm": 0.34827089309692383, "learning_rate": 8e-05, "loss": 1.5976, "step": 5428 }, { "epoch": 1.194215330473991, "grad_norm": 0.3154602646827698, "learning_rate": 8e-05, "loss": 1.5333, "step": 5429 }, { "epoch": 1.1944352798856264, "grad_norm": 0.3336210250854492, "learning_rate": 8e-05, "loss": 1.3452, "step": 5430 }, { "epoch": 1.1946552292972616, "grad_norm": 0.3689917325973511, "learning_rate": 8e-05, "loss": 1.7322, "step": 5431 }, { "epoch": 1.194875178708897, "grad_norm": 0.3229157626628876, "learning_rate": 8e-05, "loss": 1.4399, "step": 5432 }, { "epoch": 1.1950951281205322, "grad_norm": 0.3320709764957428, "learning_rate": 8e-05, "loss": 1.4726, "step": 5433 }, { "epoch": 1.1953150775321677, "grad_norm": 0.32017216086387634, "learning_rate": 8e-05, "loss": 1.4565, "step": 5434 }, { "epoch": 1.195535026943803, "grad_norm": 0.3112264573574066, "learning_rate": 8e-05, "loss": 1.5286, "step": 5435 }, { "epoch": 1.1957549763554383, "grad_norm": 0.3197689354419708, "learning_rate": 8e-05, "loss": 1.5228, "step": 5436 }, { "epoch": 1.1959749257670735, "grad_norm": 0.31938067078590393, "learning_rate": 8e-05, "loss": 1.5009, "step": 5437 }, { "epoch": 1.1961948751787088, "grad_norm": 0.33675986528396606, "learning_rate": 8e-05, "loss": 1.4413, "step": 5438 }, { "epoch": 1.1964148245903443, "grad_norm": 0.3291521370410919, "learning_rate": 8e-05, "loss": 1.5733, "step": 5439 }, { "epoch": 1.1966347740019796, "grad_norm": 0.3322976529598236, "learning_rate": 8e-05, "loss": 1.6749, "step": 5440 }, { "epoch": 1.1968547234136149, "grad_norm": 0.3417357802391052, "learning_rate": 8e-05, "loss": 1.6567, "step": 5441 }, { "epoch": 1.1970746728252502, "grad_norm": 0.3179970383644104, "learning_rate": 8e-05, "loss": 1.5336, "step": 5442 }, { "epoch": 1.1972946222368854, "grad_norm": 0.36857733130455017, "learning_rate": 8e-05, "loss": 1.5369, "step": 5443 }, { "epoch": 1.197514571648521, "grad_norm": 0.3383985459804535, "learning_rate": 8e-05, "loss": 1.5702, "step": 5444 }, { "epoch": 1.1977345210601562, "grad_norm": 0.36165478825569153, "learning_rate": 8e-05, "loss": 1.4607, "step": 5445 }, { "epoch": 1.1979544704717915, "grad_norm": 0.32188406586647034, "learning_rate": 8e-05, "loss": 1.5355, "step": 5446 }, { "epoch": 1.1981744198834268, "grad_norm": 0.320975124835968, "learning_rate": 8e-05, "loss": 1.5207, "step": 5447 }, { "epoch": 1.198394369295062, "grad_norm": 0.33258509635925293, "learning_rate": 8e-05, "loss": 1.5584, "step": 5448 }, { "epoch": 1.1986143187066975, "grad_norm": 0.34548890590667725, "learning_rate": 8e-05, "loss": 1.6466, "step": 5449 }, { "epoch": 1.1988342681183328, "grad_norm": 0.3282851576805115, "learning_rate": 8e-05, "loss": 1.5185, "step": 5450 }, { "epoch": 1.199054217529968, "grad_norm": 0.3402653932571411, "learning_rate": 8e-05, "loss": 1.6335, "step": 5451 }, { "epoch": 1.1992741669416034, "grad_norm": 0.33617302775382996, "learning_rate": 8e-05, "loss": 1.5806, "step": 5452 }, { "epoch": 1.1994941163532387, "grad_norm": 0.33767545223236084, "learning_rate": 8e-05, "loss": 1.6904, "step": 5453 }, { "epoch": 1.1997140657648742, "grad_norm": 0.3474002480506897, "learning_rate": 8e-05, "loss": 1.5858, "step": 5454 }, { "epoch": 1.1999340151765094, "grad_norm": 0.3207785189151764, "learning_rate": 8e-05, "loss": 1.4694, "step": 5455 }, { "epoch": 1.2001539645881447, "grad_norm": 0.34237292408943176, "learning_rate": 8e-05, "loss": 1.611, "step": 5456 }, { "epoch": 1.20037391399978, "grad_norm": 0.32801660895347595, "learning_rate": 8e-05, "loss": 1.4732, "step": 5457 }, { "epoch": 1.2005938634114153, "grad_norm": 0.3178023397922516, "learning_rate": 8e-05, "loss": 1.3567, "step": 5458 }, { "epoch": 1.2008138128230508, "grad_norm": 0.3182491958141327, "learning_rate": 8e-05, "loss": 1.4328, "step": 5459 }, { "epoch": 1.201033762234686, "grad_norm": 0.33574482798576355, "learning_rate": 8e-05, "loss": 1.6948, "step": 5460 }, { "epoch": 1.2012537116463213, "grad_norm": 0.33535581827163696, "learning_rate": 8e-05, "loss": 1.5764, "step": 5461 }, { "epoch": 1.2014736610579566, "grad_norm": 0.3104236423969269, "learning_rate": 8e-05, "loss": 1.4468, "step": 5462 }, { "epoch": 1.2016936104695919, "grad_norm": 0.33877041935920715, "learning_rate": 8e-05, "loss": 1.715, "step": 5463 }, { "epoch": 1.2019135598812274, "grad_norm": 0.3108452260494232, "learning_rate": 8e-05, "loss": 1.5088, "step": 5464 }, { "epoch": 1.2021335092928627, "grad_norm": 0.3338935673236847, "learning_rate": 8e-05, "loss": 1.6693, "step": 5465 }, { "epoch": 1.202353458704498, "grad_norm": 0.32279300689697266, "learning_rate": 8e-05, "loss": 1.657, "step": 5466 }, { "epoch": 1.2025734081161332, "grad_norm": 0.3512457013130188, "learning_rate": 8e-05, "loss": 1.4445, "step": 5467 }, { "epoch": 1.2027933575277685, "grad_norm": 0.3267289102077484, "learning_rate": 8e-05, "loss": 1.6085, "step": 5468 }, { "epoch": 1.203013306939404, "grad_norm": 0.33776775002479553, "learning_rate": 8e-05, "loss": 1.4444, "step": 5469 }, { "epoch": 1.2032332563510393, "grad_norm": 0.35299593210220337, "learning_rate": 8e-05, "loss": 1.7164, "step": 5470 }, { "epoch": 1.2034532057626746, "grad_norm": 0.33513110876083374, "learning_rate": 8e-05, "loss": 1.5246, "step": 5471 }, { "epoch": 1.2036731551743098, "grad_norm": 0.34809640049934387, "learning_rate": 8e-05, "loss": 1.5435, "step": 5472 }, { "epoch": 1.2038931045859451, "grad_norm": 0.3289783298969269, "learning_rate": 8e-05, "loss": 1.6345, "step": 5473 }, { "epoch": 1.2041130539975806, "grad_norm": 0.33451980352401733, "learning_rate": 8e-05, "loss": 1.4346, "step": 5474 }, { "epoch": 1.204333003409216, "grad_norm": 0.34490063786506653, "learning_rate": 8e-05, "loss": 1.6235, "step": 5475 }, { "epoch": 1.2045529528208512, "grad_norm": 0.3607546091079712, "learning_rate": 8e-05, "loss": 1.6486, "step": 5476 }, { "epoch": 1.2047729022324865, "grad_norm": 0.4446365535259247, "learning_rate": 8e-05, "loss": 1.8015, "step": 5477 }, { "epoch": 1.2049928516441217, "grad_norm": 0.31937286257743835, "learning_rate": 8e-05, "loss": 1.5561, "step": 5478 }, { "epoch": 1.2052128010557572, "grad_norm": 0.37078163027763367, "learning_rate": 8e-05, "loss": 1.7454, "step": 5479 }, { "epoch": 1.2054327504673925, "grad_norm": 0.30821090936660767, "learning_rate": 8e-05, "loss": 1.4115, "step": 5480 }, { "epoch": 1.2056526998790278, "grad_norm": 0.32840245962142944, "learning_rate": 8e-05, "loss": 1.8099, "step": 5481 }, { "epoch": 1.205872649290663, "grad_norm": 0.351441353559494, "learning_rate": 8e-05, "loss": 1.527, "step": 5482 }, { "epoch": 1.2060925987022986, "grad_norm": 0.34991344809532166, "learning_rate": 8e-05, "loss": 1.5652, "step": 5483 }, { "epoch": 1.2063125481139338, "grad_norm": 0.32807400822639465, "learning_rate": 8e-05, "loss": 1.5689, "step": 5484 }, { "epoch": 1.2065324975255691, "grad_norm": 0.3555041253566742, "learning_rate": 8e-05, "loss": 1.5705, "step": 5485 }, { "epoch": 1.2067524469372044, "grad_norm": 0.33731478452682495, "learning_rate": 8e-05, "loss": 1.5962, "step": 5486 }, { "epoch": 1.2069723963488397, "grad_norm": 0.34947067499160767, "learning_rate": 8e-05, "loss": 1.6261, "step": 5487 }, { "epoch": 1.2071923457604752, "grad_norm": 0.34859874844551086, "learning_rate": 8e-05, "loss": 1.7607, "step": 5488 }, { "epoch": 1.2074122951721105, "grad_norm": 0.35389456152915955, "learning_rate": 8e-05, "loss": 1.7361, "step": 5489 }, { "epoch": 1.2076322445837457, "grad_norm": 0.30824750661849976, "learning_rate": 8e-05, "loss": 1.4678, "step": 5490 }, { "epoch": 1.207852193995381, "grad_norm": 0.3394816219806671, "learning_rate": 8e-05, "loss": 1.6456, "step": 5491 }, { "epoch": 1.2080721434070163, "grad_norm": 0.3217008113861084, "learning_rate": 8e-05, "loss": 1.5273, "step": 5492 }, { "epoch": 1.2082920928186518, "grad_norm": 0.3242091238498688, "learning_rate": 8e-05, "loss": 1.587, "step": 5493 }, { "epoch": 1.208512042230287, "grad_norm": 0.3438120484352112, "learning_rate": 8e-05, "loss": 1.675, "step": 5494 }, { "epoch": 1.2087319916419224, "grad_norm": 0.3167682886123657, "learning_rate": 8e-05, "loss": 1.6168, "step": 5495 }, { "epoch": 1.2089519410535576, "grad_norm": 0.32584360241889954, "learning_rate": 8e-05, "loss": 1.5965, "step": 5496 }, { "epoch": 1.209171890465193, "grad_norm": 0.3319784998893738, "learning_rate": 8e-05, "loss": 1.496, "step": 5497 }, { "epoch": 1.2093918398768284, "grad_norm": 0.33543023467063904, "learning_rate": 8e-05, "loss": 1.565, "step": 5498 }, { "epoch": 1.2096117892884637, "grad_norm": 0.34996581077575684, "learning_rate": 8e-05, "loss": 1.6822, "step": 5499 }, { "epoch": 1.209831738700099, "grad_norm": 0.3253658711910248, "learning_rate": 8e-05, "loss": 1.6138, "step": 5500 }, { "epoch": 1.2100516881117342, "grad_norm": 0.32803401350975037, "learning_rate": 8e-05, "loss": 1.5853, "step": 5501 }, { "epoch": 1.2102716375233697, "grad_norm": 0.35597774386405945, "learning_rate": 8e-05, "loss": 1.6006, "step": 5502 }, { "epoch": 1.210491586935005, "grad_norm": 0.3260975182056427, "learning_rate": 8e-05, "loss": 1.6433, "step": 5503 }, { "epoch": 1.2107115363466403, "grad_norm": 0.3374732732772827, "learning_rate": 8e-05, "loss": 1.6547, "step": 5504 }, { "epoch": 1.2109314857582756, "grad_norm": 0.3221224546432495, "learning_rate": 8e-05, "loss": 1.4919, "step": 5505 }, { "epoch": 1.2111514351699109, "grad_norm": 0.33132049441337585, "learning_rate": 8e-05, "loss": 1.4938, "step": 5506 }, { "epoch": 1.2113713845815464, "grad_norm": 0.3319770097732544, "learning_rate": 8e-05, "loss": 1.6204, "step": 5507 }, { "epoch": 1.2115913339931816, "grad_norm": 0.3579905331134796, "learning_rate": 8e-05, "loss": 1.7033, "step": 5508 }, { "epoch": 1.211811283404817, "grad_norm": 0.31753915548324585, "learning_rate": 8e-05, "loss": 1.4994, "step": 5509 }, { "epoch": 1.2120312328164522, "grad_norm": 0.3174467086791992, "learning_rate": 8e-05, "loss": 1.6744, "step": 5510 }, { "epoch": 1.2122511822280875, "grad_norm": 0.3449244797229767, "learning_rate": 8e-05, "loss": 1.6727, "step": 5511 }, { "epoch": 1.212471131639723, "grad_norm": 0.3286067247390747, "learning_rate": 8e-05, "loss": 1.5787, "step": 5512 }, { "epoch": 1.2126910810513583, "grad_norm": 0.3375146985054016, "learning_rate": 8e-05, "loss": 1.5847, "step": 5513 }, { "epoch": 1.2129110304629935, "grad_norm": 0.324686735868454, "learning_rate": 8e-05, "loss": 1.6279, "step": 5514 }, { "epoch": 1.2131309798746288, "grad_norm": 0.3295452296733856, "learning_rate": 8e-05, "loss": 1.4627, "step": 5515 }, { "epoch": 1.213350929286264, "grad_norm": 0.3450112044811249, "learning_rate": 8e-05, "loss": 1.7488, "step": 5516 }, { "epoch": 1.2135708786978996, "grad_norm": 0.31478506326675415, "learning_rate": 8e-05, "loss": 1.5387, "step": 5517 }, { "epoch": 1.2137908281095349, "grad_norm": 0.35676828026771545, "learning_rate": 8e-05, "loss": 1.5927, "step": 5518 }, { "epoch": 1.2140107775211701, "grad_norm": 0.33322224020957947, "learning_rate": 8e-05, "loss": 1.7109, "step": 5519 }, { "epoch": 1.2142307269328054, "grad_norm": 0.3144984245300293, "learning_rate": 8e-05, "loss": 1.484, "step": 5520 }, { "epoch": 1.2144506763444407, "grad_norm": 0.3174855709075928, "learning_rate": 8e-05, "loss": 1.5161, "step": 5521 }, { "epoch": 1.2146706257560762, "grad_norm": 0.3244873881340027, "learning_rate": 8e-05, "loss": 1.6034, "step": 5522 }, { "epoch": 1.2148905751677115, "grad_norm": 0.3396502733230591, "learning_rate": 8e-05, "loss": 1.7516, "step": 5523 }, { "epoch": 1.2151105245793468, "grad_norm": 0.31188029050827026, "learning_rate": 8e-05, "loss": 1.4417, "step": 5524 }, { "epoch": 1.215330473990982, "grad_norm": 0.32375115156173706, "learning_rate": 8e-05, "loss": 1.4215, "step": 5525 }, { "epoch": 1.2155504234026173, "grad_norm": 0.3344710171222687, "learning_rate": 8e-05, "loss": 1.4482, "step": 5526 }, { "epoch": 1.2157703728142528, "grad_norm": 0.3282887041568756, "learning_rate": 8e-05, "loss": 1.6539, "step": 5527 }, { "epoch": 1.215990322225888, "grad_norm": 0.3495618999004364, "learning_rate": 8e-05, "loss": 1.747, "step": 5528 }, { "epoch": 1.2162102716375234, "grad_norm": 0.33851638436317444, "learning_rate": 8e-05, "loss": 1.7059, "step": 5529 }, { "epoch": 1.2164302210491587, "grad_norm": 0.3228831887245178, "learning_rate": 8e-05, "loss": 1.545, "step": 5530 }, { "epoch": 1.216650170460794, "grad_norm": 0.333201140165329, "learning_rate": 8e-05, "loss": 1.5199, "step": 5531 }, { "epoch": 1.2168701198724294, "grad_norm": 0.33130356669425964, "learning_rate": 8e-05, "loss": 1.6003, "step": 5532 }, { "epoch": 1.2170900692840647, "grad_norm": 0.33794403076171875, "learning_rate": 8e-05, "loss": 1.7324, "step": 5533 }, { "epoch": 1.2173100186957, "grad_norm": 0.34786710143089294, "learning_rate": 8e-05, "loss": 1.6922, "step": 5534 }, { "epoch": 1.2175299681073353, "grad_norm": 0.33372312784194946, "learning_rate": 8e-05, "loss": 1.5489, "step": 5535 }, { "epoch": 1.2177499175189705, "grad_norm": 0.3203335106372833, "learning_rate": 8e-05, "loss": 1.4409, "step": 5536 }, { "epoch": 1.217969866930606, "grad_norm": 0.33578673005104065, "learning_rate": 8e-05, "loss": 1.5322, "step": 5537 }, { "epoch": 1.2181898163422413, "grad_norm": 0.36111313104629517, "learning_rate": 8e-05, "loss": 1.6268, "step": 5538 }, { "epoch": 1.2184097657538766, "grad_norm": 0.3561267554759979, "learning_rate": 8e-05, "loss": 1.638, "step": 5539 }, { "epoch": 1.2186297151655119, "grad_norm": 0.32917550206184387, "learning_rate": 8e-05, "loss": 1.5741, "step": 5540 }, { "epoch": 1.2188496645771472, "grad_norm": 0.3381870687007904, "learning_rate": 8e-05, "loss": 1.6491, "step": 5541 }, { "epoch": 1.2190696139887827, "grad_norm": 0.31703197956085205, "learning_rate": 8e-05, "loss": 1.3953, "step": 5542 }, { "epoch": 1.219289563400418, "grad_norm": 0.32328978180885315, "learning_rate": 8e-05, "loss": 1.5567, "step": 5543 }, { "epoch": 1.2195095128120532, "grad_norm": 0.33085402846336365, "learning_rate": 8e-05, "loss": 1.601, "step": 5544 }, { "epoch": 1.2197294622236885, "grad_norm": 0.40923452377319336, "learning_rate": 8e-05, "loss": 1.5919, "step": 5545 }, { "epoch": 1.2199494116353238, "grad_norm": 0.3201553225517273, "learning_rate": 8e-05, "loss": 1.4663, "step": 5546 }, { "epoch": 1.2201693610469593, "grad_norm": 0.3876069188117981, "learning_rate": 8e-05, "loss": 1.5674, "step": 5547 }, { "epoch": 1.2203893104585946, "grad_norm": 0.3367742896080017, "learning_rate": 8e-05, "loss": 1.5871, "step": 5548 }, { "epoch": 1.2206092598702298, "grad_norm": 0.3308364152908325, "learning_rate": 8e-05, "loss": 1.5009, "step": 5549 }, { "epoch": 1.2208292092818651, "grad_norm": 0.33701252937316895, "learning_rate": 8e-05, "loss": 1.6998, "step": 5550 }, { "epoch": 1.2210491586935004, "grad_norm": 0.34394705295562744, "learning_rate": 8e-05, "loss": 1.6734, "step": 5551 }, { "epoch": 1.221269108105136, "grad_norm": 0.34140682220458984, "learning_rate": 8e-05, "loss": 1.6425, "step": 5552 }, { "epoch": 1.2214890575167712, "grad_norm": 0.34208136796951294, "learning_rate": 8e-05, "loss": 1.6753, "step": 5553 }, { "epoch": 1.2217090069284064, "grad_norm": 0.3169998526573181, "learning_rate": 8e-05, "loss": 1.5569, "step": 5554 }, { "epoch": 1.2219289563400417, "grad_norm": 0.3218975067138672, "learning_rate": 8e-05, "loss": 1.5679, "step": 5555 }, { "epoch": 1.222148905751677, "grad_norm": 0.3389599919319153, "learning_rate": 8e-05, "loss": 1.6421, "step": 5556 }, { "epoch": 1.2223688551633125, "grad_norm": 0.33677709102630615, "learning_rate": 8e-05, "loss": 1.6299, "step": 5557 }, { "epoch": 1.2225888045749478, "grad_norm": 0.3428082764148712, "learning_rate": 8e-05, "loss": 1.5582, "step": 5558 }, { "epoch": 1.222808753986583, "grad_norm": 0.3308615982532501, "learning_rate": 8e-05, "loss": 1.4824, "step": 5559 }, { "epoch": 1.2230287033982183, "grad_norm": 0.35097357630729675, "learning_rate": 8e-05, "loss": 1.6181, "step": 5560 }, { "epoch": 1.2232486528098536, "grad_norm": 0.3451812267303467, "learning_rate": 8e-05, "loss": 1.4872, "step": 5561 }, { "epoch": 1.2234686022214891, "grad_norm": 0.32337817549705505, "learning_rate": 8e-05, "loss": 1.5148, "step": 5562 }, { "epoch": 1.2236885516331244, "grad_norm": 0.3391774296760559, "learning_rate": 8e-05, "loss": 1.5453, "step": 5563 }, { "epoch": 1.2239085010447597, "grad_norm": 0.3410683572292328, "learning_rate": 8e-05, "loss": 1.6294, "step": 5564 }, { "epoch": 1.224128450456395, "grad_norm": 0.35200121998786926, "learning_rate": 8e-05, "loss": 1.5821, "step": 5565 }, { "epoch": 1.2243483998680302, "grad_norm": 0.37338927388191223, "learning_rate": 8e-05, "loss": 1.7834, "step": 5566 }, { "epoch": 1.2245683492796657, "grad_norm": 0.3418629765510559, "learning_rate": 8e-05, "loss": 1.6785, "step": 5567 }, { "epoch": 1.224788298691301, "grad_norm": 0.3559189438819885, "learning_rate": 8e-05, "loss": 1.6214, "step": 5568 }, { "epoch": 1.2250082481029363, "grad_norm": 0.3109119236469269, "learning_rate": 8e-05, "loss": 1.5801, "step": 5569 }, { "epoch": 1.2252281975145716, "grad_norm": 0.34502172470092773, "learning_rate": 8e-05, "loss": 1.8178, "step": 5570 }, { "epoch": 1.2254481469262068, "grad_norm": 0.3402266502380371, "learning_rate": 8e-05, "loss": 1.6485, "step": 5571 }, { "epoch": 1.2256680963378423, "grad_norm": 0.32304519414901733, "learning_rate": 8e-05, "loss": 1.4978, "step": 5572 }, { "epoch": 1.2258880457494776, "grad_norm": 0.3267800509929657, "learning_rate": 8e-05, "loss": 1.6852, "step": 5573 }, { "epoch": 1.226107995161113, "grad_norm": 0.32492130994796753, "learning_rate": 8e-05, "loss": 1.5453, "step": 5574 }, { "epoch": 1.2263279445727482, "grad_norm": 0.3426954746246338, "learning_rate": 8e-05, "loss": 1.6208, "step": 5575 }, { "epoch": 1.2265478939843837, "grad_norm": 0.33102431893348694, "learning_rate": 8e-05, "loss": 1.57, "step": 5576 }, { "epoch": 1.226767843396019, "grad_norm": 0.339023232460022, "learning_rate": 8e-05, "loss": 1.5593, "step": 5577 }, { "epoch": 1.2269877928076542, "grad_norm": 0.33789271116256714, "learning_rate": 8e-05, "loss": 1.637, "step": 5578 }, { "epoch": 1.2272077422192895, "grad_norm": 0.35140010714530945, "learning_rate": 8e-05, "loss": 1.7328, "step": 5579 }, { "epoch": 1.2274276916309248, "grad_norm": 0.33062443137168884, "learning_rate": 8e-05, "loss": 1.4441, "step": 5580 }, { "epoch": 1.2276476410425603, "grad_norm": 0.34730130434036255, "learning_rate": 8e-05, "loss": 1.7119, "step": 5581 }, { "epoch": 1.2278675904541956, "grad_norm": 0.3437526226043701, "learning_rate": 8e-05, "loss": 1.523, "step": 5582 }, { "epoch": 1.2280875398658309, "grad_norm": 0.3205513060092926, "learning_rate": 8e-05, "loss": 1.4009, "step": 5583 }, { "epoch": 1.2283074892774661, "grad_norm": 0.31829169392585754, "learning_rate": 8e-05, "loss": 1.4912, "step": 5584 }, { "epoch": 1.2285274386891014, "grad_norm": 0.353466272354126, "learning_rate": 8e-05, "loss": 1.6339, "step": 5585 }, { "epoch": 1.228747388100737, "grad_norm": 0.3424278795719147, "learning_rate": 8e-05, "loss": 1.652, "step": 5586 }, { "epoch": 1.2289673375123722, "grad_norm": 0.3184290826320648, "learning_rate": 8e-05, "loss": 1.5972, "step": 5587 }, { "epoch": 1.2291872869240075, "grad_norm": 0.33908766508102417, "learning_rate": 8e-05, "loss": 1.6518, "step": 5588 }, { "epoch": 1.2294072363356428, "grad_norm": 0.3335396647453308, "learning_rate": 8e-05, "loss": 1.6273, "step": 5589 }, { "epoch": 1.2296271857472783, "grad_norm": 0.328533798456192, "learning_rate": 8e-05, "loss": 1.6086, "step": 5590 }, { "epoch": 1.2298471351589135, "grad_norm": 0.37026166915893555, "learning_rate": 8e-05, "loss": 1.7016, "step": 5591 }, { "epoch": 1.2300670845705488, "grad_norm": 0.3394378125667572, "learning_rate": 8e-05, "loss": 1.5722, "step": 5592 }, { "epoch": 1.230287033982184, "grad_norm": 0.349181592464447, "learning_rate": 8e-05, "loss": 1.5593, "step": 5593 }, { "epoch": 1.2305069833938194, "grad_norm": 0.33427590131759644, "learning_rate": 8e-05, "loss": 1.6204, "step": 5594 }, { "epoch": 1.2307269328054549, "grad_norm": 0.34059473872184753, "learning_rate": 8e-05, "loss": 1.6989, "step": 5595 }, { "epoch": 1.2309468822170901, "grad_norm": 0.3326999247074127, "learning_rate": 8e-05, "loss": 1.5525, "step": 5596 }, { "epoch": 1.2311668316287254, "grad_norm": 0.3436473309993744, "learning_rate": 8e-05, "loss": 1.6468, "step": 5597 }, { "epoch": 1.2313867810403607, "grad_norm": 0.3459080755710602, "learning_rate": 8e-05, "loss": 1.7944, "step": 5598 }, { "epoch": 1.231606730451996, "grad_norm": 0.33405137062072754, "learning_rate": 8e-05, "loss": 1.6136, "step": 5599 }, { "epoch": 1.2318266798636315, "grad_norm": 0.35209086537361145, "learning_rate": 8e-05, "loss": 1.5015, "step": 5600 }, { "epoch": 1.2320466292752668, "grad_norm": 0.31778302788734436, "learning_rate": 8e-05, "loss": 1.5567, "step": 5601 }, { "epoch": 1.232266578686902, "grad_norm": 0.32216718792915344, "learning_rate": 8e-05, "loss": 1.5686, "step": 5602 }, { "epoch": 1.2324865280985373, "grad_norm": 0.340175986289978, "learning_rate": 8e-05, "loss": 1.6784, "step": 5603 }, { "epoch": 1.2327064775101726, "grad_norm": 0.3142259120941162, "learning_rate": 8e-05, "loss": 1.5699, "step": 5604 }, { "epoch": 1.232926426921808, "grad_norm": 0.3377830684185028, "learning_rate": 8e-05, "loss": 1.6721, "step": 5605 }, { "epoch": 1.2331463763334434, "grad_norm": 0.3371240198612213, "learning_rate": 8e-05, "loss": 1.53, "step": 5606 }, { "epoch": 1.2333663257450787, "grad_norm": 0.31657126545906067, "learning_rate": 8e-05, "loss": 1.458, "step": 5607 }, { "epoch": 1.233586275156714, "grad_norm": 0.3316138982772827, "learning_rate": 8e-05, "loss": 1.6352, "step": 5608 }, { "epoch": 1.2338062245683492, "grad_norm": 0.3483010530471802, "learning_rate": 8e-05, "loss": 1.6709, "step": 5609 }, { "epoch": 1.2340261739799847, "grad_norm": 0.32844796776771545, "learning_rate": 8e-05, "loss": 1.6613, "step": 5610 }, { "epoch": 1.23424612339162, "grad_norm": 0.3297501802444458, "learning_rate": 8e-05, "loss": 1.5264, "step": 5611 }, { "epoch": 1.2344660728032553, "grad_norm": 0.32772335410118103, "learning_rate": 8e-05, "loss": 1.6613, "step": 5612 }, { "epoch": 1.2346860222148905, "grad_norm": 0.37067434191703796, "learning_rate": 8e-05, "loss": 1.5765, "step": 5613 }, { "epoch": 1.2349059716265258, "grad_norm": 0.364818274974823, "learning_rate": 8e-05, "loss": 1.5856, "step": 5614 }, { "epoch": 1.2351259210381613, "grad_norm": 0.35764431953430176, "learning_rate": 8e-05, "loss": 1.5435, "step": 5615 }, { "epoch": 1.2353458704497966, "grad_norm": 0.34844768047332764, "learning_rate": 8e-05, "loss": 1.5523, "step": 5616 }, { "epoch": 1.2355658198614319, "grad_norm": 0.3421417772769928, "learning_rate": 8e-05, "loss": 1.681, "step": 5617 }, { "epoch": 1.2357857692730672, "grad_norm": 0.3259740471839905, "learning_rate": 8e-05, "loss": 1.5788, "step": 5618 }, { "epoch": 1.2360057186847024, "grad_norm": 0.3667384684085846, "learning_rate": 8e-05, "loss": 1.5618, "step": 5619 }, { "epoch": 1.236225668096338, "grad_norm": 0.3187921345233917, "learning_rate": 8e-05, "loss": 1.5945, "step": 5620 }, { "epoch": 1.2364456175079732, "grad_norm": 0.3462565541267395, "learning_rate": 8e-05, "loss": 1.4983, "step": 5621 }, { "epoch": 1.2366655669196085, "grad_norm": 0.35779890418052673, "learning_rate": 8e-05, "loss": 1.5833, "step": 5622 }, { "epoch": 1.2368855163312438, "grad_norm": 0.333268404006958, "learning_rate": 8e-05, "loss": 1.5756, "step": 5623 }, { "epoch": 1.237105465742879, "grad_norm": 0.3084180951118469, "learning_rate": 8e-05, "loss": 1.4466, "step": 5624 }, { "epoch": 1.2373254151545146, "grad_norm": 0.3362586498260498, "learning_rate": 8e-05, "loss": 1.5317, "step": 5625 }, { "epoch": 1.2375453645661498, "grad_norm": 0.3451949954032898, "learning_rate": 8e-05, "loss": 1.6973, "step": 5626 }, { "epoch": 1.237765313977785, "grad_norm": 0.33187639713287354, "learning_rate": 8e-05, "loss": 1.5974, "step": 5627 }, { "epoch": 1.2379852633894204, "grad_norm": 0.3374370336532593, "learning_rate": 8e-05, "loss": 1.6378, "step": 5628 }, { "epoch": 1.2382052128010557, "grad_norm": 0.32258909940719604, "learning_rate": 8e-05, "loss": 1.4557, "step": 5629 }, { "epoch": 1.2384251622126912, "grad_norm": 0.3268379271030426, "learning_rate": 8e-05, "loss": 1.4497, "step": 5630 }, { "epoch": 1.2386451116243264, "grad_norm": 0.3302018940448761, "learning_rate": 8e-05, "loss": 1.5246, "step": 5631 }, { "epoch": 1.2388650610359617, "grad_norm": 0.3260684907436371, "learning_rate": 8e-05, "loss": 1.5281, "step": 5632 }, { "epoch": 1.239085010447597, "grad_norm": 0.33316999673843384, "learning_rate": 8e-05, "loss": 1.6204, "step": 5633 }, { "epoch": 1.2393049598592323, "grad_norm": 0.3413833677768707, "learning_rate": 8e-05, "loss": 1.7486, "step": 5634 }, { "epoch": 1.2395249092708678, "grad_norm": 0.3791937828063965, "learning_rate": 8e-05, "loss": 1.6775, "step": 5635 }, { "epoch": 1.239744858682503, "grad_norm": 0.3169299066066742, "learning_rate": 8e-05, "loss": 1.5256, "step": 5636 }, { "epoch": 1.2399648080941383, "grad_norm": 0.33628177642822266, "learning_rate": 8e-05, "loss": 1.6928, "step": 5637 }, { "epoch": 1.2401847575057736, "grad_norm": 0.3233479857444763, "learning_rate": 8e-05, "loss": 1.5027, "step": 5638 }, { "epoch": 1.240404706917409, "grad_norm": 0.33468711376190186, "learning_rate": 8e-05, "loss": 1.5631, "step": 5639 }, { "epoch": 1.2406246563290444, "grad_norm": 0.3242972195148468, "learning_rate": 8e-05, "loss": 1.5068, "step": 5640 }, { "epoch": 1.2408446057406797, "grad_norm": 0.34110313653945923, "learning_rate": 8e-05, "loss": 1.5646, "step": 5641 }, { "epoch": 1.241064555152315, "grad_norm": 0.3252404034137726, "learning_rate": 8e-05, "loss": 1.6072, "step": 5642 }, { "epoch": 1.2412845045639502, "grad_norm": 0.33433806896209717, "learning_rate": 8e-05, "loss": 1.5981, "step": 5643 }, { "epoch": 1.2415044539755855, "grad_norm": 0.35784971714019775, "learning_rate": 8e-05, "loss": 1.6078, "step": 5644 }, { "epoch": 1.241724403387221, "grad_norm": 0.34111908078193665, "learning_rate": 8e-05, "loss": 1.6587, "step": 5645 }, { "epoch": 1.2419443527988563, "grad_norm": 0.32142379879951477, "learning_rate": 8e-05, "loss": 1.5945, "step": 5646 }, { "epoch": 1.2421643022104916, "grad_norm": 0.35389086604118347, "learning_rate": 8e-05, "loss": 1.8542, "step": 5647 }, { "epoch": 1.2423842516221268, "grad_norm": 0.3178441524505615, "learning_rate": 8e-05, "loss": 1.5862, "step": 5648 }, { "epoch": 1.2426042010337621, "grad_norm": 0.3342335522174835, "learning_rate": 8e-05, "loss": 1.6214, "step": 5649 }, { "epoch": 1.2428241504453976, "grad_norm": 0.33119621872901917, "learning_rate": 8e-05, "loss": 1.6162, "step": 5650 }, { "epoch": 1.243044099857033, "grad_norm": 0.3190183937549591, "learning_rate": 8e-05, "loss": 1.4455, "step": 5651 }, { "epoch": 1.2432640492686682, "grad_norm": 0.3339254558086395, "learning_rate": 8e-05, "loss": 1.6084, "step": 5652 }, { "epoch": 1.2434839986803035, "grad_norm": 0.3202483355998993, "learning_rate": 8e-05, "loss": 1.4814, "step": 5653 }, { "epoch": 1.2437039480919387, "grad_norm": 0.33712533116340637, "learning_rate": 8e-05, "loss": 1.5822, "step": 5654 }, { "epoch": 1.2439238975035742, "grad_norm": 0.3227181136608124, "learning_rate": 8e-05, "loss": 1.562, "step": 5655 }, { "epoch": 1.2441438469152095, "grad_norm": 0.3361814618110657, "learning_rate": 8e-05, "loss": 1.5438, "step": 5656 }, { "epoch": 1.2443637963268448, "grad_norm": 0.3519279658794403, "learning_rate": 8e-05, "loss": 1.5037, "step": 5657 }, { "epoch": 1.24458374573848, "grad_norm": 0.3199707865715027, "learning_rate": 8e-05, "loss": 1.4589, "step": 5658 }, { "epoch": 1.2448036951501154, "grad_norm": 0.32631179690361023, "learning_rate": 8e-05, "loss": 1.5006, "step": 5659 }, { "epoch": 1.2450236445617509, "grad_norm": 0.39477720856666565, "learning_rate": 8e-05, "loss": 1.5927, "step": 5660 }, { "epoch": 1.2452435939733861, "grad_norm": 0.3386424481868744, "learning_rate": 8e-05, "loss": 1.5121, "step": 5661 }, { "epoch": 1.2454635433850214, "grad_norm": 0.3382517695426941, "learning_rate": 8e-05, "loss": 1.6616, "step": 5662 }, { "epoch": 1.2456834927966567, "grad_norm": 0.3430521488189697, "learning_rate": 8e-05, "loss": 1.6861, "step": 5663 }, { "epoch": 1.2459034422082922, "grad_norm": 0.3269794285297394, "learning_rate": 8e-05, "loss": 1.5341, "step": 5664 }, { "epoch": 1.2461233916199275, "grad_norm": 0.35102149844169617, "learning_rate": 8e-05, "loss": 1.7615, "step": 5665 }, { "epoch": 1.2463433410315627, "grad_norm": 0.3551831841468811, "learning_rate": 8e-05, "loss": 1.6246, "step": 5666 }, { "epoch": 1.246563290443198, "grad_norm": 0.34092190861701965, "learning_rate": 8e-05, "loss": 1.6821, "step": 5667 }, { "epoch": 1.2467832398548333, "grad_norm": 0.3749929368495941, "learning_rate": 8e-05, "loss": 1.6069, "step": 5668 }, { "epoch": 1.2470031892664688, "grad_norm": 0.35616886615753174, "learning_rate": 8e-05, "loss": 1.5552, "step": 5669 }, { "epoch": 1.247223138678104, "grad_norm": 0.3592993915081024, "learning_rate": 8e-05, "loss": 1.4655, "step": 5670 }, { "epoch": 1.2474430880897394, "grad_norm": 0.3468477725982666, "learning_rate": 8e-05, "loss": 1.6737, "step": 5671 }, { "epoch": 1.2476630375013746, "grad_norm": 0.3244820535182953, "learning_rate": 8e-05, "loss": 1.5628, "step": 5672 }, { "epoch": 1.24788298691301, "grad_norm": 0.3261267840862274, "learning_rate": 8e-05, "loss": 1.5993, "step": 5673 }, { "epoch": 1.2481029363246454, "grad_norm": 0.33319368958473206, "learning_rate": 8e-05, "loss": 1.5029, "step": 5674 }, { "epoch": 1.2483228857362807, "grad_norm": 0.33900022506713867, "learning_rate": 8e-05, "loss": 1.6287, "step": 5675 }, { "epoch": 1.248542835147916, "grad_norm": 0.3529195785522461, "learning_rate": 8e-05, "loss": 1.654, "step": 5676 }, { "epoch": 1.2487627845595513, "grad_norm": 0.3262398838996887, "learning_rate": 8e-05, "loss": 1.51, "step": 5677 }, { "epoch": 1.2489827339711865, "grad_norm": 0.3486205041408539, "learning_rate": 8e-05, "loss": 1.5699, "step": 5678 }, { "epoch": 1.249202683382822, "grad_norm": 0.33117419481277466, "learning_rate": 8e-05, "loss": 1.5188, "step": 5679 }, { "epoch": 1.2494226327944573, "grad_norm": 0.33656859397888184, "learning_rate": 8e-05, "loss": 1.4485, "step": 5680 }, { "epoch": 1.2496425822060926, "grad_norm": 0.33655208349227905, "learning_rate": 8e-05, "loss": 1.6083, "step": 5681 }, { "epoch": 1.2498625316177279, "grad_norm": 0.32219189405441284, "learning_rate": 8e-05, "loss": 1.4316, "step": 5682 }, { "epoch": 1.2500824810293634, "grad_norm": 0.33567866683006287, "learning_rate": 8e-05, "loss": 1.6642, "step": 5683 }, { "epoch": 1.2503024304409986, "grad_norm": 0.31820785999298096, "learning_rate": 8e-05, "loss": 1.506, "step": 5684 }, { "epoch": 1.250522379852634, "grad_norm": 0.335056334733963, "learning_rate": 8e-05, "loss": 1.5295, "step": 5685 }, { "epoch": 1.2507423292642692, "grad_norm": 0.32729852199554443, "learning_rate": 8e-05, "loss": 1.601, "step": 5686 }, { "epoch": 1.2509622786759045, "grad_norm": 0.3292050361633301, "learning_rate": 8e-05, "loss": 1.5416, "step": 5687 }, { "epoch": 1.25118222808754, "grad_norm": 0.31196436285972595, "learning_rate": 8e-05, "loss": 1.5215, "step": 5688 }, { "epoch": 1.2514021774991753, "grad_norm": 0.3736610412597656, "learning_rate": 8e-05, "loss": 1.5142, "step": 5689 }, { "epoch": 1.2516221269108105, "grad_norm": 0.33862683176994324, "learning_rate": 8e-05, "loss": 1.5962, "step": 5690 }, { "epoch": 1.2518420763224458, "grad_norm": 0.34158089756965637, "learning_rate": 8e-05, "loss": 1.5585, "step": 5691 }, { "epoch": 1.252062025734081, "grad_norm": 0.3352968096733093, "learning_rate": 8e-05, "loss": 1.5988, "step": 5692 }, { "epoch": 1.2522819751457166, "grad_norm": 0.34869134426116943, "learning_rate": 8e-05, "loss": 1.6372, "step": 5693 }, { "epoch": 1.2525019245573519, "grad_norm": 0.34786170721054077, "learning_rate": 8e-05, "loss": 1.7258, "step": 5694 }, { "epoch": 1.2527218739689872, "grad_norm": 0.3308630585670471, "learning_rate": 8e-05, "loss": 1.462, "step": 5695 }, { "epoch": 1.2529418233806224, "grad_norm": 0.318513423204422, "learning_rate": 8e-05, "loss": 1.3967, "step": 5696 }, { "epoch": 1.2531617727922577, "grad_norm": 0.3599640130996704, "learning_rate": 8e-05, "loss": 1.7616, "step": 5697 }, { "epoch": 1.2533817222038932, "grad_norm": 0.3447583019733429, "learning_rate": 8e-05, "loss": 1.6415, "step": 5698 }, { "epoch": 1.2536016716155285, "grad_norm": 0.3410969376564026, "learning_rate": 8e-05, "loss": 1.689, "step": 5699 }, { "epoch": 1.2538216210271638, "grad_norm": 0.3068464994430542, "learning_rate": 8e-05, "loss": 1.536, "step": 5700 }, { "epoch": 1.254041570438799, "grad_norm": 0.33735471963882446, "learning_rate": 8e-05, "loss": 1.6115, "step": 5701 }, { "epoch": 1.2542615198504343, "grad_norm": 0.32334697246551514, "learning_rate": 8e-05, "loss": 1.5469, "step": 5702 }, { "epoch": 1.2544814692620698, "grad_norm": 0.33375993371009827, "learning_rate": 8e-05, "loss": 1.5233, "step": 5703 }, { "epoch": 1.254701418673705, "grad_norm": 0.33172744512557983, "learning_rate": 8e-05, "loss": 1.7076, "step": 5704 }, { "epoch": 1.2549213680853404, "grad_norm": 0.37090012431144714, "learning_rate": 8e-05, "loss": 1.6561, "step": 5705 }, { "epoch": 1.2551413174969757, "grad_norm": 0.34135085344314575, "learning_rate": 8e-05, "loss": 1.6791, "step": 5706 }, { "epoch": 1.255361266908611, "grad_norm": 0.34571292996406555, "learning_rate": 8e-05, "loss": 1.6573, "step": 5707 }, { "epoch": 1.2555812163202464, "grad_norm": 0.3257502615451813, "learning_rate": 8e-05, "loss": 1.5814, "step": 5708 }, { "epoch": 1.2558011657318817, "grad_norm": 0.34923064708709717, "learning_rate": 8e-05, "loss": 1.6963, "step": 5709 }, { "epoch": 1.256021115143517, "grad_norm": 0.36867716908454895, "learning_rate": 8e-05, "loss": 1.7231, "step": 5710 }, { "epoch": 1.2562410645551523, "grad_norm": 0.3395102918148041, "learning_rate": 8e-05, "loss": 1.5523, "step": 5711 }, { "epoch": 1.2564610139667876, "grad_norm": 0.3299650549888611, "learning_rate": 8e-05, "loss": 1.6502, "step": 5712 }, { "epoch": 1.256680963378423, "grad_norm": 0.3328086733818054, "learning_rate": 8e-05, "loss": 1.6282, "step": 5713 }, { "epoch": 1.2569009127900583, "grad_norm": 0.33379966020584106, "learning_rate": 8e-05, "loss": 1.5678, "step": 5714 }, { "epoch": 1.2571208622016936, "grad_norm": 0.3252149224281311, "learning_rate": 8e-05, "loss": 1.5963, "step": 5715 }, { "epoch": 1.257340811613329, "grad_norm": 0.34433087706565857, "learning_rate": 8e-05, "loss": 1.6991, "step": 5716 }, { "epoch": 1.2575607610249642, "grad_norm": 0.3063054084777832, "learning_rate": 8e-05, "loss": 1.4946, "step": 5717 }, { "epoch": 1.2577807104365997, "grad_norm": 0.3114505112171173, "learning_rate": 8e-05, "loss": 1.2805, "step": 5718 }, { "epoch": 1.258000659848235, "grad_norm": 0.3430366814136505, "learning_rate": 8e-05, "loss": 1.67, "step": 5719 }, { "epoch": 1.2582206092598702, "grad_norm": 0.32265177369117737, "learning_rate": 8e-05, "loss": 1.519, "step": 5720 }, { "epoch": 1.2584405586715055, "grad_norm": 0.3417562246322632, "learning_rate": 8e-05, "loss": 1.7289, "step": 5721 }, { "epoch": 1.2586605080831408, "grad_norm": 0.32057350873947144, "learning_rate": 8e-05, "loss": 1.5453, "step": 5722 }, { "epoch": 1.2588804574947763, "grad_norm": 0.3304172456264496, "learning_rate": 8e-05, "loss": 1.6544, "step": 5723 }, { "epoch": 1.2591004069064116, "grad_norm": 0.3300001621246338, "learning_rate": 8e-05, "loss": 1.6171, "step": 5724 }, { "epoch": 1.2593203563180468, "grad_norm": 0.3529702425003052, "learning_rate": 8e-05, "loss": 1.4927, "step": 5725 }, { "epoch": 1.2595403057296821, "grad_norm": 0.3493426442146301, "learning_rate": 8e-05, "loss": 1.6292, "step": 5726 }, { "epoch": 1.2597602551413174, "grad_norm": 0.3253966271877289, "learning_rate": 8e-05, "loss": 1.4913, "step": 5727 }, { "epoch": 1.259980204552953, "grad_norm": 0.32938602566719055, "learning_rate": 8e-05, "loss": 1.6371, "step": 5728 }, { "epoch": 1.2602001539645882, "grad_norm": 0.33309897780418396, "learning_rate": 8e-05, "loss": 1.6479, "step": 5729 }, { "epoch": 1.2604201033762235, "grad_norm": 0.3398410677909851, "learning_rate": 8e-05, "loss": 1.5993, "step": 5730 }, { "epoch": 1.2606400527878587, "grad_norm": 0.3318745493888855, "learning_rate": 8e-05, "loss": 1.4726, "step": 5731 }, { "epoch": 1.260860002199494, "grad_norm": 0.3479224443435669, "learning_rate": 8e-05, "loss": 1.7334, "step": 5732 }, { "epoch": 1.2610799516111295, "grad_norm": 0.3396929204463959, "learning_rate": 8e-05, "loss": 1.6004, "step": 5733 }, { "epoch": 1.2612999010227648, "grad_norm": 0.33504292368888855, "learning_rate": 8e-05, "loss": 1.6189, "step": 5734 }, { "epoch": 1.2615198504344, "grad_norm": 0.331068754196167, "learning_rate": 8e-05, "loss": 1.6025, "step": 5735 }, { "epoch": 1.2617397998460353, "grad_norm": 0.3567442297935486, "learning_rate": 8e-05, "loss": 1.6865, "step": 5736 }, { "epoch": 1.2619597492576706, "grad_norm": 0.32938671112060547, "learning_rate": 8e-05, "loss": 1.4861, "step": 5737 }, { "epoch": 1.2621796986693061, "grad_norm": 0.3283194601535797, "learning_rate": 8e-05, "loss": 1.4589, "step": 5738 }, { "epoch": 1.2623996480809414, "grad_norm": 0.3385762870311737, "learning_rate": 8e-05, "loss": 1.6102, "step": 5739 }, { "epoch": 1.2626195974925767, "grad_norm": 0.3276645839214325, "learning_rate": 8e-05, "loss": 1.5589, "step": 5740 }, { "epoch": 1.262839546904212, "grad_norm": 0.34477272629737854, "learning_rate": 8e-05, "loss": 1.624, "step": 5741 }, { "epoch": 1.2630594963158472, "grad_norm": 0.36201030015945435, "learning_rate": 8e-05, "loss": 1.5186, "step": 5742 }, { "epoch": 1.2632794457274827, "grad_norm": 0.31587764620780945, "learning_rate": 8e-05, "loss": 1.4144, "step": 5743 }, { "epoch": 1.263499395139118, "grad_norm": 0.32143643498420715, "learning_rate": 8e-05, "loss": 1.5279, "step": 5744 }, { "epoch": 1.2637193445507533, "grad_norm": 0.3379841446876526, "learning_rate": 8e-05, "loss": 1.6029, "step": 5745 }, { "epoch": 1.2639392939623886, "grad_norm": 0.33502984046936035, "learning_rate": 8e-05, "loss": 1.6304, "step": 5746 }, { "epoch": 1.2641592433740239, "grad_norm": 0.34659984707832336, "learning_rate": 8e-05, "loss": 1.7422, "step": 5747 }, { "epoch": 1.2643791927856594, "grad_norm": 0.3354898989200592, "learning_rate": 8e-05, "loss": 1.5823, "step": 5748 }, { "epoch": 1.2645991421972946, "grad_norm": 0.3458346128463745, "learning_rate": 8e-05, "loss": 1.692, "step": 5749 }, { "epoch": 1.26481909160893, "grad_norm": 0.3335098922252655, "learning_rate": 8e-05, "loss": 1.6085, "step": 5750 }, { "epoch": 1.2650390410205652, "grad_norm": 0.31742867827415466, "learning_rate": 8e-05, "loss": 1.4346, "step": 5751 }, { "epoch": 1.2652589904322005, "grad_norm": 0.34154289960861206, "learning_rate": 8e-05, "loss": 1.5719, "step": 5752 }, { "epoch": 1.265478939843836, "grad_norm": 0.3809439241886139, "learning_rate": 8e-05, "loss": 1.8626, "step": 5753 }, { "epoch": 1.2656988892554712, "grad_norm": 0.31967952847480774, "learning_rate": 8e-05, "loss": 1.6678, "step": 5754 }, { "epoch": 1.2659188386671065, "grad_norm": 0.3379282057285309, "learning_rate": 8e-05, "loss": 1.5958, "step": 5755 }, { "epoch": 1.266138788078742, "grad_norm": 0.3195663094520569, "learning_rate": 8e-05, "loss": 1.4887, "step": 5756 }, { "epoch": 1.266358737490377, "grad_norm": 0.32915961742401123, "learning_rate": 8e-05, "loss": 1.445, "step": 5757 }, { "epoch": 1.2665786869020126, "grad_norm": 0.3462459444999695, "learning_rate": 8e-05, "loss": 1.6659, "step": 5758 }, { "epoch": 1.2667986363136479, "grad_norm": 0.3534404933452606, "learning_rate": 8e-05, "loss": 1.5256, "step": 5759 }, { "epoch": 1.2670185857252831, "grad_norm": 0.34299665689468384, "learning_rate": 8e-05, "loss": 1.5902, "step": 5760 }, { "epoch": 1.2672385351369186, "grad_norm": 0.3383926749229431, "learning_rate": 8e-05, "loss": 1.6551, "step": 5761 }, { "epoch": 1.2674584845485537, "grad_norm": 0.3526313602924347, "learning_rate": 8e-05, "loss": 1.6685, "step": 5762 }, { "epoch": 1.2676784339601892, "grad_norm": 0.3610459864139557, "learning_rate": 8e-05, "loss": 1.6952, "step": 5763 }, { "epoch": 1.2678983833718245, "grad_norm": 0.32312828302383423, "learning_rate": 8e-05, "loss": 1.5748, "step": 5764 }, { "epoch": 1.2681183327834598, "grad_norm": 0.3208003640174866, "learning_rate": 8e-05, "loss": 1.5428, "step": 5765 }, { "epoch": 1.2683382821950953, "grad_norm": 0.3384496569633484, "learning_rate": 8e-05, "loss": 1.6866, "step": 5766 }, { "epoch": 1.2685582316067303, "grad_norm": 0.3159599304199219, "learning_rate": 8e-05, "loss": 1.5148, "step": 5767 }, { "epoch": 1.2687781810183658, "grad_norm": 0.3415992259979248, "learning_rate": 8e-05, "loss": 1.6993, "step": 5768 }, { "epoch": 1.268998130430001, "grad_norm": 0.3889319896697998, "learning_rate": 8e-05, "loss": 1.6654, "step": 5769 }, { "epoch": 1.2692180798416364, "grad_norm": 0.34157922863960266, "learning_rate": 8e-05, "loss": 1.6064, "step": 5770 }, { "epoch": 1.2694380292532719, "grad_norm": 0.3280433118343353, "learning_rate": 8e-05, "loss": 1.5937, "step": 5771 }, { "epoch": 1.2696579786649071, "grad_norm": 0.3196824789047241, "learning_rate": 8e-05, "loss": 1.5382, "step": 5772 }, { "epoch": 1.2698779280765424, "grad_norm": 0.33472612500190735, "learning_rate": 8e-05, "loss": 1.4842, "step": 5773 }, { "epoch": 1.2700978774881777, "grad_norm": 0.31075823307037354, "learning_rate": 8e-05, "loss": 1.4177, "step": 5774 }, { "epoch": 1.270317826899813, "grad_norm": 0.3174769878387451, "learning_rate": 8e-05, "loss": 1.5067, "step": 5775 }, { "epoch": 1.2705377763114485, "grad_norm": 0.3373272716999054, "learning_rate": 8e-05, "loss": 1.5921, "step": 5776 }, { "epoch": 1.2707577257230838, "grad_norm": 0.327604740858078, "learning_rate": 8e-05, "loss": 1.5702, "step": 5777 }, { "epoch": 1.270977675134719, "grad_norm": 0.3138866126537323, "learning_rate": 8e-05, "loss": 1.5496, "step": 5778 }, { "epoch": 1.2711976245463543, "grad_norm": 0.3237580358982086, "learning_rate": 8e-05, "loss": 1.4749, "step": 5779 }, { "epoch": 1.2714175739579896, "grad_norm": 0.32673195004463196, "learning_rate": 8e-05, "loss": 1.4819, "step": 5780 }, { "epoch": 1.271637523369625, "grad_norm": 0.32615265250205994, "learning_rate": 8e-05, "loss": 1.5661, "step": 5781 }, { "epoch": 1.2718574727812604, "grad_norm": 0.33582961559295654, "learning_rate": 8e-05, "loss": 1.7018, "step": 5782 }, { "epoch": 1.2720774221928957, "grad_norm": 0.33420315384864807, "learning_rate": 8e-05, "loss": 1.745, "step": 5783 }, { "epoch": 1.272297371604531, "grad_norm": 0.34552863240242004, "learning_rate": 8e-05, "loss": 1.5608, "step": 5784 }, { "epoch": 1.2725173210161662, "grad_norm": 0.3157491683959961, "learning_rate": 8e-05, "loss": 1.5407, "step": 5785 }, { "epoch": 1.2727372704278017, "grad_norm": 0.3380187153816223, "learning_rate": 8e-05, "loss": 1.5758, "step": 5786 }, { "epoch": 1.272957219839437, "grad_norm": 0.3173157274723053, "learning_rate": 8e-05, "loss": 1.5148, "step": 5787 }, { "epoch": 1.2731771692510723, "grad_norm": 0.32248708605766296, "learning_rate": 8e-05, "loss": 1.4923, "step": 5788 }, { "epoch": 1.2733971186627076, "grad_norm": 0.3265584111213684, "learning_rate": 8e-05, "loss": 1.5693, "step": 5789 }, { "epoch": 1.2736170680743428, "grad_norm": 0.3398585617542267, "learning_rate": 8e-05, "loss": 1.6448, "step": 5790 }, { "epoch": 1.2738370174859783, "grad_norm": 0.34659549593925476, "learning_rate": 8e-05, "loss": 1.5303, "step": 5791 }, { "epoch": 1.2740569668976136, "grad_norm": 0.3225022554397583, "learning_rate": 8e-05, "loss": 1.6491, "step": 5792 }, { "epoch": 1.2742769163092489, "grad_norm": 0.32509443163871765, "learning_rate": 8e-05, "loss": 1.5475, "step": 5793 }, { "epoch": 1.2744968657208842, "grad_norm": 0.32910865545272827, "learning_rate": 8e-05, "loss": 1.6105, "step": 5794 }, { "epoch": 1.2747168151325194, "grad_norm": 0.3526688814163208, "learning_rate": 8e-05, "loss": 1.7675, "step": 5795 }, { "epoch": 1.274936764544155, "grad_norm": 0.3291871249675751, "learning_rate": 8e-05, "loss": 1.5497, "step": 5796 }, { "epoch": 1.2751567139557902, "grad_norm": 0.33106401562690735, "learning_rate": 8e-05, "loss": 1.6814, "step": 5797 }, { "epoch": 1.2753766633674255, "grad_norm": 0.33561843633651733, "learning_rate": 8e-05, "loss": 1.4132, "step": 5798 }, { "epoch": 1.2755966127790608, "grad_norm": 0.35269832611083984, "learning_rate": 8e-05, "loss": 1.6955, "step": 5799 }, { "epoch": 1.275816562190696, "grad_norm": 0.3453201949596405, "learning_rate": 8e-05, "loss": 1.5952, "step": 5800 }, { "epoch": 1.2760365116023316, "grad_norm": 0.34805524349212646, "learning_rate": 8e-05, "loss": 1.7234, "step": 5801 }, { "epoch": 1.2762564610139668, "grad_norm": 0.3452829122543335, "learning_rate": 8e-05, "loss": 1.8184, "step": 5802 }, { "epoch": 1.2764764104256021, "grad_norm": 0.36168619990348816, "learning_rate": 8e-05, "loss": 1.5154, "step": 5803 }, { "epoch": 1.2766963598372374, "grad_norm": 0.34174251556396484, "learning_rate": 8e-05, "loss": 1.5713, "step": 5804 }, { "epoch": 1.2769163092488727, "grad_norm": 0.31720733642578125, "learning_rate": 8e-05, "loss": 1.5914, "step": 5805 }, { "epoch": 1.2771362586605082, "grad_norm": 0.32637080550193787, "learning_rate": 8e-05, "loss": 1.4831, "step": 5806 }, { "epoch": 1.2773562080721435, "grad_norm": 0.35686230659484863, "learning_rate": 8e-05, "loss": 1.6729, "step": 5807 }, { "epoch": 1.2775761574837787, "grad_norm": 0.3505331873893738, "learning_rate": 8e-05, "loss": 1.6191, "step": 5808 }, { "epoch": 1.277796106895414, "grad_norm": 0.34981733560562134, "learning_rate": 8e-05, "loss": 1.6229, "step": 5809 }, { "epoch": 1.2780160563070493, "grad_norm": 0.34759530425071716, "learning_rate": 8e-05, "loss": 1.6676, "step": 5810 }, { "epoch": 1.2782360057186848, "grad_norm": 0.32409825921058655, "learning_rate": 8e-05, "loss": 1.5127, "step": 5811 }, { "epoch": 1.27845595513032, "grad_norm": 0.36701974272727966, "learning_rate": 8e-05, "loss": 1.7353, "step": 5812 }, { "epoch": 1.2786759045419553, "grad_norm": 0.3370521366596222, "learning_rate": 8e-05, "loss": 1.626, "step": 5813 }, { "epoch": 1.2788958539535906, "grad_norm": 0.33141398429870605, "learning_rate": 8e-05, "loss": 1.5112, "step": 5814 }, { "epoch": 1.279115803365226, "grad_norm": 0.3248597979545593, "learning_rate": 8e-05, "loss": 1.4501, "step": 5815 }, { "epoch": 1.2793357527768614, "grad_norm": 0.3217858076095581, "learning_rate": 8e-05, "loss": 1.4485, "step": 5816 }, { "epoch": 1.2795557021884967, "grad_norm": 0.33774372935295105, "learning_rate": 8e-05, "loss": 1.6402, "step": 5817 }, { "epoch": 1.279775651600132, "grad_norm": 0.36976364254951477, "learning_rate": 8e-05, "loss": 1.6359, "step": 5818 }, { "epoch": 1.2799956010117672, "grad_norm": 0.3499910831451416, "learning_rate": 8e-05, "loss": 1.5072, "step": 5819 }, { "epoch": 1.2802155504234025, "grad_norm": 0.3353627622127533, "learning_rate": 8e-05, "loss": 1.6044, "step": 5820 }, { "epoch": 1.280435499835038, "grad_norm": 0.33316507935523987, "learning_rate": 8e-05, "loss": 1.5741, "step": 5821 }, { "epoch": 1.2806554492466733, "grad_norm": 0.33303508162498474, "learning_rate": 8e-05, "loss": 1.5697, "step": 5822 }, { "epoch": 1.2808753986583086, "grad_norm": 0.34064459800720215, "learning_rate": 8e-05, "loss": 1.6154, "step": 5823 }, { "epoch": 1.2810953480699439, "grad_norm": 0.3434101343154907, "learning_rate": 8e-05, "loss": 1.6279, "step": 5824 }, { "epoch": 1.2813152974815791, "grad_norm": 0.350313276052475, "learning_rate": 8e-05, "loss": 1.4661, "step": 5825 }, { "epoch": 1.2815352468932146, "grad_norm": 0.33383291959762573, "learning_rate": 8e-05, "loss": 1.6202, "step": 5826 }, { "epoch": 1.28175519630485, "grad_norm": 0.34328338503837585, "learning_rate": 8e-05, "loss": 1.7364, "step": 5827 }, { "epoch": 1.2819751457164852, "grad_norm": 0.3742656409740448, "learning_rate": 8e-05, "loss": 1.5539, "step": 5828 }, { "epoch": 1.2821950951281205, "grad_norm": 0.3297211527824402, "learning_rate": 8e-05, "loss": 1.5865, "step": 5829 }, { "epoch": 1.2824150445397557, "grad_norm": 0.3314087688922882, "learning_rate": 8e-05, "loss": 1.5076, "step": 5830 }, { "epoch": 1.2826349939513912, "grad_norm": 0.3361984193325043, "learning_rate": 8e-05, "loss": 1.7265, "step": 5831 }, { "epoch": 1.2828549433630265, "grad_norm": 0.3405860960483551, "learning_rate": 8e-05, "loss": 1.6095, "step": 5832 }, { "epoch": 1.2830748927746618, "grad_norm": 0.3125757575035095, "learning_rate": 8e-05, "loss": 1.4864, "step": 5833 }, { "epoch": 1.283294842186297, "grad_norm": 0.3363991677761078, "learning_rate": 8e-05, "loss": 1.6635, "step": 5834 }, { "epoch": 1.2835147915979324, "grad_norm": 0.3404291868209839, "learning_rate": 8e-05, "loss": 1.6951, "step": 5835 }, { "epoch": 1.2837347410095679, "grad_norm": 0.3407963812351227, "learning_rate": 8e-05, "loss": 1.5288, "step": 5836 }, { "epoch": 1.2839546904212031, "grad_norm": 0.3440614342689514, "learning_rate": 8e-05, "loss": 1.5124, "step": 5837 }, { "epoch": 1.2841746398328384, "grad_norm": 0.3231651186943054, "learning_rate": 8e-05, "loss": 1.5796, "step": 5838 }, { "epoch": 1.2843945892444737, "grad_norm": 0.3530271649360657, "learning_rate": 8e-05, "loss": 1.5986, "step": 5839 }, { "epoch": 1.284614538656109, "grad_norm": 0.3331359028816223, "learning_rate": 8e-05, "loss": 1.6307, "step": 5840 }, { "epoch": 1.2848344880677445, "grad_norm": 0.32619747519493103, "learning_rate": 8e-05, "loss": 1.6997, "step": 5841 }, { "epoch": 1.2850544374793798, "grad_norm": 0.33588674664497375, "learning_rate": 8e-05, "loss": 1.6061, "step": 5842 }, { "epoch": 1.285274386891015, "grad_norm": 0.3240552842617035, "learning_rate": 8e-05, "loss": 1.6017, "step": 5843 }, { "epoch": 1.2854943363026503, "grad_norm": 0.32684293389320374, "learning_rate": 8e-05, "loss": 1.5564, "step": 5844 }, { "epoch": 1.2857142857142856, "grad_norm": 0.34023553133010864, "learning_rate": 8e-05, "loss": 1.6286, "step": 5845 }, { "epoch": 1.285934235125921, "grad_norm": 0.3998297452926636, "learning_rate": 8e-05, "loss": 1.2967, "step": 5846 }, { "epoch": 1.2861541845375564, "grad_norm": 0.36792194843292236, "learning_rate": 8e-05, "loss": 1.6372, "step": 5847 }, { "epoch": 1.2863741339491916, "grad_norm": 0.3371519446372986, "learning_rate": 8e-05, "loss": 1.5763, "step": 5848 }, { "epoch": 1.2865940833608271, "grad_norm": 0.3424137234687805, "learning_rate": 8e-05, "loss": 1.5175, "step": 5849 }, { "epoch": 1.2868140327724622, "grad_norm": 0.3422275483608246, "learning_rate": 8e-05, "loss": 1.5764, "step": 5850 }, { "epoch": 1.2870339821840977, "grad_norm": 0.34577250480651855, "learning_rate": 8e-05, "loss": 1.598, "step": 5851 }, { "epoch": 1.287253931595733, "grad_norm": 0.34285596013069153, "learning_rate": 8e-05, "loss": 1.5636, "step": 5852 }, { "epoch": 1.2874738810073683, "grad_norm": 0.40532851219177246, "learning_rate": 8e-05, "loss": 1.7531, "step": 5853 }, { "epoch": 1.2876938304190038, "grad_norm": 0.37148046493530273, "learning_rate": 8e-05, "loss": 1.5503, "step": 5854 }, { "epoch": 1.2879137798306388, "grad_norm": 0.329576700925827, "learning_rate": 8e-05, "loss": 1.6295, "step": 5855 }, { "epoch": 1.2881337292422743, "grad_norm": 0.3374268114566803, "learning_rate": 8e-05, "loss": 1.4654, "step": 5856 }, { "epoch": 1.2883536786539096, "grad_norm": 0.3384942412376404, "learning_rate": 8e-05, "loss": 1.4383, "step": 5857 }, { "epoch": 1.2885736280655449, "grad_norm": 0.3294598162174225, "learning_rate": 8e-05, "loss": 1.6026, "step": 5858 }, { "epoch": 1.2887935774771804, "grad_norm": 0.33762332797050476, "learning_rate": 8e-05, "loss": 1.6593, "step": 5859 }, { "epoch": 1.2890135268888157, "grad_norm": 0.3353956341743469, "learning_rate": 8e-05, "loss": 1.5161, "step": 5860 }, { "epoch": 1.289233476300451, "grad_norm": 0.4760856330394745, "learning_rate": 8e-05, "loss": 1.5239, "step": 5861 }, { "epoch": 1.2894534257120862, "grad_norm": 0.33830228447914124, "learning_rate": 8e-05, "loss": 1.6457, "step": 5862 }, { "epoch": 1.2896733751237215, "grad_norm": 0.35845375061035156, "learning_rate": 8e-05, "loss": 1.4276, "step": 5863 }, { "epoch": 1.289893324535357, "grad_norm": 0.33352896571159363, "learning_rate": 8e-05, "loss": 1.4962, "step": 5864 }, { "epoch": 1.2901132739469923, "grad_norm": 0.37244275212287903, "learning_rate": 8e-05, "loss": 1.7762, "step": 5865 }, { "epoch": 1.2903332233586275, "grad_norm": 0.3415049910545349, "learning_rate": 8e-05, "loss": 1.626, "step": 5866 }, { "epoch": 1.2905531727702628, "grad_norm": 0.34911665320396423, "learning_rate": 8e-05, "loss": 1.681, "step": 5867 }, { "epoch": 1.290773122181898, "grad_norm": 0.34773507714271545, "learning_rate": 8e-05, "loss": 1.5572, "step": 5868 }, { "epoch": 1.2909930715935336, "grad_norm": 0.33827733993530273, "learning_rate": 8e-05, "loss": 1.5948, "step": 5869 }, { "epoch": 1.2912130210051689, "grad_norm": 0.338717520236969, "learning_rate": 8e-05, "loss": 1.4728, "step": 5870 }, { "epoch": 1.2914329704168042, "grad_norm": 0.35328876972198486, "learning_rate": 8e-05, "loss": 1.5482, "step": 5871 }, { "epoch": 1.2916529198284394, "grad_norm": 0.36347275972366333, "learning_rate": 8e-05, "loss": 1.5803, "step": 5872 }, { "epoch": 1.2918728692400747, "grad_norm": 0.34006237983703613, "learning_rate": 8e-05, "loss": 1.5576, "step": 5873 }, { "epoch": 1.2920928186517102, "grad_norm": 0.3328593671321869, "learning_rate": 8e-05, "loss": 1.5646, "step": 5874 }, { "epoch": 1.2923127680633455, "grad_norm": 0.3470839858055115, "learning_rate": 8e-05, "loss": 1.6872, "step": 5875 }, { "epoch": 1.2925327174749808, "grad_norm": 0.34710055589675903, "learning_rate": 8e-05, "loss": 1.6532, "step": 5876 }, { "epoch": 1.292752666886616, "grad_norm": 0.3024667203426361, "learning_rate": 8e-05, "loss": 1.4408, "step": 5877 }, { "epoch": 1.2929726162982513, "grad_norm": 0.31272682547569275, "learning_rate": 8e-05, "loss": 1.4565, "step": 5878 }, { "epoch": 1.2931925657098868, "grad_norm": 0.3355286717414856, "learning_rate": 8e-05, "loss": 1.5444, "step": 5879 }, { "epoch": 1.293412515121522, "grad_norm": 0.33059918880462646, "learning_rate": 8e-05, "loss": 1.6431, "step": 5880 }, { "epoch": 1.2936324645331574, "grad_norm": 0.3211008906364441, "learning_rate": 8e-05, "loss": 1.5291, "step": 5881 }, { "epoch": 1.2938524139447927, "grad_norm": 0.3514041304588318, "learning_rate": 8e-05, "loss": 1.6956, "step": 5882 }, { "epoch": 1.294072363356428, "grad_norm": 0.33707863092422485, "learning_rate": 8e-05, "loss": 1.5416, "step": 5883 }, { "epoch": 1.2942923127680634, "grad_norm": 0.32268643379211426, "learning_rate": 8e-05, "loss": 1.4777, "step": 5884 }, { "epoch": 1.2945122621796987, "grad_norm": 0.3480195999145508, "learning_rate": 8e-05, "loss": 1.6735, "step": 5885 }, { "epoch": 1.294732211591334, "grad_norm": 0.36998799443244934, "learning_rate": 8e-05, "loss": 1.7508, "step": 5886 }, { "epoch": 1.2949521610029693, "grad_norm": 0.4005178213119507, "learning_rate": 8e-05, "loss": 1.6968, "step": 5887 }, { "epoch": 1.2951721104146046, "grad_norm": 0.3278454542160034, "learning_rate": 8e-05, "loss": 1.5786, "step": 5888 }, { "epoch": 1.29539205982624, "grad_norm": 0.35444578528404236, "learning_rate": 8e-05, "loss": 1.644, "step": 5889 }, { "epoch": 1.2956120092378753, "grad_norm": 0.3356686234474182, "learning_rate": 8e-05, "loss": 1.4568, "step": 5890 }, { "epoch": 1.2958319586495106, "grad_norm": 0.3335033655166626, "learning_rate": 8e-05, "loss": 1.3737, "step": 5891 }, { "epoch": 1.296051908061146, "grad_norm": 0.3359556496143341, "learning_rate": 8e-05, "loss": 1.632, "step": 5892 }, { "epoch": 1.2962718574727812, "grad_norm": 0.35924333333969116, "learning_rate": 8e-05, "loss": 1.5183, "step": 5893 }, { "epoch": 1.2964918068844167, "grad_norm": 0.32025250792503357, "learning_rate": 8e-05, "loss": 1.4273, "step": 5894 }, { "epoch": 1.296711756296052, "grad_norm": 0.3237527906894684, "learning_rate": 8e-05, "loss": 1.5184, "step": 5895 }, { "epoch": 1.2969317057076872, "grad_norm": 0.33525776863098145, "learning_rate": 8e-05, "loss": 1.521, "step": 5896 }, { "epoch": 1.2971516551193225, "grad_norm": 0.3260558545589447, "learning_rate": 8e-05, "loss": 1.3431, "step": 5897 }, { "epoch": 1.2973716045309578, "grad_norm": 0.3463820517063141, "learning_rate": 8e-05, "loss": 1.6279, "step": 5898 }, { "epoch": 1.2975915539425933, "grad_norm": 0.34793299436569214, "learning_rate": 8e-05, "loss": 1.6219, "step": 5899 }, { "epoch": 1.2978115033542286, "grad_norm": 0.34392720460891724, "learning_rate": 8e-05, "loss": 1.7517, "step": 5900 }, { "epoch": 1.2980314527658638, "grad_norm": 0.36635079979896545, "learning_rate": 8e-05, "loss": 1.5951, "step": 5901 }, { "epoch": 1.2982514021774991, "grad_norm": 0.33466577529907227, "learning_rate": 8e-05, "loss": 1.5195, "step": 5902 }, { "epoch": 1.2984713515891344, "grad_norm": 0.34258362650871277, "learning_rate": 8e-05, "loss": 1.6252, "step": 5903 }, { "epoch": 1.29869130100077, "grad_norm": 0.35498541593551636, "learning_rate": 8e-05, "loss": 1.6613, "step": 5904 }, { "epoch": 1.2989112504124052, "grad_norm": 0.35436779260635376, "learning_rate": 8e-05, "loss": 1.7664, "step": 5905 }, { "epoch": 1.2991311998240405, "grad_norm": 0.3335748016834259, "learning_rate": 8e-05, "loss": 1.5144, "step": 5906 }, { "epoch": 1.2993511492356757, "grad_norm": 0.32442042231559753, "learning_rate": 8e-05, "loss": 1.5285, "step": 5907 }, { "epoch": 1.299571098647311, "grad_norm": 0.3264372646808624, "learning_rate": 8e-05, "loss": 1.5492, "step": 5908 }, { "epoch": 1.2997910480589465, "grad_norm": 0.34280022978782654, "learning_rate": 8e-05, "loss": 1.6002, "step": 5909 }, { "epoch": 1.3000109974705818, "grad_norm": 0.3372175097465515, "learning_rate": 8e-05, "loss": 1.6081, "step": 5910 }, { "epoch": 1.300230946882217, "grad_norm": 0.3387996256351471, "learning_rate": 8e-05, "loss": 1.562, "step": 5911 }, { "epoch": 1.3004508962938524, "grad_norm": 0.3518845736980438, "learning_rate": 8e-05, "loss": 1.6278, "step": 5912 }, { "epoch": 1.3006708457054876, "grad_norm": 0.33492013812065125, "learning_rate": 8e-05, "loss": 1.5377, "step": 5913 }, { "epoch": 1.3008907951171231, "grad_norm": 0.337647944688797, "learning_rate": 8e-05, "loss": 1.5752, "step": 5914 }, { "epoch": 1.3011107445287584, "grad_norm": 0.32401975989341736, "learning_rate": 8e-05, "loss": 1.4835, "step": 5915 }, { "epoch": 1.3013306939403937, "grad_norm": 0.3245975375175476, "learning_rate": 8e-05, "loss": 1.4199, "step": 5916 }, { "epoch": 1.301550643352029, "grad_norm": 0.34633052349090576, "learning_rate": 8e-05, "loss": 1.5353, "step": 5917 }, { "epoch": 1.3017705927636642, "grad_norm": 0.3719714283943176, "learning_rate": 8e-05, "loss": 1.5607, "step": 5918 }, { "epoch": 1.3019905421752997, "grad_norm": 0.32893359661102295, "learning_rate": 8e-05, "loss": 1.5397, "step": 5919 }, { "epoch": 1.302210491586935, "grad_norm": 0.326135516166687, "learning_rate": 8e-05, "loss": 1.5054, "step": 5920 }, { "epoch": 1.3024304409985703, "grad_norm": 0.32936835289001465, "learning_rate": 8e-05, "loss": 1.6037, "step": 5921 }, { "epoch": 1.3026503904102056, "grad_norm": 0.32363584637641907, "learning_rate": 8e-05, "loss": 1.4456, "step": 5922 }, { "epoch": 1.3028703398218409, "grad_norm": 0.3920203745365143, "learning_rate": 8e-05, "loss": 1.729, "step": 5923 }, { "epoch": 1.3030902892334764, "grad_norm": 0.33076465129852295, "learning_rate": 8e-05, "loss": 1.5799, "step": 5924 }, { "epoch": 1.3033102386451116, "grad_norm": 0.35621124505996704, "learning_rate": 8e-05, "loss": 1.6936, "step": 5925 }, { "epoch": 1.303530188056747, "grad_norm": 0.3444978594779968, "learning_rate": 8e-05, "loss": 1.7067, "step": 5926 }, { "epoch": 1.3037501374683822, "grad_norm": 0.3366940915584564, "learning_rate": 8e-05, "loss": 1.5143, "step": 5927 }, { "epoch": 1.3039700868800175, "grad_norm": 0.32055747509002686, "learning_rate": 8e-05, "loss": 1.489, "step": 5928 }, { "epoch": 1.304190036291653, "grad_norm": 0.35443323850631714, "learning_rate": 8e-05, "loss": 1.6075, "step": 5929 }, { "epoch": 1.3044099857032883, "grad_norm": 0.3282393217086792, "learning_rate": 8e-05, "loss": 1.6593, "step": 5930 }, { "epoch": 1.3046299351149235, "grad_norm": 0.3338877260684967, "learning_rate": 8e-05, "loss": 1.5641, "step": 5931 }, { "epoch": 1.3048498845265588, "grad_norm": 0.351909339427948, "learning_rate": 8e-05, "loss": 1.6101, "step": 5932 }, { "epoch": 1.305069833938194, "grad_norm": 0.3434227705001831, "learning_rate": 8e-05, "loss": 1.6274, "step": 5933 }, { "epoch": 1.3052897833498296, "grad_norm": 0.35757461190223694, "learning_rate": 8e-05, "loss": 1.5097, "step": 5934 }, { "epoch": 1.3055097327614649, "grad_norm": 0.3390028476715088, "learning_rate": 8e-05, "loss": 1.6082, "step": 5935 }, { "epoch": 1.3057296821731001, "grad_norm": 0.32802119851112366, "learning_rate": 8e-05, "loss": 1.4886, "step": 5936 }, { "epoch": 1.3059496315847356, "grad_norm": 0.3224239945411682, "learning_rate": 8e-05, "loss": 1.4756, "step": 5937 }, { "epoch": 1.3061695809963707, "grad_norm": 0.3324672281742096, "learning_rate": 8e-05, "loss": 1.7266, "step": 5938 }, { "epoch": 1.3063895304080062, "grad_norm": 0.3703005313873291, "learning_rate": 8e-05, "loss": 1.6417, "step": 5939 }, { "epoch": 1.3066094798196415, "grad_norm": 0.3569459021091461, "learning_rate": 8e-05, "loss": 1.6073, "step": 5940 }, { "epoch": 1.3068294292312768, "grad_norm": 0.3246351480484009, "learning_rate": 8e-05, "loss": 1.4794, "step": 5941 }, { "epoch": 1.3070493786429123, "grad_norm": 0.3382027745246887, "learning_rate": 8e-05, "loss": 1.4517, "step": 5942 }, { "epoch": 1.3072693280545473, "grad_norm": 0.3558928072452545, "learning_rate": 8e-05, "loss": 1.7814, "step": 5943 }, { "epoch": 1.3074892774661828, "grad_norm": 0.3392329514026642, "learning_rate": 8e-05, "loss": 1.643, "step": 5944 }, { "epoch": 1.307709226877818, "grad_norm": 0.33706268668174744, "learning_rate": 8e-05, "loss": 1.5449, "step": 5945 }, { "epoch": 1.3079291762894534, "grad_norm": 0.3294903039932251, "learning_rate": 8e-05, "loss": 1.5651, "step": 5946 }, { "epoch": 1.3081491257010889, "grad_norm": 0.3251626193523407, "learning_rate": 8e-05, "loss": 1.4819, "step": 5947 }, { "epoch": 1.3083690751127242, "grad_norm": 0.351766973733902, "learning_rate": 8e-05, "loss": 1.5871, "step": 5948 }, { "epoch": 1.3085890245243594, "grad_norm": 0.337314248085022, "learning_rate": 8e-05, "loss": 1.6411, "step": 5949 }, { "epoch": 1.3088089739359947, "grad_norm": 0.3252117335796356, "learning_rate": 8e-05, "loss": 1.5951, "step": 5950 }, { "epoch": 1.30902892334763, "grad_norm": 0.324153333902359, "learning_rate": 8e-05, "loss": 1.4967, "step": 5951 }, { "epoch": 1.3092488727592655, "grad_norm": 0.3286396563053131, "learning_rate": 8e-05, "loss": 1.6705, "step": 5952 }, { "epoch": 1.3094688221709008, "grad_norm": 0.31972208619117737, "learning_rate": 8e-05, "loss": 1.4626, "step": 5953 }, { "epoch": 1.309688771582536, "grad_norm": 0.3321622610092163, "learning_rate": 8e-05, "loss": 1.5618, "step": 5954 }, { "epoch": 1.3099087209941713, "grad_norm": 0.33044907450675964, "learning_rate": 8e-05, "loss": 1.5908, "step": 5955 }, { "epoch": 1.3101286704058066, "grad_norm": 0.32111304998397827, "learning_rate": 8e-05, "loss": 1.6319, "step": 5956 }, { "epoch": 1.310348619817442, "grad_norm": 0.32864564657211304, "learning_rate": 8e-05, "loss": 1.5323, "step": 5957 }, { "epoch": 1.3105685692290774, "grad_norm": 0.36656758189201355, "learning_rate": 8e-05, "loss": 1.8201, "step": 5958 }, { "epoch": 1.3107885186407127, "grad_norm": 0.3370371162891388, "learning_rate": 8e-05, "loss": 1.5212, "step": 5959 }, { "epoch": 1.311008468052348, "grad_norm": 0.32487478852272034, "learning_rate": 8e-05, "loss": 1.4797, "step": 5960 }, { "epoch": 1.3112284174639832, "grad_norm": 0.3607526421546936, "learning_rate": 8e-05, "loss": 1.627, "step": 5961 }, { "epoch": 1.3114483668756187, "grad_norm": 0.35125812888145447, "learning_rate": 8e-05, "loss": 1.569, "step": 5962 }, { "epoch": 1.311668316287254, "grad_norm": 0.3638373911380768, "learning_rate": 8e-05, "loss": 1.54, "step": 5963 }, { "epoch": 1.3118882656988893, "grad_norm": 0.33281412720680237, "learning_rate": 8e-05, "loss": 1.4969, "step": 5964 }, { "epoch": 1.3121082151105246, "grad_norm": 0.32616186141967773, "learning_rate": 8e-05, "loss": 1.6132, "step": 5965 }, { "epoch": 1.3123281645221598, "grad_norm": 0.3462194800376892, "learning_rate": 8e-05, "loss": 1.5994, "step": 5966 }, { "epoch": 1.3125481139337953, "grad_norm": 0.32356584072113037, "learning_rate": 8e-05, "loss": 1.602, "step": 5967 }, { "epoch": 1.3127680633454306, "grad_norm": 0.3561713397502899, "learning_rate": 8e-05, "loss": 1.5421, "step": 5968 }, { "epoch": 1.312988012757066, "grad_norm": 0.34750887751579285, "learning_rate": 8e-05, "loss": 1.5941, "step": 5969 }, { "epoch": 1.3132079621687012, "grad_norm": 0.3413698673248291, "learning_rate": 8e-05, "loss": 1.6256, "step": 5970 }, { "epoch": 1.3134279115803364, "grad_norm": 0.3173705041408539, "learning_rate": 8e-05, "loss": 1.4357, "step": 5971 }, { "epoch": 1.313647860991972, "grad_norm": 0.3310370147228241, "learning_rate": 8e-05, "loss": 1.5645, "step": 5972 }, { "epoch": 1.3138678104036072, "grad_norm": 0.32835450768470764, "learning_rate": 8e-05, "loss": 1.6231, "step": 5973 }, { "epoch": 1.3140877598152425, "grad_norm": 0.31849145889282227, "learning_rate": 8e-05, "loss": 1.5162, "step": 5974 }, { "epoch": 1.3143077092268778, "grad_norm": 0.308278352022171, "learning_rate": 8e-05, "loss": 1.4638, "step": 5975 }, { "epoch": 1.314527658638513, "grad_norm": 0.32485702633857727, "learning_rate": 8e-05, "loss": 1.5829, "step": 5976 }, { "epoch": 1.3147476080501486, "grad_norm": 0.3265734910964966, "learning_rate": 8e-05, "loss": 1.5485, "step": 5977 }, { "epoch": 1.3149675574617838, "grad_norm": 0.3445983827114105, "learning_rate": 8e-05, "loss": 1.7254, "step": 5978 }, { "epoch": 1.3151875068734191, "grad_norm": 0.36030635237693787, "learning_rate": 8e-05, "loss": 1.5771, "step": 5979 }, { "epoch": 1.3154074562850544, "grad_norm": 0.3250638246536255, "learning_rate": 8e-05, "loss": 1.447, "step": 5980 }, { "epoch": 1.3156274056966897, "grad_norm": 0.33962443470954895, "learning_rate": 8e-05, "loss": 1.7122, "step": 5981 }, { "epoch": 1.3158473551083252, "grad_norm": 0.5697662234306335, "learning_rate": 8e-05, "loss": 1.5121, "step": 5982 }, { "epoch": 1.3160673045199605, "grad_norm": 0.3297199010848999, "learning_rate": 8e-05, "loss": 1.4394, "step": 5983 }, { "epoch": 1.3162872539315957, "grad_norm": 0.3547095060348511, "learning_rate": 8e-05, "loss": 1.6518, "step": 5984 }, { "epoch": 1.316507203343231, "grad_norm": 0.35449567437171936, "learning_rate": 8e-05, "loss": 1.7126, "step": 5985 }, { "epoch": 1.3167271527548663, "grad_norm": 0.3340552747249603, "learning_rate": 8e-05, "loss": 1.5415, "step": 5986 }, { "epoch": 1.3169471021665018, "grad_norm": 0.3509848713874817, "learning_rate": 8e-05, "loss": 1.7562, "step": 5987 }, { "epoch": 1.317167051578137, "grad_norm": 0.3389817774295807, "learning_rate": 8e-05, "loss": 1.6782, "step": 5988 }, { "epoch": 1.3173870009897723, "grad_norm": 0.3417060375213623, "learning_rate": 8e-05, "loss": 1.621, "step": 5989 }, { "epoch": 1.3176069504014076, "grad_norm": 0.3256407082080841, "learning_rate": 8e-05, "loss": 1.6451, "step": 5990 }, { "epoch": 1.317826899813043, "grad_norm": 0.3228999376296997, "learning_rate": 8e-05, "loss": 1.5093, "step": 5991 }, { "epoch": 1.3180468492246784, "grad_norm": 0.3274962604045868, "learning_rate": 8e-05, "loss": 1.5513, "step": 5992 }, { "epoch": 1.3182667986363137, "grad_norm": 0.3139627277851105, "learning_rate": 8e-05, "loss": 1.5235, "step": 5993 }, { "epoch": 1.318486748047949, "grad_norm": 0.3347333073616028, "learning_rate": 8e-05, "loss": 1.4233, "step": 5994 }, { "epoch": 1.3187066974595842, "grad_norm": 0.338908851146698, "learning_rate": 8e-05, "loss": 1.5628, "step": 5995 }, { "epoch": 1.3189266468712195, "grad_norm": 0.3488113284111023, "learning_rate": 8e-05, "loss": 1.619, "step": 5996 }, { "epoch": 1.319146596282855, "grad_norm": 0.3352358341217041, "learning_rate": 8e-05, "loss": 1.6169, "step": 5997 }, { "epoch": 1.3193665456944903, "grad_norm": 0.3434486389160156, "learning_rate": 8e-05, "loss": 1.5001, "step": 5998 }, { "epoch": 1.3195864951061256, "grad_norm": 0.33616381883621216, "learning_rate": 8e-05, "loss": 1.5795, "step": 5999 }, { "epoch": 1.3198064445177609, "grad_norm": 0.3394475281238556, "learning_rate": 8e-05, "loss": 1.6557, "step": 6000 }, { "epoch": 1.3200263939293961, "grad_norm": 0.3576595187187195, "learning_rate": 8e-05, "loss": 1.7195, "step": 6001 }, { "epoch": 1.3202463433410316, "grad_norm": 0.3196483850479126, "learning_rate": 8e-05, "loss": 1.5853, "step": 6002 }, { "epoch": 1.320466292752667, "grad_norm": 0.3297625184059143, "learning_rate": 8e-05, "loss": 1.6251, "step": 6003 }, { "epoch": 1.3206862421643022, "grad_norm": 0.34634798765182495, "learning_rate": 8e-05, "loss": 1.6284, "step": 6004 }, { "epoch": 1.3209061915759375, "grad_norm": 0.32295578718185425, "learning_rate": 8e-05, "loss": 1.5071, "step": 6005 }, { "epoch": 1.3211261409875728, "grad_norm": 0.3227352201938629, "learning_rate": 8e-05, "loss": 1.6812, "step": 6006 }, { "epoch": 1.3213460903992083, "grad_norm": 0.3291846513748169, "learning_rate": 8e-05, "loss": 1.4854, "step": 6007 }, { "epoch": 1.3215660398108435, "grad_norm": 0.32996663451194763, "learning_rate": 8e-05, "loss": 1.5582, "step": 6008 }, { "epoch": 1.3217859892224788, "grad_norm": 0.35022804141044617, "learning_rate": 8e-05, "loss": 1.7505, "step": 6009 }, { "epoch": 1.322005938634114, "grad_norm": 0.36292940378189087, "learning_rate": 8e-05, "loss": 1.695, "step": 6010 }, { "epoch": 1.3222258880457494, "grad_norm": 0.3400166630744934, "learning_rate": 8e-05, "loss": 1.6049, "step": 6011 }, { "epoch": 1.3224458374573849, "grad_norm": 0.3160114586353302, "learning_rate": 8e-05, "loss": 1.5042, "step": 6012 }, { "epoch": 1.3226657868690201, "grad_norm": 0.31563568115234375, "learning_rate": 8e-05, "loss": 1.5282, "step": 6013 }, { "epoch": 1.3228857362806554, "grad_norm": 0.353161096572876, "learning_rate": 8e-05, "loss": 1.6662, "step": 6014 }, { "epoch": 1.3231056856922907, "grad_norm": 0.3125203251838684, "learning_rate": 8e-05, "loss": 1.4417, "step": 6015 }, { "epoch": 1.323325635103926, "grad_norm": 0.3198567032814026, "learning_rate": 8e-05, "loss": 1.6448, "step": 6016 }, { "epoch": 1.3235455845155615, "grad_norm": 0.32752862572669983, "learning_rate": 8e-05, "loss": 1.5483, "step": 6017 }, { "epoch": 1.3237655339271968, "grad_norm": 0.3351952135562897, "learning_rate": 8e-05, "loss": 1.5752, "step": 6018 }, { "epoch": 1.323985483338832, "grad_norm": 0.3336290419101715, "learning_rate": 8e-05, "loss": 1.5702, "step": 6019 }, { "epoch": 1.3242054327504673, "grad_norm": 0.3715028762817383, "learning_rate": 8e-05, "loss": 1.6565, "step": 6020 }, { "epoch": 1.3244253821621026, "grad_norm": 0.34583404660224915, "learning_rate": 8e-05, "loss": 1.5659, "step": 6021 }, { "epoch": 1.324645331573738, "grad_norm": 0.3281787633895874, "learning_rate": 8e-05, "loss": 1.5165, "step": 6022 }, { "epoch": 1.3248652809853734, "grad_norm": 0.33023008704185486, "learning_rate": 8e-05, "loss": 1.5336, "step": 6023 }, { "epoch": 1.3250852303970087, "grad_norm": 0.330657035112381, "learning_rate": 8e-05, "loss": 1.5752, "step": 6024 }, { "epoch": 1.325305179808644, "grad_norm": 0.35736900568008423, "learning_rate": 8e-05, "loss": 1.544, "step": 6025 }, { "epoch": 1.3255251292202792, "grad_norm": 0.3257259428501129, "learning_rate": 8e-05, "loss": 1.5169, "step": 6026 }, { "epoch": 1.3257450786319147, "grad_norm": 0.34167152643203735, "learning_rate": 8e-05, "loss": 1.7091, "step": 6027 }, { "epoch": 1.32596502804355, "grad_norm": 0.33038297295570374, "learning_rate": 8e-05, "loss": 1.6024, "step": 6028 }, { "epoch": 1.3261849774551853, "grad_norm": 0.32675421237945557, "learning_rate": 8e-05, "loss": 1.5849, "step": 6029 }, { "epoch": 1.3264049268668208, "grad_norm": 0.34686848521232605, "learning_rate": 8e-05, "loss": 1.6017, "step": 6030 }, { "epoch": 1.3266248762784558, "grad_norm": 0.3207021653652191, "learning_rate": 8e-05, "loss": 1.4712, "step": 6031 }, { "epoch": 1.3268448256900913, "grad_norm": 0.3322916626930237, "learning_rate": 8e-05, "loss": 1.5916, "step": 6032 }, { "epoch": 1.3270647751017266, "grad_norm": 0.3557954430580139, "learning_rate": 8e-05, "loss": 1.6464, "step": 6033 }, { "epoch": 1.3272847245133619, "grad_norm": 0.34026825428009033, "learning_rate": 8e-05, "loss": 1.6018, "step": 6034 }, { "epoch": 1.3275046739249974, "grad_norm": 0.3432289659976959, "learning_rate": 8e-05, "loss": 1.6503, "step": 6035 }, { "epoch": 1.3277246233366324, "grad_norm": 0.32447800040245056, "learning_rate": 8e-05, "loss": 1.5388, "step": 6036 }, { "epoch": 1.327944572748268, "grad_norm": 0.3442445695400238, "learning_rate": 8e-05, "loss": 1.5852, "step": 6037 }, { "epoch": 1.3281645221599032, "grad_norm": 0.3369056284427643, "learning_rate": 8e-05, "loss": 1.5092, "step": 6038 }, { "epoch": 1.3283844715715385, "grad_norm": 0.3355618119239807, "learning_rate": 8e-05, "loss": 1.5832, "step": 6039 }, { "epoch": 1.328604420983174, "grad_norm": 0.382839173078537, "learning_rate": 8e-05, "loss": 1.4238, "step": 6040 }, { "epoch": 1.3288243703948093, "grad_norm": 0.31913673877716064, "learning_rate": 8e-05, "loss": 1.4823, "step": 6041 }, { "epoch": 1.3290443198064446, "grad_norm": 0.4159768223762512, "learning_rate": 8e-05, "loss": 1.8749, "step": 6042 }, { "epoch": 1.3292642692180798, "grad_norm": 0.32975059747695923, "learning_rate": 8e-05, "loss": 1.5852, "step": 6043 }, { "epoch": 1.329484218629715, "grad_norm": 0.3431590795516968, "learning_rate": 8e-05, "loss": 1.8105, "step": 6044 }, { "epoch": 1.3297041680413506, "grad_norm": 0.3493921160697937, "learning_rate": 8e-05, "loss": 1.6214, "step": 6045 }, { "epoch": 1.3299241174529859, "grad_norm": 0.3331540822982788, "learning_rate": 8e-05, "loss": 1.5963, "step": 6046 }, { "epoch": 1.3301440668646212, "grad_norm": 0.3274116814136505, "learning_rate": 8e-05, "loss": 1.4361, "step": 6047 }, { "epoch": 1.3303640162762564, "grad_norm": 0.35047322511672974, "learning_rate": 8e-05, "loss": 1.6493, "step": 6048 }, { "epoch": 1.3305839656878917, "grad_norm": 0.35883471369743347, "learning_rate": 8e-05, "loss": 1.7726, "step": 6049 }, { "epoch": 1.3308039150995272, "grad_norm": 0.3504885137081146, "learning_rate": 8e-05, "loss": 1.5777, "step": 6050 }, { "epoch": 1.3310238645111625, "grad_norm": 0.33563196659088135, "learning_rate": 8e-05, "loss": 1.6279, "step": 6051 }, { "epoch": 1.3312438139227978, "grad_norm": 0.3330928087234497, "learning_rate": 8e-05, "loss": 1.4173, "step": 6052 }, { "epoch": 1.331463763334433, "grad_norm": 0.3432624340057373, "learning_rate": 8e-05, "loss": 1.5736, "step": 6053 }, { "epoch": 1.3316837127460683, "grad_norm": 0.3374311923980713, "learning_rate": 8e-05, "loss": 1.6143, "step": 6054 }, { "epoch": 1.3319036621577038, "grad_norm": 0.34315142035484314, "learning_rate": 8e-05, "loss": 1.5741, "step": 6055 }, { "epoch": 1.3321236115693391, "grad_norm": 0.35554829239845276, "learning_rate": 8e-05, "loss": 1.4987, "step": 6056 }, { "epoch": 1.3323435609809744, "grad_norm": 0.34292054176330566, "learning_rate": 8e-05, "loss": 1.4617, "step": 6057 }, { "epoch": 1.3325635103926097, "grad_norm": 0.33219000697135925, "learning_rate": 8e-05, "loss": 1.5718, "step": 6058 }, { "epoch": 1.332783459804245, "grad_norm": 0.351927787065506, "learning_rate": 8e-05, "loss": 1.553, "step": 6059 }, { "epoch": 1.3330034092158805, "grad_norm": 0.36198318004608154, "learning_rate": 8e-05, "loss": 1.6553, "step": 6060 }, { "epoch": 1.3332233586275157, "grad_norm": 0.34996742010116577, "learning_rate": 8e-05, "loss": 1.5125, "step": 6061 }, { "epoch": 1.333443308039151, "grad_norm": 0.3460218012332916, "learning_rate": 8e-05, "loss": 1.6684, "step": 6062 }, { "epoch": 1.3336632574507863, "grad_norm": 0.31589436531066895, "learning_rate": 8e-05, "loss": 1.3995, "step": 6063 }, { "epoch": 1.3338832068624216, "grad_norm": 0.33238857984542847, "learning_rate": 8e-05, "loss": 1.5671, "step": 6064 }, { "epoch": 1.334103156274057, "grad_norm": 0.3485605716705322, "learning_rate": 8e-05, "loss": 1.4391, "step": 6065 }, { "epoch": 1.3343231056856923, "grad_norm": 0.32765957713127136, "learning_rate": 8e-05, "loss": 1.4905, "step": 6066 }, { "epoch": 1.3345430550973276, "grad_norm": 0.34287306666374207, "learning_rate": 8e-05, "loss": 1.6635, "step": 6067 }, { "epoch": 1.334763004508963, "grad_norm": 0.3525535762310028, "learning_rate": 8e-05, "loss": 1.5757, "step": 6068 }, { "epoch": 1.3349829539205982, "grad_norm": 0.32966187596321106, "learning_rate": 8e-05, "loss": 1.5941, "step": 6069 }, { "epoch": 1.3352029033322337, "grad_norm": 0.349669486284256, "learning_rate": 8e-05, "loss": 1.6439, "step": 6070 }, { "epoch": 1.335422852743869, "grad_norm": 0.35162270069122314, "learning_rate": 8e-05, "loss": 1.5384, "step": 6071 }, { "epoch": 1.3356428021555042, "grad_norm": 0.34690824151039124, "learning_rate": 8e-05, "loss": 1.6752, "step": 6072 }, { "epoch": 1.3358627515671395, "grad_norm": 0.3307788670063019, "learning_rate": 8e-05, "loss": 1.5784, "step": 6073 }, { "epoch": 1.3360827009787748, "grad_norm": 0.3304356038570404, "learning_rate": 8e-05, "loss": 1.4711, "step": 6074 }, { "epoch": 1.3363026503904103, "grad_norm": 0.3339150547981262, "learning_rate": 8e-05, "loss": 1.5488, "step": 6075 }, { "epoch": 1.3365225998020456, "grad_norm": 0.33577847480773926, "learning_rate": 8e-05, "loss": 1.582, "step": 6076 }, { "epoch": 1.3367425492136809, "grad_norm": 0.3249453902244568, "learning_rate": 8e-05, "loss": 1.4071, "step": 6077 }, { "epoch": 1.3369624986253161, "grad_norm": 0.3324025273323059, "learning_rate": 8e-05, "loss": 1.636, "step": 6078 }, { "epoch": 1.3371824480369514, "grad_norm": 0.3386153280735016, "learning_rate": 8e-05, "loss": 1.5795, "step": 6079 }, { "epoch": 1.337402397448587, "grad_norm": 0.3352588415145874, "learning_rate": 8e-05, "loss": 1.706, "step": 6080 }, { "epoch": 1.3376223468602222, "grad_norm": 0.3381902873516083, "learning_rate": 8e-05, "loss": 1.4928, "step": 6081 }, { "epoch": 1.3378422962718575, "grad_norm": 0.3454771041870117, "learning_rate": 8e-05, "loss": 1.6505, "step": 6082 }, { "epoch": 1.3380622456834927, "grad_norm": 0.3402385711669922, "learning_rate": 8e-05, "loss": 1.5397, "step": 6083 }, { "epoch": 1.338282195095128, "grad_norm": 0.3605969548225403, "learning_rate": 8e-05, "loss": 1.6552, "step": 6084 }, { "epoch": 1.3385021445067635, "grad_norm": 0.3446277976036072, "learning_rate": 8e-05, "loss": 1.7545, "step": 6085 }, { "epoch": 1.3387220939183988, "grad_norm": 0.34508708119392395, "learning_rate": 8e-05, "loss": 1.7267, "step": 6086 }, { "epoch": 1.338942043330034, "grad_norm": 0.32693588733673096, "learning_rate": 8e-05, "loss": 1.4771, "step": 6087 }, { "epoch": 1.3391619927416694, "grad_norm": 0.327714204788208, "learning_rate": 8e-05, "loss": 1.5234, "step": 6088 }, { "epoch": 1.3393819421533046, "grad_norm": 0.33824995160102844, "learning_rate": 8e-05, "loss": 1.6057, "step": 6089 }, { "epoch": 1.3396018915649401, "grad_norm": 0.3354800343513489, "learning_rate": 8e-05, "loss": 1.6667, "step": 6090 }, { "epoch": 1.3398218409765754, "grad_norm": 0.3188812732696533, "learning_rate": 8e-05, "loss": 1.5215, "step": 6091 }, { "epoch": 1.3400417903882107, "grad_norm": 0.330102801322937, "learning_rate": 8e-05, "loss": 1.6047, "step": 6092 }, { "epoch": 1.340261739799846, "grad_norm": 0.32886308431625366, "learning_rate": 8e-05, "loss": 1.63, "step": 6093 }, { "epoch": 1.3404816892114813, "grad_norm": 0.33530691266059875, "learning_rate": 8e-05, "loss": 1.6181, "step": 6094 }, { "epoch": 1.3407016386231168, "grad_norm": 0.33497610688209534, "learning_rate": 8e-05, "loss": 1.4983, "step": 6095 }, { "epoch": 1.340921588034752, "grad_norm": 0.3412376642227173, "learning_rate": 8e-05, "loss": 1.7632, "step": 6096 }, { "epoch": 1.3411415374463873, "grad_norm": 0.36891984939575195, "learning_rate": 8e-05, "loss": 1.7677, "step": 6097 }, { "epoch": 1.3413614868580226, "grad_norm": 0.3418654203414917, "learning_rate": 8e-05, "loss": 1.4719, "step": 6098 }, { "epoch": 1.3415814362696579, "grad_norm": 0.34084224700927734, "learning_rate": 8e-05, "loss": 1.5338, "step": 6099 }, { "epoch": 1.3418013856812934, "grad_norm": 0.3422478437423706, "learning_rate": 8e-05, "loss": 1.4998, "step": 6100 }, { "epoch": 1.3420213350929286, "grad_norm": 0.39009174704551697, "learning_rate": 8e-05, "loss": 1.5994, "step": 6101 }, { "epoch": 1.342241284504564, "grad_norm": 0.32568198442459106, "learning_rate": 8e-05, "loss": 1.4095, "step": 6102 }, { "epoch": 1.3424612339161992, "grad_norm": 0.40706753730773926, "learning_rate": 8e-05, "loss": 1.7073, "step": 6103 }, { "epoch": 1.3426811833278345, "grad_norm": 0.35202789306640625, "learning_rate": 8e-05, "loss": 1.5857, "step": 6104 }, { "epoch": 1.34290113273947, "grad_norm": 0.33614814281463623, "learning_rate": 8e-05, "loss": 1.4665, "step": 6105 }, { "epoch": 1.3431210821511053, "grad_norm": 0.33402296900749207, "learning_rate": 8e-05, "loss": 1.5633, "step": 6106 }, { "epoch": 1.3433410315627405, "grad_norm": 0.3564693033695221, "learning_rate": 8e-05, "loss": 1.6523, "step": 6107 }, { "epoch": 1.3435609809743758, "grad_norm": 0.3516257703304291, "learning_rate": 8e-05, "loss": 1.6705, "step": 6108 }, { "epoch": 1.343780930386011, "grad_norm": 0.3181980848312378, "learning_rate": 8e-05, "loss": 1.5252, "step": 6109 }, { "epoch": 1.3440008797976466, "grad_norm": 0.3314395844936371, "learning_rate": 8e-05, "loss": 1.5808, "step": 6110 }, { "epoch": 1.3442208292092819, "grad_norm": 0.33175402879714966, "learning_rate": 8e-05, "loss": 1.4106, "step": 6111 }, { "epoch": 1.3444407786209172, "grad_norm": 0.32907718420028687, "learning_rate": 8e-05, "loss": 1.5173, "step": 6112 }, { "epoch": 1.3446607280325524, "grad_norm": 0.38194674253463745, "learning_rate": 8e-05, "loss": 1.6902, "step": 6113 }, { "epoch": 1.3448806774441877, "grad_norm": 0.33970892429351807, "learning_rate": 8e-05, "loss": 1.5917, "step": 6114 }, { "epoch": 1.3451006268558232, "grad_norm": 0.3336353600025177, "learning_rate": 8e-05, "loss": 1.6453, "step": 6115 }, { "epoch": 1.3453205762674585, "grad_norm": 0.3385280668735504, "learning_rate": 8e-05, "loss": 1.4609, "step": 6116 }, { "epoch": 1.3455405256790938, "grad_norm": 0.35183897614479065, "learning_rate": 8e-05, "loss": 1.7426, "step": 6117 }, { "epoch": 1.3457604750907293, "grad_norm": 0.31406834721565247, "learning_rate": 8e-05, "loss": 1.618, "step": 6118 }, { "epoch": 1.3459804245023643, "grad_norm": 0.3404744863510132, "learning_rate": 8e-05, "loss": 1.5924, "step": 6119 }, { "epoch": 1.3462003739139998, "grad_norm": 0.34450453519821167, "learning_rate": 8e-05, "loss": 1.7079, "step": 6120 }, { "epoch": 1.346420323325635, "grad_norm": 0.3417426645755768, "learning_rate": 8e-05, "loss": 1.6454, "step": 6121 }, { "epoch": 1.3466402727372704, "grad_norm": 0.34184595942497253, "learning_rate": 8e-05, "loss": 1.5749, "step": 6122 }, { "epoch": 1.3468602221489059, "grad_norm": 0.33964279294013977, "learning_rate": 8e-05, "loss": 1.6135, "step": 6123 }, { "epoch": 1.347080171560541, "grad_norm": 0.34088191390037537, "learning_rate": 8e-05, "loss": 1.6717, "step": 6124 }, { "epoch": 1.3473001209721764, "grad_norm": 0.3342702090740204, "learning_rate": 8e-05, "loss": 1.5944, "step": 6125 }, { "epoch": 1.3475200703838117, "grad_norm": 0.33129045367240906, "learning_rate": 8e-05, "loss": 1.4946, "step": 6126 }, { "epoch": 1.347740019795447, "grad_norm": 0.3411683738231659, "learning_rate": 8e-05, "loss": 1.6419, "step": 6127 }, { "epoch": 1.3479599692070825, "grad_norm": 0.36539292335510254, "learning_rate": 8e-05, "loss": 1.5853, "step": 6128 }, { "epoch": 1.3481799186187178, "grad_norm": 0.3433106541633606, "learning_rate": 8e-05, "loss": 1.6862, "step": 6129 }, { "epoch": 1.348399868030353, "grad_norm": 0.32226404547691345, "learning_rate": 8e-05, "loss": 1.5346, "step": 6130 }, { "epoch": 1.3486198174419883, "grad_norm": 0.34784388542175293, "learning_rate": 8e-05, "loss": 1.6034, "step": 6131 }, { "epoch": 1.3488397668536236, "grad_norm": 0.3294549882411957, "learning_rate": 8e-05, "loss": 1.3449, "step": 6132 }, { "epoch": 1.3490597162652591, "grad_norm": 0.33546581864356995, "learning_rate": 8e-05, "loss": 1.5107, "step": 6133 }, { "epoch": 1.3492796656768944, "grad_norm": 0.34550395607948303, "learning_rate": 8e-05, "loss": 1.604, "step": 6134 }, { "epoch": 1.3494996150885297, "grad_norm": 0.3213731050491333, "learning_rate": 8e-05, "loss": 1.5861, "step": 6135 }, { "epoch": 1.349719564500165, "grad_norm": 0.3776575028896332, "learning_rate": 8e-05, "loss": 1.7443, "step": 6136 }, { "epoch": 1.3499395139118002, "grad_norm": 0.35273638367652893, "learning_rate": 8e-05, "loss": 1.5416, "step": 6137 }, { "epoch": 1.3501594633234357, "grad_norm": 0.34917929768562317, "learning_rate": 8e-05, "loss": 1.6412, "step": 6138 }, { "epoch": 1.350379412735071, "grad_norm": 0.37784862518310547, "learning_rate": 8e-05, "loss": 1.562, "step": 6139 }, { "epoch": 1.3505993621467063, "grad_norm": 0.36981382966041565, "learning_rate": 8e-05, "loss": 1.6569, "step": 6140 }, { "epoch": 1.3508193115583416, "grad_norm": 0.4117145538330078, "learning_rate": 8e-05, "loss": 1.6745, "step": 6141 }, { "epoch": 1.3510392609699768, "grad_norm": 0.31574690341949463, "learning_rate": 8e-05, "loss": 1.3842, "step": 6142 }, { "epoch": 1.3512592103816123, "grad_norm": 0.35847634077072144, "learning_rate": 8e-05, "loss": 1.626, "step": 6143 }, { "epoch": 1.3514791597932476, "grad_norm": 0.3471231162548065, "learning_rate": 8e-05, "loss": 1.5788, "step": 6144 }, { "epoch": 1.351699109204883, "grad_norm": 0.34332576394081116, "learning_rate": 8e-05, "loss": 1.5644, "step": 6145 }, { "epoch": 1.3519190586165182, "grad_norm": 0.34728413820266724, "learning_rate": 8e-05, "loss": 1.6439, "step": 6146 }, { "epoch": 1.3521390080281535, "grad_norm": 0.3261185586452484, "learning_rate": 8e-05, "loss": 1.5138, "step": 6147 }, { "epoch": 1.352358957439789, "grad_norm": 0.3247813284397125, "learning_rate": 8e-05, "loss": 1.6145, "step": 6148 }, { "epoch": 1.3525789068514242, "grad_norm": 0.3209856450557709, "learning_rate": 8e-05, "loss": 1.489, "step": 6149 }, { "epoch": 1.3527988562630595, "grad_norm": 0.35816818475723267, "learning_rate": 8e-05, "loss": 1.7027, "step": 6150 }, { "epoch": 1.3530188056746948, "grad_norm": 0.3334057033061981, "learning_rate": 8e-05, "loss": 1.4458, "step": 6151 }, { "epoch": 1.35323875508633, "grad_norm": 0.3271174430847168, "learning_rate": 8e-05, "loss": 1.5352, "step": 6152 }, { "epoch": 1.3534587044979656, "grad_norm": 0.33537885546684265, "learning_rate": 8e-05, "loss": 1.527, "step": 6153 }, { "epoch": 1.3536786539096008, "grad_norm": 0.3317923843860626, "learning_rate": 8e-05, "loss": 1.4132, "step": 6154 }, { "epoch": 1.3538986033212361, "grad_norm": 0.3589371144771576, "learning_rate": 8e-05, "loss": 1.6859, "step": 6155 }, { "epoch": 1.3541185527328714, "grad_norm": 0.38810229301452637, "learning_rate": 8e-05, "loss": 1.5985, "step": 6156 }, { "epoch": 1.3543385021445067, "grad_norm": 0.33283352851867676, "learning_rate": 8e-05, "loss": 1.5023, "step": 6157 }, { "epoch": 1.3545584515561422, "grad_norm": 0.3441733121871948, "learning_rate": 8e-05, "loss": 1.6131, "step": 6158 }, { "epoch": 1.3547784009677775, "grad_norm": 0.33491891622543335, "learning_rate": 8e-05, "loss": 1.6062, "step": 6159 }, { "epoch": 1.3549983503794127, "grad_norm": 0.3391541540622711, "learning_rate": 8e-05, "loss": 1.6287, "step": 6160 }, { "epoch": 1.355218299791048, "grad_norm": 0.3537774980068207, "learning_rate": 8e-05, "loss": 1.6025, "step": 6161 }, { "epoch": 1.3554382492026833, "grad_norm": 0.33382049202919006, "learning_rate": 8e-05, "loss": 1.529, "step": 6162 }, { "epoch": 1.3556581986143188, "grad_norm": 0.33442288637161255, "learning_rate": 8e-05, "loss": 1.4478, "step": 6163 }, { "epoch": 1.355878148025954, "grad_norm": 0.322768896818161, "learning_rate": 8e-05, "loss": 1.4373, "step": 6164 }, { "epoch": 1.3560980974375894, "grad_norm": 0.3699006140232086, "learning_rate": 8e-05, "loss": 1.6407, "step": 6165 }, { "epoch": 1.3563180468492246, "grad_norm": 0.33188489079475403, "learning_rate": 8e-05, "loss": 1.4876, "step": 6166 }, { "epoch": 1.35653799626086, "grad_norm": 0.32916921377182007, "learning_rate": 8e-05, "loss": 1.5689, "step": 6167 }, { "epoch": 1.3567579456724954, "grad_norm": 0.3462557792663574, "learning_rate": 8e-05, "loss": 1.6924, "step": 6168 }, { "epoch": 1.3569778950841307, "grad_norm": 0.4319455921649933, "learning_rate": 8e-05, "loss": 1.6558, "step": 6169 }, { "epoch": 1.357197844495766, "grad_norm": 0.3416346609592438, "learning_rate": 8e-05, "loss": 1.6414, "step": 6170 }, { "epoch": 1.3574177939074012, "grad_norm": 0.3414340913295746, "learning_rate": 8e-05, "loss": 1.582, "step": 6171 }, { "epoch": 1.3576377433190365, "grad_norm": 0.34703901410102844, "learning_rate": 8e-05, "loss": 1.5495, "step": 6172 }, { "epoch": 1.357857692730672, "grad_norm": 0.36174365878105164, "learning_rate": 8e-05, "loss": 1.7004, "step": 6173 }, { "epoch": 1.3580776421423073, "grad_norm": 0.34451255202293396, "learning_rate": 8e-05, "loss": 1.6568, "step": 6174 }, { "epoch": 1.3582975915539426, "grad_norm": 0.3399926722049713, "learning_rate": 8e-05, "loss": 1.5917, "step": 6175 }, { "epoch": 1.3585175409655779, "grad_norm": 0.37050920724868774, "learning_rate": 8e-05, "loss": 1.6376, "step": 6176 }, { "epoch": 1.3587374903772131, "grad_norm": 0.35580915212631226, "learning_rate": 8e-05, "loss": 1.6184, "step": 6177 }, { "epoch": 1.3589574397888486, "grad_norm": 0.34056130051612854, "learning_rate": 8e-05, "loss": 1.6404, "step": 6178 }, { "epoch": 1.359177389200484, "grad_norm": 0.34386539459228516, "learning_rate": 8e-05, "loss": 1.6432, "step": 6179 }, { "epoch": 1.3593973386121192, "grad_norm": 0.37023717164993286, "learning_rate": 8e-05, "loss": 1.7056, "step": 6180 }, { "epoch": 1.3596172880237545, "grad_norm": 0.3393137753009796, "learning_rate": 8e-05, "loss": 1.6374, "step": 6181 }, { "epoch": 1.3598372374353898, "grad_norm": 0.3516847491264343, "learning_rate": 8e-05, "loss": 1.745, "step": 6182 }, { "epoch": 1.3600571868470253, "grad_norm": 0.3341298997402191, "learning_rate": 8e-05, "loss": 1.5835, "step": 6183 }, { "epoch": 1.3602771362586605, "grad_norm": 0.36878374218940735, "learning_rate": 8e-05, "loss": 1.5703, "step": 6184 }, { "epoch": 1.3604970856702958, "grad_norm": 0.3369264006614685, "learning_rate": 8e-05, "loss": 1.566, "step": 6185 }, { "epoch": 1.360717035081931, "grad_norm": 0.3325944244861603, "learning_rate": 8e-05, "loss": 1.633, "step": 6186 }, { "epoch": 1.3609369844935664, "grad_norm": 0.3398142457008362, "learning_rate": 8e-05, "loss": 1.6014, "step": 6187 }, { "epoch": 1.3611569339052019, "grad_norm": 0.3369632661342621, "learning_rate": 8e-05, "loss": 1.5657, "step": 6188 }, { "epoch": 1.3613768833168371, "grad_norm": 0.34285077452659607, "learning_rate": 8e-05, "loss": 1.6632, "step": 6189 }, { "epoch": 1.3615968327284724, "grad_norm": 0.3380112946033478, "learning_rate": 8e-05, "loss": 1.5183, "step": 6190 }, { "epoch": 1.3618167821401077, "grad_norm": 0.336857408285141, "learning_rate": 8e-05, "loss": 1.6007, "step": 6191 }, { "epoch": 1.362036731551743, "grad_norm": 0.36375364661216736, "learning_rate": 8e-05, "loss": 1.706, "step": 6192 }, { "epoch": 1.3622566809633785, "grad_norm": 0.339153915643692, "learning_rate": 8e-05, "loss": 1.5969, "step": 6193 }, { "epoch": 1.3624766303750138, "grad_norm": 0.33684271574020386, "learning_rate": 8e-05, "loss": 1.4626, "step": 6194 }, { "epoch": 1.362696579786649, "grad_norm": 0.34184730052948, "learning_rate": 8e-05, "loss": 1.5763, "step": 6195 }, { "epoch": 1.3629165291982843, "grad_norm": 0.3394002616405487, "learning_rate": 8e-05, "loss": 1.5844, "step": 6196 }, { "epoch": 1.3631364786099196, "grad_norm": 0.35448893904685974, "learning_rate": 8e-05, "loss": 1.8175, "step": 6197 }, { "epoch": 1.363356428021555, "grad_norm": 0.3401862680912018, "learning_rate": 8e-05, "loss": 1.6144, "step": 6198 }, { "epoch": 1.3635763774331904, "grad_norm": 0.3835812509059906, "learning_rate": 8e-05, "loss": 1.8419, "step": 6199 }, { "epoch": 1.3637963268448257, "grad_norm": 0.3619830906391144, "learning_rate": 8e-05, "loss": 1.6994, "step": 6200 }, { "epoch": 1.364016276256461, "grad_norm": 0.366708368062973, "learning_rate": 8e-05, "loss": 1.5202, "step": 6201 }, { "epoch": 1.3642362256680962, "grad_norm": 0.33894166350364685, "learning_rate": 8e-05, "loss": 1.5951, "step": 6202 }, { "epoch": 1.3644561750797317, "grad_norm": 0.34377342462539673, "learning_rate": 8e-05, "loss": 1.5075, "step": 6203 }, { "epoch": 1.364676124491367, "grad_norm": 0.3292672336101532, "learning_rate": 8e-05, "loss": 1.7103, "step": 6204 }, { "epoch": 1.3648960739030023, "grad_norm": 0.3415939211845398, "learning_rate": 8e-05, "loss": 1.6963, "step": 6205 }, { "epoch": 1.3651160233146378, "grad_norm": 0.3717600107192993, "learning_rate": 8e-05, "loss": 1.6212, "step": 6206 }, { "epoch": 1.3653359727262728, "grad_norm": 0.37164464592933655, "learning_rate": 8e-05, "loss": 1.6683, "step": 6207 }, { "epoch": 1.3655559221379083, "grad_norm": 0.35787010192871094, "learning_rate": 8e-05, "loss": 1.5986, "step": 6208 }, { "epoch": 1.3657758715495436, "grad_norm": 0.3477771580219269, "learning_rate": 8e-05, "loss": 1.6184, "step": 6209 }, { "epoch": 1.3659958209611789, "grad_norm": 0.3509179949760437, "learning_rate": 8e-05, "loss": 1.6144, "step": 6210 }, { "epoch": 1.3662157703728144, "grad_norm": 0.3502595126628876, "learning_rate": 8e-05, "loss": 1.6048, "step": 6211 }, { "epoch": 1.3664357197844494, "grad_norm": 0.3375539779663086, "learning_rate": 8e-05, "loss": 1.5718, "step": 6212 }, { "epoch": 1.366655669196085, "grad_norm": 0.3316386342048645, "learning_rate": 8e-05, "loss": 1.5097, "step": 6213 }, { "epoch": 1.3668756186077202, "grad_norm": 0.3383817672729492, "learning_rate": 8e-05, "loss": 1.4851, "step": 6214 }, { "epoch": 1.3670955680193555, "grad_norm": 0.327927827835083, "learning_rate": 8e-05, "loss": 1.4641, "step": 6215 }, { "epoch": 1.367315517430991, "grad_norm": 0.3849707245826721, "learning_rate": 8e-05, "loss": 1.6369, "step": 6216 }, { "epoch": 1.367535466842626, "grad_norm": 0.3638830780982971, "learning_rate": 8e-05, "loss": 1.6881, "step": 6217 }, { "epoch": 1.3677554162542616, "grad_norm": 0.359039306640625, "learning_rate": 8e-05, "loss": 1.7788, "step": 6218 }, { "epoch": 1.3679753656658968, "grad_norm": 0.34711554646492004, "learning_rate": 8e-05, "loss": 1.6075, "step": 6219 }, { "epoch": 1.3681953150775321, "grad_norm": 0.3325296640396118, "learning_rate": 8e-05, "loss": 1.6519, "step": 6220 }, { "epoch": 1.3684152644891676, "grad_norm": 0.32509225606918335, "learning_rate": 8e-05, "loss": 1.5603, "step": 6221 }, { "epoch": 1.368635213900803, "grad_norm": 0.36574751138687134, "learning_rate": 8e-05, "loss": 1.6236, "step": 6222 }, { "epoch": 1.3688551633124382, "grad_norm": 0.3031999170780182, "learning_rate": 8e-05, "loss": 1.383, "step": 6223 }, { "epoch": 1.3690751127240735, "grad_norm": 0.3497132360935211, "learning_rate": 8e-05, "loss": 1.4652, "step": 6224 }, { "epoch": 1.3692950621357087, "grad_norm": 0.3514918386936188, "learning_rate": 8e-05, "loss": 1.606, "step": 6225 }, { "epoch": 1.3695150115473442, "grad_norm": 0.346096932888031, "learning_rate": 8e-05, "loss": 1.4982, "step": 6226 }, { "epoch": 1.3697349609589795, "grad_norm": 0.33228251338005066, "learning_rate": 8e-05, "loss": 1.4524, "step": 6227 }, { "epoch": 1.3699549103706148, "grad_norm": 0.34206870198249817, "learning_rate": 8e-05, "loss": 1.5716, "step": 6228 }, { "epoch": 1.37017485978225, "grad_norm": 0.34987732768058777, "learning_rate": 8e-05, "loss": 1.5996, "step": 6229 }, { "epoch": 1.3703948091938853, "grad_norm": 0.36433377861976624, "learning_rate": 8e-05, "loss": 1.6323, "step": 6230 }, { "epoch": 1.3706147586055208, "grad_norm": 0.33258551359176636, "learning_rate": 8e-05, "loss": 1.5219, "step": 6231 }, { "epoch": 1.3708347080171561, "grad_norm": 0.3504350781440735, "learning_rate": 8e-05, "loss": 1.6219, "step": 6232 }, { "epoch": 1.3710546574287914, "grad_norm": 0.3370988368988037, "learning_rate": 8e-05, "loss": 1.6839, "step": 6233 }, { "epoch": 1.3712746068404267, "grad_norm": 0.33251965045928955, "learning_rate": 8e-05, "loss": 1.6206, "step": 6234 }, { "epoch": 1.371494556252062, "grad_norm": 0.35078728199005127, "learning_rate": 8e-05, "loss": 1.6114, "step": 6235 }, { "epoch": 1.3717145056636975, "grad_norm": 0.33817920088768005, "learning_rate": 8e-05, "loss": 1.5523, "step": 6236 }, { "epoch": 1.3719344550753327, "grad_norm": 0.3435843884944916, "learning_rate": 8e-05, "loss": 1.533, "step": 6237 }, { "epoch": 1.372154404486968, "grad_norm": 0.34475386142730713, "learning_rate": 8e-05, "loss": 1.5621, "step": 6238 }, { "epoch": 1.3723743538986033, "grad_norm": 0.3657839000225067, "learning_rate": 8e-05, "loss": 1.5071, "step": 6239 }, { "epoch": 1.3725943033102386, "grad_norm": 0.3262460231781006, "learning_rate": 8e-05, "loss": 1.5094, "step": 6240 }, { "epoch": 1.372814252721874, "grad_norm": 0.3221539258956909, "learning_rate": 8e-05, "loss": 1.4319, "step": 6241 }, { "epoch": 1.3730342021335094, "grad_norm": 0.3705308139324188, "learning_rate": 8e-05, "loss": 1.4048, "step": 6242 }, { "epoch": 1.3732541515451446, "grad_norm": 0.3256356418132782, "learning_rate": 8e-05, "loss": 1.6476, "step": 6243 }, { "epoch": 1.37347410095678, "grad_norm": 0.3506757318973541, "learning_rate": 8e-05, "loss": 1.5438, "step": 6244 }, { "epoch": 1.3736940503684152, "grad_norm": 0.3348367214202881, "learning_rate": 8e-05, "loss": 1.5212, "step": 6245 }, { "epoch": 1.3739139997800507, "grad_norm": 0.3484991788864136, "learning_rate": 8e-05, "loss": 1.6099, "step": 6246 }, { "epoch": 1.374133949191686, "grad_norm": 0.32988908886909485, "learning_rate": 8e-05, "loss": 1.4198, "step": 6247 }, { "epoch": 1.3743538986033212, "grad_norm": 0.32388877868652344, "learning_rate": 8e-05, "loss": 1.5839, "step": 6248 }, { "epoch": 1.3745738480149565, "grad_norm": 0.3409726321697235, "learning_rate": 8e-05, "loss": 1.7286, "step": 6249 }, { "epoch": 1.3747937974265918, "grad_norm": 0.361713707447052, "learning_rate": 8e-05, "loss": 1.5543, "step": 6250 }, { "epoch": 1.3750137468382273, "grad_norm": 0.35115697979927063, "learning_rate": 8e-05, "loss": 1.7025, "step": 6251 }, { "epoch": 1.3752336962498626, "grad_norm": 0.3324168622493744, "learning_rate": 8e-05, "loss": 1.5712, "step": 6252 }, { "epoch": 1.3754536456614979, "grad_norm": 0.3380112946033478, "learning_rate": 8e-05, "loss": 1.6439, "step": 6253 }, { "epoch": 1.3756735950731331, "grad_norm": 0.33820661902427673, "learning_rate": 8e-05, "loss": 1.4764, "step": 6254 }, { "epoch": 1.3758935444847684, "grad_norm": 0.35167208313941956, "learning_rate": 8e-05, "loss": 1.6189, "step": 6255 }, { "epoch": 1.376113493896404, "grad_norm": 0.36171531677246094, "learning_rate": 8e-05, "loss": 1.587, "step": 6256 }, { "epoch": 1.3763334433080392, "grad_norm": 0.3694782853126526, "learning_rate": 8e-05, "loss": 1.6981, "step": 6257 }, { "epoch": 1.3765533927196745, "grad_norm": 0.3303854465484619, "learning_rate": 8e-05, "loss": 1.5669, "step": 6258 }, { "epoch": 1.3767733421313098, "grad_norm": 0.32498255372047424, "learning_rate": 8e-05, "loss": 1.3731, "step": 6259 }, { "epoch": 1.376993291542945, "grad_norm": 0.44123104214668274, "learning_rate": 8e-05, "loss": 1.6685, "step": 6260 }, { "epoch": 1.3772132409545805, "grad_norm": 0.4620372951030731, "learning_rate": 8e-05, "loss": 1.641, "step": 6261 }, { "epoch": 1.3774331903662158, "grad_norm": 0.32788774371147156, "learning_rate": 8e-05, "loss": 1.4266, "step": 6262 }, { "epoch": 1.377653139777851, "grad_norm": 0.39007601141929626, "learning_rate": 8e-05, "loss": 1.7429, "step": 6263 }, { "epoch": 1.3778730891894864, "grad_norm": 0.3429759740829468, "learning_rate": 8e-05, "loss": 1.5271, "step": 6264 }, { "epoch": 1.3780930386011216, "grad_norm": 0.34217944741249084, "learning_rate": 8e-05, "loss": 1.4711, "step": 6265 }, { "epoch": 1.3783129880127571, "grad_norm": 0.3530440032482147, "learning_rate": 8e-05, "loss": 1.6546, "step": 6266 }, { "epoch": 1.3785329374243924, "grad_norm": 0.36109402775764465, "learning_rate": 8e-05, "loss": 1.6728, "step": 6267 }, { "epoch": 1.3787528868360277, "grad_norm": 0.35064497590065, "learning_rate": 8e-05, "loss": 1.7829, "step": 6268 }, { "epoch": 1.378972836247663, "grad_norm": 0.335161417722702, "learning_rate": 8e-05, "loss": 1.6055, "step": 6269 }, { "epoch": 1.3791927856592983, "grad_norm": 0.3635447025299072, "learning_rate": 8e-05, "loss": 1.6058, "step": 6270 }, { "epoch": 1.3794127350709338, "grad_norm": 0.3513880968093872, "learning_rate": 8e-05, "loss": 1.5386, "step": 6271 }, { "epoch": 1.379632684482569, "grad_norm": 0.32569456100463867, "learning_rate": 8e-05, "loss": 1.5416, "step": 6272 }, { "epoch": 1.3798526338942043, "grad_norm": 0.3456653356552124, "learning_rate": 8e-05, "loss": 1.6049, "step": 6273 }, { "epoch": 1.3800725833058396, "grad_norm": 0.3664407730102539, "learning_rate": 8e-05, "loss": 1.7344, "step": 6274 }, { "epoch": 1.3802925327174749, "grad_norm": 0.3387892544269562, "learning_rate": 8e-05, "loss": 1.5703, "step": 6275 }, { "epoch": 1.3805124821291104, "grad_norm": 0.3481842577457428, "learning_rate": 8e-05, "loss": 1.6632, "step": 6276 }, { "epoch": 1.3807324315407457, "grad_norm": 0.33127450942993164, "learning_rate": 8e-05, "loss": 1.5135, "step": 6277 }, { "epoch": 1.380952380952381, "grad_norm": 0.33055350184440613, "learning_rate": 8e-05, "loss": 1.5248, "step": 6278 }, { "epoch": 1.3811723303640162, "grad_norm": 0.3485487401485443, "learning_rate": 8e-05, "loss": 1.6116, "step": 6279 }, { "epoch": 1.3813922797756515, "grad_norm": 0.3212777078151703, "learning_rate": 8e-05, "loss": 1.4052, "step": 6280 }, { "epoch": 1.381612229187287, "grad_norm": 0.3521946668624878, "learning_rate": 8e-05, "loss": 1.5901, "step": 6281 }, { "epoch": 1.3818321785989223, "grad_norm": 0.3370744287967682, "learning_rate": 8e-05, "loss": 1.4954, "step": 6282 }, { "epoch": 1.3820521280105575, "grad_norm": 0.3405938148498535, "learning_rate": 8e-05, "loss": 1.6633, "step": 6283 }, { "epoch": 1.3822720774221928, "grad_norm": 0.34075847268104553, "learning_rate": 8e-05, "loss": 1.6076, "step": 6284 }, { "epoch": 1.382492026833828, "grad_norm": 0.3250913619995117, "learning_rate": 8e-05, "loss": 1.6124, "step": 6285 }, { "epoch": 1.3827119762454636, "grad_norm": 0.3292626738548279, "learning_rate": 8e-05, "loss": 1.4267, "step": 6286 }, { "epoch": 1.3829319256570989, "grad_norm": 0.37748095393180847, "learning_rate": 8e-05, "loss": 1.6602, "step": 6287 }, { "epoch": 1.3831518750687342, "grad_norm": 0.34243136644363403, "learning_rate": 8e-05, "loss": 1.6223, "step": 6288 }, { "epoch": 1.3833718244803694, "grad_norm": 0.34067806601524353, "learning_rate": 8e-05, "loss": 1.557, "step": 6289 }, { "epoch": 1.3835917738920047, "grad_norm": 0.3378852307796478, "learning_rate": 8e-05, "loss": 1.6196, "step": 6290 }, { "epoch": 1.3838117233036402, "grad_norm": 0.3414228856563568, "learning_rate": 8e-05, "loss": 1.5248, "step": 6291 }, { "epoch": 1.3840316727152755, "grad_norm": 0.33289843797683716, "learning_rate": 8e-05, "loss": 1.4027, "step": 6292 }, { "epoch": 1.3842516221269108, "grad_norm": 0.34995776414871216, "learning_rate": 8e-05, "loss": 1.5465, "step": 6293 }, { "epoch": 1.384471571538546, "grad_norm": 0.3784621059894562, "learning_rate": 8e-05, "loss": 1.6425, "step": 6294 }, { "epoch": 1.3846915209501813, "grad_norm": 0.3414026200771332, "learning_rate": 8e-05, "loss": 1.5512, "step": 6295 }, { "epoch": 1.3849114703618168, "grad_norm": 0.33340024948120117, "learning_rate": 8e-05, "loss": 1.4191, "step": 6296 }, { "epoch": 1.385131419773452, "grad_norm": 0.32080551981925964, "learning_rate": 8e-05, "loss": 1.4264, "step": 6297 }, { "epoch": 1.3853513691850874, "grad_norm": 0.39550337195396423, "learning_rate": 8e-05, "loss": 1.5704, "step": 6298 }, { "epoch": 1.385571318596723, "grad_norm": 0.376250684261322, "learning_rate": 8e-05, "loss": 1.6011, "step": 6299 }, { "epoch": 1.385791268008358, "grad_norm": 0.3607538640499115, "learning_rate": 8e-05, "loss": 1.6804, "step": 6300 }, { "epoch": 1.3860112174199934, "grad_norm": 0.3338264524936676, "learning_rate": 8e-05, "loss": 1.5237, "step": 6301 }, { "epoch": 1.3862311668316287, "grad_norm": 0.33942732214927673, "learning_rate": 8e-05, "loss": 1.5746, "step": 6302 }, { "epoch": 1.386451116243264, "grad_norm": 0.3459453880786896, "learning_rate": 8e-05, "loss": 1.5318, "step": 6303 }, { "epoch": 1.3866710656548995, "grad_norm": 0.34469082951545715, "learning_rate": 8e-05, "loss": 1.4773, "step": 6304 }, { "epoch": 1.3868910150665346, "grad_norm": 0.3395957946777344, "learning_rate": 8e-05, "loss": 1.6573, "step": 6305 }, { "epoch": 1.38711096447817, "grad_norm": 0.34206339716911316, "learning_rate": 8e-05, "loss": 1.6555, "step": 6306 }, { "epoch": 1.3873309138898053, "grad_norm": 0.33416327834129333, "learning_rate": 8e-05, "loss": 1.536, "step": 6307 }, { "epoch": 1.3875508633014406, "grad_norm": 0.3453490138053894, "learning_rate": 8e-05, "loss": 1.5958, "step": 6308 }, { "epoch": 1.3877708127130761, "grad_norm": 0.37293457984924316, "learning_rate": 8e-05, "loss": 1.7452, "step": 6309 }, { "epoch": 1.3879907621247114, "grad_norm": 0.3417176604270935, "learning_rate": 8e-05, "loss": 1.4661, "step": 6310 }, { "epoch": 1.3882107115363467, "grad_norm": 0.3569588363170624, "learning_rate": 8e-05, "loss": 1.6901, "step": 6311 }, { "epoch": 1.388430660947982, "grad_norm": 0.34806254506111145, "learning_rate": 8e-05, "loss": 1.5801, "step": 6312 }, { "epoch": 1.3886506103596172, "grad_norm": 0.3490111231803894, "learning_rate": 8e-05, "loss": 1.6077, "step": 6313 }, { "epoch": 1.3888705597712527, "grad_norm": 0.3453671634197235, "learning_rate": 8e-05, "loss": 1.7582, "step": 6314 }, { "epoch": 1.389090509182888, "grad_norm": 0.33546939492225647, "learning_rate": 8e-05, "loss": 1.5474, "step": 6315 }, { "epoch": 1.3893104585945233, "grad_norm": 0.33536821603775024, "learning_rate": 8e-05, "loss": 1.575, "step": 6316 }, { "epoch": 1.3895304080061586, "grad_norm": 0.35957640409469604, "learning_rate": 8e-05, "loss": 1.5613, "step": 6317 }, { "epoch": 1.3897503574177938, "grad_norm": 0.3367151618003845, "learning_rate": 8e-05, "loss": 1.5289, "step": 6318 }, { "epoch": 1.3899703068294293, "grad_norm": 0.3609754741191864, "learning_rate": 8e-05, "loss": 1.7404, "step": 6319 }, { "epoch": 1.3901902562410646, "grad_norm": 0.34525349736213684, "learning_rate": 8e-05, "loss": 1.6802, "step": 6320 }, { "epoch": 1.3904102056527, "grad_norm": 0.3471965789794922, "learning_rate": 8e-05, "loss": 1.4755, "step": 6321 }, { "epoch": 1.3906301550643352, "grad_norm": 0.3470200300216675, "learning_rate": 8e-05, "loss": 1.5701, "step": 6322 }, { "epoch": 1.3908501044759705, "grad_norm": 0.3666965961456299, "learning_rate": 8e-05, "loss": 1.7152, "step": 6323 }, { "epoch": 1.391070053887606, "grad_norm": 0.3445170819759369, "learning_rate": 8e-05, "loss": 1.4264, "step": 6324 }, { "epoch": 1.3912900032992412, "grad_norm": 0.3347538113594055, "learning_rate": 8e-05, "loss": 1.6252, "step": 6325 }, { "epoch": 1.3915099527108765, "grad_norm": 0.34245362877845764, "learning_rate": 8e-05, "loss": 1.4485, "step": 6326 }, { "epoch": 1.3917299021225118, "grad_norm": 0.4012244641780853, "learning_rate": 8e-05, "loss": 1.5907, "step": 6327 }, { "epoch": 1.391949851534147, "grad_norm": 0.32273146510124207, "learning_rate": 8e-05, "loss": 1.5013, "step": 6328 }, { "epoch": 1.3921698009457826, "grad_norm": 0.3513316214084625, "learning_rate": 8e-05, "loss": 1.5516, "step": 6329 }, { "epoch": 1.3923897503574179, "grad_norm": 0.3383953869342804, "learning_rate": 8e-05, "loss": 1.6963, "step": 6330 }, { "epoch": 1.3926096997690531, "grad_norm": 0.3660552203655243, "learning_rate": 8e-05, "loss": 1.5904, "step": 6331 }, { "epoch": 1.3928296491806884, "grad_norm": 0.35031604766845703, "learning_rate": 8e-05, "loss": 1.5968, "step": 6332 }, { "epoch": 1.3930495985923237, "grad_norm": 0.35155749320983887, "learning_rate": 8e-05, "loss": 1.6023, "step": 6333 }, { "epoch": 1.3932695480039592, "grad_norm": 0.333053320646286, "learning_rate": 8e-05, "loss": 1.501, "step": 6334 }, { "epoch": 1.3934894974155945, "grad_norm": 0.35252395272254944, "learning_rate": 8e-05, "loss": 1.6859, "step": 6335 }, { "epoch": 1.3937094468272297, "grad_norm": 0.3275587856769562, "learning_rate": 8e-05, "loss": 1.5643, "step": 6336 }, { "epoch": 1.393929396238865, "grad_norm": 0.3361843526363373, "learning_rate": 8e-05, "loss": 1.549, "step": 6337 }, { "epoch": 1.3941493456505003, "grad_norm": 0.3463844060897827, "learning_rate": 8e-05, "loss": 1.7083, "step": 6338 }, { "epoch": 1.3943692950621358, "grad_norm": 0.333042174577713, "learning_rate": 8e-05, "loss": 1.4936, "step": 6339 }, { "epoch": 1.394589244473771, "grad_norm": 0.35073599219322205, "learning_rate": 8e-05, "loss": 1.8159, "step": 6340 }, { "epoch": 1.3948091938854064, "grad_norm": 0.37173566222190857, "learning_rate": 8e-05, "loss": 1.5579, "step": 6341 }, { "epoch": 1.3950291432970416, "grad_norm": 0.365573525428772, "learning_rate": 8e-05, "loss": 1.6764, "step": 6342 }, { "epoch": 1.395249092708677, "grad_norm": 0.34005996584892273, "learning_rate": 8e-05, "loss": 1.5887, "step": 6343 }, { "epoch": 1.3954690421203124, "grad_norm": 0.33627253770828247, "learning_rate": 8e-05, "loss": 1.6063, "step": 6344 }, { "epoch": 1.3956889915319477, "grad_norm": 0.34296277165412903, "learning_rate": 8e-05, "loss": 1.6154, "step": 6345 }, { "epoch": 1.395908940943583, "grad_norm": 0.3517684042453766, "learning_rate": 8e-05, "loss": 1.6131, "step": 6346 }, { "epoch": 1.3961288903552183, "grad_norm": 0.3348885774612427, "learning_rate": 8e-05, "loss": 1.5087, "step": 6347 }, { "epoch": 1.3963488397668535, "grad_norm": 0.3435637056827545, "learning_rate": 8e-05, "loss": 1.5962, "step": 6348 }, { "epoch": 1.396568789178489, "grad_norm": 0.3576011061668396, "learning_rate": 8e-05, "loss": 1.8267, "step": 6349 }, { "epoch": 1.3967887385901243, "grad_norm": 0.34611353278160095, "learning_rate": 8e-05, "loss": 1.4837, "step": 6350 }, { "epoch": 1.3970086880017596, "grad_norm": 0.34550508856773376, "learning_rate": 8e-05, "loss": 1.6436, "step": 6351 }, { "epoch": 1.3972286374133949, "grad_norm": 0.3304874300956726, "learning_rate": 8e-05, "loss": 1.4328, "step": 6352 }, { "epoch": 1.3974485868250301, "grad_norm": 0.3279494345188141, "learning_rate": 8e-05, "loss": 1.4438, "step": 6353 }, { "epoch": 1.3976685362366656, "grad_norm": 0.3173222243785858, "learning_rate": 8e-05, "loss": 1.4958, "step": 6354 }, { "epoch": 1.397888485648301, "grad_norm": 0.40323036909103394, "learning_rate": 8e-05, "loss": 1.7498, "step": 6355 }, { "epoch": 1.3981084350599362, "grad_norm": 0.3319350481033325, "learning_rate": 8e-05, "loss": 1.4261, "step": 6356 }, { "epoch": 1.3983283844715715, "grad_norm": 0.32421156764030457, "learning_rate": 8e-05, "loss": 1.4682, "step": 6357 }, { "epoch": 1.3985483338832068, "grad_norm": 0.339459627866745, "learning_rate": 8e-05, "loss": 1.5019, "step": 6358 }, { "epoch": 1.3987682832948423, "grad_norm": 0.34449145197868347, "learning_rate": 8e-05, "loss": 1.6973, "step": 6359 }, { "epoch": 1.3989882327064775, "grad_norm": 0.35296133160591125, "learning_rate": 8e-05, "loss": 1.5997, "step": 6360 }, { "epoch": 1.3992081821181128, "grad_norm": 0.3450864553451538, "learning_rate": 8e-05, "loss": 1.6134, "step": 6361 }, { "epoch": 1.399428131529748, "grad_norm": 0.3539825975894928, "learning_rate": 8e-05, "loss": 1.529, "step": 6362 }, { "epoch": 1.3996480809413834, "grad_norm": 0.31548479199409485, "learning_rate": 8e-05, "loss": 1.4178, "step": 6363 }, { "epoch": 1.3998680303530189, "grad_norm": 0.33319228887557983, "learning_rate": 8e-05, "loss": 1.6905, "step": 6364 }, { "epoch": 1.4000879797646542, "grad_norm": 0.33604133129119873, "learning_rate": 8e-05, "loss": 1.4996, "step": 6365 }, { "epoch": 1.4003079291762894, "grad_norm": 0.3619355261325836, "learning_rate": 8e-05, "loss": 1.6289, "step": 6366 }, { "epoch": 1.4005278785879247, "grad_norm": 0.3530455529689789, "learning_rate": 8e-05, "loss": 1.7262, "step": 6367 }, { "epoch": 1.40074782799956, "grad_norm": 0.3333120346069336, "learning_rate": 8e-05, "loss": 1.5941, "step": 6368 }, { "epoch": 1.4009677774111955, "grad_norm": 0.36543095111846924, "learning_rate": 8e-05, "loss": 1.7595, "step": 6369 }, { "epoch": 1.4011877268228308, "grad_norm": 0.33672404289245605, "learning_rate": 8e-05, "loss": 1.7015, "step": 6370 }, { "epoch": 1.401407676234466, "grad_norm": 0.36616283655166626, "learning_rate": 8e-05, "loss": 1.6209, "step": 6371 }, { "epoch": 1.4016276256461013, "grad_norm": 0.32179492712020874, "learning_rate": 8e-05, "loss": 1.4733, "step": 6372 }, { "epoch": 1.4018475750577366, "grad_norm": 0.5398252010345459, "learning_rate": 8e-05, "loss": 1.5306, "step": 6373 }, { "epoch": 1.402067524469372, "grad_norm": 0.34317901730537415, "learning_rate": 8e-05, "loss": 1.615, "step": 6374 }, { "epoch": 1.4022874738810074, "grad_norm": 0.3554289638996124, "learning_rate": 8e-05, "loss": 1.5693, "step": 6375 }, { "epoch": 1.4025074232926427, "grad_norm": 0.4662148356437683, "learning_rate": 8e-05, "loss": 1.8054, "step": 6376 }, { "epoch": 1.402727372704278, "grad_norm": 0.3206388056278229, "learning_rate": 8e-05, "loss": 1.398, "step": 6377 }, { "epoch": 1.4029473221159132, "grad_norm": 0.3625994324684143, "learning_rate": 8e-05, "loss": 1.6903, "step": 6378 }, { "epoch": 1.4031672715275487, "grad_norm": 0.345123827457428, "learning_rate": 8e-05, "loss": 1.5011, "step": 6379 }, { "epoch": 1.403387220939184, "grad_norm": 0.3228575587272644, "learning_rate": 8e-05, "loss": 1.5447, "step": 6380 }, { "epoch": 1.4036071703508193, "grad_norm": 0.3308219313621521, "learning_rate": 8e-05, "loss": 1.4946, "step": 6381 }, { "epoch": 1.4038271197624546, "grad_norm": 0.35518044233322144, "learning_rate": 8e-05, "loss": 1.5457, "step": 6382 }, { "epoch": 1.4040470691740898, "grad_norm": 0.4154233932495117, "learning_rate": 8e-05, "loss": 1.6452, "step": 6383 }, { "epoch": 1.4042670185857253, "grad_norm": 0.3606014847755432, "learning_rate": 8e-05, "loss": 1.5859, "step": 6384 }, { "epoch": 1.4044869679973606, "grad_norm": 0.33071061968803406, "learning_rate": 8e-05, "loss": 1.6517, "step": 6385 }, { "epoch": 1.404706917408996, "grad_norm": 0.33083856105804443, "learning_rate": 8e-05, "loss": 1.5094, "step": 6386 }, { "epoch": 1.4049268668206314, "grad_norm": 0.35438406467437744, "learning_rate": 8e-05, "loss": 1.5938, "step": 6387 }, { "epoch": 1.4051468162322664, "grad_norm": 0.36129769682884216, "learning_rate": 8e-05, "loss": 1.7212, "step": 6388 }, { "epoch": 1.405366765643902, "grad_norm": 0.3530142307281494, "learning_rate": 8e-05, "loss": 1.6188, "step": 6389 }, { "epoch": 1.4055867150555372, "grad_norm": 0.3463714122772217, "learning_rate": 8e-05, "loss": 1.522, "step": 6390 }, { "epoch": 1.4058066644671725, "grad_norm": 0.33608314394950867, "learning_rate": 8e-05, "loss": 1.5169, "step": 6391 }, { "epoch": 1.406026613878808, "grad_norm": 0.37738099694252014, "learning_rate": 8e-05, "loss": 1.6251, "step": 6392 }, { "epoch": 1.406246563290443, "grad_norm": 0.336792528629303, "learning_rate": 8e-05, "loss": 1.6402, "step": 6393 }, { "epoch": 1.4064665127020786, "grad_norm": 0.32640552520751953, "learning_rate": 8e-05, "loss": 1.6194, "step": 6394 }, { "epoch": 1.4066864621137138, "grad_norm": 0.344288170337677, "learning_rate": 8e-05, "loss": 1.5654, "step": 6395 }, { "epoch": 1.4069064115253491, "grad_norm": 0.3453962206840515, "learning_rate": 8e-05, "loss": 1.5922, "step": 6396 }, { "epoch": 1.4071263609369846, "grad_norm": 0.3463610112667084, "learning_rate": 8e-05, "loss": 1.6141, "step": 6397 }, { "epoch": 1.40734631034862, "grad_norm": 0.3326776921749115, "learning_rate": 8e-05, "loss": 1.5611, "step": 6398 }, { "epoch": 1.4075662597602552, "grad_norm": 0.35130980610847473, "learning_rate": 8e-05, "loss": 1.4439, "step": 6399 }, { "epoch": 1.4077862091718905, "grad_norm": 0.332768052816391, "learning_rate": 8e-05, "loss": 1.5148, "step": 6400 }, { "epoch": 1.4080061585835257, "grad_norm": 0.33286088705062866, "learning_rate": 8e-05, "loss": 1.5069, "step": 6401 }, { "epoch": 1.4082261079951612, "grad_norm": 0.34897884726524353, "learning_rate": 8e-05, "loss": 1.5204, "step": 6402 }, { "epoch": 1.4084460574067965, "grad_norm": 0.3402821123600006, "learning_rate": 8e-05, "loss": 1.6061, "step": 6403 }, { "epoch": 1.4086660068184318, "grad_norm": 0.34030959010124207, "learning_rate": 8e-05, "loss": 1.5703, "step": 6404 }, { "epoch": 1.408885956230067, "grad_norm": 0.3408047556877136, "learning_rate": 8e-05, "loss": 1.5034, "step": 6405 }, { "epoch": 1.4091059056417023, "grad_norm": 0.3555893898010254, "learning_rate": 8e-05, "loss": 1.5953, "step": 6406 }, { "epoch": 1.4093258550533379, "grad_norm": 0.3220095634460449, "learning_rate": 8e-05, "loss": 1.4102, "step": 6407 }, { "epoch": 1.4095458044649731, "grad_norm": 0.3259319067001343, "learning_rate": 8e-05, "loss": 1.5071, "step": 6408 }, { "epoch": 1.4097657538766084, "grad_norm": 0.3347572982311249, "learning_rate": 8e-05, "loss": 1.4709, "step": 6409 }, { "epoch": 1.4099857032882437, "grad_norm": 0.3433484435081482, "learning_rate": 8e-05, "loss": 1.6135, "step": 6410 }, { "epoch": 1.410205652699879, "grad_norm": 0.3388148546218872, "learning_rate": 8e-05, "loss": 1.5891, "step": 6411 }, { "epoch": 1.4104256021115145, "grad_norm": 0.3318370282649994, "learning_rate": 8e-05, "loss": 1.5265, "step": 6412 }, { "epoch": 1.4106455515231497, "grad_norm": 0.3464888036251068, "learning_rate": 8e-05, "loss": 1.7215, "step": 6413 }, { "epoch": 1.410865500934785, "grad_norm": 0.36354172229766846, "learning_rate": 8e-05, "loss": 1.7101, "step": 6414 }, { "epoch": 1.4110854503464203, "grad_norm": 0.33394142985343933, "learning_rate": 8e-05, "loss": 1.5308, "step": 6415 }, { "epoch": 1.4113053997580556, "grad_norm": 0.3479115962982178, "learning_rate": 8e-05, "loss": 1.6054, "step": 6416 }, { "epoch": 1.411525349169691, "grad_norm": 0.3358260989189148, "learning_rate": 8e-05, "loss": 1.5858, "step": 6417 }, { "epoch": 1.4117452985813264, "grad_norm": 0.3412998616695404, "learning_rate": 8e-05, "loss": 1.5541, "step": 6418 }, { "epoch": 1.4119652479929616, "grad_norm": 0.3573668897151947, "learning_rate": 8e-05, "loss": 1.5503, "step": 6419 }, { "epoch": 1.412185197404597, "grad_norm": 0.3365357220172882, "learning_rate": 8e-05, "loss": 1.5325, "step": 6420 }, { "epoch": 1.4124051468162322, "grad_norm": 0.35861000418663025, "learning_rate": 8e-05, "loss": 1.6341, "step": 6421 }, { "epoch": 1.4126250962278677, "grad_norm": 0.3407483398914337, "learning_rate": 8e-05, "loss": 1.5068, "step": 6422 }, { "epoch": 1.412845045639503, "grad_norm": 0.3499971032142639, "learning_rate": 8e-05, "loss": 1.7126, "step": 6423 }, { "epoch": 1.4130649950511383, "grad_norm": 0.34029635787010193, "learning_rate": 8e-05, "loss": 1.517, "step": 6424 }, { "epoch": 1.4132849444627735, "grad_norm": 0.33286088705062866, "learning_rate": 8e-05, "loss": 1.5253, "step": 6425 }, { "epoch": 1.4135048938744088, "grad_norm": 0.36515215039253235, "learning_rate": 8e-05, "loss": 1.7483, "step": 6426 }, { "epoch": 1.4137248432860443, "grad_norm": 0.35051774978637695, "learning_rate": 8e-05, "loss": 1.5019, "step": 6427 }, { "epoch": 1.4139447926976796, "grad_norm": 0.332829087972641, "learning_rate": 8e-05, "loss": 1.5608, "step": 6428 }, { "epoch": 1.4141647421093149, "grad_norm": 0.34698083996772766, "learning_rate": 8e-05, "loss": 1.7454, "step": 6429 }, { "epoch": 1.4143846915209501, "grad_norm": 0.4190134108066559, "learning_rate": 8e-05, "loss": 1.5612, "step": 6430 }, { "epoch": 1.4146046409325854, "grad_norm": 0.33199557662010193, "learning_rate": 8e-05, "loss": 1.55, "step": 6431 }, { "epoch": 1.414824590344221, "grad_norm": 0.34450453519821167, "learning_rate": 8e-05, "loss": 1.6955, "step": 6432 }, { "epoch": 1.4150445397558562, "grad_norm": 0.37000054121017456, "learning_rate": 8e-05, "loss": 1.5609, "step": 6433 }, { "epoch": 1.4152644891674915, "grad_norm": 0.3477998375892639, "learning_rate": 8e-05, "loss": 1.5681, "step": 6434 }, { "epoch": 1.4154844385791268, "grad_norm": 0.33704617619514465, "learning_rate": 8e-05, "loss": 1.5553, "step": 6435 }, { "epoch": 1.415704387990762, "grad_norm": 0.3374791443347931, "learning_rate": 8e-05, "loss": 1.5245, "step": 6436 }, { "epoch": 1.4159243374023975, "grad_norm": 0.33506548404693604, "learning_rate": 8e-05, "loss": 1.5094, "step": 6437 }, { "epoch": 1.4161442868140328, "grad_norm": 0.35477250814437866, "learning_rate": 8e-05, "loss": 1.5867, "step": 6438 }, { "epoch": 1.416364236225668, "grad_norm": 0.3589935302734375, "learning_rate": 8e-05, "loss": 1.6935, "step": 6439 }, { "epoch": 1.4165841856373034, "grad_norm": 0.3475363552570343, "learning_rate": 8e-05, "loss": 1.6857, "step": 6440 }, { "epoch": 1.4168041350489387, "grad_norm": 0.3671540319919586, "learning_rate": 8e-05, "loss": 1.7203, "step": 6441 }, { "epoch": 1.4170240844605742, "grad_norm": 0.35966354608535767, "learning_rate": 8e-05, "loss": 1.5497, "step": 6442 }, { "epoch": 1.4172440338722094, "grad_norm": 0.3150070011615753, "learning_rate": 8e-05, "loss": 1.5275, "step": 6443 }, { "epoch": 1.4174639832838447, "grad_norm": 0.3351214528083801, "learning_rate": 8e-05, "loss": 1.5676, "step": 6444 }, { "epoch": 1.41768393269548, "grad_norm": 0.3502386808395386, "learning_rate": 8e-05, "loss": 1.6549, "step": 6445 }, { "epoch": 1.4179038821071153, "grad_norm": 0.3431946933269501, "learning_rate": 8e-05, "loss": 1.6084, "step": 6446 }, { "epoch": 1.4181238315187508, "grad_norm": 0.3629715144634247, "learning_rate": 8e-05, "loss": 1.7045, "step": 6447 }, { "epoch": 1.418343780930386, "grad_norm": 0.35397958755493164, "learning_rate": 8e-05, "loss": 1.5754, "step": 6448 }, { "epoch": 1.4185637303420213, "grad_norm": 0.3375137448310852, "learning_rate": 8e-05, "loss": 1.6454, "step": 6449 }, { "epoch": 1.4187836797536566, "grad_norm": 0.3498650789260864, "learning_rate": 8e-05, "loss": 1.5099, "step": 6450 }, { "epoch": 1.4190036291652919, "grad_norm": 0.36285316944122314, "learning_rate": 8e-05, "loss": 1.5788, "step": 6451 }, { "epoch": 1.4192235785769274, "grad_norm": 0.345838338136673, "learning_rate": 8e-05, "loss": 1.6287, "step": 6452 }, { "epoch": 1.4194435279885627, "grad_norm": 0.3722645044326782, "learning_rate": 8e-05, "loss": 1.6, "step": 6453 }, { "epoch": 1.419663477400198, "grad_norm": 0.34527668356895447, "learning_rate": 8e-05, "loss": 1.5351, "step": 6454 }, { "epoch": 1.4198834268118332, "grad_norm": 0.34100085496902466, "learning_rate": 8e-05, "loss": 1.5275, "step": 6455 }, { "epoch": 1.4201033762234685, "grad_norm": 0.4090154767036438, "learning_rate": 8e-05, "loss": 1.8117, "step": 6456 }, { "epoch": 1.420323325635104, "grad_norm": 0.35736754536628723, "learning_rate": 8e-05, "loss": 1.5645, "step": 6457 }, { "epoch": 1.4205432750467393, "grad_norm": 0.3781251311302185, "learning_rate": 8e-05, "loss": 1.654, "step": 6458 }, { "epoch": 1.4207632244583746, "grad_norm": 0.3213988244533539, "learning_rate": 8e-05, "loss": 1.5562, "step": 6459 }, { "epoch": 1.4209831738700098, "grad_norm": 0.3704476058483124, "learning_rate": 8e-05, "loss": 1.6805, "step": 6460 }, { "epoch": 1.421203123281645, "grad_norm": 0.3489900529384613, "learning_rate": 8e-05, "loss": 1.6246, "step": 6461 }, { "epoch": 1.4214230726932806, "grad_norm": 0.375161349773407, "learning_rate": 8e-05, "loss": 1.6161, "step": 6462 }, { "epoch": 1.4216430221049159, "grad_norm": 0.37828099727630615, "learning_rate": 8e-05, "loss": 1.7065, "step": 6463 }, { "epoch": 1.4218629715165512, "grad_norm": 0.35025736689567566, "learning_rate": 8e-05, "loss": 1.6809, "step": 6464 }, { "epoch": 1.4220829209281864, "grad_norm": 0.3295503556728363, "learning_rate": 8e-05, "loss": 1.4886, "step": 6465 }, { "epoch": 1.4223028703398217, "grad_norm": 0.32921674847602844, "learning_rate": 8e-05, "loss": 1.4762, "step": 6466 }, { "epoch": 1.4225228197514572, "grad_norm": 0.3377435505390167, "learning_rate": 8e-05, "loss": 1.5911, "step": 6467 }, { "epoch": 1.4227427691630925, "grad_norm": 0.35865476727485657, "learning_rate": 8e-05, "loss": 1.7615, "step": 6468 }, { "epoch": 1.4229627185747278, "grad_norm": 0.3397221267223358, "learning_rate": 8e-05, "loss": 1.5211, "step": 6469 }, { "epoch": 1.423182667986363, "grad_norm": 0.339513897895813, "learning_rate": 8e-05, "loss": 1.6229, "step": 6470 }, { "epoch": 1.4234026173979983, "grad_norm": 0.3445286750793457, "learning_rate": 8e-05, "loss": 1.5106, "step": 6471 }, { "epoch": 1.4236225668096338, "grad_norm": 0.3301396071910858, "learning_rate": 8e-05, "loss": 1.5572, "step": 6472 }, { "epoch": 1.4238425162212691, "grad_norm": 0.3573041260242462, "learning_rate": 8e-05, "loss": 1.543, "step": 6473 }, { "epoch": 1.4240624656329044, "grad_norm": 0.34128129482269287, "learning_rate": 8e-05, "loss": 1.4948, "step": 6474 }, { "epoch": 1.4242824150445397, "grad_norm": 0.3289186358451843, "learning_rate": 8e-05, "loss": 1.7622, "step": 6475 }, { "epoch": 1.424502364456175, "grad_norm": 0.3312521278858185, "learning_rate": 8e-05, "loss": 1.7354, "step": 6476 }, { "epoch": 1.4247223138678105, "grad_norm": 0.3518742024898529, "learning_rate": 8e-05, "loss": 1.6793, "step": 6477 }, { "epoch": 1.4249422632794457, "grad_norm": 0.3538171648979187, "learning_rate": 8e-05, "loss": 1.6343, "step": 6478 }, { "epoch": 1.425162212691081, "grad_norm": 0.36345982551574707, "learning_rate": 8e-05, "loss": 1.668, "step": 6479 }, { "epoch": 1.4253821621027165, "grad_norm": 0.3287378251552582, "learning_rate": 8e-05, "loss": 1.4877, "step": 6480 }, { "epoch": 1.4256021115143516, "grad_norm": 0.3380626440048218, "learning_rate": 8e-05, "loss": 1.6735, "step": 6481 }, { "epoch": 1.425822060925987, "grad_norm": 0.3899190127849579, "learning_rate": 8e-05, "loss": 1.5918, "step": 6482 }, { "epoch": 1.4260420103376223, "grad_norm": 0.34014394879341125, "learning_rate": 8e-05, "loss": 1.5484, "step": 6483 }, { "epoch": 1.4262619597492576, "grad_norm": 0.3494177758693695, "learning_rate": 8e-05, "loss": 1.716, "step": 6484 }, { "epoch": 1.4264819091608931, "grad_norm": 0.3425876796245575, "learning_rate": 8e-05, "loss": 1.5783, "step": 6485 }, { "epoch": 1.4267018585725282, "grad_norm": 0.3593464195728302, "learning_rate": 8e-05, "loss": 1.555, "step": 6486 }, { "epoch": 1.4269218079841637, "grad_norm": 0.3125891089439392, "learning_rate": 8e-05, "loss": 1.3514, "step": 6487 }, { "epoch": 1.427141757395799, "grad_norm": 0.3372005522251129, "learning_rate": 8e-05, "loss": 1.5579, "step": 6488 }, { "epoch": 1.4273617068074342, "grad_norm": 0.3374946117401123, "learning_rate": 8e-05, "loss": 1.5742, "step": 6489 }, { "epoch": 1.4275816562190697, "grad_norm": 0.34286433458328247, "learning_rate": 8e-05, "loss": 1.531, "step": 6490 }, { "epoch": 1.427801605630705, "grad_norm": 0.33125749230384827, "learning_rate": 8e-05, "loss": 1.5047, "step": 6491 }, { "epoch": 1.4280215550423403, "grad_norm": 0.3242490887641907, "learning_rate": 8e-05, "loss": 1.5324, "step": 6492 }, { "epoch": 1.4282415044539756, "grad_norm": 0.3474394381046295, "learning_rate": 8e-05, "loss": 1.5501, "step": 6493 }, { "epoch": 1.4284614538656109, "grad_norm": 0.3693917989730835, "learning_rate": 8e-05, "loss": 1.4684, "step": 6494 }, { "epoch": 1.4286814032772464, "grad_norm": 0.3707519471645355, "learning_rate": 8e-05, "loss": 1.6603, "step": 6495 }, { "epoch": 1.4289013526888816, "grad_norm": 0.33868518471717834, "learning_rate": 8e-05, "loss": 1.5827, "step": 6496 }, { "epoch": 1.429121302100517, "grad_norm": 0.33857741951942444, "learning_rate": 8e-05, "loss": 1.7072, "step": 6497 }, { "epoch": 1.4293412515121522, "grad_norm": 0.3983720541000366, "learning_rate": 8e-05, "loss": 1.5186, "step": 6498 }, { "epoch": 1.4295612009237875, "grad_norm": 0.3441663682460785, "learning_rate": 8e-05, "loss": 1.6099, "step": 6499 }, { "epoch": 1.429781150335423, "grad_norm": 0.3383481800556183, "learning_rate": 8e-05, "loss": 1.443, "step": 6500 }, { "epoch": 1.4300010997470582, "grad_norm": 0.33508655428886414, "learning_rate": 8e-05, "loss": 1.4732, "step": 6501 }, { "epoch": 1.4302210491586935, "grad_norm": 0.3518548607826233, "learning_rate": 8e-05, "loss": 1.6083, "step": 6502 }, { "epoch": 1.4304409985703288, "grad_norm": 0.37076324224472046, "learning_rate": 8e-05, "loss": 1.6368, "step": 6503 }, { "epoch": 1.430660947981964, "grad_norm": 0.34742873907089233, "learning_rate": 8e-05, "loss": 1.6605, "step": 6504 }, { "epoch": 1.4308808973935996, "grad_norm": 0.342403382062912, "learning_rate": 8e-05, "loss": 1.527, "step": 6505 }, { "epoch": 1.4311008468052349, "grad_norm": 0.35646745562553406, "learning_rate": 8e-05, "loss": 1.6965, "step": 6506 }, { "epoch": 1.4313207962168701, "grad_norm": 0.3406618535518646, "learning_rate": 8e-05, "loss": 1.6039, "step": 6507 }, { "epoch": 1.4315407456285054, "grad_norm": 0.3344443142414093, "learning_rate": 8e-05, "loss": 1.4728, "step": 6508 }, { "epoch": 1.4317606950401407, "grad_norm": 0.33608725666999817, "learning_rate": 8e-05, "loss": 1.6247, "step": 6509 }, { "epoch": 1.4319806444517762, "grad_norm": 0.37324950098991394, "learning_rate": 8e-05, "loss": 1.503, "step": 6510 }, { "epoch": 1.4322005938634115, "grad_norm": 0.36823001503944397, "learning_rate": 8e-05, "loss": 1.641, "step": 6511 }, { "epoch": 1.4324205432750468, "grad_norm": 0.38138583302497864, "learning_rate": 8e-05, "loss": 1.724, "step": 6512 }, { "epoch": 1.432640492686682, "grad_norm": 0.331938236951828, "learning_rate": 8e-05, "loss": 1.4926, "step": 6513 }, { "epoch": 1.4328604420983173, "grad_norm": 0.34242814779281616, "learning_rate": 8e-05, "loss": 1.3515, "step": 6514 }, { "epoch": 1.4330803915099528, "grad_norm": 0.3448748290538788, "learning_rate": 8e-05, "loss": 1.6266, "step": 6515 }, { "epoch": 1.433300340921588, "grad_norm": 0.35393190383911133, "learning_rate": 8e-05, "loss": 1.5812, "step": 6516 }, { "epoch": 1.4335202903332234, "grad_norm": 0.36170297861099243, "learning_rate": 8e-05, "loss": 1.6311, "step": 6517 }, { "epoch": 1.4337402397448586, "grad_norm": 0.36623790860176086, "learning_rate": 8e-05, "loss": 1.5862, "step": 6518 }, { "epoch": 1.433960189156494, "grad_norm": 0.34393608570098877, "learning_rate": 8e-05, "loss": 1.6065, "step": 6519 }, { "epoch": 1.4341801385681294, "grad_norm": 0.36418071389198303, "learning_rate": 8e-05, "loss": 1.612, "step": 6520 }, { "epoch": 1.4344000879797647, "grad_norm": 0.3944697082042694, "learning_rate": 8e-05, "loss": 1.5765, "step": 6521 }, { "epoch": 1.4346200373914, "grad_norm": 0.33029985427856445, "learning_rate": 8e-05, "loss": 1.4024, "step": 6522 }, { "epoch": 1.4348399868030353, "grad_norm": 0.34310629963874817, "learning_rate": 8e-05, "loss": 1.5229, "step": 6523 }, { "epoch": 1.4350599362146705, "grad_norm": 0.3470463156700134, "learning_rate": 8e-05, "loss": 1.5688, "step": 6524 }, { "epoch": 1.435279885626306, "grad_norm": 0.3496949076652527, "learning_rate": 8e-05, "loss": 1.5875, "step": 6525 }, { "epoch": 1.4354998350379413, "grad_norm": 0.3360151946544647, "learning_rate": 8e-05, "loss": 1.5569, "step": 6526 }, { "epoch": 1.4357197844495766, "grad_norm": 0.3412640690803528, "learning_rate": 8e-05, "loss": 1.6818, "step": 6527 }, { "epoch": 1.4359397338612119, "grad_norm": 0.41767919063568115, "learning_rate": 8e-05, "loss": 1.7488, "step": 6528 }, { "epoch": 1.4361596832728472, "grad_norm": 0.34725168347358704, "learning_rate": 8e-05, "loss": 1.714, "step": 6529 }, { "epoch": 1.4363796326844827, "grad_norm": 0.3573188781738281, "learning_rate": 8e-05, "loss": 1.5823, "step": 6530 }, { "epoch": 1.436599582096118, "grad_norm": 0.33740848302841187, "learning_rate": 8e-05, "loss": 1.6459, "step": 6531 }, { "epoch": 1.4368195315077532, "grad_norm": 0.34050336480140686, "learning_rate": 8e-05, "loss": 1.6509, "step": 6532 }, { "epoch": 1.4370394809193885, "grad_norm": 0.3255702555179596, "learning_rate": 8e-05, "loss": 1.531, "step": 6533 }, { "epoch": 1.4372594303310238, "grad_norm": 0.3769923746585846, "learning_rate": 8e-05, "loss": 1.6701, "step": 6534 }, { "epoch": 1.4374793797426593, "grad_norm": 0.4238828420639038, "learning_rate": 8e-05, "loss": 1.6448, "step": 6535 }, { "epoch": 1.4376993291542945, "grad_norm": 0.33605238795280457, "learning_rate": 8e-05, "loss": 1.5409, "step": 6536 }, { "epoch": 1.4379192785659298, "grad_norm": 0.35869601368904114, "learning_rate": 8e-05, "loss": 1.539, "step": 6537 }, { "epoch": 1.438139227977565, "grad_norm": 0.3357327878475189, "learning_rate": 8e-05, "loss": 1.5153, "step": 6538 }, { "epoch": 1.4383591773892004, "grad_norm": 0.39760327339172363, "learning_rate": 8e-05, "loss": 1.556, "step": 6539 }, { "epoch": 1.4385791268008359, "grad_norm": 0.34878817200660706, "learning_rate": 8e-05, "loss": 1.5688, "step": 6540 }, { "epoch": 1.4387990762124712, "grad_norm": 0.33867260813713074, "learning_rate": 8e-05, "loss": 1.6147, "step": 6541 }, { "epoch": 1.4390190256241064, "grad_norm": 0.3408527076244354, "learning_rate": 8e-05, "loss": 1.6553, "step": 6542 }, { "epoch": 1.4392389750357417, "grad_norm": 0.37420952320098877, "learning_rate": 8e-05, "loss": 1.5975, "step": 6543 }, { "epoch": 1.439458924447377, "grad_norm": 0.3548506498336792, "learning_rate": 8e-05, "loss": 1.6323, "step": 6544 }, { "epoch": 1.4396788738590125, "grad_norm": 0.3852817416191101, "learning_rate": 8e-05, "loss": 1.4593, "step": 6545 }, { "epoch": 1.4398988232706478, "grad_norm": 0.3380104601383209, "learning_rate": 8e-05, "loss": 1.6438, "step": 6546 }, { "epoch": 1.440118772682283, "grad_norm": 0.35187506675720215, "learning_rate": 8e-05, "loss": 1.5866, "step": 6547 }, { "epoch": 1.4403387220939183, "grad_norm": 0.33331960439682007, "learning_rate": 8e-05, "loss": 1.6172, "step": 6548 }, { "epoch": 1.4405586715055536, "grad_norm": 0.36942756175994873, "learning_rate": 8e-05, "loss": 1.7219, "step": 6549 }, { "epoch": 1.4407786209171891, "grad_norm": 0.32878467440605164, "learning_rate": 8e-05, "loss": 1.6331, "step": 6550 }, { "epoch": 1.4409985703288244, "grad_norm": 0.34732091426849365, "learning_rate": 8e-05, "loss": 1.6466, "step": 6551 }, { "epoch": 1.4412185197404597, "grad_norm": 0.34576013684272766, "learning_rate": 8e-05, "loss": 1.6225, "step": 6552 }, { "epoch": 1.441438469152095, "grad_norm": 0.3446757197380066, "learning_rate": 8e-05, "loss": 1.478, "step": 6553 }, { "epoch": 1.4416584185637302, "grad_norm": 0.3350602686405182, "learning_rate": 8e-05, "loss": 1.5869, "step": 6554 }, { "epoch": 1.4418783679753657, "grad_norm": 0.3322978913784027, "learning_rate": 8e-05, "loss": 1.4877, "step": 6555 }, { "epoch": 1.442098317387001, "grad_norm": 0.39257922768592834, "learning_rate": 8e-05, "loss": 1.6689, "step": 6556 }, { "epoch": 1.4423182667986363, "grad_norm": 0.34195804595947266, "learning_rate": 8e-05, "loss": 1.6235, "step": 6557 }, { "epoch": 1.4425382162102716, "grad_norm": 0.3325009346008301, "learning_rate": 8e-05, "loss": 1.4944, "step": 6558 }, { "epoch": 1.4427581656219068, "grad_norm": 0.3511332869529724, "learning_rate": 8e-05, "loss": 1.5844, "step": 6559 }, { "epoch": 1.4429781150335423, "grad_norm": 0.3489277958869934, "learning_rate": 8e-05, "loss": 1.6885, "step": 6560 }, { "epoch": 1.4431980644451776, "grad_norm": 0.35139963030815125, "learning_rate": 8e-05, "loss": 1.6548, "step": 6561 }, { "epoch": 1.443418013856813, "grad_norm": 0.34193339943885803, "learning_rate": 8e-05, "loss": 1.644, "step": 6562 }, { "epoch": 1.4436379632684482, "grad_norm": 0.3406340181827545, "learning_rate": 8e-05, "loss": 1.6076, "step": 6563 }, { "epoch": 1.4438579126800835, "grad_norm": 0.3339712917804718, "learning_rate": 8e-05, "loss": 1.4981, "step": 6564 }, { "epoch": 1.444077862091719, "grad_norm": 0.3241324722766876, "learning_rate": 8e-05, "loss": 1.5754, "step": 6565 }, { "epoch": 1.4442978115033542, "grad_norm": 0.34674471616744995, "learning_rate": 8e-05, "loss": 1.671, "step": 6566 }, { "epoch": 1.4445177609149895, "grad_norm": 0.3340025544166565, "learning_rate": 8e-05, "loss": 1.5411, "step": 6567 }, { "epoch": 1.444737710326625, "grad_norm": 0.38187265396118164, "learning_rate": 8e-05, "loss": 1.6574, "step": 6568 }, { "epoch": 1.44495765973826, "grad_norm": 0.37301933765411377, "learning_rate": 8e-05, "loss": 1.5609, "step": 6569 }, { "epoch": 1.4451776091498956, "grad_norm": 0.3474269509315491, "learning_rate": 8e-05, "loss": 1.5078, "step": 6570 }, { "epoch": 1.4453975585615308, "grad_norm": 0.33681604266166687, "learning_rate": 8e-05, "loss": 1.5321, "step": 6571 }, { "epoch": 1.4456175079731661, "grad_norm": 0.3557574152946472, "learning_rate": 8e-05, "loss": 1.5548, "step": 6572 }, { "epoch": 1.4458374573848016, "grad_norm": 0.3498886227607727, "learning_rate": 8e-05, "loss": 1.7239, "step": 6573 }, { "epoch": 1.4460574067964367, "grad_norm": 0.35020574927330017, "learning_rate": 8e-05, "loss": 1.5504, "step": 6574 }, { "epoch": 1.4462773562080722, "grad_norm": 0.33152446150779724, "learning_rate": 8e-05, "loss": 1.6505, "step": 6575 }, { "epoch": 1.4464973056197075, "grad_norm": 0.3410642445087433, "learning_rate": 8e-05, "loss": 1.5123, "step": 6576 }, { "epoch": 1.4467172550313427, "grad_norm": 0.35926496982574463, "learning_rate": 8e-05, "loss": 1.7223, "step": 6577 }, { "epoch": 1.4469372044429782, "grad_norm": 0.32567650079727173, "learning_rate": 8e-05, "loss": 1.5721, "step": 6578 }, { "epoch": 1.4471571538546135, "grad_norm": 0.3384774923324585, "learning_rate": 8e-05, "loss": 1.5266, "step": 6579 }, { "epoch": 1.4473771032662488, "grad_norm": 0.35173794627189636, "learning_rate": 8e-05, "loss": 1.6092, "step": 6580 }, { "epoch": 1.447597052677884, "grad_norm": 0.34703880548477173, "learning_rate": 8e-05, "loss": 1.6074, "step": 6581 }, { "epoch": 1.4478170020895194, "grad_norm": 0.34333762526512146, "learning_rate": 8e-05, "loss": 1.5438, "step": 6582 }, { "epoch": 1.4480369515011549, "grad_norm": 0.33471232652664185, "learning_rate": 8e-05, "loss": 1.5611, "step": 6583 }, { "epoch": 1.4482569009127901, "grad_norm": 0.3438645005226135, "learning_rate": 8e-05, "loss": 1.3787, "step": 6584 }, { "epoch": 1.4484768503244254, "grad_norm": 0.35323765873908997, "learning_rate": 8e-05, "loss": 1.6322, "step": 6585 }, { "epoch": 1.4486967997360607, "grad_norm": 0.3442220389842987, "learning_rate": 8e-05, "loss": 1.5975, "step": 6586 }, { "epoch": 1.448916749147696, "grad_norm": 0.35019269585609436, "learning_rate": 8e-05, "loss": 1.4428, "step": 6587 }, { "epoch": 1.4491366985593315, "grad_norm": 0.3332765996456146, "learning_rate": 8e-05, "loss": 1.5102, "step": 6588 }, { "epoch": 1.4493566479709667, "grad_norm": 0.3299747109413147, "learning_rate": 8e-05, "loss": 1.4528, "step": 6589 }, { "epoch": 1.449576597382602, "grad_norm": 0.339110404253006, "learning_rate": 8e-05, "loss": 1.6285, "step": 6590 }, { "epoch": 1.4497965467942373, "grad_norm": 0.3310838043689728, "learning_rate": 8e-05, "loss": 1.4516, "step": 6591 }, { "epoch": 1.4500164962058726, "grad_norm": 0.3544165790081024, "learning_rate": 8e-05, "loss": 1.649, "step": 6592 }, { "epoch": 1.450236445617508, "grad_norm": 0.33858221769332886, "learning_rate": 8e-05, "loss": 1.7216, "step": 6593 }, { "epoch": 1.4504563950291434, "grad_norm": 0.33305293321609497, "learning_rate": 8e-05, "loss": 1.5653, "step": 6594 }, { "epoch": 1.4506763444407786, "grad_norm": 0.36752066016197205, "learning_rate": 8e-05, "loss": 1.615, "step": 6595 }, { "epoch": 1.450896293852414, "grad_norm": 0.35680580139160156, "learning_rate": 8e-05, "loss": 1.576, "step": 6596 }, { "epoch": 1.4511162432640492, "grad_norm": 0.3397970199584961, "learning_rate": 8e-05, "loss": 1.6507, "step": 6597 }, { "epoch": 1.4513361926756847, "grad_norm": 0.36618486046791077, "learning_rate": 8e-05, "loss": 1.6904, "step": 6598 }, { "epoch": 1.45155614208732, "grad_norm": 0.36554989218711853, "learning_rate": 8e-05, "loss": 1.6827, "step": 6599 }, { "epoch": 1.4517760914989553, "grad_norm": 0.3565073311328888, "learning_rate": 8e-05, "loss": 1.605, "step": 6600 }, { "epoch": 1.4519960409105905, "grad_norm": 0.3391379415988922, "learning_rate": 8e-05, "loss": 1.6428, "step": 6601 }, { "epoch": 1.4522159903222258, "grad_norm": 0.35805922746658325, "learning_rate": 8e-05, "loss": 1.615, "step": 6602 }, { "epoch": 1.4524359397338613, "grad_norm": 0.3607567846775055, "learning_rate": 8e-05, "loss": 1.6536, "step": 6603 }, { "epoch": 1.4526558891454966, "grad_norm": 0.3301038444042206, "learning_rate": 8e-05, "loss": 1.5777, "step": 6604 }, { "epoch": 1.4528758385571319, "grad_norm": 0.33840441703796387, "learning_rate": 8e-05, "loss": 1.6268, "step": 6605 }, { "epoch": 1.4530957879687671, "grad_norm": 0.3477640748023987, "learning_rate": 8e-05, "loss": 1.567, "step": 6606 }, { "epoch": 1.4533157373804024, "grad_norm": 0.3532119691371918, "learning_rate": 8e-05, "loss": 1.6204, "step": 6607 }, { "epoch": 1.453535686792038, "grad_norm": 0.3503606915473938, "learning_rate": 8e-05, "loss": 1.603, "step": 6608 }, { "epoch": 1.4537556362036732, "grad_norm": 0.33437034487724304, "learning_rate": 8e-05, "loss": 1.5743, "step": 6609 }, { "epoch": 1.4539755856153085, "grad_norm": 0.33707886934280396, "learning_rate": 8e-05, "loss": 1.5914, "step": 6610 }, { "epoch": 1.4541955350269438, "grad_norm": 0.43675491213798523, "learning_rate": 8e-05, "loss": 1.7885, "step": 6611 }, { "epoch": 1.454415484438579, "grad_norm": 0.3422887921333313, "learning_rate": 8e-05, "loss": 1.6658, "step": 6612 }, { "epoch": 1.4546354338502145, "grad_norm": 0.33841755986213684, "learning_rate": 8e-05, "loss": 1.5273, "step": 6613 }, { "epoch": 1.4548553832618498, "grad_norm": 0.31525757908821106, "learning_rate": 8e-05, "loss": 1.2025, "step": 6614 }, { "epoch": 1.455075332673485, "grad_norm": 0.33726271986961365, "learning_rate": 8e-05, "loss": 1.6353, "step": 6615 }, { "epoch": 1.4552952820851204, "grad_norm": 0.3607470393180847, "learning_rate": 8e-05, "loss": 1.6596, "step": 6616 }, { "epoch": 1.4555152314967557, "grad_norm": 0.3606870472431183, "learning_rate": 8e-05, "loss": 1.6952, "step": 6617 }, { "epoch": 1.4557351809083912, "grad_norm": 0.31921473145484924, "learning_rate": 8e-05, "loss": 1.3958, "step": 6618 }, { "epoch": 1.4559551303200264, "grad_norm": 0.3195357024669647, "learning_rate": 8e-05, "loss": 1.5248, "step": 6619 }, { "epoch": 1.4561750797316617, "grad_norm": 0.3385067880153656, "learning_rate": 8e-05, "loss": 1.576, "step": 6620 }, { "epoch": 1.456395029143297, "grad_norm": 0.34516996145248413, "learning_rate": 8e-05, "loss": 1.5283, "step": 6621 }, { "epoch": 1.4566149785549323, "grad_norm": 0.3338092863559723, "learning_rate": 8e-05, "loss": 1.462, "step": 6622 }, { "epoch": 1.4568349279665678, "grad_norm": 0.3307456076145172, "learning_rate": 8e-05, "loss": 1.5591, "step": 6623 }, { "epoch": 1.457054877378203, "grad_norm": 0.35320648550987244, "learning_rate": 8e-05, "loss": 1.6944, "step": 6624 }, { "epoch": 1.4572748267898383, "grad_norm": 0.34180405735969543, "learning_rate": 8e-05, "loss": 1.6349, "step": 6625 }, { "epoch": 1.4574947762014736, "grad_norm": 0.3415107727050781, "learning_rate": 8e-05, "loss": 1.5879, "step": 6626 }, { "epoch": 1.4577147256131089, "grad_norm": 0.324910968542099, "learning_rate": 8e-05, "loss": 1.5024, "step": 6627 }, { "epoch": 1.4579346750247444, "grad_norm": 0.3484431207180023, "learning_rate": 8e-05, "loss": 1.6406, "step": 6628 }, { "epoch": 1.4581546244363797, "grad_norm": 0.3445449471473694, "learning_rate": 8e-05, "loss": 1.638, "step": 6629 }, { "epoch": 1.458374573848015, "grad_norm": 0.3336852788925171, "learning_rate": 8e-05, "loss": 1.6401, "step": 6630 }, { "epoch": 1.4585945232596502, "grad_norm": 0.3417905867099762, "learning_rate": 8e-05, "loss": 1.5477, "step": 6631 }, { "epoch": 1.4588144726712855, "grad_norm": 0.34106189012527466, "learning_rate": 8e-05, "loss": 1.6095, "step": 6632 }, { "epoch": 1.459034422082921, "grad_norm": 0.3636377155780792, "learning_rate": 8e-05, "loss": 1.7991, "step": 6633 }, { "epoch": 1.4592543714945563, "grad_norm": 0.34858354926109314, "learning_rate": 8e-05, "loss": 1.6083, "step": 6634 }, { "epoch": 1.4594743209061916, "grad_norm": 0.3465934693813324, "learning_rate": 8e-05, "loss": 1.706, "step": 6635 }, { "epoch": 1.4596942703178268, "grad_norm": 0.31630322337150574, "learning_rate": 8e-05, "loss": 1.4166, "step": 6636 }, { "epoch": 1.4599142197294621, "grad_norm": 0.3824889361858368, "learning_rate": 8e-05, "loss": 1.5571, "step": 6637 }, { "epoch": 1.4601341691410976, "grad_norm": 0.3718869686126709, "learning_rate": 8e-05, "loss": 1.7529, "step": 6638 }, { "epoch": 1.460354118552733, "grad_norm": 0.32859155535697937, "learning_rate": 8e-05, "loss": 1.5321, "step": 6639 }, { "epoch": 1.4605740679643682, "grad_norm": 0.32916608452796936, "learning_rate": 8e-05, "loss": 1.642, "step": 6640 }, { "epoch": 1.4607940173760035, "grad_norm": 0.3437146246433258, "learning_rate": 8e-05, "loss": 1.5932, "step": 6641 }, { "epoch": 1.4610139667876387, "grad_norm": 0.339389443397522, "learning_rate": 8e-05, "loss": 1.6259, "step": 6642 }, { "epoch": 1.4612339161992742, "grad_norm": 0.33149126172065735, "learning_rate": 8e-05, "loss": 1.5327, "step": 6643 }, { "epoch": 1.4614538656109095, "grad_norm": 0.345476895570755, "learning_rate": 8e-05, "loss": 1.5563, "step": 6644 }, { "epoch": 1.4616738150225448, "grad_norm": 0.3649548888206482, "learning_rate": 8e-05, "loss": 1.5965, "step": 6645 }, { "epoch": 1.46189376443418, "grad_norm": 0.3464626967906952, "learning_rate": 8e-05, "loss": 1.6147, "step": 6646 }, { "epoch": 1.4621137138458153, "grad_norm": 0.3352545499801636, "learning_rate": 8e-05, "loss": 1.6055, "step": 6647 }, { "epoch": 1.4623336632574508, "grad_norm": 0.35759812593460083, "learning_rate": 8e-05, "loss": 1.6951, "step": 6648 }, { "epoch": 1.4625536126690861, "grad_norm": 0.34264102578163147, "learning_rate": 8e-05, "loss": 1.6532, "step": 6649 }, { "epoch": 1.4627735620807214, "grad_norm": 0.319472074508667, "learning_rate": 8e-05, "loss": 1.4813, "step": 6650 }, { "epoch": 1.4629935114923567, "grad_norm": 0.334678053855896, "learning_rate": 8e-05, "loss": 1.6371, "step": 6651 }, { "epoch": 1.463213460903992, "grad_norm": 0.33871954679489136, "learning_rate": 8e-05, "loss": 1.5932, "step": 6652 }, { "epoch": 1.4634334103156275, "grad_norm": 0.36468809843063354, "learning_rate": 8e-05, "loss": 1.5356, "step": 6653 }, { "epoch": 1.4636533597272627, "grad_norm": 0.330388605594635, "learning_rate": 8e-05, "loss": 1.6291, "step": 6654 }, { "epoch": 1.463873309138898, "grad_norm": 0.33966290950775146, "learning_rate": 8e-05, "loss": 1.7127, "step": 6655 }, { "epoch": 1.4640932585505333, "grad_norm": 0.3325870633125305, "learning_rate": 8e-05, "loss": 1.6063, "step": 6656 }, { "epoch": 1.4643132079621686, "grad_norm": 0.3427707552909851, "learning_rate": 8e-05, "loss": 1.6596, "step": 6657 }, { "epoch": 1.464533157373804, "grad_norm": 0.3287096917629242, "learning_rate": 8e-05, "loss": 1.5086, "step": 6658 }, { "epoch": 1.4647531067854394, "grad_norm": 0.3508812189102173, "learning_rate": 8e-05, "loss": 1.6438, "step": 6659 }, { "epoch": 1.4649730561970746, "grad_norm": 0.3393430709838867, "learning_rate": 8e-05, "loss": 1.6553, "step": 6660 }, { "epoch": 1.4651930056087101, "grad_norm": 0.38255348801612854, "learning_rate": 8e-05, "loss": 1.7142, "step": 6661 }, { "epoch": 1.4654129550203452, "grad_norm": 0.35637953877449036, "learning_rate": 8e-05, "loss": 1.6697, "step": 6662 }, { "epoch": 1.4656329044319807, "grad_norm": 0.3366871774196625, "learning_rate": 8e-05, "loss": 1.4593, "step": 6663 }, { "epoch": 1.465852853843616, "grad_norm": 0.32915255427360535, "learning_rate": 8e-05, "loss": 1.5369, "step": 6664 }, { "epoch": 1.4660728032552512, "grad_norm": 0.3470749855041504, "learning_rate": 8e-05, "loss": 1.5979, "step": 6665 }, { "epoch": 1.4662927526668867, "grad_norm": 0.3372727930545807, "learning_rate": 8e-05, "loss": 1.7425, "step": 6666 }, { "epoch": 1.4665127020785218, "grad_norm": 0.34282398223876953, "learning_rate": 8e-05, "loss": 1.634, "step": 6667 }, { "epoch": 1.4667326514901573, "grad_norm": 0.34355470538139343, "learning_rate": 8e-05, "loss": 1.732, "step": 6668 }, { "epoch": 1.4669526009017926, "grad_norm": 0.3251343071460724, "learning_rate": 8e-05, "loss": 1.552, "step": 6669 }, { "epoch": 1.4671725503134279, "grad_norm": 0.3605474829673767, "learning_rate": 8e-05, "loss": 1.7451, "step": 6670 }, { "epoch": 1.4673924997250634, "grad_norm": 0.34263697266578674, "learning_rate": 8e-05, "loss": 1.5053, "step": 6671 }, { "epoch": 1.4676124491366986, "grad_norm": 0.3326685130596161, "learning_rate": 8e-05, "loss": 1.5503, "step": 6672 }, { "epoch": 1.467832398548334, "grad_norm": 0.35065653920173645, "learning_rate": 8e-05, "loss": 1.693, "step": 6673 }, { "epoch": 1.4680523479599692, "grad_norm": 0.3510446846485138, "learning_rate": 8e-05, "loss": 1.7354, "step": 6674 }, { "epoch": 1.4682722973716045, "grad_norm": 0.32824891805648804, "learning_rate": 8e-05, "loss": 1.5792, "step": 6675 }, { "epoch": 1.46849224678324, "grad_norm": 0.3663065433502197, "learning_rate": 8e-05, "loss": 1.6234, "step": 6676 }, { "epoch": 1.4687121961948753, "grad_norm": 0.3461126685142517, "learning_rate": 8e-05, "loss": 1.661, "step": 6677 }, { "epoch": 1.4689321456065105, "grad_norm": 0.34511736035346985, "learning_rate": 8e-05, "loss": 1.6323, "step": 6678 }, { "epoch": 1.4691520950181458, "grad_norm": 0.32213932275772095, "learning_rate": 8e-05, "loss": 1.5232, "step": 6679 }, { "epoch": 1.469372044429781, "grad_norm": 0.3560557961463928, "learning_rate": 8e-05, "loss": 1.5557, "step": 6680 }, { "epoch": 1.4695919938414166, "grad_norm": 0.32650187611579895, "learning_rate": 8e-05, "loss": 1.4849, "step": 6681 }, { "epoch": 1.4698119432530519, "grad_norm": 0.3768904209136963, "learning_rate": 8e-05, "loss": 1.6789, "step": 6682 }, { "epoch": 1.4700318926646871, "grad_norm": 0.3518642783164978, "learning_rate": 8e-05, "loss": 1.6447, "step": 6683 }, { "epoch": 1.4702518420763224, "grad_norm": 0.33414942026138306, "learning_rate": 8e-05, "loss": 1.5864, "step": 6684 }, { "epoch": 1.4704717914879577, "grad_norm": 0.33360034227371216, "learning_rate": 8e-05, "loss": 1.5731, "step": 6685 }, { "epoch": 1.4706917408995932, "grad_norm": 0.34082719683647156, "learning_rate": 8e-05, "loss": 1.5772, "step": 6686 }, { "epoch": 1.4709116903112285, "grad_norm": 0.32970255613327026, "learning_rate": 8e-05, "loss": 1.6172, "step": 6687 }, { "epoch": 1.4711316397228638, "grad_norm": 0.32739782333374023, "learning_rate": 8e-05, "loss": 1.5443, "step": 6688 }, { "epoch": 1.471351589134499, "grad_norm": 0.34104809165000916, "learning_rate": 8e-05, "loss": 1.7173, "step": 6689 }, { "epoch": 1.4715715385461343, "grad_norm": 0.34444281458854675, "learning_rate": 8e-05, "loss": 1.6309, "step": 6690 }, { "epoch": 1.4717914879577698, "grad_norm": 0.33613845705986023, "learning_rate": 8e-05, "loss": 1.573, "step": 6691 }, { "epoch": 1.472011437369405, "grad_norm": 0.34137287735939026, "learning_rate": 8e-05, "loss": 1.6753, "step": 6692 }, { "epoch": 1.4722313867810404, "grad_norm": 0.33470743894577026, "learning_rate": 8e-05, "loss": 1.4412, "step": 6693 }, { "epoch": 1.4724513361926757, "grad_norm": 0.3536590039730072, "learning_rate": 8e-05, "loss": 1.6264, "step": 6694 }, { "epoch": 1.472671285604311, "grad_norm": 0.39308688044548035, "learning_rate": 8e-05, "loss": 1.5617, "step": 6695 }, { "epoch": 1.4728912350159464, "grad_norm": 0.342668741941452, "learning_rate": 8e-05, "loss": 1.6747, "step": 6696 }, { "epoch": 1.4731111844275817, "grad_norm": 0.33939051628112793, "learning_rate": 8e-05, "loss": 1.4316, "step": 6697 }, { "epoch": 1.473331133839217, "grad_norm": 0.3327997326850891, "learning_rate": 8e-05, "loss": 1.6875, "step": 6698 }, { "epoch": 1.4735510832508523, "grad_norm": 0.3299097418785095, "learning_rate": 8e-05, "loss": 1.4649, "step": 6699 }, { "epoch": 1.4737710326624875, "grad_norm": 0.3465713858604431, "learning_rate": 8e-05, "loss": 1.5818, "step": 6700 }, { "epoch": 1.473990982074123, "grad_norm": 0.3494397699832916, "learning_rate": 8e-05, "loss": 1.5654, "step": 6701 }, { "epoch": 1.4742109314857583, "grad_norm": 0.34921789169311523, "learning_rate": 8e-05, "loss": 1.6978, "step": 6702 }, { "epoch": 1.4744308808973936, "grad_norm": 0.3753618896007538, "learning_rate": 8e-05, "loss": 1.6007, "step": 6703 }, { "epoch": 1.4746508303090289, "grad_norm": 0.34481823444366455, "learning_rate": 8e-05, "loss": 1.6405, "step": 6704 }, { "epoch": 1.4748707797206642, "grad_norm": 0.3509564995765686, "learning_rate": 8e-05, "loss": 1.5794, "step": 6705 }, { "epoch": 1.4750907291322997, "grad_norm": 0.3331051170825958, "learning_rate": 8e-05, "loss": 1.4814, "step": 6706 }, { "epoch": 1.475310678543935, "grad_norm": 0.36281606554985046, "learning_rate": 8e-05, "loss": 1.7449, "step": 6707 }, { "epoch": 1.4755306279555702, "grad_norm": 0.33139657974243164, "learning_rate": 8e-05, "loss": 1.5766, "step": 6708 }, { "epoch": 1.4757505773672055, "grad_norm": 0.31999585032463074, "learning_rate": 8e-05, "loss": 1.5084, "step": 6709 }, { "epoch": 1.4759705267788408, "grad_norm": 0.354501873254776, "learning_rate": 8e-05, "loss": 1.4333, "step": 6710 }, { "epoch": 1.4761904761904763, "grad_norm": 0.3311834931373596, "learning_rate": 8e-05, "loss": 1.5913, "step": 6711 }, { "epoch": 1.4764104256021116, "grad_norm": 0.3386140465736389, "learning_rate": 8e-05, "loss": 1.5738, "step": 6712 }, { "epoch": 1.4766303750137468, "grad_norm": 0.33449992537498474, "learning_rate": 8e-05, "loss": 1.4643, "step": 6713 }, { "epoch": 1.476850324425382, "grad_norm": 0.3780902624130249, "learning_rate": 8e-05, "loss": 1.7068, "step": 6714 }, { "epoch": 1.4770702738370174, "grad_norm": 0.3416253924369812, "learning_rate": 8e-05, "loss": 1.6137, "step": 6715 }, { "epoch": 1.477290223248653, "grad_norm": 0.3609880805015564, "learning_rate": 8e-05, "loss": 1.726, "step": 6716 }, { "epoch": 1.4775101726602882, "grad_norm": 0.3645568788051605, "learning_rate": 8e-05, "loss": 1.4622, "step": 6717 }, { "epoch": 1.4777301220719234, "grad_norm": 0.3351905643939972, "learning_rate": 8e-05, "loss": 1.5586, "step": 6718 }, { "epoch": 1.4779500714835587, "grad_norm": 0.3507244288921356, "learning_rate": 8e-05, "loss": 1.5817, "step": 6719 }, { "epoch": 1.478170020895194, "grad_norm": 0.3643120527267456, "learning_rate": 8e-05, "loss": 1.6674, "step": 6720 }, { "epoch": 1.4783899703068295, "grad_norm": 0.3607529401779175, "learning_rate": 8e-05, "loss": 1.6357, "step": 6721 }, { "epoch": 1.4786099197184648, "grad_norm": 0.35601285099983215, "learning_rate": 8e-05, "loss": 1.7592, "step": 6722 }, { "epoch": 1.4788298691301, "grad_norm": 0.3401491045951843, "learning_rate": 8e-05, "loss": 1.6301, "step": 6723 }, { "epoch": 1.4790498185417353, "grad_norm": 0.3692731559276581, "learning_rate": 8e-05, "loss": 1.6466, "step": 6724 }, { "epoch": 1.4792697679533706, "grad_norm": 0.345241516828537, "learning_rate": 8e-05, "loss": 1.6028, "step": 6725 }, { "epoch": 1.4794897173650061, "grad_norm": 0.3515876829624176, "learning_rate": 8e-05, "loss": 1.5965, "step": 6726 }, { "epoch": 1.4797096667766414, "grad_norm": 0.3440883755683899, "learning_rate": 8e-05, "loss": 1.5078, "step": 6727 }, { "epoch": 1.4799296161882767, "grad_norm": 0.32666510343551636, "learning_rate": 8e-05, "loss": 1.5019, "step": 6728 }, { "epoch": 1.480149565599912, "grad_norm": 0.33932289481163025, "learning_rate": 8e-05, "loss": 1.6155, "step": 6729 }, { "epoch": 1.4803695150115472, "grad_norm": 0.34410643577575684, "learning_rate": 8e-05, "loss": 1.6633, "step": 6730 }, { "epoch": 1.4805894644231827, "grad_norm": 0.34209951758384705, "learning_rate": 8e-05, "loss": 1.6079, "step": 6731 }, { "epoch": 1.480809413834818, "grad_norm": 0.3499366044998169, "learning_rate": 8e-05, "loss": 1.6103, "step": 6732 }, { "epoch": 1.4810293632464533, "grad_norm": 0.3390902578830719, "learning_rate": 8e-05, "loss": 1.6891, "step": 6733 }, { "epoch": 1.4812493126580886, "grad_norm": 0.36005938053131104, "learning_rate": 8e-05, "loss": 1.6548, "step": 6734 }, { "epoch": 1.4814692620697238, "grad_norm": 0.39136001467704773, "learning_rate": 8e-05, "loss": 1.8173, "step": 6735 }, { "epoch": 1.4816892114813593, "grad_norm": 0.349483847618103, "learning_rate": 8e-05, "loss": 1.6545, "step": 6736 }, { "epoch": 1.4819091608929946, "grad_norm": 0.35496512055397034, "learning_rate": 8e-05, "loss": 1.6205, "step": 6737 }, { "epoch": 1.48212911030463, "grad_norm": 0.3643285632133484, "learning_rate": 8e-05, "loss": 1.5421, "step": 6738 }, { "epoch": 1.4823490597162652, "grad_norm": 0.36531177163124084, "learning_rate": 8e-05, "loss": 1.7073, "step": 6739 }, { "epoch": 1.4825690091279005, "grad_norm": 0.3463575541973114, "learning_rate": 8e-05, "loss": 1.6452, "step": 6740 }, { "epoch": 1.482788958539536, "grad_norm": 0.3532589077949524, "learning_rate": 8e-05, "loss": 1.6335, "step": 6741 }, { "epoch": 1.4830089079511712, "grad_norm": 0.3580513596534729, "learning_rate": 8e-05, "loss": 1.471, "step": 6742 }, { "epoch": 1.4832288573628065, "grad_norm": 0.33774009346961975, "learning_rate": 8e-05, "loss": 1.605, "step": 6743 }, { "epoch": 1.4834488067744418, "grad_norm": 0.3682481646537781, "learning_rate": 8e-05, "loss": 1.7663, "step": 6744 }, { "epoch": 1.483668756186077, "grad_norm": 0.35692253708839417, "learning_rate": 8e-05, "loss": 1.7629, "step": 6745 }, { "epoch": 1.4838887055977126, "grad_norm": 0.34018054604530334, "learning_rate": 8e-05, "loss": 1.5155, "step": 6746 }, { "epoch": 1.4841086550093479, "grad_norm": 0.33997613191604614, "learning_rate": 8e-05, "loss": 1.5958, "step": 6747 }, { "epoch": 1.4843286044209831, "grad_norm": 0.3374512493610382, "learning_rate": 8e-05, "loss": 1.6652, "step": 6748 }, { "epoch": 1.4845485538326186, "grad_norm": 0.3345198333263397, "learning_rate": 8e-05, "loss": 1.5919, "step": 6749 }, { "epoch": 1.4847685032442537, "grad_norm": 0.34792134165763855, "learning_rate": 8e-05, "loss": 1.6982, "step": 6750 }, { "epoch": 1.4849884526558892, "grad_norm": 0.35036447644233704, "learning_rate": 8e-05, "loss": 1.6437, "step": 6751 }, { "epoch": 1.4852084020675245, "grad_norm": 0.3472917973995209, "learning_rate": 8e-05, "loss": 1.4654, "step": 6752 }, { "epoch": 1.4854283514791597, "grad_norm": 0.3089427649974823, "learning_rate": 8e-05, "loss": 1.3887, "step": 6753 }, { "epoch": 1.4856483008907952, "grad_norm": 0.3452395796775818, "learning_rate": 8e-05, "loss": 1.6728, "step": 6754 }, { "epoch": 1.4858682503024303, "grad_norm": 0.37810611724853516, "learning_rate": 8e-05, "loss": 1.5713, "step": 6755 }, { "epoch": 1.4860881997140658, "grad_norm": 0.3320801854133606, "learning_rate": 8e-05, "loss": 1.4995, "step": 6756 }, { "epoch": 1.486308149125701, "grad_norm": 0.34843480587005615, "learning_rate": 8e-05, "loss": 1.739, "step": 6757 }, { "epoch": 1.4865280985373364, "grad_norm": 0.3715517520904541, "learning_rate": 8e-05, "loss": 1.6698, "step": 6758 }, { "epoch": 1.4867480479489719, "grad_norm": 0.3534032702445984, "learning_rate": 8e-05, "loss": 1.755, "step": 6759 }, { "epoch": 1.4869679973606071, "grad_norm": 0.3259899318218231, "learning_rate": 8e-05, "loss": 1.4753, "step": 6760 }, { "epoch": 1.4871879467722424, "grad_norm": 0.39341607689857483, "learning_rate": 8e-05, "loss": 1.6197, "step": 6761 }, { "epoch": 1.4874078961838777, "grad_norm": 0.3402236998081207, "learning_rate": 8e-05, "loss": 1.6296, "step": 6762 }, { "epoch": 1.487627845595513, "grad_norm": 0.37127718329429626, "learning_rate": 8e-05, "loss": 1.665, "step": 6763 }, { "epoch": 1.4878477950071485, "grad_norm": 0.3346751034259796, "learning_rate": 8e-05, "loss": 1.5259, "step": 6764 }, { "epoch": 1.4880677444187838, "grad_norm": 0.3580843508243561, "learning_rate": 8e-05, "loss": 1.6182, "step": 6765 }, { "epoch": 1.488287693830419, "grad_norm": 0.33106541633605957, "learning_rate": 8e-05, "loss": 1.6078, "step": 6766 }, { "epoch": 1.4885076432420543, "grad_norm": 0.33741769194602966, "learning_rate": 8e-05, "loss": 1.5352, "step": 6767 }, { "epoch": 1.4887275926536896, "grad_norm": 0.3573254644870758, "learning_rate": 8e-05, "loss": 1.6115, "step": 6768 }, { "epoch": 1.488947542065325, "grad_norm": 0.35215190052986145, "learning_rate": 8e-05, "loss": 1.4973, "step": 6769 }, { "epoch": 1.4891674914769604, "grad_norm": 0.3504132330417633, "learning_rate": 8e-05, "loss": 1.6022, "step": 6770 }, { "epoch": 1.4893874408885956, "grad_norm": 0.3470858633518219, "learning_rate": 8e-05, "loss": 1.6559, "step": 6771 }, { "epoch": 1.489607390300231, "grad_norm": 0.37794312834739685, "learning_rate": 8e-05, "loss": 1.5518, "step": 6772 }, { "epoch": 1.4898273397118662, "grad_norm": 0.38429495692253113, "learning_rate": 8e-05, "loss": 1.7141, "step": 6773 }, { "epoch": 1.4900472891235017, "grad_norm": 0.33934980630874634, "learning_rate": 8e-05, "loss": 1.4091, "step": 6774 }, { "epoch": 1.490267238535137, "grad_norm": 0.3451005816459656, "learning_rate": 8e-05, "loss": 1.6162, "step": 6775 }, { "epoch": 1.4904871879467723, "grad_norm": 0.31960421800613403, "learning_rate": 8e-05, "loss": 1.5312, "step": 6776 }, { "epoch": 1.4907071373584075, "grad_norm": 0.34189173579216003, "learning_rate": 8e-05, "loss": 1.5363, "step": 6777 }, { "epoch": 1.4909270867700428, "grad_norm": 0.37075039744377136, "learning_rate": 8e-05, "loss": 1.6506, "step": 6778 }, { "epoch": 1.4911470361816783, "grad_norm": 0.34948328137397766, "learning_rate": 8e-05, "loss": 1.5273, "step": 6779 }, { "epoch": 1.4913669855933136, "grad_norm": 0.3259635269641876, "learning_rate": 8e-05, "loss": 1.5359, "step": 6780 }, { "epoch": 1.4915869350049489, "grad_norm": 0.33309251070022583, "learning_rate": 8e-05, "loss": 1.5983, "step": 6781 }, { "epoch": 1.4918068844165842, "grad_norm": 0.33799853920936584, "learning_rate": 8e-05, "loss": 1.5566, "step": 6782 }, { "epoch": 1.4920268338282194, "grad_norm": 0.34642601013183594, "learning_rate": 8e-05, "loss": 1.6152, "step": 6783 }, { "epoch": 1.492246783239855, "grad_norm": 0.36262521147727966, "learning_rate": 8e-05, "loss": 1.5538, "step": 6784 }, { "epoch": 1.4924667326514902, "grad_norm": 0.34209150075912476, "learning_rate": 8e-05, "loss": 1.6334, "step": 6785 }, { "epoch": 1.4926866820631255, "grad_norm": 0.3493581712245941, "learning_rate": 8e-05, "loss": 1.6431, "step": 6786 }, { "epoch": 1.4929066314747608, "grad_norm": 0.3157857060432434, "learning_rate": 8e-05, "loss": 1.4631, "step": 6787 }, { "epoch": 1.493126580886396, "grad_norm": 0.3329522907733917, "learning_rate": 8e-05, "loss": 1.5314, "step": 6788 }, { "epoch": 1.4933465302980315, "grad_norm": 0.3760625422000885, "learning_rate": 8e-05, "loss": 1.7359, "step": 6789 }, { "epoch": 1.4935664797096668, "grad_norm": 0.35832545161247253, "learning_rate": 8e-05, "loss": 1.5801, "step": 6790 }, { "epoch": 1.493786429121302, "grad_norm": 0.3514432907104492, "learning_rate": 8e-05, "loss": 1.6947, "step": 6791 }, { "epoch": 1.4940063785329374, "grad_norm": 0.34373965859413147, "learning_rate": 8e-05, "loss": 1.7971, "step": 6792 }, { "epoch": 1.4942263279445727, "grad_norm": 0.34526917338371277, "learning_rate": 8e-05, "loss": 1.4967, "step": 6793 }, { "epoch": 1.4944462773562082, "grad_norm": 0.33216044306755066, "learning_rate": 8e-05, "loss": 1.5325, "step": 6794 }, { "epoch": 1.4946662267678434, "grad_norm": 0.34151965379714966, "learning_rate": 8e-05, "loss": 1.6706, "step": 6795 }, { "epoch": 1.4948861761794787, "grad_norm": 0.3722078502178192, "learning_rate": 8e-05, "loss": 1.5811, "step": 6796 }, { "epoch": 1.495106125591114, "grad_norm": 0.3460828363895416, "learning_rate": 8e-05, "loss": 1.518, "step": 6797 }, { "epoch": 1.4953260750027493, "grad_norm": 0.36399251222610474, "learning_rate": 8e-05, "loss": 1.6701, "step": 6798 }, { "epoch": 1.4955460244143848, "grad_norm": 0.3546920120716095, "learning_rate": 8e-05, "loss": 1.61, "step": 6799 }, { "epoch": 1.49576597382602, "grad_norm": 0.3515170216560364, "learning_rate": 8e-05, "loss": 1.6022, "step": 6800 }, { "epoch": 1.4959859232376553, "grad_norm": 0.37607526779174805, "learning_rate": 8e-05, "loss": 1.7525, "step": 6801 }, { "epoch": 1.4962058726492906, "grad_norm": 0.3612443506717682, "learning_rate": 8e-05, "loss": 1.658, "step": 6802 }, { "epoch": 1.496425822060926, "grad_norm": 0.34270742535591125, "learning_rate": 8e-05, "loss": 1.6177, "step": 6803 }, { "epoch": 1.4966457714725614, "grad_norm": 0.332968533039093, "learning_rate": 8e-05, "loss": 1.447, "step": 6804 }, { "epoch": 1.4968657208841967, "grad_norm": 0.3476378917694092, "learning_rate": 8e-05, "loss": 1.6753, "step": 6805 }, { "epoch": 1.497085670295832, "grad_norm": 0.3383030593395233, "learning_rate": 8e-05, "loss": 1.6016, "step": 6806 }, { "epoch": 1.4973056197074672, "grad_norm": 0.3485026955604553, "learning_rate": 8e-05, "loss": 1.5109, "step": 6807 }, { "epoch": 1.4975255691191025, "grad_norm": 0.33920520544052124, "learning_rate": 8e-05, "loss": 1.5571, "step": 6808 }, { "epoch": 1.497745518530738, "grad_norm": 0.37536367774009705, "learning_rate": 8e-05, "loss": 1.6265, "step": 6809 }, { "epoch": 1.4979654679423733, "grad_norm": 0.3682669401168823, "learning_rate": 8e-05, "loss": 1.7842, "step": 6810 }, { "epoch": 1.4981854173540086, "grad_norm": 0.3555244505405426, "learning_rate": 8e-05, "loss": 1.6055, "step": 6811 }, { "epoch": 1.4984053667656438, "grad_norm": 0.3479917347431183, "learning_rate": 8e-05, "loss": 1.6996, "step": 6812 }, { "epoch": 1.4986253161772791, "grad_norm": 0.3478165566921234, "learning_rate": 8e-05, "loss": 1.6313, "step": 6813 }, { "epoch": 1.4988452655889146, "grad_norm": 0.35477060079574585, "learning_rate": 8e-05, "loss": 1.5621, "step": 6814 }, { "epoch": 1.49906521500055, "grad_norm": 0.35158461332321167, "learning_rate": 8e-05, "loss": 1.5441, "step": 6815 }, { "epoch": 1.4992851644121852, "grad_norm": 0.3479302227497101, "learning_rate": 8e-05, "loss": 1.6376, "step": 6816 }, { "epoch": 1.4995051138238205, "grad_norm": 0.37993723154067993, "learning_rate": 8e-05, "loss": 1.6232, "step": 6817 }, { "epoch": 1.4997250632354557, "grad_norm": 0.3537042737007141, "learning_rate": 8e-05, "loss": 1.7094, "step": 6818 }, { "epoch": 1.4999450126470912, "grad_norm": 0.3487173020839691, "learning_rate": 8e-05, "loss": 1.6021, "step": 6819 }, { "epoch": 1.5001649620587265, "grad_norm": 0.33808884024620056, "learning_rate": 8e-05, "loss": 1.5367, "step": 6820 }, { "epoch": 1.5003849114703618, "grad_norm": 0.3314312696456909, "learning_rate": 8e-05, "loss": 1.4962, "step": 6821 }, { "epoch": 1.5006048608819973, "grad_norm": 0.36458081007003784, "learning_rate": 8e-05, "loss": 1.7505, "step": 6822 }, { "epoch": 1.5008248102936323, "grad_norm": 0.34117045998573303, "learning_rate": 8e-05, "loss": 1.5576, "step": 6823 }, { "epoch": 1.5010447597052679, "grad_norm": 0.36340492963790894, "learning_rate": 8e-05, "loss": 1.608, "step": 6824 }, { "epoch": 1.5012647091169031, "grad_norm": 0.33395877480506897, "learning_rate": 8e-05, "loss": 1.5318, "step": 6825 }, { "epoch": 1.5014846585285384, "grad_norm": 0.36964181065559387, "learning_rate": 8e-05, "loss": 1.6723, "step": 6826 }, { "epoch": 1.501704607940174, "grad_norm": 0.35111913084983826, "learning_rate": 8e-05, "loss": 1.5698, "step": 6827 }, { "epoch": 1.501924557351809, "grad_norm": 0.3482351005077362, "learning_rate": 8e-05, "loss": 1.5722, "step": 6828 }, { "epoch": 1.5021445067634445, "grad_norm": 0.3467254936695099, "learning_rate": 8e-05, "loss": 1.5825, "step": 6829 }, { "epoch": 1.5023644561750797, "grad_norm": 0.3440031409263611, "learning_rate": 8e-05, "loss": 1.5338, "step": 6830 }, { "epoch": 1.502584405586715, "grad_norm": 0.3457756042480469, "learning_rate": 8e-05, "loss": 1.555, "step": 6831 }, { "epoch": 1.5028043549983505, "grad_norm": 0.3509039282798767, "learning_rate": 8e-05, "loss": 1.6283, "step": 6832 }, { "epoch": 1.5030243044099856, "grad_norm": 0.35430124402046204, "learning_rate": 8e-05, "loss": 1.5375, "step": 6833 }, { "epoch": 1.503244253821621, "grad_norm": 0.35982853174209595, "learning_rate": 8e-05, "loss": 1.7236, "step": 6834 }, { "epoch": 1.5034642032332564, "grad_norm": 0.3292112648487091, "learning_rate": 8e-05, "loss": 1.5373, "step": 6835 }, { "epoch": 1.5036841526448916, "grad_norm": 0.35925665497779846, "learning_rate": 8e-05, "loss": 1.6834, "step": 6836 }, { "epoch": 1.5039041020565271, "grad_norm": 0.34617987275123596, "learning_rate": 8e-05, "loss": 1.556, "step": 6837 }, { "epoch": 1.5041240514681622, "grad_norm": 0.33800071477890015, "learning_rate": 8e-05, "loss": 1.6196, "step": 6838 }, { "epoch": 1.5043440008797977, "grad_norm": 0.350546658039093, "learning_rate": 8e-05, "loss": 1.5127, "step": 6839 }, { "epoch": 1.504563950291433, "grad_norm": 0.35962867736816406, "learning_rate": 8e-05, "loss": 1.6716, "step": 6840 }, { "epoch": 1.5047838997030683, "grad_norm": 0.35551249980926514, "learning_rate": 8e-05, "loss": 1.7237, "step": 6841 }, { "epoch": 1.5050038491147038, "grad_norm": 0.3298833668231964, "learning_rate": 8e-05, "loss": 1.5445, "step": 6842 }, { "epoch": 1.5052237985263388, "grad_norm": 0.34092825651168823, "learning_rate": 8e-05, "loss": 1.6627, "step": 6843 }, { "epoch": 1.5054437479379743, "grad_norm": 0.33789050579071045, "learning_rate": 8e-05, "loss": 1.6822, "step": 6844 }, { "epoch": 1.5056636973496096, "grad_norm": 0.4366711974143982, "learning_rate": 8e-05, "loss": 1.4527, "step": 6845 }, { "epoch": 1.5058836467612449, "grad_norm": 0.37162622809410095, "learning_rate": 8e-05, "loss": 1.5998, "step": 6846 }, { "epoch": 1.5061035961728804, "grad_norm": 0.3435538411140442, "learning_rate": 8e-05, "loss": 1.4603, "step": 6847 }, { "epoch": 1.5063235455845154, "grad_norm": 0.35306230187416077, "learning_rate": 8e-05, "loss": 1.7163, "step": 6848 }, { "epoch": 1.506543494996151, "grad_norm": 0.3351747393608093, "learning_rate": 8e-05, "loss": 1.5381, "step": 6849 }, { "epoch": 1.5067634444077862, "grad_norm": 0.37543919682502747, "learning_rate": 8e-05, "loss": 1.6796, "step": 6850 }, { "epoch": 1.5069833938194215, "grad_norm": 0.34225577116012573, "learning_rate": 8e-05, "loss": 1.5531, "step": 6851 }, { "epoch": 1.507203343231057, "grad_norm": 0.3308938145637512, "learning_rate": 8e-05, "loss": 1.4817, "step": 6852 }, { "epoch": 1.507423292642692, "grad_norm": 0.35619664192199707, "learning_rate": 8e-05, "loss": 1.6092, "step": 6853 }, { "epoch": 1.5076432420543275, "grad_norm": 0.3353070020675659, "learning_rate": 8e-05, "loss": 1.5384, "step": 6854 }, { "epoch": 1.5078631914659628, "grad_norm": 0.33757683634757996, "learning_rate": 8e-05, "loss": 1.5529, "step": 6855 }, { "epoch": 1.508083140877598, "grad_norm": 0.33528271317481995, "learning_rate": 8e-05, "loss": 1.5418, "step": 6856 }, { "epoch": 1.5083030902892336, "grad_norm": 0.3424232006072998, "learning_rate": 8e-05, "loss": 1.6429, "step": 6857 }, { "epoch": 1.5085230397008687, "grad_norm": 0.3782368004322052, "learning_rate": 8e-05, "loss": 1.6728, "step": 6858 }, { "epoch": 1.5087429891125042, "grad_norm": 0.327402800321579, "learning_rate": 8e-05, "loss": 1.4465, "step": 6859 }, { "epoch": 1.5089629385241394, "grad_norm": 0.3247852027416229, "learning_rate": 8e-05, "loss": 1.4607, "step": 6860 }, { "epoch": 1.5091828879357747, "grad_norm": 0.33721497654914856, "learning_rate": 8e-05, "loss": 1.5971, "step": 6861 }, { "epoch": 1.5094028373474102, "grad_norm": 0.35093677043914795, "learning_rate": 8e-05, "loss": 1.6168, "step": 6862 }, { "epoch": 1.5096227867590453, "grad_norm": 0.3348419964313507, "learning_rate": 8e-05, "loss": 1.5916, "step": 6863 }, { "epoch": 1.5098427361706808, "grad_norm": 0.3405054807662964, "learning_rate": 8e-05, "loss": 1.7091, "step": 6864 }, { "epoch": 1.510062685582316, "grad_norm": 0.3512879014015198, "learning_rate": 8e-05, "loss": 1.574, "step": 6865 }, { "epoch": 1.5102826349939513, "grad_norm": 0.3376827538013458, "learning_rate": 8e-05, "loss": 1.6341, "step": 6866 }, { "epoch": 1.5105025844055868, "grad_norm": 0.37350261211395264, "learning_rate": 8e-05, "loss": 1.7553, "step": 6867 }, { "epoch": 1.5107225338172219, "grad_norm": 0.3258723020553589, "learning_rate": 8e-05, "loss": 1.4622, "step": 6868 }, { "epoch": 1.5109424832288574, "grad_norm": 0.35089972615242004, "learning_rate": 8e-05, "loss": 1.643, "step": 6869 }, { "epoch": 1.5111624326404927, "grad_norm": 0.34187957644462585, "learning_rate": 8e-05, "loss": 1.6963, "step": 6870 }, { "epoch": 1.511382382052128, "grad_norm": 0.3467416763305664, "learning_rate": 8e-05, "loss": 1.5374, "step": 6871 }, { "epoch": 1.5116023314637634, "grad_norm": 0.35036587715148926, "learning_rate": 8e-05, "loss": 1.6767, "step": 6872 }, { "epoch": 1.5118222808753985, "grad_norm": 0.3469410240650177, "learning_rate": 8e-05, "loss": 1.5198, "step": 6873 }, { "epoch": 1.512042230287034, "grad_norm": 0.3294687867164612, "learning_rate": 8e-05, "loss": 1.4973, "step": 6874 }, { "epoch": 1.5122621796986693, "grad_norm": 0.350572407245636, "learning_rate": 8e-05, "loss": 1.6509, "step": 6875 }, { "epoch": 1.5124821291103046, "grad_norm": 0.3576448857784271, "learning_rate": 8e-05, "loss": 1.629, "step": 6876 }, { "epoch": 1.51270207852194, "grad_norm": 0.33014482259750366, "learning_rate": 8e-05, "loss": 1.5065, "step": 6877 }, { "epoch": 1.5129220279335753, "grad_norm": 0.35229480266571045, "learning_rate": 8e-05, "loss": 1.6231, "step": 6878 }, { "epoch": 1.5131419773452106, "grad_norm": 0.3529079854488373, "learning_rate": 8e-05, "loss": 1.6592, "step": 6879 }, { "epoch": 1.5133619267568459, "grad_norm": 0.3364991247653961, "learning_rate": 8e-05, "loss": 1.6395, "step": 6880 }, { "epoch": 1.5135818761684812, "grad_norm": 0.3402341306209564, "learning_rate": 8e-05, "loss": 1.6528, "step": 6881 }, { "epoch": 1.5138018255801167, "grad_norm": 0.33507731556892395, "learning_rate": 8e-05, "loss": 1.56, "step": 6882 }, { "epoch": 1.514021774991752, "grad_norm": 0.3472770154476166, "learning_rate": 8e-05, "loss": 1.6341, "step": 6883 }, { "epoch": 1.5142417244033872, "grad_norm": 0.3506062626838684, "learning_rate": 8e-05, "loss": 1.4845, "step": 6884 }, { "epoch": 1.5144616738150225, "grad_norm": 0.3333079516887665, "learning_rate": 8e-05, "loss": 1.5303, "step": 6885 }, { "epoch": 1.5146816232266578, "grad_norm": 0.3290136456489563, "learning_rate": 8e-05, "loss": 1.6473, "step": 6886 }, { "epoch": 1.5149015726382933, "grad_norm": 0.33771151304244995, "learning_rate": 8e-05, "loss": 1.605, "step": 6887 }, { "epoch": 1.5151215220499286, "grad_norm": 0.3345109820365906, "learning_rate": 8e-05, "loss": 1.6293, "step": 6888 }, { "epoch": 1.5153414714615638, "grad_norm": 0.33077195286750793, "learning_rate": 8e-05, "loss": 1.5027, "step": 6889 }, { "epoch": 1.5155614208731991, "grad_norm": 0.35811492800712585, "learning_rate": 8e-05, "loss": 1.7405, "step": 6890 }, { "epoch": 1.5157813702848344, "grad_norm": 0.32588890194892883, "learning_rate": 8e-05, "loss": 1.4616, "step": 6891 }, { "epoch": 1.51600131969647, "grad_norm": 0.3409005105495453, "learning_rate": 8e-05, "loss": 1.5875, "step": 6892 }, { "epoch": 1.5162212691081052, "grad_norm": 0.35848426818847656, "learning_rate": 8e-05, "loss": 1.6237, "step": 6893 }, { "epoch": 1.5164412185197405, "grad_norm": 0.3348177969455719, "learning_rate": 8e-05, "loss": 1.4735, "step": 6894 }, { "epoch": 1.5166611679313757, "grad_norm": 0.3442823886871338, "learning_rate": 8e-05, "loss": 1.5091, "step": 6895 }, { "epoch": 1.516881117343011, "grad_norm": 0.347066193819046, "learning_rate": 8e-05, "loss": 1.6474, "step": 6896 }, { "epoch": 1.5171010667546465, "grad_norm": 0.3629944920539856, "learning_rate": 8e-05, "loss": 1.7038, "step": 6897 }, { "epoch": 1.5173210161662818, "grad_norm": 0.3692460358142853, "learning_rate": 8e-05, "loss": 1.5944, "step": 6898 }, { "epoch": 1.517540965577917, "grad_norm": 0.33668968081474304, "learning_rate": 8e-05, "loss": 1.5562, "step": 6899 }, { "epoch": 1.5177609149895526, "grad_norm": 0.3558357357978821, "learning_rate": 8e-05, "loss": 1.5452, "step": 6900 }, { "epoch": 1.5179808644011876, "grad_norm": 0.3507601022720337, "learning_rate": 8e-05, "loss": 1.6297, "step": 6901 }, { "epoch": 1.5182008138128231, "grad_norm": 0.3487901985645294, "learning_rate": 8e-05, "loss": 1.6855, "step": 6902 }, { "epoch": 1.5184207632244584, "grad_norm": 0.34774643182754517, "learning_rate": 8e-05, "loss": 1.6098, "step": 6903 }, { "epoch": 1.5186407126360937, "grad_norm": 0.3509933650493622, "learning_rate": 8e-05, "loss": 1.5967, "step": 6904 }, { "epoch": 1.5188606620477292, "grad_norm": 0.3464374244213104, "learning_rate": 8e-05, "loss": 1.6062, "step": 6905 }, { "epoch": 1.5190806114593642, "grad_norm": 0.37376701831817627, "learning_rate": 8e-05, "loss": 1.6734, "step": 6906 }, { "epoch": 1.5193005608709997, "grad_norm": 0.37075939774513245, "learning_rate": 8e-05, "loss": 1.6186, "step": 6907 }, { "epoch": 1.519520510282635, "grad_norm": 0.3511536121368408, "learning_rate": 8e-05, "loss": 1.6077, "step": 6908 }, { "epoch": 1.5197404596942703, "grad_norm": 0.34087932109832764, "learning_rate": 8e-05, "loss": 1.6271, "step": 6909 }, { "epoch": 1.5199604091059058, "grad_norm": 0.33164292573928833, "learning_rate": 8e-05, "loss": 1.5973, "step": 6910 }, { "epoch": 1.5201803585175409, "grad_norm": 0.3478955626487732, "learning_rate": 8e-05, "loss": 1.6762, "step": 6911 }, { "epoch": 1.5204003079291764, "grad_norm": 0.3476739525794983, "learning_rate": 8e-05, "loss": 1.4791, "step": 6912 }, { "epoch": 1.5206202573408116, "grad_norm": 0.35154807567596436, "learning_rate": 8e-05, "loss": 1.6934, "step": 6913 }, { "epoch": 1.520840206752447, "grad_norm": 0.34852108359336853, "learning_rate": 8e-05, "loss": 1.5102, "step": 6914 }, { "epoch": 1.5210601561640824, "grad_norm": 0.33500948548316956, "learning_rate": 8e-05, "loss": 1.576, "step": 6915 }, { "epoch": 1.5212801055757175, "grad_norm": 0.3446676433086395, "learning_rate": 8e-05, "loss": 1.6024, "step": 6916 }, { "epoch": 1.521500054987353, "grad_norm": 0.33995944261550903, "learning_rate": 8e-05, "loss": 1.5055, "step": 6917 }, { "epoch": 1.5217200043989882, "grad_norm": 0.32799360156059265, "learning_rate": 8e-05, "loss": 1.4597, "step": 6918 }, { "epoch": 1.5219399538106235, "grad_norm": 0.37319743633270264, "learning_rate": 8e-05, "loss": 1.6203, "step": 6919 }, { "epoch": 1.522159903222259, "grad_norm": 0.35828468203544617, "learning_rate": 8e-05, "loss": 1.5943, "step": 6920 }, { "epoch": 1.522379852633894, "grad_norm": 0.34003111720085144, "learning_rate": 8e-05, "loss": 1.6267, "step": 6921 }, { "epoch": 1.5225998020455296, "grad_norm": 0.34617534279823303, "learning_rate": 8e-05, "loss": 1.6425, "step": 6922 }, { "epoch": 1.5228197514571649, "grad_norm": 0.3407582938671112, "learning_rate": 8e-05, "loss": 1.5593, "step": 6923 }, { "epoch": 1.5230397008688001, "grad_norm": 0.35850071907043457, "learning_rate": 8e-05, "loss": 1.6649, "step": 6924 }, { "epoch": 1.5232596502804356, "grad_norm": 0.3511766791343689, "learning_rate": 8e-05, "loss": 1.7152, "step": 6925 }, { "epoch": 1.5234795996920707, "grad_norm": 0.34268978238105774, "learning_rate": 8e-05, "loss": 1.573, "step": 6926 }, { "epoch": 1.5236995491037062, "grad_norm": 0.3326343894004822, "learning_rate": 8e-05, "loss": 1.5186, "step": 6927 }, { "epoch": 1.5239194985153415, "grad_norm": 0.3577531576156616, "learning_rate": 8e-05, "loss": 1.716, "step": 6928 }, { "epoch": 1.5241394479269768, "grad_norm": 0.33629605174064636, "learning_rate": 8e-05, "loss": 1.4818, "step": 6929 }, { "epoch": 1.5243593973386123, "grad_norm": 0.3348085284233093, "learning_rate": 8e-05, "loss": 1.5482, "step": 6930 }, { "epoch": 1.5245793467502473, "grad_norm": 0.333825021982193, "learning_rate": 8e-05, "loss": 1.5105, "step": 6931 }, { "epoch": 1.5247992961618828, "grad_norm": 0.34642264246940613, "learning_rate": 8e-05, "loss": 1.5781, "step": 6932 }, { "epoch": 1.525019245573518, "grad_norm": 0.36821919679641724, "learning_rate": 8e-05, "loss": 1.5692, "step": 6933 }, { "epoch": 1.5252391949851534, "grad_norm": 0.34043025970458984, "learning_rate": 8e-05, "loss": 1.7334, "step": 6934 }, { "epoch": 1.5254591443967889, "grad_norm": 0.3392779529094696, "learning_rate": 8e-05, "loss": 1.5553, "step": 6935 }, { "epoch": 1.525679093808424, "grad_norm": 0.34243783354759216, "learning_rate": 8e-05, "loss": 1.6875, "step": 6936 }, { "epoch": 1.5258990432200594, "grad_norm": 0.3594472110271454, "learning_rate": 8e-05, "loss": 1.7005, "step": 6937 }, { "epoch": 1.5261189926316947, "grad_norm": 0.3486342132091522, "learning_rate": 8e-05, "loss": 1.6925, "step": 6938 }, { "epoch": 1.52633894204333, "grad_norm": 0.3412095606327057, "learning_rate": 8e-05, "loss": 1.6139, "step": 6939 }, { "epoch": 1.5265588914549655, "grad_norm": 0.33900925517082214, "learning_rate": 8e-05, "loss": 1.5114, "step": 6940 }, { "epoch": 1.5267788408666005, "grad_norm": 0.3572038412094116, "learning_rate": 8e-05, "loss": 1.5266, "step": 6941 }, { "epoch": 1.526998790278236, "grad_norm": 0.34324926137924194, "learning_rate": 8e-05, "loss": 1.6706, "step": 6942 }, { "epoch": 1.5272187396898713, "grad_norm": 0.3378064036369324, "learning_rate": 8e-05, "loss": 1.5683, "step": 6943 }, { "epoch": 1.5274386891015066, "grad_norm": 0.34111854434013367, "learning_rate": 8e-05, "loss": 1.5565, "step": 6944 }, { "epoch": 1.527658638513142, "grad_norm": 0.3344578444957733, "learning_rate": 8e-05, "loss": 1.5611, "step": 6945 }, { "epoch": 1.5278785879247772, "grad_norm": 0.3473542332649231, "learning_rate": 8e-05, "loss": 1.627, "step": 6946 }, { "epoch": 1.5280985373364127, "grad_norm": 0.3659147024154663, "learning_rate": 8e-05, "loss": 1.7304, "step": 6947 }, { "epoch": 1.528318486748048, "grad_norm": 0.36184337735176086, "learning_rate": 8e-05, "loss": 1.6083, "step": 6948 }, { "epoch": 1.5285384361596832, "grad_norm": 0.3444657325744629, "learning_rate": 8e-05, "loss": 1.6204, "step": 6949 }, { "epoch": 1.5287583855713187, "grad_norm": 0.36015865206718445, "learning_rate": 8e-05, "loss": 1.7018, "step": 6950 }, { "epoch": 1.5289783349829538, "grad_norm": 0.36784297227859497, "learning_rate": 8e-05, "loss": 1.3908, "step": 6951 }, { "epoch": 1.5291982843945893, "grad_norm": 0.31958112120628357, "learning_rate": 8e-05, "loss": 1.4822, "step": 6952 }, { "epoch": 1.5294182338062245, "grad_norm": 0.32958391308784485, "learning_rate": 8e-05, "loss": 1.4606, "step": 6953 }, { "epoch": 1.5296381832178598, "grad_norm": 0.35938677191734314, "learning_rate": 8e-05, "loss": 1.7059, "step": 6954 }, { "epoch": 1.5298581326294953, "grad_norm": 0.3251970708370209, "learning_rate": 8e-05, "loss": 1.5635, "step": 6955 }, { "epoch": 1.5300780820411304, "grad_norm": 0.3642410933971405, "learning_rate": 8e-05, "loss": 1.5798, "step": 6956 }, { "epoch": 1.5302980314527659, "grad_norm": 0.3325061500072479, "learning_rate": 8e-05, "loss": 1.5768, "step": 6957 }, { "epoch": 1.5305179808644012, "grad_norm": 0.3373423218727112, "learning_rate": 8e-05, "loss": 1.6401, "step": 6958 }, { "epoch": 1.5307379302760364, "grad_norm": 0.34439921379089355, "learning_rate": 8e-05, "loss": 1.5323, "step": 6959 }, { "epoch": 1.530957879687672, "grad_norm": 0.3245604634284973, "learning_rate": 8e-05, "loss": 1.4502, "step": 6960 }, { "epoch": 1.531177829099307, "grad_norm": 0.3605942130088806, "learning_rate": 8e-05, "loss": 1.6082, "step": 6961 }, { "epoch": 1.5313977785109425, "grad_norm": 0.3613329231739044, "learning_rate": 8e-05, "loss": 1.5195, "step": 6962 }, { "epoch": 1.5316177279225778, "grad_norm": 0.34805238246917725, "learning_rate": 8e-05, "loss": 1.5975, "step": 6963 }, { "epoch": 1.531837677334213, "grad_norm": 0.33522942662239075, "learning_rate": 8e-05, "loss": 1.5567, "step": 6964 }, { "epoch": 1.5320576267458486, "grad_norm": 0.33841249346733093, "learning_rate": 8e-05, "loss": 1.5637, "step": 6965 }, { "epoch": 1.5322775761574838, "grad_norm": 0.34547656774520874, "learning_rate": 8e-05, "loss": 1.5851, "step": 6966 }, { "epoch": 1.5324975255691191, "grad_norm": 0.32290545105934143, "learning_rate": 8e-05, "loss": 1.4211, "step": 6967 }, { "epoch": 1.5327174749807544, "grad_norm": 0.33402034640312195, "learning_rate": 8e-05, "loss": 1.5666, "step": 6968 }, { "epoch": 1.5329374243923897, "grad_norm": 0.35844454169273376, "learning_rate": 8e-05, "loss": 1.4761, "step": 6969 }, { "epoch": 1.5331573738040252, "grad_norm": 0.3536825478076935, "learning_rate": 8e-05, "loss": 1.6139, "step": 6970 }, { "epoch": 1.5333773232156604, "grad_norm": 0.3557147681713104, "learning_rate": 8e-05, "loss": 1.6062, "step": 6971 }, { "epoch": 1.5335972726272957, "grad_norm": 0.3383100926876068, "learning_rate": 8e-05, "loss": 1.5074, "step": 6972 }, { "epoch": 1.533817222038931, "grad_norm": 0.35654696822166443, "learning_rate": 8e-05, "loss": 1.5389, "step": 6973 }, { "epoch": 1.5340371714505663, "grad_norm": 0.33108365535736084, "learning_rate": 8e-05, "loss": 1.5193, "step": 6974 }, { "epoch": 1.5342571208622018, "grad_norm": 0.34586119651794434, "learning_rate": 8e-05, "loss": 1.5042, "step": 6975 }, { "epoch": 1.534477070273837, "grad_norm": 0.33299633860588074, "learning_rate": 8e-05, "loss": 1.5781, "step": 6976 }, { "epoch": 1.5346970196854723, "grad_norm": 0.36826568841934204, "learning_rate": 8e-05, "loss": 1.6044, "step": 6977 }, { "epoch": 1.5349169690971076, "grad_norm": 0.32781368494033813, "learning_rate": 8e-05, "loss": 1.5598, "step": 6978 }, { "epoch": 1.535136918508743, "grad_norm": 0.40465953946113586, "learning_rate": 8e-05, "loss": 1.6719, "step": 6979 }, { "epoch": 1.5353568679203784, "grad_norm": 0.3683792054653168, "learning_rate": 8e-05, "loss": 1.647, "step": 6980 }, { "epoch": 1.5355768173320137, "grad_norm": 0.36399969458580017, "learning_rate": 8e-05, "loss": 1.67, "step": 6981 }, { "epoch": 1.535796766743649, "grad_norm": 0.34425899386405945, "learning_rate": 8e-05, "loss": 1.5672, "step": 6982 }, { "epoch": 1.5360167161552842, "grad_norm": 0.3418682813644409, "learning_rate": 8e-05, "loss": 1.4254, "step": 6983 }, { "epoch": 1.5362366655669195, "grad_norm": 0.3359885513782501, "learning_rate": 8e-05, "loss": 1.5103, "step": 6984 }, { "epoch": 1.536456614978555, "grad_norm": 0.3228916823863983, "learning_rate": 8e-05, "loss": 1.4675, "step": 6985 }, { "epoch": 1.5366765643901903, "grad_norm": 0.32157376408576965, "learning_rate": 8e-05, "loss": 1.426, "step": 6986 }, { "epoch": 1.5368965138018256, "grad_norm": 0.34531882405281067, "learning_rate": 8e-05, "loss": 1.6361, "step": 6987 }, { "epoch": 1.5371164632134608, "grad_norm": 0.3327736258506775, "learning_rate": 8e-05, "loss": 1.4672, "step": 6988 }, { "epoch": 1.5373364126250961, "grad_norm": 0.3480343818664551, "learning_rate": 8e-05, "loss": 1.6068, "step": 6989 }, { "epoch": 1.5375563620367316, "grad_norm": 0.3546999394893646, "learning_rate": 8e-05, "loss": 1.6164, "step": 6990 }, { "epoch": 1.537776311448367, "grad_norm": 0.34107792377471924, "learning_rate": 8e-05, "loss": 1.6146, "step": 6991 }, { "epoch": 1.5379962608600022, "grad_norm": 0.3656320571899414, "learning_rate": 8e-05, "loss": 1.6461, "step": 6992 }, { "epoch": 1.5382162102716377, "grad_norm": 0.3599179983139038, "learning_rate": 8e-05, "loss": 1.6951, "step": 6993 }, { "epoch": 1.5384361596832727, "grad_norm": 0.3809260427951813, "learning_rate": 8e-05, "loss": 1.7704, "step": 6994 }, { "epoch": 1.5386561090949082, "grad_norm": 0.34450051188468933, "learning_rate": 8e-05, "loss": 1.4936, "step": 6995 }, { "epoch": 1.5388760585065435, "grad_norm": 0.3381282389163971, "learning_rate": 8e-05, "loss": 1.6474, "step": 6996 }, { "epoch": 1.5390960079181788, "grad_norm": 0.3657110929489136, "learning_rate": 8e-05, "loss": 1.5214, "step": 6997 }, { "epoch": 1.5393159573298143, "grad_norm": 0.34095335006713867, "learning_rate": 8e-05, "loss": 1.6274, "step": 6998 }, { "epoch": 1.5395359067414494, "grad_norm": 0.34200048446655273, "learning_rate": 8e-05, "loss": 1.6092, "step": 6999 }, { "epoch": 1.5397558561530849, "grad_norm": 0.3332284688949585, "learning_rate": 8e-05, "loss": 1.5489, "step": 7000 }, { "epoch": 1.5399758055647201, "grad_norm": 0.3637189269065857, "learning_rate": 8e-05, "loss": 1.5923, "step": 7001 }, { "epoch": 1.5401957549763554, "grad_norm": 0.36726921796798706, "learning_rate": 8e-05, "loss": 1.7235, "step": 7002 }, { "epoch": 1.540415704387991, "grad_norm": 0.34504783153533936, "learning_rate": 8e-05, "loss": 1.5741, "step": 7003 }, { "epoch": 1.540635653799626, "grad_norm": 0.34649813175201416, "learning_rate": 8e-05, "loss": 1.4814, "step": 7004 }, { "epoch": 1.5408556032112615, "grad_norm": 0.35015416145324707, "learning_rate": 8e-05, "loss": 1.5187, "step": 7005 }, { "epoch": 1.5410755526228967, "grad_norm": 0.3623475134372711, "learning_rate": 8e-05, "loss": 1.6897, "step": 7006 }, { "epoch": 1.541295502034532, "grad_norm": 0.3884846270084381, "learning_rate": 8e-05, "loss": 1.7049, "step": 7007 }, { "epoch": 1.5415154514461675, "grad_norm": 0.3347770571708679, "learning_rate": 8e-05, "loss": 1.4958, "step": 7008 }, { "epoch": 1.5417354008578026, "grad_norm": 0.3441554009914398, "learning_rate": 8e-05, "loss": 1.7135, "step": 7009 }, { "epoch": 1.541955350269438, "grad_norm": 0.3530106246471405, "learning_rate": 8e-05, "loss": 1.5112, "step": 7010 }, { "epoch": 1.5421752996810734, "grad_norm": 0.36142295598983765, "learning_rate": 8e-05, "loss": 1.559, "step": 7011 }, { "epoch": 1.5423952490927086, "grad_norm": 0.34351012110710144, "learning_rate": 8e-05, "loss": 1.7045, "step": 7012 }, { "epoch": 1.5426151985043441, "grad_norm": 0.3365376889705658, "learning_rate": 8e-05, "loss": 1.497, "step": 7013 }, { "epoch": 1.5428351479159792, "grad_norm": 0.4351274371147156, "learning_rate": 8e-05, "loss": 1.6336, "step": 7014 }, { "epoch": 1.5430550973276147, "grad_norm": 0.3351084887981415, "learning_rate": 8e-05, "loss": 1.64, "step": 7015 }, { "epoch": 1.54327504673925, "grad_norm": 0.35444220900535583, "learning_rate": 8e-05, "loss": 1.6653, "step": 7016 }, { "epoch": 1.5434949961508853, "grad_norm": 0.3464105427265167, "learning_rate": 8e-05, "loss": 1.6573, "step": 7017 }, { "epoch": 1.5437149455625208, "grad_norm": 0.3394041955471039, "learning_rate": 8e-05, "loss": 1.6139, "step": 7018 }, { "epoch": 1.5439348949741558, "grad_norm": 0.3467590808868408, "learning_rate": 8e-05, "loss": 1.6251, "step": 7019 }, { "epoch": 1.5441548443857913, "grad_norm": 0.3579201102256775, "learning_rate": 8e-05, "loss": 1.7087, "step": 7020 }, { "epoch": 1.5443747937974266, "grad_norm": 0.34546273946762085, "learning_rate": 8e-05, "loss": 1.55, "step": 7021 }, { "epoch": 1.5445947432090619, "grad_norm": 0.3595946133136749, "learning_rate": 8e-05, "loss": 1.6743, "step": 7022 }, { "epoch": 1.5448146926206974, "grad_norm": 0.3530957102775574, "learning_rate": 8e-05, "loss": 1.6065, "step": 7023 }, { "epoch": 1.5450346420323324, "grad_norm": 0.35392120480537415, "learning_rate": 8e-05, "loss": 1.7884, "step": 7024 }, { "epoch": 1.545254591443968, "grad_norm": 0.34644198417663574, "learning_rate": 8e-05, "loss": 1.4851, "step": 7025 }, { "epoch": 1.5454745408556032, "grad_norm": 0.3424748182296753, "learning_rate": 8e-05, "loss": 1.5347, "step": 7026 }, { "epoch": 1.5456944902672385, "grad_norm": 0.3352739214897156, "learning_rate": 8e-05, "loss": 1.5493, "step": 7027 }, { "epoch": 1.545914439678874, "grad_norm": 0.3402861952781677, "learning_rate": 8e-05, "loss": 1.696, "step": 7028 }, { "epoch": 1.546134389090509, "grad_norm": 0.3479061424732208, "learning_rate": 8e-05, "loss": 1.6602, "step": 7029 }, { "epoch": 1.5463543385021445, "grad_norm": 0.34846562147140503, "learning_rate": 8e-05, "loss": 1.5164, "step": 7030 }, { "epoch": 1.5465742879137798, "grad_norm": 0.3721713423728943, "learning_rate": 8e-05, "loss": 1.6427, "step": 7031 }, { "epoch": 1.546794237325415, "grad_norm": 0.3367082476615906, "learning_rate": 8e-05, "loss": 1.5736, "step": 7032 }, { "epoch": 1.5470141867370506, "grad_norm": 0.32471325993537903, "learning_rate": 8e-05, "loss": 1.5888, "step": 7033 }, { "epoch": 1.5472341361486857, "grad_norm": 0.3458797335624695, "learning_rate": 8e-05, "loss": 1.5686, "step": 7034 }, { "epoch": 1.5474540855603212, "grad_norm": 0.35171860456466675, "learning_rate": 8e-05, "loss": 1.7824, "step": 7035 }, { "epoch": 1.5476740349719564, "grad_norm": 0.3337644338607788, "learning_rate": 8e-05, "loss": 1.5774, "step": 7036 }, { "epoch": 1.5478939843835917, "grad_norm": 0.34575554728507996, "learning_rate": 8e-05, "loss": 1.4733, "step": 7037 }, { "epoch": 1.5481139337952272, "grad_norm": 0.3767777681350708, "learning_rate": 8e-05, "loss": 1.522, "step": 7038 }, { "epoch": 1.5483338832068623, "grad_norm": 0.3338122069835663, "learning_rate": 8e-05, "loss": 1.5821, "step": 7039 }, { "epoch": 1.5485538326184978, "grad_norm": 0.33284181356430054, "learning_rate": 8e-05, "loss": 1.653, "step": 7040 }, { "epoch": 1.548773782030133, "grad_norm": 0.36062926054000854, "learning_rate": 8e-05, "loss": 1.7027, "step": 7041 }, { "epoch": 1.5489937314417683, "grad_norm": 0.3434266448020935, "learning_rate": 8e-05, "loss": 1.5295, "step": 7042 }, { "epoch": 1.5492136808534038, "grad_norm": 0.345896452665329, "learning_rate": 8e-05, "loss": 1.6378, "step": 7043 }, { "epoch": 1.5494336302650389, "grad_norm": 0.3212950527667999, "learning_rate": 8e-05, "loss": 1.4959, "step": 7044 }, { "epoch": 1.5496535796766744, "grad_norm": 0.35252103209495544, "learning_rate": 8e-05, "loss": 1.7178, "step": 7045 }, { "epoch": 1.5498735290883097, "grad_norm": 0.34411361813545227, "learning_rate": 8e-05, "loss": 1.5999, "step": 7046 }, { "epoch": 1.550093478499945, "grad_norm": 0.3679974675178528, "learning_rate": 8e-05, "loss": 1.4614, "step": 7047 }, { "epoch": 1.5503134279115804, "grad_norm": 0.38004031777381897, "learning_rate": 8e-05, "loss": 1.6179, "step": 7048 }, { "epoch": 1.5505333773232155, "grad_norm": 0.362578421831131, "learning_rate": 8e-05, "loss": 1.7187, "step": 7049 }, { "epoch": 1.550753326734851, "grad_norm": 0.34759625792503357, "learning_rate": 8e-05, "loss": 1.6183, "step": 7050 }, { "epoch": 1.5509732761464863, "grad_norm": 0.35028231143951416, "learning_rate": 8e-05, "loss": 1.5276, "step": 7051 }, { "epoch": 1.5511932255581216, "grad_norm": 0.36650606989860535, "learning_rate": 8e-05, "loss": 1.8326, "step": 7052 }, { "epoch": 1.551413174969757, "grad_norm": 0.40062233805656433, "learning_rate": 8e-05, "loss": 1.6357, "step": 7053 }, { "epoch": 1.5516331243813921, "grad_norm": 0.3425028324127197, "learning_rate": 8e-05, "loss": 1.6301, "step": 7054 }, { "epoch": 1.5518530737930276, "grad_norm": 0.37011367082595825, "learning_rate": 8e-05, "loss": 1.6044, "step": 7055 }, { "epoch": 1.552073023204663, "grad_norm": 0.35089150071144104, "learning_rate": 8e-05, "loss": 1.6094, "step": 7056 }, { "epoch": 1.5522929726162982, "grad_norm": 0.35157760977745056, "learning_rate": 8e-05, "loss": 1.5483, "step": 7057 }, { "epoch": 1.5525129220279337, "grad_norm": 0.3449578285217285, "learning_rate": 8e-05, "loss": 1.5756, "step": 7058 }, { "epoch": 1.552732871439569, "grad_norm": 0.3463907539844513, "learning_rate": 8e-05, "loss": 1.5877, "step": 7059 }, { "epoch": 1.5529528208512042, "grad_norm": 0.37291303277015686, "learning_rate": 8e-05, "loss": 1.6679, "step": 7060 }, { "epoch": 1.5531727702628395, "grad_norm": 0.3425915837287903, "learning_rate": 8e-05, "loss": 1.7301, "step": 7061 }, { "epoch": 1.5533927196744748, "grad_norm": 0.3805132806301117, "learning_rate": 8e-05, "loss": 1.6631, "step": 7062 }, { "epoch": 1.5536126690861103, "grad_norm": 0.39177343249320984, "learning_rate": 8e-05, "loss": 1.7059, "step": 7063 }, { "epoch": 1.5538326184977456, "grad_norm": 0.34290894865989685, "learning_rate": 8e-05, "loss": 1.596, "step": 7064 }, { "epoch": 1.5540525679093808, "grad_norm": 0.36788371205329895, "learning_rate": 8e-05, "loss": 1.7294, "step": 7065 }, { "epoch": 1.5542725173210161, "grad_norm": 0.34456467628479004, "learning_rate": 8e-05, "loss": 1.557, "step": 7066 }, { "epoch": 1.5544924667326514, "grad_norm": 0.3326578438282013, "learning_rate": 8e-05, "loss": 1.439, "step": 7067 }, { "epoch": 1.554712416144287, "grad_norm": 0.34050145745277405, "learning_rate": 8e-05, "loss": 1.5437, "step": 7068 }, { "epoch": 1.5549323655559222, "grad_norm": 0.34869468212127686, "learning_rate": 8e-05, "loss": 1.5642, "step": 7069 }, { "epoch": 1.5551523149675575, "grad_norm": 0.32963672280311584, "learning_rate": 8e-05, "loss": 1.458, "step": 7070 }, { "epoch": 1.5553722643791927, "grad_norm": 0.35526415705680847, "learning_rate": 8e-05, "loss": 1.6025, "step": 7071 }, { "epoch": 1.555592213790828, "grad_norm": 0.3589567244052887, "learning_rate": 8e-05, "loss": 1.5217, "step": 7072 }, { "epoch": 1.5558121632024635, "grad_norm": 0.37806087732315063, "learning_rate": 8e-05, "loss": 1.7238, "step": 7073 }, { "epoch": 1.5560321126140988, "grad_norm": 0.3590231239795685, "learning_rate": 8e-05, "loss": 1.5416, "step": 7074 }, { "epoch": 1.556252062025734, "grad_norm": 0.32058629393577576, "learning_rate": 8e-05, "loss": 1.4483, "step": 7075 }, { "epoch": 1.5564720114373694, "grad_norm": 0.39100372791290283, "learning_rate": 8e-05, "loss": 1.6051, "step": 7076 }, { "epoch": 1.5566919608490046, "grad_norm": 0.3489615321159363, "learning_rate": 8e-05, "loss": 1.5317, "step": 7077 }, { "epoch": 1.5569119102606401, "grad_norm": 0.3447747230529785, "learning_rate": 8e-05, "loss": 1.6818, "step": 7078 }, { "epoch": 1.5571318596722754, "grad_norm": 0.3863576352596283, "learning_rate": 8e-05, "loss": 1.7309, "step": 7079 }, { "epoch": 1.5573518090839107, "grad_norm": 0.3539600968360901, "learning_rate": 8e-05, "loss": 1.5218, "step": 7080 }, { "epoch": 1.5575717584955462, "grad_norm": 0.3678518235683441, "learning_rate": 8e-05, "loss": 1.6682, "step": 7081 }, { "epoch": 1.5577917079071812, "grad_norm": 0.35988232493400574, "learning_rate": 8e-05, "loss": 1.7184, "step": 7082 }, { "epoch": 1.5580116573188167, "grad_norm": 0.35903623700141907, "learning_rate": 8e-05, "loss": 1.6448, "step": 7083 }, { "epoch": 1.558231606730452, "grad_norm": 0.3380880653858185, "learning_rate": 8e-05, "loss": 1.4487, "step": 7084 }, { "epoch": 1.5584515561420873, "grad_norm": 0.34797918796539307, "learning_rate": 8e-05, "loss": 1.6632, "step": 7085 }, { "epoch": 1.5586715055537228, "grad_norm": 0.3584599792957306, "learning_rate": 8e-05, "loss": 1.651, "step": 7086 }, { "epoch": 1.5588914549653579, "grad_norm": 0.34418028593063354, "learning_rate": 8e-05, "loss": 1.6386, "step": 7087 }, { "epoch": 1.5591114043769934, "grad_norm": 0.3517943322658539, "learning_rate": 8e-05, "loss": 1.571, "step": 7088 }, { "epoch": 1.5593313537886286, "grad_norm": 0.3382527232170105, "learning_rate": 8e-05, "loss": 1.5094, "step": 7089 }, { "epoch": 1.559551303200264, "grad_norm": 0.3561500012874603, "learning_rate": 8e-05, "loss": 1.6491, "step": 7090 }, { "epoch": 1.5597712526118994, "grad_norm": 0.33356204628944397, "learning_rate": 8e-05, "loss": 1.6268, "step": 7091 }, { "epoch": 1.5599912020235345, "grad_norm": 0.33503445982933044, "learning_rate": 8e-05, "loss": 1.6496, "step": 7092 }, { "epoch": 1.56021115143517, "grad_norm": 0.3374323546886444, "learning_rate": 8e-05, "loss": 1.5581, "step": 7093 }, { "epoch": 1.5604311008468053, "grad_norm": 0.3360513150691986, "learning_rate": 8e-05, "loss": 1.5604, "step": 7094 }, { "epoch": 1.5606510502584405, "grad_norm": 0.3350540101528168, "learning_rate": 8e-05, "loss": 1.5145, "step": 7095 }, { "epoch": 1.560870999670076, "grad_norm": 0.39198821783065796, "learning_rate": 8e-05, "loss": 1.8234, "step": 7096 }, { "epoch": 1.561090949081711, "grad_norm": 0.33092570304870605, "learning_rate": 8e-05, "loss": 1.6018, "step": 7097 }, { "epoch": 1.5613108984933466, "grad_norm": 0.368995726108551, "learning_rate": 8e-05, "loss": 1.5902, "step": 7098 }, { "epoch": 1.5615308479049819, "grad_norm": 0.36250847578048706, "learning_rate": 8e-05, "loss": 1.5567, "step": 7099 }, { "epoch": 1.5617507973166171, "grad_norm": 0.34518226981163025, "learning_rate": 8e-05, "loss": 1.4928, "step": 7100 }, { "epoch": 1.5619707467282526, "grad_norm": 0.36045128107070923, "learning_rate": 8e-05, "loss": 1.6159, "step": 7101 }, { "epoch": 1.5621906961398877, "grad_norm": 0.3695918023586273, "learning_rate": 8e-05, "loss": 1.5646, "step": 7102 }, { "epoch": 1.5624106455515232, "grad_norm": 0.36813026666641235, "learning_rate": 8e-05, "loss": 1.6876, "step": 7103 }, { "epoch": 1.5626305949631585, "grad_norm": 0.36714938282966614, "learning_rate": 8e-05, "loss": 1.5966, "step": 7104 }, { "epoch": 1.5628505443747938, "grad_norm": 0.33665919303894043, "learning_rate": 8e-05, "loss": 1.4825, "step": 7105 }, { "epoch": 1.5630704937864293, "grad_norm": 0.3344520628452301, "learning_rate": 8e-05, "loss": 1.5845, "step": 7106 }, { "epoch": 1.5632904431980643, "grad_norm": 0.3314379155635834, "learning_rate": 8e-05, "loss": 1.513, "step": 7107 }, { "epoch": 1.5635103926096998, "grad_norm": 0.347854346036911, "learning_rate": 8e-05, "loss": 1.5947, "step": 7108 }, { "epoch": 1.563730342021335, "grad_norm": 0.39210936427116394, "learning_rate": 8e-05, "loss": 1.5939, "step": 7109 }, { "epoch": 1.5639502914329704, "grad_norm": 0.35417407751083374, "learning_rate": 8e-05, "loss": 1.5512, "step": 7110 }, { "epoch": 1.5641702408446059, "grad_norm": 0.33948925137519836, "learning_rate": 8e-05, "loss": 1.6142, "step": 7111 }, { "epoch": 1.564390190256241, "grad_norm": 0.37921157479286194, "learning_rate": 8e-05, "loss": 1.6019, "step": 7112 }, { "epoch": 1.5646101396678764, "grad_norm": 0.35170260071754456, "learning_rate": 8e-05, "loss": 1.6662, "step": 7113 }, { "epoch": 1.5648300890795117, "grad_norm": 0.3793729841709137, "learning_rate": 8e-05, "loss": 1.6495, "step": 7114 }, { "epoch": 1.565050038491147, "grad_norm": 0.34262242913246155, "learning_rate": 8e-05, "loss": 1.6691, "step": 7115 }, { "epoch": 1.5652699879027825, "grad_norm": 0.3372538089752197, "learning_rate": 8e-05, "loss": 1.7444, "step": 7116 }, { "epoch": 1.5654899373144175, "grad_norm": 0.33903810381889343, "learning_rate": 8e-05, "loss": 1.5142, "step": 7117 }, { "epoch": 1.565709886726053, "grad_norm": 0.33734360337257385, "learning_rate": 8e-05, "loss": 1.403, "step": 7118 }, { "epoch": 1.5659298361376883, "grad_norm": 0.3643967807292938, "learning_rate": 8e-05, "loss": 1.6414, "step": 7119 }, { "epoch": 1.5661497855493236, "grad_norm": 0.3470838963985443, "learning_rate": 8e-05, "loss": 1.6243, "step": 7120 }, { "epoch": 1.566369734960959, "grad_norm": 0.34744367003440857, "learning_rate": 8e-05, "loss": 1.6223, "step": 7121 }, { "epoch": 1.5665896843725942, "grad_norm": 0.345938116312027, "learning_rate": 8e-05, "loss": 1.517, "step": 7122 }, { "epoch": 1.5668096337842297, "grad_norm": 0.3397997319698334, "learning_rate": 8e-05, "loss": 1.6056, "step": 7123 }, { "epoch": 1.567029583195865, "grad_norm": 0.341140478849411, "learning_rate": 8e-05, "loss": 1.5289, "step": 7124 }, { "epoch": 1.5672495326075002, "grad_norm": 0.34139782190322876, "learning_rate": 8e-05, "loss": 1.6057, "step": 7125 }, { "epoch": 1.5674694820191357, "grad_norm": 0.34360867738723755, "learning_rate": 8e-05, "loss": 1.5439, "step": 7126 }, { "epoch": 1.5676894314307708, "grad_norm": 0.3535831570625305, "learning_rate": 8e-05, "loss": 1.6664, "step": 7127 }, { "epoch": 1.5679093808424063, "grad_norm": 0.34051769971847534, "learning_rate": 8e-05, "loss": 1.5532, "step": 7128 }, { "epoch": 1.5681293302540416, "grad_norm": 0.34564119577407837, "learning_rate": 8e-05, "loss": 1.6542, "step": 7129 }, { "epoch": 1.5683492796656768, "grad_norm": 0.3367798626422882, "learning_rate": 8e-05, "loss": 1.5791, "step": 7130 }, { "epoch": 1.5685692290773123, "grad_norm": 0.3499099016189575, "learning_rate": 8e-05, "loss": 1.6575, "step": 7131 }, { "epoch": 1.5687891784889474, "grad_norm": 0.3449278175830841, "learning_rate": 8e-05, "loss": 1.6208, "step": 7132 }, { "epoch": 1.569009127900583, "grad_norm": 0.3458971083164215, "learning_rate": 8e-05, "loss": 1.6829, "step": 7133 }, { "epoch": 1.5692290773122182, "grad_norm": 0.36181560158729553, "learning_rate": 8e-05, "loss": 1.5871, "step": 7134 }, { "epoch": 1.5694490267238534, "grad_norm": 0.3260992765426636, "learning_rate": 8e-05, "loss": 1.4753, "step": 7135 }, { "epoch": 1.569668976135489, "grad_norm": 0.3622412085533142, "learning_rate": 8e-05, "loss": 1.5216, "step": 7136 }, { "epoch": 1.569888925547124, "grad_norm": 0.33729612827301025, "learning_rate": 8e-05, "loss": 1.5962, "step": 7137 }, { "epoch": 1.5701088749587595, "grad_norm": 0.34679049253463745, "learning_rate": 8e-05, "loss": 1.6478, "step": 7138 }, { "epoch": 1.5703288243703948, "grad_norm": 0.4408310651779175, "learning_rate": 8e-05, "loss": 1.6321, "step": 7139 }, { "epoch": 1.57054877378203, "grad_norm": 0.3496437966823578, "learning_rate": 8e-05, "loss": 1.6144, "step": 7140 }, { "epoch": 1.5707687231936656, "grad_norm": 0.3485501706600189, "learning_rate": 8e-05, "loss": 1.7028, "step": 7141 }, { "epoch": 1.5709886726053006, "grad_norm": 0.3518679738044739, "learning_rate": 8e-05, "loss": 1.7212, "step": 7142 }, { "epoch": 1.5712086220169361, "grad_norm": 0.33076414465904236, "learning_rate": 8e-05, "loss": 1.6075, "step": 7143 }, { "epoch": 1.5714285714285714, "grad_norm": 0.35183218121528625, "learning_rate": 8e-05, "loss": 1.5874, "step": 7144 }, { "epoch": 1.5716485208402067, "grad_norm": 0.34682831168174744, "learning_rate": 8e-05, "loss": 1.5662, "step": 7145 }, { "epoch": 1.5718684702518422, "grad_norm": 0.37010976672172546, "learning_rate": 8e-05, "loss": 1.6315, "step": 7146 }, { "epoch": 1.5720884196634775, "grad_norm": 0.36660051345825195, "learning_rate": 8e-05, "loss": 1.6434, "step": 7147 }, { "epoch": 1.5723083690751127, "grad_norm": 0.33051127195358276, "learning_rate": 8e-05, "loss": 1.5558, "step": 7148 }, { "epoch": 1.572528318486748, "grad_norm": 0.3519337475299835, "learning_rate": 8e-05, "loss": 1.4806, "step": 7149 }, { "epoch": 1.5727482678983833, "grad_norm": 0.35317865014076233, "learning_rate": 8e-05, "loss": 1.7162, "step": 7150 }, { "epoch": 1.5729682173100188, "grad_norm": 0.3526594042778015, "learning_rate": 8e-05, "loss": 1.5249, "step": 7151 }, { "epoch": 1.573188166721654, "grad_norm": 0.3479490578174591, "learning_rate": 8e-05, "loss": 1.687, "step": 7152 }, { "epoch": 1.5734081161332893, "grad_norm": 0.36917969584465027, "learning_rate": 8e-05, "loss": 1.7036, "step": 7153 }, { "epoch": 1.5736280655449246, "grad_norm": 0.3321349322795868, "learning_rate": 8e-05, "loss": 1.4279, "step": 7154 }, { "epoch": 1.57384801495656, "grad_norm": 0.33076152205467224, "learning_rate": 8e-05, "loss": 1.544, "step": 7155 }, { "epoch": 1.5740679643681954, "grad_norm": 0.3588312268257141, "learning_rate": 8e-05, "loss": 1.697, "step": 7156 }, { "epoch": 1.5742879137798307, "grad_norm": 0.3561347424983978, "learning_rate": 8e-05, "loss": 1.6277, "step": 7157 }, { "epoch": 1.574507863191466, "grad_norm": 0.3867541551589966, "learning_rate": 8e-05, "loss": 1.6357, "step": 7158 }, { "epoch": 1.5747278126031012, "grad_norm": 0.3436915874481201, "learning_rate": 8e-05, "loss": 1.6014, "step": 7159 }, { "epoch": 1.5749477620147365, "grad_norm": 0.33482062816619873, "learning_rate": 8e-05, "loss": 1.5436, "step": 7160 }, { "epoch": 1.575167711426372, "grad_norm": 0.35630127787590027, "learning_rate": 8e-05, "loss": 1.5382, "step": 7161 }, { "epoch": 1.5753876608380073, "grad_norm": 0.3346439599990845, "learning_rate": 8e-05, "loss": 1.6496, "step": 7162 }, { "epoch": 1.5756076102496426, "grad_norm": 0.3354346752166748, "learning_rate": 8e-05, "loss": 1.5779, "step": 7163 }, { "epoch": 1.5758275596612779, "grad_norm": 0.3321080803871155, "learning_rate": 8e-05, "loss": 1.5236, "step": 7164 }, { "epoch": 1.5760475090729131, "grad_norm": 0.35478687286376953, "learning_rate": 8e-05, "loss": 1.5103, "step": 7165 }, { "epoch": 1.5762674584845486, "grad_norm": 0.35751616954803467, "learning_rate": 8e-05, "loss": 1.699, "step": 7166 }, { "epoch": 1.576487407896184, "grad_norm": 0.34246498346328735, "learning_rate": 8e-05, "loss": 1.4626, "step": 7167 }, { "epoch": 1.5767073573078192, "grad_norm": 0.35325533151626587, "learning_rate": 8e-05, "loss": 1.632, "step": 7168 }, { "epoch": 1.5769273067194547, "grad_norm": 0.3400254547595978, "learning_rate": 8e-05, "loss": 1.5277, "step": 7169 }, { "epoch": 1.5771472561310897, "grad_norm": 0.3375821113586426, "learning_rate": 8e-05, "loss": 1.5149, "step": 7170 }, { "epoch": 1.5773672055427252, "grad_norm": 0.3491111993789673, "learning_rate": 8e-05, "loss": 1.7181, "step": 7171 }, { "epoch": 1.5775871549543605, "grad_norm": 0.3898676335811615, "learning_rate": 8e-05, "loss": 1.5943, "step": 7172 }, { "epoch": 1.5778071043659958, "grad_norm": 0.3414188623428345, "learning_rate": 8e-05, "loss": 1.5403, "step": 7173 }, { "epoch": 1.5780270537776313, "grad_norm": 0.32562801241874695, "learning_rate": 8e-05, "loss": 1.4489, "step": 7174 }, { "epoch": 1.5782470031892664, "grad_norm": 0.34717682003974915, "learning_rate": 8e-05, "loss": 1.588, "step": 7175 }, { "epoch": 1.5784669526009019, "grad_norm": 0.35878923535346985, "learning_rate": 8e-05, "loss": 1.611, "step": 7176 }, { "epoch": 1.5786869020125371, "grad_norm": 0.3335525691509247, "learning_rate": 8e-05, "loss": 1.5199, "step": 7177 }, { "epoch": 1.5789068514241724, "grad_norm": 0.3273618519306183, "learning_rate": 8e-05, "loss": 1.5175, "step": 7178 }, { "epoch": 1.579126800835808, "grad_norm": 0.340351402759552, "learning_rate": 8e-05, "loss": 1.5782, "step": 7179 }, { "epoch": 1.579346750247443, "grad_norm": 0.3474087119102478, "learning_rate": 8e-05, "loss": 1.578, "step": 7180 }, { "epoch": 1.5795666996590785, "grad_norm": 0.36922940611839294, "learning_rate": 8e-05, "loss": 1.6969, "step": 7181 }, { "epoch": 1.5797866490707138, "grad_norm": 0.3397550582885742, "learning_rate": 8e-05, "loss": 1.5928, "step": 7182 }, { "epoch": 1.580006598482349, "grad_norm": 0.3391132652759552, "learning_rate": 8e-05, "loss": 1.5291, "step": 7183 }, { "epoch": 1.5802265478939845, "grad_norm": 0.35304561257362366, "learning_rate": 8e-05, "loss": 1.6616, "step": 7184 }, { "epoch": 1.5804464973056196, "grad_norm": 0.36622098088264465, "learning_rate": 8e-05, "loss": 1.7786, "step": 7185 }, { "epoch": 1.580666446717255, "grad_norm": 0.3502393066883087, "learning_rate": 8e-05, "loss": 1.5598, "step": 7186 }, { "epoch": 1.5808863961288904, "grad_norm": 0.34195923805236816, "learning_rate": 8e-05, "loss": 1.5811, "step": 7187 }, { "epoch": 1.5811063455405256, "grad_norm": 0.380433052778244, "learning_rate": 8e-05, "loss": 1.5164, "step": 7188 }, { "epoch": 1.5813262949521611, "grad_norm": 0.3544248938560486, "learning_rate": 8e-05, "loss": 1.449, "step": 7189 }, { "epoch": 1.5815462443637962, "grad_norm": 0.3556905686855316, "learning_rate": 8e-05, "loss": 1.5813, "step": 7190 }, { "epoch": 1.5817661937754317, "grad_norm": 0.35380855202674866, "learning_rate": 8e-05, "loss": 1.5306, "step": 7191 }, { "epoch": 1.581986143187067, "grad_norm": 0.34805819392204285, "learning_rate": 8e-05, "loss": 1.6354, "step": 7192 }, { "epoch": 1.5822060925987023, "grad_norm": 0.3511526584625244, "learning_rate": 8e-05, "loss": 1.5202, "step": 7193 }, { "epoch": 1.5824260420103378, "grad_norm": 0.35779768228530884, "learning_rate": 8e-05, "loss": 1.5938, "step": 7194 }, { "epoch": 1.5826459914219728, "grad_norm": 0.3188896179199219, "learning_rate": 8e-05, "loss": 1.4048, "step": 7195 }, { "epoch": 1.5828659408336083, "grad_norm": 0.34854015707969666, "learning_rate": 8e-05, "loss": 1.6286, "step": 7196 }, { "epoch": 1.5830858902452436, "grad_norm": 0.3547763228416443, "learning_rate": 8e-05, "loss": 1.5131, "step": 7197 }, { "epoch": 1.5833058396568789, "grad_norm": 0.31795597076416016, "learning_rate": 8e-05, "loss": 1.5008, "step": 7198 }, { "epoch": 1.5835257890685144, "grad_norm": 0.35328438878059387, "learning_rate": 8e-05, "loss": 1.636, "step": 7199 }, { "epoch": 1.5837457384801494, "grad_norm": 0.3513997197151184, "learning_rate": 8e-05, "loss": 1.5002, "step": 7200 }, { "epoch": 1.583965687891785, "grad_norm": 0.34078308939933777, "learning_rate": 8e-05, "loss": 1.6444, "step": 7201 }, { "epoch": 1.5841856373034202, "grad_norm": 0.334715873003006, "learning_rate": 8e-05, "loss": 1.4781, "step": 7202 }, { "epoch": 1.5844055867150555, "grad_norm": 0.341130793094635, "learning_rate": 8e-05, "loss": 1.6272, "step": 7203 }, { "epoch": 1.584625536126691, "grad_norm": 0.3377307057380676, "learning_rate": 8e-05, "loss": 1.556, "step": 7204 }, { "epoch": 1.584845485538326, "grad_norm": 0.34748244285583496, "learning_rate": 8e-05, "loss": 1.5995, "step": 7205 }, { "epoch": 1.5850654349499615, "grad_norm": 0.38268405199050903, "learning_rate": 8e-05, "loss": 1.4237, "step": 7206 }, { "epoch": 1.5852853843615968, "grad_norm": 0.3548670709133148, "learning_rate": 8e-05, "loss": 1.4666, "step": 7207 }, { "epoch": 1.585505333773232, "grad_norm": 0.330289363861084, "learning_rate": 8e-05, "loss": 1.3511, "step": 7208 }, { "epoch": 1.5857252831848676, "grad_norm": 0.34244197607040405, "learning_rate": 8e-05, "loss": 1.5658, "step": 7209 }, { "epoch": 1.5859452325965027, "grad_norm": 0.35047149658203125, "learning_rate": 8e-05, "loss": 1.523, "step": 7210 }, { "epoch": 1.5861651820081382, "grad_norm": 0.3660067319869995, "learning_rate": 8e-05, "loss": 1.5978, "step": 7211 }, { "epoch": 1.5863851314197734, "grad_norm": 0.3882088363170624, "learning_rate": 8e-05, "loss": 1.6688, "step": 7212 }, { "epoch": 1.5866050808314087, "grad_norm": 0.34372031688690186, "learning_rate": 8e-05, "loss": 1.5075, "step": 7213 }, { "epoch": 1.5868250302430442, "grad_norm": 0.34752005338668823, "learning_rate": 8e-05, "loss": 1.6924, "step": 7214 }, { "epoch": 1.5870449796546793, "grad_norm": 0.349151611328125, "learning_rate": 8e-05, "loss": 1.5627, "step": 7215 }, { "epoch": 1.5872649290663148, "grad_norm": 0.34376049041748047, "learning_rate": 8e-05, "loss": 1.6252, "step": 7216 }, { "epoch": 1.58748487847795, "grad_norm": 0.3491290211677551, "learning_rate": 8e-05, "loss": 1.6525, "step": 7217 }, { "epoch": 1.5877048278895853, "grad_norm": 0.34608444571495056, "learning_rate": 8e-05, "loss": 1.6798, "step": 7218 }, { "epoch": 1.5879247773012208, "grad_norm": 0.33058491349220276, "learning_rate": 8e-05, "loss": 1.5375, "step": 7219 }, { "epoch": 1.588144726712856, "grad_norm": 0.3423958718776703, "learning_rate": 8e-05, "loss": 1.4829, "step": 7220 }, { "epoch": 1.5883646761244914, "grad_norm": 0.3463810980319977, "learning_rate": 8e-05, "loss": 1.668, "step": 7221 }, { "epoch": 1.5885846255361267, "grad_norm": 0.359178364276886, "learning_rate": 8e-05, "loss": 1.6644, "step": 7222 }, { "epoch": 1.588804574947762, "grad_norm": 0.34609201550483704, "learning_rate": 8e-05, "loss": 1.6044, "step": 7223 }, { "epoch": 1.5890245243593975, "grad_norm": 0.3406553566455841, "learning_rate": 8e-05, "loss": 1.4336, "step": 7224 }, { "epoch": 1.5892444737710325, "grad_norm": 0.3401428759098053, "learning_rate": 8e-05, "loss": 1.4724, "step": 7225 }, { "epoch": 1.589464423182668, "grad_norm": 0.3479742109775543, "learning_rate": 8e-05, "loss": 1.5271, "step": 7226 }, { "epoch": 1.5896843725943033, "grad_norm": 0.3250786364078522, "learning_rate": 8e-05, "loss": 1.5234, "step": 7227 }, { "epoch": 1.5899043220059386, "grad_norm": 0.3240611255168915, "learning_rate": 8e-05, "loss": 1.4373, "step": 7228 }, { "epoch": 1.590124271417574, "grad_norm": 0.36100974678993225, "learning_rate": 8e-05, "loss": 1.639, "step": 7229 }, { "epoch": 1.5903442208292091, "grad_norm": 0.3665464222431183, "learning_rate": 8e-05, "loss": 1.6762, "step": 7230 }, { "epoch": 1.5905641702408446, "grad_norm": 0.39021581411361694, "learning_rate": 8e-05, "loss": 1.6376, "step": 7231 }, { "epoch": 1.59078411965248, "grad_norm": 0.3702135682106018, "learning_rate": 8e-05, "loss": 1.5179, "step": 7232 }, { "epoch": 1.5910040690641152, "grad_norm": 0.3601369559764862, "learning_rate": 8e-05, "loss": 1.5537, "step": 7233 }, { "epoch": 1.5912240184757507, "grad_norm": 0.3713776767253876, "learning_rate": 8e-05, "loss": 1.6425, "step": 7234 }, { "epoch": 1.5914439678873857, "grad_norm": 0.33962202072143555, "learning_rate": 8e-05, "loss": 1.5058, "step": 7235 }, { "epoch": 1.5916639172990212, "grad_norm": 0.4278941750526428, "learning_rate": 8e-05, "loss": 1.4896, "step": 7236 }, { "epoch": 1.5918838667106565, "grad_norm": 0.3914378583431244, "learning_rate": 8e-05, "loss": 1.5183, "step": 7237 }, { "epoch": 1.5921038161222918, "grad_norm": 0.345930814743042, "learning_rate": 8e-05, "loss": 1.4661, "step": 7238 }, { "epoch": 1.5923237655339273, "grad_norm": 0.3633204400539398, "learning_rate": 8e-05, "loss": 1.6982, "step": 7239 }, { "epoch": 1.5925437149455626, "grad_norm": 0.3637048304080963, "learning_rate": 8e-05, "loss": 1.6138, "step": 7240 }, { "epoch": 1.5927636643571979, "grad_norm": 0.375039279460907, "learning_rate": 8e-05, "loss": 1.6831, "step": 7241 }, { "epoch": 1.5929836137688331, "grad_norm": 0.32754218578338623, "learning_rate": 8e-05, "loss": 1.5211, "step": 7242 }, { "epoch": 1.5932035631804684, "grad_norm": 0.3421103358268738, "learning_rate": 8e-05, "loss": 1.5375, "step": 7243 }, { "epoch": 1.593423512592104, "grad_norm": 0.3624960482120514, "learning_rate": 8e-05, "loss": 1.5199, "step": 7244 }, { "epoch": 1.5936434620037392, "grad_norm": 0.3470909893512726, "learning_rate": 8e-05, "loss": 1.4652, "step": 7245 }, { "epoch": 1.5938634114153745, "grad_norm": 0.374307245016098, "learning_rate": 8e-05, "loss": 1.4937, "step": 7246 }, { "epoch": 1.5940833608270097, "grad_norm": 0.35723182559013367, "learning_rate": 8e-05, "loss": 1.7372, "step": 7247 }, { "epoch": 1.594303310238645, "grad_norm": 0.36192354559898376, "learning_rate": 8e-05, "loss": 1.5776, "step": 7248 }, { "epoch": 1.5945232596502805, "grad_norm": 0.36252060532569885, "learning_rate": 8e-05, "loss": 1.7369, "step": 7249 }, { "epoch": 1.5947432090619158, "grad_norm": 0.33922290802001953, "learning_rate": 8e-05, "loss": 1.4568, "step": 7250 }, { "epoch": 1.594963158473551, "grad_norm": 0.3504681885242462, "learning_rate": 8e-05, "loss": 1.5165, "step": 7251 }, { "epoch": 1.5951831078851864, "grad_norm": 0.35567498207092285, "learning_rate": 8e-05, "loss": 1.5728, "step": 7252 }, { "epoch": 1.5954030572968216, "grad_norm": 0.33285123109817505, "learning_rate": 8e-05, "loss": 1.5285, "step": 7253 }, { "epoch": 1.5956230067084571, "grad_norm": 0.340334415435791, "learning_rate": 8e-05, "loss": 1.6383, "step": 7254 }, { "epoch": 1.5958429561200924, "grad_norm": 0.34641945362091064, "learning_rate": 8e-05, "loss": 1.5589, "step": 7255 }, { "epoch": 1.5960629055317277, "grad_norm": 0.35851553082466125, "learning_rate": 8e-05, "loss": 1.5809, "step": 7256 }, { "epoch": 1.596282854943363, "grad_norm": 0.35106992721557617, "learning_rate": 8e-05, "loss": 1.514, "step": 7257 }, { "epoch": 1.5965028043549983, "grad_norm": 0.38542672991752625, "learning_rate": 8e-05, "loss": 1.7067, "step": 7258 }, { "epoch": 1.5967227537666338, "grad_norm": 0.34641674160957336, "learning_rate": 8e-05, "loss": 1.6161, "step": 7259 }, { "epoch": 1.596942703178269, "grad_norm": 0.3341663181781769, "learning_rate": 8e-05, "loss": 1.5811, "step": 7260 }, { "epoch": 1.5971626525899043, "grad_norm": 0.3466171324253082, "learning_rate": 8e-05, "loss": 1.4532, "step": 7261 }, { "epoch": 1.5973826020015398, "grad_norm": 0.32633909583091736, "learning_rate": 8e-05, "loss": 1.5594, "step": 7262 }, { "epoch": 1.5976025514131749, "grad_norm": 0.4083344042301178, "learning_rate": 8e-05, "loss": 1.4278, "step": 7263 }, { "epoch": 1.5978225008248104, "grad_norm": 0.34566327929496765, "learning_rate": 8e-05, "loss": 1.7581, "step": 7264 }, { "epoch": 1.5980424502364456, "grad_norm": 0.33801624178886414, "learning_rate": 8e-05, "loss": 1.5905, "step": 7265 }, { "epoch": 1.598262399648081, "grad_norm": 0.35149914026260376, "learning_rate": 8e-05, "loss": 1.5336, "step": 7266 }, { "epoch": 1.5984823490597164, "grad_norm": 0.3278074562549591, "learning_rate": 8e-05, "loss": 1.4499, "step": 7267 }, { "epoch": 1.5987022984713515, "grad_norm": 0.3583834767341614, "learning_rate": 8e-05, "loss": 1.5363, "step": 7268 }, { "epoch": 1.598922247882987, "grad_norm": 0.3378975987434387, "learning_rate": 8e-05, "loss": 1.5515, "step": 7269 }, { "epoch": 1.5991421972946223, "grad_norm": 0.3475666344165802, "learning_rate": 8e-05, "loss": 1.5319, "step": 7270 }, { "epoch": 1.5993621467062575, "grad_norm": 0.3537213206291199, "learning_rate": 8e-05, "loss": 1.5474, "step": 7271 }, { "epoch": 1.599582096117893, "grad_norm": 0.32548391819000244, "learning_rate": 8e-05, "loss": 1.4559, "step": 7272 }, { "epoch": 1.599802045529528, "grad_norm": 0.3515409529209137, "learning_rate": 8e-05, "loss": 1.6311, "step": 7273 }, { "epoch": 1.6000219949411636, "grad_norm": 0.3460778594017029, "learning_rate": 8e-05, "loss": 1.5365, "step": 7274 }, { "epoch": 1.6002419443527989, "grad_norm": 0.3451859951019287, "learning_rate": 8e-05, "loss": 1.7142, "step": 7275 }, { "epoch": 1.6004618937644342, "grad_norm": 0.3592124581336975, "learning_rate": 8e-05, "loss": 1.6518, "step": 7276 }, { "epoch": 1.6006818431760697, "grad_norm": 0.36190110445022583, "learning_rate": 8e-05, "loss": 1.6901, "step": 7277 }, { "epoch": 1.6009017925877047, "grad_norm": 0.3509790897369385, "learning_rate": 8e-05, "loss": 1.6204, "step": 7278 }, { "epoch": 1.6011217419993402, "grad_norm": 0.3542995750904083, "learning_rate": 8e-05, "loss": 1.681, "step": 7279 }, { "epoch": 1.6013416914109755, "grad_norm": 0.3763824701309204, "learning_rate": 8e-05, "loss": 1.6455, "step": 7280 }, { "epoch": 1.6015616408226108, "grad_norm": 0.3636530637741089, "learning_rate": 8e-05, "loss": 1.7203, "step": 7281 }, { "epoch": 1.6017815902342463, "grad_norm": 0.3652912974357605, "learning_rate": 8e-05, "loss": 1.5865, "step": 7282 }, { "epoch": 1.6020015396458813, "grad_norm": 0.33475157618522644, "learning_rate": 8e-05, "loss": 1.5775, "step": 7283 }, { "epoch": 1.6022214890575168, "grad_norm": 0.33972933888435364, "learning_rate": 8e-05, "loss": 1.5143, "step": 7284 }, { "epoch": 1.602441438469152, "grad_norm": 0.3456435799598694, "learning_rate": 8e-05, "loss": 1.5105, "step": 7285 }, { "epoch": 1.6026613878807874, "grad_norm": 0.397530198097229, "learning_rate": 8e-05, "loss": 1.5182, "step": 7286 }, { "epoch": 1.6028813372924229, "grad_norm": 0.3304464519023895, "learning_rate": 8e-05, "loss": 1.4693, "step": 7287 }, { "epoch": 1.603101286704058, "grad_norm": 0.35964104533195496, "learning_rate": 8e-05, "loss": 1.6491, "step": 7288 }, { "epoch": 1.6033212361156934, "grad_norm": 0.3421217203140259, "learning_rate": 8e-05, "loss": 1.5362, "step": 7289 }, { "epoch": 1.6035411855273287, "grad_norm": 0.3692905902862549, "learning_rate": 8e-05, "loss": 1.6164, "step": 7290 }, { "epoch": 1.603761134938964, "grad_norm": 0.34769442677497864, "learning_rate": 8e-05, "loss": 1.6405, "step": 7291 }, { "epoch": 1.6039810843505995, "grad_norm": 0.3402644991874695, "learning_rate": 8e-05, "loss": 1.6076, "step": 7292 }, { "epoch": 1.6042010337622346, "grad_norm": 0.3504069447517395, "learning_rate": 8e-05, "loss": 1.668, "step": 7293 }, { "epoch": 1.60442098317387, "grad_norm": 0.34672001004219055, "learning_rate": 8e-05, "loss": 1.6585, "step": 7294 }, { "epoch": 1.6046409325855053, "grad_norm": 0.32536306977272034, "learning_rate": 8e-05, "loss": 1.3822, "step": 7295 }, { "epoch": 1.6048608819971406, "grad_norm": 0.3208390176296234, "learning_rate": 8e-05, "loss": 1.5555, "step": 7296 }, { "epoch": 1.605080831408776, "grad_norm": 0.3268856108188629, "learning_rate": 8e-05, "loss": 1.4552, "step": 7297 }, { "epoch": 1.6053007808204112, "grad_norm": 0.36255425214767456, "learning_rate": 8e-05, "loss": 1.6078, "step": 7298 }, { "epoch": 1.6055207302320467, "grad_norm": 0.3356975018978119, "learning_rate": 8e-05, "loss": 1.4995, "step": 7299 }, { "epoch": 1.605740679643682, "grad_norm": 0.35402852296829224, "learning_rate": 8e-05, "loss": 1.7335, "step": 7300 }, { "epoch": 1.6059606290553172, "grad_norm": 0.3467678725719452, "learning_rate": 8e-05, "loss": 1.6858, "step": 7301 }, { "epoch": 1.6061805784669527, "grad_norm": 0.3474421799182892, "learning_rate": 8e-05, "loss": 1.4794, "step": 7302 }, { "epoch": 1.6064005278785878, "grad_norm": 0.33954015374183655, "learning_rate": 8e-05, "loss": 1.5794, "step": 7303 }, { "epoch": 1.6066204772902233, "grad_norm": 0.36678892374038696, "learning_rate": 8e-05, "loss": 1.6351, "step": 7304 }, { "epoch": 1.6068404267018586, "grad_norm": 0.3636288046836853, "learning_rate": 8e-05, "loss": 1.7021, "step": 7305 }, { "epoch": 1.6070603761134938, "grad_norm": 0.3536776006221771, "learning_rate": 8e-05, "loss": 1.6179, "step": 7306 }, { "epoch": 1.6072803255251293, "grad_norm": 0.3369203209877014, "learning_rate": 8e-05, "loss": 1.4754, "step": 7307 }, { "epoch": 1.6075002749367644, "grad_norm": 0.3340332508087158, "learning_rate": 8e-05, "loss": 1.3831, "step": 7308 }, { "epoch": 1.6077202243484, "grad_norm": 0.3479556143283844, "learning_rate": 8e-05, "loss": 1.5855, "step": 7309 }, { "epoch": 1.6079401737600352, "grad_norm": 0.35474708676338196, "learning_rate": 8e-05, "loss": 1.5665, "step": 7310 }, { "epoch": 1.6081601231716705, "grad_norm": 0.34578758478164673, "learning_rate": 8e-05, "loss": 1.5502, "step": 7311 }, { "epoch": 1.608380072583306, "grad_norm": 0.33842918276786804, "learning_rate": 8e-05, "loss": 1.5062, "step": 7312 }, { "epoch": 1.608600021994941, "grad_norm": 0.35063037276268005, "learning_rate": 8e-05, "loss": 1.6101, "step": 7313 }, { "epoch": 1.6088199714065765, "grad_norm": 0.31392213702201843, "learning_rate": 8e-05, "loss": 1.2509, "step": 7314 }, { "epoch": 1.6090399208182118, "grad_norm": 0.3409064710140228, "learning_rate": 8e-05, "loss": 1.6057, "step": 7315 }, { "epoch": 1.609259870229847, "grad_norm": 0.3296225070953369, "learning_rate": 8e-05, "loss": 1.5381, "step": 7316 }, { "epoch": 1.6094798196414826, "grad_norm": 0.33820274472236633, "learning_rate": 8e-05, "loss": 1.51, "step": 7317 }, { "epoch": 1.6096997690531176, "grad_norm": 0.30834639072418213, "learning_rate": 8e-05, "loss": 1.418, "step": 7318 }, { "epoch": 1.6099197184647531, "grad_norm": 0.3295885920524597, "learning_rate": 8e-05, "loss": 1.4204, "step": 7319 }, { "epoch": 1.6101396678763884, "grad_norm": 0.33393704891204834, "learning_rate": 8e-05, "loss": 1.5733, "step": 7320 }, { "epoch": 1.6103596172880237, "grad_norm": 0.3638695180416107, "learning_rate": 8e-05, "loss": 1.5941, "step": 7321 }, { "epoch": 1.6105795666996592, "grad_norm": 0.3963548243045807, "learning_rate": 8e-05, "loss": 1.7292, "step": 7322 }, { "epoch": 1.6107995161112942, "grad_norm": 0.3648855686187744, "learning_rate": 8e-05, "loss": 1.6601, "step": 7323 }, { "epoch": 1.6110194655229297, "grad_norm": 0.34779202938079834, "learning_rate": 8e-05, "loss": 1.6264, "step": 7324 }, { "epoch": 1.611239414934565, "grad_norm": 0.33896639943122864, "learning_rate": 8e-05, "loss": 1.6302, "step": 7325 }, { "epoch": 1.6114593643462003, "grad_norm": 0.3649331033229828, "learning_rate": 8e-05, "loss": 1.7024, "step": 7326 }, { "epoch": 1.6116793137578358, "grad_norm": 0.3513335883617401, "learning_rate": 8e-05, "loss": 1.633, "step": 7327 }, { "epoch": 1.611899263169471, "grad_norm": 0.33497634530067444, "learning_rate": 8e-05, "loss": 1.5016, "step": 7328 }, { "epoch": 1.6121192125811064, "grad_norm": 0.35629206895828247, "learning_rate": 8e-05, "loss": 1.5914, "step": 7329 }, { "epoch": 1.6123391619927416, "grad_norm": 0.33938008546829224, "learning_rate": 8e-05, "loss": 1.5889, "step": 7330 }, { "epoch": 1.612559111404377, "grad_norm": 0.3349902927875519, "learning_rate": 8e-05, "loss": 1.6262, "step": 7331 }, { "epoch": 1.6127790608160124, "grad_norm": 0.36035633087158203, "learning_rate": 8e-05, "loss": 1.4417, "step": 7332 }, { "epoch": 1.6129990102276477, "grad_norm": 0.337649941444397, "learning_rate": 8e-05, "loss": 1.5229, "step": 7333 }, { "epoch": 1.613218959639283, "grad_norm": 0.35085952281951904, "learning_rate": 8e-05, "loss": 1.5416, "step": 7334 }, { "epoch": 1.6134389090509182, "grad_norm": 0.3534447252750397, "learning_rate": 8e-05, "loss": 1.6357, "step": 7335 }, { "epoch": 1.6136588584625535, "grad_norm": 0.33431822061538696, "learning_rate": 8e-05, "loss": 1.6178, "step": 7336 }, { "epoch": 1.613878807874189, "grad_norm": 0.349330872297287, "learning_rate": 8e-05, "loss": 1.6579, "step": 7337 }, { "epoch": 1.6140987572858243, "grad_norm": 0.34993672370910645, "learning_rate": 8e-05, "loss": 1.6064, "step": 7338 }, { "epoch": 1.6143187066974596, "grad_norm": 0.35765430331230164, "learning_rate": 8e-05, "loss": 1.5939, "step": 7339 }, { "epoch": 1.6145386561090949, "grad_norm": 0.35499534010887146, "learning_rate": 8e-05, "loss": 1.4801, "step": 7340 }, { "epoch": 1.6147586055207301, "grad_norm": 0.3562004268169403, "learning_rate": 8e-05, "loss": 1.7272, "step": 7341 }, { "epoch": 1.6149785549323656, "grad_norm": 0.35336798429489136, "learning_rate": 8e-05, "loss": 1.6301, "step": 7342 }, { "epoch": 1.615198504344001, "grad_norm": 0.3726556897163391, "learning_rate": 8e-05, "loss": 1.5391, "step": 7343 }, { "epoch": 1.6154184537556362, "grad_norm": 0.3433026075363159, "learning_rate": 8e-05, "loss": 1.5462, "step": 7344 }, { "epoch": 1.6156384031672715, "grad_norm": 0.3630391061306, "learning_rate": 8e-05, "loss": 1.6461, "step": 7345 }, { "epoch": 1.6158583525789068, "grad_norm": 0.365519255399704, "learning_rate": 8e-05, "loss": 1.6309, "step": 7346 }, { "epoch": 1.6160783019905423, "grad_norm": 0.34080734848976135, "learning_rate": 8e-05, "loss": 1.6324, "step": 7347 }, { "epoch": 1.6162982514021775, "grad_norm": 0.36356642842292786, "learning_rate": 8e-05, "loss": 1.5214, "step": 7348 }, { "epoch": 1.6165182008138128, "grad_norm": 0.33889076113700867, "learning_rate": 8e-05, "loss": 1.6005, "step": 7349 }, { "epoch": 1.6167381502254483, "grad_norm": 0.3468557894229889, "learning_rate": 8e-05, "loss": 1.5337, "step": 7350 }, { "epoch": 1.6169580996370834, "grad_norm": 0.3635764718055725, "learning_rate": 8e-05, "loss": 1.5348, "step": 7351 }, { "epoch": 1.6171780490487189, "grad_norm": 0.33496642112731934, "learning_rate": 8e-05, "loss": 1.6369, "step": 7352 }, { "epoch": 1.6173979984603541, "grad_norm": 0.36198675632476807, "learning_rate": 8e-05, "loss": 1.6511, "step": 7353 }, { "epoch": 1.6176179478719894, "grad_norm": 0.3369227945804596, "learning_rate": 8e-05, "loss": 1.4788, "step": 7354 }, { "epoch": 1.617837897283625, "grad_norm": 0.3565879762172699, "learning_rate": 8e-05, "loss": 1.6826, "step": 7355 }, { "epoch": 1.61805784669526, "grad_norm": 0.36610060930252075, "learning_rate": 8e-05, "loss": 1.7108, "step": 7356 }, { "epoch": 1.6182777961068955, "grad_norm": 0.3607473373413086, "learning_rate": 8e-05, "loss": 1.6657, "step": 7357 }, { "epoch": 1.6184977455185308, "grad_norm": 0.3617018461227417, "learning_rate": 8e-05, "loss": 1.5927, "step": 7358 }, { "epoch": 1.618717694930166, "grad_norm": 0.3475934565067291, "learning_rate": 8e-05, "loss": 1.5489, "step": 7359 }, { "epoch": 1.6189376443418015, "grad_norm": 0.36653825640678406, "learning_rate": 8e-05, "loss": 1.6736, "step": 7360 }, { "epoch": 1.6191575937534366, "grad_norm": 0.3424033522605896, "learning_rate": 8e-05, "loss": 1.6316, "step": 7361 }, { "epoch": 1.619377543165072, "grad_norm": 0.34286367893218994, "learning_rate": 8e-05, "loss": 1.5899, "step": 7362 }, { "epoch": 1.6195974925767074, "grad_norm": 0.32881078124046326, "learning_rate": 8e-05, "loss": 1.5208, "step": 7363 }, { "epoch": 1.6198174419883427, "grad_norm": 0.3607919216156006, "learning_rate": 8e-05, "loss": 1.4825, "step": 7364 }, { "epoch": 1.6200373913999782, "grad_norm": 0.343148410320282, "learning_rate": 8e-05, "loss": 1.6266, "step": 7365 }, { "epoch": 1.6202573408116132, "grad_norm": 0.34463566541671753, "learning_rate": 8e-05, "loss": 1.6966, "step": 7366 }, { "epoch": 1.6204772902232487, "grad_norm": 0.36737263202667236, "learning_rate": 8e-05, "loss": 1.638, "step": 7367 }, { "epoch": 1.620697239634884, "grad_norm": 0.3518025875091553, "learning_rate": 8e-05, "loss": 1.7254, "step": 7368 }, { "epoch": 1.6209171890465193, "grad_norm": 0.3639848828315735, "learning_rate": 8e-05, "loss": 1.6868, "step": 7369 }, { "epoch": 1.6211371384581548, "grad_norm": 0.34554919600486755, "learning_rate": 8e-05, "loss": 1.5702, "step": 7370 }, { "epoch": 1.6213570878697898, "grad_norm": 0.33507826924324036, "learning_rate": 8e-05, "loss": 1.4153, "step": 7371 }, { "epoch": 1.6215770372814253, "grad_norm": 0.354388028383255, "learning_rate": 8e-05, "loss": 1.6659, "step": 7372 }, { "epoch": 1.6217969866930606, "grad_norm": 0.3400000035762787, "learning_rate": 8e-05, "loss": 1.5159, "step": 7373 }, { "epoch": 1.6220169361046959, "grad_norm": 0.34793993830680847, "learning_rate": 8e-05, "loss": 1.5237, "step": 7374 }, { "epoch": 1.6222368855163314, "grad_norm": 0.3557305335998535, "learning_rate": 8e-05, "loss": 1.6078, "step": 7375 }, { "epoch": 1.6224568349279664, "grad_norm": 0.3646417260169983, "learning_rate": 8e-05, "loss": 1.6702, "step": 7376 }, { "epoch": 1.622676784339602, "grad_norm": 0.368301123380661, "learning_rate": 8e-05, "loss": 1.5237, "step": 7377 }, { "epoch": 1.6228967337512372, "grad_norm": 0.34088167548179626, "learning_rate": 8e-05, "loss": 1.52, "step": 7378 }, { "epoch": 1.6231166831628725, "grad_norm": 0.37170594930648804, "learning_rate": 8e-05, "loss": 1.5738, "step": 7379 }, { "epoch": 1.623336632574508, "grad_norm": 0.34634119272232056, "learning_rate": 8e-05, "loss": 1.6908, "step": 7380 }, { "epoch": 1.623556581986143, "grad_norm": 0.338987797498703, "learning_rate": 8e-05, "loss": 1.6874, "step": 7381 }, { "epoch": 1.6237765313977786, "grad_norm": 0.34802037477493286, "learning_rate": 8e-05, "loss": 1.6166, "step": 7382 }, { "epoch": 1.6239964808094138, "grad_norm": 0.36313050985336304, "learning_rate": 8e-05, "loss": 1.7549, "step": 7383 }, { "epoch": 1.6242164302210491, "grad_norm": 0.35427454113960266, "learning_rate": 8e-05, "loss": 1.6044, "step": 7384 }, { "epoch": 1.6244363796326846, "grad_norm": 0.3567980229854584, "learning_rate": 8e-05, "loss": 1.5955, "step": 7385 }, { "epoch": 1.6246563290443197, "grad_norm": 0.3741201162338257, "learning_rate": 8e-05, "loss": 1.4553, "step": 7386 }, { "epoch": 1.6248762784559552, "grad_norm": 0.34070706367492676, "learning_rate": 8e-05, "loss": 1.5722, "step": 7387 }, { "epoch": 1.6250962278675904, "grad_norm": 0.3514935076236725, "learning_rate": 8e-05, "loss": 1.6244, "step": 7388 }, { "epoch": 1.6253161772792257, "grad_norm": 0.35075074434280396, "learning_rate": 8e-05, "loss": 1.5514, "step": 7389 }, { "epoch": 1.6255361266908612, "grad_norm": 0.3427281677722931, "learning_rate": 8e-05, "loss": 1.6158, "step": 7390 }, { "epoch": 1.6257560761024963, "grad_norm": 0.34929710626602173, "learning_rate": 8e-05, "loss": 1.6871, "step": 7391 }, { "epoch": 1.6259760255141318, "grad_norm": 0.3499550521373749, "learning_rate": 8e-05, "loss": 1.7183, "step": 7392 }, { "epoch": 1.626195974925767, "grad_norm": 0.3462086617946625, "learning_rate": 8e-05, "loss": 1.6137, "step": 7393 }, { "epoch": 1.6264159243374023, "grad_norm": 0.35572314262390137, "learning_rate": 8e-05, "loss": 1.5264, "step": 7394 }, { "epoch": 1.6266358737490378, "grad_norm": 0.3303050100803375, "learning_rate": 8e-05, "loss": 1.5774, "step": 7395 }, { "epoch": 1.626855823160673, "grad_norm": 0.34964194893836975, "learning_rate": 8e-05, "loss": 1.4964, "step": 7396 }, { "epoch": 1.6270757725723084, "grad_norm": 0.3477221429347992, "learning_rate": 8e-05, "loss": 1.6341, "step": 7397 }, { "epoch": 1.6272957219839437, "grad_norm": 0.33939552307128906, "learning_rate": 8e-05, "loss": 1.5052, "step": 7398 }, { "epoch": 1.627515671395579, "grad_norm": 0.34513628482818604, "learning_rate": 8e-05, "loss": 1.6778, "step": 7399 }, { "epoch": 1.6277356208072145, "grad_norm": 0.34977173805236816, "learning_rate": 8e-05, "loss": 1.5245, "step": 7400 }, { "epoch": 1.6279555702188495, "grad_norm": 0.36601611971855164, "learning_rate": 8e-05, "loss": 1.7073, "step": 7401 }, { "epoch": 1.628175519630485, "grad_norm": 0.35568591952323914, "learning_rate": 8e-05, "loss": 1.5189, "step": 7402 }, { "epoch": 1.6283954690421203, "grad_norm": 0.33364221453666687, "learning_rate": 8e-05, "loss": 1.6222, "step": 7403 }, { "epoch": 1.6286154184537556, "grad_norm": 0.34395307302474976, "learning_rate": 8e-05, "loss": 1.5248, "step": 7404 }, { "epoch": 1.628835367865391, "grad_norm": 0.3639351427555084, "learning_rate": 8e-05, "loss": 1.6348, "step": 7405 }, { "epoch": 1.6290553172770261, "grad_norm": 0.33809980750083923, "learning_rate": 8e-05, "loss": 1.5322, "step": 7406 }, { "epoch": 1.6292752666886616, "grad_norm": 0.347369521856308, "learning_rate": 8e-05, "loss": 1.5891, "step": 7407 }, { "epoch": 1.629495216100297, "grad_norm": 0.3571222126483917, "learning_rate": 8e-05, "loss": 1.6139, "step": 7408 }, { "epoch": 1.6297151655119322, "grad_norm": 0.3478272557258606, "learning_rate": 8e-05, "loss": 1.5425, "step": 7409 }, { "epoch": 1.6299351149235677, "grad_norm": 0.31865862011909485, "learning_rate": 8e-05, "loss": 1.4634, "step": 7410 }, { "epoch": 1.6301550643352027, "grad_norm": 0.3607088029384613, "learning_rate": 8e-05, "loss": 1.6702, "step": 7411 }, { "epoch": 1.6303750137468382, "grad_norm": 0.34535321593284607, "learning_rate": 8e-05, "loss": 1.7119, "step": 7412 }, { "epoch": 1.6305949631584735, "grad_norm": 0.3476288914680481, "learning_rate": 8e-05, "loss": 1.6827, "step": 7413 }, { "epoch": 1.6308149125701088, "grad_norm": 0.32325246930122375, "learning_rate": 8e-05, "loss": 1.3218, "step": 7414 }, { "epoch": 1.6310348619817443, "grad_norm": 0.3286079168319702, "learning_rate": 8e-05, "loss": 1.494, "step": 7415 }, { "epoch": 1.6312548113933796, "grad_norm": 0.36489957571029663, "learning_rate": 8e-05, "loss": 1.7003, "step": 7416 }, { "epoch": 1.6314747608050149, "grad_norm": 0.3323022723197937, "learning_rate": 8e-05, "loss": 1.4753, "step": 7417 }, { "epoch": 1.6316947102166501, "grad_norm": 0.3368074893951416, "learning_rate": 8e-05, "loss": 1.5202, "step": 7418 }, { "epoch": 1.6319146596282854, "grad_norm": 0.3265618085861206, "learning_rate": 8e-05, "loss": 1.428, "step": 7419 }, { "epoch": 1.632134609039921, "grad_norm": 0.3246419429779053, "learning_rate": 8e-05, "loss": 1.4323, "step": 7420 }, { "epoch": 1.6323545584515562, "grad_norm": 0.3550354242324829, "learning_rate": 8e-05, "loss": 1.5894, "step": 7421 }, { "epoch": 1.6325745078631915, "grad_norm": 0.3658076822757721, "learning_rate": 8e-05, "loss": 1.6968, "step": 7422 }, { "epoch": 1.6327944572748267, "grad_norm": 0.34953486919403076, "learning_rate": 8e-05, "loss": 1.5456, "step": 7423 }, { "epoch": 1.633014406686462, "grad_norm": 0.34616169333457947, "learning_rate": 8e-05, "loss": 1.5601, "step": 7424 }, { "epoch": 1.6332343560980975, "grad_norm": 0.34249401092529297, "learning_rate": 8e-05, "loss": 1.5614, "step": 7425 }, { "epoch": 1.6334543055097328, "grad_norm": 0.3340318202972412, "learning_rate": 8e-05, "loss": 1.6248, "step": 7426 }, { "epoch": 1.633674254921368, "grad_norm": 0.33460670709609985, "learning_rate": 8e-05, "loss": 1.5974, "step": 7427 }, { "epoch": 1.6338942043330034, "grad_norm": 0.324657678604126, "learning_rate": 8e-05, "loss": 1.4784, "step": 7428 }, { "epoch": 1.6341141537446386, "grad_norm": 0.35302308201789856, "learning_rate": 8e-05, "loss": 1.6447, "step": 7429 }, { "epoch": 1.6343341031562741, "grad_norm": 0.33596208691596985, "learning_rate": 8e-05, "loss": 1.5903, "step": 7430 }, { "epoch": 1.6345540525679094, "grad_norm": 0.3537057936191559, "learning_rate": 8e-05, "loss": 1.3988, "step": 7431 }, { "epoch": 1.6347740019795447, "grad_norm": 0.3332783281803131, "learning_rate": 8e-05, "loss": 1.4989, "step": 7432 }, { "epoch": 1.63499395139118, "grad_norm": 0.33840546011924744, "learning_rate": 8e-05, "loss": 1.5321, "step": 7433 }, { "epoch": 1.6352139008028153, "grad_norm": 0.3417434096336365, "learning_rate": 8e-05, "loss": 1.584, "step": 7434 }, { "epoch": 1.6354338502144508, "grad_norm": 0.3549880385398865, "learning_rate": 8e-05, "loss": 1.4418, "step": 7435 }, { "epoch": 1.635653799626086, "grad_norm": 0.3755201995372772, "learning_rate": 8e-05, "loss": 1.7505, "step": 7436 }, { "epoch": 1.6358737490377213, "grad_norm": 0.38725370168685913, "learning_rate": 8e-05, "loss": 1.6678, "step": 7437 }, { "epoch": 1.6360936984493566, "grad_norm": 0.34561362862586975, "learning_rate": 8e-05, "loss": 1.5612, "step": 7438 }, { "epoch": 1.6363136478609919, "grad_norm": 0.3658071458339691, "learning_rate": 8e-05, "loss": 1.702, "step": 7439 }, { "epoch": 1.6365335972726274, "grad_norm": 0.3592982292175293, "learning_rate": 8e-05, "loss": 1.6475, "step": 7440 }, { "epoch": 1.6367535466842627, "grad_norm": 0.3237157166004181, "learning_rate": 8e-05, "loss": 1.5137, "step": 7441 }, { "epoch": 1.636973496095898, "grad_norm": 0.35826730728149414, "learning_rate": 8e-05, "loss": 1.7031, "step": 7442 }, { "epoch": 1.6371934455075334, "grad_norm": 0.3712642788887024, "learning_rate": 8e-05, "loss": 1.7233, "step": 7443 }, { "epoch": 1.6374133949191685, "grad_norm": 0.33850058913230896, "learning_rate": 8e-05, "loss": 1.6113, "step": 7444 }, { "epoch": 1.637633344330804, "grad_norm": 0.36649537086486816, "learning_rate": 8e-05, "loss": 1.5942, "step": 7445 }, { "epoch": 1.6378532937424393, "grad_norm": 0.35413795709609985, "learning_rate": 8e-05, "loss": 1.6198, "step": 7446 }, { "epoch": 1.6380732431540745, "grad_norm": 0.3482341170310974, "learning_rate": 8e-05, "loss": 1.5325, "step": 7447 }, { "epoch": 1.63829319256571, "grad_norm": 0.3386768698692322, "learning_rate": 8e-05, "loss": 1.6404, "step": 7448 }, { "epoch": 1.638513141977345, "grad_norm": 0.3311508893966675, "learning_rate": 8e-05, "loss": 1.5507, "step": 7449 }, { "epoch": 1.6387330913889806, "grad_norm": 0.3398387134075165, "learning_rate": 8e-05, "loss": 1.6054, "step": 7450 }, { "epoch": 1.6389530408006159, "grad_norm": 0.3461743891239166, "learning_rate": 8e-05, "loss": 1.659, "step": 7451 }, { "epoch": 1.6391729902122512, "grad_norm": 0.3451715111732483, "learning_rate": 8e-05, "loss": 1.4445, "step": 7452 }, { "epoch": 1.6393929396238867, "grad_norm": 0.32375138998031616, "learning_rate": 8e-05, "loss": 1.5651, "step": 7453 }, { "epoch": 1.6396128890355217, "grad_norm": 0.35528451204299927, "learning_rate": 8e-05, "loss": 1.6843, "step": 7454 }, { "epoch": 1.6398328384471572, "grad_norm": 0.35188382863998413, "learning_rate": 8e-05, "loss": 1.5426, "step": 7455 }, { "epoch": 1.6400527878587925, "grad_norm": 0.36305519938468933, "learning_rate": 8e-05, "loss": 1.7563, "step": 7456 }, { "epoch": 1.6402727372704278, "grad_norm": 0.3602429926395416, "learning_rate": 8e-05, "loss": 1.6335, "step": 7457 }, { "epoch": 1.6404926866820633, "grad_norm": 0.3496570885181427, "learning_rate": 8e-05, "loss": 1.5569, "step": 7458 }, { "epoch": 1.6407126360936983, "grad_norm": 0.357592910528183, "learning_rate": 8e-05, "loss": 1.6338, "step": 7459 }, { "epoch": 1.6409325855053338, "grad_norm": 0.3440350890159607, "learning_rate": 8e-05, "loss": 1.6484, "step": 7460 }, { "epoch": 1.641152534916969, "grad_norm": 0.3401230275630951, "learning_rate": 8e-05, "loss": 1.5784, "step": 7461 }, { "epoch": 1.6413724843286044, "grad_norm": 0.32542070746421814, "learning_rate": 8e-05, "loss": 1.6349, "step": 7462 }, { "epoch": 1.6415924337402399, "grad_norm": 0.35469067096710205, "learning_rate": 8e-05, "loss": 1.5831, "step": 7463 }, { "epoch": 1.641812383151875, "grad_norm": 0.34359994530677795, "learning_rate": 8e-05, "loss": 1.5779, "step": 7464 }, { "epoch": 1.6420323325635104, "grad_norm": 0.34120601415634155, "learning_rate": 8e-05, "loss": 1.4886, "step": 7465 }, { "epoch": 1.6422522819751457, "grad_norm": 0.3313623368740082, "learning_rate": 8e-05, "loss": 1.5332, "step": 7466 }, { "epoch": 1.642472231386781, "grad_norm": 0.34690678119659424, "learning_rate": 8e-05, "loss": 1.6407, "step": 7467 }, { "epoch": 1.6426921807984165, "grad_norm": 0.3398810625076294, "learning_rate": 8e-05, "loss": 1.6867, "step": 7468 }, { "epoch": 1.6429121302100516, "grad_norm": 0.3494749069213867, "learning_rate": 8e-05, "loss": 1.464, "step": 7469 }, { "epoch": 1.643132079621687, "grad_norm": 0.37120521068573, "learning_rate": 8e-05, "loss": 1.6268, "step": 7470 }, { "epoch": 1.6433520290333223, "grad_norm": 0.3611384928226471, "learning_rate": 8e-05, "loss": 1.6479, "step": 7471 }, { "epoch": 1.6435719784449576, "grad_norm": 0.36218443512916565, "learning_rate": 8e-05, "loss": 1.5302, "step": 7472 }, { "epoch": 1.6437919278565931, "grad_norm": 0.39995941519737244, "learning_rate": 8e-05, "loss": 1.7109, "step": 7473 }, { "epoch": 1.6440118772682282, "grad_norm": 0.3870525658130646, "learning_rate": 8e-05, "loss": 1.7433, "step": 7474 }, { "epoch": 1.6442318266798637, "grad_norm": 0.3534999191761017, "learning_rate": 8e-05, "loss": 1.5314, "step": 7475 }, { "epoch": 1.644451776091499, "grad_norm": 0.3510439097881317, "learning_rate": 8e-05, "loss": 1.6058, "step": 7476 }, { "epoch": 1.6446717255031342, "grad_norm": 0.35384464263916016, "learning_rate": 8e-05, "loss": 1.6558, "step": 7477 }, { "epoch": 1.6448916749147697, "grad_norm": 0.34956666827201843, "learning_rate": 8e-05, "loss": 1.7156, "step": 7478 }, { "epoch": 1.6451116243264048, "grad_norm": 0.3423210382461548, "learning_rate": 8e-05, "loss": 1.6061, "step": 7479 }, { "epoch": 1.6453315737380403, "grad_norm": 0.3628540337085724, "learning_rate": 8e-05, "loss": 1.5584, "step": 7480 }, { "epoch": 1.6455515231496756, "grad_norm": 0.3312741816043854, "learning_rate": 8e-05, "loss": 1.4816, "step": 7481 }, { "epoch": 1.6457714725613108, "grad_norm": 0.3387748599052429, "learning_rate": 8e-05, "loss": 1.5415, "step": 7482 }, { "epoch": 1.6459914219729463, "grad_norm": 0.33808577060699463, "learning_rate": 8e-05, "loss": 1.6166, "step": 7483 }, { "epoch": 1.6462113713845814, "grad_norm": 0.3380921483039856, "learning_rate": 8e-05, "loss": 1.6032, "step": 7484 }, { "epoch": 1.646431320796217, "grad_norm": 0.33434656262397766, "learning_rate": 8e-05, "loss": 1.5902, "step": 7485 }, { "epoch": 1.6466512702078522, "grad_norm": 0.3546171486377716, "learning_rate": 8e-05, "loss": 1.6301, "step": 7486 }, { "epoch": 1.6468712196194875, "grad_norm": 0.31969425082206726, "learning_rate": 8e-05, "loss": 1.344, "step": 7487 }, { "epoch": 1.647091169031123, "grad_norm": 0.39814329147338867, "learning_rate": 8e-05, "loss": 1.6142, "step": 7488 }, { "epoch": 1.647311118442758, "grad_norm": 0.3444506824016571, "learning_rate": 8e-05, "loss": 1.6504, "step": 7489 }, { "epoch": 1.6475310678543935, "grad_norm": 0.33703792095184326, "learning_rate": 8e-05, "loss": 1.4461, "step": 7490 }, { "epoch": 1.6477510172660288, "grad_norm": 0.34956085681915283, "learning_rate": 8e-05, "loss": 1.5006, "step": 7491 }, { "epoch": 1.647970966677664, "grad_norm": 0.34821397066116333, "learning_rate": 8e-05, "loss": 1.5997, "step": 7492 }, { "epoch": 1.6481909160892996, "grad_norm": 0.34046122431755066, "learning_rate": 8e-05, "loss": 1.5231, "step": 7493 }, { "epoch": 1.6484108655009346, "grad_norm": 0.3597094714641571, "learning_rate": 8e-05, "loss": 1.6632, "step": 7494 }, { "epoch": 1.6486308149125701, "grad_norm": 0.33090028166770935, "learning_rate": 8e-05, "loss": 1.5831, "step": 7495 }, { "epoch": 1.6488507643242054, "grad_norm": 0.34698915481567383, "learning_rate": 8e-05, "loss": 1.5397, "step": 7496 }, { "epoch": 1.6490707137358407, "grad_norm": 0.3223899304866791, "learning_rate": 8e-05, "loss": 1.3066, "step": 7497 }, { "epoch": 1.6492906631474762, "grad_norm": 0.34941035509109497, "learning_rate": 8e-05, "loss": 1.5899, "step": 7498 }, { "epoch": 1.6495106125591112, "grad_norm": 0.3517231345176697, "learning_rate": 8e-05, "loss": 1.7412, "step": 7499 }, { "epoch": 1.6497305619707467, "grad_norm": 0.355789452791214, "learning_rate": 8e-05, "loss": 1.6964, "step": 7500 }, { "epoch": 1.649950511382382, "grad_norm": 0.36255064606666565, "learning_rate": 8e-05, "loss": 1.7452, "step": 7501 }, { "epoch": 1.6501704607940173, "grad_norm": 0.3349359333515167, "learning_rate": 8e-05, "loss": 1.5055, "step": 7502 }, { "epoch": 1.6503904102056528, "grad_norm": 0.35435745120048523, "learning_rate": 8e-05, "loss": 1.6859, "step": 7503 }, { "epoch": 1.6506103596172879, "grad_norm": 0.34384238719940186, "learning_rate": 8e-05, "loss": 1.561, "step": 7504 }, { "epoch": 1.6508303090289234, "grad_norm": 0.35307642817497253, "learning_rate": 8e-05, "loss": 1.7436, "step": 7505 }, { "epoch": 1.6510502584405586, "grad_norm": 0.35387226939201355, "learning_rate": 8e-05, "loss": 1.6537, "step": 7506 }, { "epoch": 1.651270207852194, "grad_norm": 0.3663555085659027, "learning_rate": 8e-05, "loss": 1.6033, "step": 7507 }, { "epoch": 1.6514901572638294, "grad_norm": 0.3706420660018921, "learning_rate": 8e-05, "loss": 1.6379, "step": 7508 }, { "epoch": 1.6517101066754647, "grad_norm": 0.3337075710296631, "learning_rate": 8e-05, "loss": 1.5229, "step": 7509 }, { "epoch": 1.6519300560871, "grad_norm": 0.36228495836257935, "learning_rate": 8e-05, "loss": 1.5057, "step": 7510 }, { "epoch": 1.6521500054987353, "grad_norm": 0.35344070196151733, "learning_rate": 8e-05, "loss": 1.5386, "step": 7511 }, { "epoch": 1.6523699549103705, "grad_norm": 0.3585485816001892, "learning_rate": 8e-05, "loss": 1.5688, "step": 7512 }, { "epoch": 1.652589904322006, "grad_norm": 0.35986772179603577, "learning_rate": 8e-05, "loss": 1.5612, "step": 7513 }, { "epoch": 1.6528098537336413, "grad_norm": 0.3567938208580017, "learning_rate": 8e-05, "loss": 1.5845, "step": 7514 }, { "epoch": 1.6530298031452766, "grad_norm": 0.3423861861228943, "learning_rate": 8e-05, "loss": 1.47, "step": 7515 }, { "epoch": 1.6532497525569119, "grad_norm": 0.3586333394050598, "learning_rate": 8e-05, "loss": 1.6797, "step": 7516 }, { "epoch": 1.6534697019685471, "grad_norm": 0.3399210572242737, "learning_rate": 8e-05, "loss": 1.566, "step": 7517 }, { "epoch": 1.6536896513801826, "grad_norm": 0.37481075525283813, "learning_rate": 8e-05, "loss": 1.6758, "step": 7518 }, { "epoch": 1.653909600791818, "grad_norm": 0.37330350279808044, "learning_rate": 8e-05, "loss": 1.6436, "step": 7519 }, { "epoch": 1.6541295502034532, "grad_norm": 0.3201678693294525, "learning_rate": 8e-05, "loss": 1.4216, "step": 7520 }, { "epoch": 1.6543494996150885, "grad_norm": 0.35735848546028137, "learning_rate": 8e-05, "loss": 1.6204, "step": 7521 }, { "epoch": 1.6545694490267238, "grad_norm": 0.36018598079681396, "learning_rate": 8e-05, "loss": 1.5563, "step": 7522 }, { "epoch": 1.6547893984383593, "grad_norm": 0.3720005750656128, "learning_rate": 8e-05, "loss": 1.7081, "step": 7523 }, { "epoch": 1.6550093478499945, "grad_norm": 0.36933058500289917, "learning_rate": 8e-05, "loss": 1.6673, "step": 7524 }, { "epoch": 1.6552292972616298, "grad_norm": 0.3595059812068939, "learning_rate": 8e-05, "loss": 1.6158, "step": 7525 }, { "epoch": 1.655449246673265, "grad_norm": 0.35346463322639465, "learning_rate": 8e-05, "loss": 1.6527, "step": 7526 }, { "epoch": 1.6556691960849004, "grad_norm": 0.34317392110824585, "learning_rate": 8e-05, "loss": 1.4657, "step": 7527 }, { "epoch": 1.6558891454965359, "grad_norm": 0.34805169701576233, "learning_rate": 8e-05, "loss": 1.6248, "step": 7528 }, { "epoch": 1.6561090949081712, "grad_norm": 0.3366924524307251, "learning_rate": 8e-05, "loss": 1.6628, "step": 7529 }, { "epoch": 1.6563290443198064, "grad_norm": 0.33718881011009216, "learning_rate": 8e-05, "loss": 1.5786, "step": 7530 }, { "epoch": 1.656548993731442, "grad_norm": 0.3346345126628876, "learning_rate": 8e-05, "loss": 1.603, "step": 7531 }, { "epoch": 1.656768943143077, "grad_norm": 0.3731262683868408, "learning_rate": 8e-05, "loss": 1.5864, "step": 7532 }, { "epoch": 1.6569888925547125, "grad_norm": 0.33304840326309204, "learning_rate": 8e-05, "loss": 1.4235, "step": 7533 }, { "epoch": 1.6572088419663478, "grad_norm": 0.36336222290992737, "learning_rate": 8e-05, "loss": 1.6247, "step": 7534 }, { "epoch": 1.657428791377983, "grad_norm": 0.33156618475914, "learning_rate": 8e-05, "loss": 1.5243, "step": 7535 }, { "epoch": 1.6576487407896185, "grad_norm": 0.3235434293746948, "learning_rate": 8e-05, "loss": 1.4847, "step": 7536 }, { "epoch": 1.6578686902012536, "grad_norm": 0.3428158760070801, "learning_rate": 8e-05, "loss": 1.5849, "step": 7537 }, { "epoch": 1.658088639612889, "grad_norm": 0.3362279236316681, "learning_rate": 8e-05, "loss": 1.5104, "step": 7538 }, { "epoch": 1.6583085890245244, "grad_norm": 0.3386055529117584, "learning_rate": 8e-05, "loss": 1.537, "step": 7539 }, { "epoch": 1.6585285384361597, "grad_norm": 0.33806055784225464, "learning_rate": 8e-05, "loss": 1.5602, "step": 7540 }, { "epoch": 1.6587484878477952, "grad_norm": 0.35748037695884705, "learning_rate": 8e-05, "loss": 1.6109, "step": 7541 }, { "epoch": 1.6589684372594302, "grad_norm": 0.36311739683151245, "learning_rate": 8e-05, "loss": 1.6461, "step": 7542 }, { "epoch": 1.6591883866710657, "grad_norm": 0.337113618850708, "learning_rate": 8e-05, "loss": 1.606, "step": 7543 }, { "epoch": 1.659408336082701, "grad_norm": 0.34898507595062256, "learning_rate": 8e-05, "loss": 1.5949, "step": 7544 }, { "epoch": 1.6596282854943363, "grad_norm": 0.34912657737731934, "learning_rate": 8e-05, "loss": 1.5531, "step": 7545 }, { "epoch": 1.6598482349059718, "grad_norm": 0.3349672853946686, "learning_rate": 8e-05, "loss": 1.4853, "step": 7546 }, { "epoch": 1.6600681843176068, "grad_norm": 0.3366422653198242, "learning_rate": 8e-05, "loss": 1.5632, "step": 7547 }, { "epoch": 1.6602881337292423, "grad_norm": 0.35964590311050415, "learning_rate": 8e-05, "loss": 1.6277, "step": 7548 }, { "epoch": 1.6605080831408776, "grad_norm": 0.3346412181854248, "learning_rate": 8e-05, "loss": 1.4984, "step": 7549 }, { "epoch": 1.660728032552513, "grad_norm": 0.35949745774269104, "learning_rate": 8e-05, "loss": 1.659, "step": 7550 }, { "epoch": 1.6609479819641484, "grad_norm": 0.36303624510765076, "learning_rate": 8e-05, "loss": 1.6078, "step": 7551 }, { "epoch": 1.6611679313757834, "grad_norm": 0.34074535965919495, "learning_rate": 8e-05, "loss": 1.5481, "step": 7552 }, { "epoch": 1.661387880787419, "grad_norm": 0.35166946053504944, "learning_rate": 8e-05, "loss": 1.539, "step": 7553 }, { "epoch": 1.6616078301990542, "grad_norm": 0.35178396105766296, "learning_rate": 8e-05, "loss": 1.5898, "step": 7554 }, { "epoch": 1.6618277796106895, "grad_norm": 0.35289016366004944, "learning_rate": 8e-05, "loss": 1.4929, "step": 7555 }, { "epoch": 1.662047729022325, "grad_norm": 0.337489515542984, "learning_rate": 8e-05, "loss": 1.564, "step": 7556 }, { "epoch": 1.66226767843396, "grad_norm": 0.3517208993434906, "learning_rate": 8e-05, "loss": 1.543, "step": 7557 }, { "epoch": 1.6624876278455956, "grad_norm": 0.33007141947746277, "learning_rate": 8e-05, "loss": 1.6521, "step": 7558 }, { "epoch": 1.6627075772572308, "grad_norm": 0.3575291335582733, "learning_rate": 8e-05, "loss": 1.5306, "step": 7559 }, { "epoch": 1.6629275266688661, "grad_norm": 0.34711265563964844, "learning_rate": 8e-05, "loss": 1.652, "step": 7560 }, { "epoch": 1.6631474760805016, "grad_norm": 0.3387552797794342, "learning_rate": 8e-05, "loss": 1.6124, "step": 7561 }, { "epoch": 1.6633674254921367, "grad_norm": 0.35134533047676086, "learning_rate": 8e-05, "loss": 1.664, "step": 7562 }, { "epoch": 1.6635873749037722, "grad_norm": 0.35335835814476013, "learning_rate": 8e-05, "loss": 1.6584, "step": 7563 }, { "epoch": 1.6638073243154075, "grad_norm": 0.34525778889656067, "learning_rate": 8e-05, "loss": 1.5925, "step": 7564 }, { "epoch": 1.6640272737270427, "grad_norm": 0.37951335310935974, "learning_rate": 8e-05, "loss": 1.6176, "step": 7565 }, { "epoch": 1.6642472231386782, "grad_norm": 0.36407074332237244, "learning_rate": 8e-05, "loss": 1.528, "step": 7566 }, { "epoch": 1.6644671725503133, "grad_norm": 0.35214969515800476, "learning_rate": 8e-05, "loss": 1.6392, "step": 7567 }, { "epoch": 1.6646871219619488, "grad_norm": 0.3570637106895447, "learning_rate": 8e-05, "loss": 1.6957, "step": 7568 }, { "epoch": 1.664907071373584, "grad_norm": 0.34562286734580994, "learning_rate": 8e-05, "loss": 1.6099, "step": 7569 }, { "epoch": 1.6651270207852193, "grad_norm": 0.350987046957016, "learning_rate": 8e-05, "loss": 1.6146, "step": 7570 }, { "epoch": 1.6653469701968548, "grad_norm": 0.33313173055648804, "learning_rate": 8e-05, "loss": 1.502, "step": 7571 }, { "epoch": 1.66556691960849, "grad_norm": 0.3585159480571747, "learning_rate": 8e-05, "loss": 1.4284, "step": 7572 }, { "epoch": 1.6657868690201254, "grad_norm": 0.386651873588562, "learning_rate": 8e-05, "loss": 1.6477, "step": 7573 }, { "epoch": 1.6660068184317607, "grad_norm": 0.33219578862190247, "learning_rate": 8e-05, "loss": 1.5252, "step": 7574 }, { "epoch": 1.666226767843396, "grad_norm": 0.32161766290664673, "learning_rate": 8e-05, "loss": 1.4454, "step": 7575 }, { "epoch": 1.6664467172550315, "grad_norm": 0.3519505560398102, "learning_rate": 8e-05, "loss": 1.6539, "step": 7576 }, { "epoch": 1.6666666666666665, "grad_norm": 0.3499259650707245, "learning_rate": 8e-05, "loss": 1.5299, "step": 7577 }, { "epoch": 1.666886616078302, "grad_norm": 0.33787304162979126, "learning_rate": 8e-05, "loss": 1.4703, "step": 7578 }, { "epoch": 1.6671065654899373, "grad_norm": 0.3327706456184387, "learning_rate": 8e-05, "loss": 1.5973, "step": 7579 }, { "epoch": 1.6673265149015726, "grad_norm": 0.33336660265922546, "learning_rate": 8e-05, "loss": 1.5258, "step": 7580 }, { "epoch": 1.667546464313208, "grad_norm": 0.3463980555534363, "learning_rate": 8e-05, "loss": 1.6719, "step": 7581 }, { "epoch": 1.6677664137248431, "grad_norm": 0.35184457898139954, "learning_rate": 8e-05, "loss": 1.5724, "step": 7582 }, { "epoch": 1.6679863631364786, "grad_norm": 0.3702712655067444, "learning_rate": 8e-05, "loss": 1.648, "step": 7583 }, { "epoch": 1.668206312548114, "grad_norm": 0.3660614490509033, "learning_rate": 8e-05, "loss": 1.6738, "step": 7584 }, { "epoch": 1.6684262619597492, "grad_norm": 0.3461119830608368, "learning_rate": 8e-05, "loss": 1.6009, "step": 7585 }, { "epoch": 1.6686462113713847, "grad_norm": 0.3452329933643341, "learning_rate": 8e-05, "loss": 1.5294, "step": 7586 }, { "epoch": 1.6688661607830197, "grad_norm": 0.3298177719116211, "learning_rate": 8e-05, "loss": 1.3912, "step": 7587 }, { "epoch": 1.6690861101946552, "grad_norm": 0.3376080095767975, "learning_rate": 8e-05, "loss": 1.5186, "step": 7588 }, { "epoch": 1.6693060596062905, "grad_norm": 0.35834628343582153, "learning_rate": 8e-05, "loss": 1.7235, "step": 7589 }, { "epoch": 1.6695260090179258, "grad_norm": 0.34377264976501465, "learning_rate": 8e-05, "loss": 1.5705, "step": 7590 }, { "epoch": 1.6697459584295613, "grad_norm": 0.35771444439888, "learning_rate": 8e-05, "loss": 1.5157, "step": 7591 }, { "epoch": 1.6699659078411964, "grad_norm": 0.333508163690567, "learning_rate": 8e-05, "loss": 1.5499, "step": 7592 }, { "epoch": 1.6701858572528319, "grad_norm": 0.331186443567276, "learning_rate": 8e-05, "loss": 1.6034, "step": 7593 }, { "epoch": 1.6704058066644671, "grad_norm": 0.3267330229282379, "learning_rate": 8e-05, "loss": 1.4153, "step": 7594 }, { "epoch": 1.6706257560761024, "grad_norm": 0.3420249819755554, "learning_rate": 8e-05, "loss": 1.5966, "step": 7595 }, { "epoch": 1.670845705487738, "grad_norm": 0.329901784658432, "learning_rate": 8e-05, "loss": 1.4626, "step": 7596 }, { "epoch": 1.6710656548993732, "grad_norm": 0.3596947491168976, "learning_rate": 8e-05, "loss": 1.6466, "step": 7597 }, { "epoch": 1.6712856043110085, "grad_norm": 0.33629316091537476, "learning_rate": 8e-05, "loss": 1.5827, "step": 7598 }, { "epoch": 1.6715055537226438, "grad_norm": 0.3332720696926117, "learning_rate": 8e-05, "loss": 1.6263, "step": 7599 }, { "epoch": 1.671725503134279, "grad_norm": 0.4437011778354645, "learning_rate": 8e-05, "loss": 1.6046, "step": 7600 }, { "epoch": 1.6719454525459145, "grad_norm": 0.32602447271347046, "learning_rate": 8e-05, "loss": 1.4119, "step": 7601 }, { "epoch": 1.6721654019575498, "grad_norm": 0.3389072120189667, "learning_rate": 8e-05, "loss": 1.5725, "step": 7602 }, { "epoch": 1.672385351369185, "grad_norm": 0.34616774320602417, "learning_rate": 8e-05, "loss": 1.5908, "step": 7603 }, { "epoch": 1.6726053007808204, "grad_norm": 0.33606410026550293, "learning_rate": 8e-05, "loss": 1.5795, "step": 7604 }, { "epoch": 1.6728252501924556, "grad_norm": 0.3320625126361847, "learning_rate": 8e-05, "loss": 1.5364, "step": 7605 }, { "epoch": 1.6730451996040911, "grad_norm": 0.3394157588481903, "learning_rate": 8e-05, "loss": 1.5889, "step": 7606 }, { "epoch": 1.6732651490157264, "grad_norm": 0.3410903513431549, "learning_rate": 8e-05, "loss": 1.4222, "step": 7607 }, { "epoch": 1.6734850984273617, "grad_norm": 0.34357231855392456, "learning_rate": 8e-05, "loss": 1.6176, "step": 7608 }, { "epoch": 1.673705047838997, "grad_norm": 0.3466140925884247, "learning_rate": 8e-05, "loss": 1.6191, "step": 7609 }, { "epoch": 1.6739249972506323, "grad_norm": 0.3401027023792267, "learning_rate": 8e-05, "loss": 1.4659, "step": 7610 }, { "epoch": 1.6741449466622678, "grad_norm": 0.33620861172676086, "learning_rate": 8e-05, "loss": 1.4374, "step": 7611 }, { "epoch": 1.674364896073903, "grad_norm": 0.35475507378578186, "learning_rate": 8e-05, "loss": 1.5514, "step": 7612 }, { "epoch": 1.6745848454855383, "grad_norm": 0.34837767481803894, "learning_rate": 8e-05, "loss": 1.6389, "step": 7613 }, { "epoch": 1.6748047948971736, "grad_norm": 0.3331213891506195, "learning_rate": 8e-05, "loss": 1.368, "step": 7614 }, { "epoch": 1.6750247443088089, "grad_norm": 0.3450992703437805, "learning_rate": 8e-05, "loss": 1.4777, "step": 7615 }, { "epoch": 1.6752446937204444, "grad_norm": 0.3874543607234955, "learning_rate": 8e-05, "loss": 1.6673, "step": 7616 }, { "epoch": 1.6754646431320797, "grad_norm": 0.35173657536506653, "learning_rate": 8e-05, "loss": 1.565, "step": 7617 }, { "epoch": 1.675684592543715, "grad_norm": 0.34773820638656616, "learning_rate": 8e-05, "loss": 1.6911, "step": 7618 }, { "epoch": 1.6759045419553504, "grad_norm": 0.3544262647628784, "learning_rate": 8e-05, "loss": 1.7422, "step": 7619 }, { "epoch": 1.6761244913669855, "grad_norm": 0.3530780076980591, "learning_rate": 8e-05, "loss": 1.5492, "step": 7620 }, { "epoch": 1.676344440778621, "grad_norm": 0.34259817004203796, "learning_rate": 8e-05, "loss": 1.6034, "step": 7621 }, { "epoch": 1.6765643901902563, "grad_norm": 0.34721362590789795, "learning_rate": 8e-05, "loss": 1.5797, "step": 7622 }, { "epoch": 1.6767843396018915, "grad_norm": 0.3470019996166229, "learning_rate": 8e-05, "loss": 1.5553, "step": 7623 }, { "epoch": 1.677004289013527, "grad_norm": 0.3512800335884094, "learning_rate": 8e-05, "loss": 1.6084, "step": 7624 }, { "epoch": 1.677224238425162, "grad_norm": 0.35458195209503174, "learning_rate": 8e-05, "loss": 1.7599, "step": 7625 }, { "epoch": 1.6774441878367976, "grad_norm": 0.3621503412723541, "learning_rate": 8e-05, "loss": 1.7146, "step": 7626 }, { "epoch": 1.6776641372484329, "grad_norm": 0.361889123916626, "learning_rate": 8e-05, "loss": 1.6027, "step": 7627 }, { "epoch": 1.6778840866600682, "grad_norm": 0.33397144079208374, "learning_rate": 8e-05, "loss": 1.5029, "step": 7628 }, { "epoch": 1.6781040360717037, "grad_norm": 0.36191293597221375, "learning_rate": 8e-05, "loss": 1.5267, "step": 7629 }, { "epoch": 1.6783239854833387, "grad_norm": 0.3690151572227478, "learning_rate": 8e-05, "loss": 1.6921, "step": 7630 }, { "epoch": 1.6785439348949742, "grad_norm": 0.33848339319229126, "learning_rate": 8e-05, "loss": 1.3949, "step": 7631 }, { "epoch": 1.6787638843066095, "grad_norm": 0.3602207899093628, "learning_rate": 8e-05, "loss": 1.55, "step": 7632 }, { "epoch": 1.6789838337182448, "grad_norm": 0.35579246282577515, "learning_rate": 8e-05, "loss": 1.5237, "step": 7633 }, { "epoch": 1.6792037831298803, "grad_norm": 0.35253387689590454, "learning_rate": 8e-05, "loss": 1.5322, "step": 7634 }, { "epoch": 1.6794237325415153, "grad_norm": 0.3585261404514313, "learning_rate": 8e-05, "loss": 1.5788, "step": 7635 }, { "epoch": 1.6796436819531508, "grad_norm": 0.36290520429611206, "learning_rate": 8e-05, "loss": 1.6087, "step": 7636 }, { "epoch": 1.6798636313647861, "grad_norm": 0.36509591341018677, "learning_rate": 8e-05, "loss": 1.5957, "step": 7637 }, { "epoch": 1.6800835807764214, "grad_norm": 0.3329887390136719, "learning_rate": 8e-05, "loss": 1.3779, "step": 7638 }, { "epoch": 1.680303530188057, "grad_norm": 0.3644077479839325, "learning_rate": 8e-05, "loss": 1.502, "step": 7639 }, { "epoch": 1.680523479599692, "grad_norm": 0.34599965810775757, "learning_rate": 8e-05, "loss": 1.6234, "step": 7640 }, { "epoch": 1.6807434290113274, "grad_norm": 0.35095739364624023, "learning_rate": 8e-05, "loss": 1.6428, "step": 7641 }, { "epoch": 1.6809633784229627, "grad_norm": 0.3475542366504669, "learning_rate": 8e-05, "loss": 1.606, "step": 7642 }, { "epoch": 1.681183327834598, "grad_norm": 0.3340604305267334, "learning_rate": 8e-05, "loss": 1.5915, "step": 7643 }, { "epoch": 1.6814032772462335, "grad_norm": 0.34424611926078796, "learning_rate": 8e-05, "loss": 1.7233, "step": 7644 }, { "epoch": 1.6816232266578686, "grad_norm": 0.3398914337158203, "learning_rate": 8e-05, "loss": 1.5845, "step": 7645 }, { "epoch": 1.681843176069504, "grad_norm": 0.35905542969703674, "learning_rate": 8e-05, "loss": 1.5799, "step": 7646 }, { "epoch": 1.6820631254811393, "grad_norm": 0.34302419424057007, "learning_rate": 8e-05, "loss": 1.6647, "step": 7647 }, { "epoch": 1.6822830748927746, "grad_norm": 0.3458414077758789, "learning_rate": 8e-05, "loss": 1.6211, "step": 7648 }, { "epoch": 1.6825030243044101, "grad_norm": 0.3447232246398926, "learning_rate": 8e-05, "loss": 1.5239, "step": 7649 }, { "epoch": 1.6827229737160452, "grad_norm": 0.3388119339942932, "learning_rate": 8e-05, "loss": 1.5203, "step": 7650 }, { "epoch": 1.6829429231276807, "grad_norm": 0.34632331132888794, "learning_rate": 8e-05, "loss": 1.634, "step": 7651 }, { "epoch": 1.683162872539316, "grad_norm": 0.3421694040298462, "learning_rate": 8e-05, "loss": 1.6827, "step": 7652 }, { "epoch": 1.6833828219509512, "grad_norm": 0.33624517917633057, "learning_rate": 8e-05, "loss": 1.5477, "step": 7653 }, { "epoch": 1.6836027713625867, "grad_norm": 0.3346371352672577, "learning_rate": 8e-05, "loss": 1.4832, "step": 7654 }, { "epoch": 1.6838227207742218, "grad_norm": 0.3644462823867798, "learning_rate": 8e-05, "loss": 1.4707, "step": 7655 }, { "epoch": 1.6840426701858573, "grad_norm": 0.3637033998966217, "learning_rate": 8e-05, "loss": 1.64, "step": 7656 }, { "epoch": 1.6842626195974926, "grad_norm": 0.3667287230491638, "learning_rate": 8e-05, "loss": 1.6293, "step": 7657 }, { "epoch": 1.6844825690091279, "grad_norm": 0.37338683009147644, "learning_rate": 8e-05, "loss": 1.6156, "step": 7658 }, { "epoch": 1.6847025184207634, "grad_norm": 0.3363918364048004, "learning_rate": 8e-05, "loss": 1.3309, "step": 7659 }, { "epoch": 1.6849224678323984, "grad_norm": 0.3639584183692932, "learning_rate": 8e-05, "loss": 1.4548, "step": 7660 }, { "epoch": 1.685142417244034, "grad_norm": 0.3434525728225708, "learning_rate": 8e-05, "loss": 1.6476, "step": 7661 }, { "epoch": 1.6853623666556692, "grad_norm": 0.3390900492668152, "learning_rate": 8e-05, "loss": 1.517, "step": 7662 }, { "epoch": 1.6855823160673045, "grad_norm": 0.3601070046424866, "learning_rate": 8e-05, "loss": 1.723, "step": 7663 }, { "epoch": 1.68580226547894, "grad_norm": 0.3458713889122009, "learning_rate": 8e-05, "loss": 1.4211, "step": 7664 }, { "epoch": 1.686022214890575, "grad_norm": 0.3444063365459442, "learning_rate": 8e-05, "loss": 1.5632, "step": 7665 }, { "epoch": 1.6862421643022105, "grad_norm": 0.34388598799705505, "learning_rate": 8e-05, "loss": 1.5215, "step": 7666 }, { "epoch": 1.6864621137138458, "grad_norm": 0.3510953187942505, "learning_rate": 8e-05, "loss": 1.5834, "step": 7667 }, { "epoch": 1.686682063125481, "grad_norm": 0.34482333064079285, "learning_rate": 8e-05, "loss": 1.5328, "step": 7668 }, { "epoch": 1.6869020125371166, "grad_norm": 0.3465597629547119, "learning_rate": 8e-05, "loss": 1.5793, "step": 7669 }, { "epoch": 1.6871219619487516, "grad_norm": 0.36673063039779663, "learning_rate": 8e-05, "loss": 1.7018, "step": 7670 }, { "epoch": 1.6873419113603871, "grad_norm": 0.3431655764579773, "learning_rate": 8e-05, "loss": 1.6419, "step": 7671 }, { "epoch": 1.6875618607720224, "grad_norm": 0.36044740676879883, "learning_rate": 8e-05, "loss": 1.4832, "step": 7672 }, { "epoch": 1.6877818101836577, "grad_norm": 0.3603752851486206, "learning_rate": 8e-05, "loss": 1.7607, "step": 7673 }, { "epoch": 1.6880017595952932, "grad_norm": 0.36751630902290344, "learning_rate": 8e-05, "loss": 1.71, "step": 7674 }, { "epoch": 1.6882217090069283, "grad_norm": 0.36104369163513184, "learning_rate": 8e-05, "loss": 1.599, "step": 7675 }, { "epoch": 1.6884416584185638, "grad_norm": 0.33976805210113525, "learning_rate": 8e-05, "loss": 1.4751, "step": 7676 }, { "epoch": 1.688661607830199, "grad_norm": 0.33883416652679443, "learning_rate": 8e-05, "loss": 1.5676, "step": 7677 }, { "epoch": 1.6888815572418343, "grad_norm": 0.3576543927192688, "learning_rate": 8e-05, "loss": 1.612, "step": 7678 }, { "epoch": 1.6891015066534698, "grad_norm": 0.35499855875968933, "learning_rate": 8e-05, "loss": 1.6922, "step": 7679 }, { "epoch": 1.6893214560651049, "grad_norm": 0.3758505582809448, "learning_rate": 8e-05, "loss": 1.5734, "step": 7680 }, { "epoch": 1.6895414054767404, "grad_norm": 0.3587852120399475, "learning_rate": 8e-05, "loss": 1.7364, "step": 7681 }, { "epoch": 1.6897613548883756, "grad_norm": 0.33592450618743896, "learning_rate": 8e-05, "loss": 1.5806, "step": 7682 }, { "epoch": 1.689981304300011, "grad_norm": 0.44680455327033997, "learning_rate": 8e-05, "loss": 1.6459, "step": 7683 }, { "epoch": 1.6902012537116464, "grad_norm": 0.32252466678619385, "learning_rate": 8e-05, "loss": 1.4969, "step": 7684 }, { "epoch": 1.6904212031232815, "grad_norm": 0.32827988266944885, "learning_rate": 8e-05, "loss": 1.4667, "step": 7685 }, { "epoch": 1.690641152534917, "grad_norm": 0.3466763496398926, "learning_rate": 8e-05, "loss": 1.5456, "step": 7686 }, { "epoch": 1.6908611019465523, "grad_norm": 0.37088483572006226, "learning_rate": 8e-05, "loss": 1.6245, "step": 7687 }, { "epoch": 1.6910810513581875, "grad_norm": 0.3619740903377533, "learning_rate": 8e-05, "loss": 1.4171, "step": 7688 }, { "epoch": 1.691301000769823, "grad_norm": 0.34023022651672363, "learning_rate": 8e-05, "loss": 1.5672, "step": 7689 }, { "epoch": 1.6915209501814583, "grad_norm": 0.34109410643577576, "learning_rate": 8e-05, "loss": 1.5617, "step": 7690 }, { "epoch": 1.6917408995930936, "grad_norm": 0.34547850489616394, "learning_rate": 8e-05, "loss": 1.5971, "step": 7691 }, { "epoch": 1.6919608490047289, "grad_norm": 0.3409948945045471, "learning_rate": 8e-05, "loss": 1.4874, "step": 7692 }, { "epoch": 1.6921807984163642, "grad_norm": 0.3351404070854187, "learning_rate": 8e-05, "loss": 1.5076, "step": 7693 }, { "epoch": 1.6924007478279997, "grad_norm": 0.36004939675331116, "learning_rate": 8e-05, "loss": 1.5987, "step": 7694 }, { "epoch": 1.692620697239635, "grad_norm": 0.3515600562095642, "learning_rate": 8e-05, "loss": 1.5879, "step": 7695 }, { "epoch": 1.6928406466512702, "grad_norm": 0.3365083932876587, "learning_rate": 8e-05, "loss": 1.64, "step": 7696 }, { "epoch": 1.6930605960629055, "grad_norm": 0.3529248535633087, "learning_rate": 8e-05, "loss": 1.5608, "step": 7697 }, { "epoch": 1.6932805454745408, "grad_norm": 0.3460243344306946, "learning_rate": 8e-05, "loss": 1.5093, "step": 7698 }, { "epoch": 1.6935004948861763, "grad_norm": 0.4227786064147949, "learning_rate": 8e-05, "loss": 1.7166, "step": 7699 }, { "epoch": 1.6937204442978115, "grad_norm": 0.3608701527118683, "learning_rate": 8e-05, "loss": 1.6863, "step": 7700 }, { "epoch": 1.6939403937094468, "grad_norm": 0.33734041452407837, "learning_rate": 8e-05, "loss": 1.5284, "step": 7701 }, { "epoch": 1.694160343121082, "grad_norm": 0.36296725273132324, "learning_rate": 8e-05, "loss": 1.6612, "step": 7702 }, { "epoch": 1.6943802925327174, "grad_norm": 0.3545825183391571, "learning_rate": 8e-05, "loss": 1.6193, "step": 7703 }, { "epoch": 1.6946002419443529, "grad_norm": 0.3215986490249634, "learning_rate": 8e-05, "loss": 1.469, "step": 7704 }, { "epoch": 1.6948201913559882, "grad_norm": 0.3710196912288666, "learning_rate": 8e-05, "loss": 1.8176, "step": 7705 }, { "epoch": 1.6950401407676234, "grad_norm": 0.3482315242290497, "learning_rate": 8e-05, "loss": 1.6571, "step": 7706 }, { "epoch": 1.6952600901792587, "grad_norm": 0.3878231346607208, "learning_rate": 8e-05, "loss": 1.6715, "step": 7707 }, { "epoch": 1.695480039590894, "grad_norm": 0.3578043282032013, "learning_rate": 8e-05, "loss": 1.5938, "step": 7708 }, { "epoch": 1.6956999890025295, "grad_norm": 0.3378647267818451, "learning_rate": 8e-05, "loss": 1.4129, "step": 7709 }, { "epoch": 1.6959199384141648, "grad_norm": 0.35368889570236206, "learning_rate": 8e-05, "loss": 1.7443, "step": 7710 }, { "epoch": 1.6961398878258, "grad_norm": 0.3366484045982361, "learning_rate": 8e-05, "loss": 1.5872, "step": 7711 }, { "epoch": 1.6963598372374356, "grad_norm": 0.35220828652381897, "learning_rate": 8e-05, "loss": 1.5769, "step": 7712 }, { "epoch": 1.6965797866490706, "grad_norm": 0.3617173135280609, "learning_rate": 8e-05, "loss": 1.6589, "step": 7713 }, { "epoch": 1.696799736060706, "grad_norm": 0.3647673428058624, "learning_rate": 8e-05, "loss": 1.6482, "step": 7714 }, { "epoch": 1.6970196854723414, "grad_norm": 0.3416147530078888, "learning_rate": 8e-05, "loss": 1.515, "step": 7715 }, { "epoch": 1.6972396348839767, "grad_norm": 0.3531167209148407, "learning_rate": 8e-05, "loss": 1.5494, "step": 7716 }, { "epoch": 1.6974595842956122, "grad_norm": 0.3509884476661682, "learning_rate": 8e-05, "loss": 1.6307, "step": 7717 }, { "epoch": 1.6976795337072472, "grad_norm": 0.35767534375190735, "learning_rate": 8e-05, "loss": 1.5927, "step": 7718 }, { "epoch": 1.6978994831188827, "grad_norm": 0.3738901615142822, "learning_rate": 8e-05, "loss": 1.7652, "step": 7719 }, { "epoch": 1.698119432530518, "grad_norm": 0.3489720821380615, "learning_rate": 8e-05, "loss": 1.6218, "step": 7720 }, { "epoch": 1.6983393819421533, "grad_norm": 0.32581770420074463, "learning_rate": 8e-05, "loss": 1.5449, "step": 7721 }, { "epoch": 1.6985593313537888, "grad_norm": 0.35932430624961853, "learning_rate": 8e-05, "loss": 1.4721, "step": 7722 }, { "epoch": 1.6987792807654238, "grad_norm": 0.3305474519729614, "learning_rate": 8e-05, "loss": 1.5179, "step": 7723 }, { "epoch": 1.6989992301770593, "grad_norm": 0.35860341787338257, "learning_rate": 8e-05, "loss": 1.6497, "step": 7724 }, { "epoch": 1.6992191795886946, "grad_norm": 0.3502061367034912, "learning_rate": 8e-05, "loss": 1.6381, "step": 7725 }, { "epoch": 1.69943912900033, "grad_norm": 0.340102881193161, "learning_rate": 8e-05, "loss": 1.5772, "step": 7726 }, { "epoch": 1.6996590784119654, "grad_norm": 0.34431204199790955, "learning_rate": 8e-05, "loss": 1.5004, "step": 7727 }, { "epoch": 1.6998790278236005, "grad_norm": 0.3771098554134369, "learning_rate": 8e-05, "loss": 1.5702, "step": 7728 }, { "epoch": 1.700098977235236, "grad_norm": 0.35402974486351013, "learning_rate": 8e-05, "loss": 1.629, "step": 7729 }, { "epoch": 1.7003189266468712, "grad_norm": 0.328931599855423, "learning_rate": 8e-05, "loss": 1.5164, "step": 7730 }, { "epoch": 1.7005388760585065, "grad_norm": 0.3532305955886841, "learning_rate": 8e-05, "loss": 1.674, "step": 7731 }, { "epoch": 1.700758825470142, "grad_norm": 0.3946567177772522, "learning_rate": 8e-05, "loss": 1.6493, "step": 7732 }, { "epoch": 1.700978774881777, "grad_norm": 0.3357256054878235, "learning_rate": 8e-05, "loss": 1.5049, "step": 7733 }, { "epoch": 1.7011987242934126, "grad_norm": 0.34998443722724915, "learning_rate": 8e-05, "loss": 1.6938, "step": 7734 }, { "epoch": 1.7014186737050478, "grad_norm": 0.3363568186759949, "learning_rate": 8e-05, "loss": 1.551, "step": 7735 }, { "epoch": 1.7016386231166831, "grad_norm": 0.3563581705093384, "learning_rate": 8e-05, "loss": 1.6522, "step": 7736 }, { "epoch": 1.7018585725283186, "grad_norm": 0.32991161942481995, "learning_rate": 8e-05, "loss": 1.5063, "step": 7737 }, { "epoch": 1.7020785219399537, "grad_norm": 0.37370792031288147, "learning_rate": 8e-05, "loss": 1.7343, "step": 7738 }, { "epoch": 1.7022984713515892, "grad_norm": 0.34982678294181824, "learning_rate": 8e-05, "loss": 1.6371, "step": 7739 }, { "epoch": 1.7025184207632245, "grad_norm": 0.33817631006240845, "learning_rate": 8e-05, "loss": 1.5401, "step": 7740 }, { "epoch": 1.7027383701748597, "grad_norm": 0.3767993450164795, "learning_rate": 8e-05, "loss": 1.5531, "step": 7741 }, { "epoch": 1.7029583195864952, "grad_norm": 0.34466353058815, "learning_rate": 8e-05, "loss": 1.5158, "step": 7742 }, { "epoch": 1.7031782689981303, "grad_norm": 0.34066420793533325, "learning_rate": 8e-05, "loss": 1.559, "step": 7743 }, { "epoch": 1.7033982184097658, "grad_norm": 0.3427136540412903, "learning_rate": 8e-05, "loss": 1.7434, "step": 7744 }, { "epoch": 1.703618167821401, "grad_norm": 0.33775994181632996, "learning_rate": 8e-05, "loss": 1.5892, "step": 7745 }, { "epoch": 1.7038381172330364, "grad_norm": 0.34740275144577026, "learning_rate": 8e-05, "loss": 1.7122, "step": 7746 }, { "epoch": 1.7040580666446719, "grad_norm": 0.3429790437221527, "learning_rate": 8e-05, "loss": 1.6039, "step": 7747 }, { "epoch": 1.704278016056307, "grad_norm": 0.34647807478904724, "learning_rate": 8e-05, "loss": 1.5403, "step": 7748 }, { "epoch": 1.7044979654679424, "grad_norm": 0.34227967262268066, "learning_rate": 8e-05, "loss": 1.4166, "step": 7749 }, { "epoch": 1.7047179148795777, "grad_norm": 0.3692344129085541, "learning_rate": 8e-05, "loss": 1.3192, "step": 7750 }, { "epoch": 1.704937864291213, "grad_norm": 0.3530147671699524, "learning_rate": 8e-05, "loss": 1.5864, "step": 7751 }, { "epoch": 1.7051578137028485, "grad_norm": 0.35274988412857056, "learning_rate": 8e-05, "loss": 1.6525, "step": 7752 }, { "epoch": 1.7053777631144835, "grad_norm": 0.33769187331199646, "learning_rate": 8e-05, "loss": 1.4496, "step": 7753 }, { "epoch": 1.705597712526119, "grad_norm": 0.3667060136795044, "learning_rate": 8e-05, "loss": 1.6382, "step": 7754 }, { "epoch": 1.7058176619377543, "grad_norm": 0.36774295568466187, "learning_rate": 8e-05, "loss": 1.462, "step": 7755 }, { "epoch": 1.7060376113493896, "grad_norm": 0.37426048517227173, "learning_rate": 8e-05, "loss": 1.63, "step": 7756 }, { "epoch": 1.706257560761025, "grad_norm": 0.35134580731391907, "learning_rate": 8e-05, "loss": 1.6216, "step": 7757 }, { "epoch": 1.7064775101726601, "grad_norm": 0.34836113452911377, "learning_rate": 8e-05, "loss": 1.4297, "step": 7758 }, { "epoch": 1.7066974595842956, "grad_norm": 0.3489657938480377, "learning_rate": 8e-05, "loss": 1.557, "step": 7759 }, { "epoch": 1.706917408995931, "grad_norm": 0.3450574278831482, "learning_rate": 8e-05, "loss": 1.634, "step": 7760 }, { "epoch": 1.7071373584075662, "grad_norm": 0.3414944112300873, "learning_rate": 8e-05, "loss": 1.4602, "step": 7761 }, { "epoch": 1.7073573078192017, "grad_norm": 0.3658798038959503, "learning_rate": 8e-05, "loss": 1.6834, "step": 7762 }, { "epoch": 1.7075772572308368, "grad_norm": 0.36673933267593384, "learning_rate": 8e-05, "loss": 1.7, "step": 7763 }, { "epoch": 1.7077972066424723, "grad_norm": 0.3474530577659607, "learning_rate": 8e-05, "loss": 1.5059, "step": 7764 }, { "epoch": 1.7080171560541075, "grad_norm": 0.359688401222229, "learning_rate": 8e-05, "loss": 1.5702, "step": 7765 }, { "epoch": 1.7082371054657428, "grad_norm": 0.3471611738204956, "learning_rate": 8e-05, "loss": 1.521, "step": 7766 }, { "epoch": 1.7084570548773783, "grad_norm": 0.36128219962120056, "learning_rate": 8e-05, "loss": 1.5053, "step": 7767 }, { "epoch": 1.7086770042890134, "grad_norm": 0.3539958596229553, "learning_rate": 8e-05, "loss": 1.6126, "step": 7768 }, { "epoch": 1.7088969537006489, "grad_norm": 0.33570200204849243, "learning_rate": 8e-05, "loss": 1.5333, "step": 7769 }, { "epoch": 1.7091169031122841, "grad_norm": 0.3666260540485382, "learning_rate": 8e-05, "loss": 1.5321, "step": 7770 }, { "epoch": 1.7093368525239194, "grad_norm": 0.3419763743877411, "learning_rate": 8e-05, "loss": 1.5277, "step": 7771 }, { "epoch": 1.709556801935555, "grad_norm": 0.3419716954231262, "learning_rate": 8e-05, "loss": 1.7044, "step": 7772 }, { "epoch": 1.70977675134719, "grad_norm": 0.3493288457393646, "learning_rate": 8e-05, "loss": 1.4769, "step": 7773 }, { "epoch": 1.7099967007588255, "grad_norm": 0.33529606461524963, "learning_rate": 8e-05, "loss": 1.4739, "step": 7774 }, { "epoch": 1.7102166501704608, "grad_norm": 0.36344289779663086, "learning_rate": 8e-05, "loss": 1.6125, "step": 7775 }, { "epoch": 1.710436599582096, "grad_norm": 0.3962048888206482, "learning_rate": 8e-05, "loss": 1.6139, "step": 7776 }, { "epoch": 1.7106565489937315, "grad_norm": 0.35410889983177185, "learning_rate": 8e-05, "loss": 1.6246, "step": 7777 }, { "epoch": 1.7108764984053668, "grad_norm": 0.34323057532310486, "learning_rate": 8e-05, "loss": 1.6743, "step": 7778 }, { "epoch": 1.711096447817002, "grad_norm": 0.332224041223526, "learning_rate": 8e-05, "loss": 1.6377, "step": 7779 }, { "epoch": 1.7113163972286374, "grad_norm": 0.3297925293445587, "learning_rate": 8e-05, "loss": 1.5849, "step": 7780 }, { "epoch": 1.7115363466402727, "grad_norm": 0.34456944465637207, "learning_rate": 8e-05, "loss": 1.6061, "step": 7781 }, { "epoch": 1.7117562960519082, "grad_norm": 0.3560841381549835, "learning_rate": 8e-05, "loss": 1.6452, "step": 7782 }, { "epoch": 1.7119762454635434, "grad_norm": 0.3385053873062134, "learning_rate": 8e-05, "loss": 1.5012, "step": 7783 }, { "epoch": 1.7121961948751787, "grad_norm": 0.3615650236606598, "learning_rate": 8e-05, "loss": 1.4784, "step": 7784 }, { "epoch": 1.712416144286814, "grad_norm": 0.345405250787735, "learning_rate": 8e-05, "loss": 1.6377, "step": 7785 }, { "epoch": 1.7126360936984493, "grad_norm": 0.35640987753868103, "learning_rate": 8e-05, "loss": 1.6291, "step": 7786 }, { "epoch": 1.7128560431100848, "grad_norm": 0.3619750440120697, "learning_rate": 8e-05, "loss": 1.5742, "step": 7787 }, { "epoch": 1.71307599252172, "grad_norm": 0.367191344499588, "learning_rate": 8e-05, "loss": 1.7717, "step": 7788 }, { "epoch": 1.7132959419333553, "grad_norm": 0.33721840381622314, "learning_rate": 8e-05, "loss": 1.5146, "step": 7789 }, { "epoch": 1.7135158913449906, "grad_norm": 0.3617967963218689, "learning_rate": 8e-05, "loss": 1.5809, "step": 7790 }, { "epoch": 1.7137358407566259, "grad_norm": 0.3473452627658844, "learning_rate": 8e-05, "loss": 1.6685, "step": 7791 }, { "epoch": 1.7139557901682614, "grad_norm": 0.3493829369544983, "learning_rate": 8e-05, "loss": 1.4468, "step": 7792 }, { "epoch": 1.7141757395798967, "grad_norm": 0.33838924765586853, "learning_rate": 8e-05, "loss": 1.6047, "step": 7793 }, { "epoch": 1.714395688991532, "grad_norm": 0.3695904016494751, "learning_rate": 8e-05, "loss": 1.6672, "step": 7794 }, { "epoch": 1.7146156384031672, "grad_norm": 0.3348905146121979, "learning_rate": 8e-05, "loss": 1.6177, "step": 7795 }, { "epoch": 1.7148355878148025, "grad_norm": 0.35554322600364685, "learning_rate": 8e-05, "loss": 1.7369, "step": 7796 }, { "epoch": 1.715055537226438, "grad_norm": 0.34476399421691895, "learning_rate": 8e-05, "loss": 1.6547, "step": 7797 }, { "epoch": 1.7152754866380733, "grad_norm": 0.35872918367385864, "learning_rate": 8e-05, "loss": 1.6519, "step": 7798 }, { "epoch": 1.7154954360497086, "grad_norm": 0.3418973684310913, "learning_rate": 8e-05, "loss": 1.3963, "step": 7799 }, { "epoch": 1.715715385461344, "grad_norm": 0.36529600620269775, "learning_rate": 8e-05, "loss": 1.4988, "step": 7800 }, { "epoch": 1.7159353348729791, "grad_norm": 0.33505308628082275, "learning_rate": 8e-05, "loss": 1.5366, "step": 7801 }, { "epoch": 1.7161552842846146, "grad_norm": 0.35838159918785095, "learning_rate": 8e-05, "loss": 1.6136, "step": 7802 }, { "epoch": 1.71637523369625, "grad_norm": 0.34196028113365173, "learning_rate": 8e-05, "loss": 1.6101, "step": 7803 }, { "epoch": 1.7165951831078852, "grad_norm": 0.35280483961105347, "learning_rate": 8e-05, "loss": 1.6194, "step": 7804 }, { "epoch": 1.7168151325195207, "grad_norm": 0.35943806171417236, "learning_rate": 8e-05, "loss": 1.7032, "step": 7805 }, { "epoch": 1.7170350819311557, "grad_norm": 0.35729214549064636, "learning_rate": 8e-05, "loss": 1.4864, "step": 7806 }, { "epoch": 1.7172550313427912, "grad_norm": 0.3597615361213684, "learning_rate": 8e-05, "loss": 1.5602, "step": 7807 }, { "epoch": 1.7174749807544265, "grad_norm": 0.346787691116333, "learning_rate": 8e-05, "loss": 1.6529, "step": 7808 }, { "epoch": 1.7176949301660618, "grad_norm": 0.3508087694644928, "learning_rate": 8e-05, "loss": 1.5481, "step": 7809 }, { "epoch": 1.7179148795776973, "grad_norm": 0.34571418166160583, "learning_rate": 8e-05, "loss": 1.6169, "step": 7810 }, { "epoch": 1.7181348289893323, "grad_norm": 0.35204780101776123, "learning_rate": 8e-05, "loss": 1.5379, "step": 7811 }, { "epoch": 1.7183547784009678, "grad_norm": 0.34475111961364746, "learning_rate": 8e-05, "loss": 1.4448, "step": 7812 }, { "epoch": 1.7185747278126031, "grad_norm": 0.3491145670413971, "learning_rate": 8e-05, "loss": 1.5898, "step": 7813 }, { "epoch": 1.7187946772242384, "grad_norm": 0.37362223863601685, "learning_rate": 8e-05, "loss": 1.7339, "step": 7814 }, { "epoch": 1.719014626635874, "grad_norm": 0.35229745507240295, "learning_rate": 8e-05, "loss": 1.5568, "step": 7815 }, { "epoch": 1.719234576047509, "grad_norm": 0.3532031774520874, "learning_rate": 8e-05, "loss": 1.6393, "step": 7816 }, { "epoch": 1.7194545254591445, "grad_norm": 0.35474610328674316, "learning_rate": 8e-05, "loss": 1.5553, "step": 7817 }, { "epoch": 1.7196744748707797, "grad_norm": 0.36398574709892273, "learning_rate": 8e-05, "loss": 1.6263, "step": 7818 }, { "epoch": 1.719894424282415, "grad_norm": 0.43163037300109863, "learning_rate": 8e-05, "loss": 1.6243, "step": 7819 }, { "epoch": 1.7201143736940505, "grad_norm": 0.37369412183761597, "learning_rate": 8e-05, "loss": 1.6773, "step": 7820 }, { "epoch": 1.7203343231056856, "grad_norm": 0.36406105756759644, "learning_rate": 8e-05, "loss": 1.6252, "step": 7821 }, { "epoch": 1.720554272517321, "grad_norm": 0.34812697768211365, "learning_rate": 8e-05, "loss": 1.6057, "step": 7822 }, { "epoch": 1.7207742219289563, "grad_norm": 0.34681203961372375, "learning_rate": 8e-05, "loss": 1.5092, "step": 7823 }, { "epoch": 1.7209941713405916, "grad_norm": 0.40526044368743896, "learning_rate": 8e-05, "loss": 1.5708, "step": 7824 }, { "epoch": 1.7212141207522271, "grad_norm": 0.3921516239643097, "learning_rate": 8e-05, "loss": 1.6746, "step": 7825 }, { "epoch": 1.7214340701638622, "grad_norm": 0.34016990661621094, "learning_rate": 8e-05, "loss": 1.5927, "step": 7826 }, { "epoch": 1.7216540195754977, "grad_norm": 0.37025198340415955, "learning_rate": 8e-05, "loss": 1.5905, "step": 7827 }, { "epoch": 1.721873968987133, "grad_norm": 0.367813378572464, "learning_rate": 8e-05, "loss": 1.5694, "step": 7828 }, { "epoch": 1.7220939183987682, "grad_norm": 0.356509268283844, "learning_rate": 8e-05, "loss": 1.7333, "step": 7829 }, { "epoch": 1.7223138678104037, "grad_norm": 0.36265692114830017, "learning_rate": 8e-05, "loss": 1.5921, "step": 7830 }, { "epoch": 1.7225338172220388, "grad_norm": 0.35546553134918213, "learning_rate": 8e-05, "loss": 1.4618, "step": 7831 }, { "epoch": 1.7227537666336743, "grad_norm": 0.39496615529060364, "learning_rate": 8e-05, "loss": 1.7735, "step": 7832 }, { "epoch": 1.7229737160453096, "grad_norm": 0.3474973142147064, "learning_rate": 8e-05, "loss": 1.5738, "step": 7833 }, { "epoch": 1.7231936654569449, "grad_norm": 0.3293529748916626, "learning_rate": 8e-05, "loss": 1.4847, "step": 7834 }, { "epoch": 1.7234136148685804, "grad_norm": 0.33390429615974426, "learning_rate": 8e-05, "loss": 1.4683, "step": 7835 }, { "epoch": 1.7236335642802154, "grad_norm": 0.3572830259799957, "learning_rate": 8e-05, "loss": 1.5789, "step": 7836 }, { "epoch": 1.723853513691851, "grad_norm": 0.3391188383102417, "learning_rate": 8e-05, "loss": 1.5588, "step": 7837 }, { "epoch": 1.7240734631034862, "grad_norm": 0.34888502955436707, "learning_rate": 8e-05, "loss": 1.5542, "step": 7838 }, { "epoch": 1.7242934125151215, "grad_norm": 0.34110400080680847, "learning_rate": 8e-05, "loss": 1.5649, "step": 7839 }, { "epoch": 1.724513361926757, "grad_norm": 0.3803638517856598, "learning_rate": 8e-05, "loss": 1.6757, "step": 7840 }, { "epoch": 1.724733311338392, "grad_norm": 0.3603278696537018, "learning_rate": 8e-05, "loss": 1.6592, "step": 7841 }, { "epoch": 1.7249532607500275, "grad_norm": 0.36675170063972473, "learning_rate": 8e-05, "loss": 1.8492, "step": 7842 }, { "epoch": 1.7251732101616628, "grad_norm": 0.3173739016056061, "learning_rate": 8e-05, "loss": 1.2593, "step": 7843 }, { "epoch": 1.725393159573298, "grad_norm": 0.3419693112373352, "learning_rate": 8e-05, "loss": 1.5742, "step": 7844 }, { "epoch": 1.7256131089849336, "grad_norm": 0.3470190465450287, "learning_rate": 8e-05, "loss": 1.469, "step": 7845 }, { "epoch": 1.7258330583965686, "grad_norm": 0.3415330946445465, "learning_rate": 8e-05, "loss": 1.5418, "step": 7846 }, { "epoch": 1.7260530078082041, "grad_norm": 0.3489103615283966, "learning_rate": 8e-05, "loss": 1.5936, "step": 7847 }, { "epoch": 1.7262729572198394, "grad_norm": 0.33945247530937195, "learning_rate": 8e-05, "loss": 1.5115, "step": 7848 }, { "epoch": 1.7264929066314747, "grad_norm": 0.33075830340385437, "learning_rate": 8e-05, "loss": 1.6041, "step": 7849 }, { "epoch": 1.7267128560431102, "grad_norm": 0.3232467472553253, "learning_rate": 8e-05, "loss": 1.5328, "step": 7850 }, { "epoch": 1.7269328054547453, "grad_norm": 0.3401769697666168, "learning_rate": 8e-05, "loss": 1.5658, "step": 7851 }, { "epoch": 1.7271527548663808, "grad_norm": 0.34787148237228394, "learning_rate": 8e-05, "loss": 1.59, "step": 7852 }, { "epoch": 1.727372704278016, "grad_norm": 0.35189884901046753, "learning_rate": 8e-05, "loss": 1.5368, "step": 7853 }, { "epoch": 1.7275926536896513, "grad_norm": 0.38421642780303955, "learning_rate": 8e-05, "loss": 1.7876, "step": 7854 }, { "epoch": 1.7278126031012868, "grad_norm": 0.35964158177375793, "learning_rate": 8e-05, "loss": 1.5424, "step": 7855 }, { "epoch": 1.7280325525129219, "grad_norm": 0.3539362847805023, "learning_rate": 8e-05, "loss": 1.5728, "step": 7856 }, { "epoch": 1.7282525019245574, "grad_norm": 0.3650361895561218, "learning_rate": 8e-05, "loss": 1.6141, "step": 7857 }, { "epoch": 1.7284724513361927, "grad_norm": 0.3370092809200287, "learning_rate": 8e-05, "loss": 1.4955, "step": 7858 }, { "epoch": 1.728692400747828, "grad_norm": 0.365807443857193, "learning_rate": 8e-05, "loss": 1.7166, "step": 7859 }, { "epoch": 1.7289123501594634, "grad_norm": 0.38822993636131287, "learning_rate": 8e-05, "loss": 1.5902, "step": 7860 }, { "epoch": 1.7291322995710985, "grad_norm": 0.36028799414634705, "learning_rate": 8e-05, "loss": 1.6042, "step": 7861 }, { "epoch": 1.729352248982734, "grad_norm": 0.357787162065506, "learning_rate": 8e-05, "loss": 1.6409, "step": 7862 }, { "epoch": 1.7295721983943693, "grad_norm": 0.32934266328811646, "learning_rate": 8e-05, "loss": 1.4488, "step": 7863 }, { "epoch": 1.7297921478060045, "grad_norm": 0.34515997767448425, "learning_rate": 8e-05, "loss": 1.599, "step": 7864 }, { "epoch": 1.73001209721764, "grad_norm": 0.3542462885379791, "learning_rate": 8e-05, "loss": 1.5943, "step": 7865 }, { "epoch": 1.7302320466292753, "grad_norm": 0.33389803767204285, "learning_rate": 8e-05, "loss": 1.5504, "step": 7866 }, { "epoch": 1.7304519960409106, "grad_norm": 0.3840927481651306, "learning_rate": 8e-05, "loss": 1.8017, "step": 7867 }, { "epoch": 1.7306719454525459, "grad_norm": 0.37774398922920227, "learning_rate": 8e-05, "loss": 1.5666, "step": 7868 }, { "epoch": 1.7308918948641812, "grad_norm": 0.34967049956321716, "learning_rate": 8e-05, "loss": 1.592, "step": 7869 }, { "epoch": 1.7311118442758167, "grad_norm": 0.383025586605072, "learning_rate": 8e-05, "loss": 1.6896, "step": 7870 }, { "epoch": 1.731331793687452, "grad_norm": 0.35885441303253174, "learning_rate": 8e-05, "loss": 1.6397, "step": 7871 }, { "epoch": 1.7315517430990872, "grad_norm": 0.36957621574401855, "learning_rate": 8e-05, "loss": 1.704, "step": 7872 }, { "epoch": 1.7317716925107225, "grad_norm": 0.34623825550079346, "learning_rate": 8e-05, "loss": 1.6981, "step": 7873 }, { "epoch": 1.7319916419223578, "grad_norm": 0.3668917119503021, "learning_rate": 8e-05, "loss": 1.5943, "step": 7874 }, { "epoch": 1.7322115913339933, "grad_norm": 0.37584593892097473, "learning_rate": 8e-05, "loss": 1.5055, "step": 7875 }, { "epoch": 1.7324315407456286, "grad_norm": 0.3367500603199005, "learning_rate": 8e-05, "loss": 1.59, "step": 7876 }, { "epoch": 1.7326514901572638, "grad_norm": 0.34857890009880066, "learning_rate": 8e-05, "loss": 1.5282, "step": 7877 }, { "epoch": 1.732871439568899, "grad_norm": 0.34203341603279114, "learning_rate": 8e-05, "loss": 1.5861, "step": 7878 }, { "epoch": 1.7330913889805344, "grad_norm": 0.34288841485977173, "learning_rate": 8e-05, "loss": 1.6084, "step": 7879 }, { "epoch": 1.7333113383921699, "grad_norm": 0.3319281339645386, "learning_rate": 8e-05, "loss": 1.5982, "step": 7880 }, { "epoch": 1.7335312878038052, "grad_norm": 0.3457512855529785, "learning_rate": 8e-05, "loss": 1.5853, "step": 7881 }, { "epoch": 1.7337512372154404, "grad_norm": 0.35210123658180237, "learning_rate": 8e-05, "loss": 1.7016, "step": 7882 }, { "epoch": 1.7339711866270757, "grad_norm": 0.32268011569976807, "learning_rate": 8e-05, "loss": 1.4875, "step": 7883 }, { "epoch": 1.734191136038711, "grad_norm": 0.35991212725639343, "learning_rate": 8e-05, "loss": 1.5753, "step": 7884 }, { "epoch": 1.7344110854503465, "grad_norm": 0.36137259006500244, "learning_rate": 8e-05, "loss": 1.7679, "step": 7885 }, { "epoch": 1.7346310348619818, "grad_norm": 0.335427463054657, "learning_rate": 8e-05, "loss": 1.4731, "step": 7886 }, { "epoch": 1.734850984273617, "grad_norm": 0.345337837934494, "learning_rate": 8e-05, "loss": 1.5851, "step": 7887 }, { "epoch": 1.7350709336852523, "grad_norm": 0.476873517036438, "learning_rate": 8e-05, "loss": 1.6014, "step": 7888 }, { "epoch": 1.7352908830968876, "grad_norm": 0.34463798999786377, "learning_rate": 8e-05, "loss": 1.4861, "step": 7889 }, { "epoch": 1.7355108325085231, "grad_norm": 0.3821248710155487, "learning_rate": 8e-05, "loss": 1.6839, "step": 7890 }, { "epoch": 1.7357307819201584, "grad_norm": 0.34933704137802124, "learning_rate": 8e-05, "loss": 1.7067, "step": 7891 }, { "epoch": 1.7359507313317937, "grad_norm": 0.381144642829895, "learning_rate": 8e-05, "loss": 1.708, "step": 7892 }, { "epoch": 1.7361706807434292, "grad_norm": 0.3577573299407959, "learning_rate": 8e-05, "loss": 1.776, "step": 7893 }, { "epoch": 1.7363906301550642, "grad_norm": 0.3684750199317932, "learning_rate": 8e-05, "loss": 1.6826, "step": 7894 }, { "epoch": 1.7366105795666997, "grad_norm": 0.3623203933238983, "learning_rate": 8e-05, "loss": 1.5821, "step": 7895 }, { "epoch": 1.736830528978335, "grad_norm": 0.3776145577430725, "learning_rate": 8e-05, "loss": 1.6678, "step": 7896 }, { "epoch": 1.7370504783899703, "grad_norm": 0.3955874443054199, "learning_rate": 8e-05, "loss": 1.4434, "step": 7897 }, { "epoch": 1.7372704278016058, "grad_norm": 0.33973976969718933, "learning_rate": 8e-05, "loss": 1.52, "step": 7898 }, { "epoch": 1.7374903772132408, "grad_norm": 0.3331288993358612, "learning_rate": 8e-05, "loss": 1.521, "step": 7899 }, { "epoch": 1.7377103266248763, "grad_norm": 0.3575076758861542, "learning_rate": 8e-05, "loss": 1.7096, "step": 7900 }, { "epoch": 1.7379302760365116, "grad_norm": 0.34671491384506226, "learning_rate": 8e-05, "loss": 1.5599, "step": 7901 }, { "epoch": 1.738150225448147, "grad_norm": 0.3410821855068207, "learning_rate": 8e-05, "loss": 1.5576, "step": 7902 }, { "epoch": 1.7383701748597824, "grad_norm": 0.3565903604030609, "learning_rate": 8e-05, "loss": 1.5294, "step": 7903 }, { "epoch": 1.7385901242714175, "grad_norm": 0.3541129231452942, "learning_rate": 8e-05, "loss": 1.4841, "step": 7904 }, { "epoch": 1.738810073683053, "grad_norm": 0.4308437705039978, "learning_rate": 8e-05, "loss": 1.7901, "step": 7905 }, { "epoch": 1.7390300230946882, "grad_norm": 0.3442606031894684, "learning_rate": 8e-05, "loss": 1.5137, "step": 7906 }, { "epoch": 1.7392499725063235, "grad_norm": 0.36624211072921753, "learning_rate": 8e-05, "loss": 1.6006, "step": 7907 }, { "epoch": 1.739469921917959, "grad_norm": 0.32359573245048523, "learning_rate": 8e-05, "loss": 1.4603, "step": 7908 }, { "epoch": 1.739689871329594, "grad_norm": 0.3635515570640564, "learning_rate": 8e-05, "loss": 1.5882, "step": 7909 }, { "epoch": 1.7399098207412296, "grad_norm": 0.3601096272468567, "learning_rate": 8e-05, "loss": 1.6078, "step": 7910 }, { "epoch": 1.7401297701528649, "grad_norm": 0.336860328912735, "learning_rate": 8e-05, "loss": 1.4761, "step": 7911 }, { "epoch": 1.7403497195645001, "grad_norm": 0.3369137644767761, "learning_rate": 8e-05, "loss": 1.4661, "step": 7912 }, { "epoch": 1.7405696689761356, "grad_norm": 0.34379833936691284, "learning_rate": 8e-05, "loss": 1.5632, "step": 7913 }, { "epoch": 1.7407896183877707, "grad_norm": 0.3572351932525635, "learning_rate": 8e-05, "loss": 1.6742, "step": 7914 }, { "epoch": 1.7410095677994062, "grad_norm": 0.34679514169692993, "learning_rate": 8e-05, "loss": 1.6858, "step": 7915 }, { "epoch": 1.7412295172110415, "grad_norm": 0.3664378225803375, "learning_rate": 8e-05, "loss": 1.61, "step": 7916 }, { "epoch": 1.7414494666226767, "grad_norm": 0.35504642128944397, "learning_rate": 8e-05, "loss": 1.5353, "step": 7917 }, { "epoch": 1.7416694160343122, "grad_norm": 0.3425410985946655, "learning_rate": 8e-05, "loss": 1.4514, "step": 7918 }, { "epoch": 1.7418893654459473, "grad_norm": 0.34990036487579346, "learning_rate": 8e-05, "loss": 1.6621, "step": 7919 }, { "epoch": 1.7421093148575828, "grad_norm": 0.3625769019126892, "learning_rate": 8e-05, "loss": 1.4235, "step": 7920 }, { "epoch": 1.742329264269218, "grad_norm": 0.33893001079559326, "learning_rate": 8e-05, "loss": 1.5003, "step": 7921 }, { "epoch": 1.7425492136808534, "grad_norm": 0.3550249934196472, "learning_rate": 8e-05, "loss": 1.6594, "step": 7922 }, { "epoch": 1.7427691630924889, "grad_norm": 0.35974204540252686, "learning_rate": 8e-05, "loss": 1.7804, "step": 7923 }, { "epoch": 1.742989112504124, "grad_norm": 0.34874427318573, "learning_rate": 8e-05, "loss": 1.6448, "step": 7924 }, { "epoch": 1.7432090619157594, "grad_norm": 0.3487253785133362, "learning_rate": 8e-05, "loss": 1.5805, "step": 7925 }, { "epoch": 1.7434290113273947, "grad_norm": 0.3251039683818817, "learning_rate": 8e-05, "loss": 1.4275, "step": 7926 }, { "epoch": 1.74364896073903, "grad_norm": 0.36586251854896545, "learning_rate": 8e-05, "loss": 1.7247, "step": 7927 }, { "epoch": 1.7438689101506655, "grad_norm": 0.3744462728500366, "learning_rate": 8e-05, "loss": 1.6698, "step": 7928 }, { "epoch": 1.7440888595623005, "grad_norm": 0.3497055768966675, "learning_rate": 8e-05, "loss": 1.579, "step": 7929 }, { "epoch": 1.744308808973936, "grad_norm": 0.35381507873535156, "learning_rate": 8e-05, "loss": 1.6207, "step": 7930 }, { "epoch": 1.7445287583855713, "grad_norm": 0.33449801802635193, "learning_rate": 8e-05, "loss": 1.532, "step": 7931 }, { "epoch": 1.7447487077972066, "grad_norm": 0.34029582142829895, "learning_rate": 8e-05, "loss": 1.5437, "step": 7932 }, { "epoch": 1.744968657208842, "grad_norm": 0.35937413573265076, "learning_rate": 8e-05, "loss": 1.4756, "step": 7933 }, { "epoch": 1.7451886066204771, "grad_norm": 0.34804102778434753, "learning_rate": 8e-05, "loss": 1.5688, "step": 7934 }, { "epoch": 1.7454085560321126, "grad_norm": 0.36396536231040955, "learning_rate": 8e-05, "loss": 1.584, "step": 7935 }, { "epoch": 1.745628505443748, "grad_norm": 0.3658052682876587, "learning_rate": 8e-05, "loss": 1.6738, "step": 7936 }, { "epoch": 1.7458484548553832, "grad_norm": 0.3434094786643982, "learning_rate": 8e-05, "loss": 1.5886, "step": 7937 }, { "epoch": 1.7460684042670187, "grad_norm": 0.3775242567062378, "learning_rate": 8e-05, "loss": 1.5737, "step": 7938 }, { "epoch": 1.7462883536786538, "grad_norm": 0.3519570231437683, "learning_rate": 8e-05, "loss": 1.6672, "step": 7939 }, { "epoch": 1.7465083030902893, "grad_norm": 0.3512040674686432, "learning_rate": 8e-05, "loss": 1.532, "step": 7940 }, { "epoch": 1.7467282525019245, "grad_norm": 0.3413964509963989, "learning_rate": 8e-05, "loss": 1.376, "step": 7941 }, { "epoch": 1.7469482019135598, "grad_norm": 0.33032524585723877, "learning_rate": 8e-05, "loss": 1.4664, "step": 7942 }, { "epoch": 1.7471681513251953, "grad_norm": 0.3503587245941162, "learning_rate": 8e-05, "loss": 1.6585, "step": 7943 }, { "epoch": 1.7473881007368304, "grad_norm": 0.3558482229709625, "learning_rate": 8e-05, "loss": 1.4895, "step": 7944 }, { "epoch": 1.7476080501484659, "grad_norm": 0.34020620584487915, "learning_rate": 8e-05, "loss": 1.5243, "step": 7945 }, { "epoch": 1.7478279995601012, "grad_norm": 0.35632142424583435, "learning_rate": 8e-05, "loss": 1.4965, "step": 7946 }, { "epoch": 1.7480479489717364, "grad_norm": 0.3349021375179291, "learning_rate": 8e-05, "loss": 1.5878, "step": 7947 }, { "epoch": 1.748267898383372, "grad_norm": 0.3433366119861603, "learning_rate": 8e-05, "loss": 1.4771, "step": 7948 }, { "epoch": 1.748487847795007, "grad_norm": 0.3456018269062042, "learning_rate": 8e-05, "loss": 1.4231, "step": 7949 }, { "epoch": 1.7487077972066425, "grad_norm": 0.3500257432460785, "learning_rate": 8e-05, "loss": 1.4885, "step": 7950 }, { "epoch": 1.7489277466182778, "grad_norm": 0.3766205906867981, "learning_rate": 8e-05, "loss": 1.5456, "step": 7951 }, { "epoch": 1.749147696029913, "grad_norm": 0.34870249032974243, "learning_rate": 8e-05, "loss": 1.7438, "step": 7952 }, { "epoch": 1.7493676454415485, "grad_norm": 0.3662711977958679, "learning_rate": 8e-05, "loss": 1.6039, "step": 7953 }, { "epoch": 1.7495875948531836, "grad_norm": 0.38237541913986206, "learning_rate": 8e-05, "loss": 1.6035, "step": 7954 }, { "epoch": 1.749807544264819, "grad_norm": 0.4088335931301117, "learning_rate": 8e-05, "loss": 1.5414, "step": 7955 }, { "epoch": 1.7500274936764544, "grad_norm": 0.36536529660224915, "learning_rate": 8e-05, "loss": 1.5442, "step": 7956 }, { "epoch": 1.7502474430880897, "grad_norm": 0.34003692865371704, "learning_rate": 8e-05, "loss": 1.6036, "step": 7957 }, { "epoch": 1.7504673924997252, "grad_norm": 0.34233033657073975, "learning_rate": 8e-05, "loss": 1.5698, "step": 7958 }, { "epoch": 1.7506873419113604, "grad_norm": 0.3301025331020355, "learning_rate": 8e-05, "loss": 1.5997, "step": 7959 }, { "epoch": 1.7509072913229957, "grad_norm": 0.33631059527397156, "learning_rate": 8e-05, "loss": 1.4276, "step": 7960 }, { "epoch": 1.751127240734631, "grad_norm": 0.3597167730331421, "learning_rate": 8e-05, "loss": 1.6462, "step": 7961 }, { "epoch": 1.7513471901462663, "grad_norm": 0.3461626172065735, "learning_rate": 8e-05, "loss": 1.5604, "step": 7962 }, { "epoch": 1.7515671395579018, "grad_norm": 0.3671514689922333, "learning_rate": 8e-05, "loss": 1.6492, "step": 7963 }, { "epoch": 1.751787088969537, "grad_norm": 0.3526211977005005, "learning_rate": 8e-05, "loss": 1.6232, "step": 7964 }, { "epoch": 1.7520070383811723, "grad_norm": 0.37618961930274963, "learning_rate": 8e-05, "loss": 1.6916, "step": 7965 }, { "epoch": 1.7522269877928076, "grad_norm": 0.3503290116786957, "learning_rate": 8e-05, "loss": 1.571, "step": 7966 }, { "epoch": 1.752446937204443, "grad_norm": 0.3252408802509308, "learning_rate": 8e-05, "loss": 1.4224, "step": 7967 }, { "epoch": 1.7526668866160784, "grad_norm": 0.3436092734336853, "learning_rate": 8e-05, "loss": 1.4258, "step": 7968 }, { "epoch": 1.7528868360277137, "grad_norm": 0.3433966338634491, "learning_rate": 8e-05, "loss": 1.4811, "step": 7969 }, { "epoch": 1.753106785439349, "grad_norm": 0.3561773896217346, "learning_rate": 8e-05, "loss": 1.6046, "step": 7970 }, { "epoch": 1.7533267348509842, "grad_norm": 0.37637168169021606, "learning_rate": 8e-05, "loss": 1.7521, "step": 7971 }, { "epoch": 1.7535466842626195, "grad_norm": 0.35408321022987366, "learning_rate": 8e-05, "loss": 1.6294, "step": 7972 }, { "epoch": 1.753766633674255, "grad_norm": 0.347351998090744, "learning_rate": 8e-05, "loss": 1.4515, "step": 7973 }, { "epoch": 1.7539865830858903, "grad_norm": 0.3497125804424286, "learning_rate": 8e-05, "loss": 1.6235, "step": 7974 }, { "epoch": 1.7542065324975256, "grad_norm": 0.35236796736717224, "learning_rate": 8e-05, "loss": 1.5209, "step": 7975 }, { "epoch": 1.7544264819091608, "grad_norm": 0.34796035289764404, "learning_rate": 8e-05, "loss": 1.6913, "step": 7976 }, { "epoch": 1.7546464313207961, "grad_norm": 0.36231517791748047, "learning_rate": 8e-05, "loss": 1.4603, "step": 7977 }, { "epoch": 1.7548663807324316, "grad_norm": 0.33842164278030396, "learning_rate": 8e-05, "loss": 1.4795, "step": 7978 }, { "epoch": 1.755086330144067, "grad_norm": 0.3390091359615326, "learning_rate": 8e-05, "loss": 1.4714, "step": 7979 }, { "epoch": 1.7553062795557022, "grad_norm": 0.36939457058906555, "learning_rate": 8e-05, "loss": 1.647, "step": 7980 }, { "epoch": 1.7555262289673377, "grad_norm": 0.3471515476703644, "learning_rate": 8e-05, "loss": 1.6681, "step": 7981 }, { "epoch": 1.7557461783789727, "grad_norm": 0.36616408824920654, "learning_rate": 8e-05, "loss": 1.4655, "step": 7982 }, { "epoch": 1.7559661277906082, "grad_norm": 0.36517783999443054, "learning_rate": 8e-05, "loss": 1.7571, "step": 7983 }, { "epoch": 1.7561860772022435, "grad_norm": 0.3458475172519684, "learning_rate": 8e-05, "loss": 1.6751, "step": 7984 }, { "epoch": 1.7564060266138788, "grad_norm": 0.33926698565483093, "learning_rate": 8e-05, "loss": 1.4545, "step": 7985 }, { "epoch": 1.7566259760255143, "grad_norm": 0.3569273352622986, "learning_rate": 8e-05, "loss": 1.6073, "step": 7986 }, { "epoch": 1.7568459254371493, "grad_norm": 0.3936539590358734, "learning_rate": 8e-05, "loss": 1.6451, "step": 7987 }, { "epoch": 1.7570658748487848, "grad_norm": 0.35348644852638245, "learning_rate": 8e-05, "loss": 1.5086, "step": 7988 }, { "epoch": 1.7572858242604201, "grad_norm": 0.35798293352127075, "learning_rate": 8e-05, "loss": 1.4494, "step": 7989 }, { "epoch": 1.7575057736720554, "grad_norm": 0.3562926948070526, "learning_rate": 8e-05, "loss": 1.729, "step": 7990 }, { "epoch": 1.757725723083691, "grad_norm": 0.3411732614040375, "learning_rate": 8e-05, "loss": 1.438, "step": 7991 }, { "epoch": 1.757945672495326, "grad_norm": 0.3618207573890686, "learning_rate": 8e-05, "loss": 1.6084, "step": 7992 }, { "epoch": 1.7581656219069615, "grad_norm": 0.36162880063056946, "learning_rate": 8e-05, "loss": 1.6887, "step": 7993 }, { "epoch": 1.7583855713185967, "grad_norm": 0.3642195761203766, "learning_rate": 8e-05, "loss": 1.6395, "step": 7994 }, { "epoch": 1.758605520730232, "grad_norm": 0.34289413690567017, "learning_rate": 8e-05, "loss": 1.5219, "step": 7995 }, { "epoch": 1.7588254701418675, "grad_norm": 0.35375306010246277, "learning_rate": 8e-05, "loss": 1.6616, "step": 7996 }, { "epoch": 1.7590454195535026, "grad_norm": 0.345392107963562, "learning_rate": 8e-05, "loss": 1.5851, "step": 7997 }, { "epoch": 1.759265368965138, "grad_norm": 0.3328472077846527, "learning_rate": 8e-05, "loss": 1.5275, "step": 7998 }, { "epoch": 1.7594853183767734, "grad_norm": 0.3481276333332062, "learning_rate": 8e-05, "loss": 1.6008, "step": 7999 }, { "epoch": 1.7597052677884086, "grad_norm": 0.35142120718955994, "learning_rate": 8e-05, "loss": 1.693, "step": 8000 }, { "epoch": 1.7599252172000441, "grad_norm": 0.3514852821826935, "learning_rate": 8e-05, "loss": 1.5638, "step": 8001 }, { "epoch": 1.7601451666116792, "grad_norm": 0.33791211247444153, "learning_rate": 8e-05, "loss": 1.5224, "step": 8002 }, { "epoch": 1.7603651160233147, "grad_norm": 0.35534292459487915, "learning_rate": 8e-05, "loss": 1.7591, "step": 8003 }, { "epoch": 1.76058506543495, "grad_norm": 0.3370087444782257, "learning_rate": 8e-05, "loss": 1.6265, "step": 8004 }, { "epoch": 1.7608050148465852, "grad_norm": 0.35399237275123596, "learning_rate": 8e-05, "loss": 1.5232, "step": 8005 }, { "epoch": 1.7610249642582207, "grad_norm": 0.3479229807853699, "learning_rate": 8e-05, "loss": 1.5244, "step": 8006 }, { "epoch": 1.7612449136698558, "grad_norm": 0.3437436819076538, "learning_rate": 8e-05, "loss": 1.4968, "step": 8007 }, { "epoch": 1.7614648630814913, "grad_norm": 0.33858802914619446, "learning_rate": 8e-05, "loss": 1.5781, "step": 8008 }, { "epoch": 1.7616848124931266, "grad_norm": 0.3602173924446106, "learning_rate": 8e-05, "loss": 1.6647, "step": 8009 }, { "epoch": 1.7619047619047619, "grad_norm": 0.35099318623542786, "learning_rate": 8e-05, "loss": 1.5881, "step": 8010 }, { "epoch": 1.7621247113163974, "grad_norm": 0.3687792122364044, "learning_rate": 8e-05, "loss": 1.5604, "step": 8011 }, { "epoch": 1.7623446607280324, "grad_norm": 0.34395793080329895, "learning_rate": 8e-05, "loss": 1.6128, "step": 8012 }, { "epoch": 1.762564610139668, "grad_norm": 0.3543858826160431, "learning_rate": 8e-05, "loss": 1.611, "step": 8013 }, { "epoch": 1.7627845595513032, "grad_norm": 0.3414851129055023, "learning_rate": 8e-05, "loss": 1.6549, "step": 8014 }, { "epoch": 1.7630045089629385, "grad_norm": 0.31122151017189026, "learning_rate": 8e-05, "loss": 1.2833, "step": 8015 }, { "epoch": 1.763224458374574, "grad_norm": 0.3702350854873657, "learning_rate": 8e-05, "loss": 1.4989, "step": 8016 }, { "epoch": 1.763444407786209, "grad_norm": 0.37367334961891174, "learning_rate": 8e-05, "loss": 1.6651, "step": 8017 }, { "epoch": 1.7636643571978445, "grad_norm": 0.36758720874786377, "learning_rate": 8e-05, "loss": 1.5746, "step": 8018 }, { "epoch": 1.7638843066094798, "grad_norm": 0.3585798740386963, "learning_rate": 8e-05, "loss": 1.6838, "step": 8019 }, { "epoch": 1.764104256021115, "grad_norm": 0.3457159101963043, "learning_rate": 8e-05, "loss": 1.5432, "step": 8020 }, { "epoch": 1.7643242054327506, "grad_norm": 0.3488755226135254, "learning_rate": 8e-05, "loss": 1.5595, "step": 8021 }, { "epoch": 1.7645441548443856, "grad_norm": 0.35188889503479004, "learning_rate": 8e-05, "loss": 1.6629, "step": 8022 }, { "epoch": 1.7647641042560211, "grad_norm": 0.3423612117767334, "learning_rate": 8e-05, "loss": 1.517, "step": 8023 }, { "epoch": 1.7649840536676564, "grad_norm": 0.34943726658821106, "learning_rate": 8e-05, "loss": 1.5551, "step": 8024 }, { "epoch": 1.7652040030792917, "grad_norm": 0.3582036793231964, "learning_rate": 8e-05, "loss": 1.4525, "step": 8025 }, { "epoch": 1.7654239524909272, "grad_norm": 0.33552539348602295, "learning_rate": 8e-05, "loss": 1.6239, "step": 8026 }, { "epoch": 1.7656439019025623, "grad_norm": 0.3742694854736328, "learning_rate": 8e-05, "loss": 1.666, "step": 8027 }, { "epoch": 1.7658638513141978, "grad_norm": 0.3681190013885498, "learning_rate": 8e-05, "loss": 1.6675, "step": 8028 }, { "epoch": 1.766083800725833, "grad_norm": 0.3518717288970947, "learning_rate": 8e-05, "loss": 1.6197, "step": 8029 }, { "epoch": 1.7663037501374683, "grad_norm": 0.3546901345252991, "learning_rate": 8e-05, "loss": 1.6014, "step": 8030 }, { "epoch": 1.7665236995491038, "grad_norm": 0.3454040288925171, "learning_rate": 8e-05, "loss": 1.6191, "step": 8031 }, { "epoch": 1.7667436489607389, "grad_norm": 0.33608338236808777, "learning_rate": 8e-05, "loss": 1.4878, "step": 8032 }, { "epoch": 1.7669635983723744, "grad_norm": 0.36491164565086365, "learning_rate": 8e-05, "loss": 1.6239, "step": 8033 }, { "epoch": 1.7671835477840097, "grad_norm": 0.3351801335811615, "learning_rate": 8e-05, "loss": 1.4885, "step": 8034 }, { "epoch": 1.767403497195645, "grad_norm": 0.3532918095588684, "learning_rate": 8e-05, "loss": 1.5422, "step": 8035 }, { "epoch": 1.7676234466072804, "grad_norm": 0.334595263004303, "learning_rate": 8e-05, "loss": 1.4714, "step": 8036 }, { "epoch": 1.7678433960189155, "grad_norm": 0.3361404240131378, "learning_rate": 8e-05, "loss": 1.6269, "step": 8037 }, { "epoch": 1.768063345430551, "grad_norm": 0.40012240409851074, "learning_rate": 8e-05, "loss": 1.5772, "step": 8038 }, { "epoch": 1.7682832948421863, "grad_norm": 0.3543211817741394, "learning_rate": 8e-05, "loss": 1.7284, "step": 8039 }, { "epoch": 1.7685032442538215, "grad_norm": 0.3404897451400757, "learning_rate": 8e-05, "loss": 1.5345, "step": 8040 }, { "epoch": 1.768723193665457, "grad_norm": 0.37472373247146606, "learning_rate": 8e-05, "loss": 1.623, "step": 8041 }, { "epoch": 1.768943143077092, "grad_norm": 0.35402846336364746, "learning_rate": 8e-05, "loss": 1.5517, "step": 8042 }, { "epoch": 1.7691630924887276, "grad_norm": 0.3459567725658417, "learning_rate": 8e-05, "loss": 1.5717, "step": 8043 }, { "epoch": 1.7693830419003629, "grad_norm": 0.34237217903137207, "learning_rate": 8e-05, "loss": 1.6492, "step": 8044 }, { "epoch": 1.7696029913119982, "grad_norm": 0.3574886918067932, "learning_rate": 8e-05, "loss": 1.6017, "step": 8045 }, { "epoch": 1.7698229407236337, "grad_norm": 0.3636225759983063, "learning_rate": 8e-05, "loss": 1.7509, "step": 8046 }, { "epoch": 1.770042890135269, "grad_norm": 0.3644945025444031, "learning_rate": 8e-05, "loss": 1.6899, "step": 8047 }, { "epoch": 1.7702628395469042, "grad_norm": 0.3546747863292694, "learning_rate": 8e-05, "loss": 1.7401, "step": 8048 }, { "epoch": 1.7704827889585395, "grad_norm": 0.37387987971305847, "learning_rate": 8e-05, "loss": 1.6926, "step": 8049 }, { "epoch": 1.7707027383701748, "grad_norm": 0.3461603820323944, "learning_rate": 8e-05, "loss": 1.5629, "step": 8050 }, { "epoch": 1.7709226877818103, "grad_norm": 0.3409280776977539, "learning_rate": 8e-05, "loss": 1.4185, "step": 8051 }, { "epoch": 1.7711426371934456, "grad_norm": 0.3512961268424988, "learning_rate": 8e-05, "loss": 1.6504, "step": 8052 }, { "epoch": 1.7713625866050808, "grad_norm": 0.33746880292892456, "learning_rate": 8e-05, "loss": 1.5784, "step": 8053 }, { "epoch": 1.7715825360167161, "grad_norm": 0.33945372700691223, "learning_rate": 8e-05, "loss": 1.5952, "step": 8054 }, { "epoch": 1.7718024854283514, "grad_norm": 0.35662370920181274, "learning_rate": 8e-05, "loss": 1.659, "step": 8055 }, { "epoch": 1.772022434839987, "grad_norm": 0.35356733202934265, "learning_rate": 8e-05, "loss": 1.6629, "step": 8056 }, { "epoch": 1.7722423842516222, "grad_norm": 0.3586866855621338, "learning_rate": 8e-05, "loss": 1.5733, "step": 8057 }, { "epoch": 1.7724623336632574, "grad_norm": 0.3347083032131195, "learning_rate": 8e-05, "loss": 1.5942, "step": 8058 }, { "epoch": 1.7726822830748927, "grad_norm": 0.34790411591529846, "learning_rate": 8e-05, "loss": 1.7314, "step": 8059 }, { "epoch": 1.772902232486528, "grad_norm": 0.35196614265441895, "learning_rate": 8e-05, "loss": 1.4948, "step": 8060 }, { "epoch": 1.7731221818981635, "grad_norm": 0.3387988209724426, "learning_rate": 8e-05, "loss": 1.517, "step": 8061 }, { "epoch": 1.7733421313097988, "grad_norm": 0.36483559012413025, "learning_rate": 8e-05, "loss": 1.6476, "step": 8062 }, { "epoch": 1.773562080721434, "grad_norm": 0.344097375869751, "learning_rate": 8e-05, "loss": 1.5468, "step": 8063 }, { "epoch": 1.7737820301330693, "grad_norm": 0.32822081446647644, "learning_rate": 8e-05, "loss": 1.4768, "step": 8064 }, { "epoch": 1.7740019795447046, "grad_norm": 0.38136056065559387, "learning_rate": 8e-05, "loss": 1.5841, "step": 8065 }, { "epoch": 1.7742219289563401, "grad_norm": 0.36675310134887695, "learning_rate": 8e-05, "loss": 1.603, "step": 8066 }, { "epoch": 1.7744418783679754, "grad_norm": 0.34042128920555115, "learning_rate": 8e-05, "loss": 1.5967, "step": 8067 }, { "epoch": 1.7746618277796107, "grad_norm": 0.384797602891922, "learning_rate": 8e-05, "loss": 1.6158, "step": 8068 }, { "epoch": 1.7748817771912462, "grad_norm": 0.3330877125263214, "learning_rate": 8e-05, "loss": 1.5132, "step": 8069 }, { "epoch": 1.7751017266028812, "grad_norm": 0.3313467502593994, "learning_rate": 8e-05, "loss": 1.5265, "step": 8070 }, { "epoch": 1.7753216760145167, "grad_norm": 0.35080575942993164, "learning_rate": 8e-05, "loss": 1.7868, "step": 8071 }, { "epoch": 1.775541625426152, "grad_norm": 0.3375006914138794, "learning_rate": 8e-05, "loss": 1.4599, "step": 8072 }, { "epoch": 1.7757615748377873, "grad_norm": 0.3376653492450714, "learning_rate": 8e-05, "loss": 1.4892, "step": 8073 }, { "epoch": 1.7759815242494228, "grad_norm": 0.35610392689704895, "learning_rate": 8e-05, "loss": 1.671, "step": 8074 }, { "epoch": 1.7762014736610579, "grad_norm": 0.3397158682346344, "learning_rate": 8e-05, "loss": 1.5751, "step": 8075 }, { "epoch": 1.7764214230726934, "grad_norm": 0.3595004081726074, "learning_rate": 8e-05, "loss": 1.6893, "step": 8076 }, { "epoch": 1.7766413724843286, "grad_norm": 0.34863248467445374, "learning_rate": 8e-05, "loss": 1.5046, "step": 8077 }, { "epoch": 1.776861321895964, "grad_norm": 0.3548707664012909, "learning_rate": 8e-05, "loss": 1.6235, "step": 8078 }, { "epoch": 1.7770812713075994, "grad_norm": 0.328831285238266, "learning_rate": 8e-05, "loss": 1.5677, "step": 8079 }, { "epoch": 1.7773012207192345, "grad_norm": 0.3427900969982147, "learning_rate": 8e-05, "loss": 1.6444, "step": 8080 }, { "epoch": 1.77752117013087, "grad_norm": 0.36054977774620056, "learning_rate": 8e-05, "loss": 1.6362, "step": 8081 }, { "epoch": 1.7777411195425052, "grad_norm": 0.3534294366836548, "learning_rate": 8e-05, "loss": 1.5537, "step": 8082 }, { "epoch": 1.7779610689541405, "grad_norm": 0.3543678820133209, "learning_rate": 8e-05, "loss": 1.4591, "step": 8083 }, { "epoch": 1.778181018365776, "grad_norm": 0.34642189741134644, "learning_rate": 8e-05, "loss": 1.4734, "step": 8084 }, { "epoch": 1.778400967777411, "grad_norm": 0.35649994015693665, "learning_rate": 8e-05, "loss": 1.6803, "step": 8085 }, { "epoch": 1.7786209171890466, "grad_norm": 0.3621672987937927, "learning_rate": 8e-05, "loss": 1.6456, "step": 8086 }, { "epoch": 1.7788408666006819, "grad_norm": 0.35645154118537903, "learning_rate": 8e-05, "loss": 1.5901, "step": 8087 }, { "epoch": 1.7790608160123171, "grad_norm": 0.3597817122936249, "learning_rate": 8e-05, "loss": 1.7081, "step": 8088 }, { "epoch": 1.7792807654239526, "grad_norm": 0.3771290183067322, "learning_rate": 8e-05, "loss": 1.728, "step": 8089 }, { "epoch": 1.7795007148355877, "grad_norm": 0.354893296957016, "learning_rate": 8e-05, "loss": 1.4311, "step": 8090 }, { "epoch": 1.7797206642472232, "grad_norm": 0.37802037596702576, "learning_rate": 8e-05, "loss": 1.7336, "step": 8091 }, { "epoch": 1.7799406136588585, "grad_norm": 0.33496230840682983, "learning_rate": 8e-05, "loss": 1.6348, "step": 8092 }, { "epoch": 1.7801605630704938, "grad_norm": 0.36919352412223816, "learning_rate": 8e-05, "loss": 1.5873, "step": 8093 }, { "epoch": 1.7803805124821293, "grad_norm": 0.3559930920600891, "learning_rate": 8e-05, "loss": 1.5446, "step": 8094 }, { "epoch": 1.7806004618937643, "grad_norm": 0.3303779363632202, "learning_rate": 8e-05, "loss": 1.4474, "step": 8095 }, { "epoch": 1.7808204113053998, "grad_norm": 0.32555273175239563, "learning_rate": 8e-05, "loss": 1.4045, "step": 8096 }, { "epoch": 1.781040360717035, "grad_norm": 0.35271722078323364, "learning_rate": 8e-05, "loss": 1.4691, "step": 8097 }, { "epoch": 1.7812603101286704, "grad_norm": 0.357705682516098, "learning_rate": 8e-05, "loss": 1.7419, "step": 8098 }, { "epoch": 1.7814802595403059, "grad_norm": 0.3428994119167328, "learning_rate": 8e-05, "loss": 1.6067, "step": 8099 }, { "epoch": 1.781700208951941, "grad_norm": 0.3833044171333313, "learning_rate": 8e-05, "loss": 1.6274, "step": 8100 }, { "epoch": 1.7819201583635764, "grad_norm": 0.34632837772369385, "learning_rate": 8e-05, "loss": 1.5447, "step": 8101 }, { "epoch": 1.7821401077752117, "grad_norm": 0.34765833616256714, "learning_rate": 8e-05, "loss": 1.629, "step": 8102 }, { "epoch": 1.782360057186847, "grad_norm": 0.36964571475982666, "learning_rate": 8e-05, "loss": 1.7431, "step": 8103 }, { "epoch": 1.7825800065984825, "grad_norm": 0.3623444139957428, "learning_rate": 8e-05, "loss": 1.6646, "step": 8104 }, { "epoch": 1.7827999560101175, "grad_norm": 0.3495367467403412, "learning_rate": 8e-05, "loss": 1.5699, "step": 8105 }, { "epoch": 1.783019905421753, "grad_norm": 0.35841959714889526, "learning_rate": 8e-05, "loss": 1.5794, "step": 8106 }, { "epoch": 1.7832398548333883, "grad_norm": 0.4081127345561981, "learning_rate": 8e-05, "loss": 1.5751, "step": 8107 }, { "epoch": 1.7834598042450236, "grad_norm": 0.3534809648990631, "learning_rate": 8e-05, "loss": 1.6001, "step": 8108 }, { "epoch": 1.783679753656659, "grad_norm": 0.3549629747867584, "learning_rate": 8e-05, "loss": 1.5385, "step": 8109 }, { "epoch": 1.7838997030682942, "grad_norm": 0.3566421866416931, "learning_rate": 8e-05, "loss": 1.6754, "step": 8110 }, { "epoch": 1.7841196524799297, "grad_norm": 0.39274877309799194, "learning_rate": 8e-05, "loss": 1.5284, "step": 8111 }, { "epoch": 1.784339601891565, "grad_norm": 0.3670341670513153, "learning_rate": 8e-05, "loss": 1.6774, "step": 8112 }, { "epoch": 1.7845595513032002, "grad_norm": 0.3517700433731079, "learning_rate": 8e-05, "loss": 1.6362, "step": 8113 }, { "epoch": 1.7847795007148357, "grad_norm": 0.3678564131259918, "learning_rate": 8e-05, "loss": 1.7288, "step": 8114 }, { "epoch": 1.7849994501264708, "grad_norm": 0.35685041546821594, "learning_rate": 8e-05, "loss": 1.58, "step": 8115 }, { "epoch": 1.7852193995381063, "grad_norm": 0.34022822976112366, "learning_rate": 8e-05, "loss": 1.5531, "step": 8116 }, { "epoch": 1.7854393489497415, "grad_norm": 0.33045637607574463, "learning_rate": 8e-05, "loss": 1.57, "step": 8117 }, { "epoch": 1.7856592983613768, "grad_norm": 0.33897387981414795, "learning_rate": 8e-05, "loss": 1.5388, "step": 8118 }, { "epoch": 1.7858792477730123, "grad_norm": 0.3615911304950714, "learning_rate": 8e-05, "loss": 1.7019, "step": 8119 }, { "epoch": 1.7860991971846474, "grad_norm": 0.3631792962551117, "learning_rate": 8e-05, "loss": 1.6709, "step": 8120 }, { "epoch": 1.7863191465962829, "grad_norm": 0.35865819454193115, "learning_rate": 8e-05, "loss": 1.6726, "step": 8121 }, { "epoch": 1.7865390960079182, "grad_norm": 0.35454097390174866, "learning_rate": 8e-05, "loss": 1.6295, "step": 8122 }, { "epoch": 1.7867590454195534, "grad_norm": 0.33873650431632996, "learning_rate": 8e-05, "loss": 1.5175, "step": 8123 }, { "epoch": 1.786978994831189, "grad_norm": 0.3657936751842499, "learning_rate": 8e-05, "loss": 1.6269, "step": 8124 }, { "epoch": 1.787198944242824, "grad_norm": 0.3451533913612366, "learning_rate": 8e-05, "loss": 1.6442, "step": 8125 }, { "epoch": 1.7874188936544595, "grad_norm": 0.3451402485370636, "learning_rate": 8e-05, "loss": 1.5606, "step": 8126 }, { "epoch": 1.7876388430660948, "grad_norm": 0.35906869173049927, "learning_rate": 8e-05, "loss": 1.6328, "step": 8127 }, { "epoch": 1.78785879247773, "grad_norm": 0.37079161405563354, "learning_rate": 8e-05, "loss": 1.5873, "step": 8128 }, { "epoch": 1.7880787418893656, "grad_norm": 0.373675137758255, "learning_rate": 8e-05, "loss": 1.6839, "step": 8129 }, { "epoch": 1.7882986913010006, "grad_norm": 0.3626086115837097, "learning_rate": 8e-05, "loss": 1.5876, "step": 8130 }, { "epoch": 1.788518640712636, "grad_norm": 0.34216952323913574, "learning_rate": 8e-05, "loss": 1.5923, "step": 8131 }, { "epoch": 1.7887385901242714, "grad_norm": 0.34041163325309753, "learning_rate": 8e-05, "loss": 1.7052, "step": 8132 }, { "epoch": 1.7889585395359067, "grad_norm": 0.37129124999046326, "learning_rate": 8e-05, "loss": 1.5355, "step": 8133 }, { "epoch": 1.7891784889475422, "grad_norm": 0.3701925575733185, "learning_rate": 8e-05, "loss": 1.6186, "step": 8134 }, { "epoch": 1.7893984383591772, "grad_norm": 0.3338787257671356, "learning_rate": 8e-05, "loss": 1.5322, "step": 8135 }, { "epoch": 1.7896183877708127, "grad_norm": 0.3690315783023834, "learning_rate": 8e-05, "loss": 1.5435, "step": 8136 }, { "epoch": 1.789838337182448, "grad_norm": 0.35232481360435486, "learning_rate": 8e-05, "loss": 1.7064, "step": 8137 }, { "epoch": 1.7900582865940833, "grad_norm": 0.3273647725582123, "learning_rate": 8e-05, "loss": 1.4681, "step": 8138 }, { "epoch": 1.7902782360057188, "grad_norm": 0.34732019901275635, "learning_rate": 8e-05, "loss": 1.5851, "step": 8139 }, { "epoch": 1.790498185417354, "grad_norm": 0.3419243395328522, "learning_rate": 8e-05, "loss": 1.619, "step": 8140 }, { "epoch": 1.7907181348289893, "grad_norm": 0.35172393918037415, "learning_rate": 8e-05, "loss": 1.543, "step": 8141 }, { "epoch": 1.7909380842406246, "grad_norm": 0.3520934283733368, "learning_rate": 8e-05, "loss": 1.9173, "step": 8142 }, { "epoch": 1.79115803365226, "grad_norm": 0.35150524973869324, "learning_rate": 8e-05, "loss": 1.6026, "step": 8143 }, { "epoch": 1.7913779830638954, "grad_norm": 0.3600110709667206, "learning_rate": 8e-05, "loss": 1.503, "step": 8144 }, { "epoch": 1.7915979324755307, "grad_norm": 0.3499416708946228, "learning_rate": 8e-05, "loss": 1.6172, "step": 8145 }, { "epoch": 1.791817881887166, "grad_norm": 0.3446977436542511, "learning_rate": 8e-05, "loss": 1.5097, "step": 8146 }, { "epoch": 1.7920378312988012, "grad_norm": 0.3440687954425812, "learning_rate": 8e-05, "loss": 1.7131, "step": 8147 }, { "epoch": 1.7922577807104365, "grad_norm": 0.35714083909988403, "learning_rate": 8e-05, "loss": 1.6734, "step": 8148 }, { "epoch": 1.792477730122072, "grad_norm": 0.3297533392906189, "learning_rate": 8e-05, "loss": 1.3956, "step": 8149 }, { "epoch": 1.7926976795337073, "grad_norm": 0.3855193555355072, "learning_rate": 8e-05, "loss": 1.8072, "step": 8150 }, { "epoch": 1.7929176289453426, "grad_norm": 0.33720481395721436, "learning_rate": 8e-05, "loss": 1.3651, "step": 8151 }, { "epoch": 1.7931375783569778, "grad_norm": 0.3372405469417572, "learning_rate": 8e-05, "loss": 1.5464, "step": 8152 }, { "epoch": 1.7933575277686131, "grad_norm": 0.3626309931278229, "learning_rate": 8e-05, "loss": 1.5514, "step": 8153 }, { "epoch": 1.7935774771802486, "grad_norm": 0.3606731593608856, "learning_rate": 8e-05, "loss": 1.6443, "step": 8154 }, { "epoch": 1.793797426591884, "grad_norm": 0.35035181045532227, "learning_rate": 8e-05, "loss": 1.5467, "step": 8155 }, { "epoch": 1.7940173760035192, "grad_norm": 0.350807249546051, "learning_rate": 8e-05, "loss": 1.5834, "step": 8156 }, { "epoch": 1.7942373254151545, "grad_norm": 0.35024294257164, "learning_rate": 8e-05, "loss": 1.6412, "step": 8157 }, { "epoch": 1.7944572748267897, "grad_norm": 0.343058317899704, "learning_rate": 8e-05, "loss": 1.5621, "step": 8158 }, { "epoch": 1.7946772242384252, "grad_norm": 0.3441781997680664, "learning_rate": 8e-05, "loss": 1.5798, "step": 8159 }, { "epoch": 1.7948971736500605, "grad_norm": 0.351632684469223, "learning_rate": 8e-05, "loss": 1.6452, "step": 8160 }, { "epoch": 1.7951171230616958, "grad_norm": 0.3445560038089752, "learning_rate": 8e-05, "loss": 1.4811, "step": 8161 }, { "epoch": 1.7953370724733313, "grad_norm": 0.33792176842689514, "learning_rate": 8e-05, "loss": 1.5497, "step": 8162 }, { "epoch": 1.7955570218849664, "grad_norm": 0.33720123767852783, "learning_rate": 8e-05, "loss": 1.5994, "step": 8163 }, { "epoch": 1.7957769712966019, "grad_norm": 0.37754809856414795, "learning_rate": 8e-05, "loss": 1.6607, "step": 8164 }, { "epoch": 1.7959969207082371, "grad_norm": 0.3573611080646515, "learning_rate": 8e-05, "loss": 1.6968, "step": 8165 }, { "epoch": 1.7962168701198724, "grad_norm": 0.3580203950405121, "learning_rate": 8e-05, "loss": 1.6288, "step": 8166 }, { "epoch": 1.796436819531508, "grad_norm": 0.3726503849029541, "learning_rate": 8e-05, "loss": 1.6133, "step": 8167 }, { "epoch": 1.796656768943143, "grad_norm": 0.3658290505409241, "learning_rate": 8e-05, "loss": 1.5187, "step": 8168 }, { "epoch": 1.7968767183547785, "grad_norm": 0.3336944282054901, "learning_rate": 8e-05, "loss": 1.5738, "step": 8169 }, { "epoch": 1.7970966677664137, "grad_norm": 0.36058011651039124, "learning_rate": 8e-05, "loss": 1.5903, "step": 8170 }, { "epoch": 1.797316617178049, "grad_norm": 0.34980934858322144, "learning_rate": 8e-05, "loss": 1.6129, "step": 8171 }, { "epoch": 1.7975365665896845, "grad_norm": 0.3781343102455139, "learning_rate": 8e-05, "loss": 1.671, "step": 8172 }, { "epoch": 1.7977565160013196, "grad_norm": 0.3407975435256958, "learning_rate": 8e-05, "loss": 1.5874, "step": 8173 }, { "epoch": 1.797976465412955, "grad_norm": 0.34981194138526917, "learning_rate": 8e-05, "loss": 1.6322, "step": 8174 }, { "epoch": 1.7981964148245904, "grad_norm": 0.34200000762939453, "learning_rate": 8e-05, "loss": 1.5096, "step": 8175 }, { "epoch": 1.7984163642362256, "grad_norm": 0.3573704659938812, "learning_rate": 8e-05, "loss": 1.6239, "step": 8176 }, { "epoch": 1.7986363136478611, "grad_norm": 0.3465389907360077, "learning_rate": 8e-05, "loss": 1.6189, "step": 8177 }, { "epoch": 1.7988562630594962, "grad_norm": 0.33864277601242065, "learning_rate": 8e-05, "loss": 1.5409, "step": 8178 }, { "epoch": 1.7990762124711317, "grad_norm": 0.38658738136291504, "learning_rate": 8e-05, "loss": 1.5845, "step": 8179 }, { "epoch": 1.799296161882767, "grad_norm": 0.3457169830799103, "learning_rate": 8e-05, "loss": 1.6767, "step": 8180 }, { "epoch": 1.7995161112944023, "grad_norm": 0.37256622314453125, "learning_rate": 8e-05, "loss": 1.6717, "step": 8181 }, { "epoch": 1.7997360607060378, "grad_norm": 0.3719678819179535, "learning_rate": 8e-05, "loss": 1.5068, "step": 8182 }, { "epoch": 1.7999560101176728, "grad_norm": 0.37221449613571167, "learning_rate": 8e-05, "loss": 1.6419, "step": 8183 }, { "epoch": 1.8001759595293083, "grad_norm": 0.3236304521560669, "learning_rate": 8e-05, "loss": 1.4504, "step": 8184 }, { "epoch": 1.8003959089409436, "grad_norm": 0.3499648869037628, "learning_rate": 8e-05, "loss": 1.6569, "step": 8185 }, { "epoch": 1.8006158583525789, "grad_norm": 0.3435712158679962, "learning_rate": 8e-05, "loss": 1.548, "step": 8186 }, { "epoch": 1.8008358077642144, "grad_norm": 0.3518131673336029, "learning_rate": 8e-05, "loss": 1.6137, "step": 8187 }, { "epoch": 1.8010557571758494, "grad_norm": 0.3352228105068207, "learning_rate": 8e-05, "loss": 1.4436, "step": 8188 }, { "epoch": 1.801275706587485, "grad_norm": 0.3626515865325928, "learning_rate": 8e-05, "loss": 1.49, "step": 8189 }, { "epoch": 1.8014956559991202, "grad_norm": 0.36127617955207825, "learning_rate": 8e-05, "loss": 1.5245, "step": 8190 }, { "epoch": 1.8017156054107555, "grad_norm": 0.3666304647922516, "learning_rate": 8e-05, "loss": 1.5528, "step": 8191 }, { "epoch": 1.801935554822391, "grad_norm": 0.34664225578308105, "learning_rate": 8e-05, "loss": 1.5523, "step": 8192 }, { "epoch": 1.802155504234026, "grad_norm": 0.33589839935302734, "learning_rate": 8e-05, "loss": 1.5904, "step": 8193 }, { "epoch": 1.8023754536456615, "grad_norm": 0.3904838562011719, "learning_rate": 8e-05, "loss": 1.6562, "step": 8194 }, { "epoch": 1.8025954030572968, "grad_norm": 0.3625240623950958, "learning_rate": 8e-05, "loss": 1.622, "step": 8195 }, { "epoch": 1.802815352468932, "grad_norm": 0.34606918692588806, "learning_rate": 8e-05, "loss": 1.602, "step": 8196 }, { "epoch": 1.8030353018805676, "grad_norm": 0.3433910012245178, "learning_rate": 8e-05, "loss": 1.599, "step": 8197 }, { "epoch": 1.8032552512922027, "grad_norm": 0.3680276870727539, "learning_rate": 8e-05, "loss": 1.6977, "step": 8198 }, { "epoch": 1.8034752007038382, "grad_norm": 0.3498164713382721, "learning_rate": 8e-05, "loss": 1.7549, "step": 8199 }, { "epoch": 1.8036951501154734, "grad_norm": 0.35044124722480774, "learning_rate": 8e-05, "loss": 1.5532, "step": 8200 }, { "epoch": 1.8039150995271087, "grad_norm": 0.35026785731315613, "learning_rate": 8e-05, "loss": 1.7101, "step": 8201 }, { "epoch": 1.8041350489387442, "grad_norm": 0.33114132285118103, "learning_rate": 8e-05, "loss": 1.5363, "step": 8202 }, { "epoch": 1.8043549983503793, "grad_norm": 0.3345635235309601, "learning_rate": 8e-05, "loss": 1.4541, "step": 8203 }, { "epoch": 1.8045749477620148, "grad_norm": 0.35965657234191895, "learning_rate": 8e-05, "loss": 1.5977, "step": 8204 }, { "epoch": 1.80479489717365, "grad_norm": 0.33824896812438965, "learning_rate": 8e-05, "loss": 1.6125, "step": 8205 }, { "epoch": 1.8050148465852853, "grad_norm": 0.37090617418289185, "learning_rate": 8e-05, "loss": 1.6909, "step": 8206 }, { "epoch": 1.8052347959969208, "grad_norm": 0.3435189127922058, "learning_rate": 8e-05, "loss": 1.6284, "step": 8207 }, { "epoch": 1.8054547454085559, "grad_norm": 0.35556021332740784, "learning_rate": 8e-05, "loss": 1.5904, "step": 8208 }, { "epoch": 1.8056746948201914, "grad_norm": 0.33825594186782837, "learning_rate": 8e-05, "loss": 1.44, "step": 8209 }, { "epoch": 1.8058946442318267, "grad_norm": 0.3580138683319092, "learning_rate": 8e-05, "loss": 1.6929, "step": 8210 }, { "epoch": 1.806114593643462, "grad_norm": 0.3479258120059967, "learning_rate": 8e-05, "loss": 1.6199, "step": 8211 }, { "epoch": 1.8063345430550974, "grad_norm": 0.36011970043182373, "learning_rate": 8e-05, "loss": 1.6507, "step": 8212 }, { "epoch": 1.8065544924667325, "grad_norm": 0.345497190952301, "learning_rate": 8e-05, "loss": 1.632, "step": 8213 }, { "epoch": 1.806774441878368, "grad_norm": 0.32173842191696167, "learning_rate": 8e-05, "loss": 1.5126, "step": 8214 }, { "epoch": 1.8069943912900033, "grad_norm": 0.3459157943725586, "learning_rate": 8e-05, "loss": 1.6586, "step": 8215 }, { "epoch": 1.8072143407016386, "grad_norm": 0.3527529835700989, "learning_rate": 8e-05, "loss": 1.6638, "step": 8216 }, { "epoch": 1.807434290113274, "grad_norm": 0.33832433819770813, "learning_rate": 8e-05, "loss": 1.5884, "step": 8217 }, { "epoch": 1.8076542395249091, "grad_norm": 0.35155513882637024, "learning_rate": 8e-05, "loss": 1.6707, "step": 8218 }, { "epoch": 1.8078741889365446, "grad_norm": 0.39134401082992554, "learning_rate": 8e-05, "loss": 1.5131, "step": 8219 }, { "epoch": 1.80809413834818, "grad_norm": 0.3564075231552124, "learning_rate": 8e-05, "loss": 1.5632, "step": 8220 }, { "epoch": 1.8083140877598152, "grad_norm": 0.3224605917930603, "learning_rate": 8e-05, "loss": 1.2072, "step": 8221 }, { "epoch": 1.8085340371714507, "grad_norm": 0.36345216631889343, "learning_rate": 8e-05, "loss": 1.6153, "step": 8222 }, { "epoch": 1.8087539865830857, "grad_norm": 0.34599027037620544, "learning_rate": 8e-05, "loss": 1.5766, "step": 8223 }, { "epoch": 1.8089739359947212, "grad_norm": 0.3726230561733246, "learning_rate": 8e-05, "loss": 1.538, "step": 8224 }, { "epoch": 1.8091938854063565, "grad_norm": 0.3748787045478821, "learning_rate": 8e-05, "loss": 1.5154, "step": 8225 }, { "epoch": 1.8094138348179918, "grad_norm": 0.35886263847351074, "learning_rate": 8e-05, "loss": 1.5472, "step": 8226 }, { "epoch": 1.8096337842296273, "grad_norm": 0.34070369601249695, "learning_rate": 8e-05, "loss": 1.5239, "step": 8227 }, { "epoch": 1.8098537336412626, "grad_norm": 0.3349750339984894, "learning_rate": 8e-05, "loss": 1.531, "step": 8228 }, { "epoch": 1.8100736830528978, "grad_norm": 0.35915419459342957, "learning_rate": 8e-05, "loss": 1.7034, "step": 8229 }, { "epoch": 1.8102936324645331, "grad_norm": 0.34955593943595886, "learning_rate": 8e-05, "loss": 1.5488, "step": 8230 }, { "epoch": 1.8105135818761684, "grad_norm": 0.3539005219936371, "learning_rate": 8e-05, "loss": 1.4317, "step": 8231 }, { "epoch": 1.810733531287804, "grad_norm": 0.33863013982772827, "learning_rate": 8e-05, "loss": 1.5846, "step": 8232 }, { "epoch": 1.8109534806994392, "grad_norm": 0.3372872471809387, "learning_rate": 8e-05, "loss": 1.5795, "step": 8233 }, { "epoch": 1.8111734301110745, "grad_norm": 0.3504577577114105, "learning_rate": 8e-05, "loss": 1.6269, "step": 8234 }, { "epoch": 1.8113933795227097, "grad_norm": 0.3850012719631195, "learning_rate": 8e-05, "loss": 1.6395, "step": 8235 }, { "epoch": 1.811613328934345, "grad_norm": 0.3358573615550995, "learning_rate": 8e-05, "loss": 1.5696, "step": 8236 }, { "epoch": 1.8118332783459805, "grad_norm": 0.3354940712451935, "learning_rate": 8e-05, "loss": 1.5422, "step": 8237 }, { "epoch": 1.8120532277576158, "grad_norm": 0.34727364778518677, "learning_rate": 8e-05, "loss": 1.5353, "step": 8238 }, { "epoch": 1.812273177169251, "grad_norm": 0.34442847967147827, "learning_rate": 8e-05, "loss": 1.5285, "step": 8239 }, { "epoch": 1.8124931265808863, "grad_norm": 0.36464986205101013, "learning_rate": 8e-05, "loss": 1.6581, "step": 8240 }, { "epoch": 1.8127130759925216, "grad_norm": 0.3483176827430725, "learning_rate": 8e-05, "loss": 1.6388, "step": 8241 }, { "epoch": 1.8129330254041571, "grad_norm": 0.3637672960758209, "learning_rate": 8e-05, "loss": 1.7044, "step": 8242 }, { "epoch": 1.8131529748157924, "grad_norm": 0.3719935417175293, "learning_rate": 8e-05, "loss": 1.6832, "step": 8243 }, { "epoch": 1.8133729242274277, "grad_norm": 0.34789058566093445, "learning_rate": 8e-05, "loss": 1.5002, "step": 8244 }, { "epoch": 1.813592873639063, "grad_norm": 0.3525252640247345, "learning_rate": 8e-05, "loss": 1.6459, "step": 8245 }, { "epoch": 1.8138128230506982, "grad_norm": 0.3516945242881775, "learning_rate": 8e-05, "loss": 1.6539, "step": 8246 }, { "epoch": 1.8140327724623337, "grad_norm": 0.3715672194957733, "learning_rate": 8e-05, "loss": 1.6142, "step": 8247 }, { "epoch": 1.814252721873969, "grad_norm": 0.3375639319419861, "learning_rate": 8e-05, "loss": 1.4715, "step": 8248 }, { "epoch": 1.8144726712856043, "grad_norm": 0.3464340567588806, "learning_rate": 8e-05, "loss": 1.5605, "step": 8249 }, { "epoch": 1.8146926206972398, "grad_norm": 0.3512703776359558, "learning_rate": 8e-05, "loss": 1.5991, "step": 8250 }, { "epoch": 1.8149125701088749, "grad_norm": 0.364594042301178, "learning_rate": 8e-05, "loss": 1.6571, "step": 8251 }, { "epoch": 1.8151325195205104, "grad_norm": 0.3518291711807251, "learning_rate": 8e-05, "loss": 1.4851, "step": 8252 }, { "epoch": 1.8153524689321456, "grad_norm": 0.3212834894657135, "learning_rate": 8e-05, "loss": 1.4744, "step": 8253 }, { "epoch": 1.815572418343781, "grad_norm": 0.35395127534866333, "learning_rate": 8e-05, "loss": 1.5564, "step": 8254 }, { "epoch": 1.8157923677554164, "grad_norm": 0.35152629017829895, "learning_rate": 8e-05, "loss": 1.5555, "step": 8255 }, { "epoch": 1.8160123171670515, "grad_norm": 0.3432895541191101, "learning_rate": 8e-05, "loss": 1.5812, "step": 8256 }, { "epoch": 1.816232266578687, "grad_norm": 0.3222370743751526, "learning_rate": 8e-05, "loss": 1.4308, "step": 8257 }, { "epoch": 1.8164522159903222, "grad_norm": 0.3510958254337311, "learning_rate": 8e-05, "loss": 1.5063, "step": 8258 }, { "epoch": 1.8166721654019575, "grad_norm": 0.35373151302337646, "learning_rate": 8e-05, "loss": 1.6669, "step": 8259 }, { "epoch": 1.816892114813593, "grad_norm": 0.365554541349411, "learning_rate": 8e-05, "loss": 1.61, "step": 8260 }, { "epoch": 1.817112064225228, "grad_norm": 0.34181681275367737, "learning_rate": 8e-05, "loss": 1.4908, "step": 8261 }, { "epoch": 1.8173320136368636, "grad_norm": 0.34299781918525696, "learning_rate": 8e-05, "loss": 1.5824, "step": 8262 }, { "epoch": 1.8175519630484989, "grad_norm": 0.3738696277141571, "learning_rate": 8e-05, "loss": 1.6323, "step": 8263 }, { "epoch": 1.8177719124601341, "grad_norm": 0.43398287892341614, "learning_rate": 8e-05, "loss": 1.7438, "step": 8264 }, { "epoch": 1.8179918618717696, "grad_norm": 0.3684593737125397, "learning_rate": 8e-05, "loss": 1.6542, "step": 8265 }, { "epoch": 1.8182118112834047, "grad_norm": 0.3527638912200928, "learning_rate": 8e-05, "loss": 1.6324, "step": 8266 }, { "epoch": 1.8184317606950402, "grad_norm": 0.3408486247062683, "learning_rate": 8e-05, "loss": 1.539, "step": 8267 }, { "epoch": 1.8186517101066755, "grad_norm": 0.3470326364040375, "learning_rate": 8e-05, "loss": 1.4483, "step": 8268 }, { "epoch": 1.8188716595183108, "grad_norm": 0.34465155005455017, "learning_rate": 8e-05, "loss": 1.5993, "step": 8269 }, { "epoch": 1.8190916089299463, "grad_norm": 0.3364056944847107, "learning_rate": 8e-05, "loss": 1.5699, "step": 8270 }, { "epoch": 1.8193115583415813, "grad_norm": 0.3137265145778656, "learning_rate": 8e-05, "loss": 1.4222, "step": 8271 }, { "epoch": 1.8195315077532168, "grad_norm": 0.36566510796546936, "learning_rate": 8e-05, "loss": 1.5685, "step": 8272 }, { "epoch": 1.819751457164852, "grad_norm": 0.3372780680656433, "learning_rate": 8e-05, "loss": 1.5109, "step": 8273 }, { "epoch": 1.8199714065764874, "grad_norm": 0.3310719430446625, "learning_rate": 8e-05, "loss": 1.4307, "step": 8274 }, { "epoch": 1.8201913559881229, "grad_norm": 0.3530833423137665, "learning_rate": 8e-05, "loss": 1.5872, "step": 8275 }, { "epoch": 1.820411305399758, "grad_norm": 0.45232364535331726, "learning_rate": 8e-05, "loss": 1.6084, "step": 8276 }, { "epoch": 1.8206312548113934, "grad_norm": 0.3725367784500122, "learning_rate": 8e-05, "loss": 1.7179, "step": 8277 }, { "epoch": 1.8208512042230287, "grad_norm": 0.34884992241859436, "learning_rate": 8e-05, "loss": 1.5803, "step": 8278 }, { "epoch": 1.821071153634664, "grad_norm": 0.34694337844848633, "learning_rate": 8e-05, "loss": 1.6167, "step": 8279 }, { "epoch": 1.8212911030462995, "grad_norm": 0.3709671199321747, "learning_rate": 8e-05, "loss": 1.6089, "step": 8280 }, { "epoch": 1.8215110524579345, "grad_norm": 0.34999990463256836, "learning_rate": 8e-05, "loss": 1.6284, "step": 8281 }, { "epoch": 1.82173100186957, "grad_norm": 0.3470289409160614, "learning_rate": 8e-05, "loss": 1.5762, "step": 8282 }, { "epoch": 1.8219509512812053, "grad_norm": 0.34903913736343384, "learning_rate": 8e-05, "loss": 1.6153, "step": 8283 }, { "epoch": 1.8221709006928406, "grad_norm": 0.34808164834976196, "learning_rate": 8e-05, "loss": 1.6297, "step": 8284 }, { "epoch": 1.822390850104476, "grad_norm": 0.35319873690605164, "learning_rate": 8e-05, "loss": 1.6105, "step": 8285 }, { "epoch": 1.8226107995161112, "grad_norm": 0.3589746356010437, "learning_rate": 8e-05, "loss": 1.6992, "step": 8286 }, { "epoch": 1.8228307489277467, "grad_norm": 0.3658294081687927, "learning_rate": 8e-05, "loss": 1.6534, "step": 8287 }, { "epoch": 1.823050698339382, "grad_norm": 0.3452360928058624, "learning_rate": 8e-05, "loss": 1.4777, "step": 8288 }, { "epoch": 1.8232706477510172, "grad_norm": 0.37405478954315186, "learning_rate": 8e-05, "loss": 1.7659, "step": 8289 }, { "epoch": 1.8234905971626527, "grad_norm": 0.36962729692459106, "learning_rate": 8e-05, "loss": 1.7898, "step": 8290 }, { "epoch": 1.8237105465742878, "grad_norm": 0.35337451100349426, "learning_rate": 8e-05, "loss": 1.5021, "step": 8291 }, { "epoch": 1.8239304959859233, "grad_norm": 0.38405507802963257, "learning_rate": 8e-05, "loss": 1.5007, "step": 8292 }, { "epoch": 1.8241504453975586, "grad_norm": 0.3619679808616638, "learning_rate": 8e-05, "loss": 1.6051, "step": 8293 }, { "epoch": 1.8243703948091938, "grad_norm": 0.344767689704895, "learning_rate": 8e-05, "loss": 1.5941, "step": 8294 }, { "epoch": 1.8245903442208293, "grad_norm": 0.345346599817276, "learning_rate": 8e-05, "loss": 1.6681, "step": 8295 }, { "epoch": 1.8248102936324644, "grad_norm": 0.35795947909355164, "learning_rate": 8e-05, "loss": 1.419, "step": 8296 }, { "epoch": 1.8250302430440999, "grad_norm": 0.3476424813270569, "learning_rate": 8e-05, "loss": 1.6046, "step": 8297 }, { "epoch": 1.8252501924557352, "grad_norm": 0.351480096578598, "learning_rate": 8e-05, "loss": 1.5367, "step": 8298 }, { "epoch": 1.8254701418673704, "grad_norm": 0.3470073342323303, "learning_rate": 8e-05, "loss": 1.7146, "step": 8299 }, { "epoch": 1.825690091279006, "grad_norm": 0.34598833322525024, "learning_rate": 8e-05, "loss": 1.5982, "step": 8300 }, { "epoch": 1.825910040690641, "grad_norm": 0.35829341411590576, "learning_rate": 8e-05, "loss": 1.6495, "step": 8301 }, { "epoch": 1.8261299901022765, "grad_norm": 0.33089494705200195, "learning_rate": 8e-05, "loss": 1.4132, "step": 8302 }, { "epoch": 1.8263499395139118, "grad_norm": 0.35650819540023804, "learning_rate": 8e-05, "loss": 1.5441, "step": 8303 }, { "epoch": 1.826569888925547, "grad_norm": 0.34123703837394714, "learning_rate": 8e-05, "loss": 1.5256, "step": 8304 }, { "epoch": 1.8267898383371826, "grad_norm": 0.32282060384750366, "learning_rate": 8e-05, "loss": 1.5695, "step": 8305 }, { "epoch": 1.8270097877488176, "grad_norm": 0.3419753313064575, "learning_rate": 8e-05, "loss": 1.5891, "step": 8306 }, { "epoch": 1.8272297371604531, "grad_norm": 0.36107558012008667, "learning_rate": 8e-05, "loss": 1.6801, "step": 8307 }, { "epoch": 1.8274496865720884, "grad_norm": 0.35754168033599854, "learning_rate": 8e-05, "loss": 1.6792, "step": 8308 }, { "epoch": 1.8276696359837237, "grad_norm": 0.3402301073074341, "learning_rate": 8e-05, "loss": 1.4676, "step": 8309 }, { "epoch": 1.8278895853953592, "grad_norm": 0.3603980839252472, "learning_rate": 8e-05, "loss": 1.6353, "step": 8310 }, { "epoch": 1.8281095348069942, "grad_norm": 0.36547985672950745, "learning_rate": 8e-05, "loss": 1.5092, "step": 8311 }, { "epoch": 1.8283294842186297, "grad_norm": 0.3341686427593231, "learning_rate": 8e-05, "loss": 1.3607, "step": 8312 }, { "epoch": 1.828549433630265, "grad_norm": 0.3557389974594116, "learning_rate": 8e-05, "loss": 1.6322, "step": 8313 }, { "epoch": 1.8287693830419003, "grad_norm": 0.360391765832901, "learning_rate": 8e-05, "loss": 1.6488, "step": 8314 }, { "epoch": 1.8289893324535358, "grad_norm": 0.36478474736213684, "learning_rate": 8e-05, "loss": 1.3947, "step": 8315 }, { "epoch": 1.829209281865171, "grad_norm": 0.37550467252731323, "learning_rate": 8e-05, "loss": 1.5873, "step": 8316 }, { "epoch": 1.8294292312768063, "grad_norm": 0.3222680687904358, "learning_rate": 8e-05, "loss": 1.5329, "step": 8317 }, { "epoch": 1.8296491806884416, "grad_norm": 0.35202357172966003, "learning_rate": 8e-05, "loss": 1.6021, "step": 8318 }, { "epoch": 1.829869130100077, "grad_norm": 0.34992989897727966, "learning_rate": 8e-05, "loss": 1.5304, "step": 8319 }, { "epoch": 1.8300890795117124, "grad_norm": 0.33819815516471863, "learning_rate": 8e-05, "loss": 1.6074, "step": 8320 }, { "epoch": 1.8303090289233477, "grad_norm": 0.3540601432323456, "learning_rate": 8e-05, "loss": 1.5534, "step": 8321 }, { "epoch": 1.830528978334983, "grad_norm": 0.3417036831378937, "learning_rate": 8e-05, "loss": 1.3579, "step": 8322 }, { "epoch": 1.8307489277466182, "grad_norm": 0.3631857633590698, "learning_rate": 8e-05, "loss": 1.5298, "step": 8323 }, { "epoch": 1.8309688771582535, "grad_norm": 0.3724561333656311, "learning_rate": 8e-05, "loss": 1.6591, "step": 8324 }, { "epoch": 1.831188826569889, "grad_norm": 0.3490220904350281, "learning_rate": 8e-05, "loss": 1.5922, "step": 8325 }, { "epoch": 1.8314087759815243, "grad_norm": 0.3588115870952606, "learning_rate": 8e-05, "loss": 1.7988, "step": 8326 }, { "epoch": 1.8316287253931596, "grad_norm": 0.338979035615921, "learning_rate": 8e-05, "loss": 1.6138, "step": 8327 }, { "epoch": 1.8318486748047949, "grad_norm": 0.3363072872161865, "learning_rate": 8e-05, "loss": 1.4369, "step": 8328 }, { "epoch": 1.8320686242164301, "grad_norm": 0.3327852189540863, "learning_rate": 8e-05, "loss": 1.5166, "step": 8329 }, { "epoch": 1.8322885736280656, "grad_norm": 0.35179466009140015, "learning_rate": 8e-05, "loss": 1.5264, "step": 8330 }, { "epoch": 1.832508523039701, "grad_norm": 0.3327235281467438, "learning_rate": 8e-05, "loss": 1.46, "step": 8331 }, { "epoch": 1.8327284724513362, "grad_norm": 0.3667081892490387, "learning_rate": 8e-05, "loss": 1.4085, "step": 8332 }, { "epoch": 1.8329484218629715, "grad_norm": 0.4030098021030426, "learning_rate": 8e-05, "loss": 1.7506, "step": 8333 }, { "epoch": 1.8331683712746067, "grad_norm": 0.3515334725379944, "learning_rate": 8e-05, "loss": 1.5155, "step": 8334 }, { "epoch": 1.8333883206862422, "grad_norm": 0.3440896272659302, "learning_rate": 8e-05, "loss": 1.5672, "step": 8335 }, { "epoch": 1.8336082700978775, "grad_norm": 0.3419632613658905, "learning_rate": 8e-05, "loss": 1.472, "step": 8336 }, { "epoch": 1.8338282195095128, "grad_norm": 0.35277101397514343, "learning_rate": 8e-05, "loss": 1.3416, "step": 8337 }, { "epoch": 1.834048168921148, "grad_norm": 0.34924304485321045, "learning_rate": 8e-05, "loss": 1.5696, "step": 8338 }, { "epoch": 1.8342681183327834, "grad_norm": 0.3409399092197418, "learning_rate": 8e-05, "loss": 1.5077, "step": 8339 }, { "epoch": 1.8344880677444189, "grad_norm": 0.35770556330680847, "learning_rate": 8e-05, "loss": 1.5756, "step": 8340 }, { "epoch": 1.8347080171560541, "grad_norm": 0.34413596987724304, "learning_rate": 8e-05, "loss": 1.554, "step": 8341 }, { "epoch": 1.8349279665676894, "grad_norm": 0.36171993613243103, "learning_rate": 8e-05, "loss": 1.6378, "step": 8342 }, { "epoch": 1.835147915979325, "grad_norm": 0.3530934154987335, "learning_rate": 8e-05, "loss": 1.6612, "step": 8343 }, { "epoch": 1.83536786539096, "grad_norm": 0.3402676582336426, "learning_rate": 8e-05, "loss": 1.5553, "step": 8344 }, { "epoch": 1.8355878148025955, "grad_norm": 0.345180869102478, "learning_rate": 8e-05, "loss": 1.49, "step": 8345 }, { "epoch": 1.8358077642142308, "grad_norm": 0.32836243510246277, "learning_rate": 8e-05, "loss": 1.4873, "step": 8346 }, { "epoch": 1.836027713625866, "grad_norm": 0.3626638650894165, "learning_rate": 8e-05, "loss": 1.7049, "step": 8347 }, { "epoch": 1.8362476630375015, "grad_norm": 0.3549340069293976, "learning_rate": 8e-05, "loss": 1.6584, "step": 8348 }, { "epoch": 1.8364676124491366, "grad_norm": 0.34377989172935486, "learning_rate": 8e-05, "loss": 1.6215, "step": 8349 }, { "epoch": 1.836687561860772, "grad_norm": 0.3681532144546509, "learning_rate": 8e-05, "loss": 1.546, "step": 8350 }, { "epoch": 1.8369075112724074, "grad_norm": 0.35074663162231445, "learning_rate": 8e-05, "loss": 1.607, "step": 8351 }, { "epoch": 1.8371274606840426, "grad_norm": 0.3562186062335968, "learning_rate": 8e-05, "loss": 1.6259, "step": 8352 }, { "epoch": 1.8373474100956781, "grad_norm": 0.33790117502212524, "learning_rate": 8e-05, "loss": 1.6089, "step": 8353 }, { "epoch": 1.8375673595073132, "grad_norm": 0.3405204117298126, "learning_rate": 8e-05, "loss": 1.6178, "step": 8354 }, { "epoch": 1.8377873089189487, "grad_norm": 0.3692281246185303, "learning_rate": 8e-05, "loss": 1.6234, "step": 8355 }, { "epoch": 1.838007258330584, "grad_norm": 0.3564271628856659, "learning_rate": 8e-05, "loss": 1.61, "step": 8356 }, { "epoch": 1.8382272077422193, "grad_norm": 0.35094380378723145, "learning_rate": 8e-05, "loss": 1.5371, "step": 8357 }, { "epoch": 1.8384471571538548, "grad_norm": 0.36122363805770874, "learning_rate": 8e-05, "loss": 1.535, "step": 8358 }, { "epoch": 1.8386671065654898, "grad_norm": 0.41289475560188293, "learning_rate": 8e-05, "loss": 1.7651, "step": 8359 }, { "epoch": 1.8388870559771253, "grad_norm": 0.37235742807388306, "learning_rate": 8e-05, "loss": 1.718, "step": 8360 }, { "epoch": 1.8391070053887606, "grad_norm": 0.3482290506362915, "learning_rate": 8e-05, "loss": 1.7004, "step": 8361 }, { "epoch": 1.8393269548003959, "grad_norm": 0.37005141377449036, "learning_rate": 8e-05, "loss": 1.6549, "step": 8362 }, { "epoch": 1.8395469042120314, "grad_norm": 0.3493669927120209, "learning_rate": 8e-05, "loss": 1.6487, "step": 8363 }, { "epoch": 1.8397668536236664, "grad_norm": 0.36097395420074463, "learning_rate": 8e-05, "loss": 1.7683, "step": 8364 }, { "epoch": 1.839986803035302, "grad_norm": 0.3424735963344574, "learning_rate": 8e-05, "loss": 1.6418, "step": 8365 }, { "epoch": 1.8402067524469372, "grad_norm": 0.3583991229534149, "learning_rate": 8e-05, "loss": 1.6797, "step": 8366 }, { "epoch": 1.8404267018585725, "grad_norm": 0.34763458371162415, "learning_rate": 8e-05, "loss": 1.656, "step": 8367 }, { "epoch": 1.840646651270208, "grad_norm": 0.3418714106082916, "learning_rate": 8e-05, "loss": 1.47, "step": 8368 }, { "epoch": 1.840866600681843, "grad_norm": 0.35786813497543335, "learning_rate": 8e-05, "loss": 1.6196, "step": 8369 }, { "epoch": 1.8410865500934785, "grad_norm": 0.36169207096099854, "learning_rate": 8e-05, "loss": 1.5582, "step": 8370 }, { "epoch": 1.8413064995051138, "grad_norm": 0.35233792662620544, "learning_rate": 8e-05, "loss": 1.622, "step": 8371 }, { "epoch": 1.841526448916749, "grad_norm": 0.354062557220459, "learning_rate": 8e-05, "loss": 1.5025, "step": 8372 }, { "epoch": 1.8417463983283846, "grad_norm": 0.3319818675518036, "learning_rate": 8e-05, "loss": 1.5458, "step": 8373 }, { "epoch": 1.8419663477400197, "grad_norm": 0.3574357032775879, "learning_rate": 8e-05, "loss": 1.6999, "step": 8374 }, { "epoch": 1.8421862971516552, "grad_norm": 0.3484939932823181, "learning_rate": 8e-05, "loss": 1.5633, "step": 8375 }, { "epoch": 1.8424062465632904, "grad_norm": 0.3441633880138397, "learning_rate": 8e-05, "loss": 1.5202, "step": 8376 }, { "epoch": 1.8426261959749257, "grad_norm": 0.3537423610687256, "learning_rate": 8e-05, "loss": 1.6012, "step": 8377 }, { "epoch": 1.8428461453865612, "grad_norm": 0.3613201975822449, "learning_rate": 8e-05, "loss": 1.7812, "step": 8378 }, { "epoch": 1.8430660947981963, "grad_norm": 0.33259281516075134, "learning_rate": 8e-05, "loss": 1.463, "step": 8379 }, { "epoch": 1.8432860442098318, "grad_norm": 0.35580378770828247, "learning_rate": 8e-05, "loss": 1.554, "step": 8380 }, { "epoch": 1.843505993621467, "grad_norm": 0.34884363412857056, "learning_rate": 8e-05, "loss": 1.5015, "step": 8381 }, { "epoch": 1.8437259430331023, "grad_norm": 0.36919304728507996, "learning_rate": 8e-05, "loss": 1.4272, "step": 8382 }, { "epoch": 1.8439458924447378, "grad_norm": 0.36319422721862793, "learning_rate": 8e-05, "loss": 1.6828, "step": 8383 }, { "epoch": 1.844165841856373, "grad_norm": 0.34961768984794617, "learning_rate": 8e-05, "loss": 1.6868, "step": 8384 }, { "epoch": 1.8443857912680084, "grad_norm": 0.33452945947647095, "learning_rate": 8e-05, "loss": 1.519, "step": 8385 }, { "epoch": 1.8446057406796437, "grad_norm": 0.3947720229625702, "learning_rate": 8e-05, "loss": 1.5814, "step": 8386 }, { "epoch": 1.844825690091279, "grad_norm": 0.3773600459098816, "learning_rate": 8e-05, "loss": 1.6952, "step": 8387 }, { "epoch": 1.8450456395029144, "grad_norm": 0.37697145342826843, "learning_rate": 8e-05, "loss": 1.6868, "step": 8388 }, { "epoch": 1.8452655889145495, "grad_norm": 0.36214807629585266, "learning_rate": 8e-05, "loss": 1.5867, "step": 8389 }, { "epoch": 1.845485538326185, "grad_norm": 0.3473251760005951, "learning_rate": 8e-05, "loss": 1.6244, "step": 8390 }, { "epoch": 1.8457054877378203, "grad_norm": 0.37740665674209595, "learning_rate": 8e-05, "loss": 1.6717, "step": 8391 }, { "epoch": 1.8459254371494556, "grad_norm": 0.35415226221084595, "learning_rate": 8e-05, "loss": 1.6187, "step": 8392 }, { "epoch": 1.846145386561091, "grad_norm": 0.3552388846874237, "learning_rate": 8e-05, "loss": 1.5858, "step": 8393 }, { "epoch": 1.8463653359727261, "grad_norm": 0.323958158493042, "learning_rate": 8e-05, "loss": 1.5107, "step": 8394 }, { "epoch": 1.8465852853843616, "grad_norm": 0.34967559576034546, "learning_rate": 8e-05, "loss": 1.6786, "step": 8395 }, { "epoch": 1.846805234795997, "grad_norm": 0.35034048557281494, "learning_rate": 8e-05, "loss": 1.462, "step": 8396 }, { "epoch": 1.8470251842076322, "grad_norm": 0.3417314887046814, "learning_rate": 8e-05, "loss": 1.4247, "step": 8397 }, { "epoch": 1.8472451336192677, "grad_norm": 0.33848315477371216, "learning_rate": 8e-05, "loss": 1.6053, "step": 8398 }, { "epoch": 1.8474650830309027, "grad_norm": 0.3443315327167511, "learning_rate": 8e-05, "loss": 1.6338, "step": 8399 }, { "epoch": 1.8476850324425382, "grad_norm": 0.3386288285255432, "learning_rate": 8e-05, "loss": 1.5581, "step": 8400 }, { "epoch": 1.8479049818541735, "grad_norm": 0.3426578640937805, "learning_rate": 8e-05, "loss": 1.518, "step": 8401 }, { "epoch": 1.8481249312658088, "grad_norm": 0.3465237021446228, "learning_rate": 8e-05, "loss": 1.589, "step": 8402 }, { "epoch": 1.8483448806774443, "grad_norm": 0.3414596915245056, "learning_rate": 8e-05, "loss": 1.5196, "step": 8403 }, { "epoch": 1.8485648300890793, "grad_norm": 0.34851619601249695, "learning_rate": 8e-05, "loss": 1.6131, "step": 8404 }, { "epoch": 1.8487847795007148, "grad_norm": 0.3423817753791809, "learning_rate": 8e-05, "loss": 1.604, "step": 8405 }, { "epoch": 1.8490047289123501, "grad_norm": 0.3538081645965576, "learning_rate": 8e-05, "loss": 1.6116, "step": 8406 }, { "epoch": 1.8492246783239854, "grad_norm": 0.33287760615348816, "learning_rate": 8e-05, "loss": 1.5312, "step": 8407 }, { "epoch": 1.849444627735621, "grad_norm": 0.3567682206630707, "learning_rate": 8e-05, "loss": 1.5719, "step": 8408 }, { "epoch": 1.8496645771472562, "grad_norm": 0.36918750405311584, "learning_rate": 8e-05, "loss": 1.7162, "step": 8409 }, { "epoch": 1.8498845265588915, "grad_norm": 0.35779884457588196, "learning_rate": 8e-05, "loss": 1.5599, "step": 8410 }, { "epoch": 1.8501044759705267, "grad_norm": 0.34980669617652893, "learning_rate": 8e-05, "loss": 1.601, "step": 8411 }, { "epoch": 1.850324425382162, "grad_norm": 0.335906982421875, "learning_rate": 8e-05, "loss": 1.5279, "step": 8412 }, { "epoch": 1.8505443747937975, "grad_norm": 0.3273554742336273, "learning_rate": 8e-05, "loss": 1.4935, "step": 8413 }, { "epoch": 1.8507643242054328, "grad_norm": 0.3821386396884918, "learning_rate": 8e-05, "loss": 1.6353, "step": 8414 }, { "epoch": 1.850984273617068, "grad_norm": 0.35810238122940063, "learning_rate": 8e-05, "loss": 1.5979, "step": 8415 }, { "epoch": 1.8512042230287034, "grad_norm": 0.3377922773361206, "learning_rate": 8e-05, "loss": 1.5148, "step": 8416 }, { "epoch": 1.8514241724403386, "grad_norm": 0.3543638586997986, "learning_rate": 8e-05, "loss": 1.5146, "step": 8417 }, { "epoch": 1.8516441218519741, "grad_norm": 0.33928385376930237, "learning_rate": 8e-05, "loss": 1.6919, "step": 8418 }, { "epoch": 1.8518640712636094, "grad_norm": 0.34730416536331177, "learning_rate": 8e-05, "loss": 1.674, "step": 8419 }, { "epoch": 1.8520840206752447, "grad_norm": 0.3518086373806, "learning_rate": 8e-05, "loss": 1.7031, "step": 8420 }, { "epoch": 1.85230397008688, "grad_norm": 0.33621442317962646, "learning_rate": 8e-05, "loss": 1.5779, "step": 8421 }, { "epoch": 1.8525239194985152, "grad_norm": 0.3391902446746826, "learning_rate": 8e-05, "loss": 1.6612, "step": 8422 }, { "epoch": 1.8527438689101507, "grad_norm": 0.34444528818130493, "learning_rate": 8e-05, "loss": 1.4529, "step": 8423 }, { "epoch": 1.852963818321786, "grad_norm": 0.3608117699623108, "learning_rate": 8e-05, "loss": 1.5223, "step": 8424 }, { "epoch": 1.8531837677334213, "grad_norm": 0.33459973335266113, "learning_rate": 8e-05, "loss": 1.6377, "step": 8425 }, { "epoch": 1.8534037171450566, "grad_norm": 0.3400326669216156, "learning_rate": 8e-05, "loss": 1.4499, "step": 8426 }, { "epoch": 1.8536236665566919, "grad_norm": 0.35244467854499817, "learning_rate": 8e-05, "loss": 1.5789, "step": 8427 }, { "epoch": 1.8538436159683274, "grad_norm": 0.334883451461792, "learning_rate": 8e-05, "loss": 1.6882, "step": 8428 }, { "epoch": 1.8540635653799626, "grad_norm": 0.33941981196403503, "learning_rate": 8e-05, "loss": 1.4803, "step": 8429 }, { "epoch": 1.854283514791598, "grad_norm": 0.33939656615257263, "learning_rate": 8e-05, "loss": 1.5856, "step": 8430 }, { "epoch": 1.8545034642032334, "grad_norm": 0.3765183091163635, "learning_rate": 8e-05, "loss": 1.6692, "step": 8431 }, { "epoch": 1.8547234136148685, "grad_norm": 0.3837161958217621, "learning_rate": 8e-05, "loss": 1.7826, "step": 8432 }, { "epoch": 1.854943363026504, "grad_norm": 0.36652496457099915, "learning_rate": 8e-05, "loss": 1.6782, "step": 8433 }, { "epoch": 1.8551633124381393, "grad_norm": 0.35756972432136536, "learning_rate": 8e-05, "loss": 1.6441, "step": 8434 }, { "epoch": 1.8553832618497745, "grad_norm": 0.3437850773334503, "learning_rate": 8e-05, "loss": 1.4964, "step": 8435 }, { "epoch": 1.85560321126141, "grad_norm": 0.3410530090332031, "learning_rate": 8e-05, "loss": 1.5131, "step": 8436 }, { "epoch": 1.855823160673045, "grad_norm": 0.3601214587688446, "learning_rate": 8e-05, "loss": 1.529, "step": 8437 }, { "epoch": 1.8560431100846806, "grad_norm": 0.36779680848121643, "learning_rate": 8e-05, "loss": 1.6782, "step": 8438 }, { "epoch": 1.8562630594963159, "grad_norm": 0.3402427136898041, "learning_rate": 8e-05, "loss": 1.5113, "step": 8439 }, { "epoch": 1.8564830089079511, "grad_norm": 0.35444116592407227, "learning_rate": 8e-05, "loss": 1.6017, "step": 8440 }, { "epoch": 1.8567029583195866, "grad_norm": 0.35217025876045227, "learning_rate": 8e-05, "loss": 1.6623, "step": 8441 }, { "epoch": 1.8569229077312217, "grad_norm": 0.3640884459018707, "learning_rate": 8e-05, "loss": 1.712, "step": 8442 }, { "epoch": 1.8571428571428572, "grad_norm": 0.34922075271606445, "learning_rate": 8e-05, "loss": 1.5352, "step": 8443 }, { "epoch": 1.8573628065544925, "grad_norm": 0.3609481751918793, "learning_rate": 8e-05, "loss": 1.7049, "step": 8444 }, { "epoch": 1.8575827559661278, "grad_norm": 0.35875123739242554, "learning_rate": 8e-05, "loss": 1.6438, "step": 8445 }, { "epoch": 1.8578027053777633, "grad_norm": 0.3470298945903778, "learning_rate": 8e-05, "loss": 1.5996, "step": 8446 }, { "epoch": 1.8580226547893983, "grad_norm": 0.34263935685157776, "learning_rate": 8e-05, "loss": 1.5216, "step": 8447 }, { "epoch": 1.8582426042010338, "grad_norm": 0.3381158411502838, "learning_rate": 8e-05, "loss": 1.4053, "step": 8448 }, { "epoch": 1.858462553612669, "grad_norm": 0.34235697984695435, "learning_rate": 8e-05, "loss": 1.5637, "step": 8449 }, { "epoch": 1.8586825030243044, "grad_norm": 0.3352741003036499, "learning_rate": 8e-05, "loss": 1.4549, "step": 8450 }, { "epoch": 1.8589024524359399, "grad_norm": 0.3474292457103729, "learning_rate": 8e-05, "loss": 1.6044, "step": 8451 }, { "epoch": 1.859122401847575, "grad_norm": 0.3575131595134735, "learning_rate": 8e-05, "loss": 1.6434, "step": 8452 }, { "epoch": 1.8593423512592104, "grad_norm": 0.3549056053161621, "learning_rate": 8e-05, "loss": 1.62, "step": 8453 }, { "epoch": 1.8595623006708457, "grad_norm": 0.3578340709209442, "learning_rate": 8e-05, "loss": 1.5763, "step": 8454 }, { "epoch": 1.859782250082481, "grad_norm": 0.36537405848503113, "learning_rate": 8e-05, "loss": 1.4271, "step": 8455 }, { "epoch": 1.8600021994941165, "grad_norm": 0.3517743647098541, "learning_rate": 8e-05, "loss": 1.6935, "step": 8456 }, { "epoch": 1.8602221489057515, "grad_norm": 0.3577127456665039, "learning_rate": 8e-05, "loss": 1.7733, "step": 8457 }, { "epoch": 1.860442098317387, "grad_norm": 0.3388976454734802, "learning_rate": 8e-05, "loss": 1.4849, "step": 8458 }, { "epoch": 1.8606620477290223, "grad_norm": 0.3498600721359253, "learning_rate": 8e-05, "loss": 1.6325, "step": 8459 }, { "epoch": 1.8608819971406576, "grad_norm": 0.372713178396225, "learning_rate": 8e-05, "loss": 1.7196, "step": 8460 }, { "epoch": 1.861101946552293, "grad_norm": 0.35683417320251465, "learning_rate": 8e-05, "loss": 1.6314, "step": 8461 }, { "epoch": 1.8613218959639282, "grad_norm": 0.3543721139431, "learning_rate": 8e-05, "loss": 1.643, "step": 8462 }, { "epoch": 1.8615418453755637, "grad_norm": 0.35761141777038574, "learning_rate": 8e-05, "loss": 1.4729, "step": 8463 }, { "epoch": 1.861761794787199, "grad_norm": 0.3634766638278961, "learning_rate": 8e-05, "loss": 1.4706, "step": 8464 }, { "epoch": 1.8619817441988342, "grad_norm": 0.3535664677619934, "learning_rate": 8e-05, "loss": 1.6509, "step": 8465 }, { "epoch": 1.8622016936104697, "grad_norm": 0.3835866153240204, "learning_rate": 8e-05, "loss": 1.5842, "step": 8466 }, { "epoch": 1.8624216430221048, "grad_norm": 0.3550441563129425, "learning_rate": 8e-05, "loss": 1.5724, "step": 8467 }, { "epoch": 1.8626415924337403, "grad_norm": 0.3427978754043579, "learning_rate": 8e-05, "loss": 1.6087, "step": 8468 }, { "epoch": 1.8628615418453756, "grad_norm": 0.3748904764652252, "learning_rate": 8e-05, "loss": 1.5408, "step": 8469 }, { "epoch": 1.8630814912570108, "grad_norm": 0.35588517785072327, "learning_rate": 8e-05, "loss": 1.6186, "step": 8470 }, { "epoch": 1.8633014406686463, "grad_norm": 0.34410360455513, "learning_rate": 8e-05, "loss": 1.4607, "step": 8471 }, { "epoch": 1.8635213900802814, "grad_norm": 0.32779136300086975, "learning_rate": 8e-05, "loss": 1.4233, "step": 8472 }, { "epoch": 1.863741339491917, "grad_norm": 0.33997806906700134, "learning_rate": 8e-05, "loss": 1.462, "step": 8473 }, { "epoch": 1.8639612889035522, "grad_norm": 0.38309547305107117, "learning_rate": 8e-05, "loss": 1.6032, "step": 8474 }, { "epoch": 1.8641812383151874, "grad_norm": 0.33410051465034485, "learning_rate": 8e-05, "loss": 1.5453, "step": 8475 }, { "epoch": 1.864401187726823, "grad_norm": 0.3369615375995636, "learning_rate": 8e-05, "loss": 1.5169, "step": 8476 }, { "epoch": 1.864621137138458, "grad_norm": 0.3326999545097351, "learning_rate": 8e-05, "loss": 1.4568, "step": 8477 }, { "epoch": 1.8648410865500935, "grad_norm": 0.42403680086135864, "learning_rate": 8e-05, "loss": 1.6062, "step": 8478 }, { "epoch": 1.8650610359617288, "grad_norm": 0.336761474609375, "learning_rate": 8e-05, "loss": 1.3741, "step": 8479 }, { "epoch": 1.865280985373364, "grad_norm": 0.3325824737548828, "learning_rate": 8e-05, "loss": 1.2364, "step": 8480 }, { "epoch": 1.8655009347849996, "grad_norm": 0.4072360098361969, "learning_rate": 8e-05, "loss": 1.5573, "step": 8481 }, { "epoch": 1.8657208841966346, "grad_norm": 0.3620467185974121, "learning_rate": 8e-05, "loss": 1.6348, "step": 8482 }, { "epoch": 1.8659408336082701, "grad_norm": 0.3749293088912964, "learning_rate": 8e-05, "loss": 1.6038, "step": 8483 }, { "epoch": 1.8661607830199054, "grad_norm": 0.37319445610046387, "learning_rate": 8e-05, "loss": 1.7189, "step": 8484 }, { "epoch": 1.8663807324315407, "grad_norm": 0.3366425335407257, "learning_rate": 8e-05, "loss": 1.6039, "step": 8485 }, { "epoch": 1.8666006818431762, "grad_norm": 0.34869202971458435, "learning_rate": 8e-05, "loss": 1.6221, "step": 8486 }, { "epoch": 1.8668206312548112, "grad_norm": 0.3563818335533142, "learning_rate": 8e-05, "loss": 1.5123, "step": 8487 }, { "epoch": 1.8670405806664467, "grad_norm": 0.334749311208725, "learning_rate": 8e-05, "loss": 1.5889, "step": 8488 }, { "epoch": 1.867260530078082, "grad_norm": 0.34603309631347656, "learning_rate": 8e-05, "loss": 1.6121, "step": 8489 }, { "epoch": 1.8674804794897173, "grad_norm": 0.35474324226379395, "learning_rate": 8e-05, "loss": 1.6158, "step": 8490 }, { "epoch": 1.8677004289013528, "grad_norm": 0.35138440132141113, "learning_rate": 8e-05, "loss": 1.5572, "step": 8491 }, { "epoch": 1.8679203783129879, "grad_norm": 0.35291755199432373, "learning_rate": 8e-05, "loss": 1.6454, "step": 8492 }, { "epoch": 1.8681403277246234, "grad_norm": 0.363491952419281, "learning_rate": 8e-05, "loss": 1.4923, "step": 8493 }, { "epoch": 1.8683602771362586, "grad_norm": 0.3499024510383606, "learning_rate": 8e-05, "loss": 1.8191, "step": 8494 }, { "epoch": 1.868580226547894, "grad_norm": 0.36235424876213074, "learning_rate": 8e-05, "loss": 1.5459, "step": 8495 }, { "epoch": 1.8688001759595294, "grad_norm": 0.3406020998954773, "learning_rate": 8e-05, "loss": 1.5474, "step": 8496 }, { "epoch": 1.8690201253711647, "grad_norm": 0.36117592453956604, "learning_rate": 8e-05, "loss": 1.5239, "step": 8497 }, { "epoch": 1.8692400747828, "grad_norm": 0.41369375586509705, "learning_rate": 8e-05, "loss": 1.5681, "step": 8498 }, { "epoch": 1.8694600241944352, "grad_norm": 0.38760435581207275, "learning_rate": 8e-05, "loss": 1.7111, "step": 8499 }, { "epoch": 1.8696799736060705, "grad_norm": 0.353497177362442, "learning_rate": 8e-05, "loss": 1.576, "step": 8500 }, { "epoch": 1.869899923017706, "grad_norm": 0.3522055447101593, "learning_rate": 8e-05, "loss": 1.574, "step": 8501 }, { "epoch": 1.8701198724293413, "grad_norm": 0.3667009770870209, "learning_rate": 8e-05, "loss": 1.5858, "step": 8502 }, { "epoch": 1.8703398218409766, "grad_norm": 0.34122756123542786, "learning_rate": 8e-05, "loss": 1.4553, "step": 8503 }, { "epoch": 1.8705597712526119, "grad_norm": 0.36050090193748474, "learning_rate": 8e-05, "loss": 1.596, "step": 8504 }, { "epoch": 1.8707797206642471, "grad_norm": 0.36429500579833984, "learning_rate": 8e-05, "loss": 1.7846, "step": 8505 }, { "epoch": 1.8709996700758826, "grad_norm": 0.32850897312164307, "learning_rate": 8e-05, "loss": 1.5278, "step": 8506 }, { "epoch": 1.871219619487518, "grad_norm": 0.34682387113571167, "learning_rate": 8e-05, "loss": 1.5063, "step": 8507 }, { "epoch": 1.8714395688991532, "grad_norm": 0.3568514585494995, "learning_rate": 8e-05, "loss": 1.6062, "step": 8508 }, { "epoch": 1.8716595183107885, "grad_norm": 0.34793561697006226, "learning_rate": 8e-05, "loss": 1.6021, "step": 8509 }, { "epoch": 1.8718794677224238, "grad_norm": 0.3552246391773224, "learning_rate": 8e-05, "loss": 1.5865, "step": 8510 }, { "epoch": 1.8720994171340593, "grad_norm": 0.3828712999820709, "learning_rate": 8e-05, "loss": 1.6171, "step": 8511 }, { "epoch": 1.8723193665456945, "grad_norm": 0.3443067967891693, "learning_rate": 8e-05, "loss": 1.5636, "step": 8512 }, { "epoch": 1.8725393159573298, "grad_norm": 0.36816492676734924, "learning_rate": 8e-05, "loss": 1.5434, "step": 8513 }, { "epoch": 1.872759265368965, "grad_norm": 0.33412396907806396, "learning_rate": 8e-05, "loss": 1.6277, "step": 8514 }, { "epoch": 1.8729792147806004, "grad_norm": 0.3468528389930725, "learning_rate": 8e-05, "loss": 1.6083, "step": 8515 }, { "epoch": 1.8731991641922359, "grad_norm": 0.3594553768634796, "learning_rate": 8e-05, "loss": 1.6019, "step": 8516 }, { "epoch": 1.8734191136038711, "grad_norm": 0.34333133697509766, "learning_rate": 8e-05, "loss": 1.5782, "step": 8517 }, { "epoch": 1.8736390630155064, "grad_norm": 0.3294256329536438, "learning_rate": 8e-05, "loss": 1.4445, "step": 8518 }, { "epoch": 1.873859012427142, "grad_norm": 0.3616041839122772, "learning_rate": 8e-05, "loss": 1.6183, "step": 8519 }, { "epoch": 1.874078961838777, "grad_norm": 0.3514867424964905, "learning_rate": 8e-05, "loss": 1.4677, "step": 8520 }, { "epoch": 1.8742989112504125, "grad_norm": 0.3949829339981079, "learning_rate": 8e-05, "loss": 1.6918, "step": 8521 }, { "epoch": 1.8745188606620478, "grad_norm": 0.34189262986183167, "learning_rate": 8e-05, "loss": 1.4775, "step": 8522 }, { "epoch": 1.874738810073683, "grad_norm": 0.3461339473724365, "learning_rate": 8e-05, "loss": 1.5828, "step": 8523 }, { "epoch": 1.8749587594853185, "grad_norm": 0.3491937518119812, "learning_rate": 8e-05, "loss": 1.61, "step": 8524 }, { "epoch": 1.8751787088969536, "grad_norm": 0.34880203008651733, "learning_rate": 8e-05, "loss": 1.5588, "step": 8525 }, { "epoch": 1.875398658308589, "grad_norm": 0.3660779595375061, "learning_rate": 8e-05, "loss": 1.6375, "step": 8526 }, { "epoch": 1.8756186077202244, "grad_norm": 0.3430674970149994, "learning_rate": 8e-05, "loss": 1.5792, "step": 8527 }, { "epoch": 1.8758385571318597, "grad_norm": 0.37628546357154846, "learning_rate": 8e-05, "loss": 1.6749, "step": 8528 }, { "epoch": 1.8760585065434952, "grad_norm": 0.34860697388648987, "learning_rate": 8e-05, "loss": 1.6322, "step": 8529 }, { "epoch": 1.8762784559551302, "grad_norm": 0.34227004647254944, "learning_rate": 8e-05, "loss": 1.5589, "step": 8530 }, { "epoch": 1.8764984053667657, "grad_norm": 0.35455766320228577, "learning_rate": 8e-05, "loss": 1.6745, "step": 8531 }, { "epoch": 1.876718354778401, "grad_norm": 0.3523593544960022, "learning_rate": 8e-05, "loss": 1.5214, "step": 8532 }, { "epoch": 1.8769383041900363, "grad_norm": 0.492441326379776, "learning_rate": 8e-05, "loss": 1.6382, "step": 8533 }, { "epoch": 1.8771582536016718, "grad_norm": 0.3621513247489929, "learning_rate": 8e-05, "loss": 1.6988, "step": 8534 }, { "epoch": 1.8773782030133068, "grad_norm": 0.3542983829975128, "learning_rate": 8e-05, "loss": 1.6435, "step": 8535 }, { "epoch": 1.8775981524249423, "grad_norm": 0.344458669424057, "learning_rate": 8e-05, "loss": 1.5803, "step": 8536 }, { "epoch": 1.8778181018365776, "grad_norm": 0.34372666478157043, "learning_rate": 8e-05, "loss": 1.4953, "step": 8537 }, { "epoch": 1.8780380512482129, "grad_norm": 0.3361336886882782, "learning_rate": 8e-05, "loss": 1.5304, "step": 8538 }, { "epoch": 1.8782580006598484, "grad_norm": 0.3336019515991211, "learning_rate": 8e-05, "loss": 1.4663, "step": 8539 }, { "epoch": 1.8784779500714834, "grad_norm": 0.3492833077907562, "learning_rate": 8e-05, "loss": 1.6027, "step": 8540 }, { "epoch": 1.878697899483119, "grad_norm": 0.3628414273262024, "learning_rate": 8e-05, "loss": 1.7667, "step": 8541 }, { "epoch": 1.8789178488947542, "grad_norm": 0.3744181990623474, "learning_rate": 8e-05, "loss": 1.6031, "step": 8542 }, { "epoch": 1.8791377983063895, "grad_norm": 0.39551621675491333, "learning_rate": 8e-05, "loss": 1.8239, "step": 8543 }, { "epoch": 1.879357747718025, "grad_norm": 0.357281893491745, "learning_rate": 8e-05, "loss": 1.5924, "step": 8544 }, { "epoch": 1.87957769712966, "grad_norm": 0.37327834963798523, "learning_rate": 8e-05, "loss": 1.6365, "step": 8545 }, { "epoch": 1.8797976465412956, "grad_norm": 0.351214736700058, "learning_rate": 8e-05, "loss": 1.6635, "step": 8546 }, { "epoch": 1.8800175959529308, "grad_norm": 0.32363608479499817, "learning_rate": 8e-05, "loss": 1.5169, "step": 8547 }, { "epoch": 1.880237545364566, "grad_norm": 0.36602672934532166, "learning_rate": 8e-05, "loss": 1.5976, "step": 8548 }, { "epoch": 1.8804574947762016, "grad_norm": 0.3557812571525574, "learning_rate": 8e-05, "loss": 1.4931, "step": 8549 }, { "epoch": 1.8806774441878367, "grad_norm": 0.354444295167923, "learning_rate": 8e-05, "loss": 1.5463, "step": 8550 }, { "epoch": 1.8808973935994722, "grad_norm": 0.3568587005138397, "learning_rate": 8e-05, "loss": 1.642, "step": 8551 }, { "epoch": 1.8811173430111074, "grad_norm": 0.3713572323322296, "learning_rate": 8e-05, "loss": 1.5233, "step": 8552 }, { "epoch": 1.8813372924227427, "grad_norm": 0.35463210940361023, "learning_rate": 8e-05, "loss": 1.6425, "step": 8553 }, { "epoch": 1.8815572418343782, "grad_norm": 0.351248174905777, "learning_rate": 8e-05, "loss": 1.5365, "step": 8554 }, { "epoch": 1.8817771912460133, "grad_norm": 0.34389594197273254, "learning_rate": 8e-05, "loss": 1.4396, "step": 8555 }, { "epoch": 1.8819971406576488, "grad_norm": 0.3744233548641205, "learning_rate": 8e-05, "loss": 1.7967, "step": 8556 }, { "epoch": 1.882217090069284, "grad_norm": 0.3955702483654022, "learning_rate": 8e-05, "loss": 1.4757, "step": 8557 }, { "epoch": 1.8824370394809193, "grad_norm": 0.33794745802879333, "learning_rate": 8e-05, "loss": 1.5085, "step": 8558 }, { "epoch": 1.8826569888925548, "grad_norm": 0.3992448151111603, "learning_rate": 8e-05, "loss": 1.5871, "step": 8559 }, { "epoch": 1.88287693830419, "grad_norm": 0.3436852693557739, "learning_rate": 8e-05, "loss": 1.6108, "step": 8560 }, { "epoch": 1.8830968877158254, "grad_norm": 0.3508576452732086, "learning_rate": 8e-05, "loss": 1.6185, "step": 8561 }, { "epoch": 1.8833168371274607, "grad_norm": 0.34975898265838623, "learning_rate": 8e-05, "loss": 1.5667, "step": 8562 }, { "epoch": 1.883536786539096, "grad_norm": 0.33988022804260254, "learning_rate": 8e-05, "loss": 1.6091, "step": 8563 }, { "epoch": 1.8837567359507315, "grad_norm": 0.3228721022605896, "learning_rate": 8e-05, "loss": 1.5321, "step": 8564 }, { "epoch": 1.8839766853623665, "grad_norm": 0.33702006936073303, "learning_rate": 8e-05, "loss": 1.5077, "step": 8565 }, { "epoch": 1.884196634774002, "grad_norm": 0.35996943712234497, "learning_rate": 8e-05, "loss": 1.6693, "step": 8566 }, { "epoch": 1.8844165841856373, "grad_norm": 0.36731666326522827, "learning_rate": 8e-05, "loss": 1.5912, "step": 8567 }, { "epoch": 1.8846365335972726, "grad_norm": 0.34897664189338684, "learning_rate": 8e-05, "loss": 1.5495, "step": 8568 }, { "epoch": 1.884856483008908, "grad_norm": 0.39743104577064514, "learning_rate": 8e-05, "loss": 1.5633, "step": 8569 }, { "epoch": 1.8850764324205431, "grad_norm": 0.38562965393066406, "learning_rate": 8e-05, "loss": 1.7301, "step": 8570 }, { "epoch": 1.8852963818321786, "grad_norm": 0.3501058518886566, "learning_rate": 8e-05, "loss": 1.5118, "step": 8571 }, { "epoch": 1.885516331243814, "grad_norm": 0.3627305030822754, "learning_rate": 8e-05, "loss": 1.6309, "step": 8572 }, { "epoch": 1.8857362806554492, "grad_norm": 0.34874409437179565, "learning_rate": 8e-05, "loss": 1.4213, "step": 8573 }, { "epoch": 1.8859562300670847, "grad_norm": 0.3639465868473053, "learning_rate": 8e-05, "loss": 1.6491, "step": 8574 }, { "epoch": 1.8861761794787197, "grad_norm": 0.35457155108451843, "learning_rate": 8e-05, "loss": 1.6138, "step": 8575 }, { "epoch": 1.8863961288903552, "grad_norm": 0.3285234272480011, "learning_rate": 8e-05, "loss": 1.4991, "step": 8576 }, { "epoch": 1.8866160783019905, "grad_norm": 0.3553959131240845, "learning_rate": 8e-05, "loss": 1.6009, "step": 8577 }, { "epoch": 1.8868360277136258, "grad_norm": 0.3670818507671356, "learning_rate": 8e-05, "loss": 1.6225, "step": 8578 }, { "epoch": 1.8870559771252613, "grad_norm": 0.3641408681869507, "learning_rate": 8e-05, "loss": 1.6197, "step": 8579 }, { "epoch": 1.8872759265368964, "grad_norm": 0.3552996516227722, "learning_rate": 8e-05, "loss": 1.5702, "step": 8580 }, { "epoch": 1.8874958759485319, "grad_norm": 0.33626431226730347, "learning_rate": 8e-05, "loss": 1.4903, "step": 8581 }, { "epoch": 1.8877158253601671, "grad_norm": 0.33589887619018555, "learning_rate": 8e-05, "loss": 1.6401, "step": 8582 }, { "epoch": 1.8879357747718024, "grad_norm": 0.3862936496734619, "learning_rate": 8e-05, "loss": 1.4927, "step": 8583 }, { "epoch": 1.888155724183438, "grad_norm": 0.3450622260570526, "learning_rate": 8e-05, "loss": 1.6604, "step": 8584 }, { "epoch": 1.888375673595073, "grad_norm": 0.3854106664657593, "learning_rate": 8e-05, "loss": 1.6399, "step": 8585 }, { "epoch": 1.8885956230067085, "grad_norm": 0.3533109426498413, "learning_rate": 8e-05, "loss": 1.5274, "step": 8586 }, { "epoch": 1.8888155724183437, "grad_norm": 0.34595951437950134, "learning_rate": 8e-05, "loss": 1.5852, "step": 8587 }, { "epoch": 1.889035521829979, "grad_norm": 0.34952613711357117, "learning_rate": 8e-05, "loss": 1.6089, "step": 8588 }, { "epoch": 1.8892554712416145, "grad_norm": 0.3480660617351532, "learning_rate": 8e-05, "loss": 1.6009, "step": 8589 }, { "epoch": 1.8894754206532498, "grad_norm": 0.36702290177345276, "learning_rate": 8e-05, "loss": 1.6206, "step": 8590 }, { "epoch": 1.889695370064885, "grad_norm": 0.3584274649620056, "learning_rate": 8e-05, "loss": 1.5373, "step": 8591 }, { "epoch": 1.8899153194765204, "grad_norm": 0.37661638855934143, "learning_rate": 8e-05, "loss": 1.5922, "step": 8592 }, { "epoch": 1.8901352688881556, "grad_norm": 0.34954559803009033, "learning_rate": 8e-05, "loss": 1.609, "step": 8593 }, { "epoch": 1.8903552182997911, "grad_norm": 0.36163270473480225, "learning_rate": 8e-05, "loss": 1.5985, "step": 8594 }, { "epoch": 1.8905751677114264, "grad_norm": 0.38490691781044006, "learning_rate": 8e-05, "loss": 1.6075, "step": 8595 }, { "epoch": 1.8907951171230617, "grad_norm": 0.3745092451572418, "learning_rate": 8e-05, "loss": 1.6487, "step": 8596 }, { "epoch": 1.891015066534697, "grad_norm": 0.34269043803215027, "learning_rate": 8e-05, "loss": 1.5478, "step": 8597 }, { "epoch": 1.8912350159463323, "grad_norm": 0.34895211458206177, "learning_rate": 8e-05, "loss": 1.6257, "step": 8598 }, { "epoch": 1.8914549653579678, "grad_norm": 0.360059529542923, "learning_rate": 8e-05, "loss": 1.7074, "step": 8599 }, { "epoch": 1.891674914769603, "grad_norm": 0.35238057374954224, "learning_rate": 8e-05, "loss": 1.6196, "step": 8600 }, { "epoch": 1.8918948641812383, "grad_norm": 0.3613724410533905, "learning_rate": 8e-05, "loss": 1.6173, "step": 8601 }, { "epoch": 1.8921148135928736, "grad_norm": 0.36275872588157654, "learning_rate": 8e-05, "loss": 1.5362, "step": 8602 }, { "epoch": 1.8923347630045089, "grad_norm": 0.35288161039352417, "learning_rate": 8e-05, "loss": 1.6111, "step": 8603 }, { "epoch": 1.8925547124161444, "grad_norm": 0.35617268085479736, "learning_rate": 8e-05, "loss": 1.6401, "step": 8604 }, { "epoch": 1.8927746618277796, "grad_norm": 0.345429003238678, "learning_rate": 8e-05, "loss": 1.5566, "step": 8605 }, { "epoch": 1.892994611239415, "grad_norm": 0.338281512260437, "learning_rate": 8e-05, "loss": 1.6249, "step": 8606 }, { "epoch": 1.8932145606510502, "grad_norm": 0.3772258758544922, "learning_rate": 8e-05, "loss": 1.7104, "step": 8607 }, { "epoch": 1.8934345100626855, "grad_norm": 0.34460484981536865, "learning_rate": 8e-05, "loss": 1.5214, "step": 8608 }, { "epoch": 1.893654459474321, "grad_norm": 0.3546852171421051, "learning_rate": 8e-05, "loss": 1.6681, "step": 8609 }, { "epoch": 1.8938744088859563, "grad_norm": 0.3739391267299652, "learning_rate": 8e-05, "loss": 1.7089, "step": 8610 }, { "epoch": 1.8940943582975915, "grad_norm": 0.3596203923225403, "learning_rate": 8e-05, "loss": 1.542, "step": 8611 }, { "epoch": 1.894314307709227, "grad_norm": 0.33163151144981384, "learning_rate": 8e-05, "loss": 1.4362, "step": 8612 }, { "epoch": 1.894534257120862, "grad_norm": 0.3536735773086548, "learning_rate": 8e-05, "loss": 1.5168, "step": 8613 }, { "epoch": 1.8947542065324976, "grad_norm": 0.3259502649307251, "learning_rate": 8e-05, "loss": 1.478, "step": 8614 }, { "epoch": 1.8949741559441329, "grad_norm": 0.3487584590911865, "learning_rate": 8e-05, "loss": 1.6856, "step": 8615 }, { "epoch": 1.8951941053557682, "grad_norm": 0.35309258103370667, "learning_rate": 8e-05, "loss": 1.549, "step": 8616 }, { "epoch": 1.8954140547674037, "grad_norm": 0.35206395387649536, "learning_rate": 8e-05, "loss": 1.7442, "step": 8617 }, { "epoch": 1.8956340041790387, "grad_norm": 0.35160836577415466, "learning_rate": 8e-05, "loss": 1.5761, "step": 8618 }, { "epoch": 1.8958539535906742, "grad_norm": 0.36469635367393494, "learning_rate": 8e-05, "loss": 1.5606, "step": 8619 }, { "epoch": 1.8960739030023095, "grad_norm": 0.3267655670642853, "learning_rate": 8e-05, "loss": 1.4542, "step": 8620 }, { "epoch": 1.8962938524139448, "grad_norm": 0.34343937039375305, "learning_rate": 8e-05, "loss": 1.5402, "step": 8621 }, { "epoch": 1.8965138018255803, "grad_norm": 0.35604771971702576, "learning_rate": 8e-05, "loss": 1.6599, "step": 8622 }, { "epoch": 1.8967337512372153, "grad_norm": 0.3330933451652527, "learning_rate": 8e-05, "loss": 1.5093, "step": 8623 }, { "epoch": 1.8969537006488508, "grad_norm": 0.38253161311149597, "learning_rate": 8e-05, "loss": 1.6441, "step": 8624 }, { "epoch": 1.897173650060486, "grad_norm": 0.3369846045970917, "learning_rate": 8e-05, "loss": 1.5267, "step": 8625 }, { "epoch": 1.8973935994721214, "grad_norm": 0.3807891011238098, "learning_rate": 8e-05, "loss": 1.6107, "step": 8626 }, { "epoch": 1.8976135488837569, "grad_norm": 0.34208282828330994, "learning_rate": 8e-05, "loss": 1.6208, "step": 8627 }, { "epoch": 1.897833498295392, "grad_norm": 0.3943811357021332, "learning_rate": 8e-05, "loss": 1.6548, "step": 8628 }, { "epoch": 1.8980534477070274, "grad_norm": 0.35708415508270264, "learning_rate": 8e-05, "loss": 1.6658, "step": 8629 }, { "epoch": 1.8982733971186627, "grad_norm": 0.3386535048484802, "learning_rate": 8e-05, "loss": 1.4018, "step": 8630 }, { "epoch": 1.898493346530298, "grad_norm": 0.354107141494751, "learning_rate": 8e-05, "loss": 1.6707, "step": 8631 }, { "epoch": 1.8987132959419335, "grad_norm": 0.37325426936149597, "learning_rate": 8e-05, "loss": 1.6973, "step": 8632 }, { "epoch": 1.8989332453535686, "grad_norm": 0.3464842736721039, "learning_rate": 8e-05, "loss": 1.5627, "step": 8633 }, { "epoch": 1.899153194765204, "grad_norm": 0.3717272877693176, "learning_rate": 8e-05, "loss": 1.7123, "step": 8634 }, { "epoch": 1.8993731441768393, "grad_norm": 0.39987075328826904, "learning_rate": 8e-05, "loss": 1.5184, "step": 8635 }, { "epoch": 1.8995930935884746, "grad_norm": 0.35707584023475647, "learning_rate": 8e-05, "loss": 1.6008, "step": 8636 }, { "epoch": 1.8998130430001101, "grad_norm": 0.378002792596817, "learning_rate": 8e-05, "loss": 1.5845, "step": 8637 }, { "epoch": 1.9000329924117452, "grad_norm": 0.34597140550613403, "learning_rate": 8e-05, "loss": 1.6625, "step": 8638 }, { "epoch": 1.9002529418233807, "grad_norm": 0.36270326375961304, "learning_rate": 8e-05, "loss": 1.59, "step": 8639 }, { "epoch": 1.900472891235016, "grad_norm": 0.34567883610725403, "learning_rate": 8e-05, "loss": 1.6536, "step": 8640 }, { "epoch": 1.9006928406466512, "grad_norm": 0.34758204221725464, "learning_rate": 8e-05, "loss": 1.7142, "step": 8641 }, { "epoch": 1.9009127900582867, "grad_norm": 0.33344802260398865, "learning_rate": 8e-05, "loss": 1.4528, "step": 8642 }, { "epoch": 1.9011327394699218, "grad_norm": 0.3414006233215332, "learning_rate": 8e-05, "loss": 1.5573, "step": 8643 }, { "epoch": 1.9013526888815573, "grad_norm": 0.37247493863105774, "learning_rate": 8e-05, "loss": 1.7883, "step": 8644 }, { "epoch": 1.9015726382931926, "grad_norm": 0.3509793281555176, "learning_rate": 8e-05, "loss": 1.5279, "step": 8645 }, { "epoch": 1.9017925877048278, "grad_norm": 0.3342786431312561, "learning_rate": 8e-05, "loss": 1.5368, "step": 8646 }, { "epoch": 1.9020125371164633, "grad_norm": 0.3356292247772217, "learning_rate": 8e-05, "loss": 1.487, "step": 8647 }, { "epoch": 1.9022324865280984, "grad_norm": 0.36335641145706177, "learning_rate": 8e-05, "loss": 1.6907, "step": 8648 }, { "epoch": 1.902452435939734, "grad_norm": 0.379402756690979, "learning_rate": 8e-05, "loss": 1.6384, "step": 8649 }, { "epoch": 1.9026723853513692, "grad_norm": 0.3579609990119934, "learning_rate": 8e-05, "loss": 1.5097, "step": 8650 }, { "epoch": 1.9028923347630045, "grad_norm": 0.3324252665042877, "learning_rate": 8e-05, "loss": 1.3913, "step": 8651 }, { "epoch": 1.90311228417464, "grad_norm": 0.3281343877315521, "learning_rate": 8e-05, "loss": 1.3044, "step": 8652 }, { "epoch": 1.903332233586275, "grad_norm": 0.3636898398399353, "learning_rate": 8e-05, "loss": 1.6084, "step": 8653 }, { "epoch": 1.9035521829979105, "grad_norm": 0.3348677158355713, "learning_rate": 8e-05, "loss": 1.414, "step": 8654 }, { "epoch": 1.9037721324095458, "grad_norm": 0.3698630928993225, "learning_rate": 8e-05, "loss": 1.573, "step": 8655 }, { "epoch": 1.903992081821181, "grad_norm": 0.37396499514579773, "learning_rate": 8e-05, "loss": 1.5151, "step": 8656 }, { "epoch": 1.9042120312328166, "grad_norm": 0.32690924406051636, "learning_rate": 8e-05, "loss": 1.5667, "step": 8657 }, { "epoch": 1.9044319806444516, "grad_norm": 0.3584427833557129, "learning_rate": 8e-05, "loss": 1.5468, "step": 8658 }, { "epoch": 1.9046519300560871, "grad_norm": 0.3522416353225708, "learning_rate": 8e-05, "loss": 1.7147, "step": 8659 }, { "epoch": 1.9048718794677224, "grad_norm": 0.3747839629650116, "learning_rate": 8e-05, "loss": 1.589, "step": 8660 }, { "epoch": 1.9050918288793577, "grad_norm": 0.3373335599899292, "learning_rate": 8e-05, "loss": 1.6366, "step": 8661 }, { "epoch": 1.9053117782909932, "grad_norm": 0.3850070834159851, "learning_rate": 8e-05, "loss": 1.7934, "step": 8662 }, { "epoch": 1.9055317277026282, "grad_norm": 0.37318000197410583, "learning_rate": 8e-05, "loss": 1.636, "step": 8663 }, { "epoch": 1.9057516771142637, "grad_norm": 0.34374067187309265, "learning_rate": 8e-05, "loss": 1.5583, "step": 8664 }, { "epoch": 1.905971626525899, "grad_norm": 0.44519659876823425, "learning_rate": 8e-05, "loss": 1.706, "step": 8665 }, { "epoch": 1.9061915759375343, "grad_norm": 0.37146902084350586, "learning_rate": 8e-05, "loss": 1.7853, "step": 8666 }, { "epoch": 1.9064115253491698, "grad_norm": 0.35080650448799133, "learning_rate": 8e-05, "loss": 1.6352, "step": 8667 }, { "epoch": 1.9066314747608049, "grad_norm": 0.3496462404727936, "learning_rate": 8e-05, "loss": 1.5636, "step": 8668 }, { "epoch": 1.9068514241724404, "grad_norm": 0.3475762903690338, "learning_rate": 8e-05, "loss": 1.5742, "step": 8669 }, { "epoch": 1.9070713735840756, "grad_norm": 0.35934433341026306, "learning_rate": 8e-05, "loss": 1.6324, "step": 8670 }, { "epoch": 1.907291322995711, "grad_norm": 0.38186606764793396, "learning_rate": 8e-05, "loss": 1.4844, "step": 8671 }, { "epoch": 1.9075112724073464, "grad_norm": 0.37227699160575867, "learning_rate": 8e-05, "loss": 1.6032, "step": 8672 }, { "epoch": 1.9077312218189815, "grad_norm": 0.3417484164237976, "learning_rate": 8e-05, "loss": 1.7138, "step": 8673 }, { "epoch": 1.907951171230617, "grad_norm": 0.3453039824962616, "learning_rate": 8e-05, "loss": 1.6222, "step": 8674 }, { "epoch": 1.9081711206422522, "grad_norm": 0.34745290875434875, "learning_rate": 8e-05, "loss": 1.5409, "step": 8675 }, { "epoch": 1.9083910700538875, "grad_norm": 0.3643205463886261, "learning_rate": 8e-05, "loss": 1.7515, "step": 8676 }, { "epoch": 1.908611019465523, "grad_norm": 0.33868148922920227, "learning_rate": 8e-05, "loss": 1.375, "step": 8677 }, { "epoch": 1.9088309688771583, "grad_norm": 0.34891122579574585, "learning_rate": 8e-05, "loss": 1.7063, "step": 8678 }, { "epoch": 1.9090509182887936, "grad_norm": 0.3535442650318146, "learning_rate": 8e-05, "loss": 1.7298, "step": 8679 }, { "epoch": 1.9092708677004289, "grad_norm": 0.3414861261844635, "learning_rate": 8e-05, "loss": 1.5806, "step": 8680 }, { "epoch": 1.9094908171120641, "grad_norm": 0.35192784667015076, "learning_rate": 8e-05, "loss": 1.5707, "step": 8681 }, { "epoch": 1.9097107665236996, "grad_norm": 0.3467879891395569, "learning_rate": 8e-05, "loss": 1.6764, "step": 8682 }, { "epoch": 1.909930715935335, "grad_norm": 0.3183901906013489, "learning_rate": 8e-05, "loss": 1.3168, "step": 8683 }, { "epoch": 1.9101506653469702, "grad_norm": 0.3820887804031372, "learning_rate": 8e-05, "loss": 1.664, "step": 8684 }, { "epoch": 1.9103706147586055, "grad_norm": 0.333193302154541, "learning_rate": 8e-05, "loss": 1.4804, "step": 8685 }, { "epoch": 1.9105905641702408, "grad_norm": 0.33881473541259766, "learning_rate": 8e-05, "loss": 1.6089, "step": 8686 }, { "epoch": 1.9108105135818763, "grad_norm": 0.3485545516014099, "learning_rate": 8e-05, "loss": 1.7147, "step": 8687 }, { "epoch": 1.9110304629935115, "grad_norm": 0.31680867075920105, "learning_rate": 8e-05, "loss": 1.3723, "step": 8688 }, { "epoch": 1.9112504124051468, "grad_norm": 0.3740747272968292, "learning_rate": 8e-05, "loss": 1.5026, "step": 8689 }, { "epoch": 1.911470361816782, "grad_norm": 0.3407575488090515, "learning_rate": 8e-05, "loss": 1.427, "step": 8690 }, { "epoch": 1.9116903112284174, "grad_norm": 0.3769496977329254, "learning_rate": 8e-05, "loss": 1.6613, "step": 8691 }, { "epoch": 1.9119102606400529, "grad_norm": 0.3582764267921448, "learning_rate": 8e-05, "loss": 1.5206, "step": 8692 }, { "epoch": 1.9121302100516882, "grad_norm": 0.3435664474964142, "learning_rate": 8e-05, "loss": 1.3819, "step": 8693 }, { "epoch": 1.9123501594633234, "grad_norm": 0.3632771372795105, "learning_rate": 8e-05, "loss": 1.5858, "step": 8694 }, { "epoch": 1.9125701088749587, "grad_norm": 0.3435814082622528, "learning_rate": 8e-05, "loss": 1.491, "step": 8695 }, { "epoch": 1.912790058286594, "grad_norm": 0.3350723087787628, "learning_rate": 8e-05, "loss": 1.3703, "step": 8696 }, { "epoch": 1.9130100076982295, "grad_norm": 0.3454703092575073, "learning_rate": 8e-05, "loss": 1.5279, "step": 8697 }, { "epoch": 1.9132299571098648, "grad_norm": 0.3548125922679901, "learning_rate": 8e-05, "loss": 1.6911, "step": 8698 }, { "epoch": 1.9134499065215, "grad_norm": 0.3740186095237732, "learning_rate": 8e-05, "loss": 1.4774, "step": 8699 }, { "epoch": 1.9136698559331355, "grad_norm": 0.35770508646965027, "learning_rate": 8e-05, "loss": 1.4453, "step": 8700 }, { "epoch": 1.9138898053447706, "grad_norm": 0.34198448061943054, "learning_rate": 8e-05, "loss": 1.5171, "step": 8701 }, { "epoch": 1.914109754756406, "grad_norm": 0.3421904146671295, "learning_rate": 8e-05, "loss": 1.5062, "step": 8702 }, { "epoch": 1.9143297041680414, "grad_norm": 0.35132572054862976, "learning_rate": 8e-05, "loss": 1.6114, "step": 8703 }, { "epoch": 1.9145496535796767, "grad_norm": 0.3541097939014435, "learning_rate": 8e-05, "loss": 1.522, "step": 8704 }, { "epoch": 1.9147696029913122, "grad_norm": 0.36982080340385437, "learning_rate": 8e-05, "loss": 1.5674, "step": 8705 }, { "epoch": 1.9149895524029472, "grad_norm": 0.364587664604187, "learning_rate": 8e-05, "loss": 1.5552, "step": 8706 }, { "epoch": 1.9152095018145827, "grad_norm": 0.3559548854827881, "learning_rate": 8e-05, "loss": 1.654, "step": 8707 }, { "epoch": 1.915429451226218, "grad_norm": 0.3694092631340027, "learning_rate": 8e-05, "loss": 1.563, "step": 8708 }, { "epoch": 1.9156494006378533, "grad_norm": 0.36284688115119934, "learning_rate": 8e-05, "loss": 1.6472, "step": 8709 }, { "epoch": 1.9158693500494888, "grad_norm": 0.36165139079093933, "learning_rate": 8e-05, "loss": 1.7753, "step": 8710 }, { "epoch": 1.9160892994611238, "grad_norm": 0.3483758270740509, "learning_rate": 8e-05, "loss": 1.6365, "step": 8711 }, { "epoch": 1.9163092488727593, "grad_norm": 0.3502418100833893, "learning_rate": 8e-05, "loss": 1.62, "step": 8712 }, { "epoch": 1.9165291982843946, "grad_norm": 0.3685944974422455, "learning_rate": 8e-05, "loss": 1.4839, "step": 8713 }, { "epoch": 1.9167491476960299, "grad_norm": 0.366924524307251, "learning_rate": 8e-05, "loss": 1.6899, "step": 8714 }, { "epoch": 1.9169690971076654, "grad_norm": 0.3868931531906128, "learning_rate": 8e-05, "loss": 1.4292, "step": 8715 }, { "epoch": 1.9171890465193004, "grad_norm": 0.3586063086986542, "learning_rate": 8e-05, "loss": 1.5515, "step": 8716 }, { "epoch": 1.917408995930936, "grad_norm": 0.35634368658065796, "learning_rate": 8e-05, "loss": 1.5676, "step": 8717 }, { "epoch": 1.9176289453425712, "grad_norm": 0.3542625904083252, "learning_rate": 8e-05, "loss": 1.6377, "step": 8718 }, { "epoch": 1.9178488947542065, "grad_norm": 0.3478613495826721, "learning_rate": 8e-05, "loss": 1.5651, "step": 8719 }, { "epoch": 1.918068844165842, "grad_norm": 0.3452889025211334, "learning_rate": 8e-05, "loss": 1.4912, "step": 8720 }, { "epoch": 1.918288793577477, "grad_norm": 0.3466256558895111, "learning_rate": 8e-05, "loss": 1.5541, "step": 8721 }, { "epoch": 1.9185087429891126, "grad_norm": 0.39916518330574036, "learning_rate": 8e-05, "loss": 1.6252, "step": 8722 }, { "epoch": 1.9187286924007478, "grad_norm": 0.3480759561061859, "learning_rate": 8e-05, "loss": 1.5618, "step": 8723 }, { "epoch": 1.9189486418123831, "grad_norm": 0.38727644085884094, "learning_rate": 8e-05, "loss": 1.619, "step": 8724 }, { "epoch": 1.9191685912240186, "grad_norm": 0.32972970604896545, "learning_rate": 8e-05, "loss": 1.3434, "step": 8725 }, { "epoch": 1.9193885406356537, "grad_norm": 0.33868175745010376, "learning_rate": 8e-05, "loss": 1.534, "step": 8726 }, { "epoch": 1.9196084900472892, "grad_norm": 0.35998445749282837, "learning_rate": 8e-05, "loss": 1.7347, "step": 8727 }, { "epoch": 1.9198284394589245, "grad_norm": 0.3809187114238739, "learning_rate": 8e-05, "loss": 1.6792, "step": 8728 }, { "epoch": 1.9200483888705597, "grad_norm": 0.3639715611934662, "learning_rate": 8e-05, "loss": 1.546, "step": 8729 }, { "epoch": 1.9202683382821952, "grad_norm": 0.3677319586277008, "learning_rate": 8e-05, "loss": 1.6862, "step": 8730 }, { "epoch": 1.9204882876938303, "grad_norm": 0.35736680030822754, "learning_rate": 8e-05, "loss": 1.5797, "step": 8731 }, { "epoch": 1.9207082371054658, "grad_norm": 0.36226096749305725, "learning_rate": 8e-05, "loss": 1.7404, "step": 8732 }, { "epoch": 1.920928186517101, "grad_norm": 0.3513631820678711, "learning_rate": 8e-05, "loss": 1.602, "step": 8733 }, { "epoch": 1.9211481359287363, "grad_norm": 0.36688971519470215, "learning_rate": 8e-05, "loss": 1.6484, "step": 8734 }, { "epoch": 1.9213680853403718, "grad_norm": 0.32838132977485657, "learning_rate": 8e-05, "loss": 1.5634, "step": 8735 }, { "epoch": 1.921588034752007, "grad_norm": 0.3372204303741455, "learning_rate": 8e-05, "loss": 1.4833, "step": 8736 }, { "epoch": 1.9218079841636424, "grad_norm": 0.38286951184272766, "learning_rate": 8e-05, "loss": 1.6223, "step": 8737 }, { "epoch": 1.9220279335752777, "grad_norm": 0.35622796416282654, "learning_rate": 8e-05, "loss": 1.5554, "step": 8738 }, { "epoch": 1.922247882986913, "grad_norm": 0.3512731194496155, "learning_rate": 8e-05, "loss": 1.5964, "step": 8739 }, { "epoch": 1.9224678323985485, "grad_norm": 0.3412168622016907, "learning_rate": 8e-05, "loss": 1.5587, "step": 8740 }, { "epoch": 1.9226877818101835, "grad_norm": 0.3518720865249634, "learning_rate": 8e-05, "loss": 1.6271, "step": 8741 }, { "epoch": 1.922907731221819, "grad_norm": 0.3455604016780853, "learning_rate": 8e-05, "loss": 1.5507, "step": 8742 }, { "epoch": 1.9231276806334543, "grad_norm": 0.3779161274433136, "learning_rate": 8e-05, "loss": 1.6826, "step": 8743 }, { "epoch": 1.9233476300450896, "grad_norm": 0.37897393107414246, "learning_rate": 8e-05, "loss": 1.5618, "step": 8744 }, { "epoch": 1.923567579456725, "grad_norm": 0.39188459515571594, "learning_rate": 8e-05, "loss": 1.6629, "step": 8745 }, { "epoch": 1.9237875288683601, "grad_norm": 0.34960925579071045, "learning_rate": 8e-05, "loss": 1.6044, "step": 8746 }, { "epoch": 1.9240074782799956, "grad_norm": 0.34190958738327026, "learning_rate": 8e-05, "loss": 1.6165, "step": 8747 }, { "epoch": 1.924227427691631, "grad_norm": 0.33161982893943787, "learning_rate": 8e-05, "loss": 1.5823, "step": 8748 }, { "epoch": 1.9244473771032662, "grad_norm": 0.37069171667099, "learning_rate": 8e-05, "loss": 1.655, "step": 8749 }, { "epoch": 1.9246673265149017, "grad_norm": 0.3424525558948517, "learning_rate": 8e-05, "loss": 1.4061, "step": 8750 }, { "epoch": 1.9248872759265367, "grad_norm": 0.34674400091171265, "learning_rate": 8e-05, "loss": 1.5595, "step": 8751 }, { "epoch": 1.9251072253381722, "grad_norm": 0.37755855917930603, "learning_rate": 8e-05, "loss": 1.6209, "step": 8752 }, { "epoch": 1.9253271747498075, "grad_norm": 0.36165881156921387, "learning_rate": 8e-05, "loss": 1.6627, "step": 8753 }, { "epoch": 1.9255471241614428, "grad_norm": 0.38140466809272766, "learning_rate": 8e-05, "loss": 1.7284, "step": 8754 }, { "epoch": 1.9257670735730783, "grad_norm": 0.3842235803604126, "learning_rate": 8e-05, "loss": 1.5225, "step": 8755 }, { "epoch": 1.9259870229847134, "grad_norm": 0.34013113379478455, "learning_rate": 8e-05, "loss": 1.4652, "step": 8756 }, { "epoch": 1.9262069723963489, "grad_norm": 0.39697521924972534, "learning_rate": 8e-05, "loss": 1.6859, "step": 8757 }, { "epoch": 1.9264269218079841, "grad_norm": 0.329135924577713, "learning_rate": 8e-05, "loss": 1.3912, "step": 8758 }, { "epoch": 1.9266468712196194, "grad_norm": 0.3591449558734894, "learning_rate": 8e-05, "loss": 1.631, "step": 8759 }, { "epoch": 1.926866820631255, "grad_norm": 0.3444785177707672, "learning_rate": 8e-05, "loss": 1.5303, "step": 8760 }, { "epoch": 1.92708677004289, "grad_norm": 0.36487212777137756, "learning_rate": 8e-05, "loss": 1.71, "step": 8761 }, { "epoch": 1.9273067194545255, "grad_norm": 0.36711740493774414, "learning_rate": 8e-05, "loss": 1.709, "step": 8762 }, { "epoch": 1.9275266688661608, "grad_norm": 0.36211925745010376, "learning_rate": 8e-05, "loss": 1.6259, "step": 8763 }, { "epoch": 1.927746618277796, "grad_norm": 0.36453011631965637, "learning_rate": 8e-05, "loss": 1.6514, "step": 8764 }, { "epoch": 1.9279665676894315, "grad_norm": 0.35789453983306885, "learning_rate": 8e-05, "loss": 1.5673, "step": 8765 }, { "epoch": 1.9281865171010666, "grad_norm": 0.3368869721889496, "learning_rate": 8e-05, "loss": 1.5423, "step": 8766 }, { "epoch": 1.928406466512702, "grad_norm": 0.35390496253967285, "learning_rate": 8e-05, "loss": 1.4951, "step": 8767 }, { "epoch": 1.9286264159243374, "grad_norm": 0.3559854030609131, "learning_rate": 8e-05, "loss": 1.7192, "step": 8768 }, { "epoch": 1.9288463653359726, "grad_norm": 0.35452792048454285, "learning_rate": 8e-05, "loss": 1.5532, "step": 8769 }, { "epoch": 1.9290663147476081, "grad_norm": 0.342581570148468, "learning_rate": 8e-05, "loss": 1.4479, "step": 8770 }, { "epoch": 1.9292862641592434, "grad_norm": 0.34720492362976074, "learning_rate": 8e-05, "loss": 1.5465, "step": 8771 }, { "epoch": 1.9295062135708787, "grad_norm": 0.3399903178215027, "learning_rate": 8e-05, "loss": 1.523, "step": 8772 }, { "epoch": 1.929726162982514, "grad_norm": 0.3383594751358032, "learning_rate": 8e-05, "loss": 1.4639, "step": 8773 }, { "epoch": 1.9299461123941493, "grad_norm": 0.3522770404815674, "learning_rate": 8e-05, "loss": 1.5943, "step": 8774 }, { "epoch": 1.9301660618057848, "grad_norm": 0.3535779118537903, "learning_rate": 8e-05, "loss": 1.5827, "step": 8775 }, { "epoch": 1.93038601121742, "grad_norm": 0.3356887102127075, "learning_rate": 8e-05, "loss": 1.5109, "step": 8776 }, { "epoch": 1.9306059606290553, "grad_norm": 0.35572293400764465, "learning_rate": 8e-05, "loss": 1.6809, "step": 8777 }, { "epoch": 1.9308259100406906, "grad_norm": 0.3553368151187897, "learning_rate": 8e-05, "loss": 1.5061, "step": 8778 }, { "epoch": 1.9310458594523259, "grad_norm": 0.3398902118206024, "learning_rate": 8e-05, "loss": 1.4678, "step": 8779 }, { "epoch": 1.9312658088639614, "grad_norm": 0.37486907839775085, "learning_rate": 8e-05, "loss": 1.6223, "step": 8780 }, { "epoch": 1.9314857582755967, "grad_norm": 0.34998708963394165, "learning_rate": 8e-05, "loss": 1.6176, "step": 8781 }, { "epoch": 1.931705707687232, "grad_norm": 0.36233240365982056, "learning_rate": 8e-05, "loss": 1.6365, "step": 8782 }, { "epoch": 1.9319256570988672, "grad_norm": 0.34372398257255554, "learning_rate": 8e-05, "loss": 1.5734, "step": 8783 }, { "epoch": 1.9321456065105025, "grad_norm": 0.3303294777870178, "learning_rate": 8e-05, "loss": 1.4401, "step": 8784 }, { "epoch": 1.932365555922138, "grad_norm": 0.3654303252696991, "learning_rate": 8e-05, "loss": 1.6516, "step": 8785 }, { "epoch": 1.9325855053337733, "grad_norm": 0.36054620146751404, "learning_rate": 8e-05, "loss": 1.6347, "step": 8786 }, { "epoch": 1.9328054547454085, "grad_norm": 0.348164439201355, "learning_rate": 8e-05, "loss": 1.7076, "step": 8787 }, { "epoch": 1.9330254041570438, "grad_norm": 0.3481413424015045, "learning_rate": 8e-05, "loss": 1.5271, "step": 8788 }, { "epoch": 1.933245353568679, "grad_norm": 0.36737436056137085, "learning_rate": 8e-05, "loss": 1.7164, "step": 8789 }, { "epoch": 1.9334653029803146, "grad_norm": 0.39033088088035583, "learning_rate": 8e-05, "loss": 1.598, "step": 8790 }, { "epoch": 1.9336852523919499, "grad_norm": 0.3350699841976166, "learning_rate": 8e-05, "loss": 1.4909, "step": 8791 }, { "epoch": 1.9339052018035852, "grad_norm": 0.35445040464401245, "learning_rate": 8e-05, "loss": 1.5299, "step": 8792 }, { "epoch": 1.9341251512152207, "grad_norm": 0.35245200991630554, "learning_rate": 8e-05, "loss": 1.65, "step": 8793 }, { "epoch": 1.9343451006268557, "grad_norm": 0.38161107897758484, "learning_rate": 8e-05, "loss": 1.7746, "step": 8794 }, { "epoch": 1.9345650500384912, "grad_norm": 0.33682981133461, "learning_rate": 8e-05, "loss": 1.4582, "step": 8795 }, { "epoch": 1.9347849994501265, "grad_norm": 0.34766414761543274, "learning_rate": 8e-05, "loss": 1.5116, "step": 8796 }, { "epoch": 1.9350049488617618, "grad_norm": 0.3655867576599121, "learning_rate": 8e-05, "loss": 1.5886, "step": 8797 }, { "epoch": 1.9352248982733973, "grad_norm": 0.36127153038978577, "learning_rate": 8e-05, "loss": 1.6507, "step": 8798 }, { "epoch": 1.9354448476850323, "grad_norm": 0.3869102895259857, "learning_rate": 8e-05, "loss": 1.6695, "step": 8799 }, { "epoch": 1.9356647970966678, "grad_norm": 0.3515879511833191, "learning_rate": 8e-05, "loss": 1.6629, "step": 8800 }, { "epoch": 1.9358847465083031, "grad_norm": 0.3467347323894501, "learning_rate": 8e-05, "loss": 1.6429, "step": 8801 }, { "epoch": 1.9361046959199384, "grad_norm": 0.3430335819721222, "learning_rate": 8e-05, "loss": 1.5921, "step": 8802 }, { "epoch": 1.936324645331574, "grad_norm": 0.3464425802230835, "learning_rate": 8e-05, "loss": 1.4153, "step": 8803 }, { "epoch": 1.936544594743209, "grad_norm": 0.3412613868713379, "learning_rate": 8e-05, "loss": 1.6321, "step": 8804 }, { "epoch": 1.9367645441548444, "grad_norm": 0.35201138257980347, "learning_rate": 8e-05, "loss": 1.5685, "step": 8805 }, { "epoch": 1.9369844935664797, "grad_norm": 0.35568127036094666, "learning_rate": 8e-05, "loss": 1.6382, "step": 8806 }, { "epoch": 1.937204442978115, "grad_norm": 0.36014941334724426, "learning_rate": 8e-05, "loss": 1.5723, "step": 8807 }, { "epoch": 1.9374243923897505, "grad_norm": 0.3397732377052307, "learning_rate": 8e-05, "loss": 1.4525, "step": 8808 }, { "epoch": 1.9376443418013856, "grad_norm": 0.35272231698036194, "learning_rate": 8e-05, "loss": 1.5794, "step": 8809 }, { "epoch": 1.937864291213021, "grad_norm": 0.3433220088481903, "learning_rate": 8e-05, "loss": 1.4424, "step": 8810 }, { "epoch": 1.9380842406246563, "grad_norm": 0.33675339818000793, "learning_rate": 8e-05, "loss": 1.5232, "step": 8811 }, { "epoch": 1.9383041900362916, "grad_norm": 0.33594658970832825, "learning_rate": 8e-05, "loss": 1.4953, "step": 8812 }, { "epoch": 1.9385241394479271, "grad_norm": 0.33639612793922424, "learning_rate": 8e-05, "loss": 1.5065, "step": 8813 }, { "epoch": 1.9387440888595622, "grad_norm": 0.3377024531364441, "learning_rate": 8e-05, "loss": 1.5821, "step": 8814 }, { "epoch": 1.9389640382711977, "grad_norm": 0.3442724347114563, "learning_rate": 8e-05, "loss": 1.3899, "step": 8815 }, { "epoch": 1.939183987682833, "grad_norm": 0.3504984378814697, "learning_rate": 8e-05, "loss": 1.6184, "step": 8816 }, { "epoch": 1.9394039370944682, "grad_norm": 0.35575971007347107, "learning_rate": 8e-05, "loss": 1.6176, "step": 8817 }, { "epoch": 1.9396238865061037, "grad_norm": 0.3610246479511261, "learning_rate": 8e-05, "loss": 1.5432, "step": 8818 }, { "epoch": 1.9398438359177388, "grad_norm": 0.4106481373310089, "learning_rate": 8e-05, "loss": 1.6843, "step": 8819 }, { "epoch": 1.9400637853293743, "grad_norm": 0.33963558077812195, "learning_rate": 8e-05, "loss": 1.5106, "step": 8820 }, { "epoch": 1.9402837347410096, "grad_norm": 0.34201106429100037, "learning_rate": 8e-05, "loss": 1.4633, "step": 8821 }, { "epoch": 1.9405036841526448, "grad_norm": 0.3399169445037842, "learning_rate": 8e-05, "loss": 1.5428, "step": 8822 }, { "epoch": 1.9407236335642803, "grad_norm": 0.33397582173347473, "learning_rate": 8e-05, "loss": 1.385, "step": 8823 }, { "epoch": 1.9409435829759154, "grad_norm": 0.3645774722099304, "learning_rate": 8e-05, "loss": 1.5113, "step": 8824 }, { "epoch": 1.941163532387551, "grad_norm": 0.34826454520225525, "learning_rate": 8e-05, "loss": 1.6284, "step": 8825 }, { "epoch": 1.9413834817991862, "grad_norm": 0.3537718653678894, "learning_rate": 8e-05, "loss": 1.6687, "step": 8826 }, { "epoch": 1.9416034312108215, "grad_norm": 0.37240806221961975, "learning_rate": 8e-05, "loss": 1.5773, "step": 8827 }, { "epoch": 1.941823380622457, "grad_norm": 0.35009777545928955, "learning_rate": 8e-05, "loss": 1.5056, "step": 8828 }, { "epoch": 1.942043330034092, "grad_norm": 0.36187753081321716, "learning_rate": 8e-05, "loss": 1.4666, "step": 8829 }, { "epoch": 1.9422632794457275, "grad_norm": 0.3372364342212677, "learning_rate": 8e-05, "loss": 1.4852, "step": 8830 }, { "epoch": 1.9424832288573628, "grad_norm": 0.34617382287979126, "learning_rate": 8e-05, "loss": 1.4081, "step": 8831 }, { "epoch": 1.942703178268998, "grad_norm": 0.3661060631275177, "learning_rate": 8e-05, "loss": 1.651, "step": 8832 }, { "epoch": 1.9429231276806336, "grad_norm": 0.3770606815814972, "learning_rate": 8e-05, "loss": 1.6624, "step": 8833 }, { "epoch": 1.9431430770922686, "grad_norm": 0.3685109615325928, "learning_rate": 8e-05, "loss": 1.6692, "step": 8834 }, { "epoch": 1.9433630265039041, "grad_norm": 0.3720225393772125, "learning_rate": 8e-05, "loss": 1.5764, "step": 8835 }, { "epoch": 1.9435829759155394, "grad_norm": 0.3555021286010742, "learning_rate": 8e-05, "loss": 1.621, "step": 8836 }, { "epoch": 1.9438029253271747, "grad_norm": 0.3426794707775116, "learning_rate": 8e-05, "loss": 1.5658, "step": 8837 }, { "epoch": 1.9440228747388102, "grad_norm": 0.3800704777240753, "learning_rate": 8e-05, "loss": 1.703, "step": 8838 }, { "epoch": 1.9442428241504452, "grad_norm": 0.35194095969200134, "learning_rate": 8e-05, "loss": 1.6168, "step": 8839 }, { "epoch": 1.9444627735620807, "grad_norm": 0.3336625397205353, "learning_rate": 8e-05, "loss": 1.4526, "step": 8840 }, { "epoch": 1.944682722973716, "grad_norm": 0.3658522963523865, "learning_rate": 8e-05, "loss": 1.645, "step": 8841 }, { "epoch": 1.9449026723853513, "grad_norm": 0.3524208068847656, "learning_rate": 8e-05, "loss": 1.6243, "step": 8842 }, { "epoch": 1.9451226217969868, "grad_norm": 0.36973556876182556, "learning_rate": 8e-05, "loss": 1.5839, "step": 8843 }, { "epoch": 1.9453425712086219, "grad_norm": 0.37631791830062866, "learning_rate": 8e-05, "loss": 1.6555, "step": 8844 }, { "epoch": 1.9455625206202574, "grad_norm": 0.3489704728126526, "learning_rate": 8e-05, "loss": 1.6248, "step": 8845 }, { "epoch": 1.9457824700318926, "grad_norm": 0.3713797330856323, "learning_rate": 8e-05, "loss": 1.6978, "step": 8846 }, { "epoch": 1.946002419443528, "grad_norm": 0.35772866010665894, "learning_rate": 8e-05, "loss": 1.6395, "step": 8847 }, { "epoch": 1.9462223688551634, "grad_norm": 0.3457537591457367, "learning_rate": 8e-05, "loss": 1.5292, "step": 8848 }, { "epoch": 1.9464423182667985, "grad_norm": 0.35673823952674866, "learning_rate": 8e-05, "loss": 1.7022, "step": 8849 }, { "epoch": 1.946662267678434, "grad_norm": 0.35736164450645447, "learning_rate": 8e-05, "loss": 1.6324, "step": 8850 }, { "epoch": 1.9468822170900693, "grad_norm": 0.34501636028289795, "learning_rate": 8e-05, "loss": 1.4431, "step": 8851 }, { "epoch": 1.9471021665017045, "grad_norm": 0.3531106114387512, "learning_rate": 8e-05, "loss": 1.5135, "step": 8852 }, { "epoch": 1.94732211591334, "grad_norm": 0.3563367426395416, "learning_rate": 8e-05, "loss": 1.6354, "step": 8853 }, { "epoch": 1.947542065324975, "grad_norm": 0.3824024796485901, "learning_rate": 8e-05, "loss": 1.6109, "step": 8854 }, { "epoch": 1.9477620147366106, "grad_norm": 0.34939131140708923, "learning_rate": 8e-05, "loss": 1.5661, "step": 8855 }, { "epoch": 1.9479819641482459, "grad_norm": 0.359634667634964, "learning_rate": 8e-05, "loss": 1.6362, "step": 8856 }, { "epoch": 1.9482019135598811, "grad_norm": 0.3437758684158325, "learning_rate": 8e-05, "loss": 1.477, "step": 8857 }, { "epoch": 1.9484218629715166, "grad_norm": 0.3532315492630005, "learning_rate": 8e-05, "loss": 1.7416, "step": 8858 }, { "epoch": 1.948641812383152, "grad_norm": 0.3346468210220337, "learning_rate": 8e-05, "loss": 1.562, "step": 8859 }, { "epoch": 1.9488617617947872, "grad_norm": 0.3506799340248108, "learning_rate": 8e-05, "loss": 1.4366, "step": 8860 }, { "epoch": 1.9490817112064225, "grad_norm": 0.36108464002609253, "learning_rate": 8e-05, "loss": 1.5183, "step": 8861 }, { "epoch": 1.9493016606180578, "grad_norm": 0.3516443967819214, "learning_rate": 8e-05, "loss": 1.663, "step": 8862 }, { "epoch": 1.9495216100296933, "grad_norm": 0.34996840357780457, "learning_rate": 8e-05, "loss": 1.5962, "step": 8863 }, { "epoch": 1.9497415594413285, "grad_norm": 0.3563414514064789, "learning_rate": 8e-05, "loss": 1.7172, "step": 8864 }, { "epoch": 1.9499615088529638, "grad_norm": 0.3849164843559265, "learning_rate": 8e-05, "loss": 1.5325, "step": 8865 }, { "epoch": 1.950181458264599, "grad_norm": 0.356579452753067, "learning_rate": 8e-05, "loss": 1.6088, "step": 8866 }, { "epoch": 1.9504014076762344, "grad_norm": 0.3645315170288086, "learning_rate": 8e-05, "loss": 1.5713, "step": 8867 }, { "epoch": 1.9506213570878699, "grad_norm": 0.3508133590221405, "learning_rate": 8e-05, "loss": 1.4636, "step": 8868 }, { "epoch": 1.9508413064995052, "grad_norm": 0.34337541460990906, "learning_rate": 8e-05, "loss": 1.5926, "step": 8869 }, { "epoch": 1.9510612559111404, "grad_norm": 0.35989367961883545, "learning_rate": 8e-05, "loss": 1.744, "step": 8870 }, { "epoch": 1.9512812053227757, "grad_norm": 0.35032209753990173, "learning_rate": 8e-05, "loss": 1.6864, "step": 8871 }, { "epoch": 1.951501154734411, "grad_norm": 0.36909136176109314, "learning_rate": 8e-05, "loss": 1.6806, "step": 8872 }, { "epoch": 1.9517211041460465, "grad_norm": 0.3570422828197479, "learning_rate": 8e-05, "loss": 1.5776, "step": 8873 }, { "epoch": 1.9519410535576818, "grad_norm": 0.3341403305530548, "learning_rate": 8e-05, "loss": 1.5893, "step": 8874 }, { "epoch": 1.952161002969317, "grad_norm": 0.3531411588191986, "learning_rate": 8e-05, "loss": 1.6568, "step": 8875 }, { "epoch": 1.9523809523809523, "grad_norm": 0.35394975543022156, "learning_rate": 8e-05, "loss": 1.4787, "step": 8876 }, { "epoch": 1.9526009017925876, "grad_norm": 0.34714722633361816, "learning_rate": 8e-05, "loss": 1.5261, "step": 8877 }, { "epoch": 1.952820851204223, "grad_norm": 0.3464389145374298, "learning_rate": 8e-05, "loss": 1.5576, "step": 8878 }, { "epoch": 1.9530408006158584, "grad_norm": 0.359610915184021, "learning_rate": 8e-05, "loss": 1.7518, "step": 8879 }, { "epoch": 1.9532607500274937, "grad_norm": 0.3581307530403137, "learning_rate": 8e-05, "loss": 1.6343, "step": 8880 }, { "epoch": 1.9534806994391292, "grad_norm": 0.37795203924179077, "learning_rate": 8e-05, "loss": 1.4956, "step": 8881 }, { "epoch": 1.9537006488507642, "grad_norm": 0.3566480875015259, "learning_rate": 8e-05, "loss": 1.5359, "step": 8882 }, { "epoch": 1.9539205982623997, "grad_norm": 0.35219112038612366, "learning_rate": 8e-05, "loss": 1.6124, "step": 8883 }, { "epoch": 1.954140547674035, "grad_norm": 0.36871206760406494, "learning_rate": 8e-05, "loss": 1.6954, "step": 8884 }, { "epoch": 1.9543604970856703, "grad_norm": 0.3478564918041229, "learning_rate": 8e-05, "loss": 1.6179, "step": 8885 }, { "epoch": 1.9545804464973058, "grad_norm": 0.34233686327934265, "learning_rate": 8e-05, "loss": 1.5661, "step": 8886 }, { "epoch": 1.9548003959089408, "grad_norm": 0.33830469846725464, "learning_rate": 8e-05, "loss": 1.636, "step": 8887 }, { "epoch": 1.9550203453205763, "grad_norm": 0.3288985788822174, "learning_rate": 8e-05, "loss": 1.3433, "step": 8888 }, { "epoch": 1.9552402947322116, "grad_norm": 0.3736639618873596, "learning_rate": 8e-05, "loss": 1.7625, "step": 8889 }, { "epoch": 1.955460244143847, "grad_norm": 0.3774413764476776, "learning_rate": 8e-05, "loss": 1.833, "step": 8890 }, { "epoch": 1.9556801935554824, "grad_norm": 0.36018306016921997, "learning_rate": 8e-05, "loss": 1.595, "step": 8891 }, { "epoch": 1.9559001429671174, "grad_norm": 0.3760347068309784, "learning_rate": 8e-05, "loss": 1.5803, "step": 8892 }, { "epoch": 1.956120092378753, "grad_norm": 0.35958901047706604, "learning_rate": 8e-05, "loss": 1.6413, "step": 8893 }, { "epoch": 1.9563400417903882, "grad_norm": 0.36903706192970276, "learning_rate": 8e-05, "loss": 1.6235, "step": 8894 }, { "epoch": 1.9565599912020235, "grad_norm": 0.3471037745475769, "learning_rate": 8e-05, "loss": 1.6026, "step": 8895 }, { "epoch": 1.956779940613659, "grad_norm": 0.3501448631286621, "learning_rate": 8e-05, "loss": 1.6292, "step": 8896 }, { "epoch": 1.956999890025294, "grad_norm": 0.3638828694820404, "learning_rate": 8e-05, "loss": 1.538, "step": 8897 }, { "epoch": 1.9572198394369296, "grad_norm": 0.34630006551742554, "learning_rate": 8e-05, "loss": 1.6387, "step": 8898 }, { "epoch": 1.9574397888485648, "grad_norm": 0.37027230858802795, "learning_rate": 8e-05, "loss": 1.5745, "step": 8899 }, { "epoch": 1.9576597382602001, "grad_norm": 0.33498772978782654, "learning_rate": 8e-05, "loss": 1.4592, "step": 8900 }, { "epoch": 1.9578796876718356, "grad_norm": 0.3362079858779907, "learning_rate": 8e-05, "loss": 1.4979, "step": 8901 }, { "epoch": 1.9580996370834707, "grad_norm": 0.3289307951927185, "learning_rate": 8e-05, "loss": 1.4636, "step": 8902 }, { "epoch": 1.9583195864951062, "grad_norm": 0.35234934091567993, "learning_rate": 8e-05, "loss": 1.5402, "step": 8903 }, { "epoch": 1.9585395359067415, "grad_norm": 0.3415764272212982, "learning_rate": 8e-05, "loss": 1.4789, "step": 8904 }, { "epoch": 1.9587594853183767, "grad_norm": 0.3702290952205658, "learning_rate": 8e-05, "loss": 1.6578, "step": 8905 }, { "epoch": 1.9589794347300122, "grad_norm": 0.3454723060131073, "learning_rate": 8e-05, "loss": 1.5736, "step": 8906 }, { "epoch": 1.9591993841416473, "grad_norm": 0.3450675904750824, "learning_rate": 8e-05, "loss": 1.5237, "step": 8907 }, { "epoch": 1.9594193335532828, "grad_norm": 0.3511849641799927, "learning_rate": 8e-05, "loss": 1.5737, "step": 8908 }, { "epoch": 1.959639282964918, "grad_norm": 0.3570360541343689, "learning_rate": 8e-05, "loss": 1.5568, "step": 8909 }, { "epoch": 1.9598592323765534, "grad_norm": 0.3513273000717163, "learning_rate": 8e-05, "loss": 1.6082, "step": 8910 }, { "epoch": 1.9600791817881889, "grad_norm": 0.3458811938762665, "learning_rate": 8e-05, "loss": 1.6375, "step": 8911 }, { "epoch": 1.960299131199824, "grad_norm": 0.3500448763370514, "learning_rate": 8e-05, "loss": 1.5501, "step": 8912 }, { "epoch": 1.9605190806114594, "grad_norm": 0.350324809551239, "learning_rate": 8e-05, "loss": 1.6363, "step": 8913 }, { "epoch": 1.9607390300230947, "grad_norm": 0.36923840641975403, "learning_rate": 8e-05, "loss": 1.6274, "step": 8914 }, { "epoch": 1.96095897943473, "grad_norm": 0.42686665058135986, "learning_rate": 8e-05, "loss": 1.4855, "step": 8915 }, { "epoch": 1.9611789288463655, "grad_norm": 0.3762720823287964, "learning_rate": 8e-05, "loss": 1.5638, "step": 8916 }, { "epoch": 1.9613988782580005, "grad_norm": 0.33415770530700684, "learning_rate": 8e-05, "loss": 1.4201, "step": 8917 }, { "epoch": 1.961618827669636, "grad_norm": 0.35713276267051697, "learning_rate": 8e-05, "loss": 1.7104, "step": 8918 }, { "epoch": 1.9618387770812713, "grad_norm": 0.42678380012512207, "learning_rate": 8e-05, "loss": 1.7991, "step": 8919 }, { "epoch": 1.9620587264929066, "grad_norm": 0.34843993186950684, "learning_rate": 8e-05, "loss": 1.6887, "step": 8920 }, { "epoch": 1.962278675904542, "grad_norm": 0.34778910875320435, "learning_rate": 8e-05, "loss": 1.4583, "step": 8921 }, { "epoch": 1.9624986253161771, "grad_norm": 0.34595900774002075, "learning_rate": 8e-05, "loss": 1.5754, "step": 8922 }, { "epoch": 1.9627185747278126, "grad_norm": 0.35056477785110474, "learning_rate": 8e-05, "loss": 1.5783, "step": 8923 }, { "epoch": 1.962938524139448, "grad_norm": 0.38312003016471863, "learning_rate": 8e-05, "loss": 1.6431, "step": 8924 }, { "epoch": 1.9631584735510832, "grad_norm": 0.36121872067451477, "learning_rate": 8e-05, "loss": 1.6563, "step": 8925 }, { "epoch": 1.9633784229627187, "grad_norm": 0.34448957443237305, "learning_rate": 8e-05, "loss": 1.4965, "step": 8926 }, { "epoch": 1.9635983723743538, "grad_norm": 0.3497809171676636, "learning_rate": 8e-05, "loss": 1.5393, "step": 8927 }, { "epoch": 1.9638183217859893, "grad_norm": 0.3393605649471283, "learning_rate": 8e-05, "loss": 1.5682, "step": 8928 }, { "epoch": 1.9640382711976245, "grad_norm": 0.39003100991249084, "learning_rate": 8e-05, "loss": 1.656, "step": 8929 }, { "epoch": 1.9642582206092598, "grad_norm": 0.36270347237586975, "learning_rate": 8e-05, "loss": 1.6062, "step": 8930 }, { "epoch": 1.9644781700208953, "grad_norm": 0.3376460671424866, "learning_rate": 8e-05, "loss": 1.4635, "step": 8931 }, { "epoch": 1.9646981194325304, "grad_norm": 0.3696404695510864, "learning_rate": 8e-05, "loss": 1.6209, "step": 8932 }, { "epoch": 1.9649180688441659, "grad_norm": 0.39508286118507385, "learning_rate": 8e-05, "loss": 1.6133, "step": 8933 }, { "epoch": 1.9651380182558011, "grad_norm": 0.3448786437511444, "learning_rate": 8e-05, "loss": 1.4316, "step": 8934 }, { "epoch": 1.9653579676674364, "grad_norm": 0.3394922614097595, "learning_rate": 8e-05, "loss": 1.5091, "step": 8935 }, { "epoch": 1.965577917079072, "grad_norm": 0.33367565274238586, "learning_rate": 8e-05, "loss": 1.4062, "step": 8936 }, { "epoch": 1.965797866490707, "grad_norm": 0.34023424983024597, "learning_rate": 8e-05, "loss": 1.3978, "step": 8937 }, { "epoch": 1.9660178159023425, "grad_norm": 0.33490315079689026, "learning_rate": 8e-05, "loss": 1.6025, "step": 8938 }, { "epoch": 1.9662377653139778, "grad_norm": 0.3389790952205658, "learning_rate": 8e-05, "loss": 1.5986, "step": 8939 }, { "epoch": 1.966457714725613, "grad_norm": 0.3524009883403778, "learning_rate": 8e-05, "loss": 1.3635, "step": 8940 }, { "epoch": 1.9666776641372485, "grad_norm": 0.3672630488872528, "learning_rate": 8e-05, "loss": 1.6549, "step": 8941 }, { "epoch": 1.9668976135488836, "grad_norm": 0.4024285674095154, "learning_rate": 8e-05, "loss": 1.7017, "step": 8942 }, { "epoch": 1.967117562960519, "grad_norm": 0.35012489557266235, "learning_rate": 8e-05, "loss": 1.6988, "step": 8943 }, { "epoch": 1.9673375123721544, "grad_norm": 0.34132716059684753, "learning_rate": 8e-05, "loss": 1.6537, "step": 8944 }, { "epoch": 1.9675574617837897, "grad_norm": 0.3462089002132416, "learning_rate": 8e-05, "loss": 1.4864, "step": 8945 }, { "epoch": 1.9677774111954252, "grad_norm": 0.3421478569507599, "learning_rate": 8e-05, "loss": 1.6043, "step": 8946 }, { "epoch": 1.9679973606070604, "grad_norm": 0.341802716255188, "learning_rate": 8e-05, "loss": 1.6562, "step": 8947 }, { "epoch": 1.9682173100186957, "grad_norm": 0.34783053398132324, "learning_rate": 8e-05, "loss": 1.5333, "step": 8948 }, { "epoch": 1.968437259430331, "grad_norm": 0.3163313567638397, "learning_rate": 8e-05, "loss": 1.3533, "step": 8949 }, { "epoch": 1.9686572088419663, "grad_norm": 0.37208688259124756, "learning_rate": 8e-05, "loss": 1.6414, "step": 8950 }, { "epoch": 1.9688771582536018, "grad_norm": 0.3464539349079132, "learning_rate": 8e-05, "loss": 1.6135, "step": 8951 }, { "epoch": 1.969097107665237, "grad_norm": 0.39514413475990295, "learning_rate": 8e-05, "loss": 1.7043, "step": 8952 }, { "epoch": 1.9693170570768723, "grad_norm": 0.3525865077972412, "learning_rate": 8e-05, "loss": 1.5564, "step": 8953 }, { "epoch": 1.9695370064885076, "grad_norm": 0.34147584438323975, "learning_rate": 8e-05, "loss": 1.5415, "step": 8954 }, { "epoch": 1.9697569559001429, "grad_norm": 0.3425617218017578, "learning_rate": 8e-05, "loss": 1.5518, "step": 8955 }, { "epoch": 1.9699769053117784, "grad_norm": 0.3490561842918396, "learning_rate": 8e-05, "loss": 1.5366, "step": 8956 }, { "epoch": 1.9701968547234137, "grad_norm": 0.37528690695762634, "learning_rate": 8e-05, "loss": 1.5445, "step": 8957 }, { "epoch": 1.970416804135049, "grad_norm": 0.36405813694000244, "learning_rate": 8e-05, "loss": 1.7244, "step": 8958 }, { "epoch": 1.9706367535466842, "grad_norm": 0.32823044061660767, "learning_rate": 8e-05, "loss": 1.5277, "step": 8959 }, { "epoch": 1.9708567029583195, "grad_norm": 0.3704482913017273, "learning_rate": 8e-05, "loss": 1.6169, "step": 8960 }, { "epoch": 1.971076652369955, "grad_norm": 0.3801151514053345, "learning_rate": 8e-05, "loss": 1.6856, "step": 8961 }, { "epoch": 1.9712966017815903, "grad_norm": 0.3585366904735565, "learning_rate": 8e-05, "loss": 1.6012, "step": 8962 }, { "epoch": 1.9715165511932256, "grad_norm": 0.35014763474464417, "learning_rate": 8e-05, "loss": 1.4641, "step": 8963 }, { "epoch": 1.9717365006048608, "grad_norm": 0.34013253450393677, "learning_rate": 8e-05, "loss": 1.5441, "step": 8964 }, { "epoch": 1.971956450016496, "grad_norm": 0.3456547260284424, "learning_rate": 8e-05, "loss": 1.3901, "step": 8965 }, { "epoch": 1.9721763994281316, "grad_norm": 0.37640276551246643, "learning_rate": 8e-05, "loss": 1.5914, "step": 8966 }, { "epoch": 1.9723963488397669, "grad_norm": 0.33193156123161316, "learning_rate": 8e-05, "loss": 1.5918, "step": 8967 }, { "epoch": 1.9726162982514022, "grad_norm": 0.3583889305591583, "learning_rate": 8e-05, "loss": 1.7984, "step": 8968 }, { "epoch": 1.9728362476630377, "grad_norm": 0.4172826111316681, "learning_rate": 8e-05, "loss": 1.6304, "step": 8969 }, { "epoch": 1.9730561970746727, "grad_norm": 0.35078462958335876, "learning_rate": 8e-05, "loss": 1.5589, "step": 8970 }, { "epoch": 1.9732761464863082, "grad_norm": 0.3869638442993164, "learning_rate": 8e-05, "loss": 1.57, "step": 8971 }, { "epoch": 1.9734960958979435, "grad_norm": 0.37030133605003357, "learning_rate": 8e-05, "loss": 1.5808, "step": 8972 }, { "epoch": 1.9737160453095788, "grad_norm": 0.3652060925960541, "learning_rate": 8e-05, "loss": 1.5319, "step": 8973 }, { "epoch": 1.9739359947212143, "grad_norm": 0.3938333988189697, "learning_rate": 8e-05, "loss": 1.6726, "step": 8974 }, { "epoch": 1.9741559441328493, "grad_norm": 0.362783282995224, "learning_rate": 8e-05, "loss": 1.7772, "step": 8975 }, { "epoch": 1.9743758935444848, "grad_norm": 0.38929077982902527, "learning_rate": 8e-05, "loss": 1.5582, "step": 8976 }, { "epoch": 1.9745958429561201, "grad_norm": 0.3289644122123718, "learning_rate": 8e-05, "loss": 1.4755, "step": 8977 }, { "epoch": 1.9748157923677554, "grad_norm": 0.3423946797847748, "learning_rate": 8e-05, "loss": 1.6087, "step": 8978 }, { "epoch": 1.975035741779391, "grad_norm": 0.3447459042072296, "learning_rate": 8e-05, "loss": 1.541, "step": 8979 }, { "epoch": 1.975255691191026, "grad_norm": 0.3582029342651367, "learning_rate": 8e-05, "loss": 1.6396, "step": 8980 }, { "epoch": 1.9754756406026615, "grad_norm": 0.3636465072631836, "learning_rate": 8e-05, "loss": 1.5505, "step": 8981 }, { "epoch": 1.9756955900142967, "grad_norm": 0.3789360523223877, "learning_rate": 8e-05, "loss": 1.5362, "step": 8982 }, { "epoch": 1.975915539425932, "grad_norm": 0.3441831171512604, "learning_rate": 8e-05, "loss": 1.5661, "step": 8983 }, { "epoch": 1.9761354888375675, "grad_norm": 0.3678434491157532, "learning_rate": 8e-05, "loss": 1.6014, "step": 8984 }, { "epoch": 1.9763554382492026, "grad_norm": 0.3868325650691986, "learning_rate": 8e-05, "loss": 1.5318, "step": 8985 }, { "epoch": 1.976575387660838, "grad_norm": 0.376130074262619, "learning_rate": 8e-05, "loss": 1.5971, "step": 8986 }, { "epoch": 1.9767953370724733, "grad_norm": 0.3484027087688446, "learning_rate": 8e-05, "loss": 1.5247, "step": 8987 }, { "epoch": 1.9770152864841086, "grad_norm": 0.36819878220558167, "learning_rate": 8e-05, "loss": 1.6567, "step": 8988 }, { "epoch": 1.9772352358957441, "grad_norm": 0.3701108694076538, "learning_rate": 8e-05, "loss": 1.6822, "step": 8989 }, { "epoch": 1.9774551853073792, "grad_norm": 0.35930711030960083, "learning_rate": 8e-05, "loss": 1.5647, "step": 8990 }, { "epoch": 1.9776751347190147, "grad_norm": 0.37957409024238586, "learning_rate": 8e-05, "loss": 1.522, "step": 8991 }, { "epoch": 1.97789508413065, "grad_norm": 0.34872594475746155, "learning_rate": 8e-05, "loss": 1.6368, "step": 8992 }, { "epoch": 1.9781150335422852, "grad_norm": 0.3734714090824127, "learning_rate": 8e-05, "loss": 1.6743, "step": 8993 }, { "epoch": 1.9783349829539207, "grad_norm": 0.35892876982688904, "learning_rate": 8e-05, "loss": 1.5557, "step": 8994 }, { "epoch": 1.9785549323655558, "grad_norm": 0.3539820611476898, "learning_rate": 8e-05, "loss": 1.6133, "step": 8995 }, { "epoch": 1.9787748817771913, "grad_norm": 0.3691904544830322, "learning_rate": 8e-05, "loss": 1.6576, "step": 8996 }, { "epoch": 1.9789948311888266, "grad_norm": 0.3481293320655823, "learning_rate": 8e-05, "loss": 1.5832, "step": 8997 }, { "epoch": 1.9792147806004619, "grad_norm": 0.3633832335472107, "learning_rate": 8e-05, "loss": 1.6474, "step": 8998 }, { "epoch": 1.9794347300120974, "grad_norm": 0.3723769783973694, "learning_rate": 8e-05, "loss": 1.5798, "step": 8999 }, { "epoch": 1.9796546794237324, "grad_norm": 0.3624495267868042, "learning_rate": 8e-05, "loss": 1.5427, "step": 9000 }, { "epoch": 1.979874628835368, "grad_norm": 0.3746635615825653, "learning_rate": 8e-05, "loss": 1.5107, "step": 9001 }, { "epoch": 1.9800945782470032, "grad_norm": 0.34711867570877075, "learning_rate": 8e-05, "loss": 1.5943, "step": 9002 }, { "epoch": 1.9803145276586385, "grad_norm": 0.41190215945243835, "learning_rate": 8e-05, "loss": 1.5608, "step": 9003 }, { "epoch": 1.980534477070274, "grad_norm": 0.36933666467666626, "learning_rate": 8e-05, "loss": 1.6003, "step": 9004 }, { "epoch": 1.980754426481909, "grad_norm": 0.37891051173210144, "learning_rate": 8e-05, "loss": 1.635, "step": 9005 }, { "epoch": 1.9809743758935445, "grad_norm": 0.35259121656417847, "learning_rate": 8e-05, "loss": 1.7151, "step": 9006 }, { "epoch": 1.9811943253051798, "grad_norm": 0.3594352900981903, "learning_rate": 8e-05, "loss": 1.7349, "step": 9007 }, { "epoch": 1.981414274716815, "grad_norm": 0.3426830768585205, "learning_rate": 8e-05, "loss": 1.5083, "step": 9008 }, { "epoch": 1.9816342241284506, "grad_norm": 0.3745259940624237, "learning_rate": 8e-05, "loss": 1.7626, "step": 9009 }, { "epoch": 1.9818541735400856, "grad_norm": 0.36304426193237305, "learning_rate": 8e-05, "loss": 1.6862, "step": 9010 }, { "epoch": 1.9820741229517211, "grad_norm": 0.3667314052581787, "learning_rate": 8e-05, "loss": 1.7473, "step": 9011 }, { "epoch": 1.9822940723633564, "grad_norm": 0.3291901648044586, "learning_rate": 8e-05, "loss": 1.2847, "step": 9012 }, { "epoch": 1.9825140217749917, "grad_norm": 0.3544999957084656, "learning_rate": 8e-05, "loss": 1.5576, "step": 9013 }, { "epoch": 1.9827339711866272, "grad_norm": 0.385477751493454, "learning_rate": 8e-05, "loss": 1.7292, "step": 9014 }, { "epoch": 1.9829539205982623, "grad_norm": 0.34754636883735657, "learning_rate": 8e-05, "loss": 1.5466, "step": 9015 }, { "epoch": 1.9831738700098978, "grad_norm": 0.353657066822052, "learning_rate": 8e-05, "loss": 1.5523, "step": 9016 }, { "epoch": 1.983393819421533, "grad_norm": 0.34339675307273865, "learning_rate": 8e-05, "loss": 1.5258, "step": 9017 }, { "epoch": 1.9836137688331683, "grad_norm": 0.3284718096256256, "learning_rate": 8e-05, "loss": 1.5843, "step": 9018 }, { "epoch": 1.9838337182448038, "grad_norm": 0.360424667596817, "learning_rate": 8e-05, "loss": 1.6107, "step": 9019 }, { "epoch": 1.9840536676564389, "grad_norm": 0.3561060428619385, "learning_rate": 8e-05, "loss": 1.5663, "step": 9020 }, { "epoch": 1.9842736170680744, "grad_norm": 0.34420838952064514, "learning_rate": 8e-05, "loss": 1.521, "step": 9021 }, { "epoch": 1.9844935664797096, "grad_norm": 0.3827893137931824, "learning_rate": 8e-05, "loss": 1.4097, "step": 9022 }, { "epoch": 1.984713515891345, "grad_norm": 0.3573097586631775, "learning_rate": 8e-05, "loss": 1.6091, "step": 9023 }, { "epoch": 1.9849334653029804, "grad_norm": 0.35268348455429077, "learning_rate": 8e-05, "loss": 1.4971, "step": 9024 }, { "epoch": 1.9851534147146155, "grad_norm": 0.35729315876960754, "learning_rate": 8e-05, "loss": 1.6278, "step": 9025 }, { "epoch": 1.985373364126251, "grad_norm": 0.36463046073913574, "learning_rate": 8e-05, "loss": 1.6105, "step": 9026 }, { "epoch": 1.9855933135378863, "grad_norm": 0.37124982476234436, "learning_rate": 8e-05, "loss": 1.7779, "step": 9027 }, { "epoch": 1.9858132629495215, "grad_norm": 0.3834069073200226, "learning_rate": 8e-05, "loss": 1.7166, "step": 9028 }, { "epoch": 1.986033212361157, "grad_norm": 0.3773687481880188, "learning_rate": 8e-05, "loss": 1.6398, "step": 9029 }, { "epoch": 1.986253161772792, "grad_norm": 0.344521701335907, "learning_rate": 8e-05, "loss": 1.5746, "step": 9030 }, { "epoch": 1.9864731111844276, "grad_norm": 0.3463936746120453, "learning_rate": 8e-05, "loss": 1.6132, "step": 9031 }, { "epoch": 1.9866930605960629, "grad_norm": 0.3424704074859619, "learning_rate": 8e-05, "loss": 1.427, "step": 9032 }, { "epoch": 1.9869130100076982, "grad_norm": 0.3598610758781433, "learning_rate": 8e-05, "loss": 1.6034, "step": 9033 }, { "epoch": 1.9871329594193337, "grad_norm": 0.32463210821151733, "learning_rate": 8e-05, "loss": 1.4145, "step": 9034 }, { "epoch": 1.9873529088309687, "grad_norm": 0.34233948588371277, "learning_rate": 8e-05, "loss": 1.5858, "step": 9035 }, { "epoch": 1.9875728582426042, "grad_norm": 0.3414950966835022, "learning_rate": 8e-05, "loss": 1.5785, "step": 9036 }, { "epoch": 1.9877928076542395, "grad_norm": 0.3588337302207947, "learning_rate": 8e-05, "loss": 1.5778, "step": 9037 }, { "epoch": 1.9880127570658748, "grad_norm": 0.35715219378471375, "learning_rate": 8e-05, "loss": 1.6615, "step": 9038 }, { "epoch": 1.9882327064775103, "grad_norm": 0.3722425103187561, "learning_rate": 8e-05, "loss": 1.7314, "step": 9039 }, { "epoch": 1.9884526558891455, "grad_norm": 0.3630906045436859, "learning_rate": 8e-05, "loss": 1.6965, "step": 9040 }, { "epoch": 1.9886726053007808, "grad_norm": 0.3343754708766937, "learning_rate": 8e-05, "loss": 1.5283, "step": 9041 }, { "epoch": 1.988892554712416, "grad_norm": 0.34916242957115173, "learning_rate": 8e-05, "loss": 1.5703, "step": 9042 }, { "epoch": 1.9891125041240514, "grad_norm": 0.3460412919521332, "learning_rate": 8e-05, "loss": 1.5763, "step": 9043 }, { "epoch": 1.9893324535356869, "grad_norm": 0.3574906587600708, "learning_rate": 8e-05, "loss": 1.6169, "step": 9044 }, { "epoch": 1.9895524029473222, "grad_norm": 0.34426239132881165, "learning_rate": 8e-05, "loss": 1.5613, "step": 9045 }, { "epoch": 1.9897723523589574, "grad_norm": 0.3451346158981323, "learning_rate": 8e-05, "loss": 1.5548, "step": 9046 }, { "epoch": 1.9899923017705927, "grad_norm": 0.3435182571411133, "learning_rate": 8e-05, "loss": 1.6322, "step": 9047 }, { "epoch": 1.990212251182228, "grad_norm": 0.3573644757270813, "learning_rate": 8e-05, "loss": 1.6606, "step": 9048 }, { "epoch": 1.9904322005938635, "grad_norm": 0.39593827724456787, "learning_rate": 8e-05, "loss": 1.6012, "step": 9049 }, { "epoch": 1.9906521500054988, "grad_norm": 0.36734482645988464, "learning_rate": 8e-05, "loss": 1.6438, "step": 9050 }, { "epoch": 1.990872099417134, "grad_norm": 0.451093465089798, "learning_rate": 8e-05, "loss": 1.6751, "step": 9051 }, { "epoch": 1.9910920488287693, "grad_norm": 0.34433048963546753, "learning_rate": 8e-05, "loss": 1.6303, "step": 9052 }, { "epoch": 1.9913119982404046, "grad_norm": 0.3456341624259949, "learning_rate": 8e-05, "loss": 1.4934, "step": 9053 }, { "epoch": 1.9915319476520401, "grad_norm": 0.3573620915412903, "learning_rate": 8e-05, "loss": 1.6986, "step": 9054 }, { "epoch": 1.9917518970636754, "grad_norm": 0.35553643107414246, "learning_rate": 8e-05, "loss": 1.5506, "step": 9055 }, { "epoch": 1.9919718464753107, "grad_norm": 0.3726920783519745, "learning_rate": 8e-05, "loss": 1.4916, "step": 9056 }, { "epoch": 1.992191795886946, "grad_norm": 0.36144542694091797, "learning_rate": 8e-05, "loss": 1.7281, "step": 9057 }, { "epoch": 1.9924117452985812, "grad_norm": 0.4092212915420532, "learning_rate": 8e-05, "loss": 1.5165, "step": 9058 }, { "epoch": 1.9926316947102167, "grad_norm": 0.3463384807109833, "learning_rate": 8e-05, "loss": 1.5124, "step": 9059 }, { "epoch": 1.992851644121852, "grad_norm": 0.3370697796344757, "learning_rate": 8e-05, "loss": 1.4894, "step": 9060 }, { "epoch": 1.9930715935334873, "grad_norm": 0.35062646865844727, "learning_rate": 8e-05, "loss": 1.4927, "step": 9061 }, { "epoch": 1.9932915429451228, "grad_norm": 0.34355610609054565, "learning_rate": 8e-05, "loss": 1.5269, "step": 9062 }, { "epoch": 1.9935114923567578, "grad_norm": 0.3496817946434021, "learning_rate": 8e-05, "loss": 1.5765, "step": 9063 }, { "epoch": 1.9937314417683933, "grad_norm": 0.34196415543556213, "learning_rate": 8e-05, "loss": 1.5421, "step": 9064 }, { "epoch": 1.9939513911800286, "grad_norm": 0.3537460267543793, "learning_rate": 8e-05, "loss": 1.6186, "step": 9065 }, { "epoch": 1.994171340591664, "grad_norm": 0.3337237536907196, "learning_rate": 8e-05, "loss": 1.563, "step": 9066 }, { "epoch": 1.9943912900032994, "grad_norm": 0.3468160331249237, "learning_rate": 8e-05, "loss": 1.5751, "step": 9067 }, { "epoch": 1.9946112394149345, "grad_norm": 0.35053637623786926, "learning_rate": 8e-05, "loss": 1.6242, "step": 9068 }, { "epoch": 1.99483118882657, "grad_norm": 0.3475688695907593, "learning_rate": 8e-05, "loss": 1.7177, "step": 9069 }, { "epoch": 1.9950511382382052, "grad_norm": 0.3412964344024658, "learning_rate": 8e-05, "loss": 1.4844, "step": 9070 }, { "epoch": 1.9952710876498405, "grad_norm": 0.32939761877059937, "learning_rate": 8e-05, "loss": 1.447, "step": 9071 }, { "epoch": 1.995491037061476, "grad_norm": 0.34977617859840393, "learning_rate": 8e-05, "loss": 1.7354, "step": 9072 }, { "epoch": 1.995710986473111, "grad_norm": 0.3301527500152588, "learning_rate": 8e-05, "loss": 1.5103, "step": 9073 }, { "epoch": 1.9959309358847466, "grad_norm": 0.34539759159088135, "learning_rate": 8e-05, "loss": 1.6716, "step": 9074 }, { "epoch": 1.9961508852963818, "grad_norm": 0.35377198457717896, "learning_rate": 8e-05, "loss": 1.6709, "step": 9075 }, { "epoch": 1.9963708347080171, "grad_norm": 0.36640143394470215, "learning_rate": 8e-05, "loss": 1.7484, "step": 9076 }, { "epoch": 1.9965907841196526, "grad_norm": 0.32511720061302185, "learning_rate": 8e-05, "loss": 1.4246, "step": 9077 }, { "epoch": 1.9968107335312877, "grad_norm": 0.32862409949302673, "learning_rate": 8e-05, "loss": 1.4547, "step": 9078 }, { "epoch": 1.9970306829429232, "grad_norm": 0.36281830072402954, "learning_rate": 8e-05, "loss": 1.6095, "step": 9079 }, { "epoch": 1.9972506323545585, "grad_norm": 0.3405216634273529, "learning_rate": 8e-05, "loss": 1.4778, "step": 9080 }, { "epoch": 1.9974705817661937, "grad_norm": 0.3578289747238159, "learning_rate": 8e-05, "loss": 1.4925, "step": 9081 }, { "epoch": 1.9976905311778292, "grad_norm": 0.36427584290504456, "learning_rate": 8e-05, "loss": 1.65, "step": 9082 }, { "epoch": 1.9979104805894643, "grad_norm": 0.3563012182712555, "learning_rate": 8e-05, "loss": 1.5433, "step": 9083 }, { "epoch": 1.9981304300010998, "grad_norm": 0.36771148443222046, "learning_rate": 8e-05, "loss": 1.5739, "step": 9084 }, { "epoch": 1.998350379412735, "grad_norm": 0.34824132919311523, "learning_rate": 8e-05, "loss": 1.5757, "step": 9085 }, { "epoch": 1.9985703288243704, "grad_norm": 0.34271854162216187, "learning_rate": 8e-05, "loss": 1.6358, "step": 9086 }, { "epoch": 1.9987902782360059, "grad_norm": 0.3409704267978668, "learning_rate": 8e-05, "loss": 1.4726, "step": 9087 }, { "epoch": 1.999010227647641, "grad_norm": 0.3302672803401947, "learning_rate": 8e-05, "loss": 1.5475, "step": 9088 }, { "epoch": 1.9992301770592764, "grad_norm": 0.35327091813087463, "learning_rate": 8e-05, "loss": 1.5196, "step": 9089 }, { "epoch": 1.9994501264709117, "grad_norm": 0.3525160849094391, "learning_rate": 8e-05, "loss": 1.4678, "step": 9090 }, { "epoch": 1.999670075882547, "grad_norm": 0.44579142332077026, "learning_rate": 8e-05, "loss": 1.6509, "step": 9091 }, { "epoch": 1.9998900252941825, "grad_norm": 0.35603219270706177, "learning_rate": 8e-05, "loss": 1.6378, "step": 9092 } ], "logging_steps": 1, "max_steps": 9092, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.7144256822446653e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }