{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 4141, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002414875633904854, "grad_norm": 0.49120378494262695, "learning_rate": 1.2048192771084338e-06, "loss": 2.1193, "step": 1 }, { "epoch": 0.0004829751267809708, "grad_norm": 0.4886572062969208, "learning_rate": 2.4096385542168676e-06, "loss": 1.9434, "step": 2 }, { "epoch": 0.0007244626901714562, "grad_norm": 2.381781578063965, "learning_rate": 3.614457831325301e-06, "loss": 2.3871, "step": 3 }, { "epoch": 0.0009659502535619416, "grad_norm": 0.5716086030006409, "learning_rate": 4.819277108433735e-06, "loss": 2.0503, "step": 4 }, { "epoch": 0.001207437816952427, "grad_norm": 1.3921763896942139, "learning_rate": 6.024096385542169e-06, "loss": 1.8545, "step": 5 }, { "epoch": 0.0014489253803429123, "grad_norm": 0.6292646527290344, "learning_rate": 7.228915662650602e-06, "loss": 2.0255, "step": 6 }, { "epoch": 0.0016904129437333977, "grad_norm": 0.49030834436416626, "learning_rate": 8.433734939759036e-06, "loss": 1.994, "step": 7 }, { "epoch": 0.001931900507123883, "grad_norm": 0.47718971967697144, "learning_rate": 9.63855421686747e-06, "loss": 2.0012, "step": 8 }, { "epoch": 0.0021733880705143687, "grad_norm": 0.6809853315353394, "learning_rate": 1.0843373493975904e-05, "loss": 1.9578, "step": 9 }, { "epoch": 0.002414875633904854, "grad_norm": 0.4382535517215729, "learning_rate": 1.2048192771084338e-05, "loss": 2.0007, "step": 10 }, { "epoch": 0.0026563631972953395, "grad_norm": 0.40765514969825745, "learning_rate": 1.3253012048192772e-05, "loss": 1.8652, "step": 11 }, { "epoch": 0.0028978507606858247, "grad_norm": 0.47924724221229553, "learning_rate": 1.4457831325301205e-05, "loss": 1.9401, "step": 12 }, { "epoch": 0.0031393383240763103, "grad_norm": 0.40365689992904663, "learning_rate": 1.566265060240964e-05, "loss": 2.016, "step": 13 }, { "epoch": 0.0033808258874667954, "grad_norm": 0.46215543150901794, "learning_rate": 1.6867469879518073e-05, "loss": 1.9878, "step": 14 }, { "epoch": 0.003622313450857281, "grad_norm": 0.41287553310394287, "learning_rate": 1.8072289156626505e-05, "loss": 1.8019, "step": 15 }, { "epoch": 0.003863801014247766, "grad_norm": 0.4323561191558838, "learning_rate": 1.927710843373494e-05, "loss": 1.9787, "step": 16 }, { "epoch": 0.004105288577638252, "grad_norm": 0.4473198652267456, "learning_rate": 2.0481927710843373e-05, "loss": 1.9847, "step": 17 }, { "epoch": 0.004346776141028737, "grad_norm": 0.6705676317214966, "learning_rate": 2.168674698795181e-05, "loss": 2.2759, "step": 18 }, { "epoch": 0.004588263704419222, "grad_norm": 0.4056943655014038, "learning_rate": 2.289156626506024e-05, "loss": 2.0165, "step": 19 }, { "epoch": 0.004829751267809708, "grad_norm": 0.36379876732826233, "learning_rate": 2.4096385542168677e-05, "loss": 1.896, "step": 20 }, { "epoch": 0.005071238831200193, "grad_norm": 0.35651546716690063, "learning_rate": 2.530120481927711e-05, "loss": 1.6728, "step": 21 }, { "epoch": 0.005312726394590679, "grad_norm": 0.3706895411014557, "learning_rate": 2.6506024096385545e-05, "loss": 1.9103, "step": 22 }, { "epoch": 0.005554213957981164, "grad_norm": 0.355390727519989, "learning_rate": 2.7710843373493977e-05, "loss": 1.7729, "step": 23 }, { "epoch": 0.005795701521371649, "grad_norm": 0.4054727256298065, "learning_rate": 2.891566265060241e-05, "loss": 1.999, "step": 24 }, { "epoch": 0.006037189084762135, "grad_norm": 0.3950248062610626, "learning_rate": 3.012048192771085e-05, "loss": 1.8431, "step": 25 }, { "epoch": 0.0062786766481526205, "grad_norm": 0.35835927724838257, "learning_rate": 3.132530120481928e-05, "loss": 1.713, "step": 26 }, { "epoch": 0.006520164211543105, "grad_norm": 0.42847758531570435, "learning_rate": 3.253012048192771e-05, "loss": 1.859, "step": 27 }, { "epoch": 0.006761651774933591, "grad_norm": 0.3585743010044098, "learning_rate": 3.3734939759036146e-05, "loss": 1.7553, "step": 28 }, { "epoch": 0.0070031393383240765, "grad_norm": 0.35875406861305237, "learning_rate": 3.4939759036144585e-05, "loss": 1.9552, "step": 29 }, { "epoch": 0.007244626901714562, "grad_norm": 0.3377876281738281, "learning_rate": 3.614457831325301e-05, "loss": 1.7706, "step": 30 }, { "epoch": 0.007486114465105047, "grad_norm": 0.3570369780063629, "learning_rate": 3.734939759036144e-05, "loss": 1.8494, "step": 31 }, { "epoch": 0.007727602028495532, "grad_norm": 0.3713253438472748, "learning_rate": 3.855421686746988e-05, "loss": 2.0651, "step": 32 }, { "epoch": 0.007969089591886018, "grad_norm": 0.3895401656627655, "learning_rate": 3.9759036144578314e-05, "loss": 1.9954, "step": 33 }, { "epoch": 0.008210577155276504, "grad_norm": 0.3511655032634735, "learning_rate": 4.0963855421686746e-05, "loss": 1.7921, "step": 34 }, { "epoch": 0.00845206471866699, "grad_norm": 0.32599174976348877, "learning_rate": 4.2168674698795186e-05, "loss": 1.7539, "step": 35 }, { "epoch": 0.008693552282057475, "grad_norm": 0.32991236448287964, "learning_rate": 4.337349397590362e-05, "loss": 1.7223, "step": 36 }, { "epoch": 0.008935039845447959, "grad_norm": 0.3855891227722168, "learning_rate": 4.457831325301205e-05, "loss": 1.7663, "step": 37 }, { "epoch": 0.009176527408838444, "grad_norm": 0.36162635684013367, "learning_rate": 4.578313253012048e-05, "loss": 1.9458, "step": 38 }, { "epoch": 0.00941801497222893, "grad_norm": 0.33777403831481934, "learning_rate": 4.698795180722892e-05, "loss": 1.854, "step": 39 }, { "epoch": 0.009659502535619416, "grad_norm": 0.41380664706230164, "learning_rate": 4.8192771084337354e-05, "loss": 1.9908, "step": 40 }, { "epoch": 0.009900990099009901, "grad_norm": 0.34837591648101807, "learning_rate": 4.9397590361445786e-05, "loss": 1.9103, "step": 41 }, { "epoch": 0.010142477662400387, "grad_norm": 0.3593843877315521, "learning_rate": 5.060240963855422e-05, "loss": 2.0174, "step": 42 }, { "epoch": 0.010383965225790872, "grad_norm": 0.3512469530105591, "learning_rate": 5.180722891566265e-05, "loss": 1.8854, "step": 43 }, { "epoch": 0.010625452789181358, "grad_norm": 0.3344199061393738, "learning_rate": 5.301204819277109e-05, "loss": 1.771, "step": 44 }, { "epoch": 0.010866940352571842, "grad_norm": 0.31484276056289673, "learning_rate": 5.4216867469879516e-05, "loss": 1.7191, "step": 45 }, { "epoch": 0.011108427915962327, "grad_norm": 0.33984580636024475, "learning_rate": 5.5421686746987955e-05, "loss": 1.8153, "step": 46 }, { "epoch": 0.011349915479352813, "grad_norm": 0.37721672654151917, "learning_rate": 5.6626506024096394e-05, "loss": 1.932, "step": 47 }, { "epoch": 0.011591403042743299, "grad_norm": 0.32222846150398254, "learning_rate": 5.783132530120482e-05, "loss": 1.7607, "step": 48 }, { "epoch": 0.011832890606133784, "grad_norm": 0.35319098830223083, "learning_rate": 5.903614457831326e-05, "loss": 1.9254, "step": 49 }, { "epoch": 0.01207437816952427, "grad_norm": 0.37980917096138, "learning_rate": 6.02409638554217e-05, "loss": 1.9552, "step": 50 }, { "epoch": 0.012315865732914755, "grad_norm": 0.3551584780216217, "learning_rate": 6.144578313253012e-05, "loss": 2.0627, "step": 51 }, { "epoch": 0.012557353296305241, "grad_norm": 0.3481822609901428, "learning_rate": 6.265060240963856e-05, "loss": 1.9323, "step": 52 }, { "epoch": 0.012798840859695725, "grad_norm": 0.34489256143569946, "learning_rate": 6.385542168674698e-05, "loss": 1.863, "step": 53 }, { "epoch": 0.01304032842308621, "grad_norm": 0.34144216775894165, "learning_rate": 6.506024096385543e-05, "loss": 1.9177, "step": 54 }, { "epoch": 0.013281815986476696, "grad_norm": 0.3356941044330597, "learning_rate": 6.626506024096386e-05, "loss": 1.8424, "step": 55 }, { "epoch": 0.013523303549867182, "grad_norm": 0.3361498713493347, "learning_rate": 6.746987951807229e-05, "loss": 1.7362, "step": 56 }, { "epoch": 0.013764791113257667, "grad_norm": 0.37799155712127686, "learning_rate": 6.867469879518072e-05, "loss": 2.1137, "step": 57 }, { "epoch": 0.014006278676648153, "grad_norm": 0.3507687747478485, "learning_rate": 6.987951807228917e-05, "loss": 1.9008, "step": 58 }, { "epoch": 0.014247766240038639, "grad_norm": 0.3450543284416199, "learning_rate": 7.108433734939759e-05, "loss": 1.868, "step": 59 }, { "epoch": 0.014489253803429124, "grad_norm": 0.33962032198905945, "learning_rate": 7.228915662650602e-05, "loss": 1.7856, "step": 60 }, { "epoch": 0.014730741366819608, "grad_norm": 0.4293012022972107, "learning_rate": 7.349397590361447e-05, "loss": 2.1798, "step": 61 }, { "epoch": 0.014972228930210094, "grad_norm": 0.35071665048599243, "learning_rate": 7.469879518072289e-05, "loss": 1.8269, "step": 62 }, { "epoch": 0.01521371649360058, "grad_norm": 0.3420826494693756, "learning_rate": 7.590361445783133e-05, "loss": 1.8103, "step": 63 }, { "epoch": 0.015455204056991065, "grad_norm": 0.31659749150276184, "learning_rate": 7.710843373493976e-05, "loss": 1.7773, "step": 64 }, { "epoch": 0.01569669162038155, "grad_norm": 0.3348800837993622, "learning_rate": 7.83132530120482e-05, "loss": 1.8194, "step": 65 }, { "epoch": 0.015938179183772036, "grad_norm": 0.3300272524356842, "learning_rate": 7.951807228915663e-05, "loss": 1.8126, "step": 66 }, { "epoch": 0.01617966674716252, "grad_norm": 0.3315858542919159, "learning_rate": 8.072289156626507e-05, "loss": 1.8791, "step": 67 }, { "epoch": 0.016421154310553007, "grad_norm": 0.34572306275367737, "learning_rate": 8.192771084337349e-05, "loss": 1.8424, "step": 68 }, { "epoch": 0.016662641873943493, "grad_norm": 0.31780093908309937, "learning_rate": 8.313253012048194e-05, "loss": 1.7617, "step": 69 }, { "epoch": 0.01690412943733398, "grad_norm": 0.35894426703453064, "learning_rate": 8.433734939759037e-05, "loss": 1.9024, "step": 70 }, { "epoch": 0.017145617000724464, "grad_norm": 0.32306596636772156, "learning_rate": 8.55421686746988e-05, "loss": 1.8297, "step": 71 }, { "epoch": 0.01738710456411495, "grad_norm": 0.3225422203540802, "learning_rate": 8.674698795180724e-05, "loss": 1.8403, "step": 72 }, { "epoch": 0.017628592127505432, "grad_norm": 0.31123486161231995, "learning_rate": 8.795180722891567e-05, "loss": 1.7744, "step": 73 }, { "epoch": 0.017870079690895917, "grad_norm": 0.32267558574676514, "learning_rate": 8.91566265060241e-05, "loss": 1.8111, "step": 74 }, { "epoch": 0.018111567254286403, "grad_norm": 0.3544601798057556, "learning_rate": 9.036144578313253e-05, "loss": 2.1014, "step": 75 }, { "epoch": 0.01835305481767689, "grad_norm": 0.30508580803871155, "learning_rate": 9.156626506024096e-05, "loss": 1.7758, "step": 76 }, { "epoch": 0.018594542381067374, "grad_norm": 0.321807324886322, "learning_rate": 9.27710843373494e-05, "loss": 1.7328, "step": 77 }, { "epoch": 0.01883602994445786, "grad_norm": 0.3235253393650055, "learning_rate": 9.397590361445784e-05, "loss": 1.8749, "step": 78 }, { "epoch": 0.019077517507848345, "grad_norm": 0.29553085565567017, "learning_rate": 9.518072289156626e-05, "loss": 1.6523, "step": 79 }, { "epoch": 0.01931900507123883, "grad_norm": 0.33573049306869507, "learning_rate": 9.638554216867471e-05, "loss": 1.9299, "step": 80 }, { "epoch": 0.019560492634629317, "grad_norm": 0.32041892409324646, "learning_rate": 9.759036144578314e-05, "loss": 1.8036, "step": 81 }, { "epoch": 0.019801980198019802, "grad_norm": 0.3136705458164215, "learning_rate": 9.879518072289157e-05, "loss": 1.7463, "step": 82 }, { "epoch": 0.020043467761410288, "grad_norm": 0.31231194734573364, "learning_rate": 0.0001, "loss": 1.7531, "step": 83 }, { "epoch": 0.020284955324800773, "grad_norm": 0.3328460454940796, "learning_rate": 9.999998501641807e-05, "loss": 1.9319, "step": 84 }, { "epoch": 0.02052644288819126, "grad_norm": 0.31596639752388, "learning_rate": 9.999994006568125e-05, "loss": 1.8929, "step": 85 }, { "epoch": 0.020767930451581745, "grad_norm": 0.42943811416625977, "learning_rate": 9.99998651478165e-05, "loss": 2.0316, "step": 86 }, { "epoch": 0.02100941801497223, "grad_norm": 0.3127952516078949, "learning_rate": 9.99997602628687e-05, "loss": 1.7111, "step": 87 }, { "epoch": 0.021250905578362716, "grad_norm": 0.33090245723724365, "learning_rate": 9.999962541090071e-05, "loss": 1.7233, "step": 88 }, { "epoch": 0.021492393141753198, "grad_norm": 0.3207198679447174, "learning_rate": 9.999946059199337e-05, "loss": 1.8777, "step": 89 }, { "epoch": 0.021733880705143684, "grad_norm": 0.3291267156600952, "learning_rate": 9.999926580624545e-05, "loss": 1.883, "step": 90 }, { "epoch": 0.02197536826853417, "grad_norm": 0.3264913856983185, "learning_rate": 9.999904105377372e-05, "loss": 1.8955, "step": 91 }, { "epoch": 0.022216855831924655, "grad_norm": 0.33100128173828125, "learning_rate": 9.999878633471285e-05, "loss": 2.0645, "step": 92 }, { "epoch": 0.02245834339531514, "grad_norm": 0.3285035789012909, "learning_rate": 9.999850164921551e-05, "loss": 1.9464, "step": 93 }, { "epoch": 0.022699830958705626, "grad_norm": 0.31835636496543884, "learning_rate": 9.999818699745234e-05, "loss": 1.834, "step": 94 }, { "epoch": 0.02294131852209611, "grad_norm": 0.2967548966407776, "learning_rate": 9.999784237961192e-05, "loss": 1.7508, "step": 95 }, { "epoch": 0.023182806085486597, "grad_norm": 0.37682053446769714, "learning_rate": 9.999746779590078e-05, "loss": 1.9776, "step": 96 }, { "epoch": 0.023424293648877083, "grad_norm": 0.29876354336738586, "learning_rate": 9.999706324654343e-05, "loss": 1.5886, "step": 97 }, { "epoch": 0.02366578121226757, "grad_norm": 0.32981058955192566, "learning_rate": 9.999662873178236e-05, "loss": 1.9233, "step": 98 }, { "epoch": 0.023907268775658054, "grad_norm": 0.33086058497428894, "learning_rate": 9.999616425187796e-05, "loss": 1.9578, "step": 99 }, { "epoch": 0.02414875633904854, "grad_norm": 0.307336688041687, "learning_rate": 9.999566980710863e-05, "loss": 1.783, "step": 100 }, { "epoch": 0.024390243902439025, "grad_norm": 0.3072247803211212, "learning_rate": 9.999514539777071e-05, "loss": 1.8958, "step": 101 }, { "epoch": 0.02463173146582951, "grad_norm": 0.2904624044895172, "learning_rate": 9.999459102417847e-05, "loss": 1.7192, "step": 102 }, { "epoch": 0.024873219029219996, "grad_norm": 0.31059005856513977, "learning_rate": 9.999400668666423e-05, "loss": 1.8329, "step": 103 }, { "epoch": 0.025114706592610482, "grad_norm": 0.30321741104125977, "learning_rate": 9.999339238557816e-05, "loss": 2.0038, "step": 104 }, { "epoch": 0.025356194156000968, "grad_norm": 0.3106807768344879, "learning_rate": 9.999274812128847e-05, "loss": 1.8293, "step": 105 }, { "epoch": 0.02559768171939145, "grad_norm": 0.304987370967865, "learning_rate": 9.999207389418129e-05, "loss": 1.8311, "step": 106 }, { "epoch": 0.025839169282781935, "grad_norm": 0.31797996163368225, "learning_rate": 9.999136970466067e-05, "loss": 1.9911, "step": 107 }, { "epoch": 0.02608065684617242, "grad_norm": 0.2915087044239044, "learning_rate": 9.999063555314872e-05, "loss": 1.7453, "step": 108 }, { "epoch": 0.026322144409562907, "grad_norm": 0.31751200556755066, "learning_rate": 9.99898714400854e-05, "loss": 1.7372, "step": 109 }, { "epoch": 0.026563631972953392, "grad_norm": 0.2950197458267212, "learning_rate": 9.998907736592873e-05, "loss": 1.7576, "step": 110 }, { "epoch": 0.026805119536343878, "grad_norm": 0.3286873996257782, "learning_rate": 9.998825333115458e-05, "loss": 2.0046, "step": 111 }, { "epoch": 0.027046607099734363, "grad_norm": 0.3000810146331787, "learning_rate": 9.998739933625686e-05, "loss": 1.6593, "step": 112 }, { "epoch": 0.02728809466312485, "grad_norm": 0.33321690559387207, "learning_rate": 9.99865153817474e-05, "loss": 2.0058, "step": 113 }, { "epoch": 0.027529582226515335, "grad_norm": 0.3196820020675659, "learning_rate": 9.998560146815599e-05, "loss": 2.0243, "step": 114 }, { "epoch": 0.02777106978990582, "grad_norm": 0.31422409415245056, "learning_rate": 9.998465759603037e-05, "loss": 1.8517, "step": 115 }, { "epoch": 0.028012557353296306, "grad_norm": 0.29709357023239136, "learning_rate": 9.998368376593627e-05, "loss": 1.7622, "step": 116 }, { "epoch": 0.02825404491668679, "grad_norm": 0.29982519149780273, "learning_rate": 9.998267997845732e-05, "loss": 1.6767, "step": 117 }, { "epoch": 0.028495532480077277, "grad_norm": 0.2890397906303406, "learning_rate": 9.998164623419513e-05, "loss": 1.8896, "step": 118 }, { "epoch": 0.028737020043467763, "grad_norm": 0.28895923495292664, "learning_rate": 9.99805825337693e-05, "loss": 1.8441, "step": 119 }, { "epoch": 0.028978507606858248, "grad_norm": 0.3054564893245697, "learning_rate": 9.997948887781732e-05, "loss": 1.8496, "step": 120 }, { "epoch": 0.029219995170248734, "grad_norm": 0.3080004155635834, "learning_rate": 9.99783652669947e-05, "loss": 1.7408, "step": 121 }, { "epoch": 0.029461482733639216, "grad_norm": 0.2757675349712372, "learning_rate": 9.997721170197482e-05, "loss": 1.7709, "step": 122 }, { "epoch": 0.0297029702970297, "grad_norm": 0.32389089465141296, "learning_rate": 9.997602818344911e-05, "loss": 1.8518, "step": 123 }, { "epoch": 0.029944457860420187, "grad_norm": 0.3012002408504486, "learning_rate": 9.997481471212688e-05, "loss": 1.7495, "step": 124 }, { "epoch": 0.030185945423810673, "grad_norm": 0.3027832508087158, "learning_rate": 9.997357128873541e-05, "loss": 1.8043, "step": 125 }, { "epoch": 0.03042743298720116, "grad_norm": 0.28534793853759766, "learning_rate": 9.997229791401993e-05, "loss": 1.7346, "step": 126 }, { "epoch": 0.030668920550591644, "grad_norm": 0.31898003816604614, "learning_rate": 9.997099458874369e-05, "loss": 1.8055, "step": 127 }, { "epoch": 0.03091040811398213, "grad_norm": 0.31042003631591797, "learning_rate": 9.996966131368775e-05, "loss": 1.7288, "step": 128 }, { "epoch": 0.031151895677372615, "grad_norm": 0.30200737714767456, "learning_rate": 9.996829808965122e-05, "loss": 1.814, "step": 129 }, { "epoch": 0.0313933832407631, "grad_norm": 0.3008061647415161, "learning_rate": 9.996690491745118e-05, "loss": 1.8479, "step": 130 }, { "epoch": 0.031634870804153586, "grad_norm": 0.3122195601463318, "learning_rate": 9.996548179792257e-05, "loss": 1.7699, "step": 131 }, { "epoch": 0.03187635836754407, "grad_norm": 0.3094089925289154, "learning_rate": 9.996402873191837e-05, "loss": 1.8328, "step": 132 }, { "epoch": 0.03211784593093456, "grad_norm": 0.3003721237182617, "learning_rate": 9.996254572030943e-05, "loss": 1.7759, "step": 133 }, { "epoch": 0.03235933349432504, "grad_norm": 0.32589980959892273, "learning_rate": 9.996103276398461e-05, "loss": 1.8535, "step": 134 }, { "epoch": 0.03260082105771553, "grad_norm": 0.3078492283821106, "learning_rate": 9.995948986385067e-05, "loss": 1.8138, "step": 135 }, { "epoch": 0.032842308621106014, "grad_norm": 0.28524380922317505, "learning_rate": 9.995791702083232e-05, "loss": 1.7354, "step": 136 }, { "epoch": 0.0330837961844965, "grad_norm": 0.3418903648853302, "learning_rate": 9.995631423587228e-05, "loss": 1.9039, "step": 137 }, { "epoch": 0.033325283747886986, "grad_norm": 0.29714182019233704, "learning_rate": 9.995468150993114e-05, "loss": 1.8988, "step": 138 }, { "epoch": 0.03356677131127747, "grad_norm": 0.34489384293556213, "learning_rate": 9.995301884398747e-05, "loss": 1.9888, "step": 139 }, { "epoch": 0.03380825887466796, "grad_norm": 0.2920478284358978, "learning_rate": 9.995132623903777e-05, "loss": 1.6168, "step": 140 }, { "epoch": 0.03404974643805844, "grad_norm": 0.2992519438266754, "learning_rate": 9.994960369609649e-05, "loss": 1.8919, "step": 141 }, { "epoch": 0.03429123400144893, "grad_norm": 0.29365047812461853, "learning_rate": 9.994785121619605e-05, "loss": 1.7853, "step": 142 }, { "epoch": 0.034532721564839414, "grad_norm": 0.30724966526031494, "learning_rate": 9.994606880038676e-05, "loss": 1.9399, "step": 143 }, { "epoch": 0.0347742091282299, "grad_norm": 0.41282230615615845, "learning_rate": 9.994425644973691e-05, "loss": 2.0693, "step": 144 }, { "epoch": 0.035015696691620385, "grad_norm": 0.30199870467185974, "learning_rate": 9.99424141653327e-05, "loss": 1.8273, "step": 145 }, { "epoch": 0.035257184255010864, "grad_norm": 0.28794577717781067, "learning_rate": 9.994054194827832e-05, "loss": 1.8246, "step": 146 }, { "epoch": 0.03549867181840135, "grad_norm": 0.2896375358104706, "learning_rate": 9.993863979969587e-05, "loss": 1.8404, "step": 147 }, { "epoch": 0.035740159381791835, "grad_norm": 0.32256078720092773, "learning_rate": 9.993670772072537e-05, "loss": 1.8969, "step": 148 }, { "epoch": 0.03598164694518232, "grad_norm": 0.27402758598327637, "learning_rate": 9.993474571252479e-05, "loss": 1.7122, "step": 149 }, { "epoch": 0.036223134508572806, "grad_norm": 0.32481542229652405, "learning_rate": 9.993275377627008e-05, "loss": 2.0565, "step": 150 }, { "epoch": 0.03646462207196329, "grad_norm": 0.26531848311424255, "learning_rate": 9.993073191315507e-05, "loss": 1.5844, "step": 151 }, { "epoch": 0.03670610963535378, "grad_norm": 0.2869581878185272, "learning_rate": 9.992868012439159e-05, "loss": 1.7028, "step": 152 }, { "epoch": 0.03694759719874426, "grad_norm": 0.2975616753101349, "learning_rate": 9.992659841120929e-05, "loss": 1.9539, "step": 153 }, { "epoch": 0.03718908476213475, "grad_norm": 0.322748064994812, "learning_rate": 9.99244867748559e-05, "loss": 2.0498, "step": 154 }, { "epoch": 0.037430572325525234, "grad_norm": 0.2980214059352875, "learning_rate": 9.992234521659698e-05, "loss": 1.9035, "step": 155 }, { "epoch": 0.03767205988891572, "grad_norm": 0.3023289144039154, "learning_rate": 9.992017373771607e-05, "loss": 1.8374, "step": 156 }, { "epoch": 0.037913547452306205, "grad_norm": 0.2790420949459076, "learning_rate": 9.991797233951462e-05, "loss": 1.4934, "step": 157 }, { "epoch": 0.03815503501569669, "grad_norm": 0.2871580719947815, "learning_rate": 9.991574102331204e-05, "loss": 1.7839, "step": 158 }, { "epoch": 0.038396522579087176, "grad_norm": 0.2993938624858856, "learning_rate": 9.991347979044566e-05, "loss": 1.8198, "step": 159 }, { "epoch": 0.03863801014247766, "grad_norm": 0.30797529220581055, "learning_rate": 9.99111886422707e-05, "loss": 1.8831, "step": 160 }, { "epoch": 0.03887949770586815, "grad_norm": 0.30790555477142334, "learning_rate": 9.990886758016038e-05, "loss": 1.9753, "step": 161 }, { "epoch": 0.03912098526925863, "grad_norm": 0.2984273135662079, "learning_rate": 9.990651660550579e-05, "loss": 1.656, "step": 162 }, { "epoch": 0.03936247283264912, "grad_norm": 0.30045798420906067, "learning_rate": 9.990413571971598e-05, "loss": 1.7617, "step": 163 }, { "epoch": 0.039603960396039604, "grad_norm": 0.2841191291809082, "learning_rate": 9.990172492421791e-05, "loss": 1.7359, "step": 164 }, { "epoch": 0.03984544795943009, "grad_norm": 0.363344669342041, "learning_rate": 9.98992842204565e-05, "loss": 2.0574, "step": 165 }, { "epoch": 0.040086935522820576, "grad_norm": 0.2936279773712158, "learning_rate": 9.989681360989454e-05, "loss": 1.8597, "step": 166 }, { "epoch": 0.04032842308621106, "grad_norm": 0.31181731820106506, "learning_rate": 9.989431309401277e-05, "loss": 2.0107, "step": 167 }, { "epoch": 0.04056991064960155, "grad_norm": 0.28021594882011414, "learning_rate": 9.989178267430988e-05, "loss": 1.5916, "step": 168 }, { "epoch": 0.04081139821299203, "grad_norm": 0.3215208053588867, "learning_rate": 9.988922235230247e-05, "loss": 1.9517, "step": 169 }, { "epoch": 0.04105288577638252, "grad_norm": 0.2947755753993988, "learning_rate": 9.988663212952502e-05, "loss": 1.8771, "step": 170 }, { "epoch": 0.041294373339773004, "grad_norm": 0.28413596749305725, "learning_rate": 9.988401200752997e-05, "loss": 1.8199, "step": 171 }, { "epoch": 0.04153586090316349, "grad_norm": 0.3089015781879425, "learning_rate": 9.988136198788769e-05, "loss": 1.7263, "step": 172 }, { "epoch": 0.041777348466553975, "grad_norm": 0.30298158526420593, "learning_rate": 9.987868207218645e-05, "loss": 1.7411, "step": 173 }, { "epoch": 0.04201883602994446, "grad_norm": 0.2908902168273926, "learning_rate": 9.98759722620324e-05, "loss": 1.7757, "step": 174 }, { "epoch": 0.042260323593334946, "grad_norm": 0.3005327880382538, "learning_rate": 9.98732325590497e-05, "loss": 1.7623, "step": 175 }, { "epoch": 0.04250181115672543, "grad_norm": 0.31084272265434265, "learning_rate": 9.987046296488034e-05, "loss": 1.8229, "step": 176 }, { "epoch": 0.04274329872011592, "grad_norm": 0.3010302186012268, "learning_rate": 9.986766348118428e-05, "loss": 1.847, "step": 177 }, { "epoch": 0.042984786283506396, "grad_norm": 0.5154542922973633, "learning_rate": 9.986483410963934e-05, "loss": 1.9374, "step": 178 }, { "epoch": 0.04322627384689688, "grad_norm": 0.2857230603694916, "learning_rate": 9.98619748519413e-05, "loss": 1.6965, "step": 179 }, { "epoch": 0.04346776141028737, "grad_norm": 0.29431548714637756, "learning_rate": 9.985908570980387e-05, "loss": 1.8925, "step": 180 }, { "epoch": 0.04370924897367785, "grad_norm": 0.2822973430156708, "learning_rate": 9.985616668495859e-05, "loss": 1.7492, "step": 181 }, { "epoch": 0.04395073653706834, "grad_norm": 0.28814923763275146, "learning_rate": 9.985321777915497e-05, "loss": 1.669, "step": 182 }, { "epoch": 0.044192224100458824, "grad_norm": 0.30659806728363037, "learning_rate": 9.985023899416041e-05, "loss": 1.7437, "step": 183 }, { "epoch": 0.04443371166384931, "grad_norm": 0.4332549273967743, "learning_rate": 9.984723033176026e-05, "loss": 2.4215, "step": 184 }, { "epoch": 0.044675199227239795, "grad_norm": 0.29047444462776184, "learning_rate": 9.984419179375773e-05, "loss": 1.8027, "step": 185 }, { "epoch": 0.04491668679063028, "grad_norm": 0.3238014876842499, "learning_rate": 9.984112338197392e-05, "loss": 1.9597, "step": 186 }, { "epoch": 0.045158174354020766, "grad_norm": 0.29162389039993286, "learning_rate": 9.983802509824786e-05, "loss": 1.8871, "step": 187 }, { "epoch": 0.04539966191741125, "grad_norm": 0.3024207651615143, "learning_rate": 9.983489694443652e-05, "loss": 1.7083, "step": 188 }, { "epoch": 0.04564114948080174, "grad_norm": 0.31812480092048645, "learning_rate": 9.983173892241472e-05, "loss": 1.8127, "step": 189 }, { "epoch": 0.04588263704419222, "grad_norm": 0.2975609004497528, "learning_rate": 9.982855103407521e-05, "loss": 1.8012, "step": 190 }, { "epoch": 0.04612412460758271, "grad_norm": 0.29555127024650574, "learning_rate": 9.982533328132862e-05, "loss": 1.8123, "step": 191 }, { "epoch": 0.046365612170973194, "grad_norm": 0.35191309452056885, "learning_rate": 9.982208566610348e-05, "loss": 1.8852, "step": 192 }, { "epoch": 0.04660709973436368, "grad_norm": 0.2966957688331604, "learning_rate": 9.981880819034624e-05, "loss": 1.807, "step": 193 }, { "epoch": 0.046848587297754166, "grad_norm": 0.3028441071510315, "learning_rate": 9.981550085602123e-05, "loss": 1.8438, "step": 194 }, { "epoch": 0.04709007486114465, "grad_norm": 0.2984451353549957, "learning_rate": 9.981216366511067e-05, "loss": 1.8438, "step": 195 }, { "epoch": 0.04733156242453514, "grad_norm": 0.30961814522743225, "learning_rate": 9.980879661961469e-05, "loss": 2.0188, "step": 196 }, { "epoch": 0.04757304998792562, "grad_norm": 0.2839982807636261, "learning_rate": 9.98053997215513e-05, "loss": 1.6599, "step": 197 }, { "epoch": 0.04781453755131611, "grad_norm": 0.3275313377380371, "learning_rate": 9.980197297295644e-05, "loss": 2.0287, "step": 198 }, { "epoch": 0.048056025114706594, "grad_norm": 0.27704620361328125, "learning_rate": 9.979851637588386e-05, "loss": 1.67, "step": 199 }, { "epoch": 0.04829751267809708, "grad_norm": 0.3120492696762085, "learning_rate": 9.979502993240528e-05, "loss": 2.0147, "step": 200 }, { "epoch": 0.048539000241487565, "grad_norm": 1.1097334623336792, "learning_rate": 9.979151364461025e-05, "loss": 1.5357, "step": 201 }, { "epoch": 0.04878048780487805, "grad_norm": 0.3085632622241974, "learning_rate": 9.978796751460627e-05, "loss": 1.8097, "step": 202 }, { "epoch": 0.049021975368268536, "grad_norm": 0.2954683303833008, "learning_rate": 9.978439154451867e-05, "loss": 1.7476, "step": 203 }, { "epoch": 0.04926346293165902, "grad_norm": 0.36696991324424744, "learning_rate": 9.978078573649068e-05, "loss": 1.8692, "step": 204 }, { "epoch": 0.04950495049504951, "grad_norm": 0.2759391963481903, "learning_rate": 9.977715009268343e-05, "loss": 1.8649, "step": 205 }, { "epoch": 0.04974643805843999, "grad_norm": 0.28329887986183167, "learning_rate": 9.97734846152759e-05, "loss": 1.7726, "step": 206 }, { "epoch": 0.04998792562183048, "grad_norm": 0.2912246286869049, "learning_rate": 9.976978930646497e-05, "loss": 1.7696, "step": 207 }, { "epoch": 0.050229413185220964, "grad_norm": 0.2736220061779022, "learning_rate": 9.976606416846542e-05, "loss": 1.6264, "step": 208 }, { "epoch": 0.05047090074861145, "grad_norm": 0.3122333586215973, "learning_rate": 9.976230920350985e-05, "loss": 1.7356, "step": 209 }, { "epoch": 0.050712388312001935, "grad_norm": 0.3272218406200409, "learning_rate": 9.975852441384882e-05, "loss": 1.8347, "step": 210 }, { "epoch": 0.050953875875392414, "grad_norm": 0.28549280762672424, "learning_rate": 9.975470980175068e-05, "loss": 1.8455, "step": 211 }, { "epoch": 0.0511953634387829, "grad_norm": 0.3388885259628296, "learning_rate": 9.975086536950171e-05, "loss": 1.7048, "step": 212 }, { "epoch": 0.051436851002173385, "grad_norm": 0.2744766175746918, "learning_rate": 9.974699111940605e-05, "loss": 1.7144, "step": 213 }, { "epoch": 0.05167833856556387, "grad_norm": 0.28307104110717773, "learning_rate": 9.97430870537857e-05, "loss": 1.6624, "step": 214 }, { "epoch": 0.051919826128954356, "grad_norm": 0.31305596232414246, "learning_rate": 9.973915317498053e-05, "loss": 1.9547, "step": 215 }, { "epoch": 0.05216131369234484, "grad_norm": 0.28607913851737976, "learning_rate": 9.973518948534826e-05, "loss": 1.7163, "step": 216 }, { "epoch": 0.05240280125573533, "grad_norm": 0.28641268610954285, "learning_rate": 9.973119598726456e-05, "loss": 1.7557, "step": 217 }, { "epoch": 0.05264428881912581, "grad_norm": 0.2846461534500122, "learning_rate": 9.972717268312286e-05, "loss": 1.835, "step": 218 }, { "epoch": 0.0528857763825163, "grad_norm": 0.30998507142066956, "learning_rate": 9.972311957533452e-05, "loss": 1.8987, "step": 219 }, { "epoch": 0.053127263945906784, "grad_norm": 0.28558000922203064, "learning_rate": 9.971903666632873e-05, "loss": 1.8831, "step": 220 }, { "epoch": 0.05336875150929727, "grad_norm": 0.3230333924293518, "learning_rate": 9.971492395855256e-05, "loss": 1.9183, "step": 221 }, { "epoch": 0.053610239072687756, "grad_norm": 0.2764125466346741, "learning_rate": 9.971078145447092e-05, "loss": 1.6978, "step": 222 }, { "epoch": 0.05385172663607824, "grad_norm": 0.2793998122215271, "learning_rate": 9.970660915656664e-05, "loss": 1.8111, "step": 223 }, { "epoch": 0.05409321419946873, "grad_norm": 0.30916768312454224, "learning_rate": 9.970240706734031e-05, "loss": 1.8777, "step": 224 }, { "epoch": 0.05433470176285921, "grad_norm": 0.28356650471687317, "learning_rate": 9.969817518931043e-05, "loss": 1.8503, "step": 225 }, { "epoch": 0.0545761893262497, "grad_norm": 0.28640586137771606, "learning_rate": 9.969391352501336e-05, "loss": 1.9375, "step": 226 }, { "epoch": 0.054817676889640184, "grad_norm": 0.2708923816680908, "learning_rate": 9.96896220770033e-05, "loss": 1.693, "step": 227 }, { "epoch": 0.05505916445303067, "grad_norm": 0.3123991787433624, "learning_rate": 9.96853008478523e-05, "loss": 1.9336, "step": 228 }, { "epoch": 0.055300652016421155, "grad_norm": 0.2919258773326874, "learning_rate": 9.968094984015025e-05, "loss": 1.8187, "step": 229 }, { "epoch": 0.05554213957981164, "grad_norm": 0.2894388735294342, "learning_rate": 9.96765690565049e-05, "loss": 1.7333, "step": 230 }, { "epoch": 0.055783627143202126, "grad_norm": 0.30706238746643066, "learning_rate": 9.967215849954185e-05, "loss": 2.0766, "step": 231 }, { "epoch": 0.05602511470659261, "grad_norm": 0.2915318012237549, "learning_rate": 9.966771817190455e-05, "loss": 1.8268, "step": 232 }, { "epoch": 0.0562666022699831, "grad_norm": 0.28697696328163147, "learning_rate": 9.966324807625424e-05, "loss": 1.8361, "step": 233 }, { "epoch": 0.05650808983337358, "grad_norm": 0.3011184334754944, "learning_rate": 9.965874821527008e-05, "loss": 1.9076, "step": 234 }, { "epoch": 0.05674957739676407, "grad_norm": 0.2749825119972229, "learning_rate": 9.965421859164903e-05, "loss": 1.6789, "step": 235 }, { "epoch": 0.056991064960154554, "grad_norm": 0.281316876411438, "learning_rate": 9.964965920810586e-05, "loss": 1.735, "step": 236 }, { "epoch": 0.05723255252354504, "grad_norm": 0.2849574387073517, "learning_rate": 9.964507006737322e-05, "loss": 1.8415, "step": 237 }, { "epoch": 0.057474040086935525, "grad_norm": 0.26775339245796204, "learning_rate": 9.964045117220161e-05, "loss": 1.7411, "step": 238 }, { "epoch": 0.05771552765032601, "grad_norm": 0.3151552379131317, "learning_rate": 9.96358025253593e-05, "loss": 1.7063, "step": 239 }, { "epoch": 0.057957015213716497, "grad_norm": 0.2982829213142395, "learning_rate": 9.963112412963242e-05, "loss": 1.8773, "step": 240 }, { "epoch": 0.05819850277710698, "grad_norm": 0.2909030020236969, "learning_rate": 9.962641598782495e-05, "loss": 1.891, "step": 241 }, { "epoch": 0.05843999034049747, "grad_norm": 0.3036896288394928, "learning_rate": 9.962167810275869e-05, "loss": 1.9357, "step": 242 }, { "epoch": 0.058681477903887946, "grad_norm": 0.2985946536064148, "learning_rate": 9.961691047727324e-05, "loss": 1.837, "step": 243 }, { "epoch": 0.05892296546727843, "grad_norm": 0.278512567281723, "learning_rate": 9.961211311422607e-05, "loss": 1.7386, "step": 244 }, { "epoch": 0.05916445303066892, "grad_norm": 0.2755807340145111, "learning_rate": 9.960728601649242e-05, "loss": 1.6836, "step": 245 }, { "epoch": 0.0594059405940594, "grad_norm": 0.293917715549469, "learning_rate": 9.960242918696539e-05, "loss": 1.8625, "step": 246 }, { "epoch": 0.05964742815744989, "grad_norm": 0.28448382019996643, "learning_rate": 9.95975426285559e-05, "loss": 1.7909, "step": 247 }, { "epoch": 0.059888915720840374, "grad_norm": 0.2795000970363617, "learning_rate": 9.959262634419266e-05, "loss": 1.6246, "step": 248 }, { "epoch": 0.06013040328423086, "grad_norm": 0.29517507553100586, "learning_rate": 9.958768033682222e-05, "loss": 1.7155, "step": 249 }, { "epoch": 0.060371890847621346, "grad_norm": 0.27212411165237427, "learning_rate": 9.958270460940893e-05, "loss": 1.7492, "step": 250 }, { "epoch": 0.06061337841101183, "grad_norm": 0.29364678263664246, "learning_rate": 9.957769916493496e-05, "loss": 1.7754, "step": 251 }, { "epoch": 0.06085486597440232, "grad_norm": 0.29167085886001587, "learning_rate": 9.95726640064003e-05, "loss": 1.5348, "step": 252 }, { "epoch": 0.0610963535377928, "grad_norm": 0.2989068925380707, "learning_rate": 9.956759913682272e-05, "loss": 1.7296, "step": 253 }, { "epoch": 0.06133784110118329, "grad_norm": 0.2934739887714386, "learning_rate": 9.956250455923784e-05, "loss": 1.8387, "step": 254 }, { "epoch": 0.061579328664573774, "grad_norm": 0.2777479588985443, "learning_rate": 9.955738027669903e-05, "loss": 1.7859, "step": 255 }, { "epoch": 0.06182081622796426, "grad_norm": 0.3208652436733246, "learning_rate": 9.955222629227752e-05, "loss": 1.8396, "step": 256 }, { "epoch": 0.062062303791354745, "grad_norm": 0.2889414131641388, "learning_rate": 9.95470426090623e-05, "loss": 1.7689, "step": 257 }, { "epoch": 0.06230379135474523, "grad_norm": 0.29929113388061523, "learning_rate": 9.95418292301602e-05, "loss": 1.7447, "step": 258 }, { "epoch": 0.06254527891813572, "grad_norm": 0.304668664932251, "learning_rate": 9.953658615869579e-05, "loss": 1.8408, "step": 259 }, { "epoch": 0.0627867664815262, "grad_norm": 0.26577287912368774, "learning_rate": 9.953131339781149e-05, "loss": 1.7391, "step": 260 }, { "epoch": 0.06302825404491669, "grad_norm": 0.26456665992736816, "learning_rate": 9.952601095066749e-05, "loss": 1.588, "step": 261 }, { "epoch": 0.06326974160830717, "grad_norm": 0.2887117862701416, "learning_rate": 9.952067882044177e-05, "loss": 1.7707, "step": 262 }, { "epoch": 0.06351122917169766, "grad_norm": 0.2779730260372162, "learning_rate": 9.951531701033012e-05, "loss": 1.7844, "step": 263 }, { "epoch": 0.06375271673508814, "grad_norm": 0.27247482538223267, "learning_rate": 9.95099255235461e-05, "loss": 1.6796, "step": 264 }, { "epoch": 0.06399420429847863, "grad_norm": 0.2854093611240387, "learning_rate": 9.950450436332103e-05, "loss": 1.7844, "step": 265 }, { "epoch": 0.06423569186186912, "grad_norm": 0.3516107201576233, "learning_rate": 9.94990535329041e-05, "loss": 2.0682, "step": 266 }, { "epoch": 0.0644771794252596, "grad_norm": 0.2657569944858551, "learning_rate": 9.94935730355622e-05, "loss": 1.796, "step": 267 }, { "epoch": 0.06471866698865009, "grad_norm": 0.29638686776161194, "learning_rate": 9.948806287458003e-05, "loss": 1.7465, "step": 268 }, { "epoch": 0.06496015455204057, "grad_norm": 0.28134340047836304, "learning_rate": 9.948252305326008e-05, "loss": 1.5347, "step": 269 }, { "epoch": 0.06520164211543106, "grad_norm": 0.2832060158252716, "learning_rate": 9.947695357492258e-05, "loss": 1.8296, "step": 270 }, { "epoch": 0.06544312967882154, "grad_norm": 0.3167952597141266, "learning_rate": 9.947135444290558e-05, "loss": 2.1621, "step": 271 }, { "epoch": 0.06568461724221203, "grad_norm": 0.28933095932006836, "learning_rate": 9.946572566056487e-05, "loss": 1.764, "step": 272 }, { "epoch": 0.06592610480560251, "grad_norm": 0.30958911776542664, "learning_rate": 9.946006723127403e-05, "loss": 1.9259, "step": 273 }, { "epoch": 0.066167592368993, "grad_norm": 0.27871811389923096, "learning_rate": 9.945437915842441e-05, "loss": 1.8828, "step": 274 }, { "epoch": 0.06640907993238349, "grad_norm": 0.2978667914867401, "learning_rate": 9.944866144542511e-05, "loss": 1.8764, "step": 275 }, { "epoch": 0.06665056749577397, "grad_norm": 0.27303117513656616, "learning_rate": 9.9442914095703e-05, "loss": 1.7793, "step": 276 }, { "epoch": 0.06689205505916446, "grad_norm": 0.28109732270240784, "learning_rate": 9.94371371127027e-05, "loss": 1.7233, "step": 277 }, { "epoch": 0.06713354262255494, "grad_norm": 0.29150810837745667, "learning_rate": 9.943133049988663e-05, "loss": 1.8243, "step": 278 }, { "epoch": 0.06737503018594543, "grad_norm": 0.3005611300468445, "learning_rate": 9.942549426073494e-05, "loss": 1.8877, "step": 279 }, { "epoch": 0.06761651774933591, "grad_norm": 0.2712879478931427, "learning_rate": 9.941962839874555e-05, "loss": 1.8297, "step": 280 }, { "epoch": 0.0678580053127264, "grad_norm": 0.29506200551986694, "learning_rate": 9.94137329174341e-05, "loss": 1.746, "step": 281 }, { "epoch": 0.06809949287611688, "grad_norm": 0.2840847671031952, "learning_rate": 9.940780782033401e-05, "loss": 1.7606, "step": 282 }, { "epoch": 0.06834098043950737, "grad_norm": 0.2767079472541809, "learning_rate": 9.940185311099648e-05, "loss": 1.7573, "step": 283 }, { "epoch": 0.06858246800289786, "grad_norm": 0.28070494532585144, "learning_rate": 9.939586879299038e-05, "loss": 1.7798, "step": 284 }, { "epoch": 0.06882395556628834, "grad_norm": 0.2791160047054291, "learning_rate": 9.938985486990241e-05, "loss": 1.8631, "step": 285 }, { "epoch": 0.06906544312967883, "grad_norm": 0.27911651134490967, "learning_rate": 9.938381134533695e-05, "loss": 1.7894, "step": 286 }, { "epoch": 0.06930693069306931, "grad_norm": 0.2858154773712158, "learning_rate": 9.937773822291617e-05, "loss": 1.8625, "step": 287 }, { "epoch": 0.0695484182564598, "grad_norm": 0.2822221517562866, "learning_rate": 9.93716355062799e-05, "loss": 1.8163, "step": 288 }, { "epoch": 0.06978990581985028, "grad_norm": 0.28583386540412903, "learning_rate": 9.936550319908583e-05, "loss": 1.7072, "step": 289 }, { "epoch": 0.07003139338324077, "grad_norm": 0.27519020438194275, "learning_rate": 9.935934130500927e-05, "loss": 1.7768, "step": 290 }, { "epoch": 0.07027288094663126, "grad_norm": 0.2802446186542511, "learning_rate": 9.935314982774335e-05, "loss": 1.8043, "step": 291 }, { "epoch": 0.07051436851002173, "grad_norm": 0.2888818085193634, "learning_rate": 9.934692877099885e-05, "loss": 1.8006, "step": 292 }, { "epoch": 0.07075585607341221, "grad_norm": 0.2915632128715515, "learning_rate": 9.934067813850434e-05, "loss": 1.8006, "step": 293 }, { "epoch": 0.0709973436368027, "grad_norm": 0.281309574842453, "learning_rate": 9.93343979340061e-05, "loss": 1.8207, "step": 294 }, { "epoch": 0.07123883120019318, "grad_norm": 0.2658507525920868, "learning_rate": 9.93280881612681e-05, "loss": 1.8466, "step": 295 }, { "epoch": 0.07148031876358367, "grad_norm": 0.2695015072822571, "learning_rate": 9.932174882407209e-05, "loss": 1.7938, "step": 296 }, { "epoch": 0.07172180632697416, "grad_norm": 0.28267964720726013, "learning_rate": 9.931537992621749e-05, "loss": 1.859, "step": 297 }, { "epoch": 0.07196329389036464, "grad_norm": 0.3001064360141754, "learning_rate": 9.930898147152148e-05, "loss": 1.8308, "step": 298 }, { "epoch": 0.07220478145375513, "grad_norm": 0.2865334153175354, "learning_rate": 9.93025534638189e-05, "loss": 1.8846, "step": 299 }, { "epoch": 0.07244626901714561, "grad_norm": 0.29224735498428345, "learning_rate": 9.929609590696236e-05, "loss": 1.974, "step": 300 }, { "epoch": 0.0726877565805361, "grad_norm": 0.29061272740364075, "learning_rate": 9.928960880482214e-05, "loss": 1.891, "step": 301 }, { "epoch": 0.07292924414392658, "grad_norm": 0.27877870202064514, "learning_rate": 9.928309216128623e-05, "loss": 1.6629, "step": 302 }, { "epoch": 0.07317073170731707, "grad_norm": 0.27829432487487793, "learning_rate": 9.927654598026035e-05, "loss": 1.9232, "step": 303 }, { "epoch": 0.07341221927070755, "grad_norm": 0.28853073716163635, "learning_rate": 9.926997026566793e-05, "loss": 1.6872, "step": 304 }, { "epoch": 0.07365370683409804, "grad_norm": 0.3017938435077667, "learning_rate": 9.926336502145004e-05, "loss": 1.834, "step": 305 }, { "epoch": 0.07389519439748853, "grad_norm": 0.3079606294631958, "learning_rate": 9.92567302515655e-05, "loss": 1.9919, "step": 306 }, { "epoch": 0.07413668196087901, "grad_norm": 0.284600168466568, "learning_rate": 9.925006595999083e-05, "loss": 1.8107, "step": 307 }, { "epoch": 0.0743781695242695, "grad_norm": 0.3207036852836609, "learning_rate": 9.924337215072023e-05, "loss": 1.8985, "step": 308 }, { "epoch": 0.07461965708765998, "grad_norm": 0.27426770329475403, "learning_rate": 9.923664882776557e-05, "loss": 1.7316, "step": 309 }, { "epoch": 0.07486114465105047, "grad_norm": 0.30607980489730835, "learning_rate": 9.922989599515644e-05, "loss": 1.9892, "step": 310 }, { "epoch": 0.07510263221444095, "grad_norm": 0.2927166819572449, "learning_rate": 9.922311365694011e-05, "loss": 1.8746, "step": 311 }, { "epoch": 0.07534411977783144, "grad_norm": 0.2725332975387573, "learning_rate": 9.921630181718152e-05, "loss": 1.7773, "step": 312 }, { "epoch": 0.07558560734122192, "grad_norm": 0.2871837317943573, "learning_rate": 9.92094604799633e-05, "loss": 1.9641, "step": 313 }, { "epoch": 0.07582709490461241, "grad_norm": 0.284435510635376, "learning_rate": 9.920258964938576e-05, "loss": 1.8081, "step": 314 }, { "epoch": 0.0760685824680029, "grad_norm": 0.27971702814102173, "learning_rate": 9.91956893295669e-05, "loss": 1.8659, "step": 315 }, { "epoch": 0.07631007003139338, "grad_norm": 0.29805755615234375, "learning_rate": 9.918875952464234e-05, "loss": 2.0772, "step": 316 }, { "epoch": 0.07655155759478387, "grad_norm": 0.2688223421573639, "learning_rate": 9.918180023876548e-05, "loss": 1.7697, "step": 317 }, { "epoch": 0.07679304515817435, "grad_norm": 0.2922739088535309, "learning_rate": 9.917481147610725e-05, "loss": 1.8815, "step": 318 }, { "epoch": 0.07703453272156484, "grad_norm": 0.2920701801776886, "learning_rate": 9.916779324085636e-05, "loss": 1.6683, "step": 319 }, { "epoch": 0.07727602028495532, "grad_norm": 0.2858518958091736, "learning_rate": 9.916074553721915e-05, "loss": 1.7205, "step": 320 }, { "epoch": 0.07751750784834581, "grad_norm": 0.2882474660873413, "learning_rate": 9.915366836941957e-05, "loss": 1.8518, "step": 321 }, { "epoch": 0.0777589954117363, "grad_norm": 0.2800799310207367, "learning_rate": 9.91465617416993e-05, "loss": 1.7858, "step": 322 }, { "epoch": 0.07800048297512678, "grad_norm": 0.30860286951065063, "learning_rate": 9.913942565831766e-05, "loss": 2.0491, "step": 323 }, { "epoch": 0.07824197053851727, "grad_norm": 0.2933506965637207, "learning_rate": 9.91322601235516e-05, "loss": 1.8936, "step": 324 }, { "epoch": 0.07848345810190775, "grad_norm": 0.2754722237586975, "learning_rate": 9.912506514169573e-05, "loss": 1.7845, "step": 325 }, { "epoch": 0.07872494566529824, "grad_norm": 0.2775117754936218, "learning_rate": 9.911784071706233e-05, "loss": 1.6883, "step": 326 }, { "epoch": 0.07896643322868872, "grad_norm": 0.27762269973754883, "learning_rate": 9.91105868539813e-05, "loss": 1.7333, "step": 327 }, { "epoch": 0.07920792079207921, "grad_norm": 0.27511125802993774, "learning_rate": 9.910330355680018e-05, "loss": 1.7933, "step": 328 }, { "epoch": 0.0794494083554697, "grad_norm": 0.27266359329223633, "learning_rate": 9.909599082988419e-05, "loss": 1.8126, "step": 329 }, { "epoch": 0.07969089591886018, "grad_norm": 0.28638163208961487, "learning_rate": 9.908864867761616e-05, "loss": 1.7719, "step": 330 }, { "epoch": 0.07993238348225067, "grad_norm": 0.287117063999176, "learning_rate": 9.908127710439654e-05, "loss": 1.7776, "step": 331 }, { "epoch": 0.08017387104564115, "grad_norm": 0.2886422574520111, "learning_rate": 9.907387611464347e-05, "loss": 1.7783, "step": 332 }, { "epoch": 0.08041535860903164, "grad_norm": 0.28264063596725464, "learning_rate": 9.906644571279265e-05, "loss": 1.7754, "step": 333 }, { "epoch": 0.08065684617242212, "grad_norm": 0.2666427493095398, "learning_rate": 9.905898590329744e-05, "loss": 1.5789, "step": 334 }, { "epoch": 0.08089833373581261, "grad_norm": 0.3525300621986389, "learning_rate": 9.905149669062884e-05, "loss": 2.2147, "step": 335 }, { "epoch": 0.0811398212992031, "grad_norm": 0.28527653217315674, "learning_rate": 9.904397807927546e-05, "loss": 1.6824, "step": 336 }, { "epoch": 0.08138130886259358, "grad_norm": 0.3258638381958008, "learning_rate": 9.903643007374352e-05, "loss": 1.9276, "step": 337 }, { "epoch": 0.08162279642598406, "grad_norm": 0.2898681163787842, "learning_rate": 9.902885267855689e-05, "loss": 1.8823, "step": 338 }, { "epoch": 0.08186428398937455, "grad_norm": 0.28534162044525146, "learning_rate": 9.902124589825701e-05, "loss": 1.7955, "step": 339 }, { "epoch": 0.08210577155276504, "grad_norm": 0.29654279351234436, "learning_rate": 9.901360973740295e-05, "loss": 1.8217, "step": 340 }, { "epoch": 0.08234725911615552, "grad_norm": 0.27160125970840454, "learning_rate": 9.900594420057139e-05, "loss": 1.7292, "step": 341 }, { "epoch": 0.08258874667954601, "grad_norm": 0.28140437602996826, "learning_rate": 9.899824929235663e-05, "loss": 1.883, "step": 342 }, { "epoch": 0.08283023424293649, "grad_norm": 0.2921160161495209, "learning_rate": 9.899052501737056e-05, "loss": 1.9536, "step": 343 }, { "epoch": 0.08307172180632698, "grad_norm": 0.2749026417732239, "learning_rate": 9.898277138024268e-05, "loss": 1.7172, "step": 344 }, { "epoch": 0.08331320936971746, "grad_norm": 0.2894373834133148, "learning_rate": 9.897498838562006e-05, "loss": 1.9665, "step": 345 }, { "epoch": 0.08355469693310795, "grad_norm": 0.312502920627594, "learning_rate": 9.89671760381674e-05, "loss": 1.8277, "step": 346 }, { "epoch": 0.08379618449649844, "grad_norm": 0.281454861164093, "learning_rate": 9.895933434256696e-05, "loss": 1.6576, "step": 347 }, { "epoch": 0.08403767205988892, "grad_norm": 0.280256062746048, "learning_rate": 9.895146330351865e-05, "loss": 1.9035, "step": 348 }, { "epoch": 0.0842791596232794, "grad_norm": 0.2698526382446289, "learning_rate": 9.894356292573988e-05, "loss": 1.7014, "step": 349 }, { "epoch": 0.08452064718666989, "grad_norm": 0.2966681718826294, "learning_rate": 9.893563321396568e-05, "loss": 1.9057, "step": 350 }, { "epoch": 0.08476213475006038, "grad_norm": 0.2751750946044922, "learning_rate": 9.892767417294872e-05, "loss": 1.9109, "step": 351 }, { "epoch": 0.08500362231345086, "grad_norm": 0.27698180079460144, "learning_rate": 9.891968580745917e-05, "loss": 1.7143, "step": 352 }, { "epoch": 0.08524510987684135, "grad_norm": 0.30382975935935974, "learning_rate": 9.89116681222848e-05, "loss": 1.8529, "step": 353 }, { "epoch": 0.08548659744023183, "grad_norm": 0.28126025199890137, "learning_rate": 9.890362112223094e-05, "loss": 1.8243, "step": 354 }, { "epoch": 0.08572808500362232, "grad_norm": 0.26443248987197876, "learning_rate": 9.889554481212056e-05, "loss": 1.5424, "step": 355 }, { "epoch": 0.08596957256701279, "grad_norm": 0.302097886800766, "learning_rate": 9.888743919679409e-05, "loss": 1.7993, "step": 356 }, { "epoch": 0.08621106013040328, "grad_norm": 0.2876918911933899, "learning_rate": 9.887930428110959e-05, "loss": 1.8625, "step": 357 }, { "epoch": 0.08645254769379376, "grad_norm": 0.2955901324748993, "learning_rate": 9.887114006994267e-05, "loss": 1.7146, "step": 358 }, { "epoch": 0.08669403525718425, "grad_norm": 0.2705610394477844, "learning_rate": 9.886294656818651e-05, "loss": 1.8299, "step": 359 }, { "epoch": 0.08693552282057473, "grad_norm": 0.2755456566810608, "learning_rate": 9.88547237807518e-05, "loss": 1.8367, "step": 360 }, { "epoch": 0.08717701038396522, "grad_norm": 0.29839128255844116, "learning_rate": 9.884647171256683e-05, "loss": 1.7689, "step": 361 }, { "epoch": 0.0874184979473557, "grad_norm": 0.27627283334732056, "learning_rate": 9.883819036857743e-05, "loss": 1.7107, "step": 362 }, { "epoch": 0.08765998551074619, "grad_norm": 0.2801618278026581, "learning_rate": 9.882987975374696e-05, "loss": 1.8248, "step": 363 }, { "epoch": 0.08790147307413668, "grad_norm": 0.2711143493652344, "learning_rate": 9.882153987305632e-05, "loss": 1.6291, "step": 364 }, { "epoch": 0.08814296063752716, "grad_norm": 0.2989214062690735, "learning_rate": 9.881317073150396e-05, "loss": 1.8169, "step": 365 }, { "epoch": 0.08838444820091765, "grad_norm": 0.27399736642837524, "learning_rate": 9.880477233410587e-05, "loss": 1.7848, "step": 366 }, { "epoch": 0.08862593576430813, "grad_norm": 0.29605332016944885, "learning_rate": 9.879634468589561e-05, "loss": 1.8587, "step": 367 }, { "epoch": 0.08886742332769862, "grad_norm": 0.3036874234676361, "learning_rate": 9.87878877919242e-05, "loss": 1.8376, "step": 368 }, { "epoch": 0.0891089108910891, "grad_norm": 0.27524060010910034, "learning_rate": 9.877940165726022e-05, "loss": 1.7785, "step": 369 }, { "epoch": 0.08935039845447959, "grad_norm": 0.29827842116355896, "learning_rate": 9.87708862869898e-05, "loss": 1.8899, "step": 370 }, { "epoch": 0.08959188601787008, "grad_norm": 0.28207167983055115, "learning_rate": 9.876234168621654e-05, "loss": 1.818, "step": 371 }, { "epoch": 0.08983337358126056, "grad_norm": 0.2751006782054901, "learning_rate": 9.875376786006161e-05, "loss": 1.8081, "step": 372 }, { "epoch": 0.09007486114465105, "grad_norm": 0.2890884280204773, "learning_rate": 9.874516481366367e-05, "loss": 1.719, "step": 373 }, { "epoch": 0.09031634870804153, "grad_norm": 0.29088371992111206, "learning_rate": 9.87365325521789e-05, "loss": 1.7928, "step": 374 }, { "epoch": 0.09055783627143202, "grad_norm": 0.287952721118927, "learning_rate": 9.872787108078098e-05, "loss": 1.8332, "step": 375 }, { "epoch": 0.0907993238348225, "grad_norm": 0.2880597710609436, "learning_rate": 9.871918040466113e-05, "loss": 1.8148, "step": 376 }, { "epoch": 0.09104081139821299, "grad_norm": 0.2922658622264862, "learning_rate": 9.871046052902802e-05, "loss": 1.8887, "step": 377 }, { "epoch": 0.09128229896160348, "grad_norm": 0.3039756119251251, "learning_rate": 9.870171145910784e-05, "loss": 1.7143, "step": 378 }, { "epoch": 0.09152378652499396, "grad_norm": 0.27396541833877563, "learning_rate": 9.869293320014433e-05, "loss": 1.7626, "step": 379 }, { "epoch": 0.09176527408838445, "grad_norm": 0.30368801951408386, "learning_rate": 9.868412575739864e-05, "loss": 1.9307, "step": 380 }, { "epoch": 0.09200676165177493, "grad_norm": 0.2743094563484192, "learning_rate": 9.867528913614947e-05, "loss": 1.6689, "step": 381 }, { "epoch": 0.09224824921516542, "grad_norm": 0.2829776406288147, "learning_rate": 9.866642334169298e-05, "loss": 1.7646, "step": 382 }, { "epoch": 0.0924897367785559, "grad_norm": 0.27822762727737427, "learning_rate": 9.865752837934283e-05, "loss": 1.8207, "step": 383 }, { "epoch": 0.09273122434194639, "grad_norm": 0.2803753614425659, "learning_rate": 9.864860425443016e-05, "loss": 1.7833, "step": 384 }, { "epoch": 0.09297271190533687, "grad_norm": 0.28696754574775696, "learning_rate": 9.86396509723036e-05, "loss": 1.8914, "step": 385 }, { "epoch": 0.09321419946872736, "grad_norm": 0.27866771817207336, "learning_rate": 9.863066853832921e-05, "loss": 1.8337, "step": 386 }, { "epoch": 0.09345568703211785, "grad_norm": 0.2905522584915161, "learning_rate": 9.862165695789054e-05, "loss": 1.8121, "step": 387 }, { "epoch": 0.09369717459550833, "grad_norm": 0.2901919186115265, "learning_rate": 9.861261623638867e-05, "loss": 1.6512, "step": 388 }, { "epoch": 0.09393866215889882, "grad_norm": 0.29584378004074097, "learning_rate": 9.860354637924204e-05, "loss": 1.9358, "step": 389 }, { "epoch": 0.0941801497222893, "grad_norm": 0.27355754375457764, "learning_rate": 9.859444739188665e-05, "loss": 1.7889, "step": 390 }, { "epoch": 0.09442163728567979, "grad_norm": 0.28625258803367615, "learning_rate": 9.85853192797759e-05, "loss": 1.7989, "step": 391 }, { "epoch": 0.09466312484907027, "grad_norm": 0.2943178415298462, "learning_rate": 9.857616204838065e-05, "loss": 1.8327, "step": 392 }, { "epoch": 0.09490461241246076, "grad_norm": 0.2840021550655365, "learning_rate": 9.856697570318927e-05, "loss": 1.8274, "step": 393 }, { "epoch": 0.09514609997585124, "grad_norm": 0.28078603744506836, "learning_rate": 9.855776024970749e-05, "loss": 1.7316, "step": 394 }, { "epoch": 0.09538758753924173, "grad_norm": 0.27110886573791504, "learning_rate": 9.854851569345854e-05, "loss": 1.6548, "step": 395 }, { "epoch": 0.09562907510263222, "grad_norm": 0.2627575695514679, "learning_rate": 9.853924203998308e-05, "loss": 1.7053, "step": 396 }, { "epoch": 0.0958705626660227, "grad_norm": 0.28886744379997253, "learning_rate": 9.852993929483922e-05, "loss": 1.6716, "step": 397 }, { "epoch": 0.09611205022941319, "grad_norm": 0.2770794928073883, "learning_rate": 9.85206074636025e-05, "loss": 1.7156, "step": 398 }, { "epoch": 0.09635353779280367, "grad_norm": 0.2742690443992615, "learning_rate": 9.85112465518659e-05, "loss": 1.6733, "step": 399 }, { "epoch": 0.09659502535619416, "grad_norm": 0.28999531269073486, "learning_rate": 9.850185656523978e-05, "loss": 1.8832, "step": 400 }, { "epoch": 0.09683651291958464, "grad_norm": 0.2888575494289398, "learning_rate": 9.849243750935202e-05, "loss": 1.8296, "step": 401 }, { "epoch": 0.09707800048297513, "grad_norm": 0.2710040807723999, "learning_rate": 9.848298938984781e-05, "loss": 1.7656, "step": 402 }, { "epoch": 0.09731948804636562, "grad_norm": 0.2809413969516754, "learning_rate": 9.847351221238987e-05, "loss": 1.7886, "step": 403 }, { "epoch": 0.0975609756097561, "grad_norm": 0.2801872789859772, "learning_rate": 9.846400598265824e-05, "loss": 1.7092, "step": 404 }, { "epoch": 0.09780246317314659, "grad_norm": 0.2998405396938324, "learning_rate": 9.845447070635046e-05, "loss": 1.8165, "step": 405 }, { "epoch": 0.09804395073653707, "grad_norm": 0.33238905668258667, "learning_rate": 9.844490638918138e-05, "loss": 2.0468, "step": 406 }, { "epoch": 0.09828543829992756, "grad_norm": 0.2929340600967407, "learning_rate": 9.843531303688335e-05, "loss": 1.8788, "step": 407 }, { "epoch": 0.09852692586331804, "grad_norm": 0.2869876027107239, "learning_rate": 9.842569065520606e-05, "loss": 1.8086, "step": 408 }, { "epoch": 0.09876841342670853, "grad_norm": 0.28498587012290955, "learning_rate": 9.841603924991663e-05, "loss": 1.6979, "step": 409 }, { "epoch": 0.09900990099009901, "grad_norm": 0.2809644341468811, "learning_rate": 9.840635882679956e-05, "loss": 1.7499, "step": 410 }, { "epoch": 0.0992513885534895, "grad_norm": 0.2747531533241272, "learning_rate": 9.839664939165675e-05, "loss": 1.7549, "step": 411 }, { "epoch": 0.09949287611687999, "grad_norm": 0.2902829349040985, "learning_rate": 9.83869109503075e-05, "loss": 1.8995, "step": 412 }, { "epoch": 0.09973436368027047, "grad_norm": 0.2756800949573517, "learning_rate": 9.837714350858846e-05, "loss": 1.7057, "step": 413 }, { "epoch": 0.09997585124366096, "grad_norm": 0.289368599653244, "learning_rate": 9.836734707235366e-05, "loss": 1.933, "step": 414 }, { "epoch": 0.10021733880705144, "grad_norm": 0.29384687542915344, "learning_rate": 9.835752164747457e-05, "loss": 1.9249, "step": 415 }, { "epoch": 0.10045882637044193, "grad_norm": 0.2748492360115051, "learning_rate": 9.834766723983997e-05, "loss": 1.8197, "step": 416 }, { "epoch": 0.10070031393383241, "grad_norm": 0.29578012228012085, "learning_rate": 9.833778385535603e-05, "loss": 1.9807, "step": 417 }, { "epoch": 0.1009418014972229, "grad_norm": 0.2878894507884979, "learning_rate": 9.832787149994631e-05, "loss": 1.6623, "step": 418 }, { "epoch": 0.10118328906061339, "grad_norm": 0.28690582513809204, "learning_rate": 9.831793017955168e-05, "loss": 1.8893, "step": 419 }, { "epoch": 0.10142477662400387, "grad_norm": 0.27932021021842957, "learning_rate": 9.830795990013046e-05, "loss": 1.7858, "step": 420 }, { "epoch": 0.10166626418739434, "grad_norm": 0.27387359738349915, "learning_rate": 9.82979606676582e-05, "loss": 1.8473, "step": 421 }, { "epoch": 0.10190775175078483, "grad_norm": 0.26410290598869324, "learning_rate": 9.828793248812792e-05, "loss": 1.6531, "step": 422 }, { "epoch": 0.10214923931417531, "grad_norm": 0.2906973361968994, "learning_rate": 9.827787536754994e-05, "loss": 1.8561, "step": 423 }, { "epoch": 0.1023907268775658, "grad_norm": 0.28027960658073425, "learning_rate": 9.82677893119519e-05, "loss": 1.7624, "step": 424 }, { "epoch": 0.10263221444095628, "grad_norm": 0.26871955394744873, "learning_rate": 9.825767432737883e-05, "loss": 1.6901, "step": 425 }, { "epoch": 0.10287370200434677, "grad_norm": 0.28086093068122864, "learning_rate": 9.824753041989308e-05, "loss": 1.8209, "step": 426 }, { "epoch": 0.10311518956773726, "grad_norm": 0.2760232388973236, "learning_rate": 9.823735759557431e-05, "loss": 1.7899, "step": 427 }, { "epoch": 0.10335667713112774, "grad_norm": 0.27731311321258545, "learning_rate": 9.822715586051957e-05, "loss": 1.7834, "step": 428 }, { "epoch": 0.10359816469451823, "grad_norm": 0.26586011052131653, "learning_rate": 9.821692522084316e-05, "loss": 1.7721, "step": 429 }, { "epoch": 0.10383965225790871, "grad_norm": 0.2568017542362213, "learning_rate": 9.820666568267678e-05, "loss": 1.5376, "step": 430 }, { "epoch": 0.1040811398212992, "grad_norm": 0.29189375042915344, "learning_rate": 9.819637725216941e-05, "loss": 1.8376, "step": 431 }, { "epoch": 0.10432262738468968, "grad_norm": 0.28995007276535034, "learning_rate": 9.818605993548733e-05, "loss": 1.8012, "step": 432 }, { "epoch": 0.10456411494808017, "grad_norm": 0.2857837378978729, "learning_rate": 9.817571373881417e-05, "loss": 1.9442, "step": 433 }, { "epoch": 0.10480560251147066, "grad_norm": 0.28825604915618896, "learning_rate": 9.816533866835087e-05, "loss": 1.8442, "step": 434 }, { "epoch": 0.10504709007486114, "grad_norm": 0.28988251090049744, "learning_rate": 9.815493473031563e-05, "loss": 1.7784, "step": 435 }, { "epoch": 0.10528857763825163, "grad_norm": 0.2709440290927887, "learning_rate": 9.814450193094398e-05, "loss": 1.7199, "step": 436 }, { "epoch": 0.10553006520164211, "grad_norm": 0.2710571587085724, "learning_rate": 9.813404027648877e-05, "loss": 1.6485, "step": 437 }, { "epoch": 0.1057715527650326, "grad_norm": 0.29289767146110535, "learning_rate": 9.812354977322011e-05, "loss": 1.6867, "step": 438 }, { "epoch": 0.10601304032842308, "grad_norm": 0.27464166283607483, "learning_rate": 9.811303042742542e-05, "loss": 1.54, "step": 439 }, { "epoch": 0.10625452789181357, "grad_norm": 0.27407106757164, "learning_rate": 9.81024822454094e-05, "loss": 1.6428, "step": 440 }, { "epoch": 0.10649601545520405, "grad_norm": 0.29014259576797485, "learning_rate": 9.809190523349399e-05, "loss": 1.7654, "step": 441 }, { "epoch": 0.10673750301859454, "grad_norm": 0.2837926149368286, "learning_rate": 9.808129939801851e-05, "loss": 1.8607, "step": 442 }, { "epoch": 0.10697899058198503, "grad_norm": 0.3138989508152008, "learning_rate": 9.807066474533946e-05, "loss": 1.9194, "step": 443 }, { "epoch": 0.10722047814537551, "grad_norm": 0.2894023060798645, "learning_rate": 9.806000128183069e-05, "loss": 1.8462, "step": 444 }, { "epoch": 0.107461965708766, "grad_norm": 0.27192845940589905, "learning_rate": 9.80493090138832e-05, "loss": 1.8293, "step": 445 }, { "epoch": 0.10770345327215648, "grad_norm": 0.2810100018978119, "learning_rate": 9.803858794790541e-05, "loss": 1.8501, "step": 446 }, { "epoch": 0.10794494083554697, "grad_norm": 0.2710225284099579, "learning_rate": 9.802783809032288e-05, "loss": 1.7248, "step": 447 }, { "epoch": 0.10818642839893745, "grad_norm": 0.28958362340927124, "learning_rate": 9.801705944757846e-05, "loss": 1.7555, "step": 448 }, { "epoch": 0.10842791596232794, "grad_norm": 0.27755028009414673, "learning_rate": 9.800625202613225e-05, "loss": 1.8206, "step": 449 }, { "epoch": 0.10866940352571842, "grad_norm": 0.2801584005355835, "learning_rate": 9.799541583246163e-05, "loss": 1.764, "step": 450 }, { "epoch": 0.10891089108910891, "grad_norm": 0.2989168167114258, "learning_rate": 9.79845508730612e-05, "loss": 1.8531, "step": 451 }, { "epoch": 0.1091523786524994, "grad_norm": 0.28735461831092834, "learning_rate": 9.797365715444278e-05, "loss": 1.9467, "step": 452 }, { "epoch": 0.10939386621588988, "grad_norm": 0.2826710641384125, "learning_rate": 9.796273468313546e-05, "loss": 1.6993, "step": 453 }, { "epoch": 0.10963535377928037, "grad_norm": 0.3037896156311035, "learning_rate": 9.795178346568555e-05, "loss": 1.7925, "step": 454 }, { "epoch": 0.10987684134267085, "grad_norm": 0.27401891350746155, "learning_rate": 9.794080350865658e-05, "loss": 1.8314, "step": 455 }, { "epoch": 0.11011832890606134, "grad_norm": 0.2809910774230957, "learning_rate": 9.792979481862933e-05, "loss": 1.8314, "step": 456 }, { "epoch": 0.11035981646945182, "grad_norm": 0.27638763189315796, "learning_rate": 9.791875740220175e-05, "loss": 1.702, "step": 457 }, { "epoch": 0.11060130403284231, "grad_norm": 0.2893375754356384, "learning_rate": 9.79076912659891e-05, "loss": 1.8864, "step": 458 }, { "epoch": 0.1108427915962328, "grad_norm": 0.2902871072292328, "learning_rate": 9.789659641662373e-05, "loss": 1.8047, "step": 459 }, { "epoch": 0.11108427915962328, "grad_norm": 0.2804037928581238, "learning_rate": 9.788547286075531e-05, "loss": 1.7456, "step": 460 }, { "epoch": 0.11132576672301377, "grad_norm": 0.2759838402271271, "learning_rate": 9.787432060505064e-05, "loss": 1.7112, "step": 461 }, { "epoch": 0.11156725428640425, "grad_norm": 0.28496286273002625, "learning_rate": 9.786313965619376e-05, "loss": 1.7838, "step": 462 }, { "epoch": 0.11180874184979474, "grad_norm": 0.28257259726524353, "learning_rate": 9.78519300208859e-05, "loss": 1.8171, "step": 463 }, { "epoch": 0.11205022941318522, "grad_norm": 0.2747897505760193, "learning_rate": 9.784069170584547e-05, "loss": 1.7337, "step": 464 }, { "epoch": 0.11229171697657571, "grad_norm": 0.29796814918518066, "learning_rate": 9.78294247178081e-05, "loss": 1.977, "step": 465 }, { "epoch": 0.1125332045399662, "grad_norm": 0.2776499390602112, "learning_rate": 9.781812906352657e-05, "loss": 1.78, "step": 466 }, { "epoch": 0.11277469210335668, "grad_norm": 0.3614672124385834, "learning_rate": 9.780680474977083e-05, "loss": 1.7773, "step": 467 }, { "epoch": 0.11301617966674717, "grad_norm": 0.28927668929100037, "learning_rate": 9.779545178332807e-05, "loss": 1.8433, "step": 468 }, { "epoch": 0.11325766723013765, "grad_norm": 0.2913784086704254, "learning_rate": 9.778407017100259e-05, "loss": 1.9236, "step": 469 }, { "epoch": 0.11349915479352814, "grad_norm": 0.30389007925987244, "learning_rate": 9.77726599196159e-05, "loss": 1.7944, "step": 470 }, { "epoch": 0.11374064235691862, "grad_norm": 0.2783800959587097, "learning_rate": 9.776122103600665e-05, "loss": 1.7486, "step": 471 }, { "epoch": 0.11398212992030911, "grad_norm": 0.2779453992843628, "learning_rate": 9.774975352703065e-05, "loss": 1.6498, "step": 472 }, { "epoch": 0.1142236174836996, "grad_norm": 0.2728698253631592, "learning_rate": 9.773825739956089e-05, "loss": 1.7032, "step": 473 }, { "epoch": 0.11446510504709008, "grad_norm": 0.4645857512950897, "learning_rate": 9.772673266048747e-05, "loss": 1.9575, "step": 474 }, { "epoch": 0.11470659261048056, "grad_norm": 0.2809775471687317, "learning_rate": 9.771517931671769e-05, "loss": 1.7906, "step": 475 }, { "epoch": 0.11494808017387105, "grad_norm": 0.2997235953807831, "learning_rate": 9.770359737517599e-05, "loss": 1.8539, "step": 476 }, { "epoch": 0.11518956773726154, "grad_norm": 0.2767902910709381, "learning_rate": 9.769198684280387e-05, "loss": 1.7607, "step": 477 }, { "epoch": 0.11543105530065202, "grad_norm": 0.312592089176178, "learning_rate": 9.768034772656006e-05, "loss": 1.9156, "step": 478 }, { "epoch": 0.11567254286404251, "grad_norm": 0.2808191180229187, "learning_rate": 9.766868003342038e-05, "loss": 1.7381, "step": 479 }, { "epoch": 0.11591403042743299, "grad_norm": 0.28857922554016113, "learning_rate": 9.765698377037778e-05, "loss": 1.8443, "step": 480 }, { "epoch": 0.11615551799082348, "grad_norm": 0.28472453355789185, "learning_rate": 9.764525894444235e-05, "loss": 1.816, "step": 481 }, { "epoch": 0.11639700555421396, "grad_norm": 0.2768532335758209, "learning_rate": 9.763350556264128e-05, "loss": 1.7091, "step": 482 }, { "epoch": 0.11663849311760445, "grad_norm": 0.3824092447757721, "learning_rate": 9.762172363201887e-05, "loss": 1.6476, "step": 483 }, { "epoch": 0.11687998068099494, "grad_norm": 0.28541287779808044, "learning_rate": 9.760991315963655e-05, "loss": 1.7566, "step": 484 }, { "epoch": 0.11712146824438542, "grad_norm": 0.2787253260612488, "learning_rate": 9.759807415257285e-05, "loss": 1.8362, "step": 485 }, { "epoch": 0.11736295580777589, "grad_norm": 0.2788980305194855, "learning_rate": 9.75862066179234e-05, "loss": 1.7361, "step": 486 }, { "epoch": 0.11760444337116638, "grad_norm": 0.30717000365257263, "learning_rate": 9.75743105628009e-05, "loss": 2.0405, "step": 487 }, { "epoch": 0.11784593093455686, "grad_norm": 0.28835999965667725, "learning_rate": 9.756238599433521e-05, "loss": 1.8115, "step": 488 }, { "epoch": 0.11808741849794735, "grad_norm": 0.26419007778167725, "learning_rate": 9.755043291967321e-05, "loss": 1.5545, "step": 489 }, { "epoch": 0.11832890606133784, "grad_norm": 0.2989584505558014, "learning_rate": 9.753845134597892e-05, "loss": 1.9747, "step": 490 }, { "epoch": 0.11857039362472832, "grad_norm": 0.27482473850250244, "learning_rate": 9.752644128043338e-05, "loss": 1.7562, "step": 491 }, { "epoch": 0.1188118811881188, "grad_norm": 0.26998311281204224, "learning_rate": 9.751440273023478e-05, "loss": 1.6722, "step": 492 }, { "epoch": 0.11905336875150929, "grad_norm": 0.2935127019882202, "learning_rate": 9.750233570259833e-05, "loss": 1.8002, "step": 493 }, { "epoch": 0.11929485631489978, "grad_norm": 0.2886837422847748, "learning_rate": 9.749024020475632e-05, "loss": 1.8251, "step": 494 }, { "epoch": 0.11953634387829026, "grad_norm": 0.27557656168937683, "learning_rate": 9.747811624395811e-05, "loss": 1.6533, "step": 495 }, { "epoch": 0.11977783144168075, "grad_norm": 0.2803681194782257, "learning_rate": 9.746596382747011e-05, "loss": 1.7325, "step": 496 }, { "epoch": 0.12001931900507123, "grad_norm": 0.3031095862388611, "learning_rate": 9.745378296257579e-05, "loss": 1.8844, "step": 497 }, { "epoch": 0.12026080656846172, "grad_norm": 0.28464269638061523, "learning_rate": 9.744157365657567e-05, "loss": 1.8316, "step": 498 }, { "epoch": 0.1205022941318522, "grad_norm": 0.29071709513664246, "learning_rate": 9.742933591678731e-05, "loss": 1.7188, "step": 499 }, { "epoch": 0.12074378169524269, "grad_norm": 0.29018208384513855, "learning_rate": 9.741706975054532e-05, "loss": 1.6691, "step": 500 }, { "epoch": 0.12098526925863318, "grad_norm": 0.3093045949935913, "learning_rate": 9.740477516520135e-05, "loss": 1.9989, "step": 501 }, { "epoch": 0.12122675682202366, "grad_norm": 0.3274978697299957, "learning_rate": 9.739245216812407e-05, "loss": 1.9223, "step": 502 }, { "epoch": 0.12146824438541415, "grad_norm": 0.2763381004333496, "learning_rate": 9.738010076669919e-05, "loss": 1.8004, "step": 503 }, { "epoch": 0.12170973194880463, "grad_norm": 0.27695074677467346, "learning_rate": 9.736772096832945e-05, "loss": 1.8704, "step": 504 }, { "epoch": 0.12195121951219512, "grad_norm": 0.28391584753990173, "learning_rate": 9.735531278043457e-05, "loss": 1.8722, "step": 505 }, { "epoch": 0.1221927070755856, "grad_norm": 0.2790353000164032, "learning_rate": 9.734287621045133e-05, "loss": 1.7409, "step": 506 }, { "epoch": 0.12243419463897609, "grad_norm": 0.31203344464302063, "learning_rate": 9.733041126583351e-05, "loss": 1.8611, "step": 507 }, { "epoch": 0.12267568220236658, "grad_norm": 0.28117474913597107, "learning_rate": 9.73179179540519e-05, "loss": 1.7635, "step": 508 }, { "epoch": 0.12291716976575706, "grad_norm": 0.2965591251850128, "learning_rate": 9.730539628259425e-05, "loss": 1.7152, "step": 509 }, { "epoch": 0.12315865732914755, "grad_norm": 0.2742213010787964, "learning_rate": 9.729284625896536e-05, "loss": 1.7753, "step": 510 }, { "epoch": 0.12340014489253803, "grad_norm": 0.2835814952850342, "learning_rate": 9.7280267890687e-05, "loss": 1.8058, "step": 511 }, { "epoch": 0.12364163245592852, "grad_norm": 0.27745455503463745, "learning_rate": 9.726766118529795e-05, "loss": 1.614, "step": 512 }, { "epoch": 0.123883120019319, "grad_norm": 0.2921353876590729, "learning_rate": 9.725502615035393e-05, "loss": 2.0323, "step": 513 }, { "epoch": 0.12412460758270949, "grad_norm": 0.278221994638443, "learning_rate": 9.724236279342764e-05, "loss": 1.6944, "step": 514 }, { "epoch": 0.12436609514609998, "grad_norm": 0.2910671532154083, "learning_rate": 9.722967112210883e-05, "loss": 1.8207, "step": 515 }, { "epoch": 0.12460758270949046, "grad_norm": 0.28481143712997437, "learning_rate": 9.721695114400416e-05, "loss": 1.5918, "step": 516 }, { "epoch": 0.12484907027288095, "grad_norm": 0.2787121832370758, "learning_rate": 9.720420286673722e-05, "loss": 1.6884, "step": 517 }, { "epoch": 0.12509055783627143, "grad_norm": 0.2825336158275604, "learning_rate": 9.719142629794864e-05, "loss": 1.7568, "step": 518 }, { "epoch": 0.12533204539966192, "grad_norm": 0.30035796761512756, "learning_rate": 9.717862144529596e-05, "loss": 1.7803, "step": 519 }, { "epoch": 0.1255735329630524, "grad_norm": 0.2755514085292816, "learning_rate": 9.716578831645368e-05, "loss": 1.8127, "step": 520 }, { "epoch": 0.1258150205264429, "grad_norm": 0.27801671624183655, "learning_rate": 9.715292691911326e-05, "loss": 1.8203, "step": 521 }, { "epoch": 0.12605650808983337, "grad_norm": 0.27820059657096863, "learning_rate": 9.71400372609831e-05, "loss": 1.8601, "step": 522 }, { "epoch": 0.12629799565322386, "grad_norm": 0.3146127760410309, "learning_rate": 9.712711934978851e-05, "loss": 1.896, "step": 523 }, { "epoch": 0.12653948321661435, "grad_norm": 0.2702305018901825, "learning_rate": 9.711417319327174e-05, "loss": 1.7463, "step": 524 }, { "epoch": 0.12678097078000483, "grad_norm": 0.27754876017570496, "learning_rate": 9.710119879919201e-05, "loss": 1.7289, "step": 525 }, { "epoch": 0.12702245834339532, "grad_norm": 0.27290308475494385, "learning_rate": 9.708819617532546e-05, "loss": 1.715, "step": 526 }, { "epoch": 0.1272639459067858, "grad_norm": 0.2878723740577698, "learning_rate": 9.707516532946506e-05, "loss": 1.782, "step": 527 }, { "epoch": 0.1275054334701763, "grad_norm": 0.29204249382019043, "learning_rate": 9.706210626942078e-05, "loss": 1.7625, "step": 528 }, { "epoch": 0.12774692103356677, "grad_norm": 0.28486737608909607, "learning_rate": 9.704901900301952e-05, "loss": 1.8964, "step": 529 }, { "epoch": 0.12798840859695726, "grad_norm": 0.28245386481285095, "learning_rate": 9.703590353810501e-05, "loss": 1.7631, "step": 530 }, { "epoch": 0.12822989616034774, "grad_norm": 0.28806790709495544, "learning_rate": 9.702275988253791e-05, "loss": 1.7705, "step": 531 }, { "epoch": 0.12847138372373823, "grad_norm": 0.2874543070793152, "learning_rate": 9.700958804419581e-05, "loss": 1.6925, "step": 532 }, { "epoch": 0.12871287128712872, "grad_norm": 0.2873690128326416, "learning_rate": 9.699638803097313e-05, "loss": 1.7817, "step": 533 }, { "epoch": 0.1289543588505192, "grad_norm": 0.2811433970928192, "learning_rate": 9.698315985078124e-05, "loss": 1.6764, "step": 534 }, { "epoch": 0.1291958464139097, "grad_norm": 0.2692316174507141, "learning_rate": 9.696990351154836e-05, "loss": 1.654, "step": 535 }, { "epoch": 0.12943733397730017, "grad_norm": 0.27523520588874817, "learning_rate": 9.695661902121955e-05, "loss": 1.7807, "step": 536 }, { "epoch": 0.12967882154069066, "grad_norm": 0.27212703227996826, "learning_rate": 9.694330638775681e-05, "loss": 1.588, "step": 537 }, { "epoch": 0.12992030910408114, "grad_norm": 0.2714000940322876, "learning_rate": 9.692996561913899e-05, "loss": 1.7409, "step": 538 }, { "epoch": 0.13016179666747163, "grad_norm": 0.2825371026992798, "learning_rate": 9.691659672336175e-05, "loss": 1.6836, "step": 539 }, { "epoch": 0.13040328423086212, "grad_norm": 0.27138465642929077, "learning_rate": 9.690319970843768e-05, "loss": 1.6183, "step": 540 }, { "epoch": 0.1306447717942526, "grad_norm": 0.2688143253326416, "learning_rate": 9.688977458239619e-05, "loss": 1.7082, "step": 541 }, { "epoch": 0.1308862593576431, "grad_norm": 0.27580389380455017, "learning_rate": 9.687632135328353e-05, "loss": 1.804, "step": 542 }, { "epoch": 0.13112774692103357, "grad_norm": 0.2860567569732666, "learning_rate": 9.686284002916279e-05, "loss": 1.7159, "step": 543 }, { "epoch": 0.13136923448442406, "grad_norm": 0.28194060921669006, "learning_rate": 9.684933061811393e-05, "loss": 1.8439, "step": 544 }, { "epoch": 0.13161072204781454, "grad_norm": 0.28102830052375793, "learning_rate": 9.683579312823372e-05, "loss": 1.8587, "step": 545 }, { "epoch": 0.13185220961120503, "grad_norm": 0.28536900877952576, "learning_rate": 9.682222756763576e-05, "loss": 1.5734, "step": 546 }, { "epoch": 0.13209369717459551, "grad_norm": 0.28035515546798706, "learning_rate": 9.680863394445048e-05, "loss": 1.6984, "step": 547 }, { "epoch": 0.132335184737986, "grad_norm": 0.2830459773540497, "learning_rate": 9.679501226682513e-05, "loss": 1.7716, "step": 548 }, { "epoch": 0.13257667230137649, "grad_norm": 0.312172532081604, "learning_rate": 9.678136254292376e-05, "loss": 1.9346, "step": 549 }, { "epoch": 0.13281815986476697, "grad_norm": 0.2721990644931793, "learning_rate": 9.676768478092724e-05, "loss": 1.6447, "step": 550 }, { "epoch": 0.13305964742815746, "grad_norm": 0.31954988837242126, "learning_rate": 9.675397898903327e-05, "loss": 2.0035, "step": 551 }, { "epoch": 0.13330113499154794, "grad_norm": 0.27436164021492004, "learning_rate": 9.674024517545628e-05, "loss": 1.7928, "step": 552 }, { "epoch": 0.13354262255493843, "grad_norm": 0.2752690017223358, "learning_rate": 9.67264833484276e-05, "loss": 1.7733, "step": 553 }, { "epoch": 0.1337841101183289, "grad_norm": 0.28749632835388184, "learning_rate": 9.671269351619523e-05, "loss": 1.8311, "step": 554 }, { "epoch": 0.1340255976817194, "grad_norm": 0.28053197264671326, "learning_rate": 9.669887568702405e-05, "loss": 1.778, "step": 555 }, { "epoch": 0.13426708524510989, "grad_norm": 0.2810681462287903, "learning_rate": 9.668502986919566e-05, "loss": 1.7463, "step": 556 }, { "epoch": 0.13450857280850037, "grad_norm": 0.2764918804168701, "learning_rate": 9.667115607100846e-05, "loss": 1.7359, "step": 557 }, { "epoch": 0.13475006037189086, "grad_norm": 0.28768905997276306, "learning_rate": 9.665725430077764e-05, "loss": 1.6103, "step": 558 }, { "epoch": 0.13499154793528134, "grad_norm": 0.29200369119644165, "learning_rate": 9.664332456683511e-05, "loss": 1.6831, "step": 559 }, { "epoch": 0.13523303549867183, "grad_norm": 0.31363463401794434, "learning_rate": 9.662936687752958e-05, "loss": 2.0143, "step": 560 }, { "epoch": 0.1354745230620623, "grad_norm": 0.2834382653236389, "learning_rate": 9.661538124122648e-05, "loss": 1.7506, "step": 561 }, { "epoch": 0.1357160106254528, "grad_norm": 0.29083454608917236, "learning_rate": 9.660136766630801e-05, "loss": 1.8281, "step": 562 }, { "epoch": 0.13595749818884328, "grad_norm": 0.27659663558006287, "learning_rate": 9.658732616117312e-05, "loss": 1.6351, "step": 563 }, { "epoch": 0.13619898575223377, "grad_norm": 0.28885459899902344, "learning_rate": 9.657325673423749e-05, "loss": 1.8258, "step": 564 }, { "epoch": 0.13644047331562426, "grad_norm": 0.2921139895915985, "learning_rate": 9.655915939393354e-05, "loss": 2.0592, "step": 565 }, { "epoch": 0.13668196087901474, "grad_norm": 0.28540095686912537, "learning_rate": 9.654503414871041e-05, "loss": 1.8457, "step": 566 }, { "epoch": 0.13692344844240523, "grad_norm": 0.291238933801651, "learning_rate": 9.653088100703395e-05, "loss": 1.7909, "step": 567 }, { "epoch": 0.1371649360057957, "grad_norm": 0.28494998812675476, "learning_rate": 9.65166999773868e-05, "loss": 1.7552, "step": 568 }, { "epoch": 0.1374064235691862, "grad_norm": 0.2778211832046509, "learning_rate": 9.650249106826823e-05, "loss": 1.7522, "step": 569 }, { "epoch": 0.13764791113257668, "grad_norm": 0.2810254395008087, "learning_rate": 9.648825428819426e-05, "loss": 1.616, "step": 570 }, { "epoch": 0.13788939869596717, "grad_norm": 0.2753067910671234, "learning_rate": 9.647398964569761e-05, "loss": 1.6586, "step": 571 }, { "epoch": 0.13813088625935765, "grad_norm": 0.28924110531806946, "learning_rate": 9.645969714932768e-05, "loss": 1.848, "step": 572 }, { "epoch": 0.13837237382274814, "grad_norm": 0.29547610878944397, "learning_rate": 9.644537680765061e-05, "loss": 1.8143, "step": 573 }, { "epoch": 0.13861386138613863, "grad_norm": 0.2900772988796234, "learning_rate": 9.64310286292492e-05, "loss": 1.9223, "step": 574 }, { "epoch": 0.1388553489495291, "grad_norm": 0.30921927094459534, "learning_rate": 9.64166526227229e-05, "loss": 1.6421, "step": 575 }, { "epoch": 0.1390968365129196, "grad_norm": 0.2801254987716675, "learning_rate": 9.640224879668791e-05, "loss": 1.6986, "step": 576 }, { "epoch": 0.13933832407631008, "grad_norm": 0.3085472583770752, "learning_rate": 9.638781715977704e-05, "loss": 1.961, "step": 577 }, { "epoch": 0.13957981163970057, "grad_norm": 0.29188886284828186, "learning_rate": 9.637335772063983e-05, "loss": 1.6771, "step": 578 }, { "epoch": 0.13982129920309105, "grad_norm": 0.30945637822151184, "learning_rate": 9.635887048794242e-05, "loss": 2.1181, "step": 579 }, { "epoch": 0.14006278676648154, "grad_norm": 0.2721393406391144, "learning_rate": 9.634435547036763e-05, "loss": 1.7944, "step": 580 }, { "epoch": 0.14030427432987203, "grad_norm": 0.2815290093421936, "learning_rate": 9.632981267661496e-05, "loss": 1.5639, "step": 581 }, { "epoch": 0.1405457618932625, "grad_norm": 0.2875358760356903, "learning_rate": 9.631524211540051e-05, "loss": 1.8578, "step": 582 }, { "epoch": 0.14078724945665297, "grad_norm": 0.3072429597377777, "learning_rate": 9.630064379545707e-05, "loss": 1.6873, "step": 583 }, { "epoch": 0.14102873702004345, "grad_norm": 0.3023555278778076, "learning_rate": 9.628601772553403e-05, "loss": 1.7457, "step": 584 }, { "epoch": 0.14127022458343394, "grad_norm": 0.3081570267677307, "learning_rate": 9.627136391439744e-05, "loss": 1.7089, "step": 585 }, { "epoch": 0.14151171214682443, "grad_norm": 0.2926265299320221, "learning_rate": 9.625668237082995e-05, "loss": 1.8465, "step": 586 }, { "epoch": 0.1417531997102149, "grad_norm": 0.30102863907814026, "learning_rate": 9.624197310363085e-05, "loss": 1.779, "step": 587 }, { "epoch": 0.1419946872736054, "grad_norm": 0.3015720844268799, "learning_rate": 9.622723612161606e-05, "loss": 1.8684, "step": 588 }, { "epoch": 0.14223617483699588, "grad_norm": 0.2746000587940216, "learning_rate": 9.621247143361806e-05, "loss": 1.734, "step": 589 }, { "epoch": 0.14247766240038637, "grad_norm": 0.2910473048686981, "learning_rate": 9.619767904848598e-05, "loss": 1.9269, "step": 590 }, { "epoch": 0.14271914996377685, "grad_norm": 0.27721109986305237, "learning_rate": 9.618285897508553e-05, "loss": 1.6957, "step": 591 }, { "epoch": 0.14296063752716734, "grad_norm": 0.2783820629119873, "learning_rate": 9.616801122229904e-05, "loss": 1.7603, "step": 592 }, { "epoch": 0.14320212509055782, "grad_norm": 0.2656041979789734, "learning_rate": 9.615313579902538e-05, "loss": 1.6831, "step": 593 }, { "epoch": 0.1434436126539483, "grad_norm": 0.2902359068393707, "learning_rate": 9.613823271418006e-05, "loss": 1.7025, "step": 594 }, { "epoch": 0.1436851002173388, "grad_norm": 0.2851390838623047, "learning_rate": 9.612330197669514e-05, "loss": 1.6937, "step": 595 }, { "epoch": 0.14392658778072928, "grad_norm": 0.2772185206413269, "learning_rate": 9.610834359551926e-05, "loss": 1.6978, "step": 596 }, { "epoch": 0.14416807534411977, "grad_norm": 0.2872218191623688, "learning_rate": 9.609335757961761e-05, "loss": 1.8004, "step": 597 }, { "epoch": 0.14440956290751025, "grad_norm": 0.27167755365371704, "learning_rate": 9.607834393797197e-05, "loss": 1.6695, "step": 598 }, { "epoch": 0.14465105047090074, "grad_norm": 0.275107741355896, "learning_rate": 9.606330267958066e-05, "loss": 1.8051, "step": 599 }, { "epoch": 0.14489253803429122, "grad_norm": 0.2851438522338867, "learning_rate": 9.604823381345856e-05, "loss": 1.863, "step": 600 }, { "epoch": 0.1451340255976817, "grad_norm": 0.2878473103046417, "learning_rate": 9.603313734863708e-05, "loss": 1.8116, "step": 601 }, { "epoch": 0.1453755131610722, "grad_norm": 0.2840920388698578, "learning_rate": 9.601801329416422e-05, "loss": 1.6956, "step": 602 }, { "epoch": 0.14561700072446268, "grad_norm": 0.27009525895118713, "learning_rate": 9.600286165910445e-05, "loss": 1.6875, "step": 603 }, { "epoch": 0.14585848828785317, "grad_norm": 0.31799811124801636, "learning_rate": 9.598768245253879e-05, "loss": 1.8778, "step": 604 }, { "epoch": 0.14609997585124365, "grad_norm": 0.2943333685398102, "learning_rate": 9.597247568356484e-05, "loss": 1.6921, "step": 605 }, { "epoch": 0.14634146341463414, "grad_norm": 0.2869589030742645, "learning_rate": 9.595724136129665e-05, "loss": 1.8085, "step": 606 }, { "epoch": 0.14658295097802462, "grad_norm": 0.2950487434864044, "learning_rate": 9.594197949486478e-05, "loss": 1.8381, "step": 607 }, { "epoch": 0.1468244385414151, "grad_norm": 0.2995087206363678, "learning_rate": 9.592669009341637e-05, "loss": 1.7566, "step": 608 }, { "epoch": 0.1470659261048056, "grad_norm": 0.302502304315567, "learning_rate": 9.5911373166115e-05, "loss": 1.8809, "step": 609 }, { "epoch": 0.14730741366819608, "grad_norm": 0.30947795510292053, "learning_rate": 9.589602872214079e-05, "loss": 1.9043, "step": 610 }, { "epoch": 0.14754890123158657, "grad_norm": 0.29188233613967896, "learning_rate": 9.588065677069029e-05, "loss": 1.7712, "step": 611 }, { "epoch": 0.14779038879497705, "grad_norm": 0.29185009002685547, "learning_rate": 9.58652573209766e-05, "loss": 1.907, "step": 612 }, { "epoch": 0.14803187635836754, "grad_norm": 0.2943671941757202, "learning_rate": 9.584983038222928e-05, "loss": 2.0036, "step": 613 }, { "epoch": 0.14827336392175802, "grad_norm": 0.2962663471698761, "learning_rate": 9.583437596369434e-05, "loss": 1.8042, "step": 614 }, { "epoch": 0.1485148514851485, "grad_norm": 0.28282076120376587, "learning_rate": 9.581889407463431e-05, "loss": 1.8427, "step": 615 }, { "epoch": 0.148756339048539, "grad_norm": 0.27998435497283936, "learning_rate": 9.580338472432815e-05, "loss": 1.8312, "step": 616 }, { "epoch": 0.14899782661192948, "grad_norm": 0.285490483045578, "learning_rate": 9.578784792207127e-05, "loss": 1.745, "step": 617 }, { "epoch": 0.14923931417531996, "grad_norm": 0.2885023057460785, "learning_rate": 9.577228367717554e-05, "loss": 1.6725, "step": 618 }, { "epoch": 0.14948080173871045, "grad_norm": 0.3022843897342682, "learning_rate": 9.575669199896933e-05, "loss": 1.7719, "step": 619 }, { "epoch": 0.14972228930210094, "grad_norm": 0.27724575996398926, "learning_rate": 9.574107289679737e-05, "loss": 1.6429, "step": 620 }, { "epoch": 0.14996377686549142, "grad_norm": 0.3123320937156677, "learning_rate": 9.572542638002087e-05, "loss": 1.9326, "step": 621 }, { "epoch": 0.1502052644288819, "grad_norm": 0.33049526810646057, "learning_rate": 9.570975245801746e-05, "loss": 2.183, "step": 622 }, { "epoch": 0.1504467519922724, "grad_norm": 0.2905597686767578, "learning_rate": 9.569405114018122e-05, "loss": 1.838, "step": 623 }, { "epoch": 0.15068823955566288, "grad_norm": 0.2684404253959656, "learning_rate": 9.567832243592261e-05, "loss": 1.6479, "step": 624 }, { "epoch": 0.15092972711905336, "grad_norm": 0.2934153378009796, "learning_rate": 9.566256635466852e-05, "loss": 1.8059, "step": 625 }, { "epoch": 0.15117121468244385, "grad_norm": 0.28340545296669006, "learning_rate": 9.564678290586229e-05, "loss": 1.9517, "step": 626 }, { "epoch": 0.15141270224583434, "grad_norm": 0.3298986852169037, "learning_rate": 9.563097209896357e-05, "loss": 1.6387, "step": 627 }, { "epoch": 0.15165418980922482, "grad_norm": 0.2907513380050659, "learning_rate": 9.56151339434485e-05, "loss": 1.8626, "step": 628 }, { "epoch": 0.1518956773726153, "grad_norm": 0.29203519225120544, "learning_rate": 9.559926844880956e-05, "loss": 1.8376, "step": 629 }, { "epoch": 0.1521371649360058, "grad_norm": 0.29703041911125183, "learning_rate": 9.558337562455561e-05, "loss": 1.7033, "step": 630 }, { "epoch": 0.15237865249939628, "grad_norm": 0.28163012862205505, "learning_rate": 9.556745548021193e-05, "loss": 1.7325, "step": 631 }, { "epoch": 0.15262014006278676, "grad_norm": 0.3157091736793518, "learning_rate": 9.555150802532016e-05, "loss": 1.77, "step": 632 }, { "epoch": 0.15286162762617725, "grad_norm": 0.295043021440506, "learning_rate": 9.553553326943826e-05, "loss": 1.8613, "step": 633 }, { "epoch": 0.15310311518956773, "grad_norm": 0.29055267572402954, "learning_rate": 9.551953122214064e-05, "loss": 1.8194, "step": 634 }, { "epoch": 0.15334460275295822, "grad_norm": 0.31686437129974365, "learning_rate": 9.550350189301797e-05, "loss": 2.034, "step": 635 }, { "epoch": 0.1535860903163487, "grad_norm": 0.30802491307258606, "learning_rate": 9.548744529167737e-05, "loss": 1.8503, "step": 636 }, { "epoch": 0.1538275778797392, "grad_norm": 0.2771475613117218, "learning_rate": 9.547136142774222e-05, "loss": 1.6846, "step": 637 }, { "epoch": 0.15406906544312968, "grad_norm": 0.3178752362728119, "learning_rate": 9.54552503108523e-05, "loss": 1.8496, "step": 638 }, { "epoch": 0.15431055300652016, "grad_norm": 0.29422762989997864, "learning_rate": 9.543911195066367e-05, "loss": 1.7731, "step": 639 }, { "epoch": 0.15455204056991065, "grad_norm": 0.27632033824920654, "learning_rate": 9.542294635684878e-05, "loss": 1.6732, "step": 640 }, { "epoch": 0.15479352813330113, "grad_norm": 0.2904300093650818, "learning_rate": 9.540675353909636e-05, "loss": 1.6868, "step": 641 }, { "epoch": 0.15503501569669162, "grad_norm": 0.30907413363456726, "learning_rate": 9.539053350711145e-05, "loss": 2.053, "step": 642 }, { "epoch": 0.1552765032600821, "grad_norm": 0.3101508319377899, "learning_rate": 9.537428627061545e-05, "loss": 1.9812, "step": 643 }, { "epoch": 0.1555179908234726, "grad_norm": 0.3084372878074646, "learning_rate": 9.535801183934598e-05, "loss": 1.9835, "step": 644 }, { "epoch": 0.15575947838686308, "grad_norm": 0.2710745632648468, "learning_rate": 9.534171022305706e-05, "loss": 1.6607, "step": 645 }, { "epoch": 0.15600096595025356, "grad_norm": 0.2810233235359192, "learning_rate": 9.532538143151895e-05, "loss": 1.7959, "step": 646 }, { "epoch": 0.15624245351364405, "grad_norm": 0.2879350483417511, "learning_rate": 9.530902547451817e-05, "loss": 1.7773, "step": 647 }, { "epoch": 0.15648394107703453, "grad_norm": 0.2893047332763672, "learning_rate": 9.529264236185758e-05, "loss": 1.6512, "step": 648 }, { "epoch": 0.15672542864042502, "grad_norm": 0.29465562105178833, "learning_rate": 9.527623210335628e-05, "loss": 1.9087, "step": 649 }, { "epoch": 0.1569669162038155, "grad_norm": 0.2800419330596924, "learning_rate": 9.525979470884963e-05, "loss": 1.8204, "step": 650 }, { "epoch": 0.157208403767206, "grad_norm": 0.27463045716285706, "learning_rate": 9.524333018818931e-05, "loss": 1.6238, "step": 651 }, { "epoch": 0.15744989133059648, "grad_norm": 0.2779013216495514, "learning_rate": 9.522683855124319e-05, "loss": 1.7749, "step": 652 }, { "epoch": 0.15769137889398696, "grad_norm": 0.28004294633865356, "learning_rate": 9.521031980789543e-05, "loss": 1.7464, "step": 653 }, { "epoch": 0.15793286645737745, "grad_norm": 0.2746169865131378, "learning_rate": 9.519377396804644e-05, "loss": 1.7046, "step": 654 }, { "epoch": 0.15817435402076793, "grad_norm": 0.2713209092617035, "learning_rate": 9.517720104161284e-05, "loss": 1.6901, "step": 655 }, { "epoch": 0.15841584158415842, "grad_norm": 0.2889479696750641, "learning_rate": 9.51606010385275e-05, "loss": 1.8167, "step": 656 }, { "epoch": 0.1586573291475489, "grad_norm": 0.4551263451576233, "learning_rate": 9.514397396873956e-05, "loss": 2.2537, "step": 657 }, { "epoch": 0.1588988167109394, "grad_norm": 0.2781124413013458, "learning_rate": 9.512731984221428e-05, "loss": 1.6174, "step": 658 }, { "epoch": 0.15914030427432987, "grad_norm": 0.30449262261390686, "learning_rate": 9.511063866893324e-05, "loss": 1.8033, "step": 659 }, { "epoch": 0.15938179183772036, "grad_norm": 0.2778739035129547, "learning_rate": 9.509393045889417e-05, "loss": 1.7579, "step": 660 }, { "epoch": 0.15962327940111085, "grad_norm": 0.31164950132369995, "learning_rate": 9.507719522211104e-05, "loss": 1.8609, "step": 661 }, { "epoch": 0.15986476696450133, "grad_norm": 0.28629255294799805, "learning_rate": 9.506043296861399e-05, "loss": 1.8606, "step": 662 }, { "epoch": 0.16010625452789182, "grad_norm": 0.29869675636291504, "learning_rate": 9.504364370844938e-05, "loss": 1.9861, "step": 663 }, { "epoch": 0.1603477420912823, "grad_norm": 0.2828952968120575, "learning_rate": 9.502682745167972e-05, "loss": 1.6205, "step": 664 }, { "epoch": 0.1605892296546728, "grad_norm": 0.30284184217453003, "learning_rate": 9.500998420838374e-05, "loss": 1.8012, "step": 665 }, { "epoch": 0.16083071721806327, "grad_norm": 0.2756812274456024, "learning_rate": 9.49931139886563e-05, "loss": 1.7166, "step": 666 }, { "epoch": 0.16107220478145376, "grad_norm": 0.2983384132385254, "learning_rate": 9.497621680260848e-05, "loss": 1.7543, "step": 667 }, { "epoch": 0.16131369234484425, "grad_norm": 0.32105186581611633, "learning_rate": 9.495929266036747e-05, "loss": 1.6919, "step": 668 }, { "epoch": 0.16155517990823473, "grad_norm": 0.2885781526565552, "learning_rate": 9.494234157207665e-05, "loss": 1.723, "step": 669 }, { "epoch": 0.16179666747162522, "grad_norm": 0.28944310545921326, "learning_rate": 9.492536354789554e-05, "loss": 1.7258, "step": 670 }, { "epoch": 0.1620381550350157, "grad_norm": 0.31029340624809265, "learning_rate": 9.490835859799983e-05, "loss": 1.7345, "step": 671 }, { "epoch": 0.1622796425984062, "grad_norm": 0.2986032962799072, "learning_rate": 9.489132673258128e-05, "loss": 1.6178, "step": 672 }, { "epoch": 0.16252113016179667, "grad_norm": 0.2751445770263672, "learning_rate": 9.487426796184785e-05, "loss": 1.7689, "step": 673 }, { "epoch": 0.16276261772518716, "grad_norm": 0.34798896312713623, "learning_rate": 9.485718229602359e-05, "loss": 2.016, "step": 674 }, { "epoch": 0.16300410528857764, "grad_norm": 0.31324881315231323, "learning_rate": 9.484006974534869e-05, "loss": 1.831, "step": 675 }, { "epoch": 0.16324559285196813, "grad_norm": 0.28822532296180725, "learning_rate": 9.482293032007943e-05, "loss": 1.8226, "step": 676 }, { "epoch": 0.16348708041535862, "grad_norm": 0.2884620726108551, "learning_rate": 9.480576403048822e-05, "loss": 1.6349, "step": 677 }, { "epoch": 0.1637285679787491, "grad_norm": 0.29445016384124756, "learning_rate": 9.478857088686355e-05, "loss": 1.6545, "step": 678 }, { "epoch": 0.1639700555421396, "grad_norm": 0.28695762157440186, "learning_rate": 9.477135089951001e-05, "loss": 1.6071, "step": 679 }, { "epoch": 0.16421154310553007, "grad_norm": 0.2973484694957733, "learning_rate": 9.47541040787483e-05, "loss": 1.9545, "step": 680 }, { "epoch": 0.16445303066892056, "grad_norm": 0.3042725920677185, "learning_rate": 9.473683043491518e-05, "loss": 1.8942, "step": 681 }, { "epoch": 0.16469451823231104, "grad_norm": 0.2901724874973297, "learning_rate": 9.471952997836348e-05, "loss": 1.6541, "step": 682 }, { "epoch": 0.16493600579570153, "grad_norm": 0.27054041624069214, "learning_rate": 9.470220271946215e-05, "loss": 1.6211, "step": 683 }, { "epoch": 0.16517749335909201, "grad_norm": 0.28445035219192505, "learning_rate": 9.468484866859612e-05, "loss": 1.7979, "step": 684 }, { "epoch": 0.1654189809224825, "grad_norm": 0.29308000206947327, "learning_rate": 9.466746783616645e-05, "loss": 1.7483, "step": 685 }, { "epoch": 0.16566046848587299, "grad_norm": 0.2914656698703766, "learning_rate": 9.465006023259022e-05, "loss": 1.7282, "step": 686 }, { "epoch": 0.16590195604926347, "grad_norm": 0.2674042284488678, "learning_rate": 9.463262586830056e-05, "loss": 1.7118, "step": 687 }, { "epoch": 0.16614344361265396, "grad_norm": 0.2843051552772522, "learning_rate": 9.461516475374663e-05, "loss": 1.8195, "step": 688 }, { "epoch": 0.16638493117604444, "grad_norm": 0.29515916109085083, "learning_rate": 9.459767689939365e-05, "loss": 1.7215, "step": 689 }, { "epoch": 0.16662641873943493, "grad_norm": 0.2934631109237671, "learning_rate": 9.458016231572283e-05, "loss": 1.7785, "step": 690 }, { "epoch": 0.16686790630282541, "grad_norm": 0.28389841318130493, "learning_rate": 9.45626210132314e-05, "loss": 1.6819, "step": 691 }, { "epoch": 0.1671093938662159, "grad_norm": 0.30072394013404846, "learning_rate": 9.45450530024327e-05, "loss": 1.7743, "step": 692 }, { "epoch": 0.16735088142960639, "grad_norm": 0.28151774406433105, "learning_rate": 9.452745829385592e-05, "loss": 1.7631, "step": 693 }, { "epoch": 0.16759236899299687, "grad_norm": 0.2804087698459625, "learning_rate": 9.450983689804636e-05, "loss": 1.82, "step": 694 }, { "epoch": 0.16783385655638736, "grad_norm": 0.2899933457374573, "learning_rate": 9.449218882556528e-05, "loss": 1.7561, "step": 695 }, { "epoch": 0.16807534411977784, "grad_norm": 0.3071437180042267, "learning_rate": 9.447451408698993e-05, "loss": 1.8007, "step": 696 }, { "epoch": 0.16831683168316833, "grad_norm": 0.2796712815761566, "learning_rate": 9.445681269291354e-05, "loss": 1.6631, "step": 697 }, { "epoch": 0.1685583192465588, "grad_norm": 0.3034926652908325, "learning_rate": 9.443908465394535e-05, "loss": 1.879, "step": 698 }, { "epoch": 0.1687998068099493, "grad_norm": 0.2936783730983734, "learning_rate": 9.442132998071052e-05, "loss": 1.7325, "step": 699 }, { "epoch": 0.16904129437333978, "grad_norm": 0.2850625813007355, "learning_rate": 9.44035486838502e-05, "loss": 1.7287, "step": 700 }, { "epoch": 0.16928278193673027, "grad_norm": 0.28181543946266174, "learning_rate": 9.43857407740215e-05, "loss": 1.7932, "step": 701 }, { "epoch": 0.16952426950012076, "grad_norm": 0.2871532142162323, "learning_rate": 9.436790626189745e-05, "loss": 1.7514, "step": 702 }, { "epoch": 0.16976575706351124, "grad_norm": 0.28033384680747986, "learning_rate": 9.435004515816703e-05, "loss": 1.7618, "step": 703 }, { "epoch": 0.17000724462690173, "grad_norm": 0.28677472472190857, "learning_rate": 9.433215747353522e-05, "loss": 1.7955, "step": 704 }, { "epoch": 0.1702487321902922, "grad_norm": 0.28708505630493164, "learning_rate": 9.431424321872286e-05, "loss": 1.7104, "step": 705 }, { "epoch": 0.1704902197536827, "grad_norm": 0.2901146113872528, "learning_rate": 9.429630240446674e-05, "loss": 1.7043, "step": 706 }, { "epoch": 0.17073170731707318, "grad_norm": 0.3041292130947113, "learning_rate": 9.427833504151955e-05, "loss": 1.645, "step": 707 }, { "epoch": 0.17097319488046367, "grad_norm": 0.2819676995277405, "learning_rate": 9.426034114064994e-05, "loss": 1.7287, "step": 708 }, { "epoch": 0.17121468244385415, "grad_norm": 0.2727168798446655, "learning_rate": 9.424232071264241e-05, "loss": 1.6778, "step": 709 }, { "epoch": 0.17145617000724464, "grad_norm": 0.28442737460136414, "learning_rate": 9.422427376829739e-05, "loss": 1.7041, "step": 710 }, { "epoch": 0.17169765757063513, "grad_norm": 0.49621260166168213, "learning_rate": 9.42062003184312e-05, "loss": 1.558, "step": 711 }, { "epoch": 0.17193914513402558, "grad_norm": 0.2837514877319336, "learning_rate": 9.418810037387603e-05, "loss": 1.7101, "step": 712 }, { "epoch": 0.17218063269741607, "grad_norm": 0.27921706438064575, "learning_rate": 9.416997394547996e-05, "loss": 1.8319, "step": 713 }, { "epoch": 0.17242212026080656, "grad_norm": 0.2778397798538208, "learning_rate": 9.415182104410693e-05, "loss": 1.724, "step": 714 }, { "epoch": 0.17266360782419704, "grad_norm": 0.28968319296836853, "learning_rate": 9.413364168063681e-05, "loss": 1.6833, "step": 715 }, { "epoch": 0.17290509538758753, "grad_norm": 0.2843336760997772, "learning_rate": 9.411543586596523e-05, "loss": 1.7913, "step": 716 }, { "epoch": 0.173146582950978, "grad_norm": 0.2921014428138733, "learning_rate": 9.409720361100374e-05, "loss": 1.7845, "step": 717 }, { "epoch": 0.1733880705143685, "grad_norm": 0.3008785545825958, "learning_rate": 9.407894492667972e-05, "loss": 1.7576, "step": 718 }, { "epoch": 0.17362955807775898, "grad_norm": 0.2838538587093353, "learning_rate": 9.406065982393639e-05, "loss": 1.7193, "step": 719 }, { "epoch": 0.17387104564114947, "grad_norm": 0.3217141628265381, "learning_rate": 9.40423483137328e-05, "loss": 1.87, "step": 720 }, { "epoch": 0.17411253320453995, "grad_norm": 0.29201528429985046, "learning_rate": 9.402401040704383e-05, "loss": 1.7704, "step": 721 }, { "epoch": 0.17435402076793044, "grad_norm": 0.2874560058116913, "learning_rate": 9.400564611486017e-05, "loss": 1.7274, "step": 722 }, { "epoch": 0.17459550833132093, "grad_norm": 0.27963292598724365, "learning_rate": 9.398725544818838e-05, "loss": 1.6752, "step": 723 }, { "epoch": 0.1748369958947114, "grad_norm": 0.2946629524230957, "learning_rate": 9.396883841805073e-05, "loss": 1.9184, "step": 724 }, { "epoch": 0.1750784834581019, "grad_norm": 0.2995729446411133, "learning_rate": 9.395039503548537e-05, "loss": 1.7893, "step": 725 }, { "epoch": 0.17531997102149238, "grad_norm": 0.292082816362381, "learning_rate": 9.393192531154621e-05, "loss": 1.8141, "step": 726 }, { "epoch": 0.17556145858488287, "grad_norm": 0.27504876255989075, "learning_rate": 9.391342925730295e-05, "loss": 1.6477, "step": 727 }, { "epoch": 0.17580294614827335, "grad_norm": 0.2799321413040161, "learning_rate": 9.389490688384109e-05, "loss": 1.6797, "step": 728 }, { "epoch": 0.17604443371166384, "grad_norm": 0.2817777991294861, "learning_rate": 9.387635820226189e-05, "loss": 1.7789, "step": 729 }, { "epoch": 0.17628592127505432, "grad_norm": 0.266647607088089, "learning_rate": 9.385778322368235e-05, "loss": 1.6869, "step": 730 }, { "epoch": 0.1765274088384448, "grad_norm": 0.2966921925544739, "learning_rate": 9.383918195923527e-05, "loss": 1.8977, "step": 731 }, { "epoch": 0.1767688964018353, "grad_norm": 0.2740517258644104, "learning_rate": 9.382055442006923e-05, "loss": 1.7156, "step": 732 }, { "epoch": 0.17701038396522578, "grad_norm": 0.27741682529449463, "learning_rate": 9.380190061734847e-05, "loss": 1.6714, "step": 733 }, { "epoch": 0.17725187152861627, "grad_norm": 0.2760551869869232, "learning_rate": 9.378322056225305e-05, "loss": 1.7512, "step": 734 }, { "epoch": 0.17749335909200675, "grad_norm": 0.2780584394931793, "learning_rate": 9.376451426597872e-05, "loss": 1.7036, "step": 735 }, { "epoch": 0.17773484665539724, "grad_norm": 0.2876451909542084, "learning_rate": 9.374578173973698e-05, "loss": 1.6687, "step": 736 }, { "epoch": 0.17797633421878772, "grad_norm": 0.29878392815589905, "learning_rate": 9.372702299475506e-05, "loss": 1.772, "step": 737 }, { "epoch": 0.1782178217821782, "grad_norm": 0.28005820512771606, "learning_rate": 9.370823804227586e-05, "loss": 1.7543, "step": 738 }, { "epoch": 0.1784593093455687, "grad_norm": 0.26662278175354004, "learning_rate": 9.368942689355804e-05, "loss": 1.6537, "step": 739 }, { "epoch": 0.17870079690895918, "grad_norm": 0.2788659632205963, "learning_rate": 9.367058955987591e-05, "loss": 1.7021, "step": 740 }, { "epoch": 0.17894228447234967, "grad_norm": 0.2913663685321808, "learning_rate": 9.36517260525195e-05, "loss": 1.8001, "step": 741 }, { "epoch": 0.17918377203574015, "grad_norm": 0.2656884789466858, "learning_rate": 9.363283638279457e-05, "loss": 1.6377, "step": 742 }, { "epoch": 0.17942525959913064, "grad_norm": 0.3025050759315491, "learning_rate": 9.361392056202246e-05, "loss": 1.8939, "step": 743 }, { "epoch": 0.17966674716252112, "grad_norm": 0.30079180002212524, "learning_rate": 9.359497860154027e-05, "loss": 1.8976, "step": 744 }, { "epoch": 0.1799082347259116, "grad_norm": 0.28714820742607117, "learning_rate": 9.357601051270074e-05, "loss": 1.7784, "step": 745 }, { "epoch": 0.1801497222893021, "grad_norm": 0.29758569598197937, "learning_rate": 9.355701630687225e-05, "loss": 1.7182, "step": 746 }, { "epoch": 0.18039120985269258, "grad_norm": 0.26696479320526123, "learning_rate": 9.353799599543887e-05, "loss": 1.6108, "step": 747 }, { "epoch": 0.18063269741608307, "grad_norm": 0.2863609790802002, "learning_rate": 9.351894958980026e-05, "loss": 1.8426, "step": 748 }, { "epoch": 0.18087418497947355, "grad_norm": 0.2715589702129364, "learning_rate": 9.349987710137179e-05, "loss": 1.7442, "step": 749 }, { "epoch": 0.18111567254286404, "grad_norm": 0.2864156663417816, "learning_rate": 9.348077854158441e-05, "loss": 1.7882, "step": 750 }, { "epoch": 0.18135716010625452, "grad_norm": 0.30364513397216797, "learning_rate": 9.346165392188473e-05, "loss": 1.64, "step": 751 }, { "epoch": 0.181598647669645, "grad_norm": 0.3006398379802704, "learning_rate": 9.344250325373495e-05, "loss": 1.901, "step": 752 }, { "epoch": 0.1818401352330355, "grad_norm": 0.2953626811504364, "learning_rate": 9.34233265486129e-05, "loss": 1.797, "step": 753 }, { "epoch": 0.18208162279642598, "grad_norm": 0.2777319550514221, "learning_rate": 9.3404123818012e-05, "loss": 1.7733, "step": 754 }, { "epoch": 0.18232311035981646, "grad_norm": 0.2769542932510376, "learning_rate": 9.338489507344128e-05, "loss": 1.7931, "step": 755 }, { "epoch": 0.18256459792320695, "grad_norm": 0.28098264336586, "learning_rate": 9.336564032642536e-05, "loss": 1.7173, "step": 756 }, { "epoch": 0.18280608548659744, "grad_norm": 0.278130441904068, "learning_rate": 9.334635958850446e-05, "loss": 1.7329, "step": 757 }, { "epoch": 0.18304757304998792, "grad_norm": 0.28257855772972107, "learning_rate": 9.332705287123434e-05, "loss": 1.7744, "step": 758 }, { "epoch": 0.1832890606133784, "grad_norm": 0.28167739510536194, "learning_rate": 9.330772018618634e-05, "loss": 1.821, "step": 759 }, { "epoch": 0.1835305481767689, "grad_norm": 0.29732733964920044, "learning_rate": 9.32883615449474e-05, "loss": 1.7651, "step": 760 }, { "epoch": 0.18377203574015938, "grad_norm": 0.27904653549194336, "learning_rate": 9.326897695912e-05, "loss": 1.884, "step": 761 }, { "epoch": 0.18401352330354986, "grad_norm": 0.29639023542404175, "learning_rate": 9.324956644032213e-05, "loss": 1.9033, "step": 762 }, { "epoch": 0.18425501086694035, "grad_norm": 0.28502506017684937, "learning_rate": 9.323013000018736e-05, "loss": 1.8383, "step": 763 }, { "epoch": 0.18449649843033084, "grad_norm": 0.2888893187046051, "learning_rate": 9.32106676503648e-05, "loss": 1.7931, "step": 764 }, { "epoch": 0.18473798599372132, "grad_norm": 0.29819416999816895, "learning_rate": 9.31911794025191e-05, "loss": 1.8012, "step": 765 }, { "epoch": 0.1849794735571118, "grad_norm": 0.2912543714046478, "learning_rate": 9.317166526833035e-05, "loss": 1.8254, "step": 766 }, { "epoch": 0.1852209611205023, "grad_norm": 0.2873326539993286, "learning_rate": 9.315212525949428e-05, "loss": 1.8009, "step": 767 }, { "epoch": 0.18546244868389278, "grad_norm": 0.27630600333213806, "learning_rate": 9.313255938772202e-05, "loss": 1.7497, "step": 768 }, { "epoch": 0.18570393624728326, "grad_norm": 0.28622114658355713, "learning_rate": 9.311296766474027e-05, "loss": 1.6059, "step": 769 }, { "epoch": 0.18594542381067375, "grad_norm": 0.2943037748336792, "learning_rate": 9.309335010229118e-05, "loss": 1.728, "step": 770 }, { "epoch": 0.18618691137406423, "grad_norm": 0.29202359914779663, "learning_rate": 9.30737067121324e-05, "loss": 1.7766, "step": 771 }, { "epoch": 0.18642839893745472, "grad_norm": 0.2959134578704834, "learning_rate": 9.30540375060371e-05, "loss": 1.8195, "step": 772 }, { "epoch": 0.1866698865008452, "grad_norm": 0.28923317790031433, "learning_rate": 9.303434249579384e-05, "loss": 1.6646, "step": 773 }, { "epoch": 0.1869113740642357, "grad_norm": 0.2839503288269043, "learning_rate": 9.301462169320672e-05, "loss": 1.7667, "step": 774 }, { "epoch": 0.18715286162762618, "grad_norm": 0.28493398427963257, "learning_rate": 9.299487511009525e-05, "loss": 1.8927, "step": 775 }, { "epoch": 0.18739434919101666, "grad_norm": 0.3119465112686157, "learning_rate": 9.297510275829444e-05, "loss": 1.8846, "step": 776 }, { "epoch": 0.18763583675440715, "grad_norm": 0.29498162865638733, "learning_rate": 9.29553046496547e-05, "loss": 1.7002, "step": 777 }, { "epoch": 0.18787732431779763, "grad_norm": 0.28077542781829834, "learning_rate": 9.293548079604189e-05, "loss": 1.7079, "step": 778 }, { "epoch": 0.18811881188118812, "grad_norm": 0.2838478982448578, "learning_rate": 9.291563120933731e-05, "loss": 1.6531, "step": 779 }, { "epoch": 0.1883602994445786, "grad_norm": 0.29093652963638306, "learning_rate": 9.289575590143766e-05, "loss": 1.6198, "step": 780 }, { "epoch": 0.1886017870079691, "grad_norm": 0.28098905086517334, "learning_rate": 9.28758548842551e-05, "loss": 1.7282, "step": 781 }, { "epoch": 0.18884327457135958, "grad_norm": 0.31896719336509705, "learning_rate": 9.285592816971716e-05, "loss": 2.0032, "step": 782 }, { "epoch": 0.18908476213475006, "grad_norm": 0.2947934865951538, "learning_rate": 9.283597576976677e-05, "loss": 1.8173, "step": 783 }, { "epoch": 0.18932624969814055, "grad_norm": 0.2865394949913025, "learning_rate": 9.281599769636226e-05, "loss": 1.7014, "step": 784 }, { "epoch": 0.18956773726153103, "grad_norm": 0.2883725166320801, "learning_rate": 9.27959939614774e-05, "loss": 1.9053, "step": 785 }, { "epoch": 0.18980922482492152, "grad_norm": 0.3068401515483856, "learning_rate": 9.277596457710125e-05, "loss": 1.7265, "step": 786 }, { "epoch": 0.190050712388312, "grad_norm": 0.29906126856803894, "learning_rate": 9.275590955523831e-05, "loss": 1.7202, "step": 787 }, { "epoch": 0.1902921999517025, "grad_norm": 0.2799694240093231, "learning_rate": 9.273582890790842e-05, "loss": 1.7297, "step": 788 }, { "epoch": 0.19053368751509298, "grad_norm": 0.29376327991485596, "learning_rate": 9.271572264714675e-05, "loss": 1.7631, "step": 789 }, { "epoch": 0.19077517507848346, "grad_norm": 0.29661813378334045, "learning_rate": 9.269559078500389e-05, "loss": 1.7929, "step": 790 }, { "epoch": 0.19101666264187395, "grad_norm": 0.28860270977020264, "learning_rate": 9.267543333354571e-05, "loss": 1.8858, "step": 791 }, { "epoch": 0.19125815020526443, "grad_norm": 0.29066187143325806, "learning_rate": 9.265525030485348e-05, "loss": 1.7584, "step": 792 }, { "epoch": 0.19149963776865492, "grad_norm": 0.29545891284942627, "learning_rate": 9.263504171102372e-05, "loss": 1.6921, "step": 793 }, { "epoch": 0.1917411253320454, "grad_norm": 0.3049885034561157, "learning_rate": 9.261480756416835e-05, "loss": 1.778, "step": 794 }, { "epoch": 0.1919826128954359, "grad_norm": 0.31124168634414673, "learning_rate": 9.259454787641454e-05, "loss": 1.9121, "step": 795 }, { "epoch": 0.19222410045882637, "grad_norm": 0.3002880811691284, "learning_rate": 9.257426265990481e-05, "loss": 1.8388, "step": 796 }, { "epoch": 0.19246558802221686, "grad_norm": 0.32877475023269653, "learning_rate": 9.255395192679697e-05, "loss": 2.0557, "step": 797 }, { "epoch": 0.19270707558560735, "grad_norm": 0.28469905257225037, "learning_rate": 9.253361568926413e-05, "loss": 1.8209, "step": 798 }, { "epoch": 0.19294856314899783, "grad_norm": 0.2947460114955902, "learning_rate": 9.251325395949466e-05, "loss": 1.7037, "step": 799 }, { "epoch": 0.19319005071238832, "grad_norm": 0.2917262017726898, "learning_rate": 9.249286674969224e-05, "loss": 1.7924, "step": 800 }, { "epoch": 0.1934315382757788, "grad_norm": 0.2889077961444855, "learning_rate": 9.247245407207579e-05, "loss": 1.7546, "step": 801 }, { "epoch": 0.1936730258391693, "grad_norm": 0.3094068169593811, "learning_rate": 9.245201593887952e-05, "loss": 1.8725, "step": 802 }, { "epoch": 0.19391451340255977, "grad_norm": 0.29328641295433044, "learning_rate": 9.24315523623529e-05, "loss": 1.8439, "step": 803 }, { "epoch": 0.19415600096595026, "grad_norm": 0.2785046696662903, "learning_rate": 9.241106335476061e-05, "loss": 1.6542, "step": 804 }, { "epoch": 0.19439748852934075, "grad_norm": 0.2787037193775177, "learning_rate": 9.239054892838264e-05, "loss": 1.675, "step": 805 }, { "epoch": 0.19463897609273123, "grad_norm": 0.2860700488090515, "learning_rate": 9.237000909551415e-05, "loss": 1.7113, "step": 806 }, { "epoch": 0.19488046365612172, "grad_norm": 0.2860807776451111, "learning_rate": 9.234944386846552e-05, "loss": 1.7551, "step": 807 }, { "epoch": 0.1951219512195122, "grad_norm": 0.2986462414264679, "learning_rate": 9.232885325956243e-05, "loss": 1.7758, "step": 808 }, { "epoch": 0.1953634387829027, "grad_norm": 0.9163442254066467, "learning_rate": 9.23082372811457e-05, "loss": 2.1011, "step": 809 }, { "epoch": 0.19560492634629317, "grad_norm": 0.3197281062602997, "learning_rate": 9.22875959455714e-05, "loss": 1.9184, "step": 810 }, { "epoch": 0.19584641390968366, "grad_norm": 0.288152813911438, "learning_rate": 9.226692926521072e-05, "loss": 1.5923, "step": 811 }, { "epoch": 0.19608790147307414, "grad_norm": 0.3113833963871002, "learning_rate": 9.224623725245016e-05, "loss": 1.7094, "step": 812 }, { "epoch": 0.19632938903646463, "grad_norm": 0.28525310754776, "learning_rate": 9.22255199196913e-05, "loss": 1.7011, "step": 813 }, { "epoch": 0.19657087659985512, "grad_norm": 0.28587856888771057, "learning_rate": 9.220477727935095e-05, "loss": 1.5106, "step": 814 }, { "epoch": 0.1968123641632456, "grad_norm": 0.324881911277771, "learning_rate": 9.218400934386106e-05, "loss": 1.934, "step": 815 }, { "epoch": 0.1970538517266361, "grad_norm": 0.28722596168518066, "learning_rate": 9.216321612566879e-05, "loss": 1.8429, "step": 816 }, { "epoch": 0.19729533929002657, "grad_norm": 0.298000693321228, "learning_rate": 9.214239763723636e-05, "loss": 1.6637, "step": 817 }, { "epoch": 0.19753682685341706, "grad_norm": 0.29165783524513245, "learning_rate": 9.21215538910412e-05, "loss": 1.7989, "step": 818 }, { "epoch": 0.19777831441680754, "grad_norm": 0.28170880675315857, "learning_rate": 9.210068489957591e-05, "loss": 1.6985, "step": 819 }, { "epoch": 0.19801980198019803, "grad_norm": 0.29324185848236084, "learning_rate": 9.207979067534815e-05, "loss": 1.7398, "step": 820 }, { "epoch": 0.19826128954358851, "grad_norm": 0.29249295592308044, "learning_rate": 9.205887123088075e-05, "loss": 1.7616, "step": 821 }, { "epoch": 0.198502777106979, "grad_norm": 0.2885819375514984, "learning_rate": 9.203792657871163e-05, "loss": 1.6781, "step": 822 }, { "epoch": 0.19874426467036949, "grad_norm": 0.2888461947441101, "learning_rate": 9.201695673139381e-05, "loss": 1.7416, "step": 823 }, { "epoch": 0.19898575223375997, "grad_norm": 0.27995526790618896, "learning_rate": 9.199596170149543e-05, "loss": 1.7762, "step": 824 }, { "epoch": 0.19922723979715046, "grad_norm": 0.30026763677597046, "learning_rate": 9.197494150159976e-05, "loss": 1.8803, "step": 825 }, { "epoch": 0.19946872736054094, "grad_norm": 0.2981962561607361, "learning_rate": 9.195389614430508e-05, "loss": 1.6778, "step": 826 }, { "epoch": 0.19971021492393143, "grad_norm": 0.29537999629974365, "learning_rate": 9.193282564222477e-05, "loss": 1.6503, "step": 827 }, { "epoch": 0.19995170248732191, "grad_norm": 0.3164536654949188, "learning_rate": 9.191173000798732e-05, "loss": 1.9035, "step": 828 }, { "epoch": 0.2001931900507124, "grad_norm": 0.36678019165992737, "learning_rate": 9.189060925423624e-05, "loss": 2.0678, "step": 829 }, { "epoch": 0.20043467761410289, "grad_norm": 0.29730290174484253, "learning_rate": 9.186946339363013e-05, "loss": 1.9599, "step": 830 }, { "epoch": 0.20067616517749337, "grad_norm": 0.2871508002281189, "learning_rate": 9.184829243884264e-05, "loss": 1.748, "step": 831 }, { "epoch": 0.20091765274088386, "grad_norm": 0.2855934500694275, "learning_rate": 9.182709640256238e-05, "loss": 1.7618, "step": 832 }, { "epoch": 0.20115914030427434, "grad_norm": 0.2966662049293518, "learning_rate": 9.180587529749309e-05, "loss": 1.7395, "step": 833 }, { "epoch": 0.20140062786766483, "grad_norm": 0.29292529821395874, "learning_rate": 9.178462913635349e-05, "loss": 1.7003, "step": 834 }, { "epoch": 0.2016421154310553, "grad_norm": 0.29136964678764343, "learning_rate": 9.176335793187732e-05, "loss": 1.8498, "step": 835 }, { "epoch": 0.2018836029944458, "grad_norm": 0.3045142590999603, "learning_rate": 9.174206169681334e-05, "loss": 1.6792, "step": 836 }, { "epoch": 0.20212509055783628, "grad_norm": 0.3062690198421478, "learning_rate": 9.172074044392529e-05, "loss": 1.8017, "step": 837 }, { "epoch": 0.20236657812122677, "grad_norm": 0.27704912424087524, "learning_rate": 9.169939418599194e-05, "loss": 1.6993, "step": 838 }, { "epoch": 0.20260806568461726, "grad_norm": 0.29494112730026245, "learning_rate": 9.167802293580703e-05, "loss": 1.6878, "step": 839 }, { "epoch": 0.20284955324800774, "grad_norm": 0.3410157561302185, "learning_rate": 9.165662670617925e-05, "loss": 2.0581, "step": 840 }, { "epoch": 0.20309104081139823, "grad_norm": 0.28066882491111755, "learning_rate": 9.16352055099323e-05, "loss": 1.7619, "step": 841 }, { "epoch": 0.20333252837478868, "grad_norm": 0.27249637246131897, "learning_rate": 9.161375935990482e-05, "loss": 1.5283, "step": 842 }, { "epoch": 0.20357401593817917, "grad_norm": 0.3004692494869232, "learning_rate": 9.159228826895043e-05, "loss": 1.8467, "step": 843 }, { "epoch": 0.20381550350156966, "grad_norm": 0.2757605016231537, "learning_rate": 9.15707922499377e-05, "loss": 1.6736, "step": 844 }, { "epoch": 0.20405699106496014, "grad_norm": 0.281103253364563, "learning_rate": 9.154927131575009e-05, "loss": 1.7212, "step": 845 }, { "epoch": 0.20429847862835063, "grad_norm": 0.28760358691215515, "learning_rate": 9.152772547928601e-05, "loss": 1.6699, "step": 846 }, { "epoch": 0.2045399661917411, "grad_norm": 0.2798704504966736, "learning_rate": 9.150615475345888e-05, "loss": 1.6742, "step": 847 }, { "epoch": 0.2047814537551316, "grad_norm": 0.29464927315711975, "learning_rate": 9.148455915119691e-05, "loss": 1.8128, "step": 848 }, { "epoch": 0.20502294131852208, "grad_norm": 0.30707064270973206, "learning_rate": 9.146293868544333e-05, "loss": 1.916, "step": 849 }, { "epoch": 0.20526442888191257, "grad_norm": 0.2912966012954712, "learning_rate": 9.144129336915616e-05, "loss": 1.7689, "step": 850 }, { "epoch": 0.20550591644530306, "grad_norm": 0.28413230180740356, "learning_rate": 9.141962321530842e-05, "loss": 1.7297, "step": 851 }, { "epoch": 0.20574740400869354, "grad_norm": 0.2876090109348297, "learning_rate": 9.139792823688797e-05, "loss": 1.6406, "step": 852 }, { "epoch": 0.20598889157208403, "grad_norm": 0.3216966390609741, "learning_rate": 9.137620844689752e-05, "loss": 2.0118, "step": 853 }, { "epoch": 0.2062303791354745, "grad_norm": 0.28935253620147705, "learning_rate": 9.135446385835471e-05, "loss": 1.9161, "step": 854 }, { "epoch": 0.206471866698865, "grad_norm": 0.2942180931568146, "learning_rate": 9.133269448429199e-05, "loss": 1.7412, "step": 855 }, { "epoch": 0.20671335426225548, "grad_norm": 0.3060719072818756, "learning_rate": 9.13109003377567e-05, "loss": 1.9438, "step": 856 }, { "epoch": 0.20695484182564597, "grad_norm": 0.2826623320579529, "learning_rate": 9.128908143181101e-05, "loss": 1.7717, "step": 857 }, { "epoch": 0.20719632938903645, "grad_norm": 0.30230000615119934, "learning_rate": 9.126723777953195e-05, "loss": 1.8519, "step": 858 }, { "epoch": 0.20743781695242694, "grad_norm": 0.305400013923645, "learning_rate": 9.124536939401135e-05, "loss": 1.7981, "step": 859 }, { "epoch": 0.20767930451581743, "grad_norm": 0.2971997559070587, "learning_rate": 9.122347628835589e-05, "loss": 1.9272, "step": 860 }, { "epoch": 0.2079207920792079, "grad_norm": 0.2872750461101532, "learning_rate": 9.120155847568704e-05, "loss": 1.5787, "step": 861 }, { "epoch": 0.2081622796425984, "grad_norm": 0.2825159728527069, "learning_rate": 9.117961596914109e-05, "loss": 1.729, "step": 862 }, { "epoch": 0.20840376720598888, "grad_norm": 0.2875017821788788, "learning_rate": 9.115764878186917e-05, "loss": 1.7238, "step": 863 }, { "epoch": 0.20864525476937937, "grad_norm": 0.2908998727798462, "learning_rate": 9.113565692703714e-05, "loss": 1.6799, "step": 864 }, { "epoch": 0.20888674233276985, "grad_norm": 0.28806325793266296, "learning_rate": 9.111364041782566e-05, "loss": 1.7153, "step": 865 }, { "epoch": 0.20912822989616034, "grad_norm": 0.29157885909080505, "learning_rate": 9.10915992674302e-05, "loss": 1.6363, "step": 866 }, { "epoch": 0.20936971745955082, "grad_norm": 0.29479745030403137, "learning_rate": 9.106953348906097e-05, "loss": 1.9421, "step": 867 }, { "epoch": 0.2096112050229413, "grad_norm": 0.3266473412513733, "learning_rate": 9.104744309594294e-05, "loss": 1.9736, "step": 868 }, { "epoch": 0.2098526925863318, "grad_norm": 0.2955428659915924, "learning_rate": 9.102532810131584e-05, "loss": 1.945, "step": 869 }, { "epoch": 0.21009418014972228, "grad_norm": 0.29601460695266724, "learning_rate": 9.100318851843415e-05, "loss": 1.8434, "step": 870 }, { "epoch": 0.21033566771311277, "grad_norm": 0.2987689673900604, "learning_rate": 9.098102436056706e-05, "loss": 1.9485, "step": 871 }, { "epoch": 0.21057715527650325, "grad_norm": 0.27123886346817017, "learning_rate": 9.095883564099854e-05, "loss": 1.6151, "step": 872 }, { "epoch": 0.21081864283989374, "grad_norm": 0.28310680389404297, "learning_rate": 9.093662237302722e-05, "loss": 1.731, "step": 873 }, { "epoch": 0.21106013040328422, "grad_norm": 0.281590074300766, "learning_rate": 9.091438456996648e-05, "loss": 1.7374, "step": 874 }, { "epoch": 0.2113016179666747, "grad_norm": 0.2950079143047333, "learning_rate": 9.089212224514441e-05, "loss": 1.7526, "step": 875 }, { "epoch": 0.2115431055300652, "grad_norm": 0.2879007160663605, "learning_rate": 9.086983541190379e-05, "loss": 1.7355, "step": 876 }, { "epoch": 0.21178459309345568, "grad_norm": 0.2673324644565582, "learning_rate": 9.084752408360206e-05, "loss": 1.6847, "step": 877 }, { "epoch": 0.21202608065684617, "grad_norm": 0.2840292751789093, "learning_rate": 9.082518827361137e-05, "loss": 1.6642, "step": 878 }, { "epoch": 0.21226756822023665, "grad_norm": 0.2851640582084656, "learning_rate": 9.080282799531855e-05, "loss": 1.6895, "step": 879 }, { "epoch": 0.21250905578362714, "grad_norm": 0.30625519156455994, "learning_rate": 9.078044326212507e-05, "loss": 1.8807, "step": 880 }, { "epoch": 0.21275054334701762, "grad_norm": 0.2955930233001709, "learning_rate": 9.07580340874471e-05, "loss": 1.8051, "step": 881 }, { "epoch": 0.2129920309104081, "grad_norm": 0.3044041395187378, "learning_rate": 9.073560048471538e-05, "loss": 1.8127, "step": 882 }, { "epoch": 0.2132335184737986, "grad_norm": 0.26778867840766907, "learning_rate": 9.071314246737536e-05, "loss": 1.6614, "step": 883 }, { "epoch": 0.21347500603718908, "grad_norm": 0.28641968965530396, "learning_rate": 9.069066004888712e-05, "loss": 1.8822, "step": 884 }, { "epoch": 0.21371649360057957, "grad_norm": 0.2820203900337219, "learning_rate": 9.066815324272533e-05, "loss": 1.625, "step": 885 }, { "epoch": 0.21395798116397005, "grad_norm": 0.2761995792388916, "learning_rate": 9.064562206237929e-05, "loss": 1.6245, "step": 886 }, { "epoch": 0.21419946872736054, "grad_norm": 0.2768464684486389, "learning_rate": 9.062306652135292e-05, "loss": 1.7488, "step": 887 }, { "epoch": 0.21444095629075102, "grad_norm": 0.2869665026664734, "learning_rate": 9.060048663316472e-05, "loss": 1.7725, "step": 888 }, { "epoch": 0.2146824438541415, "grad_norm": 0.28295812010765076, "learning_rate": 9.05778824113478e-05, "loss": 1.724, "step": 889 }, { "epoch": 0.214923931417532, "grad_norm": 0.2879597246646881, "learning_rate": 9.055525386944985e-05, "loss": 1.7923, "step": 890 }, { "epoch": 0.21516541898092248, "grad_norm": 0.2905023396015167, "learning_rate": 9.053260102103315e-05, "loss": 1.6549, "step": 891 }, { "epoch": 0.21540690654431296, "grad_norm": 0.2903376817703247, "learning_rate": 9.05099238796745e-05, "loss": 1.7865, "step": 892 }, { "epoch": 0.21564839410770345, "grad_norm": 0.2778032422065735, "learning_rate": 9.048722245896531e-05, "loss": 1.6382, "step": 893 }, { "epoch": 0.21588988167109394, "grad_norm": 0.3031550943851471, "learning_rate": 9.046449677251155e-05, "loss": 1.7798, "step": 894 }, { "epoch": 0.21613136923448442, "grad_norm": 0.2920134365558624, "learning_rate": 9.044174683393365e-05, "loss": 1.7833, "step": 895 }, { "epoch": 0.2163728567978749, "grad_norm": 0.27324342727661133, "learning_rate": 9.041897265686666e-05, "loss": 1.6678, "step": 896 }, { "epoch": 0.2166143443612654, "grad_norm": 0.289196640253067, "learning_rate": 9.039617425496015e-05, "loss": 1.7339, "step": 897 }, { "epoch": 0.21685583192465588, "grad_norm": 0.26956117153167725, "learning_rate": 9.037335164187817e-05, "loss": 1.7743, "step": 898 }, { "epoch": 0.21709731948804636, "grad_norm": 0.2876354455947876, "learning_rate": 9.035050483129928e-05, "loss": 1.7449, "step": 899 }, { "epoch": 0.21733880705143685, "grad_norm": 0.29619812965393066, "learning_rate": 9.032763383691662e-05, "loss": 1.8701, "step": 900 }, { "epoch": 0.21758029461482734, "grad_norm": 0.27593353390693665, "learning_rate": 9.03047386724377e-05, "loss": 1.6191, "step": 901 }, { "epoch": 0.21782178217821782, "grad_norm": 0.2756110727787018, "learning_rate": 9.028181935158463e-05, "loss": 1.7214, "step": 902 }, { "epoch": 0.2180632697416083, "grad_norm": 0.29085442423820496, "learning_rate": 9.025887588809394e-05, "loss": 1.8983, "step": 903 }, { "epoch": 0.2183047573049988, "grad_norm": 0.30053767561912537, "learning_rate": 9.023590829571663e-05, "loss": 2.027, "step": 904 }, { "epoch": 0.21854624486838928, "grad_norm": 0.27619075775146484, "learning_rate": 9.021291658821815e-05, "loss": 1.6083, "step": 905 }, { "epoch": 0.21878773243177976, "grad_norm": 0.2857961356639862, "learning_rate": 9.018990077937848e-05, "loss": 1.7759, "step": 906 }, { "epoch": 0.21902921999517025, "grad_norm": 0.2999686598777771, "learning_rate": 9.016686088299196e-05, "loss": 1.7725, "step": 907 }, { "epoch": 0.21927070755856073, "grad_norm": 0.2941741943359375, "learning_rate": 9.01437969128674e-05, "loss": 1.7253, "step": 908 }, { "epoch": 0.21951219512195122, "grad_norm": 0.2836509644985199, "learning_rate": 9.012070888282803e-05, "loss": 1.7763, "step": 909 }, { "epoch": 0.2197536826853417, "grad_norm": 0.2917303740978241, "learning_rate": 9.009759680671151e-05, "loss": 1.6533, "step": 910 }, { "epoch": 0.2199951702487322, "grad_norm": 0.27767467498779297, "learning_rate": 9.007446069836991e-05, "loss": 1.7, "step": 911 }, { "epoch": 0.22023665781212268, "grad_norm": 0.27494391798973083, "learning_rate": 9.005130057166971e-05, "loss": 1.6589, "step": 912 }, { "epoch": 0.22047814537551316, "grad_norm": 0.3042466342449188, "learning_rate": 9.002811644049176e-05, "loss": 1.8579, "step": 913 }, { "epoch": 0.22071963293890365, "grad_norm": 0.308646023273468, "learning_rate": 9.00049083187313e-05, "loss": 1.911, "step": 914 }, { "epoch": 0.22096112050229413, "grad_norm": 0.2836478054523468, "learning_rate": 8.998167622029801e-05, "loss": 1.8678, "step": 915 }, { "epoch": 0.22120260806568462, "grad_norm": 0.3099924325942993, "learning_rate": 8.995842015911584e-05, "loss": 1.8014, "step": 916 }, { "epoch": 0.2214440956290751, "grad_norm": 0.2966037094593048, "learning_rate": 8.993514014912318e-05, "loss": 1.8265, "step": 917 }, { "epoch": 0.2216855831924656, "grad_norm": 0.2903251051902771, "learning_rate": 8.991183620427275e-05, "loss": 1.7539, "step": 918 }, { "epoch": 0.22192707075585608, "grad_norm": 0.28033512830734253, "learning_rate": 8.98885083385316e-05, "loss": 1.5729, "step": 919 }, { "epoch": 0.22216855831924656, "grad_norm": 0.29812338948249817, "learning_rate": 8.986515656588114e-05, "loss": 1.6498, "step": 920 }, { "epoch": 0.22241004588263705, "grad_norm": 0.29040244221687317, "learning_rate": 8.98417809003171e-05, "loss": 1.8032, "step": 921 }, { "epoch": 0.22265153344602753, "grad_norm": 0.2998603880405426, "learning_rate": 8.981838135584953e-05, "loss": 1.7509, "step": 922 }, { "epoch": 0.22289302100941802, "grad_norm": 0.2951640188694, "learning_rate": 8.979495794650276e-05, "loss": 1.8133, "step": 923 }, { "epoch": 0.2231345085728085, "grad_norm": 0.2719232439994812, "learning_rate": 8.977151068631548e-05, "loss": 1.6761, "step": 924 }, { "epoch": 0.223375996136199, "grad_norm": 0.29052817821502686, "learning_rate": 8.974803958934064e-05, "loss": 1.7221, "step": 925 }, { "epoch": 0.22361748369958948, "grad_norm": 0.2881939113140106, "learning_rate": 8.972454466964551e-05, "loss": 1.6036, "step": 926 }, { "epoch": 0.22385897126297996, "grad_norm": 0.2801560163497925, "learning_rate": 8.970102594131156e-05, "loss": 1.59, "step": 927 }, { "epoch": 0.22410045882637045, "grad_norm": 0.289957195520401, "learning_rate": 8.967748341843462e-05, "loss": 1.9019, "step": 928 }, { "epoch": 0.22434194638976093, "grad_norm": 0.29187458753585815, "learning_rate": 8.965391711512475e-05, "loss": 1.6741, "step": 929 }, { "epoch": 0.22458343395315142, "grad_norm": 0.2863827049732208, "learning_rate": 8.96303270455062e-05, "loss": 1.8284, "step": 930 }, { "epoch": 0.2248249215165419, "grad_norm": 0.29024264216423035, "learning_rate": 8.960671322371759e-05, "loss": 1.7697, "step": 931 }, { "epoch": 0.2250664090799324, "grad_norm": 0.273242712020874, "learning_rate": 8.958307566391166e-05, "loss": 1.7714, "step": 932 }, { "epoch": 0.22530789664332287, "grad_norm": 0.298308789730072, "learning_rate": 8.955941438025545e-05, "loss": 1.7871, "step": 933 }, { "epoch": 0.22554938420671336, "grad_norm": 0.28436291217803955, "learning_rate": 8.953572938693016e-05, "loss": 1.7225, "step": 934 }, { "epoch": 0.22579087177010385, "grad_norm": 0.2844787836074829, "learning_rate": 8.951202069813126e-05, "loss": 1.6326, "step": 935 }, { "epoch": 0.22603235933349433, "grad_norm": 0.28223082423210144, "learning_rate": 8.948828832806838e-05, "loss": 1.8086, "step": 936 }, { "epoch": 0.22627384689688482, "grad_norm": 0.3171670734882355, "learning_rate": 8.946453229096537e-05, "loss": 1.7749, "step": 937 }, { "epoch": 0.2265153344602753, "grad_norm": 0.27827945351600647, "learning_rate": 8.944075260106022e-05, "loss": 1.7798, "step": 938 }, { "epoch": 0.2267568220236658, "grad_norm": 0.3321087062358856, "learning_rate": 8.941694927260516e-05, "loss": 2.0057, "step": 939 }, { "epoch": 0.22699830958705627, "grad_norm": 0.3078181743621826, "learning_rate": 8.939312231986655e-05, "loss": 1.7606, "step": 940 }, { "epoch": 0.22723979715044676, "grad_norm": 0.3000055253505707, "learning_rate": 8.936927175712488e-05, "loss": 1.8077, "step": 941 }, { "epoch": 0.22748128471383725, "grad_norm": 0.3055233955383301, "learning_rate": 8.934539759867487e-05, "loss": 1.8196, "step": 942 }, { "epoch": 0.22772277227722773, "grad_norm": 0.30723288655281067, "learning_rate": 8.932149985882529e-05, "loss": 1.7099, "step": 943 }, { "epoch": 0.22796425984061822, "grad_norm": 0.2921900153160095, "learning_rate": 8.929757855189915e-05, "loss": 1.7364, "step": 944 }, { "epoch": 0.2282057474040087, "grad_norm": 0.3061923384666443, "learning_rate": 8.927363369223348e-05, "loss": 1.6376, "step": 945 }, { "epoch": 0.2284472349673992, "grad_norm": 0.2868843078613281, "learning_rate": 8.924966529417949e-05, "loss": 1.7675, "step": 946 }, { "epoch": 0.22868872253078967, "grad_norm": 0.2890067994594574, "learning_rate": 8.922567337210245e-05, "loss": 1.7781, "step": 947 }, { "epoch": 0.22893021009418016, "grad_norm": 0.30553463101387024, "learning_rate": 8.920165794038177e-05, "loss": 1.8075, "step": 948 }, { "epoch": 0.22917169765757064, "grad_norm": 0.27772581577301025, "learning_rate": 8.917761901341097e-05, "loss": 1.6267, "step": 949 }, { "epoch": 0.22941318522096113, "grad_norm": 0.3060512840747833, "learning_rate": 8.915355660559758e-05, "loss": 1.8912, "step": 950 }, { "epoch": 0.22965467278435162, "grad_norm": 0.30589228868484497, "learning_rate": 8.912947073136327e-05, "loss": 1.8268, "step": 951 }, { "epoch": 0.2298961603477421, "grad_norm": 0.29312169551849365, "learning_rate": 8.910536140514372e-05, "loss": 1.6905, "step": 952 }, { "epoch": 0.2301376479111326, "grad_norm": 0.29430586099624634, "learning_rate": 8.908122864138872e-05, "loss": 1.7801, "step": 953 }, { "epoch": 0.23037913547452307, "grad_norm": 0.28983962535858154, "learning_rate": 8.905707245456206e-05, "loss": 1.8132, "step": 954 }, { "epoch": 0.23062062303791356, "grad_norm": 0.3049819767475128, "learning_rate": 8.903289285914159e-05, "loss": 1.8283, "step": 955 }, { "epoch": 0.23086211060130404, "grad_norm": 0.2923794090747833, "learning_rate": 8.900868986961919e-05, "loss": 1.8282, "step": 956 }, { "epoch": 0.23110359816469453, "grad_norm": 0.2887613773345947, "learning_rate": 8.898446350050076e-05, "loss": 1.7958, "step": 957 }, { "epoch": 0.23134508572808501, "grad_norm": 0.2974589169025421, "learning_rate": 8.896021376630621e-05, "loss": 1.8724, "step": 958 }, { "epoch": 0.2315865732914755, "grad_norm": 0.2794170081615448, "learning_rate": 8.893594068156947e-05, "loss": 1.7259, "step": 959 }, { "epoch": 0.23182806085486599, "grad_norm": 0.2978573441505432, "learning_rate": 8.891164426083842e-05, "loss": 1.7693, "step": 960 }, { "epoch": 0.23206954841825647, "grad_norm": 0.30293652415275574, "learning_rate": 8.888732451867498e-05, "loss": 1.7213, "step": 961 }, { "epoch": 0.23231103598164696, "grad_norm": 0.2683996558189392, "learning_rate": 8.886298146965501e-05, "loss": 1.7209, "step": 962 }, { "epoch": 0.23255252354503744, "grad_norm": 0.2920992970466614, "learning_rate": 8.883861512836839e-05, "loss": 1.5465, "step": 963 }, { "epoch": 0.23279401110842793, "grad_norm": 0.2909085154533386, "learning_rate": 8.881422550941887e-05, "loss": 1.723, "step": 964 }, { "epoch": 0.23303549867181841, "grad_norm": 0.28865692019462585, "learning_rate": 8.878981262742423e-05, "loss": 1.8612, "step": 965 }, { "epoch": 0.2332769862352089, "grad_norm": 0.27150455117225647, "learning_rate": 8.876537649701618e-05, "loss": 1.5894, "step": 966 }, { "epoch": 0.23351847379859939, "grad_norm": 0.2986278533935547, "learning_rate": 8.87409171328403e-05, "loss": 1.8431, "step": 967 }, { "epoch": 0.23375996136198987, "grad_norm": 0.29470667243003845, "learning_rate": 8.87164345495562e-05, "loss": 1.9004, "step": 968 }, { "epoch": 0.23400144892538036, "grad_norm": 0.2897234559059143, "learning_rate": 8.869192876183734e-05, "loss": 1.78, "step": 969 }, { "epoch": 0.23424293648877084, "grad_norm": 0.29372528195381165, "learning_rate": 8.866739978437109e-05, "loss": 1.8018, "step": 970 }, { "epoch": 0.2344844240521613, "grad_norm": 0.27670425176620483, "learning_rate": 8.864284763185873e-05, "loss": 1.6661, "step": 971 }, { "epoch": 0.23472591161555179, "grad_norm": 0.2911158800125122, "learning_rate": 8.861827231901541e-05, "loss": 1.9302, "step": 972 }, { "epoch": 0.23496739917894227, "grad_norm": 0.27287617325782776, "learning_rate": 8.85936738605702e-05, "loss": 1.6371, "step": 973 }, { "epoch": 0.23520888674233276, "grad_norm": 0.2705206871032715, "learning_rate": 8.856905227126602e-05, "loss": 1.6918, "step": 974 }, { "epoch": 0.23545037430572324, "grad_norm": 0.28230130672454834, "learning_rate": 8.854440756585964e-05, "loss": 1.7005, "step": 975 }, { "epoch": 0.23569186186911373, "grad_norm": 0.2819898724555969, "learning_rate": 8.851973975912173e-05, "loss": 1.7495, "step": 976 }, { "epoch": 0.2359333494325042, "grad_norm": 0.2814916670322418, "learning_rate": 8.849504886583674e-05, "loss": 1.6623, "step": 977 }, { "epoch": 0.2361748369958947, "grad_norm": 0.2934419512748718, "learning_rate": 8.847033490080301e-05, "loss": 1.7376, "step": 978 }, { "epoch": 0.23641632455928518, "grad_norm": 0.3402636647224426, "learning_rate": 8.844559787883266e-05, "loss": 2.0732, "step": 979 }, { "epoch": 0.23665781212267567, "grad_norm": 0.2801571786403656, "learning_rate": 8.84208378147517e-05, "loss": 1.6666, "step": 980 }, { "epoch": 0.23689929968606616, "grad_norm": 0.31147944927215576, "learning_rate": 8.839605472339988e-05, "loss": 1.7559, "step": 981 }, { "epoch": 0.23714078724945664, "grad_norm": 0.2805061936378479, "learning_rate": 8.83712486196308e-05, "loss": 1.7002, "step": 982 }, { "epoch": 0.23738227481284713, "grad_norm": 0.2966560423374176, "learning_rate": 8.83464195183118e-05, "loss": 1.7956, "step": 983 }, { "epoch": 0.2376237623762376, "grad_norm": 0.29622846841812134, "learning_rate": 8.832156743432405e-05, "loss": 1.7916, "step": 984 }, { "epoch": 0.2378652499396281, "grad_norm": 0.29022935032844543, "learning_rate": 8.829669238256251e-05, "loss": 1.7386, "step": 985 }, { "epoch": 0.23810673750301858, "grad_norm": 0.2820836007595062, "learning_rate": 8.827179437793582e-05, "loss": 1.7125, "step": 986 }, { "epoch": 0.23834822506640907, "grad_norm": 0.2786775529384613, "learning_rate": 8.824687343536648e-05, "loss": 1.9048, "step": 987 }, { "epoch": 0.23858971262979956, "grad_norm": 0.273759663105011, "learning_rate": 8.822192956979066e-05, "loss": 1.6711, "step": 988 }, { "epoch": 0.23883120019319004, "grad_norm": 0.2763570249080658, "learning_rate": 8.81969627961583e-05, "loss": 1.6444, "step": 989 }, { "epoch": 0.23907268775658053, "grad_norm": 0.2795482277870178, "learning_rate": 8.817197312943308e-05, "loss": 1.7041, "step": 990 }, { "epoch": 0.239314175319971, "grad_norm": 0.28513190150260925, "learning_rate": 8.81469605845924e-05, "loss": 1.74, "step": 991 }, { "epoch": 0.2395556628833615, "grad_norm": 0.3679279088973999, "learning_rate": 8.812192517662732e-05, "loss": 1.6976, "step": 992 }, { "epoch": 0.23979715044675198, "grad_norm": 0.3091084957122803, "learning_rate": 8.809686692054268e-05, "loss": 1.9301, "step": 993 }, { "epoch": 0.24003863801014247, "grad_norm": 0.3088282644748688, "learning_rate": 8.807178583135697e-05, "loss": 1.8976, "step": 994 }, { "epoch": 0.24028012557353295, "grad_norm": 0.275089293718338, "learning_rate": 8.804668192410235e-05, "loss": 1.7477, "step": 995 }, { "epoch": 0.24052161313692344, "grad_norm": 0.2925816476345062, "learning_rate": 8.802155521382471e-05, "loss": 1.9009, "step": 996 }, { "epoch": 0.24076310070031393, "grad_norm": 0.2990473806858063, "learning_rate": 8.799640571558357e-05, "loss": 1.796, "step": 997 }, { "epoch": 0.2410045882637044, "grad_norm": 0.2928846776485443, "learning_rate": 8.797123344445208e-05, "loss": 1.7707, "step": 998 }, { "epoch": 0.2412460758270949, "grad_norm": 0.28296223282814026, "learning_rate": 8.794603841551712e-05, "loss": 1.7324, "step": 999 }, { "epoch": 0.24148756339048538, "grad_norm": 0.291424036026001, "learning_rate": 8.79208206438791e-05, "loss": 1.9164, "step": 1000 }, { "epoch": 0.24172905095387587, "grad_norm": 0.2974405884742737, "learning_rate": 8.789558014465218e-05, "loss": 1.8217, "step": 1001 }, { "epoch": 0.24197053851726635, "grad_norm": 0.2871314287185669, "learning_rate": 8.787031693296407e-05, "loss": 1.7372, "step": 1002 }, { "epoch": 0.24221202608065684, "grad_norm": 0.28517138957977295, "learning_rate": 8.784503102395609e-05, "loss": 1.7268, "step": 1003 }, { "epoch": 0.24245351364404732, "grad_norm": 0.2855314612388611, "learning_rate": 8.781972243278319e-05, "loss": 1.7284, "step": 1004 }, { "epoch": 0.2426950012074378, "grad_norm": 0.30058592557907104, "learning_rate": 8.779439117461388e-05, "loss": 1.676, "step": 1005 }, { "epoch": 0.2429364887708283, "grad_norm": 0.2805761694908142, "learning_rate": 8.776903726463031e-05, "loss": 1.6762, "step": 1006 }, { "epoch": 0.24317797633421878, "grad_norm": 0.2890090346336365, "learning_rate": 8.774366071802817e-05, "loss": 1.772, "step": 1007 }, { "epoch": 0.24341946389760927, "grad_norm": 0.287696897983551, "learning_rate": 8.771826155001673e-05, "loss": 1.8927, "step": 1008 }, { "epoch": 0.24366095146099975, "grad_norm": 0.28548726439476013, "learning_rate": 8.769283977581878e-05, "loss": 1.7648, "step": 1009 }, { "epoch": 0.24390243902439024, "grad_norm": 0.28890261054039, "learning_rate": 8.766739541067072e-05, "loss": 1.8778, "step": 1010 }, { "epoch": 0.24414392658778072, "grad_norm": 0.28177410364151, "learning_rate": 8.764192846982244e-05, "loss": 1.7047, "step": 1011 }, { "epoch": 0.2443854141511712, "grad_norm": 0.2893199026584625, "learning_rate": 8.761643896853739e-05, "loss": 1.6994, "step": 1012 }, { "epoch": 0.2446269017145617, "grad_norm": 0.2756044566631317, "learning_rate": 8.759092692209251e-05, "loss": 1.7141, "step": 1013 }, { "epoch": 0.24486838927795218, "grad_norm": 0.29576975107192993, "learning_rate": 8.756539234577834e-05, "loss": 1.7438, "step": 1014 }, { "epoch": 0.24510987684134267, "grad_norm": 0.31498026847839355, "learning_rate": 8.753983525489878e-05, "loss": 1.9633, "step": 1015 }, { "epoch": 0.24535136440473315, "grad_norm": 0.28949326276779175, "learning_rate": 8.751425566477131e-05, "loss": 1.7926, "step": 1016 }, { "epoch": 0.24559285196812364, "grad_norm": 0.2963774502277374, "learning_rate": 8.748865359072694e-05, "loss": 1.7185, "step": 1017 }, { "epoch": 0.24583433953151412, "grad_norm": 0.27066588401794434, "learning_rate": 8.746302904811004e-05, "loss": 1.7726, "step": 1018 }, { "epoch": 0.2460758270949046, "grad_norm": 0.29991111159324646, "learning_rate": 8.743738205227854e-05, "loss": 1.9028, "step": 1019 }, { "epoch": 0.2463173146582951, "grad_norm": 0.3058822453022003, "learning_rate": 8.74117126186038e-05, "loss": 1.7786, "step": 1020 }, { "epoch": 0.24655880222168558, "grad_norm": 0.28362077474594116, "learning_rate": 8.73860207624706e-05, "loss": 1.6691, "step": 1021 }, { "epoch": 0.24680028978507607, "grad_norm": 0.29726213216781616, "learning_rate": 8.736030649927717e-05, "loss": 1.7436, "step": 1022 }, { "epoch": 0.24704177734846655, "grad_norm": 0.278328001499176, "learning_rate": 8.733456984443522e-05, "loss": 1.6813, "step": 1023 }, { "epoch": 0.24728326491185704, "grad_norm": 0.30932289361953735, "learning_rate": 8.730881081336984e-05, "loss": 1.9074, "step": 1024 }, { "epoch": 0.24752475247524752, "grad_norm": 0.28863757848739624, "learning_rate": 8.728302942151949e-05, "loss": 1.7997, "step": 1025 }, { "epoch": 0.247766240038638, "grad_norm": 0.28880390524864197, "learning_rate": 8.725722568433612e-05, "loss": 1.7873, "step": 1026 }, { "epoch": 0.2480077276020285, "grad_norm": 0.27976545691490173, "learning_rate": 8.723139961728499e-05, "loss": 1.6941, "step": 1027 }, { "epoch": 0.24824921516541898, "grad_norm": 0.2951839864253998, "learning_rate": 8.72055512358448e-05, "loss": 1.8255, "step": 1028 }, { "epoch": 0.24849070272880946, "grad_norm": 0.29703572392463684, "learning_rate": 8.717968055550759e-05, "loss": 1.6139, "step": 1029 }, { "epoch": 0.24873219029219995, "grad_norm": 0.3209615647792816, "learning_rate": 8.71537875917788e-05, "loss": 1.8684, "step": 1030 }, { "epoch": 0.24897367785559044, "grad_norm": 0.32092544436454773, "learning_rate": 8.712787236017719e-05, "loss": 1.8249, "step": 1031 }, { "epoch": 0.24921516541898092, "grad_norm": 0.28598418831825256, "learning_rate": 8.710193487623488e-05, "loss": 1.7439, "step": 1032 }, { "epoch": 0.2494566529823714, "grad_norm": 0.30598631501197815, "learning_rate": 8.707597515549733e-05, "loss": 1.7287, "step": 1033 }, { "epoch": 0.2496981405457619, "grad_norm": 0.31710535287857056, "learning_rate": 8.704999321352331e-05, "loss": 1.9099, "step": 1034 }, { "epoch": 0.24993962810915238, "grad_norm": 0.3359716236591339, "learning_rate": 8.702398906588494e-05, "loss": 1.9935, "step": 1035 }, { "epoch": 0.25018111567254286, "grad_norm": 0.3071852922439575, "learning_rate": 8.699796272816762e-05, "loss": 1.7408, "step": 1036 }, { "epoch": 0.25042260323593335, "grad_norm": 0.3323792815208435, "learning_rate": 8.697191421597008e-05, "loss": 1.8794, "step": 1037 }, { "epoch": 0.25066409079932384, "grad_norm": 0.29584386944770813, "learning_rate": 8.69458435449043e-05, "loss": 1.7439, "step": 1038 }, { "epoch": 0.2509055783627143, "grad_norm": 0.28571659326553345, "learning_rate": 8.691975073059555e-05, "loss": 1.6339, "step": 1039 }, { "epoch": 0.2511470659261048, "grad_norm": 0.2837628722190857, "learning_rate": 8.689363578868241e-05, "loss": 1.717, "step": 1040 }, { "epoch": 0.2513885534894953, "grad_norm": 0.2912003993988037, "learning_rate": 8.68674987348167e-05, "loss": 1.648, "step": 1041 }, { "epoch": 0.2516300410528858, "grad_norm": 0.30284371972084045, "learning_rate": 8.684133958466347e-05, "loss": 1.8681, "step": 1042 }, { "epoch": 0.25187152861627626, "grad_norm": 0.27904531359672546, "learning_rate": 8.681515835390101e-05, "loss": 1.6447, "step": 1043 }, { "epoch": 0.25211301617966675, "grad_norm": 0.300796240568161, "learning_rate": 8.678895505822091e-05, "loss": 1.6384, "step": 1044 }, { "epoch": 0.25235450374305723, "grad_norm": 0.2925683557987213, "learning_rate": 8.676272971332791e-05, "loss": 1.7102, "step": 1045 }, { "epoch": 0.2525959913064477, "grad_norm": 0.29522907733917236, "learning_rate": 8.673648233494e-05, "loss": 1.7875, "step": 1046 }, { "epoch": 0.2528374788698382, "grad_norm": 0.2745014429092407, "learning_rate": 8.671021293878838e-05, "loss": 1.7905, "step": 1047 }, { "epoch": 0.2530789664332287, "grad_norm": 0.28517085313796997, "learning_rate": 8.668392154061741e-05, "loss": 1.7684, "step": 1048 }, { "epoch": 0.2533204539966192, "grad_norm": 0.29019320011138916, "learning_rate": 8.66576081561847e-05, "loss": 1.8825, "step": 1049 }, { "epoch": 0.25356194156000966, "grad_norm": 0.2811944782733917, "learning_rate": 8.663127280126096e-05, "loss": 1.7097, "step": 1050 }, { "epoch": 0.25380342912340015, "grad_norm": 0.2872609496116638, "learning_rate": 8.660491549163014e-05, "loss": 1.7745, "step": 1051 }, { "epoch": 0.25404491668679063, "grad_norm": 0.27062755823135376, "learning_rate": 8.65785362430893e-05, "loss": 1.536, "step": 1052 }, { "epoch": 0.2542864042501811, "grad_norm": 0.28564590215682983, "learning_rate": 8.655213507144868e-05, "loss": 1.6888, "step": 1053 }, { "epoch": 0.2545278918135716, "grad_norm": 0.295981228351593, "learning_rate": 8.652571199253161e-05, "loss": 1.8429, "step": 1054 }, { "epoch": 0.2547693793769621, "grad_norm": 0.27825748920440674, "learning_rate": 8.649926702217463e-05, "loss": 1.7498, "step": 1055 }, { "epoch": 0.2550108669403526, "grad_norm": 0.26681530475616455, "learning_rate": 8.647280017622733e-05, "loss": 1.7221, "step": 1056 }, { "epoch": 0.25525235450374306, "grad_norm": 0.2864789664745331, "learning_rate": 8.644631147055244e-05, "loss": 1.7624, "step": 1057 }, { "epoch": 0.25549384206713355, "grad_norm": 0.28378716111183167, "learning_rate": 8.641980092102579e-05, "loss": 1.7936, "step": 1058 }, { "epoch": 0.25573532963052403, "grad_norm": 0.2861391007900238, "learning_rate": 8.639326854353628e-05, "loss": 1.7771, "step": 1059 }, { "epoch": 0.2559768171939145, "grad_norm": 0.29720035195350647, "learning_rate": 8.636671435398595e-05, "loss": 1.7993, "step": 1060 }, { "epoch": 0.256218304757305, "grad_norm": 0.28660377860069275, "learning_rate": 8.634013836828985e-05, "loss": 1.6742, "step": 1061 }, { "epoch": 0.2564597923206955, "grad_norm": 0.2879665195941925, "learning_rate": 8.631354060237614e-05, "loss": 1.7474, "step": 1062 }, { "epoch": 0.256701279884086, "grad_norm": 0.30121028423309326, "learning_rate": 8.628692107218596e-05, "loss": 1.7211, "step": 1063 }, { "epoch": 0.25694276744747646, "grad_norm": 0.2766074538230896, "learning_rate": 8.62602797936736e-05, "loss": 1.6623, "step": 1064 }, { "epoch": 0.25718425501086695, "grad_norm": 0.2935093939304352, "learning_rate": 8.62336167828063e-05, "loss": 1.6995, "step": 1065 }, { "epoch": 0.25742574257425743, "grad_norm": 0.2767849862575531, "learning_rate": 8.62069320555644e-05, "loss": 1.794, "step": 1066 }, { "epoch": 0.2576672301376479, "grad_norm": 0.29953351616859436, "learning_rate": 8.618022562794114e-05, "loss": 1.8499, "step": 1067 }, { "epoch": 0.2579087177010384, "grad_norm": 0.4503518342971802, "learning_rate": 8.615349751594291e-05, "loss": 1.7237, "step": 1068 }, { "epoch": 0.2581502052644289, "grad_norm": 0.278839647769928, "learning_rate": 8.612674773558899e-05, "loss": 1.6459, "step": 1069 }, { "epoch": 0.2583916928278194, "grad_norm": 0.2816259264945984, "learning_rate": 8.609997630291167e-05, "loss": 1.7863, "step": 1070 }, { "epoch": 0.25863318039120986, "grad_norm": 0.30857524275779724, "learning_rate": 8.607318323395626e-05, "loss": 1.8289, "step": 1071 }, { "epoch": 0.25887466795460035, "grad_norm": 0.288631796836853, "learning_rate": 8.604636854478098e-05, "loss": 1.7618, "step": 1072 }, { "epoch": 0.25911615551799083, "grad_norm": 0.29358530044555664, "learning_rate": 8.601953225145704e-05, "loss": 1.9032, "step": 1073 }, { "epoch": 0.2593576430813813, "grad_norm": 0.29718780517578125, "learning_rate": 8.599267437006859e-05, "loss": 1.8647, "step": 1074 }, { "epoch": 0.2595991306447718, "grad_norm": 0.2939820885658264, "learning_rate": 8.596579491671273e-05, "loss": 1.7947, "step": 1075 }, { "epoch": 0.2598406182081623, "grad_norm": 0.29942068457603455, "learning_rate": 8.593889390749947e-05, "loss": 1.7313, "step": 1076 }, { "epoch": 0.2600821057715528, "grad_norm": 0.28265923261642456, "learning_rate": 8.591197135855175e-05, "loss": 1.8637, "step": 1077 }, { "epoch": 0.26032359333494326, "grad_norm": 0.28485462069511414, "learning_rate": 8.588502728600544e-05, "loss": 1.7327, "step": 1078 }, { "epoch": 0.26056508089833375, "grad_norm": 0.2807444632053375, "learning_rate": 8.585806170600926e-05, "loss": 1.784, "step": 1079 }, { "epoch": 0.26080656846172423, "grad_norm": 0.3553418219089508, "learning_rate": 8.583107463472484e-05, "loss": 1.6733, "step": 1080 }, { "epoch": 0.2610480560251147, "grad_norm": 0.27911999821662903, "learning_rate": 8.580406608832675e-05, "loss": 1.7077, "step": 1081 }, { "epoch": 0.2612895435885052, "grad_norm": 0.28452926874160767, "learning_rate": 8.577703608300234e-05, "loss": 1.7534, "step": 1082 }, { "epoch": 0.2615310311518957, "grad_norm": 0.3004327118396759, "learning_rate": 8.574998463495187e-05, "loss": 1.7699, "step": 1083 }, { "epoch": 0.2617725187152862, "grad_norm": 0.28714221715927124, "learning_rate": 8.572291176038845e-05, "loss": 1.7046, "step": 1084 }, { "epoch": 0.26201400627867666, "grad_norm": 0.2888624668121338, "learning_rate": 8.569581747553801e-05, "loss": 1.764, "step": 1085 }, { "epoch": 0.26225549384206714, "grad_norm": 0.2808159291744232, "learning_rate": 8.566870179663935e-05, "loss": 1.6601, "step": 1086 }, { "epoch": 0.26249698140545763, "grad_norm": 0.2881624102592468, "learning_rate": 8.564156473994405e-05, "loss": 1.8495, "step": 1087 }, { "epoch": 0.2627384689688481, "grad_norm": 0.28251782059669495, "learning_rate": 8.561440632171653e-05, "loss": 1.6594, "step": 1088 }, { "epoch": 0.2629799565322386, "grad_norm": 0.3031906187534332, "learning_rate": 8.558722655823401e-05, "loss": 1.8254, "step": 1089 }, { "epoch": 0.2632214440956291, "grad_norm": 0.31372591853141785, "learning_rate": 8.55600254657865e-05, "loss": 1.7989, "step": 1090 }, { "epoch": 0.2634629316590196, "grad_norm": 0.2720772922039032, "learning_rate": 8.553280306067678e-05, "loss": 1.6419, "step": 1091 }, { "epoch": 0.26370441922241006, "grad_norm": 0.26789841055870056, "learning_rate": 8.550555935922042e-05, "loss": 1.6165, "step": 1092 }, { "epoch": 0.26394590678580054, "grad_norm": 0.2883191406726837, "learning_rate": 8.547829437774577e-05, "loss": 1.6455, "step": 1093 }, { "epoch": 0.26418739434919103, "grad_norm": 0.29379168152809143, "learning_rate": 8.545100813259387e-05, "loss": 1.785, "step": 1094 }, { "epoch": 0.2644288819125815, "grad_norm": 0.2879287004470825, "learning_rate": 8.542370064011858e-05, "loss": 1.8095, "step": 1095 }, { "epoch": 0.264670369475972, "grad_norm": 0.29632505774497986, "learning_rate": 8.539637191668646e-05, "loss": 1.7733, "step": 1096 }, { "epoch": 0.2649118570393625, "grad_norm": 0.28169459104537964, "learning_rate": 8.53690219786768e-05, "loss": 1.6289, "step": 1097 }, { "epoch": 0.26515334460275297, "grad_norm": 0.31244707107543945, "learning_rate": 8.534165084248157e-05, "loss": 1.9603, "step": 1098 }, { "epoch": 0.26539483216614346, "grad_norm": 0.29472532868385315, "learning_rate": 8.531425852450552e-05, "loss": 1.8188, "step": 1099 }, { "epoch": 0.26563631972953394, "grad_norm": 0.27472302317619324, "learning_rate": 8.528684504116601e-05, "loss": 1.5912, "step": 1100 }, { "epoch": 0.26587780729292443, "grad_norm": 0.2806059420108795, "learning_rate": 8.525941040889315e-05, "loss": 1.5864, "step": 1101 }, { "epoch": 0.2661192948563149, "grad_norm": 0.3072028160095215, "learning_rate": 8.523195464412972e-05, "loss": 1.7789, "step": 1102 }, { "epoch": 0.2663607824197054, "grad_norm": 0.2739306390285492, "learning_rate": 8.520447776333113e-05, "loss": 1.631, "step": 1103 }, { "epoch": 0.2666022699830959, "grad_norm": 0.3055983781814575, "learning_rate": 8.517697978296544e-05, "loss": 1.8486, "step": 1104 }, { "epoch": 0.26684375754648637, "grad_norm": 0.2936011552810669, "learning_rate": 8.514946071951342e-05, "loss": 1.6725, "step": 1105 }, { "epoch": 0.26708524510987686, "grad_norm": 0.28947994112968445, "learning_rate": 8.51219205894684e-05, "loss": 1.7471, "step": 1106 }, { "epoch": 0.26732673267326734, "grad_norm": 0.3159331679344177, "learning_rate": 8.50943594093364e-05, "loss": 1.8664, "step": 1107 }, { "epoch": 0.2675682202366578, "grad_norm": 0.2945443093776703, "learning_rate": 8.5066777195636e-05, "loss": 1.7574, "step": 1108 }, { "epoch": 0.2678097078000483, "grad_norm": 0.29109689593315125, "learning_rate": 8.503917396489847e-05, "loss": 1.6845, "step": 1109 }, { "epoch": 0.2680511953634388, "grad_norm": 0.2907101809978485, "learning_rate": 8.501154973366754e-05, "loss": 1.7214, "step": 1110 }, { "epoch": 0.2682926829268293, "grad_norm": 0.27306222915649414, "learning_rate": 8.498390451849967e-05, "loss": 1.6801, "step": 1111 }, { "epoch": 0.26853417049021977, "grad_norm": 0.28223463892936707, "learning_rate": 8.495623833596382e-05, "loss": 1.7371, "step": 1112 }, { "epoch": 0.26877565805361026, "grad_norm": 0.29571643471717834, "learning_rate": 8.492855120264151e-05, "loss": 1.6157, "step": 1113 }, { "epoch": 0.26901714561700074, "grad_norm": 0.27164462208747864, "learning_rate": 8.490084313512685e-05, "loss": 1.6855, "step": 1114 }, { "epoch": 0.2692586331803912, "grad_norm": 0.28996163606643677, "learning_rate": 8.48731141500265e-05, "loss": 1.7985, "step": 1115 }, { "epoch": 0.2695001207437817, "grad_norm": 0.2976994812488556, "learning_rate": 8.484536426395962e-05, "loss": 1.7614, "step": 1116 }, { "epoch": 0.2697416083071722, "grad_norm": 0.2953455448150635, "learning_rate": 8.481759349355791e-05, "loss": 1.8129, "step": 1117 }, { "epoch": 0.2699830958705627, "grad_norm": 0.30153965950012207, "learning_rate": 8.478980185546562e-05, "loss": 1.9054, "step": 1118 }, { "epoch": 0.27022458343395317, "grad_norm": 0.29038190841674805, "learning_rate": 8.476198936633946e-05, "loss": 1.7608, "step": 1119 }, { "epoch": 0.27046607099734365, "grad_norm": 0.295154869556427, "learning_rate": 8.473415604284869e-05, "loss": 1.8574, "step": 1120 }, { "epoch": 0.27070755856073414, "grad_norm": 0.30491289496421814, "learning_rate": 8.470630190167499e-05, "loss": 1.5952, "step": 1121 }, { "epoch": 0.2709490461241246, "grad_norm": 0.28542447090148926, "learning_rate": 8.467842695951256e-05, "loss": 1.6682, "step": 1122 }, { "epoch": 0.2711905336875151, "grad_norm": 0.2852567434310913, "learning_rate": 8.465053123306806e-05, "loss": 1.7543, "step": 1123 }, { "epoch": 0.2714320212509056, "grad_norm": 0.2947034537792206, "learning_rate": 8.462261473906063e-05, "loss": 1.778, "step": 1124 }, { "epoch": 0.2716735088142961, "grad_norm": 0.3023829758167267, "learning_rate": 8.45946774942218e-05, "loss": 1.6253, "step": 1125 }, { "epoch": 0.27191499637768657, "grad_norm": 0.2906095087528229, "learning_rate": 8.456671951529559e-05, "loss": 1.6695, "step": 1126 }, { "epoch": 0.27215648394107705, "grad_norm": 0.28810620307922363, "learning_rate": 8.453874081903841e-05, "loss": 1.8087, "step": 1127 }, { "epoch": 0.27239797150446754, "grad_norm": 0.2939632833003998, "learning_rate": 8.451074142221913e-05, "loss": 1.8691, "step": 1128 }, { "epoch": 0.272639459067858, "grad_norm": 0.47220274806022644, "learning_rate": 8.448272134161896e-05, "loss": 1.8504, "step": 1129 }, { "epoch": 0.2728809466312485, "grad_norm": 0.26881143450737, "learning_rate": 8.445468059403159e-05, "loss": 1.6685, "step": 1130 }, { "epoch": 0.273122434194639, "grad_norm": 0.29267653822898865, "learning_rate": 8.442661919626305e-05, "loss": 1.814, "step": 1131 }, { "epoch": 0.2733639217580295, "grad_norm": 0.29204314947128296, "learning_rate": 8.439853716513171e-05, "loss": 1.7541, "step": 1132 }, { "epoch": 0.27360540932141997, "grad_norm": 0.28699198365211487, "learning_rate": 8.437043451746837e-05, "loss": 1.7636, "step": 1133 }, { "epoch": 0.27384689688481045, "grad_norm": 0.2884604334831238, "learning_rate": 8.434231127011617e-05, "loss": 1.7746, "step": 1134 }, { "epoch": 0.27408838444820094, "grad_norm": 0.275660902261734, "learning_rate": 8.431416743993059e-05, "loss": 1.5836, "step": 1135 }, { "epoch": 0.2743298720115914, "grad_norm": 0.2811957597732544, "learning_rate": 8.428600304377942e-05, "loss": 1.7617, "step": 1136 }, { "epoch": 0.2745713595749819, "grad_norm": 0.28643205761909485, "learning_rate": 8.425781809854285e-05, "loss": 1.6944, "step": 1137 }, { "epoch": 0.2748128471383724, "grad_norm": 0.286495566368103, "learning_rate": 8.42296126211133e-05, "loss": 1.7086, "step": 1138 }, { "epoch": 0.2750543347017629, "grad_norm": 0.28585320711135864, "learning_rate": 8.420138662839552e-05, "loss": 1.6291, "step": 1139 }, { "epoch": 0.27529582226515337, "grad_norm": 0.2841216027736664, "learning_rate": 8.417314013730662e-05, "loss": 1.6028, "step": 1140 }, { "epoch": 0.27553730982854385, "grad_norm": 0.2992970645427704, "learning_rate": 8.414487316477589e-05, "loss": 1.6879, "step": 1141 }, { "epoch": 0.27577879739193434, "grad_norm": 0.2880629897117615, "learning_rate": 8.411658572774498e-05, "loss": 1.799, "step": 1142 }, { "epoch": 0.2760202849553248, "grad_norm": 0.286931574344635, "learning_rate": 8.408827784316777e-05, "loss": 1.8065, "step": 1143 }, { "epoch": 0.2762617725187153, "grad_norm": 0.29781198501586914, "learning_rate": 8.405994952801042e-05, "loss": 1.7046, "step": 1144 }, { "epoch": 0.2765032600821058, "grad_norm": 0.2892191708087921, "learning_rate": 8.403160079925127e-05, "loss": 1.8779, "step": 1145 }, { "epoch": 0.2767447476454963, "grad_norm": 0.2801336646080017, "learning_rate": 8.400323167388098e-05, "loss": 1.6652, "step": 1146 }, { "epoch": 0.27698623520888677, "grad_norm": 0.28385961055755615, "learning_rate": 8.397484216890237e-05, "loss": 1.5174, "step": 1147 }, { "epoch": 0.27722772277227725, "grad_norm": 0.2902655005455017, "learning_rate": 8.39464323013305e-05, "loss": 1.7628, "step": 1148 }, { "epoch": 0.27746921033566774, "grad_norm": 0.31906628608703613, "learning_rate": 8.391800208819267e-05, "loss": 1.9554, "step": 1149 }, { "epoch": 0.2777106978990582, "grad_norm": 0.2920599579811096, "learning_rate": 8.38895515465283e-05, "loss": 1.8078, "step": 1150 }, { "epoch": 0.2779521854624487, "grad_norm": 0.2755699157714844, "learning_rate": 8.386108069338903e-05, "loss": 1.6479, "step": 1151 }, { "epoch": 0.2781936730258392, "grad_norm": 0.31938937306404114, "learning_rate": 8.383258954583868e-05, "loss": 1.7485, "step": 1152 }, { "epoch": 0.2784351605892297, "grad_norm": 0.3132666349411011, "learning_rate": 8.380407812095325e-05, "loss": 1.7489, "step": 1153 }, { "epoch": 0.27867664815262017, "grad_norm": 0.2854362428188324, "learning_rate": 8.377554643582084e-05, "loss": 1.8807, "step": 1154 }, { "epoch": 0.27891813571601065, "grad_norm": 0.3173444867134094, "learning_rate": 8.374699450754174e-05, "loss": 1.8614, "step": 1155 }, { "epoch": 0.27915962327940114, "grad_norm": 0.27949345111846924, "learning_rate": 8.371842235322836e-05, "loss": 1.731, "step": 1156 }, { "epoch": 0.2794011108427916, "grad_norm": 0.27879253029823303, "learning_rate": 8.368982999000521e-05, "loss": 1.6843, "step": 1157 }, { "epoch": 0.2796425984061821, "grad_norm": 0.2896178364753723, "learning_rate": 8.366121743500895e-05, "loss": 1.589, "step": 1158 }, { "epoch": 0.2798840859695726, "grad_norm": 0.32527682185173035, "learning_rate": 8.363258470538832e-05, "loss": 1.8235, "step": 1159 }, { "epoch": 0.2801255735329631, "grad_norm": 0.3003365695476532, "learning_rate": 8.360393181830414e-05, "loss": 1.9417, "step": 1160 }, { "epoch": 0.28036706109635356, "grad_norm": 0.30865755677223206, "learning_rate": 8.357525879092933e-05, "loss": 1.759, "step": 1161 }, { "epoch": 0.28060854865974405, "grad_norm": 0.29030731320381165, "learning_rate": 8.35465656404489e-05, "loss": 1.7726, "step": 1162 }, { "epoch": 0.28085003622313454, "grad_norm": 0.2997666895389557, "learning_rate": 8.351785238405985e-05, "loss": 1.6101, "step": 1163 }, { "epoch": 0.281091523786525, "grad_norm": 0.2943105399608612, "learning_rate": 8.348911903897132e-05, "loss": 1.7153, "step": 1164 }, { "epoch": 0.28133301134991545, "grad_norm": 0.2931126058101654, "learning_rate": 8.346036562240444e-05, "loss": 1.6673, "step": 1165 }, { "epoch": 0.28157449891330594, "grad_norm": 0.31237348914146423, "learning_rate": 8.343159215159235e-05, "loss": 1.965, "step": 1166 }, { "epoch": 0.2818159864766964, "grad_norm": 0.30502235889434814, "learning_rate": 8.340279864378026e-05, "loss": 1.6968, "step": 1167 }, { "epoch": 0.2820574740400869, "grad_norm": 0.28124526143074036, "learning_rate": 8.337398511622536e-05, "loss": 1.7541, "step": 1168 }, { "epoch": 0.2822989616034774, "grad_norm": 0.2825506031513214, "learning_rate": 8.334515158619685e-05, "loss": 1.625, "step": 1169 }, { "epoch": 0.2825404491668679, "grad_norm": 0.29996126890182495, "learning_rate": 8.33162980709759e-05, "loss": 1.8162, "step": 1170 }, { "epoch": 0.28278193673025837, "grad_norm": 0.28735971450805664, "learning_rate": 8.328742458785568e-05, "loss": 1.6568, "step": 1171 }, { "epoch": 0.28302342429364885, "grad_norm": 0.30295073986053467, "learning_rate": 8.325853115414132e-05, "loss": 1.768, "step": 1172 }, { "epoch": 0.28326491185703934, "grad_norm": 0.29709091782569885, "learning_rate": 8.322961778714989e-05, "loss": 1.7637, "step": 1173 }, { "epoch": 0.2835063994204298, "grad_norm": 0.29945021867752075, "learning_rate": 8.320068450421044e-05, "loss": 1.7821, "step": 1174 }, { "epoch": 0.2837478869838203, "grad_norm": 0.29444846510887146, "learning_rate": 8.317173132266392e-05, "loss": 1.8555, "step": 1175 }, { "epoch": 0.2839893745472108, "grad_norm": 0.27773210406303406, "learning_rate": 8.314275825986325e-05, "loss": 1.6424, "step": 1176 }, { "epoch": 0.2842308621106013, "grad_norm": 0.31503719091415405, "learning_rate": 8.311376533317321e-05, "loss": 1.985, "step": 1177 }, { "epoch": 0.28447234967399176, "grad_norm": 0.3152500092983246, "learning_rate": 8.308475255997055e-05, "loss": 1.8576, "step": 1178 }, { "epoch": 0.28471383723738225, "grad_norm": 0.28431928157806396, "learning_rate": 8.305571995764385e-05, "loss": 1.6745, "step": 1179 }, { "epoch": 0.28495532480077274, "grad_norm": 0.29776814579963684, "learning_rate": 8.302666754359363e-05, "loss": 1.6738, "step": 1180 }, { "epoch": 0.2851968123641632, "grad_norm": 0.29094064235687256, "learning_rate": 8.299759533523222e-05, "loss": 1.7761, "step": 1181 }, { "epoch": 0.2854382999275537, "grad_norm": 0.28861159086227417, "learning_rate": 8.296850334998391e-05, "loss": 1.7584, "step": 1182 }, { "epoch": 0.2856797874909442, "grad_norm": 0.32065683603286743, "learning_rate": 8.293939160528474e-05, "loss": 1.9342, "step": 1183 }, { "epoch": 0.2859212750543347, "grad_norm": 0.3114961087703705, "learning_rate": 8.291026011858266e-05, "loss": 1.9161, "step": 1184 }, { "epoch": 0.28616276261772516, "grad_norm": 0.2836478650569916, "learning_rate": 8.288110890733741e-05, "loss": 1.6908, "step": 1185 }, { "epoch": 0.28640425018111565, "grad_norm": 0.289570689201355, "learning_rate": 8.28519379890206e-05, "loss": 1.7737, "step": 1186 }, { "epoch": 0.28664573774450613, "grad_norm": 0.2810715138912201, "learning_rate": 8.28227473811156e-05, "loss": 1.7923, "step": 1187 }, { "epoch": 0.2868872253078966, "grad_norm": 0.2763075828552246, "learning_rate": 8.279353710111761e-05, "loss": 1.6779, "step": 1188 }, { "epoch": 0.2871287128712871, "grad_norm": 0.28934016823768616, "learning_rate": 8.276430716653363e-05, "loss": 1.7764, "step": 1189 }, { "epoch": 0.2873702004346776, "grad_norm": 0.3152405619621277, "learning_rate": 8.273505759488241e-05, "loss": 1.9608, "step": 1190 }, { "epoch": 0.2876116879980681, "grad_norm": 0.3540990948677063, "learning_rate": 8.270578840369449e-05, "loss": 1.817, "step": 1191 }, { "epoch": 0.28785317556145856, "grad_norm": 0.27829819917678833, "learning_rate": 8.267649961051219e-05, "loss": 1.632, "step": 1192 }, { "epoch": 0.28809466312484905, "grad_norm": 0.3032492995262146, "learning_rate": 8.264719123288949e-05, "loss": 1.853, "step": 1193 }, { "epoch": 0.28833615068823953, "grad_norm": 0.2742372453212738, "learning_rate": 8.26178632883922e-05, "loss": 1.6385, "step": 1194 }, { "epoch": 0.28857763825163, "grad_norm": 0.27677586674690247, "learning_rate": 8.258851579459783e-05, "loss": 1.6423, "step": 1195 }, { "epoch": 0.2888191258150205, "grad_norm": 0.2932434380054474, "learning_rate": 8.25591487690956e-05, "loss": 1.8024, "step": 1196 }, { "epoch": 0.289060613378411, "grad_norm": 0.2936798632144928, "learning_rate": 8.252976222948647e-05, "loss": 1.8136, "step": 1197 }, { "epoch": 0.2893021009418015, "grad_norm": 0.3236442804336548, "learning_rate": 8.250035619338302e-05, "loss": 1.7561, "step": 1198 }, { "epoch": 0.28954358850519196, "grad_norm": 0.3159981966018677, "learning_rate": 8.247093067840956e-05, "loss": 1.9135, "step": 1199 }, { "epoch": 0.28978507606858245, "grad_norm": 0.2588784098625183, "learning_rate": 8.244148570220211e-05, "loss": 1.5257, "step": 1200 }, { "epoch": 0.29002656363197293, "grad_norm": 0.29745063185691833, "learning_rate": 8.241202128240829e-05, "loss": 1.833, "step": 1201 }, { "epoch": 0.2902680511953634, "grad_norm": 0.28102126717567444, "learning_rate": 8.23825374366874e-05, "loss": 1.673, "step": 1202 }, { "epoch": 0.2905095387587539, "grad_norm": 0.27232810854911804, "learning_rate": 8.23530341827104e-05, "loss": 1.6953, "step": 1203 }, { "epoch": 0.2907510263221444, "grad_norm": 0.28626495599746704, "learning_rate": 8.232351153815988e-05, "loss": 1.695, "step": 1204 }, { "epoch": 0.2909925138855349, "grad_norm": 0.30019521713256836, "learning_rate": 8.229396952073001e-05, "loss": 1.6836, "step": 1205 }, { "epoch": 0.29123400144892536, "grad_norm": 0.2773890197277069, "learning_rate": 8.226440814812662e-05, "loss": 1.7156, "step": 1206 }, { "epoch": 0.29147548901231585, "grad_norm": 0.3097532093524933, "learning_rate": 8.223482743806709e-05, "loss": 1.8985, "step": 1207 }, { "epoch": 0.29171697657570633, "grad_norm": 0.3096452057361603, "learning_rate": 8.220522740828046e-05, "loss": 1.9616, "step": 1208 }, { "epoch": 0.2919584641390968, "grad_norm": 0.2948162853717804, "learning_rate": 8.217560807650728e-05, "loss": 1.7219, "step": 1209 }, { "epoch": 0.2921999517024873, "grad_norm": 0.2883644700050354, "learning_rate": 8.21459694604997e-05, "loss": 1.7977, "step": 1210 }, { "epoch": 0.2924414392658778, "grad_norm": 0.3046281635761261, "learning_rate": 8.211631157802144e-05, "loss": 2.0905, "step": 1211 }, { "epoch": 0.2926829268292683, "grad_norm": 0.31257206201553345, "learning_rate": 8.208663444684776e-05, "loss": 1.9364, "step": 1212 }, { "epoch": 0.29292441439265876, "grad_norm": 0.29145514965057373, "learning_rate": 8.20569380847654e-05, "loss": 1.7626, "step": 1213 }, { "epoch": 0.29316590195604925, "grad_norm": 0.2752838730812073, "learning_rate": 8.202722250957273e-05, "loss": 1.5806, "step": 1214 }, { "epoch": 0.29340738951943973, "grad_norm": 0.2736091911792755, "learning_rate": 8.199748773907956e-05, "loss": 1.6547, "step": 1215 }, { "epoch": 0.2936488770828302, "grad_norm": 0.3060460090637207, "learning_rate": 8.196773379110722e-05, "loss": 1.8482, "step": 1216 }, { "epoch": 0.2938903646462207, "grad_norm": 0.2833315134048462, "learning_rate": 8.193796068348853e-05, "loss": 1.6437, "step": 1217 }, { "epoch": 0.2941318522096112, "grad_norm": 0.2926206588745117, "learning_rate": 8.190816843406783e-05, "loss": 1.8218, "step": 1218 }, { "epoch": 0.2943733397730017, "grad_norm": 0.29806745052337646, "learning_rate": 8.187835706070089e-05, "loss": 1.8769, "step": 1219 }, { "epoch": 0.29461482733639216, "grad_norm": 0.2936302125453949, "learning_rate": 8.184852658125494e-05, "loss": 1.7622, "step": 1220 }, { "epoch": 0.29485631489978265, "grad_norm": 0.30911576747894287, "learning_rate": 8.18186770136087e-05, "loss": 1.9, "step": 1221 }, { "epoch": 0.29509780246317313, "grad_norm": 0.2942495048046112, "learning_rate": 8.178880837565228e-05, "loss": 1.7852, "step": 1222 }, { "epoch": 0.2953392900265636, "grad_norm": 0.2813624441623688, "learning_rate": 8.17589206852873e-05, "loss": 1.738, "step": 1223 }, { "epoch": 0.2955807775899541, "grad_norm": 0.2759395241737366, "learning_rate": 8.172901396042669e-05, "loss": 1.6633, "step": 1224 }, { "epoch": 0.2958222651533446, "grad_norm": 0.28530681133270264, "learning_rate": 8.169908821899489e-05, "loss": 1.6496, "step": 1225 }, { "epoch": 0.2960637527167351, "grad_norm": 0.28899961709976196, "learning_rate": 8.166914347892764e-05, "loss": 1.7069, "step": 1226 }, { "epoch": 0.29630524028012556, "grad_norm": 0.3118133842945099, "learning_rate": 8.163917975817217e-05, "loss": 1.7776, "step": 1227 }, { "epoch": 0.29654672784351604, "grad_norm": 0.288650244474411, "learning_rate": 8.1609197074687e-05, "loss": 1.6574, "step": 1228 }, { "epoch": 0.29678821540690653, "grad_norm": 0.28079408407211304, "learning_rate": 8.157919544644206e-05, "loss": 1.6564, "step": 1229 }, { "epoch": 0.297029702970297, "grad_norm": 0.28103119134902954, "learning_rate": 8.154917489141865e-05, "loss": 1.6368, "step": 1230 }, { "epoch": 0.2972711905336875, "grad_norm": 0.2799832820892334, "learning_rate": 8.151913542760934e-05, "loss": 1.6049, "step": 1231 }, { "epoch": 0.297512678097078, "grad_norm": 0.2864444851875305, "learning_rate": 8.148907707301811e-05, "loss": 1.6147, "step": 1232 }, { "epoch": 0.2977541656604685, "grad_norm": 0.28280696272850037, "learning_rate": 8.145899984566024e-05, "loss": 1.8078, "step": 1233 }, { "epoch": 0.29799565322385896, "grad_norm": 0.3180987238883972, "learning_rate": 8.142890376356229e-05, "loss": 1.8723, "step": 1234 }, { "epoch": 0.29823714078724944, "grad_norm": 0.28342029452323914, "learning_rate": 8.139878884476216e-05, "loss": 1.6105, "step": 1235 }, { "epoch": 0.29847862835063993, "grad_norm": 0.2965908944606781, "learning_rate": 8.136865510730903e-05, "loss": 1.7005, "step": 1236 }, { "epoch": 0.2987201159140304, "grad_norm": 0.28203895688056946, "learning_rate": 8.133850256926335e-05, "loss": 1.7559, "step": 1237 }, { "epoch": 0.2989616034774209, "grad_norm": 0.29620274901390076, "learning_rate": 8.130833124869681e-05, "loss": 1.8065, "step": 1238 }, { "epoch": 0.2992030910408114, "grad_norm": 0.2921755909919739, "learning_rate": 8.127814116369242e-05, "loss": 1.7224, "step": 1239 }, { "epoch": 0.29944457860420187, "grad_norm": 0.3076702356338501, "learning_rate": 8.12479323323444e-05, "loss": 1.9597, "step": 1240 }, { "epoch": 0.29968606616759236, "grad_norm": 0.3073546290397644, "learning_rate": 8.121770477275821e-05, "loss": 1.7661, "step": 1241 }, { "epoch": 0.29992755373098284, "grad_norm": 0.27899089455604553, "learning_rate": 8.118745850305054e-05, "loss": 1.6729, "step": 1242 }, { "epoch": 0.30016904129437333, "grad_norm": 0.28602051734924316, "learning_rate": 8.115719354134926e-05, "loss": 1.6565, "step": 1243 }, { "epoch": 0.3004105288577638, "grad_norm": 0.3022647798061371, "learning_rate": 8.11269099057935e-05, "loss": 1.7554, "step": 1244 }, { "epoch": 0.3006520164211543, "grad_norm": 0.2899514436721802, "learning_rate": 8.109660761453355e-05, "loss": 1.7334, "step": 1245 }, { "epoch": 0.3008935039845448, "grad_norm": 0.2962121367454529, "learning_rate": 8.106628668573087e-05, "loss": 1.769, "step": 1246 }, { "epoch": 0.30113499154793527, "grad_norm": 0.29414042830467224, "learning_rate": 8.103594713755813e-05, "loss": 1.8094, "step": 1247 }, { "epoch": 0.30137647911132576, "grad_norm": 0.33522823452949524, "learning_rate": 8.100558898819912e-05, "loss": 1.9885, "step": 1248 }, { "epoch": 0.30161796667471624, "grad_norm": 0.32740721106529236, "learning_rate": 8.097521225584876e-05, "loss": 1.8678, "step": 1249 }, { "epoch": 0.30185945423810673, "grad_norm": 0.2977084815502167, "learning_rate": 8.094481695871319e-05, "loss": 1.6237, "step": 1250 }, { "epoch": 0.3021009418014972, "grad_norm": 0.29849734902381897, "learning_rate": 8.091440311500963e-05, "loss": 1.8449, "step": 1251 }, { "epoch": 0.3023424293648877, "grad_norm": 0.2919251620769501, "learning_rate": 8.088397074296636e-05, "loss": 1.7264, "step": 1252 }, { "epoch": 0.3025839169282782, "grad_norm": 0.28110143542289734, "learning_rate": 8.085351986082287e-05, "loss": 1.6551, "step": 1253 }, { "epoch": 0.30282540449166867, "grad_norm": 0.28460079431533813, "learning_rate": 8.082305048682966e-05, "loss": 1.7723, "step": 1254 }, { "epoch": 0.30306689205505916, "grad_norm": 0.2959478795528412, "learning_rate": 8.079256263924836e-05, "loss": 1.7073, "step": 1255 }, { "epoch": 0.30330837961844964, "grad_norm": 0.30130937695503235, "learning_rate": 8.076205633635165e-05, "loss": 1.8099, "step": 1256 }, { "epoch": 0.3035498671818401, "grad_norm": 0.2832283675670624, "learning_rate": 8.073153159642328e-05, "loss": 1.7325, "step": 1257 }, { "epoch": 0.3037913547452306, "grad_norm": 0.27280157804489136, "learning_rate": 8.070098843775804e-05, "loss": 1.7815, "step": 1258 }, { "epoch": 0.3040328423086211, "grad_norm": 0.3622336983680725, "learning_rate": 8.067042687866178e-05, "loss": 2.2295, "step": 1259 }, { "epoch": 0.3042743298720116, "grad_norm": 0.29381340742111206, "learning_rate": 8.063984693745136e-05, "loss": 1.8125, "step": 1260 }, { "epoch": 0.30451581743540207, "grad_norm": 0.28132760524749756, "learning_rate": 8.060924863245466e-05, "loss": 1.4983, "step": 1261 }, { "epoch": 0.30475730499879256, "grad_norm": 0.30654898285865784, "learning_rate": 8.057863198201056e-05, "loss": 1.7907, "step": 1262 }, { "epoch": 0.30499879256218304, "grad_norm": 0.2809448838233948, "learning_rate": 8.054799700446896e-05, "loss": 1.7044, "step": 1263 }, { "epoch": 0.3052402801255735, "grad_norm": 0.2905556559562683, "learning_rate": 8.051734371819072e-05, "loss": 1.7571, "step": 1264 }, { "epoch": 0.305481767688964, "grad_norm": 0.30511733889579773, "learning_rate": 8.048667214154769e-05, "loss": 1.8195, "step": 1265 }, { "epoch": 0.3057232552523545, "grad_norm": 0.2864232659339905, "learning_rate": 8.045598229292265e-05, "loss": 1.5109, "step": 1266 }, { "epoch": 0.305964742815745, "grad_norm": 0.2991337776184082, "learning_rate": 8.042527419070938e-05, "loss": 1.568, "step": 1267 }, { "epoch": 0.30620623037913547, "grad_norm": 0.327675998210907, "learning_rate": 8.039454785331256e-05, "loss": 2.1194, "step": 1268 }, { "epoch": 0.30644771794252595, "grad_norm": 0.28206297755241394, "learning_rate": 8.036380329914781e-05, "loss": 1.6651, "step": 1269 }, { "epoch": 0.30668920550591644, "grad_norm": 0.286390483379364, "learning_rate": 8.03330405466417e-05, "loss": 1.695, "step": 1270 }, { "epoch": 0.3069306930693069, "grad_norm": 0.3134172260761261, "learning_rate": 8.030225961423165e-05, "loss": 1.8522, "step": 1271 }, { "epoch": 0.3071721806326974, "grad_norm": 0.30529677867889404, "learning_rate": 8.0271460520366e-05, "loss": 1.7846, "step": 1272 }, { "epoch": 0.3074136681960879, "grad_norm": 0.2850496768951416, "learning_rate": 8.0240643283504e-05, "loss": 1.8515, "step": 1273 }, { "epoch": 0.3076551557594784, "grad_norm": 0.31650790572166443, "learning_rate": 8.020980792211576e-05, "loss": 1.7635, "step": 1274 }, { "epoch": 0.30789664332286887, "grad_norm": 0.3554689288139343, "learning_rate": 8.017895445468222e-05, "loss": 1.6921, "step": 1275 }, { "epoch": 0.30813813088625935, "grad_norm": 0.28742724657058716, "learning_rate": 8.014808289969523e-05, "loss": 1.8524, "step": 1276 }, { "epoch": 0.30837961844964984, "grad_norm": 0.27519968152046204, "learning_rate": 8.01171932756574e-05, "loss": 1.7443, "step": 1277 }, { "epoch": 0.3086211060130403, "grad_norm": 0.288105845451355, "learning_rate": 8.008628560108227e-05, "loss": 1.6894, "step": 1278 }, { "epoch": 0.3088625935764308, "grad_norm": 0.29678773880004883, "learning_rate": 8.005535989449411e-05, "loss": 1.6458, "step": 1279 }, { "epoch": 0.3091040811398213, "grad_norm": 0.2868390679359436, "learning_rate": 8.002441617442807e-05, "loss": 1.5688, "step": 1280 }, { "epoch": 0.3093455687032118, "grad_norm": 0.302274614572525, "learning_rate": 7.999345445943003e-05, "loss": 1.6713, "step": 1281 }, { "epoch": 0.30958705626660227, "grad_norm": 0.30040210485458374, "learning_rate": 7.99624747680567e-05, "loss": 1.7323, "step": 1282 }, { "epoch": 0.30982854382999275, "grad_norm": 0.2940504550933838, "learning_rate": 7.993147711887554e-05, "loss": 1.776, "step": 1283 }, { "epoch": 0.31007003139338324, "grad_norm": 0.31216347217559814, "learning_rate": 7.99004615304648e-05, "loss": 1.6848, "step": 1284 }, { "epoch": 0.3103115189567737, "grad_norm": 0.3009435534477234, "learning_rate": 7.986942802141346e-05, "loss": 1.8094, "step": 1285 }, { "epoch": 0.3105530065201642, "grad_norm": 0.31435760855674744, "learning_rate": 7.983837661032123e-05, "loss": 1.6859, "step": 1286 }, { "epoch": 0.3107944940835547, "grad_norm": 0.2916475236415863, "learning_rate": 7.980730731579856e-05, "loss": 1.7225, "step": 1287 }, { "epoch": 0.3110359816469452, "grad_norm": 0.29502925276756287, "learning_rate": 7.977622015646667e-05, "loss": 1.7175, "step": 1288 }, { "epoch": 0.31127746921033567, "grad_norm": 0.3051729202270508, "learning_rate": 7.974511515095738e-05, "loss": 1.6778, "step": 1289 }, { "epoch": 0.31151895677372615, "grad_norm": 0.29863834381103516, "learning_rate": 7.971399231791328e-05, "loss": 1.6819, "step": 1290 }, { "epoch": 0.31176044433711664, "grad_norm": 0.27716687321662903, "learning_rate": 7.968285167598766e-05, "loss": 1.6873, "step": 1291 }, { "epoch": 0.3120019319005071, "grad_norm": 0.30108094215393066, "learning_rate": 7.965169324384445e-05, "loss": 1.6907, "step": 1292 }, { "epoch": 0.3122434194638976, "grad_norm": 0.3081267774105072, "learning_rate": 7.96205170401582e-05, "loss": 1.6812, "step": 1293 }, { "epoch": 0.3124849070272881, "grad_norm": 0.2833951711654663, "learning_rate": 7.958932308361422e-05, "loss": 1.6024, "step": 1294 }, { "epoch": 0.3127263945906786, "grad_norm": 0.3181832730770111, "learning_rate": 7.955811139290837e-05, "loss": 1.896, "step": 1295 }, { "epoch": 0.31296788215406907, "grad_norm": 0.29816505312919617, "learning_rate": 7.952688198674714e-05, "loss": 1.8003, "step": 1296 }, { "epoch": 0.31320936971745955, "grad_norm": 0.33011528849601746, "learning_rate": 7.949563488384772e-05, "loss": 1.9278, "step": 1297 }, { "epoch": 0.31345085728085004, "grad_norm": 0.30737966299057007, "learning_rate": 7.946437010293781e-05, "loss": 1.8645, "step": 1298 }, { "epoch": 0.3136923448442405, "grad_norm": 0.26630595326423645, "learning_rate": 7.943308766275577e-05, "loss": 1.494, "step": 1299 }, { "epoch": 0.313933832407631, "grad_norm": 0.298896849155426, "learning_rate": 7.94017875820505e-05, "loss": 1.8095, "step": 1300 }, { "epoch": 0.3141753199710215, "grad_norm": 0.2950008809566498, "learning_rate": 7.93704698795815e-05, "loss": 1.8302, "step": 1301 }, { "epoch": 0.314416807534412, "grad_norm": 0.29186734557151794, "learning_rate": 7.933913457411884e-05, "loss": 1.7945, "step": 1302 }, { "epoch": 0.31465829509780247, "grad_norm": 0.29058417677879333, "learning_rate": 7.93077816844431e-05, "loss": 1.6974, "step": 1303 }, { "epoch": 0.31489978266119295, "grad_norm": 0.3064737021923065, "learning_rate": 7.927641122934543e-05, "loss": 1.8599, "step": 1304 }, { "epoch": 0.31514127022458344, "grad_norm": 0.3019355535507202, "learning_rate": 7.924502322762752e-05, "loss": 1.8489, "step": 1305 }, { "epoch": 0.3153827577879739, "grad_norm": 0.27845993638038635, "learning_rate": 7.921361769810153e-05, "loss": 1.6851, "step": 1306 }, { "epoch": 0.3156242453513644, "grad_norm": 0.28601062297821045, "learning_rate": 7.918219465959018e-05, "loss": 1.7533, "step": 1307 }, { "epoch": 0.3158657329147549, "grad_norm": 0.28513553738594055, "learning_rate": 7.915075413092664e-05, "loss": 1.6407, "step": 1308 }, { "epoch": 0.3161072204781454, "grad_norm": 0.29098600149154663, "learning_rate": 7.91192961309546e-05, "loss": 1.6635, "step": 1309 }, { "epoch": 0.31634870804153586, "grad_norm": 0.3085253834724426, "learning_rate": 7.908782067852816e-05, "loss": 1.7706, "step": 1310 }, { "epoch": 0.31659019560492635, "grad_norm": 0.29994043707847595, "learning_rate": 7.905632779251195e-05, "loss": 1.7929, "step": 1311 }, { "epoch": 0.31683168316831684, "grad_norm": 0.3095788061618805, "learning_rate": 7.902481749178101e-05, "loss": 1.8123, "step": 1312 }, { "epoch": 0.3170731707317073, "grad_norm": 0.30506932735443115, "learning_rate": 7.899328979522085e-05, "loss": 1.6084, "step": 1313 }, { "epoch": 0.3173146582950978, "grad_norm": 0.28293487429618835, "learning_rate": 7.896174472172735e-05, "loss": 1.7082, "step": 1314 }, { "epoch": 0.3175561458584883, "grad_norm": 0.3032687306404114, "learning_rate": 7.893018229020686e-05, "loss": 1.7877, "step": 1315 }, { "epoch": 0.3177976334218788, "grad_norm": 0.3086721897125244, "learning_rate": 7.889860251957609e-05, "loss": 1.8977, "step": 1316 }, { "epoch": 0.31803912098526926, "grad_norm": 0.3076464831829071, "learning_rate": 7.886700542876218e-05, "loss": 1.783, "step": 1317 }, { "epoch": 0.31828060854865975, "grad_norm": 0.30208855867385864, "learning_rate": 7.883539103670264e-05, "loss": 1.7573, "step": 1318 }, { "epoch": 0.31852209611205023, "grad_norm": 0.3009052574634552, "learning_rate": 7.880375936234534e-05, "loss": 1.7234, "step": 1319 }, { "epoch": 0.3187635836754407, "grad_norm": 0.3098020553588867, "learning_rate": 7.87721104246485e-05, "loss": 1.8425, "step": 1320 }, { "epoch": 0.3190050712388312, "grad_norm": 0.2842791974544525, "learning_rate": 7.874044424258069e-05, "loss": 1.7381, "step": 1321 }, { "epoch": 0.3192465588022217, "grad_norm": 0.3024597764015198, "learning_rate": 7.870876083512084e-05, "loss": 1.6135, "step": 1322 }, { "epoch": 0.3194880463656122, "grad_norm": 0.2864610254764557, "learning_rate": 7.867706022125819e-05, "loss": 1.6423, "step": 1323 }, { "epoch": 0.31972953392900266, "grad_norm": 0.3033413887023926, "learning_rate": 7.864534241999228e-05, "loss": 1.8562, "step": 1324 }, { "epoch": 0.31997102149239315, "grad_norm": 0.2795904874801636, "learning_rate": 7.861360745033297e-05, "loss": 1.6943, "step": 1325 }, { "epoch": 0.32021250905578363, "grad_norm": 0.2918721139431, "learning_rate": 7.85818553313004e-05, "loss": 1.7442, "step": 1326 }, { "epoch": 0.3204539966191741, "grad_norm": 0.2953207194805145, "learning_rate": 7.855008608192498e-05, "loss": 1.7553, "step": 1327 }, { "epoch": 0.3206954841825646, "grad_norm": 0.29291972517967224, "learning_rate": 7.85182997212474e-05, "loss": 1.738, "step": 1328 }, { "epoch": 0.3209369717459551, "grad_norm": 0.28523069620132446, "learning_rate": 7.848649626831862e-05, "loss": 1.8182, "step": 1329 }, { "epoch": 0.3211784593093456, "grad_norm": 0.2963908910751343, "learning_rate": 7.845467574219978e-05, "loss": 1.7844, "step": 1330 }, { "epoch": 0.32141994687273606, "grad_norm": 0.293582022190094, "learning_rate": 7.842283816196232e-05, "loss": 1.7315, "step": 1331 }, { "epoch": 0.32166143443612655, "grad_norm": 0.287688672542572, "learning_rate": 7.839098354668791e-05, "loss": 1.5484, "step": 1332 }, { "epoch": 0.32190292199951703, "grad_norm": 0.3031911253929138, "learning_rate": 7.835911191546836e-05, "loss": 1.8595, "step": 1333 }, { "epoch": 0.3221444095629075, "grad_norm": 0.3043416738510132, "learning_rate": 7.832722328740575e-05, "loss": 1.8035, "step": 1334 }, { "epoch": 0.322385897126298, "grad_norm": 0.284494549036026, "learning_rate": 7.82953176816123e-05, "loss": 1.6982, "step": 1335 }, { "epoch": 0.3226273846896885, "grad_norm": 0.28933319449424744, "learning_rate": 7.82633951172104e-05, "loss": 1.7081, "step": 1336 }, { "epoch": 0.322868872253079, "grad_norm": 0.2961520850658417, "learning_rate": 7.823145561333266e-05, "loss": 1.8237, "step": 1337 }, { "epoch": 0.32311035981646946, "grad_norm": 0.33161845803260803, "learning_rate": 7.819949918912178e-05, "loss": 1.9155, "step": 1338 }, { "epoch": 0.32335184737985995, "grad_norm": 0.28777819871902466, "learning_rate": 7.816752586373064e-05, "loss": 1.7327, "step": 1339 }, { "epoch": 0.32359333494325043, "grad_norm": 0.28206440806388855, "learning_rate": 7.813553565632222e-05, "loss": 1.7053, "step": 1340 }, { "epoch": 0.3238348225066409, "grad_norm": 0.2921142578125, "learning_rate": 7.810352858606966e-05, "loss": 1.7221, "step": 1341 }, { "epoch": 0.3240763100700314, "grad_norm": 0.30265289545059204, "learning_rate": 7.807150467215619e-05, "loss": 1.8833, "step": 1342 }, { "epoch": 0.3243177976334219, "grad_norm": 0.3193051218986511, "learning_rate": 7.803946393377511e-05, "loss": 1.9729, "step": 1343 }, { "epoch": 0.3245592851968124, "grad_norm": 0.2971660792827606, "learning_rate": 7.80074063901298e-05, "loss": 1.8479, "step": 1344 }, { "epoch": 0.32480077276020286, "grad_norm": 0.28120097517967224, "learning_rate": 7.797533206043377e-05, "loss": 1.6295, "step": 1345 }, { "epoch": 0.32504226032359335, "grad_norm": 0.2861529588699341, "learning_rate": 7.794324096391055e-05, "loss": 1.571, "step": 1346 }, { "epoch": 0.32528374788698383, "grad_norm": 0.28205791115760803, "learning_rate": 7.79111331197937e-05, "loss": 1.6045, "step": 1347 }, { "epoch": 0.3255252354503743, "grad_norm": 0.3102561831474304, "learning_rate": 7.787900854732686e-05, "loss": 1.9812, "step": 1348 }, { "epoch": 0.3257667230137648, "grad_norm": 0.29614007472991943, "learning_rate": 7.784686726576364e-05, "loss": 1.7906, "step": 1349 }, { "epoch": 0.3260082105771553, "grad_norm": 0.27276650071144104, "learning_rate": 7.781470929436776e-05, "loss": 1.5792, "step": 1350 }, { "epoch": 0.3262496981405458, "grad_norm": 0.30475497245788574, "learning_rate": 7.778253465241286e-05, "loss": 1.7098, "step": 1351 }, { "epoch": 0.32649118570393626, "grad_norm": 0.28294622898101807, "learning_rate": 7.775034335918256e-05, "loss": 1.711, "step": 1352 }, { "epoch": 0.32673267326732675, "grad_norm": 0.30347758531570435, "learning_rate": 7.771813543397055e-05, "loss": 1.8165, "step": 1353 }, { "epoch": 0.32697416083071723, "grad_norm": 0.2811405062675476, "learning_rate": 7.768591089608042e-05, "loss": 1.7231, "step": 1354 }, { "epoch": 0.3272156483941077, "grad_norm": 0.2834404408931732, "learning_rate": 7.765366976482568e-05, "loss": 1.6664, "step": 1355 }, { "epoch": 0.3274571359574982, "grad_norm": 0.30318307876586914, "learning_rate": 7.762141205952991e-05, "loss": 1.8148, "step": 1356 }, { "epoch": 0.3276986235208887, "grad_norm": 0.30411288142204285, "learning_rate": 7.75891377995265e-05, "loss": 1.7338, "step": 1357 }, { "epoch": 0.3279401110842792, "grad_norm": 0.28736481070518494, "learning_rate": 7.755684700415881e-05, "loss": 1.6882, "step": 1358 }, { "epoch": 0.32818159864766966, "grad_norm": 0.2920001149177551, "learning_rate": 7.752453969278014e-05, "loss": 1.7787, "step": 1359 }, { "epoch": 0.32842308621106014, "grad_norm": 0.30056503415107727, "learning_rate": 7.749221588475363e-05, "loss": 1.6943, "step": 1360 }, { "epoch": 0.32866457377445063, "grad_norm": 0.2995055913925171, "learning_rate": 7.745987559945236e-05, "loss": 1.7648, "step": 1361 }, { "epoch": 0.3289060613378411, "grad_norm": 0.29763472080230713, "learning_rate": 7.742751885625926e-05, "loss": 1.8365, "step": 1362 }, { "epoch": 0.3291475489012316, "grad_norm": 0.2970298230648041, "learning_rate": 7.739514567456712e-05, "loss": 1.8233, "step": 1363 }, { "epoch": 0.3293890364646221, "grad_norm": 0.28750601410865784, "learning_rate": 7.736275607377859e-05, "loss": 1.7402, "step": 1364 }, { "epoch": 0.3296305240280126, "grad_norm": 0.29735836386680603, "learning_rate": 7.733035007330615e-05, "loss": 1.5324, "step": 1365 }, { "epoch": 0.32987201159140306, "grad_norm": 0.2967546284198761, "learning_rate": 7.729792769257214e-05, "loss": 1.7626, "step": 1366 }, { "epoch": 0.33011349915479354, "grad_norm": 0.29205527901649475, "learning_rate": 7.726548895100867e-05, "loss": 1.7793, "step": 1367 }, { "epoch": 0.33035498671818403, "grad_norm": 0.2876708507537842, "learning_rate": 7.72330338680577e-05, "loss": 1.8089, "step": 1368 }, { "epoch": 0.3305964742815745, "grad_norm": 0.3068663477897644, "learning_rate": 7.720056246317096e-05, "loss": 1.918, "step": 1369 }, { "epoch": 0.330837961844965, "grad_norm": 0.29179996252059937, "learning_rate": 7.716807475580997e-05, "loss": 1.6605, "step": 1370 }, { "epoch": 0.3310794494083555, "grad_norm": 0.30181172490119934, "learning_rate": 7.713557076544601e-05, "loss": 1.6363, "step": 1371 }, { "epoch": 0.33132093697174597, "grad_norm": 0.29619401693344116, "learning_rate": 7.710305051156015e-05, "loss": 1.7194, "step": 1372 }, { "epoch": 0.33156242453513646, "grad_norm": 0.2949178218841553, "learning_rate": 7.707051401364318e-05, "loss": 1.803, "step": 1373 }, { "epoch": 0.33180391209852694, "grad_norm": 0.2937702536582947, "learning_rate": 7.70379612911956e-05, "loss": 1.7161, "step": 1374 }, { "epoch": 0.33204539966191743, "grad_norm": 0.30445921421051025, "learning_rate": 7.70053923637277e-05, "loss": 1.8072, "step": 1375 }, { "epoch": 0.3322868872253079, "grad_norm": 0.3168744146823883, "learning_rate": 7.697280725075944e-05, "loss": 1.7361, "step": 1376 }, { "epoch": 0.3325283747886984, "grad_norm": 0.31621459126472473, "learning_rate": 7.694020597182048e-05, "loss": 1.763, "step": 1377 }, { "epoch": 0.3327698623520889, "grad_norm": 0.2877269685268402, "learning_rate": 7.690758854645018e-05, "loss": 1.7322, "step": 1378 }, { "epoch": 0.33301134991547937, "grad_norm": 0.2892703115940094, "learning_rate": 7.687495499419757e-05, "loss": 1.6159, "step": 1379 }, { "epoch": 0.33325283747886986, "grad_norm": 0.29529622197151184, "learning_rate": 7.684230533462138e-05, "loss": 1.7719, "step": 1380 }, { "epoch": 0.33349432504226034, "grad_norm": 0.2938997745513916, "learning_rate": 7.680963958728993e-05, "loss": 1.8697, "step": 1381 }, { "epoch": 0.33373581260565083, "grad_norm": 0.2826845645904541, "learning_rate": 7.67769577717812e-05, "loss": 1.7505, "step": 1382 }, { "epoch": 0.3339773001690413, "grad_norm": 0.29490166902542114, "learning_rate": 7.674425990768288e-05, "loss": 1.8577, "step": 1383 }, { "epoch": 0.3342187877324318, "grad_norm": 0.29277679324150085, "learning_rate": 7.671154601459215e-05, "loss": 1.6305, "step": 1384 }, { "epoch": 0.3344602752958223, "grad_norm": 0.28395023941993713, "learning_rate": 7.667881611211592e-05, "loss": 1.7382, "step": 1385 }, { "epoch": 0.33470176285921277, "grad_norm": 0.30212682485580444, "learning_rate": 7.664607021987058e-05, "loss": 1.8324, "step": 1386 }, { "epoch": 0.33494325042260326, "grad_norm": 0.28818273544311523, "learning_rate": 7.661330835748219e-05, "loss": 1.8119, "step": 1387 }, { "epoch": 0.33518473798599374, "grad_norm": 0.30034732818603516, "learning_rate": 7.658053054458636e-05, "loss": 1.7632, "step": 1388 }, { "epoch": 0.3354262255493842, "grad_norm": 0.2828637659549713, "learning_rate": 7.654773680082823e-05, "loss": 1.734, "step": 1389 }, { "epoch": 0.3356677131127747, "grad_norm": 0.2823232114315033, "learning_rate": 7.651492714586253e-05, "loss": 1.7447, "step": 1390 }, { "epoch": 0.3359092006761652, "grad_norm": 0.3083249032497406, "learning_rate": 7.64821015993535e-05, "loss": 1.855, "step": 1391 }, { "epoch": 0.3361506882395557, "grad_norm": 0.30150431394577026, "learning_rate": 7.644926018097491e-05, "loss": 1.8758, "step": 1392 }, { "epoch": 0.33639217580294617, "grad_norm": 0.3692246079444885, "learning_rate": 7.641640291041004e-05, "loss": 1.7376, "step": 1393 }, { "epoch": 0.33663366336633666, "grad_norm": 0.28748929500579834, "learning_rate": 7.638352980735167e-05, "loss": 1.6032, "step": 1394 }, { "epoch": 0.33687515092972714, "grad_norm": 0.30165570974349976, "learning_rate": 7.635064089150209e-05, "loss": 1.8655, "step": 1395 }, { "epoch": 0.3371166384931176, "grad_norm": 0.280122309923172, "learning_rate": 7.631773618257301e-05, "loss": 1.5955, "step": 1396 }, { "epoch": 0.3373581260565081, "grad_norm": 0.30419206619262695, "learning_rate": 7.62848157002857e-05, "loss": 1.8109, "step": 1397 }, { "epoch": 0.3375996136198986, "grad_norm": 0.3030405342578888, "learning_rate": 7.62518794643708e-05, "loss": 1.6872, "step": 1398 }, { "epoch": 0.3378411011832891, "grad_norm": 0.29341575503349304, "learning_rate": 7.62189274945684e-05, "loss": 1.7672, "step": 1399 }, { "epoch": 0.33808258874667957, "grad_norm": 0.28583914041519165, "learning_rate": 7.61859598106281e-05, "loss": 1.6706, "step": 1400 }, { "epoch": 0.33832407631007005, "grad_norm": 0.2980149984359741, "learning_rate": 7.615297643230883e-05, "loss": 1.7647, "step": 1401 }, { "epoch": 0.33856556387346054, "grad_norm": 0.2835332453250885, "learning_rate": 7.611997737937895e-05, "loss": 1.6417, "step": 1402 }, { "epoch": 0.338807051436851, "grad_norm": 0.29867133498191833, "learning_rate": 7.608696267161621e-05, "loss": 1.7539, "step": 1403 }, { "epoch": 0.3390485390002415, "grad_norm": 0.2891360819339752, "learning_rate": 7.60539323288078e-05, "loss": 1.6868, "step": 1404 }, { "epoch": 0.339290026563632, "grad_norm": 0.2839174270629883, "learning_rate": 7.602088637075016e-05, "loss": 1.7187, "step": 1405 }, { "epoch": 0.3395315141270225, "grad_norm": 0.28129449486732483, "learning_rate": 7.598782481724923e-05, "loss": 1.5241, "step": 1406 }, { "epoch": 0.33977300169041297, "grad_norm": 0.30046090483665466, "learning_rate": 7.595474768812021e-05, "loss": 1.7827, "step": 1407 }, { "epoch": 0.34001448925380345, "grad_norm": 0.3015134632587433, "learning_rate": 7.592165500318761e-05, "loss": 1.656, "step": 1408 }, { "epoch": 0.34025597681719394, "grad_norm": 0.28390491008758545, "learning_rate": 7.588854678228539e-05, "loss": 1.5548, "step": 1409 }, { "epoch": 0.3404974643805844, "grad_norm": 0.29424551129341125, "learning_rate": 7.585542304525667e-05, "loss": 1.5973, "step": 1410 }, { "epoch": 0.3407389519439749, "grad_norm": 0.3038700819015503, "learning_rate": 7.5822283811954e-05, "loss": 1.7428, "step": 1411 }, { "epoch": 0.3409804395073654, "grad_norm": 0.2812209129333496, "learning_rate": 7.57891291022391e-05, "loss": 1.603, "step": 1412 }, { "epoch": 0.3412219270707559, "grad_norm": 0.28015992045402527, "learning_rate": 7.575595893598304e-05, "loss": 1.6477, "step": 1413 }, { "epoch": 0.34146341463414637, "grad_norm": 0.2998206913471222, "learning_rate": 7.572277333306614e-05, "loss": 1.7594, "step": 1414 }, { "epoch": 0.34170490219753685, "grad_norm": 0.3038783073425293, "learning_rate": 7.568957231337799e-05, "loss": 1.9082, "step": 1415 }, { "epoch": 0.34194638976092734, "grad_norm": 0.2797698378562927, "learning_rate": 7.565635589681737e-05, "loss": 1.6403, "step": 1416 }, { "epoch": 0.3421878773243178, "grad_norm": 0.2895340025424957, "learning_rate": 7.56231241032923e-05, "loss": 1.7442, "step": 1417 }, { "epoch": 0.3424293648877083, "grad_norm": 0.28286561369895935, "learning_rate": 7.558987695272009e-05, "loss": 1.8109, "step": 1418 }, { "epoch": 0.3426708524510988, "grad_norm": 0.2754989564418793, "learning_rate": 7.555661446502714e-05, "loss": 1.6567, "step": 1419 }, { "epoch": 0.3429123400144893, "grad_norm": 0.2939462959766388, "learning_rate": 7.552333666014913e-05, "loss": 1.7232, "step": 1420 }, { "epoch": 0.34315382757787977, "grad_norm": 0.3267726004123688, "learning_rate": 7.549004355803086e-05, "loss": 1.9031, "step": 1421 }, { "epoch": 0.34339531514127025, "grad_norm": 0.3062046766281128, "learning_rate": 7.545673517862637e-05, "loss": 1.8201, "step": 1422 }, { "epoch": 0.34363680270466074, "grad_norm": 0.2913348376750946, "learning_rate": 7.542341154189878e-05, "loss": 1.6796, "step": 1423 }, { "epoch": 0.34387829026805117, "grad_norm": 0.29292842745780945, "learning_rate": 7.539007266782039e-05, "loss": 1.8692, "step": 1424 }, { "epoch": 0.34411977783144165, "grad_norm": 0.2910190224647522, "learning_rate": 7.535671857637265e-05, "loss": 1.7212, "step": 1425 }, { "epoch": 0.34436126539483214, "grad_norm": 0.30508074164390564, "learning_rate": 7.532334928754608e-05, "loss": 1.6877, "step": 1426 }, { "epoch": 0.3446027529582226, "grad_norm": 0.30790939927101135, "learning_rate": 7.528996482134038e-05, "loss": 1.8494, "step": 1427 }, { "epoch": 0.3448442405216131, "grad_norm": 0.30237263441085815, "learning_rate": 7.525656519776427e-05, "loss": 1.7703, "step": 1428 }, { "epoch": 0.3450857280850036, "grad_norm": 0.2987865209579468, "learning_rate": 7.522315043683559e-05, "loss": 1.7415, "step": 1429 }, { "epoch": 0.3453272156483941, "grad_norm": 0.2867589294910431, "learning_rate": 7.518972055858127e-05, "loss": 1.7095, "step": 1430 }, { "epoch": 0.34556870321178457, "grad_norm": 0.2851780652999878, "learning_rate": 7.515627558303728e-05, "loss": 1.6923, "step": 1431 }, { "epoch": 0.34581019077517505, "grad_norm": 0.3218798339366913, "learning_rate": 7.512281553024863e-05, "loss": 1.9602, "step": 1432 }, { "epoch": 0.34605167833856554, "grad_norm": 0.30771544575691223, "learning_rate": 7.508934042026938e-05, "loss": 1.968, "step": 1433 }, { "epoch": 0.346293165901956, "grad_norm": 0.28604716062545776, "learning_rate": 7.505585027316265e-05, "loss": 1.7667, "step": 1434 }, { "epoch": 0.3465346534653465, "grad_norm": 0.29297733306884766, "learning_rate": 7.502234510900046e-05, "loss": 1.7588, "step": 1435 }, { "epoch": 0.346776141028737, "grad_norm": 0.29484325647354126, "learning_rate": 7.498882494786396e-05, "loss": 1.8345, "step": 1436 }, { "epoch": 0.3470176285921275, "grad_norm": 0.30629557371139526, "learning_rate": 7.495528980984324e-05, "loss": 1.8114, "step": 1437 }, { "epoch": 0.34725911615551797, "grad_norm": 0.2843991219997406, "learning_rate": 7.492173971503732e-05, "loss": 1.7903, "step": 1438 }, { "epoch": 0.34750060371890845, "grad_norm": 0.29854124784469604, "learning_rate": 7.488817468355426e-05, "loss": 1.7366, "step": 1439 }, { "epoch": 0.34774209128229894, "grad_norm": 0.2846215069293976, "learning_rate": 7.485459473551101e-05, "loss": 1.6711, "step": 1440 }, { "epoch": 0.3479835788456894, "grad_norm": 0.2946016192436218, "learning_rate": 7.482099989103349e-05, "loss": 1.7971, "step": 1441 }, { "epoch": 0.3482250664090799, "grad_norm": 0.2970103621482849, "learning_rate": 7.478739017025654e-05, "loss": 1.8869, "step": 1442 }, { "epoch": 0.3484665539724704, "grad_norm": 0.30067914724349976, "learning_rate": 7.475376559332396e-05, "loss": 1.6867, "step": 1443 }, { "epoch": 0.3487080415358609, "grad_norm": 0.2954392731189728, "learning_rate": 7.472012618038835e-05, "loss": 1.8272, "step": 1444 }, { "epoch": 0.34894952909925137, "grad_norm": 0.29047641158103943, "learning_rate": 7.468647195161132e-05, "loss": 1.6909, "step": 1445 }, { "epoch": 0.34919101666264185, "grad_norm": 0.2894633710384369, "learning_rate": 7.465280292716329e-05, "loss": 1.747, "step": 1446 }, { "epoch": 0.34943250422603234, "grad_norm": 0.28781044483184814, "learning_rate": 7.461911912722355e-05, "loss": 1.5743, "step": 1447 }, { "epoch": 0.3496739917894228, "grad_norm": 0.30114755034446716, "learning_rate": 7.458542057198027e-05, "loss": 1.8015, "step": 1448 }, { "epoch": 0.3499154793528133, "grad_norm": 0.2965746819972992, "learning_rate": 7.455170728163045e-05, "loss": 1.7565, "step": 1449 }, { "epoch": 0.3501569669162038, "grad_norm": 0.32243311405181885, "learning_rate": 7.451797927637992e-05, "loss": 1.8916, "step": 1450 }, { "epoch": 0.3503984544795943, "grad_norm": 0.3025962710380554, "learning_rate": 7.448423657644336e-05, "loss": 1.6573, "step": 1451 }, { "epoch": 0.35063994204298476, "grad_norm": 0.2864728271961212, "learning_rate": 7.445047920204418e-05, "loss": 1.6866, "step": 1452 }, { "epoch": 0.35088142960637525, "grad_norm": 0.3004949986934662, "learning_rate": 7.441670717341466e-05, "loss": 1.7636, "step": 1453 }, { "epoch": 0.35112291716976574, "grad_norm": 0.2901725471019745, "learning_rate": 7.438292051079588e-05, "loss": 1.6909, "step": 1454 }, { "epoch": 0.3513644047331562, "grad_norm": 0.3008805513381958, "learning_rate": 7.434911923443757e-05, "loss": 1.8379, "step": 1455 }, { "epoch": 0.3516058922965467, "grad_norm": 0.2823113799095154, "learning_rate": 7.431530336459837e-05, "loss": 1.7156, "step": 1456 }, { "epoch": 0.3518473798599372, "grad_norm": 0.2956325113773346, "learning_rate": 7.428147292154554e-05, "loss": 1.819, "step": 1457 }, { "epoch": 0.3520888674233277, "grad_norm": 0.2923109233379364, "learning_rate": 7.424762792555516e-05, "loss": 1.7036, "step": 1458 }, { "epoch": 0.35233035498671816, "grad_norm": 0.28335970640182495, "learning_rate": 7.4213768396912e-05, "loss": 1.6259, "step": 1459 }, { "epoch": 0.35257184255010865, "grad_norm": 0.29262620210647583, "learning_rate": 7.417989435590953e-05, "loss": 1.6768, "step": 1460 }, { "epoch": 0.35281333011349914, "grad_norm": 0.3107213079929352, "learning_rate": 7.414600582284992e-05, "loss": 1.9349, "step": 1461 }, { "epoch": 0.3530548176768896, "grad_norm": 0.2939004600048065, "learning_rate": 7.411210281804407e-05, "loss": 1.7732, "step": 1462 }, { "epoch": 0.3532963052402801, "grad_norm": 0.28268906474113464, "learning_rate": 7.407818536181148e-05, "loss": 1.6477, "step": 1463 }, { "epoch": 0.3535377928036706, "grad_norm": 0.2945854365825653, "learning_rate": 7.404425347448036e-05, "loss": 1.756, "step": 1464 }, { "epoch": 0.3537792803670611, "grad_norm": 0.29590821266174316, "learning_rate": 7.401030717638758e-05, "loss": 1.716, "step": 1465 }, { "epoch": 0.35402076793045156, "grad_norm": 0.278075248003006, "learning_rate": 7.397634648787859e-05, "loss": 1.6162, "step": 1466 }, { "epoch": 0.35426225549384205, "grad_norm": 0.3146173655986786, "learning_rate": 7.394237142930751e-05, "loss": 1.794, "step": 1467 }, { "epoch": 0.35450374305723253, "grad_norm": 0.28644752502441406, "learning_rate": 7.390838202103709e-05, "loss": 1.7197, "step": 1468 }, { "epoch": 0.354745230620623, "grad_norm": 0.3296287953853607, "learning_rate": 7.387437828343864e-05, "loss": 1.6563, "step": 1469 }, { "epoch": 0.3549867181840135, "grad_norm": 0.2927229106426239, "learning_rate": 7.384036023689204e-05, "loss": 1.783, "step": 1470 }, { "epoch": 0.355228205747404, "grad_norm": 0.3050139546394348, "learning_rate": 7.380632790178583e-05, "loss": 1.6854, "step": 1471 }, { "epoch": 0.3554696933107945, "grad_norm": 0.3071853816509247, "learning_rate": 7.377228129851703e-05, "loss": 1.8396, "step": 1472 }, { "epoch": 0.35571118087418496, "grad_norm": 0.3828865587711334, "learning_rate": 7.373822044749124e-05, "loss": 1.6513, "step": 1473 }, { "epoch": 0.35595266843757545, "grad_norm": 0.31001341342926025, "learning_rate": 7.370414536912261e-05, "loss": 1.9578, "step": 1474 }, { "epoch": 0.35619415600096593, "grad_norm": 0.3114062249660492, "learning_rate": 7.367005608383382e-05, "loss": 1.8608, "step": 1475 }, { "epoch": 0.3564356435643564, "grad_norm": 0.29749569296836853, "learning_rate": 7.363595261205602e-05, "loss": 1.8161, "step": 1476 }, { "epoch": 0.3566771311277469, "grad_norm": 0.3024919629096985, "learning_rate": 7.360183497422895e-05, "loss": 1.8777, "step": 1477 }, { "epoch": 0.3569186186911374, "grad_norm": 0.2933084964752197, "learning_rate": 7.356770319080074e-05, "loss": 1.685, "step": 1478 }, { "epoch": 0.3571601062545279, "grad_norm": 0.30307063460350037, "learning_rate": 7.353355728222804e-05, "loss": 1.8085, "step": 1479 }, { "epoch": 0.35740159381791836, "grad_norm": 0.3066220283508301, "learning_rate": 7.349939726897604e-05, "loss": 1.7865, "step": 1480 }, { "epoch": 0.35764308138130885, "grad_norm": 0.30603402853012085, "learning_rate": 7.346522317151823e-05, "loss": 1.8973, "step": 1481 }, { "epoch": 0.35788456894469933, "grad_norm": 0.3032185137271881, "learning_rate": 7.343103501033667e-05, "loss": 1.7836, "step": 1482 }, { "epoch": 0.3581260565080898, "grad_norm": 0.2852492332458496, "learning_rate": 7.33968328059218e-05, "loss": 1.7213, "step": 1483 }, { "epoch": 0.3583675440714803, "grad_norm": 0.2975196838378906, "learning_rate": 7.336261657877247e-05, "loss": 1.8799, "step": 1484 }, { "epoch": 0.3586090316348708, "grad_norm": 0.2893730103969574, "learning_rate": 7.332838634939597e-05, "loss": 1.7396, "step": 1485 }, { "epoch": 0.3588505191982613, "grad_norm": 0.28530189394950867, "learning_rate": 7.329414213830793e-05, "loss": 1.7311, "step": 1486 }, { "epoch": 0.35909200676165176, "grad_norm": 0.3011293411254883, "learning_rate": 7.325988396603241e-05, "loss": 1.8305, "step": 1487 }, { "epoch": 0.35933349432504225, "grad_norm": 0.30045387148857117, "learning_rate": 7.322561185310178e-05, "loss": 1.7941, "step": 1488 }, { "epoch": 0.35957498188843273, "grad_norm": 0.30552494525909424, "learning_rate": 7.319132582005685e-05, "loss": 1.624, "step": 1489 }, { "epoch": 0.3598164694518232, "grad_norm": 0.3145065903663635, "learning_rate": 7.315702588744668e-05, "loss": 1.8659, "step": 1490 }, { "epoch": 0.3600579570152137, "grad_norm": 0.28973329067230225, "learning_rate": 7.312271207582873e-05, "loss": 1.8122, "step": 1491 }, { "epoch": 0.3602994445786042, "grad_norm": 0.2974574863910675, "learning_rate": 7.308838440576875e-05, "loss": 1.9315, "step": 1492 }, { "epoch": 0.3605409321419947, "grad_norm": 0.310698539018631, "learning_rate": 7.305404289784079e-05, "loss": 2.0308, "step": 1493 }, { "epoch": 0.36078241970538516, "grad_norm": 0.27670934796333313, "learning_rate": 7.301968757262721e-05, "loss": 1.6081, "step": 1494 }, { "epoch": 0.36102390726877565, "grad_norm": 0.27521172165870667, "learning_rate": 7.298531845071861e-05, "loss": 1.6248, "step": 1495 }, { "epoch": 0.36126539483216613, "grad_norm": 0.2828007638454437, "learning_rate": 7.295093555271395e-05, "loss": 1.7688, "step": 1496 }, { "epoch": 0.3615068823955566, "grad_norm": 0.29078638553619385, "learning_rate": 7.291653889922035e-05, "loss": 1.7569, "step": 1497 }, { "epoch": 0.3617483699589471, "grad_norm": 0.2831771969795227, "learning_rate": 7.288212851085322e-05, "loss": 1.6375, "step": 1498 }, { "epoch": 0.3619898575223376, "grad_norm": 0.307668000459671, "learning_rate": 7.28477044082362e-05, "loss": 1.9078, "step": 1499 }, { "epoch": 0.3622313450857281, "grad_norm": 0.2992483675479889, "learning_rate": 7.281326661200114e-05, "loss": 1.8321, "step": 1500 }, { "epoch": 0.36247283264911856, "grad_norm": 0.3095793128013611, "learning_rate": 7.277881514278811e-05, "loss": 1.8003, "step": 1501 }, { "epoch": 0.36271432021250904, "grad_norm": 0.29581761360168457, "learning_rate": 7.274435002124534e-05, "loss": 1.8205, "step": 1502 }, { "epoch": 0.36295580777589953, "grad_norm": 0.2918574810028076, "learning_rate": 7.27098712680293e-05, "loss": 1.7362, "step": 1503 }, { "epoch": 0.36319729533929, "grad_norm": 0.29628437757492065, "learning_rate": 7.267537890380459e-05, "loss": 1.6339, "step": 1504 }, { "epoch": 0.3634387829026805, "grad_norm": 0.2891917824745178, "learning_rate": 7.264087294924397e-05, "loss": 1.7657, "step": 1505 }, { "epoch": 0.363680270466071, "grad_norm": 0.28467097878456116, "learning_rate": 7.260635342502837e-05, "loss": 1.6824, "step": 1506 }, { "epoch": 0.3639217580294615, "grad_norm": 0.2941914200782776, "learning_rate": 7.25718203518468e-05, "loss": 1.6893, "step": 1507 }, { "epoch": 0.36416324559285196, "grad_norm": 0.28581947088241577, "learning_rate": 7.253727375039646e-05, "loss": 1.7599, "step": 1508 }, { "epoch": 0.36440473315624244, "grad_norm": 0.29284581542015076, "learning_rate": 7.250271364138259e-05, "loss": 1.8434, "step": 1509 }, { "epoch": 0.36464622071963293, "grad_norm": 0.3031042516231537, "learning_rate": 7.24681400455186e-05, "loss": 1.7305, "step": 1510 }, { "epoch": 0.3648877082830234, "grad_norm": 0.29574331641197205, "learning_rate": 7.24335529835259e-05, "loss": 1.6791, "step": 1511 }, { "epoch": 0.3651291958464139, "grad_norm": 0.3356575667858124, "learning_rate": 7.239895247613404e-05, "loss": 1.99, "step": 1512 }, { "epoch": 0.3653706834098044, "grad_norm": 0.28749579191207886, "learning_rate": 7.236433854408061e-05, "loss": 1.564, "step": 1513 }, { "epoch": 0.36561217097319487, "grad_norm": 0.2926652729511261, "learning_rate": 7.232971120811121e-05, "loss": 1.7391, "step": 1514 }, { "epoch": 0.36585365853658536, "grad_norm": 0.28827694058418274, "learning_rate": 7.229507048897952e-05, "loss": 1.7495, "step": 1515 }, { "epoch": 0.36609514609997584, "grad_norm": 0.29912760853767395, "learning_rate": 7.226041640744721e-05, "loss": 1.6743, "step": 1516 }, { "epoch": 0.36633663366336633, "grad_norm": 0.29889270663261414, "learning_rate": 7.222574898428398e-05, "loss": 1.8151, "step": 1517 }, { "epoch": 0.3665781212267568, "grad_norm": 0.30631643533706665, "learning_rate": 7.219106824026751e-05, "loss": 1.879, "step": 1518 }, { "epoch": 0.3668196087901473, "grad_norm": 0.303406298160553, "learning_rate": 7.215637419618347e-05, "loss": 1.6758, "step": 1519 }, { "epoch": 0.3670610963535378, "grad_norm": 0.2735936939716339, "learning_rate": 7.212166687282551e-05, "loss": 1.6312, "step": 1520 }, { "epoch": 0.36730258391692827, "grad_norm": 0.2837960720062256, "learning_rate": 7.208694629099523e-05, "loss": 1.6888, "step": 1521 }, { "epoch": 0.36754407148031876, "grad_norm": 0.2823384404182434, "learning_rate": 7.205221247150218e-05, "loss": 1.7177, "step": 1522 }, { "epoch": 0.36778555904370924, "grad_norm": 0.2750958204269409, "learning_rate": 7.201746543516382e-05, "loss": 1.5654, "step": 1523 }, { "epoch": 0.36802704660709973, "grad_norm": 0.29475295543670654, "learning_rate": 7.198270520280559e-05, "loss": 1.7343, "step": 1524 }, { "epoch": 0.3682685341704902, "grad_norm": 0.26865559816360474, "learning_rate": 7.194793179526078e-05, "loss": 1.5128, "step": 1525 }, { "epoch": 0.3685100217338807, "grad_norm": 0.28918734192848206, "learning_rate": 7.191314523337058e-05, "loss": 1.8486, "step": 1526 }, { "epoch": 0.3687515092972712, "grad_norm": 0.279786616563797, "learning_rate": 7.18783455379841e-05, "loss": 1.7636, "step": 1527 }, { "epoch": 0.36899299686066167, "grad_norm": 0.289742648601532, "learning_rate": 7.184353272995831e-05, "loss": 1.7157, "step": 1528 }, { "epoch": 0.36923448442405216, "grad_norm": 0.2943384647369385, "learning_rate": 7.180870683015802e-05, "loss": 1.8054, "step": 1529 }, { "epoch": 0.36947597198744264, "grad_norm": 0.2821480929851532, "learning_rate": 7.177386785945591e-05, "loss": 1.7366, "step": 1530 }, { "epoch": 0.3697174595508331, "grad_norm": 0.289713978767395, "learning_rate": 7.173901583873248e-05, "loss": 1.8236, "step": 1531 }, { "epoch": 0.3699589471142236, "grad_norm": 0.29103440046310425, "learning_rate": 7.170415078887604e-05, "loss": 1.6997, "step": 1532 }, { "epoch": 0.3702004346776141, "grad_norm": 0.28201615810394287, "learning_rate": 7.166927273078275e-05, "loss": 1.6628, "step": 1533 }, { "epoch": 0.3704419222410046, "grad_norm": 0.290673166513443, "learning_rate": 7.163438168535652e-05, "loss": 1.6033, "step": 1534 }, { "epoch": 0.37068340980439507, "grad_norm": 0.28538063168525696, "learning_rate": 7.159947767350907e-05, "loss": 1.6938, "step": 1535 }, { "epoch": 0.37092489736778556, "grad_norm": 0.2745797634124756, "learning_rate": 7.156456071615989e-05, "loss": 1.6851, "step": 1536 }, { "epoch": 0.37116638493117604, "grad_norm": 0.2976955473423004, "learning_rate": 7.15296308342362e-05, "loss": 1.8967, "step": 1537 }, { "epoch": 0.3714078724945665, "grad_norm": 0.27819645404815674, "learning_rate": 7.149468804867303e-05, "loss": 1.6837, "step": 1538 }, { "epoch": 0.371649360057957, "grad_norm": 0.2899627387523651, "learning_rate": 7.145973238041306e-05, "loss": 1.7605, "step": 1539 }, { "epoch": 0.3718908476213475, "grad_norm": 0.2979569733142853, "learning_rate": 7.142476385040677e-05, "loss": 1.7937, "step": 1540 }, { "epoch": 0.372132335184738, "grad_norm": 0.28263163566589355, "learning_rate": 7.138978247961228e-05, "loss": 1.7896, "step": 1541 }, { "epoch": 0.37237382274812847, "grad_norm": 0.28612080216407776, "learning_rate": 7.135478828899546e-05, "loss": 1.6894, "step": 1542 }, { "epoch": 0.37261531031151895, "grad_norm": 0.29869383573532104, "learning_rate": 7.131978129952985e-05, "loss": 1.658, "step": 1543 }, { "epoch": 0.37285679787490944, "grad_norm": 0.29021137952804565, "learning_rate": 7.128476153219665e-05, "loss": 1.8043, "step": 1544 }, { "epoch": 0.3730982854382999, "grad_norm": 0.29576247930526733, "learning_rate": 7.124972900798471e-05, "loss": 1.8886, "step": 1545 }, { "epoch": 0.3733397730016904, "grad_norm": 0.30245786905288696, "learning_rate": 7.121468374789055e-05, "loss": 1.6748, "step": 1546 }, { "epoch": 0.3735812605650809, "grad_norm": 0.2910307049751282, "learning_rate": 7.11796257729183e-05, "loss": 1.6754, "step": 1547 }, { "epoch": 0.3738227481284714, "grad_norm": 0.3084256052970886, "learning_rate": 7.114455510407974e-05, "loss": 1.7797, "step": 1548 }, { "epoch": 0.37406423569186187, "grad_norm": 0.29425761103630066, "learning_rate": 7.11094717623942e-05, "loss": 1.657, "step": 1549 }, { "epoch": 0.37430572325525235, "grad_norm": 0.29802206158638, "learning_rate": 7.107437576888869e-05, "loss": 1.7229, "step": 1550 }, { "epoch": 0.37454721081864284, "grad_norm": 0.3134438693523407, "learning_rate": 7.103926714459774e-05, "loss": 1.8207, "step": 1551 }, { "epoch": 0.3747886983820333, "grad_norm": 0.2775018513202667, "learning_rate": 7.100414591056347e-05, "loss": 1.6175, "step": 1552 }, { "epoch": 0.3750301859454238, "grad_norm": 0.2902098000049591, "learning_rate": 7.096901208783554e-05, "loss": 1.7434, "step": 1553 }, { "epoch": 0.3752716735088143, "grad_norm": 0.28874725103378296, "learning_rate": 7.093386569747121e-05, "loss": 1.6254, "step": 1554 }, { "epoch": 0.3755131610722048, "grad_norm": 0.2993185520172119, "learning_rate": 7.089870676053519e-05, "loss": 1.738, "step": 1555 }, { "epoch": 0.37575464863559527, "grad_norm": 0.29154321551322937, "learning_rate": 7.086353529809975e-05, "loss": 1.8475, "step": 1556 }, { "epoch": 0.37599613619898575, "grad_norm": 0.2712036073207855, "learning_rate": 7.082835133124472e-05, "loss": 1.6923, "step": 1557 }, { "epoch": 0.37623762376237624, "grad_norm": 0.2821873128414154, "learning_rate": 7.079315488105732e-05, "loss": 1.652, "step": 1558 }, { "epoch": 0.3764791113257667, "grad_norm": 0.29745790362358093, "learning_rate": 7.075794596863234e-05, "loss": 1.7248, "step": 1559 }, { "epoch": 0.3767205988891572, "grad_norm": 0.2806411385536194, "learning_rate": 7.072272461507198e-05, "loss": 1.6761, "step": 1560 }, { "epoch": 0.3769620864525477, "grad_norm": 0.2909418046474457, "learning_rate": 7.068749084148596e-05, "loss": 1.6984, "step": 1561 }, { "epoch": 0.3772035740159382, "grad_norm": 0.29692786931991577, "learning_rate": 7.065224466899137e-05, "loss": 1.7558, "step": 1562 }, { "epoch": 0.37744506157932867, "grad_norm": 0.3164198100566864, "learning_rate": 7.061698611871276e-05, "loss": 1.7037, "step": 1563 }, { "epoch": 0.37768654914271915, "grad_norm": 0.28125908970832825, "learning_rate": 7.058171521178214e-05, "loss": 1.6706, "step": 1564 }, { "epoch": 0.37792803670610964, "grad_norm": 0.30474957823753357, "learning_rate": 7.054643196933885e-05, "loss": 1.8193, "step": 1565 }, { "epoch": 0.3781695242695001, "grad_norm": 0.30764004588127136, "learning_rate": 7.051113641252971e-05, "loss": 1.8221, "step": 1566 }, { "epoch": 0.3784110118328906, "grad_norm": 0.27815955877304077, "learning_rate": 7.047582856250885e-05, "loss": 1.6797, "step": 1567 }, { "epoch": 0.3786524993962811, "grad_norm": 0.3051069974899292, "learning_rate": 7.044050844043778e-05, "loss": 1.8329, "step": 1568 }, { "epoch": 0.3788939869596716, "grad_norm": 0.29128777980804443, "learning_rate": 7.040517606748541e-05, "loss": 1.5598, "step": 1569 }, { "epoch": 0.37913547452306207, "grad_norm": 0.3010561466217041, "learning_rate": 7.036983146482794e-05, "loss": 1.7792, "step": 1570 }, { "epoch": 0.37937696208645255, "grad_norm": 0.32075226306915283, "learning_rate": 7.033447465364891e-05, "loss": 1.8827, "step": 1571 }, { "epoch": 0.37961844964984304, "grad_norm": 0.29643532633781433, "learning_rate": 7.029910565513919e-05, "loss": 1.8178, "step": 1572 }, { "epoch": 0.3798599372132335, "grad_norm": 0.312639445066452, "learning_rate": 7.026372449049699e-05, "loss": 1.7738, "step": 1573 }, { "epoch": 0.380101424776624, "grad_norm": 0.3031156659126282, "learning_rate": 7.022833118092773e-05, "loss": 1.7509, "step": 1574 }, { "epoch": 0.3803429123400145, "grad_norm": 0.28569579124450684, "learning_rate": 7.019292574764415e-05, "loss": 1.6701, "step": 1575 }, { "epoch": 0.380584399903405, "grad_norm": 0.29795435070991516, "learning_rate": 7.015750821186628e-05, "loss": 1.7447, "step": 1576 }, { "epoch": 0.38082588746679547, "grad_norm": 0.3067462742328644, "learning_rate": 7.012207859482138e-05, "loss": 1.8176, "step": 1577 }, { "epoch": 0.38106737503018595, "grad_norm": 0.2866966426372528, "learning_rate": 7.008663691774396e-05, "loss": 1.64, "step": 1578 }, { "epoch": 0.38130886259357644, "grad_norm": 0.28150928020477295, "learning_rate": 7.005118320187573e-05, "loss": 1.6103, "step": 1579 }, { "epoch": 0.3815503501569669, "grad_norm": 0.28614646196365356, "learning_rate": 7.001571746846565e-05, "loss": 1.7165, "step": 1580 }, { "epoch": 0.3817918377203574, "grad_norm": 0.29418009519577026, "learning_rate": 6.998023973876987e-05, "loss": 1.7223, "step": 1581 }, { "epoch": 0.3820333252837479, "grad_norm": 0.2841942310333252, "learning_rate": 6.99447500340517e-05, "loss": 1.5126, "step": 1582 }, { "epoch": 0.3822748128471384, "grad_norm": 0.30715563893318176, "learning_rate": 6.990924837558171e-05, "loss": 1.8462, "step": 1583 }, { "epoch": 0.38251630041052886, "grad_norm": 0.2975330948829651, "learning_rate": 6.987373478463754e-05, "loss": 1.83, "step": 1584 }, { "epoch": 0.38275778797391935, "grad_norm": 0.28202196955680847, "learning_rate": 6.983820928250402e-05, "loss": 1.6507, "step": 1585 }, { "epoch": 0.38299927553730984, "grad_norm": 0.28683874011039734, "learning_rate": 6.980267189047314e-05, "loss": 1.6067, "step": 1586 }, { "epoch": 0.3832407631007003, "grad_norm": 0.2874414622783661, "learning_rate": 6.976712262984401e-05, "loss": 1.5705, "step": 1587 }, { "epoch": 0.3834822506640908, "grad_norm": 0.29029643535614014, "learning_rate": 6.973156152192279e-05, "loss": 1.6031, "step": 1588 }, { "epoch": 0.3837237382274813, "grad_norm": 0.2800596356391907, "learning_rate": 6.969598858802285e-05, "loss": 1.7132, "step": 1589 }, { "epoch": 0.3839652257908718, "grad_norm": 0.28613579273223877, "learning_rate": 6.966040384946454e-05, "loss": 1.7108, "step": 1590 }, { "epoch": 0.38420671335426226, "grad_norm": 0.30830395221710205, "learning_rate": 6.962480732757535e-05, "loss": 1.6721, "step": 1591 }, { "epoch": 0.38444820091765275, "grad_norm": 0.29725363850593567, "learning_rate": 6.958919904368983e-05, "loss": 1.8466, "step": 1592 }, { "epoch": 0.38468968848104323, "grad_norm": 0.37713706493377686, "learning_rate": 6.955357901914957e-05, "loss": 1.9359, "step": 1593 }, { "epoch": 0.3849311760444337, "grad_norm": 0.29880592226982117, "learning_rate": 6.951794727530315e-05, "loss": 1.6493, "step": 1594 }, { "epoch": 0.3851726636078242, "grad_norm": 0.29536154866218567, "learning_rate": 6.948230383350626e-05, "loss": 1.714, "step": 1595 }, { "epoch": 0.3854141511712147, "grad_norm": 0.2920708954334259, "learning_rate": 6.944664871512154e-05, "loss": 1.7116, "step": 1596 }, { "epoch": 0.3856556387346052, "grad_norm": 0.29399019479751587, "learning_rate": 6.941098194151864e-05, "loss": 1.708, "step": 1597 }, { "epoch": 0.38589712629799566, "grad_norm": 0.29361873865127563, "learning_rate": 6.93753035340742e-05, "loss": 1.6959, "step": 1598 }, { "epoch": 0.38613861386138615, "grad_norm": 0.28866901993751526, "learning_rate": 6.933961351417186e-05, "loss": 1.6782, "step": 1599 }, { "epoch": 0.38638010142477663, "grad_norm": 0.2916489243507385, "learning_rate": 6.930391190320217e-05, "loss": 1.6748, "step": 1600 }, { "epoch": 0.3866215889881671, "grad_norm": 0.2900312840938568, "learning_rate": 6.926819872256264e-05, "loss": 1.7323, "step": 1601 }, { "epoch": 0.3868630765515576, "grad_norm": 0.2876428961753845, "learning_rate": 6.923247399365776e-05, "loss": 1.5876, "step": 1602 }, { "epoch": 0.3871045641149481, "grad_norm": 0.3108595609664917, "learning_rate": 6.919673773789888e-05, "loss": 1.7827, "step": 1603 }, { "epoch": 0.3873460516783386, "grad_norm": 0.30719828605651855, "learning_rate": 6.91609899767043e-05, "loss": 1.7054, "step": 1604 }, { "epoch": 0.38758753924172906, "grad_norm": 0.2866917848587036, "learning_rate": 6.912523073149917e-05, "loss": 1.7833, "step": 1605 }, { "epoch": 0.38782902680511955, "grad_norm": 0.28398698568344116, "learning_rate": 6.908946002371559e-05, "loss": 1.7662, "step": 1606 }, { "epoch": 0.38807051436851003, "grad_norm": 0.30139845609664917, "learning_rate": 6.905367787479248e-05, "loss": 1.7251, "step": 1607 }, { "epoch": 0.3883120019319005, "grad_norm": 0.30310872197151184, "learning_rate": 6.901788430617562e-05, "loss": 1.6172, "step": 1608 }, { "epoch": 0.388553489495291, "grad_norm": 0.30176272988319397, "learning_rate": 6.898207933931764e-05, "loss": 1.7503, "step": 1609 }, { "epoch": 0.3887949770586815, "grad_norm": 0.2845175266265869, "learning_rate": 6.894626299567804e-05, "loss": 1.7345, "step": 1610 }, { "epoch": 0.389036464622072, "grad_norm": 0.29985424876213074, "learning_rate": 6.891043529672308e-05, "loss": 1.7745, "step": 1611 }, { "epoch": 0.38927795218546246, "grad_norm": 0.2928183674812317, "learning_rate": 6.887459626392585e-05, "loss": 1.739, "step": 1612 }, { "epoch": 0.38951943974885295, "grad_norm": 0.3028491735458374, "learning_rate": 6.883874591876624e-05, "loss": 1.7981, "step": 1613 }, { "epoch": 0.38976092731224343, "grad_norm": 0.2912960350513458, "learning_rate": 6.880288428273091e-05, "loss": 1.7531, "step": 1614 }, { "epoch": 0.3900024148756339, "grad_norm": 0.29823577404022217, "learning_rate": 6.876701137731329e-05, "loss": 1.7662, "step": 1615 }, { "epoch": 0.3902439024390244, "grad_norm": 0.283291220664978, "learning_rate": 6.873112722401357e-05, "loss": 1.6241, "step": 1616 }, { "epoch": 0.3904853900024149, "grad_norm": 0.2959177792072296, "learning_rate": 6.869523184433867e-05, "loss": 1.8359, "step": 1617 }, { "epoch": 0.3907268775658054, "grad_norm": 0.28464576601982117, "learning_rate": 6.865932525980224e-05, "loss": 1.8036, "step": 1618 }, { "epoch": 0.39096836512919586, "grad_norm": 0.3036990463733673, "learning_rate": 6.862340749192467e-05, "loss": 1.9036, "step": 1619 }, { "epoch": 0.39120985269258635, "grad_norm": 0.29203951358795166, "learning_rate": 6.8587478562233e-05, "loss": 1.8115, "step": 1620 }, { "epoch": 0.39145134025597683, "grad_norm": 0.3313468098640442, "learning_rate": 6.855153849226102e-05, "loss": 1.7969, "step": 1621 }, { "epoch": 0.3916928278193673, "grad_norm": 0.27904316782951355, "learning_rate": 6.851558730354917e-05, "loss": 1.6888, "step": 1622 }, { "epoch": 0.3919343153827578, "grad_norm": 0.2998868227005005, "learning_rate": 6.847962501764454e-05, "loss": 1.7044, "step": 1623 }, { "epoch": 0.3921758029461483, "grad_norm": 0.29342302680015564, "learning_rate": 6.844365165610089e-05, "loss": 1.8111, "step": 1624 }, { "epoch": 0.3924172905095388, "grad_norm": 0.30202576518058777, "learning_rate": 6.840766724047862e-05, "loss": 1.7587, "step": 1625 }, { "epoch": 0.39265877807292926, "grad_norm": 0.2904617488384247, "learning_rate": 6.837167179234472e-05, "loss": 1.7509, "step": 1626 }, { "epoch": 0.39290026563631975, "grad_norm": 0.29565051198005676, "learning_rate": 6.833566533327285e-05, "loss": 1.7288, "step": 1627 }, { "epoch": 0.39314175319971023, "grad_norm": 0.28997743129730225, "learning_rate": 6.829964788484322e-05, "loss": 1.6617, "step": 1628 }, { "epoch": 0.3933832407631007, "grad_norm": 0.3013928532600403, "learning_rate": 6.826361946864266e-05, "loss": 1.6874, "step": 1629 }, { "epoch": 0.3936247283264912, "grad_norm": 0.30889344215393066, "learning_rate": 6.822758010626456e-05, "loss": 1.7683, "step": 1630 }, { "epoch": 0.3938662158898817, "grad_norm": 0.30394911766052246, "learning_rate": 6.819152981930883e-05, "loss": 1.6567, "step": 1631 }, { "epoch": 0.3941077034532722, "grad_norm": 0.3007825016975403, "learning_rate": 6.815546862938202e-05, "loss": 1.8263, "step": 1632 }, { "epoch": 0.39434919101666266, "grad_norm": 0.2870945632457733, "learning_rate": 6.811939655809715e-05, "loss": 1.7453, "step": 1633 }, { "epoch": 0.39459067858005314, "grad_norm": 0.30396586656570435, "learning_rate": 6.808331362707377e-05, "loss": 1.9653, "step": 1634 }, { "epoch": 0.39483216614344363, "grad_norm": 0.29016467928886414, "learning_rate": 6.804721985793792e-05, "loss": 1.7476, "step": 1635 }, { "epoch": 0.3950736537068341, "grad_norm": 0.28695061802864075, "learning_rate": 6.801111527232217e-05, "loss": 1.6255, "step": 1636 }, { "epoch": 0.3953151412702246, "grad_norm": 0.29611697793006897, "learning_rate": 6.797499989186558e-05, "loss": 1.7653, "step": 1637 }, { "epoch": 0.3955566288336151, "grad_norm": 0.29037386178970337, "learning_rate": 6.793887373821363e-05, "loss": 1.7265, "step": 1638 }, { "epoch": 0.3957981163970056, "grad_norm": 0.30254122614860535, "learning_rate": 6.790273683301831e-05, "loss": 1.6677, "step": 1639 }, { "epoch": 0.39603960396039606, "grad_norm": 0.30332258343696594, "learning_rate": 6.786658919793803e-05, "loss": 1.6915, "step": 1640 }, { "epoch": 0.39628109152378654, "grad_norm": 0.30312207341194153, "learning_rate": 6.783043085463762e-05, "loss": 1.8231, "step": 1641 }, { "epoch": 0.39652257908717703, "grad_norm": 0.29442471265792847, "learning_rate": 6.779426182478833e-05, "loss": 1.772, "step": 1642 }, { "epoch": 0.3967640666505675, "grad_norm": 0.2834378182888031, "learning_rate": 6.775808213006787e-05, "loss": 1.5351, "step": 1643 }, { "epoch": 0.397005554213958, "grad_norm": 0.2885272204875946, "learning_rate": 6.772189179216024e-05, "loss": 1.7361, "step": 1644 }, { "epoch": 0.3972470417773485, "grad_norm": 0.29504212737083435, "learning_rate": 6.768569083275594e-05, "loss": 1.5897, "step": 1645 }, { "epoch": 0.39748852934073897, "grad_norm": 0.28233128786087036, "learning_rate": 6.76494792735517e-05, "loss": 1.682, "step": 1646 }, { "epoch": 0.39773001690412946, "grad_norm": 0.2948426902294159, "learning_rate": 6.761325713625072e-05, "loss": 1.8264, "step": 1647 }, { "epoch": 0.39797150446751994, "grad_norm": 0.3149384558200836, "learning_rate": 6.75770244425625e-05, "loss": 1.8042, "step": 1648 }, { "epoch": 0.39821299203091043, "grad_norm": 0.5162968039512634, "learning_rate": 6.754078121420283e-05, "loss": 1.7922, "step": 1649 }, { "epoch": 0.3984544795943009, "grad_norm": 0.3127591013908386, "learning_rate": 6.750452747289386e-05, "loss": 1.7433, "step": 1650 }, { "epoch": 0.3986959671576914, "grad_norm": 0.3073781430721283, "learning_rate": 6.746826324036403e-05, "loss": 1.9478, "step": 1651 }, { "epoch": 0.3989374547210819, "grad_norm": 0.29865652322769165, "learning_rate": 6.743198853834806e-05, "loss": 1.6781, "step": 1652 }, { "epoch": 0.39917894228447237, "grad_norm": 0.29269179701805115, "learning_rate": 6.739570338858693e-05, "loss": 1.6779, "step": 1653 }, { "epoch": 0.39942042984786286, "grad_norm": 0.29973065853118896, "learning_rate": 6.735940781282796e-05, "loss": 1.6762, "step": 1654 }, { "epoch": 0.39966191741125334, "grad_norm": 0.29493921995162964, "learning_rate": 6.732310183282461e-05, "loss": 1.7821, "step": 1655 }, { "epoch": 0.39990340497464383, "grad_norm": 0.2976763844490051, "learning_rate": 6.728678547033662e-05, "loss": 1.7797, "step": 1656 }, { "epoch": 0.4001448925380343, "grad_norm": 0.3011676073074341, "learning_rate": 6.725045874712999e-05, "loss": 1.8517, "step": 1657 }, { "epoch": 0.4003863801014248, "grad_norm": 0.2780913710594177, "learning_rate": 6.721412168497686e-05, "loss": 1.6398, "step": 1658 }, { "epoch": 0.4006278676648153, "grad_norm": 0.3012704849243164, "learning_rate": 6.717777430565563e-05, "loss": 1.7828, "step": 1659 }, { "epoch": 0.40086935522820577, "grad_norm": 0.2912366986274719, "learning_rate": 6.714141663095085e-05, "loss": 1.6365, "step": 1660 }, { "epoch": 0.40111084279159626, "grad_norm": 0.29877522587776184, "learning_rate": 6.710504868265326e-05, "loss": 1.7992, "step": 1661 }, { "epoch": 0.40135233035498674, "grad_norm": 0.2767972946166992, "learning_rate": 6.706867048255972e-05, "loss": 1.7815, "step": 1662 }, { "epoch": 0.4015938179183772, "grad_norm": 0.3042682707309723, "learning_rate": 6.703228205247329e-05, "loss": 1.7805, "step": 1663 }, { "epoch": 0.4018353054817677, "grad_norm": 0.30128151178359985, "learning_rate": 6.699588341420308e-05, "loss": 1.5885, "step": 1664 }, { "epoch": 0.4020767930451582, "grad_norm": 0.3044366240501404, "learning_rate": 6.695947458956443e-05, "loss": 1.7927, "step": 1665 }, { "epoch": 0.4023182806085487, "grad_norm": 0.31940144300460815, "learning_rate": 6.692305560037871e-05, "loss": 1.7555, "step": 1666 }, { "epoch": 0.40255976817193917, "grad_norm": 0.29257991909980774, "learning_rate": 6.688662646847337e-05, "loss": 1.7296, "step": 1667 }, { "epoch": 0.40280125573532966, "grad_norm": 0.2922995090484619, "learning_rate": 6.685018721568197e-05, "loss": 1.859, "step": 1668 }, { "epoch": 0.40304274329872014, "grad_norm": 0.29321998357772827, "learning_rate": 6.681373786384416e-05, "loss": 1.8717, "step": 1669 }, { "epoch": 0.4032842308621106, "grad_norm": 0.2926195561885834, "learning_rate": 6.67772784348056e-05, "loss": 1.744, "step": 1670 }, { "epoch": 0.4035257184255011, "grad_norm": 0.2928321659564972, "learning_rate": 6.674080895041798e-05, "loss": 1.6524, "step": 1671 }, { "epoch": 0.4037672059888916, "grad_norm": 0.27669477462768555, "learning_rate": 6.670432943253906e-05, "loss": 1.6994, "step": 1672 }, { "epoch": 0.4040086935522821, "grad_norm": 0.2832506597042084, "learning_rate": 6.66678399030326e-05, "loss": 1.656, "step": 1673 }, { "epoch": 0.40425018111567257, "grad_norm": 0.2894239127635956, "learning_rate": 6.663134038376835e-05, "loss": 1.7739, "step": 1674 }, { "epoch": 0.40449166867906305, "grad_norm": 0.3202664852142334, "learning_rate": 6.659483089662203e-05, "loss": 1.8945, "step": 1675 }, { "epoch": 0.40473315624245354, "grad_norm": 0.30392253398895264, "learning_rate": 6.65583114634754e-05, "loss": 1.8153, "step": 1676 }, { "epoch": 0.404974643805844, "grad_norm": 0.2862154543399811, "learning_rate": 6.652178210621609e-05, "loss": 1.5981, "step": 1677 }, { "epoch": 0.4052161313692345, "grad_norm": 0.2755371630191803, "learning_rate": 6.648524284673777e-05, "loss": 1.5838, "step": 1678 }, { "epoch": 0.405457618932625, "grad_norm": 0.2834921181201935, "learning_rate": 6.644869370693995e-05, "loss": 1.7238, "step": 1679 }, { "epoch": 0.4056991064960155, "grad_norm": 0.3005118668079376, "learning_rate": 6.641213470872814e-05, "loss": 1.8139, "step": 1680 }, { "epoch": 0.40594059405940597, "grad_norm": 0.2832629084587097, "learning_rate": 6.637556587401374e-05, "loss": 1.6167, "step": 1681 }, { "epoch": 0.40618208162279645, "grad_norm": 0.28953537344932556, "learning_rate": 6.633898722471402e-05, "loss": 1.7748, "step": 1682 }, { "epoch": 0.4064235691861869, "grad_norm": 0.3091205358505249, "learning_rate": 6.630239878275214e-05, "loss": 1.8577, "step": 1683 }, { "epoch": 0.40666505674957737, "grad_norm": 0.29937854409217834, "learning_rate": 6.626580057005716e-05, "loss": 1.7232, "step": 1684 }, { "epoch": 0.40690654431296785, "grad_norm": 0.3058737516403198, "learning_rate": 6.622919260856392e-05, "loss": 1.7831, "step": 1685 }, { "epoch": 0.40714803187635834, "grad_norm": 0.3024592399597168, "learning_rate": 6.619257492021324e-05, "loss": 1.7563, "step": 1686 }, { "epoch": 0.4073895194397488, "grad_norm": 0.28710269927978516, "learning_rate": 6.615594752695163e-05, "loss": 1.5895, "step": 1687 }, { "epoch": 0.4076310070031393, "grad_norm": 0.2993939220905304, "learning_rate": 6.611931045073147e-05, "loss": 1.7654, "step": 1688 }, { "epoch": 0.4078724945665298, "grad_norm": 0.2898264527320862, "learning_rate": 6.608266371351094e-05, "loss": 1.6705, "step": 1689 }, { "epoch": 0.4081139821299203, "grad_norm": 0.3162345886230469, "learning_rate": 6.604600733725407e-05, "loss": 1.9323, "step": 1690 }, { "epoch": 0.40835546969331077, "grad_norm": 0.29158738255500793, "learning_rate": 6.600934134393055e-05, "loss": 1.6548, "step": 1691 }, { "epoch": 0.40859695725670125, "grad_norm": 0.3384915888309479, "learning_rate": 6.597266575551593e-05, "loss": 1.9476, "step": 1692 }, { "epoch": 0.40883844482009174, "grad_norm": 0.3284933269023895, "learning_rate": 6.593598059399148e-05, "loss": 2.0154, "step": 1693 }, { "epoch": 0.4090799323834822, "grad_norm": 0.2912117838859558, "learning_rate": 6.589928588134418e-05, "loss": 1.5946, "step": 1694 }, { "epoch": 0.4093214199468727, "grad_norm": 0.29039883613586426, "learning_rate": 6.586258163956677e-05, "loss": 1.6569, "step": 1695 }, { "epoch": 0.4095629075102632, "grad_norm": 0.29182976484298706, "learning_rate": 6.582586789065768e-05, "loss": 1.7651, "step": 1696 }, { "epoch": 0.4098043950736537, "grad_norm": 0.287898987531662, "learning_rate": 6.578914465662109e-05, "loss": 1.8019, "step": 1697 }, { "epoch": 0.41004588263704417, "grad_norm": 0.28599241375923157, "learning_rate": 6.575241195946679e-05, "loss": 1.7098, "step": 1698 }, { "epoch": 0.41028737020043465, "grad_norm": 0.29934966564178467, "learning_rate": 6.571566982121027e-05, "loss": 1.772, "step": 1699 }, { "epoch": 0.41052885776382514, "grad_norm": 0.28729212284088135, "learning_rate": 6.56789182638727e-05, "loss": 1.7225, "step": 1700 }, { "epoch": 0.4107703453272156, "grad_norm": 0.29589328169822693, "learning_rate": 6.564215730948089e-05, "loss": 1.7541, "step": 1701 }, { "epoch": 0.4110118328906061, "grad_norm": 0.30237460136413574, "learning_rate": 6.560538698006725e-05, "loss": 1.7909, "step": 1702 }, { "epoch": 0.4112533204539966, "grad_norm": 0.2996138036251068, "learning_rate": 6.55686072976698e-05, "loss": 1.7098, "step": 1703 }, { "epoch": 0.4114948080173871, "grad_norm": 0.31643104553222656, "learning_rate": 6.553181828433225e-05, "loss": 1.7339, "step": 1704 }, { "epoch": 0.41173629558077757, "grad_norm": 0.26532241702079773, "learning_rate": 6.549501996210385e-05, "loss": 1.5175, "step": 1705 }, { "epoch": 0.41197778314416805, "grad_norm": 0.2829447090625763, "learning_rate": 6.545821235303938e-05, "loss": 1.7096, "step": 1706 }, { "epoch": 0.41221927070755854, "grad_norm": 0.2905785143375397, "learning_rate": 6.542139547919927e-05, "loss": 1.6579, "step": 1707 }, { "epoch": 0.412460758270949, "grad_norm": 0.28984224796295166, "learning_rate": 6.538456936264943e-05, "loss": 1.6448, "step": 1708 }, { "epoch": 0.4127022458343395, "grad_norm": 0.27520114183425903, "learning_rate": 6.534773402546138e-05, "loss": 1.5512, "step": 1709 }, { "epoch": 0.41294373339773, "grad_norm": 0.3095702528953552, "learning_rate": 6.531088948971212e-05, "loss": 1.7525, "step": 1710 }, { "epoch": 0.4131852209611205, "grad_norm": 0.2894667387008667, "learning_rate": 6.527403577748419e-05, "loss": 1.63, "step": 1711 }, { "epoch": 0.41342670852451097, "grad_norm": 0.28379741311073303, "learning_rate": 6.523717291086558e-05, "loss": 1.6679, "step": 1712 }, { "epoch": 0.41366819608790145, "grad_norm": 0.2986467480659485, "learning_rate": 6.520030091194984e-05, "loss": 1.7661, "step": 1713 }, { "epoch": 0.41390968365129194, "grad_norm": 0.3017704486846924, "learning_rate": 6.516341980283593e-05, "loss": 1.7938, "step": 1714 }, { "epoch": 0.4141511712146824, "grad_norm": 0.30169060826301575, "learning_rate": 6.512652960562829e-05, "loss": 1.7313, "step": 1715 }, { "epoch": 0.4143926587780729, "grad_norm": 0.30517810583114624, "learning_rate": 6.508963034243683e-05, "loss": 1.8192, "step": 1716 }, { "epoch": 0.4146341463414634, "grad_norm": 0.2784494161605835, "learning_rate": 6.505272203537687e-05, "loss": 1.6988, "step": 1717 }, { "epoch": 0.4148756339048539, "grad_norm": 0.6281806826591492, "learning_rate": 6.501580470656913e-05, "loss": 1.6437, "step": 1718 }, { "epoch": 0.41511712146824437, "grad_norm": 0.28869011998176575, "learning_rate": 6.497887837813981e-05, "loss": 1.7, "step": 1719 }, { "epoch": 0.41535860903163485, "grad_norm": 0.30811411142349243, "learning_rate": 6.494194307222043e-05, "loss": 1.8591, "step": 1720 }, { "epoch": 0.41560009659502534, "grad_norm": 0.2911246716976166, "learning_rate": 6.49049988109479e-05, "loss": 1.7103, "step": 1721 }, { "epoch": 0.4158415841584158, "grad_norm": 0.27908068895339966, "learning_rate": 6.486804561646456e-05, "loss": 1.6012, "step": 1722 }, { "epoch": 0.4160830717218063, "grad_norm": 0.2703852951526642, "learning_rate": 6.483108351091802e-05, "loss": 1.3896, "step": 1723 }, { "epoch": 0.4163245592851968, "grad_norm": 0.2969891130924225, "learning_rate": 6.479411251646128e-05, "loss": 1.8627, "step": 1724 }, { "epoch": 0.4165660468485873, "grad_norm": 0.31090447306632996, "learning_rate": 6.475713265525267e-05, "loss": 1.8897, "step": 1725 }, { "epoch": 0.41680753441197776, "grad_norm": 0.293481707572937, "learning_rate": 6.472014394945582e-05, "loss": 1.7154, "step": 1726 }, { "epoch": 0.41704902197536825, "grad_norm": 0.28805553913116455, "learning_rate": 6.468314642123963e-05, "loss": 1.8052, "step": 1727 }, { "epoch": 0.41729050953875874, "grad_norm": 0.3230077624320984, "learning_rate": 6.464614009277837e-05, "loss": 1.8545, "step": 1728 }, { "epoch": 0.4175319971021492, "grad_norm": 0.29030686616897583, "learning_rate": 6.460912498625147e-05, "loss": 1.7401, "step": 1729 }, { "epoch": 0.4177734846655397, "grad_norm": 0.30003416538238525, "learning_rate": 6.457210112384376e-05, "loss": 1.7905, "step": 1730 }, { "epoch": 0.4180149722289302, "grad_norm": 0.2969563901424408, "learning_rate": 6.453506852774518e-05, "loss": 1.6655, "step": 1731 }, { "epoch": 0.4182564597923207, "grad_norm": 0.29960548877716064, "learning_rate": 6.4498027220151e-05, "loss": 1.7215, "step": 1732 }, { "epoch": 0.41849794735571116, "grad_norm": 0.2943073511123657, "learning_rate": 6.446097722326166e-05, "loss": 1.8138, "step": 1733 }, { "epoch": 0.41873943491910165, "grad_norm": 0.30021798610687256, "learning_rate": 6.442391855928285e-05, "loss": 1.6481, "step": 1734 }, { "epoch": 0.41898092248249214, "grad_norm": 0.30978235602378845, "learning_rate": 6.438685125042543e-05, "loss": 1.7445, "step": 1735 }, { "epoch": 0.4192224100458826, "grad_norm": 0.2761182487010956, "learning_rate": 6.434977531890542e-05, "loss": 1.6566, "step": 1736 }, { "epoch": 0.4194638976092731, "grad_norm": 0.3045799434185028, "learning_rate": 6.431269078694405e-05, "loss": 1.7833, "step": 1737 }, { "epoch": 0.4197053851726636, "grad_norm": 0.32096797227859497, "learning_rate": 6.427559767676767e-05, "loss": 1.6964, "step": 1738 }, { "epoch": 0.4199468727360541, "grad_norm": 0.31089526414871216, "learning_rate": 6.423849601060778e-05, "loss": 1.7477, "step": 1739 }, { "epoch": 0.42018836029944456, "grad_norm": 0.2792001962661743, "learning_rate": 6.420138581070106e-05, "loss": 1.5561, "step": 1740 }, { "epoch": 0.42042984786283505, "grad_norm": 0.3000398874282837, "learning_rate": 6.41642670992892e-05, "loss": 1.8557, "step": 1741 }, { "epoch": 0.42067133542622553, "grad_norm": 0.2896421253681183, "learning_rate": 6.412713989861907e-05, "loss": 1.6449, "step": 1742 }, { "epoch": 0.420912822989616, "grad_norm": 0.29571399092674255, "learning_rate": 6.409000423094264e-05, "loss": 1.9293, "step": 1743 }, { "epoch": 0.4211543105530065, "grad_norm": 0.3130645155906677, "learning_rate": 6.405286011851689e-05, "loss": 1.926, "step": 1744 }, { "epoch": 0.421395798116397, "grad_norm": 0.31119149923324585, "learning_rate": 6.40157075836039e-05, "loss": 1.7332, "step": 1745 }, { "epoch": 0.4216372856797875, "grad_norm": 0.2839961051940918, "learning_rate": 6.397854664847078e-05, "loss": 1.7123, "step": 1746 }, { "epoch": 0.42187877324317796, "grad_norm": 0.29415860772132874, "learning_rate": 6.394137733538972e-05, "loss": 1.7624, "step": 1747 }, { "epoch": 0.42212026080656845, "grad_norm": 0.2900754511356354, "learning_rate": 6.390419966663785e-05, "loss": 1.7625, "step": 1748 }, { "epoch": 0.42236174836995893, "grad_norm": 0.2831285297870636, "learning_rate": 6.386701366449741e-05, "loss": 1.8242, "step": 1749 }, { "epoch": 0.4226032359333494, "grad_norm": 0.29872801899909973, "learning_rate": 6.382981935125555e-05, "loss": 1.8123, "step": 1750 }, { "epoch": 0.4228447234967399, "grad_norm": 0.2907567620277405, "learning_rate": 6.379261674920443e-05, "loss": 1.8015, "step": 1751 }, { "epoch": 0.4230862110601304, "grad_norm": 0.28774112462997437, "learning_rate": 6.375540588064118e-05, "loss": 1.6881, "step": 1752 }, { "epoch": 0.4233276986235209, "grad_norm": 0.2798502445220947, "learning_rate": 6.37181867678679e-05, "loss": 1.6737, "step": 1753 }, { "epoch": 0.42356918618691136, "grad_norm": 0.3050326704978943, "learning_rate": 6.368095943319157e-05, "loss": 1.9083, "step": 1754 }, { "epoch": 0.42381067375030185, "grad_norm": 0.27888479828834534, "learning_rate": 6.364372389892423e-05, "loss": 1.6069, "step": 1755 }, { "epoch": 0.42405216131369233, "grad_norm": 0.2976183295249939, "learning_rate": 6.360648018738267e-05, "loss": 1.6508, "step": 1756 }, { "epoch": 0.4242936488770828, "grad_norm": 0.2891102433204651, "learning_rate": 6.356922832088867e-05, "loss": 1.6793, "step": 1757 }, { "epoch": 0.4245351364404733, "grad_norm": 0.29795247316360474, "learning_rate": 6.35319683217689e-05, "loss": 1.846, "step": 1758 }, { "epoch": 0.4247766240038638, "grad_norm": 0.3060801029205322, "learning_rate": 6.349470021235491e-05, "loss": 1.7774, "step": 1759 }, { "epoch": 0.4250181115672543, "grad_norm": 0.2822737395763397, "learning_rate": 6.345742401498305e-05, "loss": 1.6143, "step": 1760 }, { "epoch": 0.42525959913064476, "grad_norm": 0.30945831537246704, "learning_rate": 6.342013975199459e-05, "loss": 1.8053, "step": 1761 }, { "epoch": 0.42550108669403525, "grad_norm": 0.32073312997817993, "learning_rate": 6.338284744573557e-05, "loss": 1.8592, "step": 1762 }, { "epoch": 0.42574257425742573, "grad_norm": 0.28697794675827026, "learning_rate": 6.334554711855692e-05, "loss": 1.7174, "step": 1763 }, { "epoch": 0.4259840618208162, "grad_norm": 0.27200600504875183, "learning_rate": 6.330823879281431e-05, "loss": 1.5327, "step": 1764 }, { "epoch": 0.4262255493842067, "grad_norm": 0.31434571743011475, "learning_rate": 6.327092249086825e-05, "loss": 2.0499, "step": 1765 }, { "epoch": 0.4264670369475972, "grad_norm": 0.3031150996685028, "learning_rate": 6.323359823508401e-05, "loss": 1.8122, "step": 1766 }, { "epoch": 0.4267085245109877, "grad_norm": 0.31416165828704834, "learning_rate": 6.319626604783166e-05, "loss": 1.6704, "step": 1767 }, { "epoch": 0.42695001207437816, "grad_norm": 0.2872917354106903, "learning_rate": 6.315892595148593e-05, "loss": 1.6125, "step": 1768 }, { "epoch": 0.42719149963776865, "grad_norm": 0.27335771918296814, "learning_rate": 6.312157796842641e-05, "loss": 1.7027, "step": 1769 }, { "epoch": 0.42743298720115913, "grad_norm": 0.29706230759620667, "learning_rate": 6.308422212103736e-05, "loss": 1.9199, "step": 1770 }, { "epoch": 0.4276744747645496, "grad_norm": 0.29793912172317505, "learning_rate": 6.304685843170772e-05, "loss": 1.7714, "step": 1771 }, { "epoch": 0.4279159623279401, "grad_norm": 0.2788400650024414, "learning_rate": 6.300948692283122e-05, "loss": 1.6532, "step": 1772 }, { "epoch": 0.4281574498913306, "grad_norm": 0.2901703715324402, "learning_rate": 6.297210761680618e-05, "loss": 1.6379, "step": 1773 }, { "epoch": 0.4283989374547211, "grad_norm": 0.3157692849636078, "learning_rate": 6.293472053603563e-05, "loss": 1.9983, "step": 1774 }, { "epoch": 0.42864042501811156, "grad_norm": 0.2885459363460541, "learning_rate": 6.289732570292728e-05, "loss": 1.7734, "step": 1775 }, { "epoch": 0.42888191258150204, "grad_norm": 0.2981279492378235, "learning_rate": 6.28599231398935e-05, "loss": 1.75, "step": 1776 }, { "epoch": 0.42912340014489253, "grad_norm": 0.2864707410335541, "learning_rate": 6.282251286935121e-05, "loss": 1.6159, "step": 1777 }, { "epoch": 0.429364887708283, "grad_norm": 0.28859418630599976, "learning_rate": 6.278509491372207e-05, "loss": 1.6682, "step": 1778 }, { "epoch": 0.4296063752716735, "grad_norm": 0.3026987612247467, "learning_rate": 6.274766929543221e-05, "loss": 1.7819, "step": 1779 }, { "epoch": 0.429847862835064, "grad_norm": 0.29130101203918457, "learning_rate": 6.271023603691248e-05, "loss": 1.7521, "step": 1780 }, { "epoch": 0.4300893503984545, "grad_norm": 0.28929030895233154, "learning_rate": 6.267279516059823e-05, "loss": 1.8133, "step": 1781 }, { "epoch": 0.43033083796184496, "grad_norm": 0.2978893220424652, "learning_rate": 6.263534668892938e-05, "loss": 1.7392, "step": 1782 }, { "epoch": 0.43057232552523544, "grad_norm": 0.2781008183956146, "learning_rate": 6.259789064435044e-05, "loss": 1.7491, "step": 1783 }, { "epoch": 0.43081381308862593, "grad_norm": 0.2996642291545868, "learning_rate": 6.256042704931045e-05, "loss": 1.7415, "step": 1784 }, { "epoch": 0.4310553006520164, "grad_norm": 0.2978329360485077, "learning_rate": 6.252295592626293e-05, "loss": 1.783, "step": 1785 }, { "epoch": 0.4312967882154069, "grad_norm": 0.29910266399383545, "learning_rate": 6.248547729766596e-05, "loss": 1.8312, "step": 1786 }, { "epoch": 0.4315382757787974, "grad_norm": 0.2915440499782562, "learning_rate": 6.244799118598211e-05, "loss": 1.8281, "step": 1787 }, { "epoch": 0.43177976334218787, "grad_norm": 0.2931154668331146, "learning_rate": 6.241049761367845e-05, "loss": 1.8002, "step": 1788 }, { "epoch": 0.43202125090557836, "grad_norm": 0.2912518382072449, "learning_rate": 6.237299660322645e-05, "loss": 1.8242, "step": 1789 }, { "epoch": 0.43226273846896884, "grad_norm": 0.2783883810043335, "learning_rate": 6.233548817710213e-05, "loss": 1.6564, "step": 1790 }, { "epoch": 0.43250422603235933, "grad_norm": 0.29445287585258484, "learning_rate": 6.22979723577859e-05, "loss": 1.7016, "step": 1791 }, { "epoch": 0.4327457135957498, "grad_norm": 0.27962037920951843, "learning_rate": 6.22604491677626e-05, "loss": 1.5158, "step": 1792 }, { "epoch": 0.4329872011591403, "grad_norm": 0.28880858421325684, "learning_rate": 6.222291862952153e-05, "loss": 1.6425, "step": 1793 }, { "epoch": 0.4332286887225308, "grad_norm": 0.2967377007007599, "learning_rate": 6.218538076555635e-05, "loss": 1.8272, "step": 1794 }, { "epoch": 0.43347017628592127, "grad_norm": 0.29760682582855225, "learning_rate": 6.214783559836512e-05, "loss": 1.7123, "step": 1795 }, { "epoch": 0.43371166384931176, "grad_norm": 0.26291459798812866, "learning_rate": 6.211028315045027e-05, "loss": 1.4946, "step": 1796 }, { "epoch": 0.43395315141270224, "grad_norm": 0.3057827651500702, "learning_rate": 6.207272344431866e-05, "loss": 1.7002, "step": 1797 }, { "epoch": 0.43419463897609273, "grad_norm": 0.3144783079624176, "learning_rate": 6.20351565024814e-05, "loss": 1.8826, "step": 1798 }, { "epoch": 0.4344361265394832, "grad_norm": 0.29561561346054077, "learning_rate": 6.199758234745401e-05, "loss": 1.7635, "step": 1799 }, { "epoch": 0.4346776141028737, "grad_norm": 0.2834794521331787, "learning_rate": 6.196000100175629e-05, "loss": 1.7412, "step": 1800 }, { "epoch": 0.4349191016662642, "grad_norm": 0.29855379462242126, "learning_rate": 6.192241248791236e-05, "loss": 1.8944, "step": 1801 }, { "epoch": 0.43516058922965467, "grad_norm": 0.30267342925071716, "learning_rate": 6.188481682845068e-05, "loss": 1.7823, "step": 1802 }, { "epoch": 0.43540207679304516, "grad_norm": 0.30372753739356995, "learning_rate": 6.184721404590392e-05, "loss": 1.871, "step": 1803 }, { "epoch": 0.43564356435643564, "grad_norm": 0.2923888862133026, "learning_rate": 6.180960416280906e-05, "loss": 1.7681, "step": 1804 }, { "epoch": 0.4358850519198261, "grad_norm": 0.3047378361225128, "learning_rate": 6.177198720170735e-05, "loss": 1.8215, "step": 1805 }, { "epoch": 0.4361265394832166, "grad_norm": 0.30522942543029785, "learning_rate": 6.173436318514425e-05, "loss": 1.8184, "step": 1806 }, { "epoch": 0.4363680270466071, "grad_norm": 0.29536402225494385, "learning_rate": 6.169673213566946e-05, "loss": 1.7936, "step": 1807 }, { "epoch": 0.4366095146099976, "grad_norm": 0.29366007447242737, "learning_rate": 6.165909407583687e-05, "loss": 1.8101, "step": 1808 }, { "epoch": 0.43685100217338807, "grad_norm": 0.3027951419353485, "learning_rate": 6.162144902820467e-05, "loss": 1.8094, "step": 1809 }, { "epoch": 0.43709248973677856, "grad_norm": 0.30503374338150024, "learning_rate": 6.15837970153351e-05, "loss": 1.5617, "step": 1810 }, { "epoch": 0.43733397730016904, "grad_norm": 0.279544472694397, "learning_rate": 6.154613805979467e-05, "loss": 1.6252, "step": 1811 }, { "epoch": 0.4375754648635595, "grad_norm": 0.300307959318161, "learning_rate": 6.150847218415401e-05, "loss": 1.8711, "step": 1812 }, { "epoch": 0.43781695242695, "grad_norm": 0.2886538505554199, "learning_rate": 6.14707994109879e-05, "loss": 1.5643, "step": 1813 }, { "epoch": 0.4380584399903405, "grad_norm": 0.2958506643772125, "learning_rate": 6.143311976287531e-05, "loss": 1.7646, "step": 1814 }, { "epoch": 0.438299927553731, "grad_norm": 0.29687565565109253, "learning_rate": 6.139543326239926e-05, "loss": 1.642, "step": 1815 }, { "epoch": 0.43854141511712147, "grad_norm": 0.2901788353919983, "learning_rate": 6.135773993214685e-05, "loss": 1.5288, "step": 1816 }, { "epoch": 0.43878290268051195, "grad_norm": 0.29023823142051697, "learning_rate": 6.132003979470939e-05, "loss": 1.6135, "step": 1817 }, { "epoch": 0.43902439024390244, "grad_norm": 0.2979198098182678, "learning_rate": 6.128233287268216e-05, "loss": 1.7553, "step": 1818 }, { "epoch": 0.4392658778072929, "grad_norm": 0.28564679622650146, "learning_rate": 6.124461918866459e-05, "loss": 1.6581, "step": 1819 }, { "epoch": 0.4395073653706834, "grad_norm": 0.2789301872253418, "learning_rate": 6.12068987652601e-05, "loss": 1.6782, "step": 1820 }, { "epoch": 0.4397488529340739, "grad_norm": 0.2890610992908478, "learning_rate": 6.116917162507618e-05, "loss": 1.6198, "step": 1821 }, { "epoch": 0.4399903404974644, "grad_norm": 0.29789113998413086, "learning_rate": 6.113143779072432e-05, "loss": 1.8246, "step": 1822 }, { "epoch": 0.44023182806085487, "grad_norm": 0.2954878807067871, "learning_rate": 6.109369728482006e-05, "loss": 1.8309, "step": 1823 }, { "epoch": 0.44047331562424535, "grad_norm": 0.29515141248703003, "learning_rate": 6.105595012998293e-05, "loss": 1.8473, "step": 1824 }, { "epoch": 0.44071480318763584, "grad_norm": 0.28120049834251404, "learning_rate": 6.1018196348836384e-05, "loss": 1.6964, "step": 1825 }, { "epoch": 0.4409562907510263, "grad_norm": 0.2976786196231842, "learning_rate": 6.098043596400795e-05, "loss": 1.7344, "step": 1826 }, { "epoch": 0.4411977783144168, "grad_norm": 0.27497339248657227, "learning_rate": 6.094266899812904e-05, "loss": 1.5725, "step": 1827 }, { "epoch": 0.4414392658778073, "grad_norm": 0.29393255710601807, "learning_rate": 6.090489547383502e-05, "loss": 1.7235, "step": 1828 }, { "epoch": 0.4416807534411978, "grad_norm": 0.28918004035949707, "learning_rate": 6.0867115413765206e-05, "loss": 1.6561, "step": 1829 }, { "epoch": 0.44192224100458827, "grad_norm": 0.2973797917366028, "learning_rate": 6.0829328840562836e-05, "loss": 1.8733, "step": 1830 }, { "epoch": 0.44216372856797875, "grad_norm": 0.2833927571773529, "learning_rate": 6.079153577687502e-05, "loss": 1.6147, "step": 1831 }, { "epoch": 0.44240521613136924, "grad_norm": 0.2931411862373352, "learning_rate": 6.075373624535279e-05, "loss": 1.7791, "step": 1832 }, { "epoch": 0.4426467036947597, "grad_norm": 0.30425596237182617, "learning_rate": 6.071593026865102e-05, "loss": 1.8605, "step": 1833 }, { "epoch": 0.4428881912581502, "grad_norm": 0.2918083369731903, "learning_rate": 6.067811786942851e-05, "loss": 1.814, "step": 1834 }, { "epoch": 0.4431296788215407, "grad_norm": 0.2972649931907654, "learning_rate": 6.064029907034783e-05, "loss": 1.6886, "step": 1835 }, { "epoch": 0.4433711663849312, "grad_norm": 0.2869977653026581, "learning_rate": 6.0602473894075416e-05, "loss": 1.6742, "step": 1836 }, { "epoch": 0.44361265394832167, "grad_norm": 0.2964032292366028, "learning_rate": 6.0564642363281577e-05, "loss": 1.6807, "step": 1837 }, { "epoch": 0.44385414151171215, "grad_norm": 0.4216827154159546, "learning_rate": 6.052680450064036e-05, "loss": 1.8796, "step": 1838 }, { "epoch": 0.44409562907510264, "grad_norm": 0.2836448848247528, "learning_rate": 6.0488960328829615e-05, "loss": 1.8391, "step": 1839 }, { "epoch": 0.4443371166384931, "grad_norm": 0.2807202637195587, "learning_rate": 6.045110987053102e-05, "loss": 1.6688, "step": 1840 }, { "epoch": 0.4445786042018836, "grad_norm": 0.2864443361759186, "learning_rate": 6.0413253148429986e-05, "loss": 1.7344, "step": 1841 }, { "epoch": 0.4448200917652741, "grad_norm": 0.31129157543182373, "learning_rate": 6.0375390185215664e-05, "loss": 1.8526, "step": 1842 }, { "epoch": 0.4450615793286646, "grad_norm": 0.3081095218658447, "learning_rate": 6.033752100358101e-05, "loss": 1.8352, "step": 1843 }, { "epoch": 0.44530306689205507, "grad_norm": 0.28839313983917236, "learning_rate": 6.0299645626222636e-05, "loss": 1.7259, "step": 1844 }, { "epoch": 0.44554455445544555, "grad_norm": 0.28718453645706177, "learning_rate": 6.026176407584088e-05, "loss": 1.8188, "step": 1845 }, { "epoch": 0.44578604201883604, "grad_norm": 0.29002997279167175, "learning_rate": 6.022387637513983e-05, "loss": 1.6686, "step": 1846 }, { "epoch": 0.4460275295822265, "grad_norm": 0.29619693756103516, "learning_rate": 6.0185982546827205e-05, "loss": 1.8598, "step": 1847 }, { "epoch": 0.446269017145617, "grad_norm": 0.29213395714759827, "learning_rate": 6.014808261361441e-05, "loss": 1.6915, "step": 1848 }, { "epoch": 0.4465105047090075, "grad_norm": 0.2871073782444, "learning_rate": 6.0110176598216525e-05, "loss": 1.6989, "step": 1849 }, { "epoch": 0.446751992272398, "grad_norm": 0.2969827353954315, "learning_rate": 6.0072264523352275e-05, "loss": 1.6611, "step": 1850 }, { "epoch": 0.44699347983578847, "grad_norm": 0.28079450130462646, "learning_rate": 6.003434641174398e-05, "loss": 1.6835, "step": 1851 }, { "epoch": 0.44723496739917895, "grad_norm": 0.3196025490760803, "learning_rate": 5.999642228611761e-05, "loss": 1.9651, "step": 1852 }, { "epoch": 0.44747645496256944, "grad_norm": 0.3039729595184326, "learning_rate": 5.9958492169202776e-05, "loss": 1.6505, "step": 1853 }, { "epoch": 0.4477179425259599, "grad_norm": 0.2899879217147827, "learning_rate": 5.9920556083732585e-05, "loss": 1.6848, "step": 1854 }, { "epoch": 0.4479594300893504, "grad_norm": 0.2984606921672821, "learning_rate": 5.988261405244382e-05, "loss": 1.833, "step": 1855 }, { "epoch": 0.4482009176527409, "grad_norm": 0.3010087311267853, "learning_rate": 5.984466609807675e-05, "loss": 1.7083, "step": 1856 }, { "epoch": 0.4484424052161314, "grad_norm": 0.2915293872356415, "learning_rate": 5.980671224337523e-05, "loss": 1.75, "step": 1857 }, { "epoch": 0.44868389277952186, "grad_norm": 0.2936859428882599, "learning_rate": 5.976875251108667e-05, "loss": 1.7717, "step": 1858 }, { "epoch": 0.44892538034291235, "grad_norm": 0.2996618151664734, "learning_rate": 5.9730786923961946e-05, "loss": 1.7103, "step": 1859 }, { "epoch": 0.44916686790630284, "grad_norm": 0.2864699363708496, "learning_rate": 5.9692815504755505e-05, "loss": 1.686, "step": 1860 }, { "epoch": 0.4494083554696933, "grad_norm": 0.2949967086315155, "learning_rate": 5.965483827622526e-05, "loss": 1.6495, "step": 1861 }, { "epoch": 0.4496498430330838, "grad_norm": 0.29744744300842285, "learning_rate": 5.9616855261132586e-05, "loss": 1.7475, "step": 1862 }, { "epoch": 0.4498913305964743, "grad_norm": 0.2873002290725708, "learning_rate": 5.957886648224238e-05, "loss": 1.6152, "step": 1863 }, { "epoch": 0.4501328181598648, "grad_norm": 0.3030773401260376, "learning_rate": 5.954087196232292e-05, "loss": 1.9727, "step": 1864 }, { "epoch": 0.45037430572325526, "grad_norm": 0.31282663345336914, "learning_rate": 5.9502871724146014e-05, "loss": 1.7839, "step": 1865 }, { "epoch": 0.45061579328664575, "grad_norm": 0.2903457283973694, "learning_rate": 5.9464865790486815e-05, "loss": 1.7714, "step": 1866 }, { "epoch": 0.45085728085003623, "grad_norm": 0.2956843376159668, "learning_rate": 5.9426854184123935e-05, "loss": 1.7269, "step": 1867 }, { "epoch": 0.4510987684134267, "grad_norm": 0.30317896604537964, "learning_rate": 5.938883692783938e-05, "loss": 1.8399, "step": 1868 }, { "epoch": 0.4513402559768172, "grad_norm": 0.3119915723800659, "learning_rate": 5.9350814044418515e-05, "loss": 1.7026, "step": 1869 }, { "epoch": 0.4515817435402077, "grad_norm": 0.29706525802612305, "learning_rate": 5.931278555665013e-05, "loss": 1.7608, "step": 1870 }, { "epoch": 0.4518232311035982, "grad_norm": 0.30105337500572205, "learning_rate": 5.927475148732633e-05, "loss": 1.9063, "step": 1871 }, { "epoch": 0.45206471866698866, "grad_norm": 0.28921782970428467, "learning_rate": 5.923671185924255e-05, "loss": 1.5752, "step": 1872 }, { "epoch": 0.45230620623037915, "grad_norm": 0.29720595479011536, "learning_rate": 5.919866669519765e-05, "loss": 1.7688, "step": 1873 }, { "epoch": 0.45254769379376963, "grad_norm": 0.30140766501426697, "learning_rate": 5.9160616017993685e-05, "loss": 1.7024, "step": 1874 }, { "epoch": 0.4527891813571601, "grad_norm": 0.29524683952331543, "learning_rate": 5.9122559850436085e-05, "loss": 1.9465, "step": 1875 }, { "epoch": 0.4530306689205506, "grad_norm": 0.30466195940971375, "learning_rate": 5.90844982153336e-05, "loss": 1.905, "step": 1876 }, { "epoch": 0.4532721564839411, "grad_norm": 0.3002755641937256, "learning_rate": 5.904643113549816e-05, "loss": 1.7018, "step": 1877 }, { "epoch": 0.4535136440473316, "grad_norm": 0.2979101836681366, "learning_rate": 5.900835863374503e-05, "loss": 1.709, "step": 1878 }, { "epoch": 0.45375513161072206, "grad_norm": 0.2791266441345215, "learning_rate": 5.897028073289272e-05, "loss": 1.6399, "step": 1879 }, { "epoch": 0.45399661917411255, "grad_norm": 0.3100603222846985, "learning_rate": 5.8932197455762974e-05, "loss": 1.7315, "step": 1880 }, { "epoch": 0.45423810673750303, "grad_norm": 0.29227718710899353, "learning_rate": 5.889410882518071e-05, "loss": 1.8037, "step": 1881 }, { "epoch": 0.4544795943008935, "grad_norm": 0.2878359258174896, "learning_rate": 5.885601486397412e-05, "loss": 1.6143, "step": 1882 }, { "epoch": 0.454721081864284, "grad_norm": 0.30222880840301514, "learning_rate": 5.881791559497456e-05, "loss": 1.6901, "step": 1883 }, { "epoch": 0.4549625694276745, "grad_norm": 0.32026219367980957, "learning_rate": 5.877981104101657e-05, "loss": 1.8235, "step": 1884 }, { "epoch": 0.455204056991065, "grad_norm": 0.3001136779785156, "learning_rate": 5.874170122493784e-05, "loss": 1.9317, "step": 1885 }, { "epoch": 0.45544554455445546, "grad_norm": 0.27894434332847595, "learning_rate": 5.870358616957926e-05, "loss": 1.5573, "step": 1886 }, { "epoch": 0.45568703211784595, "grad_norm": 0.2884087860584259, "learning_rate": 5.8665465897784824e-05, "loss": 1.6473, "step": 1887 }, { "epoch": 0.45592851968123643, "grad_norm": 0.2925865054130554, "learning_rate": 5.8627340432401656e-05, "loss": 1.7021, "step": 1888 }, { "epoch": 0.4561700072446269, "grad_norm": 0.30170294642448425, "learning_rate": 5.8589209796279996e-05, "loss": 1.8159, "step": 1889 }, { "epoch": 0.4564114948080174, "grad_norm": 0.3007247745990753, "learning_rate": 5.855107401227319e-05, "loss": 1.8116, "step": 1890 }, { "epoch": 0.4566529823714079, "grad_norm": 0.28724128007888794, "learning_rate": 5.851293310323767e-05, "loss": 1.7403, "step": 1891 }, { "epoch": 0.4568944699347984, "grad_norm": 0.28229671716690063, "learning_rate": 5.847478709203291e-05, "loss": 1.7085, "step": 1892 }, { "epoch": 0.45713595749818886, "grad_norm": 0.3043849766254425, "learning_rate": 5.843663600152148e-05, "loss": 1.7451, "step": 1893 }, { "epoch": 0.45737744506157935, "grad_norm": 0.2898251712322235, "learning_rate": 5.8398479854569e-05, "loss": 1.82, "step": 1894 }, { "epoch": 0.45761893262496983, "grad_norm": 0.3063147962093353, "learning_rate": 5.836031867404406e-05, "loss": 1.9651, "step": 1895 }, { "epoch": 0.4578604201883603, "grad_norm": 0.2852920889854431, "learning_rate": 5.832215248281834e-05, "loss": 1.6368, "step": 1896 }, { "epoch": 0.4581019077517508, "grad_norm": 0.30048584938049316, "learning_rate": 5.828398130376645e-05, "loss": 1.738, "step": 1897 }, { "epoch": 0.4583433953151413, "grad_norm": 0.2849442660808563, "learning_rate": 5.824580515976608e-05, "loss": 1.6016, "step": 1898 }, { "epoch": 0.4585848828785318, "grad_norm": 0.28128960728645325, "learning_rate": 5.8207624073697816e-05, "loss": 1.6994, "step": 1899 }, { "epoch": 0.45882637044192226, "grad_norm": 0.29790449142456055, "learning_rate": 5.8169438068445246e-05, "loss": 1.7874, "step": 1900 }, { "epoch": 0.45906785800531275, "grad_norm": 0.28888222575187683, "learning_rate": 5.813124716689488e-05, "loss": 1.7332, "step": 1901 }, { "epoch": 0.45930934556870323, "grad_norm": 0.2996605932712555, "learning_rate": 5.809305139193619e-05, "loss": 1.8353, "step": 1902 }, { "epoch": 0.4595508331320937, "grad_norm": 0.29596322774887085, "learning_rate": 5.805485076646156e-05, "loss": 1.682, "step": 1903 }, { "epoch": 0.4597923206954842, "grad_norm": 0.3038746416568756, "learning_rate": 5.801664531336628e-05, "loss": 1.6489, "step": 1904 }, { "epoch": 0.4600338082588747, "grad_norm": 0.29957544803619385, "learning_rate": 5.797843505554852e-05, "loss": 1.7067, "step": 1905 }, { "epoch": 0.4602752958222652, "grad_norm": 0.29159390926361084, "learning_rate": 5.794022001590935e-05, "loss": 1.7833, "step": 1906 }, { "epoch": 0.46051678338565566, "grad_norm": 0.2858039140701294, "learning_rate": 5.790200021735268e-05, "loss": 1.6577, "step": 1907 }, { "epoch": 0.46075827094904614, "grad_norm": 0.3013875484466553, "learning_rate": 5.78637756827853e-05, "loss": 1.771, "step": 1908 }, { "epoch": 0.46099975851243663, "grad_norm": 0.30382275581359863, "learning_rate": 5.782554643511685e-05, "loss": 1.7432, "step": 1909 }, { "epoch": 0.4612412460758271, "grad_norm": 0.28172066807746887, "learning_rate": 5.778731249725974e-05, "loss": 1.6688, "step": 1910 }, { "epoch": 0.4614827336392176, "grad_norm": 0.28853094577789307, "learning_rate": 5.7749073892129244e-05, "loss": 1.6464, "step": 1911 }, { "epoch": 0.4617242212026081, "grad_norm": 0.29917535185813904, "learning_rate": 5.77108306426434e-05, "loss": 1.7856, "step": 1912 }, { "epoch": 0.4619657087659986, "grad_norm": 0.2910739779472351, "learning_rate": 5.767258277172305e-05, "loss": 1.7508, "step": 1913 }, { "epoch": 0.46220719632938906, "grad_norm": 0.29367324709892273, "learning_rate": 5.7634330302291795e-05, "loss": 1.6982, "step": 1914 }, { "epoch": 0.46244868389277954, "grad_norm": 0.297710657119751, "learning_rate": 5.759607325727601e-05, "loss": 1.679, "step": 1915 }, { "epoch": 0.46269017145617003, "grad_norm": 0.316152423620224, "learning_rate": 5.7557811659604766e-05, "loss": 2.0318, "step": 1916 }, { "epoch": 0.4629316590195605, "grad_norm": 0.28985244035720825, "learning_rate": 5.751954553220992e-05, "loss": 1.767, "step": 1917 }, { "epoch": 0.463173146582951, "grad_norm": 0.28868818283081055, "learning_rate": 5.748127489802601e-05, "loss": 1.7489, "step": 1918 }, { "epoch": 0.4634146341463415, "grad_norm": 0.2903616428375244, "learning_rate": 5.744299977999027e-05, "loss": 1.663, "step": 1919 }, { "epoch": 0.46365612170973197, "grad_norm": 0.289998859167099, "learning_rate": 5.7404720201042664e-05, "loss": 1.796, "step": 1920 }, { "epoch": 0.46389760927312246, "grad_norm": 0.3104515075683594, "learning_rate": 5.736643618412579e-05, "loss": 1.8674, "step": 1921 }, { "epoch": 0.46413909683651294, "grad_norm": 0.30258458852767944, "learning_rate": 5.73281477521849e-05, "loss": 1.7722, "step": 1922 }, { "epoch": 0.46438058439990343, "grad_norm": 0.2958606779575348, "learning_rate": 5.728985492816793e-05, "loss": 1.7653, "step": 1923 }, { "epoch": 0.4646220719632939, "grad_norm": 0.2986416220664978, "learning_rate": 5.725155773502542e-05, "loss": 1.7348, "step": 1924 }, { "epoch": 0.4648635595266844, "grad_norm": 0.28947514295578003, "learning_rate": 5.7213256195710516e-05, "loss": 1.7064, "step": 1925 }, { "epoch": 0.4651050470900749, "grad_norm": 0.29144924879074097, "learning_rate": 5.717495033317901e-05, "loss": 1.7251, "step": 1926 }, { "epoch": 0.46534653465346537, "grad_norm": 0.2858518660068512, "learning_rate": 5.7136640170389263e-05, "loss": 1.5974, "step": 1927 }, { "epoch": 0.46558802221685586, "grad_norm": 0.3050241768360138, "learning_rate": 5.7098325730302195e-05, "loss": 1.6366, "step": 1928 }, { "epoch": 0.46582950978024634, "grad_norm": 0.3163412809371948, "learning_rate": 5.7060007035881324e-05, "loss": 1.7627, "step": 1929 }, { "epoch": 0.46607099734363683, "grad_norm": 0.31081393361091614, "learning_rate": 5.702168411009271e-05, "loss": 1.9454, "step": 1930 }, { "epoch": 0.4663124849070273, "grad_norm": 0.2910256087779999, "learning_rate": 5.698335697590491e-05, "loss": 1.665, "step": 1931 }, { "epoch": 0.4665539724704178, "grad_norm": 0.3363916277885437, "learning_rate": 5.6945025656289076e-05, "loss": 1.8586, "step": 1932 }, { "epoch": 0.4667954600338083, "grad_norm": 0.29299837350845337, "learning_rate": 5.690669017421879e-05, "loss": 1.714, "step": 1933 }, { "epoch": 0.46703694759719877, "grad_norm": 0.2883269488811493, "learning_rate": 5.686835055267018e-05, "loss": 1.7377, "step": 1934 }, { "epoch": 0.46727843516058926, "grad_norm": 0.2988598644733429, "learning_rate": 5.683000681462185e-05, "loss": 1.8745, "step": 1935 }, { "epoch": 0.46751992272397974, "grad_norm": 0.28724566102027893, "learning_rate": 5.679165898305485e-05, "loss": 1.6266, "step": 1936 }, { "epoch": 0.4677614102873702, "grad_norm": 0.3093075752258301, "learning_rate": 5.675330708095268e-05, "loss": 1.7257, "step": 1937 }, { "epoch": 0.4680028978507607, "grad_norm": 0.2962707579135895, "learning_rate": 5.671495113130133e-05, "loss": 1.7822, "step": 1938 }, { "epoch": 0.4682443854141512, "grad_norm": 0.29928648471832275, "learning_rate": 5.6676591157089166e-05, "loss": 1.8773, "step": 1939 }, { "epoch": 0.4684858729775417, "grad_norm": 0.29388922452926636, "learning_rate": 5.663822718130696e-05, "loss": 1.812, "step": 1940 }, { "epoch": 0.46872736054093217, "grad_norm": 0.30245307087898254, "learning_rate": 5.659985922694792e-05, "loss": 1.8397, "step": 1941 }, { "epoch": 0.4689688481043226, "grad_norm": 0.2745733857154846, "learning_rate": 5.656148731700763e-05, "loss": 1.537, "step": 1942 }, { "epoch": 0.4692103356677131, "grad_norm": 0.3115253150463104, "learning_rate": 5.6523111474484016e-05, "loss": 1.7199, "step": 1943 }, { "epoch": 0.46945182323110357, "grad_norm": 0.29406628012657166, "learning_rate": 5.64847317223774e-05, "loss": 1.5904, "step": 1944 }, { "epoch": 0.46969331079449406, "grad_norm": 0.28329282999038696, "learning_rate": 5.644634808369044e-05, "loss": 1.5586, "step": 1945 }, { "epoch": 0.46993479835788454, "grad_norm": 0.2810218334197998, "learning_rate": 5.6407960581428064e-05, "loss": 1.6041, "step": 1946 }, { "epoch": 0.47017628592127503, "grad_norm": 0.29527994990348816, "learning_rate": 5.6369569238597606e-05, "loss": 1.7334, "step": 1947 }, { "epoch": 0.4704177734846655, "grad_norm": 0.2843940556049347, "learning_rate": 5.633117407820865e-05, "loss": 1.7189, "step": 1948 }, { "epoch": 0.470659261048056, "grad_norm": 0.2926733195781708, "learning_rate": 5.629277512327306e-05, "loss": 1.6232, "step": 1949 }, { "epoch": 0.4709007486114465, "grad_norm": 0.2885076403617859, "learning_rate": 5.625437239680501e-05, "loss": 1.613, "step": 1950 }, { "epoch": 0.47114223617483697, "grad_norm": 0.2991282045841217, "learning_rate": 5.621596592182091e-05, "loss": 1.8069, "step": 1951 }, { "epoch": 0.47138372373822746, "grad_norm": 0.28910577297210693, "learning_rate": 5.617755572133942e-05, "loss": 1.6859, "step": 1952 }, { "epoch": 0.47162521130161794, "grad_norm": 0.30354511737823486, "learning_rate": 5.6139141818381446e-05, "loss": 1.8275, "step": 1953 }, { "epoch": 0.4718666988650084, "grad_norm": 0.31015917658805847, "learning_rate": 5.610072423597008e-05, "loss": 1.7884, "step": 1954 }, { "epoch": 0.4721081864283989, "grad_norm": 0.29766714572906494, "learning_rate": 5.606230299713068e-05, "loss": 1.7359, "step": 1955 }, { "epoch": 0.4723496739917894, "grad_norm": 0.28761327266693115, "learning_rate": 5.6023878124890716e-05, "loss": 1.6276, "step": 1956 }, { "epoch": 0.4725911615551799, "grad_norm": 0.2790350317955017, "learning_rate": 5.59854496422799e-05, "loss": 1.668, "step": 1957 }, { "epoch": 0.47283264911857037, "grad_norm": 0.2867283225059509, "learning_rate": 5.5947017572330084e-05, "loss": 1.6826, "step": 1958 }, { "epoch": 0.47307413668196086, "grad_norm": 0.295149028301239, "learning_rate": 5.5908581938075264e-05, "loss": 1.7293, "step": 1959 }, { "epoch": 0.47331562424535134, "grad_norm": 0.28681740164756775, "learning_rate": 5.587014276255157e-05, "loss": 1.6374, "step": 1960 }, { "epoch": 0.4735571118087418, "grad_norm": 0.30145061016082764, "learning_rate": 5.5831700068797286e-05, "loss": 1.7667, "step": 1961 }, { "epoch": 0.4737985993721323, "grad_norm": 0.28270459175109863, "learning_rate": 5.579325387985277e-05, "loss": 1.5991, "step": 1962 }, { "epoch": 0.4740400869355228, "grad_norm": 0.3064315915107727, "learning_rate": 5.575480421876048e-05, "loss": 1.6987, "step": 1963 }, { "epoch": 0.4742815744989133, "grad_norm": 0.3125905692577362, "learning_rate": 5.5716351108564965e-05, "loss": 1.829, "step": 1964 }, { "epoch": 0.47452306206230377, "grad_norm": 0.27573850750923157, "learning_rate": 5.567789457231286e-05, "loss": 1.5916, "step": 1965 }, { "epoch": 0.47476454962569425, "grad_norm": 0.31593382358551025, "learning_rate": 5.56394346330528e-05, "loss": 1.7972, "step": 1966 }, { "epoch": 0.47500603718908474, "grad_norm": 0.31528183817863464, "learning_rate": 5.560097131383552e-05, "loss": 1.7862, "step": 1967 }, { "epoch": 0.4752475247524752, "grad_norm": 0.29852205514907837, "learning_rate": 5.556250463771373e-05, "loss": 1.7726, "step": 1968 }, { "epoch": 0.4754890123158657, "grad_norm": 0.2947467863559723, "learning_rate": 5.5524034627742174e-05, "loss": 1.5666, "step": 1969 }, { "epoch": 0.4757304998792562, "grad_norm": 0.2866770923137665, "learning_rate": 5.5485561306977604e-05, "loss": 1.683, "step": 1970 }, { "epoch": 0.4759719874426467, "grad_norm": 0.29122698307037354, "learning_rate": 5.5447084698478746e-05, "loss": 1.5568, "step": 1971 }, { "epoch": 0.47621347500603717, "grad_norm": 0.29016897082328796, "learning_rate": 5.540860482530629e-05, "loss": 1.8156, "step": 1972 }, { "epoch": 0.47645496256942765, "grad_norm": 0.30687373876571655, "learning_rate": 5.537012171052289e-05, "loss": 1.9504, "step": 1973 }, { "epoch": 0.47669645013281814, "grad_norm": 0.2921820282936096, "learning_rate": 5.533163537719315e-05, "loss": 1.7727, "step": 1974 }, { "epoch": 0.4769379376962086, "grad_norm": 0.28375446796417236, "learning_rate": 5.5293145848383576e-05, "loss": 1.6884, "step": 1975 }, { "epoch": 0.4771794252595991, "grad_norm": 0.2858161926269531, "learning_rate": 5.5254653147162626e-05, "loss": 1.6366, "step": 1976 }, { "epoch": 0.4774209128229896, "grad_norm": 0.2857281565666199, "learning_rate": 5.521615729660066e-05, "loss": 1.6584, "step": 1977 }, { "epoch": 0.4776624003863801, "grad_norm": 0.37559181451797485, "learning_rate": 5.5177658319769856e-05, "loss": 1.7549, "step": 1978 }, { "epoch": 0.47790388794977057, "grad_norm": 0.2909361720085144, "learning_rate": 5.513915623974436e-05, "loss": 1.7672, "step": 1979 }, { "epoch": 0.47814537551316105, "grad_norm": 0.28941604495048523, "learning_rate": 5.510065107960013e-05, "loss": 1.7656, "step": 1980 }, { "epoch": 0.47838686307655154, "grad_norm": 0.31622543931007385, "learning_rate": 5.506214286241495e-05, "loss": 1.9301, "step": 1981 }, { "epoch": 0.478628350639942, "grad_norm": 0.3030341565608978, "learning_rate": 5.5023631611268486e-05, "loss": 1.7637, "step": 1982 }, { "epoch": 0.4788698382033325, "grad_norm": 0.31502172350883484, "learning_rate": 5.498511734924219e-05, "loss": 1.735, "step": 1983 }, { "epoch": 0.479111325766723, "grad_norm": 0.2744358479976654, "learning_rate": 5.4946600099419306e-05, "loss": 1.7905, "step": 1984 }, { "epoch": 0.4793528133301135, "grad_norm": 0.2697215974330902, "learning_rate": 5.4908079884884925e-05, "loss": 1.5816, "step": 1985 }, { "epoch": 0.47959430089350397, "grad_norm": 0.27491477131843567, "learning_rate": 5.486955672872584e-05, "loss": 1.6728, "step": 1986 }, { "epoch": 0.47983578845689445, "grad_norm": 0.2729800343513489, "learning_rate": 5.483103065403069e-05, "loss": 1.4615, "step": 1987 }, { "epoch": 0.48007727602028494, "grad_norm": 0.2833001911640167, "learning_rate": 5.4792501683889784e-05, "loss": 1.6446, "step": 1988 }, { "epoch": 0.4803187635836754, "grad_norm": 0.2760225832462311, "learning_rate": 5.4753969841395224e-05, "loss": 1.6635, "step": 1989 }, { "epoch": 0.4805602511470659, "grad_norm": 0.3052677512168884, "learning_rate": 5.471543514964078e-05, "loss": 1.6142, "step": 1990 }, { "epoch": 0.4808017387104564, "grad_norm": 0.2885996997356415, "learning_rate": 5.4676897631721993e-05, "loss": 1.6738, "step": 1991 }, { "epoch": 0.4810432262738469, "grad_norm": 0.28482338786125183, "learning_rate": 5.4638357310736054e-05, "loss": 1.71, "step": 1992 }, { "epoch": 0.48128471383723737, "grad_norm": 0.2904854118824005, "learning_rate": 5.459981420978183e-05, "loss": 1.6673, "step": 1993 }, { "epoch": 0.48152620140062785, "grad_norm": 0.3114590346813202, "learning_rate": 5.456126835195989e-05, "loss": 1.9513, "step": 1994 }, { "epoch": 0.48176768896401834, "grad_norm": 0.30344778299331665, "learning_rate": 5.4522719760372435e-05, "loss": 1.779, "step": 1995 }, { "epoch": 0.4820091765274088, "grad_norm": 0.29543033242225647, "learning_rate": 5.448416845812329e-05, "loss": 1.6585, "step": 1996 }, { "epoch": 0.4822506640907993, "grad_norm": 0.29839324951171875, "learning_rate": 5.444561446831792e-05, "loss": 1.7803, "step": 1997 }, { "epoch": 0.4824921516541898, "grad_norm": 0.2827812731266022, "learning_rate": 5.440705781406341e-05, "loss": 1.7993, "step": 1998 }, { "epoch": 0.4827336392175803, "grad_norm": 0.292828768491745, "learning_rate": 5.4368498518468414e-05, "loss": 1.6683, "step": 1999 }, { "epoch": 0.48297512678097076, "grad_norm": 0.29965028166770935, "learning_rate": 5.4329936604643216e-05, "loss": 1.7802, "step": 2000 }, { "epoch": 0.48321661434436125, "grad_norm": 0.30213335156440735, "learning_rate": 5.429137209569962e-05, "loss": 1.7504, "step": 2001 }, { "epoch": 0.48345810190775174, "grad_norm": 0.31635966897010803, "learning_rate": 5.425280501475101e-05, "loss": 1.8073, "step": 2002 }, { "epoch": 0.4836995894711422, "grad_norm": 0.2913029193878174, "learning_rate": 5.4214235384912306e-05, "loss": 1.7328, "step": 2003 }, { "epoch": 0.4839410770345327, "grad_norm": 0.2901521325111389, "learning_rate": 5.4175663229299946e-05, "loss": 1.6408, "step": 2004 }, { "epoch": 0.4841825645979232, "grad_norm": 0.2936553955078125, "learning_rate": 5.4137088571031904e-05, "loss": 1.6108, "step": 2005 }, { "epoch": 0.4844240521613137, "grad_norm": 0.2733401954174042, "learning_rate": 5.409851143322764e-05, "loss": 1.5933, "step": 2006 }, { "epoch": 0.48466553972470416, "grad_norm": 0.3041709065437317, "learning_rate": 5.4059931839008095e-05, "loss": 1.7972, "step": 2007 }, { "epoch": 0.48490702728809465, "grad_norm": 0.2946314215660095, "learning_rate": 5.402134981149569e-05, "loss": 1.7429, "step": 2008 }, { "epoch": 0.48514851485148514, "grad_norm": 0.27985838055610657, "learning_rate": 5.3982765373814334e-05, "loss": 1.5364, "step": 2009 }, { "epoch": 0.4853900024148756, "grad_norm": 0.2916784882545471, "learning_rate": 5.394417854908932e-05, "loss": 1.7623, "step": 2010 }, { "epoch": 0.4856314899782661, "grad_norm": 0.27369216084480286, "learning_rate": 5.3905589360447394e-05, "loss": 1.6255, "step": 2011 }, { "epoch": 0.4858729775416566, "grad_norm": 0.29541531205177307, "learning_rate": 5.386699783101675e-05, "loss": 1.707, "step": 2012 }, { "epoch": 0.4861144651050471, "grad_norm": 0.30250734090805054, "learning_rate": 5.3828403983926966e-05, "loss": 1.8212, "step": 2013 }, { "epoch": 0.48635595266843756, "grad_norm": 0.2893443703651428, "learning_rate": 5.378980784230897e-05, "loss": 1.7433, "step": 2014 }, { "epoch": 0.48659744023182805, "grad_norm": 0.3142542541027069, "learning_rate": 5.3751209429295135e-05, "loss": 1.7603, "step": 2015 }, { "epoch": 0.48683892779521853, "grad_norm": 0.2962985634803772, "learning_rate": 5.3712608768019146e-05, "loss": 1.6304, "step": 2016 }, { "epoch": 0.487080415358609, "grad_norm": 0.28009456396102905, "learning_rate": 5.3674005881616044e-05, "loss": 1.4477, "step": 2017 }, { "epoch": 0.4873219029219995, "grad_norm": 0.2999042868614197, "learning_rate": 5.36354007932222e-05, "loss": 1.819, "step": 2018 }, { "epoch": 0.48756339048539, "grad_norm": 0.32176443934440613, "learning_rate": 5.359679352597534e-05, "loss": 1.671, "step": 2019 }, { "epoch": 0.4878048780487805, "grad_norm": 0.2954274117946625, "learning_rate": 5.3558184103014475e-05, "loss": 1.6826, "step": 2020 }, { "epoch": 0.48804636561217096, "grad_norm": 0.2915526032447815, "learning_rate": 5.351957254747989e-05, "loss": 1.7633, "step": 2021 }, { "epoch": 0.48828785317556145, "grad_norm": 0.2885175943374634, "learning_rate": 5.348095888251314e-05, "loss": 1.8001, "step": 2022 }, { "epoch": 0.48852934073895193, "grad_norm": 0.29986679553985596, "learning_rate": 5.344234313125711e-05, "loss": 1.7582, "step": 2023 }, { "epoch": 0.4887708283023424, "grad_norm": 0.28797924518585205, "learning_rate": 5.3403725316855856e-05, "loss": 1.6822, "step": 2024 }, { "epoch": 0.4890123158657329, "grad_norm": 0.2786951959133148, "learning_rate": 5.336510546245471e-05, "loss": 1.5085, "step": 2025 }, { "epoch": 0.4892538034291234, "grad_norm": 0.3011576533317566, "learning_rate": 5.332648359120024e-05, "loss": 1.8847, "step": 2026 }, { "epoch": 0.4894952909925139, "grad_norm": 0.2964388132095337, "learning_rate": 5.3287859726240184e-05, "loss": 1.7667, "step": 2027 }, { "epoch": 0.48973677855590436, "grad_norm": 0.28657767176628113, "learning_rate": 5.324923389072349e-05, "loss": 1.7107, "step": 2028 }, { "epoch": 0.48997826611929485, "grad_norm": 0.3085033893585205, "learning_rate": 5.321060610780033e-05, "loss": 1.8535, "step": 2029 }, { "epoch": 0.49021975368268533, "grad_norm": 0.31247368454933167, "learning_rate": 5.317197640062197e-05, "loss": 1.7139, "step": 2030 }, { "epoch": 0.4904612412460758, "grad_norm": 0.2796979546546936, "learning_rate": 5.313334479234088e-05, "loss": 1.6217, "step": 2031 }, { "epoch": 0.4907027288094663, "grad_norm": 0.32081031799316406, "learning_rate": 5.3094711306110655e-05, "loss": 1.9071, "step": 2032 }, { "epoch": 0.4909442163728568, "grad_norm": 0.3371107876300812, "learning_rate": 5.3056075965086016e-05, "loss": 1.9733, "step": 2033 }, { "epoch": 0.4911857039362473, "grad_norm": 0.29537174105644226, "learning_rate": 5.301743879242279e-05, "loss": 1.729, "step": 2034 }, { "epoch": 0.49142719149963776, "grad_norm": 0.2830136716365814, "learning_rate": 5.297879981127791e-05, "loss": 1.6829, "step": 2035 }, { "epoch": 0.49166867906302825, "grad_norm": 0.2872219681739807, "learning_rate": 5.294015904480939e-05, "loss": 1.602, "step": 2036 }, { "epoch": 0.49191016662641873, "grad_norm": 0.2806476652622223, "learning_rate": 5.290151651617631e-05, "loss": 1.6547, "step": 2037 }, { "epoch": 0.4921516541898092, "grad_norm": 0.28836601972579956, "learning_rate": 5.286287224853882e-05, "loss": 1.6271, "step": 2038 }, { "epoch": 0.4923931417531997, "grad_norm": 0.28855109214782715, "learning_rate": 5.282422626505808e-05, "loss": 1.7132, "step": 2039 }, { "epoch": 0.4926346293165902, "grad_norm": 0.275304913520813, "learning_rate": 5.278557858889632e-05, "loss": 1.3799, "step": 2040 }, { "epoch": 0.4928761168799807, "grad_norm": 0.282404363155365, "learning_rate": 5.274692924321676e-05, "loss": 1.7319, "step": 2041 }, { "epoch": 0.49311760444337116, "grad_norm": 0.3073646128177643, "learning_rate": 5.270827825118363e-05, "loss": 1.8026, "step": 2042 }, { "epoch": 0.49335909200676165, "grad_norm": 0.2814038395881653, "learning_rate": 5.2669625635962125e-05, "loss": 1.4985, "step": 2043 }, { "epoch": 0.49360057957015213, "grad_norm": 0.304783433675766, "learning_rate": 5.263097142071846e-05, "loss": 1.6709, "step": 2044 }, { "epoch": 0.4938420671335426, "grad_norm": 0.3093326985836029, "learning_rate": 5.259231562861975e-05, "loss": 1.7919, "step": 2045 }, { "epoch": 0.4940835546969331, "grad_norm": 0.29553940892219543, "learning_rate": 5.2553658282834096e-05, "loss": 1.7418, "step": 2046 }, { "epoch": 0.4943250422603236, "grad_norm": 0.28401437401771545, "learning_rate": 5.251499940653052e-05, "loss": 1.528, "step": 2047 }, { "epoch": 0.4945665298237141, "grad_norm": 0.2857573926448822, "learning_rate": 5.247633902287896e-05, "loss": 1.7674, "step": 2048 }, { "epoch": 0.49480801738710456, "grad_norm": 0.2939291298389435, "learning_rate": 5.2437677155050246e-05, "loss": 1.7797, "step": 2049 }, { "epoch": 0.49504950495049505, "grad_norm": 0.294514924287796, "learning_rate": 5.239901382621612e-05, "loss": 1.713, "step": 2050 }, { "epoch": 0.49529099251388553, "grad_norm": 0.29818663001060486, "learning_rate": 5.236034905954918e-05, "loss": 1.62, "step": 2051 }, { "epoch": 0.495532480077276, "grad_norm": 0.2945563793182373, "learning_rate": 5.23216828782229e-05, "loss": 1.7415, "step": 2052 }, { "epoch": 0.4957739676406665, "grad_norm": 0.28991132974624634, "learning_rate": 5.228301530541161e-05, "loss": 1.7203, "step": 2053 }, { "epoch": 0.496015455204057, "grad_norm": 0.307682067155838, "learning_rate": 5.224434636429043e-05, "loss": 1.8736, "step": 2054 }, { "epoch": 0.4962569427674475, "grad_norm": 0.29729771614074707, "learning_rate": 5.220567607803535e-05, "loss": 1.7664, "step": 2055 }, { "epoch": 0.49649843033083796, "grad_norm": 0.2977916896343231, "learning_rate": 5.216700446982314e-05, "loss": 1.7176, "step": 2056 }, { "epoch": 0.49673991789422844, "grad_norm": 0.31754955649375916, "learning_rate": 5.2128331562831376e-05, "loss": 1.8952, "step": 2057 }, { "epoch": 0.49698140545761893, "grad_norm": 0.2777542173862457, "learning_rate": 5.208965738023839e-05, "loss": 1.6587, "step": 2058 }, { "epoch": 0.4972228930210094, "grad_norm": 0.30027759075164795, "learning_rate": 5.205098194522331e-05, "loss": 1.7277, "step": 2059 }, { "epoch": 0.4974643805843999, "grad_norm": 0.2900436222553253, "learning_rate": 5.2012305280965976e-05, "loss": 1.7419, "step": 2060 }, { "epoch": 0.4977058681477904, "grad_norm": 0.3102516233921051, "learning_rate": 5.197362741064701e-05, "loss": 1.892, "step": 2061 }, { "epoch": 0.4979473557111809, "grad_norm": 0.31957629323005676, "learning_rate": 5.193494835744771e-05, "loss": 1.8064, "step": 2062 }, { "epoch": 0.49818884327457136, "grad_norm": 0.28344500064849854, "learning_rate": 5.189626814455013e-05, "loss": 1.6437, "step": 2063 }, { "epoch": 0.49843033083796184, "grad_norm": 0.2863394618034363, "learning_rate": 5.185758679513698e-05, "loss": 1.5393, "step": 2064 }, { "epoch": 0.49867181840135233, "grad_norm": 0.30922794342041016, "learning_rate": 5.1818904332391684e-05, "loss": 1.7618, "step": 2065 }, { "epoch": 0.4989133059647428, "grad_norm": 0.515668511390686, "learning_rate": 5.17802207794983e-05, "loss": 1.8118, "step": 2066 }, { "epoch": 0.4991547935281333, "grad_norm": 0.2919605076313019, "learning_rate": 5.174153615964156e-05, "loss": 1.7037, "step": 2067 }, { "epoch": 0.4993962810915238, "grad_norm": 0.31465280055999756, "learning_rate": 5.1702850496006826e-05, "loss": 1.959, "step": 2068 }, { "epoch": 0.49963776865491427, "grad_norm": 0.31834542751312256, "learning_rate": 5.16641638117801e-05, "loss": 1.7764, "step": 2069 }, { "epoch": 0.49987925621830476, "grad_norm": 0.2853394150733948, "learning_rate": 5.162547613014798e-05, "loss": 1.7564, "step": 2070 }, { "epoch": 0.5001207437816952, "grad_norm": 0.3017140328884125, "learning_rate": 5.158678747429766e-05, "loss": 1.8533, "step": 2071 }, { "epoch": 0.5003622313450857, "grad_norm": 0.29292136430740356, "learning_rate": 5.154809786741696e-05, "loss": 1.6889, "step": 2072 }, { "epoch": 0.5006037189084762, "grad_norm": 0.2969605028629303, "learning_rate": 5.1509407332694183e-05, "loss": 1.6546, "step": 2073 }, { "epoch": 0.5008452064718667, "grad_norm": 0.28881746530532837, "learning_rate": 5.1470715893318266e-05, "loss": 1.7994, "step": 2074 }, { "epoch": 0.5010866940352572, "grad_norm": 0.29532918334007263, "learning_rate": 5.1432023572478684e-05, "loss": 1.6729, "step": 2075 }, { "epoch": 0.5013281815986477, "grad_norm": 0.31070125102996826, "learning_rate": 5.139333039336539e-05, "loss": 1.8329, "step": 2076 }, { "epoch": 0.5015696691620382, "grad_norm": 0.29049959778785706, "learning_rate": 5.1354636379168894e-05, "loss": 1.7358, "step": 2077 }, { "epoch": 0.5018111567254286, "grad_norm": 0.2849730849266052, "learning_rate": 5.1315941553080205e-05, "loss": 1.8501, "step": 2078 }, { "epoch": 0.5020526442888191, "grad_norm": 0.2853429913520813, "learning_rate": 5.127724593829077e-05, "loss": 1.639, "step": 2079 }, { "epoch": 0.5022941318522096, "grad_norm": 0.3187100887298584, "learning_rate": 5.1238549557992585e-05, "loss": 2.1625, "step": 2080 }, { "epoch": 0.5025356194156001, "grad_norm": 0.2928568124771118, "learning_rate": 5.119985243537804e-05, "loss": 1.7106, "step": 2081 }, { "epoch": 0.5027771069789906, "grad_norm": 0.2922608256340027, "learning_rate": 5.116115459364001e-05, "loss": 1.7444, "step": 2082 }, { "epoch": 0.5030185945423811, "grad_norm": 0.3103829324245453, "learning_rate": 5.112245605597178e-05, "loss": 1.8411, "step": 2083 }, { "epoch": 0.5032600821057716, "grad_norm": 0.30781078338623047, "learning_rate": 5.108375684556705e-05, "loss": 1.8777, "step": 2084 }, { "epoch": 0.503501569669162, "grad_norm": 0.2837512195110321, "learning_rate": 5.104505698561994e-05, "loss": 1.7237, "step": 2085 }, { "epoch": 0.5037430572325525, "grad_norm": 0.2791733741760254, "learning_rate": 5.100635649932497e-05, "loss": 1.637, "step": 2086 }, { "epoch": 0.503984544795943, "grad_norm": 0.2833278477191925, "learning_rate": 5.096765540987698e-05, "loss": 1.7087, "step": 2087 }, { "epoch": 0.5042260323593335, "grad_norm": 0.2958666682243347, "learning_rate": 5.092895374047123e-05, "loss": 1.7915, "step": 2088 }, { "epoch": 0.504467519922724, "grad_norm": 0.2936709225177765, "learning_rate": 5.089025151430331e-05, "loss": 1.6704, "step": 2089 }, { "epoch": 0.5047090074861145, "grad_norm": 0.3201253414154053, "learning_rate": 5.0851548754569114e-05, "loss": 1.8818, "step": 2090 }, { "epoch": 0.504950495049505, "grad_norm": 0.3030899465084076, "learning_rate": 5.081284548446491e-05, "loss": 1.791, "step": 2091 }, { "epoch": 0.5051919826128954, "grad_norm": 0.2952651381492615, "learning_rate": 5.077414172718723e-05, "loss": 1.6815, "step": 2092 }, { "epoch": 0.5054334701762859, "grad_norm": 0.2999666929244995, "learning_rate": 5.073543750593289e-05, "loss": 1.6372, "step": 2093 }, { "epoch": 0.5056749577396764, "grad_norm": 0.47690799832344055, "learning_rate": 5.069673284389904e-05, "loss": 1.6216, "step": 2094 }, { "epoch": 0.5059164453030669, "grad_norm": 0.2998300790786743, "learning_rate": 5.065802776428304e-05, "loss": 1.8669, "step": 2095 }, { "epoch": 0.5061579328664574, "grad_norm": 0.2994838058948517, "learning_rate": 5.061932229028253e-05, "loss": 1.8474, "step": 2096 }, { "epoch": 0.5063994204298479, "grad_norm": 0.31959816813468933, "learning_rate": 5.058061644509536e-05, "loss": 1.8936, "step": 2097 }, { "epoch": 0.5066409079932384, "grad_norm": 0.2968556880950928, "learning_rate": 5.054191025191963e-05, "loss": 1.6705, "step": 2098 }, { "epoch": 0.5068823955566288, "grad_norm": 0.28798797726631165, "learning_rate": 5.050320373395364e-05, "loss": 1.7489, "step": 2099 }, { "epoch": 0.5071238831200193, "grad_norm": 0.3115372359752655, "learning_rate": 5.046449691439588e-05, "loss": 1.8516, "step": 2100 }, { "epoch": 0.5073653706834098, "grad_norm": 0.30434104800224304, "learning_rate": 5.0425789816445e-05, "loss": 1.7521, "step": 2101 }, { "epoch": 0.5076068582468003, "grad_norm": 0.3264348804950714, "learning_rate": 5.0387082463299864e-05, "loss": 1.8269, "step": 2102 }, { "epoch": 0.5078483458101908, "grad_norm": 0.2886433005332947, "learning_rate": 5.034837487815945e-05, "loss": 1.5941, "step": 2103 }, { "epoch": 0.5080898333735813, "grad_norm": 0.2878704071044922, "learning_rate": 5.030966708422291e-05, "loss": 1.553, "step": 2104 }, { "epoch": 0.5083313209369718, "grad_norm": 0.2795392572879791, "learning_rate": 5.027095910468947e-05, "loss": 1.5734, "step": 2105 }, { "epoch": 0.5085728085003622, "grad_norm": 0.2933345437049866, "learning_rate": 5.023225096275852e-05, "loss": 1.7157, "step": 2106 }, { "epoch": 0.5088142960637527, "grad_norm": 0.3096478581428528, "learning_rate": 5.019354268162951e-05, "loss": 1.7194, "step": 2107 }, { "epoch": 0.5090557836271432, "grad_norm": 0.3165113627910614, "learning_rate": 5.015483428450199e-05, "loss": 1.7633, "step": 2108 }, { "epoch": 0.5092972711905337, "grad_norm": 0.3086087703704834, "learning_rate": 5.0116125794575595e-05, "loss": 1.8371, "step": 2109 }, { "epoch": 0.5095387587539242, "grad_norm": 0.28862884640693665, "learning_rate": 5.0077417235049985e-05, "loss": 1.6616, "step": 2110 }, { "epoch": 0.5097802463173147, "grad_norm": 0.314456045627594, "learning_rate": 5.003870862912487e-05, "loss": 1.9459, "step": 2111 }, { "epoch": 0.5100217338807052, "grad_norm": 0.3061044216156006, "learning_rate": 5e-05, "loss": 1.8326, "step": 2112 }, { "epoch": 0.5102632214440956, "grad_norm": 0.29353898763656616, "learning_rate": 4.996129137087513e-05, "loss": 1.6994, "step": 2113 }, { "epoch": 0.5105047090074861, "grad_norm": 0.2928066551685333, "learning_rate": 4.992258276495003e-05, "loss": 1.7293, "step": 2114 }, { "epoch": 0.5107461965708766, "grad_norm": 0.2927094101905823, "learning_rate": 4.9883874205424417e-05, "loss": 1.7618, "step": 2115 }, { "epoch": 0.5109876841342671, "grad_norm": 0.2999373972415924, "learning_rate": 4.984516571549801e-05, "loss": 1.835, "step": 2116 }, { "epoch": 0.5112291716976576, "grad_norm": 0.2926490008831024, "learning_rate": 4.98064573183705e-05, "loss": 1.6406, "step": 2117 }, { "epoch": 0.5114706592610481, "grad_norm": 0.2959551513195038, "learning_rate": 4.976774903724149e-05, "loss": 1.7653, "step": 2118 }, { "epoch": 0.5117121468244386, "grad_norm": 0.3198218047618866, "learning_rate": 4.972904089531054e-05, "loss": 1.7139, "step": 2119 }, { "epoch": 0.511953634387829, "grad_norm": 0.2835996747016907, "learning_rate": 4.969033291577711e-05, "loss": 1.7756, "step": 2120 }, { "epoch": 0.5121951219512195, "grad_norm": 0.28976184129714966, "learning_rate": 4.965162512184056e-05, "loss": 1.7377, "step": 2121 }, { "epoch": 0.51243660951461, "grad_norm": 0.2830486595630646, "learning_rate": 4.961291753670014e-05, "loss": 1.7808, "step": 2122 }, { "epoch": 0.5126780970780005, "grad_norm": 0.2898695170879364, "learning_rate": 4.957421018355502e-05, "loss": 1.7918, "step": 2123 }, { "epoch": 0.512919584641391, "grad_norm": 0.27601760625839233, "learning_rate": 4.953550308560414e-05, "loss": 1.6371, "step": 2124 }, { "epoch": 0.5131610722047815, "grad_norm": 0.2738933563232422, "learning_rate": 4.9496796266046367e-05, "loss": 1.6111, "step": 2125 }, { "epoch": 0.513402559768172, "grad_norm": 0.2990981340408325, "learning_rate": 4.945808974808038e-05, "loss": 1.7059, "step": 2126 }, { "epoch": 0.5136440473315624, "grad_norm": 0.2851322293281555, "learning_rate": 4.941938355490465e-05, "loss": 1.6349, "step": 2127 }, { "epoch": 0.5138855348949529, "grad_norm": 0.28323864936828613, "learning_rate": 4.938067770971748e-05, "loss": 1.6369, "step": 2128 }, { "epoch": 0.5141270224583434, "grad_norm": 0.2917851507663727, "learning_rate": 4.934197223571697e-05, "loss": 1.8279, "step": 2129 }, { "epoch": 0.5143685100217339, "grad_norm": 0.2805134356021881, "learning_rate": 4.9303267156100965e-05, "loss": 1.7407, "step": 2130 }, { "epoch": 0.5146099975851244, "grad_norm": 0.29183030128479004, "learning_rate": 4.926456249406711e-05, "loss": 1.7376, "step": 2131 }, { "epoch": 0.5148514851485149, "grad_norm": 0.2907525599002838, "learning_rate": 4.9225858272812796e-05, "loss": 1.7321, "step": 2132 }, { "epoch": 0.5150929727119054, "grad_norm": 0.2907441556453705, "learning_rate": 4.918715451553511e-05, "loss": 1.8091, "step": 2133 }, { "epoch": 0.5153344602752958, "grad_norm": 0.2903591990470886, "learning_rate": 4.9148451245430885e-05, "loss": 1.6849, "step": 2134 }, { "epoch": 0.5155759478386863, "grad_norm": 0.3057684004306793, "learning_rate": 4.9109748485696715e-05, "loss": 1.7193, "step": 2135 }, { "epoch": 0.5158174354020768, "grad_norm": 0.2820757031440735, "learning_rate": 4.907104625952878e-05, "loss": 1.6808, "step": 2136 }, { "epoch": 0.5160589229654673, "grad_norm": 0.28923845291137695, "learning_rate": 4.9032344590123025e-05, "loss": 1.7343, "step": 2137 }, { "epoch": 0.5163004105288578, "grad_norm": 0.2927323579788208, "learning_rate": 4.8993643500675045e-05, "loss": 1.7495, "step": 2138 }, { "epoch": 0.5165418980922483, "grad_norm": 0.32093602418899536, "learning_rate": 4.8954943014380064e-05, "loss": 1.7346, "step": 2139 }, { "epoch": 0.5167833856556387, "grad_norm": 0.28218767046928406, "learning_rate": 4.891624315443296e-05, "loss": 1.7227, "step": 2140 }, { "epoch": 0.5170248732190292, "grad_norm": 0.2791909873485565, "learning_rate": 4.887754394402824e-05, "loss": 1.5317, "step": 2141 }, { "epoch": 0.5172663607824197, "grad_norm": 0.2936622202396393, "learning_rate": 4.883884540636e-05, "loss": 1.7714, "step": 2142 }, { "epoch": 0.5175078483458102, "grad_norm": 0.2987155318260193, "learning_rate": 4.880014756462196e-05, "loss": 1.8333, "step": 2143 }, { "epoch": 0.5177493359092007, "grad_norm": 0.29716813564300537, "learning_rate": 4.876145044200744e-05, "loss": 1.7744, "step": 2144 }, { "epoch": 0.5179908234725912, "grad_norm": 0.28382810950279236, "learning_rate": 4.872275406170924e-05, "loss": 1.7451, "step": 2145 }, { "epoch": 0.5182323110359817, "grad_norm": 0.30451273918151855, "learning_rate": 4.868405844691981e-05, "loss": 1.8383, "step": 2146 }, { "epoch": 0.5184737985993721, "grad_norm": 0.27715468406677246, "learning_rate": 4.8645363620831124e-05, "loss": 1.5774, "step": 2147 }, { "epoch": 0.5187152861627626, "grad_norm": 0.2944760024547577, "learning_rate": 4.860666960663463e-05, "loss": 1.6889, "step": 2148 }, { "epoch": 0.5189567737261531, "grad_norm": 0.2942114472389221, "learning_rate": 4.856797642752133e-05, "loss": 1.7268, "step": 2149 }, { "epoch": 0.5191982612895436, "grad_norm": 0.29451313614845276, "learning_rate": 4.8529284106681746e-05, "loss": 1.7929, "step": 2150 }, { "epoch": 0.5194397488529341, "grad_norm": 0.32748252153396606, "learning_rate": 4.8490592667305835e-05, "loss": 1.9073, "step": 2151 }, { "epoch": 0.5196812364163246, "grad_norm": 0.29063066840171814, "learning_rate": 4.845190213258306e-05, "loss": 1.7321, "step": 2152 }, { "epoch": 0.5199227239797151, "grad_norm": 0.29514726996421814, "learning_rate": 4.841321252570236e-05, "loss": 1.6949, "step": 2153 }, { "epoch": 0.5201642115431055, "grad_norm": 0.28942957520484924, "learning_rate": 4.837452386985204e-05, "loss": 1.6283, "step": 2154 }, { "epoch": 0.520405699106496, "grad_norm": 0.29086393117904663, "learning_rate": 4.833583618821991e-05, "loss": 1.6983, "step": 2155 }, { "epoch": 0.5206471866698865, "grad_norm": 0.2910321354866028, "learning_rate": 4.829714950399319e-05, "loss": 1.6577, "step": 2156 }, { "epoch": 0.520888674233277, "grad_norm": 0.28043949604034424, "learning_rate": 4.825846384035846e-05, "loss": 1.5504, "step": 2157 }, { "epoch": 0.5211301617966675, "grad_norm": 0.2987416684627533, "learning_rate": 4.821977922050171e-05, "loss": 1.7115, "step": 2158 }, { "epoch": 0.521371649360058, "grad_norm": 0.29055726528167725, "learning_rate": 4.818109566760834e-05, "loss": 1.6575, "step": 2159 }, { "epoch": 0.5216131369234485, "grad_norm": 0.3061610460281372, "learning_rate": 4.814241320486303e-05, "loss": 1.797, "step": 2160 }, { "epoch": 0.521854624486839, "grad_norm": 0.2867301106452942, "learning_rate": 4.810373185544988e-05, "loss": 1.7559, "step": 2161 }, { "epoch": 0.5220961120502294, "grad_norm": 0.2962303161621094, "learning_rate": 4.80650516425523e-05, "loss": 1.6002, "step": 2162 }, { "epoch": 0.5223375996136199, "grad_norm": 0.29246842861175537, "learning_rate": 4.802637258935301e-05, "loss": 1.7973, "step": 2163 }, { "epoch": 0.5225790871770104, "grad_norm": 0.3058999180793762, "learning_rate": 4.798769471903403e-05, "loss": 1.8058, "step": 2164 }, { "epoch": 0.5228205747404009, "grad_norm": 0.3095049262046814, "learning_rate": 4.794901805477672e-05, "loss": 1.726, "step": 2165 }, { "epoch": 0.5230620623037914, "grad_norm": 0.27616173028945923, "learning_rate": 4.791034261976162e-05, "loss": 1.4841, "step": 2166 }, { "epoch": 0.5233035498671819, "grad_norm": 0.29235804080963135, "learning_rate": 4.787166843716863e-05, "loss": 1.6561, "step": 2167 }, { "epoch": 0.5235450374305723, "grad_norm": 0.3048040568828583, "learning_rate": 4.7832995530176875e-05, "loss": 1.7102, "step": 2168 }, { "epoch": 0.5237865249939628, "grad_norm": 0.29728803038597107, "learning_rate": 4.779432392196466e-05, "loss": 1.7071, "step": 2169 }, { "epoch": 0.5240280125573533, "grad_norm": 0.283037006855011, "learning_rate": 4.7755653635709576e-05, "loss": 1.6172, "step": 2170 }, { "epoch": 0.5242695001207438, "grad_norm": 0.29151520133018494, "learning_rate": 4.771698469458841e-05, "loss": 1.6439, "step": 2171 }, { "epoch": 0.5245109876841343, "grad_norm": 0.29279422760009766, "learning_rate": 4.7678317121777114e-05, "loss": 1.6553, "step": 2172 }, { "epoch": 0.5247524752475248, "grad_norm": 0.3101252317428589, "learning_rate": 4.763965094045083e-05, "loss": 1.747, "step": 2173 }, { "epoch": 0.5249939628109153, "grad_norm": 0.309313029050827, "learning_rate": 4.76009861737839e-05, "loss": 2.0366, "step": 2174 }, { "epoch": 0.5252354503743057, "grad_norm": 0.3047976493835449, "learning_rate": 4.756232284494977e-05, "loss": 1.8299, "step": 2175 }, { "epoch": 0.5254769379376962, "grad_norm": 0.29255151748657227, "learning_rate": 4.752366097712105e-05, "loss": 1.7837, "step": 2176 }, { "epoch": 0.5257184255010867, "grad_norm": 0.28311216831207275, "learning_rate": 4.748500059346948e-05, "loss": 1.8098, "step": 2177 }, { "epoch": 0.5259599130644772, "grad_norm": 0.30000126361846924, "learning_rate": 4.744634171716592e-05, "loss": 1.6604, "step": 2178 }, { "epoch": 0.5262014006278677, "grad_norm": 0.31336405873298645, "learning_rate": 4.7407684371380254e-05, "loss": 1.7967, "step": 2179 }, { "epoch": 0.5264428881912582, "grad_norm": 0.28911155462265015, "learning_rate": 4.7369028579281547e-05, "loss": 1.7033, "step": 2180 }, { "epoch": 0.5266843757546487, "grad_norm": 0.29213377833366394, "learning_rate": 4.733037436403788e-05, "loss": 1.8059, "step": 2181 }, { "epoch": 0.5269258633180391, "grad_norm": 0.2918547987937927, "learning_rate": 4.7291721748816384e-05, "loss": 1.7084, "step": 2182 }, { "epoch": 0.5271673508814296, "grad_norm": 0.30548644065856934, "learning_rate": 4.725307075678324e-05, "loss": 1.7293, "step": 2183 }, { "epoch": 0.5274088384448201, "grad_norm": 0.2997039556503296, "learning_rate": 4.721442141110369e-05, "loss": 1.7372, "step": 2184 }, { "epoch": 0.5276503260082106, "grad_norm": 0.29757243394851685, "learning_rate": 4.717577373494192e-05, "loss": 1.815, "step": 2185 }, { "epoch": 0.5278918135716011, "grad_norm": 0.2895524799823761, "learning_rate": 4.71371277514612e-05, "loss": 1.641, "step": 2186 }, { "epoch": 0.5281333011349916, "grad_norm": 0.27370625734329224, "learning_rate": 4.7098483483823705e-05, "loss": 1.5722, "step": 2187 }, { "epoch": 0.5283747886983821, "grad_norm": 0.28297242522239685, "learning_rate": 4.7059840955190614e-05, "loss": 1.6656, "step": 2188 }, { "epoch": 0.5286162762617725, "grad_norm": 0.30684277415275574, "learning_rate": 4.702120018872209e-05, "loss": 1.8562, "step": 2189 }, { "epoch": 0.528857763825163, "grad_norm": 0.28609248995780945, "learning_rate": 4.698256120757723e-05, "loss": 1.674, "step": 2190 }, { "epoch": 0.5290992513885535, "grad_norm": 0.3094319701194763, "learning_rate": 4.6943924034913995e-05, "loss": 1.9017, "step": 2191 }, { "epoch": 0.529340738951944, "grad_norm": 0.31621307134628296, "learning_rate": 4.690528869388934e-05, "loss": 1.9076, "step": 2192 }, { "epoch": 0.5295822265153345, "grad_norm": 0.33111467957496643, "learning_rate": 4.686665520765914e-05, "loss": 1.8098, "step": 2193 }, { "epoch": 0.529823714078725, "grad_norm": 0.28939831256866455, "learning_rate": 4.6828023599378046e-05, "loss": 1.7295, "step": 2194 }, { "epoch": 0.5300652016421155, "grad_norm": 0.35473617911338806, "learning_rate": 4.678939389219968e-05, "loss": 1.7657, "step": 2195 }, { "epoch": 0.5303066892055059, "grad_norm": 0.29074251651763916, "learning_rate": 4.675076610927652e-05, "loss": 1.6936, "step": 2196 }, { "epoch": 0.5305481767688964, "grad_norm": 0.28574123978614807, "learning_rate": 4.671214027375983e-05, "loss": 1.6916, "step": 2197 }, { "epoch": 0.5307896643322869, "grad_norm": 0.26639223098754883, "learning_rate": 4.667351640879977e-05, "loss": 1.5951, "step": 2198 }, { "epoch": 0.5310311518956774, "grad_norm": 0.28303614258766174, "learning_rate": 4.66348945375453e-05, "loss": 1.6207, "step": 2199 }, { "epoch": 0.5312726394590679, "grad_norm": 0.3010537028312683, "learning_rate": 4.6596274683144156e-05, "loss": 1.6644, "step": 2200 }, { "epoch": 0.5315141270224584, "grad_norm": 0.3192451298236847, "learning_rate": 4.6557656868742886e-05, "loss": 1.9587, "step": 2201 }, { "epoch": 0.5317556145858489, "grad_norm": 0.29769235849380493, "learning_rate": 4.651904111748686e-05, "loss": 1.7913, "step": 2202 }, { "epoch": 0.5319971021492393, "grad_norm": 0.29097649455070496, "learning_rate": 4.648042745252012e-05, "loss": 1.6633, "step": 2203 }, { "epoch": 0.5322385897126298, "grad_norm": 0.2759608030319214, "learning_rate": 4.644181589698552e-05, "loss": 1.7795, "step": 2204 }, { "epoch": 0.5324800772760203, "grad_norm": 0.2948470413684845, "learning_rate": 4.640320647402466e-05, "loss": 1.7897, "step": 2205 }, { "epoch": 0.5327215648394108, "grad_norm": 0.29832395911216736, "learning_rate": 4.63645992067778e-05, "loss": 1.7564, "step": 2206 }, { "epoch": 0.5329630524028013, "grad_norm": 0.2934744358062744, "learning_rate": 4.6325994118383974e-05, "loss": 1.7947, "step": 2207 }, { "epoch": 0.5332045399661918, "grad_norm": 0.2907322645187378, "learning_rate": 4.628739123198088e-05, "loss": 1.8366, "step": 2208 }, { "epoch": 0.5334460275295823, "grad_norm": 0.2789422571659088, "learning_rate": 4.6248790570704883e-05, "loss": 1.7147, "step": 2209 }, { "epoch": 0.5336875150929727, "grad_norm": 0.2942877411842346, "learning_rate": 4.621019215769103e-05, "loss": 1.6951, "step": 2210 }, { "epoch": 0.5339290026563632, "grad_norm": 0.30500897765159607, "learning_rate": 4.617159601607306e-05, "loss": 1.7644, "step": 2211 }, { "epoch": 0.5341704902197537, "grad_norm": 0.27835962176322937, "learning_rate": 4.6133002168983255e-05, "loss": 1.623, "step": 2212 }, { "epoch": 0.5344119777831442, "grad_norm": 0.30493804812431335, "learning_rate": 4.6094410639552604e-05, "loss": 1.8388, "step": 2213 }, { "epoch": 0.5346534653465347, "grad_norm": 0.2895996570587158, "learning_rate": 4.60558214509107e-05, "loss": 1.7834, "step": 2214 }, { "epoch": 0.5348949529099252, "grad_norm": 0.31576910614967346, "learning_rate": 4.601723462618568e-05, "loss": 1.8653, "step": 2215 }, { "epoch": 0.5351364404733157, "grad_norm": 0.2853288948535919, "learning_rate": 4.59786501885043e-05, "loss": 1.6741, "step": 2216 }, { "epoch": 0.5353779280367061, "grad_norm": 0.2919026017189026, "learning_rate": 4.594006816099192e-05, "loss": 1.6253, "step": 2217 }, { "epoch": 0.5356194156000966, "grad_norm": 0.2977089583873749, "learning_rate": 4.590148856677237e-05, "loss": 1.6418, "step": 2218 }, { "epoch": 0.5358609031634871, "grad_norm": 0.2856467366218567, "learning_rate": 4.5862911428968114e-05, "loss": 1.688, "step": 2219 }, { "epoch": 0.5361023907268776, "grad_norm": 0.3178512752056122, "learning_rate": 4.582433677070008e-05, "loss": 1.8809, "step": 2220 }, { "epoch": 0.5363438782902681, "grad_norm": 0.2868853211402893, "learning_rate": 4.578576461508771e-05, "loss": 1.7408, "step": 2221 }, { "epoch": 0.5365853658536586, "grad_norm": 0.30971699953079224, "learning_rate": 4.5747194985249e-05, "loss": 1.8222, "step": 2222 }, { "epoch": 0.536826853417049, "grad_norm": 0.31655246019363403, "learning_rate": 4.5708627904300396e-05, "loss": 1.8578, "step": 2223 }, { "epoch": 0.5370683409804395, "grad_norm": 0.2989317774772644, "learning_rate": 4.5670063395356796e-05, "loss": 1.8581, "step": 2224 }, { "epoch": 0.53730982854383, "grad_norm": 0.2952711582183838, "learning_rate": 4.5631501481531584e-05, "loss": 1.7061, "step": 2225 }, { "epoch": 0.5375513161072205, "grad_norm": 0.29182133078575134, "learning_rate": 4.559294218593662e-05, "loss": 1.8787, "step": 2226 }, { "epoch": 0.537792803670611, "grad_norm": 0.2831118702888489, "learning_rate": 4.55543855316821e-05, "loss": 1.5233, "step": 2227 }, { "epoch": 0.5380342912340015, "grad_norm": 0.2916523218154907, "learning_rate": 4.5515831541876724e-05, "loss": 1.7581, "step": 2228 }, { "epoch": 0.538275778797392, "grad_norm": 0.2807374894618988, "learning_rate": 4.5477280239627584e-05, "loss": 1.6607, "step": 2229 }, { "epoch": 0.5385172663607825, "grad_norm": 0.32455793023109436, "learning_rate": 4.5438731648040114e-05, "loss": 1.9755, "step": 2230 }, { "epoch": 0.5387587539241729, "grad_norm": 0.28726670145988464, "learning_rate": 4.540018579021817e-05, "loss": 1.7195, "step": 2231 }, { "epoch": 0.5390002414875634, "grad_norm": 0.3077503740787506, "learning_rate": 4.5361642689263965e-05, "loss": 1.9033, "step": 2232 }, { "epoch": 0.5392417290509539, "grad_norm": 0.293334424495697, "learning_rate": 4.532310236827801e-05, "loss": 1.8263, "step": 2233 }, { "epoch": 0.5394832166143444, "grad_norm": 0.28368473052978516, "learning_rate": 4.528456485035922e-05, "loss": 1.768, "step": 2234 }, { "epoch": 0.5397247041777349, "grad_norm": 0.2948681116104126, "learning_rate": 4.52460301586048e-05, "loss": 1.7278, "step": 2235 }, { "epoch": 0.5399661917411254, "grad_norm": 0.2870345115661621, "learning_rate": 4.520749831611023e-05, "loss": 1.8161, "step": 2236 }, { "epoch": 0.5402076793045159, "grad_norm": 0.300326406955719, "learning_rate": 4.5168969345969316e-05, "loss": 1.6916, "step": 2237 }, { "epoch": 0.5404491668679063, "grad_norm": 0.29715168476104736, "learning_rate": 4.513044327127416e-05, "loss": 1.8946, "step": 2238 }, { "epoch": 0.5406906544312968, "grad_norm": 0.2976692318916321, "learning_rate": 4.5091920115115094e-05, "loss": 1.8709, "step": 2239 }, { "epoch": 0.5409321419946873, "grad_norm": 0.2877989709377289, "learning_rate": 4.50533999005807e-05, "loss": 1.7004, "step": 2240 }, { "epoch": 0.5411736295580778, "grad_norm": 0.27564293146133423, "learning_rate": 4.5014882650757835e-05, "loss": 1.6473, "step": 2241 }, { "epoch": 0.5414151171214683, "grad_norm": 0.2953254282474518, "learning_rate": 4.4976368388731525e-05, "loss": 1.6382, "step": 2242 }, { "epoch": 0.5416566046848588, "grad_norm": 0.28585103154182434, "learning_rate": 4.493785713758505e-05, "loss": 1.6273, "step": 2243 }, { "epoch": 0.5418980922482493, "grad_norm": 0.2975962162017822, "learning_rate": 4.489934892039989e-05, "loss": 1.796, "step": 2244 }, { "epoch": 0.5421395798116397, "grad_norm": 0.30563226342201233, "learning_rate": 4.486084376025564e-05, "loss": 1.7896, "step": 2245 }, { "epoch": 0.5423810673750302, "grad_norm": 0.3066451847553253, "learning_rate": 4.482234168023014e-05, "loss": 1.7673, "step": 2246 }, { "epoch": 0.5426225549384207, "grad_norm": 0.2895350456237793, "learning_rate": 4.4783842703399366e-05, "loss": 1.5467, "step": 2247 }, { "epoch": 0.5428640425018112, "grad_norm": 0.2916082739830017, "learning_rate": 4.474534685283738e-05, "loss": 1.8583, "step": 2248 }, { "epoch": 0.5431055300652017, "grad_norm": 0.2968622148036957, "learning_rate": 4.470685415161643e-05, "loss": 1.7605, "step": 2249 }, { "epoch": 0.5433470176285922, "grad_norm": 0.29519784450531006, "learning_rate": 4.466836462280687e-05, "loss": 1.6436, "step": 2250 }, { "epoch": 0.5435885051919827, "grad_norm": 0.2979161739349365, "learning_rate": 4.462987828947712e-05, "loss": 1.9031, "step": 2251 }, { "epoch": 0.5438299927553731, "grad_norm": 0.35301947593688965, "learning_rate": 4.459139517469373e-05, "loss": 1.62, "step": 2252 }, { "epoch": 0.5440714803187636, "grad_norm": 0.28557130694389343, "learning_rate": 4.455291530152127e-05, "loss": 1.6449, "step": 2253 }, { "epoch": 0.5443129678821541, "grad_norm": 0.29312124848365784, "learning_rate": 4.451443869302241e-05, "loss": 1.697, "step": 2254 }, { "epoch": 0.5445544554455446, "grad_norm": 0.29386618733406067, "learning_rate": 4.447596537225783e-05, "loss": 1.8455, "step": 2255 }, { "epoch": 0.5447959430089351, "grad_norm": 0.27189183235168457, "learning_rate": 4.4437495362286294e-05, "loss": 1.6241, "step": 2256 }, { "epoch": 0.5450374305723256, "grad_norm": 0.29548728466033936, "learning_rate": 4.4399028686164493e-05, "loss": 1.6562, "step": 2257 }, { "epoch": 0.545278918135716, "grad_norm": 0.27528753876686096, "learning_rate": 4.4360565366947196e-05, "loss": 1.6809, "step": 2258 }, { "epoch": 0.5455204056991065, "grad_norm": 0.3114590644836426, "learning_rate": 4.432210542768716e-05, "loss": 1.7378, "step": 2259 }, { "epoch": 0.545761893262497, "grad_norm": 0.3026160001754761, "learning_rate": 4.428364889143504e-05, "loss": 1.772, "step": 2260 }, { "epoch": 0.5460033808258875, "grad_norm": 0.2879740595817566, "learning_rate": 4.424519578123953e-05, "loss": 1.7705, "step": 2261 }, { "epoch": 0.546244868389278, "grad_norm": 0.2831686735153198, "learning_rate": 4.420674612014724e-05, "loss": 1.7476, "step": 2262 }, { "epoch": 0.5464863559526685, "grad_norm": 0.292718768119812, "learning_rate": 4.4168299931202726e-05, "loss": 1.7048, "step": 2263 }, { "epoch": 0.546727843516059, "grad_norm": 0.3042823076248169, "learning_rate": 4.412985723744843e-05, "loss": 1.7876, "step": 2264 }, { "epoch": 0.5469693310794495, "grad_norm": 0.2998793125152588, "learning_rate": 4.409141806192476e-05, "loss": 1.6914, "step": 2265 }, { "epoch": 0.5472108186428399, "grad_norm": 0.2851142883300781, "learning_rate": 4.4052982427669934e-05, "loss": 1.5798, "step": 2266 }, { "epoch": 0.5474523062062304, "grad_norm": 0.28385090827941895, "learning_rate": 4.40145503577201e-05, "loss": 1.5925, "step": 2267 }, { "epoch": 0.5476937937696209, "grad_norm": 0.30428406596183777, "learning_rate": 4.39761218751093e-05, "loss": 1.8052, "step": 2268 }, { "epoch": 0.5479352813330114, "grad_norm": 0.27645143866539, "learning_rate": 4.393769700286934e-05, "loss": 1.6405, "step": 2269 }, { "epoch": 0.5481767688964019, "grad_norm": 0.31941384077072144, "learning_rate": 4.389927576402992e-05, "loss": 1.8355, "step": 2270 }, { "epoch": 0.5484182564597924, "grad_norm": 0.29175373911857605, "learning_rate": 4.386085818161857e-05, "loss": 1.6231, "step": 2271 }, { "epoch": 0.5486597440231828, "grad_norm": 0.3057311773300171, "learning_rate": 4.3822444278660595e-05, "loss": 1.6965, "step": 2272 }, { "epoch": 0.5489012315865733, "grad_norm": 0.3028654158115387, "learning_rate": 4.37840340781791e-05, "loss": 1.6836, "step": 2273 }, { "epoch": 0.5491427191499638, "grad_norm": 0.2839415669441223, "learning_rate": 4.374562760319501e-05, "loss": 1.676, "step": 2274 }, { "epoch": 0.5493842067133543, "grad_norm": 0.2853753864765167, "learning_rate": 4.370722487672695e-05, "loss": 1.7158, "step": 2275 }, { "epoch": 0.5496256942767448, "grad_norm": 0.283378541469574, "learning_rate": 4.366882592179136e-05, "loss": 1.6225, "step": 2276 }, { "epoch": 0.5498671818401353, "grad_norm": 0.30874744057655334, "learning_rate": 4.363043076140241e-05, "loss": 1.8377, "step": 2277 }, { "epoch": 0.5501086694035258, "grad_norm": 0.2817251980304718, "learning_rate": 4.359203941857195e-05, "loss": 1.6161, "step": 2278 }, { "epoch": 0.5503501569669162, "grad_norm": 0.2946871519088745, "learning_rate": 4.355365191630958e-05, "loss": 1.733, "step": 2279 }, { "epoch": 0.5505916445303067, "grad_norm": 0.29518142342567444, "learning_rate": 4.3515268277622617e-05, "loss": 1.7686, "step": 2280 }, { "epoch": 0.5508331320936972, "grad_norm": 0.31922781467437744, "learning_rate": 4.3476888525515996e-05, "loss": 2.1372, "step": 2281 }, { "epoch": 0.5510746196570877, "grad_norm": 0.2924897074699402, "learning_rate": 4.343851268299238e-05, "loss": 1.6698, "step": 2282 }, { "epoch": 0.5513161072204782, "grad_norm": 0.2899172306060791, "learning_rate": 4.3400140773052094e-05, "loss": 1.642, "step": 2283 }, { "epoch": 0.5515575947838687, "grad_norm": 0.3056505620479584, "learning_rate": 4.336177281869305e-05, "loss": 1.8164, "step": 2284 }, { "epoch": 0.5517990823472592, "grad_norm": 0.29910221695899963, "learning_rate": 4.332340884291085e-05, "loss": 1.7433, "step": 2285 }, { "epoch": 0.5520405699106496, "grad_norm": 0.30004215240478516, "learning_rate": 4.328504886869869e-05, "loss": 1.6854, "step": 2286 }, { "epoch": 0.5522820574740401, "grad_norm": 0.2739005386829376, "learning_rate": 4.324669291904733e-05, "loss": 1.5625, "step": 2287 }, { "epoch": 0.5525235450374306, "grad_norm": 0.27121537923812866, "learning_rate": 4.320834101694516e-05, "loss": 1.6249, "step": 2288 }, { "epoch": 0.5527650326008211, "grad_norm": 0.285295307636261, "learning_rate": 4.316999318537817e-05, "loss": 1.798, "step": 2289 }, { "epoch": 0.5530065201642116, "grad_norm": 0.29337432980537415, "learning_rate": 4.313164944732983e-05, "loss": 1.6031, "step": 2290 }, { "epoch": 0.5532480077276021, "grad_norm": 0.3011399507522583, "learning_rate": 4.309330982578122e-05, "loss": 1.8216, "step": 2291 }, { "epoch": 0.5534894952909926, "grad_norm": 0.29034000635147095, "learning_rate": 4.305497434371095e-05, "loss": 1.6317, "step": 2292 }, { "epoch": 0.553730982854383, "grad_norm": 0.2813529372215271, "learning_rate": 4.30166430240951e-05, "loss": 1.6489, "step": 2293 }, { "epoch": 0.5539724704177735, "grad_norm": 0.30264776945114136, "learning_rate": 4.2978315889907305e-05, "loss": 1.8002, "step": 2294 }, { "epoch": 0.554213957981164, "grad_norm": 0.31814688444137573, "learning_rate": 4.293999296411869e-05, "loss": 1.9815, "step": 2295 }, { "epoch": 0.5544554455445545, "grad_norm": 0.2931728661060333, "learning_rate": 4.290167426969781e-05, "loss": 1.8097, "step": 2296 }, { "epoch": 0.554696933107945, "grad_norm": 0.2793978452682495, "learning_rate": 4.286335982961074e-05, "loss": 1.6111, "step": 2297 }, { "epoch": 0.5549384206713355, "grad_norm": 0.28356504440307617, "learning_rate": 4.282504966682101e-05, "loss": 1.5934, "step": 2298 }, { "epoch": 0.555179908234726, "grad_norm": 0.2903590202331543, "learning_rate": 4.2786743804289495e-05, "loss": 1.7474, "step": 2299 }, { "epoch": 0.5554213957981164, "grad_norm": 0.31551313400268555, "learning_rate": 4.27484422649746e-05, "loss": 1.7758, "step": 2300 }, { "epoch": 0.5556628833615069, "grad_norm": 0.29820793867111206, "learning_rate": 4.271014507183209e-05, "loss": 1.6703, "step": 2301 }, { "epoch": 0.5559043709248974, "grad_norm": 0.29952242970466614, "learning_rate": 4.267185224781511e-05, "loss": 1.6916, "step": 2302 }, { "epoch": 0.5561458584882879, "grad_norm": 0.27568548917770386, "learning_rate": 4.263356381587422e-05, "loss": 1.7504, "step": 2303 }, { "epoch": 0.5563873460516784, "grad_norm": 0.26956528425216675, "learning_rate": 4.2595279798957334e-05, "loss": 1.5723, "step": 2304 }, { "epoch": 0.5566288336150689, "grad_norm": 0.2860077917575836, "learning_rate": 4.255700022000974e-05, "loss": 1.595, "step": 2305 }, { "epoch": 0.5568703211784594, "grad_norm": 0.2942974865436554, "learning_rate": 4.251872510197401e-05, "loss": 1.6774, "step": 2306 }, { "epoch": 0.5571118087418498, "grad_norm": 0.3224583864212036, "learning_rate": 4.2480454467790095e-05, "loss": 1.9213, "step": 2307 }, { "epoch": 0.5573532963052403, "grad_norm": 0.2924089729785919, "learning_rate": 4.244218834039525e-05, "loss": 1.739, "step": 2308 }, { "epoch": 0.5575947838686308, "grad_norm": 0.27995172142982483, "learning_rate": 4.240392674272401e-05, "loss": 1.5351, "step": 2309 }, { "epoch": 0.5578362714320213, "grad_norm": 0.28078192472457886, "learning_rate": 4.2365669697708196e-05, "loss": 1.6549, "step": 2310 }, { "epoch": 0.5580777589954118, "grad_norm": 0.2886807918548584, "learning_rate": 4.232741722827697e-05, "loss": 1.7442, "step": 2311 }, { "epoch": 0.5583192465588023, "grad_norm": 0.2938794493675232, "learning_rate": 4.2289169357356604e-05, "loss": 1.6158, "step": 2312 }, { "epoch": 0.5585607341221928, "grad_norm": 0.3002696633338928, "learning_rate": 4.225092610787076e-05, "loss": 1.749, "step": 2313 }, { "epoch": 0.5588022216855832, "grad_norm": 0.30026525259017944, "learning_rate": 4.221268750274027e-05, "loss": 1.7773, "step": 2314 }, { "epoch": 0.5590437092489737, "grad_norm": 0.30950358510017395, "learning_rate": 4.217445356488316e-05, "loss": 1.8417, "step": 2315 }, { "epoch": 0.5592851968123642, "grad_norm": 0.2927887737751007, "learning_rate": 4.2136224317214696e-05, "loss": 1.6511, "step": 2316 }, { "epoch": 0.5595266843757547, "grad_norm": 0.29351142048835754, "learning_rate": 4.209799978264733e-05, "loss": 1.6808, "step": 2317 }, { "epoch": 0.5597681719391452, "grad_norm": 0.31574076414108276, "learning_rate": 4.205977998409067e-05, "loss": 1.8953, "step": 2318 }, { "epoch": 0.5600096595025357, "grad_norm": 0.2794061005115509, "learning_rate": 4.202156494445149e-05, "loss": 1.5778, "step": 2319 }, { "epoch": 0.5602511470659262, "grad_norm": 0.3101471960544586, "learning_rate": 4.1983354686633736e-05, "loss": 1.7337, "step": 2320 }, { "epoch": 0.5604926346293166, "grad_norm": 0.2811819314956665, "learning_rate": 4.194514923353844e-05, "loss": 1.5988, "step": 2321 }, { "epoch": 0.5607341221927071, "grad_norm": 0.28594309091567993, "learning_rate": 4.19069486080638e-05, "loss": 1.7068, "step": 2322 }, { "epoch": 0.5609756097560976, "grad_norm": 0.29027509689331055, "learning_rate": 4.1868752833105134e-05, "loss": 1.6302, "step": 2323 }, { "epoch": 0.5612170973194881, "grad_norm": 0.3033696413040161, "learning_rate": 4.1830561931554766e-05, "loss": 1.8369, "step": 2324 }, { "epoch": 0.5614585848828786, "grad_norm": 0.2925368845462799, "learning_rate": 4.179237592630219e-05, "loss": 1.6798, "step": 2325 }, { "epoch": 0.5617000724462691, "grad_norm": 0.2927570343017578, "learning_rate": 4.175419484023394e-05, "loss": 1.6735, "step": 2326 }, { "epoch": 0.5619415600096596, "grad_norm": 0.29845020174980164, "learning_rate": 4.171601869623356e-05, "loss": 1.7078, "step": 2327 }, { "epoch": 0.56218304757305, "grad_norm": 0.2932681143283844, "learning_rate": 4.167784751718168e-05, "loss": 1.7334, "step": 2328 }, { "epoch": 0.5624245351364405, "grad_norm": 0.2873363792896271, "learning_rate": 4.163968132595595e-05, "loss": 1.7561, "step": 2329 }, { "epoch": 0.5626660226998309, "grad_norm": 0.2868853807449341, "learning_rate": 4.1601520145431016e-05, "loss": 1.7715, "step": 2330 }, { "epoch": 0.5629075102632214, "grad_norm": 0.2683984339237213, "learning_rate": 4.156336399847851e-05, "loss": 1.4812, "step": 2331 }, { "epoch": 0.5631489978266119, "grad_norm": 0.2749142646789551, "learning_rate": 4.15252129079671e-05, "loss": 1.5403, "step": 2332 }, { "epoch": 0.5633904853900024, "grad_norm": 0.2823919951915741, "learning_rate": 4.1487066896762345e-05, "loss": 1.6958, "step": 2333 }, { "epoch": 0.5636319729533928, "grad_norm": 0.2981533706188202, "learning_rate": 4.1448925987726804e-05, "loss": 1.6578, "step": 2334 }, { "epoch": 0.5638734605167833, "grad_norm": 0.2998083829879761, "learning_rate": 4.1410790203720015e-05, "loss": 1.6885, "step": 2335 }, { "epoch": 0.5641149480801738, "grad_norm": 0.29314276576042175, "learning_rate": 4.1372659567598356e-05, "loss": 1.6822, "step": 2336 }, { "epoch": 0.5643564356435643, "grad_norm": 0.2964651584625244, "learning_rate": 4.133453410221518e-05, "loss": 1.7131, "step": 2337 }, { "epoch": 0.5645979232069548, "grad_norm": 0.2981303036212921, "learning_rate": 4.129641383042076e-05, "loss": 1.7426, "step": 2338 }, { "epoch": 0.5648394107703453, "grad_norm": 0.29483747482299805, "learning_rate": 4.1258298775062176e-05, "loss": 1.6644, "step": 2339 }, { "epoch": 0.5650808983337358, "grad_norm": 0.28027236461639404, "learning_rate": 4.122018895898345e-05, "loss": 1.5905, "step": 2340 }, { "epoch": 0.5653223858971262, "grad_norm": 0.29557839035987854, "learning_rate": 4.118208440502546e-05, "loss": 1.6643, "step": 2341 }, { "epoch": 0.5655638734605167, "grad_norm": 0.29199713468551636, "learning_rate": 4.114398513602589e-05, "loss": 1.5863, "step": 2342 }, { "epoch": 0.5658053610239072, "grad_norm": 0.30521339178085327, "learning_rate": 4.11058911748193e-05, "loss": 1.7297, "step": 2343 }, { "epoch": 0.5660468485872977, "grad_norm": 0.321456640958786, "learning_rate": 4.106780254423706e-05, "loss": 1.886, "step": 2344 }, { "epoch": 0.5662883361506882, "grad_norm": 0.2877240777015686, "learning_rate": 4.102971926710728e-05, "loss": 1.6106, "step": 2345 }, { "epoch": 0.5665298237140787, "grad_norm": 0.2801564335823059, "learning_rate": 4.0991641366254974e-05, "loss": 1.727, "step": 2346 }, { "epoch": 0.5667713112774692, "grad_norm": 0.29211926460266113, "learning_rate": 4.095356886450187e-05, "loss": 1.7455, "step": 2347 }, { "epoch": 0.5670127988408596, "grad_norm": 0.294394850730896, "learning_rate": 4.091550178466642e-05, "loss": 1.7495, "step": 2348 }, { "epoch": 0.5672542864042501, "grad_norm": 0.29620879888534546, "learning_rate": 4.087744014956391e-05, "loss": 1.7547, "step": 2349 }, { "epoch": 0.5674957739676406, "grad_norm": 0.28298842906951904, "learning_rate": 4.083938398200633e-05, "loss": 1.6101, "step": 2350 }, { "epoch": 0.5677372615310311, "grad_norm": 0.31541067361831665, "learning_rate": 4.080133330480236e-05, "loss": 2.0024, "step": 2351 }, { "epoch": 0.5679787490944216, "grad_norm": 0.2839082181453705, "learning_rate": 4.076328814075745e-05, "loss": 1.6478, "step": 2352 }, { "epoch": 0.5682202366578121, "grad_norm": 0.30273088812828064, "learning_rate": 4.07252485126737e-05, "loss": 1.8235, "step": 2353 }, { "epoch": 0.5684617242212026, "grad_norm": 0.30257901549339294, "learning_rate": 4.068721444334988e-05, "loss": 1.7311, "step": 2354 }, { "epoch": 0.568703211784593, "grad_norm": 0.2906223237514496, "learning_rate": 4.064918595558149e-05, "loss": 1.7521, "step": 2355 }, { "epoch": 0.5689446993479835, "grad_norm": 0.3021504580974579, "learning_rate": 4.061116307216065e-05, "loss": 1.8744, "step": 2356 }, { "epoch": 0.569186186911374, "grad_norm": 0.28596988320350647, "learning_rate": 4.0573145815876076e-05, "loss": 1.6737, "step": 2357 }, { "epoch": 0.5694276744747645, "grad_norm": 0.3047086298465729, "learning_rate": 4.053513420951319e-05, "loss": 1.9388, "step": 2358 }, { "epoch": 0.569669162038155, "grad_norm": 0.2860809862613678, "learning_rate": 4.0497128275854005e-05, "loss": 1.8027, "step": 2359 }, { "epoch": 0.5699106496015455, "grad_norm": 0.2784964442253113, "learning_rate": 4.0459128037677086e-05, "loss": 1.6722, "step": 2360 }, { "epoch": 0.570152137164936, "grad_norm": 0.2995767593383789, "learning_rate": 4.042113351775764e-05, "loss": 1.8453, "step": 2361 }, { "epoch": 0.5703936247283264, "grad_norm": 0.29328471422195435, "learning_rate": 4.0383144738867426e-05, "loss": 1.8588, "step": 2362 }, { "epoch": 0.5706351122917169, "grad_norm": 0.28885993361473083, "learning_rate": 4.034516172377475e-05, "loss": 1.6593, "step": 2363 }, { "epoch": 0.5708765998551074, "grad_norm": 0.2774667739868164, "learning_rate": 4.030718449524449e-05, "loss": 1.7335, "step": 2364 }, { "epoch": 0.5711180874184979, "grad_norm": 0.30328652262687683, "learning_rate": 4.026921307603807e-05, "loss": 1.7287, "step": 2365 }, { "epoch": 0.5713595749818884, "grad_norm": 0.3014756739139557, "learning_rate": 4.023124748891335e-05, "loss": 1.678, "step": 2366 }, { "epoch": 0.5716010625452789, "grad_norm": 0.29024362564086914, "learning_rate": 4.0193287756624775e-05, "loss": 1.6659, "step": 2367 }, { "epoch": 0.5718425501086694, "grad_norm": 0.29662129282951355, "learning_rate": 4.015533390192327e-05, "loss": 1.5967, "step": 2368 }, { "epoch": 0.5720840376720598, "grad_norm": 0.3441145718097687, "learning_rate": 4.01173859475562e-05, "loss": 2.1541, "step": 2369 }, { "epoch": 0.5723255252354503, "grad_norm": 0.279339998960495, "learning_rate": 4.007944391626741e-05, "loss": 1.4705, "step": 2370 }, { "epoch": 0.5725670127988408, "grad_norm": 0.3139488399028778, "learning_rate": 4.0041507830797235e-05, "loss": 1.8637, "step": 2371 }, { "epoch": 0.5728085003622313, "grad_norm": 0.28232163190841675, "learning_rate": 4.00035777138824e-05, "loss": 1.5222, "step": 2372 }, { "epoch": 0.5730499879256218, "grad_norm": 0.314220666885376, "learning_rate": 3.9965653588256034e-05, "loss": 1.8915, "step": 2373 }, { "epoch": 0.5732914754890123, "grad_norm": 0.30938661098480225, "learning_rate": 3.992773547664775e-05, "loss": 1.9457, "step": 2374 }, { "epoch": 0.5735329630524028, "grad_norm": 0.2891864776611328, "learning_rate": 3.988982340178348e-05, "loss": 1.6559, "step": 2375 }, { "epoch": 0.5737744506157932, "grad_norm": 0.29653200507164, "learning_rate": 3.9851917386385595e-05, "loss": 1.7833, "step": 2376 }, { "epoch": 0.5740159381791837, "grad_norm": 0.2898452877998352, "learning_rate": 3.981401745317281e-05, "loss": 1.7273, "step": 2377 }, { "epoch": 0.5742574257425742, "grad_norm": 0.29258641600608826, "learning_rate": 3.977612362486018e-05, "loss": 1.7492, "step": 2378 }, { "epoch": 0.5744989133059647, "grad_norm": 0.29464229941368103, "learning_rate": 3.973823592415912e-05, "loss": 1.8544, "step": 2379 }, { "epoch": 0.5747404008693552, "grad_norm": 0.3071231544017792, "learning_rate": 3.970035437377739e-05, "loss": 1.8271, "step": 2380 }, { "epoch": 0.5749818884327457, "grad_norm": 0.2985515296459198, "learning_rate": 3.966247899641901e-05, "loss": 1.8049, "step": 2381 }, { "epoch": 0.5752233759961362, "grad_norm": 0.2996237277984619, "learning_rate": 3.9624609814784334e-05, "loss": 1.7587, "step": 2382 }, { "epoch": 0.5754648635595266, "grad_norm": 0.29162517189979553, "learning_rate": 3.958674685157003e-05, "loss": 1.7144, "step": 2383 }, { "epoch": 0.5757063511229171, "grad_norm": 0.2822064459323883, "learning_rate": 3.954889012946899e-05, "loss": 1.5611, "step": 2384 }, { "epoch": 0.5759478386863076, "grad_norm": 0.29852136969566345, "learning_rate": 3.951103967117039e-05, "loss": 1.7841, "step": 2385 }, { "epoch": 0.5761893262496981, "grad_norm": 0.28846701979637146, "learning_rate": 3.947319549935967e-05, "loss": 1.6026, "step": 2386 }, { "epoch": 0.5764308138130886, "grad_norm": 0.2942967414855957, "learning_rate": 3.9435357636718435e-05, "loss": 1.71, "step": 2387 }, { "epoch": 0.5766723013764791, "grad_norm": 0.2943750321865082, "learning_rate": 3.9397526105924575e-05, "loss": 1.7564, "step": 2388 }, { "epoch": 0.5769137889398696, "grad_norm": 0.2861891984939575, "learning_rate": 3.935970092965219e-05, "loss": 1.7191, "step": 2389 }, { "epoch": 0.57715527650326, "grad_norm": 0.31451812386512756, "learning_rate": 3.932188213057151e-05, "loss": 1.8053, "step": 2390 }, { "epoch": 0.5773967640666505, "grad_norm": 0.3218313753604889, "learning_rate": 3.9284069731348976e-05, "loss": 1.9981, "step": 2391 }, { "epoch": 0.577638251630041, "grad_norm": 0.26717546582221985, "learning_rate": 3.9246263754647236e-05, "loss": 1.5531, "step": 2392 }, { "epoch": 0.5778797391934315, "grad_norm": 0.2770186960697174, "learning_rate": 3.9208464223124996e-05, "loss": 1.6213, "step": 2393 }, { "epoch": 0.578121226756822, "grad_norm": 0.3039669096469879, "learning_rate": 3.9170671159437176e-05, "loss": 1.6494, "step": 2394 }, { "epoch": 0.5783627143202125, "grad_norm": 0.30178627371788025, "learning_rate": 3.9132884586234805e-05, "loss": 1.7509, "step": 2395 }, { "epoch": 0.578604201883603, "grad_norm": 0.28393906354904175, "learning_rate": 3.9095104526165e-05, "loss": 1.7128, "step": 2396 }, { "epoch": 0.5788456894469934, "grad_norm": 0.3065143823623657, "learning_rate": 3.905733100187097e-05, "loss": 1.6912, "step": 2397 }, { "epoch": 0.5790871770103839, "grad_norm": 0.28713905811309814, "learning_rate": 3.9019564035992066e-05, "loss": 1.6125, "step": 2398 }, { "epoch": 0.5793286645737744, "grad_norm": 0.28162071108818054, "learning_rate": 3.898180365116362e-05, "loss": 1.7051, "step": 2399 }, { "epoch": 0.5795701521371649, "grad_norm": 0.2927330732345581, "learning_rate": 3.8944049870017084e-05, "loss": 1.7898, "step": 2400 }, { "epoch": 0.5798116397005554, "grad_norm": 0.2920984625816345, "learning_rate": 3.890630271517995e-05, "loss": 1.5425, "step": 2401 }, { "epoch": 0.5800531272639459, "grad_norm": 0.2938392460346222, "learning_rate": 3.886856220927569e-05, "loss": 1.8429, "step": 2402 }, { "epoch": 0.5802946148273364, "grad_norm": 0.30278971791267395, "learning_rate": 3.883082837492383e-05, "loss": 1.8823, "step": 2403 }, { "epoch": 0.5805361023907268, "grad_norm": 0.3013326823711395, "learning_rate": 3.879310123473992e-05, "loss": 1.7736, "step": 2404 }, { "epoch": 0.5807775899541173, "grad_norm": 0.28163909912109375, "learning_rate": 3.875538081133543e-05, "loss": 1.6267, "step": 2405 }, { "epoch": 0.5810190775175078, "grad_norm": 0.29443424940109253, "learning_rate": 3.871766712731785e-05, "loss": 1.6943, "step": 2406 }, { "epoch": 0.5812605650808983, "grad_norm": 0.28833210468292236, "learning_rate": 3.867996020529063e-05, "loss": 1.7046, "step": 2407 }, { "epoch": 0.5815020526442888, "grad_norm": 0.2918473780155182, "learning_rate": 3.8642260067853165e-05, "loss": 1.7966, "step": 2408 }, { "epoch": 0.5817435402076793, "grad_norm": 0.3106781244277954, "learning_rate": 3.860456673760077e-05, "loss": 1.847, "step": 2409 }, { "epoch": 0.5819850277710698, "grad_norm": 0.29532912373542786, "learning_rate": 3.856688023712471e-05, "loss": 1.7156, "step": 2410 }, { "epoch": 0.5822265153344602, "grad_norm": 0.2847370207309723, "learning_rate": 3.852920058901209e-05, "loss": 1.6685, "step": 2411 }, { "epoch": 0.5824680028978507, "grad_norm": 0.28784051537513733, "learning_rate": 3.8491527815846e-05, "loss": 1.7316, "step": 2412 }, { "epoch": 0.5827094904612412, "grad_norm": 0.28324827551841736, "learning_rate": 3.845386194020535e-05, "loss": 1.6571, "step": 2413 }, { "epoch": 0.5829509780246317, "grad_norm": 0.28542235493659973, "learning_rate": 3.841620298466492e-05, "loss": 1.6766, "step": 2414 }, { "epoch": 0.5831924655880222, "grad_norm": 0.29939380288124084, "learning_rate": 3.8378550971795346e-05, "loss": 1.6555, "step": 2415 }, { "epoch": 0.5834339531514127, "grad_norm": 0.27919045090675354, "learning_rate": 3.834090592416313e-05, "loss": 1.7092, "step": 2416 }, { "epoch": 0.5836754407148032, "grad_norm": 0.2961058020591736, "learning_rate": 3.830326786433056e-05, "loss": 1.6037, "step": 2417 }, { "epoch": 0.5839169282781936, "grad_norm": 0.29718253016471863, "learning_rate": 3.826563681485576e-05, "loss": 1.6855, "step": 2418 }, { "epoch": 0.5841584158415841, "grad_norm": 0.32008591294288635, "learning_rate": 3.8228012798292666e-05, "loss": 1.9544, "step": 2419 }, { "epoch": 0.5843999034049746, "grad_norm": 0.29271090030670166, "learning_rate": 3.8190395837190945e-05, "loss": 1.7386, "step": 2420 }, { "epoch": 0.5846413909683651, "grad_norm": 0.2830386757850647, "learning_rate": 3.8152785954096086e-05, "loss": 1.5592, "step": 2421 }, { "epoch": 0.5848828785317556, "grad_norm": 0.30118370056152344, "learning_rate": 3.811518317154934e-05, "loss": 1.61, "step": 2422 }, { "epoch": 0.5851243660951461, "grad_norm": 0.28814879059791565, "learning_rate": 3.807758751208765e-05, "loss": 1.7294, "step": 2423 }, { "epoch": 0.5853658536585366, "grad_norm": 0.28450068831443787, "learning_rate": 3.803999899824372e-05, "loss": 1.5363, "step": 2424 }, { "epoch": 0.585607341221927, "grad_norm": 0.2805667519569397, "learning_rate": 3.800241765254602e-05, "loss": 1.675, "step": 2425 }, { "epoch": 0.5858488287853175, "grad_norm": 0.28969326615333557, "learning_rate": 3.796484349751861e-05, "loss": 1.8751, "step": 2426 }, { "epoch": 0.586090316348708, "grad_norm": 0.2742403745651245, "learning_rate": 3.792727655568135e-05, "loss": 1.634, "step": 2427 }, { "epoch": 0.5863318039120985, "grad_norm": 0.28678256273269653, "learning_rate": 3.7889716849549734e-05, "loss": 1.6174, "step": 2428 }, { "epoch": 0.586573291475489, "grad_norm": 0.2994878888130188, "learning_rate": 3.78521644016349e-05, "loss": 1.7828, "step": 2429 }, { "epoch": 0.5868147790388795, "grad_norm": 0.2877824902534485, "learning_rate": 3.781461923444366e-05, "loss": 1.7237, "step": 2430 }, { "epoch": 0.58705626660227, "grad_norm": 0.30101504921913147, "learning_rate": 3.777708137047847e-05, "loss": 1.6802, "step": 2431 }, { "epoch": 0.5872977541656604, "grad_norm": 0.28113824129104614, "learning_rate": 3.7739550832237406e-05, "loss": 1.6509, "step": 2432 }, { "epoch": 0.5875392417290509, "grad_norm": 0.29675886034965515, "learning_rate": 3.770202764221411e-05, "loss": 1.8597, "step": 2433 }, { "epoch": 0.5877807292924414, "grad_norm": 0.27202603220939636, "learning_rate": 3.766451182289787e-05, "loss": 1.6519, "step": 2434 }, { "epoch": 0.5880222168558319, "grad_norm": 0.28778839111328125, "learning_rate": 3.762700339677356e-05, "loss": 1.7648, "step": 2435 }, { "epoch": 0.5882637044192224, "grad_norm": 0.2734607458114624, "learning_rate": 3.7589502386321565e-05, "loss": 1.6916, "step": 2436 }, { "epoch": 0.5885051919826129, "grad_norm": 0.28360888361930847, "learning_rate": 3.755200881401788e-05, "loss": 1.663, "step": 2437 }, { "epoch": 0.5887466795460033, "grad_norm": 0.2922933101654053, "learning_rate": 3.751452270233405e-05, "loss": 1.6373, "step": 2438 }, { "epoch": 0.5889881671093938, "grad_norm": 0.2839530408382416, "learning_rate": 3.747704407373709e-05, "loss": 1.7324, "step": 2439 }, { "epoch": 0.5892296546727843, "grad_norm": 0.31759944558143616, "learning_rate": 3.743957295068957e-05, "loss": 1.8874, "step": 2440 }, { "epoch": 0.5894711422361748, "grad_norm": 0.28140199184417725, "learning_rate": 3.740210935564957e-05, "loss": 1.6555, "step": 2441 }, { "epoch": 0.5897126297995653, "grad_norm": 0.28111740946769714, "learning_rate": 3.7364653311070624e-05, "loss": 1.5488, "step": 2442 }, { "epoch": 0.5899541173629558, "grad_norm": 0.2938031852245331, "learning_rate": 3.732720483940177e-05, "loss": 1.6941, "step": 2443 }, { "epoch": 0.5901956049263463, "grad_norm": 0.2941116988658905, "learning_rate": 3.728976396308753e-05, "loss": 1.6027, "step": 2444 }, { "epoch": 0.5904370924897367, "grad_norm": 0.28593021631240845, "learning_rate": 3.7252330704567783e-05, "loss": 1.6309, "step": 2445 }, { "epoch": 0.5906785800531272, "grad_norm": 0.3406621217727661, "learning_rate": 3.721490508627794e-05, "loss": 1.5683, "step": 2446 }, { "epoch": 0.5909200676165177, "grad_norm": 0.2913300693035126, "learning_rate": 3.717748713064879e-05, "loss": 1.8347, "step": 2447 }, { "epoch": 0.5911615551799082, "grad_norm": 0.2758163809776306, "learning_rate": 3.7140076860106516e-05, "loss": 1.6426, "step": 2448 }, { "epoch": 0.5914030427432987, "grad_norm": 0.2935280203819275, "learning_rate": 3.710267429707271e-05, "loss": 1.7378, "step": 2449 }, { "epoch": 0.5916445303066892, "grad_norm": 0.27553948760032654, "learning_rate": 3.7065279463964386e-05, "loss": 1.7324, "step": 2450 }, { "epoch": 0.5918860178700797, "grad_norm": 0.283493310213089, "learning_rate": 3.702789238319384e-05, "loss": 1.667, "step": 2451 }, { "epoch": 0.5921275054334701, "grad_norm": 0.280386358499527, "learning_rate": 3.69905130771688e-05, "loss": 1.6037, "step": 2452 }, { "epoch": 0.5923689929968606, "grad_norm": 0.29831987619400024, "learning_rate": 3.695314156829228e-05, "loss": 1.7686, "step": 2453 }, { "epoch": 0.5926104805602511, "grad_norm": 0.2817460298538208, "learning_rate": 3.691577787896265e-05, "loss": 1.6645, "step": 2454 }, { "epoch": 0.5928519681236416, "grad_norm": 0.2845121920108795, "learning_rate": 3.6878422031573576e-05, "loss": 1.6362, "step": 2455 }, { "epoch": 0.5930934556870321, "grad_norm": 0.3131418824195862, "learning_rate": 3.684107404851408e-05, "loss": 1.8211, "step": 2456 }, { "epoch": 0.5933349432504226, "grad_norm": 0.3037307858467102, "learning_rate": 3.680373395216836e-05, "loss": 1.7217, "step": 2457 }, { "epoch": 0.5935764308138131, "grad_norm": 0.2949305474758148, "learning_rate": 3.676640176491598e-05, "loss": 1.7137, "step": 2458 }, { "epoch": 0.5938179183772035, "grad_norm": 0.2797428071498871, "learning_rate": 3.672907750913176e-05, "loss": 1.5676, "step": 2459 }, { "epoch": 0.594059405940594, "grad_norm": 0.28435376286506653, "learning_rate": 3.6691761207185695e-05, "loss": 1.6533, "step": 2460 }, { "epoch": 0.5943008935039845, "grad_norm": 0.27962738275527954, "learning_rate": 3.665445288144309e-05, "loss": 1.5085, "step": 2461 }, { "epoch": 0.594542381067375, "grad_norm": 0.31301939487457275, "learning_rate": 3.661715255426444e-05, "loss": 1.79, "step": 2462 }, { "epoch": 0.5947838686307655, "grad_norm": 0.2880149185657501, "learning_rate": 3.657986024800543e-05, "loss": 1.8574, "step": 2463 }, { "epoch": 0.595025356194156, "grad_norm": 0.2945672869682312, "learning_rate": 3.654257598501695e-05, "loss": 1.6864, "step": 2464 }, { "epoch": 0.5952668437575465, "grad_norm": 0.2866813838481903, "learning_rate": 3.650529978764511e-05, "loss": 1.7664, "step": 2465 }, { "epoch": 0.595508331320937, "grad_norm": 0.2802763283252716, "learning_rate": 3.6468031678231094e-05, "loss": 1.6141, "step": 2466 }, { "epoch": 0.5957498188843274, "grad_norm": 0.3504123389720917, "learning_rate": 3.643077167911133e-05, "loss": 2.0596, "step": 2467 }, { "epoch": 0.5959913064477179, "grad_norm": 0.31365031003952026, "learning_rate": 3.6393519812617363e-05, "loss": 1.9543, "step": 2468 }, { "epoch": 0.5962327940111084, "grad_norm": 0.29390665888786316, "learning_rate": 3.6356276101075795e-05, "loss": 1.8236, "step": 2469 }, { "epoch": 0.5964742815744989, "grad_norm": 0.31754180788993835, "learning_rate": 3.631904056680842e-05, "loss": 1.6463, "step": 2470 }, { "epoch": 0.5967157691378894, "grad_norm": 0.28453078866004944, "learning_rate": 3.6281813232132135e-05, "loss": 1.6558, "step": 2471 }, { "epoch": 0.5969572567012799, "grad_norm": 0.27454662322998047, "learning_rate": 3.624459411935884e-05, "loss": 1.7101, "step": 2472 }, { "epoch": 0.5971987442646703, "grad_norm": 0.2981424927711487, "learning_rate": 3.620738325079559e-05, "loss": 1.7915, "step": 2473 }, { "epoch": 0.5974402318280608, "grad_norm": 0.2696138024330139, "learning_rate": 3.617018064874448e-05, "loss": 1.4138, "step": 2474 }, { "epoch": 0.5976817193914513, "grad_norm": 0.2896297574043274, "learning_rate": 3.6132986335502604e-05, "loss": 1.7364, "step": 2475 }, { "epoch": 0.5979232069548418, "grad_norm": 0.3051794171333313, "learning_rate": 3.609580033336215e-05, "loss": 1.9886, "step": 2476 }, { "epoch": 0.5981646945182323, "grad_norm": 0.2891792058944702, "learning_rate": 3.605862266461031e-05, "loss": 1.7159, "step": 2477 }, { "epoch": 0.5984061820816228, "grad_norm": 0.30678558349609375, "learning_rate": 3.6021453351529236e-05, "loss": 1.9696, "step": 2478 }, { "epoch": 0.5986476696450133, "grad_norm": 0.28317415714263916, "learning_rate": 3.598429241639612e-05, "loss": 1.5766, "step": 2479 }, { "epoch": 0.5988891572084037, "grad_norm": 0.29772067070007324, "learning_rate": 3.594713988148314e-05, "loss": 1.7123, "step": 2480 }, { "epoch": 0.5991306447717942, "grad_norm": 0.2847268283367157, "learning_rate": 3.590999576905738e-05, "loss": 1.7397, "step": 2481 }, { "epoch": 0.5993721323351847, "grad_norm": 0.31777217984199524, "learning_rate": 3.5872860101380923e-05, "loss": 1.6845, "step": 2482 }, { "epoch": 0.5996136198985752, "grad_norm": 0.2729204595088959, "learning_rate": 3.5835732900710815e-05, "loss": 1.5827, "step": 2483 }, { "epoch": 0.5998551074619657, "grad_norm": 0.29481253027915955, "learning_rate": 3.579861418929895e-05, "loss": 1.6554, "step": 2484 }, { "epoch": 0.6000965950253562, "grad_norm": 0.285703182220459, "learning_rate": 3.576150398939222e-05, "loss": 1.6673, "step": 2485 }, { "epoch": 0.6003380825887467, "grad_norm": 0.2772272825241089, "learning_rate": 3.5724402323232345e-05, "loss": 1.6138, "step": 2486 }, { "epoch": 0.6005795701521371, "grad_norm": 0.2847372889518738, "learning_rate": 3.5687309213055963e-05, "loss": 1.7024, "step": 2487 }, { "epoch": 0.6008210577155276, "grad_norm": 0.273150771856308, "learning_rate": 3.565022468109458e-05, "loss": 1.6056, "step": 2488 }, { "epoch": 0.6010625452789181, "grad_norm": 0.2882135510444641, "learning_rate": 3.561314874957459e-05, "loss": 1.663, "step": 2489 }, { "epoch": 0.6013040328423086, "grad_norm": 0.2922912538051605, "learning_rate": 3.5576081440717146e-05, "loss": 1.5678, "step": 2490 }, { "epoch": 0.6015455204056991, "grad_norm": 0.2886035144329071, "learning_rate": 3.5539022776738333e-05, "loss": 1.6148, "step": 2491 }, { "epoch": 0.6017870079690896, "grad_norm": 0.2937653660774231, "learning_rate": 3.550197277984902e-05, "loss": 1.7949, "step": 2492 }, { "epoch": 0.6020284955324801, "grad_norm": 0.2867985665798187, "learning_rate": 3.5464931472254835e-05, "loss": 1.5835, "step": 2493 }, { "epoch": 0.6022699830958705, "grad_norm": 0.2932209372520447, "learning_rate": 3.5427898876156266e-05, "loss": 1.6855, "step": 2494 }, { "epoch": 0.602511470659261, "grad_norm": 0.29696425795555115, "learning_rate": 3.5390875013748536e-05, "loss": 1.7314, "step": 2495 }, { "epoch": 0.6027529582226515, "grad_norm": 0.29131823778152466, "learning_rate": 3.535385990722165e-05, "loss": 1.7477, "step": 2496 }, { "epoch": 0.602994445786042, "grad_norm": 0.2906573414802551, "learning_rate": 3.5316853578760366e-05, "loss": 1.7911, "step": 2497 }, { "epoch": 0.6032359333494325, "grad_norm": 0.2835630178451538, "learning_rate": 3.5279856050544205e-05, "loss": 1.6069, "step": 2498 }, { "epoch": 0.603477420912823, "grad_norm": 0.283893346786499, "learning_rate": 3.524286734474733e-05, "loss": 1.6057, "step": 2499 }, { "epoch": 0.6037189084762135, "grad_norm": 0.2826724052429199, "learning_rate": 3.520588748353871e-05, "loss": 1.5871, "step": 2500 }, { "epoch": 0.6039603960396039, "grad_norm": 0.2928446829319, "learning_rate": 3.516891648908199e-05, "loss": 1.6579, "step": 2501 }, { "epoch": 0.6042018836029944, "grad_norm": 0.29195818305015564, "learning_rate": 3.513195438353545e-05, "loss": 1.5969, "step": 2502 }, { "epoch": 0.6044433711663849, "grad_norm": 0.28115230798721313, "learning_rate": 3.50950011890521e-05, "loss": 1.5779, "step": 2503 }, { "epoch": 0.6046848587297754, "grad_norm": 0.28780871629714966, "learning_rate": 3.505805692777959e-05, "loss": 1.7158, "step": 2504 }, { "epoch": 0.6049263462931659, "grad_norm": 0.30311766266822815, "learning_rate": 3.502112162186021e-05, "loss": 1.8843, "step": 2505 }, { "epoch": 0.6051678338565564, "grad_norm": 0.2881084382534027, "learning_rate": 3.498419529343088e-05, "loss": 1.6563, "step": 2506 }, { "epoch": 0.6054093214199469, "grad_norm": 0.28938812017440796, "learning_rate": 3.494727796462316e-05, "loss": 1.695, "step": 2507 }, { "epoch": 0.6056508089833373, "grad_norm": 0.29259443283081055, "learning_rate": 3.491036965756319e-05, "loss": 1.7133, "step": 2508 }, { "epoch": 0.6058922965467278, "grad_norm": 0.2863447368144989, "learning_rate": 3.487347039437172e-05, "loss": 1.5392, "step": 2509 }, { "epoch": 0.6061337841101183, "grad_norm": 0.3094983696937561, "learning_rate": 3.48365801971641e-05, "loss": 1.7958, "step": 2510 }, { "epoch": 0.6063752716735088, "grad_norm": 0.29001355171203613, "learning_rate": 3.4799699088050175e-05, "loss": 1.6547, "step": 2511 }, { "epoch": 0.6066167592368993, "grad_norm": 0.2896830141544342, "learning_rate": 3.476282708913442e-05, "loss": 1.6018, "step": 2512 }, { "epoch": 0.6068582468002898, "grad_norm": 0.28078269958496094, "learning_rate": 3.472596422251583e-05, "loss": 1.6191, "step": 2513 }, { "epoch": 0.6070997343636803, "grad_norm": 0.29417359828948975, "learning_rate": 3.4689110510287884e-05, "loss": 1.7853, "step": 2514 }, { "epoch": 0.6073412219270707, "grad_norm": 0.2910160422325134, "learning_rate": 3.4652265974538624e-05, "loss": 1.7013, "step": 2515 }, { "epoch": 0.6075827094904612, "grad_norm": 0.2800171375274658, "learning_rate": 3.4615430637350574e-05, "loss": 1.7375, "step": 2516 }, { "epoch": 0.6078241970538517, "grad_norm": 0.30251628160476685, "learning_rate": 3.457860452080075e-05, "loss": 1.6859, "step": 2517 }, { "epoch": 0.6080656846172422, "grad_norm": 0.31701117753982544, "learning_rate": 3.454178764696062e-05, "loss": 1.6141, "step": 2518 }, { "epoch": 0.6083071721806327, "grad_norm": 0.299738347530365, "learning_rate": 3.4504980037896174e-05, "loss": 1.6891, "step": 2519 }, { "epoch": 0.6085486597440232, "grad_norm": 0.32146161794662476, "learning_rate": 3.4468181715667744e-05, "loss": 1.81, "step": 2520 }, { "epoch": 0.6087901473074137, "grad_norm": 0.28754496574401855, "learning_rate": 3.443139270233019e-05, "loss": 1.7752, "step": 2521 }, { "epoch": 0.6090316348708041, "grad_norm": 0.2981823682785034, "learning_rate": 3.439461301993278e-05, "loss": 1.7541, "step": 2522 }, { "epoch": 0.6092731224341946, "grad_norm": 0.2895716428756714, "learning_rate": 3.435784269051913e-05, "loss": 1.5271, "step": 2523 }, { "epoch": 0.6095146099975851, "grad_norm": 0.28530311584472656, "learning_rate": 3.432108173612729e-05, "loss": 1.6378, "step": 2524 }, { "epoch": 0.6097560975609756, "grad_norm": 0.2870151102542877, "learning_rate": 3.4284330178789744e-05, "loss": 1.7852, "step": 2525 }, { "epoch": 0.6099975851243661, "grad_norm": 0.2743149697780609, "learning_rate": 3.4247588040533225e-05, "loss": 1.7404, "step": 2526 }, { "epoch": 0.6102390726877566, "grad_norm": 0.3176242411136627, "learning_rate": 3.4210855343378915e-05, "loss": 1.9476, "step": 2527 }, { "epoch": 0.610480560251147, "grad_norm": 0.28857824206352234, "learning_rate": 3.417413210934232e-05, "loss": 1.6602, "step": 2528 }, { "epoch": 0.6107220478145375, "grad_norm": 0.2840351462364197, "learning_rate": 3.4137418360433246e-05, "loss": 1.6683, "step": 2529 }, { "epoch": 0.610963535377928, "grad_norm": 0.2995029091835022, "learning_rate": 3.410071411865583e-05, "loss": 1.7435, "step": 2530 }, { "epoch": 0.6112050229413185, "grad_norm": 0.2946648895740509, "learning_rate": 3.406401940600855e-05, "loss": 1.8661, "step": 2531 }, { "epoch": 0.611446510504709, "grad_norm": 0.27369895577430725, "learning_rate": 3.402733424448408e-05, "loss": 1.6897, "step": 2532 }, { "epoch": 0.6116879980680995, "grad_norm": 0.27926602959632874, "learning_rate": 3.399065865606945e-05, "loss": 1.6643, "step": 2533 }, { "epoch": 0.61192948563149, "grad_norm": 0.2990940511226654, "learning_rate": 3.395399266274596e-05, "loss": 1.6767, "step": 2534 }, { "epoch": 0.6121709731948805, "grad_norm": 0.2834193706512451, "learning_rate": 3.391733628648907e-05, "loss": 1.7052, "step": 2535 }, { "epoch": 0.6124124607582709, "grad_norm": 0.28575870394706726, "learning_rate": 3.388068954926855e-05, "loss": 1.6699, "step": 2536 }, { "epoch": 0.6126539483216614, "grad_norm": 0.30421075224876404, "learning_rate": 3.384405247304839e-05, "loss": 1.8825, "step": 2537 }, { "epoch": 0.6128954358850519, "grad_norm": 0.280814528465271, "learning_rate": 3.380742507978678e-05, "loss": 1.5337, "step": 2538 }, { "epoch": 0.6131369234484424, "grad_norm": 0.2863577902317047, "learning_rate": 3.3770807391436074e-05, "loss": 1.7573, "step": 2539 }, { "epoch": 0.6133784110118329, "grad_norm": 0.2878822982311249, "learning_rate": 3.373419942994287e-05, "loss": 1.6626, "step": 2540 }, { "epoch": 0.6136198985752234, "grad_norm": 0.29794904589653015, "learning_rate": 3.369760121724787e-05, "loss": 1.7652, "step": 2541 }, { "epoch": 0.6138613861386139, "grad_norm": 0.28335195779800415, "learning_rate": 3.3661012775285985e-05, "loss": 1.7923, "step": 2542 }, { "epoch": 0.6141028737020043, "grad_norm": 0.2980829179286957, "learning_rate": 3.362443412598628e-05, "loss": 1.8541, "step": 2543 }, { "epoch": 0.6143443612653948, "grad_norm": 0.27349144220352173, "learning_rate": 3.358786529127187e-05, "loss": 1.5869, "step": 2544 }, { "epoch": 0.6145858488287853, "grad_norm": 0.3009245693683624, "learning_rate": 3.355130629306006e-05, "loss": 1.8052, "step": 2545 }, { "epoch": 0.6148273363921758, "grad_norm": 0.298391729593277, "learning_rate": 3.3514757153262266e-05, "loss": 1.7874, "step": 2546 }, { "epoch": 0.6150688239555663, "grad_norm": 0.2899065315723419, "learning_rate": 3.347821789378393e-05, "loss": 1.7291, "step": 2547 }, { "epoch": 0.6153103115189568, "grad_norm": 0.2932477295398712, "learning_rate": 3.344168853652462e-05, "loss": 1.8334, "step": 2548 }, { "epoch": 0.6155517990823473, "grad_norm": 0.2859414517879486, "learning_rate": 3.340516910337798e-05, "loss": 1.7236, "step": 2549 }, { "epoch": 0.6157932866457377, "grad_norm": 0.28934478759765625, "learning_rate": 3.336865961623167e-05, "loss": 1.6699, "step": 2550 }, { "epoch": 0.6160347742091282, "grad_norm": 0.2862183749675751, "learning_rate": 3.33321600969674e-05, "loss": 1.7199, "step": 2551 }, { "epoch": 0.6162762617725187, "grad_norm": 0.3068394064903259, "learning_rate": 3.329567056746096e-05, "loss": 1.7657, "step": 2552 }, { "epoch": 0.6165177493359092, "grad_norm": 0.27572986483573914, "learning_rate": 3.325919104958204e-05, "loss": 1.5948, "step": 2553 }, { "epoch": 0.6167592368992997, "grad_norm": 0.295484721660614, "learning_rate": 3.322272156519442e-05, "loss": 1.7904, "step": 2554 }, { "epoch": 0.6170007244626902, "grad_norm": 0.28706488013267517, "learning_rate": 3.318626213615586e-05, "loss": 1.7572, "step": 2555 }, { "epoch": 0.6172422120260806, "grad_norm": 0.2871039807796478, "learning_rate": 3.314981278431804e-05, "loss": 1.6473, "step": 2556 }, { "epoch": 0.6174836995894711, "grad_norm": 0.28321489691734314, "learning_rate": 3.3113373531526646e-05, "loss": 1.6641, "step": 2557 }, { "epoch": 0.6177251871528616, "grad_norm": 0.2938781678676605, "learning_rate": 3.30769443996213e-05, "loss": 1.5486, "step": 2558 }, { "epoch": 0.6179666747162521, "grad_norm": 0.27972105145454407, "learning_rate": 3.304052541043558e-05, "loss": 1.5644, "step": 2559 }, { "epoch": 0.6182081622796426, "grad_norm": 0.2967609763145447, "learning_rate": 3.3004116585796916e-05, "loss": 1.7207, "step": 2560 }, { "epoch": 0.6184496498430331, "grad_norm": 0.27557387948036194, "learning_rate": 3.296771794752673e-05, "loss": 1.4993, "step": 2561 }, { "epoch": 0.6186911374064236, "grad_norm": 0.2900605797767639, "learning_rate": 3.293132951744029e-05, "loss": 1.6753, "step": 2562 }, { "epoch": 0.618932624969814, "grad_norm": 0.29671502113342285, "learning_rate": 3.289495131734676e-05, "loss": 1.7206, "step": 2563 }, { "epoch": 0.6191741125332045, "grad_norm": 0.27957433462142944, "learning_rate": 3.285858336904914e-05, "loss": 1.7577, "step": 2564 }, { "epoch": 0.619415600096595, "grad_norm": 0.2918657958507538, "learning_rate": 3.282222569434439e-05, "loss": 1.7567, "step": 2565 }, { "epoch": 0.6196570876599855, "grad_norm": 0.3243831396102905, "learning_rate": 3.278587831502315e-05, "loss": 1.6618, "step": 2566 }, { "epoch": 0.619898575223376, "grad_norm": 0.2882837653160095, "learning_rate": 3.274954125287002e-05, "loss": 1.7469, "step": 2567 }, { "epoch": 0.6201400627867665, "grad_norm": 0.2877691686153412, "learning_rate": 3.271321452966339e-05, "loss": 1.7546, "step": 2568 }, { "epoch": 0.620381550350157, "grad_norm": 0.291064590215683, "learning_rate": 3.267689816717541e-05, "loss": 1.6374, "step": 2569 }, { "epoch": 0.6206230379135474, "grad_norm": 0.2818000912666321, "learning_rate": 3.2640592187172033e-05, "loss": 1.6859, "step": 2570 }, { "epoch": 0.6208645254769379, "grad_norm": 0.2997339069843292, "learning_rate": 3.2604296611413065e-05, "loss": 1.7283, "step": 2571 }, { "epoch": 0.6211060130403284, "grad_norm": 0.296236515045166, "learning_rate": 3.256801146165195e-05, "loss": 1.7916, "step": 2572 }, { "epoch": 0.6213475006037189, "grad_norm": 0.2987402379512787, "learning_rate": 3.253173675963598e-05, "loss": 1.8163, "step": 2573 }, { "epoch": 0.6215889881671094, "grad_norm": 0.2863186299800873, "learning_rate": 3.2495472527106154e-05, "loss": 1.6833, "step": 2574 }, { "epoch": 0.6218304757304999, "grad_norm": 0.27705222368240356, "learning_rate": 3.245921878579718e-05, "loss": 1.544, "step": 2575 }, { "epoch": 0.6220719632938904, "grad_norm": 0.29465770721435547, "learning_rate": 3.242297555743751e-05, "loss": 1.7578, "step": 2576 }, { "epoch": 0.6223134508572808, "grad_norm": 0.27233806252479553, "learning_rate": 3.2386742863749286e-05, "loss": 1.4936, "step": 2577 }, { "epoch": 0.6225549384206713, "grad_norm": 0.29390379786491394, "learning_rate": 3.235052072644831e-05, "loss": 1.6453, "step": 2578 }, { "epoch": 0.6227964259840618, "grad_norm": 0.28615322709083557, "learning_rate": 3.2314309167244075e-05, "loss": 1.6399, "step": 2579 }, { "epoch": 0.6230379135474523, "grad_norm": 0.29714158177375793, "learning_rate": 3.227810820783976e-05, "loss": 1.8235, "step": 2580 }, { "epoch": 0.6232794011108428, "grad_norm": 0.28728702664375305, "learning_rate": 3.2241917869932144e-05, "loss": 1.7948, "step": 2581 }, { "epoch": 0.6235208886742333, "grad_norm": 0.28320935368537903, "learning_rate": 3.2205738175211665e-05, "loss": 1.6333, "step": 2582 }, { "epoch": 0.6237623762376238, "grad_norm": 0.2972544729709625, "learning_rate": 3.21695691453624e-05, "loss": 1.8516, "step": 2583 }, { "epoch": 0.6240038638010142, "grad_norm": 0.29611480236053467, "learning_rate": 3.213341080206198e-05, "loss": 1.5637, "step": 2584 }, { "epoch": 0.6242453513644047, "grad_norm": 0.305329829454422, "learning_rate": 3.20972631669817e-05, "loss": 1.8579, "step": 2585 }, { "epoch": 0.6244868389277952, "grad_norm": 0.29886186122894287, "learning_rate": 3.206112626178639e-05, "loss": 1.6933, "step": 2586 }, { "epoch": 0.6247283264911857, "grad_norm": 0.27526891231536865, "learning_rate": 3.202500010813444e-05, "loss": 1.5552, "step": 2587 }, { "epoch": 0.6249698140545762, "grad_norm": 0.31331107020378113, "learning_rate": 3.198888472767784e-05, "loss": 2.0463, "step": 2588 }, { "epoch": 0.6252113016179667, "grad_norm": 0.29151982069015503, "learning_rate": 3.1952780142062105e-05, "loss": 1.5956, "step": 2589 }, { "epoch": 0.6254527891813572, "grad_norm": 0.2870679199695587, "learning_rate": 3.191668637292625e-05, "loss": 1.6477, "step": 2590 }, { "epoch": 0.6256942767447476, "grad_norm": 0.2879990339279175, "learning_rate": 3.188060344190286e-05, "loss": 1.6848, "step": 2591 }, { "epoch": 0.6259357643081381, "grad_norm": 0.2850443124771118, "learning_rate": 3.1844531370617993e-05, "loss": 1.7737, "step": 2592 }, { "epoch": 0.6261772518715286, "grad_norm": 0.296732634305954, "learning_rate": 3.180847018069117e-05, "loss": 1.76, "step": 2593 }, { "epoch": 0.6264187394349191, "grad_norm": 0.28823122382164, "learning_rate": 3.1772419893735464e-05, "loss": 1.6884, "step": 2594 }, { "epoch": 0.6266602269983096, "grad_norm": 0.31147027015686035, "learning_rate": 3.1736380531357356e-05, "loss": 1.9074, "step": 2595 }, { "epoch": 0.6269017145617001, "grad_norm": 0.29306653141975403, "learning_rate": 3.1700352115156786e-05, "loss": 1.7151, "step": 2596 }, { "epoch": 0.6271432021250906, "grad_norm": 0.2906031608581543, "learning_rate": 3.166433466672716e-05, "loss": 1.7093, "step": 2597 }, { "epoch": 0.627384689688481, "grad_norm": 0.2861940562725067, "learning_rate": 3.16283282076553e-05, "loss": 1.6589, "step": 2598 }, { "epoch": 0.6276261772518715, "grad_norm": 0.27941274642944336, "learning_rate": 3.1592332759521406e-05, "loss": 1.549, "step": 2599 }, { "epoch": 0.627867664815262, "grad_norm": 0.2811283767223358, "learning_rate": 3.155634834389911e-05, "loss": 1.6513, "step": 2600 }, { "epoch": 0.6281091523786525, "grad_norm": 0.27396464347839355, "learning_rate": 3.152037498235548e-05, "loss": 1.5598, "step": 2601 }, { "epoch": 0.628350639942043, "grad_norm": 0.27273061871528625, "learning_rate": 3.148441269645084e-05, "loss": 1.478, "step": 2602 }, { "epoch": 0.6285921275054335, "grad_norm": 0.30087143182754517, "learning_rate": 3.144846150773898e-05, "loss": 1.6847, "step": 2603 }, { "epoch": 0.628833615068824, "grad_norm": 0.30136755108833313, "learning_rate": 3.1412521437767005e-05, "loss": 1.6842, "step": 2604 }, { "epoch": 0.6290751026322144, "grad_norm": 0.3245827555656433, "learning_rate": 3.137659250807535e-05, "loss": 2.0499, "step": 2605 }, { "epoch": 0.6293165901956049, "grad_norm": 0.29533663392066956, "learning_rate": 3.134067474019777e-05, "loss": 1.6967, "step": 2606 }, { "epoch": 0.6295580777589954, "grad_norm": 0.29607248306274414, "learning_rate": 3.130476815566134e-05, "loss": 1.7638, "step": 2607 }, { "epoch": 0.6297995653223859, "grad_norm": 0.29233407974243164, "learning_rate": 3.1268872775986444e-05, "loss": 1.7201, "step": 2608 }, { "epoch": 0.6300410528857764, "grad_norm": 0.29236966371536255, "learning_rate": 3.123298862268671e-05, "loss": 1.6817, "step": 2609 }, { "epoch": 0.6302825404491669, "grad_norm": 0.28770408034324646, "learning_rate": 3.1197115717269107e-05, "loss": 1.7639, "step": 2610 }, { "epoch": 0.6305240280125574, "grad_norm": 0.29003438353538513, "learning_rate": 3.116125408123377e-05, "loss": 1.7106, "step": 2611 }, { "epoch": 0.6307655155759478, "grad_norm": 0.2912517488002777, "learning_rate": 3.112540373607415e-05, "loss": 1.6966, "step": 2612 }, { "epoch": 0.6310070031393383, "grad_norm": 0.2724066972732544, "learning_rate": 3.1089564703276944e-05, "loss": 1.6665, "step": 2613 }, { "epoch": 0.6312484907027288, "grad_norm": 0.2684726417064667, "learning_rate": 3.105373700432197e-05, "loss": 1.6248, "step": 2614 }, { "epoch": 0.6314899782661193, "grad_norm": 0.28684425354003906, "learning_rate": 3.101792066068235e-05, "loss": 1.6181, "step": 2615 }, { "epoch": 0.6317314658295098, "grad_norm": 0.31175005435943604, "learning_rate": 3.0982115693824396e-05, "loss": 1.7954, "step": 2616 }, { "epoch": 0.6319729533929003, "grad_norm": 0.29426541924476624, "learning_rate": 3.094632212520753e-05, "loss": 1.7507, "step": 2617 }, { "epoch": 0.6322144409562908, "grad_norm": 0.28397175669670105, "learning_rate": 3.091053997628442e-05, "loss": 1.6525, "step": 2618 }, { "epoch": 0.6324559285196812, "grad_norm": 0.2837525010108948, "learning_rate": 3.087476926850084e-05, "loss": 1.643, "step": 2619 }, { "epoch": 0.6326974160830717, "grad_norm": 0.303946316242218, "learning_rate": 3.083901002329571e-05, "loss": 1.6161, "step": 2620 }, { "epoch": 0.6329389036464622, "grad_norm": 0.2878418564796448, "learning_rate": 3.080326226210112e-05, "loss": 1.6927, "step": 2621 }, { "epoch": 0.6331803912098527, "grad_norm": 0.3076441287994385, "learning_rate": 3.076752600634225e-05, "loss": 1.7669, "step": 2622 }, { "epoch": 0.6334218787732432, "grad_norm": 0.31105130910873413, "learning_rate": 3.073180127743735e-05, "loss": 1.7484, "step": 2623 }, { "epoch": 0.6336633663366337, "grad_norm": 0.28631484508514404, "learning_rate": 3.0696088096797834e-05, "loss": 1.7043, "step": 2624 }, { "epoch": 0.6339048539000242, "grad_norm": 0.28444620966911316, "learning_rate": 3.066038648582816e-05, "loss": 1.585, "step": 2625 }, { "epoch": 0.6341463414634146, "grad_norm": 0.37340953946113586, "learning_rate": 3.062469646592581e-05, "loss": 1.9321, "step": 2626 }, { "epoch": 0.6343878290268051, "grad_norm": 0.3058568835258484, "learning_rate": 3.0589018058481375e-05, "loss": 1.6385, "step": 2627 }, { "epoch": 0.6346293165901956, "grad_norm": 0.29442429542541504, "learning_rate": 3.055335128487848e-05, "loss": 1.7678, "step": 2628 }, { "epoch": 0.6348708041535861, "grad_norm": 0.3029642403125763, "learning_rate": 3.051769616649375e-05, "loss": 1.6958, "step": 2629 }, { "epoch": 0.6351122917169766, "grad_norm": 0.28807470202445984, "learning_rate": 3.0482052724696852e-05, "loss": 1.7145, "step": 2630 }, { "epoch": 0.6353537792803671, "grad_norm": 0.2913690507411957, "learning_rate": 3.0446420980850453e-05, "loss": 1.6712, "step": 2631 }, { "epoch": 0.6355952668437576, "grad_norm": 0.2830966114997864, "learning_rate": 3.0410800956310176e-05, "loss": 1.7044, "step": 2632 }, { "epoch": 0.635836754407148, "grad_norm": 0.2860349118709564, "learning_rate": 3.0375192672424645e-05, "loss": 1.7258, "step": 2633 }, { "epoch": 0.6360782419705385, "grad_norm": 0.2756868898868561, "learning_rate": 3.033959615053548e-05, "loss": 1.6017, "step": 2634 }, { "epoch": 0.636319729533929, "grad_norm": 0.3024400472640991, "learning_rate": 3.0304011411977174e-05, "loss": 1.7423, "step": 2635 }, { "epoch": 0.6365612170973195, "grad_norm": 0.30431923270225525, "learning_rate": 3.026843847807721e-05, "loss": 1.7332, "step": 2636 }, { "epoch": 0.63680270466071, "grad_norm": 0.2827305793762207, "learning_rate": 3.0232877370156004e-05, "loss": 1.8066, "step": 2637 }, { "epoch": 0.6370441922241005, "grad_norm": 0.27585330605506897, "learning_rate": 3.0197328109526862e-05, "loss": 1.74, "step": 2638 }, { "epoch": 0.637285679787491, "grad_norm": 0.29698488116264343, "learning_rate": 3.016179071749598e-05, "loss": 1.765, "step": 2639 }, { "epoch": 0.6375271673508814, "grad_norm": 0.3048704266548157, "learning_rate": 3.012626521536247e-05, "loss": 1.8214, "step": 2640 }, { "epoch": 0.6377686549142719, "grad_norm": 0.2927187383174896, "learning_rate": 3.0090751624418295e-05, "loss": 1.8029, "step": 2641 }, { "epoch": 0.6380101424776624, "grad_norm": 0.29611796140670776, "learning_rate": 3.0055249965948286e-05, "loss": 1.6614, "step": 2642 }, { "epoch": 0.6382516300410529, "grad_norm": 0.29900723695755005, "learning_rate": 3.0019760261230145e-05, "loss": 1.6778, "step": 2643 }, { "epoch": 0.6384931176044434, "grad_norm": 0.29869556427001953, "learning_rate": 2.9984282531534358e-05, "loss": 1.8344, "step": 2644 }, { "epoch": 0.6387346051678339, "grad_norm": 0.2877620458602905, "learning_rate": 2.994881679812427e-05, "loss": 1.6512, "step": 2645 }, { "epoch": 0.6389760927312244, "grad_norm": 0.2879989743232727, "learning_rate": 2.9913363082256057e-05, "loss": 1.6892, "step": 2646 }, { "epoch": 0.6392175802946148, "grad_norm": 0.28560197353363037, "learning_rate": 2.9877921405178622e-05, "loss": 1.7606, "step": 2647 }, { "epoch": 0.6394590678580053, "grad_norm": 0.2680649757385254, "learning_rate": 2.9842491788133718e-05, "loss": 1.416, "step": 2648 }, { "epoch": 0.6397005554213958, "grad_norm": 0.2798928916454315, "learning_rate": 2.980707425235586e-05, "loss": 1.6681, "step": 2649 }, { "epoch": 0.6399420429847863, "grad_norm": 0.3013726770877838, "learning_rate": 2.977166881907229e-05, "loss": 1.8712, "step": 2650 }, { "epoch": 0.6401835305481768, "grad_norm": 0.28890109062194824, "learning_rate": 2.973627550950302e-05, "loss": 1.6756, "step": 2651 }, { "epoch": 0.6404250181115673, "grad_norm": 0.31177380681037903, "learning_rate": 2.970089434486082e-05, "loss": 1.8275, "step": 2652 }, { "epoch": 0.6406665056749578, "grad_norm": 0.29367968440055847, "learning_rate": 2.966552534635111e-05, "loss": 1.6863, "step": 2653 }, { "epoch": 0.6409079932383482, "grad_norm": 0.2822515070438385, "learning_rate": 2.9630168535172074e-05, "loss": 1.7251, "step": 2654 }, { "epoch": 0.6411494808017387, "grad_norm": 0.2834995687007904, "learning_rate": 2.9594823932514614e-05, "loss": 1.7278, "step": 2655 }, { "epoch": 0.6413909683651292, "grad_norm": 0.2786793112754822, "learning_rate": 2.9559491559562226e-05, "loss": 1.7239, "step": 2656 }, { "epoch": 0.6416324559285197, "grad_norm": 0.2937695384025574, "learning_rate": 2.952417143749116e-05, "loss": 1.5963, "step": 2657 }, { "epoch": 0.6418739434919102, "grad_norm": 0.2992871403694153, "learning_rate": 2.9488863587470306e-05, "loss": 1.7809, "step": 2658 }, { "epoch": 0.6421154310553007, "grad_norm": 0.28979378938674927, "learning_rate": 2.945356803066116e-05, "loss": 1.7272, "step": 2659 }, { "epoch": 0.6423569186186912, "grad_norm": 0.29563620686531067, "learning_rate": 2.9418284788217875e-05, "loss": 1.603, "step": 2660 }, { "epoch": 0.6425984061820816, "grad_norm": 0.28843438625335693, "learning_rate": 2.9383013881287256e-05, "loss": 1.739, "step": 2661 }, { "epoch": 0.6428398937454721, "grad_norm": 0.2856224775314331, "learning_rate": 2.9347755331008652e-05, "loss": 1.7827, "step": 2662 }, { "epoch": 0.6430813813088626, "grad_norm": 0.28124549984931946, "learning_rate": 2.931250915851405e-05, "loss": 1.6795, "step": 2663 }, { "epoch": 0.6433228688722531, "grad_norm": 0.2995930314064026, "learning_rate": 2.9277275384928028e-05, "loss": 1.6768, "step": 2664 }, { "epoch": 0.6435643564356436, "grad_norm": 0.3006519675254822, "learning_rate": 2.9242054031367673e-05, "loss": 1.6554, "step": 2665 }, { "epoch": 0.6438058439990341, "grad_norm": 0.2931753098964691, "learning_rate": 2.9206845118942683e-05, "loss": 1.684, "step": 2666 }, { "epoch": 0.6440473315624246, "grad_norm": 0.2884087860584259, "learning_rate": 2.9171648668755302e-05, "loss": 1.6289, "step": 2667 }, { "epoch": 0.644288819125815, "grad_norm": 0.29300758242607117, "learning_rate": 2.9136464701900258e-05, "loss": 1.7902, "step": 2668 }, { "epoch": 0.6445303066892055, "grad_norm": 0.3091634511947632, "learning_rate": 2.9101293239464823e-05, "loss": 1.8098, "step": 2669 }, { "epoch": 0.644771794252596, "grad_norm": 0.2898050844669342, "learning_rate": 2.9066134302528824e-05, "loss": 1.6409, "step": 2670 }, { "epoch": 0.6450132818159865, "grad_norm": 0.30081823468208313, "learning_rate": 2.9030987912164463e-05, "loss": 1.8338, "step": 2671 }, { "epoch": 0.645254769379377, "grad_norm": 0.281520277261734, "learning_rate": 2.8995854089436548e-05, "loss": 1.5724, "step": 2672 }, { "epoch": 0.6454962569427675, "grad_norm": 0.28424423933029175, "learning_rate": 2.8960732855402273e-05, "loss": 1.6813, "step": 2673 }, { "epoch": 0.645737744506158, "grad_norm": 0.29240164160728455, "learning_rate": 2.8925624231111326e-05, "loss": 1.6981, "step": 2674 }, { "epoch": 0.6459792320695484, "grad_norm": 0.29602888226509094, "learning_rate": 2.8890528237605797e-05, "loss": 1.7719, "step": 2675 }, { "epoch": 0.6462207196329389, "grad_norm": 0.2806943356990814, "learning_rate": 2.885544489592028e-05, "loss": 1.5705, "step": 2676 }, { "epoch": 0.6464622071963294, "grad_norm": 0.29543089866638184, "learning_rate": 2.8820374227081715e-05, "loss": 1.7142, "step": 2677 }, { "epoch": 0.6467036947597199, "grad_norm": 0.2848484516143799, "learning_rate": 2.8785316252109468e-05, "loss": 1.5637, "step": 2678 }, { "epoch": 0.6469451823231104, "grad_norm": 0.2866585850715637, "learning_rate": 2.8750270992015304e-05, "loss": 1.6714, "step": 2679 }, { "epoch": 0.6471866698865009, "grad_norm": 0.28150928020477295, "learning_rate": 2.871523846780335e-05, "loss": 1.6729, "step": 2680 }, { "epoch": 0.6474281574498914, "grad_norm": 0.3050326704978943, "learning_rate": 2.868021870047015e-05, "loss": 1.8211, "step": 2681 }, { "epoch": 0.6476696450132818, "grad_norm": 0.3062339425086975, "learning_rate": 2.864521171100455e-05, "loss": 1.7446, "step": 2682 }, { "epoch": 0.6479111325766723, "grad_norm": 0.2874729037284851, "learning_rate": 2.8610217520387738e-05, "loss": 1.6337, "step": 2683 }, { "epoch": 0.6481526201400628, "grad_norm": 0.2956065833568573, "learning_rate": 2.8575236149593244e-05, "loss": 1.7639, "step": 2684 }, { "epoch": 0.6483941077034533, "grad_norm": 0.2926720678806305, "learning_rate": 2.8540267619586947e-05, "loss": 1.8428, "step": 2685 }, { "epoch": 0.6486355952668438, "grad_norm": 0.2935812771320343, "learning_rate": 2.8505311951326992e-05, "loss": 1.6786, "step": 2686 }, { "epoch": 0.6488770828302343, "grad_norm": 0.29179954528808594, "learning_rate": 2.8470369165763812e-05, "loss": 1.686, "step": 2687 }, { "epoch": 0.6491185703936247, "grad_norm": 0.3037336766719818, "learning_rate": 2.8435439283840116e-05, "loss": 1.8595, "step": 2688 }, { "epoch": 0.6493600579570152, "grad_norm": 0.2783583104610443, "learning_rate": 2.840052232649093e-05, "loss": 1.6229, "step": 2689 }, { "epoch": 0.6496015455204057, "grad_norm": 0.30440038442611694, "learning_rate": 2.8365618314643484e-05, "loss": 2.021, "step": 2690 }, { "epoch": 0.6498430330837962, "grad_norm": 0.28783807158470154, "learning_rate": 2.833072726921724e-05, "loss": 1.6471, "step": 2691 }, { "epoch": 0.6500845206471867, "grad_norm": 0.2774794399738312, "learning_rate": 2.829584921112397e-05, "loss": 1.5241, "step": 2692 }, { "epoch": 0.6503260082105772, "grad_norm": 0.3046042025089264, "learning_rate": 2.8260984161267523e-05, "loss": 1.7674, "step": 2693 }, { "epoch": 0.6505674957739677, "grad_norm": 0.2924773097038269, "learning_rate": 2.8226132140544092e-05, "loss": 1.708, "step": 2694 }, { "epoch": 0.6508089833373581, "grad_norm": 0.297407865524292, "learning_rate": 2.819129316984199e-05, "loss": 1.7601, "step": 2695 }, { "epoch": 0.6510504709007486, "grad_norm": 0.30786001682281494, "learning_rate": 2.815646727004171e-05, "loss": 1.816, "step": 2696 }, { "epoch": 0.6512919584641391, "grad_norm": 0.29902076721191406, "learning_rate": 2.81216544620159e-05, "loss": 1.8452, "step": 2697 }, { "epoch": 0.6515334460275296, "grad_norm": 0.2809820771217346, "learning_rate": 2.8086854766629457e-05, "loss": 1.6145, "step": 2698 }, { "epoch": 0.6517749335909201, "grad_norm": 0.28967392444610596, "learning_rate": 2.8052068204739245e-05, "loss": 1.634, "step": 2699 }, { "epoch": 0.6520164211543106, "grad_norm": 0.29875287413597107, "learning_rate": 2.8017294797194427e-05, "loss": 1.7046, "step": 2700 }, { "epoch": 0.6522579087177011, "grad_norm": 0.28991883993148804, "learning_rate": 2.7982534564836187e-05, "loss": 1.6097, "step": 2701 }, { "epoch": 0.6524993962810915, "grad_norm": 0.28873929381370544, "learning_rate": 2.794778752849782e-05, "loss": 1.7679, "step": 2702 }, { "epoch": 0.652740883844482, "grad_norm": 0.32991576194763184, "learning_rate": 2.7913053709004765e-05, "loss": 1.7982, "step": 2703 }, { "epoch": 0.6529823714078725, "grad_norm": 0.2825528085231781, "learning_rate": 2.7878333127174493e-05, "loss": 1.6175, "step": 2704 }, { "epoch": 0.653223858971263, "grad_norm": 0.295487642288208, "learning_rate": 2.784362580381654e-05, "loss": 1.7463, "step": 2705 }, { "epoch": 0.6534653465346535, "grad_norm": 0.2789340019226074, "learning_rate": 2.7808931759732492e-05, "loss": 1.6146, "step": 2706 }, { "epoch": 0.653706834098044, "grad_norm": 0.3013107180595398, "learning_rate": 2.7774251015716048e-05, "loss": 1.8198, "step": 2707 }, { "epoch": 0.6539483216614345, "grad_norm": 0.2862893044948578, "learning_rate": 2.77395835925528e-05, "loss": 1.6245, "step": 2708 }, { "epoch": 0.654189809224825, "grad_norm": 0.28606319427490234, "learning_rate": 2.7704929511020495e-05, "loss": 1.6733, "step": 2709 }, { "epoch": 0.6544312967882154, "grad_norm": 0.2939609885215759, "learning_rate": 2.7670288791888797e-05, "loss": 1.6891, "step": 2710 }, { "epoch": 0.6546727843516059, "grad_norm": 0.29280340671539307, "learning_rate": 2.7635661455919408e-05, "loss": 1.7564, "step": 2711 }, { "epoch": 0.6549142719149964, "grad_norm": 0.28173065185546875, "learning_rate": 2.760104752386595e-05, "loss": 1.6877, "step": 2712 }, { "epoch": 0.6551557594783869, "grad_norm": 0.28756988048553467, "learning_rate": 2.7566447016474106e-05, "loss": 1.6177, "step": 2713 }, { "epoch": 0.6553972470417774, "grad_norm": 0.29960814118385315, "learning_rate": 2.753185995448142e-05, "loss": 1.6857, "step": 2714 }, { "epoch": 0.6556387346051679, "grad_norm": 0.2856014668941498, "learning_rate": 2.7497286358617402e-05, "loss": 1.7174, "step": 2715 }, { "epoch": 0.6558802221685583, "grad_norm": 0.2883416414260864, "learning_rate": 2.7462726249603572e-05, "loss": 1.6354, "step": 2716 }, { "epoch": 0.6561217097319488, "grad_norm": 0.29868680238723755, "learning_rate": 2.7428179648153208e-05, "loss": 1.7685, "step": 2717 }, { "epoch": 0.6563631972953393, "grad_norm": 0.28661859035491943, "learning_rate": 2.7393646574971654e-05, "loss": 1.8062, "step": 2718 }, { "epoch": 0.6566046848587298, "grad_norm": 0.2813142240047455, "learning_rate": 2.7359127050756038e-05, "loss": 1.6258, "step": 2719 }, { "epoch": 0.6568461724221203, "grad_norm": 0.2795809209346771, "learning_rate": 2.732462109619543e-05, "loss": 1.6217, "step": 2720 }, { "epoch": 0.6570876599855108, "grad_norm": 0.2935608923435211, "learning_rate": 2.72901287319707e-05, "loss": 1.6851, "step": 2721 }, { "epoch": 0.6573291475489013, "grad_norm": 0.28350740671157837, "learning_rate": 2.7255649978754666e-05, "loss": 1.589, "step": 2722 }, { "epoch": 0.6575706351122917, "grad_norm": 0.28539225459098816, "learning_rate": 2.722118485721191e-05, "loss": 1.6009, "step": 2723 }, { "epoch": 0.6578121226756822, "grad_norm": 0.30798783898353577, "learning_rate": 2.718673338799886e-05, "loss": 1.7514, "step": 2724 }, { "epoch": 0.6580536102390727, "grad_norm": 0.284024178981781, "learning_rate": 2.715229559176382e-05, "loss": 1.6093, "step": 2725 }, { "epoch": 0.6582950978024632, "grad_norm": 0.29781007766723633, "learning_rate": 2.7117871489146786e-05, "loss": 1.7696, "step": 2726 }, { "epoch": 0.6585365853658537, "grad_norm": 0.2936575710773468, "learning_rate": 2.708346110077966e-05, "loss": 1.8158, "step": 2727 }, { "epoch": 0.6587780729292442, "grad_norm": 0.30539223551750183, "learning_rate": 2.7049064447286064e-05, "loss": 1.8357, "step": 2728 }, { "epoch": 0.6590195604926347, "grad_norm": 0.28084996342658997, "learning_rate": 2.7014681549281397e-05, "loss": 1.5934, "step": 2729 }, { "epoch": 0.6592610480560251, "grad_norm": 0.29941701889038086, "learning_rate": 2.69803124273728e-05, "loss": 1.6693, "step": 2730 }, { "epoch": 0.6595025356194156, "grad_norm": 0.2934275269508362, "learning_rate": 2.6945957102159236e-05, "loss": 1.7669, "step": 2731 }, { "epoch": 0.6597440231828061, "grad_norm": 0.28211256861686707, "learning_rate": 2.691161559423126e-05, "loss": 1.6845, "step": 2732 }, { "epoch": 0.6599855107461966, "grad_norm": 0.5288688540458679, "learning_rate": 2.687728792417128e-05, "loss": 1.8411, "step": 2733 }, { "epoch": 0.6602269983095871, "grad_norm": 0.2935761511325836, "learning_rate": 2.684297411255333e-05, "loss": 1.7228, "step": 2734 }, { "epoch": 0.6604684858729776, "grad_norm": 0.28827327489852905, "learning_rate": 2.680867417994316e-05, "loss": 1.708, "step": 2735 }, { "epoch": 0.6607099734363681, "grad_norm": 0.30115723609924316, "learning_rate": 2.6774388146898223e-05, "loss": 1.7706, "step": 2736 }, { "epoch": 0.6609514609997585, "grad_norm": 0.2931334972381592, "learning_rate": 2.674011603396761e-05, "loss": 1.6886, "step": 2737 }, { "epoch": 0.661192948563149, "grad_norm": 0.29254698753356934, "learning_rate": 2.6705857861692086e-05, "loss": 1.7348, "step": 2738 }, { "epoch": 0.6614344361265395, "grad_norm": 0.2890866994857788, "learning_rate": 2.667161365060403e-05, "loss": 1.7158, "step": 2739 }, { "epoch": 0.66167592368993, "grad_norm": 0.2868228554725647, "learning_rate": 2.6637383421227547e-05, "loss": 1.6048, "step": 2740 }, { "epoch": 0.6619174112533205, "grad_norm": 0.2857600450515747, "learning_rate": 2.6603167194078206e-05, "loss": 1.5711, "step": 2741 }, { "epoch": 0.662158898816711, "grad_norm": 0.29845717549324036, "learning_rate": 2.6568964989663335e-05, "loss": 1.8061, "step": 2742 }, { "epoch": 0.6624003863801015, "grad_norm": 0.31675419211387634, "learning_rate": 2.653477682848179e-05, "loss": 1.8324, "step": 2743 }, { "epoch": 0.6626418739434919, "grad_norm": 0.2821733057498932, "learning_rate": 2.6500602731023993e-05, "loss": 1.6558, "step": 2744 }, { "epoch": 0.6628833615068824, "grad_norm": 0.2879883050918579, "learning_rate": 2.646644271777195e-05, "loss": 1.6916, "step": 2745 }, { "epoch": 0.6631248490702729, "grad_norm": 0.27586695551872253, "learning_rate": 2.6432296809199275e-05, "loss": 1.5031, "step": 2746 }, { "epoch": 0.6633663366336634, "grad_norm": 0.29299119114875793, "learning_rate": 2.639816502577107e-05, "loss": 1.7643, "step": 2747 }, { "epoch": 0.6636078241970539, "grad_norm": 0.289853036403656, "learning_rate": 2.636404738794397e-05, "loss": 1.6881, "step": 2748 }, { "epoch": 0.6638493117604444, "grad_norm": 0.2892284691333771, "learning_rate": 2.6329943916166212e-05, "loss": 1.8007, "step": 2749 }, { "epoch": 0.6640907993238349, "grad_norm": 0.2801893651485443, "learning_rate": 2.629585463087739e-05, "loss": 1.7164, "step": 2750 }, { "epoch": 0.6643322868872253, "grad_norm": 0.2808801829814911, "learning_rate": 2.626177955250877e-05, "loss": 1.6035, "step": 2751 }, { "epoch": 0.6645737744506158, "grad_norm": 0.30470070242881775, "learning_rate": 2.622771870148299e-05, "loss": 1.721, "step": 2752 }, { "epoch": 0.6648152620140063, "grad_norm": 0.28783950209617615, "learning_rate": 2.6193672098214185e-05, "loss": 1.6124, "step": 2753 }, { "epoch": 0.6650567495773968, "grad_norm": 0.30088046193122864, "learning_rate": 2.615963976310795e-05, "loss": 1.6576, "step": 2754 }, { "epoch": 0.6652982371407873, "grad_norm": 0.2879355549812317, "learning_rate": 2.6125621716561376e-05, "loss": 1.5887, "step": 2755 }, { "epoch": 0.6655397247041778, "grad_norm": 0.2904006540775299, "learning_rate": 2.609161797896292e-05, "loss": 1.7002, "step": 2756 }, { "epoch": 0.6657812122675683, "grad_norm": 0.2809663414955139, "learning_rate": 2.6057628570692476e-05, "loss": 1.6963, "step": 2757 }, { "epoch": 0.6660226998309587, "grad_norm": 0.2866246998310089, "learning_rate": 2.6023653512121437e-05, "loss": 1.4988, "step": 2758 }, { "epoch": 0.6662641873943492, "grad_norm": 0.29889899492263794, "learning_rate": 2.5989692823612433e-05, "loss": 1.875, "step": 2759 }, { "epoch": 0.6665056749577397, "grad_norm": 0.2869660556316376, "learning_rate": 2.5955746525519646e-05, "loss": 1.6752, "step": 2760 }, { "epoch": 0.6667471625211302, "grad_norm": 0.33077922463417053, "learning_rate": 2.5921814638188536e-05, "loss": 2.0008, "step": 2761 }, { "epoch": 0.6669886500845207, "grad_norm": 0.28912678360939026, "learning_rate": 2.588789718195595e-05, "loss": 1.6892, "step": 2762 }, { "epoch": 0.6672301376479112, "grad_norm": 0.2845942974090576, "learning_rate": 2.585399417715007e-05, "loss": 1.605, "step": 2763 }, { "epoch": 0.6674716252113017, "grad_norm": 0.28622081875801086, "learning_rate": 2.5820105644090498e-05, "loss": 1.7582, "step": 2764 }, { "epoch": 0.6677131127746921, "grad_norm": 0.29499760270118713, "learning_rate": 2.5786231603088006e-05, "loss": 1.8148, "step": 2765 }, { "epoch": 0.6679546003380826, "grad_norm": 0.30665308237075806, "learning_rate": 2.5752372074444842e-05, "loss": 1.8085, "step": 2766 }, { "epoch": 0.6681960879014731, "grad_norm": 0.29079926013946533, "learning_rate": 2.571852707845447e-05, "loss": 1.7032, "step": 2767 }, { "epoch": 0.6684375754648636, "grad_norm": 0.28521308302879333, "learning_rate": 2.568469663540164e-05, "loss": 1.6796, "step": 2768 }, { "epoch": 0.6686790630282541, "grad_norm": 0.2907143533229828, "learning_rate": 2.5650880765562426e-05, "loss": 1.655, "step": 2769 }, { "epoch": 0.6689205505916446, "grad_norm": 0.31864815950393677, "learning_rate": 2.5617079489204145e-05, "loss": 1.9504, "step": 2770 }, { "epoch": 0.669162038155035, "grad_norm": 0.28453752398490906, "learning_rate": 2.5583292826585342e-05, "loss": 1.6718, "step": 2771 }, { "epoch": 0.6694035257184255, "grad_norm": 0.27700912952423096, "learning_rate": 2.5549520797955816e-05, "loss": 1.5478, "step": 2772 }, { "epoch": 0.669645013281816, "grad_norm": 0.28177401423454285, "learning_rate": 2.5515763423556677e-05, "loss": 1.7419, "step": 2773 }, { "epoch": 0.6698865008452065, "grad_norm": 0.2821323871612549, "learning_rate": 2.548202072362008e-05, "loss": 1.7676, "step": 2774 }, { "epoch": 0.670127988408597, "grad_norm": 0.29551199078559875, "learning_rate": 2.5448292718369564e-05, "loss": 1.6599, "step": 2775 }, { "epoch": 0.6703694759719875, "grad_norm": 0.28306227922439575, "learning_rate": 2.5414579428019746e-05, "loss": 1.6682, "step": 2776 }, { "epoch": 0.670610963535378, "grad_norm": 0.2981685698032379, "learning_rate": 2.5380880872776452e-05, "loss": 1.6846, "step": 2777 }, { "epoch": 0.6708524510987685, "grad_norm": 0.2757263779640198, "learning_rate": 2.534719707283672e-05, "loss": 1.6345, "step": 2778 }, { "epoch": 0.6710939386621589, "grad_norm": 0.286205917596817, "learning_rate": 2.5313528048388686e-05, "loss": 1.6353, "step": 2779 }, { "epoch": 0.6713354262255494, "grad_norm": 0.2774648666381836, "learning_rate": 2.5279873819611656e-05, "loss": 1.6256, "step": 2780 }, { "epoch": 0.6715769137889399, "grad_norm": 0.28847214579582214, "learning_rate": 2.524623440667605e-05, "loss": 1.6521, "step": 2781 }, { "epoch": 0.6718184013523304, "grad_norm": 0.29158225655555725, "learning_rate": 2.521260982974347e-05, "loss": 1.7226, "step": 2782 }, { "epoch": 0.6720598889157209, "grad_norm": 0.33863604068756104, "learning_rate": 2.517900010896652e-05, "loss": 1.6735, "step": 2783 }, { "epoch": 0.6723013764791114, "grad_norm": 0.2932925522327423, "learning_rate": 2.5145405264489008e-05, "loss": 1.7371, "step": 2784 }, { "epoch": 0.6725428640425019, "grad_norm": 0.30178266763687134, "learning_rate": 2.5111825316445764e-05, "loss": 1.5816, "step": 2785 }, { "epoch": 0.6727843516058923, "grad_norm": 0.28754690289497375, "learning_rate": 2.5078260284962697e-05, "loss": 1.6914, "step": 2786 }, { "epoch": 0.6730258391692828, "grad_norm": 0.2963842451572418, "learning_rate": 2.5044710190156767e-05, "loss": 1.7703, "step": 2787 }, { "epoch": 0.6732673267326733, "grad_norm": 0.2981335520744324, "learning_rate": 2.5011175052136037e-05, "loss": 1.7432, "step": 2788 }, { "epoch": 0.6735088142960638, "grad_norm": 0.2900048494338989, "learning_rate": 2.4977654890999548e-05, "loss": 1.6751, "step": 2789 }, { "epoch": 0.6737503018594543, "grad_norm": 0.27933359146118164, "learning_rate": 2.4944149726837364e-05, "loss": 1.7394, "step": 2790 }, { "epoch": 0.6739917894228448, "grad_norm": 0.3018084168434143, "learning_rate": 2.4910659579730633e-05, "loss": 1.7255, "step": 2791 }, { "epoch": 0.6742332769862353, "grad_norm": 0.2868594825267792, "learning_rate": 2.487718446975138e-05, "loss": 1.7346, "step": 2792 }, { "epoch": 0.6744747645496257, "grad_norm": 0.2942366898059845, "learning_rate": 2.484372441696274e-05, "loss": 1.7086, "step": 2793 }, { "epoch": 0.6747162521130162, "grad_norm": 0.28912457823753357, "learning_rate": 2.481027944141875e-05, "loss": 1.6767, "step": 2794 }, { "epoch": 0.6749577396764067, "grad_norm": 0.3127002716064453, "learning_rate": 2.4776849563164435e-05, "loss": 1.7032, "step": 2795 }, { "epoch": 0.6751992272397972, "grad_norm": 0.2829575836658478, "learning_rate": 2.4743434802235744e-05, "loss": 1.6551, "step": 2796 }, { "epoch": 0.6754407148031877, "grad_norm": 0.2902395725250244, "learning_rate": 2.4710035178659657e-05, "loss": 1.8044, "step": 2797 }, { "epoch": 0.6756822023665782, "grad_norm": 0.2812401056289673, "learning_rate": 2.4676650712453924e-05, "loss": 1.6648, "step": 2798 }, { "epoch": 0.6759236899299687, "grad_norm": 0.291409432888031, "learning_rate": 2.4643281423627367e-05, "loss": 1.6908, "step": 2799 }, { "epoch": 0.6761651774933591, "grad_norm": 0.2796201705932617, "learning_rate": 2.4609927332179624e-05, "loss": 1.7502, "step": 2800 }, { "epoch": 0.6764066650567496, "grad_norm": 0.2864021360874176, "learning_rate": 2.4576588458101224e-05, "loss": 1.8332, "step": 2801 }, { "epoch": 0.6766481526201401, "grad_norm": 0.2680753469467163, "learning_rate": 2.4543264821373633e-05, "loss": 1.5199, "step": 2802 }, { "epoch": 0.6768896401835306, "grad_norm": 0.29651933908462524, "learning_rate": 2.450995644196914e-05, "loss": 1.7485, "step": 2803 }, { "epoch": 0.6771311277469211, "grad_norm": 0.2806631326675415, "learning_rate": 2.447666333985089e-05, "loss": 1.6797, "step": 2804 }, { "epoch": 0.6773726153103116, "grad_norm": 0.2897458076477051, "learning_rate": 2.4443385534972863e-05, "loss": 1.7084, "step": 2805 }, { "epoch": 0.677614102873702, "grad_norm": 0.2659716308116913, "learning_rate": 2.4410123047279937e-05, "loss": 1.4833, "step": 2806 }, { "epoch": 0.6778555904370925, "grad_norm": 0.28540828824043274, "learning_rate": 2.4376875896707694e-05, "loss": 1.6145, "step": 2807 }, { "epoch": 0.678097078000483, "grad_norm": 0.3037117123603821, "learning_rate": 2.434364410318265e-05, "loss": 1.9319, "step": 2808 }, { "epoch": 0.6783385655638735, "grad_norm": 0.2886074483394623, "learning_rate": 2.4310427686622028e-05, "loss": 1.7199, "step": 2809 }, { "epoch": 0.678580053127264, "grad_norm": 0.30575114488601685, "learning_rate": 2.4277226666933854e-05, "loss": 1.8807, "step": 2810 }, { "epoch": 0.6788215406906545, "grad_norm": 0.29991793632507324, "learning_rate": 2.4244041064016964e-05, "loss": 1.7737, "step": 2811 }, { "epoch": 0.679063028254045, "grad_norm": 0.28687959909439087, "learning_rate": 2.4210870897760917e-05, "loss": 1.6495, "step": 2812 }, { "epoch": 0.6793045158174355, "grad_norm": 0.2801690399646759, "learning_rate": 2.4177716188046025e-05, "loss": 1.4793, "step": 2813 }, { "epoch": 0.6795460033808259, "grad_norm": 0.2892257869243622, "learning_rate": 2.4144576954743314e-05, "loss": 1.6759, "step": 2814 }, { "epoch": 0.6797874909442164, "grad_norm": 0.28120866417884827, "learning_rate": 2.4111453217714612e-05, "loss": 1.6697, "step": 2815 }, { "epoch": 0.6800289785076069, "grad_norm": 0.2849350869655609, "learning_rate": 2.4078344996812384e-05, "loss": 1.6999, "step": 2816 }, { "epoch": 0.6802704660709974, "grad_norm": 0.2832583487033844, "learning_rate": 2.4045252311879812e-05, "loss": 1.6921, "step": 2817 }, { "epoch": 0.6805119536343879, "grad_norm": 0.31813693046569824, "learning_rate": 2.401217518275077e-05, "loss": 1.8634, "step": 2818 }, { "epoch": 0.6807534411977784, "grad_norm": 0.2877633571624756, "learning_rate": 2.3979113629249854e-05, "loss": 1.7367, "step": 2819 }, { "epoch": 0.6809949287611688, "grad_norm": 0.2952253222465515, "learning_rate": 2.394606767119222e-05, "loss": 1.6737, "step": 2820 }, { "epoch": 0.6812364163245593, "grad_norm": 0.3123829662799835, "learning_rate": 2.3913037328383792e-05, "loss": 1.8997, "step": 2821 }, { "epoch": 0.6814779038879498, "grad_norm": 0.29048192501068115, "learning_rate": 2.3880022620621066e-05, "loss": 1.7051, "step": 2822 }, { "epoch": 0.6817193914513403, "grad_norm": 0.3822292983531952, "learning_rate": 2.3847023567691168e-05, "loss": 1.7814, "step": 2823 }, { "epoch": 0.6819608790147308, "grad_norm": 0.2761949896812439, "learning_rate": 2.3814040189371893e-05, "loss": 1.6193, "step": 2824 }, { "epoch": 0.6822023665781213, "grad_norm": 0.2803972661495209, "learning_rate": 2.378107250543159e-05, "loss": 1.659, "step": 2825 }, { "epoch": 0.6824438541415118, "grad_norm": 0.2949616014957428, "learning_rate": 2.374812053562922e-05, "loss": 1.7682, "step": 2826 }, { "epoch": 0.6826853417049022, "grad_norm": 0.27776139974594116, "learning_rate": 2.3715184299714303e-05, "loss": 1.6747, "step": 2827 }, { "epoch": 0.6829268292682927, "grad_norm": 0.28865471482276917, "learning_rate": 2.3682263817427015e-05, "loss": 1.6654, "step": 2828 }, { "epoch": 0.6831683168316832, "grad_norm": 0.2939668893814087, "learning_rate": 2.3649359108497936e-05, "loss": 1.739, "step": 2829 }, { "epoch": 0.6834098043950737, "grad_norm": 0.2937363088130951, "learning_rate": 2.3616470192648348e-05, "loss": 1.7261, "step": 2830 }, { "epoch": 0.6836512919584642, "grad_norm": 0.310563862323761, "learning_rate": 2.358359708958998e-05, "loss": 1.8524, "step": 2831 }, { "epoch": 0.6838927795218547, "grad_norm": 0.30350807309150696, "learning_rate": 2.3550739819025113e-05, "loss": 1.6683, "step": 2832 }, { "epoch": 0.6841342670852452, "grad_norm": 0.2919284403324127, "learning_rate": 2.3517898400646498e-05, "loss": 1.8232, "step": 2833 }, { "epoch": 0.6843757546486356, "grad_norm": 0.28464943170547485, "learning_rate": 2.348507285413747e-05, "loss": 1.5885, "step": 2834 }, { "epoch": 0.6846172422120261, "grad_norm": 0.2740783989429474, "learning_rate": 2.3452263199171776e-05, "loss": 1.6082, "step": 2835 }, { "epoch": 0.6848587297754166, "grad_norm": 0.2829441428184509, "learning_rate": 2.3419469455413634e-05, "loss": 1.7187, "step": 2836 }, { "epoch": 0.6851002173388071, "grad_norm": 0.2870274782180786, "learning_rate": 2.3386691642517828e-05, "loss": 1.6927, "step": 2837 }, { "epoch": 0.6853417049021976, "grad_norm": 0.29190465807914734, "learning_rate": 2.335392978012943e-05, "loss": 1.815, "step": 2838 }, { "epoch": 0.6855831924655881, "grad_norm": 0.2841876149177551, "learning_rate": 2.3321183887884103e-05, "loss": 1.6733, "step": 2839 }, { "epoch": 0.6858246800289786, "grad_norm": 0.2819958031177521, "learning_rate": 2.328845398540786e-05, "loss": 1.5887, "step": 2840 }, { "epoch": 0.686066167592369, "grad_norm": 0.2854621410369873, "learning_rate": 2.3255740092317146e-05, "loss": 1.644, "step": 2841 }, { "epoch": 0.6863076551557595, "grad_norm": 0.2923666834831238, "learning_rate": 2.3223042228218793e-05, "loss": 1.709, "step": 2842 }, { "epoch": 0.68654914271915, "grad_norm": 0.26582345366477966, "learning_rate": 2.319036041271009e-05, "loss": 1.509, "step": 2843 }, { "epoch": 0.6867906302825405, "grad_norm": 0.291042685508728, "learning_rate": 2.3157694665378638e-05, "loss": 1.6484, "step": 2844 }, { "epoch": 0.687032117845931, "grad_norm": 0.28579196333885193, "learning_rate": 2.3125045005802436e-05, "loss": 1.7326, "step": 2845 }, { "epoch": 0.6872736054093215, "grad_norm": 0.298378050327301, "learning_rate": 2.309241145354984e-05, "loss": 1.705, "step": 2846 }, { "epoch": 0.6875150929727118, "grad_norm": 0.2950095534324646, "learning_rate": 2.305979402817952e-05, "loss": 1.8238, "step": 2847 }, { "epoch": 0.6877565805361023, "grad_norm": 0.2783282399177551, "learning_rate": 2.3027192749240568e-05, "loss": 1.7191, "step": 2848 }, { "epoch": 0.6879980680994928, "grad_norm": 0.29382461309432983, "learning_rate": 2.2994607636272303e-05, "loss": 1.788, "step": 2849 }, { "epoch": 0.6882395556628833, "grad_norm": 0.2981439232826233, "learning_rate": 2.2962038708804406e-05, "loss": 1.8092, "step": 2850 }, { "epoch": 0.6884810432262738, "grad_norm": 0.2787988781929016, "learning_rate": 2.2929485986356823e-05, "loss": 1.7334, "step": 2851 }, { "epoch": 0.6887225307896643, "grad_norm": 0.2923099994659424, "learning_rate": 2.289694948843986e-05, "loss": 1.7616, "step": 2852 }, { "epoch": 0.6889640183530548, "grad_norm": 0.2824375629425049, "learning_rate": 2.2864429234553984e-05, "loss": 1.6833, "step": 2853 }, { "epoch": 0.6892055059164452, "grad_norm": 0.2824935019016266, "learning_rate": 2.2831925244190033e-05, "loss": 1.7743, "step": 2854 }, { "epoch": 0.6894469934798357, "grad_norm": 0.284822016954422, "learning_rate": 2.279943753682905e-05, "loss": 1.6294, "step": 2855 }, { "epoch": 0.6896884810432262, "grad_norm": 0.29414719343185425, "learning_rate": 2.2766966131942296e-05, "loss": 1.6617, "step": 2856 }, { "epoch": 0.6899299686066167, "grad_norm": 0.2843882143497467, "learning_rate": 2.273451104899133e-05, "loss": 1.6879, "step": 2857 }, { "epoch": 0.6901714561700072, "grad_norm": 0.2787471115589142, "learning_rate": 2.2702072307427875e-05, "loss": 1.6709, "step": 2858 }, { "epoch": 0.6904129437333977, "grad_norm": 0.2877217233181, "learning_rate": 2.266964992669386e-05, "loss": 1.6645, "step": 2859 }, { "epoch": 0.6906544312967882, "grad_norm": 0.2946026027202606, "learning_rate": 2.263724392622141e-05, "loss": 1.7734, "step": 2860 }, { "epoch": 0.6908959188601786, "grad_norm": 0.28651416301727295, "learning_rate": 2.26048543254329e-05, "loss": 1.6474, "step": 2861 }, { "epoch": 0.6911374064235691, "grad_norm": 0.2742198407649994, "learning_rate": 2.257248114374074e-05, "loss": 1.5986, "step": 2862 }, { "epoch": 0.6913788939869596, "grad_norm": 0.2896310091018677, "learning_rate": 2.254012440054764e-05, "loss": 1.7155, "step": 2863 }, { "epoch": 0.6916203815503501, "grad_norm": 0.2856878638267517, "learning_rate": 2.250778411524638e-05, "loss": 1.7578, "step": 2864 }, { "epoch": 0.6918618691137406, "grad_norm": 0.27023351192474365, "learning_rate": 2.2475460307219882e-05, "loss": 1.4724, "step": 2865 }, { "epoch": 0.6921033566771311, "grad_norm": 0.2988951802253723, "learning_rate": 2.2443152995841194e-05, "loss": 1.7878, "step": 2866 }, { "epoch": 0.6923448442405216, "grad_norm": 0.2986622154712677, "learning_rate": 2.2410862200473514e-05, "loss": 1.7905, "step": 2867 }, { "epoch": 0.692586331803912, "grad_norm": 0.28369197249412537, "learning_rate": 2.2378587940470108e-05, "loss": 1.6016, "step": 2868 }, { "epoch": 0.6928278193673025, "grad_norm": 0.28344225883483887, "learning_rate": 2.2346330235174305e-05, "loss": 1.573, "step": 2869 }, { "epoch": 0.693069306930693, "grad_norm": 0.28998225927352905, "learning_rate": 2.231408910391961e-05, "loss": 1.7242, "step": 2870 }, { "epoch": 0.6933107944940835, "grad_norm": 0.28246867656707764, "learning_rate": 2.2281864566029448e-05, "loss": 1.7316, "step": 2871 }, { "epoch": 0.693552282057474, "grad_norm": 0.2980121970176697, "learning_rate": 2.2249656640817434e-05, "loss": 1.7163, "step": 2872 }, { "epoch": 0.6937937696208645, "grad_norm": 0.3020820617675781, "learning_rate": 2.221746534758716e-05, "loss": 1.6339, "step": 2873 }, { "epoch": 0.694035257184255, "grad_norm": 0.32497313618659973, "learning_rate": 2.218529070563225e-05, "loss": 1.4859, "step": 2874 }, { "epoch": 0.6942767447476454, "grad_norm": 0.2932278513908386, "learning_rate": 2.215313273423635e-05, "loss": 1.7974, "step": 2875 }, { "epoch": 0.6945182323110359, "grad_norm": 0.2885207235813141, "learning_rate": 2.2120991452673156e-05, "loss": 1.6916, "step": 2876 }, { "epoch": 0.6947597198744264, "grad_norm": 0.29947951436042786, "learning_rate": 2.2088866880206317e-05, "loss": 1.8383, "step": 2877 }, { "epoch": 0.6950012074378169, "grad_norm": 0.29510560631752014, "learning_rate": 2.2056759036089468e-05, "loss": 1.7765, "step": 2878 }, { "epoch": 0.6952426950012074, "grad_norm": 0.2835674583911896, "learning_rate": 2.202466793956624e-05, "loss": 1.6373, "step": 2879 }, { "epoch": 0.6954841825645979, "grad_norm": 0.314756840467453, "learning_rate": 2.1992593609870198e-05, "loss": 1.8121, "step": 2880 }, { "epoch": 0.6957256701279884, "grad_norm": 0.30321574211120605, "learning_rate": 2.1960536066224908e-05, "loss": 1.6039, "step": 2881 }, { "epoch": 0.6959671576913788, "grad_norm": 0.29521188139915466, "learning_rate": 2.192849532784382e-05, "loss": 1.7304, "step": 2882 }, { "epoch": 0.6962086452547693, "grad_norm": 0.2963431179523468, "learning_rate": 2.189647141393034e-05, "loss": 1.6935, "step": 2883 }, { "epoch": 0.6964501328181598, "grad_norm": 0.3004039525985718, "learning_rate": 2.1864464343677776e-05, "loss": 1.7687, "step": 2884 }, { "epoch": 0.6966916203815503, "grad_norm": 0.2856568694114685, "learning_rate": 2.1832474136269394e-05, "loss": 1.6993, "step": 2885 }, { "epoch": 0.6969331079449408, "grad_norm": 0.2841035723686218, "learning_rate": 2.180050081087824e-05, "loss": 1.534, "step": 2886 }, { "epoch": 0.6971745955083313, "grad_norm": 0.2901994287967682, "learning_rate": 2.1768544386667363e-05, "loss": 1.6598, "step": 2887 }, { "epoch": 0.6974160830717218, "grad_norm": 0.28846752643585205, "learning_rate": 2.173660488278962e-05, "loss": 1.6666, "step": 2888 }, { "epoch": 0.6976575706351122, "grad_norm": 0.31415724754333496, "learning_rate": 2.1704682318387708e-05, "loss": 1.7523, "step": 2889 }, { "epoch": 0.6978990581985027, "grad_norm": 0.2872610092163086, "learning_rate": 2.1672776712594258e-05, "loss": 1.6344, "step": 2890 }, { "epoch": 0.6981405457618932, "grad_norm": 0.29688772559165955, "learning_rate": 2.164088808453164e-05, "loss": 1.8173, "step": 2891 }, { "epoch": 0.6983820333252837, "grad_norm": 0.2823156714439392, "learning_rate": 2.1609016453312104e-05, "loss": 1.6636, "step": 2892 }, { "epoch": 0.6986235208886742, "grad_norm": 0.28919029235839844, "learning_rate": 2.1577161838037668e-05, "loss": 1.7281, "step": 2893 }, { "epoch": 0.6988650084520647, "grad_norm": 0.2961343228816986, "learning_rate": 2.154532425780025e-05, "loss": 1.8594, "step": 2894 }, { "epoch": 0.6991064960154552, "grad_norm": 0.2904617190361023, "learning_rate": 2.15135037316814e-05, "loss": 1.7012, "step": 2895 }, { "epoch": 0.6993479835788456, "grad_norm": 0.30171629786491394, "learning_rate": 2.1481700278752608e-05, "loss": 1.6433, "step": 2896 }, { "epoch": 0.6995894711422361, "grad_norm": 0.39154428243637085, "learning_rate": 2.1449913918075032e-05, "loss": 1.8341, "step": 2897 }, { "epoch": 0.6998309587056266, "grad_norm": 0.2805890142917633, "learning_rate": 2.1418144668699614e-05, "loss": 1.6159, "step": 2898 }, { "epoch": 0.7000724462690171, "grad_norm": 0.2880823314189911, "learning_rate": 2.138639254966703e-05, "loss": 1.7342, "step": 2899 }, { "epoch": 0.7003139338324076, "grad_norm": 0.2888581156730652, "learning_rate": 2.135465758000772e-05, "loss": 1.6567, "step": 2900 }, { "epoch": 0.7005554213957981, "grad_norm": 0.2792738378047943, "learning_rate": 2.132293977874182e-05, "loss": 1.627, "step": 2901 }, { "epoch": 0.7007969089591886, "grad_norm": 0.28728029131889343, "learning_rate": 2.129123916487915e-05, "loss": 1.6944, "step": 2902 }, { "epoch": 0.701038396522579, "grad_norm": 0.2757112681865692, "learning_rate": 2.125955575741933e-05, "loss": 1.4853, "step": 2903 }, { "epoch": 0.7012798840859695, "grad_norm": 0.2849709689617157, "learning_rate": 2.1227889575351517e-05, "loss": 1.8135, "step": 2904 }, { "epoch": 0.70152137164936, "grad_norm": 0.2988278865814209, "learning_rate": 2.1196240637654678e-05, "loss": 1.741, "step": 2905 }, { "epoch": 0.7017628592127505, "grad_norm": 0.2675354480743408, "learning_rate": 2.1164608963297373e-05, "loss": 1.6658, "step": 2906 }, { "epoch": 0.702004346776141, "grad_norm": 0.3158419728279114, "learning_rate": 2.113299457123783e-05, "loss": 1.9596, "step": 2907 }, { "epoch": 0.7022458343395315, "grad_norm": 0.28420954942703247, "learning_rate": 2.110139748042391e-05, "loss": 1.794, "step": 2908 }, { "epoch": 0.702487321902922, "grad_norm": 0.28654980659484863, "learning_rate": 2.1069817709793155e-05, "loss": 1.7874, "step": 2909 }, { "epoch": 0.7027288094663124, "grad_norm": 0.29616981744766235, "learning_rate": 2.1038255278272666e-05, "loss": 1.7427, "step": 2910 }, { "epoch": 0.7029702970297029, "grad_norm": 0.2863653302192688, "learning_rate": 2.1006710204779174e-05, "loss": 1.6713, "step": 2911 }, { "epoch": 0.7032117845930934, "grad_norm": 0.29114967584609985, "learning_rate": 2.0975182508219005e-05, "loss": 1.7114, "step": 2912 }, { "epoch": 0.7034532721564839, "grad_norm": 0.3041740357875824, "learning_rate": 2.0943672207488057e-05, "loss": 1.7098, "step": 2913 }, { "epoch": 0.7036947597198744, "grad_norm": 0.28644055128097534, "learning_rate": 2.091217932147186e-05, "loss": 1.6648, "step": 2914 }, { "epoch": 0.7039362472832649, "grad_norm": 0.2934432625770569, "learning_rate": 2.0880703869045426e-05, "loss": 1.7399, "step": 2915 }, { "epoch": 0.7041777348466554, "grad_norm": 0.28152257204055786, "learning_rate": 2.0849245869073375e-05, "loss": 1.6608, "step": 2916 }, { "epoch": 0.7044192224100458, "grad_norm": 0.2937367260456085, "learning_rate": 2.081780534040982e-05, "loss": 1.7698, "step": 2917 }, { "epoch": 0.7046607099734363, "grad_norm": 0.29193854331970215, "learning_rate": 2.078638230189849e-05, "loss": 1.7386, "step": 2918 }, { "epoch": 0.7049021975368268, "grad_norm": 0.28087425231933594, "learning_rate": 2.0754976772372493e-05, "loss": 1.7737, "step": 2919 }, { "epoch": 0.7051436851002173, "grad_norm": 0.2874087393283844, "learning_rate": 2.0723588770654577e-05, "loss": 1.6664, "step": 2920 }, { "epoch": 0.7053851726636078, "grad_norm": 0.27303439378738403, "learning_rate": 2.069221831555692e-05, "loss": 1.605, "step": 2921 }, { "epoch": 0.7056266602269983, "grad_norm": 0.28451457619667053, "learning_rate": 2.066086542588117e-05, "loss": 1.6653, "step": 2922 }, { "epoch": 0.7058681477903888, "grad_norm": 0.2788723111152649, "learning_rate": 2.0629530120418505e-05, "loss": 1.7068, "step": 2923 }, { "epoch": 0.7061096353537792, "grad_norm": 0.27940279245376587, "learning_rate": 2.0598212417949514e-05, "loss": 1.6425, "step": 2924 }, { "epoch": 0.7063511229171697, "grad_norm": 0.2720695734024048, "learning_rate": 2.0566912337244248e-05, "loss": 1.5881, "step": 2925 }, { "epoch": 0.7065926104805602, "grad_norm": 0.2986637055873871, "learning_rate": 2.0535629897062186e-05, "loss": 1.7799, "step": 2926 }, { "epoch": 0.7068340980439507, "grad_norm": 0.29969969391822815, "learning_rate": 2.05043651161523e-05, "loss": 1.7627, "step": 2927 }, { "epoch": 0.7070755856073412, "grad_norm": 0.2902384102344513, "learning_rate": 2.0473118013252855e-05, "loss": 1.7182, "step": 2928 }, { "epoch": 0.7073170731707317, "grad_norm": 0.28835979104042053, "learning_rate": 2.0441888607091653e-05, "loss": 1.7554, "step": 2929 }, { "epoch": 0.7075585607341222, "grad_norm": 0.28563472628593445, "learning_rate": 2.0410676916385794e-05, "loss": 1.65, "step": 2930 }, { "epoch": 0.7078000482975126, "grad_norm": 0.27859583497047424, "learning_rate": 2.0379482959841813e-05, "loss": 1.5977, "step": 2931 }, { "epoch": 0.7080415358609031, "grad_norm": 0.2961641252040863, "learning_rate": 2.0348306756155568e-05, "loss": 1.7207, "step": 2932 }, { "epoch": 0.7082830234242936, "grad_norm": 0.30430692434310913, "learning_rate": 2.0317148324012346e-05, "loss": 1.7681, "step": 2933 }, { "epoch": 0.7085245109876841, "grad_norm": 0.31176868081092834, "learning_rate": 2.0286007682086733e-05, "loss": 2.0429, "step": 2934 }, { "epoch": 0.7087659985510746, "grad_norm": 0.28726351261138916, "learning_rate": 2.0254884849042632e-05, "loss": 1.7747, "step": 2935 }, { "epoch": 0.7090074861144651, "grad_norm": 0.2892892360687256, "learning_rate": 2.022377984353337e-05, "loss": 1.7164, "step": 2936 }, { "epoch": 0.7092489736778556, "grad_norm": 0.29728206992149353, "learning_rate": 2.0192692684201443e-05, "loss": 1.7069, "step": 2937 }, { "epoch": 0.709490461241246, "grad_norm": 0.2823873460292816, "learning_rate": 2.0161623389678785e-05, "loss": 1.6182, "step": 2938 }, { "epoch": 0.7097319488046365, "grad_norm": 0.2865864932537079, "learning_rate": 2.0130571978586543e-05, "loss": 1.605, "step": 2939 }, { "epoch": 0.709973436368027, "grad_norm": 0.27976319193840027, "learning_rate": 2.009953846953521e-05, "loss": 1.577, "step": 2940 }, { "epoch": 0.7102149239314175, "grad_norm": 0.2827305793762207, "learning_rate": 2.0068522881124458e-05, "loss": 1.7148, "step": 2941 }, { "epoch": 0.710456411494808, "grad_norm": 0.2937701940536499, "learning_rate": 2.0037525231943304e-05, "loss": 1.8237, "step": 2942 }, { "epoch": 0.7106978990581985, "grad_norm": 0.28721216320991516, "learning_rate": 2.000654554056998e-05, "loss": 1.7646, "step": 2943 }, { "epoch": 0.710939386621589, "grad_norm": 0.282754123210907, "learning_rate": 1.9975583825571947e-05, "loss": 1.7393, "step": 2944 }, { "epoch": 0.7111808741849794, "grad_norm": 0.2946651875972748, "learning_rate": 1.9944640105505887e-05, "loss": 1.7832, "step": 2945 }, { "epoch": 0.7114223617483699, "grad_norm": 0.3008608818054199, "learning_rate": 1.9913714398917743e-05, "loss": 1.6468, "step": 2946 }, { "epoch": 0.7116638493117604, "grad_norm": 0.5344873666763306, "learning_rate": 1.9882806724342612e-05, "loss": 1.5665, "step": 2947 }, { "epoch": 0.7119053368751509, "grad_norm": 0.2858935594558716, "learning_rate": 1.9851917100304783e-05, "loss": 1.6571, "step": 2948 }, { "epoch": 0.7121468244385414, "grad_norm": 0.2995569705963135, "learning_rate": 1.98210455453178e-05, "loss": 1.7285, "step": 2949 }, { "epoch": 0.7123883120019319, "grad_norm": 0.2990138828754425, "learning_rate": 1.979019207788425e-05, "loss": 1.7486, "step": 2950 }, { "epoch": 0.7126297995653224, "grad_norm": 0.2903957664966583, "learning_rate": 1.9759356716496004e-05, "loss": 1.6629, "step": 2951 }, { "epoch": 0.7128712871287128, "grad_norm": 0.2701815068721771, "learning_rate": 1.972853947963401e-05, "loss": 1.4822, "step": 2952 }, { "epoch": 0.7131127746921033, "grad_norm": 0.348672479391098, "learning_rate": 1.969774038576837e-05, "loss": 1.9282, "step": 2953 }, { "epoch": 0.7133542622554938, "grad_norm": 0.27178075909614563, "learning_rate": 1.9666959453358303e-05, "loss": 1.579, "step": 2954 }, { "epoch": 0.7135957498188843, "grad_norm": 0.2898945212364197, "learning_rate": 1.963619670085219e-05, "loss": 1.624, "step": 2955 }, { "epoch": 0.7138372373822748, "grad_norm": 0.3034437596797943, "learning_rate": 1.9605452146687447e-05, "loss": 1.7628, "step": 2956 }, { "epoch": 0.7140787249456653, "grad_norm": 0.3096582293510437, "learning_rate": 1.957472580929062e-05, "loss": 1.8434, "step": 2957 }, { "epoch": 0.7143202125090558, "grad_norm": 0.2902306020259857, "learning_rate": 1.9544017707077362e-05, "loss": 1.742, "step": 2958 }, { "epoch": 0.7145617000724462, "grad_norm": 0.2872730791568756, "learning_rate": 1.9513327858452312e-05, "loss": 1.8039, "step": 2959 }, { "epoch": 0.7148031876358367, "grad_norm": 0.27818602323532104, "learning_rate": 1.948265628180928e-05, "loss": 1.7027, "step": 2960 }, { "epoch": 0.7150446751992272, "grad_norm": 0.28905409574508667, "learning_rate": 1.9452002995531045e-05, "loss": 1.6897, "step": 2961 }, { "epoch": 0.7152861627626177, "grad_norm": 0.28164663910865784, "learning_rate": 1.942136801798945e-05, "loss": 1.6795, "step": 2962 }, { "epoch": 0.7155276503260082, "grad_norm": 0.2943513095378876, "learning_rate": 1.939075136754534e-05, "loss": 1.6836, "step": 2963 }, { "epoch": 0.7157691378893987, "grad_norm": 0.28039127588272095, "learning_rate": 1.9360153062548663e-05, "loss": 1.6606, "step": 2964 }, { "epoch": 0.7160106254527892, "grad_norm": 0.2990105152130127, "learning_rate": 1.9329573121338223e-05, "loss": 1.7438, "step": 2965 }, { "epoch": 0.7162521130161796, "grad_norm": 0.2838454246520996, "learning_rate": 1.929901156224197e-05, "loss": 1.6559, "step": 2966 }, { "epoch": 0.7164936005795701, "grad_norm": 0.29628652334213257, "learning_rate": 1.9268468403576738e-05, "loss": 1.8589, "step": 2967 }, { "epoch": 0.7167350881429606, "grad_norm": 0.28628724813461304, "learning_rate": 1.923794366364835e-05, "loss": 1.7211, "step": 2968 }, { "epoch": 0.7169765757063511, "grad_norm": 0.29965096712112427, "learning_rate": 1.9207437360751647e-05, "loss": 1.8358, "step": 2969 }, { "epoch": 0.7172180632697416, "grad_norm": 0.293460875749588, "learning_rate": 1.9176949513170344e-05, "loss": 1.6512, "step": 2970 }, { "epoch": 0.7174595508331321, "grad_norm": 0.2852771282196045, "learning_rate": 1.914648013917714e-05, "loss": 1.5923, "step": 2971 }, { "epoch": 0.7177010383965226, "grad_norm": 0.2926000952720642, "learning_rate": 1.911602925703363e-05, "loss": 1.704, "step": 2972 }, { "epoch": 0.717942525959913, "grad_norm": 0.3050646483898163, "learning_rate": 1.90855968849904e-05, "loss": 1.8044, "step": 2973 }, { "epoch": 0.7181840135233035, "grad_norm": 0.3000335693359375, "learning_rate": 1.9055183041286805e-05, "loss": 1.6992, "step": 2974 }, { "epoch": 0.718425501086694, "grad_norm": 0.2953099310398102, "learning_rate": 1.902478774415124e-05, "loss": 1.7251, "step": 2975 }, { "epoch": 0.7186669886500845, "grad_norm": 0.34150078892707825, "learning_rate": 1.8994411011800906e-05, "loss": 1.6863, "step": 2976 }, { "epoch": 0.718908476213475, "grad_norm": 0.29604944586753845, "learning_rate": 1.896405286244189e-05, "loss": 1.8018, "step": 2977 }, { "epoch": 0.7191499637768655, "grad_norm": 0.2912079095840454, "learning_rate": 1.8933713314269126e-05, "loss": 1.7717, "step": 2978 }, { "epoch": 0.719391451340256, "grad_norm": 0.30076339840888977, "learning_rate": 1.8903392385466457e-05, "loss": 1.7101, "step": 2979 }, { "epoch": 0.7196329389036464, "grad_norm": 0.29625046253204346, "learning_rate": 1.887309009420651e-05, "loss": 1.7215, "step": 2980 }, { "epoch": 0.7198744264670369, "grad_norm": 0.2872282564640045, "learning_rate": 1.8842806458650735e-05, "loss": 1.6604, "step": 2981 }, { "epoch": 0.7201159140304274, "grad_norm": 0.2833510935306549, "learning_rate": 1.881254149694949e-05, "loss": 1.6415, "step": 2982 }, { "epoch": 0.7203574015938179, "grad_norm": 0.276692271232605, "learning_rate": 1.8782295227241797e-05, "loss": 1.6489, "step": 2983 }, { "epoch": 0.7205988891572084, "grad_norm": 0.30272284150123596, "learning_rate": 1.875206766765561e-05, "loss": 1.7359, "step": 2984 }, { "epoch": 0.7208403767205989, "grad_norm": 0.2910572588443756, "learning_rate": 1.8721858836307592e-05, "loss": 1.6977, "step": 2985 }, { "epoch": 0.7210818642839893, "grad_norm": 0.27629804611206055, "learning_rate": 1.869166875130321e-05, "loss": 1.5098, "step": 2986 }, { "epoch": 0.7213233518473798, "grad_norm": 0.2985958456993103, "learning_rate": 1.866149743073667e-05, "loss": 1.6716, "step": 2987 }, { "epoch": 0.7215648394107703, "grad_norm": 0.2775895297527313, "learning_rate": 1.8631344892690977e-05, "loss": 1.729, "step": 2988 }, { "epoch": 0.7218063269741608, "grad_norm": 0.47814857959747314, "learning_rate": 1.860121115523784e-05, "loss": 1.7696, "step": 2989 }, { "epoch": 0.7220478145375513, "grad_norm": 0.28711941838264465, "learning_rate": 1.8571096236437702e-05, "loss": 1.7315, "step": 2990 }, { "epoch": 0.7222893021009418, "grad_norm": 0.28282269835472107, "learning_rate": 1.854100015433978e-05, "loss": 1.5714, "step": 2991 }, { "epoch": 0.7225307896643323, "grad_norm": 0.3048555254936218, "learning_rate": 1.851092292698189e-05, "loss": 1.5682, "step": 2992 }, { "epoch": 0.7227722772277227, "grad_norm": 0.2775809168815613, "learning_rate": 1.8480864572390665e-05, "loss": 1.6345, "step": 2993 }, { "epoch": 0.7230137647911132, "grad_norm": 0.2935445308685303, "learning_rate": 1.845082510858137e-05, "loss": 1.7516, "step": 2994 }, { "epoch": 0.7232552523545037, "grad_norm": 0.297763854265213, "learning_rate": 1.8420804553557948e-05, "loss": 1.6333, "step": 2995 }, { "epoch": 0.7234967399178942, "grad_norm": 0.2975766062736511, "learning_rate": 1.8390802925313004e-05, "loss": 1.7363, "step": 2996 }, { "epoch": 0.7237382274812847, "grad_norm": 0.28656622767448425, "learning_rate": 1.836082024182786e-05, "loss": 1.5584, "step": 2997 }, { "epoch": 0.7239797150446752, "grad_norm": 0.3044784665107727, "learning_rate": 1.833085652107237e-05, "loss": 1.945, "step": 2998 }, { "epoch": 0.7242212026080657, "grad_norm": 0.2924712598323822, "learning_rate": 1.8300911781005136e-05, "loss": 1.7682, "step": 2999 }, { "epoch": 0.7244626901714561, "grad_norm": 0.2927478849887848, "learning_rate": 1.8270986039573322e-05, "loss": 1.8581, "step": 3000 }, { "epoch": 0.7247041777348466, "grad_norm": 0.28625860810279846, "learning_rate": 1.8241079314712713e-05, "loss": 1.7517, "step": 3001 }, { "epoch": 0.7249456652982371, "grad_norm": 0.2834714651107788, "learning_rate": 1.821119162434772e-05, "loss": 1.7576, "step": 3002 }, { "epoch": 0.7251871528616276, "grad_norm": 0.3014164865016937, "learning_rate": 1.818132298639132e-05, "loss": 1.8837, "step": 3003 }, { "epoch": 0.7254286404250181, "grad_norm": 0.28558745980262756, "learning_rate": 1.815147341874508e-05, "loss": 1.5874, "step": 3004 }, { "epoch": 0.7256701279884086, "grad_norm": 0.293117880821228, "learning_rate": 1.8121642939299122e-05, "loss": 1.6249, "step": 3005 }, { "epoch": 0.7259116155517991, "grad_norm": 0.29122838377952576, "learning_rate": 1.8091831565932192e-05, "loss": 1.8675, "step": 3006 }, { "epoch": 0.7261531031151895, "grad_norm": 0.28399384021759033, "learning_rate": 1.8062039316511477e-05, "loss": 1.8138, "step": 3007 }, { "epoch": 0.72639459067858, "grad_norm": 0.29282137751579285, "learning_rate": 1.8032266208892796e-05, "loss": 1.7707, "step": 3008 }, { "epoch": 0.7266360782419705, "grad_norm": 0.28476330637931824, "learning_rate": 1.8002512260920458e-05, "loss": 1.6447, "step": 3009 }, { "epoch": 0.726877565805361, "grad_norm": 0.30083781480789185, "learning_rate": 1.7972777490427268e-05, "loss": 1.8612, "step": 3010 }, { "epoch": 0.7271190533687515, "grad_norm": 0.28111714124679565, "learning_rate": 1.7943061915234598e-05, "loss": 1.6507, "step": 3011 }, { "epoch": 0.727360540932142, "grad_norm": 0.27549225091934204, "learning_rate": 1.7913365553152255e-05, "loss": 1.4946, "step": 3012 }, { "epoch": 0.7276020284955325, "grad_norm": 0.27884441614151, "learning_rate": 1.7883688421978567e-05, "loss": 1.5774, "step": 3013 }, { "epoch": 0.727843516058923, "grad_norm": 0.2881878614425659, "learning_rate": 1.7854030539500293e-05, "loss": 1.7279, "step": 3014 }, { "epoch": 0.7280850036223134, "grad_norm": 0.2941453456878662, "learning_rate": 1.7824391923492738e-05, "loss": 1.6974, "step": 3015 }, { "epoch": 0.7283264911857039, "grad_norm": 0.2675376236438751, "learning_rate": 1.779477259171955e-05, "loss": 1.523, "step": 3016 }, { "epoch": 0.7285679787490944, "grad_norm": 0.2774561643600464, "learning_rate": 1.7765172561932913e-05, "loss": 1.5871, "step": 3017 }, { "epoch": 0.7288094663124849, "grad_norm": 0.3005836308002472, "learning_rate": 1.77355918518734e-05, "loss": 1.6879, "step": 3018 }, { "epoch": 0.7290509538758754, "grad_norm": 0.28712427616119385, "learning_rate": 1.7706030479270003e-05, "loss": 1.7204, "step": 3019 }, { "epoch": 0.7292924414392659, "grad_norm": 0.2803870737552643, "learning_rate": 1.7676488461840123e-05, "loss": 1.8017, "step": 3020 }, { "epoch": 0.7295339290026563, "grad_norm": 0.2653166353702545, "learning_rate": 1.7646965817289595e-05, "loss": 1.4845, "step": 3021 }, { "epoch": 0.7297754165660468, "grad_norm": 0.28554877638816833, "learning_rate": 1.7617462563312602e-05, "loss": 1.713, "step": 3022 }, { "epoch": 0.7300169041294373, "grad_norm": 0.28936368227005005, "learning_rate": 1.758797871759171e-05, "loss": 1.6205, "step": 3023 }, { "epoch": 0.7302583916928278, "grad_norm": 0.30533939599990845, "learning_rate": 1.755851429779792e-05, "loss": 1.8041, "step": 3024 }, { "epoch": 0.7304998792562183, "grad_norm": 0.2854176163673401, "learning_rate": 1.752906932159044e-05, "loss": 1.5408, "step": 3025 }, { "epoch": 0.7307413668196088, "grad_norm": 0.2792026400566101, "learning_rate": 1.7499643806617e-05, "loss": 1.6757, "step": 3026 }, { "epoch": 0.7309828543829993, "grad_norm": 0.2907997965812683, "learning_rate": 1.7470237770513548e-05, "loss": 1.6434, "step": 3027 }, { "epoch": 0.7312243419463897, "grad_norm": 0.29097244143486023, "learning_rate": 1.7440851230904397e-05, "loss": 1.6614, "step": 3028 }, { "epoch": 0.7314658295097802, "grad_norm": 0.3114943206310272, "learning_rate": 1.7411484205402166e-05, "loss": 1.8196, "step": 3029 }, { "epoch": 0.7317073170731707, "grad_norm": 0.2941250503063202, "learning_rate": 1.738213671160782e-05, "loss": 1.707, "step": 3030 }, { "epoch": 0.7319488046365612, "grad_norm": 0.291372150182724, "learning_rate": 1.735280876711053e-05, "loss": 1.7419, "step": 3031 }, { "epoch": 0.7321902921999517, "grad_norm": 0.2872284948825836, "learning_rate": 1.732350038948784e-05, "loss": 1.7158, "step": 3032 }, { "epoch": 0.7324317797633422, "grad_norm": 0.2803144156932831, "learning_rate": 1.7294211596305514e-05, "loss": 1.6141, "step": 3033 }, { "epoch": 0.7326732673267327, "grad_norm": 0.5381343364715576, "learning_rate": 1.7264942405117586e-05, "loss": 1.6275, "step": 3034 }, { "epoch": 0.7329147548901231, "grad_norm": 0.2898636758327484, "learning_rate": 1.7235692833466373e-05, "loss": 1.6739, "step": 3035 }, { "epoch": 0.7331562424535136, "grad_norm": 0.28221726417541504, "learning_rate": 1.7206462898882397e-05, "loss": 1.7224, "step": 3036 }, { "epoch": 0.7333977300169041, "grad_norm": 0.2817859947681427, "learning_rate": 1.717725261888442e-05, "loss": 1.6742, "step": 3037 }, { "epoch": 0.7336392175802946, "grad_norm": 0.28856009244918823, "learning_rate": 1.714806201097941e-05, "loss": 1.7727, "step": 3038 }, { "epoch": 0.7338807051436851, "grad_norm": 0.2776874303817749, "learning_rate": 1.7118891092662614e-05, "loss": 1.5477, "step": 3039 }, { "epoch": 0.7341221927070756, "grad_norm": 0.31398606300354004, "learning_rate": 1.708973988141736e-05, "loss": 1.8539, "step": 3040 }, { "epoch": 0.7343636802704661, "grad_norm": 0.28254368901252747, "learning_rate": 1.7060608394715276e-05, "loss": 1.7903, "step": 3041 }, { "epoch": 0.7346051678338565, "grad_norm": 0.2785940170288086, "learning_rate": 1.703149665001611e-05, "loss": 1.504, "step": 3042 }, { "epoch": 0.734846655397247, "grad_norm": 0.2761463224887848, "learning_rate": 1.7002404664767774e-05, "loss": 1.5761, "step": 3043 }, { "epoch": 0.7350881429606375, "grad_norm": 0.29854467511177063, "learning_rate": 1.6973332456406383e-05, "loss": 1.7886, "step": 3044 }, { "epoch": 0.735329630524028, "grad_norm": 0.28894439339637756, "learning_rate": 1.694428004235616e-05, "loss": 1.7557, "step": 3045 }, { "epoch": 0.7355711180874185, "grad_norm": 0.28462332487106323, "learning_rate": 1.6915247440029464e-05, "loss": 1.7399, "step": 3046 }, { "epoch": 0.735812605650809, "grad_norm": 0.28917616605758667, "learning_rate": 1.6886234666826783e-05, "loss": 1.7092, "step": 3047 }, { "epoch": 0.7360540932141995, "grad_norm": 0.2863120138645172, "learning_rate": 1.685724174013677e-05, "loss": 1.6699, "step": 3048 }, { "epoch": 0.7362955807775899, "grad_norm": 0.2809310853481293, "learning_rate": 1.6828268677336085e-05, "loss": 1.6557, "step": 3049 }, { "epoch": 0.7365370683409804, "grad_norm": 0.29641038179397583, "learning_rate": 1.6799315495789575e-05, "loss": 1.8007, "step": 3050 }, { "epoch": 0.7367785559043709, "grad_norm": 0.2900030314922333, "learning_rate": 1.6770382212850122e-05, "loss": 1.7137, "step": 3051 }, { "epoch": 0.7370200434677614, "grad_norm": 0.29697611927986145, "learning_rate": 1.6741468845858703e-05, "loss": 1.7126, "step": 3052 }, { "epoch": 0.7372615310311519, "grad_norm": 0.29003795981407166, "learning_rate": 1.6712575412144326e-05, "loss": 1.7755, "step": 3053 }, { "epoch": 0.7375030185945424, "grad_norm": 0.27872002124786377, "learning_rate": 1.668370192902411e-05, "loss": 1.6429, "step": 3054 }, { "epoch": 0.7377445061579329, "grad_norm": 0.2773483097553253, "learning_rate": 1.6654848413803164e-05, "loss": 1.7595, "step": 3055 }, { "epoch": 0.7379859937213233, "grad_norm": 0.2997627854347229, "learning_rate": 1.662601488377464e-05, "loss": 1.7456, "step": 3056 }, { "epoch": 0.7382274812847138, "grad_norm": 0.2864185869693756, "learning_rate": 1.659720135621976e-05, "loss": 1.6767, "step": 3057 }, { "epoch": 0.7384689688481043, "grad_norm": 0.2858075797557831, "learning_rate": 1.656840784840766e-05, "loss": 1.7058, "step": 3058 }, { "epoch": 0.7387104564114948, "grad_norm": 0.28780397772789, "learning_rate": 1.653963437759558e-05, "loss": 1.6804, "step": 3059 }, { "epoch": 0.7389519439748853, "grad_norm": 0.3009507954120636, "learning_rate": 1.651088096102869e-05, "loss": 1.7636, "step": 3060 }, { "epoch": 0.7391934315382758, "grad_norm": 0.29653480648994446, "learning_rate": 1.6482147615940165e-05, "loss": 1.8417, "step": 3061 }, { "epoch": 0.7394349191016663, "grad_norm": 0.2834724485874176, "learning_rate": 1.6453434359551117e-05, "loss": 1.6444, "step": 3062 }, { "epoch": 0.7396764066650567, "grad_norm": 0.2846713066101074, "learning_rate": 1.6424741209070694e-05, "loss": 1.598, "step": 3063 }, { "epoch": 0.7399178942284472, "grad_norm": 0.2759416103363037, "learning_rate": 1.639606818169587e-05, "loss": 1.6323, "step": 3064 }, { "epoch": 0.7401593817918377, "grad_norm": 0.2958802282810211, "learning_rate": 1.63674152946117e-05, "loss": 1.6934, "step": 3065 }, { "epoch": 0.7404008693552282, "grad_norm": 0.28621771931648254, "learning_rate": 1.633878256499105e-05, "loss": 1.6516, "step": 3066 }, { "epoch": 0.7406423569186187, "grad_norm": 0.30480894446372986, "learning_rate": 1.6310170009994792e-05, "loss": 1.9149, "step": 3067 }, { "epoch": 0.7408838444820092, "grad_norm": 0.29486343264579773, "learning_rate": 1.6281577646771652e-05, "loss": 1.7708, "step": 3068 }, { "epoch": 0.7411253320453997, "grad_norm": 0.2811615765094757, "learning_rate": 1.6253005492458253e-05, "loss": 1.6379, "step": 3069 }, { "epoch": 0.7413668196087901, "grad_norm": 0.2905067801475525, "learning_rate": 1.6224453564179175e-05, "loss": 1.6988, "step": 3070 }, { "epoch": 0.7416083071721806, "grad_norm": 0.2873300313949585, "learning_rate": 1.6195921879046753e-05, "loss": 1.5925, "step": 3071 }, { "epoch": 0.7418497947355711, "grad_norm": 0.33222946524620056, "learning_rate": 1.616741045416132e-05, "loss": 1.8501, "step": 3072 }, { "epoch": 0.7420912822989616, "grad_norm": 0.2834736108779907, "learning_rate": 1.6138919306610983e-05, "loss": 1.7429, "step": 3073 }, { "epoch": 0.7423327698623521, "grad_norm": 0.2861657738685608, "learning_rate": 1.611044845347172e-05, "loss": 1.6765, "step": 3074 }, { "epoch": 0.7425742574257426, "grad_norm": 0.2855978012084961, "learning_rate": 1.6081997911807333e-05, "loss": 1.5943, "step": 3075 }, { "epoch": 0.742815744989133, "grad_norm": 0.29351118206977844, "learning_rate": 1.605356769866949e-05, "loss": 1.838, "step": 3076 }, { "epoch": 0.7430572325525235, "grad_norm": 0.2739328444004059, "learning_rate": 1.6025157831097638e-05, "loss": 1.6354, "step": 3077 }, { "epoch": 0.743298720115914, "grad_norm": 0.29537075757980347, "learning_rate": 1.599676832611904e-05, "loss": 1.7167, "step": 3078 }, { "epoch": 0.7435402076793045, "grad_norm": 0.30688053369522095, "learning_rate": 1.5968399200748745e-05, "loss": 1.9005, "step": 3079 }, { "epoch": 0.743781695242695, "grad_norm": 0.29034462571144104, "learning_rate": 1.5940050471989594e-05, "loss": 1.8667, "step": 3080 }, { "epoch": 0.7440231828060855, "grad_norm": 0.2946985960006714, "learning_rate": 1.5911722156832225e-05, "loss": 1.8504, "step": 3081 }, { "epoch": 0.744264670369476, "grad_norm": 0.27625375986099243, "learning_rate": 1.5883414272255027e-05, "loss": 1.5605, "step": 3082 }, { "epoch": 0.7445061579328665, "grad_norm": 0.3007868826389313, "learning_rate": 1.585512683522412e-05, "loss": 1.8901, "step": 3083 }, { "epoch": 0.7447476454962569, "grad_norm": 0.28775554895401, "learning_rate": 1.5826859862693387e-05, "loss": 1.6705, "step": 3084 }, { "epoch": 0.7449891330596474, "grad_norm": 0.30867692828178406, "learning_rate": 1.579861337160449e-05, "loss": 1.6515, "step": 3085 }, { "epoch": 0.7452306206230379, "grad_norm": 0.28869694471359253, "learning_rate": 1.5770387378886713e-05, "loss": 1.6076, "step": 3086 }, { "epoch": 0.7454721081864284, "grad_norm": 0.2880997359752655, "learning_rate": 1.574218190145716e-05, "loss": 1.656, "step": 3087 }, { "epoch": 0.7457135957498189, "grad_norm": 0.29413697123527527, "learning_rate": 1.5713996956220578e-05, "loss": 1.7752, "step": 3088 }, { "epoch": 0.7459550833132094, "grad_norm": 0.2936905026435852, "learning_rate": 1.568583256006941e-05, "loss": 1.52, "step": 3089 }, { "epoch": 0.7461965708765999, "grad_norm": 0.2842849791049957, "learning_rate": 1.565768872988383e-05, "loss": 1.6886, "step": 3090 }, { "epoch": 0.7464380584399903, "grad_norm": 0.2809235751628876, "learning_rate": 1.5629565482531632e-05, "loss": 1.5969, "step": 3091 }, { "epoch": 0.7466795460033808, "grad_norm": 0.3016565144062042, "learning_rate": 1.5601462834868306e-05, "loss": 1.618, "step": 3092 }, { "epoch": 0.7469210335667713, "grad_norm": 0.29342415928840637, "learning_rate": 1.5573380803736957e-05, "loss": 1.6963, "step": 3093 }, { "epoch": 0.7471625211301618, "grad_norm": 0.2991015911102295, "learning_rate": 1.554531940596842e-05, "loss": 1.7581, "step": 3094 }, { "epoch": 0.7474040086935523, "grad_norm": 0.29637351632118225, "learning_rate": 1.5517278658381035e-05, "loss": 1.6698, "step": 3095 }, { "epoch": 0.7476454962569428, "grad_norm": 0.2995702922344208, "learning_rate": 1.5489258577780885e-05, "loss": 1.6229, "step": 3096 }, { "epoch": 0.7478869838203333, "grad_norm": 0.286895751953125, "learning_rate": 1.54612591809616e-05, "loss": 1.701, "step": 3097 }, { "epoch": 0.7481284713837237, "grad_norm": 0.2805013656616211, "learning_rate": 1.543328048470443e-05, "loss": 1.6485, "step": 3098 }, { "epoch": 0.7483699589471142, "grad_norm": 0.27921825647354126, "learning_rate": 1.540532250577821e-05, "loss": 1.579, "step": 3099 }, { "epoch": 0.7486114465105047, "grad_norm": 0.2843914330005646, "learning_rate": 1.537738526093938e-05, "loss": 1.7283, "step": 3100 }, { "epoch": 0.7488529340738952, "grad_norm": 0.278006911277771, "learning_rate": 1.5349468766931947e-05, "loss": 1.662, "step": 3101 }, { "epoch": 0.7490944216372857, "grad_norm": 0.28066810965538025, "learning_rate": 1.532157304048744e-05, "loss": 1.6109, "step": 3102 }, { "epoch": 0.7493359092006762, "grad_norm": 0.28360772132873535, "learning_rate": 1.5293698098325038e-05, "loss": 1.7584, "step": 3103 }, { "epoch": 0.7495773967640667, "grad_norm": 0.2857002019882202, "learning_rate": 1.526584395715132e-05, "loss": 1.6592, "step": 3104 }, { "epoch": 0.7498188843274571, "grad_norm": 0.2857431173324585, "learning_rate": 1.5238010633660543e-05, "loss": 1.79, "step": 3105 }, { "epoch": 0.7500603718908476, "grad_norm": 0.29060348868370056, "learning_rate": 1.5210198144534394e-05, "loss": 1.7208, "step": 3106 }, { "epoch": 0.7503018594542381, "grad_norm": 0.29606837034225464, "learning_rate": 1.5182406506442104e-05, "loss": 1.7321, "step": 3107 }, { "epoch": 0.7505433470176286, "grad_norm": 0.2827746868133545, "learning_rate": 1.515463573604039e-05, "loss": 1.5039, "step": 3108 }, { "epoch": 0.7507848345810191, "grad_norm": 0.28397637605667114, "learning_rate": 1.512688584997351e-05, "loss": 1.7428, "step": 3109 }, { "epoch": 0.7510263221444096, "grad_norm": 0.2833336889743805, "learning_rate": 1.5099156864873154e-05, "loss": 1.7371, "step": 3110 }, { "epoch": 0.7512678097078, "grad_norm": 0.28867974877357483, "learning_rate": 1.5071448797358506e-05, "loss": 1.634, "step": 3111 }, { "epoch": 0.7515092972711905, "grad_norm": 0.27735328674316406, "learning_rate": 1.5043761664036205e-05, "loss": 1.5874, "step": 3112 }, { "epoch": 0.751750784834581, "grad_norm": 0.28543347120285034, "learning_rate": 1.501609548150033e-05, "loss": 1.7446, "step": 3113 }, { "epoch": 0.7519922723979715, "grad_norm": 0.30940118432044983, "learning_rate": 1.498845026633246e-05, "loss": 1.7605, "step": 3114 }, { "epoch": 0.752233759961362, "grad_norm": 0.2953583300113678, "learning_rate": 1.4960826035101556e-05, "loss": 1.7172, "step": 3115 }, { "epoch": 0.7524752475247525, "grad_norm": 0.2987193167209625, "learning_rate": 1.4933222804364e-05, "loss": 1.7557, "step": 3116 }, { "epoch": 0.752716735088143, "grad_norm": 0.2855208218097687, "learning_rate": 1.4905640590663606e-05, "loss": 1.8188, "step": 3117 }, { "epoch": 0.7529582226515334, "grad_norm": 0.30061909556388855, "learning_rate": 1.487807941053162e-05, "loss": 1.886, "step": 3118 }, { "epoch": 0.7531997102149239, "grad_norm": 0.27971214056015015, "learning_rate": 1.4850539280486596e-05, "loss": 1.6974, "step": 3119 }, { "epoch": 0.7534411977783144, "grad_norm": 0.2844647765159607, "learning_rate": 1.4823020217034566e-05, "loss": 1.7015, "step": 3120 }, { "epoch": 0.7536826853417049, "grad_norm": 0.27971306443214417, "learning_rate": 1.479552223666889e-05, "loss": 1.4961, "step": 3121 }, { "epoch": 0.7539241729050954, "grad_norm": 0.29338502883911133, "learning_rate": 1.4768045355870274e-05, "loss": 1.6573, "step": 3122 }, { "epoch": 0.7541656604684859, "grad_norm": 0.29643264412879944, "learning_rate": 1.474058959110684e-05, "loss": 1.6598, "step": 3123 }, { "epoch": 0.7544071480318764, "grad_norm": 0.2982703447341919, "learning_rate": 1.471315495883399e-05, "loss": 1.7551, "step": 3124 }, { "epoch": 0.7546486355952668, "grad_norm": 0.29914137721061707, "learning_rate": 1.4685741475494497e-05, "loss": 1.809, "step": 3125 }, { "epoch": 0.7548901231586573, "grad_norm": 0.29608285427093506, "learning_rate": 1.4658349157518425e-05, "loss": 1.733, "step": 3126 }, { "epoch": 0.7551316107220478, "grad_norm": 0.27723485231399536, "learning_rate": 1.4630978021323227e-05, "loss": 1.5983, "step": 3127 }, { "epoch": 0.7553730982854383, "grad_norm": 0.2844619154930115, "learning_rate": 1.4603628083313543e-05, "loss": 1.7424, "step": 3128 }, { "epoch": 0.7556145858488288, "grad_norm": 0.2863559126853943, "learning_rate": 1.4576299359881423e-05, "loss": 1.7432, "step": 3129 }, { "epoch": 0.7558560734122193, "grad_norm": 0.3061801791191101, "learning_rate": 1.4548991867406142e-05, "loss": 1.7387, "step": 3130 }, { "epoch": 0.7560975609756098, "grad_norm": 0.2797708809375763, "learning_rate": 1.4521705622254256e-05, "loss": 1.7408, "step": 3131 }, { "epoch": 0.7563390485390002, "grad_norm": 0.28697487711906433, "learning_rate": 1.4494440640779577e-05, "loss": 1.7039, "step": 3132 }, { "epoch": 0.7565805361023907, "grad_norm": 0.2965584695339203, "learning_rate": 1.4467196939323225e-05, "loss": 1.8164, "step": 3133 }, { "epoch": 0.7568220236657812, "grad_norm": 0.2774227559566498, "learning_rate": 1.4439974534213513e-05, "loss": 1.6039, "step": 3134 }, { "epoch": 0.7570635112291717, "grad_norm": 0.31710150837898254, "learning_rate": 1.4412773441765986e-05, "loss": 1.9236, "step": 3135 }, { "epoch": 0.7573049987925622, "grad_norm": 0.2914585769176483, "learning_rate": 1.4385593678283482e-05, "loss": 1.586, "step": 3136 }, { "epoch": 0.7575464863559527, "grad_norm": 0.2858920097351074, "learning_rate": 1.4358435260055958e-05, "loss": 1.727, "step": 3137 }, { "epoch": 0.7577879739193432, "grad_norm": 0.31640803813934326, "learning_rate": 1.4331298203360666e-05, "loss": 2.0242, "step": 3138 }, { "epoch": 0.7580294614827336, "grad_norm": 0.28446534276008606, "learning_rate": 1.4304182524462001e-05, "loss": 1.5983, "step": 3139 }, { "epoch": 0.7582709490461241, "grad_norm": 0.29239290952682495, "learning_rate": 1.4277088239611575e-05, "loss": 1.7614, "step": 3140 }, { "epoch": 0.7585124366095146, "grad_norm": 0.2873914837837219, "learning_rate": 1.4250015365048136e-05, "loss": 1.7005, "step": 3141 }, { "epoch": 0.7587539241729051, "grad_norm": 0.2990749180316925, "learning_rate": 1.4222963916997667e-05, "loss": 1.7283, "step": 3142 }, { "epoch": 0.7589954117362956, "grad_norm": 0.2803306579589844, "learning_rate": 1.419593391167326e-05, "loss": 1.6648, "step": 3143 }, { "epoch": 0.7592368992996861, "grad_norm": 0.27686238288879395, "learning_rate": 1.416892536527516e-05, "loss": 1.5488, "step": 3144 }, { "epoch": 0.7594783868630766, "grad_norm": 0.28388309478759766, "learning_rate": 1.4141938293990763e-05, "loss": 1.5762, "step": 3145 }, { "epoch": 0.759719874426467, "grad_norm": 0.287249356508255, "learning_rate": 1.4114972713994567e-05, "loss": 1.6871, "step": 3146 }, { "epoch": 0.7599613619898575, "grad_norm": 0.28766515851020813, "learning_rate": 1.4088028641448247e-05, "loss": 1.6785, "step": 3147 }, { "epoch": 0.760202849553248, "grad_norm": 0.28722694516181946, "learning_rate": 1.4061106092500536e-05, "loss": 1.6654, "step": 3148 }, { "epoch": 0.7604443371166385, "grad_norm": 0.2890714406967163, "learning_rate": 1.4034205083287283e-05, "loss": 1.6766, "step": 3149 }, { "epoch": 0.760685824680029, "grad_norm": 0.2804972231388092, "learning_rate": 1.4007325629931412e-05, "loss": 1.7154, "step": 3150 }, { "epoch": 0.7609273122434195, "grad_norm": 0.2925170361995697, "learning_rate": 1.3980467748542986e-05, "loss": 1.626, "step": 3151 }, { "epoch": 0.76116879980681, "grad_norm": 0.29058411717414856, "learning_rate": 1.3953631455219034e-05, "loss": 1.7097, "step": 3152 }, { "epoch": 0.7614102873702004, "grad_norm": 0.2916194498538971, "learning_rate": 1.3926816766043754e-05, "loss": 1.7151, "step": 3153 }, { "epoch": 0.7616517749335909, "grad_norm": 0.2868903875350952, "learning_rate": 1.3900023697088338e-05, "loss": 1.7487, "step": 3154 }, { "epoch": 0.7618932624969814, "grad_norm": 0.2841212749481201, "learning_rate": 1.3873252264411013e-05, "loss": 1.5429, "step": 3155 }, { "epoch": 0.7621347500603719, "grad_norm": 0.28481557965278625, "learning_rate": 1.3846502484057094e-05, "loss": 1.6972, "step": 3156 }, { "epoch": 0.7623762376237624, "grad_norm": 0.3077322244644165, "learning_rate": 1.3819774372058858e-05, "loss": 1.763, "step": 3157 }, { "epoch": 0.7626177251871529, "grad_norm": 0.29711633920669556, "learning_rate": 1.3793067944435622e-05, "loss": 1.735, "step": 3158 }, { "epoch": 0.7628592127505434, "grad_norm": 0.28862911462783813, "learning_rate": 1.376638321719369e-05, "loss": 1.7224, "step": 3159 }, { "epoch": 0.7631007003139338, "grad_norm": 0.2817215621471405, "learning_rate": 1.3739720206326417e-05, "loss": 1.6972, "step": 3160 }, { "epoch": 0.7633421878773243, "grad_norm": 0.31388506293296814, "learning_rate": 1.3713078927814043e-05, "loss": 1.7658, "step": 3161 }, { "epoch": 0.7635836754407148, "grad_norm": 0.2770969867706299, "learning_rate": 1.3686459397623886e-05, "loss": 1.6603, "step": 3162 }, { "epoch": 0.7638251630041053, "grad_norm": 0.28280574083328247, "learning_rate": 1.3659861631710163e-05, "loss": 1.7169, "step": 3163 }, { "epoch": 0.7640666505674958, "grad_norm": 0.2775695323944092, "learning_rate": 1.3633285646014066e-05, "loss": 1.6607, "step": 3164 }, { "epoch": 0.7643081381308863, "grad_norm": 0.3132243752479553, "learning_rate": 1.3606731456463717e-05, "loss": 1.8768, "step": 3165 }, { "epoch": 0.7645496256942768, "grad_norm": 0.2730356454849243, "learning_rate": 1.3580199078974221e-05, "loss": 1.5169, "step": 3166 }, { "epoch": 0.7647911132576672, "grad_norm": 0.29339200258255005, "learning_rate": 1.3553688529447572e-05, "loss": 1.6291, "step": 3167 }, { "epoch": 0.7650326008210577, "grad_norm": 0.29365086555480957, "learning_rate": 1.3527199823772667e-05, "loss": 1.7219, "step": 3168 }, { "epoch": 0.7652740883844482, "grad_norm": 0.30693885684013367, "learning_rate": 1.3500732977825387e-05, "loss": 1.6672, "step": 3169 }, { "epoch": 0.7655155759478387, "grad_norm": 0.29114919900894165, "learning_rate": 1.3474288007468389e-05, "loss": 1.663, "step": 3170 }, { "epoch": 0.7657570635112292, "grad_norm": 0.35604071617126465, "learning_rate": 1.344786492855134e-05, "loss": 2.0797, "step": 3171 }, { "epoch": 0.7659985510746197, "grad_norm": 0.3010037839412689, "learning_rate": 1.3421463756910712e-05, "loss": 1.8108, "step": 3172 }, { "epoch": 0.7662400386380102, "grad_norm": 0.296786367893219, "learning_rate": 1.3395084508369876e-05, "loss": 1.8169, "step": 3173 }, { "epoch": 0.7664815262014006, "grad_norm": 0.27886155247688293, "learning_rate": 1.3368727198739046e-05, "loss": 1.6247, "step": 3174 }, { "epoch": 0.7667230137647911, "grad_norm": 0.28021758794784546, "learning_rate": 1.3342391843815316e-05, "loss": 1.5002, "step": 3175 }, { "epoch": 0.7669645013281816, "grad_norm": 0.2799939215183258, "learning_rate": 1.3316078459382597e-05, "loss": 1.6948, "step": 3176 }, { "epoch": 0.7672059888915721, "grad_norm": 0.2892768085002899, "learning_rate": 1.328978706121164e-05, "loss": 1.7276, "step": 3177 }, { "epoch": 0.7674474764549626, "grad_norm": 0.2675463557243347, "learning_rate": 1.3263517665060015e-05, "loss": 1.5687, "step": 3178 }, { "epoch": 0.7676889640183531, "grad_norm": 0.296813040971756, "learning_rate": 1.3237270286672098e-05, "loss": 1.6419, "step": 3179 }, { "epoch": 0.7679304515817436, "grad_norm": 0.3036782741546631, "learning_rate": 1.32110449417791e-05, "loss": 1.7433, "step": 3180 }, { "epoch": 0.768171939145134, "grad_norm": 0.29037466645240784, "learning_rate": 1.3184841646099e-05, "loss": 1.7653, "step": 3181 }, { "epoch": 0.7684134267085245, "grad_norm": 0.2794424593448639, "learning_rate": 1.3158660415336555e-05, "loss": 1.5486, "step": 3182 }, { "epoch": 0.768654914271915, "grad_norm": 0.2844526469707489, "learning_rate": 1.3132501265183306e-05, "loss": 1.7376, "step": 3183 }, { "epoch": 0.7688964018353055, "grad_norm": 0.30006957054138184, "learning_rate": 1.31063642113176e-05, "loss": 1.6812, "step": 3184 }, { "epoch": 0.769137889398696, "grad_norm": 0.28671425580978394, "learning_rate": 1.3080249269404454e-05, "loss": 1.6261, "step": 3185 }, { "epoch": 0.7693793769620865, "grad_norm": 0.28891482949256897, "learning_rate": 1.3054156455095723e-05, "loss": 1.5923, "step": 3186 }, { "epoch": 0.769620864525477, "grad_norm": 0.3181431293487549, "learning_rate": 1.3028085784029943e-05, "loss": 1.8079, "step": 3187 }, { "epoch": 0.7698623520888674, "grad_norm": 0.27865979075431824, "learning_rate": 1.300203727183238e-05, "loss": 1.6451, "step": 3188 }, { "epoch": 0.7701038396522579, "grad_norm": 0.27682119607925415, "learning_rate": 1.2976010934115068e-05, "loss": 1.5495, "step": 3189 }, { "epoch": 0.7703453272156484, "grad_norm": 0.2913415729999542, "learning_rate": 1.2950006786476688e-05, "loss": 1.7481, "step": 3190 }, { "epoch": 0.7705868147790389, "grad_norm": 0.2849331498146057, "learning_rate": 1.2924024844502691e-05, "loss": 1.6748, "step": 3191 }, { "epoch": 0.7708283023424294, "grad_norm": 0.2848500907421112, "learning_rate": 1.2898065123765124e-05, "loss": 1.74, "step": 3192 }, { "epoch": 0.7710697899058199, "grad_norm": 0.29086798429489136, "learning_rate": 1.2872127639822817e-05, "loss": 1.6632, "step": 3193 }, { "epoch": 0.7713112774692104, "grad_norm": 0.3069058954715729, "learning_rate": 1.2846212408221204e-05, "loss": 1.866, "step": 3194 }, { "epoch": 0.7715527650326008, "grad_norm": 0.2816770672798157, "learning_rate": 1.2820319444492413e-05, "loss": 1.6053, "step": 3195 }, { "epoch": 0.7717942525959913, "grad_norm": 0.26773592829704285, "learning_rate": 1.2794448764155204e-05, "loss": 1.5483, "step": 3196 }, { "epoch": 0.7720357401593818, "grad_norm": 0.2855380177497864, "learning_rate": 1.2768600382715029e-05, "loss": 1.6927, "step": 3197 }, { "epoch": 0.7722772277227723, "grad_norm": 0.2751227915287018, "learning_rate": 1.2742774315663892e-05, "loss": 1.5896, "step": 3198 }, { "epoch": 0.7725187152861628, "grad_norm": 0.2863582372665405, "learning_rate": 1.2716970578480514e-05, "loss": 1.7249, "step": 3199 }, { "epoch": 0.7727602028495533, "grad_norm": 0.2789832353591919, "learning_rate": 1.269118918663018e-05, "loss": 1.6993, "step": 3200 }, { "epoch": 0.7730016904129438, "grad_norm": 0.29148024320602417, "learning_rate": 1.2665430155564773e-05, "loss": 1.7887, "step": 3201 }, { "epoch": 0.7732431779763342, "grad_norm": 0.28856217861175537, "learning_rate": 1.2639693500722827e-05, "loss": 1.7012, "step": 3202 }, { "epoch": 0.7734846655397247, "grad_norm": 0.2841264605522156, "learning_rate": 1.2613979237529422e-05, "loss": 1.6671, "step": 3203 }, { "epoch": 0.7737261531031152, "grad_norm": 0.27530044317245483, "learning_rate": 1.2588287381396224e-05, "loss": 1.5728, "step": 3204 }, { "epoch": 0.7739676406665057, "grad_norm": 0.27895525097846985, "learning_rate": 1.2562617947721462e-05, "loss": 1.6097, "step": 3205 }, { "epoch": 0.7742091282298962, "grad_norm": 0.29398396611213684, "learning_rate": 1.2536970951889982e-05, "loss": 1.7301, "step": 3206 }, { "epoch": 0.7744506157932867, "grad_norm": 0.2903694808483124, "learning_rate": 1.2511346409273078e-05, "loss": 1.6795, "step": 3207 }, { "epoch": 0.7746921033566772, "grad_norm": 0.29402676224708557, "learning_rate": 1.2485744335228695e-05, "loss": 1.7845, "step": 3208 }, { "epoch": 0.7749335909200676, "grad_norm": 0.2845938205718994, "learning_rate": 1.2460164745101244e-05, "loss": 1.675, "step": 3209 }, { "epoch": 0.7751750784834581, "grad_norm": 0.30662643909454346, "learning_rate": 1.2434607654221686e-05, "loss": 1.8337, "step": 3210 }, { "epoch": 0.7754165660468486, "grad_norm": 0.30462169647216797, "learning_rate": 1.2409073077907474e-05, "loss": 1.89, "step": 3211 }, { "epoch": 0.7756580536102391, "grad_norm": 0.2823159396648407, "learning_rate": 1.2383561031462615e-05, "loss": 1.6591, "step": 3212 }, { "epoch": 0.7758995411736296, "grad_norm": 0.3022060692310333, "learning_rate": 1.2358071530177573e-05, "loss": 1.6732, "step": 3213 }, { "epoch": 0.7761410287370201, "grad_norm": 0.29048630595207214, "learning_rate": 1.2332604589329282e-05, "loss": 1.7083, "step": 3214 }, { "epoch": 0.7763825163004106, "grad_norm": 0.2948112487792969, "learning_rate": 1.2307160224181236e-05, "loss": 1.8011, "step": 3215 }, { "epoch": 0.776624003863801, "grad_norm": 0.28570568561553955, "learning_rate": 1.2281738449983282e-05, "loss": 1.6478, "step": 3216 }, { "epoch": 0.7768654914271915, "grad_norm": 0.2771815359592438, "learning_rate": 1.225633928197183e-05, "loss": 1.7081, "step": 3217 }, { "epoch": 0.777106978990582, "grad_norm": 0.29521358013153076, "learning_rate": 1.2230962735369695e-05, "loss": 1.8419, "step": 3218 }, { "epoch": 0.7773484665539725, "grad_norm": 0.29237470030784607, "learning_rate": 1.2205608825386133e-05, "loss": 1.7326, "step": 3219 }, { "epoch": 0.777589954117363, "grad_norm": 0.298209547996521, "learning_rate": 1.2180277567216824e-05, "loss": 1.5866, "step": 3220 }, { "epoch": 0.7778314416807535, "grad_norm": 0.26703622937202454, "learning_rate": 1.2154968976043923e-05, "loss": 1.5421, "step": 3221 }, { "epoch": 0.778072929244144, "grad_norm": 0.2806689143180847, "learning_rate": 1.2129683067035941e-05, "loss": 1.6482, "step": 3222 }, { "epoch": 0.7783144168075344, "grad_norm": 0.29450640082359314, "learning_rate": 1.210441985534781e-05, "loss": 1.7079, "step": 3223 }, { "epoch": 0.7785559043709249, "grad_norm": 0.29023537039756775, "learning_rate": 1.2079179356120907e-05, "loss": 1.7055, "step": 3224 }, { "epoch": 0.7787973919343154, "grad_norm": 0.28531989455223083, "learning_rate": 1.2053961584482893e-05, "loss": 1.6496, "step": 3225 }, { "epoch": 0.7790388794977059, "grad_norm": 0.29664355516433716, "learning_rate": 1.2028766555547922e-05, "loss": 1.818, "step": 3226 }, { "epoch": 0.7792803670610964, "grad_norm": 0.2795257866382599, "learning_rate": 1.2003594284416447e-05, "loss": 1.667, "step": 3227 }, { "epoch": 0.7795218546244869, "grad_norm": 0.27589744329452515, "learning_rate": 1.1978444786175298e-05, "loss": 1.6667, "step": 3228 }, { "epoch": 0.7797633421878774, "grad_norm": 0.29114487767219543, "learning_rate": 1.195331807589764e-05, "loss": 1.735, "step": 3229 }, { "epoch": 0.7800048297512678, "grad_norm": 0.2899205982685089, "learning_rate": 1.1928214168643049e-05, "loss": 1.7336, "step": 3230 }, { "epoch": 0.7802463173146583, "grad_norm": 0.2912905514240265, "learning_rate": 1.1903133079457323e-05, "loss": 1.6066, "step": 3231 }, { "epoch": 0.7804878048780488, "grad_norm": 0.2899705469608307, "learning_rate": 1.1878074823372687e-05, "loss": 1.6714, "step": 3232 }, { "epoch": 0.7807292924414393, "grad_norm": 0.2928745448589325, "learning_rate": 1.1853039415407619e-05, "loss": 1.7063, "step": 3233 }, { "epoch": 0.7809707800048298, "grad_norm": 0.29191234707832336, "learning_rate": 1.1828026870566916e-05, "loss": 1.6632, "step": 3234 }, { "epoch": 0.7812122675682203, "grad_norm": 0.29038006067276, "learning_rate": 1.1803037203841699e-05, "loss": 1.576, "step": 3235 }, { "epoch": 0.7814537551316107, "grad_norm": 0.2763194143772125, "learning_rate": 1.1778070430209353e-05, "loss": 1.733, "step": 3236 }, { "epoch": 0.7816952426950012, "grad_norm": 0.27849137783050537, "learning_rate": 1.1753126564633537e-05, "loss": 1.5304, "step": 3237 }, { "epoch": 0.7819367302583917, "grad_norm": 0.29204875230789185, "learning_rate": 1.1728205622064176e-05, "loss": 1.6763, "step": 3238 }, { "epoch": 0.7821782178217822, "grad_norm": 0.2888553738594055, "learning_rate": 1.1703307617437514e-05, "loss": 1.6508, "step": 3239 }, { "epoch": 0.7824197053851727, "grad_norm": 0.2893203794956207, "learning_rate": 1.1678432565675945e-05, "loss": 1.6373, "step": 3240 }, { "epoch": 0.7826611929485632, "grad_norm": 0.29634320735931396, "learning_rate": 1.165358048168821e-05, "loss": 1.6828, "step": 3241 }, { "epoch": 0.7829026805119537, "grad_norm": 0.2999733090400696, "learning_rate": 1.1628751380369224e-05, "loss": 1.7992, "step": 3242 }, { "epoch": 0.7831441680753441, "grad_norm": 0.2619905471801758, "learning_rate": 1.1603945276600136e-05, "loss": 1.5119, "step": 3243 }, { "epoch": 0.7833856556387346, "grad_norm": 0.2907295823097229, "learning_rate": 1.1579162185248305e-05, "loss": 1.6952, "step": 3244 }, { "epoch": 0.7836271432021251, "grad_norm": 0.3023069202899933, "learning_rate": 1.1554402121167341e-05, "loss": 1.9187, "step": 3245 }, { "epoch": 0.7838686307655156, "grad_norm": 0.2990405559539795, "learning_rate": 1.1529665099197013e-05, "loss": 1.786, "step": 3246 }, { "epoch": 0.7841101183289061, "grad_norm": 0.29923105239868164, "learning_rate": 1.1504951134163261e-05, "loss": 1.8546, "step": 3247 }, { "epoch": 0.7843516058922966, "grad_norm": 0.2835824489593506, "learning_rate": 1.1480260240878287e-05, "loss": 1.6679, "step": 3248 }, { "epoch": 0.7845930934556871, "grad_norm": 0.2835007607936859, "learning_rate": 1.1455592434140355e-05, "loss": 1.6104, "step": 3249 }, { "epoch": 0.7848345810190775, "grad_norm": 0.30193471908569336, "learning_rate": 1.1430947728733988e-05, "loss": 1.8154, "step": 3250 }, { "epoch": 0.785076068582468, "grad_norm": 0.2950722277164459, "learning_rate": 1.1406326139429813e-05, "loss": 1.7746, "step": 3251 }, { "epoch": 0.7853175561458585, "grad_norm": 0.28022435307502747, "learning_rate": 1.1381727680984605e-05, "loss": 1.5873, "step": 3252 }, { "epoch": 0.785559043709249, "grad_norm": 0.2895815372467041, "learning_rate": 1.1357152368141289e-05, "loss": 1.7739, "step": 3253 }, { "epoch": 0.7858005312726395, "grad_norm": 0.28453055024147034, "learning_rate": 1.1332600215628919e-05, "loss": 1.7852, "step": 3254 }, { "epoch": 0.78604201883603, "grad_norm": 0.2981914281845093, "learning_rate": 1.1308071238162671e-05, "loss": 1.6995, "step": 3255 }, { "epoch": 0.7862835063994205, "grad_norm": 0.2992986738681793, "learning_rate": 1.1283565450443789e-05, "loss": 1.757, "step": 3256 }, { "epoch": 0.786524993962811, "grad_norm": 0.2941392660140991, "learning_rate": 1.125908286715971e-05, "loss": 1.5906, "step": 3257 }, { "epoch": 0.7867664815262014, "grad_norm": 0.2926832437515259, "learning_rate": 1.1234623502983838e-05, "loss": 1.7143, "step": 3258 }, { "epoch": 0.7870079690895919, "grad_norm": 0.27814507484436035, "learning_rate": 1.1210187372575776e-05, "loss": 1.5896, "step": 3259 }, { "epoch": 0.7872494566529824, "grad_norm": 0.298307865858078, "learning_rate": 1.1185774490581142e-05, "loss": 1.8445, "step": 3260 }, { "epoch": 0.7874909442163729, "grad_norm": 0.2890680134296417, "learning_rate": 1.116138487163163e-05, "loss": 1.8276, "step": 3261 }, { "epoch": 0.7877324317797634, "grad_norm": 0.2923431694507599, "learning_rate": 1.1137018530344978e-05, "loss": 1.6615, "step": 3262 }, { "epoch": 0.7879739193431539, "grad_norm": 0.3103092312812805, "learning_rate": 1.1112675481325035e-05, "loss": 1.8, "step": 3263 }, { "epoch": 0.7882154069065443, "grad_norm": 0.29589107632637024, "learning_rate": 1.1088355739161588e-05, "loss": 1.7776, "step": 3264 }, { "epoch": 0.7884568944699348, "grad_norm": 0.2966512441635132, "learning_rate": 1.1064059318430548e-05, "loss": 1.8251, "step": 3265 }, { "epoch": 0.7886983820333253, "grad_norm": 0.302327036857605, "learning_rate": 1.1039786233693806e-05, "loss": 1.6147, "step": 3266 }, { "epoch": 0.7889398695967158, "grad_norm": 0.27197715640068054, "learning_rate": 1.1015536499499246e-05, "loss": 1.5721, "step": 3267 }, { "epoch": 0.7891813571601063, "grad_norm": 0.2948354482650757, "learning_rate": 1.0991310130380822e-05, "loss": 1.7469, "step": 3268 }, { "epoch": 0.7894228447234968, "grad_norm": 0.2939530313014984, "learning_rate": 1.0967107140858424e-05, "loss": 1.6843, "step": 3269 }, { "epoch": 0.7896643322868873, "grad_norm": 0.290088415145874, "learning_rate": 1.094292754543796e-05, "loss": 1.7287, "step": 3270 }, { "epoch": 0.7899058198502777, "grad_norm": 0.2943084239959717, "learning_rate": 1.0918771358611285e-05, "loss": 1.7648, "step": 3271 }, { "epoch": 0.7901473074136682, "grad_norm": 0.2849422097206116, "learning_rate": 1.0894638594856288e-05, "loss": 1.6574, "step": 3272 }, { "epoch": 0.7903887949770587, "grad_norm": 0.2878836393356323, "learning_rate": 1.0870529268636736e-05, "loss": 1.6888, "step": 3273 }, { "epoch": 0.7906302825404492, "grad_norm": 0.2860524654388428, "learning_rate": 1.0846443394402417e-05, "loss": 1.5591, "step": 3274 }, { "epoch": 0.7908717701038397, "grad_norm": 0.3130074739456177, "learning_rate": 1.0822380986589037e-05, "loss": 1.8431, "step": 3275 }, { "epoch": 0.7911132576672302, "grad_norm": 0.28435906767845154, "learning_rate": 1.0798342059618215e-05, "loss": 1.7706, "step": 3276 }, { "epoch": 0.7913547452306207, "grad_norm": 0.2775307297706604, "learning_rate": 1.0774326627897557e-05, "loss": 1.6174, "step": 3277 }, { "epoch": 0.7915962327940111, "grad_norm": 0.2893638610839844, "learning_rate": 1.0750334705820531e-05, "loss": 1.7105, "step": 3278 }, { "epoch": 0.7918377203574016, "grad_norm": 0.30694591999053955, "learning_rate": 1.0726366307766533e-05, "loss": 1.7873, "step": 3279 }, { "epoch": 0.7920792079207921, "grad_norm": 0.2769351005554199, "learning_rate": 1.0702421448100853e-05, "loss": 1.5906, "step": 3280 }, { "epoch": 0.7923206954841826, "grad_norm": 0.3117207884788513, "learning_rate": 1.0678500141174713e-05, "loss": 1.8216, "step": 3281 }, { "epoch": 0.7925621830475731, "grad_norm": 0.2866639494895935, "learning_rate": 1.0654602401325147e-05, "loss": 1.7335, "step": 3282 }, { "epoch": 0.7928036706109636, "grad_norm": 0.28938835859298706, "learning_rate": 1.0630728242875132e-05, "loss": 1.6822, "step": 3283 }, { "epoch": 0.7930451581743541, "grad_norm": 0.2999311685562134, "learning_rate": 1.0606877680133481e-05, "loss": 1.8154, "step": 3284 }, { "epoch": 0.7932866457377445, "grad_norm": 0.28304949402809143, "learning_rate": 1.0583050727394861e-05, "loss": 1.7293, "step": 3285 }, { "epoch": 0.793528133301135, "grad_norm": 0.2820422649383545, "learning_rate": 1.0559247398939786e-05, "loss": 1.6999, "step": 3286 }, { "epoch": 0.7937696208645255, "grad_norm": 0.2734607458114624, "learning_rate": 1.0535467709034647e-05, "loss": 1.5438, "step": 3287 }, { "epoch": 0.794011108427916, "grad_norm": 0.2850915491580963, "learning_rate": 1.051171167193163e-05, "loss": 1.6129, "step": 3288 }, { "epoch": 0.7942525959913065, "grad_norm": 0.300689697265625, "learning_rate": 1.0487979301868739e-05, "loss": 1.889, "step": 3289 }, { "epoch": 0.794494083554697, "grad_norm": 0.282624751329422, "learning_rate": 1.0464270613069849e-05, "loss": 1.5902, "step": 3290 }, { "epoch": 0.7947355711180875, "grad_norm": 0.3154928386211395, "learning_rate": 1.0440585619744558e-05, "loss": 1.9566, "step": 3291 }, { "epoch": 0.7949770586814779, "grad_norm": 0.3213886320590973, "learning_rate": 1.041692433608834e-05, "loss": 1.9176, "step": 3292 }, { "epoch": 0.7952185462448684, "grad_norm": 0.2840975821018219, "learning_rate": 1.0393286776282419e-05, "loss": 1.6191, "step": 3293 }, { "epoch": 0.7954600338082589, "grad_norm": 0.275286465883255, "learning_rate": 1.0369672954493803e-05, "loss": 1.5905, "step": 3294 }, { "epoch": 0.7957015213716494, "grad_norm": 0.2848682701587677, "learning_rate": 1.0346082884875269e-05, "loss": 1.675, "step": 3295 }, { "epoch": 0.7959430089350399, "grad_norm": 0.27860069274902344, "learning_rate": 1.03225165815654e-05, "loss": 1.6886, "step": 3296 }, { "epoch": 0.7961844964984304, "grad_norm": 0.28801366686820984, "learning_rate": 1.029897405868845e-05, "loss": 1.6584, "step": 3297 }, { "epoch": 0.7964259840618209, "grad_norm": 0.2972698509693146, "learning_rate": 1.0275455330354512e-05, "loss": 1.8388, "step": 3298 }, { "epoch": 0.7966674716252113, "grad_norm": 0.28386732935905457, "learning_rate": 1.0251960410659367e-05, "loss": 1.7367, "step": 3299 }, { "epoch": 0.7969089591886018, "grad_norm": 0.2872152626514435, "learning_rate": 1.0228489313684519e-05, "loss": 1.6663, "step": 3300 }, { "epoch": 0.7971504467519923, "grad_norm": 0.2880032956600189, "learning_rate": 1.0205042053497249e-05, "loss": 1.6189, "step": 3301 }, { "epoch": 0.7973919343153828, "grad_norm": 0.28714028000831604, "learning_rate": 1.0181618644150488e-05, "loss": 1.65, "step": 3302 }, { "epoch": 0.7976334218787733, "grad_norm": 0.2771432101726532, "learning_rate": 1.0158219099682909e-05, "loss": 1.5128, "step": 3303 }, { "epoch": 0.7978749094421638, "grad_norm": 0.2889125645160675, "learning_rate": 1.0134843434118852e-05, "loss": 1.6493, "step": 3304 }, { "epoch": 0.7981163970055543, "grad_norm": 0.28715401887893677, "learning_rate": 1.0111491661468408e-05, "loss": 1.7849, "step": 3305 }, { "epoch": 0.7983578845689447, "grad_norm": 0.28724607825279236, "learning_rate": 1.0088163795727252e-05, "loss": 1.6569, "step": 3306 }, { "epoch": 0.7985993721323352, "grad_norm": 0.29266589879989624, "learning_rate": 1.0064859850876824e-05, "loss": 1.6414, "step": 3307 }, { "epoch": 0.7988408596957257, "grad_norm": 0.305501252412796, "learning_rate": 1.0041579840884169e-05, "loss": 1.7197, "step": 3308 }, { "epoch": 0.7990823472591162, "grad_norm": 0.2808031141757965, "learning_rate": 1.0018323779701998e-05, "loss": 1.6267, "step": 3309 }, { "epoch": 0.7993238348225067, "grad_norm": 0.28568148612976074, "learning_rate": 9.995091681268698e-06, "loss": 1.7157, "step": 3310 }, { "epoch": 0.7995653223858972, "grad_norm": 0.2838073670864105, "learning_rate": 9.971883559508256e-06, "loss": 1.6894, "step": 3311 }, { "epoch": 0.7998068099492877, "grad_norm": 0.2802664041519165, "learning_rate": 9.948699428330305e-06, "loss": 1.6598, "step": 3312 }, { "epoch": 0.8000482975126781, "grad_norm": 0.2997002601623535, "learning_rate": 9.925539301630088e-06, "loss": 1.8612, "step": 3313 }, { "epoch": 0.8002897850760686, "grad_norm": 0.3086371123790741, "learning_rate": 9.9024031932885e-06, "loss": 1.8849, "step": 3314 }, { "epoch": 0.8005312726394591, "grad_norm": 0.31059250235557556, "learning_rate": 9.87929111717198e-06, "loss": 1.7897, "step": 3315 }, { "epoch": 0.8007727602028496, "grad_norm": 0.30091264843940735, "learning_rate": 9.85620308713261e-06, "loss": 1.7329, "step": 3316 }, { "epoch": 0.8010142477662401, "grad_norm": 0.27469906210899353, "learning_rate": 9.83313911700804e-06, "loss": 1.6194, "step": 3317 }, { "epoch": 0.8012557353296306, "grad_norm": 0.31146568059921265, "learning_rate": 9.810099220621532e-06, "loss": 2.0038, "step": 3318 }, { "epoch": 0.801497222893021, "grad_norm": 0.2786215841770172, "learning_rate": 9.78708341178185e-06, "loss": 1.5996, "step": 3319 }, { "epoch": 0.8017387104564115, "grad_norm": 0.4017598032951355, "learning_rate": 9.76409170428339e-06, "loss": 1.6637, "step": 3320 }, { "epoch": 0.801980198019802, "grad_norm": 0.3601098656654358, "learning_rate": 9.741124111906081e-06, "loss": 1.8182, "step": 3321 }, { "epoch": 0.8022216855831925, "grad_norm": 0.2967846393585205, "learning_rate": 9.718180648415371e-06, "loss": 1.7739, "step": 3322 }, { "epoch": 0.802463173146583, "grad_norm": 0.2796829640865326, "learning_rate": 9.6952613275623e-06, "loss": 1.6086, "step": 3323 }, { "epoch": 0.8027046607099735, "grad_norm": 0.2965974807739258, "learning_rate": 9.672366163083396e-06, "loss": 1.7279, "step": 3324 }, { "epoch": 0.802946148273364, "grad_norm": 0.2807267904281616, "learning_rate": 9.649495168700717e-06, "loss": 1.6443, "step": 3325 }, { "epoch": 0.8031876358367545, "grad_norm": 0.28872230648994446, "learning_rate": 9.626648358121837e-06, "loss": 1.716, "step": 3326 }, { "epoch": 0.8034291234001449, "grad_norm": 0.29388031363487244, "learning_rate": 9.603825745039868e-06, "loss": 1.7489, "step": 3327 }, { "epoch": 0.8036706109635354, "grad_norm": 0.2864110469818115, "learning_rate": 9.581027343133342e-06, "loss": 1.7397, "step": 3328 }, { "epoch": 0.8039120985269259, "grad_norm": 0.2774543762207031, "learning_rate": 9.558253166066367e-06, "loss": 1.7122, "step": 3329 }, { "epoch": 0.8041535860903164, "grad_norm": 0.2785169184207916, "learning_rate": 9.535503227488474e-06, "loss": 1.5871, "step": 3330 }, { "epoch": 0.8043950736537069, "grad_norm": 0.283530056476593, "learning_rate": 9.512777541034696e-06, "loss": 1.71, "step": 3331 }, { "epoch": 0.8046365612170974, "grad_norm": 0.2865777611732483, "learning_rate": 9.490076120325503e-06, "loss": 1.745, "step": 3332 }, { "epoch": 0.8048780487804879, "grad_norm": 0.29725387692451477, "learning_rate": 9.467398978966856e-06, "loss": 1.7237, "step": 3333 }, { "epoch": 0.8051195363438783, "grad_norm": 0.2947281301021576, "learning_rate": 9.444746130550153e-06, "loss": 1.6848, "step": 3334 }, { "epoch": 0.8053610239072688, "grad_norm": 0.2953648567199707, "learning_rate": 9.422117588652197e-06, "loss": 1.6566, "step": 3335 }, { "epoch": 0.8056025114706593, "grad_norm": 0.29047906398773193, "learning_rate": 9.399513366835295e-06, "loss": 1.7636, "step": 3336 }, { "epoch": 0.8058439990340498, "grad_norm": 0.30120185017585754, "learning_rate": 9.376933478647087e-06, "loss": 1.7569, "step": 3337 }, { "epoch": 0.8060854865974403, "grad_norm": 0.29411962628364563, "learning_rate": 9.354377937620717e-06, "loss": 1.7641, "step": 3338 }, { "epoch": 0.8063269741608308, "grad_norm": 0.2840220034122467, "learning_rate": 9.331846757274682e-06, "loss": 1.664, "step": 3339 }, { "epoch": 0.8065684617242213, "grad_norm": 0.29821398854255676, "learning_rate": 9.309339951112889e-06, "loss": 1.6336, "step": 3340 }, { "epoch": 0.8068099492876117, "grad_norm": 0.28731051087379456, "learning_rate": 9.286857532624632e-06, "loss": 1.6755, "step": 3341 }, { "epoch": 0.8070514368510022, "grad_norm": 0.2926536202430725, "learning_rate": 9.264399515284627e-06, "loss": 1.7932, "step": 3342 }, { "epoch": 0.8072929244143927, "grad_norm": 0.30503982305526733, "learning_rate": 9.24196591255292e-06, "loss": 1.8566, "step": 3343 }, { "epoch": 0.8075344119777832, "grad_norm": 0.2795906960964203, "learning_rate": 9.219556737874934e-06, "loss": 1.5399, "step": 3344 }, { "epoch": 0.8077758995411737, "grad_norm": 0.2772836983203888, "learning_rate": 9.197172004681465e-06, "loss": 1.6282, "step": 3345 }, { "epoch": 0.8080173871045642, "grad_norm": 0.2740544378757477, "learning_rate": 9.174811726388632e-06, "loss": 1.6298, "step": 3346 }, { "epoch": 0.8082588746679547, "grad_norm": 0.2901799976825714, "learning_rate": 9.152475916397952e-06, "loss": 1.7716, "step": 3347 }, { "epoch": 0.8085003622313451, "grad_norm": 0.28366419672966003, "learning_rate": 9.13016458809623e-06, "loss": 1.7034, "step": 3348 }, { "epoch": 0.8087418497947356, "grad_norm": 0.29443198442459106, "learning_rate": 9.107877754855598e-06, "loss": 1.7238, "step": 3349 }, { "epoch": 0.8089833373581261, "grad_norm": 0.2834337651729584, "learning_rate": 9.085615430033522e-06, "loss": 1.6691, "step": 3350 }, { "epoch": 0.8092248249215166, "grad_norm": 0.28951868414878845, "learning_rate": 9.063377626972802e-06, "loss": 1.6224, "step": 3351 }, { "epoch": 0.8094663124849071, "grad_norm": 0.2872212529182434, "learning_rate": 9.041164359001475e-06, "loss": 1.5824, "step": 3352 }, { "epoch": 0.8097078000482976, "grad_norm": 0.2743586003780365, "learning_rate": 9.018975639432947e-06, "loss": 1.4758, "step": 3353 }, { "epoch": 0.809949287611688, "grad_norm": 0.2793242335319519, "learning_rate": 8.996811481565864e-06, "loss": 1.7197, "step": 3354 }, { "epoch": 0.8101907751750785, "grad_norm": 0.27659285068511963, "learning_rate": 8.974671898684155e-06, "loss": 1.534, "step": 3355 }, { "epoch": 0.810432262738469, "grad_norm": 0.29640308022499084, "learning_rate": 8.952556904057058e-06, "loss": 1.797, "step": 3356 }, { "epoch": 0.8106737503018595, "grad_norm": 0.42030230164527893, "learning_rate": 8.930466510939028e-06, "loss": 1.7252, "step": 3357 }, { "epoch": 0.81091523786525, "grad_norm": 0.2717979848384857, "learning_rate": 8.9084007325698e-06, "loss": 1.486, "step": 3358 }, { "epoch": 0.8111567254286405, "grad_norm": 0.2941747009754181, "learning_rate": 8.88635958217433e-06, "loss": 1.8005, "step": 3359 }, { "epoch": 0.811398212992031, "grad_norm": 0.3059873580932617, "learning_rate": 8.864343072962878e-06, "loss": 1.8894, "step": 3360 }, { "epoch": 0.8116397005554215, "grad_norm": 0.2849557101726532, "learning_rate": 8.842351218130835e-06, "loss": 1.7134, "step": 3361 }, { "epoch": 0.8118811881188119, "grad_norm": 0.2986636161804199, "learning_rate": 8.820384030858909e-06, "loss": 1.7521, "step": 3362 }, { "epoch": 0.8121226756822024, "grad_norm": 0.30249664187431335, "learning_rate": 8.798441524312978e-06, "loss": 1.8185, "step": 3363 }, { "epoch": 0.8123641632455929, "grad_norm": 0.282980352640152, "learning_rate": 8.776523711644135e-06, "loss": 1.6128, "step": 3364 }, { "epoch": 0.8126056508089833, "grad_norm": 0.317222535610199, "learning_rate": 8.754630605988656e-06, "loss": 2.0867, "step": 3365 }, { "epoch": 0.8128471383723738, "grad_norm": 0.2663152515888214, "learning_rate": 8.732762220468055e-06, "loss": 1.6344, "step": 3366 }, { "epoch": 0.8130886259357643, "grad_norm": 0.2748333513736725, "learning_rate": 8.710918568188992e-06, "loss": 1.5279, "step": 3367 }, { "epoch": 0.8133301134991547, "grad_norm": 0.29500123858451843, "learning_rate": 8.6890996622433e-06, "loss": 1.6194, "step": 3368 }, { "epoch": 0.8135716010625452, "grad_norm": 0.28941595554351807, "learning_rate": 8.667305515708024e-06, "loss": 1.7666, "step": 3369 }, { "epoch": 0.8138130886259357, "grad_norm": 0.2898358702659607, "learning_rate": 8.6455361416453e-06, "loss": 1.5518, "step": 3370 }, { "epoch": 0.8140545761893262, "grad_norm": 0.2836954891681671, "learning_rate": 8.623791553102483e-06, "loss": 1.6483, "step": 3371 }, { "epoch": 0.8142960637527167, "grad_norm": 0.29418620467185974, "learning_rate": 8.602071763112046e-06, "loss": 1.6417, "step": 3372 }, { "epoch": 0.8145375513161072, "grad_norm": 0.2983822524547577, "learning_rate": 8.580376784691584e-06, "loss": 1.7356, "step": 3373 }, { "epoch": 0.8147790388794977, "grad_norm": 0.2827969789505005, "learning_rate": 8.558706630843838e-06, "loss": 1.5887, "step": 3374 }, { "epoch": 0.8150205264428881, "grad_norm": 0.2951814532279968, "learning_rate": 8.537061314556683e-06, "loss": 1.8056, "step": 3375 }, { "epoch": 0.8152620140062786, "grad_norm": 0.29535600543022156, "learning_rate": 8.515440848803086e-06, "loss": 1.6918, "step": 3376 }, { "epoch": 0.8155035015696691, "grad_norm": 0.30457353591918945, "learning_rate": 8.493845246541133e-06, "loss": 1.712, "step": 3377 }, { "epoch": 0.8157449891330596, "grad_norm": 0.2859930098056793, "learning_rate": 8.472274520713991e-06, "loss": 1.643, "step": 3378 }, { "epoch": 0.8159864766964501, "grad_norm": 0.288379967212677, "learning_rate": 8.450728684249925e-06, "loss": 1.6968, "step": 3379 }, { "epoch": 0.8162279642598406, "grad_norm": 0.2931092381477356, "learning_rate": 8.429207750062313e-06, "loss": 1.7067, "step": 3380 }, { "epoch": 0.816469451823231, "grad_norm": 0.29980191588401794, "learning_rate": 8.407711731049567e-06, "loss": 1.8051, "step": 3381 }, { "epoch": 0.8167109393866215, "grad_norm": 0.29367756843566895, "learning_rate": 8.386240640095184e-06, "loss": 1.7483, "step": 3382 }, { "epoch": 0.816952426950012, "grad_norm": 0.28324583172798157, "learning_rate": 8.364794490067702e-06, "loss": 1.7055, "step": 3383 }, { "epoch": 0.8171939145134025, "grad_norm": 0.2929769456386566, "learning_rate": 8.343373293820767e-06, "loss": 1.6644, "step": 3384 }, { "epoch": 0.817435402076793, "grad_norm": 0.27553340792655945, "learning_rate": 8.321977064192976e-06, "loss": 1.7438, "step": 3385 }, { "epoch": 0.8176768896401835, "grad_norm": 0.27319297194480896, "learning_rate": 8.300605814008061e-06, "loss": 1.4832, "step": 3386 }, { "epoch": 0.817918377203574, "grad_norm": 0.29181429743766785, "learning_rate": 8.279259556074714e-06, "loss": 1.7727, "step": 3387 }, { "epoch": 0.8181598647669645, "grad_norm": 0.2925823926925659, "learning_rate": 8.257938303186663e-06, "loss": 1.7503, "step": 3388 }, { "epoch": 0.8184013523303549, "grad_norm": 0.27817782759666443, "learning_rate": 8.236642068122686e-06, "loss": 1.8721, "step": 3389 }, { "epoch": 0.8186428398937454, "grad_norm": 0.3001486659049988, "learning_rate": 8.215370863646515e-06, "loss": 1.74, "step": 3390 }, { "epoch": 0.8188843274571359, "grad_norm": 0.29377415776252747, "learning_rate": 8.194124702506916e-06, "loss": 1.6675, "step": 3391 }, { "epoch": 0.8191258150205264, "grad_norm": 0.2930455207824707, "learning_rate": 8.17290359743762e-06, "loss": 1.8254, "step": 3392 }, { "epoch": 0.8193673025839169, "grad_norm": 0.29683515429496765, "learning_rate": 8.151707561157379e-06, "loss": 1.7252, "step": 3393 }, { "epoch": 0.8196087901473074, "grad_norm": 0.3075847923755646, "learning_rate": 8.13053660636986e-06, "loss": 1.8765, "step": 3394 }, { "epoch": 0.8198502777106979, "grad_norm": 0.288847416639328, "learning_rate": 8.109390745763756e-06, "loss": 1.6931, "step": 3395 }, { "epoch": 0.8200917652740883, "grad_norm": 0.30925628542900085, "learning_rate": 8.088269992012692e-06, "loss": 1.8384, "step": 3396 }, { "epoch": 0.8203332528374788, "grad_norm": 0.2875019311904907, "learning_rate": 8.067174357775248e-06, "loss": 1.7044, "step": 3397 }, { "epoch": 0.8205747404008693, "grad_norm": 0.2871297001838684, "learning_rate": 8.046103855694942e-06, "loss": 1.6674, "step": 3398 }, { "epoch": 0.8208162279642598, "grad_norm": 0.2761847972869873, "learning_rate": 8.025058498400251e-06, "loss": 1.5616, "step": 3399 }, { "epoch": 0.8210577155276503, "grad_norm": 0.2757909893989563, "learning_rate": 8.004038298504569e-06, "loss": 1.4904, "step": 3400 }, { "epoch": 0.8212992030910408, "grad_norm": 0.29062095284461975, "learning_rate": 7.983043268606195e-06, "loss": 1.5498, "step": 3401 }, { "epoch": 0.8215406906544312, "grad_norm": 0.2779589593410492, "learning_rate": 7.962073421288391e-06, "loss": 1.6617, "step": 3402 }, { "epoch": 0.8217821782178217, "grad_norm": 0.2934403717517853, "learning_rate": 7.94112876911925e-06, "loss": 1.7207, "step": 3403 }, { "epoch": 0.8220236657812122, "grad_norm": 0.29695427417755127, "learning_rate": 7.920209324651846e-06, "loss": 1.796, "step": 3404 }, { "epoch": 0.8222651533446027, "grad_norm": 0.3044850826263428, "learning_rate": 7.899315100424093e-06, "loss": 1.9633, "step": 3405 }, { "epoch": 0.8225066409079932, "grad_norm": 0.28858712315559387, "learning_rate": 7.878446108958803e-06, "loss": 1.6384, "step": 3406 }, { "epoch": 0.8227481284713837, "grad_norm": 0.275881826877594, "learning_rate": 7.857602362763655e-06, "loss": 1.6978, "step": 3407 }, { "epoch": 0.8229896160347742, "grad_norm": 0.3458348512649536, "learning_rate": 7.836783874331233e-06, "loss": 2.1767, "step": 3408 }, { "epoch": 0.8232311035981646, "grad_norm": 0.28876399993896484, "learning_rate": 7.815990656138943e-06, "loss": 1.7108, "step": 3409 }, { "epoch": 0.8234725911615551, "grad_norm": 0.2833706736564636, "learning_rate": 7.795222720649065e-06, "loss": 1.6326, "step": 3410 }, { "epoch": 0.8237140787249456, "grad_norm": 0.30351823568344116, "learning_rate": 7.774480080308715e-06, "loss": 2.035, "step": 3411 }, { "epoch": 0.8239555662883361, "grad_norm": 0.28744566440582275, "learning_rate": 7.753762747549847e-06, "loss": 1.6541, "step": 3412 }, { "epoch": 0.8241970538517266, "grad_norm": 0.28277140855789185, "learning_rate": 7.733070734789283e-06, "loss": 1.5859, "step": 3413 }, { "epoch": 0.8244385414151171, "grad_norm": 0.2819003164768219, "learning_rate": 7.712404054428623e-06, "loss": 1.7276, "step": 3414 }, { "epoch": 0.8246800289785076, "grad_norm": 0.2814866900444031, "learning_rate": 7.691762718854312e-06, "loss": 1.7448, "step": 3415 }, { "epoch": 0.824921516541898, "grad_norm": 0.30012404918670654, "learning_rate": 7.671146740437569e-06, "loss": 1.6757, "step": 3416 }, { "epoch": 0.8251630041052885, "grad_norm": 0.28455424308776855, "learning_rate": 7.650556131534492e-06, "loss": 1.784, "step": 3417 }, { "epoch": 0.825404491668679, "grad_norm": 0.3023732602596283, "learning_rate": 7.629990904485868e-06, "loss": 1.7423, "step": 3418 }, { "epoch": 0.8256459792320695, "grad_norm": 0.3087024986743927, "learning_rate": 7.609451071617368e-06, "loss": 1.7693, "step": 3419 }, { "epoch": 0.82588746679546, "grad_norm": 0.2780247628688812, "learning_rate": 7.588936645239392e-06, "loss": 1.607, "step": 3420 }, { "epoch": 0.8261289543588505, "grad_norm": 0.28824329376220703, "learning_rate": 7.568447637647103e-06, "loss": 1.6424, "step": 3421 }, { "epoch": 0.826370441922241, "grad_norm": 0.2860906422138214, "learning_rate": 7.547984061120483e-06, "loss": 1.7018, "step": 3422 }, { "epoch": 0.8266119294856314, "grad_norm": 0.29281318187713623, "learning_rate": 7.527545927924224e-06, "loss": 1.6711, "step": 3423 }, { "epoch": 0.8268534170490219, "grad_norm": 0.2928633987903595, "learning_rate": 7.507133250307779e-06, "loss": 1.553, "step": 3424 }, { "epoch": 0.8270949046124124, "grad_norm": 0.28861743211746216, "learning_rate": 7.486746040505343e-06, "loss": 1.7755, "step": 3425 }, { "epoch": 0.8273363921758029, "grad_norm": 0.2941233515739441, "learning_rate": 7.4663843107358915e-06, "loss": 1.5538, "step": 3426 }, { "epoch": 0.8275778797391934, "grad_norm": 0.281040221452713, "learning_rate": 7.446048073203033e-06, "loss": 1.6009, "step": 3427 }, { "epoch": 0.8278193673025839, "grad_norm": 0.28734269738197327, "learning_rate": 7.425737340095196e-06, "loss": 1.6083, "step": 3428 }, { "epoch": 0.8280608548659744, "grad_norm": 0.29222771525382996, "learning_rate": 7.4054521235854765e-06, "loss": 1.6689, "step": 3429 }, { "epoch": 0.8283023424293648, "grad_norm": 0.2768942713737488, "learning_rate": 7.385192435831673e-06, "loss": 1.6528, "step": 3430 }, { "epoch": 0.8285438299927553, "grad_norm": 0.28925302624702454, "learning_rate": 7.364958288976281e-06, "loss": 1.6402, "step": 3431 }, { "epoch": 0.8287853175561458, "grad_norm": 0.3015986979007721, "learning_rate": 7.34474969514653e-06, "loss": 1.773, "step": 3432 }, { "epoch": 0.8290268051195363, "grad_norm": 0.3007544279098511, "learning_rate": 7.324566666454291e-06, "loss": 1.675, "step": 3433 }, { "epoch": 0.8292682926829268, "grad_norm": 0.29716983437538147, "learning_rate": 7.30440921499611e-06, "loss": 1.7674, "step": 3434 }, { "epoch": 0.8295097802463173, "grad_norm": 0.2843656837940216, "learning_rate": 7.284277352853264e-06, "loss": 1.8757, "step": 3435 }, { "epoch": 0.8297512678097078, "grad_norm": 0.2758077085018158, "learning_rate": 7.2641710920915986e-06, "loss": 1.6086, "step": 3436 }, { "epoch": 0.8299927553730982, "grad_norm": 0.2896127998828888, "learning_rate": 7.244090444761697e-06, "loss": 1.7197, "step": 3437 }, { "epoch": 0.8302342429364887, "grad_norm": 0.28539329767227173, "learning_rate": 7.224035422898756e-06, "loss": 1.5659, "step": 3438 }, { "epoch": 0.8304757304998792, "grad_norm": 0.30481478571891785, "learning_rate": 7.204006038522609e-06, "loss": 1.773, "step": 3439 }, { "epoch": 0.8307172180632697, "grad_norm": 0.2793824374675751, "learning_rate": 7.1840023036377325e-06, "loss": 1.5323, "step": 3440 }, { "epoch": 0.8309587056266602, "grad_norm": 0.3147065043449402, "learning_rate": 7.16402423023324e-06, "loss": 1.8955, "step": 3441 }, { "epoch": 0.8312001931900507, "grad_norm": 0.285542368888855, "learning_rate": 7.144071830282861e-06, "loss": 1.6913, "step": 3442 }, { "epoch": 0.8314416807534412, "grad_norm": 0.29227393865585327, "learning_rate": 7.1241451157449125e-06, "loss": 1.566, "step": 3443 }, { "epoch": 0.8316831683168316, "grad_norm": 0.2829485833644867, "learning_rate": 7.10424409856234e-06, "loss": 1.5929, "step": 3444 }, { "epoch": 0.8319246558802221, "grad_norm": 0.2756589949131012, "learning_rate": 7.084368790662704e-06, "loss": 1.6546, "step": 3445 }, { "epoch": 0.8321661434436126, "grad_norm": 0.28917887806892395, "learning_rate": 7.064519203958126e-06, "loss": 1.6386, "step": 3446 }, { "epoch": 0.8324076310070031, "grad_norm": 0.37987956404685974, "learning_rate": 7.044695350345304e-06, "loss": 1.7416, "step": 3447 }, { "epoch": 0.8326491185703936, "grad_norm": 0.2843257486820221, "learning_rate": 7.024897241705575e-06, "loss": 1.7021, "step": 3448 }, { "epoch": 0.8328906061337841, "grad_norm": 0.31252092123031616, "learning_rate": 7.0051248899047515e-06, "loss": 2.0006, "step": 3449 }, { "epoch": 0.8331320936971746, "grad_norm": 0.3045337200164795, "learning_rate": 6.9853783067932946e-06, "loss": 1.8593, "step": 3450 }, { "epoch": 0.833373581260565, "grad_norm": 0.3015189468860626, "learning_rate": 6.965657504206175e-06, "loss": 1.7762, "step": 3451 }, { "epoch": 0.8336150688239555, "grad_norm": 0.28390347957611084, "learning_rate": 6.945962493962921e-06, "loss": 1.7287, "step": 3452 }, { "epoch": 0.833856556387346, "grad_norm": 0.28913745284080505, "learning_rate": 6.926293287867597e-06, "loss": 1.7217, "step": 3453 }, { "epoch": 0.8340980439507365, "grad_norm": 0.2955908179283142, "learning_rate": 6.906649897708828e-06, "loss": 1.777, "step": 3454 }, { "epoch": 0.834339531514127, "grad_norm": 0.28371500968933105, "learning_rate": 6.887032335259741e-06, "loss": 1.6873, "step": 3455 }, { "epoch": 0.8345810190775175, "grad_norm": 0.29068586230278015, "learning_rate": 6.867440612277976e-06, "loss": 1.641, "step": 3456 }, { "epoch": 0.834822506640908, "grad_norm": 0.2878762483596802, "learning_rate": 6.847874740505733e-06, "loss": 1.6861, "step": 3457 }, { "epoch": 0.8350639942042984, "grad_norm": 0.33472228050231934, "learning_rate": 6.82833473166965e-06, "loss": 1.8475, "step": 3458 }, { "epoch": 0.8353054817676889, "grad_norm": 0.28831198811531067, "learning_rate": 6.808820597480919e-06, "loss": 1.6685, "step": 3459 }, { "epoch": 0.8355469693310794, "grad_norm": 0.2853391170501709, "learning_rate": 6.789332349635203e-06, "loss": 1.6969, "step": 3460 }, { "epoch": 0.8357884568944699, "grad_norm": 0.28601667284965515, "learning_rate": 6.7698699998126505e-06, "loss": 1.6259, "step": 3461 }, { "epoch": 0.8360299444578604, "grad_norm": 0.27889561653137207, "learning_rate": 6.750433559677882e-06, "loss": 1.6954, "step": 3462 }, { "epoch": 0.8362714320212509, "grad_norm": 0.3119412958621979, "learning_rate": 6.731023040880024e-06, "loss": 1.8028, "step": 3463 }, { "epoch": 0.8365129195846414, "grad_norm": 0.33762526512145996, "learning_rate": 6.711638455052605e-06, "loss": 1.7111, "step": 3464 }, { "epoch": 0.8367544071480318, "grad_norm": 0.2838843762874603, "learning_rate": 6.692279813813668e-06, "loss": 1.639, "step": 3465 }, { "epoch": 0.8369958947114223, "grad_norm": 0.305677592754364, "learning_rate": 6.672947128765683e-06, "loss": 1.5554, "step": 3466 }, { "epoch": 0.8372373822748128, "grad_norm": 0.28526613116264343, "learning_rate": 6.6536404114955484e-06, "loss": 1.7968, "step": 3467 }, { "epoch": 0.8374788698382033, "grad_norm": 0.2730858325958252, "learning_rate": 6.634359673574636e-06, "loss": 1.6149, "step": 3468 }, { "epoch": 0.8377203574015938, "grad_norm": 0.2923898994922638, "learning_rate": 6.615104926558724e-06, "loss": 1.716, "step": 3469 }, { "epoch": 0.8379618449649843, "grad_norm": 0.30661213397979736, "learning_rate": 6.595876181988009e-06, "loss": 1.8459, "step": 3470 }, { "epoch": 0.8382033325283748, "grad_norm": 0.28740110993385315, "learning_rate": 6.5766734513870996e-06, "loss": 1.6709, "step": 3471 }, { "epoch": 0.8384448200917652, "grad_norm": 0.2901138961315155, "learning_rate": 6.557496746265057e-06, "loss": 1.7219, "step": 3472 }, { "epoch": 0.8386863076551557, "grad_norm": 0.2874894142150879, "learning_rate": 6.53834607811527e-06, "loss": 1.6067, "step": 3473 }, { "epoch": 0.8389277952185462, "grad_norm": 0.303545743227005, "learning_rate": 6.5192214584155864e-06, "loss": 1.831, "step": 3474 }, { "epoch": 0.8391692827819367, "grad_norm": 0.2856457829475403, "learning_rate": 6.5001228986282194e-06, "loss": 1.6695, "step": 3475 }, { "epoch": 0.8394107703453272, "grad_norm": 0.2787890136241913, "learning_rate": 6.4810504101997515e-06, "loss": 1.7909, "step": 3476 }, { "epoch": 0.8396522579087177, "grad_norm": 0.28715819120407104, "learning_rate": 6.462004004561145e-06, "loss": 1.6821, "step": 3477 }, { "epoch": 0.8398937454721082, "grad_norm": 0.2893564701080322, "learning_rate": 6.442983693127758e-06, "loss": 1.6961, "step": 3478 }, { "epoch": 0.8401352330354986, "grad_norm": 0.2894771993160248, "learning_rate": 6.42398948729927e-06, "loss": 1.7755, "step": 3479 }, { "epoch": 0.8403767205988891, "grad_norm": 0.28807491064071655, "learning_rate": 6.405021398459726e-06, "loss": 1.6594, "step": 3480 }, { "epoch": 0.8406182081622796, "grad_norm": 0.28036779165267944, "learning_rate": 6.386079437977549e-06, "loss": 1.6076, "step": 3481 }, { "epoch": 0.8408596957256701, "grad_norm": 0.2892704904079437, "learning_rate": 6.3671636172054405e-06, "loss": 1.7329, "step": 3482 }, { "epoch": 0.8411011832890606, "grad_norm": 0.28481730818748474, "learning_rate": 6.348273947480493e-06, "loss": 1.6885, "step": 3483 }, { "epoch": 0.8413426708524511, "grad_norm": 0.2941053807735443, "learning_rate": 6.3294104401241e-06, "loss": 1.7411, "step": 3484 }, { "epoch": 0.8415841584158416, "grad_norm": 0.27901408076286316, "learning_rate": 6.310573106441975e-06, "loss": 1.3937, "step": 3485 }, { "epoch": 0.841825645979232, "grad_norm": 0.296093225479126, "learning_rate": 6.291761957724141e-06, "loss": 1.7963, "step": 3486 }, { "epoch": 0.8420671335426225, "grad_norm": 0.2775656580924988, "learning_rate": 6.272977005244945e-06, "loss": 1.5445, "step": 3487 }, { "epoch": 0.842308621106013, "grad_norm": 0.27906760573387146, "learning_rate": 6.254218260263017e-06, "loss": 1.7398, "step": 3488 }, { "epoch": 0.8425501086694035, "grad_norm": 0.2945201098918915, "learning_rate": 6.235485734021279e-06, "loss": 1.6905, "step": 3489 }, { "epoch": 0.842791596232794, "grad_norm": 0.3001944124698639, "learning_rate": 6.216779437746967e-06, "loss": 1.8119, "step": 3490 }, { "epoch": 0.8430330837961845, "grad_norm": 0.28580793738365173, "learning_rate": 6.198099382651534e-06, "loss": 1.6401, "step": 3491 }, { "epoch": 0.843274571359575, "grad_norm": 0.28856202960014343, "learning_rate": 6.1794455799307875e-06, "loss": 1.5049, "step": 3492 }, { "epoch": 0.8435160589229654, "grad_norm": 0.30331215262413025, "learning_rate": 6.1608180407647324e-06, "loss": 1.8694, "step": 3493 }, { "epoch": 0.8437575464863559, "grad_norm": 0.29826226830482483, "learning_rate": 6.142216776317666e-06, "loss": 1.7555, "step": 3494 }, { "epoch": 0.8439990340497464, "grad_norm": 0.27153995633125305, "learning_rate": 6.123641797738127e-06, "loss": 1.6205, "step": 3495 }, { "epoch": 0.8442405216131369, "grad_norm": 0.28906306624412537, "learning_rate": 6.105093116158927e-06, "loss": 1.6384, "step": 3496 }, { "epoch": 0.8444820091765274, "grad_norm": 0.28365620970726013, "learning_rate": 6.086570742697057e-06, "loss": 1.7598, "step": 3497 }, { "epoch": 0.8447234967399179, "grad_norm": 0.2755637764930725, "learning_rate": 6.0680746884538005e-06, "loss": 1.5731, "step": 3498 }, { "epoch": 0.8449649843033084, "grad_norm": 0.29565727710723877, "learning_rate": 6.049604964514644e-06, "loss": 1.6236, "step": 3499 }, { "epoch": 0.8452064718666988, "grad_norm": 0.2774536609649658, "learning_rate": 6.031161581949268e-06, "loss": 1.7424, "step": 3500 }, { "epoch": 0.8454479594300893, "grad_norm": 0.29315805435180664, "learning_rate": 6.012744551811627e-06, "loss": 1.7142, "step": 3501 }, { "epoch": 0.8456894469934798, "grad_norm": 0.28174543380737305, "learning_rate": 5.99435388513982e-06, "loss": 1.6779, "step": 3502 }, { "epoch": 0.8459309345568703, "grad_norm": 0.2805562913417816, "learning_rate": 5.975989592956177e-06, "loss": 1.7292, "step": 3503 }, { "epoch": 0.8461724221202608, "grad_norm": 0.2782053053379059, "learning_rate": 5.957651686267202e-06, "loss": 1.6408, "step": 3504 }, { "epoch": 0.8464139096836513, "grad_norm": 0.28644683957099915, "learning_rate": 5.939340176063623e-06, "loss": 1.6612, "step": 3505 }, { "epoch": 0.8466553972470418, "grad_norm": 0.2986832559108734, "learning_rate": 5.921055073320281e-06, "loss": 1.7914, "step": 3506 }, { "epoch": 0.8468968848104322, "grad_norm": 0.2919190227985382, "learning_rate": 5.902796388996262e-06, "loss": 1.6913, "step": 3507 }, { "epoch": 0.8471383723738227, "grad_norm": 0.29307764768600464, "learning_rate": 5.884564134034776e-06, "loss": 1.8325, "step": 3508 }, { "epoch": 0.8473798599372132, "grad_norm": 0.3066037893295288, "learning_rate": 5.866358319363191e-06, "loss": 1.5635, "step": 3509 }, { "epoch": 0.8476213475006037, "grad_norm": 0.28856411576271057, "learning_rate": 5.8481789558930634e-06, "loss": 1.6504, "step": 3510 }, { "epoch": 0.8478628350639942, "grad_norm": 0.3006540834903717, "learning_rate": 5.8300260545200545e-06, "loss": 1.864, "step": 3511 }, { "epoch": 0.8481043226273847, "grad_norm": 0.29206767678260803, "learning_rate": 5.81189962612399e-06, "loss": 1.7415, "step": 3512 }, { "epoch": 0.8483458101907752, "grad_norm": 0.28227290511131287, "learning_rate": 5.793799681568807e-06, "loss": 1.5805, "step": 3513 }, { "epoch": 0.8485872977541656, "grad_norm": 0.3129713237285614, "learning_rate": 5.775726231702622e-06, "loss": 1.9378, "step": 3514 }, { "epoch": 0.8488287853175561, "grad_norm": 0.27972254157066345, "learning_rate": 5.757679287357598e-06, "loss": 1.6597, "step": 3515 }, { "epoch": 0.8490702728809466, "grad_norm": 0.2852473855018616, "learning_rate": 5.739658859350066e-06, "loss": 1.6855, "step": 3516 }, { "epoch": 0.8493117604443371, "grad_norm": 0.29172688722610474, "learning_rate": 5.721664958480455e-06, "loss": 1.7677, "step": 3517 }, { "epoch": 0.8495532480077276, "grad_norm": 0.27997854351997375, "learning_rate": 5.703697595533275e-06, "loss": 1.692, "step": 3518 }, { "epoch": 0.8497947355711181, "grad_norm": 0.2789522409439087, "learning_rate": 5.685756781277146e-06, "loss": 1.6757, "step": 3519 }, { "epoch": 0.8500362231345086, "grad_norm": 0.2771410048007965, "learning_rate": 5.6678425264647836e-06, "loss": 1.5492, "step": 3520 }, { "epoch": 0.850277710697899, "grad_norm": 0.2927265167236328, "learning_rate": 5.6499548418329714e-06, "loss": 1.6318, "step": 3521 }, { "epoch": 0.8505191982612895, "grad_norm": 0.27935269474983215, "learning_rate": 5.632093738102567e-06, "loss": 1.6649, "step": 3522 }, { "epoch": 0.85076068582468, "grad_norm": 0.29220134019851685, "learning_rate": 5.614259225978524e-06, "loss": 1.748, "step": 3523 }, { "epoch": 0.8510021733880705, "grad_norm": 0.2834475338459015, "learning_rate": 5.596451316149803e-06, "loss": 1.69, "step": 3524 }, { "epoch": 0.851243660951461, "grad_norm": 0.2882104218006134, "learning_rate": 5.578670019289484e-06, "loss": 1.5744, "step": 3525 }, { "epoch": 0.8514851485148515, "grad_norm": 0.3032463788986206, "learning_rate": 5.560915346054657e-06, "loss": 1.8577, "step": 3526 }, { "epoch": 0.851726636078242, "grad_norm": 0.2897374927997589, "learning_rate": 5.543187307086467e-06, "loss": 1.6271, "step": 3527 }, { "epoch": 0.8519681236416324, "grad_norm": 0.2905126214027405, "learning_rate": 5.525485913010081e-06, "loss": 1.7439, "step": 3528 }, { "epoch": 0.8522096112050229, "grad_norm": 0.2841590344905853, "learning_rate": 5.507811174434746e-06, "loss": 1.6666, "step": 3529 }, { "epoch": 0.8524510987684134, "grad_norm": 0.3335350453853607, "learning_rate": 5.49016310195366e-06, "loss": 1.8109, "step": 3530 }, { "epoch": 0.8526925863318039, "grad_norm": 0.2881028354167938, "learning_rate": 5.472541706144096e-06, "loss": 1.7412, "step": 3531 }, { "epoch": 0.8529340738951944, "grad_norm": 0.2817363440990448, "learning_rate": 5.4549469975673174e-06, "loss": 1.7455, "step": 3532 }, { "epoch": 0.8531755614585849, "grad_norm": 0.36139267683029175, "learning_rate": 5.437378986768582e-06, "loss": 1.6258, "step": 3533 }, { "epoch": 0.8534170490219753, "grad_norm": 0.2768571972846985, "learning_rate": 5.419837684277185e-06, "loss": 1.6229, "step": 3534 }, { "epoch": 0.8536585365853658, "grad_norm": 0.28267383575439453, "learning_rate": 5.402323100606366e-06, "loss": 1.635, "step": 3535 }, { "epoch": 0.8539000241487563, "grad_norm": 0.2810875475406647, "learning_rate": 5.384835246253384e-06, "loss": 1.4505, "step": 3536 }, { "epoch": 0.8541415117121468, "grad_norm": 0.2848489284515381, "learning_rate": 5.367374131699448e-06, "loss": 1.6605, "step": 3537 }, { "epoch": 0.8543829992755373, "grad_norm": 0.2918657958507538, "learning_rate": 5.349939767409795e-06, "loss": 1.6647, "step": 3538 }, { "epoch": 0.8546244868389278, "grad_norm": 0.28559190034866333, "learning_rate": 5.332532163833553e-06, "loss": 1.6731, "step": 3539 }, { "epoch": 0.8548659744023183, "grad_norm": 0.288510799407959, "learning_rate": 5.315151331403884e-06, "loss": 1.6906, "step": 3540 }, { "epoch": 0.8551074619657087, "grad_norm": 0.2950460910797119, "learning_rate": 5.297797280537864e-06, "loss": 1.6865, "step": 3541 }, { "epoch": 0.8553489495290992, "grad_norm": 0.28459155559539795, "learning_rate": 5.280470021636513e-06, "loss": 1.6816, "step": 3542 }, { "epoch": 0.8555904370924897, "grad_norm": 0.2667498290538788, "learning_rate": 5.263169565084825e-06, "loss": 1.5251, "step": 3543 }, { "epoch": 0.8558319246558802, "grad_norm": 0.28919240832328796, "learning_rate": 5.2458959212517085e-06, "loss": 1.5304, "step": 3544 }, { "epoch": 0.8560734122192707, "grad_norm": 0.2731263339519501, "learning_rate": 5.228649100490002e-06, "loss": 1.6409, "step": 3545 }, { "epoch": 0.8563148997826612, "grad_norm": 0.28352221846580505, "learning_rate": 5.211429113136462e-06, "loss": 1.6555, "step": 3546 }, { "epoch": 0.8565563873460517, "grad_norm": 0.3019556403160095, "learning_rate": 5.1942359695118016e-06, "loss": 1.7653, "step": 3547 }, { "epoch": 0.8567978749094421, "grad_norm": 0.27253273129463196, "learning_rate": 5.177069679920576e-06, "loss": 1.6479, "step": 3548 }, { "epoch": 0.8570393624728326, "grad_norm": 0.30235016345977783, "learning_rate": 5.1599302546513186e-06, "loss": 1.7307, "step": 3549 }, { "epoch": 0.8572808500362231, "grad_norm": 0.30397292971611023, "learning_rate": 5.142817703976416e-06, "loss": 1.6984, "step": 3550 }, { "epoch": 0.8575223375996136, "grad_norm": 0.29515138268470764, "learning_rate": 5.125732038152165e-06, "loss": 1.7683, "step": 3551 }, { "epoch": 0.8577638251630041, "grad_norm": 0.30275678634643555, "learning_rate": 5.1086732674187235e-06, "loss": 1.6125, "step": 3552 }, { "epoch": 0.8580053127263946, "grad_norm": 0.2966776490211487, "learning_rate": 5.09164140200018e-06, "loss": 1.9035, "step": 3553 }, { "epoch": 0.8582468002897851, "grad_norm": 0.30367422103881836, "learning_rate": 5.074636452104459e-06, "loss": 1.8248, "step": 3554 }, { "epoch": 0.8584882878531755, "grad_norm": 0.2948387861251831, "learning_rate": 5.0576584279233506e-06, "loss": 1.7842, "step": 3555 }, { "epoch": 0.858729775416566, "grad_norm": 0.2952226400375366, "learning_rate": 5.040707339632544e-06, "loss": 1.7137, "step": 3556 }, { "epoch": 0.8589712629799565, "grad_norm": 0.3093251585960388, "learning_rate": 5.023783197391529e-06, "loss": 1.6151, "step": 3557 }, { "epoch": 0.859212750543347, "grad_norm": 0.2945159375667572, "learning_rate": 5.0068860113437036e-06, "loss": 1.7464, "step": 3558 }, { "epoch": 0.8594542381067375, "grad_norm": 0.29917454719543457, "learning_rate": 4.9900157916162705e-06, "loss": 1.7665, "step": 3559 }, { "epoch": 0.859695725670128, "grad_norm": 0.2902509570121765, "learning_rate": 4.973172548320288e-06, "loss": 1.7194, "step": 3560 }, { "epoch": 0.8599372132335185, "grad_norm": 0.3060162365436554, "learning_rate": 4.956356291550618e-06, "loss": 1.7651, "step": 3561 }, { "epoch": 0.860178700796909, "grad_norm": 0.2775404453277588, "learning_rate": 4.939567031386016e-06, "loss": 1.6402, "step": 3562 }, { "epoch": 0.8604201883602994, "grad_norm": 0.2733018100261688, "learning_rate": 4.9228047778889615e-06, "loss": 1.705, "step": 3563 }, { "epoch": 0.8606616759236899, "grad_norm": 0.29368650913238525, "learning_rate": 4.906069541105834e-06, "loss": 1.6573, "step": 3564 }, { "epoch": 0.8609031634870804, "grad_norm": 0.2898925840854645, "learning_rate": 4.889361331066777e-06, "loss": 1.7358, "step": 3565 }, { "epoch": 0.8611446510504709, "grad_norm": 0.2868831753730774, "learning_rate": 4.8726801577857305e-06, "loss": 1.6845, "step": 3566 }, { "epoch": 0.8613861386138614, "grad_norm": 0.2895372211933136, "learning_rate": 4.856026031260457e-06, "loss": 1.6342, "step": 3567 }, { "epoch": 0.8616276261772519, "grad_norm": 0.2928381562232971, "learning_rate": 4.839398961472497e-06, "loss": 1.7245, "step": 3568 }, { "epoch": 0.8618691137406423, "grad_norm": 0.28802043199539185, "learning_rate": 4.82279895838717e-06, "loss": 1.7636, "step": 3569 }, { "epoch": 0.8621106013040328, "grad_norm": 0.29191848635673523, "learning_rate": 4.806226031953564e-06, "loss": 1.7149, "step": 3570 }, { "epoch": 0.8623520888674233, "grad_norm": 0.28406769037246704, "learning_rate": 4.789680192104567e-06, "loss": 1.6755, "step": 3571 }, { "epoch": 0.8625935764308138, "grad_norm": 0.28282323479652405, "learning_rate": 4.773161448756813e-06, "loss": 1.651, "step": 3572 }, { "epoch": 0.8628350639942043, "grad_norm": 0.2820364236831665, "learning_rate": 4.756669811810699e-06, "loss": 1.7381, "step": 3573 }, { "epoch": 0.8630765515575948, "grad_norm": 0.2893783152103424, "learning_rate": 4.7402052911503625e-06, "loss": 1.5578, "step": 3574 }, { "epoch": 0.8633180391209853, "grad_norm": 0.2835719883441925, "learning_rate": 4.723767896643732e-06, "loss": 1.7331, "step": 3575 }, { "epoch": 0.8635595266843757, "grad_norm": 0.28912338614463806, "learning_rate": 4.707357638142429e-06, "loss": 1.681, "step": 3576 }, { "epoch": 0.8638010142477662, "grad_norm": 0.28847548365592957, "learning_rate": 4.690974525481839e-06, "loss": 1.6416, "step": 3577 }, { "epoch": 0.8640425018111567, "grad_norm": 0.2889868915081024, "learning_rate": 4.674618568481065e-06, "loss": 1.7727, "step": 3578 }, { "epoch": 0.8642839893745472, "grad_norm": 0.3136354088783264, "learning_rate": 4.658289776942937e-06, "loss": 1.8924, "step": 3579 }, { "epoch": 0.8645254769379377, "grad_norm": 0.31432783603668213, "learning_rate": 4.64198816065402e-06, "loss": 1.8953, "step": 3580 }, { "epoch": 0.8647669645013282, "grad_norm": 0.2839619517326355, "learning_rate": 4.625713729384568e-06, "loss": 1.6365, "step": 3581 }, { "epoch": 0.8650084520647187, "grad_norm": 0.29579365253448486, "learning_rate": 4.609466492888559e-06, "loss": 1.7411, "step": 3582 }, { "epoch": 0.8652499396281091, "grad_norm": 0.2694234848022461, "learning_rate": 4.593246460903644e-06, "loss": 1.5635, "step": 3583 }, { "epoch": 0.8654914271914996, "grad_norm": 0.2970821261405945, "learning_rate": 4.57705364315123e-06, "loss": 1.7972, "step": 3584 }, { "epoch": 0.8657329147548901, "grad_norm": 0.28647640347480774, "learning_rate": 4.560888049336331e-06, "loss": 1.5809, "step": 3585 }, { "epoch": 0.8659744023182806, "grad_norm": 0.2875736355781555, "learning_rate": 4.544749689147715e-06, "loss": 1.7593, "step": 3586 }, { "epoch": 0.8662158898816711, "grad_norm": 0.28044697642326355, "learning_rate": 4.528638572257793e-06, "loss": 1.6717, "step": 3587 }, { "epoch": 0.8664573774450616, "grad_norm": 0.3051677942276001, "learning_rate": 4.512554708322636e-06, "loss": 1.8281, "step": 3588 }, { "epoch": 0.8666988650084521, "grad_norm": 0.2829369008541107, "learning_rate": 4.496498106982034e-06, "loss": 1.6521, "step": 3589 }, { "epoch": 0.8669403525718425, "grad_norm": 0.2796487808227539, "learning_rate": 4.480468777859375e-06, "loss": 1.5651, "step": 3590 }, { "epoch": 0.867181840135233, "grad_norm": 0.3046395182609558, "learning_rate": 4.464466730561745e-06, "loss": 1.8054, "step": 3591 }, { "epoch": 0.8674233276986235, "grad_norm": 0.28046467900276184, "learning_rate": 4.448491974679853e-06, "loss": 1.7227, "step": 3592 }, { "epoch": 0.867664815262014, "grad_norm": 0.2821386754512787, "learning_rate": 4.432544519788079e-06, "loss": 1.6205, "step": 3593 }, { "epoch": 0.8679063028254045, "grad_norm": 0.3085712492465973, "learning_rate": 4.416624375444395e-06, "loss": 1.8391, "step": 3594 }, { "epoch": 0.868147790388795, "grad_norm": 0.266520231962204, "learning_rate": 4.400731551190451e-06, "loss": 1.5565, "step": 3595 }, { "epoch": 0.8683892779521855, "grad_norm": 0.28883063793182373, "learning_rate": 4.384866056551506e-06, "loss": 1.7738, "step": 3596 }, { "epoch": 0.8686307655155759, "grad_norm": 0.2957184910774231, "learning_rate": 4.369027901036432e-06, "loss": 1.8211, "step": 3597 }, { "epoch": 0.8688722530789664, "grad_norm": 0.3059837818145752, "learning_rate": 4.353217094137718e-06, "loss": 1.873, "step": 3598 }, { "epoch": 0.8691137406423569, "grad_norm": 0.30132120847702026, "learning_rate": 4.337433645331474e-06, "loss": 1.8072, "step": 3599 }, { "epoch": 0.8693552282057474, "grad_norm": 0.27254632115364075, "learning_rate": 4.321677564077398e-06, "loss": 1.4992, "step": 3600 }, { "epoch": 0.8695967157691379, "grad_norm": 0.2872336804866791, "learning_rate": 4.305948859818781e-06, "loss": 1.7752, "step": 3601 }, { "epoch": 0.8698382033325284, "grad_norm": 0.2907829284667969, "learning_rate": 4.290247541982545e-06, "loss": 1.8192, "step": 3602 }, { "epoch": 0.8700796908959189, "grad_norm": 0.3041236698627472, "learning_rate": 4.2745736199791386e-06, "loss": 1.8647, "step": 3603 }, { "epoch": 0.8703211784593093, "grad_norm": 0.28655949234962463, "learning_rate": 4.258927103202637e-06, "loss": 1.5526, "step": 3604 }, { "epoch": 0.8705626660226998, "grad_norm": 0.3033926784992218, "learning_rate": 4.243308001030677e-06, "loss": 1.7125, "step": 3605 }, { "epoch": 0.8708041535860903, "grad_norm": 0.2886970043182373, "learning_rate": 4.227716322824459e-06, "loss": 1.6728, "step": 3606 }, { "epoch": 0.8710456411494808, "grad_norm": 0.29204967617988586, "learning_rate": 4.212152077928738e-06, "loss": 1.6524, "step": 3607 }, { "epoch": 0.8712871287128713, "grad_norm": 0.28188931941986084, "learning_rate": 4.196615275671856e-06, "loss": 1.5715, "step": 3608 }, { "epoch": 0.8715286162762618, "grad_norm": 0.2849540114402771, "learning_rate": 4.181105925365691e-06, "loss": 1.7302, "step": 3609 }, { "epoch": 0.8717701038396523, "grad_norm": 0.27933409810066223, "learning_rate": 4.165624036305659e-06, "loss": 1.6012, "step": 3610 }, { "epoch": 0.8720115914030427, "grad_norm": 0.2898978888988495, "learning_rate": 4.150169617770733e-06, "loss": 1.7171, "step": 3611 }, { "epoch": 0.8722530789664332, "grad_norm": 0.2901992201805115, "learning_rate": 4.134742679023401e-06, "loss": 1.6086, "step": 3612 }, { "epoch": 0.8724945665298237, "grad_norm": 0.3406147360801697, "learning_rate": 4.119343229309719e-06, "loss": 1.848, "step": 3613 }, { "epoch": 0.8727360540932142, "grad_norm": 0.2773250341415405, "learning_rate": 4.103971277859225e-06, "loss": 1.6724, "step": 3614 }, { "epoch": 0.8729775416566047, "grad_norm": 0.30533015727996826, "learning_rate": 4.088626833885001e-06, "loss": 1.9755, "step": 3615 }, { "epoch": 0.8732190292199952, "grad_norm": 0.2960030436515808, "learning_rate": 4.0733099065836275e-06, "loss": 1.7833, "step": 3616 }, { "epoch": 0.8734605167833857, "grad_norm": 0.28000757098197937, "learning_rate": 4.05802050513523e-06, "loss": 1.6431, "step": 3617 }, { "epoch": 0.8737020043467761, "grad_norm": 0.29284968972206116, "learning_rate": 4.042758638703364e-06, "loss": 1.6229, "step": 3618 }, { "epoch": 0.8739434919101666, "grad_norm": 0.2796996533870697, "learning_rate": 4.027524316435166e-06, "loss": 1.5696, "step": 3619 }, { "epoch": 0.8741849794735571, "grad_norm": 0.29490870237350464, "learning_rate": 4.0123175474612085e-06, "loss": 1.4853, "step": 3620 }, { "epoch": 0.8744264670369476, "grad_norm": 0.3215023875236511, "learning_rate": 3.997138340895557e-06, "loss": 1.859, "step": 3621 }, { "epoch": 0.8746679546003381, "grad_norm": 0.2848607301712036, "learning_rate": 3.981986705835783e-06, "loss": 1.7621, "step": 3622 }, { "epoch": 0.8749094421637286, "grad_norm": 0.2749423682689667, "learning_rate": 3.966862651362918e-06, "loss": 1.6605, "step": 3623 }, { "epoch": 0.875150929727119, "grad_norm": 0.2763156592845917, "learning_rate": 3.9517661865414505e-06, "loss": 1.6116, "step": 3624 }, { "epoch": 0.8753924172905095, "grad_norm": 0.31394004821777344, "learning_rate": 3.936697320419341e-06, "loss": 1.7444, "step": 3625 }, { "epoch": 0.8756339048539, "grad_norm": 0.3056226968765259, "learning_rate": 3.921656062028045e-06, "loss": 1.798, "step": 3626 }, { "epoch": 0.8758753924172905, "grad_norm": 0.3123130798339844, "learning_rate": 3.906642420382395e-06, "loss": 1.6911, "step": 3627 }, { "epoch": 0.876116879980681, "grad_norm": 0.29763081669807434, "learning_rate": 3.891656404480748e-06, "loss": 1.5814, "step": 3628 }, { "epoch": 0.8763583675440715, "grad_norm": 0.28500983119010925, "learning_rate": 3.876698023304865e-06, "loss": 1.7173, "step": 3629 }, { "epoch": 0.876599855107462, "grad_norm": 0.29087314009666443, "learning_rate": 3.861767285819945e-06, "loss": 1.7377, "step": 3630 }, { "epoch": 0.8768413426708525, "grad_norm": 0.5176892280578613, "learning_rate": 3.84686420097462e-06, "loss": 1.7903, "step": 3631 }, { "epoch": 0.8770828302342429, "grad_norm": 0.2725013792514801, "learning_rate": 3.831988777700973e-06, "loss": 1.6015, "step": 3632 }, { "epoch": 0.8773243177976334, "grad_norm": 0.2676476240158081, "learning_rate": 3.817141024914478e-06, "loss": 1.6007, "step": 3633 }, { "epoch": 0.8775658053610239, "grad_norm": 0.28349554538726807, "learning_rate": 3.802320951514027e-06, "loss": 1.7179, "step": 3634 }, { "epoch": 0.8778072929244144, "grad_norm": 0.3007166087627411, "learning_rate": 3.7875285663819536e-06, "loss": 1.6629, "step": 3635 }, { "epoch": 0.8780487804878049, "grad_norm": 0.2894110083580017, "learning_rate": 3.772763878383945e-06, "loss": 1.6623, "step": 3636 }, { "epoch": 0.8782902680511954, "grad_norm": 0.2908770442008972, "learning_rate": 3.758026896369149e-06, "loss": 1.74, "step": 3637 }, { "epoch": 0.8785317556145859, "grad_norm": 0.27949637174606323, "learning_rate": 3.743317629170057e-06, "loss": 1.6076, "step": 3638 }, { "epoch": 0.8787732431779763, "grad_norm": 0.28380855917930603, "learning_rate": 3.7286360856025713e-06, "loss": 1.7442, "step": 3639 }, { "epoch": 0.8790147307413668, "grad_norm": 0.2948709726333618, "learning_rate": 3.7139822744659713e-06, "loss": 1.7345, "step": 3640 }, { "epoch": 0.8792562183047573, "grad_norm": 0.2999047040939331, "learning_rate": 3.6993562045429363e-06, "loss": 1.7659, "step": 3641 }, { "epoch": 0.8794977058681478, "grad_norm": 0.31496214866638184, "learning_rate": 3.684757884599499e-06, "loss": 1.6543, "step": 3642 }, { "epoch": 0.8797391934315383, "grad_norm": 0.2878762483596802, "learning_rate": 3.6701873233850547e-06, "loss": 1.6697, "step": 3643 }, { "epoch": 0.8799806809949288, "grad_norm": 0.2935860753059387, "learning_rate": 3.6556445296323794e-06, "loss": 1.6649, "step": 3644 }, { "epoch": 0.8802221685583193, "grad_norm": 0.2822621762752533, "learning_rate": 3.641129512057584e-06, "loss": 1.7232, "step": 3645 }, { "epoch": 0.8804636561217097, "grad_norm": 0.2823951244354248, "learning_rate": 3.6266422793601706e-06, "loss": 1.661, "step": 3646 }, { "epoch": 0.8807051436851002, "grad_norm": 0.29623061418533325, "learning_rate": 3.6121828402229496e-06, "loss": 1.7347, "step": 3647 }, { "epoch": 0.8809466312484907, "grad_norm": 0.2824953496456146, "learning_rate": 3.597751203312094e-06, "loss": 1.7623, "step": 3648 }, { "epoch": 0.8811881188118812, "grad_norm": 0.2898530960083008, "learning_rate": 3.5833473772770965e-06, "loss": 1.7143, "step": 3649 }, { "epoch": 0.8814296063752717, "grad_norm": 0.29235631227493286, "learning_rate": 3.5689713707508233e-06, "loss": 1.6894, "step": 3650 }, { "epoch": 0.8816710939386622, "grad_norm": 0.28454381227493286, "learning_rate": 3.554623192349393e-06, "loss": 1.7559, "step": 3651 }, { "epoch": 0.8819125815020527, "grad_norm": 0.29639336466789246, "learning_rate": 3.540302850672328e-06, "loss": 1.6573, "step": 3652 }, { "epoch": 0.8821540690654431, "grad_norm": 0.2831573784351349, "learning_rate": 3.5260103543024114e-06, "loss": 1.6139, "step": 3653 }, { "epoch": 0.8823955566288336, "grad_norm": 0.29325729608535767, "learning_rate": 3.511745711805753e-06, "loss": 1.7383, "step": 3654 }, { "epoch": 0.8826370441922241, "grad_norm": 0.2697634696960449, "learning_rate": 3.49750893173178e-06, "loss": 1.5719, "step": 3655 }, { "epoch": 0.8828785317556146, "grad_norm": 0.2780228555202484, "learning_rate": 3.483300022613206e-06, "loss": 1.7957, "step": 3656 }, { "epoch": 0.8831200193190051, "grad_norm": 0.2855720520019531, "learning_rate": 3.4691189929660472e-06, "loss": 1.7727, "step": 3657 }, { "epoch": 0.8833615068823956, "grad_norm": 0.2958657741546631, "learning_rate": 3.4549658512895987e-06, "loss": 1.724, "step": 3658 }, { "epoch": 0.883602994445786, "grad_norm": 0.2907499670982361, "learning_rate": 3.4408406060664754e-06, "loss": 1.6969, "step": 3659 }, { "epoch": 0.8838444820091765, "grad_norm": 0.28889361023902893, "learning_rate": 3.426743265762511e-06, "loss": 1.763, "step": 3660 }, { "epoch": 0.884085969572567, "grad_norm": 0.28712794184684753, "learning_rate": 3.412673838826885e-06, "loss": 1.6704, "step": 3661 }, { "epoch": 0.8843274571359575, "grad_norm": 0.28045400977134705, "learning_rate": 3.3986323336919978e-06, "loss": 1.6242, "step": 3662 }, { "epoch": 0.884568944699348, "grad_norm": 0.28906765580177307, "learning_rate": 3.3846187587735336e-06, "loss": 1.6951, "step": 3663 }, { "epoch": 0.8848104322627385, "grad_norm": 0.2904246151447296, "learning_rate": 3.370633122470429e-06, "loss": 1.6312, "step": 3664 }, { "epoch": 0.885051919826129, "grad_norm": 0.2904842793941498, "learning_rate": 3.3566754331648963e-06, "loss": 1.8774, "step": 3665 }, { "epoch": 0.8852934073895194, "grad_norm": 0.28442999720573425, "learning_rate": 3.3427456992223716e-06, "loss": 1.7342, "step": 3666 }, { "epoch": 0.8855348949529099, "grad_norm": 0.28631576895713806, "learning_rate": 3.3288439289915375e-06, "loss": 1.6828, "step": 3667 }, { "epoch": 0.8857763825163004, "grad_norm": 0.2903057038784027, "learning_rate": 3.3149701308043568e-06, "loss": 1.7221, "step": 3668 }, { "epoch": 0.8860178700796909, "grad_norm": 0.2785232961177826, "learning_rate": 3.301124312975967e-06, "loss": 1.5806, "step": 3669 }, { "epoch": 0.8862593576430814, "grad_norm": 0.2787090539932251, "learning_rate": 3.2873064838047806e-06, "loss": 1.6033, "step": 3670 }, { "epoch": 0.8865008452064719, "grad_norm": 0.2764091491699219, "learning_rate": 3.273516651572417e-06, "loss": 1.5415, "step": 3671 }, { "epoch": 0.8867423327698624, "grad_norm": 0.3111838400363922, "learning_rate": 3.2597548245437203e-06, "loss": 1.8512, "step": 3672 }, { "epoch": 0.8869838203332528, "grad_norm": 0.283963143825531, "learning_rate": 3.2460210109667434e-06, "loss": 1.6712, "step": 3673 }, { "epoch": 0.8872253078966433, "grad_norm": 0.2889876663684845, "learning_rate": 3.2323152190727625e-06, "loss": 1.6137, "step": 3674 }, { "epoch": 0.8874667954600338, "grad_norm": 0.281650185585022, "learning_rate": 3.218637457076257e-06, "loss": 1.654, "step": 3675 }, { "epoch": 0.8877082830234243, "grad_norm": 0.27170509099960327, "learning_rate": 3.2049877331748912e-06, "loss": 1.5488, "step": 3676 }, { "epoch": 0.8879497705868148, "grad_norm": 0.3009040057659149, "learning_rate": 3.191366055549533e-06, "loss": 1.6728, "step": 3677 }, { "epoch": 0.8881912581502053, "grad_norm": 0.2903916835784912, "learning_rate": 3.1777724323642456e-06, "loss": 1.6838, "step": 3678 }, { "epoch": 0.8884327457135958, "grad_norm": 0.2689923346042633, "learning_rate": 3.1642068717662897e-06, "loss": 1.5638, "step": 3679 }, { "epoch": 0.8886742332769862, "grad_norm": 0.29266276955604553, "learning_rate": 3.150669381886079e-06, "loss": 1.8412, "step": 3680 }, { "epoch": 0.8889157208403767, "grad_norm": 0.3078247904777527, "learning_rate": 3.1371599708372167e-06, "loss": 1.8111, "step": 3681 }, { "epoch": 0.8891572084037672, "grad_norm": 0.2779790163040161, "learning_rate": 3.123678646716477e-06, "loss": 1.7199, "step": 3682 }, { "epoch": 0.8893986959671577, "grad_norm": 0.29290392994880676, "learning_rate": 3.110225417603818e-06, "loss": 1.7965, "step": 3683 }, { "epoch": 0.8896401835305482, "grad_norm": 0.2934718430042267, "learning_rate": 3.0968002915623185e-06, "loss": 1.7689, "step": 3684 }, { "epoch": 0.8898816710939387, "grad_norm": 0.2683970034122467, "learning_rate": 3.0834032766382525e-06, "loss": 1.5235, "step": 3685 }, { "epoch": 0.8901231586573292, "grad_norm": 0.2945578396320343, "learning_rate": 3.07003438086102e-06, "loss": 1.771, "step": 3686 }, { "epoch": 0.8903646462207196, "grad_norm": 0.2975611090660095, "learning_rate": 3.056693612243183e-06, "loss": 1.7231, "step": 3687 }, { "epoch": 0.8906061337841101, "grad_norm": 0.30733227729797363, "learning_rate": 3.0433809787804525e-06, "loss": 1.711, "step": 3688 }, { "epoch": 0.8908476213475006, "grad_norm": 0.2985078692436218, "learning_rate": 3.0300964884516557e-06, "loss": 1.7603, "step": 3689 }, { "epoch": 0.8910891089108911, "grad_norm": 0.29112014174461365, "learning_rate": 3.0168401492187635e-06, "loss": 1.7592, "step": 3690 }, { "epoch": 0.8913305964742816, "grad_norm": 0.29112181067466736, "learning_rate": 3.003611969026865e-06, "loss": 1.7306, "step": 3691 }, { "epoch": 0.8915720840376721, "grad_norm": 0.2815907895565033, "learning_rate": 2.9904119558042066e-06, "loss": 1.7418, "step": 3692 }, { "epoch": 0.8918135716010626, "grad_norm": 0.2853732407093048, "learning_rate": 2.9772401174620935e-06, "loss": 1.8023, "step": 3693 }, { "epoch": 0.892055059164453, "grad_norm": 0.28228485584259033, "learning_rate": 2.964096461895005e-06, "loss": 1.6704, "step": 3694 }, { "epoch": 0.8922965467278435, "grad_norm": 0.31012919545173645, "learning_rate": 2.950980996980496e-06, "loss": 1.9585, "step": 3695 }, { "epoch": 0.892538034291234, "grad_norm": 0.2896401584148407, "learning_rate": 2.937893730579222e-06, "loss": 1.7102, "step": 3696 }, { "epoch": 0.8927795218546245, "grad_norm": 0.2976386845111847, "learning_rate": 2.924834670534954e-06, "loss": 1.6954, "step": 3697 }, { "epoch": 0.893021009418015, "grad_norm": 0.29057908058166504, "learning_rate": 2.9118038246745636e-06, "loss": 1.6988, "step": 3698 }, { "epoch": 0.8932624969814055, "grad_norm": 0.277283638715744, "learning_rate": 2.898801200807988e-06, "loss": 1.5941, "step": 3699 }, { "epoch": 0.893503984544796, "grad_norm": 0.3489179313182831, "learning_rate": 2.8858268067282535e-06, "loss": 1.6184, "step": 3700 }, { "epoch": 0.8937454721081864, "grad_norm": 0.2937741279602051, "learning_rate": 2.872880650211501e-06, "loss": 1.62, "step": 3701 }, { "epoch": 0.8939869596715769, "grad_norm": 0.28105103969573975, "learning_rate": 2.8599627390169082e-06, "loss": 1.7256, "step": 3702 }, { "epoch": 0.8942284472349674, "grad_norm": 0.2642846405506134, "learning_rate": 2.8470730808867363e-06, "loss": 1.3655, "step": 3703 }, { "epoch": 0.8944699347983579, "grad_norm": 0.30971407890319824, "learning_rate": 2.8342116835463117e-06, "loss": 1.8151, "step": 3704 }, { "epoch": 0.8947114223617484, "grad_norm": 0.277342289686203, "learning_rate": 2.821378554704052e-06, "loss": 1.5223, "step": 3705 }, { "epoch": 0.8949529099251389, "grad_norm": 0.28479868173599243, "learning_rate": 2.8085737020513647e-06, "loss": 1.5905, "step": 3706 }, { "epoch": 0.8951943974885294, "grad_norm": 0.30333152413368225, "learning_rate": 2.7957971332627895e-06, "loss": 1.9046, "step": 3707 }, { "epoch": 0.8954358850519198, "grad_norm": 0.27579569816589355, "learning_rate": 2.783048855995857e-06, "loss": 1.6768, "step": 3708 }, { "epoch": 0.8956773726153103, "grad_norm": 0.28334715962409973, "learning_rate": 2.7703288778911706e-06, "loss": 1.6294, "step": 3709 }, { "epoch": 0.8959188601787008, "grad_norm": 0.30097290873527527, "learning_rate": 2.7576372065723556e-06, "loss": 1.7061, "step": 3710 }, { "epoch": 0.8961603477420913, "grad_norm": 0.2801850438117981, "learning_rate": 2.7449738496460874e-06, "loss": 1.6197, "step": 3711 }, { "epoch": 0.8964018353054818, "grad_norm": 0.2738920748233795, "learning_rate": 2.732338814702057e-06, "loss": 1.5063, "step": 3712 }, { "epoch": 0.8966433228688723, "grad_norm": 0.2827121317386627, "learning_rate": 2.719732109312989e-06, "loss": 1.6853, "step": 3713 }, { "epoch": 0.8968848104322628, "grad_norm": 0.3043137490749359, "learning_rate": 2.707153741034646e-06, "loss": 1.8222, "step": 3714 }, { "epoch": 0.8971262979956532, "grad_norm": 0.2979698181152344, "learning_rate": 2.6946037174057527e-06, "loss": 1.7522, "step": 3715 }, { "epoch": 0.8973677855590437, "grad_norm": 0.28279903531074524, "learning_rate": 2.6820820459481146e-06, "loss": 1.4597, "step": 3716 }, { "epoch": 0.8976092731224342, "grad_norm": 0.2726062834262848, "learning_rate": 2.6695887341664947e-06, "loss": 1.5925, "step": 3717 }, { "epoch": 0.8978507606858247, "grad_norm": 0.28320199251174927, "learning_rate": 2.6571237895486768e-06, "loss": 1.7275, "step": 3718 }, { "epoch": 0.8980922482492152, "grad_norm": 0.2890796363353729, "learning_rate": 2.644687219565439e-06, "loss": 1.8677, "step": 3719 }, { "epoch": 0.8983337358126057, "grad_norm": 0.2925397455692291, "learning_rate": 2.6322790316705647e-06, "loss": 1.7388, "step": 3720 }, { "epoch": 0.8985752233759962, "grad_norm": 0.2931881546974182, "learning_rate": 2.6198992333008155e-06, "loss": 1.7072, "step": 3721 }, { "epoch": 0.8988167109393866, "grad_norm": 0.3081625699996948, "learning_rate": 2.607547831875928e-06, "loss": 1.9215, "step": 3722 }, { "epoch": 0.8990581985027771, "grad_norm": 0.27621951699256897, "learning_rate": 2.5952248347986596e-06, "loss": 1.5997, "step": 3723 }, { "epoch": 0.8992996860661676, "grad_norm": 0.27733898162841797, "learning_rate": 2.582930249454685e-06, "loss": 1.6312, "step": 3724 }, { "epoch": 0.8995411736295581, "grad_norm": 0.3431466221809387, "learning_rate": 2.570664083212698e-06, "loss": 1.6439, "step": 3725 }, { "epoch": 0.8997826611929486, "grad_norm": 0.2937251329421997, "learning_rate": 2.558426343424342e-06, "loss": 1.8034, "step": 3726 }, { "epoch": 0.9000241487563391, "grad_norm": 0.26996278762817383, "learning_rate": 2.54621703742422e-06, "loss": 1.6587, "step": 3727 }, { "epoch": 0.9002656363197296, "grad_norm": 0.30015066266059875, "learning_rate": 2.5340361725298912e-06, "loss": 1.7535, "step": 3728 }, { "epoch": 0.90050712388312, "grad_norm": 0.29198890924453735, "learning_rate": 2.5218837560418984e-06, "loss": 1.6503, "step": 3729 }, { "epoch": 0.9007486114465105, "grad_norm": 0.29858389496803284, "learning_rate": 2.509759795243677e-06, "loss": 1.7874, "step": 3730 }, { "epoch": 0.900990099009901, "grad_norm": 0.28021368384361267, "learning_rate": 2.4976642974016696e-06, "loss": 1.7122, "step": 3731 }, { "epoch": 0.9012315865732915, "grad_norm": 0.28410908579826355, "learning_rate": 2.485597269765222e-06, "loss": 1.7054, "step": 3732 }, { "epoch": 0.901473074136682, "grad_norm": 0.277952641248703, "learning_rate": 2.4735587195666153e-06, "loss": 1.6176, "step": 3733 }, { "epoch": 0.9017145617000725, "grad_norm": 0.2899259030818939, "learning_rate": 2.461548654021095e-06, "loss": 1.6055, "step": 3734 }, { "epoch": 0.901956049263463, "grad_norm": 0.28117528557777405, "learning_rate": 2.449567080326798e-06, "loss": 1.7165, "step": 3735 }, { "epoch": 0.9021975368268534, "grad_norm": 0.3105177581310272, "learning_rate": 2.4376140056648066e-06, "loss": 1.8328, "step": 3736 }, { "epoch": 0.9024390243902439, "grad_norm": 0.28931888937950134, "learning_rate": 2.425689437199108e-06, "loss": 1.6223, "step": 3737 }, { "epoch": 0.9026805119536344, "grad_norm": 0.2876199781894684, "learning_rate": 2.4137933820766235e-06, "loss": 1.7588, "step": 3738 }, { "epoch": 0.9029219995170249, "grad_norm": 0.27212607860565186, "learning_rate": 2.401925847427161e-06, "loss": 1.6519, "step": 3739 }, { "epoch": 0.9031634870804154, "grad_norm": 0.29444506764411926, "learning_rate": 2.390086840363459e-06, "loss": 1.8825, "step": 3740 }, { "epoch": 0.9034049746438059, "grad_norm": 0.27534398436546326, "learning_rate": 2.378276367981136e-06, "loss": 1.6509, "step": 3741 }, { "epoch": 0.9036464622071964, "grad_norm": 0.2755245268344879, "learning_rate": 2.3664944373587296e-06, "loss": 1.6497, "step": 3742 }, { "epoch": 0.9038879497705868, "grad_norm": 0.28756704926490784, "learning_rate": 2.3547410555576522e-06, "loss": 1.7498, "step": 3743 }, { "epoch": 0.9041294373339773, "grad_norm": 0.2990304231643677, "learning_rate": 2.343016229622219e-06, "loss": 1.7188, "step": 3744 }, { "epoch": 0.9043709248973678, "grad_norm": 0.2797776758670807, "learning_rate": 2.3313199665796317e-06, "loss": 1.676, "step": 3745 }, { "epoch": 0.9046124124607583, "grad_norm": 0.2969546914100647, "learning_rate": 2.3196522734399485e-06, "loss": 1.7262, "step": 3746 }, { "epoch": 0.9048539000241488, "grad_norm": 0.3097802400588989, "learning_rate": 2.308013157196148e-06, "loss": 1.9066, "step": 3747 }, { "epoch": 0.9050953875875393, "grad_norm": 0.2961737811565399, "learning_rate": 2.2964026248240287e-06, "loss": 1.6393, "step": 3748 }, { "epoch": 0.9053368751509298, "grad_norm": 0.2888379395008087, "learning_rate": 2.284820683282307e-06, "loss": 1.6906, "step": 3749 }, { "epoch": 0.9055783627143202, "grad_norm": 0.2969198226928711, "learning_rate": 2.2732673395125304e-06, "loss": 1.7629, "step": 3750 }, { "epoch": 0.9058198502777107, "grad_norm": 0.2836252450942993, "learning_rate": 2.2617426004391263e-06, "loss": 1.7987, "step": 3751 }, { "epoch": 0.9060613378411012, "grad_norm": 0.29032793641090393, "learning_rate": 2.250246472969353e-06, "loss": 1.7476, "step": 3752 }, { "epoch": 0.9063028254044917, "grad_norm": 0.2925272583961487, "learning_rate": 2.238778963993354e-06, "loss": 1.7374, "step": 3753 }, { "epoch": 0.9065443129678822, "grad_norm": 0.266701877117157, "learning_rate": 2.227340080384105e-06, "loss": 1.6693, "step": 3754 }, { "epoch": 0.9067858005312727, "grad_norm": 0.28594446182250977, "learning_rate": 2.2159298289974038e-06, "loss": 1.7237, "step": 3755 }, { "epoch": 0.9070272880946632, "grad_norm": 0.2878924608230591, "learning_rate": 2.204548216671942e-06, "loss": 1.6927, "step": 3756 }, { "epoch": 0.9072687756580536, "grad_norm": 0.2949213683605194, "learning_rate": 2.193195250229174e-06, "loss": 1.6293, "step": 3757 }, { "epoch": 0.9075102632214441, "grad_norm": 0.29727548360824585, "learning_rate": 2.181870936473446e-06, "loss": 1.8721, "step": 3758 }, { "epoch": 0.9077517507848346, "grad_norm": 0.2732381820678711, "learning_rate": 2.170575282191906e-06, "loss": 1.5579, "step": 3759 }, { "epoch": 0.9079932383482251, "grad_norm": 0.278227299451828, "learning_rate": 2.1593082941545327e-06, "loss": 1.6583, "step": 3760 }, { "epoch": 0.9082347259116156, "grad_norm": 0.27635660767555237, "learning_rate": 2.1480699791141037e-06, "loss": 1.6148, "step": 3761 }, { "epoch": 0.9084762134750061, "grad_norm": 0.2797464430332184, "learning_rate": 2.1368603438062518e-06, "loss": 1.6219, "step": 3762 }, { "epoch": 0.9087177010383966, "grad_norm": 0.2997564375400543, "learning_rate": 2.1256793949493704e-06, "loss": 1.8035, "step": 3763 }, { "epoch": 0.908959188601787, "grad_norm": 0.2835381329059601, "learning_rate": 2.114527139244704e-06, "loss": 1.6885, "step": 3764 }, { "epoch": 0.9092006761651775, "grad_norm": 0.2909567654132843, "learning_rate": 2.1034035833762787e-06, "loss": 1.6199, "step": 3765 }, { "epoch": 0.909442163728568, "grad_norm": 0.29690560698509216, "learning_rate": 2.0923087340109115e-06, "loss": 1.7686, "step": 3766 }, { "epoch": 0.9096836512919585, "grad_norm": 0.28672075271606445, "learning_rate": 2.081242597798244e-06, "loss": 1.7726, "step": 3767 }, { "epoch": 0.909925138855349, "grad_norm": 0.2886490523815155, "learning_rate": 2.0702051813706825e-06, "loss": 1.6471, "step": 3768 }, { "epoch": 0.9101666264187395, "grad_norm": 0.27859172224998474, "learning_rate": 2.0591964913434238e-06, "loss": 1.7475, "step": 3769 }, { "epoch": 0.91040811398213, "grad_norm": 0.2758730351924896, "learning_rate": 2.048216534314451e-06, "loss": 1.5618, "step": 3770 }, { "epoch": 0.9106496015455204, "grad_norm": 0.2859569489955902, "learning_rate": 2.037265316864545e-06, "loss": 1.7833, "step": 3771 }, { "epoch": 0.9108910891089109, "grad_norm": 0.30485159158706665, "learning_rate": 2.026342845557222e-06, "loss": 1.8197, "step": 3772 }, { "epoch": 0.9111325766723014, "grad_norm": 0.2852356433868408, "learning_rate": 2.0154491269388063e-06, "loss": 1.7482, "step": 3773 }, { "epoch": 0.9113740642356919, "grad_norm": 0.28762078285217285, "learning_rate": 2.0045841675383704e-06, "loss": 1.5435, "step": 3774 }, { "epoch": 0.9116155517990824, "grad_norm": 0.286262184381485, "learning_rate": 1.9937479738677488e-06, "loss": 1.7981, "step": 3775 }, { "epoch": 0.9118570393624729, "grad_norm": 0.3011592924594879, "learning_rate": 1.982940552421553e-06, "loss": 1.7707, "step": 3776 }, { "epoch": 0.9120985269258634, "grad_norm": 0.29348087310791016, "learning_rate": 1.972161909677134e-06, "loss": 1.7805, "step": 3777 }, { "epoch": 0.9123400144892538, "grad_norm": 0.2972409129142761, "learning_rate": 1.9614120520945967e-06, "loss": 1.817, "step": 3778 }, { "epoch": 0.9125815020526443, "grad_norm": 0.29411864280700684, "learning_rate": 1.950690986116793e-06, "loss": 1.7039, "step": 3779 }, { "epoch": 0.9128229896160348, "grad_norm": 0.29106858372688293, "learning_rate": 1.9399987181693336e-06, "loss": 1.7798, "step": 3780 }, { "epoch": 0.9130644771794253, "grad_norm": 0.28918927907943726, "learning_rate": 1.929335254660536e-06, "loss": 1.7439, "step": 3781 }, { "epoch": 0.9133059647428158, "grad_norm": 0.2888122797012329, "learning_rate": 1.918700601981499e-06, "loss": 1.6759, "step": 3782 }, { "epoch": 0.9135474523062063, "grad_norm": 0.31226205825805664, "learning_rate": 1.908094766506019e-06, "loss": 2.0557, "step": 3783 }, { "epoch": 0.9137889398695968, "grad_norm": 0.29069530963897705, "learning_rate": 1.8975177545906275e-06, "loss": 1.8395, "step": 3784 }, { "epoch": 0.9140304274329872, "grad_norm": 0.29319819808006287, "learning_rate": 1.8869695725745872e-06, "loss": 1.6923, "step": 3785 }, { "epoch": 0.9142719149963777, "grad_norm": 0.2889004945755005, "learning_rate": 1.8764502267798967e-06, "loss": 1.686, "step": 3786 }, { "epoch": 0.9145134025597682, "grad_norm": 0.2927665412425995, "learning_rate": 1.8659597235112347e-06, "loss": 1.6664, "step": 3787 }, { "epoch": 0.9147548901231587, "grad_norm": 0.28432509303092957, "learning_rate": 1.8554980690560163e-06, "loss": 1.7847, "step": 3788 }, { "epoch": 0.9149963776865492, "grad_norm": 0.28083717823028564, "learning_rate": 1.8450652696843817e-06, "loss": 1.5961, "step": 3789 }, { "epoch": 0.9152378652499397, "grad_norm": 0.2860347330570221, "learning_rate": 1.8346613316491345e-06, "loss": 1.7269, "step": 3790 }, { "epoch": 0.9154793528133301, "grad_norm": 0.29229259490966797, "learning_rate": 1.8242862611858258e-06, "loss": 1.76, "step": 3791 }, { "epoch": 0.9157208403767206, "grad_norm": 0.2943476140499115, "learning_rate": 1.8139400645126703e-06, "loss": 1.756, "step": 3792 }, { "epoch": 0.9159623279401111, "grad_norm": 0.2993648648262024, "learning_rate": 1.8036227478306022e-06, "loss": 1.7881, "step": 3793 }, { "epoch": 0.9162038155035016, "grad_norm": 0.28937676548957825, "learning_rate": 1.7933343173232199e-06, "loss": 1.779, "step": 3794 }, { "epoch": 0.9164453030668921, "grad_norm": 0.28702759742736816, "learning_rate": 1.7830747791568457e-06, "loss": 1.6789, "step": 3795 }, { "epoch": 0.9166867906302826, "grad_norm": 0.27720916271209717, "learning_rate": 1.772844139480445e-06, "loss": 1.5913, "step": 3796 }, { "epoch": 0.9169282781936731, "grad_norm": 0.305535227060318, "learning_rate": 1.7626424044256963e-06, "loss": 1.7974, "step": 3797 }, { "epoch": 0.9171697657570635, "grad_norm": 0.29275986552238464, "learning_rate": 1.7524695801069412e-06, "loss": 1.7324, "step": 3798 }, { "epoch": 0.917411253320454, "grad_norm": 0.2831110954284668, "learning_rate": 1.7423256726211811e-06, "loss": 1.7081, "step": 3799 }, { "epoch": 0.9176527408838445, "grad_norm": 0.28903546929359436, "learning_rate": 1.7322106880481137e-06, "loss": 1.7348, "step": 3800 }, { "epoch": 0.917894228447235, "grad_norm": 0.3020871579647064, "learning_rate": 1.7221246324500784e-06, "loss": 1.813, "step": 3801 }, { "epoch": 0.9181357160106255, "grad_norm": 0.2730100452899933, "learning_rate": 1.7120675118720842e-06, "loss": 1.6626, "step": 3802 }, { "epoch": 0.918377203574016, "grad_norm": 0.28377363085746765, "learning_rate": 1.7020393323417983e-06, "loss": 1.7281, "step": 3803 }, { "epoch": 0.9186186911374065, "grad_norm": 0.2961227595806122, "learning_rate": 1.6920400998695574e-06, "loss": 1.7109, "step": 3804 }, { "epoch": 0.918860178700797, "grad_norm": 0.294062077999115, "learning_rate": 1.682069820448312e-06, "loss": 1.7016, "step": 3805 }, { "epoch": 0.9191016662641874, "grad_norm": 0.28122201561927795, "learning_rate": 1.672128500053699e-06, "loss": 1.6506, "step": 3806 }, { "epoch": 0.9193431538275779, "grad_norm": 0.2936934530735016, "learning_rate": 1.6622161446439744e-06, "loss": 1.5704, "step": 3807 }, { "epoch": 0.9195846413909684, "grad_norm": 0.2706112563610077, "learning_rate": 1.6523327601600357e-06, "loss": 1.645, "step": 3808 }, { "epoch": 0.9198261289543589, "grad_norm": 0.2646171748638153, "learning_rate": 1.642478352525434e-06, "loss": 1.6279, "step": 3809 }, { "epoch": 0.9200676165177494, "grad_norm": 0.287831574678421, "learning_rate": 1.6326529276463387e-06, "loss": 1.6784, "step": 3810 }, { "epoch": 0.9203091040811399, "grad_norm": 0.2851034998893738, "learning_rate": 1.6228564914115564e-06, "loss": 1.6375, "step": 3811 }, { "epoch": 0.9205505916445303, "grad_norm": 0.28739801049232483, "learning_rate": 1.6130890496925011e-06, "loss": 1.6467, "step": 3812 }, { "epoch": 0.9207920792079208, "grad_norm": 0.29712507128715515, "learning_rate": 1.6033506083432459e-06, "loss": 1.7607, "step": 3813 }, { "epoch": 0.9210335667713113, "grad_norm": 0.2879638671875, "learning_rate": 1.593641173200433e-06, "loss": 1.734, "step": 3814 }, { "epoch": 0.9212750543347018, "grad_norm": 0.290679931640625, "learning_rate": 1.5839607500833687e-06, "loss": 1.5929, "step": 3815 }, { "epoch": 0.9215165418980923, "grad_norm": 0.31854555010795593, "learning_rate": 1.574309344793945e-06, "loss": 1.8794, "step": 3816 }, { "epoch": 0.9217580294614828, "grad_norm": 0.2682536840438843, "learning_rate": 1.5646869631166573e-06, "loss": 1.598, "step": 3817 }, { "epoch": 0.9219995170248733, "grad_norm": 0.26716476678848267, "learning_rate": 1.55509361081862e-06, "loss": 1.5477, "step": 3818 }, { "epoch": 0.9222410045882637, "grad_norm": 0.2830582559108734, "learning_rate": 1.5455292936495558e-06, "loss": 1.6642, "step": 3819 }, { "epoch": 0.9224824921516542, "grad_norm": 0.29494237899780273, "learning_rate": 1.5359940173417575e-06, "loss": 1.6055, "step": 3820 }, { "epoch": 0.9227239797150447, "grad_norm": 0.2776681184768677, "learning_rate": 1.5264877876101314e-06, "loss": 1.6369, "step": 3821 }, { "epoch": 0.9229654672784352, "grad_norm": 0.29700928926467896, "learning_rate": 1.517010610152192e-06, "loss": 1.7536, "step": 3822 }, { "epoch": 0.9232069548418257, "grad_norm": 0.27303093671798706, "learning_rate": 1.5075624906479902e-06, "loss": 1.5837, "step": 3823 }, { "epoch": 0.9234484424052162, "grad_norm": 0.2823008894920349, "learning_rate": 1.4981434347602185e-06, "loss": 1.736, "step": 3824 }, { "epoch": 0.9236899299686067, "grad_norm": 0.27433207631111145, "learning_rate": 1.4887534481341058e-06, "loss": 1.577, "step": 3825 }, { "epoch": 0.9239314175319971, "grad_norm": 0.27874621748924255, "learning_rate": 1.4793925363975058e-06, "loss": 1.7509, "step": 3826 }, { "epoch": 0.9241729050953876, "grad_norm": 0.2738780081272125, "learning_rate": 1.4700607051607806e-06, "loss": 1.5793, "step": 3827 }, { "epoch": 0.9244143926587781, "grad_norm": 0.2975897192955017, "learning_rate": 1.4607579600169286e-06, "loss": 1.7944, "step": 3828 }, { "epoch": 0.9246558802221686, "grad_norm": 0.29397067427635193, "learning_rate": 1.4514843065414785e-06, "loss": 1.6425, "step": 3829 }, { "epoch": 0.9248973677855591, "grad_norm": 0.28670430183410645, "learning_rate": 1.4422397502925289e-06, "loss": 1.7348, "step": 3830 }, { "epoch": 0.9251388553489496, "grad_norm": 0.2757027745246887, "learning_rate": 1.4330242968107422e-06, "loss": 1.5882, "step": 3831 }, { "epoch": 0.9253803429123401, "grad_norm": 0.35739678144454956, "learning_rate": 1.4238379516193445e-06, "loss": 1.8124, "step": 3832 }, { "epoch": 0.9256218304757305, "grad_norm": 0.27541524171829224, "learning_rate": 1.4146807202241096e-06, "loss": 1.5513, "step": 3833 }, { "epoch": 0.925863318039121, "grad_norm": 0.2763047516345978, "learning_rate": 1.4055526081133529e-06, "loss": 1.552, "step": 3834 }, { "epoch": 0.9261048056025115, "grad_norm": 0.2798953354358673, "learning_rate": 1.3964536207579704e-06, "loss": 1.6226, "step": 3835 }, { "epoch": 0.926346293165902, "grad_norm": 0.276141881942749, "learning_rate": 1.387383763611344e-06, "loss": 1.711, "step": 3836 }, { "epoch": 0.9265877807292925, "grad_norm": 0.2811782658100128, "learning_rate": 1.3783430421094645e-06, "loss": 1.5793, "step": 3837 }, { "epoch": 0.926829268292683, "grad_norm": 0.28732654452323914, "learning_rate": 1.3693314616708086e-06, "loss": 1.8873, "step": 3838 }, { "epoch": 0.9270707558560735, "grad_norm": 0.29729849100112915, "learning_rate": 1.3603490276964114e-06, "loss": 1.6249, "step": 3839 }, { "epoch": 0.9273122434194639, "grad_norm": 0.2793877124786377, "learning_rate": 1.3513957455698279e-06, "loss": 1.6443, "step": 3840 }, { "epoch": 0.9275537309828544, "grad_norm": 0.29684704542160034, "learning_rate": 1.3424716206571652e-06, "loss": 1.7706, "step": 3841 }, { "epoch": 0.9277952185462449, "grad_norm": 0.2845028042793274, "learning_rate": 1.3335766583070175e-06, "loss": 1.5303, "step": 3842 }, { "epoch": 0.9280367061096354, "grad_norm": 0.2891022861003876, "learning_rate": 1.3247108638505368e-06, "loss": 1.752, "step": 3843 }, { "epoch": 0.9282781936730259, "grad_norm": 0.28825807571411133, "learning_rate": 1.315874242601367e-06, "loss": 1.8039, "step": 3844 }, { "epoch": 0.9285196812364164, "grad_norm": 0.2929156720638275, "learning_rate": 1.3070667998556774e-06, "loss": 1.7352, "step": 3845 }, { "epoch": 0.9287611687998069, "grad_norm": 0.28524020314216614, "learning_rate": 1.2982885408921564e-06, "loss": 1.69, "step": 3846 }, { "epoch": 0.9290026563631973, "grad_norm": 0.26982516050338745, "learning_rate": 1.2895394709719954e-06, "loss": 1.5886, "step": 3847 }, { "epoch": 0.9292441439265878, "grad_norm": 0.2917701005935669, "learning_rate": 1.2808195953388836e-06, "loss": 1.6977, "step": 3848 }, { "epoch": 0.9294856314899783, "grad_norm": 0.2904724180698395, "learning_rate": 1.2721289192190123e-06, "loss": 1.7169, "step": 3849 }, { "epoch": 0.9297271190533688, "grad_norm": 0.2919880151748657, "learning_rate": 1.263467447821104e-06, "loss": 1.7694, "step": 3850 }, { "epoch": 0.9299686066167593, "grad_norm": 0.2819439172744751, "learning_rate": 1.254835186336334e-06, "loss": 1.7126, "step": 3851 }, { "epoch": 0.9302100941801498, "grad_norm": 0.3009864091873169, "learning_rate": 1.246232139938397e-06, "loss": 1.8709, "step": 3852 }, { "epoch": 0.9304515817435403, "grad_norm": 0.3153412938117981, "learning_rate": 1.2376583137834685e-06, "loss": 1.7909, "step": 3853 }, { "epoch": 0.9306930693069307, "grad_norm": 0.28017258644104004, "learning_rate": 1.2291137130102104e-06, "loss": 1.6289, "step": 3854 }, { "epoch": 0.9309345568703212, "grad_norm": 0.3065062165260315, "learning_rate": 1.2205983427397816e-06, "loss": 1.7834, "step": 3855 }, { "epoch": 0.9311760444337117, "grad_norm": 0.28788235783576965, "learning_rate": 1.2121122080758052e-06, "loss": 1.6829, "step": 3856 }, { "epoch": 0.9314175319971022, "grad_norm": 0.27349546551704407, "learning_rate": 1.2036553141043906e-06, "loss": 1.4902, "step": 3857 }, { "epoch": 0.9316590195604927, "grad_norm": 0.2849350571632385, "learning_rate": 1.1952276658941164e-06, "loss": 1.6421, "step": 3858 }, { "epoch": 0.9319005071238832, "grad_norm": 0.2708853483200073, "learning_rate": 1.1868292684960535e-06, "loss": 1.5191, "step": 3859 }, { "epoch": 0.9321419946872737, "grad_norm": 0.2766154706478119, "learning_rate": 1.1784601269436978e-06, "loss": 1.5937, "step": 3860 }, { "epoch": 0.9323834822506641, "grad_norm": 0.2782767415046692, "learning_rate": 1.1701202462530536e-06, "loss": 1.7261, "step": 3861 }, { "epoch": 0.9326249698140546, "grad_norm": 0.2797935903072357, "learning_rate": 1.1618096314225779e-06, "loss": 1.6266, "step": 3862 }, { "epoch": 0.9328664573774451, "grad_norm": 0.2885272800922394, "learning_rate": 1.15352828743317e-06, "loss": 1.693, "step": 3863 }, { "epoch": 0.9331079449408356, "grad_norm": 0.28160789608955383, "learning_rate": 1.1452762192481981e-06, "loss": 1.6304, "step": 3864 }, { "epoch": 0.9333494325042261, "grad_norm": 0.26970502734184265, "learning_rate": 1.1370534318134951e-06, "loss": 1.6356, "step": 3865 }, { "epoch": 0.9335909200676166, "grad_norm": 0.2736988663673401, "learning_rate": 1.12885993005733e-06, "loss": 1.5794, "step": 3866 }, { "epoch": 0.933832407631007, "grad_norm": 0.2958070933818817, "learning_rate": 1.1206957188904077e-06, "loss": 1.8218, "step": 3867 }, { "epoch": 0.9340738951943975, "grad_norm": 0.26863253116607666, "learning_rate": 1.1125608032059198e-06, "loss": 1.6391, "step": 3868 }, { "epoch": 0.934315382757788, "grad_norm": 0.29687753319740295, "learning_rate": 1.104455187879444e-06, "loss": 1.8236, "step": 3869 }, { "epoch": 0.9345568703211785, "grad_norm": 0.28738346695899963, "learning_rate": 1.0963788777690497e-06, "loss": 1.6503, "step": 3870 }, { "epoch": 0.934798357884569, "grad_norm": 0.29304075241088867, "learning_rate": 1.0883318777152096e-06, "loss": 1.7738, "step": 3871 }, { "epoch": 0.9350398454479595, "grad_norm": 0.27741140127182007, "learning_rate": 1.0803141925408378e-06, "loss": 1.6146, "step": 3872 }, { "epoch": 0.93528133301135, "grad_norm": 0.30497393012046814, "learning_rate": 1.0723258270512792e-06, "loss": 1.8853, "step": 3873 }, { "epoch": 0.9355228205747405, "grad_norm": 0.28605154156684875, "learning_rate": 1.0643667860343154e-06, "loss": 1.7662, "step": 3874 }, { "epoch": 0.9357643081381309, "grad_norm": 0.28958433866500854, "learning_rate": 1.0564370742601359e-06, "loss": 1.7578, "step": 3875 }, { "epoch": 0.9360057957015214, "grad_norm": 0.28037330508232117, "learning_rate": 1.048536696481367e-06, "loss": 1.6424, "step": 3876 }, { "epoch": 0.9362472832649119, "grad_norm": 0.2637234926223755, "learning_rate": 1.0406656574330376e-06, "loss": 1.5186, "step": 3877 }, { "epoch": 0.9364887708283024, "grad_norm": 0.2860293388366699, "learning_rate": 1.0328239618326074e-06, "loss": 1.5581, "step": 3878 }, { "epoch": 0.9367302583916929, "grad_norm": 0.3104569613933563, "learning_rate": 1.0250116143799448e-06, "loss": 1.8877, "step": 3879 }, { "epoch": 0.9369717459550834, "grad_norm": 0.284689724445343, "learning_rate": 1.0172286197573266e-06, "loss": 1.6588, "step": 3880 }, { "epoch": 0.9372132335184739, "grad_norm": 0.28726133704185486, "learning_rate": 1.0094749826294381e-06, "loss": 1.7134, "step": 3881 }, { "epoch": 0.9374547210818643, "grad_norm": 0.31140753626823425, "learning_rate": 1.0017507076433674e-06, "loss": 1.6876, "step": 3882 }, { "epoch": 0.9376962086452547, "grad_norm": 0.30054032802581787, "learning_rate": 9.94055799428617e-07, "loss": 1.7297, "step": 3883 }, { "epoch": 0.9379376962086452, "grad_norm": 0.29753032326698303, "learning_rate": 9.863902625970644e-07, "loss": 1.7521, "step": 3884 }, { "epoch": 0.9381791837720357, "grad_norm": 0.2824787199497223, "learning_rate": 9.787541017430068e-07, "loss": 1.69, "step": 3885 }, { "epoch": 0.9384206713354262, "grad_norm": 0.2984289228916168, "learning_rate": 9.711473214431222e-07, "loss": 1.9633, "step": 3886 }, { "epoch": 0.9386621588988167, "grad_norm": 0.2809242904186249, "learning_rate": 9.635699262564745e-07, "loss": 1.6255, "step": 3887 }, { "epoch": 0.9389036464622071, "grad_norm": 0.2918394207954407, "learning_rate": 9.560219207245479e-07, "loss": 1.7538, "step": 3888 }, { "epoch": 0.9391451340255976, "grad_norm": 0.3060348629951477, "learning_rate": 9.485033093711682e-07, "loss": 1.8433, "step": 3889 }, { "epoch": 0.9393866215889881, "grad_norm": 0.2847803831100464, "learning_rate": 9.410140967025749e-07, "loss": 1.6937, "step": 3890 }, { "epoch": 0.9396281091523786, "grad_norm": 0.2865715026855469, "learning_rate": 9.335542872073666e-07, "loss": 1.7634, "step": 3891 }, { "epoch": 0.9398695967157691, "grad_norm": 0.3000757396221161, "learning_rate": 9.261238853565446e-07, "loss": 1.9596, "step": 3892 }, { "epoch": 0.9401110842791596, "grad_norm": 0.2781744599342346, "learning_rate": 9.187228956034577e-07, "loss": 1.5145, "step": 3893 }, { "epoch": 0.9403525718425501, "grad_norm": 0.28826314210891724, "learning_rate": 9.113513223838465e-07, "loss": 1.7792, "step": 3894 }, { "epoch": 0.9405940594059405, "grad_norm": 0.2929260730743408, "learning_rate": 9.040091701158104e-07, "loss": 1.7191, "step": 3895 }, { "epoch": 0.940835546969331, "grad_norm": 0.286565899848938, "learning_rate": 8.966964431998237e-07, "loss": 1.6206, "step": 3896 }, { "epoch": 0.9410770345327215, "grad_norm": 0.27345308661460876, "learning_rate": 8.894131460187083e-07, "loss": 1.5865, "step": 3897 }, { "epoch": 0.941318522096112, "grad_norm": 0.2961215674877167, "learning_rate": 8.82159282937678e-07, "loss": 1.7189, "step": 3898 }, { "epoch": 0.9415600096595025, "grad_norm": 0.27731019258499146, "learning_rate": 8.749348583042716e-07, "loss": 1.6043, "step": 3899 }, { "epoch": 0.941801497222893, "grad_norm": 0.29379338026046753, "learning_rate": 8.677398764484035e-07, "loss": 1.7162, "step": 3900 }, { "epoch": 0.9420429847862835, "grad_norm": 0.28957459330558777, "learning_rate": 8.605743416823464e-07, "loss": 1.7266, "step": 3901 }, { "epoch": 0.9422844723496739, "grad_norm": 0.27559155225753784, "learning_rate": 8.534382583006983e-07, "loss": 1.6083, "step": 3902 }, { "epoch": 0.9425259599130644, "grad_norm": 0.28408434987068176, "learning_rate": 8.463316305804325e-07, "loss": 1.6909, "step": 3903 }, { "epoch": 0.9427674474764549, "grad_norm": 0.30094483494758606, "learning_rate": 8.39254462780864e-07, "loss": 1.7717, "step": 3904 }, { "epoch": 0.9430089350398454, "grad_norm": 0.291727215051651, "learning_rate": 8.322067591436389e-07, "loss": 1.6767, "step": 3905 }, { "epoch": 0.9432504226032359, "grad_norm": 0.2792252004146576, "learning_rate": 8.251885238927448e-07, "loss": 1.5777, "step": 3906 }, { "epoch": 0.9434919101666264, "grad_norm": 0.2789081037044525, "learning_rate": 8.181997612345282e-07, "loss": 1.5874, "step": 3907 }, { "epoch": 0.9437333977300169, "grad_norm": 0.2843283414840698, "learning_rate": 8.112404753576496e-07, "loss": 1.5954, "step": 3908 }, { "epoch": 0.9439748852934073, "grad_norm": 0.2803763747215271, "learning_rate": 8.043106704331116e-07, "loss": 1.6252, "step": 3909 }, { "epoch": 0.9442163728567978, "grad_norm": 0.28027307987213135, "learning_rate": 7.974103506142417e-07, "loss": 1.5233, "step": 3910 }, { "epoch": 0.9444578604201883, "grad_norm": 0.308739572763443, "learning_rate": 7.905395200367038e-07, "loss": 1.8099, "step": 3911 }, { "epoch": 0.9446993479835788, "grad_norm": 0.29119589924812317, "learning_rate": 7.83698182818482e-07, "loss": 1.6833, "step": 3912 }, { "epoch": 0.9449408355469693, "grad_norm": 0.28837186098098755, "learning_rate": 7.768863430598905e-07, "loss": 1.8165, "step": 3913 }, { "epoch": 0.9451823231103598, "grad_norm": 0.2852427065372467, "learning_rate": 7.701040048435582e-07, "loss": 1.7777, "step": 3914 }, { "epoch": 0.9454238106737503, "grad_norm": 0.29468417167663574, "learning_rate": 7.633511722344278e-07, "loss": 1.7889, "step": 3915 }, { "epoch": 0.9456652982371407, "grad_norm": 0.28421550989151, "learning_rate": 7.566278492797785e-07, "loss": 1.6566, "step": 3916 }, { "epoch": 0.9459067858005312, "grad_norm": 0.2829783260822296, "learning_rate": 7.499340400091703e-07, "loss": 1.7162, "step": 3917 }, { "epoch": 0.9461482733639217, "grad_norm": 0.2955000698566437, "learning_rate": 7.432697484345052e-07, "loss": 1.7325, "step": 3918 }, { "epoch": 0.9463897609273122, "grad_norm": 0.2817592918872833, "learning_rate": 7.366349785499771e-07, "loss": 1.7062, "step": 3919 }, { "epoch": 0.9466312484907027, "grad_norm": 0.2831808924674988, "learning_rate": 7.300297343320827e-07, "loss": 1.6383, "step": 3920 }, { "epoch": 0.9468727360540932, "grad_norm": 0.3093096613883972, "learning_rate": 7.2345401973965e-07, "loss": 1.7668, "step": 3921 }, { "epoch": 0.9471142236174837, "grad_norm": 0.27752164006233215, "learning_rate": 7.169078387137762e-07, "loss": 1.6342, "step": 3922 }, { "epoch": 0.9473557111808741, "grad_norm": 0.32786694169044495, "learning_rate": 7.10391195177873e-07, "loss": 1.9515, "step": 3923 }, { "epoch": 0.9475971987442646, "grad_norm": 0.267087459564209, "learning_rate": 7.039040930376495e-07, "loss": 1.4224, "step": 3924 }, { "epoch": 0.9478386863076551, "grad_norm": 0.2860214412212372, "learning_rate": 6.974465361811067e-07, "loss": 1.6075, "step": 3925 }, { "epoch": 0.9480801738710456, "grad_norm": 0.27244532108306885, "learning_rate": 6.910185284785264e-07, "loss": 1.5565, "step": 3926 }, { "epoch": 0.9483216614344361, "grad_norm": 0.2986495792865753, "learning_rate": 6.846200737825104e-07, "loss": 1.6445, "step": 3927 }, { "epoch": 0.9485631489978266, "grad_norm": 0.2798667550086975, "learning_rate": 6.782511759279186e-07, "loss": 1.7094, "step": 3928 }, { "epoch": 0.948804636561217, "grad_norm": 0.2928531765937805, "learning_rate": 6.71911838731909e-07, "loss": 1.765, "step": 3929 }, { "epoch": 0.9490461241246075, "grad_norm": 0.26900431513786316, "learning_rate": 6.656020659939144e-07, "loss": 1.5384, "step": 3930 }, { "epoch": 0.949287611687998, "grad_norm": 0.30054423213005066, "learning_rate": 6.593218614956653e-07, "loss": 1.8153, "step": 3931 }, { "epoch": 0.9495290992513885, "grad_norm": 0.31609615683555603, "learning_rate": 6.530712290011565e-07, "loss": 2.003, "step": 3932 }, { "epoch": 0.949770586814779, "grad_norm": 0.285409152507782, "learning_rate": 6.468501722566577e-07, "loss": 1.5697, "step": 3933 }, { "epoch": 0.9500120743781695, "grad_norm": 0.2756394147872925, "learning_rate": 6.406586949907312e-07, "loss": 1.656, "step": 3934 }, { "epoch": 0.95025356194156, "grad_norm": 0.28793469071388245, "learning_rate": 6.34496800914175e-07, "loss": 1.692, "step": 3935 }, { "epoch": 0.9504950495049505, "grad_norm": 0.310819149017334, "learning_rate": 6.283644937201016e-07, "loss": 1.8459, "step": 3936 }, { "epoch": 0.9507365370683409, "grad_norm": 0.2961222529411316, "learning_rate": 6.222617770838546e-07, "loss": 1.8205, "step": 3937 }, { "epoch": 0.9509780246317314, "grad_norm": 0.2894151508808136, "learning_rate": 6.16188654663058e-07, "loss": 1.6931, "step": 3938 }, { "epoch": 0.9512195121951219, "grad_norm": 0.2774762213230133, "learning_rate": 6.101451300975946e-07, "loss": 1.5855, "step": 3939 }, { "epoch": 0.9514609997585124, "grad_norm": 0.29174312949180603, "learning_rate": 6.041312070096172e-07, "loss": 1.6132, "step": 3940 }, { "epoch": 0.9517024873219029, "grad_norm": 0.27548885345458984, "learning_rate": 5.981468890035314e-07, "loss": 1.5677, "step": 3941 }, { "epoch": 0.9519439748852934, "grad_norm": 0.2879452407360077, "learning_rate": 5.921921796659902e-07, "loss": 1.6664, "step": 3942 }, { "epoch": 0.9521854624486839, "grad_norm": 0.3094227612018585, "learning_rate": 5.862670825659111e-07, "loss": 1.8452, "step": 3943 }, { "epoch": 0.9524269500120743, "grad_norm": 0.286016047000885, "learning_rate": 5.803716012544647e-07, "loss": 1.6296, "step": 3944 }, { "epoch": 0.9526684375754648, "grad_norm": 0.28221380710601807, "learning_rate": 5.745057392650632e-07, "loss": 1.6588, "step": 3945 }, { "epoch": 0.9529099251388553, "grad_norm": 0.2891559302806854, "learning_rate": 5.686695001133724e-07, "loss": 1.7216, "step": 3946 }, { "epoch": 0.9531514127022458, "grad_norm": 0.2880290746688843, "learning_rate": 5.628628872973052e-07, "loss": 1.7129, "step": 3947 }, { "epoch": 0.9533929002656363, "grad_norm": 0.2738760709762573, "learning_rate": 5.570859042970167e-07, "loss": 1.5445, "step": 3948 }, { "epoch": 0.9536343878290268, "grad_norm": 0.2887960970401764, "learning_rate": 5.513385545749039e-07, "loss": 1.7553, "step": 3949 }, { "epoch": 0.9538758753924172, "grad_norm": 0.275007426738739, "learning_rate": 5.456208415755948e-07, "loss": 1.687, "step": 3950 }, { "epoch": 0.9541173629558077, "grad_norm": 0.2679949998855591, "learning_rate": 5.399327687259703e-07, "loss": 1.6526, "step": 3951 }, { "epoch": 0.9543588505191982, "grad_norm": 0.2911270558834076, "learning_rate": 5.342743394351313e-07, "loss": 1.6715, "step": 3952 }, { "epoch": 0.9546003380825887, "grad_norm": 0.2992401421070099, "learning_rate": 5.28645557094426e-07, "loss": 1.7486, "step": 3953 }, { "epoch": 0.9548418256459792, "grad_norm": 0.2793290913105011, "learning_rate": 5.230464250774225e-07, "loss": 1.6231, "step": 3954 }, { "epoch": 0.9550833132093697, "grad_norm": 0.29168450832366943, "learning_rate": 5.174769467399255e-07, "loss": 1.7562, "step": 3955 }, { "epoch": 0.9553248007727602, "grad_norm": 0.27952778339385986, "learning_rate": 5.119371254199701e-07, "loss": 1.6488, "step": 3956 }, { "epoch": 0.9555662883361506, "grad_norm": 0.2867637574672699, "learning_rate": 5.06426964437795e-07, "loss": 1.5891, "step": 3957 }, { "epoch": 0.9558077758995411, "grad_norm": 0.28503820300102234, "learning_rate": 5.009464670958974e-07, "loss": 1.6239, "step": 3958 }, { "epoch": 0.9560492634629316, "grad_norm": 0.2997734248638153, "learning_rate": 4.95495636678961e-07, "loss": 1.815, "step": 3959 }, { "epoch": 0.9562907510263221, "grad_norm": 0.28257283568382263, "learning_rate": 4.900744764539167e-07, "loss": 1.6809, "step": 3960 }, { "epoch": 0.9565322385897126, "grad_norm": 0.286513090133667, "learning_rate": 4.846829896698879e-07, "loss": 1.7903, "step": 3961 }, { "epoch": 0.9567737261531031, "grad_norm": 0.2749822437763214, "learning_rate": 4.793211795582398e-07, "loss": 1.555, "step": 3962 }, { "epoch": 0.9570152137164936, "grad_norm": 0.2722713351249695, "learning_rate": 4.739890493325183e-07, "loss": 1.5878, "step": 3963 }, { "epoch": 0.957256701279884, "grad_norm": 0.30979806184768677, "learning_rate": 4.6868660218851725e-07, "loss": 1.6961, "step": 3964 }, { "epoch": 0.9574981888432745, "grad_norm": 0.29073143005371094, "learning_rate": 4.6341384130422216e-07, "loss": 1.7803, "step": 3965 }, { "epoch": 0.957739676406665, "grad_norm": 0.2786603569984436, "learning_rate": 4.581707698398108e-07, "loss": 1.5712, "step": 3966 }, { "epoch": 0.9579811639700555, "grad_norm": 0.28374600410461426, "learning_rate": 4.529573909376972e-07, "loss": 1.7517, "step": 3967 }, { "epoch": 0.958222651533446, "grad_norm": 0.2832549214363098, "learning_rate": 4.477737077224875e-07, "loss": 1.7312, "step": 3968 }, { "epoch": 0.9584641390968365, "grad_norm": 0.28513893485069275, "learning_rate": 4.4261972330097434e-07, "loss": 1.7116, "step": 3969 }, { "epoch": 0.958705626660227, "grad_norm": 0.300758957862854, "learning_rate": 4.3749544076217006e-07, "loss": 1.6529, "step": 3970 }, { "epoch": 0.9589471142236174, "grad_norm": 0.291905015707016, "learning_rate": 4.3240086317728467e-07, "loss": 1.7491, "step": 3971 }, { "epoch": 0.9591886017870079, "grad_norm": 0.2884933650493622, "learning_rate": 4.27335993599709e-07, "loss": 1.8187, "step": 3972 }, { "epoch": 0.9594300893503984, "grad_norm": 0.30293458700180054, "learning_rate": 4.223008350650426e-07, "loss": 1.5803, "step": 3973 }, { "epoch": 0.9596715769137889, "grad_norm": 0.27663272619247437, "learning_rate": 4.17295390591077e-07, "loss": 1.7036, "step": 3974 }, { "epoch": 0.9599130644771794, "grad_norm": 0.28130343556404114, "learning_rate": 4.123196631777904e-07, "loss": 1.6463, "step": 3975 }, { "epoch": 0.9601545520405699, "grad_norm": 0.28887301683425903, "learning_rate": 4.073736558073471e-07, "loss": 1.757, "step": 3976 }, { "epoch": 0.9603960396039604, "grad_norm": 0.2934032082557678, "learning_rate": 4.0245737144410934e-07, "loss": 1.5656, "step": 3977 }, { "epoch": 0.9606375271673508, "grad_norm": 0.27440911531448364, "learning_rate": 3.975708130346145e-07, "loss": 1.5496, "step": 3978 }, { "epoch": 0.9608790147307413, "grad_norm": 0.2799850404262543, "learning_rate": 3.927139835075866e-07, "loss": 1.684, "step": 3979 }, { "epoch": 0.9611205022941318, "grad_norm": 0.28903862833976746, "learning_rate": 3.878868857739415e-07, "loss": 1.5812, "step": 3980 }, { "epoch": 0.9613619898575223, "grad_norm": 0.28020167350769043, "learning_rate": 3.83089522726765e-07, "loss": 1.7196, "step": 3981 }, { "epoch": 0.9616034774209128, "grad_norm": 0.2839907109737396, "learning_rate": 3.7832189724131827e-07, "loss": 1.7121, "step": 3982 }, { "epoch": 0.9618449649843033, "grad_norm": 0.28018689155578613, "learning_rate": 3.7358401217505443e-07, "loss": 1.5548, "step": 3983 }, { "epoch": 0.9620864525476938, "grad_norm": 0.28298869729042053, "learning_rate": 3.6887587036759097e-07, "loss": 1.6588, "step": 3984 }, { "epoch": 0.9623279401110842, "grad_norm": 0.28870052099227905, "learning_rate": 3.641974746407151e-07, "loss": 1.896, "step": 3985 }, { "epoch": 0.9625694276744747, "grad_norm": 0.2799728810787201, "learning_rate": 3.5954882779839493e-07, "loss": 1.6041, "step": 3986 }, { "epoch": 0.9628109152378652, "grad_norm": 0.2967098355293274, "learning_rate": 3.549299326267741e-07, "loss": 1.7802, "step": 3987 }, { "epoch": 0.9630524028012557, "grad_norm": 0.30360880494117737, "learning_rate": 3.5034079189413815e-07, "loss": 1.8652, "step": 3988 }, { "epoch": 0.9632938903646462, "grad_norm": 0.278390496969223, "learning_rate": 3.457814083509814e-07, "loss": 1.4824, "step": 3989 }, { "epoch": 0.9635353779280367, "grad_norm": 0.28068608045578003, "learning_rate": 3.412517847299179e-07, "loss": 1.6422, "step": 3990 }, { "epoch": 0.9637768654914272, "grad_norm": 0.2911674678325653, "learning_rate": 3.3675192374575394e-07, "loss": 1.6614, "step": 3991 }, { "epoch": 0.9640183530548176, "grad_norm": 0.28525784611701965, "learning_rate": 3.3228182809545984e-07, "loss": 1.7905, "step": 3992 }, { "epoch": 0.9642598406182081, "grad_norm": 0.29642611742019653, "learning_rate": 3.278415004581481e-07, "loss": 1.6817, "step": 3993 }, { "epoch": 0.9645013281815986, "grad_norm": 0.27766481041908264, "learning_rate": 3.2343094349510104e-07, "loss": 1.6619, "step": 3994 }, { "epoch": 0.9647428157449891, "grad_norm": 0.27751415967941284, "learning_rate": 3.1905015984975974e-07, "loss": 1.5445, "step": 3995 }, { "epoch": 0.9649843033083796, "grad_norm": 0.27992838621139526, "learning_rate": 3.1469915214770717e-07, "loss": 1.6712, "step": 3996 }, { "epoch": 0.9652257908717701, "grad_norm": 0.2856504023075104, "learning_rate": 3.103779229967074e-07, "loss": 1.6732, "step": 3997 }, { "epoch": 0.9654672784351606, "grad_norm": 0.28135809302330017, "learning_rate": 3.0608647498664434e-07, "loss": 1.6533, "step": 3998 }, { "epoch": 0.965708765998551, "grad_norm": 0.27954551577568054, "learning_rate": 3.018248106895771e-07, "loss": 1.5896, "step": 3999 }, { "epoch": 0.9659502535619415, "grad_norm": 0.31844958662986755, "learning_rate": 2.9759293265970155e-07, "loss": 2.1066, "step": 4000 }, { "epoch": 0.966191741125332, "grad_norm": 0.29068711400032043, "learning_rate": 2.9339084343337207e-07, "loss": 1.7638, "step": 4001 }, { "epoch": 0.9664332286887225, "grad_norm": 0.27710625529289246, "learning_rate": 2.8921854552907415e-07, "loss": 1.6426, "step": 4002 }, { "epoch": 0.966674716252113, "grad_norm": 0.28776073455810547, "learning_rate": 2.8507604144744626e-07, "loss": 1.739, "step": 4003 }, { "epoch": 0.9669162038155035, "grad_norm": 0.28407105803489685, "learning_rate": 2.8096333367128024e-07, "loss": 1.6652, "step": 4004 }, { "epoch": 0.967157691378894, "grad_norm": 0.29590874910354614, "learning_rate": 2.768804246654932e-07, "loss": 1.5099, "step": 4005 }, { "epoch": 0.9673991789422844, "grad_norm": 0.2906011939048767, "learning_rate": 2.728273168771445e-07, "loss": 1.6932, "step": 4006 }, { "epoch": 0.9676406665056749, "grad_norm": 0.29354578256607056, "learning_rate": 2.6880401273544633e-07, "loss": 1.818, "step": 4007 }, { "epoch": 0.9678821540690654, "grad_norm": 0.2981238067150116, "learning_rate": 2.6481051465173654e-07, "loss": 1.7239, "step": 4008 }, { "epoch": 0.9681236416324559, "grad_norm": 0.2789437770843506, "learning_rate": 2.6084682501948396e-07, "loss": 1.6351, "step": 4009 }, { "epoch": 0.9683651291958464, "grad_norm": 0.283161997795105, "learning_rate": 2.569129462143105e-07, "loss": 1.8456, "step": 4010 }, { "epoch": 0.9686066167592369, "grad_norm": 0.2940691113471985, "learning_rate": 2.5300888059395234e-07, "loss": 1.7471, "step": 4011 }, { "epoch": 0.9688481043226274, "grad_norm": 0.29267147183418274, "learning_rate": 2.491346304982878e-07, "loss": 1.7974, "step": 4012 }, { "epoch": 0.9690895918860178, "grad_norm": 0.280164897441864, "learning_rate": 2.452901982493261e-07, "loss": 1.6318, "step": 4013 }, { "epoch": 0.9693310794494083, "grad_norm": 0.28439733386039734, "learning_rate": 2.414755861511908e-07, "loss": 1.5752, "step": 4014 }, { "epoch": 0.9695725670127988, "grad_norm": 0.29540446400642395, "learning_rate": 2.3769079649015315e-07, "loss": 1.5721, "step": 4015 }, { "epoch": 0.9698140545761893, "grad_norm": 0.2872696816921234, "learning_rate": 2.3393583153459853e-07, "loss": 1.6942, "step": 4016 }, { "epoch": 0.9700555421395798, "grad_norm": 0.28481537103652954, "learning_rate": 2.3021069353504344e-07, "loss": 1.7121, "step": 4017 }, { "epoch": 0.9702970297029703, "grad_norm": 0.2834312319755554, "learning_rate": 2.265153847241186e-07, "loss": 1.7231, "step": 4018 }, { "epoch": 0.9705385172663608, "grad_norm": 0.36638566851615906, "learning_rate": 2.2284990731658573e-07, "loss": 1.7904, "step": 4019 }, { "epoch": 0.9707800048297512, "grad_norm": 0.29458123445510864, "learning_rate": 2.1921426350932639e-07, "loss": 1.7091, "step": 4020 }, { "epoch": 0.9710214923931417, "grad_norm": 0.2861957550048828, "learning_rate": 2.1560845548133647e-07, "loss": 1.6589, "step": 4021 }, { "epoch": 0.9712629799565322, "grad_norm": 0.2898642420768738, "learning_rate": 2.1203248539373166e-07, "loss": 1.6233, "step": 4022 }, { "epoch": 0.9715044675199227, "grad_norm": 0.300205796957016, "learning_rate": 2.084863553897476e-07, "loss": 1.7665, "step": 4023 }, { "epoch": 0.9717459550833132, "grad_norm": 0.2905106544494629, "learning_rate": 2.0497006759472859e-07, "loss": 1.7173, "step": 4024 }, { "epoch": 0.9719874426467037, "grad_norm": 0.3021292984485626, "learning_rate": 2.0148362411614997e-07, "loss": 1.7901, "step": 4025 }, { "epoch": 0.9722289302100942, "grad_norm": 0.2960629165172577, "learning_rate": 1.9802702704357357e-07, "loss": 1.9125, "step": 4026 }, { "epoch": 0.9724704177734846, "grad_norm": 0.2897185981273651, "learning_rate": 1.9460027844869777e-07, "loss": 1.7529, "step": 4027 }, { "epoch": 0.9727119053368751, "grad_norm": 0.2937391996383667, "learning_rate": 1.912033803853186e-07, "loss": 1.5834, "step": 4028 }, { "epoch": 0.9729533929002656, "grad_norm": 0.28485971689224243, "learning_rate": 1.8783633488934082e-07, "loss": 1.8083, "step": 4029 }, { "epoch": 0.9731948804636561, "grad_norm": 0.2996341288089752, "learning_rate": 1.8449914397878353e-07, "loss": 1.819, "step": 4030 }, { "epoch": 0.9734363680270466, "grad_norm": 0.2845102846622467, "learning_rate": 1.8119180965376903e-07, "loss": 1.6396, "step": 4031 }, { "epoch": 0.9736778555904371, "grad_norm": 0.2807163596153259, "learning_rate": 1.7791433389652834e-07, "loss": 1.6056, "step": 4032 }, { "epoch": 0.9739193431538276, "grad_norm": 0.2934764325618744, "learning_rate": 1.7466671867139018e-07, "loss": 1.8104, "step": 4033 }, { "epoch": 0.974160830717218, "grad_norm": 0.2934281826019287, "learning_rate": 1.7144896592479197e-07, "loss": 1.6943, "step": 4034 }, { "epoch": 0.9744023182806085, "grad_norm": 0.30650898814201355, "learning_rate": 1.6826107758527998e-07, "loss": 1.6751, "step": 4035 }, { "epoch": 0.974643805843999, "grad_norm": 0.2871125340461731, "learning_rate": 1.6510305556348137e-07, "loss": 1.6884, "step": 4036 }, { "epoch": 0.9748852934073895, "grad_norm": 0.2861565053462982, "learning_rate": 1.6197490175214324e-07, "loss": 1.7488, "step": 4037 }, { "epoch": 0.97512678097078, "grad_norm": 0.29717275500297546, "learning_rate": 1.5887661802609922e-07, "loss": 1.7399, "step": 4038 }, { "epoch": 0.9753682685341705, "grad_norm": 0.29214397072792053, "learning_rate": 1.558082062422861e-07, "loss": 1.7829, "step": 4039 }, { "epoch": 0.975609756097561, "grad_norm": 0.2873966693878174, "learning_rate": 1.5276966823973838e-07, "loss": 1.6397, "step": 4040 }, { "epoch": 0.9758512436609514, "grad_norm": 0.28294169902801514, "learning_rate": 1.4976100583958264e-07, "loss": 1.5401, "step": 4041 }, { "epoch": 0.9760927312243419, "grad_norm": 0.28937163949012756, "learning_rate": 1.4678222084503758e-07, "loss": 1.7171, "step": 4042 }, { "epoch": 0.9763342187877324, "grad_norm": 0.27437540888786316, "learning_rate": 1.4383331504142507e-07, "loss": 1.597, "step": 4043 }, { "epoch": 0.9765757063511229, "grad_norm": 0.2892947494983673, "learning_rate": 1.4091429019614245e-07, "loss": 1.7393, "step": 4044 }, { "epoch": 0.9768171939145134, "grad_norm": 0.28885194659233093, "learning_rate": 1.3802514805869028e-07, "loss": 1.6814, "step": 4045 }, { "epoch": 0.9770586814779039, "grad_norm": 0.28577619791030884, "learning_rate": 1.3516589036066118e-07, "loss": 1.7054, "step": 4046 }, { "epoch": 0.9773001690412944, "grad_norm": 0.28375500440597534, "learning_rate": 1.3233651881572883e-07, "loss": 1.7147, "step": 4047 }, { "epoch": 0.9775416566046848, "grad_norm": 0.29194769263267517, "learning_rate": 1.2953703511965897e-07, "loss": 1.7326, "step": 4048 }, { "epoch": 0.9777831441680753, "grad_norm": 0.2715654969215393, "learning_rate": 1.267674409503039e-07, "loss": 1.5483, "step": 4049 }, { "epoch": 0.9780246317314658, "grad_norm": 0.29190412163734436, "learning_rate": 1.2402773796759692e-07, "loss": 1.8056, "step": 4050 }, { "epoch": 0.9782661192948563, "grad_norm": 0.2856195867061615, "learning_rate": 1.2131792781356345e-07, "loss": 1.6973, "step": 4051 }, { "epoch": 0.9785076068582468, "grad_norm": 0.2819734811782837, "learning_rate": 1.1863801211230984e-07, "loss": 1.6848, "step": 4052 }, { "epoch": 0.9787490944216373, "grad_norm": 0.2744895815849304, "learning_rate": 1.1598799247002912e-07, "loss": 1.6642, "step": 4053 }, { "epoch": 0.9789905819850278, "grad_norm": 0.2842368483543396, "learning_rate": 1.1336787047498409e-07, "loss": 1.7056, "step": 4054 }, { "epoch": 0.9792320695484182, "grad_norm": 0.29554474353790283, "learning_rate": 1.1077764769754085e-07, "loss": 1.8154, "step": 4055 }, { "epoch": 0.9794735571118087, "grad_norm": 0.27927619218826294, "learning_rate": 1.0821732569011866e-07, "loss": 1.5707, "step": 4056 }, { "epoch": 0.9797150446751992, "grad_norm": 0.2810564935207367, "learning_rate": 1.0568690598723451e-07, "loss": 1.7296, "step": 4057 }, { "epoch": 0.9799565322385897, "grad_norm": 0.2852177619934082, "learning_rate": 1.0318639010547526e-07, "loss": 1.6966, "step": 4058 }, { "epoch": 0.9801980198019802, "grad_norm": 0.2877807021141052, "learning_rate": 1.0071577954351985e-07, "loss": 1.6191, "step": 4059 }, { "epoch": 0.9804395073653707, "grad_norm": 0.2901590168476105, "learning_rate": 9.827507578209493e-08, "loss": 1.7534, "step": 4060 }, { "epoch": 0.9806809949287612, "grad_norm": 0.27793508768081665, "learning_rate": 9.586428028403593e-08, "loss": 1.5497, "step": 4061 }, { "epoch": 0.9809224824921516, "grad_norm": 0.2883639335632324, "learning_rate": 9.348339449422594e-08, "loss": 1.7365, "step": 4062 }, { "epoch": 0.9811639700555421, "grad_norm": 0.2988511621952057, "learning_rate": 9.113241983964016e-08, "loss": 1.5809, "step": 4063 }, { "epoch": 0.9814054576189326, "grad_norm": 0.28930485248565674, "learning_rate": 8.88113577293126e-08, "loss": 1.7983, "step": 4064 }, { "epoch": 0.9816469451823231, "grad_norm": 0.28008490800857544, "learning_rate": 8.652020955435825e-08, "loss": 1.6522, "step": 4065 }, { "epoch": 0.9818884327457136, "grad_norm": 0.3092018663883209, "learning_rate": 8.4258976687962e-08, "loss": 1.8771, "step": 4066 }, { "epoch": 0.9821299203091041, "grad_norm": 0.2992875277996063, "learning_rate": 8.202766048538423e-08, "loss": 1.6963, "step": 4067 }, { "epoch": 0.9823714078724946, "grad_norm": 0.30520957708358765, "learning_rate": 7.982626228394408e-08, "loss": 1.8652, "step": 4068 }, { "epoch": 0.982612895435885, "grad_norm": 0.29011887311935425, "learning_rate": 7.765478340303057e-08, "loss": 1.7634, "step": 4069 }, { "epoch": 0.9828543829992755, "grad_norm": 0.27978020906448364, "learning_rate": 7.551322514411374e-08, "loss": 1.6892, "step": 4070 }, { "epoch": 0.983095870562666, "grad_norm": 0.3046594262123108, "learning_rate": 7.340158879071136e-08, "loss": 1.773, "step": 4071 }, { "epoch": 0.9833373581260565, "grad_norm": 0.279558390378952, "learning_rate": 7.131987560843323e-08, "loss": 1.6129, "step": 4072 }, { "epoch": 0.983578845689447, "grad_norm": 0.29142114520072937, "learning_rate": 6.926808684492581e-08, "loss": 1.7623, "step": 4073 }, { "epoch": 0.9838203332528375, "grad_norm": 0.27713072299957275, "learning_rate": 6.724622372992206e-08, "loss": 1.5533, "step": 4074 }, { "epoch": 0.984061820816228, "grad_norm": 0.282428115606308, "learning_rate": 6.525428747520823e-08, "loss": 1.6276, "step": 4075 }, { "epoch": 0.9843033083796184, "grad_norm": 0.2742138206958771, "learning_rate": 6.329227927464599e-08, "loss": 1.7086, "step": 4076 }, { "epoch": 0.9845447959430089, "grad_norm": 0.2692698538303375, "learning_rate": 6.136020030414469e-08, "loss": 1.4676, "step": 4077 }, { "epoch": 0.9847862835063994, "grad_norm": 0.2890666425228119, "learning_rate": 5.9458051721683614e-08, "loss": 1.7142, "step": 4078 }, { "epoch": 0.9850277710697899, "grad_norm": 0.29658809304237366, "learning_rate": 5.758583466730638e-08, "loss": 1.841, "step": 4079 }, { "epoch": 0.9852692586331804, "grad_norm": 0.2951255142688751, "learning_rate": 5.57435502631043e-08, "loss": 1.6484, "step": 4080 }, { "epoch": 0.9855107461965709, "grad_norm": 0.28006869554519653, "learning_rate": 5.393119961324966e-08, "loss": 1.7655, "step": 4081 }, { "epoch": 0.9857522337599613, "grad_norm": 0.286891371011734, "learning_rate": 5.2148783803956934e-08, "loss": 1.7461, "step": 4082 }, { "epoch": 0.9859937213233518, "grad_norm": 0.2748337984085083, "learning_rate": 5.039630390351047e-08, "loss": 1.5034, "step": 4083 }, { "epoch": 0.9862352088867423, "grad_norm": 0.2836664021015167, "learning_rate": 4.8673760962236746e-08, "loss": 1.6507, "step": 4084 }, { "epoch": 0.9864766964501328, "grad_norm": 0.2878413200378418, "learning_rate": 4.698115601253772e-08, "loss": 1.6787, "step": 4085 }, { "epoch": 0.9867181840135233, "grad_norm": 0.2947593629360199, "learning_rate": 4.531849006886302e-08, "loss": 1.8441, "step": 4086 }, { "epoch": 0.9869596715769138, "grad_norm": 0.2792850732803345, "learning_rate": 4.3685764127721076e-08, "loss": 1.5852, "step": 4087 }, { "epoch": 0.9872011591403043, "grad_norm": 0.29473429918289185, "learning_rate": 4.2082979167673564e-08, "loss": 1.7987, "step": 4088 }, { "epoch": 0.9874426467036947, "grad_norm": 0.2965838611125946, "learning_rate": 4.051013614934096e-08, "loss": 1.7967, "step": 4089 }, { "epoch": 0.9876841342670852, "grad_norm": 0.2892136573791504, "learning_rate": 3.896723601539698e-08, "loss": 1.8447, "step": 4090 }, { "epoch": 0.9879256218304757, "grad_norm": 0.2873370945453644, "learning_rate": 3.745427969056858e-08, "loss": 1.784, "step": 4091 }, { "epoch": 0.9881671093938662, "grad_norm": 0.2879502475261688, "learning_rate": 3.597126808163043e-08, "loss": 1.7381, "step": 4092 }, { "epoch": 0.9884085969572567, "grad_norm": 0.28820958733558655, "learning_rate": 3.4518202077421515e-08, "loss": 1.5987, "step": 4093 }, { "epoch": 0.9886500845206472, "grad_norm": 0.2767198085784912, "learning_rate": 3.309508254882299e-08, "loss": 1.7025, "step": 4094 }, { "epoch": 0.9888915720840377, "grad_norm": 0.28355032205581665, "learning_rate": 3.170191034877479e-08, "loss": 1.6779, "step": 4095 }, { "epoch": 0.9891330596474281, "grad_norm": 0.27833428978919983, "learning_rate": 3.033868631226455e-08, "loss": 1.6684, "step": 4096 }, { "epoch": 0.9893745472108186, "grad_norm": 0.2811650037765503, "learning_rate": 2.9005411256327608e-08, "loss": 1.6721, "step": 4097 }, { "epoch": 0.9896160347742091, "grad_norm": 0.27860844135284424, "learning_rate": 2.770208598006363e-08, "loss": 1.7799, "step": 4098 }, { "epoch": 0.9898575223375996, "grad_norm": 0.30014029145240784, "learning_rate": 2.6428711264597782e-08, "loss": 1.7932, "step": 4099 }, { "epoch": 0.9900990099009901, "grad_norm": 0.29555508494377136, "learning_rate": 2.518528787313068e-08, "loss": 1.7893, "step": 4100 }, { "epoch": 0.9903404974643806, "grad_norm": 0.2952914535999298, "learning_rate": 2.397181655089398e-08, "loss": 1.6317, "step": 4101 }, { "epoch": 0.9905819850277711, "grad_norm": 0.28481733798980713, "learning_rate": 2.278829802517812e-08, "loss": 1.5957, "step": 4102 }, { "epoch": 0.9908234725911615, "grad_norm": 0.2788567543029785, "learning_rate": 2.1634733005310158e-08, "loss": 1.6752, "step": 4103 }, { "epoch": 0.991064960154552, "grad_norm": 0.27494505047798157, "learning_rate": 2.0511122182675923e-08, "loss": 1.7471, "step": 4104 }, { "epoch": 0.9913064477179425, "grad_norm": 0.283424973487854, "learning_rate": 1.9417466230703397e-08, "loss": 1.5881, "step": 4105 }, { "epoch": 0.991547935281333, "grad_norm": 0.2902548015117645, "learning_rate": 1.835376580486825e-08, "loss": 1.6312, "step": 4106 }, { "epoch": 0.9917894228447235, "grad_norm": 0.2776944041252136, "learning_rate": 1.7320021542693853e-08, "loss": 1.6578, "step": 4107 }, { "epoch": 0.992030910408114, "grad_norm": 0.2889111638069153, "learning_rate": 1.6316234063745718e-08, "loss": 1.6646, "step": 4108 }, { "epoch": 0.9922723979715045, "grad_norm": 0.2885117530822754, "learning_rate": 1.5342403969631492e-08, "loss": 1.6566, "step": 4109 }, { "epoch": 0.992513885534895, "grad_norm": 0.298667848110199, "learning_rate": 1.4398531844017627e-08, "loss": 1.7558, "step": 4110 }, { "epoch": 0.9927553730982854, "grad_norm": 0.2731683552265167, "learning_rate": 1.3484618252607162e-08, "loss": 1.4108, "step": 4111 }, { "epoch": 0.9929968606616759, "grad_norm": 0.2888652980327606, "learning_rate": 1.260066374314528e-08, "loss": 1.6021, "step": 4112 }, { "epoch": 0.9932383482250664, "grad_norm": 0.2830989360809326, "learning_rate": 1.1746668845424858e-08, "loss": 1.6842, "step": 4113 }, { "epoch": 0.9934798357884569, "grad_norm": 0.28587043285369873, "learning_rate": 1.0922634071280913e-08, "loss": 1.616, "step": 4114 }, { "epoch": 0.9937213233518474, "grad_norm": 0.27678626775741577, "learning_rate": 1.0128559914596159e-08, "loss": 1.7028, "step": 4115 }, { "epoch": 0.9939628109152379, "grad_norm": 0.28679585456848145, "learning_rate": 9.364446851289899e-09, "loss": 1.6563, "step": 4116 }, { "epoch": 0.9942042984786283, "grad_norm": 0.2937733829021454, "learning_rate": 8.630295339329131e-09, "loss": 1.7667, "step": 4117 }, { "epoch": 0.9944457860420188, "grad_norm": 0.29052743315696716, "learning_rate": 7.926105818728547e-09, "loss": 1.7307, "step": 4118 }, { "epoch": 0.9946872736054093, "grad_norm": 0.29469937086105347, "learning_rate": 7.251878711528326e-09, "loss": 1.6921, "step": 4119 }, { "epoch": 0.9949287611687998, "grad_norm": 0.27424147725105286, "learning_rate": 6.607614421832997e-09, "loss": 1.5347, "step": 4120 }, { "epoch": 0.9951702487321903, "grad_norm": 0.2801046073436737, "learning_rate": 5.993313335772577e-09, "loss": 1.6008, "step": 4121 }, { "epoch": 0.9954117362955808, "grad_norm": 0.2919626832008362, "learning_rate": 5.408975821524775e-09, "loss": 1.8283, "step": 4122 }, { "epoch": 0.9956532238589713, "grad_norm": 0.2922534644603729, "learning_rate": 4.854602229303895e-09, "loss": 1.7255, "step": 4123 }, { "epoch": 0.9958947114223617, "grad_norm": 0.2929743826389313, "learning_rate": 4.3301928913774825e-09, "loss": 1.9228, "step": 4124 }, { "epoch": 0.9961361989857522, "grad_norm": 0.2770915925502777, "learning_rate": 3.835748122044125e-09, "loss": 1.5987, "step": 4125 }, { "epoch": 0.9963776865491427, "grad_norm": 0.28377971053123474, "learning_rate": 3.371268217644552e-09, "loss": 1.6763, "step": 4126 }, { "epoch": 0.9966191741125332, "grad_norm": 0.30905595421791077, "learning_rate": 2.936753456567187e-09, "loss": 1.796, "step": 4127 }, { "epoch": 0.9968606616759237, "grad_norm": 0.29452139139175415, "learning_rate": 2.532204099225943e-09, "loss": 1.7421, "step": 4128 }, { "epoch": 0.9971021492393142, "grad_norm": 0.27836665511131287, "learning_rate": 2.1576203880879775e-09, "loss": 1.6496, "step": 4129 }, { "epoch": 0.9973436368027047, "grad_norm": 0.30187124013900757, "learning_rate": 1.81300254766259e-09, "loss": 1.8534, "step": 4130 }, { "epoch": 0.9975851243660951, "grad_norm": 0.2946297526359558, "learning_rate": 1.4983507844956723e-09, "loss": 1.788, "step": 4131 }, { "epoch": 0.9978266119294856, "grad_norm": 0.28089439868927, "learning_rate": 1.2136652871586051e-09, "loss": 1.5319, "step": 4132 }, { "epoch": 0.9980680994928761, "grad_norm": 0.2959831953048706, "learning_rate": 9.589462262871163e-10, "loss": 1.8031, "step": 4133 }, { "epoch": 0.9983095870562666, "grad_norm": 0.2971201539039612, "learning_rate": 7.341937545479738e-10, "loss": 1.7426, "step": 4134 }, { "epoch": 0.9985510746196571, "grad_norm": 0.28376510739326477, "learning_rate": 5.39408006633435e-10, "loss": 1.6963, "step": 4135 }, { "epoch": 0.9987925621830476, "grad_norm": 0.2785360515117645, "learning_rate": 3.745890992945533e-10, "loss": 1.7218, "step": 4136 }, { "epoch": 0.9990340497464381, "grad_norm": 0.2925812900066376, "learning_rate": 2.397371313134222e-10, "loss": 1.5044, "step": 4137 }, { "epoch": 0.9992755373098285, "grad_norm": 0.2958856225013733, "learning_rate": 1.3485218350872686e-10, "loss": 1.7866, "step": 4138 }, { "epoch": 0.999517024873219, "grad_norm": 0.2930704951286316, "learning_rate": 5.993431874684596e-11, "loss": 1.7573, "step": 4139 }, { "epoch": 0.9997585124366095, "grad_norm": 0.2895719110965729, "learning_rate": 1.498358193074978e-11, "loss": 1.7456, "step": 4140 }, { "epoch": 1.0, "grad_norm": 0.28141534328460693, "learning_rate": 0.0, "loss": 1.619, "step": 4141 }, { "epoch": 1.0, "step": 4141, "total_flos": 6.383347059105202e+18, "train_loss": 1.7380205562396258, "train_runtime": 122926.7048, "train_samples_per_second": 0.135, "train_steps_per_second": 0.034 } ], "logging_steps": 1.0, "max_steps": 4141, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.383347059105202e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }