{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9896907216494846, "eval_steps": 500, "global_step": 435, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006872852233676976, "grad_norm": 0.8787163686521929, "learning_rate": 4.5454545454545455e-06, "loss": 1.3132, "step": 1 }, { "epoch": 0.013745704467353952, "grad_norm": 0.8745142001431168, "learning_rate": 9.090909090909091e-06, "loss": 1.3159, "step": 2 }, { "epoch": 0.020618556701030927, "grad_norm": 0.9087297378930481, "learning_rate": 1.3636363636363637e-05, "loss": 1.3225, "step": 3 }, { "epoch": 0.027491408934707903, "grad_norm": 0.8399095237469762, "learning_rate": 1.8181818181818182e-05, "loss": 1.2664, "step": 4 }, { "epoch": 0.03436426116838488, "grad_norm": 0.8812518157356586, "learning_rate": 2.272727272727273e-05, "loss": 1.3062, "step": 5 }, { "epoch": 0.041237113402061855, "grad_norm": 0.8188461846417425, "learning_rate": 2.7272727272727273e-05, "loss": 1.2826, "step": 6 }, { "epoch": 0.048109965635738834, "grad_norm": 0.6673163495229003, "learning_rate": 3.181818181818182e-05, "loss": 1.173, "step": 7 }, { "epoch": 0.054982817869415807, "grad_norm": 0.5776728278756518, "learning_rate": 3.6363636363636364e-05, "loss": 1.1413, "step": 8 }, { "epoch": 0.061855670103092786, "grad_norm": 0.5099173550442497, "learning_rate": 4.0909090909090915e-05, "loss": 1.0567, "step": 9 }, { "epoch": 0.06872852233676977, "grad_norm": 0.5261602414277569, "learning_rate": 4.545454545454546e-05, "loss": 1.0071, "step": 10 }, { "epoch": 0.07560137457044673, "grad_norm": 0.5814275777211757, "learning_rate": 5e-05, "loss": 0.9513, "step": 11 }, { "epoch": 0.08247422680412371, "grad_norm": 0.5747558406218217, "learning_rate": 5.4545454545454546e-05, "loss": 0.8876, "step": 12 }, { "epoch": 0.08934707903780069, "grad_norm": 0.6073769267220694, "learning_rate": 5.90909090909091e-05, "loss": 0.8372, "step": 13 }, { "epoch": 0.09621993127147767, "grad_norm": 0.5248086576375874, "learning_rate": 6.363636363636364e-05, "loss": 0.7688, "step": 14 }, { "epoch": 0.10309278350515463, "grad_norm": 0.5235010503505939, "learning_rate": 6.818181818181818e-05, "loss": 0.701, "step": 15 }, { "epoch": 0.10996563573883161, "grad_norm": 0.4181074290208477, "learning_rate": 7.272727272727273e-05, "loss": 0.6258, "step": 16 }, { "epoch": 0.11683848797250859, "grad_norm": 0.36050872954123153, "learning_rate": 7.727272727272727e-05, "loss": 0.5911, "step": 17 }, { "epoch": 0.12371134020618557, "grad_norm": 0.26606954132597654, "learning_rate": 8.181818181818183e-05, "loss": 0.5537, "step": 18 }, { "epoch": 0.13058419243986255, "grad_norm": 0.2512795993534893, "learning_rate": 8.636363636363637e-05, "loss": 0.5466, "step": 19 }, { "epoch": 0.13745704467353953, "grad_norm": 0.2152533409896265, "learning_rate": 9.090909090909092e-05, "loss": 0.55, "step": 20 }, { "epoch": 0.14432989690721648, "grad_norm": 0.2337951137318842, "learning_rate": 9.545454545454546e-05, "loss": 0.5106, "step": 21 }, { "epoch": 0.15120274914089346, "grad_norm": 0.2107755578982072, "learning_rate": 0.0001, "loss": 0.5171, "step": 22 }, { "epoch": 0.15807560137457044, "grad_norm": 0.25196498727335587, "learning_rate": 0.00010454545454545455, "loss": 0.4914, "step": 23 }, { "epoch": 0.16494845360824742, "grad_norm": 0.19453194298455792, "learning_rate": 0.00010909090909090909, "loss": 0.4751, "step": 24 }, { "epoch": 0.1718213058419244, "grad_norm": 0.22560806544365505, "learning_rate": 0.00011363636363636365, "loss": 0.4924, "step": 25 }, { "epoch": 0.17869415807560138, "grad_norm": 0.14641932896583643, "learning_rate": 0.0001181818181818182, "loss": 0.4866, "step": 26 }, { "epoch": 0.18556701030927836, "grad_norm": 0.1499236725930378, "learning_rate": 0.00012272727272727272, "loss": 0.4812, "step": 27 }, { "epoch": 0.19243986254295534, "grad_norm": 0.1356769970969474, "learning_rate": 0.00012727272727272728, "loss": 0.4499, "step": 28 }, { "epoch": 0.19931271477663232, "grad_norm": 0.1194925615955461, "learning_rate": 0.0001318181818181818, "loss": 0.4582, "step": 29 }, { "epoch": 0.20618556701030927, "grad_norm": 0.1192278167152467, "learning_rate": 0.00013636363636363637, "loss": 0.4517, "step": 30 }, { "epoch": 0.21305841924398625, "grad_norm": 0.11847931284495589, "learning_rate": 0.00014090909090909093, "loss": 0.4518, "step": 31 }, { "epoch": 0.21993127147766323, "grad_norm": 0.12060222593960905, "learning_rate": 0.00014545454545454546, "loss": 0.4454, "step": 32 }, { "epoch": 0.2268041237113402, "grad_norm": 0.11278362242668162, "learning_rate": 0.00015000000000000001, "loss": 0.45, "step": 33 }, { "epoch": 0.23367697594501718, "grad_norm": 0.11991229318193297, "learning_rate": 0.00015454545454545454, "loss": 0.4513, "step": 34 }, { "epoch": 0.24054982817869416, "grad_norm": 0.10011835943787034, "learning_rate": 0.0001590909090909091, "loss": 0.4327, "step": 35 }, { "epoch": 0.24742268041237114, "grad_norm": 0.10181283297982383, "learning_rate": 0.00016363636363636366, "loss": 0.4358, "step": 36 }, { "epoch": 0.2542955326460481, "grad_norm": 0.0963166583351938, "learning_rate": 0.0001681818181818182, "loss": 0.4296, "step": 37 }, { "epoch": 0.2611683848797251, "grad_norm": 0.10467681911478774, "learning_rate": 0.00017272727272727275, "loss": 0.4262, "step": 38 }, { "epoch": 0.26804123711340205, "grad_norm": 0.10947746121065367, "learning_rate": 0.00017727272727272728, "loss": 0.4474, "step": 39 }, { "epoch": 0.27491408934707906, "grad_norm": 0.10512828010917714, "learning_rate": 0.00018181818181818183, "loss": 0.4196, "step": 40 }, { "epoch": 0.281786941580756, "grad_norm": 0.09354211923076758, "learning_rate": 0.00018636363636363636, "loss": 0.4073, "step": 41 }, { "epoch": 0.28865979381443296, "grad_norm": 0.09623053036305426, "learning_rate": 0.00019090909090909092, "loss": 0.4127, "step": 42 }, { "epoch": 0.29553264604810997, "grad_norm": 0.09513290936780307, "learning_rate": 0.00019545454545454548, "loss": 0.4121, "step": 43 }, { "epoch": 0.3024054982817869, "grad_norm": 0.1009667554598984, "learning_rate": 0.0002, "loss": 0.4336, "step": 44 }, { "epoch": 0.30927835051546393, "grad_norm": 0.10826196086367862, "learning_rate": 0.00019999677214588312, "loss": 0.4093, "step": 45 }, { "epoch": 0.3161512027491409, "grad_norm": 0.10525019938447876, "learning_rate": 0.00019998708879191335, "loss": 0.4093, "step": 46 }, { "epoch": 0.3230240549828179, "grad_norm": 0.0975596947832485, "learning_rate": 0.00019997095056321971, "loss": 0.4166, "step": 47 }, { "epoch": 0.32989690721649484, "grad_norm": 0.0990329131686243, "learning_rate": 0.00019994835850163924, "loss": 0.4023, "step": 48 }, { "epoch": 0.33676975945017185, "grad_norm": 0.10763088614071568, "learning_rate": 0.00019991931406564944, "loss": 0.4173, "step": 49 }, { "epoch": 0.3436426116838488, "grad_norm": 0.09892826810424134, "learning_rate": 0.00019988381913027442, "loss": 0.3931, "step": 50 }, { "epoch": 0.35051546391752575, "grad_norm": 0.10034165970288787, "learning_rate": 0.00019984187598696363, "loss": 0.3916, "step": 51 }, { "epoch": 0.35738831615120276, "grad_norm": 0.0963929784015873, "learning_rate": 0.00019979348734344398, "loss": 0.3959, "step": 52 }, { "epoch": 0.3642611683848797, "grad_norm": 0.09716524463965583, "learning_rate": 0.00019973865632354516, "loss": 0.3895, "step": 53 }, { "epoch": 0.3711340206185567, "grad_norm": 0.09165795939759437, "learning_rate": 0.0001996773864669978, "loss": 0.397, "step": 54 }, { "epoch": 0.37800687285223367, "grad_norm": 0.09285948678897281, "learning_rate": 0.00019960968172920516, "loss": 0.4075, "step": 55 }, { "epoch": 0.3848797250859107, "grad_norm": 0.09548722575257236, "learning_rate": 0.00019953554648098748, "loss": 0.4013, "step": 56 }, { "epoch": 0.3917525773195876, "grad_norm": 0.09024676675535273, "learning_rate": 0.0001994549855083001, "loss": 0.3915, "step": 57 }, { "epoch": 0.39862542955326463, "grad_norm": 0.092474650175165, "learning_rate": 0.0001993680040119244, "loss": 0.3967, "step": 58 }, { "epoch": 0.4054982817869416, "grad_norm": 0.08989161688777235, "learning_rate": 0.00019927460760713197, "loss": 0.3977, "step": 59 }, { "epoch": 0.41237113402061853, "grad_norm": 0.08919287484996163, "learning_rate": 0.00019917480232332224, "loss": 0.3844, "step": 60 }, { "epoch": 0.41924398625429554, "grad_norm": 0.09181573832814811, "learning_rate": 0.00019906859460363307, "loss": 0.3805, "step": 61 }, { "epoch": 0.4261168384879725, "grad_norm": 0.09806715470399115, "learning_rate": 0.00019895599130452505, "loss": 0.3953, "step": 62 }, { "epoch": 0.4329896907216495, "grad_norm": 0.08993856841618529, "learning_rate": 0.0001988369996953386, "loss": 0.393, "step": 63 }, { "epoch": 0.43986254295532645, "grad_norm": 0.0849484655192697, "learning_rate": 0.00019871162745782478, "loss": 0.3859, "step": 64 }, { "epoch": 0.44673539518900346, "grad_norm": 0.09431581902975165, "learning_rate": 0.00019857988268564953, "loss": 0.3944, "step": 65 }, { "epoch": 0.4536082474226804, "grad_norm": 0.08841348606045497, "learning_rate": 0.0001984417738838709, "loss": 0.3976, "step": 66 }, { "epoch": 0.46048109965635736, "grad_norm": 0.09534816645722197, "learning_rate": 0.0001982973099683902, "loss": 0.3944, "step": 67 }, { "epoch": 0.46735395189003437, "grad_norm": 0.0877563011548879, "learning_rate": 0.0001981465002653763, "loss": 0.381, "step": 68 }, { "epoch": 0.4742268041237113, "grad_norm": 0.0984361061012536, "learning_rate": 0.00019798935451066361, "loss": 0.3993, "step": 69 }, { "epoch": 0.48109965635738833, "grad_norm": 0.09454729070699702, "learning_rate": 0.0001978258828491236, "loss": 0.3823, "step": 70 }, { "epoch": 0.4879725085910653, "grad_norm": 0.08765299270457577, "learning_rate": 0.00019765609583400977, "loss": 0.3806, "step": 71 }, { "epoch": 0.4948453608247423, "grad_norm": 0.09292671023355425, "learning_rate": 0.0001974800044262764, "loss": 0.3748, "step": 72 }, { "epoch": 0.5017182130584192, "grad_norm": 0.10005400473357809, "learning_rate": 0.00019729761999387103, "loss": 0.3783, "step": 73 }, { "epoch": 0.5085910652920962, "grad_norm": 0.09046360550664917, "learning_rate": 0.00019710895431100046, "loss": 0.3892, "step": 74 }, { "epoch": 0.5154639175257731, "grad_norm": 0.10040173176660709, "learning_rate": 0.00019691401955737072, "loss": 0.3812, "step": 75 }, { "epoch": 0.5223367697594502, "grad_norm": 0.09259312688442031, "learning_rate": 0.00019671282831740076, "loss": 0.3825, "step": 76 }, { "epoch": 0.5292096219931272, "grad_norm": 0.09838379897477743, "learning_rate": 0.00019650539357941003, "loss": 0.372, "step": 77 }, { "epoch": 0.5360824742268041, "grad_norm": 0.09020050588363886, "learning_rate": 0.00019629172873477995, "loss": 0.3653, "step": 78 }, { "epoch": 0.5429553264604811, "grad_norm": 0.09409473862649323, "learning_rate": 0.00019607184757708951, "loss": 0.3733, "step": 79 }, { "epoch": 0.5498281786941581, "grad_norm": 0.0904887628828706, "learning_rate": 0.00019584576430122473, "loss": 0.3731, "step": 80 }, { "epoch": 0.5567010309278351, "grad_norm": 0.09889299233384473, "learning_rate": 0.00019561349350246226, "loss": 0.3771, "step": 81 }, { "epoch": 0.563573883161512, "grad_norm": 0.09140051020188593, "learning_rate": 0.00019537505017552716, "loss": 0.3851, "step": 82 }, { "epoch": 0.570446735395189, "grad_norm": 0.09351557402961665, "learning_rate": 0.00019513044971362494, "loss": 0.3669, "step": 83 }, { "epoch": 0.5773195876288659, "grad_norm": 0.0942116546600885, "learning_rate": 0.00019487970790744774, "loss": 0.3737, "step": 84 }, { "epoch": 0.584192439862543, "grad_norm": 0.09048511666973631, "learning_rate": 0.000194622840944155, "loss": 0.376, "step": 85 }, { "epoch": 0.5910652920962199, "grad_norm": 0.09109493630227013, "learning_rate": 0.00019435986540632843, "loss": 0.3833, "step": 86 }, { "epoch": 0.5979381443298969, "grad_norm": 0.0917062224160209, "learning_rate": 0.00019409079827090145, "loss": 0.3765, "step": 87 }, { "epoch": 0.6048109965635738, "grad_norm": 0.087016348439853, "learning_rate": 0.00019381565690806328, "loss": 0.3769, "step": 88 }, { "epoch": 0.6116838487972509, "grad_norm": 0.08529182950833275, "learning_rate": 0.00019353445908013755, "loss": 0.3604, "step": 89 }, { "epoch": 0.6185567010309279, "grad_norm": 0.09163176752655496, "learning_rate": 0.00019324722294043558, "loss": 0.3664, "step": 90 }, { "epoch": 0.6254295532646048, "grad_norm": 0.0893557825896007, "learning_rate": 0.00019295396703208453, "loss": 0.3536, "step": 91 }, { "epoch": 0.6323024054982818, "grad_norm": 0.08880871478677001, "learning_rate": 0.00019265471028683014, "loss": 0.367, "step": 92 }, { "epoch": 0.6391752577319587, "grad_norm": 0.09174194526074796, "learning_rate": 0.00019234947202381486, "loss": 0.3671, "step": 93 }, { "epoch": 0.6460481099656358, "grad_norm": 0.08525381739806115, "learning_rate": 0.00019203827194833026, "loss": 0.3669, "step": 94 }, { "epoch": 0.6529209621993127, "grad_norm": 0.09142431339142947, "learning_rate": 0.00019172113015054532, "loss": 0.3659, "step": 95 }, { "epoch": 0.6597938144329897, "grad_norm": 0.08531323439210271, "learning_rate": 0.00019139806710420914, "loss": 0.3658, "step": 96 }, { "epoch": 0.6666666666666666, "grad_norm": 0.08799641190602743, "learning_rate": 0.00019106910366532942, "loss": 0.3747, "step": 97 }, { "epoch": 0.6735395189003437, "grad_norm": 0.08743449501178935, "learning_rate": 0.000190734261070826, "loss": 0.36, "step": 98 }, { "epoch": 0.6804123711340206, "grad_norm": 0.09113119767243852, "learning_rate": 0.00019039356093715975, "loss": 0.3704, "step": 99 }, { "epoch": 0.6872852233676976, "grad_norm": 0.09180424917650232, "learning_rate": 0.00019004702525893732, "loss": 0.3724, "step": 100 }, { "epoch": 0.6941580756013745, "grad_norm": 0.08318596234383352, "learning_rate": 0.000189694676407491, "loss": 0.3519, "step": 101 }, { "epoch": 0.7010309278350515, "grad_norm": 0.09859998180030405, "learning_rate": 0.0001893365371294346, "loss": 0.3581, "step": 102 }, { "epoch": 0.7079037800687286, "grad_norm": 0.09515518526828363, "learning_rate": 0.00018897263054519498, "loss": 0.3593, "step": 103 }, { "epoch": 0.7147766323024055, "grad_norm": 0.09409102959516591, "learning_rate": 0.00018860298014751944, "loss": 0.3545, "step": 104 }, { "epoch": 0.7216494845360825, "grad_norm": 0.08944322265598344, "learning_rate": 0.0001882276097999592, "loss": 0.3605, "step": 105 }, { "epoch": 0.7285223367697594, "grad_norm": 0.09154417497792565, "learning_rate": 0.00018784654373532866, "loss": 0.3598, "step": 106 }, { "epoch": 0.7353951890034365, "grad_norm": 0.09170643146752601, "learning_rate": 0.00018745980655414114, "loss": 0.3709, "step": 107 }, { "epoch": 0.7422680412371134, "grad_norm": 0.09183342351937492, "learning_rate": 0.00018706742322302064, "loss": 0.3763, "step": 108 }, { "epoch": 0.7491408934707904, "grad_norm": 0.08939732803130215, "learning_rate": 0.00018666941907309026, "loss": 0.3739, "step": 109 }, { "epoch": 0.7560137457044673, "grad_norm": 0.09688998084801166, "learning_rate": 0.0001862658197983366, "loss": 0.366, "step": 110 }, { "epoch": 0.7628865979381443, "grad_norm": 0.09721818711988676, "learning_rate": 0.0001858566514539513, "loss": 0.3532, "step": 111 }, { "epoch": 0.7697594501718213, "grad_norm": 0.08784658510431156, "learning_rate": 0.00018544194045464886, "loss": 0.3547, "step": 112 }, { "epoch": 0.7766323024054983, "grad_norm": 0.09550822822389522, "learning_rate": 0.00018502171357296144, "loss": 0.3693, "step": 113 }, { "epoch": 0.7835051546391752, "grad_norm": 0.08612356641715116, "learning_rate": 0.0001845959979375104, "loss": 0.3503, "step": 114 }, { "epoch": 0.7903780068728522, "grad_norm": 0.08387105339613918, "learning_rate": 0.00018416482103125506, "loss": 0.3486, "step": 115 }, { "epoch": 0.7972508591065293, "grad_norm": 0.09144753629798742, "learning_rate": 0.0001837282106897185, "loss": 0.3679, "step": 116 }, { "epoch": 0.8041237113402062, "grad_norm": 0.08610892743550637, "learning_rate": 0.00018328619509919044, "loss": 0.3685, "step": 117 }, { "epoch": 0.8109965635738832, "grad_norm": 0.08198302051539944, "learning_rate": 0.0001828388027949078, "loss": 0.3512, "step": 118 }, { "epoch": 0.8178694158075601, "grad_norm": 0.09234301966476734, "learning_rate": 0.00018238606265921238, "loss": 0.3521, "step": 119 }, { "epoch": 0.8247422680412371, "grad_norm": 0.08780732306223378, "learning_rate": 0.00018192800391968642, "loss": 0.3646, "step": 120 }, { "epoch": 0.8316151202749141, "grad_norm": 0.09000209673595402, "learning_rate": 0.00018146465614726567, "loss": 0.3583, "step": 121 }, { "epoch": 0.8384879725085911, "grad_norm": 0.09313824941726875, "learning_rate": 0.00018099604925433043, "loss": 0.3625, "step": 122 }, { "epoch": 0.845360824742268, "grad_norm": 0.08718919445098797, "learning_rate": 0.00018052221349277442, "loss": 0.357, "step": 123 }, { "epoch": 0.852233676975945, "grad_norm": 0.08734120489767262, "learning_rate": 0.00018004317945205197, "loss": 0.3553, "step": 124 }, { "epoch": 0.8591065292096219, "grad_norm": 0.08532866925543212, "learning_rate": 0.0001795589780572031, "loss": 0.3474, "step": 125 }, { "epoch": 0.865979381443299, "grad_norm": 0.0948957721406938, "learning_rate": 0.00017906964056685706, "loss": 0.3716, "step": 126 }, { "epoch": 0.872852233676976, "grad_norm": 0.09185305719945758, "learning_rate": 0.00017857519857121458, "loss": 0.365, "step": 127 }, { "epoch": 0.8797250859106529, "grad_norm": 0.08387486486184584, "learning_rate": 0.00017807568399000822, "loss": 0.3731, "step": 128 }, { "epoch": 0.8865979381443299, "grad_norm": 0.08958752134232016, "learning_rate": 0.000177571129070442, "loss": 0.3535, "step": 129 }, { "epoch": 0.8934707903780069, "grad_norm": 0.09017595258335488, "learning_rate": 0.0001770615663851093, "loss": 0.3539, "step": 130 }, { "epoch": 0.9003436426116839, "grad_norm": 0.0913540820915045, "learning_rate": 0.0001765470288298905, "loss": 0.363, "step": 131 }, { "epoch": 0.9072164948453608, "grad_norm": 0.0840320790229592, "learning_rate": 0.0001760275496218288, "loss": 0.3629, "step": 132 }, { "epoch": 0.9140893470790378, "grad_norm": 0.08241622611056106, "learning_rate": 0.0001755031622969862, "loss": 0.3554, "step": 133 }, { "epoch": 0.9209621993127147, "grad_norm": 0.09347547363374605, "learning_rate": 0.00017497390070827848, "loss": 0.3596, "step": 134 }, { "epoch": 0.9278350515463918, "grad_norm": 0.08743537290376564, "learning_rate": 0.00017443979902328956, "loss": 0.3683, "step": 135 }, { "epoch": 0.9347079037800687, "grad_norm": 0.08753141534190809, "learning_rate": 0.00017390089172206592, "loss": 0.363, "step": 136 }, { "epoch": 0.9415807560137457, "grad_norm": 0.09416474808423252, "learning_rate": 0.00017335721359489057, "loss": 0.3638, "step": 137 }, { "epoch": 0.9484536082474226, "grad_norm": 0.08703582006975921, "learning_rate": 0.00017280879974003707, "loss": 0.3635, "step": 138 }, { "epoch": 0.9553264604810997, "grad_norm": 0.08869681149969279, "learning_rate": 0.0001722556855615039, "loss": 0.3704, "step": 139 }, { "epoch": 0.9621993127147767, "grad_norm": 0.08463564854075577, "learning_rate": 0.00017169790676672858, "loss": 0.3527, "step": 140 }, { "epoch": 0.9690721649484536, "grad_norm": 0.08340544460840134, "learning_rate": 0.0001711354993642827, "loss": 0.3353, "step": 141 }, { "epoch": 0.9759450171821306, "grad_norm": 0.083884920865059, "learning_rate": 0.0001705684996615472, "loss": 0.3593, "step": 142 }, { "epoch": 0.9828178694158075, "grad_norm": 0.08488282695769025, "learning_rate": 0.0001699969442623686, "loss": 0.3576, "step": 143 }, { "epoch": 0.9896907216494846, "grad_norm": 0.0884325438987741, "learning_rate": 0.00016942087006469592, "loss": 0.3488, "step": 144 }, { "epoch": 0.9965635738831615, "grad_norm": 0.08900573019638774, "learning_rate": 0.00016884031425819853, "loss": 0.3484, "step": 145 }, { "epoch": 0.9965635738831615, "eval_loss": 0.35606497526168823, "eval_runtime": 34.8675, "eval_samples_per_second": 28.078, "eval_steps_per_second": 0.889, "step": 145 }, { "epoch": 1.0034364261168385, "grad_norm": 0.08196549921566786, "learning_rate": 0.00016825531432186543, "loss": 0.3433, "step": 146 }, { "epoch": 1.0103092783505154, "grad_norm": 0.09480776610698287, "learning_rate": 0.00016766590802158566, "loss": 0.3474, "step": 147 }, { "epoch": 1.0171821305841924, "grad_norm": 0.08982094766748744, "learning_rate": 0.0001670721334077103, "loss": 0.3451, "step": 148 }, { "epoch": 1.0240549828178693, "grad_norm": 0.08398184726912637, "learning_rate": 0.00016647402881259598, "loss": 0.3334, "step": 149 }, { "epoch": 1.0309278350515463, "grad_norm": 0.08328598821509355, "learning_rate": 0.00016587163284813032, "loss": 0.3498, "step": 150 }, { "epoch": 1.0378006872852235, "grad_norm": 0.08677022423834434, "learning_rate": 0.00016526498440323914, "loss": 0.3308, "step": 151 }, { "epoch": 1.0446735395189004, "grad_norm": 0.09149509263746641, "learning_rate": 0.0001646541226413761, "loss": 0.3484, "step": 152 }, { "epoch": 1.0515463917525774, "grad_norm": 0.08566567561323409, "learning_rate": 0.00016403908699799425, "loss": 0.3322, "step": 153 }, { "epoch": 1.0584192439862543, "grad_norm": 0.09092939447329713, "learning_rate": 0.00016341991717800023, "loss": 0.3354, "step": 154 }, { "epoch": 1.0652920962199313, "grad_norm": 0.09591841092657519, "learning_rate": 0.00016279665315319114, "loss": 0.3437, "step": 155 }, { "epoch": 1.0721649484536082, "grad_norm": 0.08459426448823766, "learning_rate": 0.0001621693351596739, "loss": 0.3318, "step": 156 }, { "epoch": 1.0790378006872852, "grad_norm": 0.0898288755251788, "learning_rate": 0.00016153800369526788, "loss": 0.3443, "step": 157 }, { "epoch": 1.0859106529209621, "grad_norm": 0.09386604225118941, "learning_rate": 0.0001609026995168904, "loss": 0.3591, "step": 158 }, { "epoch": 1.0927835051546393, "grad_norm": 0.08825560777370975, "learning_rate": 0.00016026346363792567, "loss": 0.339, "step": 159 }, { "epoch": 1.0996563573883162, "grad_norm": 0.0946307420552798, "learning_rate": 0.00015962033732557686, "loss": 0.3448, "step": 160 }, { "epoch": 1.1065292096219932, "grad_norm": 0.09319504059891445, "learning_rate": 0.00015897336209820239, "loss": 0.3424, "step": 161 }, { "epoch": 1.1134020618556701, "grad_norm": 0.08990355758678513, "learning_rate": 0.00015832257972263523, "loss": 0.3288, "step": 162 }, { "epoch": 1.120274914089347, "grad_norm": 0.09322483505864229, "learning_rate": 0.00015766803221148673, "loss": 0.3477, "step": 163 }, { "epoch": 1.127147766323024, "grad_norm": 0.09272419855505644, "learning_rate": 0.0001570097618204345, "loss": 0.3473, "step": 164 }, { "epoch": 1.134020618556701, "grad_norm": 0.0928011828910675, "learning_rate": 0.00015634781104549442, "loss": 0.3448, "step": 165 }, { "epoch": 1.140893470790378, "grad_norm": 0.08706895131837093, "learning_rate": 0.00015568222262027717, "loss": 0.3272, "step": 166 }, { "epoch": 1.147766323024055, "grad_norm": 0.09258528015337668, "learning_rate": 0.00015501303951322943, "loss": 0.3534, "step": 167 }, { "epoch": 1.1546391752577319, "grad_norm": 0.09009575864515591, "learning_rate": 0.00015434030492486023, "loss": 0.3335, "step": 168 }, { "epoch": 1.161512027491409, "grad_norm": 0.09142107904123438, "learning_rate": 0.00015366406228495172, "loss": 0.3422, "step": 169 }, { "epoch": 1.168384879725086, "grad_norm": 0.0906744126369693, "learning_rate": 0.00015298435524975572, "loss": 0.3367, "step": 170 }, { "epoch": 1.175257731958763, "grad_norm": 0.09418258936552298, "learning_rate": 0.00015230122769917527, "loss": 0.3408, "step": 171 }, { "epoch": 1.1821305841924399, "grad_norm": 0.09024132061907325, "learning_rate": 0.00015161472373393186, "loss": 0.3395, "step": 172 }, { "epoch": 1.1890034364261168, "grad_norm": 0.09712592520855806, "learning_rate": 0.00015092488767271857, "loss": 0.3395, "step": 173 }, { "epoch": 1.1958762886597938, "grad_norm": 0.09145493125421761, "learning_rate": 0.00015023176404933874, "loss": 0.3468, "step": 174 }, { "epoch": 1.2027491408934707, "grad_norm": 0.09050083673868306, "learning_rate": 0.00014953539760983122, "loss": 0.3489, "step": 175 }, { "epoch": 1.2096219931271477, "grad_norm": 0.08871170021022631, "learning_rate": 0.0001488358333095816, "loss": 0.342, "step": 176 }, { "epoch": 1.2164948453608249, "grad_norm": 0.09610489537144708, "learning_rate": 0.00014813311631041995, "loss": 0.3398, "step": 177 }, { "epoch": 1.2233676975945018, "grad_norm": 0.09534796071973317, "learning_rate": 0.00014742729197770552, "loss": 0.3385, "step": 178 }, { "epoch": 1.2302405498281788, "grad_norm": 0.08847329869865773, "learning_rate": 0.00014671840587739783, "loss": 0.3292, "step": 179 }, { "epoch": 1.2371134020618557, "grad_norm": 0.10633331915522143, "learning_rate": 0.00014600650377311522, "loss": 0.3594, "step": 180 }, { "epoch": 1.2439862542955327, "grad_norm": 0.092799430553948, "learning_rate": 0.0001452916316231805, "loss": 0.3407, "step": 181 }, { "epoch": 1.2508591065292096, "grad_norm": 0.09674557611171644, "learning_rate": 0.00014457383557765386, "loss": 0.3477, "step": 182 }, { "epoch": 1.2577319587628866, "grad_norm": 0.09123613793045929, "learning_rate": 0.00014385316197535372, "loss": 0.3349, "step": 183 }, { "epoch": 1.2646048109965635, "grad_norm": 0.09065276134504362, "learning_rate": 0.00014312965734086518, "loss": 0.3389, "step": 184 }, { "epoch": 1.2714776632302405, "grad_norm": 0.0961637917464615, "learning_rate": 0.0001424033683815365, "loss": 0.337, "step": 185 }, { "epoch": 1.2783505154639174, "grad_norm": 0.09687324812097148, "learning_rate": 0.00014167434198446383, "loss": 0.332, "step": 186 }, { "epoch": 1.2852233676975944, "grad_norm": 0.09291726456391641, "learning_rate": 0.00014094262521346427, "loss": 0.3286, "step": 187 }, { "epoch": 1.2920962199312716, "grad_norm": 0.092544941090839, "learning_rate": 0.00014020826530603776, "loss": 0.3341, "step": 188 }, { "epoch": 1.2989690721649485, "grad_norm": 0.09931171836372284, "learning_rate": 0.00013947130967031717, "loss": 0.3372, "step": 189 }, { "epoch": 1.3058419243986255, "grad_norm": 0.09700494142400526, "learning_rate": 0.00013873180588200827, "loss": 0.3437, "step": 190 }, { "epoch": 1.3127147766323024, "grad_norm": 0.09119541578806489, "learning_rate": 0.00013798980168131794, "loss": 0.3433, "step": 191 }, { "epoch": 1.3195876288659794, "grad_norm": 0.08944637651096048, "learning_rate": 0.00013724534496987247, "loss": 0.3207, "step": 192 }, { "epoch": 1.3264604810996563, "grad_norm": 0.0950462508734682, "learning_rate": 0.00013649848380762513, "loss": 0.3452, "step": 193 }, { "epoch": 1.3333333333333333, "grad_norm": 0.09138145448814565, "learning_rate": 0.0001357492664097534, "loss": 0.3334, "step": 194 }, { "epoch": 1.3402061855670104, "grad_norm": 0.08772751482009418, "learning_rate": 0.00013499774114354655, "loss": 0.33, "step": 195 }, { "epoch": 1.3470790378006874, "grad_norm": 0.09409705441370414, "learning_rate": 0.0001342439565252831, "loss": 0.3434, "step": 196 }, { "epoch": 1.3539518900343643, "grad_norm": 0.08673978467409582, "learning_rate": 0.00013348796121709862, "loss": 0.3196, "step": 197 }, { "epoch": 1.3608247422680413, "grad_norm": 0.09064872337097117, "learning_rate": 0.0001327298040238446, "loss": 0.342, "step": 198 }, { "epoch": 1.3676975945017182, "grad_norm": 0.08690156462091585, "learning_rate": 0.00013196953388993726, "loss": 0.3189, "step": 199 }, { "epoch": 1.3745704467353952, "grad_norm": 0.0883092380252361, "learning_rate": 0.00013120719989619833, "loss": 0.3192, "step": 200 }, { "epoch": 1.3814432989690721, "grad_norm": 0.08705036057054863, "learning_rate": 0.00013044285125668614, "loss": 0.3313, "step": 201 }, { "epoch": 1.388316151202749, "grad_norm": 0.08967338120146973, "learning_rate": 0.0001296765373155188, "loss": 0.3328, "step": 202 }, { "epoch": 1.395189003436426, "grad_norm": 0.09514202153989075, "learning_rate": 0.00012890830754368855, "loss": 0.34, "step": 203 }, { "epoch": 1.402061855670103, "grad_norm": 0.09685510036699954, "learning_rate": 0.0001281382115358679, "loss": 0.3302, "step": 204 }, { "epoch": 1.40893470790378, "grad_norm": 0.08977741862683598, "learning_rate": 0.0001273662990072083, "loss": 0.3389, "step": 205 }, { "epoch": 1.4158075601374571, "grad_norm": 0.08557176910720292, "learning_rate": 0.00012659261979013043, "loss": 0.3304, "step": 206 }, { "epoch": 1.422680412371134, "grad_norm": 0.09181328919723801, "learning_rate": 0.00012581722383110718, "loss": 0.3379, "step": 207 }, { "epoch": 1.429553264604811, "grad_norm": 0.09304158736479076, "learning_rate": 0.00012504016118743935, "loss": 0.3448, "step": 208 }, { "epoch": 1.436426116838488, "grad_norm": 0.0894776893437783, "learning_rate": 0.00012426148202402404, "loss": 0.3304, "step": 209 }, { "epoch": 1.443298969072165, "grad_norm": 0.0845928369792518, "learning_rate": 0.00012348123661011601, "loss": 0.329, "step": 210 }, { "epoch": 1.4501718213058419, "grad_norm": 0.08616671483865138, "learning_rate": 0.00012269947531608276, "loss": 0.3359, "step": 211 }, { "epoch": 1.4570446735395188, "grad_norm": 0.09066326423464864, "learning_rate": 0.00012191624861015254, "loss": 0.3465, "step": 212 }, { "epoch": 1.463917525773196, "grad_norm": 0.08698014221698906, "learning_rate": 0.00012113160705515625, "loss": 0.3433, "step": 213 }, { "epoch": 1.470790378006873, "grad_norm": 0.09136539363522148, "learning_rate": 0.0001203456013052634, "loss": 0.3206, "step": 214 }, { "epoch": 1.47766323024055, "grad_norm": 0.09142998051446428, "learning_rate": 0.00011955828210271187, "loss": 0.322, "step": 215 }, { "epoch": 1.4845360824742269, "grad_norm": 0.09497325730270861, "learning_rate": 0.00011876970027453222, "loss": 0.3354, "step": 216 }, { "epoch": 1.4914089347079038, "grad_norm": 0.08907032478482627, "learning_rate": 0.00011797990672926652, "loss": 0.3284, "step": 217 }, { "epoch": 1.4982817869415808, "grad_norm": 0.08960473573204909, "learning_rate": 0.00011718895245368167, "loss": 0.3441, "step": 218 }, { "epoch": 1.5051546391752577, "grad_norm": 0.08696717394605924, "learning_rate": 0.00011639688850947799, "loss": 0.3286, "step": 219 }, { "epoch": 1.5120274914089347, "grad_norm": 0.0863709611120778, "learning_rate": 0.00011560376602999272, "loss": 0.3265, "step": 220 }, { "epoch": 1.5189003436426116, "grad_norm": 0.0871856101257185, "learning_rate": 0.00011480963621689905, "loss": 0.3324, "step": 221 }, { "epoch": 1.5257731958762886, "grad_norm": 0.08759213829704489, "learning_rate": 0.00011401455033690076, "loss": 0.3259, "step": 222 }, { "epoch": 1.5326460481099655, "grad_norm": 0.09024137226684803, "learning_rate": 0.00011321855971842243, "loss": 0.3276, "step": 223 }, { "epoch": 1.5395189003436425, "grad_norm": 0.08692746205690999, "learning_rate": 0.00011242171574829599, "loss": 0.3302, "step": 224 }, { "epoch": 1.5463917525773194, "grad_norm": 0.08752363912789099, "learning_rate": 0.00011162406986844323, "loss": 0.3286, "step": 225 }, { "epoch": 1.5532646048109966, "grad_norm": 0.0869952556307819, "learning_rate": 0.00011082567357255484, "loss": 0.3356, "step": 226 }, { "epoch": 1.5601374570446735, "grad_norm": 0.09252005231748338, "learning_rate": 0.00011002657840276627, "loss": 0.3186, "step": 227 }, { "epoch": 1.5670103092783505, "grad_norm": 0.09038751832017329, "learning_rate": 0.00010922683594633021, "loss": 0.3302, "step": 228 }, { "epoch": 1.5738831615120275, "grad_norm": 0.0873753567344157, "learning_rate": 0.00010842649783228624, "loss": 0.3271, "step": 229 }, { "epoch": 1.5807560137457046, "grad_norm": 0.08573086002729792, "learning_rate": 0.00010762561572812788, "loss": 0.3371, "step": 230 }, { "epoch": 1.5876288659793816, "grad_norm": 0.08886072392248996, "learning_rate": 0.0001068242413364671, "loss": 0.3385, "step": 231 }, { "epoch": 1.5945017182130585, "grad_norm": 0.09071815826585726, "learning_rate": 0.00010602242639169648, "loss": 0.3331, "step": 232 }, { "epoch": 1.6013745704467355, "grad_norm": 0.09259047720970519, "learning_rate": 0.0001052202226566494, "loss": 0.3408, "step": 233 }, { "epoch": 1.6082474226804124, "grad_norm": 0.09141451703161811, "learning_rate": 0.00010441768191925847, "loss": 0.3323, "step": 234 }, { "epoch": 1.6151202749140894, "grad_norm": 0.08745446473269225, "learning_rate": 0.00010361485598921212, "loss": 0.3246, "step": 235 }, { "epoch": 1.6219931271477663, "grad_norm": 0.0901150348236022, "learning_rate": 0.00010281179669461005, "loss": 0.3284, "step": 236 }, { "epoch": 1.6288659793814433, "grad_norm": 0.08512873333856283, "learning_rate": 0.00010200855587861724, "loss": 0.3308, "step": 237 }, { "epoch": 1.6357388316151202, "grad_norm": 0.08840435492240208, "learning_rate": 0.0001012051853961172, "loss": 0.3324, "step": 238 }, { "epoch": 1.6426116838487972, "grad_norm": 0.08581819785864932, "learning_rate": 0.00010040173711036431, "loss": 0.3317, "step": 239 }, { "epoch": 1.6494845360824741, "grad_norm": 0.08542520226750239, "learning_rate": 9.959826288963571e-05, "loss": 0.3326, "step": 240 }, { "epoch": 1.656357388316151, "grad_norm": 0.08568785124185577, "learning_rate": 9.879481460388282e-05, "loss": 0.3309, "step": 241 }, { "epoch": 1.663230240549828, "grad_norm": 0.08493559228806066, "learning_rate": 9.799144412138275e-05, "loss": 0.3433, "step": 242 }, { "epoch": 1.670103092783505, "grad_norm": 0.0927706606246952, "learning_rate": 9.718820330538998e-05, "loss": 0.3384, "step": 243 }, { "epoch": 1.6769759450171822, "grad_norm": 0.08847927125248477, "learning_rate": 9.638514401078788e-05, "loss": 0.3363, "step": 244 }, { "epoch": 1.6838487972508591, "grad_norm": 0.0861526333400457, "learning_rate": 9.558231808074156e-05, "loss": 0.3267, "step": 245 }, { "epoch": 1.690721649484536, "grad_norm": 0.08416812339435613, "learning_rate": 9.477977734335061e-05, "loss": 0.3192, "step": 246 }, { "epoch": 1.697594501718213, "grad_norm": 0.09232813296047053, "learning_rate": 9.397757360830353e-05, "loss": 0.3324, "step": 247 }, { "epoch": 1.7044673539518902, "grad_norm": 0.09150718904790608, "learning_rate": 9.317575866353292e-05, "loss": 0.3306, "step": 248 }, { "epoch": 1.7113402061855671, "grad_norm": 0.08446766316839585, "learning_rate": 9.23743842718721e-05, "loss": 0.3255, "step": 249 }, { "epoch": 1.718213058419244, "grad_norm": 0.08544243696334995, "learning_rate": 9.157350216771378e-05, "loss": 0.3175, "step": 250 }, { "epoch": 1.725085910652921, "grad_norm": 0.09117767233559447, "learning_rate": 9.077316405366981e-05, "loss": 0.3285, "step": 251 }, { "epoch": 1.731958762886598, "grad_norm": 0.0898857039389826, "learning_rate": 8.997342159723371e-05, "loss": 0.3347, "step": 252 }, { "epoch": 1.738831615120275, "grad_norm": 0.08753370082587175, "learning_rate": 8.917432642744518e-05, "loss": 0.3201, "step": 253 }, { "epoch": 1.745704467353952, "grad_norm": 0.09086816195894529, "learning_rate": 8.83759301315568e-05, "loss": 0.3418, "step": 254 }, { "epoch": 1.7525773195876289, "grad_norm": 0.08642157429244211, "learning_rate": 8.757828425170404e-05, "loss": 0.3309, "step": 255 }, { "epoch": 1.7594501718213058, "grad_norm": 0.08764811098838521, "learning_rate": 8.678144028157759e-05, "loss": 0.3259, "step": 256 }, { "epoch": 1.7663230240549828, "grad_norm": 0.08613745660845641, "learning_rate": 8.598544966309925e-05, "loss": 0.3334, "step": 257 }, { "epoch": 1.7731958762886597, "grad_norm": 0.0891943095591871, "learning_rate": 8.519036378310096e-05, "loss": 0.3294, "step": 258 }, { "epoch": 1.7800687285223367, "grad_norm": 0.08878752115347828, "learning_rate": 8.43962339700073e-05, "loss": 0.3336, "step": 259 }, { "epoch": 1.7869415807560136, "grad_norm": 0.08898557508638621, "learning_rate": 8.360311149052205e-05, "loss": 0.3378, "step": 260 }, { "epoch": 1.7938144329896906, "grad_norm": 0.08783526771292885, "learning_rate": 8.281104754631835e-05, "loss": 0.3308, "step": 261 }, { "epoch": 1.8006872852233677, "grad_norm": 0.0860469379949689, "learning_rate": 8.20200932707335e-05, "loss": 0.3267, "step": 262 }, { "epoch": 1.8075601374570447, "grad_norm": 0.0880852007113171, "learning_rate": 8.123029972546781e-05, "loss": 0.318, "step": 263 }, { "epoch": 1.8144329896907216, "grad_norm": 0.08860147425115192, "learning_rate": 8.044171789728816e-05, "loss": 0.3329, "step": 264 }, { "epoch": 1.8213058419243986, "grad_norm": 0.0903494006074923, "learning_rate": 7.965439869473664e-05, "loss": 0.3274, "step": 265 }, { "epoch": 1.8281786941580758, "grad_norm": 0.08903567129909708, "learning_rate": 7.886839294484377e-05, "loss": 0.3236, "step": 266 }, { "epoch": 1.8350515463917527, "grad_norm": 0.08773778151725498, "learning_rate": 7.808375138984745e-05, "loss": 0.325, "step": 267 }, { "epoch": 1.8419243986254297, "grad_norm": 0.08561976416011557, "learning_rate": 7.730052468391725e-05, "loss": 0.3353, "step": 268 }, { "epoch": 1.8487972508591066, "grad_norm": 0.083966948367127, "learning_rate": 7.6518763389884e-05, "loss": 0.3133, "step": 269 }, { "epoch": 1.8556701030927836, "grad_norm": 0.08497881973677633, "learning_rate": 7.573851797597602e-05, "loss": 0.3266, "step": 270 }, { "epoch": 1.8625429553264605, "grad_norm": 0.08580964041061652, "learning_rate": 7.495983881256067e-05, "loss": 0.324, "step": 271 }, { "epoch": 1.8694158075601375, "grad_norm": 0.08889053750453073, "learning_rate": 7.418277616889282e-05, "loss": 0.336, "step": 272 }, { "epoch": 1.8762886597938144, "grad_norm": 0.08973009616819638, "learning_rate": 7.340738020986961e-05, "loss": 0.3246, "step": 273 }, { "epoch": 1.8831615120274914, "grad_norm": 0.08634366628003234, "learning_rate": 7.263370099279172e-05, "loss": 0.3322, "step": 274 }, { "epoch": 1.8900343642611683, "grad_norm": 0.08921445885421862, "learning_rate": 7.186178846413214e-05, "loss": 0.338, "step": 275 }, { "epoch": 1.8969072164948453, "grad_norm": 0.09056566489149884, "learning_rate": 7.109169245631149e-05, "loss": 0.3283, "step": 276 }, { "epoch": 1.9037800687285222, "grad_norm": 0.09143054225370065, "learning_rate": 7.032346268448118e-05, "loss": 0.3328, "step": 277 }, { "epoch": 1.9106529209621992, "grad_norm": 0.08916475575468534, "learning_rate": 6.955714874331387e-05, "loss": 0.3389, "step": 278 }, { "epoch": 1.9175257731958761, "grad_norm": 0.08689028605536298, "learning_rate": 6.87928001038017e-05, "loss": 0.3198, "step": 279 }, { "epoch": 1.9243986254295533, "grad_norm": 0.08721784819232929, "learning_rate": 6.803046611006278e-05, "loss": 0.3287, "step": 280 }, { "epoch": 1.9312714776632303, "grad_norm": 0.0861327292824577, "learning_rate": 6.727019597615545e-05, "loss": 0.3172, "step": 281 }, { "epoch": 1.9381443298969072, "grad_norm": 0.08960018283734845, "learning_rate": 6.651203878290139e-05, "loss": 0.3147, "step": 282 }, { "epoch": 1.9450171821305842, "grad_norm": 0.0859949131259282, "learning_rate": 6.575604347471695e-05, "loss": 0.326, "step": 283 }, { "epoch": 1.9518900343642611, "grad_norm": 0.0899609330948954, "learning_rate": 6.500225885645346e-05, "loss": 0.324, "step": 284 }, { "epoch": 1.9587628865979383, "grad_norm": 0.08780936818528311, "learning_rate": 6.425073359024663e-05, "loss": 0.327, "step": 285 }, { "epoch": 1.9656357388316152, "grad_norm": 0.08893339129332559, "learning_rate": 6.350151619237488e-05, "loss": 0.3421, "step": 286 }, { "epoch": 1.9725085910652922, "grad_norm": 0.08896802523779608, "learning_rate": 6.275465503012751e-05, "loss": 0.3227, "step": 287 }, { "epoch": 1.9793814432989691, "grad_norm": 0.087523152727361, "learning_rate": 6.201019831868208e-05, "loss": 0.3285, "step": 288 }, { "epoch": 1.986254295532646, "grad_norm": 0.08541538675399138, "learning_rate": 6.126819411799175e-05, "loss": 0.3185, "step": 289 }, { "epoch": 1.993127147766323, "grad_norm": 0.09078696627480189, "learning_rate": 6.052869032968285e-05, "loss": 0.3358, "step": 290 }, { "epoch": 2.0, "grad_norm": 0.08553151813043743, "learning_rate": 5.979173469396227e-05, "loss": 0.3202, "step": 291 }, { "epoch": 2.0, "eval_loss": 0.3387638330459595, "eval_runtime": 30.297, "eval_samples_per_second": 32.313, "eval_steps_per_second": 1.023, "step": 291 }, { "epoch": 2.006872852233677, "grad_norm": 0.08858238128409665, "learning_rate": 5.905737478653572e-05, "loss": 0.317, "step": 292 }, { "epoch": 2.013745704467354, "grad_norm": 0.08871142136939832, "learning_rate": 5.83256580155362e-05, "loss": 0.3185, "step": 293 }, { "epoch": 2.020618556701031, "grad_norm": 0.08783036643679949, "learning_rate": 5.7596631618463514e-05, "loss": 0.3071, "step": 294 }, { "epoch": 2.027491408934708, "grad_norm": 0.0870793992748591, "learning_rate": 5.687034265913485e-05, "loss": 0.3091, "step": 295 }, { "epoch": 2.0343642611683848, "grad_norm": 0.08957657180081456, "learning_rate": 5.614683802464631e-05, "loss": 0.3073, "step": 296 }, { "epoch": 2.0412371134020617, "grad_norm": 0.09169488873413334, "learning_rate": 5.542616442234618e-05, "loss": 0.3122, "step": 297 }, { "epoch": 2.0481099656357387, "grad_norm": 0.08794744061942562, "learning_rate": 5.470836837681954e-05, "loss": 0.2972, "step": 298 }, { "epoch": 2.0549828178694156, "grad_norm": 0.08727587338428755, "learning_rate": 5.399349622688479e-05, "loss": 0.3055, "step": 299 }, { "epoch": 2.0618556701030926, "grad_norm": 0.0913122883717448, "learning_rate": 5.32815941226022e-05, "loss": 0.305, "step": 300 }, { "epoch": 2.06872852233677, "grad_norm": 0.09098580933788052, "learning_rate": 5.2572708022294504e-05, "loss": 0.3052, "step": 301 }, { "epoch": 2.075601374570447, "grad_norm": 0.08915793232234448, "learning_rate": 5.1866883689580056e-05, "loss": 0.3066, "step": 302 }, { "epoch": 2.082474226804124, "grad_norm": 0.09096622308376123, "learning_rate": 5.116416669041843e-05, "loss": 0.3102, "step": 303 }, { "epoch": 2.089347079037801, "grad_norm": 0.09116939560588368, "learning_rate": 5.046460239016879e-05, "loss": 0.3063, "step": 304 }, { "epoch": 2.0962199312714778, "grad_norm": 0.09580234555387002, "learning_rate": 4.976823595066128e-05, "loss": 0.3156, "step": 305 }, { "epoch": 2.1030927835051547, "grad_norm": 0.08883597363260062, "learning_rate": 4.907511232728145e-05, "loss": 0.3104, "step": 306 }, { "epoch": 2.1099656357388317, "grad_norm": 0.08821074905983495, "learning_rate": 4.8385276266068146e-05, "loss": 0.315, "step": 307 }, { "epoch": 2.1168384879725086, "grad_norm": 0.09061922572205756, "learning_rate": 4.7698772300824756e-05, "loss": 0.2935, "step": 308 }, { "epoch": 2.1237113402061856, "grad_norm": 0.08924368572192466, "learning_rate": 4.7015644750244306e-05, "loss": 0.2982, "step": 309 }, { "epoch": 2.1305841924398625, "grad_norm": 0.08972847264239034, "learning_rate": 4.6335937715048306e-05, "loss": 0.2931, "step": 310 }, { "epoch": 2.1374570446735395, "grad_norm": 0.09063578898786588, "learning_rate": 4.565969507513981e-05, "loss": 0.3051, "step": 311 }, { "epoch": 2.1443298969072164, "grad_norm": 0.09288697220498662, "learning_rate": 4.498696048677059e-05, "loss": 0.3069, "step": 312 }, { "epoch": 2.1512027491408934, "grad_norm": 0.09200416050627463, "learning_rate": 4.4317777379722866e-05, "loss": 0.2972, "step": 313 }, { "epoch": 2.1580756013745703, "grad_norm": 0.10052000654398745, "learning_rate": 4.365218895450558e-05, "loss": 0.3182, "step": 314 }, { "epoch": 2.1649484536082473, "grad_norm": 0.09538404372680304, "learning_rate": 4.29902381795655e-05, "loss": 0.3091, "step": 315 }, { "epoch": 2.1718213058419242, "grad_norm": 0.091440461220329, "learning_rate": 4.2331967788513295e-05, "loss": 0.309, "step": 316 }, { "epoch": 2.178694158075601, "grad_norm": 0.09320451118652054, "learning_rate": 4.167742027736482e-05, "loss": 0.3011, "step": 317 }, { "epoch": 2.1855670103092786, "grad_norm": 0.09415713175422614, "learning_rate": 4.102663790179764e-05, "loss": 0.2983, "step": 318 }, { "epoch": 2.1924398625429555, "grad_norm": 0.09474272012171304, "learning_rate": 4.037966267442315e-05, "loss": 0.3062, "step": 319 }, { "epoch": 2.1993127147766325, "grad_norm": 0.09608160629906436, "learning_rate": 3.973653636207437e-05, "loss": 0.314, "step": 320 }, { "epoch": 2.2061855670103094, "grad_norm": 0.09248540483276455, "learning_rate": 3.909730048310962e-05, "loss": 0.3086, "step": 321 }, { "epoch": 2.2130584192439864, "grad_norm": 0.09224550748807564, "learning_rate": 3.846199630473216e-05, "loss": 0.2976, "step": 322 }, { "epoch": 2.2199312714776633, "grad_norm": 0.09089249170006114, "learning_rate": 3.7830664840326145e-05, "loss": 0.3121, "step": 323 }, { "epoch": 2.2268041237113403, "grad_norm": 0.09298061525920483, "learning_rate": 3.720334684680889e-05, "loss": 0.3096, "step": 324 }, { "epoch": 2.2336769759450172, "grad_norm": 0.09395581456690184, "learning_rate": 3.6580082821999786e-05, "loss": 0.3105, "step": 325 }, { "epoch": 2.240549828178694, "grad_norm": 0.08928864085223644, "learning_rate": 3.596091300200578e-05, "loss": 0.2912, "step": 326 }, { "epoch": 2.247422680412371, "grad_norm": 0.09093778839066968, "learning_rate": 3.534587735862391e-05, "loss": 0.3136, "step": 327 }, { "epoch": 2.254295532646048, "grad_norm": 0.09160578676115573, "learning_rate": 3.473501559676088e-05, "loss": 0.3258, "step": 328 }, { "epoch": 2.261168384879725, "grad_norm": 0.09189783799811771, "learning_rate": 3.4128367151869714e-05, "loss": 0.3114, "step": 329 }, { "epoch": 2.268041237113402, "grad_norm": 0.09119317661103386, "learning_rate": 3.352597118740404e-05, "loss": 0.3164, "step": 330 }, { "epoch": 2.274914089347079, "grad_norm": 0.09122554724875623, "learning_rate": 3.292786659228973e-05, "loss": 0.3135, "step": 331 }, { "epoch": 2.281786941580756, "grad_norm": 0.09150268699640655, "learning_rate": 3.233409197841437e-05, "loss": 0.3063, "step": 332 }, { "epoch": 2.288659793814433, "grad_norm": 0.09113932444478898, "learning_rate": 3.174468567813461e-05, "loss": 0.2966, "step": 333 }, { "epoch": 2.29553264604811, "grad_norm": 0.09422573121160913, "learning_rate": 3.115968574180149e-05, "loss": 0.3157, "step": 334 }, { "epoch": 2.3024054982817868, "grad_norm": 0.08951733292349671, "learning_rate": 3.0579129935304066e-05, "loss": 0.3049, "step": 335 }, { "epoch": 2.3092783505154637, "grad_norm": 0.09044653189120008, "learning_rate": 3.0003055737631403e-05, "loss": 0.306, "step": 336 }, { "epoch": 2.3161512027491407, "grad_norm": 0.09170697769106743, "learning_rate": 2.9431500338452832e-05, "loss": 0.3122, "step": 337 }, { "epoch": 2.323024054982818, "grad_norm": 0.09275886216547184, "learning_rate": 2.886450063571735e-05, "loss": 0.3151, "step": 338 }, { "epoch": 2.329896907216495, "grad_norm": 0.09547690962435015, "learning_rate": 2.8302093233271453e-05, "loss": 0.3118, "step": 339 }, { "epoch": 2.336769759450172, "grad_norm": 0.09582144684676756, "learning_rate": 2.7744314438496088e-05, "loss": 0.3232, "step": 340 }, { "epoch": 2.343642611683849, "grad_norm": 0.09182783327747177, "learning_rate": 2.7191200259962934e-05, "loss": 0.3037, "step": 341 }, { "epoch": 2.350515463917526, "grad_norm": 0.09139471113236357, "learning_rate": 2.6642786405109475e-05, "loss": 0.3077, "step": 342 }, { "epoch": 2.357388316151203, "grad_norm": 0.09035874101942142, "learning_rate": 2.6099108277934103e-05, "loss": 0.3037, "step": 343 }, { "epoch": 2.3642611683848798, "grad_norm": 0.09021764526352836, "learning_rate": 2.556020097671046e-05, "loss": 0.3004, "step": 344 }, { "epoch": 2.3711340206185567, "grad_norm": 0.0932732899584461, "learning_rate": 2.5026099291721516e-05, "loss": 0.2961, "step": 345 }, { "epoch": 2.3780068728522337, "grad_norm": 0.09148994978232605, "learning_rate": 2.449683770301382e-05, "loss": 0.3079, "step": 346 }, { "epoch": 2.3848797250859106, "grad_norm": 0.09324259218888263, "learning_rate": 2.397245037817125e-05, "loss": 0.3147, "step": 347 }, { "epoch": 2.3917525773195876, "grad_norm": 0.09734475749317849, "learning_rate": 2.345297117010954e-05, "loss": 0.317, "step": 348 }, { "epoch": 2.3986254295532645, "grad_norm": 0.09100711571073897, "learning_rate": 2.2938433614890697e-05, "loss": 0.2986, "step": 349 }, { "epoch": 2.4054982817869415, "grad_norm": 0.0910354069221194, "learning_rate": 2.242887092955801e-05, "loss": 0.3008, "step": 350 }, { "epoch": 2.4123711340206184, "grad_norm": 0.0942509318825381, "learning_rate": 2.1924316009991787e-05, "loss": 0.3135, "step": 351 }, { "epoch": 2.4192439862542954, "grad_norm": 0.09258806539293742, "learning_rate": 2.1424801428785447e-05, "loss": 0.3055, "step": 352 }, { "epoch": 2.4261168384879723, "grad_norm": 0.09319658440067498, "learning_rate": 2.0930359433142932e-05, "loss": 0.307, "step": 353 }, { "epoch": 2.4329896907216497, "grad_norm": 0.09439720176336394, "learning_rate": 2.0441021942796944e-05, "loss": 0.3132, "step": 354 }, { "epoch": 2.4398625429553267, "grad_norm": 0.09361360081337755, "learning_rate": 1.995682054794803e-05, "loss": 0.3193, "step": 355 }, { "epoch": 2.4467353951890036, "grad_norm": 0.0930084945680393, "learning_rate": 1.9477786507225616e-05, "loss": 0.3114, "step": 356 }, { "epoch": 2.4536082474226806, "grad_norm": 0.0908146101447993, "learning_rate": 1.900395074566962e-05, "loss": 0.3163, "step": 357 }, { "epoch": 2.4604810996563575, "grad_norm": 0.09488135432829335, "learning_rate": 1.8535343852734332e-05, "loss": 0.3034, "step": 358 }, { "epoch": 2.4673539518900345, "grad_norm": 0.0898503027420635, "learning_rate": 1.8071996080313602e-05, "loss": 0.31, "step": 359 }, { "epoch": 2.4742268041237114, "grad_norm": 0.09252961562907165, "learning_rate": 1.76139373407876e-05, "loss": 0.3091, "step": 360 }, { "epoch": 2.4810996563573884, "grad_norm": 0.09174502065107969, "learning_rate": 1.7161197205092216e-05, "loss": 0.3085, "step": 361 }, { "epoch": 2.4879725085910653, "grad_norm": 0.09316919210140392, "learning_rate": 1.6713804900809582e-05, "loss": 0.3155, "step": 362 }, { "epoch": 2.4948453608247423, "grad_norm": 0.09099707702810458, "learning_rate": 1.6271789310281517e-05, "loss": 0.2998, "step": 363 }, { "epoch": 2.5017182130584192, "grad_norm": 0.09278684184375971, "learning_rate": 1.583517896874498e-05, "loss": 0.3051, "step": 364 }, { "epoch": 2.508591065292096, "grad_norm": 0.09051739377036547, "learning_rate": 1.540400206248963e-05, "loss": 0.3029, "step": 365 }, { "epoch": 2.515463917525773, "grad_norm": 0.09329916464271287, "learning_rate": 1.4978286427038601e-05, "loss": 0.3094, "step": 366 }, { "epoch": 2.52233676975945, "grad_norm": 0.09148761608373268, "learning_rate": 1.4558059545351143e-05, "loss": 0.304, "step": 367 }, { "epoch": 2.529209621993127, "grad_norm": 0.09421689754467912, "learning_rate": 1.4143348546048707e-05, "loss": 0.3067, "step": 368 }, { "epoch": 2.536082474226804, "grad_norm": 0.09160109557858281, "learning_rate": 1.3734180201663439e-05, "loss": 0.3043, "step": 369 }, { "epoch": 2.542955326460481, "grad_norm": 0.09180788950797147, "learning_rate": 1.3330580926909763e-05, "loss": 0.3077, "step": 370 }, { "epoch": 2.549828178694158, "grad_norm": 0.09372519989473548, "learning_rate": 1.2932576776979377e-05, "loss": 0.3109, "step": 371 }, { "epoch": 2.556701030927835, "grad_norm": 0.09446944007408255, "learning_rate": 1.2540193445858883e-05, "loss": 0.3067, "step": 372 }, { "epoch": 2.563573883161512, "grad_norm": 0.09447559767362776, "learning_rate": 1.2153456264671337e-05, "loss": 0.301, "step": 373 }, { "epoch": 2.5704467353951888, "grad_norm": 0.09090726623042243, "learning_rate": 1.1772390200040817e-05, "loss": 0.3015, "step": 374 }, { "epoch": 2.5773195876288657, "grad_norm": 0.0955835608141768, "learning_rate": 1.139701985248055e-05, "loss": 0.3038, "step": 375 }, { "epoch": 2.584192439862543, "grad_norm": 0.09674510160940122, "learning_rate": 1.1027369454805058e-05, "loss": 0.323, "step": 376 }, { "epoch": 2.59106529209622, "grad_norm": 0.09246843820414014, "learning_rate": 1.0663462870565411e-05, "loss": 0.2965, "step": 377 }, { "epoch": 2.597938144329897, "grad_norm": 0.0911382349041746, "learning_rate": 1.0305323592509009e-05, "loss": 0.309, "step": 378 }, { "epoch": 2.604810996563574, "grad_norm": 0.09028073131867864, "learning_rate": 9.952974741062703e-06, "loss": 0.3052, "step": 379 }, { "epoch": 2.611683848797251, "grad_norm": 0.09305044978342818, "learning_rate": 9.606439062840256e-06, "loss": 0.3109, "step": 380 }, { "epoch": 2.618556701030928, "grad_norm": 0.09355689582069324, "learning_rate": 9.265738929174051e-06, "loss": 0.3119, "step": 381 }, { "epoch": 2.625429553264605, "grad_norm": 0.09182829501492683, "learning_rate": 8.93089633467058e-06, "loss": 0.3049, "step": 382 }, { "epoch": 2.6323024054982818, "grad_norm": 0.09098818342879997, "learning_rate": 8.601932895790877e-06, "loss": 0.2957, "step": 383 }, { "epoch": 2.6391752577319587, "grad_norm": 0.0904906584480114, "learning_rate": 8.278869849454718e-06, "loss": 0.3007, "step": 384 }, { "epoch": 2.6460481099656357, "grad_norm": 0.0905792844368664, "learning_rate": 7.961728051669737e-06, "loss": 0.3051, "step": 385 }, { "epoch": 2.6529209621993126, "grad_norm": 0.09053848941318468, "learning_rate": 7.650527976185173e-06, "loss": 0.3024, "step": 386 }, { "epoch": 2.6597938144329896, "grad_norm": 0.095071037340431, "learning_rate": 7.3452897131698564e-06, "loss": 0.3024, "step": 387 }, { "epoch": 2.6666666666666665, "grad_norm": 0.09101793743260775, "learning_rate": 7.046032967915483e-06, "loss": 0.3063, "step": 388 }, { "epoch": 2.673539518900344, "grad_norm": 0.09513393631374653, "learning_rate": 6.75277705956443e-06, "loss": 0.3061, "step": 389 }, { "epoch": 2.680412371134021, "grad_norm": 0.09138433427685168, "learning_rate": 6.465540919862456e-06, "loss": 0.3011, "step": 390 }, { "epoch": 2.687285223367698, "grad_norm": 0.09331732527586893, "learning_rate": 6.184343091936751e-06, "loss": 0.3104, "step": 391 }, { "epoch": 2.6941580756013748, "grad_norm": 0.09509119959060863, "learning_rate": 5.909201729098579e-06, "loss": 0.3007, "step": 392 }, { "epoch": 2.7010309278350517, "grad_norm": 0.09381700531573253, "learning_rate": 5.640134593671598e-06, "loss": 0.3078, "step": 393 }, { "epoch": 2.7079037800687287, "grad_norm": 0.09280079893783877, "learning_rate": 5.3771590558450265e-06, "loss": 0.3104, "step": 394 }, { "epoch": 2.7147766323024056, "grad_norm": 0.09333406593374688, "learning_rate": 5.12029209255227e-06, "loss": 0.3012, "step": 395 }, { "epoch": 2.7216494845360826, "grad_norm": 0.09105023068223807, "learning_rate": 4.869550286375091e-06, "loss": 0.3092, "step": 396 }, { "epoch": 2.7285223367697595, "grad_norm": 0.09463942585438903, "learning_rate": 4.624949824472858e-06, "loss": 0.3115, "step": 397 }, { "epoch": 2.7353951890034365, "grad_norm": 0.09430754898630415, "learning_rate": 4.386506497537757e-06, "loss": 0.3061, "step": 398 }, { "epoch": 2.7422680412371134, "grad_norm": 0.09352429892925276, "learning_rate": 4.154235698775277e-06, "loss": 0.3074, "step": 399 }, { "epoch": 2.7491408934707904, "grad_norm": 0.09556948713182102, "learning_rate": 3.928152422910491e-06, "loss": 0.3162, "step": 400 }, { "epoch": 2.7560137457044673, "grad_norm": 0.09326761578566435, "learning_rate": 3.7082712652200867e-06, "loss": 0.3068, "step": 401 }, { "epoch": 2.7628865979381443, "grad_norm": 0.09134209603164739, "learning_rate": 3.4946064205899965e-06, "loss": 0.2996, "step": 402 }, { "epoch": 2.7697594501718212, "grad_norm": 0.09444902133595975, "learning_rate": 3.287171682599255e-06, "loss": 0.3081, "step": 403 }, { "epoch": 2.776632302405498, "grad_norm": 0.09279018937918222, "learning_rate": 3.085980442629288e-06, "loss": 0.3059, "step": 404 }, { "epoch": 2.783505154639175, "grad_norm": 0.08924896615534708, "learning_rate": 2.8910456889995498e-06, "loss": 0.2972, "step": 405 }, { "epoch": 2.790378006872852, "grad_norm": 0.09425301008285972, "learning_rate": 2.7023800061289907e-06, "loss": 0.3098, "step": 406 }, { "epoch": 2.797250859106529, "grad_norm": 0.09239675778317553, "learning_rate": 2.5199955737236104e-06, "loss": 0.3044, "step": 407 }, { "epoch": 2.804123711340206, "grad_norm": 0.09184163394359819, "learning_rate": 2.3439041659902407e-06, "loss": 0.3097, "step": 408 }, { "epoch": 2.810996563573883, "grad_norm": 0.09226772961301682, "learning_rate": 2.174117150876398e-06, "loss": 0.3048, "step": 409 }, { "epoch": 2.81786941580756, "grad_norm": 0.09219190526437263, "learning_rate": 2.010645489336382e-06, "loss": 0.3035, "step": 410 }, { "epoch": 2.824742268041237, "grad_norm": 0.09215977672671191, "learning_rate": 1.8534997346237093e-06, "loss": 0.3003, "step": 411 }, { "epoch": 2.8316151202749142, "grad_norm": 0.09424915722583155, "learning_rate": 1.7026900316098215e-06, "loss": 0.3191, "step": 412 }, { "epoch": 2.838487972508591, "grad_norm": 0.09036000251671127, "learning_rate": 1.5582261161291245e-06, "loss": 0.3042, "step": 413 }, { "epoch": 2.845360824742268, "grad_norm": 0.09293334858564611, "learning_rate": 1.4201173143504888e-06, "loss": 0.3014, "step": 414 }, { "epoch": 2.852233676975945, "grad_norm": 0.09566829239969535, "learning_rate": 1.2883725421752201e-06, "loss": 0.307, "step": 415 }, { "epoch": 2.859106529209622, "grad_norm": 0.09518964464057635, "learning_rate": 1.1630003046614323e-06, "loss": 0.3128, "step": 416 }, { "epoch": 2.865979381443299, "grad_norm": 0.09407646982509986, "learning_rate": 1.0440086954749517e-06, "loss": 0.3054, "step": 417 }, { "epoch": 2.872852233676976, "grad_norm": 0.09089170678998365, "learning_rate": 9.314053963669245e-07, "loss": 0.3085, "step": 418 }, { "epoch": 2.879725085910653, "grad_norm": 0.09244683910430469, "learning_rate": 8.251976766777913e-07, "loss": 0.308, "step": 419 }, { "epoch": 2.88659793814433, "grad_norm": 0.0916294208696784, "learning_rate": 7.253923928680406e-07, "loss": 0.3068, "step": 420 }, { "epoch": 2.893470790378007, "grad_norm": 0.09049174926564073, "learning_rate": 6.319959880756177e-07, "loss": 0.2977, "step": 421 }, { "epoch": 2.9003436426116838, "grad_norm": 0.09225040195333954, "learning_rate": 5.450144916999134e-07, "loss": 0.3065, "step": 422 }, { "epoch": 2.9072164948453607, "grad_norm": 0.09153019196126938, "learning_rate": 4.644535190125421e-07, "loss": 0.3048, "step": 423 }, { "epoch": 2.9140893470790377, "grad_norm": 0.09184602643946774, "learning_rate": 3.903182707948649e-07, "loss": 0.3082, "step": 424 }, { "epoch": 2.9209621993127146, "grad_norm": 0.09071782265424343, "learning_rate": 3.2261353300219176e-07, "loss": 0.2976, "step": 425 }, { "epoch": 2.927835051546392, "grad_norm": 0.09421947114150792, "learning_rate": 2.613436764548505e-07, "loss": 0.2974, "step": 426 }, { "epoch": 2.934707903780069, "grad_norm": 0.0895546126126861, "learning_rate": 2.0651265655603492e-07, "loss": 0.3021, "step": 427 }, { "epoch": 2.941580756013746, "grad_norm": 0.09136227060021743, "learning_rate": 1.5812401303639813e-07, "loss": 0.3014, "step": 428 }, { "epoch": 2.948453608247423, "grad_norm": 0.09222786580967714, "learning_rate": 1.1618086972559062e-07, "loss": 0.2982, "step": 429 }, { "epoch": 2.9553264604811, "grad_norm": 0.09149223979418855, "learning_rate": 8.068593435055505e-08, "loss": 0.311, "step": 430 }, { "epoch": 2.9621993127147768, "grad_norm": 0.09088402733520035, "learning_rate": 5.164149836077714e-08, "loss": 0.305, "step": 431 }, { "epoch": 2.9690721649484537, "grad_norm": 0.09451653562403552, "learning_rate": 2.9049436780281825e-08, "loss": 0.3039, "step": 432 }, { "epoch": 2.9759450171821307, "grad_norm": 0.09212740196617852, "learning_rate": 1.2911208086663351e-08, "loss": 0.3011, "step": 433 }, { "epoch": 2.9828178694158076, "grad_norm": 0.09458249212474722, "learning_rate": 3.2278541168717646e-09, "loss": 0.308, "step": 434 }, { "epoch": 2.9896907216494846, "grad_norm": 0.09670394082118725, "learning_rate": 0.0, "loss": 0.3107, "step": 435 }, { "epoch": 2.9896907216494846, "eval_loss": 0.3359842896461487, "eval_runtime": 29.9218, "eval_samples_per_second": 32.719, "eval_steps_per_second": 1.036, "step": 435 }, { "epoch": 2.9896907216494846, "step": 435, "total_flos": 1.5613530376975155e+17, "train_loss": 0.36922966947500735, "train_runtime": 4905.4793, "train_samples_per_second": 11.374, "train_steps_per_second": 0.089 } ], "logging_steps": 1, "max_steps": 435, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.5613530376975155e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }