{ "best_metric": 0.79888368, "best_model_checkpoint": "/mnt/nas1/daoze/code/swift/output/llava1_6-llama3_1-8b-instruct-my/v33-20240901-191352/checkpoint-4200", "epoch": 1.9995860070378804, "eval_steps": 300, "global_step": 4830, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00041399296211964395, "grad_norm": 68.31517175098004, "learning_rate": 0.0, "loss": 7.09566689, "memory(GiB)": 42.23, "step": 1, "train_speed(iter/s)": 0.013537 }, { "epoch": 0.0020699648105982197, "grad_norm": 75.27486731705963, "learning_rate": 3.2339240870284233e-06, "loss": 4.4965229, "memory(GiB)": 48.78, "step": 5, "train_speed(iter/s)": 0.018721 }, { "epoch": 0.004139929621196439, "grad_norm": 7.7589717582741615, "learning_rate": 4.626699381900465e-06, "loss": 1.79170246, "memory(GiB)": 52.1, "step": 10, "train_speed(iter/s)": 0.01954 }, { "epoch": 0.00620989443179466, "grad_norm": 4.8519342782653485, "learning_rate": 5.44142070133146e-06, "loss": 1.36023655, "memory(GiB)": 52.1, "step": 15, "train_speed(iter/s)": 0.019784 }, { "epoch": 0.008279859242392879, "grad_norm": 3.375818313593022, "learning_rate": 6.0194746767725065e-06, "loss": 1.18811779, "memory(GiB)": 52.1, "step": 20, "train_speed(iter/s)": 0.020008 }, { "epoch": 0.0103498240529911, "grad_norm": 3.5284378077706813, "learning_rate": 6.467848174056847e-06, "loss": 1.1600296, "memory(GiB)": 52.1, "step": 25, "train_speed(iter/s)": 0.020192 }, { "epoch": 0.01241978886358932, "grad_norm": 3.163950589670924, "learning_rate": 6.834195996203502e-06, "loss": 1.11185207, "memory(GiB)": 52.1, "step": 30, "train_speed(iter/s)": 0.020268 }, { "epoch": 0.014489753674187538, "grad_norm": 2.212271768126557, "learning_rate": 7.143938666407679e-06, "loss": 1.0538393, "memory(GiB)": 52.1, "step": 35, "train_speed(iter/s)": 0.020328 }, { "epoch": 0.016559718484785758, "grad_norm": 3.493503288451801, "learning_rate": 7.412249971644547e-06, "loss": 1.04750004, "memory(GiB)": 52.1, "step": 40, "train_speed(iter/s)": 0.020393 }, { "epoch": 0.01862968329538398, "grad_norm": 2.6291679253427644, "learning_rate": 7.648917315634497e-06, "loss": 1.04756851, "memory(GiB)": 52.1, "step": 45, "train_speed(iter/s)": 0.020444 }, { "epoch": 0.0206996481059822, "grad_norm": 3.1377476353112312, "learning_rate": 7.860623468928888e-06, "loss": 1.00407467, "memory(GiB)": 52.1, "step": 50, "train_speed(iter/s)": 0.02043 }, { "epoch": 0.022769612916580417, "grad_norm": 2.6547100112946733, "learning_rate": 8.05213497976565e-06, "loss": 1.01097145, "memory(GiB)": 52.1, "step": 55, "train_speed(iter/s)": 0.020407 }, { "epoch": 0.02483957772717864, "grad_norm": 2.6383576740248555, "learning_rate": 8.226971291075542e-06, "loss": 0.98796005, "memory(GiB)": 52.1, "step": 60, "train_speed(iter/s)": 0.020419 }, { "epoch": 0.02690954253777686, "grad_norm": 2.542613928151568, "learning_rate": 8.387805106618597e-06, "loss": 0.98850031, "memory(GiB)": 52.1, "step": 65, "train_speed(iter/s)": 0.020424 }, { "epoch": 0.028979507348375077, "grad_norm": 2.877488922903144, "learning_rate": 8.536713961279723e-06, "loss": 0.98919926, "memory(GiB)": 52.1, "step": 70, "train_speed(iter/s)": 0.020439 }, { "epoch": 0.031049472158973298, "grad_norm": 1.8237010588157114, "learning_rate": 8.675344788359883e-06, "loss": 1.012813, "memory(GiB)": 52.1, "step": 75, "train_speed(iter/s)": 0.020466 }, { "epoch": 0.033119436969571515, "grad_norm": 2.5261639534973788, "learning_rate": 8.805025266516589e-06, "loss": 0.98800411, "memory(GiB)": 52.1, "step": 80, "train_speed(iter/s)": 0.02047 }, { "epoch": 0.035189401780169736, "grad_norm": 2.2711243603226032, "learning_rate": 8.926841351029377e-06, "loss": 1.00139637, "memory(GiB)": 55.52, "step": 85, "train_speed(iter/s)": 0.020479 }, { "epoch": 0.03725936659076796, "grad_norm": 2.145948654847171, "learning_rate": 9.041692610506539e-06, "loss": 0.99127426, "memory(GiB)": 55.52, "step": 90, "train_speed(iter/s)": 0.020491 }, { "epoch": 0.03932933140136618, "grad_norm": 2.836740011107602, "learning_rate": 9.150332582159872e-06, "loss": 0.96552677, "memory(GiB)": 55.52, "step": 95, "train_speed(iter/s)": 0.020506 }, { "epoch": 0.0413992962119644, "grad_norm": 2.2398611362766006, "learning_rate": 9.25339876380093e-06, "loss": 0.98038893, "memory(GiB)": 55.52, "step": 100, "train_speed(iter/s)": 0.020501 }, { "epoch": 0.043469261022562614, "grad_norm": 2.4986121497854192, "learning_rate": 9.351435280710716e-06, "loss": 0.9604641, "memory(GiB)": 63.35, "step": 105, "train_speed(iter/s)": 0.020504 }, { "epoch": 0.045539225833160835, "grad_norm": 2.1909523678523755, "learning_rate": 9.444910274637691e-06, "loss": 0.93479166, "memory(GiB)": 63.35, "step": 110, "train_speed(iter/s)": 0.020519 }, { "epoch": 0.047609190643759056, "grad_norm": 2.5301628247983663, "learning_rate": 9.534229424247679e-06, "loss": 0.97118149, "memory(GiB)": 63.35, "step": 115, "train_speed(iter/s)": 0.020522 }, { "epoch": 0.04967915545435728, "grad_norm": 1.92102667189958, "learning_rate": 9.619746585947584e-06, "loss": 0.97038708, "memory(GiB)": 63.35, "step": 120, "train_speed(iter/s)": 0.020545 }, { "epoch": 0.0517491202649555, "grad_norm": 2.2968578337197036, "learning_rate": 9.701772261085271e-06, "loss": 0.96241703, "memory(GiB)": 63.35, "step": 125, "train_speed(iter/s)": 0.020544 }, { "epoch": 0.05381908507555372, "grad_norm": 2.110866882726078, "learning_rate": 9.780580401490638e-06, "loss": 0.95122089, "memory(GiB)": 63.35, "step": 130, "train_speed(iter/s)": 0.020551 }, { "epoch": 0.05588904988615193, "grad_norm": 2.082289679777962, "learning_rate": 9.856413929937534e-06, "loss": 0.95173302, "memory(GiB)": 63.35, "step": 135, "train_speed(iter/s)": 0.020557 }, { "epoch": 0.057959014696750154, "grad_norm": 2.6318522567639784, "learning_rate": 9.929489256151762e-06, "loss": 0.94850836, "memory(GiB)": 63.35, "step": 140, "train_speed(iter/s)": 0.020559 }, { "epoch": 0.060028979507348375, "grad_norm": 2.091985474615185, "learning_rate": 1e-05, "loss": 0.91576376, "memory(GiB)": 63.35, "step": 145, "train_speed(iter/s)": 0.020557 }, { "epoch": 0.062098944317946596, "grad_norm": 2.3841665763842705, "learning_rate": 9.991462113127002e-06, "loss": 0.93616524, "memory(GiB)": 63.35, "step": 150, "train_speed(iter/s)": 0.020563 }, { "epoch": 0.06416890912854481, "grad_norm": 2.1042524174231505, "learning_rate": 9.980789754535753e-06, "loss": 0.9277298, "memory(GiB)": 63.35, "step": 155, "train_speed(iter/s)": 0.020558 }, { "epoch": 0.06623887393914303, "grad_norm": 2.276077986025351, "learning_rate": 9.970117395944504e-06, "loss": 0.95900288, "memory(GiB)": 63.35, "step": 160, "train_speed(iter/s)": 0.020556 }, { "epoch": 0.06830883874974125, "grad_norm": 2.2459536201915387, "learning_rate": 9.959445037353256e-06, "loss": 0.94394236, "memory(GiB)": 63.35, "step": 165, "train_speed(iter/s)": 0.020552 }, { "epoch": 0.07037880356033947, "grad_norm": 2.517051668755459, "learning_rate": 9.948772678762007e-06, "loss": 0.94310379, "memory(GiB)": 63.35, "step": 170, "train_speed(iter/s)": 0.020552 }, { "epoch": 0.0724487683709377, "grad_norm": 2.2026553318124313, "learning_rate": 9.938100320170759e-06, "loss": 0.95436573, "memory(GiB)": 63.35, "step": 175, "train_speed(iter/s)": 0.020554 }, { "epoch": 0.07451873318153591, "grad_norm": 2.2476130486911465, "learning_rate": 9.92742796157951e-06, "loss": 0.93582458, "memory(GiB)": 63.35, "step": 180, "train_speed(iter/s)": 0.020553 }, { "epoch": 0.07658869799213414, "grad_norm": 2.388512026493847, "learning_rate": 9.916755602988262e-06, "loss": 0.91813745, "memory(GiB)": 63.35, "step": 185, "train_speed(iter/s)": 0.020558 }, { "epoch": 0.07865866280273236, "grad_norm": 2.3220003726970204, "learning_rate": 9.906083244397012e-06, "loss": 0.92115765, "memory(GiB)": 63.35, "step": 190, "train_speed(iter/s)": 0.020561 }, { "epoch": 0.08072862761333058, "grad_norm": 2.3768930570990805, "learning_rate": 9.895410885805764e-06, "loss": 0.91676846, "memory(GiB)": 63.35, "step": 195, "train_speed(iter/s)": 0.020562 }, { "epoch": 0.0827985924239288, "grad_norm": 4.2054479128966165, "learning_rate": 9.884738527214515e-06, "loss": 0.91852398, "memory(GiB)": 63.35, "step": 200, "train_speed(iter/s)": 0.020566 }, { "epoch": 0.084868557234527, "grad_norm": 2.14133807915419, "learning_rate": 9.874066168623266e-06, "loss": 0.91718044, "memory(GiB)": 63.35, "step": 205, "train_speed(iter/s)": 0.020563 }, { "epoch": 0.08693852204512523, "grad_norm": 1.9390816379786193, "learning_rate": 9.863393810032017e-06, "loss": 0.92340775, "memory(GiB)": 63.35, "step": 210, "train_speed(iter/s)": 0.020559 }, { "epoch": 0.08900848685572345, "grad_norm": 2.1261095243289208, "learning_rate": 9.852721451440769e-06, "loss": 0.90159931, "memory(GiB)": 63.35, "step": 215, "train_speed(iter/s)": 0.020561 }, { "epoch": 0.09107845166632167, "grad_norm": 2.9284829291227243, "learning_rate": 9.842049092849521e-06, "loss": 0.91123104, "memory(GiB)": 63.35, "step": 220, "train_speed(iter/s)": 0.020557 }, { "epoch": 0.09314841647691989, "grad_norm": 2.7049739018013583, "learning_rate": 9.831376734258272e-06, "loss": 0.941084, "memory(GiB)": 63.35, "step": 225, "train_speed(iter/s)": 0.02056 }, { "epoch": 0.09521838128751811, "grad_norm": 2.0467249315549845, "learning_rate": 9.820704375667023e-06, "loss": 0.90574436, "memory(GiB)": 63.35, "step": 230, "train_speed(iter/s)": 0.020561 }, { "epoch": 0.09728834609811633, "grad_norm": 2.1061950654127006, "learning_rate": 9.810032017075774e-06, "loss": 0.92933855, "memory(GiB)": 63.35, "step": 235, "train_speed(iter/s)": 0.020568 }, { "epoch": 0.09935831090871455, "grad_norm": 2.390745868374031, "learning_rate": 9.799359658484527e-06, "loss": 0.9383173, "memory(GiB)": 63.35, "step": 240, "train_speed(iter/s)": 0.02057 }, { "epoch": 0.10142827571931277, "grad_norm": 1.9948725536279355, "learning_rate": 9.788687299893276e-06, "loss": 0.85975437, "memory(GiB)": 63.35, "step": 245, "train_speed(iter/s)": 0.020577 }, { "epoch": 0.103498240529911, "grad_norm": 2.29936461247775, "learning_rate": 9.77801494130203e-06, "loss": 0.90683613, "memory(GiB)": 63.47, "step": 250, "train_speed(iter/s)": 0.020575 }, { "epoch": 0.10556820534050922, "grad_norm": 1.8961456174283475, "learning_rate": 9.76734258271078e-06, "loss": 0.88947477, "memory(GiB)": 63.47, "step": 255, "train_speed(iter/s)": 0.020574 }, { "epoch": 0.10763817015110744, "grad_norm": 1.8558977060217532, "learning_rate": 9.756670224119531e-06, "loss": 0.90292645, "memory(GiB)": 63.47, "step": 260, "train_speed(iter/s)": 0.020581 }, { "epoch": 0.10970813496170564, "grad_norm": 2.1017797656853725, "learning_rate": 9.745997865528282e-06, "loss": 0.92776756, "memory(GiB)": 63.47, "step": 265, "train_speed(iter/s)": 0.020579 }, { "epoch": 0.11177809977230387, "grad_norm": 1.9051226867861688, "learning_rate": 9.735325506937033e-06, "loss": 0.90565796, "memory(GiB)": 63.59, "step": 270, "train_speed(iter/s)": 0.020578 }, { "epoch": 0.11384806458290209, "grad_norm": 1.932217719000402, "learning_rate": 9.724653148345784e-06, "loss": 0.90403481, "memory(GiB)": 63.59, "step": 275, "train_speed(iter/s)": 0.020575 }, { "epoch": 0.11591802939350031, "grad_norm": 3.033913485092789, "learning_rate": 9.713980789754537e-06, "loss": 0.86916351, "memory(GiB)": 63.59, "step": 280, "train_speed(iter/s)": 0.020575 }, { "epoch": 0.11798799420409853, "grad_norm": 1.9434774416915237, "learning_rate": 9.703308431163288e-06, "loss": 0.87491503, "memory(GiB)": 63.59, "step": 285, "train_speed(iter/s)": 0.020569 }, { "epoch": 0.12005795901469675, "grad_norm": 2.051766135268311, "learning_rate": 9.69263607257204e-06, "loss": 0.89817352, "memory(GiB)": 63.59, "step": 290, "train_speed(iter/s)": 0.020572 }, { "epoch": 0.12212792382529497, "grad_norm": 2.2615047465793796, "learning_rate": 9.68196371398079e-06, "loss": 0.89681797, "memory(GiB)": 63.59, "step": 295, "train_speed(iter/s)": 0.020574 }, { "epoch": 0.12419788863589319, "grad_norm": 2.3348971307519637, "learning_rate": 9.671291355389541e-06, "loss": 0.91658554, "memory(GiB)": 63.59, "step": 300, "train_speed(iter/s)": 0.020576 }, { "epoch": 0.12419788863589319, "eval_loss": 0.9112715721130371, "eval_runtime": 338.0523, "eval_samples_per_second": 18.476, "eval_steps_per_second": 1.157, "step": 300 }, { "epoch": 0.1262678534464914, "grad_norm": 2.0234892490546614, "learning_rate": 9.660618996798294e-06, "loss": 0.89021435, "memory(GiB)": 63.59, "step": 305, "train_speed(iter/s)": 0.02005 }, { "epoch": 0.12833781825708962, "grad_norm": 2.3363054707830195, "learning_rate": 9.649946638207045e-06, "loss": 0.87254162, "memory(GiB)": 63.59, "step": 310, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.13040778306768785, "grad_norm": 1.820741232007416, "learning_rate": 9.639274279615796e-06, "loss": 0.90060663, "memory(GiB)": 63.59, "step": 315, "train_speed(iter/s)": 0.020066 }, { "epoch": 0.13247774787828606, "grad_norm": 1.9084755504752218, "learning_rate": 9.628601921024547e-06, "loss": 0.8869771, "memory(GiB)": 63.59, "step": 320, "train_speed(iter/s)": 0.020075 }, { "epoch": 0.1345477126888843, "grad_norm": 1.8856515103808584, "learning_rate": 9.617929562433298e-06, "loss": 0.87808056, "memory(GiB)": 63.59, "step": 325, "train_speed(iter/s)": 0.020084 }, { "epoch": 0.1366176774994825, "grad_norm": 1.9384672045466198, "learning_rate": 9.60725720384205e-06, "loss": 0.88911896, "memory(GiB)": 63.59, "step": 330, "train_speed(iter/s)": 0.020095 }, { "epoch": 0.13868764231008074, "grad_norm": 1.9948177737503383, "learning_rate": 9.596584845250802e-06, "loss": 0.8958828, "memory(GiB)": 63.59, "step": 335, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.14075760712067895, "grad_norm": 2.2317608749634874, "learning_rate": 9.585912486659551e-06, "loss": 0.89190845, "memory(GiB)": 63.59, "step": 340, "train_speed(iter/s)": 0.020112 }, { "epoch": 0.14282757193127718, "grad_norm": 2.1298996429991908, "learning_rate": 9.575240128068304e-06, "loss": 0.87826462, "memory(GiB)": 63.59, "step": 345, "train_speed(iter/s)": 0.02012 }, { "epoch": 0.1448975367418754, "grad_norm": 1.9367356654913552, "learning_rate": 9.564567769477055e-06, "loss": 0.88935146, "memory(GiB)": 63.59, "step": 350, "train_speed(iter/s)": 0.020128 }, { "epoch": 0.1469675015524736, "grad_norm": 2.965286627679984, "learning_rate": 9.553895410885806e-06, "loss": 0.91815538, "memory(GiB)": 63.59, "step": 355, "train_speed(iter/s)": 0.020135 }, { "epoch": 0.14903746636307183, "grad_norm": 2.322705848413725, "learning_rate": 9.543223052294557e-06, "loss": 0.89534588, "memory(GiB)": 63.59, "step": 360, "train_speed(iter/s)": 0.020143 }, { "epoch": 0.15110743117367004, "grad_norm": 1.9467115924709606, "learning_rate": 9.53255069370331e-06, "loss": 0.87547607, "memory(GiB)": 63.59, "step": 365, "train_speed(iter/s)": 0.020146 }, { "epoch": 0.15317739598426827, "grad_norm": 2.056244423727218, "learning_rate": 9.521878335112061e-06, "loss": 0.87522736, "memory(GiB)": 63.59, "step": 370, "train_speed(iter/s)": 0.02015 }, { "epoch": 0.15524736079486648, "grad_norm": 1.9570323722204157, "learning_rate": 9.511205976520812e-06, "loss": 0.88565502, "memory(GiB)": 63.59, "step": 375, "train_speed(iter/s)": 0.020154 }, { "epoch": 0.15731732560546471, "grad_norm": 1.9075809920338722, "learning_rate": 9.500533617929563e-06, "loss": 0.87662697, "memory(GiB)": 63.59, "step": 380, "train_speed(iter/s)": 0.020159 }, { "epoch": 0.15938729041606292, "grad_norm": 2.3674969791857983, "learning_rate": 9.489861259338314e-06, "loss": 0.88602619, "memory(GiB)": 63.59, "step": 385, "train_speed(iter/s)": 0.020163 }, { "epoch": 0.16145725522666116, "grad_norm": 2.157304860709474, "learning_rate": 9.479188900747067e-06, "loss": 0.86379642, "memory(GiB)": 63.59, "step": 390, "train_speed(iter/s)": 0.02017 }, { "epoch": 0.16352722003725936, "grad_norm": 1.8795995948139297, "learning_rate": 9.468516542155816e-06, "loss": 0.88137684, "memory(GiB)": 63.59, "step": 395, "train_speed(iter/s)": 0.020176 }, { "epoch": 0.1655971848478576, "grad_norm": 1.8996836205094734, "learning_rate": 9.457844183564569e-06, "loss": 0.86949444, "memory(GiB)": 63.59, "step": 400, "train_speed(iter/s)": 0.020182 }, { "epoch": 0.1676671496584558, "grad_norm": 2.0977604679128854, "learning_rate": 9.44717182497332e-06, "loss": 0.85557442, "memory(GiB)": 63.59, "step": 405, "train_speed(iter/s)": 0.020184 }, { "epoch": 0.169737114469054, "grad_norm": 2.2253818762342155, "learning_rate": 9.436499466382071e-06, "loss": 0.85497751, "memory(GiB)": 63.59, "step": 410, "train_speed(iter/s)": 0.02019 }, { "epoch": 0.17180707927965225, "grad_norm": 1.9007105346828383, "learning_rate": 9.425827107790822e-06, "loss": 0.86072025, "memory(GiB)": 63.59, "step": 415, "train_speed(iter/s)": 0.020193 }, { "epoch": 0.17387704409025045, "grad_norm": 2.1940409564689656, "learning_rate": 9.415154749199575e-06, "loss": 0.89283857, "memory(GiB)": 63.59, "step": 420, "train_speed(iter/s)": 0.020198 }, { "epoch": 0.1759470089008487, "grad_norm": 2.2711317580338912, "learning_rate": 9.404482390608326e-06, "loss": 0.8422184, "memory(GiB)": 63.59, "step": 425, "train_speed(iter/s)": 0.020202 }, { "epoch": 0.1780169737114469, "grad_norm": 1.9734252369885248, "learning_rate": 9.393810032017077e-06, "loss": 0.90587616, "memory(GiB)": 63.59, "step": 430, "train_speed(iter/s)": 0.020208 }, { "epoch": 0.18008693852204513, "grad_norm": 2.1358019624149653, "learning_rate": 9.383137673425828e-06, "loss": 0.87137203, "memory(GiB)": 63.59, "step": 435, "train_speed(iter/s)": 0.020212 }, { "epoch": 0.18215690333264334, "grad_norm": 2.136663123639741, "learning_rate": 9.372465314834579e-06, "loss": 0.84276152, "memory(GiB)": 63.59, "step": 440, "train_speed(iter/s)": 0.020218 }, { "epoch": 0.18422686814324157, "grad_norm": 2.1730526677654005, "learning_rate": 9.361792956243332e-06, "loss": 0.86879997, "memory(GiB)": 63.59, "step": 445, "train_speed(iter/s)": 0.020222 }, { "epoch": 0.18629683295383978, "grad_norm": 1.7007840288725673, "learning_rate": 9.351120597652081e-06, "loss": 0.85356216, "memory(GiB)": 63.59, "step": 450, "train_speed(iter/s)": 0.020224 }, { "epoch": 0.18836679776443802, "grad_norm": 2.333224503644692, "learning_rate": 9.340448239060834e-06, "loss": 0.87298975, "memory(GiB)": 63.59, "step": 455, "train_speed(iter/s)": 0.02023 }, { "epoch": 0.19043676257503622, "grad_norm": 1.7497884985137717, "learning_rate": 9.329775880469585e-06, "loss": 0.89487724, "memory(GiB)": 63.59, "step": 460, "train_speed(iter/s)": 0.02023 }, { "epoch": 0.19250672738563446, "grad_norm": 1.7881559718064066, "learning_rate": 9.319103521878336e-06, "loss": 0.88939381, "memory(GiB)": 63.59, "step": 465, "train_speed(iter/s)": 0.020232 }, { "epoch": 0.19457669219623266, "grad_norm": 2.5056663267756605, "learning_rate": 9.308431163287087e-06, "loss": 0.85123787, "memory(GiB)": 63.59, "step": 470, "train_speed(iter/s)": 0.020236 }, { "epoch": 0.19664665700683087, "grad_norm": 2.2621024798210403, "learning_rate": 9.29775880469584e-06, "loss": 0.87380323, "memory(GiB)": 63.59, "step": 475, "train_speed(iter/s)": 0.020238 }, { "epoch": 0.1987166218174291, "grad_norm": 1.9008868824283842, "learning_rate": 9.287086446104589e-06, "loss": 0.82988033, "memory(GiB)": 63.59, "step": 480, "train_speed(iter/s)": 0.020241 }, { "epoch": 0.2007865866280273, "grad_norm": 1.6960491773696469, "learning_rate": 9.276414087513342e-06, "loss": 0.85842476, "memory(GiB)": 63.59, "step": 485, "train_speed(iter/s)": 0.020243 }, { "epoch": 0.20285655143862555, "grad_norm": 1.8755094068628242, "learning_rate": 9.265741728922093e-06, "loss": 0.86819458, "memory(GiB)": 63.59, "step": 490, "train_speed(iter/s)": 0.020249 }, { "epoch": 0.20492651624922376, "grad_norm": 1.6029255395235227, "learning_rate": 9.255069370330844e-06, "loss": 0.86832209, "memory(GiB)": 63.59, "step": 495, "train_speed(iter/s)": 0.020253 }, { "epoch": 0.206996481059822, "grad_norm": 1.8312986531673774, "learning_rate": 9.244397011739595e-06, "loss": 0.87804108, "memory(GiB)": 63.59, "step": 500, "train_speed(iter/s)": 0.020256 }, { "epoch": 0.2090664458704202, "grad_norm": 2.2159879518724686, "learning_rate": 9.233724653148346e-06, "loss": 0.84401827, "memory(GiB)": 63.59, "step": 505, "train_speed(iter/s)": 0.020258 }, { "epoch": 0.21113641068101843, "grad_norm": 1.9489190185976173, "learning_rate": 9.223052294557098e-06, "loss": 0.83081837, "memory(GiB)": 63.59, "step": 510, "train_speed(iter/s)": 0.020262 }, { "epoch": 0.21320637549161664, "grad_norm": 1.8621375658015202, "learning_rate": 9.21237993596585e-06, "loss": 0.84451389, "memory(GiB)": 63.59, "step": 515, "train_speed(iter/s)": 0.020265 }, { "epoch": 0.21527634030221487, "grad_norm": 2.657592267470185, "learning_rate": 9.2017075773746e-06, "loss": 0.85252399, "memory(GiB)": 63.59, "step": 520, "train_speed(iter/s)": 0.020272 }, { "epoch": 0.21734630511281308, "grad_norm": 3.2134734541192556, "learning_rate": 9.191035218783352e-06, "loss": 0.85981674, "memory(GiB)": 63.59, "step": 525, "train_speed(iter/s)": 0.020278 }, { "epoch": 0.2194162699234113, "grad_norm": 1.979014196110588, "learning_rate": 9.180362860192104e-06, "loss": 0.85360508, "memory(GiB)": 63.59, "step": 530, "train_speed(iter/s)": 0.020281 }, { "epoch": 0.22148623473400952, "grad_norm": 2.263346027010783, "learning_rate": 9.169690501600854e-06, "loss": 0.8649641, "memory(GiB)": 63.59, "step": 535, "train_speed(iter/s)": 0.020285 }, { "epoch": 0.22355619954460773, "grad_norm": 2.0257701801426786, "learning_rate": 9.159018143009606e-06, "loss": 0.85079117, "memory(GiB)": 63.59, "step": 540, "train_speed(iter/s)": 0.020291 }, { "epoch": 0.22562616435520597, "grad_norm": 2.125007231598407, "learning_rate": 9.148345784418357e-06, "loss": 0.83591347, "memory(GiB)": 63.59, "step": 545, "train_speed(iter/s)": 0.020292 }, { "epoch": 0.22769612916580417, "grad_norm": 1.830826573395782, "learning_rate": 9.137673425827108e-06, "loss": 0.8597187, "memory(GiB)": 63.59, "step": 550, "train_speed(iter/s)": 0.020297 }, { "epoch": 0.2297660939764024, "grad_norm": 2.145744349446719, "learning_rate": 9.12700106723586e-06, "loss": 0.82627001, "memory(GiB)": 63.59, "step": 555, "train_speed(iter/s)": 0.020299 }, { "epoch": 0.23183605878700062, "grad_norm": 1.895819945079046, "learning_rate": 9.116328708644612e-06, "loss": 0.83418722, "memory(GiB)": 63.59, "step": 560, "train_speed(iter/s)": 0.020301 }, { "epoch": 0.23390602359759885, "grad_norm": 1.7330277741570008, "learning_rate": 9.105656350053362e-06, "loss": 0.83898754, "memory(GiB)": 63.59, "step": 565, "train_speed(iter/s)": 0.020301 }, { "epoch": 0.23597598840819706, "grad_norm": 1.873262235276853, "learning_rate": 9.094983991462114e-06, "loss": 0.86104965, "memory(GiB)": 63.59, "step": 570, "train_speed(iter/s)": 0.020307 }, { "epoch": 0.2380459532187953, "grad_norm": 1.8133555971052358, "learning_rate": 9.084311632870865e-06, "loss": 0.83760166, "memory(GiB)": 63.59, "step": 575, "train_speed(iter/s)": 0.020309 }, { "epoch": 0.2401159180293935, "grad_norm": 2.1026280097135377, "learning_rate": 9.073639274279616e-06, "loss": 0.85497513, "memory(GiB)": 63.59, "step": 580, "train_speed(iter/s)": 0.02031 }, { "epoch": 0.2421858828399917, "grad_norm": 1.7385609176743078, "learning_rate": 9.062966915688367e-06, "loss": 0.83206367, "memory(GiB)": 63.59, "step": 585, "train_speed(iter/s)": 0.02031 }, { "epoch": 0.24425584765058994, "grad_norm": 2.310226819070514, "learning_rate": 9.052294557097118e-06, "loss": 0.82417412, "memory(GiB)": 63.59, "step": 590, "train_speed(iter/s)": 0.020312 }, { "epoch": 0.24632581246118815, "grad_norm": 2.1660135303280126, "learning_rate": 9.041622198505871e-06, "loss": 0.8371232, "memory(GiB)": 63.59, "step": 595, "train_speed(iter/s)": 0.020312 }, { "epoch": 0.24839577727178638, "grad_norm": 2.0932218548460493, "learning_rate": 9.030949839914622e-06, "loss": 0.86303692, "memory(GiB)": 63.59, "step": 600, "train_speed(iter/s)": 0.020315 }, { "epoch": 0.24839577727178638, "eval_loss": 0.8761223554611206, "eval_runtime": 333.7076, "eval_samples_per_second": 18.717, "eval_steps_per_second": 1.172, "step": 600 }, { "epoch": 0.2504657420823846, "grad_norm": 1.9850417129561133, "learning_rate": 9.020277481323373e-06, "loss": 0.84975281, "memory(GiB)": 63.59, "step": 605, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.2525357068929828, "grad_norm": 1.8024791223082373, "learning_rate": 9.009605122732124e-06, "loss": 0.83602209, "memory(GiB)": 63.59, "step": 610, "train_speed(iter/s)": 0.020063 }, { "epoch": 0.25460567170358106, "grad_norm": 1.795807849269691, "learning_rate": 8.998932764140877e-06, "loss": 0.84287434, "memory(GiB)": 63.59, "step": 615, "train_speed(iter/s)": 0.020066 }, { "epoch": 0.25667563651417924, "grad_norm": 2.8462701485274855, "learning_rate": 8.988260405549626e-06, "loss": 0.86969414, "memory(GiB)": 63.59, "step": 620, "train_speed(iter/s)": 0.020069 }, { "epoch": 0.2587456013247775, "grad_norm": 1.9467464102992238, "learning_rate": 8.977588046958379e-06, "loss": 0.84205284, "memory(GiB)": 63.59, "step": 625, "train_speed(iter/s)": 0.020075 }, { "epoch": 0.2608155661353757, "grad_norm": 1.9359113111268293, "learning_rate": 8.96691568836713e-06, "loss": 0.83059912, "memory(GiB)": 63.59, "step": 630, "train_speed(iter/s)": 0.02008 }, { "epoch": 0.26288553094597394, "grad_norm": 2.269649830017561, "learning_rate": 8.956243329775881e-06, "loss": 0.85204124, "memory(GiB)": 63.59, "step": 635, "train_speed(iter/s)": 0.020082 }, { "epoch": 0.2649554957565721, "grad_norm": 2.0739328945699014, "learning_rate": 8.945570971184632e-06, "loss": 0.83141527, "memory(GiB)": 63.59, "step": 640, "train_speed(iter/s)": 0.020085 }, { "epoch": 0.26702546056717036, "grad_norm": 2.39308971265692, "learning_rate": 8.934898612593383e-06, "loss": 0.82764578, "memory(GiB)": 63.59, "step": 645, "train_speed(iter/s)": 0.020088 }, { "epoch": 0.2690954253777686, "grad_norm": 2.014414879589864, "learning_rate": 8.924226254002136e-06, "loss": 0.8550128, "memory(GiB)": 63.59, "step": 650, "train_speed(iter/s)": 0.020095 }, { "epoch": 0.2711653901883668, "grad_norm": 2.8374721965360887, "learning_rate": 8.913553895410887e-06, "loss": 0.86329079, "memory(GiB)": 63.59, "step": 655, "train_speed(iter/s)": 0.020098 }, { "epoch": 0.273235354998965, "grad_norm": 1.9583062241735367, "learning_rate": 8.902881536819638e-06, "loss": 0.83884621, "memory(GiB)": 63.59, "step": 660, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.27530531980956324, "grad_norm": 1.765892011718539, "learning_rate": 8.892209178228389e-06, "loss": 0.86448555, "memory(GiB)": 63.59, "step": 665, "train_speed(iter/s)": 0.020105 }, { "epoch": 0.2773752846201615, "grad_norm": 1.7474590269235404, "learning_rate": 8.88153681963714e-06, "loss": 0.81313992, "memory(GiB)": 63.59, "step": 670, "train_speed(iter/s)": 0.020107 }, { "epoch": 0.27944524943075966, "grad_norm": 1.9301834126279824, "learning_rate": 8.870864461045891e-06, "loss": 0.8374301, "memory(GiB)": 63.59, "step": 675, "train_speed(iter/s)": 0.02011 }, { "epoch": 0.2815152142413579, "grad_norm": 1.876790685008959, "learning_rate": 8.860192102454644e-06, "loss": 0.82494678, "memory(GiB)": 63.59, "step": 680, "train_speed(iter/s)": 0.020115 }, { "epoch": 0.2835851790519561, "grad_norm": 1.7854143077330529, "learning_rate": 8.849519743863395e-06, "loss": 0.83442841, "memory(GiB)": 63.59, "step": 685, "train_speed(iter/s)": 0.020116 }, { "epoch": 0.28565514386255436, "grad_norm": 1.8923637542669056, "learning_rate": 8.838847385272146e-06, "loss": 0.82085514, "memory(GiB)": 63.72, "step": 690, "train_speed(iter/s)": 0.020118 }, { "epoch": 0.28772510867315254, "grad_norm": 2.0170179705017066, "learning_rate": 8.828175026680897e-06, "loss": 0.85945168, "memory(GiB)": 63.72, "step": 695, "train_speed(iter/s)": 0.020121 }, { "epoch": 0.2897950734837508, "grad_norm": 3.3823068740775755, "learning_rate": 8.817502668089648e-06, "loss": 0.82407131, "memory(GiB)": 63.72, "step": 700, "train_speed(iter/s)": 0.020123 }, { "epoch": 0.291865038294349, "grad_norm": 1.9439481100139924, "learning_rate": 8.806830309498399e-06, "loss": 0.82105274, "memory(GiB)": 63.72, "step": 705, "train_speed(iter/s)": 0.020126 }, { "epoch": 0.2939350031049472, "grad_norm": 1.8993570568677929, "learning_rate": 8.796157950907152e-06, "loss": 0.83391781, "memory(GiB)": 63.72, "step": 710, "train_speed(iter/s)": 0.020131 }, { "epoch": 0.2960049679155454, "grad_norm": 2.175929579998878, "learning_rate": 8.785485592315903e-06, "loss": 0.84003325, "memory(GiB)": 63.72, "step": 715, "train_speed(iter/s)": 0.020137 }, { "epoch": 0.29807493272614366, "grad_norm": 1.6910798969618672, "learning_rate": 8.774813233724654e-06, "loss": 0.82005548, "memory(GiB)": 63.72, "step": 720, "train_speed(iter/s)": 0.02014 }, { "epoch": 0.3001448975367419, "grad_norm": 1.7577054177826072, "learning_rate": 8.764140875133405e-06, "loss": 0.84406672, "memory(GiB)": 63.72, "step": 725, "train_speed(iter/s)": 0.020144 }, { "epoch": 0.3022148623473401, "grad_norm": 1.9343300270246129, "learning_rate": 8.753468516542156e-06, "loss": 0.81861668, "memory(GiB)": 63.72, "step": 730, "train_speed(iter/s)": 0.020144 }, { "epoch": 0.3042848271579383, "grad_norm": 1.8986244788103208, "learning_rate": 8.742796157950909e-06, "loss": 0.81786537, "memory(GiB)": 63.72, "step": 735, "train_speed(iter/s)": 0.020148 }, { "epoch": 0.30635479196853654, "grad_norm": 2.095799409220846, "learning_rate": 8.73212379935966e-06, "loss": 0.83321962, "memory(GiB)": 63.72, "step": 740, "train_speed(iter/s)": 0.020151 }, { "epoch": 0.3084247567791348, "grad_norm": 1.9094006901482394, "learning_rate": 8.72145144076841e-06, "loss": 0.82653723, "memory(GiB)": 63.72, "step": 745, "train_speed(iter/s)": 0.020152 }, { "epoch": 0.31049472158973296, "grad_norm": 2.126120113530993, "learning_rate": 8.710779082177162e-06, "loss": 0.85463696, "memory(GiB)": 63.72, "step": 750, "train_speed(iter/s)": 0.020156 }, { "epoch": 0.3125646864003312, "grad_norm": 1.766780713214732, "learning_rate": 8.700106723585913e-06, "loss": 0.83797083, "memory(GiB)": 63.72, "step": 755, "train_speed(iter/s)": 0.020157 }, { "epoch": 0.31463465121092943, "grad_norm": 1.8957319688723608, "learning_rate": 8.689434364994664e-06, "loss": 0.81888847, "memory(GiB)": 63.72, "step": 760, "train_speed(iter/s)": 0.020159 }, { "epoch": 0.3167046160215276, "grad_norm": 1.9661061189594824, "learning_rate": 8.678762006403417e-06, "loss": 0.78800874, "memory(GiB)": 63.72, "step": 765, "train_speed(iter/s)": 0.020164 }, { "epoch": 0.31877458083212584, "grad_norm": 1.8837863956075926, "learning_rate": 8.668089647812166e-06, "loss": 0.84463196, "memory(GiB)": 63.72, "step": 770, "train_speed(iter/s)": 0.020167 }, { "epoch": 0.3208445456427241, "grad_norm": 2.5248078655238326, "learning_rate": 8.657417289220919e-06, "loss": 0.83094559, "memory(GiB)": 63.72, "step": 775, "train_speed(iter/s)": 0.02017 }, { "epoch": 0.3229145104533223, "grad_norm": 1.8996595550385447, "learning_rate": 8.64674493062967e-06, "loss": 0.81705608, "memory(GiB)": 63.72, "step": 780, "train_speed(iter/s)": 0.020173 }, { "epoch": 0.3249844752639205, "grad_norm": 1.8243459235808355, "learning_rate": 8.63607257203842e-06, "loss": 0.82983418, "memory(GiB)": 63.72, "step": 785, "train_speed(iter/s)": 0.020177 }, { "epoch": 0.3270544400745187, "grad_norm": 2.125198435674726, "learning_rate": 8.625400213447172e-06, "loss": 0.85153885, "memory(GiB)": 63.72, "step": 790, "train_speed(iter/s)": 0.02018 }, { "epoch": 0.32912440488511696, "grad_norm": 1.822527258966965, "learning_rate": 8.614727854855925e-06, "loss": 0.7932189, "memory(GiB)": 63.72, "step": 795, "train_speed(iter/s)": 0.020184 }, { "epoch": 0.3311943696957152, "grad_norm": 1.9585269031801074, "learning_rate": 8.604055496264676e-06, "loss": 0.80502253, "memory(GiB)": 63.72, "step": 800, "train_speed(iter/s)": 0.020188 }, { "epoch": 0.3332643345063134, "grad_norm": 1.9244862407118186, "learning_rate": 8.593383137673427e-06, "loss": 0.81400661, "memory(GiB)": 63.72, "step": 805, "train_speed(iter/s)": 0.020192 }, { "epoch": 0.3353342993169116, "grad_norm": 1.8781928942945239, "learning_rate": 8.582710779082178e-06, "loss": 0.82624207, "memory(GiB)": 63.72, "step": 810, "train_speed(iter/s)": 0.020195 }, { "epoch": 0.33740426412750985, "grad_norm": 2.4821098212553108, "learning_rate": 8.572038420490929e-06, "loss": 0.81296177, "memory(GiB)": 63.72, "step": 815, "train_speed(iter/s)": 0.020197 }, { "epoch": 0.339474228938108, "grad_norm": 3.2468832100225877, "learning_rate": 8.561366061899681e-06, "loss": 0.81447935, "memory(GiB)": 63.72, "step": 820, "train_speed(iter/s)": 0.0202 }, { "epoch": 0.34154419374870626, "grad_norm": 1.726217016729622, "learning_rate": 8.55069370330843e-06, "loss": 0.82119083, "memory(GiB)": 63.72, "step": 825, "train_speed(iter/s)": 0.020203 }, { "epoch": 0.3436141585593045, "grad_norm": 1.8200397633087098, "learning_rate": 8.540021344717184e-06, "loss": 0.80688438, "memory(GiB)": 63.72, "step": 830, "train_speed(iter/s)": 0.020205 }, { "epoch": 0.34568412336990273, "grad_norm": 1.7077741062644576, "learning_rate": 8.529348986125935e-06, "loss": 0.83244801, "memory(GiB)": 63.72, "step": 835, "train_speed(iter/s)": 0.020205 }, { "epoch": 0.3477540881805009, "grad_norm": 2.582896676288874, "learning_rate": 8.518676627534686e-06, "loss": 0.81135302, "memory(GiB)": 63.72, "step": 840, "train_speed(iter/s)": 0.020208 }, { "epoch": 0.34982405299109914, "grad_norm": 3.4613638587514033, "learning_rate": 8.508004268943437e-06, "loss": 0.80561113, "memory(GiB)": 63.72, "step": 845, "train_speed(iter/s)": 0.020209 }, { "epoch": 0.3518940178016974, "grad_norm": 1.6179386547462884, "learning_rate": 8.49733191035219e-06, "loss": 0.82198238, "memory(GiB)": 63.72, "step": 850, "train_speed(iter/s)": 0.020211 }, { "epoch": 0.3539639826122956, "grad_norm": 2.202413903162471, "learning_rate": 8.48665955176094e-06, "loss": 0.78598285, "memory(GiB)": 63.72, "step": 855, "train_speed(iter/s)": 0.020214 }, { "epoch": 0.3560339474228938, "grad_norm": 1.9513315920239633, "learning_rate": 8.475987193169691e-06, "loss": 0.80893326, "memory(GiB)": 63.72, "step": 860, "train_speed(iter/s)": 0.020219 }, { "epoch": 0.358103912233492, "grad_norm": 1.9113374189570778, "learning_rate": 8.465314834578443e-06, "loss": 0.8136569, "memory(GiB)": 63.72, "step": 865, "train_speed(iter/s)": 0.020222 }, { "epoch": 0.36017387704409026, "grad_norm": 2.084935583050277, "learning_rate": 8.454642475987194e-06, "loss": 0.81384058, "memory(GiB)": 63.72, "step": 870, "train_speed(iter/s)": 0.020226 }, { "epoch": 0.3622438418546885, "grad_norm": 1.6048226105298027, "learning_rate": 8.443970117395945e-06, "loss": 0.81689348, "memory(GiB)": 63.72, "step": 875, "train_speed(iter/s)": 0.020228 }, { "epoch": 0.3643138066652867, "grad_norm": 1.8081549724032602, "learning_rate": 8.433297758804696e-06, "loss": 0.8224082, "memory(GiB)": 63.72, "step": 880, "train_speed(iter/s)": 0.02023 }, { "epoch": 0.3663837714758849, "grad_norm": 1.8184484923663322, "learning_rate": 8.422625400213448e-06, "loss": 0.78473282, "memory(GiB)": 63.72, "step": 885, "train_speed(iter/s)": 0.020232 }, { "epoch": 0.36845373628648315, "grad_norm": 2.010882441005616, "learning_rate": 8.4119530416222e-06, "loss": 0.81135426, "memory(GiB)": 63.72, "step": 890, "train_speed(iter/s)": 0.020232 }, { "epoch": 0.3705237010970813, "grad_norm": 2.363919887534564, "learning_rate": 8.40128068303095e-06, "loss": 0.80090466, "memory(GiB)": 63.72, "step": 895, "train_speed(iter/s)": 0.020234 }, { "epoch": 0.37259366590767956, "grad_norm": 1.6332844070852461, "learning_rate": 8.390608324439701e-06, "loss": 0.81239138, "memory(GiB)": 63.72, "step": 900, "train_speed(iter/s)": 0.020237 }, { "epoch": 0.37259366590767956, "eval_loss": 0.8537026047706604, "eval_runtime": 333.2325, "eval_samples_per_second": 18.744, "eval_steps_per_second": 1.173, "step": 900 }, { "epoch": 0.3746636307182778, "grad_norm": 2.4425379037284607, "learning_rate": 8.379935965848454e-06, "loss": 0.81980333, "memory(GiB)": 63.72, "step": 905, "train_speed(iter/s)": 0.02007 }, { "epoch": 0.37673359552887603, "grad_norm": 2.2417018048030575, "learning_rate": 8.369263607257204e-06, "loss": 0.82684288, "memory(GiB)": 63.72, "step": 910, "train_speed(iter/s)": 0.020071 }, { "epoch": 0.3788035603394742, "grad_norm": 1.7587001424417825, "learning_rate": 8.358591248665956e-06, "loss": 0.83183241, "memory(GiB)": 63.72, "step": 915, "train_speed(iter/s)": 0.020074 }, { "epoch": 0.38087352515007245, "grad_norm": 1.5739208911232379, "learning_rate": 8.347918890074707e-06, "loss": 0.76680841, "memory(GiB)": 63.72, "step": 920, "train_speed(iter/s)": 0.020076 }, { "epoch": 0.3829434899606707, "grad_norm": 1.8684171867799126, "learning_rate": 8.337246531483458e-06, "loss": 0.85463772, "memory(GiB)": 63.72, "step": 925, "train_speed(iter/s)": 0.020079 }, { "epoch": 0.3850134547712689, "grad_norm": 1.8550376770414303, "learning_rate": 8.32657417289221e-06, "loss": 0.81911144, "memory(GiB)": 63.72, "step": 930, "train_speed(iter/s)": 0.020083 }, { "epoch": 0.3870834195818671, "grad_norm": 1.8194471444369447, "learning_rate": 8.31590181430096e-06, "loss": 0.83044968, "memory(GiB)": 63.72, "step": 935, "train_speed(iter/s)": 0.020086 }, { "epoch": 0.38915338439246533, "grad_norm": 2.095169413730359, "learning_rate": 8.305229455709713e-06, "loss": 0.82608871, "memory(GiB)": 63.72, "step": 940, "train_speed(iter/s)": 0.020088 }, { "epoch": 0.39122334920306356, "grad_norm": 13.855393242400119, "learning_rate": 8.294557097118464e-06, "loss": 0.81676388, "memory(GiB)": 63.72, "step": 945, "train_speed(iter/s)": 0.02009 }, { "epoch": 0.39329331401366174, "grad_norm": 2.100046441650532, "learning_rate": 8.283884738527215e-06, "loss": 0.81071377, "memory(GiB)": 63.72, "step": 950, "train_speed(iter/s)": 0.020093 }, { "epoch": 0.39536327882426, "grad_norm": 2.1608254386705594, "learning_rate": 8.273212379935966e-06, "loss": 0.78902674, "memory(GiB)": 63.72, "step": 955, "train_speed(iter/s)": 0.020095 }, { "epoch": 0.3974332436348582, "grad_norm": 1.61490095503505, "learning_rate": 8.262540021344719e-06, "loss": 0.78527632, "memory(GiB)": 63.72, "step": 960, "train_speed(iter/s)": 0.020097 }, { "epoch": 0.39950320844545645, "grad_norm": 2.589460194979307, "learning_rate": 8.251867662753468e-06, "loss": 0.81925201, "memory(GiB)": 63.72, "step": 965, "train_speed(iter/s)": 0.020098 }, { "epoch": 0.4015731732560546, "grad_norm": 1.8550853889727008, "learning_rate": 8.241195304162221e-06, "loss": 0.80688972, "memory(GiB)": 63.72, "step": 970, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.40364313806665286, "grad_norm": 2.4637090355397517, "learning_rate": 8.230522945570972e-06, "loss": 0.82061405, "memory(GiB)": 63.72, "step": 975, "train_speed(iter/s)": 0.020104 }, { "epoch": 0.4057131028772511, "grad_norm": 1.9846780448601058, "learning_rate": 8.219850586979723e-06, "loss": 0.80861397, "memory(GiB)": 63.72, "step": 980, "train_speed(iter/s)": 0.020106 }, { "epoch": 0.40778306768784933, "grad_norm": 1.9138045073506678, "learning_rate": 8.209178228388474e-06, "loss": 0.79171362, "memory(GiB)": 63.72, "step": 985, "train_speed(iter/s)": 0.020108 }, { "epoch": 0.4098530324984475, "grad_norm": 2.100897386160194, "learning_rate": 8.198505869797227e-06, "loss": 0.78543482, "memory(GiB)": 63.72, "step": 990, "train_speed(iter/s)": 0.020113 }, { "epoch": 0.41192299730904575, "grad_norm": 2.1398479736163667, "learning_rate": 8.187833511205976e-06, "loss": 0.79959226, "memory(GiB)": 63.72, "step": 995, "train_speed(iter/s)": 0.020114 }, { "epoch": 0.413992962119644, "grad_norm": 1.8543885416746075, "learning_rate": 8.177161152614729e-06, "loss": 0.80144148, "memory(GiB)": 63.72, "step": 1000, "train_speed(iter/s)": 0.020116 }, { "epoch": 0.41606292693024216, "grad_norm": 1.8895963845216188, "learning_rate": 8.16648879402348e-06, "loss": 0.78557086, "memory(GiB)": 63.72, "step": 1005, "train_speed(iter/s)": 0.020118 }, { "epoch": 0.4181328917408404, "grad_norm": 1.7583886118404264, "learning_rate": 8.155816435432231e-06, "loss": 0.81441746, "memory(GiB)": 63.72, "step": 1010, "train_speed(iter/s)": 0.02012 }, { "epoch": 0.42020285655143863, "grad_norm": 1.8640464710188405, "learning_rate": 8.145144076840982e-06, "loss": 0.76718016, "memory(GiB)": 63.72, "step": 1015, "train_speed(iter/s)": 0.020122 }, { "epoch": 0.42227282136203687, "grad_norm": 2.0754981449007084, "learning_rate": 8.134471718249733e-06, "loss": 0.78537526, "memory(GiB)": 63.72, "step": 1020, "train_speed(iter/s)": 0.020125 }, { "epoch": 0.42434278617263504, "grad_norm": 2.1358764475250105, "learning_rate": 8.123799359658486e-06, "loss": 0.82194328, "memory(GiB)": 63.72, "step": 1025, "train_speed(iter/s)": 0.020128 }, { "epoch": 0.4264127509832333, "grad_norm": 1.940572767867165, "learning_rate": 8.113127001067237e-06, "loss": 0.8162715, "memory(GiB)": 63.72, "step": 1030, "train_speed(iter/s)": 0.020131 }, { "epoch": 0.4284827157938315, "grad_norm": 1.7824953515185047, "learning_rate": 8.102454642475988e-06, "loss": 0.78834782, "memory(GiB)": 63.72, "step": 1035, "train_speed(iter/s)": 0.020133 }, { "epoch": 0.43055268060442975, "grad_norm": 1.9585541886433688, "learning_rate": 8.091782283884739e-06, "loss": 0.79206867, "memory(GiB)": 63.72, "step": 1040, "train_speed(iter/s)": 0.020136 }, { "epoch": 0.43262264541502793, "grad_norm": 1.6194935665114412, "learning_rate": 8.081109925293492e-06, "loss": 0.80889845, "memory(GiB)": 63.72, "step": 1045, "train_speed(iter/s)": 0.020139 }, { "epoch": 0.43469261022562616, "grad_norm": 1.5909296898104581, "learning_rate": 8.070437566702241e-06, "loss": 0.76998816, "memory(GiB)": 63.72, "step": 1050, "train_speed(iter/s)": 0.020141 }, { "epoch": 0.4367625750362244, "grad_norm": 1.582985265467202, "learning_rate": 8.059765208110994e-06, "loss": 0.79827099, "memory(GiB)": 63.72, "step": 1055, "train_speed(iter/s)": 0.020143 }, { "epoch": 0.4388325398468226, "grad_norm": 1.9696406410447012, "learning_rate": 8.049092849519743e-06, "loss": 0.79220991, "memory(GiB)": 63.72, "step": 1060, "train_speed(iter/s)": 0.020146 }, { "epoch": 0.4409025046574208, "grad_norm": 1.9479888997003834, "learning_rate": 8.038420490928496e-06, "loss": 0.79184585, "memory(GiB)": 63.72, "step": 1065, "train_speed(iter/s)": 0.020148 }, { "epoch": 0.44297246946801905, "grad_norm": 1.7883498032309324, "learning_rate": 8.027748132337247e-06, "loss": 0.78507504, "memory(GiB)": 63.72, "step": 1070, "train_speed(iter/s)": 0.02015 }, { "epoch": 0.4450424342786173, "grad_norm": 1.6985331753731079, "learning_rate": 8.017075773745998e-06, "loss": 0.81149197, "memory(GiB)": 63.72, "step": 1075, "train_speed(iter/s)": 0.020154 }, { "epoch": 0.44711239908921546, "grad_norm": 1.7646873640943033, "learning_rate": 8.006403415154749e-06, "loss": 0.77548814, "memory(GiB)": 63.72, "step": 1080, "train_speed(iter/s)": 0.020158 }, { "epoch": 0.4491823638998137, "grad_norm": 1.7739180508215708, "learning_rate": 7.995731056563502e-06, "loss": 0.77309542, "memory(GiB)": 63.72, "step": 1085, "train_speed(iter/s)": 0.02016 }, { "epoch": 0.45125232871041193, "grad_norm": 1.6366153166920923, "learning_rate": 7.985058697972253e-06, "loss": 0.80448093, "memory(GiB)": 63.72, "step": 1090, "train_speed(iter/s)": 0.020163 }, { "epoch": 0.45332229352101017, "grad_norm": 1.9520729703143727, "learning_rate": 7.974386339381004e-06, "loss": 0.78142538, "memory(GiB)": 63.72, "step": 1095, "train_speed(iter/s)": 0.020164 }, { "epoch": 0.45539225833160835, "grad_norm": 1.9086497373347489, "learning_rate": 7.963713980789755e-06, "loss": 0.7876678, "memory(GiB)": 63.72, "step": 1100, "train_speed(iter/s)": 0.020167 }, { "epoch": 0.4574622231422066, "grad_norm": 1.773963819363606, "learning_rate": 7.953041622198506e-06, "loss": 0.817309, "memory(GiB)": 63.72, "step": 1105, "train_speed(iter/s)": 0.020169 }, { "epoch": 0.4595321879528048, "grad_norm": 1.7610901122064158, "learning_rate": 7.942369263607259e-06, "loss": 0.81472855, "memory(GiB)": 63.72, "step": 1110, "train_speed(iter/s)": 0.020171 }, { "epoch": 0.461602152763403, "grad_norm": 1.8809865504177992, "learning_rate": 7.931696905016008e-06, "loss": 0.79487166, "memory(GiB)": 63.72, "step": 1115, "train_speed(iter/s)": 0.020174 }, { "epoch": 0.46367211757400123, "grad_norm": 1.8810868009507724, "learning_rate": 7.92102454642476e-06, "loss": 0.78505554, "memory(GiB)": 63.72, "step": 1120, "train_speed(iter/s)": 0.020176 }, { "epoch": 0.46574208238459947, "grad_norm": 1.7134608698183469, "learning_rate": 7.910352187833512e-06, "loss": 0.77789249, "memory(GiB)": 63.72, "step": 1125, "train_speed(iter/s)": 0.020179 }, { "epoch": 0.4678120471951977, "grad_norm": 1.9145455941813492, "learning_rate": 7.899679829242263e-06, "loss": 0.76669245, "memory(GiB)": 63.72, "step": 1130, "train_speed(iter/s)": 0.020182 }, { "epoch": 0.4698820120057959, "grad_norm": 1.9752917655252427, "learning_rate": 7.889007470651014e-06, "loss": 0.77915101, "memory(GiB)": 63.72, "step": 1135, "train_speed(iter/s)": 0.020184 }, { "epoch": 0.4719519768163941, "grad_norm": 1.8705706085741929, "learning_rate": 7.878335112059767e-06, "loss": 0.78053985, "memory(GiB)": 63.72, "step": 1140, "train_speed(iter/s)": 0.020187 }, { "epoch": 0.47402194162699235, "grad_norm": 1.8137417073497548, "learning_rate": 7.867662753468518e-06, "loss": 0.8304471, "memory(GiB)": 63.72, "step": 1145, "train_speed(iter/s)": 0.020189 }, { "epoch": 0.4760919064375906, "grad_norm": 1.7537064971860614, "learning_rate": 7.856990394877269e-06, "loss": 0.76652546, "memory(GiB)": 63.72, "step": 1150, "train_speed(iter/s)": 0.020192 }, { "epoch": 0.47816187124818876, "grad_norm": 1.8981437943138895, "learning_rate": 7.84631803628602e-06, "loss": 0.77384648, "memory(GiB)": 63.72, "step": 1155, "train_speed(iter/s)": 0.020194 }, { "epoch": 0.480231836058787, "grad_norm": 1.968718081590253, "learning_rate": 7.83564567769477e-06, "loss": 0.81451969, "memory(GiB)": 63.72, "step": 1160, "train_speed(iter/s)": 0.020196 }, { "epoch": 0.48230180086938523, "grad_norm": 1.9755371858466928, "learning_rate": 7.824973319103523e-06, "loss": 0.79220142, "memory(GiB)": 63.72, "step": 1165, "train_speed(iter/s)": 0.020198 }, { "epoch": 0.4843717656799834, "grad_norm": 1.8485795416766981, "learning_rate": 7.814300960512274e-06, "loss": 0.80019064, "memory(GiB)": 63.72, "step": 1170, "train_speed(iter/s)": 0.0202 }, { "epoch": 0.48644173049058165, "grad_norm": 2.388358446370589, "learning_rate": 7.803628601921026e-06, "loss": 0.8037425, "memory(GiB)": 63.72, "step": 1175, "train_speed(iter/s)": 0.020202 }, { "epoch": 0.4885116953011799, "grad_norm": 1.7963803355457697, "learning_rate": 7.792956243329777e-06, "loss": 0.76506805, "memory(GiB)": 63.72, "step": 1180, "train_speed(iter/s)": 0.020205 }, { "epoch": 0.4905816601117781, "grad_norm": 1.642582867995439, "learning_rate": 7.782283884738528e-06, "loss": 0.76571236, "memory(GiB)": 63.72, "step": 1185, "train_speed(iter/s)": 0.020206 }, { "epoch": 0.4926516249223763, "grad_norm": 1.8722199369588735, "learning_rate": 7.771611526147279e-06, "loss": 0.81547689, "memory(GiB)": 63.72, "step": 1190, "train_speed(iter/s)": 0.020209 }, { "epoch": 0.49472158973297453, "grad_norm": 1.6444393246271363, "learning_rate": 7.760939167556031e-06, "loss": 0.77196584, "memory(GiB)": 63.72, "step": 1195, "train_speed(iter/s)": 0.02021 }, { "epoch": 0.49679155454357277, "grad_norm": 1.7322851516861686, "learning_rate": 7.75026680896478e-06, "loss": 0.78245749, "memory(GiB)": 63.72, "step": 1200, "train_speed(iter/s)": 0.020211 }, { "epoch": 0.49679155454357277, "eval_loss": 0.8388283252716064, "eval_runtime": 333.9836, "eval_samples_per_second": 18.702, "eval_steps_per_second": 1.171, "step": 1200 }, { "epoch": 0.498861519354171, "grad_norm": 1.779853897918009, "learning_rate": 7.739594450373533e-06, "loss": 0.76194401, "memory(GiB)": 63.72, "step": 1205, "train_speed(iter/s)": 0.020085 }, { "epoch": 0.5009314841647692, "grad_norm": 1.6935101575699751, "learning_rate": 7.728922091782284e-06, "loss": 0.76538324, "memory(GiB)": 63.72, "step": 1210, "train_speed(iter/s)": 0.020088 }, { "epoch": 0.5030014489753675, "grad_norm": 2.1568052824101196, "learning_rate": 7.718249733191036e-06, "loss": 0.80119467, "memory(GiB)": 63.72, "step": 1215, "train_speed(iter/s)": 0.02009 }, { "epoch": 0.5050714137859657, "grad_norm": 2.0414007101619815, "learning_rate": 7.707577374599787e-06, "loss": 0.78779106, "memory(GiB)": 63.72, "step": 1220, "train_speed(iter/s)": 0.020091 }, { "epoch": 0.5071413785965638, "grad_norm": 1.6191663238961727, "learning_rate": 7.69690501600854e-06, "loss": 0.79252872, "memory(GiB)": 63.72, "step": 1225, "train_speed(iter/s)": 0.020092 }, { "epoch": 0.5092113434071621, "grad_norm": 1.8617354993121655, "learning_rate": 7.68623265741729e-06, "loss": 0.76940928, "memory(GiB)": 63.72, "step": 1230, "train_speed(iter/s)": 0.020095 }, { "epoch": 0.5112813082177603, "grad_norm": 2.1148307227706, "learning_rate": 7.675560298826041e-06, "loss": 0.81177235, "memory(GiB)": 63.72, "step": 1235, "train_speed(iter/s)": 0.020096 }, { "epoch": 0.5133512730283585, "grad_norm": 1.858296305288767, "learning_rate": 7.664887940234792e-06, "loss": 0.78712654, "memory(GiB)": 63.72, "step": 1240, "train_speed(iter/s)": 0.020097 }, { "epoch": 0.5154212378389568, "grad_norm": 2.040424767723149, "learning_rate": 7.654215581643543e-06, "loss": 0.7963089, "memory(GiB)": 63.72, "step": 1245, "train_speed(iter/s)": 0.020099 }, { "epoch": 0.517491202649555, "grad_norm": 1.7313703601186623, "learning_rate": 7.643543223052296e-06, "loss": 0.76896205, "memory(GiB)": 63.72, "step": 1250, "train_speed(iter/s)": 0.020102 }, { "epoch": 0.5195611674601531, "grad_norm": 1.6916331849372186, "learning_rate": 7.632870864461046e-06, "loss": 0.78258944, "memory(GiB)": 63.72, "step": 1255, "train_speed(iter/s)": 0.020104 }, { "epoch": 0.5216311322707514, "grad_norm": 2.1058096966812303, "learning_rate": 7.622198505869797e-06, "loss": 0.80921211, "memory(GiB)": 63.72, "step": 1260, "train_speed(iter/s)": 0.020105 }, { "epoch": 0.5237010970813496, "grad_norm": 1.7220759067410432, "learning_rate": 7.611526147278549e-06, "loss": 0.77246647, "memory(GiB)": 63.72, "step": 1265, "train_speed(iter/s)": 0.020106 }, { "epoch": 0.5257710618919479, "grad_norm": 2.084224319084108, "learning_rate": 7.6008537886873e-06, "loss": 0.7677907, "memory(GiB)": 63.72, "step": 1270, "train_speed(iter/s)": 0.020107 }, { "epoch": 0.5278410267025461, "grad_norm": 1.7928505615246706, "learning_rate": 7.590181430096052e-06, "loss": 0.78496704, "memory(GiB)": 63.72, "step": 1275, "train_speed(iter/s)": 0.020108 }, { "epoch": 0.5299109915131442, "grad_norm": 1.8397320603347174, "learning_rate": 7.579509071504803e-06, "loss": 0.77303753, "memory(GiB)": 63.72, "step": 1280, "train_speed(iter/s)": 0.02011 }, { "epoch": 0.5319809563237425, "grad_norm": 2.1479969295234187, "learning_rate": 7.568836712913554e-06, "loss": 0.75871119, "memory(GiB)": 63.72, "step": 1285, "train_speed(iter/s)": 0.020112 }, { "epoch": 0.5340509211343407, "grad_norm": 1.94767502078934, "learning_rate": 7.558164354322306e-06, "loss": 0.75106993, "memory(GiB)": 63.72, "step": 1290, "train_speed(iter/s)": 0.020114 }, { "epoch": 0.5361208859449389, "grad_norm": 1.5236425325852578, "learning_rate": 7.547491995731058e-06, "loss": 0.79110327, "memory(GiB)": 63.72, "step": 1295, "train_speed(iter/s)": 0.020116 }, { "epoch": 0.5381908507555372, "grad_norm": 1.8541149671409907, "learning_rate": 7.536819637139808e-06, "loss": 0.77403798, "memory(GiB)": 63.72, "step": 1300, "train_speed(iter/s)": 0.020116 }, { "epoch": 0.5402608155661354, "grad_norm": 1.8743174944996448, "learning_rate": 7.52614727854856e-06, "loss": 0.77032347, "memory(GiB)": 71.94, "step": 1305, "train_speed(iter/s)": 0.020118 }, { "epoch": 0.5423307803767335, "grad_norm": 2.579806479546849, "learning_rate": 7.51547491995731e-06, "loss": 0.76461482, "memory(GiB)": 71.94, "step": 1310, "train_speed(iter/s)": 0.020119 }, { "epoch": 0.5444007451873318, "grad_norm": 2.0039452129208035, "learning_rate": 7.504802561366062e-06, "loss": 0.77457762, "memory(GiB)": 71.94, "step": 1315, "train_speed(iter/s)": 0.020121 }, { "epoch": 0.54647070999793, "grad_norm": 2.060283685569936, "learning_rate": 7.494130202774814e-06, "loss": 0.77914829, "memory(GiB)": 71.94, "step": 1320, "train_speed(iter/s)": 0.020123 }, { "epoch": 0.5485406748085283, "grad_norm": 2.163132135636586, "learning_rate": 7.483457844183565e-06, "loss": 0.77322574, "memory(GiB)": 71.94, "step": 1325, "train_speed(iter/s)": 0.020125 }, { "epoch": 0.5506106396191265, "grad_norm": 1.842195467860799, "learning_rate": 7.472785485592316e-06, "loss": 0.77454052, "memory(GiB)": 71.94, "step": 1330, "train_speed(iter/s)": 0.020126 }, { "epoch": 0.5526806044297247, "grad_norm": 1.775275008552653, "learning_rate": 7.462113127001068e-06, "loss": 0.77025108, "memory(GiB)": 71.94, "step": 1335, "train_speed(iter/s)": 0.020128 }, { "epoch": 0.554750569240323, "grad_norm": 2.165651142341684, "learning_rate": 7.451440768409819e-06, "loss": 0.78470011, "memory(GiB)": 71.94, "step": 1340, "train_speed(iter/s)": 0.02013 }, { "epoch": 0.5568205340509211, "grad_norm": 1.6530168942960388, "learning_rate": 7.440768409818571e-06, "loss": 0.74261112, "memory(GiB)": 71.94, "step": 1345, "train_speed(iter/s)": 0.020131 }, { "epoch": 0.5588904988615193, "grad_norm": 2.1178890231616694, "learning_rate": 7.430096051227322e-06, "loss": 0.77076225, "memory(GiB)": 71.94, "step": 1350, "train_speed(iter/s)": 0.020132 }, { "epoch": 0.5609604636721176, "grad_norm": 1.6332209286889638, "learning_rate": 7.419423692636073e-06, "loss": 0.76129122, "memory(GiB)": 71.94, "step": 1355, "train_speed(iter/s)": 0.020134 }, { "epoch": 0.5630304284827158, "grad_norm": 1.9276105656674607, "learning_rate": 7.408751334044825e-06, "loss": 0.77616062, "memory(GiB)": 71.94, "step": 1360, "train_speed(iter/s)": 0.020136 }, { "epoch": 0.565100393293314, "grad_norm": 1.838664332126464, "learning_rate": 7.398078975453575e-06, "loss": 0.77545385, "memory(GiB)": 71.94, "step": 1365, "train_speed(iter/s)": 0.020137 }, { "epoch": 0.5671703581039123, "grad_norm": 2.090052030958157, "learning_rate": 7.387406616862327e-06, "loss": 0.7824297, "memory(GiB)": 71.94, "step": 1370, "train_speed(iter/s)": 0.020139 }, { "epoch": 0.5692403229145104, "grad_norm": 1.7799554116738177, "learning_rate": 7.376734258271079e-06, "loss": 0.77833185, "memory(GiB)": 71.94, "step": 1375, "train_speed(iter/s)": 0.020142 }, { "epoch": 0.5713102877251087, "grad_norm": 2.266691996975209, "learning_rate": 7.366061899679829e-06, "loss": 0.77535782, "memory(GiB)": 71.94, "step": 1380, "train_speed(iter/s)": 0.020144 }, { "epoch": 0.5733802525357069, "grad_norm": 1.8220471587007605, "learning_rate": 7.355389541088581e-06, "loss": 0.76158247, "memory(GiB)": 71.94, "step": 1385, "train_speed(iter/s)": 0.020146 }, { "epoch": 0.5754502173463051, "grad_norm": 1.7869060368578336, "learning_rate": 7.344717182497333e-06, "loss": 0.79457912, "memory(GiB)": 71.94, "step": 1390, "train_speed(iter/s)": 0.020147 }, { "epoch": 0.5775201821569034, "grad_norm": 2.730877403121895, "learning_rate": 7.334044823906084e-06, "loss": 0.75181475, "memory(GiB)": 71.94, "step": 1395, "train_speed(iter/s)": 0.020149 }, { "epoch": 0.5795901469675016, "grad_norm": 2.091944883020518, "learning_rate": 7.323372465314835e-06, "loss": 0.75992446, "memory(GiB)": 71.94, "step": 1400, "train_speed(iter/s)": 0.02015 }, { "epoch": 0.5816601117780997, "grad_norm": 1.5904822426334966, "learning_rate": 7.312700106723587e-06, "loss": 0.77254944, "memory(GiB)": 71.94, "step": 1405, "train_speed(iter/s)": 0.020152 }, { "epoch": 0.583730076588698, "grad_norm": 1.673083919686743, "learning_rate": 7.302027748132338e-06, "loss": 0.74836388, "memory(GiB)": 71.94, "step": 1410, "train_speed(iter/s)": 0.020152 }, { "epoch": 0.5858000413992962, "grad_norm": 2.05811523971159, "learning_rate": 7.29135538954109e-06, "loss": 0.74358282, "memory(GiB)": 71.94, "step": 1415, "train_speed(iter/s)": 0.020154 }, { "epoch": 0.5878700062098944, "grad_norm": 1.983632865952002, "learning_rate": 7.28068303094984e-06, "loss": 0.78234367, "memory(GiB)": 71.94, "step": 1420, "train_speed(iter/s)": 0.020156 }, { "epoch": 0.5899399710204927, "grad_norm": 1.6612296882759847, "learning_rate": 7.270010672358592e-06, "loss": 0.76740494, "memory(GiB)": 71.94, "step": 1425, "train_speed(iter/s)": 0.020157 }, { "epoch": 0.5920099358310908, "grad_norm": 1.8232818202515155, "learning_rate": 7.259338313767344e-06, "loss": 0.76410437, "memory(GiB)": 71.94, "step": 1430, "train_speed(iter/s)": 0.020159 }, { "epoch": 0.5940799006416891, "grad_norm": 1.6871789120586522, "learning_rate": 7.248665955176094e-06, "loss": 0.76673613, "memory(GiB)": 71.94, "step": 1435, "train_speed(iter/s)": 0.02016 }, { "epoch": 0.5961498654522873, "grad_norm": 1.9181669169557467, "learning_rate": 7.237993596584846e-06, "loss": 0.73530726, "memory(GiB)": 71.94, "step": 1440, "train_speed(iter/s)": 0.020161 }, { "epoch": 0.5982198302628855, "grad_norm": 2.0425311715534513, "learning_rate": 7.227321237993598e-06, "loss": 0.78409719, "memory(GiB)": 71.94, "step": 1445, "train_speed(iter/s)": 0.020163 }, { "epoch": 0.6002897950734838, "grad_norm": 1.725457162133973, "learning_rate": 7.216648879402348e-06, "loss": 0.74391842, "memory(GiB)": 71.94, "step": 1450, "train_speed(iter/s)": 0.020163 }, { "epoch": 0.602359759884082, "grad_norm": 2.2362927243629613, "learning_rate": 7.2059765208111e-06, "loss": 0.77035971, "memory(GiB)": 71.94, "step": 1455, "train_speed(iter/s)": 0.020166 }, { "epoch": 0.6044297246946801, "grad_norm": 2.379202645179455, "learning_rate": 7.195304162219852e-06, "loss": 0.74266062, "memory(GiB)": 71.94, "step": 1460, "train_speed(iter/s)": 0.020168 }, { "epoch": 0.6064996895052784, "grad_norm": 1.6006607749389805, "learning_rate": 7.184631803628602e-06, "loss": 0.76957574, "memory(GiB)": 71.94, "step": 1465, "train_speed(iter/s)": 0.020169 }, { "epoch": 0.6085696543158766, "grad_norm": 1.7633012594109296, "learning_rate": 7.173959445037354e-06, "loss": 0.77078071, "memory(GiB)": 71.94, "step": 1470, "train_speed(iter/s)": 0.02017 }, { "epoch": 0.6106396191264748, "grad_norm": 1.6009632285824897, "learning_rate": 7.163287086446106e-06, "loss": 0.7669549, "memory(GiB)": 71.94, "step": 1475, "train_speed(iter/s)": 0.020171 }, { "epoch": 0.6127095839370731, "grad_norm": 1.932344117154099, "learning_rate": 7.152614727854857e-06, "loss": 0.76528344, "memory(GiB)": 71.94, "step": 1480, "train_speed(iter/s)": 0.020172 }, { "epoch": 0.6147795487476713, "grad_norm": 2.02896587820159, "learning_rate": 7.141942369263608e-06, "loss": 0.75168095, "memory(GiB)": 71.94, "step": 1485, "train_speed(iter/s)": 0.020174 }, { "epoch": 0.6168495135582696, "grad_norm": 1.9974467066335662, "learning_rate": 7.131270010672359e-06, "loss": 0.7488194, "memory(GiB)": 71.94, "step": 1490, "train_speed(iter/s)": 0.020176 }, { "epoch": 0.6189194783688677, "grad_norm": 1.7902382164373858, "learning_rate": 7.120597652081111e-06, "loss": 0.75267982, "memory(GiB)": 71.94, "step": 1495, "train_speed(iter/s)": 0.020177 }, { "epoch": 0.6209894431794659, "grad_norm": 2.5929739472863838, "learning_rate": 7.1099252934898625e-06, "loss": 0.73765378, "memory(GiB)": 71.94, "step": 1500, "train_speed(iter/s)": 0.020178 }, { "epoch": 0.6209894431794659, "eval_loss": 0.8280953168869019, "eval_runtime": 333.1777, "eval_samples_per_second": 18.747, "eval_steps_per_second": 1.174, "step": 1500 }, { "epoch": 0.6230594079900642, "grad_norm": 1.9847247838289337, "learning_rate": 7.099252934898613e-06, "loss": 0.75908709, "memory(GiB)": 71.94, "step": 1505, "train_speed(iter/s)": 0.020077 }, { "epoch": 0.6251293728006624, "grad_norm": 1.801176606187135, "learning_rate": 7.088580576307365e-06, "loss": 0.78817225, "memory(GiB)": 71.94, "step": 1510, "train_speed(iter/s)": 0.020079 }, { "epoch": 0.6271993376112606, "grad_norm": 1.8685868923530962, "learning_rate": 7.0779082177161165e-06, "loss": 0.74608173, "memory(GiB)": 71.94, "step": 1515, "train_speed(iter/s)": 0.02008 }, { "epoch": 0.6292693024218589, "grad_norm": 1.9112331285520685, "learning_rate": 7.067235859124867e-06, "loss": 0.78047667, "memory(GiB)": 71.94, "step": 1520, "train_speed(iter/s)": 0.020081 }, { "epoch": 0.631339267232457, "grad_norm": 1.69066034933461, "learning_rate": 7.0565635005336185e-06, "loss": 0.75149813, "memory(GiB)": 71.94, "step": 1525, "train_speed(iter/s)": 0.02008 }, { "epoch": 0.6334092320430552, "grad_norm": 1.5394308738884603, "learning_rate": 7.0458911419423704e-06, "loss": 0.76838903, "memory(GiB)": 71.94, "step": 1530, "train_speed(iter/s)": 0.020081 }, { "epoch": 0.6354791968536535, "grad_norm": 1.7690454856119193, "learning_rate": 7.035218783351121e-06, "loss": 0.7524828, "memory(GiB)": 71.94, "step": 1535, "train_speed(iter/s)": 0.020083 }, { "epoch": 0.6375491616642517, "grad_norm": 2.0548651954814154, "learning_rate": 7.0245464247598725e-06, "loss": 0.76242485, "memory(GiB)": 71.94, "step": 1540, "train_speed(iter/s)": 0.020083 }, { "epoch": 0.63961912647485, "grad_norm": 1.5706142249866388, "learning_rate": 7.0138740661686235e-06, "loss": 0.78198986, "memory(GiB)": 71.94, "step": 1545, "train_speed(iter/s)": 0.020084 }, { "epoch": 0.6416890912854482, "grad_norm": 2.9215645341915275, "learning_rate": 7.0032017075773754e-06, "loss": 0.74943571, "memory(GiB)": 71.94, "step": 1550, "train_speed(iter/s)": 0.020086 }, { "epoch": 0.6437590560960463, "grad_norm": 1.983762828992232, "learning_rate": 6.9925293489861265e-06, "loss": 0.75862083, "memory(GiB)": 71.94, "step": 1555, "train_speed(iter/s)": 0.020087 }, { "epoch": 0.6458290209066446, "grad_norm": 1.8256568832087245, "learning_rate": 6.9818569903948775e-06, "loss": 0.75860863, "memory(GiB)": 71.94, "step": 1560, "train_speed(iter/s)": 0.020089 }, { "epoch": 0.6478989857172428, "grad_norm": 1.7082999758601491, "learning_rate": 6.971184631803629e-06, "loss": 0.78126869, "memory(GiB)": 71.94, "step": 1565, "train_speed(iter/s)": 0.02009 }, { "epoch": 0.649968950527841, "grad_norm": 2.388449730753909, "learning_rate": 6.960512273212381e-06, "loss": 0.76754818, "memory(GiB)": 71.94, "step": 1570, "train_speed(iter/s)": 0.020091 }, { "epoch": 0.6520389153384393, "grad_norm": 2.570709108294184, "learning_rate": 6.9498399146211315e-06, "loss": 0.74245424, "memory(GiB)": 71.94, "step": 1575, "train_speed(iter/s)": 0.020093 }, { "epoch": 0.6541088801490375, "grad_norm": 1.942646301485773, "learning_rate": 6.939167556029883e-06, "loss": 0.77264175, "memory(GiB)": 71.94, "step": 1580, "train_speed(iter/s)": 0.020094 }, { "epoch": 0.6561788449596356, "grad_norm": 1.7902561992868253, "learning_rate": 6.928495197438635e-06, "loss": 0.76845627, "memory(GiB)": 71.94, "step": 1585, "train_speed(iter/s)": 0.020096 }, { "epoch": 0.6582488097702339, "grad_norm": 1.8253085296468832, "learning_rate": 6.9178228388473854e-06, "loss": 0.75771255, "memory(GiB)": 71.94, "step": 1590, "train_speed(iter/s)": 0.020097 }, { "epoch": 0.6603187745808321, "grad_norm": 1.8440321320283706, "learning_rate": 6.907150480256137e-06, "loss": 0.74345121, "memory(GiB)": 71.94, "step": 1595, "train_speed(iter/s)": 0.020096 }, { "epoch": 0.6623887393914304, "grad_norm": 1.8894245153228149, "learning_rate": 6.896478121664889e-06, "loss": 0.76188393, "memory(GiB)": 71.94, "step": 1600, "train_speed(iter/s)": 0.020098 }, { "epoch": 0.6644587042020286, "grad_norm": 2.055312059883184, "learning_rate": 6.885805763073639e-06, "loss": 0.75247483, "memory(GiB)": 71.94, "step": 1605, "train_speed(iter/s)": 0.020099 }, { "epoch": 0.6665286690126268, "grad_norm": 1.8131130404445874, "learning_rate": 6.875133404482391e-06, "loss": 0.77208357, "memory(GiB)": 71.94, "step": 1610, "train_speed(iter/s)": 0.0201 }, { "epoch": 0.668598633823225, "grad_norm": 1.721876122278255, "learning_rate": 6.864461045891142e-06, "loss": 0.7271523, "memory(GiB)": 71.94, "step": 1615, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.6706685986338232, "grad_norm": 1.880195637198508, "learning_rate": 6.853788687299893e-06, "loss": 0.74756432, "memory(GiB)": 71.94, "step": 1620, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.6727385634444214, "grad_norm": 1.5895741550527986, "learning_rate": 6.843116328708645e-06, "loss": 0.77960148, "memory(GiB)": 71.94, "step": 1625, "train_speed(iter/s)": 0.020102 }, { "epoch": 0.6748085282550197, "grad_norm": 1.762407790878705, "learning_rate": 6.832443970117396e-06, "loss": 0.76564293, "memory(GiB)": 71.94, "step": 1630, "train_speed(iter/s)": 0.020104 }, { "epoch": 0.6768784930656179, "grad_norm": 1.85152265865232, "learning_rate": 6.821771611526148e-06, "loss": 0.75755472, "memory(GiB)": 71.94, "step": 1635, "train_speed(iter/s)": 0.020104 }, { "epoch": 0.678948457876216, "grad_norm": 2.4048195292609464, "learning_rate": 6.811099252934899e-06, "loss": 0.75722828, "memory(GiB)": 71.94, "step": 1640, "train_speed(iter/s)": 0.020106 }, { "epoch": 0.6810184226868143, "grad_norm": 1.7461290969223273, "learning_rate": 6.80042689434365e-06, "loss": 0.74719772, "memory(GiB)": 71.94, "step": 1645, "train_speed(iter/s)": 0.020107 }, { "epoch": 0.6830883874974125, "grad_norm": 1.7606017366047548, "learning_rate": 6.789754535752402e-06, "loss": 0.74246426, "memory(GiB)": 71.94, "step": 1650, "train_speed(iter/s)": 0.020108 }, { "epoch": 0.6851583523080108, "grad_norm": 2.3484261163252884, "learning_rate": 6.779082177161154e-06, "loss": 0.7567915, "memory(GiB)": 71.94, "step": 1655, "train_speed(iter/s)": 0.020109 }, { "epoch": 0.687228317118609, "grad_norm": 1.686698632081635, "learning_rate": 6.768409818569904e-06, "loss": 0.73500414, "memory(GiB)": 71.94, "step": 1660, "train_speed(iter/s)": 0.020111 }, { "epoch": 0.6892982819292072, "grad_norm": 1.9785908023609375, "learning_rate": 6.757737459978656e-06, "loss": 0.7035881, "memory(GiB)": 71.94, "step": 1665, "train_speed(iter/s)": 0.020112 }, { "epoch": 0.6913682467398055, "grad_norm": 1.8288827641332985, "learning_rate": 6.747065101387406e-06, "loss": 0.74135156, "memory(GiB)": 71.94, "step": 1670, "train_speed(iter/s)": 0.020113 }, { "epoch": 0.6934382115504036, "grad_norm": 2.106219884662748, "learning_rate": 6.736392742796158e-06, "loss": 0.77649341, "memory(GiB)": 71.94, "step": 1675, "train_speed(iter/s)": 0.020115 }, { "epoch": 0.6955081763610018, "grad_norm": 1.857981089347382, "learning_rate": 6.72572038420491e-06, "loss": 0.73271251, "memory(GiB)": 71.94, "step": 1680, "train_speed(iter/s)": 0.020116 }, { "epoch": 0.6975781411716001, "grad_norm": 1.8252469751324223, "learning_rate": 6.715048025613661e-06, "loss": 0.76072979, "memory(GiB)": 71.94, "step": 1685, "train_speed(iter/s)": 0.020117 }, { "epoch": 0.6996481059821983, "grad_norm": 1.9875787155351985, "learning_rate": 6.704375667022412e-06, "loss": 0.73438239, "memory(GiB)": 71.94, "step": 1690, "train_speed(iter/s)": 0.020118 }, { "epoch": 0.7017180707927966, "grad_norm": 1.685302389303672, "learning_rate": 6.693703308431164e-06, "loss": 0.76429882, "memory(GiB)": 71.94, "step": 1695, "train_speed(iter/s)": 0.020119 }, { "epoch": 0.7037880356033948, "grad_norm": 2.30374573526697, "learning_rate": 6.683030949839915e-06, "loss": 0.7437336, "memory(GiB)": 71.94, "step": 1700, "train_speed(iter/s)": 0.020119 }, { "epoch": 0.7058580004139929, "grad_norm": 2.4577356463267104, "learning_rate": 6.672358591248667e-06, "loss": 0.79406719, "memory(GiB)": 71.94, "step": 1705, "train_speed(iter/s)": 0.02012 }, { "epoch": 0.7079279652245912, "grad_norm": 1.7489028068953179, "learning_rate": 6.661686232657418e-06, "loss": 0.75482893, "memory(GiB)": 71.94, "step": 1710, "train_speed(iter/s)": 0.020121 }, { "epoch": 0.7099979300351894, "grad_norm": 1.929474801980816, "learning_rate": 6.651013874066169e-06, "loss": 0.74136767, "memory(GiB)": 71.94, "step": 1715, "train_speed(iter/s)": 0.020122 }, { "epoch": 0.7120678948457876, "grad_norm": 1.6790168198096502, "learning_rate": 6.640341515474921e-06, "loss": 0.72392588, "memory(GiB)": 71.94, "step": 1720, "train_speed(iter/s)": 0.020122 }, { "epoch": 0.7141378596563859, "grad_norm": 2.2963610149429488, "learning_rate": 6.629669156883671e-06, "loss": 0.7462635, "memory(GiB)": 71.94, "step": 1725, "train_speed(iter/s)": 0.020123 }, { "epoch": 0.716207824466984, "grad_norm": 4.289784718847475, "learning_rate": 6.618996798292423e-06, "loss": 0.73541126, "memory(GiB)": 71.94, "step": 1730, "train_speed(iter/s)": 0.020124 }, { "epoch": 0.7182777892775822, "grad_norm": 2.1972884462976263, "learning_rate": 6.608324439701175e-06, "loss": 0.7353497, "memory(GiB)": 71.94, "step": 1735, "train_speed(iter/s)": 0.020124 }, { "epoch": 0.7203477540881805, "grad_norm": 2.1738189409828377, "learning_rate": 6.597652081109925e-06, "loss": 0.71738148, "memory(GiB)": 71.94, "step": 1740, "train_speed(iter/s)": 0.020126 }, { "epoch": 0.7224177188987787, "grad_norm": 1.6342074890059992, "learning_rate": 6.586979722518677e-06, "loss": 0.75047336, "memory(GiB)": 71.94, "step": 1745, "train_speed(iter/s)": 0.020126 }, { "epoch": 0.724487683709377, "grad_norm": 1.7007570391919413, "learning_rate": 6.576307363927429e-06, "loss": 0.73253298, "memory(GiB)": 71.94, "step": 1750, "train_speed(iter/s)": 0.020127 }, { "epoch": 0.7265576485199752, "grad_norm": 1.5323053950217638, "learning_rate": 6.56563500533618e-06, "loss": 0.74062099, "memory(GiB)": 71.94, "step": 1755, "train_speed(iter/s)": 0.020128 }, { "epoch": 0.7286276133305734, "grad_norm": 1.9624071404199714, "learning_rate": 6.554962646744931e-06, "loss": 0.76860294, "memory(GiB)": 71.94, "step": 1760, "train_speed(iter/s)": 0.02013 }, { "epoch": 0.7306975781411716, "grad_norm": 1.8145041855689747, "learning_rate": 6.544290288153683e-06, "loss": 0.72403975, "memory(GiB)": 71.94, "step": 1765, "train_speed(iter/s)": 0.020131 }, { "epoch": 0.7327675429517698, "grad_norm": 1.7793196071264126, "learning_rate": 6.533617929562434e-06, "loss": 0.76407566, "memory(GiB)": 71.94, "step": 1770, "train_speed(iter/s)": 0.020132 }, { "epoch": 0.734837507762368, "grad_norm": 1.8806974947113853, "learning_rate": 6.522945570971186e-06, "loss": 0.73674612, "memory(GiB)": 71.94, "step": 1775, "train_speed(iter/s)": 0.020133 }, { "epoch": 0.7369074725729663, "grad_norm": 1.6203999356727887, "learning_rate": 6.512273212379937e-06, "loss": 0.73720975, "memory(GiB)": 71.94, "step": 1780, "train_speed(iter/s)": 0.020135 }, { "epoch": 0.7389774373835645, "grad_norm": 1.8256501665131275, "learning_rate": 6.501600853788688e-06, "loss": 0.74560528, "memory(GiB)": 71.94, "step": 1785, "train_speed(iter/s)": 0.020136 }, { "epoch": 0.7410474021941627, "grad_norm": 2.3313828860511294, "learning_rate": 6.49092849519744e-06, "loss": 0.73726311, "memory(GiB)": 71.94, "step": 1790, "train_speed(iter/s)": 0.020137 }, { "epoch": 0.7431173670047609, "grad_norm": 2.01967250245603, "learning_rate": 6.48025613660619e-06, "loss": 0.72599983, "memory(GiB)": 71.94, "step": 1795, "train_speed(iter/s)": 0.020138 }, { "epoch": 0.7451873318153591, "grad_norm": 1.5873931113082191, "learning_rate": 6.469583778014942e-06, "loss": 0.72361288, "memory(GiB)": 71.94, "step": 1800, "train_speed(iter/s)": 0.020139 }, { "epoch": 0.7451873318153591, "eval_loss": 0.821691632270813, "eval_runtime": 333.5584, "eval_samples_per_second": 18.725, "eval_steps_per_second": 1.172, "step": 1800 }, { "epoch": 0.7472572966259574, "grad_norm": 1.6284031081935817, "learning_rate": 6.458911419423694e-06, "loss": 0.73620996, "memory(GiB)": 71.94, "step": 1805, "train_speed(iter/s)": 0.020055 }, { "epoch": 0.7493272614365556, "grad_norm": 1.8143587680892548, "learning_rate": 6.448239060832444e-06, "loss": 0.73436375, "memory(GiB)": 71.94, "step": 1810, "train_speed(iter/s)": 0.020056 }, { "epoch": 0.7513972262471538, "grad_norm": 1.9502300953167138, "learning_rate": 6.437566702241196e-06, "loss": 0.74698448, "memory(GiB)": 71.94, "step": 1815, "train_speed(iter/s)": 0.020057 }, { "epoch": 0.7534671910577521, "grad_norm": 1.72820315761644, "learning_rate": 6.426894343649948e-06, "loss": 0.75661831, "memory(GiB)": 71.94, "step": 1820, "train_speed(iter/s)": 0.020059 }, { "epoch": 0.7555371558683502, "grad_norm": 1.8039931329917855, "learning_rate": 6.416221985058698e-06, "loss": 0.74954405, "memory(GiB)": 71.94, "step": 1825, "train_speed(iter/s)": 0.020059 }, { "epoch": 0.7576071206789484, "grad_norm": 2.0711566925028433, "learning_rate": 6.40554962646745e-06, "loss": 0.73745756, "memory(GiB)": 71.94, "step": 1830, "train_speed(iter/s)": 0.020061 }, { "epoch": 0.7596770854895467, "grad_norm": 1.826394447351557, "learning_rate": 6.3948772678762016e-06, "loss": 0.73249888, "memory(GiB)": 71.94, "step": 1835, "train_speed(iter/s)": 0.020062 }, { "epoch": 0.7617470503001449, "grad_norm": 2.4929262063136175, "learning_rate": 6.384204909284953e-06, "loss": 0.72507439, "memory(GiB)": 71.94, "step": 1840, "train_speed(iter/s)": 0.020062 }, { "epoch": 0.7638170151107431, "grad_norm": 1.7606115982834467, "learning_rate": 6.373532550693704e-06, "loss": 0.72618227, "memory(GiB)": 71.94, "step": 1845, "train_speed(iter/s)": 0.020064 }, { "epoch": 0.7658869799213414, "grad_norm": 1.7780633999979434, "learning_rate": 6.362860192102455e-06, "loss": 0.73577065, "memory(GiB)": 71.94, "step": 1850, "train_speed(iter/s)": 0.020064 }, { "epoch": 0.7679569447319395, "grad_norm": 2.2901271183050294, "learning_rate": 6.3521878335112066e-06, "loss": 0.75972033, "memory(GiB)": 71.94, "step": 1855, "train_speed(iter/s)": 0.020065 }, { "epoch": 0.7700269095425378, "grad_norm": 1.974709118996213, "learning_rate": 6.3415154749199585e-06, "loss": 0.7513834, "memory(GiB)": 71.94, "step": 1860, "train_speed(iter/s)": 0.020065 }, { "epoch": 0.772096874353136, "grad_norm": 1.7795619053561953, "learning_rate": 6.330843116328709e-06, "loss": 0.70549402, "memory(GiB)": 71.94, "step": 1865, "train_speed(iter/s)": 0.020066 }, { "epoch": 0.7741668391637342, "grad_norm": 1.7925210555170064, "learning_rate": 6.3201707577374605e-06, "loss": 0.75538568, "memory(GiB)": 71.94, "step": 1870, "train_speed(iter/s)": 0.020067 }, { "epoch": 0.7762368039743325, "grad_norm": 1.7592519440523378, "learning_rate": 6.309498399146212e-06, "loss": 0.74328322, "memory(GiB)": 71.94, "step": 1875, "train_speed(iter/s)": 0.020068 }, { "epoch": 0.7783067687849307, "grad_norm": 1.7630568722667896, "learning_rate": 6.298826040554963e-06, "loss": 0.73995676, "memory(GiB)": 71.94, "step": 1880, "train_speed(iter/s)": 0.020069 }, { "epoch": 0.7803767335955288, "grad_norm": 1.732047211183728, "learning_rate": 6.2881536819637145e-06, "loss": 0.70023623, "memory(GiB)": 71.94, "step": 1885, "train_speed(iter/s)": 0.02007 }, { "epoch": 0.7824466984061271, "grad_norm": 1.7004814074017778, "learning_rate": 6.277481323372466e-06, "loss": 0.73256617, "memory(GiB)": 71.94, "step": 1890, "train_speed(iter/s)": 0.020071 }, { "epoch": 0.7845166632167253, "grad_norm": 1.7170761577962488, "learning_rate": 6.2668089647812166e-06, "loss": 0.71450982, "memory(GiB)": 71.94, "step": 1895, "train_speed(iter/s)": 0.020071 }, { "epoch": 0.7865866280273235, "grad_norm": 2.0086204171681565, "learning_rate": 6.2561366061899685e-06, "loss": 0.7096673, "memory(GiB)": 71.94, "step": 1900, "train_speed(iter/s)": 0.020072 }, { "epoch": 0.7886565928379218, "grad_norm": 1.6109075949007228, "learning_rate": 6.2454642475987195e-06, "loss": 0.7170536, "memory(GiB)": 71.94, "step": 1905, "train_speed(iter/s)": 0.020073 }, { "epoch": 0.79072655764852, "grad_norm": 1.6468982825229455, "learning_rate": 6.234791889007471e-06, "loss": 0.7817646, "memory(GiB)": 71.94, "step": 1910, "train_speed(iter/s)": 0.020073 }, { "epoch": 0.7927965224591182, "grad_norm": 1.8405361482523723, "learning_rate": 6.224119530416222e-06, "loss": 0.71389322, "memory(GiB)": 71.94, "step": 1915, "train_speed(iter/s)": 0.020075 }, { "epoch": 0.7948664872697164, "grad_norm": 1.7937559729338877, "learning_rate": 6.2134471718249735e-06, "loss": 0.72494421, "memory(GiB)": 71.94, "step": 1920, "train_speed(iter/s)": 0.020076 }, { "epoch": 0.7969364520803146, "grad_norm": 1.98762799360225, "learning_rate": 6.202774813233725e-06, "loss": 0.75637407, "memory(GiB)": 71.94, "step": 1925, "train_speed(iter/s)": 0.020077 }, { "epoch": 0.7990064168909129, "grad_norm": 2.469167716565665, "learning_rate": 6.192102454642477e-06, "loss": 0.71725979, "memory(GiB)": 71.94, "step": 1930, "train_speed(iter/s)": 0.020078 }, { "epoch": 0.8010763817015111, "grad_norm": 1.6526117746871118, "learning_rate": 6.181430096051227e-06, "loss": 0.73172369, "memory(GiB)": 71.94, "step": 1935, "train_speed(iter/s)": 0.020079 }, { "epoch": 0.8031463465121093, "grad_norm": 1.8881085526929478, "learning_rate": 6.170757737459979e-06, "loss": 0.68869176, "memory(GiB)": 71.94, "step": 1940, "train_speed(iter/s)": 0.02008 }, { "epoch": 0.8052163113227075, "grad_norm": 2.1341112552467107, "learning_rate": 6.160085378868731e-06, "loss": 0.74425526, "memory(GiB)": 71.94, "step": 1945, "train_speed(iter/s)": 0.020081 }, { "epoch": 0.8072862761333057, "grad_norm": 2.1587279161379906, "learning_rate": 6.149413020277481e-06, "loss": 0.72686901, "memory(GiB)": 71.94, "step": 1950, "train_speed(iter/s)": 0.020081 }, { "epoch": 0.8093562409439039, "grad_norm": 1.7183166805211196, "learning_rate": 6.138740661686233e-06, "loss": 0.70801015, "memory(GiB)": 71.94, "step": 1955, "train_speed(iter/s)": 0.020082 }, { "epoch": 0.8114262057545022, "grad_norm": 1.6918548000232365, "learning_rate": 6.128068303094985e-06, "loss": 0.71407743, "memory(GiB)": 71.94, "step": 1960, "train_speed(iter/s)": 0.020083 }, { "epoch": 0.8134961705651004, "grad_norm": 1.6783017839137153, "learning_rate": 6.117395944503735e-06, "loss": 0.74968853, "memory(GiB)": 71.94, "step": 1965, "train_speed(iter/s)": 0.020083 }, { "epoch": 0.8155661353756987, "grad_norm": 1.7117864547494115, "learning_rate": 6.106723585912487e-06, "loss": 0.71134882, "memory(GiB)": 71.94, "step": 1970, "train_speed(iter/s)": 0.020085 }, { "epoch": 0.8176361001862968, "grad_norm": 1.9043254368072582, "learning_rate": 6.096051227321238e-06, "loss": 0.74376335, "memory(GiB)": 71.94, "step": 1975, "train_speed(iter/s)": 0.020085 }, { "epoch": 0.819706064996895, "grad_norm": 1.8416766338299921, "learning_rate": 6.08537886872999e-06, "loss": 0.74048576, "memory(GiB)": 71.94, "step": 1980, "train_speed(iter/s)": 0.020086 }, { "epoch": 0.8217760298074933, "grad_norm": 1.9993092375152783, "learning_rate": 6.074706510138741e-06, "loss": 0.73070145, "memory(GiB)": 71.94, "step": 1985, "train_speed(iter/s)": 0.020087 }, { "epoch": 0.8238459946180915, "grad_norm": 1.9255287807426156, "learning_rate": 6.064034151547492e-06, "loss": 0.71732969, "memory(GiB)": 71.94, "step": 1990, "train_speed(iter/s)": 0.020088 }, { "epoch": 0.8259159594286897, "grad_norm": 1.699758287154301, "learning_rate": 6.053361792956244e-06, "loss": 0.70977745, "memory(GiB)": 71.94, "step": 1995, "train_speed(iter/s)": 0.020089 }, { "epoch": 0.827985924239288, "grad_norm": 1.7447876663920783, "learning_rate": 6.042689434364995e-06, "loss": 0.72701621, "memory(GiB)": 71.94, "step": 2000, "train_speed(iter/s)": 0.02009 }, { "epoch": 0.8300558890498861, "grad_norm": 1.8875191736470693, "learning_rate": 6.032017075773746e-06, "loss": 0.72644696, "memory(GiB)": 71.94, "step": 2005, "train_speed(iter/s)": 0.02009 }, { "epoch": 0.8321258538604843, "grad_norm": 1.6472530019204896, "learning_rate": 6.021344717182498e-06, "loss": 0.70005097, "memory(GiB)": 71.94, "step": 2010, "train_speed(iter/s)": 0.020091 }, { "epoch": 0.8341958186710826, "grad_norm": 2.2818513681921266, "learning_rate": 6.01067235859125e-06, "loss": 0.72365198, "memory(GiB)": 71.94, "step": 2015, "train_speed(iter/s)": 0.020092 }, { "epoch": 0.8362657834816808, "grad_norm": 1.8205665836409684, "learning_rate": 6e-06, "loss": 0.7322978, "memory(GiB)": 71.94, "step": 2020, "train_speed(iter/s)": 0.020092 }, { "epoch": 0.8383357482922791, "grad_norm": 1.980690292432822, "learning_rate": 5.989327641408752e-06, "loss": 0.73382425, "memory(GiB)": 71.94, "step": 2025, "train_speed(iter/s)": 0.020093 }, { "epoch": 0.8404057131028773, "grad_norm": 1.7748584169287815, "learning_rate": 5.978655282817502e-06, "loss": 0.7543438, "memory(GiB)": 71.94, "step": 2030, "train_speed(iter/s)": 0.020094 }, { "epoch": 0.8424756779134754, "grad_norm": 1.9477910790390784, "learning_rate": 5.967982924226254e-06, "loss": 0.73893185, "memory(GiB)": 71.94, "step": 2035, "train_speed(iter/s)": 0.020095 }, { "epoch": 0.8445456427240737, "grad_norm": 1.5695526526206898, "learning_rate": 5.957310565635006e-06, "loss": 0.71403108, "memory(GiB)": 71.94, "step": 2040, "train_speed(iter/s)": 0.020096 }, { "epoch": 0.8466156075346719, "grad_norm": 2.1517602299856557, "learning_rate": 5.946638207043757e-06, "loss": 0.71713848, "memory(GiB)": 71.94, "step": 2045, "train_speed(iter/s)": 0.020097 }, { "epoch": 0.8486855723452701, "grad_norm": 2.739525728221928, "learning_rate": 5.935965848452508e-06, "loss": 0.72253714, "memory(GiB)": 71.94, "step": 2050, "train_speed(iter/s)": 0.020099 }, { "epoch": 0.8507555371558684, "grad_norm": 1.9454548994868823, "learning_rate": 5.92529348986126e-06, "loss": 0.74796634, "memory(GiB)": 71.94, "step": 2055, "train_speed(iter/s)": 0.0201 }, { "epoch": 0.8528255019664666, "grad_norm": 1.7996528216814918, "learning_rate": 5.914621131270011e-06, "loss": 0.71262107, "memory(GiB)": 71.94, "step": 2060, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.8548954667770647, "grad_norm": 1.9001952709840753, "learning_rate": 5.903948772678763e-06, "loss": 0.71814876, "memory(GiB)": 71.94, "step": 2065, "train_speed(iter/s)": 0.020101 }, { "epoch": 0.856965431587663, "grad_norm": 3.4631327121051036, "learning_rate": 5.893276414087514e-06, "loss": 0.70359154, "memory(GiB)": 71.94, "step": 2070, "train_speed(iter/s)": 0.020103 }, { "epoch": 0.8590353963982612, "grad_norm": 1.9078035163840932, "learning_rate": 5.882604055496265e-06, "loss": 0.74100571, "memory(GiB)": 71.94, "step": 2075, "train_speed(iter/s)": 0.020103 }, { "epoch": 0.8611053612088595, "grad_norm": 2.7698267455997576, "learning_rate": 5.871931696905017e-06, "loss": 0.71972656, "memory(GiB)": 71.94, "step": 2080, "train_speed(iter/s)": 0.020104 }, { "epoch": 0.8631753260194577, "grad_norm": 1.9640858230267009, "learning_rate": 5.861259338313769e-06, "loss": 0.71868258, "memory(GiB)": 71.94, "step": 2085, "train_speed(iter/s)": 0.020105 }, { "epoch": 0.8652452908300559, "grad_norm": 1.9593324236104832, "learning_rate": 5.850586979722519e-06, "loss": 0.70707326, "memory(GiB)": 71.94, "step": 2090, "train_speed(iter/s)": 0.020106 }, { "epoch": 0.8673152556406541, "grad_norm": 2.0337621679872946, "learning_rate": 5.839914621131271e-06, "loss": 0.7303226, "memory(GiB)": 71.94, "step": 2095, "train_speed(iter/s)": 0.020107 }, { "epoch": 0.8693852204512523, "grad_norm": 1.7464491411508778, "learning_rate": 5.829242262540021e-06, "loss": 0.70045071, "memory(GiB)": 71.94, "step": 2100, "train_speed(iter/s)": 0.020108 }, { "epoch": 0.8693852204512523, "eval_loss": 0.8155556321144104, "eval_runtime": 334.2741, "eval_samples_per_second": 18.685, "eval_steps_per_second": 1.17, "step": 2100 }, { "epoch": 0.8714551852618505, "grad_norm": 1.893698412703024, "learning_rate": 5.818569903948773e-06, "loss": 0.71434135, "memory(GiB)": 71.94, "step": 2105, "train_speed(iter/s)": 0.020036 }, { "epoch": 0.8735251500724488, "grad_norm": 1.8915863471870793, "learning_rate": 5.807897545357525e-06, "loss": 0.73264565, "memory(GiB)": 71.94, "step": 2110, "train_speed(iter/s)": 0.020037 }, { "epoch": 0.875595114883047, "grad_norm": 1.5909748405215243, "learning_rate": 5.797225186766276e-06, "loss": 0.6809236, "memory(GiB)": 71.94, "step": 2115, "train_speed(iter/s)": 0.020037 }, { "epoch": 0.8776650796936452, "grad_norm": 2.298029244846505, "learning_rate": 5.786552828175027e-06, "loss": 0.72660675, "memory(GiB)": 71.94, "step": 2120, "train_speed(iter/s)": 0.020038 }, { "epoch": 0.8797350445042434, "grad_norm": 2.4471148736933634, "learning_rate": 5.775880469583779e-06, "loss": 0.71458502, "memory(GiB)": 71.94, "step": 2125, "train_speed(iter/s)": 0.020039 }, { "epoch": 0.8818050093148416, "grad_norm": 2.237641446928337, "learning_rate": 5.76520811099253e-06, "loss": 0.72527971, "memory(GiB)": 71.94, "step": 2130, "train_speed(iter/s)": 0.02004 }, { "epoch": 0.8838749741254399, "grad_norm": 2.1373854033173667, "learning_rate": 5.754535752401282e-06, "loss": 0.68969226, "memory(GiB)": 71.94, "step": 2135, "train_speed(iter/s)": 0.020041 }, { "epoch": 0.8859449389360381, "grad_norm": 2.4814505236104254, "learning_rate": 5.743863393810033e-06, "loss": 0.72110472, "memory(GiB)": 71.94, "step": 2140, "train_speed(iter/s)": 0.020042 }, { "epoch": 0.8880149037466363, "grad_norm": 1.8846949427239792, "learning_rate": 5.733191035218784e-06, "loss": 0.71526957, "memory(GiB)": 71.94, "step": 2145, "train_speed(iter/s)": 0.020043 }, { "epoch": 0.8900848685572346, "grad_norm": 1.6414510600406789, "learning_rate": 5.722518676627536e-06, "loss": 0.72875662, "memory(GiB)": 71.94, "step": 2150, "train_speed(iter/s)": 0.020043 }, { "epoch": 0.8921548333678327, "grad_norm": 2.027904279012944, "learning_rate": 5.711846318036286e-06, "loss": 0.73166924, "memory(GiB)": 71.94, "step": 2155, "train_speed(iter/s)": 0.020044 }, { "epoch": 0.8942247981784309, "grad_norm": 2.0007554119149136, "learning_rate": 5.701173959445038e-06, "loss": 0.69900818, "memory(GiB)": 71.94, "step": 2160, "train_speed(iter/s)": 0.020046 }, { "epoch": 0.8962947629890292, "grad_norm": 1.732929144774659, "learning_rate": 5.69050160085379e-06, "loss": 0.70091386, "memory(GiB)": 71.94, "step": 2165, "train_speed(iter/s)": 0.020046 }, { "epoch": 0.8983647277996274, "grad_norm": 1.6575807865879337, "learning_rate": 5.67982924226254e-06, "loss": 0.70530014, "memory(GiB)": 71.94, "step": 2170, "train_speed(iter/s)": 0.020047 }, { "epoch": 0.9004346926102256, "grad_norm": 1.8402505669080536, "learning_rate": 5.669156883671292e-06, "loss": 0.72022433, "memory(GiB)": 71.94, "step": 2175, "train_speed(iter/s)": 0.020048 }, { "epoch": 0.9025046574208239, "grad_norm": 1.8122270786550385, "learning_rate": 5.6584845250800435e-06, "loss": 0.67802505, "memory(GiB)": 71.94, "step": 2180, "train_speed(iter/s)": 0.020048 }, { "epoch": 0.904574622231422, "grad_norm": 1.7969445274298348, "learning_rate": 5.647812166488794e-06, "loss": 0.70636382, "memory(GiB)": 71.94, "step": 2185, "train_speed(iter/s)": 0.020049 }, { "epoch": 0.9066445870420203, "grad_norm": 2.1142537074713412, "learning_rate": 5.637139807897546e-06, "loss": 0.71558409, "memory(GiB)": 71.94, "step": 2190, "train_speed(iter/s)": 0.02005 }, { "epoch": 0.9087145518526185, "grad_norm": 1.944413383820558, "learning_rate": 5.6264674493062975e-06, "loss": 0.72004523, "memory(GiB)": 71.94, "step": 2195, "train_speed(iter/s)": 0.020051 }, { "epoch": 0.9107845166632167, "grad_norm": 2.3051719970776934, "learning_rate": 5.6157950907150485e-06, "loss": 0.7308382, "memory(GiB)": 71.94, "step": 2200, "train_speed(iter/s)": 0.020052 }, { "epoch": 0.912854481473815, "grad_norm": 2.27411033803044, "learning_rate": 5.6051227321238e-06, "loss": 0.74844613, "memory(GiB)": 71.94, "step": 2205, "train_speed(iter/s)": 0.020053 }, { "epoch": 0.9149244462844132, "grad_norm": 1.8444250783225764, "learning_rate": 5.594450373532551e-06, "loss": 0.68941135, "memory(GiB)": 71.94, "step": 2210, "train_speed(iter/s)": 0.020053 }, { "epoch": 0.9169944110950113, "grad_norm": 1.8662325411124825, "learning_rate": 5.5837780149413025e-06, "loss": 0.73066435, "memory(GiB)": 71.94, "step": 2215, "train_speed(iter/s)": 0.020054 }, { "epoch": 0.9190643759056096, "grad_norm": 1.6833532844662813, "learning_rate": 5.573105656350054e-06, "loss": 0.7062602, "memory(GiB)": 71.94, "step": 2220, "train_speed(iter/s)": 0.020055 }, { "epoch": 0.9211343407162078, "grad_norm": 1.6070808318678096, "learning_rate": 5.562433297758805e-06, "loss": 0.69585543, "memory(GiB)": 71.94, "step": 2225, "train_speed(iter/s)": 0.020055 }, { "epoch": 0.923204305526806, "grad_norm": 2.0016548598313024, "learning_rate": 5.5517609391675565e-06, "loss": 0.7085475, "memory(GiB)": 71.94, "step": 2230, "train_speed(iter/s)": 0.020056 }, { "epoch": 0.9252742703374043, "grad_norm": 1.9201243304059477, "learning_rate": 5.541088580576308e-06, "loss": 0.71516237, "memory(GiB)": 71.94, "step": 2235, "train_speed(iter/s)": 0.020057 }, { "epoch": 0.9273442351480025, "grad_norm": 1.9055608045022892, "learning_rate": 5.5304162219850586e-06, "loss": 0.69740834, "memory(GiB)": 71.94, "step": 2240, "train_speed(iter/s)": 0.020058 }, { "epoch": 0.9294141999586008, "grad_norm": 2.0041753291659026, "learning_rate": 5.5197438633938104e-06, "loss": 0.71469507, "memory(GiB)": 71.94, "step": 2245, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.9314841647691989, "grad_norm": 2.025723530711083, "learning_rate": 5.509071504802562e-06, "loss": 0.69406614, "memory(GiB)": 71.94, "step": 2250, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.9335541295797971, "grad_norm": 1.7845786824484478, "learning_rate": 5.4983991462113125e-06, "loss": 0.7137248, "memory(GiB)": 71.94, "step": 2255, "train_speed(iter/s)": 0.020061 }, { "epoch": 0.9356240943903954, "grad_norm": 2.3504717810438196, "learning_rate": 5.487726787620064e-06, "loss": 0.70752907, "memory(GiB)": 71.94, "step": 2260, "train_speed(iter/s)": 0.020062 }, { "epoch": 0.9376940592009936, "grad_norm": 2.0225261644141797, "learning_rate": 5.477054429028816e-06, "loss": 0.70490198, "memory(GiB)": 71.94, "step": 2265, "train_speed(iter/s)": 0.020063 }, { "epoch": 0.9397640240115918, "grad_norm": 2.2107863770119747, "learning_rate": 5.466382070437567e-06, "loss": 0.71501665, "memory(GiB)": 71.94, "step": 2270, "train_speed(iter/s)": 0.020064 }, { "epoch": 0.94183398882219, "grad_norm": 1.9030684437330223, "learning_rate": 5.455709711846318e-06, "loss": 0.70587921, "memory(GiB)": 71.94, "step": 2275, "train_speed(iter/s)": 0.020064 }, { "epoch": 0.9439039536327882, "grad_norm": 1.8981878877998872, "learning_rate": 5.445037353255069e-06, "loss": 0.70294933, "memory(GiB)": 71.94, "step": 2280, "train_speed(iter/s)": 0.020064 }, { "epoch": 0.9459739184433864, "grad_norm": 1.7195321236677932, "learning_rate": 5.434364994663821e-06, "loss": 0.71657829, "memory(GiB)": 71.94, "step": 2285, "train_speed(iter/s)": 0.020065 }, { "epoch": 0.9480438832539847, "grad_norm": 1.6695574824900545, "learning_rate": 5.423692636072573e-06, "loss": 0.70917149, "memory(GiB)": 71.94, "step": 2290, "train_speed(iter/s)": 0.020065 }, { "epoch": 0.9501138480645829, "grad_norm": 1.7410897548688689, "learning_rate": 5.413020277481323e-06, "loss": 0.7276587, "memory(GiB)": 71.94, "step": 2295, "train_speed(iter/s)": 0.020066 }, { "epoch": 0.9521838128751812, "grad_norm": 1.6737135024901502, "learning_rate": 5.402347918890075e-06, "loss": 0.70822001, "memory(GiB)": 71.94, "step": 2300, "train_speed(iter/s)": 0.020066 }, { "epoch": 0.9542537776857793, "grad_norm": 1.7876076111815575, "learning_rate": 5.391675560298827e-06, "loss": 0.72815857, "memory(GiB)": 71.94, "step": 2305, "train_speed(iter/s)": 0.020067 }, { "epoch": 0.9563237424963775, "grad_norm": 1.653921158054905, "learning_rate": 5.381003201707577e-06, "loss": 0.715973, "memory(GiB)": 71.94, "step": 2310, "train_speed(iter/s)": 0.020068 }, { "epoch": 0.9583937073069758, "grad_norm": 1.7830026539914627, "learning_rate": 5.370330843116329e-06, "loss": 0.70955439, "memory(GiB)": 71.94, "step": 2315, "train_speed(iter/s)": 0.020068 }, { "epoch": 0.960463672117574, "grad_norm": 2.0794293583699712, "learning_rate": 5.359658484525081e-06, "loss": 0.71212959, "memory(GiB)": 71.94, "step": 2320, "train_speed(iter/s)": 0.020049 }, { "epoch": 0.9625336369281722, "grad_norm": 1.62651854843721, "learning_rate": 5.348986125933831e-06, "loss": 0.69347043, "memory(GiB)": 71.94, "step": 2325, "train_speed(iter/s)": 0.02005 }, { "epoch": 0.9646036017387705, "grad_norm": 1.6345937025216515, "learning_rate": 5.338313767342583e-06, "loss": 0.68745365, "memory(GiB)": 71.94, "step": 2330, "train_speed(iter/s)": 0.02005 }, { "epoch": 0.9666735665493686, "grad_norm": 2.1460062258102504, "learning_rate": 5.327641408751334e-06, "loss": 0.69751196, "memory(GiB)": 71.94, "step": 2335, "train_speed(iter/s)": 0.020051 }, { "epoch": 0.9687435313599668, "grad_norm": 1.8428729242460318, "learning_rate": 5.316969050160086e-06, "loss": 0.70150051, "memory(GiB)": 71.94, "step": 2340, "train_speed(iter/s)": 0.020052 }, { "epoch": 0.9708134961705651, "grad_norm": 1.870750640547904, "learning_rate": 5.306296691568837e-06, "loss": 0.67915797, "memory(GiB)": 71.94, "step": 2345, "train_speed(iter/s)": 0.020053 }, { "epoch": 0.9728834609811633, "grad_norm": 1.6984421405387677, "learning_rate": 5.295624332977588e-06, "loss": 0.71915126, "memory(GiB)": 71.94, "step": 2350, "train_speed(iter/s)": 0.020054 }, { "epoch": 0.9749534257917616, "grad_norm": 1.7839025594515001, "learning_rate": 5.28495197438634e-06, "loss": 0.69594321, "memory(GiB)": 71.94, "step": 2355, "train_speed(iter/s)": 0.020055 }, { "epoch": 0.9770233906023598, "grad_norm": 1.666815065361009, "learning_rate": 5.274279615795091e-06, "loss": 0.68858194, "memory(GiB)": 71.94, "step": 2360, "train_speed(iter/s)": 0.020055 }, { "epoch": 0.9790933554129579, "grad_norm": 1.6613287141536495, "learning_rate": 5.263607257203842e-06, "loss": 0.65968986, "memory(GiB)": 71.94, "step": 2365, "train_speed(iter/s)": 0.020056 }, { "epoch": 0.9811633202235562, "grad_norm": 1.5579649689164343, "learning_rate": 5.252934898612594e-06, "loss": 0.69255476, "memory(GiB)": 71.94, "step": 2370, "train_speed(iter/s)": 0.020057 }, { "epoch": 0.9832332850341544, "grad_norm": 1.7564735837589676, "learning_rate": 5.242262540021346e-06, "loss": 0.71395578, "memory(GiB)": 71.94, "step": 2375, "train_speed(iter/s)": 0.020057 }, { "epoch": 0.9853032498447526, "grad_norm": 2.0952603393058076, "learning_rate": 5.231590181430096e-06, "loss": 0.69916964, "memory(GiB)": 71.94, "step": 2380, "train_speed(iter/s)": 0.020058 }, { "epoch": 0.9873732146553509, "grad_norm": 1.851124802207451, "learning_rate": 5.220917822838848e-06, "loss": 0.6992053, "memory(GiB)": 71.94, "step": 2385, "train_speed(iter/s)": 0.020059 }, { "epoch": 0.9894431794659491, "grad_norm": 1.977421778833197, "learning_rate": 5.2102454642476e-06, "loss": 0.70937605, "memory(GiB)": 71.94, "step": 2390, "train_speed(iter/s)": 0.020059 }, { "epoch": 0.9915131442765474, "grad_norm": 1.8125821116129224, "learning_rate": 5.19957310565635e-06, "loss": 0.70189705, "memory(GiB)": 71.94, "step": 2395, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.9935831090871455, "grad_norm": 1.8812587805267227, "learning_rate": 5.188900747065102e-06, "loss": 0.68958111, "memory(GiB)": 71.94, "step": 2400, "train_speed(iter/s)": 0.02006 }, { "epoch": 0.9935831090871455, "eval_loss": 0.812356173992157, "eval_runtime": 333.5694, "eval_samples_per_second": 18.725, "eval_steps_per_second": 1.172, "step": 2400 }, { "epoch": 0.9956530738977437, "grad_norm": 1.77525759977015, "learning_rate": 5.178228388473853e-06, "loss": 0.71421309, "memory(GiB)": 71.94, "step": 2405, "train_speed(iter/s)": 0.019998 }, { "epoch": 0.997723038708342, "grad_norm": 1.9847604199741764, "learning_rate": 5.167556029882604e-06, "loss": 0.67816806, "memory(GiB)": 71.94, "step": 2410, "train_speed(iter/s)": 0.019999 }, { "epoch": 0.9997930035189402, "grad_norm": 1.8831401901251863, "learning_rate": 5.156883671291356e-06, "loss": 0.67297964, "memory(GiB)": 71.94, "step": 2415, "train_speed(iter/s)": 0.02 }, { "epoch": 1.0018629683295384, "grad_norm": 1.902600390439468, "learning_rate": 5.146211312700107e-06, "loss": 0.65476379, "memory(GiB)": 71.94, "step": 2420, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.0039329331401365, "grad_norm": 1.8958163712669238, "learning_rate": 5.135538954108859e-06, "loss": 0.69827843, "memory(GiB)": 71.94, "step": 2425, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.006002897950735, "grad_norm": 2.0893567670048774, "learning_rate": 5.12486659551761e-06, "loss": 0.6843009, "memory(GiB)": 71.94, "step": 2430, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.0080728627613331, "grad_norm": 1.7283831963515028, "learning_rate": 5.114194236926361e-06, "loss": 0.66953688, "memory(GiB)": 71.94, "step": 2435, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.0101428275719313, "grad_norm": 2.4550295961951023, "learning_rate": 5.103521878335113e-06, "loss": 0.6826447, "memory(GiB)": 71.94, "step": 2440, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.0122127923825295, "grad_norm": 2.065348792100011, "learning_rate": 5.092849519743865e-06, "loss": 0.71306543, "memory(GiB)": 71.94, "step": 2445, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.0142827571931277, "grad_norm": 1.8461278616269987, "learning_rate": 5.082177161152615e-06, "loss": 0.70268006, "memory(GiB)": 71.94, "step": 2450, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.0163527220037258, "grad_norm": 1.8474179385577107, "learning_rate": 5.071504802561367e-06, "loss": 0.68759146, "memory(GiB)": 71.94, "step": 2455, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.0184226868143242, "grad_norm": 2.1221766243412556, "learning_rate": 5.060832443970117e-06, "loss": 0.70161247, "memory(GiB)": 71.94, "step": 2460, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.0204926516249224, "grad_norm": 1.7992664488684018, "learning_rate": 5.050160085378869e-06, "loss": 0.70210953, "memory(GiB)": 71.94, "step": 2465, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.0225626164355206, "grad_norm": 2.010160051422228, "learning_rate": 5.039487726787621e-06, "loss": 0.70377779, "memory(GiB)": 71.94, "step": 2470, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.0246325812461188, "grad_norm": 1.677895969704363, "learning_rate": 5.028815368196372e-06, "loss": 0.71084547, "memory(GiB)": 71.94, "step": 2475, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.026702546056717, "grad_norm": 1.9057688334203953, "learning_rate": 5.018143009605123e-06, "loss": 0.67874904, "memory(GiB)": 71.94, "step": 2480, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.0287725108673154, "grad_norm": 1.826208938410727, "learning_rate": 5.007470651013875e-06, "loss": 0.67107067, "memory(GiB)": 71.94, "step": 2485, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.0308424756779135, "grad_norm": 2.0686042207681425, "learning_rate": 4.996798292422626e-06, "loss": 0.70525031, "memory(GiB)": 71.94, "step": 2490, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.0329124404885117, "grad_norm": 1.7733232551181717, "learning_rate": 4.986125933831378e-06, "loss": 0.68992233, "memory(GiB)": 71.94, "step": 2495, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.03498240529911, "grad_norm": 1.7950220673083168, "learning_rate": 4.975453575240129e-06, "loss": 0.67175922, "memory(GiB)": 71.94, "step": 2500, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.037052370109708, "grad_norm": 1.9232030990454307, "learning_rate": 4.96478121664888e-06, "loss": 0.70407228, "memory(GiB)": 71.94, "step": 2505, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.0391223349203063, "grad_norm": 1.708121386060437, "learning_rate": 4.9541088580576316e-06, "loss": 0.70078506, "memory(GiB)": 71.94, "step": 2510, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.0411922997309047, "grad_norm": 1.5750528842945233, "learning_rate": 4.943436499466383e-06, "loss": 0.66830945, "memory(GiB)": 71.94, "step": 2515, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.0432622645415028, "grad_norm": 1.8181518834205428, "learning_rate": 4.932764140875134e-06, "loss": 0.69259553, "memory(GiB)": 71.94, "step": 2520, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.045332229352101, "grad_norm": 1.551131375424082, "learning_rate": 4.9220917822838855e-06, "loss": 0.69300728, "memory(GiB)": 71.94, "step": 2525, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.0474021941626992, "grad_norm": 1.9636419108578083, "learning_rate": 4.9114194236926366e-06, "loss": 0.69065495, "memory(GiB)": 71.94, "step": 2530, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.0494721589732974, "grad_norm": 2.2096885596291918, "learning_rate": 4.900747065101388e-06, "loss": 0.72614369, "memory(GiB)": 71.94, "step": 2535, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.0515421237838958, "grad_norm": 1.8198162773587976, "learning_rate": 4.890074706510139e-06, "loss": 0.68779688, "memory(GiB)": 71.94, "step": 2540, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.053612088594494, "grad_norm": 1.8767827748337365, "learning_rate": 4.8794023479188905e-06, "loss": 0.65902538, "memory(GiB)": 71.94, "step": 2545, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.0556820534050921, "grad_norm": 1.774731724427948, "learning_rate": 4.8687299893276416e-06, "loss": 0.6885426, "memory(GiB)": 71.94, "step": 2550, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.0577520182156903, "grad_norm": 1.7292661960079105, "learning_rate": 4.858057630736393e-06, "loss": 0.67627816, "memory(GiB)": 71.94, "step": 2555, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.0598219830262885, "grad_norm": 2.022917994019762, "learning_rate": 4.8473852721451445e-06, "loss": 0.69524202, "memory(GiB)": 71.94, "step": 2560, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.0618919478368867, "grad_norm": 1.7503339393851076, "learning_rate": 4.8367129135538955e-06, "loss": 0.69861703, "memory(GiB)": 71.94, "step": 2565, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.063961912647485, "grad_norm": 2.3241272289849126, "learning_rate": 4.826040554962647e-06, "loss": 0.69416137, "memory(GiB)": 71.94, "step": 2570, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.0660318774580833, "grad_norm": 1.9145774620065716, "learning_rate": 4.8153681963713985e-06, "loss": 0.707304, "memory(GiB)": 71.94, "step": 2575, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.0681018422686814, "grad_norm": 2.2205440096454363, "learning_rate": 4.80469583778015e-06, "loss": 0.68690495, "memory(GiB)": 71.94, "step": 2580, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.0701718070792796, "grad_norm": 1.907681817748529, "learning_rate": 4.794023479188901e-06, "loss": 0.67836595, "memory(GiB)": 71.94, "step": 2585, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.0722417718898778, "grad_norm": 1.812113935386319, "learning_rate": 4.783351120597652e-06, "loss": 0.69949398, "memory(GiB)": 71.94, "step": 2590, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.0743117367004762, "grad_norm": 1.9890056601624129, "learning_rate": 4.7726787620064035e-06, "loss": 0.68311234, "memory(GiB)": 71.94, "step": 2595, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.0763817015110744, "grad_norm": 1.6389803143854558, "learning_rate": 4.762006403415155e-06, "loss": 0.68783474, "memory(GiB)": 71.94, "step": 2600, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.0784516663216726, "grad_norm": 1.7243055433624384, "learning_rate": 4.751334044823906e-06, "loss": 0.68109202, "memory(GiB)": 71.94, "step": 2605, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.0805216311322707, "grad_norm": 2.0564605414653356, "learning_rate": 4.740661686232657e-06, "loss": 0.67778783, "memory(GiB)": 71.94, "step": 2610, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.082591595942869, "grad_norm": 1.9673716530688552, "learning_rate": 4.729989327641409e-06, "loss": 0.67061253, "memory(GiB)": 71.94, "step": 2615, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.084661560753467, "grad_norm": 2.256366938059263, "learning_rate": 4.71931696905016e-06, "loss": 0.6971777, "memory(GiB)": 71.94, "step": 2620, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.0867315255640655, "grad_norm": 2.092500467589189, "learning_rate": 4.708644610458911e-06, "loss": 0.67864285, "memory(GiB)": 71.94, "step": 2625, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.0888014903746637, "grad_norm": 1.7063119566002638, "learning_rate": 4.697972251867663e-06, "loss": 0.66528416, "memory(GiB)": 71.94, "step": 2630, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.0908714551852619, "grad_norm": 1.7825515797938636, "learning_rate": 4.687299893276414e-06, "loss": 0.67014618, "memory(GiB)": 71.94, "step": 2635, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.09294141999586, "grad_norm": 1.7034629950636981, "learning_rate": 4.676627534685166e-06, "loss": 0.67851248, "memory(GiB)": 71.94, "step": 2640, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.0950113848064582, "grad_norm": 1.6535356357438644, "learning_rate": 4.665955176093917e-06, "loss": 0.69617391, "memory(GiB)": 71.94, "step": 2645, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.0970813496170566, "grad_norm": 1.790083800076922, "learning_rate": 4.655282817502668e-06, "loss": 0.68771133, "memory(GiB)": 71.94, "step": 2650, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.0991513144276548, "grad_norm": 1.9625451394764908, "learning_rate": 4.64461045891142e-06, "loss": 0.68705645, "memory(GiB)": 71.94, "step": 2655, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.101221279238253, "grad_norm": 2.41883263144616, "learning_rate": 4.633938100320171e-06, "loss": 0.66960602, "memory(GiB)": 71.94, "step": 2660, "train_speed(iter/s)": 0.020044 }, { "epoch": 1.1032912440488511, "grad_norm": 1.6833879907399165, "learning_rate": 4.623265741728922e-06, "loss": 0.67477846, "memory(GiB)": 71.94, "step": 2665, "train_speed(iter/s)": 0.020045 }, { "epoch": 1.1053612088594493, "grad_norm": 1.84460802825979, "learning_rate": 4.612593383137674e-06, "loss": 0.66959124, "memory(GiB)": 71.94, "step": 2670, "train_speed(iter/s)": 0.020046 }, { "epoch": 1.1074311736700475, "grad_norm": 1.8226080355090242, "learning_rate": 4.601921024546425e-06, "loss": 0.66996727, "memory(GiB)": 71.94, "step": 2675, "train_speed(iter/s)": 0.020047 }, { "epoch": 1.109501138480646, "grad_norm": 1.9273809027924724, "learning_rate": 4.591248665955176e-06, "loss": 0.69185095, "memory(GiB)": 71.94, "step": 2680, "train_speed(iter/s)": 0.020048 }, { "epoch": 1.111571103291244, "grad_norm": 2.2322742496415517, "learning_rate": 4.580576307363927e-06, "loss": 0.70724788, "memory(GiB)": 71.94, "step": 2685, "train_speed(iter/s)": 0.020049 }, { "epoch": 1.1136410681018423, "grad_norm": 2.1881226444266133, "learning_rate": 4.569903948772679e-06, "loss": 0.67986469, "memory(GiB)": 71.94, "step": 2690, "train_speed(iter/s)": 0.020049 }, { "epoch": 1.1157110329124404, "grad_norm": 1.9344503535315112, "learning_rate": 4.55923159018143e-06, "loss": 0.6932023, "memory(GiB)": 71.94, "step": 2695, "train_speed(iter/s)": 0.02005 }, { "epoch": 1.1177809977230386, "grad_norm": 2.0197719509697056, "learning_rate": 4.548559231590182e-06, "loss": 0.67475772, "memory(GiB)": 71.94, "step": 2700, "train_speed(iter/s)": 0.020051 }, { "epoch": 1.1177809977230386, "eval_loss": 0.8077359795570374, "eval_runtime": 332.9193, "eval_samples_per_second": 18.761, "eval_steps_per_second": 1.174, "step": 2700 }, { "epoch": 1.119850962533637, "grad_norm": 1.8305900241415576, "learning_rate": 4.537886872998933e-06, "loss": 0.71444244, "memory(GiB)": 71.94, "step": 2705, "train_speed(iter/s)": 0.019995 }, { "epoch": 1.1219209273442352, "grad_norm": 1.9866747607495971, "learning_rate": 4.527214514407685e-06, "loss": 0.68752618, "memory(GiB)": 71.94, "step": 2710, "train_speed(iter/s)": 0.019996 }, { "epoch": 1.1239908921548334, "grad_norm": 1.8866009316885002, "learning_rate": 4.516542155816436e-06, "loss": 0.67673769, "memory(GiB)": 71.94, "step": 2715, "train_speed(iter/s)": 0.019997 }, { "epoch": 1.1260608569654316, "grad_norm": 1.8693665648036668, "learning_rate": 4.505869797225187e-06, "loss": 0.68302212, "memory(GiB)": 71.94, "step": 2720, "train_speed(iter/s)": 0.019997 }, { "epoch": 1.1281308217760297, "grad_norm": 2.088485544417028, "learning_rate": 4.495197438633939e-06, "loss": 0.69513187, "memory(GiB)": 71.94, "step": 2725, "train_speed(iter/s)": 0.019999 }, { "epoch": 1.1302007865866281, "grad_norm": 1.83149183273408, "learning_rate": 4.48452508004269e-06, "loss": 0.67474871, "memory(GiB)": 71.94, "step": 2730, "train_speed(iter/s)": 0.02 }, { "epoch": 1.1322707513972263, "grad_norm": 1.9681727198174188, "learning_rate": 4.473852721451441e-06, "loss": 0.6727396, "memory(GiB)": 71.94, "step": 2735, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.1343407162078245, "grad_norm": 3.0386548884501288, "learning_rate": 4.463180362860193e-06, "loss": 0.69408131, "memory(GiB)": 71.94, "step": 2740, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.1364106810184227, "grad_norm": 3.156158661144904, "learning_rate": 4.452508004268944e-06, "loss": 0.6340518, "memory(GiB)": 71.94, "step": 2745, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.1384806458290209, "grad_norm": 2.005947464562104, "learning_rate": 4.441835645677695e-06, "loss": 0.68589301, "memory(GiB)": 71.94, "step": 2750, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.140550610639619, "grad_norm": 1.8910732500138312, "learning_rate": 4.431163287086446e-06, "loss": 0.68320475, "memory(GiB)": 71.94, "step": 2755, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.1426205754502172, "grad_norm": 1.5980061705957451, "learning_rate": 4.420490928495198e-06, "loss": 0.68288679, "memory(GiB)": 71.94, "step": 2760, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.1446905402608156, "grad_norm": 2.1893217377555367, "learning_rate": 4.409818569903949e-06, "loss": 0.67679596, "memory(GiB)": 71.94, "step": 2765, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.1467605050714138, "grad_norm": 1.6144531687050387, "learning_rate": 4.3991462113127e-06, "loss": 0.67341776, "memory(GiB)": 71.94, "step": 2770, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.148830469882012, "grad_norm": 1.7848487108948465, "learning_rate": 4.388473852721452e-06, "loss": 0.65265465, "memory(GiB)": 71.94, "step": 2775, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.1509004346926102, "grad_norm": 1.9198684626115081, "learning_rate": 4.377801494130203e-06, "loss": 0.65432949, "memory(GiB)": 71.94, "step": 2780, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.1529703995032086, "grad_norm": 1.8075401444295365, "learning_rate": 4.367129135538955e-06, "loss": 0.66166754, "memory(GiB)": 71.94, "step": 2785, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.1550403643138067, "grad_norm": 2.093007547124246, "learning_rate": 4.356456776947706e-06, "loss": 0.671489, "memory(GiB)": 71.94, "step": 2790, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.157110329124405, "grad_norm": 2.155911339314564, "learning_rate": 4.345784418356458e-06, "loss": 0.66719484, "memory(GiB)": 71.94, "step": 2795, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.159180293935003, "grad_norm": 1.7929721304219823, "learning_rate": 4.335112059765209e-06, "loss": 0.65916939, "memory(GiB)": 71.94, "step": 2800, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.1612502587456013, "grad_norm": 2.3382800828112695, "learning_rate": 4.32443970117396e-06, "loss": 0.68535299, "memory(GiB)": 71.94, "step": 2805, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.1633202235561995, "grad_norm": 1.9393370355158424, "learning_rate": 4.313767342582711e-06, "loss": 0.68975515, "memory(GiB)": 71.94, "step": 2810, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.1653901883667976, "grad_norm": 2.1161711862572172, "learning_rate": 4.303094983991463e-06, "loss": 0.68595347, "memory(GiB)": 71.94, "step": 2815, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.167460153177396, "grad_norm": 1.89196514815735, "learning_rate": 4.292422625400214e-06, "loss": 0.66012936, "memory(GiB)": 71.94, "step": 2820, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.1695301179879942, "grad_norm": 1.8601131110854523, "learning_rate": 4.281750266808965e-06, "loss": 0.68001904, "memory(GiB)": 71.94, "step": 2825, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.1716000827985924, "grad_norm": 1.8930363611249428, "learning_rate": 4.271077908217717e-06, "loss": 0.66193466, "memory(GiB)": 71.94, "step": 2830, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.1736700476091906, "grad_norm": 1.651757107446397, "learning_rate": 4.260405549626468e-06, "loss": 0.67280817, "memory(GiB)": 71.94, "step": 2835, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.175740012419789, "grad_norm": 1.6748291349437752, "learning_rate": 4.249733191035219e-06, "loss": 0.70100274, "memory(GiB)": 71.94, "step": 2840, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.1778099772303872, "grad_norm": 1.7834078724205271, "learning_rate": 4.239060832443971e-06, "loss": 0.64067845, "memory(GiB)": 71.94, "step": 2845, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.1798799420409853, "grad_norm": 1.9357492367842137, "learning_rate": 4.228388473852722e-06, "loss": 0.68998647, "memory(GiB)": 71.94, "step": 2850, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.1819499068515835, "grad_norm": 1.9186551723129406, "learning_rate": 4.2177161152614736e-06, "loss": 0.67889709, "memory(GiB)": 71.94, "step": 2855, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.1840198716621817, "grad_norm": 1.9166194791943714, "learning_rate": 4.207043756670225e-06, "loss": 0.67329493, "memory(GiB)": 71.94, "step": 2860, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.1860898364727799, "grad_norm": 1.9439517613212347, "learning_rate": 4.196371398078976e-06, "loss": 0.6614254, "memory(GiB)": 71.94, "step": 2865, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.188159801283378, "grad_norm": 1.6420722807797328, "learning_rate": 4.1856990394877275e-06, "loss": 0.67023277, "memory(GiB)": 71.94, "step": 2870, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.1902297660939765, "grad_norm": 1.7445706716402636, "learning_rate": 4.1750266808964786e-06, "loss": 0.67275252, "memory(GiB)": 71.94, "step": 2875, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.1922997309045746, "grad_norm": 1.8273002856280824, "learning_rate": 4.16435432230523e-06, "loss": 0.65841031, "memory(GiB)": 71.94, "step": 2880, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.1943696957151728, "grad_norm": 1.7392597436189736, "learning_rate": 4.1536819637139815e-06, "loss": 0.66225605, "memory(GiB)": 71.94, "step": 2885, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.196439660525771, "grad_norm": 1.639200530922351, "learning_rate": 4.1430096051227325e-06, "loss": 0.65032516, "memory(GiB)": 71.94, "step": 2890, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.1985096253363694, "grad_norm": 1.8960982589133293, "learning_rate": 4.1323372465314836e-06, "loss": 0.69994841, "memory(GiB)": 71.94, "step": 2895, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.2005795901469676, "grad_norm": 2.1518082070491613, "learning_rate": 4.121664887940235e-06, "loss": 0.64357581, "memory(GiB)": 71.94, "step": 2900, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.2026495549575658, "grad_norm": 2.450141804935637, "learning_rate": 4.1109925293489865e-06, "loss": 0.68345547, "memory(GiB)": 71.94, "step": 2905, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.204719519768164, "grad_norm": 1.8681725857258895, "learning_rate": 4.1003201707577375e-06, "loss": 0.66512461, "memory(GiB)": 71.94, "step": 2910, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.2067894845787621, "grad_norm": 2.2592923799668774, "learning_rate": 4.089647812166489e-06, "loss": 0.68196335, "memory(GiB)": 71.94, "step": 2915, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.2088594493893603, "grad_norm": 2.0216041168873775, "learning_rate": 4.0789754535752404e-06, "loss": 0.6808126, "memory(GiB)": 71.94, "step": 2920, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.2109294141999585, "grad_norm": 1.7659468605949793, "learning_rate": 4.0683030949839915e-06, "loss": 0.64283953, "memory(GiB)": 71.94, "step": 2925, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.2129993790105569, "grad_norm": 1.970065027416205, "learning_rate": 4.057630736392743e-06, "loss": 0.6555728, "memory(GiB)": 71.94, "step": 2930, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.215069343821155, "grad_norm": 1.998737392424493, "learning_rate": 4.046958377801494e-06, "loss": 0.6660428, "memory(GiB)": 71.94, "step": 2935, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.2171393086317532, "grad_norm": 2.1589153805149173, "learning_rate": 4.036286019210246e-06, "loss": 0.65805073, "memory(GiB)": 71.94, "step": 2940, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.2192092734423514, "grad_norm": 1.990247509684529, "learning_rate": 4.025613660618997e-06, "loss": 0.6413794, "memory(GiB)": 71.94, "step": 2945, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.2212792382529498, "grad_norm": 2.0438116812415625, "learning_rate": 4.014941302027748e-06, "loss": 0.65000477, "memory(GiB)": 71.94, "step": 2950, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.223349203063548, "grad_norm": 1.6608044995599232, "learning_rate": 4.004268943436499e-06, "loss": 0.65562267, "memory(GiB)": 71.94, "step": 2955, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.2254191678741462, "grad_norm": 1.9291845395844707, "learning_rate": 3.993596584845251e-06, "loss": 0.67264051, "memory(GiB)": 71.94, "step": 2960, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.2274891326847444, "grad_norm": 2.3618922331006753, "learning_rate": 3.982924226254002e-06, "loss": 0.68117104, "memory(GiB)": 71.94, "step": 2965, "train_speed(iter/s)": 0.020044 }, { "epoch": 1.2295590974953425, "grad_norm": 1.9561928806095756, "learning_rate": 3.972251867662753e-06, "loss": 0.6778089, "memory(GiB)": 71.94, "step": 2970, "train_speed(iter/s)": 0.020045 }, { "epoch": 1.2316290623059407, "grad_norm": 2.103021206596317, "learning_rate": 3.961579509071505e-06, "loss": 0.66529841, "memory(GiB)": 71.94, "step": 2975, "train_speed(iter/s)": 0.020046 }, { "epoch": 1.2336990271165391, "grad_norm": 1.6317138464756236, "learning_rate": 3.950907150480256e-06, "loss": 0.63005896, "memory(GiB)": 71.94, "step": 2980, "train_speed(iter/s)": 0.020047 }, { "epoch": 1.2357689919271373, "grad_norm": 1.8763154321083348, "learning_rate": 3.940234791889007e-06, "loss": 0.69782147, "memory(GiB)": 71.94, "step": 2985, "train_speed(iter/s)": 0.020048 }, { "epoch": 1.2378389567377355, "grad_norm": 2.12720513001939, "learning_rate": 3.929562433297759e-06, "loss": 0.67128716, "memory(GiB)": 71.94, "step": 2990, "train_speed(iter/s)": 0.020049 }, { "epoch": 1.2399089215483337, "grad_norm": 2.0560360745042243, "learning_rate": 3.91889007470651e-06, "loss": 0.66187463, "memory(GiB)": 71.94, "step": 2995, "train_speed(iter/s)": 0.02005 }, { "epoch": 1.2419788863589318, "grad_norm": 1.9884557277874149, "learning_rate": 3.908217716115262e-06, "loss": 0.68540969, "memory(GiB)": 71.94, "step": 3000, "train_speed(iter/s)": 0.020051 }, { "epoch": 1.2419788863589318, "eval_loss": 0.8033931255340576, "eval_runtime": 334.3532, "eval_samples_per_second": 18.681, "eval_steps_per_second": 1.169, "step": 3000 }, { "epoch": 1.2440488511695302, "grad_norm": 1.8132336089456924, "learning_rate": 3.897545357524013e-06, "loss": 0.65594606, "memory(GiB)": 71.94, "step": 3005, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.2461188159801284, "grad_norm": 1.6982582415846914, "learning_rate": 3.886872998932765e-06, "loss": 0.65645032, "memory(GiB)": 71.94, "step": 3010, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.2481887807907266, "grad_norm": 1.9942609739954071, "learning_rate": 3.876200640341516e-06, "loss": 0.66572218, "memory(GiB)": 71.94, "step": 3015, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.2502587456013248, "grad_norm": 1.7952303972411592, "learning_rate": 3.865528281750267e-06, "loss": 0.64628544, "memory(GiB)": 71.94, "step": 3020, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.252328710411923, "grad_norm": 1.9058576079441742, "learning_rate": 3.854855923159018e-06, "loss": 0.68286681, "memory(GiB)": 71.94, "step": 3025, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.2543986752225211, "grad_norm": 2.4990275874354357, "learning_rate": 3.84418356456777e-06, "loss": 0.67756252, "memory(GiB)": 71.94, "step": 3030, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.2564686400331193, "grad_norm": 2.18800214487155, "learning_rate": 3.833511205976521e-06, "loss": 0.65295424, "memory(GiB)": 71.94, "step": 3035, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.2585386048437177, "grad_norm": 2.1525831686251737, "learning_rate": 3.822838847385272e-06, "loss": 0.66122398, "memory(GiB)": 71.94, "step": 3040, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.260608569654316, "grad_norm": 2.1001976709383374, "learning_rate": 3.812166488794024e-06, "loss": 0.64958653, "memory(GiB)": 71.94, "step": 3045, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.262678534464914, "grad_norm": 1.8005403174974608, "learning_rate": 3.801494130202775e-06, "loss": 0.66733065, "memory(GiB)": 71.94, "step": 3050, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.2647484992755123, "grad_norm": 2.0392708573153255, "learning_rate": 3.7908217716115265e-06, "loss": 0.65150566, "memory(GiB)": 71.94, "step": 3055, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.2668184640861107, "grad_norm": 1.9818256506070209, "learning_rate": 3.7801494130202776e-06, "loss": 0.65937147, "memory(GiB)": 71.94, "step": 3060, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.2688884288967088, "grad_norm": 2.091473795655044, "learning_rate": 3.7694770544290294e-06, "loss": 0.65645385, "memory(GiB)": 71.94, "step": 3065, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.270958393707307, "grad_norm": 1.9887693066644563, "learning_rate": 3.7588046958377805e-06, "loss": 0.6356863, "memory(GiB)": 71.94, "step": 3070, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.2730283585179052, "grad_norm": 1.849373670863495, "learning_rate": 3.7481323372465315e-06, "loss": 0.64913082, "memory(GiB)": 71.94, "step": 3075, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.2750983233285034, "grad_norm": 2.8796661699092776, "learning_rate": 3.737459978655283e-06, "loss": 0.66612153, "memory(GiB)": 71.94, "step": 3080, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.2771682881391015, "grad_norm": 1.9206629854919683, "learning_rate": 3.7267876200640345e-06, "loss": 0.65432153, "memory(GiB)": 71.94, "step": 3085, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.2792382529496997, "grad_norm": 2.0649234898157562, "learning_rate": 3.716115261472786e-06, "loss": 0.63801923, "memory(GiB)": 71.94, "step": 3090, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.2813082177602981, "grad_norm": 1.859875670695032, "learning_rate": 3.705442902881537e-06, "loss": 0.6548945, "memory(GiB)": 71.94, "step": 3095, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.2833781825708963, "grad_norm": 1.767661952337357, "learning_rate": 3.694770544290289e-06, "loss": 0.64956121, "memory(GiB)": 71.94, "step": 3100, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.2854481473814945, "grad_norm": 1.8007833500981838, "learning_rate": 3.68409818569904e-06, "loss": 0.669453, "memory(GiB)": 71.94, "step": 3105, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.2875181121920927, "grad_norm": 2.0099069002485863, "learning_rate": 3.673425827107791e-06, "loss": 0.67178354, "memory(GiB)": 71.94, "step": 3110, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.289588077002691, "grad_norm": 2.0463019564128886, "learning_rate": 3.6627534685165424e-06, "loss": 0.65595369, "memory(GiB)": 71.94, "step": 3115, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.2916580418132892, "grad_norm": 1.9152750413831356, "learning_rate": 3.652081109925294e-06, "loss": 0.65940399, "memory(GiB)": 71.94, "step": 3120, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.2937280066238874, "grad_norm": 2.9816950881544306, "learning_rate": 3.6414087513340453e-06, "loss": 0.66090908, "memory(GiB)": 71.94, "step": 3125, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.2957979714344856, "grad_norm": 1.8591409546980415, "learning_rate": 3.6307363927427963e-06, "loss": 0.66902189, "memory(GiB)": 71.94, "step": 3130, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.2978679362450838, "grad_norm": 2.078590180758969, "learning_rate": 3.6200640341515482e-06, "loss": 0.66616745, "memory(GiB)": 71.94, "step": 3135, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.299937901055682, "grad_norm": 1.8683848031966166, "learning_rate": 3.6093916755602993e-06, "loss": 0.64836683, "memory(GiB)": 71.94, "step": 3140, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.3020078658662801, "grad_norm": 1.7924123567589454, "learning_rate": 3.5987193169690503e-06, "loss": 0.66824627, "memory(GiB)": 71.94, "step": 3145, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.3040778306768785, "grad_norm": 1.94170729576761, "learning_rate": 3.5880469583778018e-06, "loss": 0.67116079, "memory(GiB)": 71.94, "step": 3150, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.3061477954874767, "grad_norm": 2.1208196521106357, "learning_rate": 3.5773745997865532e-06, "loss": 0.66398339, "memory(GiB)": 71.94, "step": 3155, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.308217760298075, "grad_norm": 1.94319742259963, "learning_rate": 3.5667022411953047e-06, "loss": 0.66830492, "memory(GiB)": 71.94, "step": 3160, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.310287725108673, "grad_norm": 1.8986704786653348, "learning_rate": 3.5560298826040557e-06, "loss": 0.64020205, "memory(GiB)": 71.94, "step": 3165, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.3123576899192715, "grad_norm": 1.832166033780513, "learning_rate": 3.5453575240128068e-06, "loss": 0.65297813, "memory(GiB)": 71.94, "step": 3170, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.3144276547298697, "grad_norm": 1.7524632564639653, "learning_rate": 3.5346851654215586e-06, "loss": 0.65833459, "memory(GiB)": 71.94, "step": 3175, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.3164976195404678, "grad_norm": 1.784819778645554, "learning_rate": 3.5240128068303097e-06, "loss": 0.65319304, "memory(GiB)": 71.94, "step": 3180, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.318567584351066, "grad_norm": 1.7883657805218875, "learning_rate": 3.513340448239061e-06, "loss": 0.65240812, "memory(GiB)": 71.94, "step": 3185, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.3206375491616642, "grad_norm": 1.894144860689911, "learning_rate": 3.5026680896478126e-06, "loss": 0.6723794, "memory(GiB)": 71.94, "step": 3190, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.3227075139722624, "grad_norm": 2.0320221849381195, "learning_rate": 3.491995731056564e-06, "loss": 0.64077559, "memory(GiB)": 71.94, "step": 3195, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.3247774787828606, "grad_norm": 1.727824740633838, "learning_rate": 3.481323372465315e-06, "loss": 0.63609524, "memory(GiB)": 71.94, "step": 3200, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.326847443593459, "grad_norm": 1.7454884742520516, "learning_rate": 3.470651013874066e-06, "loss": 0.65425596, "memory(GiB)": 71.94, "step": 3205, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.3289174084040571, "grad_norm": 1.8911130178984759, "learning_rate": 3.459978655282818e-06, "loss": 0.66199875, "memory(GiB)": 71.94, "step": 3210, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.3309873732146553, "grad_norm": 2.3250937309778474, "learning_rate": 3.449306296691569e-06, "loss": 0.65385923, "memory(GiB)": 71.94, "step": 3215, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.3330573380252535, "grad_norm": 2.1704453467921447, "learning_rate": 3.43863393810032e-06, "loss": 0.66229916, "memory(GiB)": 71.94, "step": 3220, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.335127302835852, "grad_norm": 2.2209117698054768, "learning_rate": 3.427961579509072e-06, "loss": 0.65051212, "memory(GiB)": 71.94, "step": 3225, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.33719726764645, "grad_norm": 1.6845784156480532, "learning_rate": 3.417289220917823e-06, "loss": 0.64576015, "memory(GiB)": 71.94, "step": 3230, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.3392672324570483, "grad_norm": 2.7721561024455332, "learning_rate": 3.4066168623265745e-06, "loss": 0.68257604, "memory(GiB)": 71.94, "step": 3235, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.3413371972676464, "grad_norm": 1.9362479296558788, "learning_rate": 3.3959445037353255e-06, "loss": 0.63483586, "memory(GiB)": 71.94, "step": 3240, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.3434071620782446, "grad_norm": 1.9956347191580364, "learning_rate": 3.3852721451440774e-06, "loss": 0.65959587, "memory(GiB)": 71.94, "step": 3245, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.3454771268888428, "grad_norm": 1.7321062675450865, "learning_rate": 3.3745997865528285e-06, "loss": 0.66381845, "memory(GiB)": 71.94, "step": 3250, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.347547091699441, "grad_norm": 1.9696403539230432, "learning_rate": 3.3639274279615795e-06, "loss": 0.64812822, "memory(GiB)": 71.94, "step": 3255, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.3496170565100394, "grad_norm": 2.1319401432279053, "learning_rate": 3.353255069370331e-06, "loss": 0.64005919, "memory(GiB)": 71.94, "step": 3260, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.3516870213206376, "grad_norm": 1.8363706763408283, "learning_rate": 3.3425827107790824e-06, "loss": 0.64201698, "memory(GiB)": 71.94, "step": 3265, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.3537569861312357, "grad_norm": 2.124129454637882, "learning_rate": 3.331910352187834e-06, "loss": 0.65025196, "memory(GiB)": 71.94, "step": 3270, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.355826950941834, "grad_norm": 1.7896864503844883, "learning_rate": 3.321237993596585e-06, "loss": 0.63300438, "memory(GiB)": 71.94, "step": 3275, "train_speed(iter/s)": 0.020044 }, { "epoch": 1.3578969157524323, "grad_norm": 1.9837261215389441, "learning_rate": 3.310565635005337e-06, "loss": 0.63380709, "memory(GiB)": 71.94, "step": 3280, "train_speed(iter/s)": 0.020044 }, { "epoch": 1.3599668805630305, "grad_norm": 2.097355543637799, "learning_rate": 3.299893276414088e-06, "loss": 0.62883258, "memory(GiB)": 71.94, "step": 3285, "train_speed(iter/s)": 0.020044 }, { "epoch": 1.3620368453736287, "grad_norm": 1.9315070924851874, "learning_rate": 3.289220917822839e-06, "loss": 0.63195004, "memory(GiB)": 71.94, "step": 3290, "train_speed(iter/s)": 0.020045 }, { "epoch": 1.3641068101842269, "grad_norm": 1.7050134771385488, "learning_rate": 3.2785485592315903e-06, "loss": 0.65333614, "memory(GiB)": 71.94, "step": 3295, "train_speed(iter/s)": 0.020045 }, { "epoch": 1.366176774994825, "grad_norm": 3.6481119398862005, "learning_rate": 3.267876200640342e-06, "loss": 0.68793478, "memory(GiB)": 71.94, "step": 3300, "train_speed(iter/s)": 0.020046 }, { "epoch": 1.366176774994825, "eval_loss": 0.8023556470870972, "eval_runtime": 334.0894, "eval_samples_per_second": 18.696, "eval_steps_per_second": 1.17, "step": 3300 }, { "epoch": 1.3682467398054232, "grad_norm": 2.092330422962737, "learning_rate": 3.2572038420490933e-06, "loss": 0.63313837, "memory(GiB)": 71.94, "step": 3305, "train_speed(iter/s)": 0.02 }, { "epoch": 1.3703167046160214, "grad_norm": 1.9167879677265438, "learning_rate": 3.2465314834578443e-06, "loss": 0.65159397, "memory(GiB)": 71.94, "step": 3310, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.3723866694266198, "grad_norm": 1.8761921604953933, "learning_rate": 3.235859124866596e-06, "loss": 0.62655239, "memory(GiB)": 71.94, "step": 3315, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.374456634237218, "grad_norm": 2.178411039961288, "learning_rate": 3.2251867662753472e-06, "loss": 0.67218485, "memory(GiB)": 71.94, "step": 3320, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.3765265990478162, "grad_norm": 1.776609105623676, "learning_rate": 3.2145144076840983e-06, "loss": 0.63943739, "memory(GiB)": 71.94, "step": 3325, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.3785965638584143, "grad_norm": 1.8854963168848657, "learning_rate": 3.2038420490928497e-06, "loss": 0.65016818, "memory(GiB)": 71.94, "step": 3330, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.3806665286690127, "grad_norm": 2.3494051096524915, "learning_rate": 3.193169690501601e-06, "loss": 0.6696517, "memory(GiB)": 71.94, "step": 3335, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.382736493479611, "grad_norm": 2.390390518794705, "learning_rate": 3.1824973319103527e-06, "loss": 0.64389553, "memory(GiB)": 71.94, "step": 3340, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.384806458290209, "grad_norm": 1.8195738414381504, "learning_rate": 3.1718249733191037e-06, "loss": 0.64606419, "memory(GiB)": 71.94, "step": 3345, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.3868764231008073, "grad_norm": 2.3772901728441846, "learning_rate": 3.1611526147278547e-06, "loss": 0.66564407, "memory(GiB)": 71.94, "step": 3350, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.3889463879114055, "grad_norm": 1.9865374550113573, "learning_rate": 3.1504802561366066e-06, "loss": 0.62708969, "memory(GiB)": 71.94, "step": 3355, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.3910163527220036, "grad_norm": 1.8995100511755476, "learning_rate": 3.1398078975453577e-06, "loss": 0.64633865, "memory(GiB)": 71.94, "step": 3360, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.3930863175326018, "grad_norm": 2.3050512969189954, "learning_rate": 3.129135538954109e-06, "loss": 0.64496307, "memory(GiB)": 71.94, "step": 3365, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.3951562823432002, "grad_norm": 1.9056478303612696, "learning_rate": 3.1184631803628606e-06, "loss": 0.641465, "memory(GiB)": 71.94, "step": 3370, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.3972262471537984, "grad_norm": 2.166573531526341, "learning_rate": 3.107790821771612e-06, "loss": 0.64069824, "memory(GiB)": 71.94, "step": 3375, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.3992962119643966, "grad_norm": 1.9398811847394521, "learning_rate": 3.097118463180363e-06, "loss": 0.64301763, "memory(GiB)": 47.6, "step": 3380, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.4013661767749948, "grad_norm": 1.959159982373063, "learning_rate": 3.086446104589114e-06, "loss": 0.6362546, "memory(GiB)": 47.6, "step": 3385, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.4034361415855932, "grad_norm": 2.0530926805349083, "learning_rate": 3.075773745997866e-06, "loss": 0.63418798, "memory(GiB)": 47.6, "step": 3390, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.4055061063961913, "grad_norm": 1.9297510699914728, "learning_rate": 3.065101387406617e-06, "loss": 0.6311512, "memory(GiB)": 47.6, "step": 3395, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.4075760712067895, "grad_norm": 1.9271894163285872, "learning_rate": 3.054429028815368e-06, "loss": 0.63234644, "memory(GiB)": 47.6, "step": 3400, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.4096460360173877, "grad_norm": 1.7395603251397769, "learning_rate": 3.04375667022412e-06, "loss": 0.63525658, "memory(GiB)": 47.6, "step": 3405, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.4117160008279859, "grad_norm": 1.9592997788913435, "learning_rate": 3.033084311632871e-06, "loss": 0.62426691, "memory(GiB)": 47.6, "step": 3410, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.413785965638584, "grad_norm": 1.83244625095987, "learning_rate": 3.0224119530416225e-06, "loss": 0.64436603, "memory(GiB)": 47.6, "step": 3415, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.4158559304491822, "grad_norm": 1.9274094419667949, "learning_rate": 3.0117395944503735e-06, "loss": 0.65462785, "memory(GiB)": 47.6, "step": 3420, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.4179258952597806, "grad_norm": 1.9910550366499922, "learning_rate": 3.0010672358591254e-06, "loss": 0.62345576, "memory(GiB)": 47.6, "step": 3425, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.4199958600703788, "grad_norm": 1.9961790791051468, "learning_rate": 2.9903948772678764e-06, "loss": 0.62686663, "memory(GiB)": 48.58, "step": 3430, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.422065824880977, "grad_norm": 2.2471760043812767, "learning_rate": 2.9797225186766275e-06, "loss": 0.65531764, "memory(GiB)": 48.58, "step": 3435, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.4241357896915752, "grad_norm": 2.225238329922807, "learning_rate": 2.9690501600853794e-06, "loss": 0.64034252, "memory(GiB)": 48.58, "step": 3440, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.4262057545021736, "grad_norm": 1.6663799088887756, "learning_rate": 2.9583778014941304e-06, "loss": 0.61355238, "memory(GiB)": 48.58, "step": 3445, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.4282757193127718, "grad_norm": 1.9127892416039678, "learning_rate": 2.947705442902882e-06, "loss": 0.61548796, "memory(GiB)": 48.58, "step": 3450, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.43034568412337, "grad_norm": 1.7733393610398442, "learning_rate": 2.937033084311633e-06, "loss": 0.66395988, "memory(GiB)": 48.58, "step": 3455, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.432415648933968, "grad_norm": 2.1962868424551676, "learning_rate": 2.9263607257203848e-06, "loss": 0.61601906, "memory(GiB)": 48.58, "step": 3460, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.4344856137445663, "grad_norm": 1.8134260024404016, "learning_rate": 2.915688367129136e-06, "loss": 0.64126596, "memory(GiB)": 48.58, "step": 3465, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.4365555785551645, "grad_norm": 2.136457991499728, "learning_rate": 2.905016008537887e-06, "loss": 0.63421082, "memory(GiB)": 48.58, "step": 3470, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.4386255433657626, "grad_norm": 1.988532670614334, "learning_rate": 2.8943436499466383e-06, "loss": 0.63867459, "memory(GiB)": 48.58, "step": 3475, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.440695508176361, "grad_norm": 1.9277626836086974, "learning_rate": 2.8836712913553898e-06, "loss": 0.65572329, "memory(GiB)": 48.58, "step": 3480, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.4427654729869592, "grad_norm": 2.1432613819151247, "learning_rate": 2.8729989327641412e-06, "loss": 0.64992175, "memory(GiB)": 48.58, "step": 3485, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.4448354377975574, "grad_norm": 2.355567308350272, "learning_rate": 2.8623265741728923e-06, "loss": 0.66429825, "memory(GiB)": 48.58, "step": 3490, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.4469054026081556, "grad_norm": 2.045569698254085, "learning_rate": 2.851654215581644e-06, "loss": 0.65050926, "memory(GiB)": 48.58, "step": 3495, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.448975367418754, "grad_norm": 1.8389546498339602, "learning_rate": 2.840981856990395e-06, "loss": 0.63454714, "memory(GiB)": 48.58, "step": 3500, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.4510453322293522, "grad_norm": 2.1282462505999518, "learning_rate": 2.8303094983991462e-06, "loss": 0.61026077, "memory(GiB)": 48.58, "step": 3505, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.4531152970399503, "grad_norm": 1.9822876450283116, "learning_rate": 2.8196371398078977e-06, "loss": 0.63836317, "memory(GiB)": 48.58, "step": 3510, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.4551852618505485, "grad_norm": 2.1201754460326603, "learning_rate": 2.808964781216649e-06, "loss": 0.65191045, "memory(GiB)": 48.58, "step": 3515, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.4572552266611467, "grad_norm": 1.8520751525916404, "learning_rate": 2.7982924226254006e-06, "loss": 0.62120395, "memory(GiB)": 48.58, "step": 3520, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.4593251914717449, "grad_norm": 1.7983375554778653, "learning_rate": 2.7876200640341517e-06, "loss": 0.63227787, "memory(GiB)": 48.58, "step": 3525, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.461395156282343, "grad_norm": 2.340481282081477, "learning_rate": 2.7769477054429036e-06, "loss": 0.63372889, "memory(GiB)": 48.58, "step": 3530, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.4634651210929415, "grad_norm": 1.9262925464363927, "learning_rate": 2.7662753468516546e-06, "loss": 0.63678517, "memory(GiB)": 48.58, "step": 3535, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.4655350859035396, "grad_norm": 2.2093247947290164, "learning_rate": 2.7556029882604056e-06, "loss": 0.6558732, "memory(GiB)": 48.58, "step": 3540, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.4676050507141378, "grad_norm": 2.020333271689009, "learning_rate": 2.744930629669157e-06, "loss": 0.60733166, "memory(GiB)": 48.58, "step": 3545, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.469675015524736, "grad_norm": 2.0734276125917632, "learning_rate": 2.7342582710779086e-06, "loss": 0.66044526, "memory(GiB)": 48.58, "step": 3550, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.4717449803353344, "grad_norm": 1.8332312440056042, "learning_rate": 2.72358591248666e-06, "loss": 0.64025326, "memory(GiB)": 48.58, "step": 3555, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.4738149451459326, "grad_norm": 1.8856532947296074, "learning_rate": 2.712913553895411e-06, "loss": 0.63681345, "memory(GiB)": 48.58, "step": 3560, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.4758849099565308, "grad_norm": 1.6710776220883121, "learning_rate": 2.702241195304162e-06, "loss": 0.64032116, "memory(GiB)": 48.58, "step": 3565, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.477954874767129, "grad_norm": 2.095636589348373, "learning_rate": 2.691568836712914e-06, "loss": 0.63223238, "memory(GiB)": 48.58, "step": 3570, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.4800248395777271, "grad_norm": 1.8927929015268774, "learning_rate": 2.680896478121665e-06, "loss": 0.63061609, "memory(GiB)": 48.58, "step": 3575, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.4820948043883253, "grad_norm": 2.0929754486788137, "learning_rate": 2.670224119530416e-06, "loss": 0.62695656, "memory(GiB)": 48.58, "step": 3580, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.4841647691989235, "grad_norm": 1.7458918576306506, "learning_rate": 2.659551760939168e-06, "loss": 0.63322277, "memory(GiB)": 48.58, "step": 3585, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.4862347340095219, "grad_norm": 1.8907196532347343, "learning_rate": 2.648879402347919e-06, "loss": 0.63732347, "memory(GiB)": 48.58, "step": 3590, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.48830469882012, "grad_norm": 2.263161529685459, "learning_rate": 2.6382070437566704e-06, "loss": 0.6138607, "memory(GiB)": 48.58, "step": 3595, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.4903746636307182, "grad_norm": 2.572454412416673, "learning_rate": 2.6275346851654215e-06, "loss": 0.63049603, "memory(GiB)": 48.58, "step": 3600, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.4903746636307182, "eval_loss": 0.8007386922836304, "eval_runtime": 333.8857, "eval_samples_per_second": 18.707, "eval_steps_per_second": 1.171, "step": 3600 }, { "epoch": 1.4924446284413164, "grad_norm": 1.901029287627636, "learning_rate": 2.6168623265741734e-06, "loss": 0.65960011, "memory(GiB)": 48.58, "step": 3605, "train_speed(iter/s)": 0.019998 }, { "epoch": 1.4945145932519148, "grad_norm": 2.0108141055428423, "learning_rate": 2.6061899679829244e-06, "loss": 0.63098369, "memory(GiB)": 48.58, "step": 3610, "train_speed(iter/s)": 0.019999 }, { "epoch": 1.496584558062513, "grad_norm": 1.8385014047339334, "learning_rate": 2.5955176093916754e-06, "loss": 0.63654456, "memory(GiB)": 48.58, "step": 3615, "train_speed(iter/s)": 0.019999 }, { "epoch": 1.4986545228731112, "grad_norm": 2.502391600466987, "learning_rate": 2.5848452508004273e-06, "loss": 0.64129109, "memory(GiB)": 48.58, "step": 3620, "train_speed(iter/s)": 0.02 }, { "epoch": 1.5007244876837094, "grad_norm": 1.8078536683803752, "learning_rate": 2.5741728922091784e-06, "loss": 0.61762905, "memory(GiB)": 48.58, "step": 3625, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.5027944524943075, "grad_norm": 2.4124021719958813, "learning_rate": 2.56350053361793e-06, "loss": 0.62886133, "memory(GiB)": 48.58, "step": 3630, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.5048644173049057, "grad_norm": 2.14818078370575, "learning_rate": 2.552828175026681e-06, "loss": 0.64229774, "memory(GiB)": 48.58, "step": 3635, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.506934382115504, "grad_norm": 2.0607592121598777, "learning_rate": 2.5421558164354328e-06, "loss": 0.62922821, "memory(GiB)": 48.58, "step": 3640, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.5090043469261023, "grad_norm": 1.9537072105756903, "learning_rate": 2.531483457844184e-06, "loss": 0.6114254, "memory(GiB)": 48.58, "step": 3645, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.5110743117367005, "grad_norm": 2.076894133768813, "learning_rate": 2.520811099252935e-06, "loss": 0.65702705, "memory(GiB)": 48.58, "step": 3650, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.5131442765472987, "grad_norm": 1.8684495033846724, "learning_rate": 2.5101387406616863e-06, "loss": 0.65235605, "memory(GiB)": 48.58, "step": 3655, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.515214241357897, "grad_norm": 2.099820461610625, "learning_rate": 2.4994663820704378e-06, "loss": 0.63383131, "memory(GiB)": 48.58, "step": 3660, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.5172842061684952, "grad_norm": 2.0251264820061765, "learning_rate": 2.4887940234791892e-06, "loss": 0.63897676, "memory(GiB)": 48.58, "step": 3665, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.5193541709790934, "grad_norm": 1.8916351044126378, "learning_rate": 2.4781216648879407e-06, "loss": 0.63116732, "memory(GiB)": 48.58, "step": 3670, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.5214241357896916, "grad_norm": 1.758567214100863, "learning_rate": 2.4674493062966917e-06, "loss": 0.61664515, "memory(GiB)": 48.58, "step": 3675, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.5234941006002898, "grad_norm": 2.004985649591871, "learning_rate": 2.456776947705443e-06, "loss": 0.63878107, "memory(GiB)": 48.58, "step": 3680, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.525564065410888, "grad_norm": 2.0496876810492894, "learning_rate": 2.4461045891141942e-06, "loss": 0.62643003, "memory(GiB)": 48.58, "step": 3685, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.5276340302214861, "grad_norm": 1.9829235096697742, "learning_rate": 2.4354322305229457e-06, "loss": 0.62055197, "memory(GiB)": 48.58, "step": 3690, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.5297039950320843, "grad_norm": 1.7991062631513268, "learning_rate": 2.424759871931697e-06, "loss": 0.6605298, "memory(GiB)": 48.58, "step": 3695, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.5317739598426827, "grad_norm": 2.526220221053364, "learning_rate": 2.4140875133404486e-06, "loss": 0.65973449, "memory(GiB)": 48.58, "step": 3700, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.533843924653281, "grad_norm": 1.682540853202414, "learning_rate": 2.4034151547492e-06, "loss": 0.65421052, "memory(GiB)": 48.58, "step": 3705, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.535913889463879, "grad_norm": 1.965734649630595, "learning_rate": 2.392742796157951e-06, "loss": 0.64042482, "memory(GiB)": 48.58, "step": 3710, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.5379838542744775, "grad_norm": 2.2061994907060005, "learning_rate": 2.3820704375667026e-06, "loss": 0.61186495, "memory(GiB)": 48.58, "step": 3715, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.5400538190850757, "grad_norm": 2.2749623686643474, "learning_rate": 2.3713980789754536e-06, "loss": 0.63306904, "memory(GiB)": 48.58, "step": 3720, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.5421237838956738, "grad_norm": 1.8676661262594305, "learning_rate": 2.360725720384205e-06, "loss": 0.64988294, "memory(GiB)": 48.58, "step": 3725, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.544193748706272, "grad_norm": 2.069377222520985, "learning_rate": 2.3500533617929565e-06, "loss": 0.63958693, "memory(GiB)": 48.58, "step": 3730, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.5462637135168702, "grad_norm": 2.0148406212108574, "learning_rate": 2.339381003201708e-06, "loss": 0.60387516, "memory(GiB)": 48.58, "step": 3735, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.5483336783274684, "grad_norm": 1.8592072092273917, "learning_rate": 2.328708644610459e-06, "loss": 0.60533466, "memory(GiB)": 48.58, "step": 3740, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.5504036431380666, "grad_norm": 1.9293967097992641, "learning_rate": 2.3180362860192105e-06, "loss": 0.64885559, "memory(GiB)": 48.58, "step": 3745, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.5524736079486647, "grad_norm": 2.105150298236883, "learning_rate": 2.307363927427962e-06, "loss": 0.6291214, "memory(GiB)": 48.58, "step": 3750, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.5545435727592631, "grad_norm": 1.7706279189258218, "learning_rate": 2.296691568836713e-06, "loss": 0.62637095, "memory(GiB)": 48.58, "step": 3755, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.5566135375698613, "grad_norm": 2.565565704557741, "learning_rate": 2.2860192102454645e-06, "loss": 0.61526871, "memory(GiB)": 48.58, "step": 3760, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.5586835023804595, "grad_norm": 2.07386908713523, "learning_rate": 2.2753468516542155e-06, "loss": 0.62105417, "memory(GiB)": 48.58, "step": 3765, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.560753467191058, "grad_norm": 2.0656516897338064, "learning_rate": 2.264674493062967e-06, "loss": 0.62474051, "memory(GiB)": 48.58, "step": 3770, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.562823432001656, "grad_norm": 2.1881677769227075, "learning_rate": 2.2540021344717184e-06, "loss": 0.6300148, "memory(GiB)": 48.58, "step": 3775, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.5648933968122543, "grad_norm": 1.9586494376217993, "learning_rate": 2.24332977588047e-06, "loss": 0.59469123, "memory(GiB)": 48.58, "step": 3780, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.5669633616228524, "grad_norm": 2.440599800961732, "learning_rate": 2.232657417289221e-06, "loss": 0.62468157, "memory(GiB)": 48.58, "step": 3785, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.5690333264334506, "grad_norm": 2.0016886010949797, "learning_rate": 2.2219850586979724e-06, "loss": 0.62775316, "memory(GiB)": 48.58, "step": 3790, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.5711032912440488, "grad_norm": 1.9086880143876215, "learning_rate": 2.211312700106724e-06, "loss": 0.62862492, "memory(GiB)": 48.58, "step": 3795, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.573173256054647, "grad_norm": 2.0888610038198636, "learning_rate": 2.200640341515475e-06, "loss": 0.59845972, "memory(GiB)": 48.58, "step": 3800, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.5752432208652452, "grad_norm": 2.055969507055363, "learning_rate": 2.1899679829242263e-06, "loss": 0.61734905, "memory(GiB)": 48.58, "step": 3805, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.5773131856758436, "grad_norm": 1.9707520677944765, "learning_rate": 2.179295624332978e-06, "loss": 0.62550197, "memory(GiB)": 48.58, "step": 3810, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.5793831504864417, "grad_norm": 2.1321630269454617, "learning_rate": 2.1686232657417293e-06, "loss": 0.62156429, "memory(GiB)": 48.58, "step": 3815, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.58145311529704, "grad_norm": 2.3313380056807373, "learning_rate": 2.1579509071504803e-06, "loss": 0.62489176, "memory(GiB)": 48.58, "step": 3820, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.5835230801076383, "grad_norm": 2.0579013534602044, "learning_rate": 2.1472785485592318e-06, "loss": 0.60487609, "memory(GiB)": 48.58, "step": 3825, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.5855930449182365, "grad_norm": 2.1449642041698267, "learning_rate": 2.136606189967983e-06, "loss": 0.63105164, "memory(GiB)": 48.58, "step": 3830, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.5876630097288347, "grad_norm": 1.8838252023426636, "learning_rate": 2.1259338313767343e-06, "loss": 0.62664189, "memory(GiB)": 48.58, "step": 3835, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.5897329745394329, "grad_norm": 2.289021375023386, "learning_rate": 2.1152614727854857e-06, "loss": 0.60014114, "memory(GiB)": 48.58, "step": 3840, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.591802939350031, "grad_norm": 1.952623771952289, "learning_rate": 2.104589114194237e-06, "loss": 0.63038387, "memory(GiB)": 48.58, "step": 3845, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.5938729041606292, "grad_norm": 2.151088803547346, "learning_rate": 2.0939167556029887e-06, "loss": 0.64208837, "memory(GiB)": 48.58, "step": 3850, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.5959428689712274, "grad_norm": 2.0667970850832025, "learning_rate": 2.0832443970117397e-06, "loss": 0.62941227, "memory(GiB)": 48.58, "step": 3855, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.5980128337818256, "grad_norm": 1.8287162112237207, "learning_rate": 2.072572038420491e-06, "loss": 0.60888386, "memory(GiB)": 48.58, "step": 3860, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.600082798592424, "grad_norm": 2.083482144562111, "learning_rate": 2.061899679829242e-06, "loss": 0.60378475, "memory(GiB)": 48.58, "step": 3865, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.6021527634030222, "grad_norm": 2.3064863216709925, "learning_rate": 2.0512273212379937e-06, "loss": 0.60522232, "memory(GiB)": 48.58, "step": 3870, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.6042227282136203, "grad_norm": 2.1700368490113844, "learning_rate": 2.040554962646745e-06, "loss": 0.61363611, "memory(GiB)": 48.58, "step": 3875, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.6062926930242187, "grad_norm": 2.2038759685376843, "learning_rate": 2.0298826040554966e-06, "loss": 0.60480423, "memory(GiB)": 48.58, "step": 3880, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.608362657834817, "grad_norm": 2.1947427482623914, "learning_rate": 2.019210245464248e-06, "loss": 0.61413918, "memory(GiB)": 48.58, "step": 3885, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.610432622645415, "grad_norm": 2.022770825774821, "learning_rate": 2.008537886872999e-06, "loss": 0.61320429, "memory(GiB)": 48.58, "step": 3890, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.6125025874560133, "grad_norm": 2.052392899511488, "learning_rate": 1.9978655282817505e-06, "loss": 0.6130487, "memory(GiB)": 48.58, "step": 3895, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.6145725522666114, "grad_norm": 2.0156586912928596, "learning_rate": 1.9871931696905016e-06, "loss": 0.64238014, "memory(GiB)": 48.58, "step": 3900, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.6145725522666114, "eval_loss": 0.79938143491745, "eval_runtime": 333.9342, "eval_samples_per_second": 18.704, "eval_steps_per_second": 1.171, "step": 3900 }, { "epoch": 1.6166425170772096, "grad_norm": 2.023426003459966, "learning_rate": 1.976520811099253e-06, "loss": 0.61180491, "memory(GiB)": 48.58, "step": 3905, "train_speed(iter/s)": 0.020001 }, { "epoch": 1.6187124818878078, "grad_norm": 2.053001775702235, "learning_rate": 1.9658484525080045e-06, "loss": 0.62299452, "memory(GiB)": 48.58, "step": 3910, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.620782446698406, "grad_norm": 1.867769590995953, "learning_rate": 1.955176093916756e-06, "loss": 0.61298056, "memory(GiB)": 48.58, "step": 3915, "train_speed(iter/s)": 0.020002 }, { "epoch": 1.6228524115090044, "grad_norm": 2.3195386012179404, "learning_rate": 1.944503735325507e-06, "loss": 0.62579803, "memory(GiB)": 48.58, "step": 3920, "train_speed(iter/s)": 0.020003 }, { "epoch": 1.6249223763196026, "grad_norm": 2.0728629845134825, "learning_rate": 1.9338313767342585e-06, "loss": 0.60280285, "memory(GiB)": 48.58, "step": 3925, "train_speed(iter/s)": 0.020004 }, { "epoch": 1.6269923411302007, "grad_norm": 1.9603574858854882, "learning_rate": 1.92315901814301e-06, "loss": 0.61659031, "memory(GiB)": 48.58, "step": 3930, "train_speed(iter/s)": 0.020005 }, { "epoch": 1.6290623059407991, "grad_norm": 2.1782071337826014, "learning_rate": 1.912486659551761e-06, "loss": 0.60701981, "memory(GiB)": 48.58, "step": 3935, "train_speed(iter/s)": 0.020006 }, { "epoch": 1.6311322707513973, "grad_norm": 1.7195215240069541, "learning_rate": 1.9018143009605124e-06, "loss": 0.59961052, "memory(GiB)": 48.58, "step": 3940, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.6332022355619955, "grad_norm": 2.076700307294715, "learning_rate": 1.8911419423692637e-06, "loss": 0.61397924, "memory(GiB)": 48.58, "step": 3945, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.6352722003725937, "grad_norm": 2.1171856192990615, "learning_rate": 1.8804695837780151e-06, "loss": 0.61544151, "memory(GiB)": 48.58, "step": 3950, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.6373421651831919, "grad_norm": 2.2158147695337838, "learning_rate": 1.8697972251867664e-06, "loss": 0.62159052, "memory(GiB)": 48.58, "step": 3955, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.63941212999379, "grad_norm": 1.8635077630631116, "learning_rate": 1.8591248665955179e-06, "loss": 0.62834597, "memory(GiB)": 48.58, "step": 3960, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.6414820948043882, "grad_norm": 1.9610933756803843, "learning_rate": 1.8484525080042693e-06, "loss": 0.61779599, "memory(GiB)": 48.58, "step": 3965, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.6435520596149864, "grad_norm": 1.9029383953647434, "learning_rate": 1.8377801494130204e-06, "loss": 0.61077566, "memory(GiB)": 48.58, "step": 3970, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.6456220244255848, "grad_norm": 2.0846014405355926, "learning_rate": 1.8271077908217718e-06, "loss": 0.61311092, "memory(GiB)": 48.58, "step": 3975, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.647691989236183, "grad_norm": 2.390238831095919, "learning_rate": 1.816435432230523e-06, "loss": 0.62842617, "memory(GiB)": 48.58, "step": 3980, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.6497619540467812, "grad_norm": 1.8650578046674164, "learning_rate": 1.8057630736392745e-06, "loss": 0.62033787, "memory(GiB)": 48.58, "step": 3985, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.6518319188573796, "grad_norm": 1.9997553186688102, "learning_rate": 1.7950907150480258e-06, "loss": 0.61835356, "memory(GiB)": 48.58, "step": 3990, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.6539018836679777, "grad_norm": 2.0185275543308, "learning_rate": 1.7844183564567772e-06, "loss": 0.62564411, "memory(GiB)": 48.58, "step": 3995, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.655971848478576, "grad_norm": 1.6493866482983615, "learning_rate": 1.7737459978655283e-06, "loss": 0.62476611, "memory(GiB)": 48.58, "step": 4000, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.658041813289174, "grad_norm": 2.1258878217039157, "learning_rate": 1.7630736392742797e-06, "loss": 0.61094875, "memory(GiB)": 48.58, "step": 4005, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.6601117780997723, "grad_norm": 1.8264630776919117, "learning_rate": 1.7524012806830312e-06, "loss": 0.61455221, "memory(GiB)": 48.58, "step": 4010, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.6621817429103705, "grad_norm": 2.1834934664197148, "learning_rate": 1.7417289220917825e-06, "loss": 0.61525717, "memory(GiB)": 48.58, "step": 4015, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.6642517077209686, "grad_norm": 1.951531129714831, "learning_rate": 1.731056563500534e-06, "loss": 0.61558003, "memory(GiB)": 48.58, "step": 4020, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.6663216725315668, "grad_norm": 2.4480446682769315, "learning_rate": 1.720384204909285e-06, "loss": 0.58283405, "memory(GiB)": 48.58, "step": 4025, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.6683916373421652, "grad_norm": 1.7243360581014375, "learning_rate": 1.7097118463180364e-06, "loss": 0.60816731, "memory(GiB)": 48.58, "step": 4030, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.6704616021527634, "grad_norm": 1.935378370315751, "learning_rate": 1.6990394877267877e-06, "loss": 0.59363813, "memory(GiB)": 48.58, "step": 4035, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.6725315669633616, "grad_norm": 2.343341523870396, "learning_rate": 1.6883671291355391e-06, "loss": 0.59670277, "memory(GiB)": 48.58, "step": 4040, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.67460153177396, "grad_norm": 2.002429630691932, "learning_rate": 1.6776947705442904e-06, "loss": 0.61318674, "memory(GiB)": 48.58, "step": 4045, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.6766714965845582, "grad_norm": 2.161836907625474, "learning_rate": 1.6670224119530418e-06, "loss": 0.63845253, "memory(GiB)": 48.58, "step": 4050, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.6787414613951563, "grad_norm": 2.3095783835557993, "learning_rate": 1.6563500533617933e-06, "loss": 0.61825991, "memory(GiB)": 48.58, "step": 4055, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.6808114262057545, "grad_norm": 2.0857994277393326, "learning_rate": 1.6456776947705443e-06, "loss": 0.58271861, "memory(GiB)": 48.58, "step": 4060, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.6828813910163527, "grad_norm": 2.591242513945162, "learning_rate": 1.6350053361792958e-06, "loss": 0.59729037, "memory(GiB)": 48.58, "step": 4065, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.6849513558269509, "grad_norm": 2.2408993824641836, "learning_rate": 1.624332977588047e-06, "loss": 0.61958027, "memory(GiB)": 48.58, "step": 4070, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.687021320637549, "grad_norm": 2.07836511660639, "learning_rate": 1.6136606189967985e-06, "loss": 0.61059999, "memory(GiB)": 48.58, "step": 4075, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.6890912854481472, "grad_norm": 2.1885333872092767, "learning_rate": 1.6029882604055498e-06, "loss": 0.61813364, "memory(GiB)": 48.58, "step": 4080, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.6911612502587456, "grad_norm": 2.0563236890963075, "learning_rate": 1.5923159018143012e-06, "loss": 0.62983589, "memory(GiB)": 48.58, "step": 4085, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.6932312150693438, "grad_norm": 2.205936525088634, "learning_rate": 1.5816435432230523e-06, "loss": 0.6022356, "memory(GiB)": 48.58, "step": 4090, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.695301179879942, "grad_norm": 2.1192285983309262, "learning_rate": 1.5709711846318037e-06, "loss": 0.61405392, "memory(GiB)": 48.58, "step": 4095, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.6973711446905404, "grad_norm": 2.2747714483339676, "learning_rate": 1.5602988260405552e-06, "loss": 0.59890566, "memory(GiB)": 48.58, "step": 4100, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.6994411095011386, "grad_norm": 1.7865480667421139, "learning_rate": 1.5496264674493064e-06, "loss": 0.59438276, "memory(GiB)": 48.58, "step": 4105, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.7015110743117368, "grad_norm": 2.1136132702931953, "learning_rate": 1.538954108858058e-06, "loss": 0.58430662, "memory(GiB)": 48.58, "step": 4110, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.703581039122335, "grad_norm": 1.8888155653077559, "learning_rate": 1.528281750266809e-06, "loss": 0.61480141, "memory(GiB)": 48.58, "step": 4115, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.7056510039329331, "grad_norm": 2.1453583525948567, "learning_rate": 1.5176093916755604e-06, "loss": 0.60317545, "memory(GiB)": 48.58, "step": 4120, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.7077209687435313, "grad_norm": 2.0869705530610174, "learning_rate": 1.5069370330843117e-06, "loss": 0.57888694, "memory(GiB)": 48.58, "step": 4125, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.7097909335541295, "grad_norm": 2.0589261216950177, "learning_rate": 1.4962646744930631e-06, "loss": 0.63887987, "memory(GiB)": 48.58, "step": 4130, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.7118608983647277, "grad_norm": 2.246554976018598, "learning_rate": 1.4855923159018144e-06, "loss": 0.60080147, "memory(GiB)": 48.58, "step": 4135, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.713930863175326, "grad_norm": 2.2465140455191377, "learning_rate": 1.4749199573105658e-06, "loss": 0.62593145, "memory(GiB)": 48.58, "step": 4140, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.7160008279859242, "grad_norm": 2.1975792886927135, "learning_rate": 1.4642475987193173e-06, "loss": 0.61860814, "memory(GiB)": 48.58, "step": 4145, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.7180707927965224, "grad_norm": 1.8897839022485312, "learning_rate": 1.4535752401280683e-06, "loss": 0.62411423, "memory(GiB)": 48.58, "step": 4150, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.7201407576071208, "grad_norm": 2.4140521953157794, "learning_rate": 1.4429028815368198e-06, "loss": 0.60574412, "memory(GiB)": 48.58, "step": 4155, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.722210722417719, "grad_norm": 2.04528954072566, "learning_rate": 1.432230522945571e-06, "loss": 0.61065197, "memory(GiB)": 48.58, "step": 4160, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.7242806872283172, "grad_norm": 2.544468455409069, "learning_rate": 1.4215581643543225e-06, "loss": 0.59008269, "memory(GiB)": 48.58, "step": 4165, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.7263506520389154, "grad_norm": 1.9933726482621115, "learning_rate": 1.4108858057630738e-06, "loss": 0.61828232, "memory(GiB)": 48.58, "step": 4170, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.7284206168495135, "grad_norm": 2.0938798089462702, "learning_rate": 1.4002134471718252e-06, "loss": 0.63251686, "memory(GiB)": 48.58, "step": 4175, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.7304905816601117, "grad_norm": 2.586866443962492, "learning_rate": 1.3895410885805763e-06, "loss": 0.61843009, "memory(GiB)": 48.58, "step": 4180, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.73256054647071, "grad_norm": 2.157748885604068, "learning_rate": 1.3788687299893277e-06, "loss": 0.61311278, "memory(GiB)": 48.58, "step": 4185, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.734630511281308, "grad_norm": 2.1925495515847944, "learning_rate": 1.3681963713980792e-06, "loss": 0.61852412, "memory(GiB)": 48.58, "step": 4190, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.7367004760919065, "grad_norm": 2.0519214102963566, "learning_rate": 1.3575240128068304e-06, "loss": 0.60840869, "memory(GiB)": 48.58, "step": 4195, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.7387704409025047, "grad_norm": 2.1138169542373335, "learning_rate": 1.3468516542155819e-06, "loss": 0.60299668, "memory(GiB)": 48.58, "step": 4200, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.7387704409025047, "eval_loss": 0.7988836765289307, "eval_runtime": 333.377, "eval_samples_per_second": 18.736, "eval_steps_per_second": 1.173, "step": 4200 }, { "epoch": 1.7408404057131028, "grad_norm": 1.8176186505765208, "learning_rate": 1.336179295624333e-06, "loss": 0.59205551, "memory(GiB)": 48.58, "step": 4205, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.7429103705237012, "grad_norm": 2.097301273083197, "learning_rate": 1.3255069370330844e-06, "loss": 0.59544134, "memory(GiB)": 48.58, "step": 4210, "train_speed(iter/s)": 0.020007 }, { "epoch": 1.7449803353342994, "grad_norm": 2.1754814306718093, "learning_rate": 1.3148345784418356e-06, "loss": 0.61374321, "memory(GiB)": 48.58, "step": 4215, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.7470503001448976, "grad_norm": 2.1091284971781543, "learning_rate": 1.304162219850587e-06, "loss": 0.58602142, "memory(GiB)": 48.58, "step": 4220, "train_speed(iter/s)": 0.020008 }, { "epoch": 1.7491202649554958, "grad_norm": 2.1153060430505177, "learning_rate": 1.2934898612593383e-06, "loss": 0.59698052, "memory(GiB)": 48.58, "step": 4225, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.751190229766094, "grad_norm": 2.3476551963062193, "learning_rate": 1.2828175026680898e-06, "loss": 0.63100615, "memory(GiB)": 48.58, "step": 4230, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.7532601945766921, "grad_norm": 2.1723309903381987, "learning_rate": 1.2721451440768413e-06, "loss": 0.61266661, "memory(GiB)": 48.58, "step": 4235, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.7553301593872903, "grad_norm": 2.4855935431363267, "learning_rate": 1.2614727854855923e-06, "loss": 0.60998316, "memory(GiB)": 48.58, "step": 4240, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.7574001241978885, "grad_norm": 2.5103101197105233, "learning_rate": 1.2508004268943438e-06, "loss": 0.61193109, "memory(GiB)": 48.58, "step": 4245, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.759470089008487, "grad_norm": 2.2253279455991195, "learning_rate": 1.2401280683030952e-06, "loss": 0.58802786, "memory(GiB)": 48.58, "step": 4250, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.761540053819085, "grad_norm": 2.412011728154561, "learning_rate": 1.2294557097118465e-06, "loss": 0.6127542, "memory(GiB)": 48.58, "step": 4255, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.7636100186296833, "grad_norm": 2.186989181036045, "learning_rate": 1.2187833511205977e-06, "loss": 0.59003277, "memory(GiB)": 48.58, "step": 4260, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.7656799834402817, "grad_norm": 2.528633630931418, "learning_rate": 1.2081109925293492e-06, "loss": 0.61563702, "memory(GiB)": 48.58, "step": 4265, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.7677499482508798, "grad_norm": 1.9867035049680128, "learning_rate": 1.1974386339381004e-06, "loss": 0.60806894, "memory(GiB)": 48.58, "step": 4270, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.769819913061478, "grad_norm": 2.2593003701423755, "learning_rate": 1.1867662753468517e-06, "loss": 0.61837912, "memory(GiB)": 48.58, "step": 4275, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.7718898778720762, "grad_norm": 2.4545294029220797, "learning_rate": 1.176093916755603e-06, "loss": 0.61774817, "memory(GiB)": 48.58, "step": 4280, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.7739598426826744, "grad_norm": 1.9914997887032238, "learning_rate": 1.1654215581643544e-06, "loss": 0.58686681, "memory(GiB)": 48.58, "step": 4285, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.7760298074932725, "grad_norm": 2.0999532307692896, "learning_rate": 1.1547491995731057e-06, "loss": 0.60145164, "memory(GiB)": 48.58, "step": 4290, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.7780997723038707, "grad_norm": 2.480798626827661, "learning_rate": 1.1440768409818571e-06, "loss": 0.61948671, "memory(GiB)": 48.58, "step": 4295, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.780169737114469, "grad_norm": 2.1964203822462527, "learning_rate": 1.1334044823906084e-06, "loss": 0.5903161, "memory(GiB)": 48.58, "step": 4300, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.7822397019250673, "grad_norm": 2.2986740177948914, "learning_rate": 1.1227321237993598e-06, "loss": 0.60875359, "memory(GiB)": 48.58, "step": 4305, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.7843096667356655, "grad_norm": 1.994450780247921, "learning_rate": 1.112059765208111e-06, "loss": 0.61587105, "memory(GiB)": 48.58, "step": 4310, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.7863796315462637, "grad_norm": 2.096310617308162, "learning_rate": 1.1013874066168623e-06, "loss": 0.58633337, "memory(GiB)": 48.58, "step": 4315, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.788449596356862, "grad_norm": 1.9466363269467861, "learning_rate": 1.0907150480256138e-06, "loss": 0.60186481, "memory(GiB)": 48.58, "step": 4320, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.7905195611674602, "grad_norm": 2.308498919041305, "learning_rate": 1.080042689434365e-06, "loss": 0.59321814, "memory(GiB)": 48.58, "step": 4325, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.7925895259780584, "grad_norm": 2.6206625429667874, "learning_rate": 1.0693703308431163e-06, "loss": 0.61461964, "memory(GiB)": 48.58, "step": 4330, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.7946594907886566, "grad_norm": 2.4886244412597094, "learning_rate": 1.0586979722518678e-06, "loss": 0.61646094, "memory(GiB)": 48.58, "step": 4335, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.7967294555992548, "grad_norm": 1.9987554252944004, "learning_rate": 1.0480256136606192e-06, "loss": 0.60513401, "memory(GiB)": 48.58, "step": 4340, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.798799420409853, "grad_norm": 2.909678280995542, "learning_rate": 1.0373532550693705e-06, "loss": 0.58900928, "memory(GiB)": 48.58, "step": 4345, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.8008693852204511, "grad_norm": 2.232043207476952, "learning_rate": 1.0266808964781217e-06, "loss": 0.61423898, "memory(GiB)": 48.58, "step": 4350, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.8029393500310493, "grad_norm": 1.9536032593654695, "learning_rate": 1.0160085378868732e-06, "loss": 0.61663337, "memory(GiB)": 48.58, "step": 4355, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.8050093148416477, "grad_norm": 2.5224030653638048, "learning_rate": 1.0053361792956244e-06, "loss": 0.61187458, "memory(GiB)": 48.58, "step": 4360, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.807079279652246, "grad_norm": 1.7925531365832896, "learning_rate": 9.946638207043757e-07, "loss": 0.59958668, "memory(GiB)": 48.58, "step": 4365, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.809149244462844, "grad_norm": 2.000073758007796, "learning_rate": 9.839914621131271e-07, "loss": 0.62268171, "memory(GiB)": 48.58, "step": 4370, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.8112192092734425, "grad_norm": 1.8927383998351053, "learning_rate": 9.733191035218784e-07, "loss": 0.61844292, "memory(GiB)": 48.58, "step": 4375, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.8132891740840407, "grad_norm": 2.2603010222528708, "learning_rate": 9.626467449306296e-07, "loss": 0.60280704, "memory(GiB)": 48.58, "step": 4380, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.8153591388946388, "grad_norm": 2.0743626326418494, "learning_rate": 9.519743863393811e-07, "loss": 0.59184837, "memory(GiB)": 48.58, "step": 4385, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.817429103705237, "grad_norm": 2.5328803417607273, "learning_rate": 9.413020277481325e-07, "loss": 0.60617228, "memory(GiB)": 48.58, "step": 4390, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.8194990685158352, "grad_norm": 2.157648955819526, "learning_rate": 9.306296691568837e-07, "loss": 0.57980437, "memory(GiB)": 48.58, "step": 4395, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.8215690333264334, "grad_norm": 2.238654145184739, "learning_rate": 9.199573105656351e-07, "loss": 0.59073811, "memory(GiB)": 48.58, "step": 4400, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.8236389981370316, "grad_norm": 2.1560196555189677, "learning_rate": 9.092849519743864e-07, "loss": 0.60031624, "memory(GiB)": 48.58, "step": 4405, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.8257089629476297, "grad_norm": 1.9778064514427567, "learning_rate": 8.986125933831377e-07, "loss": 0.59476948, "memory(GiB)": 48.58, "step": 4410, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.8277789277582281, "grad_norm": 1.9062652252413357, "learning_rate": 8.87940234791889e-07, "loss": 0.60492353, "memory(GiB)": 48.58, "step": 4415, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.8298488925688263, "grad_norm": 2.20251758356039, "learning_rate": 8.772678762006404e-07, "loss": 0.57913284, "memory(GiB)": 48.58, "step": 4420, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.8319188573794245, "grad_norm": 2.374670539942745, "learning_rate": 8.665955176093919e-07, "loss": 0.61436529, "memory(GiB)": 48.58, "step": 4425, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.833988822190023, "grad_norm": 1.8014659618328237, "learning_rate": 8.559231590181431e-07, "loss": 0.58954339, "memory(GiB)": 48.58, "step": 4430, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.836058787000621, "grad_norm": 2.5328263785548706, "learning_rate": 8.452508004268945e-07, "loss": 0.62326274, "memory(GiB)": 48.58, "step": 4435, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.8381287518112193, "grad_norm": 2.201541673623922, "learning_rate": 8.345784418356458e-07, "loss": 0.61668777, "memory(GiB)": 48.58, "step": 4440, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.8401987166218174, "grad_norm": 1.9916726147288757, "learning_rate": 8.239060832443971e-07, "loss": 0.60296612, "memory(GiB)": 48.58, "step": 4445, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.8422686814324156, "grad_norm": 2.2611511913260167, "learning_rate": 8.132337246531484e-07, "loss": 0.59926748, "memory(GiB)": 48.58, "step": 4450, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.8443386462430138, "grad_norm": 1.8225104361277575, "learning_rate": 8.025613660618997e-07, "loss": 0.58590517, "memory(GiB)": 48.58, "step": 4455, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.846408611053612, "grad_norm": 2.350449897326385, "learning_rate": 7.91889007470651e-07, "loss": 0.60878654, "memory(GiB)": 48.58, "step": 4460, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.8484785758642102, "grad_norm": 2.4294512189314075, "learning_rate": 7.812166488794024e-07, "loss": 0.60740719, "memory(GiB)": 48.58, "step": 4465, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.8505485406748086, "grad_norm": 2.0078565564513413, "learning_rate": 7.705442902881538e-07, "loss": 0.61569843, "memory(GiB)": 48.58, "step": 4470, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.8526185054854067, "grad_norm": 2.416347350001525, "learning_rate": 7.598719316969051e-07, "loss": 0.59907522, "memory(GiB)": 48.58, "step": 4475, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.854688470296005, "grad_norm": 2.5546773910049523, "learning_rate": 7.491995731056565e-07, "loss": 0.58857327, "memory(GiB)": 48.58, "step": 4480, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.8567584351066033, "grad_norm": 2.1560857096799775, "learning_rate": 7.385272145144078e-07, "loss": 0.59972405, "memory(GiB)": 48.58, "step": 4485, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.8588283999172015, "grad_norm": 2.8045450674344155, "learning_rate": 7.278548559231591e-07, "loss": 0.62519865, "memory(GiB)": 48.58, "step": 4490, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.8608983647277997, "grad_norm": 1.88470841271018, "learning_rate": 7.171824973319104e-07, "loss": 0.59394321, "memory(GiB)": 48.58, "step": 4495, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.8629683295383979, "grad_norm": 1.900323108929672, "learning_rate": 7.065101387406617e-07, "loss": 0.59171925, "memory(GiB)": 48.58, "step": 4500, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.8629683295383979, "eval_loss": 0.7995001077651978, "eval_runtime": 333.5022, "eval_samples_per_second": 18.729, "eval_steps_per_second": 1.172, "step": 4500 }, { "epoch": 1.865038294348996, "grad_norm": 2.3418380686910854, "learning_rate": 6.95837780149413e-07, "loss": 0.62550597, "memory(GiB)": 48.58, "step": 4505, "train_speed(iter/s)": 0.020009 }, { "epoch": 1.8671082591595942, "grad_norm": 2.2011802954910826, "learning_rate": 6.851654215581644e-07, "loss": 0.60017891, "memory(GiB)": 48.58, "step": 4510, "train_speed(iter/s)": 0.02001 }, { "epoch": 1.8691782239701924, "grad_norm": 2.346777967941492, "learning_rate": 6.744930629669158e-07, "loss": 0.62277446, "memory(GiB)": 48.58, "step": 4515, "train_speed(iter/s)": 0.020011 }, { "epoch": 1.8712481887807906, "grad_norm": 2.3039363321887385, "learning_rate": 6.638207043756671e-07, "loss": 0.63189201, "memory(GiB)": 48.58, "step": 4520, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.873318153591389, "grad_norm": 2.4254296862394753, "learning_rate": 6.531483457844184e-07, "loss": 0.59847736, "memory(GiB)": 48.58, "step": 4525, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.8753881184019872, "grad_norm": 2.520645231015452, "learning_rate": 6.424759871931698e-07, "loss": 0.60772429, "memory(GiB)": 48.58, "step": 4530, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.8774580832125853, "grad_norm": 2.14246492812266, "learning_rate": 6.31803628601921e-07, "loss": 0.61204777, "memory(GiB)": 48.58, "step": 4535, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.8795280480231837, "grad_norm": 1.9822250514442827, "learning_rate": 6.211312700106724e-07, "loss": 0.55947261, "memory(GiB)": 48.58, "step": 4540, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.881598012833782, "grad_norm": 2.052778758307509, "learning_rate": 6.104589114194238e-07, "loss": 0.60927758, "memory(GiB)": 48.58, "step": 4545, "train_speed(iter/s)": 0.020015 }, { "epoch": 1.88366797764438, "grad_norm": 2.3908992672472222, "learning_rate": 5.997865528281751e-07, "loss": 0.61153603, "memory(GiB)": 48.58, "step": 4550, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.8857379424549783, "grad_norm": 1.9114509697538704, "learning_rate": 5.891141942369264e-07, "loss": 0.6099647, "memory(GiB)": 48.58, "step": 4555, "train_speed(iter/s)": 0.020016 }, { "epoch": 1.8878079072655765, "grad_norm": 2.063913038194293, "learning_rate": 5.784418356456777e-07, "loss": 0.61249609, "memory(GiB)": 48.58, "step": 4560, "train_speed(iter/s)": 0.020017 }, { "epoch": 1.8898778720761746, "grad_norm": 2.3274588748356404, "learning_rate": 5.677694770544291e-07, "loss": 0.60335112, "memory(GiB)": 48.58, "step": 4565, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.8919478368867728, "grad_norm": 2.5803919299014404, "learning_rate": 5.570971184631804e-07, "loss": 0.59715414, "memory(GiB)": 48.58, "step": 4570, "train_speed(iter/s)": 0.020018 }, { "epoch": 1.894017801697371, "grad_norm": 2.0982349697996727, "learning_rate": 5.464247598719318e-07, "loss": 0.59020104, "memory(GiB)": 48.58, "step": 4575, "train_speed(iter/s)": 0.020019 }, { "epoch": 1.8960877665079694, "grad_norm": 2.4156310349424492, "learning_rate": 5.35752401280683e-07, "loss": 0.58418913, "memory(GiB)": 48.58, "step": 4580, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.8981577313185676, "grad_norm": 2.1192384982329977, "learning_rate": 5.250800426894344e-07, "loss": 0.57976351, "memory(GiB)": 48.58, "step": 4585, "train_speed(iter/s)": 0.02002 }, { "epoch": 1.9002276961291658, "grad_norm": 2.363144438078418, "learning_rate": 5.144076840981858e-07, "loss": 0.59181528, "memory(GiB)": 48.58, "step": 4590, "train_speed(iter/s)": 0.020021 }, { "epoch": 1.9022976609397642, "grad_norm": 2.014916601780685, "learning_rate": 5.037353255069371e-07, "loss": 0.57993307, "memory(GiB)": 48.58, "step": 4595, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.9043676257503623, "grad_norm": 2.1798732452842615, "learning_rate": 4.930629669156884e-07, "loss": 0.61651163, "memory(GiB)": 48.58, "step": 4600, "train_speed(iter/s)": 0.020022 }, { "epoch": 1.9064375905609605, "grad_norm": 2.4067029242954847, "learning_rate": 4.823906083244397e-07, "loss": 0.60284195, "memory(GiB)": 48.58, "step": 4605, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.9085075553715587, "grad_norm": 2.1464043153264587, "learning_rate": 4.7171824973319113e-07, "loss": 0.58217282, "memory(GiB)": 48.58, "step": 4610, "train_speed(iter/s)": 0.020023 }, { "epoch": 1.9105775201821569, "grad_norm": 2.786031828765745, "learning_rate": 4.6104589114194243e-07, "loss": 0.59245019, "memory(GiB)": 48.58, "step": 4615, "train_speed(iter/s)": 0.020024 }, { "epoch": 1.912647484992755, "grad_norm": 1.971002889872401, "learning_rate": 4.5037353255069374e-07, "loss": 0.55735373, "memory(GiB)": 48.58, "step": 4620, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.9147174498033532, "grad_norm": 2.6133112222197097, "learning_rate": 4.3970117395944504e-07, "loss": 0.58715906, "memory(GiB)": 48.58, "step": 4625, "train_speed(iter/s)": 0.020025 }, { "epoch": 1.9167874146139514, "grad_norm": 2.2493774425876496, "learning_rate": 4.290288153681964e-07, "loss": 0.61737943, "memory(GiB)": 48.58, "step": 4630, "train_speed(iter/s)": 0.020026 }, { "epoch": 1.9188573794245498, "grad_norm": 2.066157101262181, "learning_rate": 4.1835645677694775e-07, "loss": 0.58552856, "memory(GiB)": 48.58, "step": 4635, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.920927344235148, "grad_norm": 2.2279403099524164, "learning_rate": 4.076840981856991e-07, "loss": 0.60430613, "memory(GiB)": 48.58, "step": 4640, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.9229973090457462, "grad_norm": 2.21706254471347, "learning_rate": 3.970117395944504e-07, "loss": 0.60209589, "memory(GiB)": 48.58, "step": 4645, "train_speed(iter/s)": 0.020027 }, { "epoch": 1.9250672738563446, "grad_norm": 1.8175472634449323, "learning_rate": 3.863393810032017e-07, "loss": 0.57723808, "memory(GiB)": 48.58, "step": 4650, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.9271372386669428, "grad_norm": 2.013736958312471, "learning_rate": 3.756670224119531e-07, "loss": 0.59187717, "memory(GiB)": 48.58, "step": 4655, "train_speed(iter/s)": 0.020028 }, { "epoch": 1.929207203477541, "grad_norm": 2.9147581447233883, "learning_rate": 3.649946638207044e-07, "loss": 0.62887087, "memory(GiB)": 48.58, "step": 4660, "train_speed(iter/s)": 0.020029 }, { "epoch": 1.9312771682881391, "grad_norm": 2.1113909828567605, "learning_rate": 3.5432230522945573e-07, "loss": 0.60369582, "memory(GiB)": 48.58, "step": 4665, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.9333471330987373, "grad_norm": 2.42442448807154, "learning_rate": 3.4364994663820703e-07, "loss": 0.62186384, "memory(GiB)": 48.58, "step": 4670, "train_speed(iter/s)": 0.02003 }, { "epoch": 1.9354170979093355, "grad_norm": 2.380814562965916, "learning_rate": 3.329775880469584e-07, "loss": 0.59275131, "memory(GiB)": 48.58, "step": 4675, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.9374870627199337, "grad_norm": 2.6584440377132363, "learning_rate": 3.2230522945570974e-07, "loss": 0.60647793, "memory(GiB)": 48.58, "step": 4680, "train_speed(iter/s)": 0.020031 }, { "epoch": 1.9395570275305318, "grad_norm": 1.9567852417003078, "learning_rate": 3.116328708644611e-07, "loss": 0.60262537, "memory(GiB)": 48.58, "step": 4685, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.9416269923411302, "grad_norm": 2.3501175824898266, "learning_rate": 3.009605122732124e-07, "loss": 0.58797078, "memory(GiB)": 48.58, "step": 4690, "train_speed(iter/s)": 0.020032 }, { "epoch": 1.9436969571517284, "grad_norm": 2.237666059037871, "learning_rate": 2.9028815368196376e-07, "loss": 0.60981102, "memory(GiB)": 48.58, "step": 4695, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.9457669219623266, "grad_norm": 2.4233069808853993, "learning_rate": 2.7961579509071506e-07, "loss": 0.60726156, "memory(GiB)": 48.58, "step": 4700, "train_speed(iter/s)": 0.020033 }, { "epoch": 1.947836886772925, "grad_norm": 2.260763775980905, "learning_rate": 2.689434364994664e-07, "loss": 0.60588284, "memory(GiB)": 48.58, "step": 4705, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.9499068515835232, "grad_norm": 2.5081221538985625, "learning_rate": 2.582710779082177e-07, "loss": 0.59843764, "memory(GiB)": 48.58, "step": 4710, "train_speed(iter/s)": 0.020034 }, { "epoch": 1.9519768163941214, "grad_norm": 2.3881113527972304, "learning_rate": 2.475987193169691e-07, "loss": 0.58006935, "memory(GiB)": 48.58, "step": 4715, "train_speed(iter/s)": 0.020035 }, { "epoch": 1.9540467812047195, "grad_norm": 2.0678535427683564, "learning_rate": 2.369263607257204e-07, "loss": 0.60157442, "memory(GiB)": 48.58, "step": 4720, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.9561167460153177, "grad_norm": 2.438506236548372, "learning_rate": 2.2625400213447176e-07, "loss": 0.61028309, "memory(GiB)": 48.58, "step": 4725, "train_speed(iter/s)": 0.020036 }, { "epoch": 1.9581867108259159, "grad_norm": 2.0110599859655482, "learning_rate": 2.1558164354322307e-07, "loss": 0.59143724, "memory(GiB)": 48.58, "step": 4730, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.960256675636514, "grad_norm": 2.0688837840068124, "learning_rate": 2.049092849519744e-07, "loss": 0.60122604, "memory(GiB)": 48.58, "step": 4735, "train_speed(iter/s)": 0.020037 }, { "epoch": 1.9623266404471122, "grad_norm": 2.1607611011042964, "learning_rate": 1.9423692636072575e-07, "loss": 0.59857554, "memory(GiB)": 48.58, "step": 4740, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.9643966052577106, "grad_norm": 1.9355873743150782, "learning_rate": 1.8356456776947706e-07, "loss": 0.58816404, "memory(GiB)": 48.58, "step": 4745, "train_speed(iter/s)": 0.020038 }, { "epoch": 1.9664665700683088, "grad_norm": 2.1028560489657915, "learning_rate": 1.728922091782284e-07, "loss": 0.59075899, "memory(GiB)": 48.58, "step": 4750, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.968536534878907, "grad_norm": 1.9716163219927025, "learning_rate": 1.6221985058697972e-07, "loss": 0.58181, "memory(GiB)": 48.58, "step": 4755, "train_speed(iter/s)": 0.020039 }, { "epoch": 1.9706064996895054, "grad_norm": 2.243337925247892, "learning_rate": 1.5154749199573107e-07, "loss": 0.58783703, "memory(GiB)": 48.58, "step": 4760, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.9726764645001036, "grad_norm": 2.4859440712948166, "learning_rate": 1.408751334044824e-07, "loss": 0.60655708, "memory(GiB)": 48.58, "step": 4765, "train_speed(iter/s)": 0.02004 }, { "epoch": 1.9747464293107018, "grad_norm": 1.9831273691126385, "learning_rate": 1.3020277481323373e-07, "loss": 0.61188507, "memory(GiB)": 48.58, "step": 4770, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.9768163941213, "grad_norm": 2.2314049587449807, "learning_rate": 1.1953041622198506e-07, "loss": 0.58718634, "memory(GiB)": 48.58, "step": 4775, "train_speed(iter/s)": 0.020041 }, { "epoch": 1.9788863589318981, "grad_norm": 1.8150041254780722, "learning_rate": 1.088580576307364e-07, "loss": 0.58888893, "memory(GiB)": 48.58, "step": 4780, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.9809563237424963, "grad_norm": 2.247396093692947, "learning_rate": 9.818569903948773e-08, "loss": 0.58732767, "memory(GiB)": 48.58, "step": 4785, "train_speed(iter/s)": 0.020042 }, { "epoch": 1.9830262885530945, "grad_norm": 2.2098523979831644, "learning_rate": 8.751334044823908e-08, "loss": 0.56731772, "memory(GiB)": 48.58, "step": 4790, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.9850962533636927, "grad_norm": 2.097191818370972, "learning_rate": 7.68409818569904e-08, "loss": 0.60409393, "memory(GiB)": 48.58, "step": 4795, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.987166218174291, "grad_norm": 2.228865758940409, "learning_rate": 6.616862326574174e-08, "loss": 0.58583736, "memory(GiB)": 48.58, "step": 4800, "train_speed(iter/s)": 0.020043 }, { "epoch": 1.987166218174291, "eval_loss": 0.7990086674690247, "eval_runtime": 335.524, "eval_samples_per_second": 18.616, "eval_steps_per_second": 1.165, "step": 4800 }, { "epoch": 1.9892361829848892, "grad_norm": 2.144126117644109, "learning_rate": 5.5496264674493065e-08, "loss": 0.59295273, "memory(GiB)": 48.58, "step": 4805, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.9913061477954874, "grad_norm": 2.020692421672203, "learning_rate": 4.48239060832444e-08, "loss": 0.58215327, "memory(GiB)": 48.58, "step": 4810, "train_speed(iter/s)": 0.020012 }, { "epoch": 1.9933761126060858, "grad_norm": 2.560575104653987, "learning_rate": 3.415154749199574e-08, "loss": 0.59236603, "memory(GiB)": 48.58, "step": 4815, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.995446077416684, "grad_norm": 2.242799455811118, "learning_rate": 2.347918890074707e-08, "loss": 0.59284697, "memory(GiB)": 48.58, "step": 4820, "train_speed(iter/s)": 0.020013 }, { "epoch": 1.9975160422272822, "grad_norm": 2.496931570288639, "learning_rate": 1.28068303094984e-08, "loss": 0.58884125, "memory(GiB)": 48.58, "step": 4825, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.9995860070378804, "grad_norm": 2.1504541545847133, "learning_rate": 2.1344717182497336e-09, "loss": 0.57651815, "memory(GiB)": 48.58, "step": 4830, "train_speed(iter/s)": 0.020014 }, { "epoch": 1.9995860070378804, "eval_loss": 0.7990483641624451, "eval_runtime": 333.2574, "eval_samples_per_second": 18.742, "eval_steps_per_second": 1.173, "step": 4830 } ], "logging_steps": 5, "max_steps": 4830, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.0363792423256064e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }