llava-llama3.1-8b / trainer_state.json
DaozeZhang
add ckpt
a0a784e
raw
history blame
241 kB
{
"best_metric": 0.79888368,
"best_model_checkpoint": "/mnt/nas1/daoze/code/swift/output/llava1_6-llama3_1-8b-instruct-my/v33-20240901-191352/checkpoint-4200",
"epoch": 1.9995860070378804,
"eval_steps": 300,
"global_step": 4830,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00041399296211964395,
"grad_norm": 68.31517175098004,
"learning_rate": 0.0,
"loss": 7.09566689,
"memory(GiB)": 42.23,
"step": 1,
"train_speed(iter/s)": 0.013537
},
{
"epoch": 0.0020699648105982197,
"grad_norm": 75.27486731705963,
"learning_rate": 3.2339240870284233e-06,
"loss": 4.4965229,
"memory(GiB)": 48.78,
"step": 5,
"train_speed(iter/s)": 0.018721
},
{
"epoch": 0.004139929621196439,
"grad_norm": 7.7589717582741615,
"learning_rate": 4.626699381900465e-06,
"loss": 1.79170246,
"memory(GiB)": 52.1,
"step": 10,
"train_speed(iter/s)": 0.01954
},
{
"epoch": 0.00620989443179466,
"grad_norm": 4.8519342782653485,
"learning_rate": 5.44142070133146e-06,
"loss": 1.36023655,
"memory(GiB)": 52.1,
"step": 15,
"train_speed(iter/s)": 0.019784
},
{
"epoch": 0.008279859242392879,
"grad_norm": 3.375818313593022,
"learning_rate": 6.0194746767725065e-06,
"loss": 1.18811779,
"memory(GiB)": 52.1,
"step": 20,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 0.0103498240529911,
"grad_norm": 3.5284378077706813,
"learning_rate": 6.467848174056847e-06,
"loss": 1.1600296,
"memory(GiB)": 52.1,
"step": 25,
"train_speed(iter/s)": 0.020192
},
{
"epoch": 0.01241978886358932,
"grad_norm": 3.163950589670924,
"learning_rate": 6.834195996203502e-06,
"loss": 1.11185207,
"memory(GiB)": 52.1,
"step": 30,
"train_speed(iter/s)": 0.020268
},
{
"epoch": 0.014489753674187538,
"grad_norm": 2.212271768126557,
"learning_rate": 7.143938666407679e-06,
"loss": 1.0538393,
"memory(GiB)": 52.1,
"step": 35,
"train_speed(iter/s)": 0.020328
},
{
"epoch": 0.016559718484785758,
"grad_norm": 3.493503288451801,
"learning_rate": 7.412249971644547e-06,
"loss": 1.04750004,
"memory(GiB)": 52.1,
"step": 40,
"train_speed(iter/s)": 0.020393
},
{
"epoch": 0.01862968329538398,
"grad_norm": 2.6291679253427644,
"learning_rate": 7.648917315634497e-06,
"loss": 1.04756851,
"memory(GiB)": 52.1,
"step": 45,
"train_speed(iter/s)": 0.020444
},
{
"epoch": 0.0206996481059822,
"grad_norm": 3.1377476353112312,
"learning_rate": 7.860623468928888e-06,
"loss": 1.00407467,
"memory(GiB)": 52.1,
"step": 50,
"train_speed(iter/s)": 0.02043
},
{
"epoch": 0.022769612916580417,
"grad_norm": 2.6547100112946733,
"learning_rate": 8.05213497976565e-06,
"loss": 1.01097145,
"memory(GiB)": 52.1,
"step": 55,
"train_speed(iter/s)": 0.020407
},
{
"epoch": 0.02483957772717864,
"grad_norm": 2.6383576740248555,
"learning_rate": 8.226971291075542e-06,
"loss": 0.98796005,
"memory(GiB)": 52.1,
"step": 60,
"train_speed(iter/s)": 0.020419
},
{
"epoch": 0.02690954253777686,
"grad_norm": 2.542613928151568,
"learning_rate": 8.387805106618597e-06,
"loss": 0.98850031,
"memory(GiB)": 52.1,
"step": 65,
"train_speed(iter/s)": 0.020424
},
{
"epoch": 0.028979507348375077,
"grad_norm": 2.877488922903144,
"learning_rate": 8.536713961279723e-06,
"loss": 0.98919926,
"memory(GiB)": 52.1,
"step": 70,
"train_speed(iter/s)": 0.020439
},
{
"epoch": 0.031049472158973298,
"grad_norm": 1.8237010588157114,
"learning_rate": 8.675344788359883e-06,
"loss": 1.012813,
"memory(GiB)": 52.1,
"step": 75,
"train_speed(iter/s)": 0.020466
},
{
"epoch": 0.033119436969571515,
"grad_norm": 2.5261639534973788,
"learning_rate": 8.805025266516589e-06,
"loss": 0.98800411,
"memory(GiB)": 52.1,
"step": 80,
"train_speed(iter/s)": 0.02047
},
{
"epoch": 0.035189401780169736,
"grad_norm": 2.2711243603226032,
"learning_rate": 8.926841351029377e-06,
"loss": 1.00139637,
"memory(GiB)": 55.52,
"step": 85,
"train_speed(iter/s)": 0.020479
},
{
"epoch": 0.03725936659076796,
"grad_norm": 2.145948654847171,
"learning_rate": 9.041692610506539e-06,
"loss": 0.99127426,
"memory(GiB)": 55.52,
"step": 90,
"train_speed(iter/s)": 0.020491
},
{
"epoch": 0.03932933140136618,
"grad_norm": 2.836740011107602,
"learning_rate": 9.150332582159872e-06,
"loss": 0.96552677,
"memory(GiB)": 55.52,
"step": 95,
"train_speed(iter/s)": 0.020506
},
{
"epoch": 0.0413992962119644,
"grad_norm": 2.2398611362766006,
"learning_rate": 9.25339876380093e-06,
"loss": 0.98038893,
"memory(GiB)": 55.52,
"step": 100,
"train_speed(iter/s)": 0.020501
},
{
"epoch": 0.043469261022562614,
"grad_norm": 2.4986121497854192,
"learning_rate": 9.351435280710716e-06,
"loss": 0.9604641,
"memory(GiB)": 63.35,
"step": 105,
"train_speed(iter/s)": 0.020504
},
{
"epoch": 0.045539225833160835,
"grad_norm": 2.1909523678523755,
"learning_rate": 9.444910274637691e-06,
"loss": 0.93479166,
"memory(GiB)": 63.35,
"step": 110,
"train_speed(iter/s)": 0.020519
},
{
"epoch": 0.047609190643759056,
"grad_norm": 2.5301628247983663,
"learning_rate": 9.534229424247679e-06,
"loss": 0.97118149,
"memory(GiB)": 63.35,
"step": 115,
"train_speed(iter/s)": 0.020522
},
{
"epoch": 0.04967915545435728,
"grad_norm": 1.92102667189958,
"learning_rate": 9.619746585947584e-06,
"loss": 0.97038708,
"memory(GiB)": 63.35,
"step": 120,
"train_speed(iter/s)": 0.020545
},
{
"epoch": 0.0517491202649555,
"grad_norm": 2.2968578337197036,
"learning_rate": 9.701772261085271e-06,
"loss": 0.96241703,
"memory(GiB)": 63.35,
"step": 125,
"train_speed(iter/s)": 0.020544
},
{
"epoch": 0.05381908507555372,
"grad_norm": 2.110866882726078,
"learning_rate": 9.780580401490638e-06,
"loss": 0.95122089,
"memory(GiB)": 63.35,
"step": 130,
"train_speed(iter/s)": 0.020551
},
{
"epoch": 0.05588904988615193,
"grad_norm": 2.082289679777962,
"learning_rate": 9.856413929937534e-06,
"loss": 0.95173302,
"memory(GiB)": 63.35,
"step": 135,
"train_speed(iter/s)": 0.020557
},
{
"epoch": 0.057959014696750154,
"grad_norm": 2.6318522567639784,
"learning_rate": 9.929489256151762e-06,
"loss": 0.94850836,
"memory(GiB)": 63.35,
"step": 140,
"train_speed(iter/s)": 0.020559
},
{
"epoch": 0.060028979507348375,
"grad_norm": 2.091985474615185,
"learning_rate": 1e-05,
"loss": 0.91576376,
"memory(GiB)": 63.35,
"step": 145,
"train_speed(iter/s)": 0.020557
},
{
"epoch": 0.062098944317946596,
"grad_norm": 2.3841665763842705,
"learning_rate": 9.991462113127002e-06,
"loss": 0.93616524,
"memory(GiB)": 63.35,
"step": 150,
"train_speed(iter/s)": 0.020563
},
{
"epoch": 0.06416890912854481,
"grad_norm": 2.1042524174231505,
"learning_rate": 9.980789754535753e-06,
"loss": 0.9277298,
"memory(GiB)": 63.35,
"step": 155,
"train_speed(iter/s)": 0.020558
},
{
"epoch": 0.06623887393914303,
"grad_norm": 2.276077986025351,
"learning_rate": 9.970117395944504e-06,
"loss": 0.95900288,
"memory(GiB)": 63.35,
"step": 160,
"train_speed(iter/s)": 0.020556
},
{
"epoch": 0.06830883874974125,
"grad_norm": 2.2459536201915387,
"learning_rate": 9.959445037353256e-06,
"loss": 0.94394236,
"memory(GiB)": 63.35,
"step": 165,
"train_speed(iter/s)": 0.020552
},
{
"epoch": 0.07037880356033947,
"grad_norm": 2.517051668755459,
"learning_rate": 9.948772678762007e-06,
"loss": 0.94310379,
"memory(GiB)": 63.35,
"step": 170,
"train_speed(iter/s)": 0.020552
},
{
"epoch": 0.0724487683709377,
"grad_norm": 2.2026553318124313,
"learning_rate": 9.938100320170759e-06,
"loss": 0.95436573,
"memory(GiB)": 63.35,
"step": 175,
"train_speed(iter/s)": 0.020554
},
{
"epoch": 0.07451873318153591,
"grad_norm": 2.2476130486911465,
"learning_rate": 9.92742796157951e-06,
"loss": 0.93582458,
"memory(GiB)": 63.35,
"step": 180,
"train_speed(iter/s)": 0.020553
},
{
"epoch": 0.07658869799213414,
"grad_norm": 2.388512026493847,
"learning_rate": 9.916755602988262e-06,
"loss": 0.91813745,
"memory(GiB)": 63.35,
"step": 185,
"train_speed(iter/s)": 0.020558
},
{
"epoch": 0.07865866280273236,
"grad_norm": 2.3220003726970204,
"learning_rate": 9.906083244397012e-06,
"loss": 0.92115765,
"memory(GiB)": 63.35,
"step": 190,
"train_speed(iter/s)": 0.020561
},
{
"epoch": 0.08072862761333058,
"grad_norm": 2.3768930570990805,
"learning_rate": 9.895410885805764e-06,
"loss": 0.91676846,
"memory(GiB)": 63.35,
"step": 195,
"train_speed(iter/s)": 0.020562
},
{
"epoch": 0.0827985924239288,
"grad_norm": 4.2054479128966165,
"learning_rate": 9.884738527214515e-06,
"loss": 0.91852398,
"memory(GiB)": 63.35,
"step": 200,
"train_speed(iter/s)": 0.020566
},
{
"epoch": 0.084868557234527,
"grad_norm": 2.14133807915419,
"learning_rate": 9.874066168623266e-06,
"loss": 0.91718044,
"memory(GiB)": 63.35,
"step": 205,
"train_speed(iter/s)": 0.020563
},
{
"epoch": 0.08693852204512523,
"grad_norm": 1.9390816379786193,
"learning_rate": 9.863393810032017e-06,
"loss": 0.92340775,
"memory(GiB)": 63.35,
"step": 210,
"train_speed(iter/s)": 0.020559
},
{
"epoch": 0.08900848685572345,
"grad_norm": 2.1261095243289208,
"learning_rate": 9.852721451440769e-06,
"loss": 0.90159931,
"memory(GiB)": 63.35,
"step": 215,
"train_speed(iter/s)": 0.020561
},
{
"epoch": 0.09107845166632167,
"grad_norm": 2.9284829291227243,
"learning_rate": 9.842049092849521e-06,
"loss": 0.91123104,
"memory(GiB)": 63.35,
"step": 220,
"train_speed(iter/s)": 0.020557
},
{
"epoch": 0.09314841647691989,
"grad_norm": 2.7049739018013583,
"learning_rate": 9.831376734258272e-06,
"loss": 0.941084,
"memory(GiB)": 63.35,
"step": 225,
"train_speed(iter/s)": 0.02056
},
{
"epoch": 0.09521838128751811,
"grad_norm": 2.0467249315549845,
"learning_rate": 9.820704375667023e-06,
"loss": 0.90574436,
"memory(GiB)": 63.35,
"step": 230,
"train_speed(iter/s)": 0.020561
},
{
"epoch": 0.09728834609811633,
"grad_norm": 2.1061950654127006,
"learning_rate": 9.810032017075774e-06,
"loss": 0.92933855,
"memory(GiB)": 63.35,
"step": 235,
"train_speed(iter/s)": 0.020568
},
{
"epoch": 0.09935831090871455,
"grad_norm": 2.390745868374031,
"learning_rate": 9.799359658484527e-06,
"loss": 0.9383173,
"memory(GiB)": 63.35,
"step": 240,
"train_speed(iter/s)": 0.02057
},
{
"epoch": 0.10142827571931277,
"grad_norm": 1.9948725536279355,
"learning_rate": 9.788687299893276e-06,
"loss": 0.85975437,
"memory(GiB)": 63.35,
"step": 245,
"train_speed(iter/s)": 0.020577
},
{
"epoch": 0.103498240529911,
"grad_norm": 2.29936461247775,
"learning_rate": 9.77801494130203e-06,
"loss": 0.90683613,
"memory(GiB)": 63.47,
"step": 250,
"train_speed(iter/s)": 0.020575
},
{
"epoch": 0.10556820534050922,
"grad_norm": 1.8961456174283475,
"learning_rate": 9.76734258271078e-06,
"loss": 0.88947477,
"memory(GiB)": 63.47,
"step": 255,
"train_speed(iter/s)": 0.020574
},
{
"epoch": 0.10763817015110744,
"grad_norm": 1.8558977060217532,
"learning_rate": 9.756670224119531e-06,
"loss": 0.90292645,
"memory(GiB)": 63.47,
"step": 260,
"train_speed(iter/s)": 0.020581
},
{
"epoch": 0.10970813496170564,
"grad_norm": 2.1017797656853725,
"learning_rate": 9.745997865528282e-06,
"loss": 0.92776756,
"memory(GiB)": 63.47,
"step": 265,
"train_speed(iter/s)": 0.020579
},
{
"epoch": 0.11177809977230387,
"grad_norm": 1.9051226867861688,
"learning_rate": 9.735325506937033e-06,
"loss": 0.90565796,
"memory(GiB)": 63.59,
"step": 270,
"train_speed(iter/s)": 0.020578
},
{
"epoch": 0.11384806458290209,
"grad_norm": 1.932217719000402,
"learning_rate": 9.724653148345784e-06,
"loss": 0.90403481,
"memory(GiB)": 63.59,
"step": 275,
"train_speed(iter/s)": 0.020575
},
{
"epoch": 0.11591802939350031,
"grad_norm": 3.033913485092789,
"learning_rate": 9.713980789754537e-06,
"loss": 0.86916351,
"memory(GiB)": 63.59,
"step": 280,
"train_speed(iter/s)": 0.020575
},
{
"epoch": 0.11798799420409853,
"grad_norm": 1.9434774416915237,
"learning_rate": 9.703308431163288e-06,
"loss": 0.87491503,
"memory(GiB)": 63.59,
"step": 285,
"train_speed(iter/s)": 0.020569
},
{
"epoch": 0.12005795901469675,
"grad_norm": 2.051766135268311,
"learning_rate": 9.69263607257204e-06,
"loss": 0.89817352,
"memory(GiB)": 63.59,
"step": 290,
"train_speed(iter/s)": 0.020572
},
{
"epoch": 0.12212792382529497,
"grad_norm": 2.2615047465793796,
"learning_rate": 9.68196371398079e-06,
"loss": 0.89681797,
"memory(GiB)": 63.59,
"step": 295,
"train_speed(iter/s)": 0.020574
},
{
"epoch": 0.12419788863589319,
"grad_norm": 2.3348971307519637,
"learning_rate": 9.671291355389541e-06,
"loss": 0.91658554,
"memory(GiB)": 63.59,
"step": 300,
"train_speed(iter/s)": 0.020576
},
{
"epoch": 0.12419788863589319,
"eval_loss": 0.9112715721130371,
"eval_runtime": 338.0523,
"eval_samples_per_second": 18.476,
"eval_steps_per_second": 1.157,
"step": 300
},
{
"epoch": 0.1262678534464914,
"grad_norm": 2.0234892490546614,
"learning_rate": 9.660618996798294e-06,
"loss": 0.89021435,
"memory(GiB)": 63.59,
"step": 305,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 0.12833781825708962,
"grad_norm": 2.3363054707830195,
"learning_rate": 9.649946638207045e-06,
"loss": 0.87254162,
"memory(GiB)": 63.59,
"step": 310,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.13040778306768785,
"grad_norm": 1.820741232007416,
"learning_rate": 9.639274279615796e-06,
"loss": 0.90060663,
"memory(GiB)": 63.59,
"step": 315,
"train_speed(iter/s)": 0.020066
},
{
"epoch": 0.13247774787828606,
"grad_norm": 1.9084755504752218,
"learning_rate": 9.628601921024547e-06,
"loss": 0.8869771,
"memory(GiB)": 63.59,
"step": 320,
"train_speed(iter/s)": 0.020075
},
{
"epoch": 0.1345477126888843,
"grad_norm": 1.8856515103808584,
"learning_rate": 9.617929562433298e-06,
"loss": 0.87808056,
"memory(GiB)": 63.59,
"step": 325,
"train_speed(iter/s)": 0.020084
},
{
"epoch": 0.1366176774994825,
"grad_norm": 1.9384672045466198,
"learning_rate": 9.60725720384205e-06,
"loss": 0.88911896,
"memory(GiB)": 63.59,
"step": 330,
"train_speed(iter/s)": 0.020095
},
{
"epoch": 0.13868764231008074,
"grad_norm": 1.9948177737503383,
"learning_rate": 9.596584845250802e-06,
"loss": 0.8958828,
"memory(GiB)": 63.59,
"step": 335,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.14075760712067895,
"grad_norm": 2.2317608749634874,
"learning_rate": 9.585912486659551e-06,
"loss": 0.89190845,
"memory(GiB)": 63.59,
"step": 340,
"train_speed(iter/s)": 0.020112
},
{
"epoch": 0.14282757193127718,
"grad_norm": 2.1298996429991908,
"learning_rate": 9.575240128068304e-06,
"loss": 0.87826462,
"memory(GiB)": 63.59,
"step": 345,
"train_speed(iter/s)": 0.02012
},
{
"epoch": 0.1448975367418754,
"grad_norm": 1.9367356654913552,
"learning_rate": 9.564567769477055e-06,
"loss": 0.88935146,
"memory(GiB)": 63.59,
"step": 350,
"train_speed(iter/s)": 0.020128
},
{
"epoch": 0.1469675015524736,
"grad_norm": 2.965286627679984,
"learning_rate": 9.553895410885806e-06,
"loss": 0.91815538,
"memory(GiB)": 63.59,
"step": 355,
"train_speed(iter/s)": 0.020135
},
{
"epoch": 0.14903746636307183,
"grad_norm": 2.322705848413725,
"learning_rate": 9.543223052294557e-06,
"loss": 0.89534588,
"memory(GiB)": 63.59,
"step": 360,
"train_speed(iter/s)": 0.020143
},
{
"epoch": 0.15110743117367004,
"grad_norm": 1.9467115924709606,
"learning_rate": 9.53255069370331e-06,
"loss": 0.87547607,
"memory(GiB)": 63.59,
"step": 365,
"train_speed(iter/s)": 0.020146
},
{
"epoch": 0.15317739598426827,
"grad_norm": 2.056244423727218,
"learning_rate": 9.521878335112061e-06,
"loss": 0.87522736,
"memory(GiB)": 63.59,
"step": 370,
"train_speed(iter/s)": 0.02015
},
{
"epoch": 0.15524736079486648,
"grad_norm": 1.9570323722204157,
"learning_rate": 9.511205976520812e-06,
"loss": 0.88565502,
"memory(GiB)": 63.59,
"step": 375,
"train_speed(iter/s)": 0.020154
},
{
"epoch": 0.15731732560546471,
"grad_norm": 1.9075809920338722,
"learning_rate": 9.500533617929563e-06,
"loss": 0.87662697,
"memory(GiB)": 63.59,
"step": 380,
"train_speed(iter/s)": 0.020159
},
{
"epoch": 0.15938729041606292,
"grad_norm": 2.3674969791857983,
"learning_rate": 9.489861259338314e-06,
"loss": 0.88602619,
"memory(GiB)": 63.59,
"step": 385,
"train_speed(iter/s)": 0.020163
},
{
"epoch": 0.16145725522666116,
"grad_norm": 2.157304860709474,
"learning_rate": 9.479188900747067e-06,
"loss": 0.86379642,
"memory(GiB)": 63.59,
"step": 390,
"train_speed(iter/s)": 0.02017
},
{
"epoch": 0.16352722003725936,
"grad_norm": 1.8795995948139297,
"learning_rate": 9.468516542155816e-06,
"loss": 0.88137684,
"memory(GiB)": 63.59,
"step": 395,
"train_speed(iter/s)": 0.020176
},
{
"epoch": 0.1655971848478576,
"grad_norm": 1.8996836205094734,
"learning_rate": 9.457844183564569e-06,
"loss": 0.86949444,
"memory(GiB)": 63.59,
"step": 400,
"train_speed(iter/s)": 0.020182
},
{
"epoch": 0.1676671496584558,
"grad_norm": 2.0977604679128854,
"learning_rate": 9.44717182497332e-06,
"loss": 0.85557442,
"memory(GiB)": 63.59,
"step": 405,
"train_speed(iter/s)": 0.020184
},
{
"epoch": 0.169737114469054,
"grad_norm": 2.2253818762342155,
"learning_rate": 9.436499466382071e-06,
"loss": 0.85497751,
"memory(GiB)": 63.59,
"step": 410,
"train_speed(iter/s)": 0.02019
},
{
"epoch": 0.17180707927965225,
"grad_norm": 1.9007105346828383,
"learning_rate": 9.425827107790822e-06,
"loss": 0.86072025,
"memory(GiB)": 63.59,
"step": 415,
"train_speed(iter/s)": 0.020193
},
{
"epoch": 0.17387704409025045,
"grad_norm": 2.1940409564689656,
"learning_rate": 9.415154749199575e-06,
"loss": 0.89283857,
"memory(GiB)": 63.59,
"step": 420,
"train_speed(iter/s)": 0.020198
},
{
"epoch": 0.1759470089008487,
"grad_norm": 2.2711317580338912,
"learning_rate": 9.404482390608326e-06,
"loss": 0.8422184,
"memory(GiB)": 63.59,
"step": 425,
"train_speed(iter/s)": 0.020202
},
{
"epoch": 0.1780169737114469,
"grad_norm": 1.9734252369885248,
"learning_rate": 9.393810032017077e-06,
"loss": 0.90587616,
"memory(GiB)": 63.59,
"step": 430,
"train_speed(iter/s)": 0.020208
},
{
"epoch": 0.18008693852204513,
"grad_norm": 2.1358019624149653,
"learning_rate": 9.383137673425828e-06,
"loss": 0.87137203,
"memory(GiB)": 63.59,
"step": 435,
"train_speed(iter/s)": 0.020212
},
{
"epoch": 0.18215690333264334,
"grad_norm": 2.136663123639741,
"learning_rate": 9.372465314834579e-06,
"loss": 0.84276152,
"memory(GiB)": 63.59,
"step": 440,
"train_speed(iter/s)": 0.020218
},
{
"epoch": 0.18422686814324157,
"grad_norm": 2.1730526677654005,
"learning_rate": 9.361792956243332e-06,
"loss": 0.86879997,
"memory(GiB)": 63.59,
"step": 445,
"train_speed(iter/s)": 0.020222
},
{
"epoch": 0.18629683295383978,
"grad_norm": 1.7007840288725673,
"learning_rate": 9.351120597652081e-06,
"loss": 0.85356216,
"memory(GiB)": 63.59,
"step": 450,
"train_speed(iter/s)": 0.020224
},
{
"epoch": 0.18836679776443802,
"grad_norm": 2.333224503644692,
"learning_rate": 9.340448239060834e-06,
"loss": 0.87298975,
"memory(GiB)": 63.59,
"step": 455,
"train_speed(iter/s)": 0.02023
},
{
"epoch": 0.19043676257503622,
"grad_norm": 1.7497884985137717,
"learning_rate": 9.329775880469585e-06,
"loss": 0.89487724,
"memory(GiB)": 63.59,
"step": 460,
"train_speed(iter/s)": 0.02023
},
{
"epoch": 0.19250672738563446,
"grad_norm": 1.7881559718064066,
"learning_rate": 9.319103521878336e-06,
"loss": 0.88939381,
"memory(GiB)": 63.59,
"step": 465,
"train_speed(iter/s)": 0.020232
},
{
"epoch": 0.19457669219623266,
"grad_norm": 2.5056663267756605,
"learning_rate": 9.308431163287087e-06,
"loss": 0.85123787,
"memory(GiB)": 63.59,
"step": 470,
"train_speed(iter/s)": 0.020236
},
{
"epoch": 0.19664665700683087,
"grad_norm": 2.2621024798210403,
"learning_rate": 9.29775880469584e-06,
"loss": 0.87380323,
"memory(GiB)": 63.59,
"step": 475,
"train_speed(iter/s)": 0.020238
},
{
"epoch": 0.1987166218174291,
"grad_norm": 1.9008868824283842,
"learning_rate": 9.287086446104589e-06,
"loss": 0.82988033,
"memory(GiB)": 63.59,
"step": 480,
"train_speed(iter/s)": 0.020241
},
{
"epoch": 0.2007865866280273,
"grad_norm": 1.6960491773696469,
"learning_rate": 9.276414087513342e-06,
"loss": 0.85842476,
"memory(GiB)": 63.59,
"step": 485,
"train_speed(iter/s)": 0.020243
},
{
"epoch": 0.20285655143862555,
"grad_norm": 1.8755094068628242,
"learning_rate": 9.265741728922093e-06,
"loss": 0.86819458,
"memory(GiB)": 63.59,
"step": 490,
"train_speed(iter/s)": 0.020249
},
{
"epoch": 0.20492651624922376,
"grad_norm": 1.6029255395235227,
"learning_rate": 9.255069370330844e-06,
"loss": 0.86832209,
"memory(GiB)": 63.59,
"step": 495,
"train_speed(iter/s)": 0.020253
},
{
"epoch": 0.206996481059822,
"grad_norm": 1.8312986531673774,
"learning_rate": 9.244397011739595e-06,
"loss": 0.87804108,
"memory(GiB)": 63.59,
"step": 500,
"train_speed(iter/s)": 0.020256
},
{
"epoch": 0.2090664458704202,
"grad_norm": 2.2159879518724686,
"learning_rate": 9.233724653148346e-06,
"loss": 0.84401827,
"memory(GiB)": 63.59,
"step": 505,
"train_speed(iter/s)": 0.020258
},
{
"epoch": 0.21113641068101843,
"grad_norm": 1.9489190185976173,
"learning_rate": 9.223052294557098e-06,
"loss": 0.83081837,
"memory(GiB)": 63.59,
"step": 510,
"train_speed(iter/s)": 0.020262
},
{
"epoch": 0.21320637549161664,
"grad_norm": 1.8621375658015202,
"learning_rate": 9.21237993596585e-06,
"loss": 0.84451389,
"memory(GiB)": 63.59,
"step": 515,
"train_speed(iter/s)": 0.020265
},
{
"epoch": 0.21527634030221487,
"grad_norm": 2.657592267470185,
"learning_rate": 9.2017075773746e-06,
"loss": 0.85252399,
"memory(GiB)": 63.59,
"step": 520,
"train_speed(iter/s)": 0.020272
},
{
"epoch": 0.21734630511281308,
"grad_norm": 3.2134734541192556,
"learning_rate": 9.191035218783352e-06,
"loss": 0.85981674,
"memory(GiB)": 63.59,
"step": 525,
"train_speed(iter/s)": 0.020278
},
{
"epoch": 0.2194162699234113,
"grad_norm": 1.979014196110588,
"learning_rate": 9.180362860192104e-06,
"loss": 0.85360508,
"memory(GiB)": 63.59,
"step": 530,
"train_speed(iter/s)": 0.020281
},
{
"epoch": 0.22148623473400952,
"grad_norm": 2.263346027010783,
"learning_rate": 9.169690501600854e-06,
"loss": 0.8649641,
"memory(GiB)": 63.59,
"step": 535,
"train_speed(iter/s)": 0.020285
},
{
"epoch": 0.22355619954460773,
"grad_norm": 2.0257701801426786,
"learning_rate": 9.159018143009606e-06,
"loss": 0.85079117,
"memory(GiB)": 63.59,
"step": 540,
"train_speed(iter/s)": 0.020291
},
{
"epoch": 0.22562616435520597,
"grad_norm": 2.125007231598407,
"learning_rate": 9.148345784418357e-06,
"loss": 0.83591347,
"memory(GiB)": 63.59,
"step": 545,
"train_speed(iter/s)": 0.020292
},
{
"epoch": 0.22769612916580417,
"grad_norm": 1.830826573395782,
"learning_rate": 9.137673425827108e-06,
"loss": 0.8597187,
"memory(GiB)": 63.59,
"step": 550,
"train_speed(iter/s)": 0.020297
},
{
"epoch": 0.2297660939764024,
"grad_norm": 2.145744349446719,
"learning_rate": 9.12700106723586e-06,
"loss": 0.82627001,
"memory(GiB)": 63.59,
"step": 555,
"train_speed(iter/s)": 0.020299
},
{
"epoch": 0.23183605878700062,
"grad_norm": 1.895819945079046,
"learning_rate": 9.116328708644612e-06,
"loss": 0.83418722,
"memory(GiB)": 63.59,
"step": 560,
"train_speed(iter/s)": 0.020301
},
{
"epoch": 0.23390602359759885,
"grad_norm": 1.7330277741570008,
"learning_rate": 9.105656350053362e-06,
"loss": 0.83898754,
"memory(GiB)": 63.59,
"step": 565,
"train_speed(iter/s)": 0.020301
},
{
"epoch": 0.23597598840819706,
"grad_norm": 1.873262235276853,
"learning_rate": 9.094983991462114e-06,
"loss": 0.86104965,
"memory(GiB)": 63.59,
"step": 570,
"train_speed(iter/s)": 0.020307
},
{
"epoch": 0.2380459532187953,
"grad_norm": 1.8133555971052358,
"learning_rate": 9.084311632870865e-06,
"loss": 0.83760166,
"memory(GiB)": 63.59,
"step": 575,
"train_speed(iter/s)": 0.020309
},
{
"epoch": 0.2401159180293935,
"grad_norm": 2.1026280097135377,
"learning_rate": 9.073639274279616e-06,
"loss": 0.85497513,
"memory(GiB)": 63.59,
"step": 580,
"train_speed(iter/s)": 0.02031
},
{
"epoch": 0.2421858828399917,
"grad_norm": 1.7385609176743078,
"learning_rate": 9.062966915688367e-06,
"loss": 0.83206367,
"memory(GiB)": 63.59,
"step": 585,
"train_speed(iter/s)": 0.02031
},
{
"epoch": 0.24425584765058994,
"grad_norm": 2.310226819070514,
"learning_rate": 9.052294557097118e-06,
"loss": 0.82417412,
"memory(GiB)": 63.59,
"step": 590,
"train_speed(iter/s)": 0.020312
},
{
"epoch": 0.24632581246118815,
"grad_norm": 2.1660135303280126,
"learning_rate": 9.041622198505871e-06,
"loss": 0.8371232,
"memory(GiB)": 63.59,
"step": 595,
"train_speed(iter/s)": 0.020312
},
{
"epoch": 0.24839577727178638,
"grad_norm": 2.0932218548460493,
"learning_rate": 9.030949839914622e-06,
"loss": 0.86303692,
"memory(GiB)": 63.59,
"step": 600,
"train_speed(iter/s)": 0.020315
},
{
"epoch": 0.24839577727178638,
"eval_loss": 0.8761223554611206,
"eval_runtime": 333.7076,
"eval_samples_per_second": 18.717,
"eval_steps_per_second": 1.172,
"step": 600
},
{
"epoch": 0.2504657420823846,
"grad_norm": 1.9850417129561133,
"learning_rate": 9.020277481323373e-06,
"loss": 0.84975281,
"memory(GiB)": 63.59,
"step": 605,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.2525357068929828,
"grad_norm": 1.8024791223082373,
"learning_rate": 9.009605122732124e-06,
"loss": 0.83602209,
"memory(GiB)": 63.59,
"step": 610,
"train_speed(iter/s)": 0.020063
},
{
"epoch": 0.25460567170358106,
"grad_norm": 1.795807849269691,
"learning_rate": 8.998932764140877e-06,
"loss": 0.84287434,
"memory(GiB)": 63.59,
"step": 615,
"train_speed(iter/s)": 0.020066
},
{
"epoch": 0.25667563651417924,
"grad_norm": 2.8462701485274855,
"learning_rate": 8.988260405549626e-06,
"loss": 0.86969414,
"memory(GiB)": 63.59,
"step": 620,
"train_speed(iter/s)": 0.020069
},
{
"epoch": 0.2587456013247775,
"grad_norm": 1.9467464102992238,
"learning_rate": 8.977588046958379e-06,
"loss": 0.84205284,
"memory(GiB)": 63.59,
"step": 625,
"train_speed(iter/s)": 0.020075
},
{
"epoch": 0.2608155661353757,
"grad_norm": 1.9359113111268293,
"learning_rate": 8.96691568836713e-06,
"loss": 0.83059912,
"memory(GiB)": 63.59,
"step": 630,
"train_speed(iter/s)": 0.02008
},
{
"epoch": 0.26288553094597394,
"grad_norm": 2.269649830017561,
"learning_rate": 8.956243329775881e-06,
"loss": 0.85204124,
"memory(GiB)": 63.59,
"step": 635,
"train_speed(iter/s)": 0.020082
},
{
"epoch": 0.2649554957565721,
"grad_norm": 2.0739328945699014,
"learning_rate": 8.945570971184632e-06,
"loss": 0.83141527,
"memory(GiB)": 63.59,
"step": 640,
"train_speed(iter/s)": 0.020085
},
{
"epoch": 0.26702546056717036,
"grad_norm": 2.39308971265692,
"learning_rate": 8.934898612593383e-06,
"loss": 0.82764578,
"memory(GiB)": 63.59,
"step": 645,
"train_speed(iter/s)": 0.020088
},
{
"epoch": 0.2690954253777686,
"grad_norm": 2.014414879589864,
"learning_rate": 8.924226254002136e-06,
"loss": 0.8550128,
"memory(GiB)": 63.59,
"step": 650,
"train_speed(iter/s)": 0.020095
},
{
"epoch": 0.2711653901883668,
"grad_norm": 2.8374721965360887,
"learning_rate": 8.913553895410887e-06,
"loss": 0.86329079,
"memory(GiB)": 63.59,
"step": 655,
"train_speed(iter/s)": 0.020098
},
{
"epoch": 0.273235354998965,
"grad_norm": 1.9583062241735367,
"learning_rate": 8.902881536819638e-06,
"loss": 0.83884621,
"memory(GiB)": 63.59,
"step": 660,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.27530531980956324,
"grad_norm": 1.765892011718539,
"learning_rate": 8.892209178228389e-06,
"loss": 0.86448555,
"memory(GiB)": 63.59,
"step": 665,
"train_speed(iter/s)": 0.020105
},
{
"epoch": 0.2773752846201615,
"grad_norm": 1.7474590269235404,
"learning_rate": 8.88153681963714e-06,
"loss": 0.81313992,
"memory(GiB)": 63.59,
"step": 670,
"train_speed(iter/s)": 0.020107
},
{
"epoch": 0.27944524943075966,
"grad_norm": 1.9301834126279824,
"learning_rate": 8.870864461045891e-06,
"loss": 0.8374301,
"memory(GiB)": 63.59,
"step": 675,
"train_speed(iter/s)": 0.02011
},
{
"epoch": 0.2815152142413579,
"grad_norm": 1.876790685008959,
"learning_rate": 8.860192102454644e-06,
"loss": 0.82494678,
"memory(GiB)": 63.59,
"step": 680,
"train_speed(iter/s)": 0.020115
},
{
"epoch": 0.2835851790519561,
"grad_norm": 1.7854143077330529,
"learning_rate": 8.849519743863395e-06,
"loss": 0.83442841,
"memory(GiB)": 63.59,
"step": 685,
"train_speed(iter/s)": 0.020116
},
{
"epoch": 0.28565514386255436,
"grad_norm": 1.8923637542669056,
"learning_rate": 8.838847385272146e-06,
"loss": 0.82085514,
"memory(GiB)": 63.72,
"step": 690,
"train_speed(iter/s)": 0.020118
},
{
"epoch": 0.28772510867315254,
"grad_norm": 2.0170179705017066,
"learning_rate": 8.828175026680897e-06,
"loss": 0.85945168,
"memory(GiB)": 63.72,
"step": 695,
"train_speed(iter/s)": 0.020121
},
{
"epoch": 0.2897950734837508,
"grad_norm": 3.3823068740775755,
"learning_rate": 8.817502668089648e-06,
"loss": 0.82407131,
"memory(GiB)": 63.72,
"step": 700,
"train_speed(iter/s)": 0.020123
},
{
"epoch": 0.291865038294349,
"grad_norm": 1.9439481100139924,
"learning_rate": 8.806830309498399e-06,
"loss": 0.82105274,
"memory(GiB)": 63.72,
"step": 705,
"train_speed(iter/s)": 0.020126
},
{
"epoch": 0.2939350031049472,
"grad_norm": 1.8993570568677929,
"learning_rate": 8.796157950907152e-06,
"loss": 0.83391781,
"memory(GiB)": 63.72,
"step": 710,
"train_speed(iter/s)": 0.020131
},
{
"epoch": 0.2960049679155454,
"grad_norm": 2.175929579998878,
"learning_rate": 8.785485592315903e-06,
"loss": 0.84003325,
"memory(GiB)": 63.72,
"step": 715,
"train_speed(iter/s)": 0.020137
},
{
"epoch": 0.29807493272614366,
"grad_norm": 1.6910798969618672,
"learning_rate": 8.774813233724654e-06,
"loss": 0.82005548,
"memory(GiB)": 63.72,
"step": 720,
"train_speed(iter/s)": 0.02014
},
{
"epoch": 0.3001448975367419,
"grad_norm": 1.7577054177826072,
"learning_rate": 8.764140875133405e-06,
"loss": 0.84406672,
"memory(GiB)": 63.72,
"step": 725,
"train_speed(iter/s)": 0.020144
},
{
"epoch": 0.3022148623473401,
"grad_norm": 1.9343300270246129,
"learning_rate": 8.753468516542156e-06,
"loss": 0.81861668,
"memory(GiB)": 63.72,
"step": 730,
"train_speed(iter/s)": 0.020144
},
{
"epoch": 0.3042848271579383,
"grad_norm": 1.8986244788103208,
"learning_rate": 8.742796157950909e-06,
"loss": 0.81786537,
"memory(GiB)": 63.72,
"step": 735,
"train_speed(iter/s)": 0.020148
},
{
"epoch": 0.30635479196853654,
"grad_norm": 2.095799409220846,
"learning_rate": 8.73212379935966e-06,
"loss": 0.83321962,
"memory(GiB)": 63.72,
"step": 740,
"train_speed(iter/s)": 0.020151
},
{
"epoch": 0.3084247567791348,
"grad_norm": 1.9094006901482394,
"learning_rate": 8.72145144076841e-06,
"loss": 0.82653723,
"memory(GiB)": 63.72,
"step": 745,
"train_speed(iter/s)": 0.020152
},
{
"epoch": 0.31049472158973296,
"grad_norm": 2.126120113530993,
"learning_rate": 8.710779082177162e-06,
"loss": 0.85463696,
"memory(GiB)": 63.72,
"step": 750,
"train_speed(iter/s)": 0.020156
},
{
"epoch": 0.3125646864003312,
"grad_norm": 1.766780713214732,
"learning_rate": 8.700106723585913e-06,
"loss": 0.83797083,
"memory(GiB)": 63.72,
"step": 755,
"train_speed(iter/s)": 0.020157
},
{
"epoch": 0.31463465121092943,
"grad_norm": 1.8957319688723608,
"learning_rate": 8.689434364994664e-06,
"loss": 0.81888847,
"memory(GiB)": 63.72,
"step": 760,
"train_speed(iter/s)": 0.020159
},
{
"epoch": 0.3167046160215276,
"grad_norm": 1.9661061189594824,
"learning_rate": 8.678762006403417e-06,
"loss": 0.78800874,
"memory(GiB)": 63.72,
"step": 765,
"train_speed(iter/s)": 0.020164
},
{
"epoch": 0.31877458083212584,
"grad_norm": 1.8837863956075926,
"learning_rate": 8.668089647812166e-06,
"loss": 0.84463196,
"memory(GiB)": 63.72,
"step": 770,
"train_speed(iter/s)": 0.020167
},
{
"epoch": 0.3208445456427241,
"grad_norm": 2.5248078655238326,
"learning_rate": 8.657417289220919e-06,
"loss": 0.83094559,
"memory(GiB)": 63.72,
"step": 775,
"train_speed(iter/s)": 0.02017
},
{
"epoch": 0.3229145104533223,
"grad_norm": 1.8996595550385447,
"learning_rate": 8.64674493062967e-06,
"loss": 0.81705608,
"memory(GiB)": 63.72,
"step": 780,
"train_speed(iter/s)": 0.020173
},
{
"epoch": 0.3249844752639205,
"grad_norm": 1.8243459235808355,
"learning_rate": 8.63607257203842e-06,
"loss": 0.82983418,
"memory(GiB)": 63.72,
"step": 785,
"train_speed(iter/s)": 0.020177
},
{
"epoch": 0.3270544400745187,
"grad_norm": 2.125198435674726,
"learning_rate": 8.625400213447172e-06,
"loss": 0.85153885,
"memory(GiB)": 63.72,
"step": 790,
"train_speed(iter/s)": 0.02018
},
{
"epoch": 0.32912440488511696,
"grad_norm": 1.822527258966965,
"learning_rate": 8.614727854855925e-06,
"loss": 0.7932189,
"memory(GiB)": 63.72,
"step": 795,
"train_speed(iter/s)": 0.020184
},
{
"epoch": 0.3311943696957152,
"grad_norm": 1.9585269031801074,
"learning_rate": 8.604055496264676e-06,
"loss": 0.80502253,
"memory(GiB)": 63.72,
"step": 800,
"train_speed(iter/s)": 0.020188
},
{
"epoch": 0.3332643345063134,
"grad_norm": 1.9244862407118186,
"learning_rate": 8.593383137673427e-06,
"loss": 0.81400661,
"memory(GiB)": 63.72,
"step": 805,
"train_speed(iter/s)": 0.020192
},
{
"epoch": 0.3353342993169116,
"grad_norm": 1.8781928942945239,
"learning_rate": 8.582710779082178e-06,
"loss": 0.82624207,
"memory(GiB)": 63.72,
"step": 810,
"train_speed(iter/s)": 0.020195
},
{
"epoch": 0.33740426412750985,
"grad_norm": 2.4821098212553108,
"learning_rate": 8.572038420490929e-06,
"loss": 0.81296177,
"memory(GiB)": 63.72,
"step": 815,
"train_speed(iter/s)": 0.020197
},
{
"epoch": 0.339474228938108,
"grad_norm": 3.2468832100225877,
"learning_rate": 8.561366061899681e-06,
"loss": 0.81447935,
"memory(GiB)": 63.72,
"step": 820,
"train_speed(iter/s)": 0.0202
},
{
"epoch": 0.34154419374870626,
"grad_norm": 1.726217016729622,
"learning_rate": 8.55069370330843e-06,
"loss": 0.82119083,
"memory(GiB)": 63.72,
"step": 825,
"train_speed(iter/s)": 0.020203
},
{
"epoch": 0.3436141585593045,
"grad_norm": 1.8200397633087098,
"learning_rate": 8.540021344717184e-06,
"loss": 0.80688438,
"memory(GiB)": 63.72,
"step": 830,
"train_speed(iter/s)": 0.020205
},
{
"epoch": 0.34568412336990273,
"grad_norm": 1.7077741062644576,
"learning_rate": 8.529348986125935e-06,
"loss": 0.83244801,
"memory(GiB)": 63.72,
"step": 835,
"train_speed(iter/s)": 0.020205
},
{
"epoch": 0.3477540881805009,
"grad_norm": 2.582896676288874,
"learning_rate": 8.518676627534686e-06,
"loss": 0.81135302,
"memory(GiB)": 63.72,
"step": 840,
"train_speed(iter/s)": 0.020208
},
{
"epoch": 0.34982405299109914,
"grad_norm": 3.4613638587514033,
"learning_rate": 8.508004268943437e-06,
"loss": 0.80561113,
"memory(GiB)": 63.72,
"step": 845,
"train_speed(iter/s)": 0.020209
},
{
"epoch": 0.3518940178016974,
"grad_norm": 1.6179386547462884,
"learning_rate": 8.49733191035219e-06,
"loss": 0.82198238,
"memory(GiB)": 63.72,
"step": 850,
"train_speed(iter/s)": 0.020211
},
{
"epoch": 0.3539639826122956,
"grad_norm": 2.202413903162471,
"learning_rate": 8.48665955176094e-06,
"loss": 0.78598285,
"memory(GiB)": 63.72,
"step": 855,
"train_speed(iter/s)": 0.020214
},
{
"epoch": 0.3560339474228938,
"grad_norm": 1.9513315920239633,
"learning_rate": 8.475987193169691e-06,
"loss": 0.80893326,
"memory(GiB)": 63.72,
"step": 860,
"train_speed(iter/s)": 0.020219
},
{
"epoch": 0.358103912233492,
"grad_norm": 1.9113374189570778,
"learning_rate": 8.465314834578443e-06,
"loss": 0.8136569,
"memory(GiB)": 63.72,
"step": 865,
"train_speed(iter/s)": 0.020222
},
{
"epoch": 0.36017387704409026,
"grad_norm": 2.084935583050277,
"learning_rate": 8.454642475987194e-06,
"loss": 0.81384058,
"memory(GiB)": 63.72,
"step": 870,
"train_speed(iter/s)": 0.020226
},
{
"epoch": 0.3622438418546885,
"grad_norm": 1.6048226105298027,
"learning_rate": 8.443970117395945e-06,
"loss": 0.81689348,
"memory(GiB)": 63.72,
"step": 875,
"train_speed(iter/s)": 0.020228
},
{
"epoch": 0.3643138066652867,
"grad_norm": 1.8081549724032602,
"learning_rate": 8.433297758804696e-06,
"loss": 0.8224082,
"memory(GiB)": 63.72,
"step": 880,
"train_speed(iter/s)": 0.02023
},
{
"epoch": 0.3663837714758849,
"grad_norm": 1.8184484923663322,
"learning_rate": 8.422625400213448e-06,
"loss": 0.78473282,
"memory(GiB)": 63.72,
"step": 885,
"train_speed(iter/s)": 0.020232
},
{
"epoch": 0.36845373628648315,
"grad_norm": 2.010882441005616,
"learning_rate": 8.4119530416222e-06,
"loss": 0.81135426,
"memory(GiB)": 63.72,
"step": 890,
"train_speed(iter/s)": 0.020232
},
{
"epoch": 0.3705237010970813,
"grad_norm": 2.363919887534564,
"learning_rate": 8.40128068303095e-06,
"loss": 0.80090466,
"memory(GiB)": 63.72,
"step": 895,
"train_speed(iter/s)": 0.020234
},
{
"epoch": 0.37259366590767956,
"grad_norm": 1.6332844070852461,
"learning_rate": 8.390608324439701e-06,
"loss": 0.81239138,
"memory(GiB)": 63.72,
"step": 900,
"train_speed(iter/s)": 0.020237
},
{
"epoch": 0.37259366590767956,
"eval_loss": 0.8537026047706604,
"eval_runtime": 333.2325,
"eval_samples_per_second": 18.744,
"eval_steps_per_second": 1.173,
"step": 900
},
{
"epoch": 0.3746636307182778,
"grad_norm": 2.4425379037284607,
"learning_rate": 8.379935965848454e-06,
"loss": 0.81980333,
"memory(GiB)": 63.72,
"step": 905,
"train_speed(iter/s)": 0.02007
},
{
"epoch": 0.37673359552887603,
"grad_norm": 2.2417018048030575,
"learning_rate": 8.369263607257204e-06,
"loss": 0.82684288,
"memory(GiB)": 63.72,
"step": 910,
"train_speed(iter/s)": 0.020071
},
{
"epoch": 0.3788035603394742,
"grad_norm": 1.7587001424417825,
"learning_rate": 8.358591248665956e-06,
"loss": 0.83183241,
"memory(GiB)": 63.72,
"step": 915,
"train_speed(iter/s)": 0.020074
},
{
"epoch": 0.38087352515007245,
"grad_norm": 1.5739208911232379,
"learning_rate": 8.347918890074707e-06,
"loss": 0.76680841,
"memory(GiB)": 63.72,
"step": 920,
"train_speed(iter/s)": 0.020076
},
{
"epoch": 0.3829434899606707,
"grad_norm": 1.8684171867799126,
"learning_rate": 8.337246531483458e-06,
"loss": 0.85463772,
"memory(GiB)": 63.72,
"step": 925,
"train_speed(iter/s)": 0.020079
},
{
"epoch": 0.3850134547712689,
"grad_norm": 1.8550376770414303,
"learning_rate": 8.32657417289221e-06,
"loss": 0.81911144,
"memory(GiB)": 63.72,
"step": 930,
"train_speed(iter/s)": 0.020083
},
{
"epoch": 0.3870834195818671,
"grad_norm": 1.8194471444369447,
"learning_rate": 8.31590181430096e-06,
"loss": 0.83044968,
"memory(GiB)": 63.72,
"step": 935,
"train_speed(iter/s)": 0.020086
},
{
"epoch": 0.38915338439246533,
"grad_norm": 2.095169413730359,
"learning_rate": 8.305229455709713e-06,
"loss": 0.82608871,
"memory(GiB)": 63.72,
"step": 940,
"train_speed(iter/s)": 0.020088
},
{
"epoch": 0.39122334920306356,
"grad_norm": 13.855393242400119,
"learning_rate": 8.294557097118464e-06,
"loss": 0.81676388,
"memory(GiB)": 63.72,
"step": 945,
"train_speed(iter/s)": 0.02009
},
{
"epoch": 0.39329331401366174,
"grad_norm": 2.100046441650532,
"learning_rate": 8.283884738527215e-06,
"loss": 0.81071377,
"memory(GiB)": 63.72,
"step": 950,
"train_speed(iter/s)": 0.020093
},
{
"epoch": 0.39536327882426,
"grad_norm": 2.1608254386705594,
"learning_rate": 8.273212379935966e-06,
"loss": 0.78902674,
"memory(GiB)": 63.72,
"step": 955,
"train_speed(iter/s)": 0.020095
},
{
"epoch": 0.3974332436348582,
"grad_norm": 1.61490095503505,
"learning_rate": 8.262540021344719e-06,
"loss": 0.78527632,
"memory(GiB)": 63.72,
"step": 960,
"train_speed(iter/s)": 0.020097
},
{
"epoch": 0.39950320844545645,
"grad_norm": 2.589460194979307,
"learning_rate": 8.251867662753468e-06,
"loss": 0.81925201,
"memory(GiB)": 63.72,
"step": 965,
"train_speed(iter/s)": 0.020098
},
{
"epoch": 0.4015731732560546,
"grad_norm": 1.8550853889727008,
"learning_rate": 8.241195304162221e-06,
"loss": 0.80688972,
"memory(GiB)": 63.72,
"step": 970,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.40364313806665286,
"grad_norm": 2.4637090355397517,
"learning_rate": 8.230522945570972e-06,
"loss": 0.82061405,
"memory(GiB)": 63.72,
"step": 975,
"train_speed(iter/s)": 0.020104
},
{
"epoch": 0.4057131028772511,
"grad_norm": 1.9846780448601058,
"learning_rate": 8.219850586979723e-06,
"loss": 0.80861397,
"memory(GiB)": 63.72,
"step": 980,
"train_speed(iter/s)": 0.020106
},
{
"epoch": 0.40778306768784933,
"grad_norm": 1.9138045073506678,
"learning_rate": 8.209178228388474e-06,
"loss": 0.79171362,
"memory(GiB)": 63.72,
"step": 985,
"train_speed(iter/s)": 0.020108
},
{
"epoch": 0.4098530324984475,
"grad_norm": 2.100897386160194,
"learning_rate": 8.198505869797227e-06,
"loss": 0.78543482,
"memory(GiB)": 63.72,
"step": 990,
"train_speed(iter/s)": 0.020113
},
{
"epoch": 0.41192299730904575,
"grad_norm": 2.1398479736163667,
"learning_rate": 8.187833511205976e-06,
"loss": 0.79959226,
"memory(GiB)": 63.72,
"step": 995,
"train_speed(iter/s)": 0.020114
},
{
"epoch": 0.413992962119644,
"grad_norm": 1.8543885416746075,
"learning_rate": 8.177161152614729e-06,
"loss": 0.80144148,
"memory(GiB)": 63.72,
"step": 1000,
"train_speed(iter/s)": 0.020116
},
{
"epoch": 0.41606292693024216,
"grad_norm": 1.8895963845216188,
"learning_rate": 8.16648879402348e-06,
"loss": 0.78557086,
"memory(GiB)": 63.72,
"step": 1005,
"train_speed(iter/s)": 0.020118
},
{
"epoch": 0.4181328917408404,
"grad_norm": 1.7583886118404264,
"learning_rate": 8.155816435432231e-06,
"loss": 0.81441746,
"memory(GiB)": 63.72,
"step": 1010,
"train_speed(iter/s)": 0.02012
},
{
"epoch": 0.42020285655143863,
"grad_norm": 1.8640464710188405,
"learning_rate": 8.145144076840982e-06,
"loss": 0.76718016,
"memory(GiB)": 63.72,
"step": 1015,
"train_speed(iter/s)": 0.020122
},
{
"epoch": 0.42227282136203687,
"grad_norm": 2.0754981449007084,
"learning_rate": 8.134471718249733e-06,
"loss": 0.78537526,
"memory(GiB)": 63.72,
"step": 1020,
"train_speed(iter/s)": 0.020125
},
{
"epoch": 0.42434278617263504,
"grad_norm": 2.1358764475250105,
"learning_rate": 8.123799359658486e-06,
"loss": 0.82194328,
"memory(GiB)": 63.72,
"step": 1025,
"train_speed(iter/s)": 0.020128
},
{
"epoch": 0.4264127509832333,
"grad_norm": 1.940572767867165,
"learning_rate": 8.113127001067237e-06,
"loss": 0.8162715,
"memory(GiB)": 63.72,
"step": 1030,
"train_speed(iter/s)": 0.020131
},
{
"epoch": 0.4284827157938315,
"grad_norm": 1.7824953515185047,
"learning_rate": 8.102454642475988e-06,
"loss": 0.78834782,
"memory(GiB)": 63.72,
"step": 1035,
"train_speed(iter/s)": 0.020133
},
{
"epoch": 0.43055268060442975,
"grad_norm": 1.9585541886433688,
"learning_rate": 8.091782283884739e-06,
"loss": 0.79206867,
"memory(GiB)": 63.72,
"step": 1040,
"train_speed(iter/s)": 0.020136
},
{
"epoch": 0.43262264541502793,
"grad_norm": 1.6194935665114412,
"learning_rate": 8.081109925293492e-06,
"loss": 0.80889845,
"memory(GiB)": 63.72,
"step": 1045,
"train_speed(iter/s)": 0.020139
},
{
"epoch": 0.43469261022562616,
"grad_norm": 1.5909296898104581,
"learning_rate": 8.070437566702241e-06,
"loss": 0.76998816,
"memory(GiB)": 63.72,
"step": 1050,
"train_speed(iter/s)": 0.020141
},
{
"epoch": 0.4367625750362244,
"grad_norm": 1.582985265467202,
"learning_rate": 8.059765208110994e-06,
"loss": 0.79827099,
"memory(GiB)": 63.72,
"step": 1055,
"train_speed(iter/s)": 0.020143
},
{
"epoch": 0.4388325398468226,
"grad_norm": 1.9696406410447012,
"learning_rate": 8.049092849519743e-06,
"loss": 0.79220991,
"memory(GiB)": 63.72,
"step": 1060,
"train_speed(iter/s)": 0.020146
},
{
"epoch": 0.4409025046574208,
"grad_norm": 1.9479888997003834,
"learning_rate": 8.038420490928496e-06,
"loss": 0.79184585,
"memory(GiB)": 63.72,
"step": 1065,
"train_speed(iter/s)": 0.020148
},
{
"epoch": 0.44297246946801905,
"grad_norm": 1.7883498032309324,
"learning_rate": 8.027748132337247e-06,
"loss": 0.78507504,
"memory(GiB)": 63.72,
"step": 1070,
"train_speed(iter/s)": 0.02015
},
{
"epoch": 0.4450424342786173,
"grad_norm": 1.6985331753731079,
"learning_rate": 8.017075773745998e-06,
"loss": 0.81149197,
"memory(GiB)": 63.72,
"step": 1075,
"train_speed(iter/s)": 0.020154
},
{
"epoch": 0.44711239908921546,
"grad_norm": 1.7646873640943033,
"learning_rate": 8.006403415154749e-06,
"loss": 0.77548814,
"memory(GiB)": 63.72,
"step": 1080,
"train_speed(iter/s)": 0.020158
},
{
"epoch": 0.4491823638998137,
"grad_norm": 1.7739180508215708,
"learning_rate": 7.995731056563502e-06,
"loss": 0.77309542,
"memory(GiB)": 63.72,
"step": 1085,
"train_speed(iter/s)": 0.02016
},
{
"epoch": 0.45125232871041193,
"grad_norm": 1.6366153166920923,
"learning_rate": 7.985058697972253e-06,
"loss": 0.80448093,
"memory(GiB)": 63.72,
"step": 1090,
"train_speed(iter/s)": 0.020163
},
{
"epoch": 0.45332229352101017,
"grad_norm": 1.9520729703143727,
"learning_rate": 7.974386339381004e-06,
"loss": 0.78142538,
"memory(GiB)": 63.72,
"step": 1095,
"train_speed(iter/s)": 0.020164
},
{
"epoch": 0.45539225833160835,
"grad_norm": 1.9086497373347489,
"learning_rate": 7.963713980789755e-06,
"loss": 0.7876678,
"memory(GiB)": 63.72,
"step": 1100,
"train_speed(iter/s)": 0.020167
},
{
"epoch": 0.4574622231422066,
"grad_norm": 1.773963819363606,
"learning_rate": 7.953041622198506e-06,
"loss": 0.817309,
"memory(GiB)": 63.72,
"step": 1105,
"train_speed(iter/s)": 0.020169
},
{
"epoch": 0.4595321879528048,
"grad_norm": 1.7610901122064158,
"learning_rate": 7.942369263607259e-06,
"loss": 0.81472855,
"memory(GiB)": 63.72,
"step": 1110,
"train_speed(iter/s)": 0.020171
},
{
"epoch": 0.461602152763403,
"grad_norm": 1.8809865504177992,
"learning_rate": 7.931696905016008e-06,
"loss": 0.79487166,
"memory(GiB)": 63.72,
"step": 1115,
"train_speed(iter/s)": 0.020174
},
{
"epoch": 0.46367211757400123,
"grad_norm": 1.8810868009507724,
"learning_rate": 7.92102454642476e-06,
"loss": 0.78505554,
"memory(GiB)": 63.72,
"step": 1120,
"train_speed(iter/s)": 0.020176
},
{
"epoch": 0.46574208238459947,
"grad_norm": 1.7134608698183469,
"learning_rate": 7.910352187833512e-06,
"loss": 0.77789249,
"memory(GiB)": 63.72,
"step": 1125,
"train_speed(iter/s)": 0.020179
},
{
"epoch": 0.4678120471951977,
"grad_norm": 1.9145455941813492,
"learning_rate": 7.899679829242263e-06,
"loss": 0.76669245,
"memory(GiB)": 63.72,
"step": 1130,
"train_speed(iter/s)": 0.020182
},
{
"epoch": 0.4698820120057959,
"grad_norm": 1.9752917655252427,
"learning_rate": 7.889007470651014e-06,
"loss": 0.77915101,
"memory(GiB)": 63.72,
"step": 1135,
"train_speed(iter/s)": 0.020184
},
{
"epoch": 0.4719519768163941,
"grad_norm": 1.8705706085741929,
"learning_rate": 7.878335112059767e-06,
"loss": 0.78053985,
"memory(GiB)": 63.72,
"step": 1140,
"train_speed(iter/s)": 0.020187
},
{
"epoch": 0.47402194162699235,
"grad_norm": 1.8137417073497548,
"learning_rate": 7.867662753468518e-06,
"loss": 0.8304471,
"memory(GiB)": 63.72,
"step": 1145,
"train_speed(iter/s)": 0.020189
},
{
"epoch": 0.4760919064375906,
"grad_norm": 1.7537064971860614,
"learning_rate": 7.856990394877269e-06,
"loss": 0.76652546,
"memory(GiB)": 63.72,
"step": 1150,
"train_speed(iter/s)": 0.020192
},
{
"epoch": 0.47816187124818876,
"grad_norm": 1.8981437943138895,
"learning_rate": 7.84631803628602e-06,
"loss": 0.77384648,
"memory(GiB)": 63.72,
"step": 1155,
"train_speed(iter/s)": 0.020194
},
{
"epoch": 0.480231836058787,
"grad_norm": 1.968718081590253,
"learning_rate": 7.83564567769477e-06,
"loss": 0.81451969,
"memory(GiB)": 63.72,
"step": 1160,
"train_speed(iter/s)": 0.020196
},
{
"epoch": 0.48230180086938523,
"grad_norm": 1.9755371858466928,
"learning_rate": 7.824973319103523e-06,
"loss": 0.79220142,
"memory(GiB)": 63.72,
"step": 1165,
"train_speed(iter/s)": 0.020198
},
{
"epoch": 0.4843717656799834,
"grad_norm": 1.8485795416766981,
"learning_rate": 7.814300960512274e-06,
"loss": 0.80019064,
"memory(GiB)": 63.72,
"step": 1170,
"train_speed(iter/s)": 0.0202
},
{
"epoch": 0.48644173049058165,
"grad_norm": 2.388358446370589,
"learning_rate": 7.803628601921026e-06,
"loss": 0.8037425,
"memory(GiB)": 63.72,
"step": 1175,
"train_speed(iter/s)": 0.020202
},
{
"epoch": 0.4885116953011799,
"grad_norm": 1.7963803355457697,
"learning_rate": 7.792956243329777e-06,
"loss": 0.76506805,
"memory(GiB)": 63.72,
"step": 1180,
"train_speed(iter/s)": 0.020205
},
{
"epoch": 0.4905816601117781,
"grad_norm": 1.642582867995439,
"learning_rate": 7.782283884738528e-06,
"loss": 0.76571236,
"memory(GiB)": 63.72,
"step": 1185,
"train_speed(iter/s)": 0.020206
},
{
"epoch": 0.4926516249223763,
"grad_norm": 1.8722199369588735,
"learning_rate": 7.771611526147279e-06,
"loss": 0.81547689,
"memory(GiB)": 63.72,
"step": 1190,
"train_speed(iter/s)": 0.020209
},
{
"epoch": 0.49472158973297453,
"grad_norm": 1.6444393246271363,
"learning_rate": 7.760939167556031e-06,
"loss": 0.77196584,
"memory(GiB)": 63.72,
"step": 1195,
"train_speed(iter/s)": 0.02021
},
{
"epoch": 0.49679155454357277,
"grad_norm": 1.7322851516861686,
"learning_rate": 7.75026680896478e-06,
"loss": 0.78245749,
"memory(GiB)": 63.72,
"step": 1200,
"train_speed(iter/s)": 0.020211
},
{
"epoch": 0.49679155454357277,
"eval_loss": 0.8388283252716064,
"eval_runtime": 333.9836,
"eval_samples_per_second": 18.702,
"eval_steps_per_second": 1.171,
"step": 1200
},
{
"epoch": 0.498861519354171,
"grad_norm": 1.779853897918009,
"learning_rate": 7.739594450373533e-06,
"loss": 0.76194401,
"memory(GiB)": 63.72,
"step": 1205,
"train_speed(iter/s)": 0.020085
},
{
"epoch": 0.5009314841647692,
"grad_norm": 1.6935101575699751,
"learning_rate": 7.728922091782284e-06,
"loss": 0.76538324,
"memory(GiB)": 63.72,
"step": 1210,
"train_speed(iter/s)": 0.020088
},
{
"epoch": 0.5030014489753675,
"grad_norm": 2.1568052824101196,
"learning_rate": 7.718249733191036e-06,
"loss": 0.80119467,
"memory(GiB)": 63.72,
"step": 1215,
"train_speed(iter/s)": 0.02009
},
{
"epoch": 0.5050714137859657,
"grad_norm": 2.0414007101619815,
"learning_rate": 7.707577374599787e-06,
"loss": 0.78779106,
"memory(GiB)": 63.72,
"step": 1220,
"train_speed(iter/s)": 0.020091
},
{
"epoch": 0.5071413785965638,
"grad_norm": 1.6191663238961727,
"learning_rate": 7.69690501600854e-06,
"loss": 0.79252872,
"memory(GiB)": 63.72,
"step": 1225,
"train_speed(iter/s)": 0.020092
},
{
"epoch": 0.5092113434071621,
"grad_norm": 1.8617354993121655,
"learning_rate": 7.68623265741729e-06,
"loss": 0.76940928,
"memory(GiB)": 63.72,
"step": 1230,
"train_speed(iter/s)": 0.020095
},
{
"epoch": 0.5112813082177603,
"grad_norm": 2.1148307227706,
"learning_rate": 7.675560298826041e-06,
"loss": 0.81177235,
"memory(GiB)": 63.72,
"step": 1235,
"train_speed(iter/s)": 0.020096
},
{
"epoch": 0.5133512730283585,
"grad_norm": 1.858296305288767,
"learning_rate": 7.664887940234792e-06,
"loss": 0.78712654,
"memory(GiB)": 63.72,
"step": 1240,
"train_speed(iter/s)": 0.020097
},
{
"epoch": 0.5154212378389568,
"grad_norm": 2.040424767723149,
"learning_rate": 7.654215581643543e-06,
"loss": 0.7963089,
"memory(GiB)": 63.72,
"step": 1245,
"train_speed(iter/s)": 0.020099
},
{
"epoch": 0.517491202649555,
"grad_norm": 1.7313703601186623,
"learning_rate": 7.643543223052296e-06,
"loss": 0.76896205,
"memory(GiB)": 63.72,
"step": 1250,
"train_speed(iter/s)": 0.020102
},
{
"epoch": 0.5195611674601531,
"grad_norm": 1.6916331849372186,
"learning_rate": 7.632870864461046e-06,
"loss": 0.78258944,
"memory(GiB)": 63.72,
"step": 1255,
"train_speed(iter/s)": 0.020104
},
{
"epoch": 0.5216311322707514,
"grad_norm": 2.1058096966812303,
"learning_rate": 7.622198505869797e-06,
"loss": 0.80921211,
"memory(GiB)": 63.72,
"step": 1260,
"train_speed(iter/s)": 0.020105
},
{
"epoch": 0.5237010970813496,
"grad_norm": 1.7220759067410432,
"learning_rate": 7.611526147278549e-06,
"loss": 0.77246647,
"memory(GiB)": 63.72,
"step": 1265,
"train_speed(iter/s)": 0.020106
},
{
"epoch": 0.5257710618919479,
"grad_norm": 2.084224319084108,
"learning_rate": 7.6008537886873e-06,
"loss": 0.7677907,
"memory(GiB)": 63.72,
"step": 1270,
"train_speed(iter/s)": 0.020107
},
{
"epoch": 0.5278410267025461,
"grad_norm": 1.7928505615246706,
"learning_rate": 7.590181430096052e-06,
"loss": 0.78496704,
"memory(GiB)": 63.72,
"step": 1275,
"train_speed(iter/s)": 0.020108
},
{
"epoch": 0.5299109915131442,
"grad_norm": 1.8397320603347174,
"learning_rate": 7.579509071504803e-06,
"loss": 0.77303753,
"memory(GiB)": 63.72,
"step": 1280,
"train_speed(iter/s)": 0.02011
},
{
"epoch": 0.5319809563237425,
"grad_norm": 2.1479969295234187,
"learning_rate": 7.568836712913554e-06,
"loss": 0.75871119,
"memory(GiB)": 63.72,
"step": 1285,
"train_speed(iter/s)": 0.020112
},
{
"epoch": 0.5340509211343407,
"grad_norm": 1.94767502078934,
"learning_rate": 7.558164354322306e-06,
"loss": 0.75106993,
"memory(GiB)": 63.72,
"step": 1290,
"train_speed(iter/s)": 0.020114
},
{
"epoch": 0.5361208859449389,
"grad_norm": 1.5236425325852578,
"learning_rate": 7.547491995731058e-06,
"loss": 0.79110327,
"memory(GiB)": 63.72,
"step": 1295,
"train_speed(iter/s)": 0.020116
},
{
"epoch": 0.5381908507555372,
"grad_norm": 1.8541149671409907,
"learning_rate": 7.536819637139808e-06,
"loss": 0.77403798,
"memory(GiB)": 63.72,
"step": 1300,
"train_speed(iter/s)": 0.020116
},
{
"epoch": 0.5402608155661354,
"grad_norm": 1.8743174944996448,
"learning_rate": 7.52614727854856e-06,
"loss": 0.77032347,
"memory(GiB)": 71.94,
"step": 1305,
"train_speed(iter/s)": 0.020118
},
{
"epoch": 0.5423307803767335,
"grad_norm": 2.579806479546849,
"learning_rate": 7.51547491995731e-06,
"loss": 0.76461482,
"memory(GiB)": 71.94,
"step": 1310,
"train_speed(iter/s)": 0.020119
},
{
"epoch": 0.5444007451873318,
"grad_norm": 2.0039452129208035,
"learning_rate": 7.504802561366062e-06,
"loss": 0.77457762,
"memory(GiB)": 71.94,
"step": 1315,
"train_speed(iter/s)": 0.020121
},
{
"epoch": 0.54647070999793,
"grad_norm": 2.060283685569936,
"learning_rate": 7.494130202774814e-06,
"loss": 0.77914829,
"memory(GiB)": 71.94,
"step": 1320,
"train_speed(iter/s)": 0.020123
},
{
"epoch": 0.5485406748085283,
"grad_norm": 2.163132135636586,
"learning_rate": 7.483457844183565e-06,
"loss": 0.77322574,
"memory(GiB)": 71.94,
"step": 1325,
"train_speed(iter/s)": 0.020125
},
{
"epoch": 0.5506106396191265,
"grad_norm": 1.842195467860799,
"learning_rate": 7.472785485592316e-06,
"loss": 0.77454052,
"memory(GiB)": 71.94,
"step": 1330,
"train_speed(iter/s)": 0.020126
},
{
"epoch": 0.5526806044297247,
"grad_norm": 1.775275008552653,
"learning_rate": 7.462113127001068e-06,
"loss": 0.77025108,
"memory(GiB)": 71.94,
"step": 1335,
"train_speed(iter/s)": 0.020128
},
{
"epoch": 0.554750569240323,
"grad_norm": 2.165651142341684,
"learning_rate": 7.451440768409819e-06,
"loss": 0.78470011,
"memory(GiB)": 71.94,
"step": 1340,
"train_speed(iter/s)": 0.02013
},
{
"epoch": 0.5568205340509211,
"grad_norm": 1.6530168942960388,
"learning_rate": 7.440768409818571e-06,
"loss": 0.74261112,
"memory(GiB)": 71.94,
"step": 1345,
"train_speed(iter/s)": 0.020131
},
{
"epoch": 0.5588904988615193,
"grad_norm": 2.1178890231616694,
"learning_rate": 7.430096051227322e-06,
"loss": 0.77076225,
"memory(GiB)": 71.94,
"step": 1350,
"train_speed(iter/s)": 0.020132
},
{
"epoch": 0.5609604636721176,
"grad_norm": 1.6332209286889638,
"learning_rate": 7.419423692636073e-06,
"loss": 0.76129122,
"memory(GiB)": 71.94,
"step": 1355,
"train_speed(iter/s)": 0.020134
},
{
"epoch": 0.5630304284827158,
"grad_norm": 1.9276105656674607,
"learning_rate": 7.408751334044825e-06,
"loss": 0.77616062,
"memory(GiB)": 71.94,
"step": 1360,
"train_speed(iter/s)": 0.020136
},
{
"epoch": 0.565100393293314,
"grad_norm": 1.838664332126464,
"learning_rate": 7.398078975453575e-06,
"loss": 0.77545385,
"memory(GiB)": 71.94,
"step": 1365,
"train_speed(iter/s)": 0.020137
},
{
"epoch": 0.5671703581039123,
"grad_norm": 2.090052030958157,
"learning_rate": 7.387406616862327e-06,
"loss": 0.7824297,
"memory(GiB)": 71.94,
"step": 1370,
"train_speed(iter/s)": 0.020139
},
{
"epoch": 0.5692403229145104,
"grad_norm": 1.7799554116738177,
"learning_rate": 7.376734258271079e-06,
"loss": 0.77833185,
"memory(GiB)": 71.94,
"step": 1375,
"train_speed(iter/s)": 0.020142
},
{
"epoch": 0.5713102877251087,
"grad_norm": 2.266691996975209,
"learning_rate": 7.366061899679829e-06,
"loss": 0.77535782,
"memory(GiB)": 71.94,
"step": 1380,
"train_speed(iter/s)": 0.020144
},
{
"epoch": 0.5733802525357069,
"grad_norm": 1.8220471587007605,
"learning_rate": 7.355389541088581e-06,
"loss": 0.76158247,
"memory(GiB)": 71.94,
"step": 1385,
"train_speed(iter/s)": 0.020146
},
{
"epoch": 0.5754502173463051,
"grad_norm": 1.7869060368578336,
"learning_rate": 7.344717182497333e-06,
"loss": 0.79457912,
"memory(GiB)": 71.94,
"step": 1390,
"train_speed(iter/s)": 0.020147
},
{
"epoch": 0.5775201821569034,
"grad_norm": 2.730877403121895,
"learning_rate": 7.334044823906084e-06,
"loss": 0.75181475,
"memory(GiB)": 71.94,
"step": 1395,
"train_speed(iter/s)": 0.020149
},
{
"epoch": 0.5795901469675016,
"grad_norm": 2.091944883020518,
"learning_rate": 7.323372465314835e-06,
"loss": 0.75992446,
"memory(GiB)": 71.94,
"step": 1400,
"train_speed(iter/s)": 0.02015
},
{
"epoch": 0.5816601117780997,
"grad_norm": 1.5904822426334966,
"learning_rate": 7.312700106723587e-06,
"loss": 0.77254944,
"memory(GiB)": 71.94,
"step": 1405,
"train_speed(iter/s)": 0.020152
},
{
"epoch": 0.583730076588698,
"grad_norm": 1.673083919686743,
"learning_rate": 7.302027748132338e-06,
"loss": 0.74836388,
"memory(GiB)": 71.94,
"step": 1410,
"train_speed(iter/s)": 0.020152
},
{
"epoch": 0.5858000413992962,
"grad_norm": 2.05811523971159,
"learning_rate": 7.29135538954109e-06,
"loss": 0.74358282,
"memory(GiB)": 71.94,
"step": 1415,
"train_speed(iter/s)": 0.020154
},
{
"epoch": 0.5878700062098944,
"grad_norm": 1.983632865952002,
"learning_rate": 7.28068303094984e-06,
"loss": 0.78234367,
"memory(GiB)": 71.94,
"step": 1420,
"train_speed(iter/s)": 0.020156
},
{
"epoch": 0.5899399710204927,
"grad_norm": 1.6612296882759847,
"learning_rate": 7.270010672358592e-06,
"loss": 0.76740494,
"memory(GiB)": 71.94,
"step": 1425,
"train_speed(iter/s)": 0.020157
},
{
"epoch": 0.5920099358310908,
"grad_norm": 1.8232818202515155,
"learning_rate": 7.259338313767344e-06,
"loss": 0.76410437,
"memory(GiB)": 71.94,
"step": 1430,
"train_speed(iter/s)": 0.020159
},
{
"epoch": 0.5940799006416891,
"grad_norm": 1.6871789120586522,
"learning_rate": 7.248665955176094e-06,
"loss": 0.76673613,
"memory(GiB)": 71.94,
"step": 1435,
"train_speed(iter/s)": 0.02016
},
{
"epoch": 0.5961498654522873,
"grad_norm": 1.9181669169557467,
"learning_rate": 7.237993596584846e-06,
"loss": 0.73530726,
"memory(GiB)": 71.94,
"step": 1440,
"train_speed(iter/s)": 0.020161
},
{
"epoch": 0.5982198302628855,
"grad_norm": 2.0425311715534513,
"learning_rate": 7.227321237993598e-06,
"loss": 0.78409719,
"memory(GiB)": 71.94,
"step": 1445,
"train_speed(iter/s)": 0.020163
},
{
"epoch": 0.6002897950734838,
"grad_norm": 1.725457162133973,
"learning_rate": 7.216648879402348e-06,
"loss": 0.74391842,
"memory(GiB)": 71.94,
"step": 1450,
"train_speed(iter/s)": 0.020163
},
{
"epoch": 0.602359759884082,
"grad_norm": 2.2362927243629613,
"learning_rate": 7.2059765208111e-06,
"loss": 0.77035971,
"memory(GiB)": 71.94,
"step": 1455,
"train_speed(iter/s)": 0.020166
},
{
"epoch": 0.6044297246946801,
"grad_norm": 2.379202645179455,
"learning_rate": 7.195304162219852e-06,
"loss": 0.74266062,
"memory(GiB)": 71.94,
"step": 1460,
"train_speed(iter/s)": 0.020168
},
{
"epoch": 0.6064996895052784,
"grad_norm": 1.6006607749389805,
"learning_rate": 7.184631803628602e-06,
"loss": 0.76957574,
"memory(GiB)": 71.94,
"step": 1465,
"train_speed(iter/s)": 0.020169
},
{
"epoch": 0.6085696543158766,
"grad_norm": 1.7633012594109296,
"learning_rate": 7.173959445037354e-06,
"loss": 0.77078071,
"memory(GiB)": 71.94,
"step": 1470,
"train_speed(iter/s)": 0.02017
},
{
"epoch": 0.6106396191264748,
"grad_norm": 1.6009632285824897,
"learning_rate": 7.163287086446106e-06,
"loss": 0.7669549,
"memory(GiB)": 71.94,
"step": 1475,
"train_speed(iter/s)": 0.020171
},
{
"epoch": 0.6127095839370731,
"grad_norm": 1.932344117154099,
"learning_rate": 7.152614727854857e-06,
"loss": 0.76528344,
"memory(GiB)": 71.94,
"step": 1480,
"train_speed(iter/s)": 0.020172
},
{
"epoch": 0.6147795487476713,
"grad_norm": 2.02896587820159,
"learning_rate": 7.141942369263608e-06,
"loss": 0.75168095,
"memory(GiB)": 71.94,
"step": 1485,
"train_speed(iter/s)": 0.020174
},
{
"epoch": 0.6168495135582696,
"grad_norm": 1.9974467066335662,
"learning_rate": 7.131270010672359e-06,
"loss": 0.7488194,
"memory(GiB)": 71.94,
"step": 1490,
"train_speed(iter/s)": 0.020176
},
{
"epoch": 0.6189194783688677,
"grad_norm": 1.7902382164373858,
"learning_rate": 7.120597652081111e-06,
"loss": 0.75267982,
"memory(GiB)": 71.94,
"step": 1495,
"train_speed(iter/s)": 0.020177
},
{
"epoch": 0.6209894431794659,
"grad_norm": 2.5929739472863838,
"learning_rate": 7.1099252934898625e-06,
"loss": 0.73765378,
"memory(GiB)": 71.94,
"step": 1500,
"train_speed(iter/s)": 0.020178
},
{
"epoch": 0.6209894431794659,
"eval_loss": 0.8280953168869019,
"eval_runtime": 333.1777,
"eval_samples_per_second": 18.747,
"eval_steps_per_second": 1.174,
"step": 1500
},
{
"epoch": 0.6230594079900642,
"grad_norm": 1.9847247838289337,
"learning_rate": 7.099252934898613e-06,
"loss": 0.75908709,
"memory(GiB)": 71.94,
"step": 1505,
"train_speed(iter/s)": 0.020077
},
{
"epoch": 0.6251293728006624,
"grad_norm": 1.801176606187135,
"learning_rate": 7.088580576307365e-06,
"loss": 0.78817225,
"memory(GiB)": 71.94,
"step": 1510,
"train_speed(iter/s)": 0.020079
},
{
"epoch": 0.6271993376112606,
"grad_norm": 1.8685868923530962,
"learning_rate": 7.0779082177161165e-06,
"loss": 0.74608173,
"memory(GiB)": 71.94,
"step": 1515,
"train_speed(iter/s)": 0.02008
},
{
"epoch": 0.6292693024218589,
"grad_norm": 1.9112331285520685,
"learning_rate": 7.067235859124867e-06,
"loss": 0.78047667,
"memory(GiB)": 71.94,
"step": 1520,
"train_speed(iter/s)": 0.020081
},
{
"epoch": 0.631339267232457,
"grad_norm": 1.69066034933461,
"learning_rate": 7.0565635005336185e-06,
"loss": 0.75149813,
"memory(GiB)": 71.94,
"step": 1525,
"train_speed(iter/s)": 0.02008
},
{
"epoch": 0.6334092320430552,
"grad_norm": 1.5394308738884603,
"learning_rate": 7.0458911419423704e-06,
"loss": 0.76838903,
"memory(GiB)": 71.94,
"step": 1530,
"train_speed(iter/s)": 0.020081
},
{
"epoch": 0.6354791968536535,
"grad_norm": 1.7690454856119193,
"learning_rate": 7.035218783351121e-06,
"loss": 0.7524828,
"memory(GiB)": 71.94,
"step": 1535,
"train_speed(iter/s)": 0.020083
},
{
"epoch": 0.6375491616642517,
"grad_norm": 2.0548651954814154,
"learning_rate": 7.0245464247598725e-06,
"loss": 0.76242485,
"memory(GiB)": 71.94,
"step": 1540,
"train_speed(iter/s)": 0.020083
},
{
"epoch": 0.63961912647485,
"grad_norm": 1.5706142249866388,
"learning_rate": 7.0138740661686235e-06,
"loss": 0.78198986,
"memory(GiB)": 71.94,
"step": 1545,
"train_speed(iter/s)": 0.020084
},
{
"epoch": 0.6416890912854482,
"grad_norm": 2.9215645341915275,
"learning_rate": 7.0032017075773754e-06,
"loss": 0.74943571,
"memory(GiB)": 71.94,
"step": 1550,
"train_speed(iter/s)": 0.020086
},
{
"epoch": 0.6437590560960463,
"grad_norm": 1.983762828992232,
"learning_rate": 6.9925293489861265e-06,
"loss": 0.75862083,
"memory(GiB)": 71.94,
"step": 1555,
"train_speed(iter/s)": 0.020087
},
{
"epoch": 0.6458290209066446,
"grad_norm": 1.8256568832087245,
"learning_rate": 6.9818569903948775e-06,
"loss": 0.75860863,
"memory(GiB)": 71.94,
"step": 1560,
"train_speed(iter/s)": 0.020089
},
{
"epoch": 0.6478989857172428,
"grad_norm": 1.7082999758601491,
"learning_rate": 6.971184631803629e-06,
"loss": 0.78126869,
"memory(GiB)": 71.94,
"step": 1565,
"train_speed(iter/s)": 0.02009
},
{
"epoch": 0.649968950527841,
"grad_norm": 2.388449730753909,
"learning_rate": 6.960512273212381e-06,
"loss": 0.76754818,
"memory(GiB)": 71.94,
"step": 1570,
"train_speed(iter/s)": 0.020091
},
{
"epoch": 0.6520389153384393,
"grad_norm": 2.570709108294184,
"learning_rate": 6.9498399146211315e-06,
"loss": 0.74245424,
"memory(GiB)": 71.94,
"step": 1575,
"train_speed(iter/s)": 0.020093
},
{
"epoch": 0.6541088801490375,
"grad_norm": 1.942646301485773,
"learning_rate": 6.939167556029883e-06,
"loss": 0.77264175,
"memory(GiB)": 71.94,
"step": 1580,
"train_speed(iter/s)": 0.020094
},
{
"epoch": 0.6561788449596356,
"grad_norm": 1.7902561992868253,
"learning_rate": 6.928495197438635e-06,
"loss": 0.76845627,
"memory(GiB)": 71.94,
"step": 1585,
"train_speed(iter/s)": 0.020096
},
{
"epoch": 0.6582488097702339,
"grad_norm": 1.8253085296468832,
"learning_rate": 6.9178228388473854e-06,
"loss": 0.75771255,
"memory(GiB)": 71.94,
"step": 1590,
"train_speed(iter/s)": 0.020097
},
{
"epoch": 0.6603187745808321,
"grad_norm": 1.8440321320283706,
"learning_rate": 6.907150480256137e-06,
"loss": 0.74345121,
"memory(GiB)": 71.94,
"step": 1595,
"train_speed(iter/s)": 0.020096
},
{
"epoch": 0.6623887393914304,
"grad_norm": 1.8894245153228149,
"learning_rate": 6.896478121664889e-06,
"loss": 0.76188393,
"memory(GiB)": 71.94,
"step": 1600,
"train_speed(iter/s)": 0.020098
},
{
"epoch": 0.6644587042020286,
"grad_norm": 2.055312059883184,
"learning_rate": 6.885805763073639e-06,
"loss": 0.75247483,
"memory(GiB)": 71.94,
"step": 1605,
"train_speed(iter/s)": 0.020099
},
{
"epoch": 0.6665286690126268,
"grad_norm": 1.8131130404445874,
"learning_rate": 6.875133404482391e-06,
"loss": 0.77208357,
"memory(GiB)": 71.94,
"step": 1610,
"train_speed(iter/s)": 0.0201
},
{
"epoch": 0.668598633823225,
"grad_norm": 1.721876122278255,
"learning_rate": 6.864461045891142e-06,
"loss": 0.7271523,
"memory(GiB)": 71.94,
"step": 1615,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.6706685986338232,
"grad_norm": 1.880195637198508,
"learning_rate": 6.853788687299893e-06,
"loss": 0.74756432,
"memory(GiB)": 71.94,
"step": 1620,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.6727385634444214,
"grad_norm": 1.5895741550527986,
"learning_rate": 6.843116328708645e-06,
"loss": 0.77960148,
"memory(GiB)": 71.94,
"step": 1625,
"train_speed(iter/s)": 0.020102
},
{
"epoch": 0.6748085282550197,
"grad_norm": 1.762407790878705,
"learning_rate": 6.832443970117396e-06,
"loss": 0.76564293,
"memory(GiB)": 71.94,
"step": 1630,
"train_speed(iter/s)": 0.020104
},
{
"epoch": 0.6768784930656179,
"grad_norm": 1.85152265865232,
"learning_rate": 6.821771611526148e-06,
"loss": 0.75755472,
"memory(GiB)": 71.94,
"step": 1635,
"train_speed(iter/s)": 0.020104
},
{
"epoch": 0.678948457876216,
"grad_norm": 2.4048195292609464,
"learning_rate": 6.811099252934899e-06,
"loss": 0.75722828,
"memory(GiB)": 71.94,
"step": 1640,
"train_speed(iter/s)": 0.020106
},
{
"epoch": 0.6810184226868143,
"grad_norm": 1.7461290969223273,
"learning_rate": 6.80042689434365e-06,
"loss": 0.74719772,
"memory(GiB)": 71.94,
"step": 1645,
"train_speed(iter/s)": 0.020107
},
{
"epoch": 0.6830883874974125,
"grad_norm": 1.7606017366047548,
"learning_rate": 6.789754535752402e-06,
"loss": 0.74246426,
"memory(GiB)": 71.94,
"step": 1650,
"train_speed(iter/s)": 0.020108
},
{
"epoch": 0.6851583523080108,
"grad_norm": 2.3484261163252884,
"learning_rate": 6.779082177161154e-06,
"loss": 0.7567915,
"memory(GiB)": 71.94,
"step": 1655,
"train_speed(iter/s)": 0.020109
},
{
"epoch": 0.687228317118609,
"grad_norm": 1.686698632081635,
"learning_rate": 6.768409818569904e-06,
"loss": 0.73500414,
"memory(GiB)": 71.94,
"step": 1660,
"train_speed(iter/s)": 0.020111
},
{
"epoch": 0.6892982819292072,
"grad_norm": 1.9785908023609375,
"learning_rate": 6.757737459978656e-06,
"loss": 0.7035881,
"memory(GiB)": 71.94,
"step": 1665,
"train_speed(iter/s)": 0.020112
},
{
"epoch": 0.6913682467398055,
"grad_norm": 1.8288827641332985,
"learning_rate": 6.747065101387406e-06,
"loss": 0.74135156,
"memory(GiB)": 71.94,
"step": 1670,
"train_speed(iter/s)": 0.020113
},
{
"epoch": 0.6934382115504036,
"grad_norm": 2.106219884662748,
"learning_rate": 6.736392742796158e-06,
"loss": 0.77649341,
"memory(GiB)": 71.94,
"step": 1675,
"train_speed(iter/s)": 0.020115
},
{
"epoch": 0.6955081763610018,
"grad_norm": 1.857981089347382,
"learning_rate": 6.72572038420491e-06,
"loss": 0.73271251,
"memory(GiB)": 71.94,
"step": 1680,
"train_speed(iter/s)": 0.020116
},
{
"epoch": 0.6975781411716001,
"grad_norm": 1.8252469751324223,
"learning_rate": 6.715048025613661e-06,
"loss": 0.76072979,
"memory(GiB)": 71.94,
"step": 1685,
"train_speed(iter/s)": 0.020117
},
{
"epoch": 0.6996481059821983,
"grad_norm": 1.9875787155351985,
"learning_rate": 6.704375667022412e-06,
"loss": 0.73438239,
"memory(GiB)": 71.94,
"step": 1690,
"train_speed(iter/s)": 0.020118
},
{
"epoch": 0.7017180707927966,
"grad_norm": 1.685302389303672,
"learning_rate": 6.693703308431164e-06,
"loss": 0.76429882,
"memory(GiB)": 71.94,
"step": 1695,
"train_speed(iter/s)": 0.020119
},
{
"epoch": 0.7037880356033948,
"grad_norm": 2.30374573526697,
"learning_rate": 6.683030949839915e-06,
"loss": 0.7437336,
"memory(GiB)": 71.94,
"step": 1700,
"train_speed(iter/s)": 0.020119
},
{
"epoch": 0.7058580004139929,
"grad_norm": 2.4577356463267104,
"learning_rate": 6.672358591248667e-06,
"loss": 0.79406719,
"memory(GiB)": 71.94,
"step": 1705,
"train_speed(iter/s)": 0.02012
},
{
"epoch": 0.7079279652245912,
"grad_norm": 1.7489028068953179,
"learning_rate": 6.661686232657418e-06,
"loss": 0.75482893,
"memory(GiB)": 71.94,
"step": 1710,
"train_speed(iter/s)": 0.020121
},
{
"epoch": 0.7099979300351894,
"grad_norm": 1.929474801980816,
"learning_rate": 6.651013874066169e-06,
"loss": 0.74136767,
"memory(GiB)": 71.94,
"step": 1715,
"train_speed(iter/s)": 0.020122
},
{
"epoch": 0.7120678948457876,
"grad_norm": 1.6790168198096502,
"learning_rate": 6.640341515474921e-06,
"loss": 0.72392588,
"memory(GiB)": 71.94,
"step": 1720,
"train_speed(iter/s)": 0.020122
},
{
"epoch": 0.7141378596563859,
"grad_norm": 2.2963610149429488,
"learning_rate": 6.629669156883671e-06,
"loss": 0.7462635,
"memory(GiB)": 71.94,
"step": 1725,
"train_speed(iter/s)": 0.020123
},
{
"epoch": 0.716207824466984,
"grad_norm": 4.289784718847475,
"learning_rate": 6.618996798292423e-06,
"loss": 0.73541126,
"memory(GiB)": 71.94,
"step": 1730,
"train_speed(iter/s)": 0.020124
},
{
"epoch": 0.7182777892775822,
"grad_norm": 2.1972884462976263,
"learning_rate": 6.608324439701175e-06,
"loss": 0.7353497,
"memory(GiB)": 71.94,
"step": 1735,
"train_speed(iter/s)": 0.020124
},
{
"epoch": 0.7203477540881805,
"grad_norm": 2.1738189409828377,
"learning_rate": 6.597652081109925e-06,
"loss": 0.71738148,
"memory(GiB)": 71.94,
"step": 1740,
"train_speed(iter/s)": 0.020126
},
{
"epoch": 0.7224177188987787,
"grad_norm": 1.6342074890059992,
"learning_rate": 6.586979722518677e-06,
"loss": 0.75047336,
"memory(GiB)": 71.94,
"step": 1745,
"train_speed(iter/s)": 0.020126
},
{
"epoch": 0.724487683709377,
"grad_norm": 1.7007570391919413,
"learning_rate": 6.576307363927429e-06,
"loss": 0.73253298,
"memory(GiB)": 71.94,
"step": 1750,
"train_speed(iter/s)": 0.020127
},
{
"epoch": 0.7265576485199752,
"grad_norm": 1.5323053950217638,
"learning_rate": 6.56563500533618e-06,
"loss": 0.74062099,
"memory(GiB)": 71.94,
"step": 1755,
"train_speed(iter/s)": 0.020128
},
{
"epoch": 0.7286276133305734,
"grad_norm": 1.9624071404199714,
"learning_rate": 6.554962646744931e-06,
"loss": 0.76860294,
"memory(GiB)": 71.94,
"step": 1760,
"train_speed(iter/s)": 0.02013
},
{
"epoch": 0.7306975781411716,
"grad_norm": 1.8145041855689747,
"learning_rate": 6.544290288153683e-06,
"loss": 0.72403975,
"memory(GiB)": 71.94,
"step": 1765,
"train_speed(iter/s)": 0.020131
},
{
"epoch": 0.7327675429517698,
"grad_norm": 1.7793196071264126,
"learning_rate": 6.533617929562434e-06,
"loss": 0.76407566,
"memory(GiB)": 71.94,
"step": 1770,
"train_speed(iter/s)": 0.020132
},
{
"epoch": 0.734837507762368,
"grad_norm": 1.8806974947113853,
"learning_rate": 6.522945570971186e-06,
"loss": 0.73674612,
"memory(GiB)": 71.94,
"step": 1775,
"train_speed(iter/s)": 0.020133
},
{
"epoch": 0.7369074725729663,
"grad_norm": 1.6203999356727887,
"learning_rate": 6.512273212379937e-06,
"loss": 0.73720975,
"memory(GiB)": 71.94,
"step": 1780,
"train_speed(iter/s)": 0.020135
},
{
"epoch": 0.7389774373835645,
"grad_norm": 1.8256501665131275,
"learning_rate": 6.501600853788688e-06,
"loss": 0.74560528,
"memory(GiB)": 71.94,
"step": 1785,
"train_speed(iter/s)": 0.020136
},
{
"epoch": 0.7410474021941627,
"grad_norm": 2.3313828860511294,
"learning_rate": 6.49092849519744e-06,
"loss": 0.73726311,
"memory(GiB)": 71.94,
"step": 1790,
"train_speed(iter/s)": 0.020137
},
{
"epoch": 0.7431173670047609,
"grad_norm": 2.01967250245603,
"learning_rate": 6.48025613660619e-06,
"loss": 0.72599983,
"memory(GiB)": 71.94,
"step": 1795,
"train_speed(iter/s)": 0.020138
},
{
"epoch": 0.7451873318153591,
"grad_norm": 1.5873931113082191,
"learning_rate": 6.469583778014942e-06,
"loss": 0.72361288,
"memory(GiB)": 71.94,
"step": 1800,
"train_speed(iter/s)": 0.020139
},
{
"epoch": 0.7451873318153591,
"eval_loss": 0.821691632270813,
"eval_runtime": 333.5584,
"eval_samples_per_second": 18.725,
"eval_steps_per_second": 1.172,
"step": 1800
},
{
"epoch": 0.7472572966259574,
"grad_norm": 1.6284031081935817,
"learning_rate": 6.458911419423694e-06,
"loss": 0.73620996,
"memory(GiB)": 71.94,
"step": 1805,
"train_speed(iter/s)": 0.020055
},
{
"epoch": 0.7493272614365556,
"grad_norm": 1.8143587680892548,
"learning_rate": 6.448239060832444e-06,
"loss": 0.73436375,
"memory(GiB)": 71.94,
"step": 1810,
"train_speed(iter/s)": 0.020056
},
{
"epoch": 0.7513972262471538,
"grad_norm": 1.9502300953167138,
"learning_rate": 6.437566702241196e-06,
"loss": 0.74698448,
"memory(GiB)": 71.94,
"step": 1815,
"train_speed(iter/s)": 0.020057
},
{
"epoch": 0.7534671910577521,
"grad_norm": 1.72820315761644,
"learning_rate": 6.426894343649948e-06,
"loss": 0.75661831,
"memory(GiB)": 71.94,
"step": 1820,
"train_speed(iter/s)": 0.020059
},
{
"epoch": 0.7555371558683502,
"grad_norm": 1.8039931329917855,
"learning_rate": 6.416221985058698e-06,
"loss": 0.74954405,
"memory(GiB)": 71.94,
"step": 1825,
"train_speed(iter/s)": 0.020059
},
{
"epoch": 0.7576071206789484,
"grad_norm": 2.0711566925028433,
"learning_rate": 6.40554962646745e-06,
"loss": 0.73745756,
"memory(GiB)": 71.94,
"step": 1830,
"train_speed(iter/s)": 0.020061
},
{
"epoch": 0.7596770854895467,
"grad_norm": 1.826394447351557,
"learning_rate": 6.3948772678762016e-06,
"loss": 0.73249888,
"memory(GiB)": 71.94,
"step": 1835,
"train_speed(iter/s)": 0.020062
},
{
"epoch": 0.7617470503001449,
"grad_norm": 2.4929262063136175,
"learning_rate": 6.384204909284953e-06,
"loss": 0.72507439,
"memory(GiB)": 71.94,
"step": 1840,
"train_speed(iter/s)": 0.020062
},
{
"epoch": 0.7638170151107431,
"grad_norm": 1.7606115982834467,
"learning_rate": 6.373532550693704e-06,
"loss": 0.72618227,
"memory(GiB)": 71.94,
"step": 1845,
"train_speed(iter/s)": 0.020064
},
{
"epoch": 0.7658869799213414,
"grad_norm": 1.7780633999979434,
"learning_rate": 6.362860192102455e-06,
"loss": 0.73577065,
"memory(GiB)": 71.94,
"step": 1850,
"train_speed(iter/s)": 0.020064
},
{
"epoch": 0.7679569447319395,
"grad_norm": 2.2901271183050294,
"learning_rate": 6.3521878335112066e-06,
"loss": 0.75972033,
"memory(GiB)": 71.94,
"step": 1855,
"train_speed(iter/s)": 0.020065
},
{
"epoch": 0.7700269095425378,
"grad_norm": 1.974709118996213,
"learning_rate": 6.3415154749199585e-06,
"loss": 0.7513834,
"memory(GiB)": 71.94,
"step": 1860,
"train_speed(iter/s)": 0.020065
},
{
"epoch": 0.772096874353136,
"grad_norm": 1.7795619053561953,
"learning_rate": 6.330843116328709e-06,
"loss": 0.70549402,
"memory(GiB)": 71.94,
"step": 1865,
"train_speed(iter/s)": 0.020066
},
{
"epoch": 0.7741668391637342,
"grad_norm": 1.7925210555170064,
"learning_rate": 6.3201707577374605e-06,
"loss": 0.75538568,
"memory(GiB)": 71.94,
"step": 1870,
"train_speed(iter/s)": 0.020067
},
{
"epoch": 0.7762368039743325,
"grad_norm": 1.7592519440523378,
"learning_rate": 6.309498399146212e-06,
"loss": 0.74328322,
"memory(GiB)": 71.94,
"step": 1875,
"train_speed(iter/s)": 0.020068
},
{
"epoch": 0.7783067687849307,
"grad_norm": 1.7630568722667896,
"learning_rate": 6.298826040554963e-06,
"loss": 0.73995676,
"memory(GiB)": 71.94,
"step": 1880,
"train_speed(iter/s)": 0.020069
},
{
"epoch": 0.7803767335955288,
"grad_norm": 1.732047211183728,
"learning_rate": 6.2881536819637145e-06,
"loss": 0.70023623,
"memory(GiB)": 71.94,
"step": 1885,
"train_speed(iter/s)": 0.02007
},
{
"epoch": 0.7824466984061271,
"grad_norm": 1.7004814074017778,
"learning_rate": 6.277481323372466e-06,
"loss": 0.73256617,
"memory(GiB)": 71.94,
"step": 1890,
"train_speed(iter/s)": 0.020071
},
{
"epoch": 0.7845166632167253,
"grad_norm": 1.7170761577962488,
"learning_rate": 6.2668089647812166e-06,
"loss": 0.71450982,
"memory(GiB)": 71.94,
"step": 1895,
"train_speed(iter/s)": 0.020071
},
{
"epoch": 0.7865866280273235,
"grad_norm": 2.0086204171681565,
"learning_rate": 6.2561366061899685e-06,
"loss": 0.7096673,
"memory(GiB)": 71.94,
"step": 1900,
"train_speed(iter/s)": 0.020072
},
{
"epoch": 0.7886565928379218,
"grad_norm": 1.6109075949007228,
"learning_rate": 6.2454642475987195e-06,
"loss": 0.7170536,
"memory(GiB)": 71.94,
"step": 1905,
"train_speed(iter/s)": 0.020073
},
{
"epoch": 0.79072655764852,
"grad_norm": 1.6468982825229455,
"learning_rate": 6.234791889007471e-06,
"loss": 0.7817646,
"memory(GiB)": 71.94,
"step": 1910,
"train_speed(iter/s)": 0.020073
},
{
"epoch": 0.7927965224591182,
"grad_norm": 1.8405361482523723,
"learning_rate": 6.224119530416222e-06,
"loss": 0.71389322,
"memory(GiB)": 71.94,
"step": 1915,
"train_speed(iter/s)": 0.020075
},
{
"epoch": 0.7948664872697164,
"grad_norm": 1.7937559729338877,
"learning_rate": 6.2134471718249735e-06,
"loss": 0.72494421,
"memory(GiB)": 71.94,
"step": 1920,
"train_speed(iter/s)": 0.020076
},
{
"epoch": 0.7969364520803146,
"grad_norm": 1.98762799360225,
"learning_rate": 6.202774813233725e-06,
"loss": 0.75637407,
"memory(GiB)": 71.94,
"step": 1925,
"train_speed(iter/s)": 0.020077
},
{
"epoch": 0.7990064168909129,
"grad_norm": 2.469167716565665,
"learning_rate": 6.192102454642477e-06,
"loss": 0.71725979,
"memory(GiB)": 71.94,
"step": 1930,
"train_speed(iter/s)": 0.020078
},
{
"epoch": 0.8010763817015111,
"grad_norm": 1.6526117746871118,
"learning_rate": 6.181430096051227e-06,
"loss": 0.73172369,
"memory(GiB)": 71.94,
"step": 1935,
"train_speed(iter/s)": 0.020079
},
{
"epoch": 0.8031463465121093,
"grad_norm": 1.8881085526929478,
"learning_rate": 6.170757737459979e-06,
"loss": 0.68869176,
"memory(GiB)": 71.94,
"step": 1940,
"train_speed(iter/s)": 0.02008
},
{
"epoch": 0.8052163113227075,
"grad_norm": 2.1341112552467107,
"learning_rate": 6.160085378868731e-06,
"loss": 0.74425526,
"memory(GiB)": 71.94,
"step": 1945,
"train_speed(iter/s)": 0.020081
},
{
"epoch": 0.8072862761333057,
"grad_norm": 2.1587279161379906,
"learning_rate": 6.149413020277481e-06,
"loss": 0.72686901,
"memory(GiB)": 71.94,
"step": 1950,
"train_speed(iter/s)": 0.020081
},
{
"epoch": 0.8093562409439039,
"grad_norm": 1.7183166805211196,
"learning_rate": 6.138740661686233e-06,
"loss": 0.70801015,
"memory(GiB)": 71.94,
"step": 1955,
"train_speed(iter/s)": 0.020082
},
{
"epoch": 0.8114262057545022,
"grad_norm": 1.6918548000232365,
"learning_rate": 6.128068303094985e-06,
"loss": 0.71407743,
"memory(GiB)": 71.94,
"step": 1960,
"train_speed(iter/s)": 0.020083
},
{
"epoch": 0.8134961705651004,
"grad_norm": 1.6783017839137153,
"learning_rate": 6.117395944503735e-06,
"loss": 0.74968853,
"memory(GiB)": 71.94,
"step": 1965,
"train_speed(iter/s)": 0.020083
},
{
"epoch": 0.8155661353756987,
"grad_norm": 1.7117864547494115,
"learning_rate": 6.106723585912487e-06,
"loss": 0.71134882,
"memory(GiB)": 71.94,
"step": 1970,
"train_speed(iter/s)": 0.020085
},
{
"epoch": 0.8176361001862968,
"grad_norm": 1.9043254368072582,
"learning_rate": 6.096051227321238e-06,
"loss": 0.74376335,
"memory(GiB)": 71.94,
"step": 1975,
"train_speed(iter/s)": 0.020085
},
{
"epoch": 0.819706064996895,
"grad_norm": 1.8416766338299921,
"learning_rate": 6.08537886872999e-06,
"loss": 0.74048576,
"memory(GiB)": 71.94,
"step": 1980,
"train_speed(iter/s)": 0.020086
},
{
"epoch": 0.8217760298074933,
"grad_norm": 1.9993092375152783,
"learning_rate": 6.074706510138741e-06,
"loss": 0.73070145,
"memory(GiB)": 71.94,
"step": 1985,
"train_speed(iter/s)": 0.020087
},
{
"epoch": 0.8238459946180915,
"grad_norm": 1.9255287807426156,
"learning_rate": 6.064034151547492e-06,
"loss": 0.71732969,
"memory(GiB)": 71.94,
"step": 1990,
"train_speed(iter/s)": 0.020088
},
{
"epoch": 0.8259159594286897,
"grad_norm": 1.699758287154301,
"learning_rate": 6.053361792956244e-06,
"loss": 0.70977745,
"memory(GiB)": 71.94,
"step": 1995,
"train_speed(iter/s)": 0.020089
},
{
"epoch": 0.827985924239288,
"grad_norm": 1.7447876663920783,
"learning_rate": 6.042689434364995e-06,
"loss": 0.72701621,
"memory(GiB)": 71.94,
"step": 2000,
"train_speed(iter/s)": 0.02009
},
{
"epoch": 0.8300558890498861,
"grad_norm": 1.8875191736470693,
"learning_rate": 6.032017075773746e-06,
"loss": 0.72644696,
"memory(GiB)": 71.94,
"step": 2005,
"train_speed(iter/s)": 0.02009
},
{
"epoch": 0.8321258538604843,
"grad_norm": 1.6472530019204896,
"learning_rate": 6.021344717182498e-06,
"loss": 0.70005097,
"memory(GiB)": 71.94,
"step": 2010,
"train_speed(iter/s)": 0.020091
},
{
"epoch": 0.8341958186710826,
"grad_norm": 2.2818513681921266,
"learning_rate": 6.01067235859125e-06,
"loss": 0.72365198,
"memory(GiB)": 71.94,
"step": 2015,
"train_speed(iter/s)": 0.020092
},
{
"epoch": 0.8362657834816808,
"grad_norm": 1.8205665836409684,
"learning_rate": 6e-06,
"loss": 0.7322978,
"memory(GiB)": 71.94,
"step": 2020,
"train_speed(iter/s)": 0.020092
},
{
"epoch": 0.8383357482922791,
"grad_norm": 1.980690292432822,
"learning_rate": 5.989327641408752e-06,
"loss": 0.73382425,
"memory(GiB)": 71.94,
"step": 2025,
"train_speed(iter/s)": 0.020093
},
{
"epoch": 0.8404057131028773,
"grad_norm": 1.7748584169287815,
"learning_rate": 5.978655282817502e-06,
"loss": 0.7543438,
"memory(GiB)": 71.94,
"step": 2030,
"train_speed(iter/s)": 0.020094
},
{
"epoch": 0.8424756779134754,
"grad_norm": 1.9477910790390784,
"learning_rate": 5.967982924226254e-06,
"loss": 0.73893185,
"memory(GiB)": 71.94,
"step": 2035,
"train_speed(iter/s)": 0.020095
},
{
"epoch": 0.8445456427240737,
"grad_norm": 1.5695526526206898,
"learning_rate": 5.957310565635006e-06,
"loss": 0.71403108,
"memory(GiB)": 71.94,
"step": 2040,
"train_speed(iter/s)": 0.020096
},
{
"epoch": 0.8466156075346719,
"grad_norm": 2.1517602299856557,
"learning_rate": 5.946638207043757e-06,
"loss": 0.71713848,
"memory(GiB)": 71.94,
"step": 2045,
"train_speed(iter/s)": 0.020097
},
{
"epoch": 0.8486855723452701,
"grad_norm": 2.739525728221928,
"learning_rate": 5.935965848452508e-06,
"loss": 0.72253714,
"memory(GiB)": 71.94,
"step": 2050,
"train_speed(iter/s)": 0.020099
},
{
"epoch": 0.8507555371558684,
"grad_norm": 1.9454548994868823,
"learning_rate": 5.92529348986126e-06,
"loss": 0.74796634,
"memory(GiB)": 71.94,
"step": 2055,
"train_speed(iter/s)": 0.0201
},
{
"epoch": 0.8528255019664666,
"grad_norm": 1.7996528216814918,
"learning_rate": 5.914621131270011e-06,
"loss": 0.71262107,
"memory(GiB)": 71.94,
"step": 2060,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.8548954667770647,
"grad_norm": 1.9001952709840753,
"learning_rate": 5.903948772678763e-06,
"loss": 0.71814876,
"memory(GiB)": 71.94,
"step": 2065,
"train_speed(iter/s)": 0.020101
},
{
"epoch": 0.856965431587663,
"grad_norm": 3.4631327121051036,
"learning_rate": 5.893276414087514e-06,
"loss": 0.70359154,
"memory(GiB)": 71.94,
"step": 2070,
"train_speed(iter/s)": 0.020103
},
{
"epoch": 0.8590353963982612,
"grad_norm": 1.9078035163840932,
"learning_rate": 5.882604055496265e-06,
"loss": 0.74100571,
"memory(GiB)": 71.94,
"step": 2075,
"train_speed(iter/s)": 0.020103
},
{
"epoch": 0.8611053612088595,
"grad_norm": 2.7698267455997576,
"learning_rate": 5.871931696905017e-06,
"loss": 0.71972656,
"memory(GiB)": 71.94,
"step": 2080,
"train_speed(iter/s)": 0.020104
},
{
"epoch": 0.8631753260194577,
"grad_norm": 1.9640858230267009,
"learning_rate": 5.861259338313769e-06,
"loss": 0.71868258,
"memory(GiB)": 71.94,
"step": 2085,
"train_speed(iter/s)": 0.020105
},
{
"epoch": 0.8652452908300559,
"grad_norm": 1.9593324236104832,
"learning_rate": 5.850586979722519e-06,
"loss": 0.70707326,
"memory(GiB)": 71.94,
"step": 2090,
"train_speed(iter/s)": 0.020106
},
{
"epoch": 0.8673152556406541,
"grad_norm": 2.0337621679872946,
"learning_rate": 5.839914621131271e-06,
"loss": 0.7303226,
"memory(GiB)": 71.94,
"step": 2095,
"train_speed(iter/s)": 0.020107
},
{
"epoch": 0.8693852204512523,
"grad_norm": 1.7464491411508778,
"learning_rate": 5.829242262540021e-06,
"loss": 0.70045071,
"memory(GiB)": 71.94,
"step": 2100,
"train_speed(iter/s)": 0.020108
},
{
"epoch": 0.8693852204512523,
"eval_loss": 0.8155556321144104,
"eval_runtime": 334.2741,
"eval_samples_per_second": 18.685,
"eval_steps_per_second": 1.17,
"step": 2100
},
{
"epoch": 0.8714551852618505,
"grad_norm": 1.893698412703024,
"learning_rate": 5.818569903948773e-06,
"loss": 0.71434135,
"memory(GiB)": 71.94,
"step": 2105,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 0.8735251500724488,
"grad_norm": 1.8915863471870793,
"learning_rate": 5.807897545357525e-06,
"loss": 0.73264565,
"memory(GiB)": 71.94,
"step": 2110,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 0.875595114883047,
"grad_norm": 1.5909748405215243,
"learning_rate": 5.797225186766276e-06,
"loss": 0.6809236,
"memory(GiB)": 71.94,
"step": 2115,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 0.8776650796936452,
"grad_norm": 2.298029244846505,
"learning_rate": 5.786552828175027e-06,
"loss": 0.72660675,
"memory(GiB)": 71.94,
"step": 2120,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 0.8797350445042434,
"grad_norm": 2.4471148736933634,
"learning_rate": 5.775880469583779e-06,
"loss": 0.71458502,
"memory(GiB)": 71.94,
"step": 2125,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 0.8818050093148416,
"grad_norm": 2.237641446928337,
"learning_rate": 5.76520811099253e-06,
"loss": 0.72527971,
"memory(GiB)": 71.94,
"step": 2130,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 0.8838749741254399,
"grad_norm": 2.1373854033173667,
"learning_rate": 5.754535752401282e-06,
"loss": 0.68969226,
"memory(GiB)": 71.94,
"step": 2135,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 0.8859449389360381,
"grad_norm": 2.4814505236104254,
"learning_rate": 5.743863393810033e-06,
"loss": 0.72110472,
"memory(GiB)": 71.94,
"step": 2140,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 0.8880149037466363,
"grad_norm": 1.8846949427239792,
"learning_rate": 5.733191035218784e-06,
"loss": 0.71526957,
"memory(GiB)": 71.94,
"step": 2145,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 0.8900848685572346,
"grad_norm": 1.6414510600406789,
"learning_rate": 5.722518676627536e-06,
"loss": 0.72875662,
"memory(GiB)": 71.94,
"step": 2150,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 0.8921548333678327,
"grad_norm": 2.027904279012944,
"learning_rate": 5.711846318036286e-06,
"loss": 0.73166924,
"memory(GiB)": 71.94,
"step": 2155,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 0.8942247981784309,
"grad_norm": 2.0007554119149136,
"learning_rate": 5.701173959445038e-06,
"loss": 0.69900818,
"memory(GiB)": 71.94,
"step": 2160,
"train_speed(iter/s)": 0.020046
},
{
"epoch": 0.8962947629890292,
"grad_norm": 1.732929144774659,
"learning_rate": 5.69050160085379e-06,
"loss": 0.70091386,
"memory(GiB)": 71.94,
"step": 2165,
"train_speed(iter/s)": 0.020046
},
{
"epoch": 0.8983647277996274,
"grad_norm": 1.6575807865879337,
"learning_rate": 5.67982924226254e-06,
"loss": 0.70530014,
"memory(GiB)": 71.94,
"step": 2170,
"train_speed(iter/s)": 0.020047
},
{
"epoch": 0.9004346926102256,
"grad_norm": 1.8402505669080536,
"learning_rate": 5.669156883671292e-06,
"loss": 0.72022433,
"memory(GiB)": 71.94,
"step": 2175,
"train_speed(iter/s)": 0.020048
},
{
"epoch": 0.9025046574208239,
"grad_norm": 1.8122270786550385,
"learning_rate": 5.6584845250800435e-06,
"loss": 0.67802505,
"memory(GiB)": 71.94,
"step": 2180,
"train_speed(iter/s)": 0.020048
},
{
"epoch": 0.904574622231422,
"grad_norm": 1.7969445274298348,
"learning_rate": 5.647812166488794e-06,
"loss": 0.70636382,
"memory(GiB)": 71.94,
"step": 2185,
"train_speed(iter/s)": 0.020049
},
{
"epoch": 0.9066445870420203,
"grad_norm": 2.1142537074713412,
"learning_rate": 5.637139807897546e-06,
"loss": 0.71558409,
"memory(GiB)": 71.94,
"step": 2190,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 0.9087145518526185,
"grad_norm": 1.944413383820558,
"learning_rate": 5.6264674493062975e-06,
"loss": 0.72004523,
"memory(GiB)": 71.94,
"step": 2195,
"train_speed(iter/s)": 0.020051
},
{
"epoch": 0.9107845166632167,
"grad_norm": 2.3051719970776934,
"learning_rate": 5.6157950907150485e-06,
"loss": 0.7308382,
"memory(GiB)": 71.94,
"step": 2200,
"train_speed(iter/s)": 0.020052
},
{
"epoch": 0.912854481473815,
"grad_norm": 2.27411033803044,
"learning_rate": 5.6051227321238e-06,
"loss": 0.74844613,
"memory(GiB)": 71.94,
"step": 2205,
"train_speed(iter/s)": 0.020053
},
{
"epoch": 0.9149244462844132,
"grad_norm": 1.8444250783225764,
"learning_rate": 5.594450373532551e-06,
"loss": 0.68941135,
"memory(GiB)": 71.94,
"step": 2210,
"train_speed(iter/s)": 0.020053
},
{
"epoch": 0.9169944110950113,
"grad_norm": 1.8662325411124825,
"learning_rate": 5.5837780149413025e-06,
"loss": 0.73066435,
"memory(GiB)": 71.94,
"step": 2215,
"train_speed(iter/s)": 0.020054
},
{
"epoch": 0.9190643759056096,
"grad_norm": 1.6833532844662813,
"learning_rate": 5.573105656350054e-06,
"loss": 0.7062602,
"memory(GiB)": 71.94,
"step": 2220,
"train_speed(iter/s)": 0.020055
},
{
"epoch": 0.9211343407162078,
"grad_norm": 1.6070808318678096,
"learning_rate": 5.562433297758805e-06,
"loss": 0.69585543,
"memory(GiB)": 71.94,
"step": 2225,
"train_speed(iter/s)": 0.020055
},
{
"epoch": 0.923204305526806,
"grad_norm": 2.0016548598313024,
"learning_rate": 5.5517609391675565e-06,
"loss": 0.7085475,
"memory(GiB)": 71.94,
"step": 2230,
"train_speed(iter/s)": 0.020056
},
{
"epoch": 0.9252742703374043,
"grad_norm": 1.9201243304059477,
"learning_rate": 5.541088580576308e-06,
"loss": 0.71516237,
"memory(GiB)": 71.94,
"step": 2235,
"train_speed(iter/s)": 0.020057
},
{
"epoch": 0.9273442351480025,
"grad_norm": 1.9055608045022892,
"learning_rate": 5.5304162219850586e-06,
"loss": 0.69740834,
"memory(GiB)": 71.94,
"step": 2240,
"train_speed(iter/s)": 0.020058
},
{
"epoch": 0.9294141999586008,
"grad_norm": 2.0041753291659026,
"learning_rate": 5.5197438633938104e-06,
"loss": 0.71469507,
"memory(GiB)": 71.94,
"step": 2245,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.9314841647691989,
"grad_norm": 2.025723530711083,
"learning_rate": 5.509071504802562e-06,
"loss": 0.69406614,
"memory(GiB)": 71.94,
"step": 2250,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.9335541295797971,
"grad_norm": 1.7845786824484478,
"learning_rate": 5.4983991462113125e-06,
"loss": 0.7137248,
"memory(GiB)": 71.94,
"step": 2255,
"train_speed(iter/s)": 0.020061
},
{
"epoch": 0.9356240943903954,
"grad_norm": 2.3504717810438196,
"learning_rate": 5.487726787620064e-06,
"loss": 0.70752907,
"memory(GiB)": 71.94,
"step": 2260,
"train_speed(iter/s)": 0.020062
},
{
"epoch": 0.9376940592009936,
"grad_norm": 2.0225261644141797,
"learning_rate": 5.477054429028816e-06,
"loss": 0.70490198,
"memory(GiB)": 71.94,
"step": 2265,
"train_speed(iter/s)": 0.020063
},
{
"epoch": 0.9397640240115918,
"grad_norm": 2.2107863770119747,
"learning_rate": 5.466382070437567e-06,
"loss": 0.71501665,
"memory(GiB)": 71.94,
"step": 2270,
"train_speed(iter/s)": 0.020064
},
{
"epoch": 0.94183398882219,
"grad_norm": 1.9030684437330223,
"learning_rate": 5.455709711846318e-06,
"loss": 0.70587921,
"memory(GiB)": 71.94,
"step": 2275,
"train_speed(iter/s)": 0.020064
},
{
"epoch": 0.9439039536327882,
"grad_norm": 1.8981878877998872,
"learning_rate": 5.445037353255069e-06,
"loss": 0.70294933,
"memory(GiB)": 71.94,
"step": 2280,
"train_speed(iter/s)": 0.020064
},
{
"epoch": 0.9459739184433864,
"grad_norm": 1.7195321236677932,
"learning_rate": 5.434364994663821e-06,
"loss": 0.71657829,
"memory(GiB)": 71.94,
"step": 2285,
"train_speed(iter/s)": 0.020065
},
{
"epoch": 0.9480438832539847,
"grad_norm": 1.6695574824900545,
"learning_rate": 5.423692636072573e-06,
"loss": 0.70917149,
"memory(GiB)": 71.94,
"step": 2290,
"train_speed(iter/s)": 0.020065
},
{
"epoch": 0.9501138480645829,
"grad_norm": 1.7410897548688689,
"learning_rate": 5.413020277481323e-06,
"loss": 0.7276587,
"memory(GiB)": 71.94,
"step": 2295,
"train_speed(iter/s)": 0.020066
},
{
"epoch": 0.9521838128751812,
"grad_norm": 1.6737135024901502,
"learning_rate": 5.402347918890075e-06,
"loss": 0.70822001,
"memory(GiB)": 71.94,
"step": 2300,
"train_speed(iter/s)": 0.020066
},
{
"epoch": 0.9542537776857793,
"grad_norm": 1.7876076111815575,
"learning_rate": 5.391675560298827e-06,
"loss": 0.72815857,
"memory(GiB)": 71.94,
"step": 2305,
"train_speed(iter/s)": 0.020067
},
{
"epoch": 0.9563237424963775,
"grad_norm": 1.653921158054905,
"learning_rate": 5.381003201707577e-06,
"loss": 0.715973,
"memory(GiB)": 71.94,
"step": 2310,
"train_speed(iter/s)": 0.020068
},
{
"epoch": 0.9583937073069758,
"grad_norm": 1.7830026539914627,
"learning_rate": 5.370330843116329e-06,
"loss": 0.70955439,
"memory(GiB)": 71.94,
"step": 2315,
"train_speed(iter/s)": 0.020068
},
{
"epoch": 0.960463672117574,
"grad_norm": 2.0794293583699712,
"learning_rate": 5.359658484525081e-06,
"loss": 0.71212959,
"memory(GiB)": 71.94,
"step": 2320,
"train_speed(iter/s)": 0.020049
},
{
"epoch": 0.9625336369281722,
"grad_norm": 1.62651854843721,
"learning_rate": 5.348986125933831e-06,
"loss": 0.69347043,
"memory(GiB)": 71.94,
"step": 2325,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 0.9646036017387705,
"grad_norm": 1.6345937025216515,
"learning_rate": 5.338313767342583e-06,
"loss": 0.68745365,
"memory(GiB)": 71.94,
"step": 2330,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 0.9666735665493686,
"grad_norm": 2.1460062258102504,
"learning_rate": 5.327641408751334e-06,
"loss": 0.69751196,
"memory(GiB)": 71.94,
"step": 2335,
"train_speed(iter/s)": 0.020051
},
{
"epoch": 0.9687435313599668,
"grad_norm": 1.8428729242460318,
"learning_rate": 5.316969050160086e-06,
"loss": 0.70150051,
"memory(GiB)": 71.94,
"step": 2340,
"train_speed(iter/s)": 0.020052
},
{
"epoch": 0.9708134961705651,
"grad_norm": 1.870750640547904,
"learning_rate": 5.306296691568837e-06,
"loss": 0.67915797,
"memory(GiB)": 71.94,
"step": 2345,
"train_speed(iter/s)": 0.020053
},
{
"epoch": 0.9728834609811633,
"grad_norm": 1.6984421405387677,
"learning_rate": 5.295624332977588e-06,
"loss": 0.71915126,
"memory(GiB)": 71.94,
"step": 2350,
"train_speed(iter/s)": 0.020054
},
{
"epoch": 0.9749534257917616,
"grad_norm": 1.7839025594515001,
"learning_rate": 5.28495197438634e-06,
"loss": 0.69594321,
"memory(GiB)": 71.94,
"step": 2355,
"train_speed(iter/s)": 0.020055
},
{
"epoch": 0.9770233906023598,
"grad_norm": 1.666815065361009,
"learning_rate": 5.274279615795091e-06,
"loss": 0.68858194,
"memory(GiB)": 71.94,
"step": 2360,
"train_speed(iter/s)": 0.020055
},
{
"epoch": 0.9790933554129579,
"grad_norm": 1.6613287141536495,
"learning_rate": 5.263607257203842e-06,
"loss": 0.65968986,
"memory(GiB)": 71.94,
"step": 2365,
"train_speed(iter/s)": 0.020056
},
{
"epoch": 0.9811633202235562,
"grad_norm": 1.5579649689164343,
"learning_rate": 5.252934898612594e-06,
"loss": 0.69255476,
"memory(GiB)": 71.94,
"step": 2370,
"train_speed(iter/s)": 0.020057
},
{
"epoch": 0.9832332850341544,
"grad_norm": 1.7564735837589676,
"learning_rate": 5.242262540021346e-06,
"loss": 0.71395578,
"memory(GiB)": 71.94,
"step": 2375,
"train_speed(iter/s)": 0.020057
},
{
"epoch": 0.9853032498447526,
"grad_norm": 2.0952603393058076,
"learning_rate": 5.231590181430096e-06,
"loss": 0.69916964,
"memory(GiB)": 71.94,
"step": 2380,
"train_speed(iter/s)": 0.020058
},
{
"epoch": 0.9873732146553509,
"grad_norm": 1.851124802207451,
"learning_rate": 5.220917822838848e-06,
"loss": 0.6992053,
"memory(GiB)": 71.94,
"step": 2385,
"train_speed(iter/s)": 0.020059
},
{
"epoch": 0.9894431794659491,
"grad_norm": 1.977421778833197,
"learning_rate": 5.2102454642476e-06,
"loss": 0.70937605,
"memory(GiB)": 71.94,
"step": 2390,
"train_speed(iter/s)": 0.020059
},
{
"epoch": 0.9915131442765474,
"grad_norm": 1.8125821116129224,
"learning_rate": 5.19957310565635e-06,
"loss": 0.70189705,
"memory(GiB)": 71.94,
"step": 2395,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.9935831090871455,
"grad_norm": 1.8812587805267227,
"learning_rate": 5.188900747065102e-06,
"loss": 0.68958111,
"memory(GiB)": 71.94,
"step": 2400,
"train_speed(iter/s)": 0.02006
},
{
"epoch": 0.9935831090871455,
"eval_loss": 0.812356173992157,
"eval_runtime": 333.5694,
"eval_samples_per_second": 18.725,
"eval_steps_per_second": 1.172,
"step": 2400
},
{
"epoch": 0.9956530738977437,
"grad_norm": 1.77525759977015,
"learning_rate": 5.178228388473853e-06,
"loss": 0.71421309,
"memory(GiB)": 71.94,
"step": 2405,
"train_speed(iter/s)": 0.019998
},
{
"epoch": 0.997723038708342,
"grad_norm": 1.9847604199741764,
"learning_rate": 5.167556029882604e-06,
"loss": 0.67816806,
"memory(GiB)": 71.94,
"step": 2410,
"train_speed(iter/s)": 0.019999
},
{
"epoch": 0.9997930035189402,
"grad_norm": 1.8831401901251863,
"learning_rate": 5.156883671291356e-06,
"loss": 0.67297964,
"memory(GiB)": 71.94,
"step": 2415,
"train_speed(iter/s)": 0.02
},
{
"epoch": 1.0018629683295384,
"grad_norm": 1.902600390439468,
"learning_rate": 5.146211312700107e-06,
"loss": 0.65476379,
"memory(GiB)": 71.94,
"step": 2420,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.0039329331401365,
"grad_norm": 1.8958163712669238,
"learning_rate": 5.135538954108859e-06,
"loss": 0.69827843,
"memory(GiB)": 71.94,
"step": 2425,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.006002897950735,
"grad_norm": 2.0893567670048774,
"learning_rate": 5.12486659551761e-06,
"loss": 0.6843009,
"memory(GiB)": 71.94,
"step": 2430,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.0080728627613331,
"grad_norm": 1.7283831963515028,
"learning_rate": 5.114194236926361e-06,
"loss": 0.66953688,
"memory(GiB)": 71.94,
"step": 2435,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.0101428275719313,
"grad_norm": 2.4550295961951023,
"learning_rate": 5.103521878335113e-06,
"loss": 0.6826447,
"memory(GiB)": 71.94,
"step": 2440,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.0122127923825295,
"grad_norm": 2.065348792100011,
"learning_rate": 5.092849519743865e-06,
"loss": 0.71306543,
"memory(GiB)": 71.94,
"step": 2445,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.0142827571931277,
"grad_norm": 1.8461278616269987,
"learning_rate": 5.082177161152615e-06,
"loss": 0.70268006,
"memory(GiB)": 71.94,
"step": 2450,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.0163527220037258,
"grad_norm": 1.8474179385577107,
"learning_rate": 5.071504802561367e-06,
"loss": 0.68759146,
"memory(GiB)": 71.94,
"step": 2455,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.0184226868143242,
"grad_norm": 2.1221766243412556,
"learning_rate": 5.060832443970117e-06,
"loss": 0.70161247,
"memory(GiB)": 71.94,
"step": 2460,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.0204926516249224,
"grad_norm": 1.7992664488684018,
"learning_rate": 5.050160085378869e-06,
"loss": 0.70210953,
"memory(GiB)": 71.94,
"step": 2465,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.0225626164355206,
"grad_norm": 2.010160051422228,
"learning_rate": 5.039487726787621e-06,
"loss": 0.70377779,
"memory(GiB)": 71.94,
"step": 2470,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.0246325812461188,
"grad_norm": 1.677895969704363,
"learning_rate": 5.028815368196372e-06,
"loss": 0.71084547,
"memory(GiB)": 71.94,
"step": 2475,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.026702546056717,
"grad_norm": 1.9057688334203953,
"learning_rate": 5.018143009605123e-06,
"loss": 0.67874904,
"memory(GiB)": 71.94,
"step": 2480,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.0287725108673154,
"grad_norm": 1.826208938410727,
"learning_rate": 5.007470651013875e-06,
"loss": 0.67107067,
"memory(GiB)": 71.94,
"step": 2485,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.0308424756779135,
"grad_norm": 2.0686042207681425,
"learning_rate": 4.996798292422626e-06,
"loss": 0.70525031,
"memory(GiB)": 71.94,
"step": 2490,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.0329124404885117,
"grad_norm": 1.7733232551181717,
"learning_rate": 4.986125933831378e-06,
"loss": 0.68992233,
"memory(GiB)": 71.94,
"step": 2495,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.03498240529911,
"grad_norm": 1.7950220673083168,
"learning_rate": 4.975453575240129e-06,
"loss": 0.67175922,
"memory(GiB)": 71.94,
"step": 2500,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.037052370109708,
"grad_norm": 1.9232030990454307,
"learning_rate": 4.96478121664888e-06,
"loss": 0.70407228,
"memory(GiB)": 71.94,
"step": 2505,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.0391223349203063,
"grad_norm": 1.708121386060437,
"learning_rate": 4.9541088580576316e-06,
"loss": 0.70078506,
"memory(GiB)": 71.94,
"step": 2510,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.0411922997309047,
"grad_norm": 1.5750528842945233,
"learning_rate": 4.943436499466383e-06,
"loss": 0.66830945,
"memory(GiB)": 71.94,
"step": 2515,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.0432622645415028,
"grad_norm": 1.8181518834205428,
"learning_rate": 4.932764140875134e-06,
"loss": 0.69259553,
"memory(GiB)": 71.94,
"step": 2520,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.045332229352101,
"grad_norm": 1.551131375424082,
"learning_rate": 4.9220917822838855e-06,
"loss": 0.69300728,
"memory(GiB)": 71.94,
"step": 2525,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.0474021941626992,
"grad_norm": 1.9636419108578083,
"learning_rate": 4.9114194236926366e-06,
"loss": 0.69065495,
"memory(GiB)": 71.94,
"step": 2530,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.0494721589732974,
"grad_norm": 2.2096885596291918,
"learning_rate": 4.900747065101388e-06,
"loss": 0.72614369,
"memory(GiB)": 71.94,
"step": 2535,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.0515421237838958,
"grad_norm": 1.8198162773587976,
"learning_rate": 4.890074706510139e-06,
"loss": 0.68779688,
"memory(GiB)": 71.94,
"step": 2540,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.053612088594494,
"grad_norm": 1.8767827748337365,
"learning_rate": 4.8794023479188905e-06,
"loss": 0.65902538,
"memory(GiB)": 71.94,
"step": 2545,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.0556820534050921,
"grad_norm": 1.774731724427948,
"learning_rate": 4.8687299893276416e-06,
"loss": 0.6885426,
"memory(GiB)": 71.94,
"step": 2550,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.0577520182156903,
"grad_norm": 1.7292661960079105,
"learning_rate": 4.858057630736393e-06,
"loss": 0.67627816,
"memory(GiB)": 71.94,
"step": 2555,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.0598219830262885,
"grad_norm": 2.022917994019762,
"learning_rate": 4.8473852721451445e-06,
"loss": 0.69524202,
"memory(GiB)": 71.94,
"step": 2560,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.0618919478368867,
"grad_norm": 1.7503339393851076,
"learning_rate": 4.8367129135538955e-06,
"loss": 0.69861703,
"memory(GiB)": 71.94,
"step": 2565,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.063961912647485,
"grad_norm": 2.3241272289849126,
"learning_rate": 4.826040554962647e-06,
"loss": 0.69416137,
"memory(GiB)": 71.94,
"step": 2570,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.0660318774580833,
"grad_norm": 1.9145774620065716,
"learning_rate": 4.8153681963713985e-06,
"loss": 0.707304,
"memory(GiB)": 71.94,
"step": 2575,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.0681018422686814,
"grad_norm": 2.2205440096454363,
"learning_rate": 4.80469583778015e-06,
"loss": 0.68690495,
"memory(GiB)": 71.94,
"step": 2580,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.0701718070792796,
"grad_norm": 1.907681817748529,
"learning_rate": 4.794023479188901e-06,
"loss": 0.67836595,
"memory(GiB)": 71.94,
"step": 2585,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.0722417718898778,
"grad_norm": 1.812113935386319,
"learning_rate": 4.783351120597652e-06,
"loss": 0.69949398,
"memory(GiB)": 71.94,
"step": 2590,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.0743117367004762,
"grad_norm": 1.9890056601624129,
"learning_rate": 4.7726787620064035e-06,
"loss": 0.68311234,
"memory(GiB)": 71.94,
"step": 2595,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.0763817015110744,
"grad_norm": 1.6389803143854558,
"learning_rate": 4.762006403415155e-06,
"loss": 0.68783474,
"memory(GiB)": 71.94,
"step": 2600,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.0784516663216726,
"grad_norm": 1.7243055433624384,
"learning_rate": 4.751334044823906e-06,
"loss": 0.68109202,
"memory(GiB)": 71.94,
"step": 2605,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.0805216311322707,
"grad_norm": 2.0564605414653356,
"learning_rate": 4.740661686232657e-06,
"loss": 0.67778783,
"memory(GiB)": 71.94,
"step": 2610,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.082591595942869,
"grad_norm": 1.9673716530688552,
"learning_rate": 4.729989327641409e-06,
"loss": 0.67061253,
"memory(GiB)": 71.94,
"step": 2615,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.084661560753467,
"grad_norm": 2.256366938059263,
"learning_rate": 4.71931696905016e-06,
"loss": 0.6971777,
"memory(GiB)": 71.94,
"step": 2620,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.0867315255640655,
"grad_norm": 2.092500467589189,
"learning_rate": 4.708644610458911e-06,
"loss": 0.67864285,
"memory(GiB)": 71.94,
"step": 2625,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.0888014903746637,
"grad_norm": 1.7063119566002638,
"learning_rate": 4.697972251867663e-06,
"loss": 0.66528416,
"memory(GiB)": 71.94,
"step": 2630,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.0908714551852619,
"grad_norm": 1.7825515797938636,
"learning_rate": 4.687299893276414e-06,
"loss": 0.67014618,
"memory(GiB)": 71.94,
"step": 2635,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.09294141999586,
"grad_norm": 1.7034629950636981,
"learning_rate": 4.676627534685166e-06,
"loss": 0.67851248,
"memory(GiB)": 71.94,
"step": 2640,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.0950113848064582,
"grad_norm": 1.6535356357438644,
"learning_rate": 4.665955176093917e-06,
"loss": 0.69617391,
"memory(GiB)": 71.94,
"step": 2645,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.0970813496170566,
"grad_norm": 1.790083800076922,
"learning_rate": 4.655282817502668e-06,
"loss": 0.68771133,
"memory(GiB)": 71.94,
"step": 2650,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.0991513144276548,
"grad_norm": 1.9625451394764908,
"learning_rate": 4.64461045891142e-06,
"loss": 0.68705645,
"memory(GiB)": 71.94,
"step": 2655,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.101221279238253,
"grad_norm": 2.41883263144616,
"learning_rate": 4.633938100320171e-06,
"loss": 0.66960602,
"memory(GiB)": 71.94,
"step": 2660,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 1.1032912440488511,
"grad_norm": 1.6833879907399165,
"learning_rate": 4.623265741728922e-06,
"loss": 0.67477846,
"memory(GiB)": 71.94,
"step": 2665,
"train_speed(iter/s)": 0.020045
},
{
"epoch": 1.1053612088594493,
"grad_norm": 1.84460802825979,
"learning_rate": 4.612593383137674e-06,
"loss": 0.66959124,
"memory(GiB)": 71.94,
"step": 2670,
"train_speed(iter/s)": 0.020046
},
{
"epoch": 1.1074311736700475,
"grad_norm": 1.8226080355090242,
"learning_rate": 4.601921024546425e-06,
"loss": 0.66996727,
"memory(GiB)": 71.94,
"step": 2675,
"train_speed(iter/s)": 0.020047
},
{
"epoch": 1.109501138480646,
"grad_norm": 1.9273809027924724,
"learning_rate": 4.591248665955176e-06,
"loss": 0.69185095,
"memory(GiB)": 71.94,
"step": 2680,
"train_speed(iter/s)": 0.020048
},
{
"epoch": 1.111571103291244,
"grad_norm": 2.2322742496415517,
"learning_rate": 4.580576307363927e-06,
"loss": 0.70724788,
"memory(GiB)": 71.94,
"step": 2685,
"train_speed(iter/s)": 0.020049
},
{
"epoch": 1.1136410681018423,
"grad_norm": 2.1881226444266133,
"learning_rate": 4.569903948772679e-06,
"loss": 0.67986469,
"memory(GiB)": 71.94,
"step": 2690,
"train_speed(iter/s)": 0.020049
},
{
"epoch": 1.1157110329124404,
"grad_norm": 1.9344503535315112,
"learning_rate": 4.55923159018143e-06,
"loss": 0.6932023,
"memory(GiB)": 71.94,
"step": 2695,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 1.1177809977230386,
"grad_norm": 2.0197719509697056,
"learning_rate": 4.548559231590182e-06,
"loss": 0.67475772,
"memory(GiB)": 71.94,
"step": 2700,
"train_speed(iter/s)": 0.020051
},
{
"epoch": 1.1177809977230386,
"eval_loss": 0.8077359795570374,
"eval_runtime": 332.9193,
"eval_samples_per_second": 18.761,
"eval_steps_per_second": 1.174,
"step": 2700
},
{
"epoch": 1.119850962533637,
"grad_norm": 1.8305900241415576,
"learning_rate": 4.537886872998933e-06,
"loss": 0.71444244,
"memory(GiB)": 71.94,
"step": 2705,
"train_speed(iter/s)": 0.019995
},
{
"epoch": 1.1219209273442352,
"grad_norm": 1.9866747607495971,
"learning_rate": 4.527214514407685e-06,
"loss": 0.68752618,
"memory(GiB)": 71.94,
"step": 2710,
"train_speed(iter/s)": 0.019996
},
{
"epoch": 1.1239908921548334,
"grad_norm": 1.8866009316885002,
"learning_rate": 4.516542155816436e-06,
"loss": 0.67673769,
"memory(GiB)": 71.94,
"step": 2715,
"train_speed(iter/s)": 0.019997
},
{
"epoch": 1.1260608569654316,
"grad_norm": 1.8693665648036668,
"learning_rate": 4.505869797225187e-06,
"loss": 0.68302212,
"memory(GiB)": 71.94,
"step": 2720,
"train_speed(iter/s)": 0.019997
},
{
"epoch": 1.1281308217760297,
"grad_norm": 2.088485544417028,
"learning_rate": 4.495197438633939e-06,
"loss": 0.69513187,
"memory(GiB)": 71.94,
"step": 2725,
"train_speed(iter/s)": 0.019999
},
{
"epoch": 1.1302007865866281,
"grad_norm": 1.83149183273408,
"learning_rate": 4.48452508004269e-06,
"loss": 0.67474871,
"memory(GiB)": 71.94,
"step": 2730,
"train_speed(iter/s)": 0.02
},
{
"epoch": 1.1322707513972263,
"grad_norm": 1.9681727198174188,
"learning_rate": 4.473852721451441e-06,
"loss": 0.6727396,
"memory(GiB)": 71.94,
"step": 2735,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.1343407162078245,
"grad_norm": 3.0386548884501288,
"learning_rate": 4.463180362860193e-06,
"loss": 0.69408131,
"memory(GiB)": 71.94,
"step": 2740,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.1364106810184227,
"grad_norm": 3.156158661144904,
"learning_rate": 4.452508004268944e-06,
"loss": 0.6340518,
"memory(GiB)": 71.94,
"step": 2745,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.1384806458290209,
"grad_norm": 2.005947464562104,
"learning_rate": 4.441835645677695e-06,
"loss": 0.68589301,
"memory(GiB)": 71.94,
"step": 2750,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.140550610639619,
"grad_norm": 1.8910732500138312,
"learning_rate": 4.431163287086446e-06,
"loss": 0.68320475,
"memory(GiB)": 71.94,
"step": 2755,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.1426205754502172,
"grad_norm": 1.5980061705957451,
"learning_rate": 4.420490928495198e-06,
"loss": 0.68288679,
"memory(GiB)": 71.94,
"step": 2760,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.1446905402608156,
"grad_norm": 2.1893217377555367,
"learning_rate": 4.409818569903949e-06,
"loss": 0.67679596,
"memory(GiB)": 71.94,
"step": 2765,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.1467605050714138,
"grad_norm": 1.6144531687050387,
"learning_rate": 4.3991462113127e-06,
"loss": 0.67341776,
"memory(GiB)": 71.94,
"step": 2770,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.148830469882012,
"grad_norm": 1.7848487108948465,
"learning_rate": 4.388473852721452e-06,
"loss": 0.65265465,
"memory(GiB)": 71.94,
"step": 2775,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.1509004346926102,
"grad_norm": 1.9198684626115081,
"learning_rate": 4.377801494130203e-06,
"loss": 0.65432949,
"memory(GiB)": 71.94,
"step": 2780,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.1529703995032086,
"grad_norm": 1.8075401444295365,
"learning_rate": 4.367129135538955e-06,
"loss": 0.66166754,
"memory(GiB)": 71.94,
"step": 2785,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.1550403643138067,
"grad_norm": 2.093007547124246,
"learning_rate": 4.356456776947706e-06,
"loss": 0.671489,
"memory(GiB)": 71.94,
"step": 2790,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.157110329124405,
"grad_norm": 2.155911339314564,
"learning_rate": 4.345784418356458e-06,
"loss": 0.66719484,
"memory(GiB)": 71.94,
"step": 2795,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.159180293935003,
"grad_norm": 1.7929721304219823,
"learning_rate": 4.335112059765209e-06,
"loss": 0.65916939,
"memory(GiB)": 71.94,
"step": 2800,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.1612502587456013,
"grad_norm": 2.3382800828112695,
"learning_rate": 4.32443970117396e-06,
"loss": 0.68535299,
"memory(GiB)": 71.94,
"step": 2805,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.1633202235561995,
"grad_norm": 1.9393370355158424,
"learning_rate": 4.313767342582711e-06,
"loss": 0.68975515,
"memory(GiB)": 71.94,
"step": 2810,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.1653901883667976,
"grad_norm": 2.1161711862572172,
"learning_rate": 4.303094983991463e-06,
"loss": 0.68595347,
"memory(GiB)": 71.94,
"step": 2815,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.167460153177396,
"grad_norm": 1.89196514815735,
"learning_rate": 4.292422625400214e-06,
"loss": 0.66012936,
"memory(GiB)": 71.94,
"step": 2820,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.1695301179879942,
"grad_norm": 1.8601131110854523,
"learning_rate": 4.281750266808965e-06,
"loss": 0.68001904,
"memory(GiB)": 71.94,
"step": 2825,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.1716000827985924,
"grad_norm": 1.8930363611249428,
"learning_rate": 4.271077908217717e-06,
"loss": 0.66193466,
"memory(GiB)": 71.94,
"step": 2830,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.1736700476091906,
"grad_norm": 1.651757107446397,
"learning_rate": 4.260405549626468e-06,
"loss": 0.67280817,
"memory(GiB)": 71.94,
"step": 2835,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.175740012419789,
"grad_norm": 1.6748291349437752,
"learning_rate": 4.249733191035219e-06,
"loss": 0.70100274,
"memory(GiB)": 71.94,
"step": 2840,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.1778099772303872,
"grad_norm": 1.7834078724205271,
"learning_rate": 4.239060832443971e-06,
"loss": 0.64067845,
"memory(GiB)": 71.94,
"step": 2845,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.1798799420409853,
"grad_norm": 1.9357492367842137,
"learning_rate": 4.228388473852722e-06,
"loss": 0.68998647,
"memory(GiB)": 71.94,
"step": 2850,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.1819499068515835,
"grad_norm": 1.9186551723129406,
"learning_rate": 4.2177161152614736e-06,
"loss": 0.67889709,
"memory(GiB)": 71.94,
"step": 2855,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.1840198716621817,
"grad_norm": 1.9166194791943714,
"learning_rate": 4.207043756670225e-06,
"loss": 0.67329493,
"memory(GiB)": 71.94,
"step": 2860,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.1860898364727799,
"grad_norm": 1.9439517613212347,
"learning_rate": 4.196371398078976e-06,
"loss": 0.6614254,
"memory(GiB)": 71.94,
"step": 2865,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.188159801283378,
"grad_norm": 1.6420722807797328,
"learning_rate": 4.1856990394877275e-06,
"loss": 0.67023277,
"memory(GiB)": 71.94,
"step": 2870,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.1902297660939765,
"grad_norm": 1.7445706716402636,
"learning_rate": 4.1750266808964786e-06,
"loss": 0.67275252,
"memory(GiB)": 71.94,
"step": 2875,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.1922997309045746,
"grad_norm": 1.8273002856280824,
"learning_rate": 4.16435432230523e-06,
"loss": 0.65841031,
"memory(GiB)": 71.94,
"step": 2880,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.1943696957151728,
"grad_norm": 1.7392597436189736,
"learning_rate": 4.1536819637139815e-06,
"loss": 0.66225605,
"memory(GiB)": 71.94,
"step": 2885,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.196439660525771,
"grad_norm": 1.639200530922351,
"learning_rate": 4.1430096051227325e-06,
"loss": 0.65032516,
"memory(GiB)": 71.94,
"step": 2890,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.1985096253363694,
"grad_norm": 1.8960982589133293,
"learning_rate": 4.1323372465314836e-06,
"loss": 0.69994841,
"memory(GiB)": 71.94,
"step": 2895,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.2005795901469676,
"grad_norm": 2.1518082070491613,
"learning_rate": 4.121664887940235e-06,
"loss": 0.64357581,
"memory(GiB)": 71.94,
"step": 2900,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.2026495549575658,
"grad_norm": 2.450141804935637,
"learning_rate": 4.1109925293489865e-06,
"loss": 0.68345547,
"memory(GiB)": 71.94,
"step": 2905,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.204719519768164,
"grad_norm": 1.8681725857258895,
"learning_rate": 4.1003201707577375e-06,
"loss": 0.66512461,
"memory(GiB)": 71.94,
"step": 2910,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.2067894845787621,
"grad_norm": 2.2592923799668774,
"learning_rate": 4.089647812166489e-06,
"loss": 0.68196335,
"memory(GiB)": 71.94,
"step": 2915,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.2088594493893603,
"grad_norm": 2.0216041168873775,
"learning_rate": 4.0789754535752404e-06,
"loss": 0.6808126,
"memory(GiB)": 71.94,
"step": 2920,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.2109294141999585,
"grad_norm": 1.7659468605949793,
"learning_rate": 4.0683030949839915e-06,
"loss": 0.64283953,
"memory(GiB)": 71.94,
"step": 2925,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.2129993790105569,
"grad_norm": 1.970065027416205,
"learning_rate": 4.057630736392743e-06,
"loss": 0.6555728,
"memory(GiB)": 71.94,
"step": 2930,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.215069343821155,
"grad_norm": 1.998737392424493,
"learning_rate": 4.046958377801494e-06,
"loss": 0.6660428,
"memory(GiB)": 71.94,
"step": 2935,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.2171393086317532,
"grad_norm": 2.1589153805149173,
"learning_rate": 4.036286019210246e-06,
"loss": 0.65805073,
"memory(GiB)": 71.94,
"step": 2940,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.2192092734423514,
"grad_norm": 1.990247509684529,
"learning_rate": 4.025613660618997e-06,
"loss": 0.6413794,
"memory(GiB)": 71.94,
"step": 2945,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.2212792382529498,
"grad_norm": 2.0438116812415625,
"learning_rate": 4.014941302027748e-06,
"loss": 0.65000477,
"memory(GiB)": 71.94,
"step": 2950,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.223349203063548,
"grad_norm": 1.6608044995599232,
"learning_rate": 4.004268943436499e-06,
"loss": 0.65562267,
"memory(GiB)": 71.94,
"step": 2955,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.2254191678741462,
"grad_norm": 1.9291845395844707,
"learning_rate": 3.993596584845251e-06,
"loss": 0.67264051,
"memory(GiB)": 71.94,
"step": 2960,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.2274891326847444,
"grad_norm": 2.3618922331006753,
"learning_rate": 3.982924226254002e-06,
"loss": 0.68117104,
"memory(GiB)": 71.94,
"step": 2965,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 1.2295590974953425,
"grad_norm": 1.9561928806095756,
"learning_rate": 3.972251867662753e-06,
"loss": 0.6778089,
"memory(GiB)": 71.94,
"step": 2970,
"train_speed(iter/s)": 0.020045
},
{
"epoch": 1.2316290623059407,
"grad_norm": 2.103021206596317,
"learning_rate": 3.961579509071505e-06,
"loss": 0.66529841,
"memory(GiB)": 71.94,
"step": 2975,
"train_speed(iter/s)": 0.020046
},
{
"epoch": 1.2336990271165391,
"grad_norm": 1.6317138464756236,
"learning_rate": 3.950907150480256e-06,
"loss": 0.63005896,
"memory(GiB)": 71.94,
"step": 2980,
"train_speed(iter/s)": 0.020047
},
{
"epoch": 1.2357689919271373,
"grad_norm": 1.8763154321083348,
"learning_rate": 3.940234791889007e-06,
"loss": 0.69782147,
"memory(GiB)": 71.94,
"step": 2985,
"train_speed(iter/s)": 0.020048
},
{
"epoch": 1.2378389567377355,
"grad_norm": 2.12720513001939,
"learning_rate": 3.929562433297759e-06,
"loss": 0.67128716,
"memory(GiB)": 71.94,
"step": 2990,
"train_speed(iter/s)": 0.020049
},
{
"epoch": 1.2399089215483337,
"grad_norm": 2.0560360745042243,
"learning_rate": 3.91889007470651e-06,
"loss": 0.66187463,
"memory(GiB)": 71.94,
"step": 2995,
"train_speed(iter/s)": 0.02005
},
{
"epoch": 1.2419788863589318,
"grad_norm": 1.9884557277874149,
"learning_rate": 3.908217716115262e-06,
"loss": 0.68540969,
"memory(GiB)": 71.94,
"step": 3000,
"train_speed(iter/s)": 0.020051
},
{
"epoch": 1.2419788863589318,
"eval_loss": 0.8033931255340576,
"eval_runtime": 334.3532,
"eval_samples_per_second": 18.681,
"eval_steps_per_second": 1.169,
"step": 3000
},
{
"epoch": 1.2440488511695302,
"grad_norm": 1.8132336089456924,
"learning_rate": 3.897545357524013e-06,
"loss": 0.65594606,
"memory(GiB)": 71.94,
"step": 3005,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.2461188159801284,
"grad_norm": 1.6982582415846914,
"learning_rate": 3.886872998932765e-06,
"loss": 0.65645032,
"memory(GiB)": 71.94,
"step": 3010,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.2481887807907266,
"grad_norm": 1.9942609739954071,
"learning_rate": 3.876200640341516e-06,
"loss": 0.66572218,
"memory(GiB)": 71.94,
"step": 3015,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.2502587456013248,
"grad_norm": 1.7952303972411592,
"learning_rate": 3.865528281750267e-06,
"loss": 0.64628544,
"memory(GiB)": 71.94,
"step": 3020,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.252328710411923,
"grad_norm": 1.9058576079441742,
"learning_rate": 3.854855923159018e-06,
"loss": 0.68286681,
"memory(GiB)": 71.94,
"step": 3025,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.2543986752225211,
"grad_norm": 2.4990275874354357,
"learning_rate": 3.84418356456777e-06,
"loss": 0.67756252,
"memory(GiB)": 71.94,
"step": 3030,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.2564686400331193,
"grad_norm": 2.18800214487155,
"learning_rate": 3.833511205976521e-06,
"loss": 0.65295424,
"memory(GiB)": 71.94,
"step": 3035,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.2585386048437177,
"grad_norm": 2.1525831686251737,
"learning_rate": 3.822838847385272e-06,
"loss": 0.66122398,
"memory(GiB)": 71.94,
"step": 3040,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.260608569654316,
"grad_norm": 2.1001976709383374,
"learning_rate": 3.812166488794024e-06,
"loss": 0.64958653,
"memory(GiB)": 71.94,
"step": 3045,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.262678534464914,
"grad_norm": 1.8005403174974608,
"learning_rate": 3.801494130202775e-06,
"loss": 0.66733065,
"memory(GiB)": 71.94,
"step": 3050,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.2647484992755123,
"grad_norm": 2.0392708573153255,
"learning_rate": 3.7908217716115265e-06,
"loss": 0.65150566,
"memory(GiB)": 71.94,
"step": 3055,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.2668184640861107,
"grad_norm": 1.9818256506070209,
"learning_rate": 3.7801494130202776e-06,
"loss": 0.65937147,
"memory(GiB)": 71.94,
"step": 3060,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.2688884288967088,
"grad_norm": 2.091473795655044,
"learning_rate": 3.7694770544290294e-06,
"loss": 0.65645385,
"memory(GiB)": 71.94,
"step": 3065,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.270958393707307,
"grad_norm": 1.9887693066644563,
"learning_rate": 3.7588046958377805e-06,
"loss": 0.6356863,
"memory(GiB)": 71.94,
"step": 3070,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.2730283585179052,
"grad_norm": 1.849373670863495,
"learning_rate": 3.7481323372465315e-06,
"loss": 0.64913082,
"memory(GiB)": 71.94,
"step": 3075,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.2750983233285034,
"grad_norm": 2.8796661699092776,
"learning_rate": 3.737459978655283e-06,
"loss": 0.66612153,
"memory(GiB)": 71.94,
"step": 3080,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.2771682881391015,
"grad_norm": 1.9206629854919683,
"learning_rate": 3.7267876200640345e-06,
"loss": 0.65432153,
"memory(GiB)": 71.94,
"step": 3085,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.2792382529496997,
"grad_norm": 2.0649234898157562,
"learning_rate": 3.716115261472786e-06,
"loss": 0.63801923,
"memory(GiB)": 71.94,
"step": 3090,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.2813082177602981,
"grad_norm": 1.859875670695032,
"learning_rate": 3.705442902881537e-06,
"loss": 0.6548945,
"memory(GiB)": 71.94,
"step": 3095,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.2833781825708963,
"grad_norm": 1.767661952337357,
"learning_rate": 3.694770544290289e-06,
"loss": 0.64956121,
"memory(GiB)": 71.94,
"step": 3100,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.2854481473814945,
"grad_norm": 1.8007833500981838,
"learning_rate": 3.68409818569904e-06,
"loss": 0.669453,
"memory(GiB)": 71.94,
"step": 3105,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.2875181121920927,
"grad_norm": 2.0099069002485863,
"learning_rate": 3.673425827107791e-06,
"loss": 0.67178354,
"memory(GiB)": 71.94,
"step": 3110,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.289588077002691,
"grad_norm": 2.0463019564128886,
"learning_rate": 3.6627534685165424e-06,
"loss": 0.65595369,
"memory(GiB)": 71.94,
"step": 3115,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.2916580418132892,
"grad_norm": 1.9152750413831356,
"learning_rate": 3.652081109925294e-06,
"loss": 0.65940399,
"memory(GiB)": 71.94,
"step": 3120,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.2937280066238874,
"grad_norm": 2.9816950881544306,
"learning_rate": 3.6414087513340453e-06,
"loss": 0.66090908,
"memory(GiB)": 71.94,
"step": 3125,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.2957979714344856,
"grad_norm": 1.8591409546980415,
"learning_rate": 3.6307363927427963e-06,
"loss": 0.66902189,
"memory(GiB)": 71.94,
"step": 3130,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.2978679362450838,
"grad_norm": 2.078590180758969,
"learning_rate": 3.6200640341515482e-06,
"loss": 0.66616745,
"memory(GiB)": 71.94,
"step": 3135,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.299937901055682,
"grad_norm": 1.8683848031966166,
"learning_rate": 3.6093916755602993e-06,
"loss": 0.64836683,
"memory(GiB)": 71.94,
"step": 3140,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.3020078658662801,
"grad_norm": 1.7924123567589454,
"learning_rate": 3.5987193169690503e-06,
"loss": 0.66824627,
"memory(GiB)": 71.94,
"step": 3145,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.3040778306768785,
"grad_norm": 1.94170729576761,
"learning_rate": 3.5880469583778018e-06,
"loss": 0.67116079,
"memory(GiB)": 71.94,
"step": 3150,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.3061477954874767,
"grad_norm": 2.1208196521106357,
"learning_rate": 3.5773745997865532e-06,
"loss": 0.66398339,
"memory(GiB)": 71.94,
"step": 3155,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.308217760298075,
"grad_norm": 1.94319742259963,
"learning_rate": 3.5667022411953047e-06,
"loss": 0.66830492,
"memory(GiB)": 71.94,
"step": 3160,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.310287725108673,
"grad_norm": 1.8986704786653348,
"learning_rate": 3.5560298826040557e-06,
"loss": 0.64020205,
"memory(GiB)": 71.94,
"step": 3165,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.3123576899192715,
"grad_norm": 1.832166033780513,
"learning_rate": 3.5453575240128068e-06,
"loss": 0.65297813,
"memory(GiB)": 71.94,
"step": 3170,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.3144276547298697,
"grad_norm": 1.7524632564639653,
"learning_rate": 3.5346851654215586e-06,
"loss": 0.65833459,
"memory(GiB)": 71.94,
"step": 3175,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.3164976195404678,
"grad_norm": 1.784819778645554,
"learning_rate": 3.5240128068303097e-06,
"loss": 0.65319304,
"memory(GiB)": 71.94,
"step": 3180,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.318567584351066,
"grad_norm": 1.7883657805218875,
"learning_rate": 3.513340448239061e-06,
"loss": 0.65240812,
"memory(GiB)": 71.94,
"step": 3185,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.3206375491616642,
"grad_norm": 1.894144860689911,
"learning_rate": 3.5026680896478126e-06,
"loss": 0.6723794,
"memory(GiB)": 71.94,
"step": 3190,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.3227075139722624,
"grad_norm": 2.0320221849381195,
"learning_rate": 3.491995731056564e-06,
"loss": 0.64077559,
"memory(GiB)": 71.94,
"step": 3195,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.3247774787828606,
"grad_norm": 1.727824740633838,
"learning_rate": 3.481323372465315e-06,
"loss": 0.63609524,
"memory(GiB)": 71.94,
"step": 3200,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.326847443593459,
"grad_norm": 1.7454884742520516,
"learning_rate": 3.470651013874066e-06,
"loss": 0.65425596,
"memory(GiB)": 71.94,
"step": 3205,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.3289174084040571,
"grad_norm": 1.8911130178984759,
"learning_rate": 3.459978655282818e-06,
"loss": 0.66199875,
"memory(GiB)": 71.94,
"step": 3210,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.3309873732146553,
"grad_norm": 2.3250937309778474,
"learning_rate": 3.449306296691569e-06,
"loss": 0.65385923,
"memory(GiB)": 71.94,
"step": 3215,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.3330573380252535,
"grad_norm": 2.1704453467921447,
"learning_rate": 3.43863393810032e-06,
"loss": 0.66229916,
"memory(GiB)": 71.94,
"step": 3220,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.335127302835852,
"grad_norm": 2.2209117698054768,
"learning_rate": 3.427961579509072e-06,
"loss": 0.65051212,
"memory(GiB)": 71.94,
"step": 3225,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.33719726764645,
"grad_norm": 1.6845784156480532,
"learning_rate": 3.417289220917823e-06,
"loss": 0.64576015,
"memory(GiB)": 71.94,
"step": 3230,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.3392672324570483,
"grad_norm": 2.7721561024455332,
"learning_rate": 3.4066168623265745e-06,
"loss": 0.68257604,
"memory(GiB)": 71.94,
"step": 3235,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.3413371972676464,
"grad_norm": 1.9362479296558788,
"learning_rate": 3.3959445037353255e-06,
"loss": 0.63483586,
"memory(GiB)": 71.94,
"step": 3240,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.3434071620782446,
"grad_norm": 1.9956347191580364,
"learning_rate": 3.3852721451440774e-06,
"loss": 0.65959587,
"memory(GiB)": 71.94,
"step": 3245,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.3454771268888428,
"grad_norm": 1.7321062675450865,
"learning_rate": 3.3745997865528285e-06,
"loss": 0.66381845,
"memory(GiB)": 71.94,
"step": 3250,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.347547091699441,
"grad_norm": 1.9696403539230432,
"learning_rate": 3.3639274279615795e-06,
"loss": 0.64812822,
"memory(GiB)": 71.94,
"step": 3255,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.3496170565100394,
"grad_norm": 2.1319401432279053,
"learning_rate": 3.353255069370331e-06,
"loss": 0.64005919,
"memory(GiB)": 71.94,
"step": 3260,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.3516870213206376,
"grad_norm": 1.8363706763408283,
"learning_rate": 3.3425827107790824e-06,
"loss": 0.64201698,
"memory(GiB)": 71.94,
"step": 3265,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.3537569861312357,
"grad_norm": 2.124129454637882,
"learning_rate": 3.331910352187834e-06,
"loss": 0.65025196,
"memory(GiB)": 71.94,
"step": 3270,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.355826950941834,
"grad_norm": 1.7896864503844883,
"learning_rate": 3.321237993596585e-06,
"loss": 0.63300438,
"memory(GiB)": 71.94,
"step": 3275,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 1.3578969157524323,
"grad_norm": 1.9837261215389441,
"learning_rate": 3.310565635005337e-06,
"loss": 0.63380709,
"memory(GiB)": 71.94,
"step": 3280,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 1.3599668805630305,
"grad_norm": 2.097355543637799,
"learning_rate": 3.299893276414088e-06,
"loss": 0.62883258,
"memory(GiB)": 71.94,
"step": 3285,
"train_speed(iter/s)": 0.020044
},
{
"epoch": 1.3620368453736287,
"grad_norm": 1.9315070924851874,
"learning_rate": 3.289220917822839e-06,
"loss": 0.63195004,
"memory(GiB)": 71.94,
"step": 3290,
"train_speed(iter/s)": 0.020045
},
{
"epoch": 1.3641068101842269,
"grad_norm": 1.7050134771385488,
"learning_rate": 3.2785485592315903e-06,
"loss": 0.65333614,
"memory(GiB)": 71.94,
"step": 3295,
"train_speed(iter/s)": 0.020045
},
{
"epoch": 1.366176774994825,
"grad_norm": 3.6481119398862005,
"learning_rate": 3.267876200640342e-06,
"loss": 0.68793478,
"memory(GiB)": 71.94,
"step": 3300,
"train_speed(iter/s)": 0.020046
},
{
"epoch": 1.366176774994825,
"eval_loss": 0.8023556470870972,
"eval_runtime": 334.0894,
"eval_samples_per_second": 18.696,
"eval_steps_per_second": 1.17,
"step": 3300
},
{
"epoch": 1.3682467398054232,
"grad_norm": 2.092330422962737,
"learning_rate": 3.2572038420490933e-06,
"loss": 0.63313837,
"memory(GiB)": 71.94,
"step": 3305,
"train_speed(iter/s)": 0.02
},
{
"epoch": 1.3703167046160214,
"grad_norm": 1.9167879677265438,
"learning_rate": 3.2465314834578443e-06,
"loss": 0.65159397,
"memory(GiB)": 71.94,
"step": 3310,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.3723866694266198,
"grad_norm": 1.8761921604953933,
"learning_rate": 3.235859124866596e-06,
"loss": 0.62655239,
"memory(GiB)": 71.94,
"step": 3315,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.374456634237218,
"grad_norm": 2.178411039961288,
"learning_rate": 3.2251867662753472e-06,
"loss": 0.67218485,
"memory(GiB)": 71.94,
"step": 3320,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.3765265990478162,
"grad_norm": 1.776609105623676,
"learning_rate": 3.2145144076840983e-06,
"loss": 0.63943739,
"memory(GiB)": 71.94,
"step": 3325,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.3785965638584143,
"grad_norm": 1.8854963168848657,
"learning_rate": 3.2038420490928497e-06,
"loss": 0.65016818,
"memory(GiB)": 71.94,
"step": 3330,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.3806665286690127,
"grad_norm": 2.3494051096524915,
"learning_rate": 3.193169690501601e-06,
"loss": 0.6696517,
"memory(GiB)": 71.94,
"step": 3335,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.382736493479611,
"grad_norm": 2.390390518794705,
"learning_rate": 3.1824973319103527e-06,
"loss": 0.64389553,
"memory(GiB)": 71.94,
"step": 3340,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.384806458290209,
"grad_norm": 1.8195738414381504,
"learning_rate": 3.1718249733191037e-06,
"loss": 0.64606419,
"memory(GiB)": 71.94,
"step": 3345,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.3868764231008073,
"grad_norm": 2.3772901728441846,
"learning_rate": 3.1611526147278547e-06,
"loss": 0.66564407,
"memory(GiB)": 71.94,
"step": 3350,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.3889463879114055,
"grad_norm": 1.9865374550113573,
"learning_rate": 3.1504802561366066e-06,
"loss": 0.62708969,
"memory(GiB)": 71.94,
"step": 3355,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.3910163527220036,
"grad_norm": 1.8995100511755476,
"learning_rate": 3.1398078975453577e-06,
"loss": 0.64633865,
"memory(GiB)": 71.94,
"step": 3360,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.3930863175326018,
"grad_norm": 2.3050512969189954,
"learning_rate": 3.129135538954109e-06,
"loss": 0.64496307,
"memory(GiB)": 71.94,
"step": 3365,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.3951562823432002,
"grad_norm": 1.9056478303612696,
"learning_rate": 3.1184631803628606e-06,
"loss": 0.641465,
"memory(GiB)": 71.94,
"step": 3370,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.3972262471537984,
"grad_norm": 2.166573531526341,
"learning_rate": 3.107790821771612e-06,
"loss": 0.64069824,
"memory(GiB)": 71.94,
"step": 3375,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.3992962119643966,
"grad_norm": 1.9398811847394521,
"learning_rate": 3.097118463180363e-06,
"loss": 0.64301763,
"memory(GiB)": 47.6,
"step": 3380,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.4013661767749948,
"grad_norm": 1.959159982373063,
"learning_rate": 3.086446104589114e-06,
"loss": 0.6362546,
"memory(GiB)": 47.6,
"step": 3385,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.4034361415855932,
"grad_norm": 2.0530926805349083,
"learning_rate": 3.075773745997866e-06,
"loss": 0.63418798,
"memory(GiB)": 47.6,
"step": 3390,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.4055061063961913,
"grad_norm": 1.9297510699914728,
"learning_rate": 3.065101387406617e-06,
"loss": 0.6311512,
"memory(GiB)": 47.6,
"step": 3395,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.4075760712067895,
"grad_norm": 1.9271894163285872,
"learning_rate": 3.054429028815368e-06,
"loss": 0.63234644,
"memory(GiB)": 47.6,
"step": 3400,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.4096460360173877,
"grad_norm": 1.7395603251397769,
"learning_rate": 3.04375667022412e-06,
"loss": 0.63525658,
"memory(GiB)": 47.6,
"step": 3405,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.4117160008279859,
"grad_norm": 1.9592997788913435,
"learning_rate": 3.033084311632871e-06,
"loss": 0.62426691,
"memory(GiB)": 47.6,
"step": 3410,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.413785965638584,
"grad_norm": 1.83244625095987,
"learning_rate": 3.0224119530416225e-06,
"loss": 0.64436603,
"memory(GiB)": 47.6,
"step": 3415,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.4158559304491822,
"grad_norm": 1.9274094419667949,
"learning_rate": 3.0117395944503735e-06,
"loss": 0.65462785,
"memory(GiB)": 47.6,
"step": 3420,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.4179258952597806,
"grad_norm": 1.9910550366499922,
"learning_rate": 3.0010672358591254e-06,
"loss": 0.62345576,
"memory(GiB)": 47.6,
"step": 3425,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.4199958600703788,
"grad_norm": 1.9961790791051468,
"learning_rate": 2.9903948772678764e-06,
"loss": 0.62686663,
"memory(GiB)": 48.58,
"step": 3430,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.422065824880977,
"grad_norm": 2.2471760043812767,
"learning_rate": 2.9797225186766275e-06,
"loss": 0.65531764,
"memory(GiB)": 48.58,
"step": 3435,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.4241357896915752,
"grad_norm": 2.225238329922807,
"learning_rate": 2.9690501600853794e-06,
"loss": 0.64034252,
"memory(GiB)": 48.58,
"step": 3440,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.4262057545021736,
"grad_norm": 1.6663799088887756,
"learning_rate": 2.9583778014941304e-06,
"loss": 0.61355238,
"memory(GiB)": 48.58,
"step": 3445,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.4282757193127718,
"grad_norm": 1.9127892416039678,
"learning_rate": 2.947705442902882e-06,
"loss": 0.61548796,
"memory(GiB)": 48.58,
"step": 3450,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.43034568412337,
"grad_norm": 1.7733393610398442,
"learning_rate": 2.937033084311633e-06,
"loss": 0.66395988,
"memory(GiB)": 48.58,
"step": 3455,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.432415648933968,
"grad_norm": 2.1962868424551676,
"learning_rate": 2.9263607257203848e-06,
"loss": 0.61601906,
"memory(GiB)": 48.58,
"step": 3460,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.4344856137445663,
"grad_norm": 1.8134260024404016,
"learning_rate": 2.915688367129136e-06,
"loss": 0.64126596,
"memory(GiB)": 48.58,
"step": 3465,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.4365555785551645,
"grad_norm": 2.136457991499728,
"learning_rate": 2.905016008537887e-06,
"loss": 0.63421082,
"memory(GiB)": 48.58,
"step": 3470,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.4386255433657626,
"grad_norm": 1.988532670614334,
"learning_rate": 2.8943436499466383e-06,
"loss": 0.63867459,
"memory(GiB)": 48.58,
"step": 3475,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.440695508176361,
"grad_norm": 1.9277626836086974,
"learning_rate": 2.8836712913553898e-06,
"loss": 0.65572329,
"memory(GiB)": 48.58,
"step": 3480,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.4427654729869592,
"grad_norm": 2.1432613819151247,
"learning_rate": 2.8729989327641412e-06,
"loss": 0.64992175,
"memory(GiB)": 48.58,
"step": 3485,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.4448354377975574,
"grad_norm": 2.355567308350272,
"learning_rate": 2.8623265741728923e-06,
"loss": 0.66429825,
"memory(GiB)": 48.58,
"step": 3490,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.4469054026081556,
"grad_norm": 2.045569698254085,
"learning_rate": 2.851654215581644e-06,
"loss": 0.65050926,
"memory(GiB)": 48.58,
"step": 3495,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.448975367418754,
"grad_norm": 1.8389546498339602,
"learning_rate": 2.840981856990395e-06,
"loss": 0.63454714,
"memory(GiB)": 48.58,
"step": 3500,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.4510453322293522,
"grad_norm": 2.1282462505999518,
"learning_rate": 2.8303094983991462e-06,
"loss": 0.61026077,
"memory(GiB)": 48.58,
"step": 3505,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.4531152970399503,
"grad_norm": 1.9822876450283116,
"learning_rate": 2.8196371398078977e-06,
"loss": 0.63836317,
"memory(GiB)": 48.58,
"step": 3510,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.4551852618505485,
"grad_norm": 2.1201754460326603,
"learning_rate": 2.808964781216649e-06,
"loss": 0.65191045,
"memory(GiB)": 48.58,
"step": 3515,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.4572552266611467,
"grad_norm": 1.8520751525916404,
"learning_rate": 2.7982924226254006e-06,
"loss": 0.62120395,
"memory(GiB)": 48.58,
"step": 3520,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.4593251914717449,
"grad_norm": 1.7983375554778653,
"learning_rate": 2.7876200640341517e-06,
"loss": 0.63227787,
"memory(GiB)": 48.58,
"step": 3525,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.461395156282343,
"grad_norm": 2.340481282081477,
"learning_rate": 2.7769477054429036e-06,
"loss": 0.63372889,
"memory(GiB)": 48.58,
"step": 3530,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.4634651210929415,
"grad_norm": 1.9262925464363927,
"learning_rate": 2.7662753468516546e-06,
"loss": 0.63678517,
"memory(GiB)": 48.58,
"step": 3535,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.4655350859035396,
"grad_norm": 2.2093247947290164,
"learning_rate": 2.7556029882604056e-06,
"loss": 0.6558732,
"memory(GiB)": 48.58,
"step": 3540,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.4676050507141378,
"grad_norm": 2.020333271689009,
"learning_rate": 2.744930629669157e-06,
"loss": 0.60733166,
"memory(GiB)": 48.58,
"step": 3545,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.469675015524736,
"grad_norm": 2.0734276125917632,
"learning_rate": 2.7342582710779086e-06,
"loss": 0.66044526,
"memory(GiB)": 48.58,
"step": 3550,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.4717449803353344,
"grad_norm": 1.8332312440056042,
"learning_rate": 2.72358591248666e-06,
"loss": 0.64025326,
"memory(GiB)": 48.58,
"step": 3555,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.4738149451459326,
"grad_norm": 1.8856532947296074,
"learning_rate": 2.712913553895411e-06,
"loss": 0.63681345,
"memory(GiB)": 48.58,
"step": 3560,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.4758849099565308,
"grad_norm": 1.6710776220883121,
"learning_rate": 2.702241195304162e-06,
"loss": 0.64032116,
"memory(GiB)": 48.58,
"step": 3565,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.477954874767129,
"grad_norm": 2.095636589348373,
"learning_rate": 2.691568836712914e-06,
"loss": 0.63223238,
"memory(GiB)": 48.58,
"step": 3570,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.4800248395777271,
"grad_norm": 1.8927929015268774,
"learning_rate": 2.680896478121665e-06,
"loss": 0.63061609,
"memory(GiB)": 48.58,
"step": 3575,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.4820948043883253,
"grad_norm": 2.0929754486788137,
"learning_rate": 2.670224119530416e-06,
"loss": 0.62695656,
"memory(GiB)": 48.58,
"step": 3580,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.4841647691989235,
"grad_norm": 1.7458918576306506,
"learning_rate": 2.659551760939168e-06,
"loss": 0.63322277,
"memory(GiB)": 48.58,
"step": 3585,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.4862347340095219,
"grad_norm": 1.8907196532347343,
"learning_rate": 2.648879402347919e-06,
"loss": 0.63732347,
"memory(GiB)": 48.58,
"step": 3590,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.48830469882012,
"grad_norm": 2.263161529685459,
"learning_rate": 2.6382070437566704e-06,
"loss": 0.6138607,
"memory(GiB)": 48.58,
"step": 3595,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.4903746636307182,
"grad_norm": 2.572454412416673,
"learning_rate": 2.6275346851654215e-06,
"loss": 0.63049603,
"memory(GiB)": 48.58,
"step": 3600,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.4903746636307182,
"eval_loss": 0.8007386922836304,
"eval_runtime": 333.8857,
"eval_samples_per_second": 18.707,
"eval_steps_per_second": 1.171,
"step": 3600
},
{
"epoch": 1.4924446284413164,
"grad_norm": 1.901029287627636,
"learning_rate": 2.6168623265741734e-06,
"loss": 0.65960011,
"memory(GiB)": 48.58,
"step": 3605,
"train_speed(iter/s)": 0.019998
},
{
"epoch": 1.4945145932519148,
"grad_norm": 2.0108141055428423,
"learning_rate": 2.6061899679829244e-06,
"loss": 0.63098369,
"memory(GiB)": 48.58,
"step": 3610,
"train_speed(iter/s)": 0.019999
},
{
"epoch": 1.496584558062513,
"grad_norm": 1.8385014047339334,
"learning_rate": 2.5955176093916754e-06,
"loss": 0.63654456,
"memory(GiB)": 48.58,
"step": 3615,
"train_speed(iter/s)": 0.019999
},
{
"epoch": 1.4986545228731112,
"grad_norm": 2.502391600466987,
"learning_rate": 2.5848452508004273e-06,
"loss": 0.64129109,
"memory(GiB)": 48.58,
"step": 3620,
"train_speed(iter/s)": 0.02
},
{
"epoch": 1.5007244876837094,
"grad_norm": 1.8078536683803752,
"learning_rate": 2.5741728922091784e-06,
"loss": 0.61762905,
"memory(GiB)": 48.58,
"step": 3625,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.5027944524943075,
"grad_norm": 2.4124021719958813,
"learning_rate": 2.56350053361793e-06,
"loss": 0.62886133,
"memory(GiB)": 48.58,
"step": 3630,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.5048644173049057,
"grad_norm": 2.14818078370575,
"learning_rate": 2.552828175026681e-06,
"loss": 0.64229774,
"memory(GiB)": 48.58,
"step": 3635,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.506934382115504,
"grad_norm": 2.0607592121598777,
"learning_rate": 2.5421558164354328e-06,
"loss": 0.62922821,
"memory(GiB)": 48.58,
"step": 3640,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.5090043469261023,
"grad_norm": 1.9537072105756903,
"learning_rate": 2.531483457844184e-06,
"loss": 0.6114254,
"memory(GiB)": 48.58,
"step": 3645,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.5110743117367005,
"grad_norm": 2.076894133768813,
"learning_rate": 2.520811099252935e-06,
"loss": 0.65702705,
"memory(GiB)": 48.58,
"step": 3650,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.5131442765472987,
"grad_norm": 1.8684495033846724,
"learning_rate": 2.5101387406616863e-06,
"loss": 0.65235605,
"memory(GiB)": 48.58,
"step": 3655,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.515214241357897,
"grad_norm": 2.099820461610625,
"learning_rate": 2.4994663820704378e-06,
"loss": 0.63383131,
"memory(GiB)": 48.58,
"step": 3660,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.5172842061684952,
"grad_norm": 2.0251264820061765,
"learning_rate": 2.4887940234791892e-06,
"loss": 0.63897676,
"memory(GiB)": 48.58,
"step": 3665,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.5193541709790934,
"grad_norm": 1.8916351044126378,
"learning_rate": 2.4781216648879407e-06,
"loss": 0.63116732,
"memory(GiB)": 48.58,
"step": 3670,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.5214241357896916,
"grad_norm": 1.758567214100863,
"learning_rate": 2.4674493062966917e-06,
"loss": 0.61664515,
"memory(GiB)": 48.58,
"step": 3675,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.5234941006002898,
"grad_norm": 2.004985649591871,
"learning_rate": 2.456776947705443e-06,
"loss": 0.63878107,
"memory(GiB)": 48.58,
"step": 3680,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.525564065410888,
"grad_norm": 2.0496876810492894,
"learning_rate": 2.4461045891141942e-06,
"loss": 0.62643003,
"memory(GiB)": 48.58,
"step": 3685,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.5276340302214861,
"grad_norm": 1.9829235096697742,
"learning_rate": 2.4354322305229457e-06,
"loss": 0.62055197,
"memory(GiB)": 48.58,
"step": 3690,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.5297039950320843,
"grad_norm": 1.7991062631513268,
"learning_rate": 2.424759871931697e-06,
"loss": 0.6605298,
"memory(GiB)": 48.58,
"step": 3695,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.5317739598426827,
"grad_norm": 2.526220221053364,
"learning_rate": 2.4140875133404486e-06,
"loss": 0.65973449,
"memory(GiB)": 48.58,
"step": 3700,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.533843924653281,
"grad_norm": 1.682540853202414,
"learning_rate": 2.4034151547492e-06,
"loss": 0.65421052,
"memory(GiB)": 48.58,
"step": 3705,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.535913889463879,
"grad_norm": 1.965734649630595,
"learning_rate": 2.392742796157951e-06,
"loss": 0.64042482,
"memory(GiB)": 48.58,
"step": 3710,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.5379838542744775,
"grad_norm": 2.2061994907060005,
"learning_rate": 2.3820704375667026e-06,
"loss": 0.61186495,
"memory(GiB)": 48.58,
"step": 3715,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.5400538190850757,
"grad_norm": 2.2749623686643474,
"learning_rate": 2.3713980789754536e-06,
"loss": 0.63306904,
"memory(GiB)": 48.58,
"step": 3720,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.5421237838956738,
"grad_norm": 1.8676661262594305,
"learning_rate": 2.360725720384205e-06,
"loss": 0.64988294,
"memory(GiB)": 48.58,
"step": 3725,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.544193748706272,
"grad_norm": 2.069377222520985,
"learning_rate": 2.3500533617929565e-06,
"loss": 0.63958693,
"memory(GiB)": 48.58,
"step": 3730,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.5462637135168702,
"grad_norm": 2.0148406212108574,
"learning_rate": 2.339381003201708e-06,
"loss": 0.60387516,
"memory(GiB)": 48.58,
"step": 3735,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.5483336783274684,
"grad_norm": 1.8592072092273917,
"learning_rate": 2.328708644610459e-06,
"loss": 0.60533466,
"memory(GiB)": 48.58,
"step": 3740,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.5504036431380666,
"grad_norm": 1.9293967097992641,
"learning_rate": 2.3180362860192105e-06,
"loss": 0.64885559,
"memory(GiB)": 48.58,
"step": 3745,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.5524736079486647,
"grad_norm": 2.105150298236883,
"learning_rate": 2.307363927427962e-06,
"loss": 0.6291214,
"memory(GiB)": 48.58,
"step": 3750,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.5545435727592631,
"grad_norm": 1.7706279189258218,
"learning_rate": 2.296691568836713e-06,
"loss": 0.62637095,
"memory(GiB)": 48.58,
"step": 3755,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.5566135375698613,
"grad_norm": 2.565565704557741,
"learning_rate": 2.2860192102454645e-06,
"loss": 0.61526871,
"memory(GiB)": 48.58,
"step": 3760,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.5586835023804595,
"grad_norm": 2.07386908713523,
"learning_rate": 2.2753468516542155e-06,
"loss": 0.62105417,
"memory(GiB)": 48.58,
"step": 3765,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.560753467191058,
"grad_norm": 2.0656516897338064,
"learning_rate": 2.264674493062967e-06,
"loss": 0.62474051,
"memory(GiB)": 48.58,
"step": 3770,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.562823432001656,
"grad_norm": 2.1881677769227075,
"learning_rate": 2.2540021344717184e-06,
"loss": 0.6300148,
"memory(GiB)": 48.58,
"step": 3775,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.5648933968122543,
"grad_norm": 1.9586494376217993,
"learning_rate": 2.24332977588047e-06,
"loss": 0.59469123,
"memory(GiB)": 48.58,
"step": 3780,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.5669633616228524,
"grad_norm": 2.440599800961732,
"learning_rate": 2.232657417289221e-06,
"loss": 0.62468157,
"memory(GiB)": 48.58,
"step": 3785,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.5690333264334506,
"grad_norm": 2.0016886010949797,
"learning_rate": 2.2219850586979724e-06,
"loss": 0.62775316,
"memory(GiB)": 48.58,
"step": 3790,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.5711032912440488,
"grad_norm": 1.9086880143876215,
"learning_rate": 2.211312700106724e-06,
"loss": 0.62862492,
"memory(GiB)": 48.58,
"step": 3795,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.573173256054647,
"grad_norm": 2.0888610038198636,
"learning_rate": 2.200640341515475e-06,
"loss": 0.59845972,
"memory(GiB)": 48.58,
"step": 3800,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.5752432208652452,
"grad_norm": 2.055969507055363,
"learning_rate": 2.1899679829242263e-06,
"loss": 0.61734905,
"memory(GiB)": 48.58,
"step": 3805,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.5773131856758436,
"grad_norm": 1.9707520677944765,
"learning_rate": 2.179295624332978e-06,
"loss": 0.62550197,
"memory(GiB)": 48.58,
"step": 3810,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.5793831504864417,
"grad_norm": 2.1321630269454617,
"learning_rate": 2.1686232657417293e-06,
"loss": 0.62156429,
"memory(GiB)": 48.58,
"step": 3815,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.58145311529704,
"grad_norm": 2.3313380056807373,
"learning_rate": 2.1579509071504803e-06,
"loss": 0.62489176,
"memory(GiB)": 48.58,
"step": 3820,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.5835230801076383,
"grad_norm": 2.0579013534602044,
"learning_rate": 2.1472785485592318e-06,
"loss": 0.60487609,
"memory(GiB)": 48.58,
"step": 3825,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.5855930449182365,
"grad_norm": 2.1449642041698267,
"learning_rate": 2.136606189967983e-06,
"loss": 0.63105164,
"memory(GiB)": 48.58,
"step": 3830,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.5876630097288347,
"grad_norm": 1.8838252023426636,
"learning_rate": 2.1259338313767343e-06,
"loss": 0.62664189,
"memory(GiB)": 48.58,
"step": 3835,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.5897329745394329,
"grad_norm": 2.289021375023386,
"learning_rate": 2.1152614727854857e-06,
"loss": 0.60014114,
"memory(GiB)": 48.58,
"step": 3840,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.591802939350031,
"grad_norm": 1.952623771952289,
"learning_rate": 2.104589114194237e-06,
"loss": 0.63038387,
"memory(GiB)": 48.58,
"step": 3845,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.5938729041606292,
"grad_norm": 2.151088803547346,
"learning_rate": 2.0939167556029887e-06,
"loss": 0.64208837,
"memory(GiB)": 48.58,
"step": 3850,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.5959428689712274,
"grad_norm": 2.0667970850832025,
"learning_rate": 2.0832443970117397e-06,
"loss": 0.62941227,
"memory(GiB)": 48.58,
"step": 3855,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.5980128337818256,
"grad_norm": 1.8287162112237207,
"learning_rate": 2.072572038420491e-06,
"loss": 0.60888386,
"memory(GiB)": 48.58,
"step": 3860,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.600082798592424,
"grad_norm": 2.083482144562111,
"learning_rate": 2.061899679829242e-06,
"loss": 0.60378475,
"memory(GiB)": 48.58,
"step": 3865,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.6021527634030222,
"grad_norm": 2.3064863216709925,
"learning_rate": 2.0512273212379937e-06,
"loss": 0.60522232,
"memory(GiB)": 48.58,
"step": 3870,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.6042227282136203,
"grad_norm": 2.1700368490113844,
"learning_rate": 2.040554962646745e-06,
"loss": 0.61363611,
"memory(GiB)": 48.58,
"step": 3875,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.6062926930242187,
"grad_norm": 2.2038759685376843,
"learning_rate": 2.0298826040554966e-06,
"loss": 0.60480423,
"memory(GiB)": 48.58,
"step": 3880,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.608362657834817,
"grad_norm": 2.1947427482623914,
"learning_rate": 2.019210245464248e-06,
"loss": 0.61413918,
"memory(GiB)": 48.58,
"step": 3885,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.610432622645415,
"grad_norm": 2.022770825774821,
"learning_rate": 2.008537886872999e-06,
"loss": 0.61320429,
"memory(GiB)": 48.58,
"step": 3890,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.6125025874560133,
"grad_norm": 2.052392899511488,
"learning_rate": 1.9978655282817505e-06,
"loss": 0.6130487,
"memory(GiB)": 48.58,
"step": 3895,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.6145725522666114,
"grad_norm": 2.0156586912928596,
"learning_rate": 1.9871931696905016e-06,
"loss": 0.64238014,
"memory(GiB)": 48.58,
"step": 3900,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.6145725522666114,
"eval_loss": 0.79938143491745,
"eval_runtime": 333.9342,
"eval_samples_per_second": 18.704,
"eval_steps_per_second": 1.171,
"step": 3900
},
{
"epoch": 1.6166425170772096,
"grad_norm": 2.023426003459966,
"learning_rate": 1.976520811099253e-06,
"loss": 0.61180491,
"memory(GiB)": 48.58,
"step": 3905,
"train_speed(iter/s)": 0.020001
},
{
"epoch": 1.6187124818878078,
"grad_norm": 2.053001775702235,
"learning_rate": 1.9658484525080045e-06,
"loss": 0.62299452,
"memory(GiB)": 48.58,
"step": 3910,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.620782446698406,
"grad_norm": 1.867769590995953,
"learning_rate": 1.955176093916756e-06,
"loss": 0.61298056,
"memory(GiB)": 48.58,
"step": 3915,
"train_speed(iter/s)": 0.020002
},
{
"epoch": 1.6228524115090044,
"grad_norm": 2.3195386012179404,
"learning_rate": 1.944503735325507e-06,
"loss": 0.62579803,
"memory(GiB)": 48.58,
"step": 3920,
"train_speed(iter/s)": 0.020003
},
{
"epoch": 1.6249223763196026,
"grad_norm": 2.0728629845134825,
"learning_rate": 1.9338313767342585e-06,
"loss": 0.60280285,
"memory(GiB)": 48.58,
"step": 3925,
"train_speed(iter/s)": 0.020004
},
{
"epoch": 1.6269923411302007,
"grad_norm": 1.9603574858854882,
"learning_rate": 1.92315901814301e-06,
"loss": 0.61659031,
"memory(GiB)": 48.58,
"step": 3930,
"train_speed(iter/s)": 0.020005
},
{
"epoch": 1.6290623059407991,
"grad_norm": 2.1782071337826014,
"learning_rate": 1.912486659551761e-06,
"loss": 0.60701981,
"memory(GiB)": 48.58,
"step": 3935,
"train_speed(iter/s)": 0.020006
},
{
"epoch": 1.6311322707513973,
"grad_norm": 1.7195215240069541,
"learning_rate": 1.9018143009605124e-06,
"loss": 0.59961052,
"memory(GiB)": 48.58,
"step": 3940,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.6332022355619955,
"grad_norm": 2.076700307294715,
"learning_rate": 1.8911419423692637e-06,
"loss": 0.61397924,
"memory(GiB)": 48.58,
"step": 3945,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.6352722003725937,
"grad_norm": 2.1171856192990615,
"learning_rate": 1.8804695837780151e-06,
"loss": 0.61544151,
"memory(GiB)": 48.58,
"step": 3950,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.6373421651831919,
"grad_norm": 2.2158147695337838,
"learning_rate": 1.8697972251867664e-06,
"loss": 0.62159052,
"memory(GiB)": 48.58,
"step": 3955,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.63941212999379,
"grad_norm": 1.8635077630631116,
"learning_rate": 1.8591248665955179e-06,
"loss": 0.62834597,
"memory(GiB)": 48.58,
"step": 3960,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.6414820948043882,
"grad_norm": 1.9610933756803843,
"learning_rate": 1.8484525080042693e-06,
"loss": 0.61779599,
"memory(GiB)": 48.58,
"step": 3965,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.6435520596149864,
"grad_norm": 1.9029383953647434,
"learning_rate": 1.8377801494130204e-06,
"loss": 0.61077566,
"memory(GiB)": 48.58,
"step": 3970,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.6456220244255848,
"grad_norm": 2.0846014405355926,
"learning_rate": 1.8271077908217718e-06,
"loss": 0.61311092,
"memory(GiB)": 48.58,
"step": 3975,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.647691989236183,
"grad_norm": 2.390238831095919,
"learning_rate": 1.816435432230523e-06,
"loss": 0.62842617,
"memory(GiB)": 48.58,
"step": 3980,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.6497619540467812,
"grad_norm": 1.8650578046674164,
"learning_rate": 1.8057630736392745e-06,
"loss": 0.62033787,
"memory(GiB)": 48.58,
"step": 3985,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.6518319188573796,
"grad_norm": 1.9997553186688102,
"learning_rate": 1.7950907150480258e-06,
"loss": 0.61835356,
"memory(GiB)": 48.58,
"step": 3990,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.6539018836679777,
"grad_norm": 2.0185275543308,
"learning_rate": 1.7844183564567772e-06,
"loss": 0.62564411,
"memory(GiB)": 48.58,
"step": 3995,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.655971848478576,
"grad_norm": 1.6493866482983615,
"learning_rate": 1.7737459978655283e-06,
"loss": 0.62476611,
"memory(GiB)": 48.58,
"step": 4000,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.658041813289174,
"grad_norm": 2.1258878217039157,
"learning_rate": 1.7630736392742797e-06,
"loss": 0.61094875,
"memory(GiB)": 48.58,
"step": 4005,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.6601117780997723,
"grad_norm": 1.8264630776919117,
"learning_rate": 1.7524012806830312e-06,
"loss": 0.61455221,
"memory(GiB)": 48.58,
"step": 4010,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.6621817429103705,
"grad_norm": 2.1834934664197148,
"learning_rate": 1.7417289220917825e-06,
"loss": 0.61525717,
"memory(GiB)": 48.58,
"step": 4015,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.6642517077209686,
"grad_norm": 1.951531129714831,
"learning_rate": 1.731056563500534e-06,
"loss": 0.61558003,
"memory(GiB)": 48.58,
"step": 4020,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.6663216725315668,
"grad_norm": 2.4480446682769315,
"learning_rate": 1.720384204909285e-06,
"loss": 0.58283405,
"memory(GiB)": 48.58,
"step": 4025,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.6683916373421652,
"grad_norm": 1.7243360581014375,
"learning_rate": 1.7097118463180364e-06,
"loss": 0.60816731,
"memory(GiB)": 48.58,
"step": 4030,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.6704616021527634,
"grad_norm": 1.935378370315751,
"learning_rate": 1.6990394877267877e-06,
"loss": 0.59363813,
"memory(GiB)": 48.58,
"step": 4035,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.6725315669633616,
"grad_norm": 2.343341523870396,
"learning_rate": 1.6883671291355391e-06,
"loss": 0.59670277,
"memory(GiB)": 48.58,
"step": 4040,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.67460153177396,
"grad_norm": 2.002429630691932,
"learning_rate": 1.6776947705442904e-06,
"loss": 0.61318674,
"memory(GiB)": 48.58,
"step": 4045,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.6766714965845582,
"grad_norm": 2.161836907625474,
"learning_rate": 1.6670224119530418e-06,
"loss": 0.63845253,
"memory(GiB)": 48.58,
"step": 4050,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.6787414613951563,
"grad_norm": 2.3095783835557993,
"learning_rate": 1.6563500533617933e-06,
"loss": 0.61825991,
"memory(GiB)": 48.58,
"step": 4055,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.6808114262057545,
"grad_norm": 2.0857994277393326,
"learning_rate": 1.6456776947705443e-06,
"loss": 0.58271861,
"memory(GiB)": 48.58,
"step": 4060,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.6828813910163527,
"grad_norm": 2.591242513945162,
"learning_rate": 1.6350053361792958e-06,
"loss": 0.59729037,
"memory(GiB)": 48.58,
"step": 4065,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.6849513558269509,
"grad_norm": 2.2408993824641836,
"learning_rate": 1.624332977588047e-06,
"loss": 0.61958027,
"memory(GiB)": 48.58,
"step": 4070,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.687021320637549,
"grad_norm": 2.07836511660639,
"learning_rate": 1.6136606189967985e-06,
"loss": 0.61059999,
"memory(GiB)": 48.58,
"step": 4075,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.6890912854481472,
"grad_norm": 2.1885333872092767,
"learning_rate": 1.6029882604055498e-06,
"loss": 0.61813364,
"memory(GiB)": 48.58,
"step": 4080,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.6911612502587456,
"grad_norm": 2.0563236890963075,
"learning_rate": 1.5923159018143012e-06,
"loss": 0.62983589,
"memory(GiB)": 48.58,
"step": 4085,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.6932312150693438,
"grad_norm": 2.205936525088634,
"learning_rate": 1.5816435432230523e-06,
"loss": 0.6022356,
"memory(GiB)": 48.58,
"step": 4090,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.695301179879942,
"grad_norm": 2.1192285983309262,
"learning_rate": 1.5709711846318037e-06,
"loss": 0.61405392,
"memory(GiB)": 48.58,
"step": 4095,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.6973711446905404,
"grad_norm": 2.2747714483339676,
"learning_rate": 1.5602988260405552e-06,
"loss": 0.59890566,
"memory(GiB)": 48.58,
"step": 4100,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.6994411095011386,
"grad_norm": 1.7865480667421139,
"learning_rate": 1.5496264674493064e-06,
"loss": 0.59438276,
"memory(GiB)": 48.58,
"step": 4105,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.7015110743117368,
"grad_norm": 2.1136132702931953,
"learning_rate": 1.538954108858058e-06,
"loss": 0.58430662,
"memory(GiB)": 48.58,
"step": 4110,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.703581039122335,
"grad_norm": 1.8888155653077559,
"learning_rate": 1.528281750266809e-06,
"loss": 0.61480141,
"memory(GiB)": 48.58,
"step": 4115,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.7056510039329331,
"grad_norm": 2.1453583525948567,
"learning_rate": 1.5176093916755604e-06,
"loss": 0.60317545,
"memory(GiB)": 48.58,
"step": 4120,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.7077209687435313,
"grad_norm": 2.0869705530610174,
"learning_rate": 1.5069370330843117e-06,
"loss": 0.57888694,
"memory(GiB)": 48.58,
"step": 4125,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.7097909335541295,
"grad_norm": 2.0589261216950177,
"learning_rate": 1.4962646744930631e-06,
"loss": 0.63887987,
"memory(GiB)": 48.58,
"step": 4130,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.7118608983647277,
"grad_norm": 2.246554976018598,
"learning_rate": 1.4855923159018144e-06,
"loss": 0.60080147,
"memory(GiB)": 48.58,
"step": 4135,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.713930863175326,
"grad_norm": 2.2465140455191377,
"learning_rate": 1.4749199573105658e-06,
"loss": 0.62593145,
"memory(GiB)": 48.58,
"step": 4140,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.7160008279859242,
"grad_norm": 2.1975792886927135,
"learning_rate": 1.4642475987193173e-06,
"loss": 0.61860814,
"memory(GiB)": 48.58,
"step": 4145,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.7180707927965224,
"grad_norm": 1.8897839022485312,
"learning_rate": 1.4535752401280683e-06,
"loss": 0.62411423,
"memory(GiB)": 48.58,
"step": 4150,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.7201407576071208,
"grad_norm": 2.4140521953157794,
"learning_rate": 1.4429028815368198e-06,
"loss": 0.60574412,
"memory(GiB)": 48.58,
"step": 4155,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.722210722417719,
"grad_norm": 2.04528954072566,
"learning_rate": 1.432230522945571e-06,
"loss": 0.61065197,
"memory(GiB)": 48.58,
"step": 4160,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.7242806872283172,
"grad_norm": 2.544468455409069,
"learning_rate": 1.4215581643543225e-06,
"loss": 0.59008269,
"memory(GiB)": 48.58,
"step": 4165,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.7263506520389154,
"grad_norm": 1.9933726482621115,
"learning_rate": 1.4108858057630738e-06,
"loss": 0.61828232,
"memory(GiB)": 48.58,
"step": 4170,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.7284206168495135,
"grad_norm": 2.0938798089462702,
"learning_rate": 1.4002134471718252e-06,
"loss": 0.63251686,
"memory(GiB)": 48.58,
"step": 4175,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.7304905816601117,
"grad_norm": 2.586866443962492,
"learning_rate": 1.3895410885805763e-06,
"loss": 0.61843009,
"memory(GiB)": 48.58,
"step": 4180,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.73256054647071,
"grad_norm": 2.157748885604068,
"learning_rate": 1.3788687299893277e-06,
"loss": 0.61311278,
"memory(GiB)": 48.58,
"step": 4185,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.734630511281308,
"grad_norm": 2.1925495515847944,
"learning_rate": 1.3681963713980792e-06,
"loss": 0.61852412,
"memory(GiB)": 48.58,
"step": 4190,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.7367004760919065,
"grad_norm": 2.0519214102963566,
"learning_rate": 1.3575240128068304e-06,
"loss": 0.60840869,
"memory(GiB)": 48.58,
"step": 4195,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.7387704409025047,
"grad_norm": 2.1138169542373335,
"learning_rate": 1.3468516542155819e-06,
"loss": 0.60299668,
"memory(GiB)": 48.58,
"step": 4200,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.7387704409025047,
"eval_loss": 0.7988836765289307,
"eval_runtime": 333.377,
"eval_samples_per_second": 18.736,
"eval_steps_per_second": 1.173,
"step": 4200
},
{
"epoch": 1.7408404057131028,
"grad_norm": 1.8176186505765208,
"learning_rate": 1.336179295624333e-06,
"loss": 0.59205551,
"memory(GiB)": 48.58,
"step": 4205,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.7429103705237012,
"grad_norm": 2.097301273083197,
"learning_rate": 1.3255069370330844e-06,
"loss": 0.59544134,
"memory(GiB)": 48.58,
"step": 4210,
"train_speed(iter/s)": 0.020007
},
{
"epoch": 1.7449803353342994,
"grad_norm": 2.1754814306718093,
"learning_rate": 1.3148345784418356e-06,
"loss": 0.61374321,
"memory(GiB)": 48.58,
"step": 4215,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.7470503001448976,
"grad_norm": 2.1091284971781543,
"learning_rate": 1.304162219850587e-06,
"loss": 0.58602142,
"memory(GiB)": 48.58,
"step": 4220,
"train_speed(iter/s)": 0.020008
},
{
"epoch": 1.7491202649554958,
"grad_norm": 2.1153060430505177,
"learning_rate": 1.2934898612593383e-06,
"loss": 0.59698052,
"memory(GiB)": 48.58,
"step": 4225,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.751190229766094,
"grad_norm": 2.3476551963062193,
"learning_rate": 1.2828175026680898e-06,
"loss": 0.63100615,
"memory(GiB)": 48.58,
"step": 4230,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.7532601945766921,
"grad_norm": 2.1723309903381987,
"learning_rate": 1.2721451440768413e-06,
"loss": 0.61266661,
"memory(GiB)": 48.58,
"step": 4235,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.7553301593872903,
"grad_norm": 2.4855935431363267,
"learning_rate": 1.2614727854855923e-06,
"loss": 0.60998316,
"memory(GiB)": 48.58,
"step": 4240,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.7574001241978885,
"grad_norm": 2.5103101197105233,
"learning_rate": 1.2508004268943438e-06,
"loss": 0.61193109,
"memory(GiB)": 48.58,
"step": 4245,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.759470089008487,
"grad_norm": 2.2253279455991195,
"learning_rate": 1.2401280683030952e-06,
"loss": 0.58802786,
"memory(GiB)": 48.58,
"step": 4250,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.761540053819085,
"grad_norm": 2.412011728154561,
"learning_rate": 1.2294557097118465e-06,
"loss": 0.6127542,
"memory(GiB)": 48.58,
"step": 4255,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.7636100186296833,
"grad_norm": 2.186989181036045,
"learning_rate": 1.2187833511205977e-06,
"loss": 0.59003277,
"memory(GiB)": 48.58,
"step": 4260,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.7656799834402817,
"grad_norm": 2.528633630931418,
"learning_rate": 1.2081109925293492e-06,
"loss": 0.61563702,
"memory(GiB)": 48.58,
"step": 4265,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.7677499482508798,
"grad_norm": 1.9867035049680128,
"learning_rate": 1.1974386339381004e-06,
"loss": 0.60806894,
"memory(GiB)": 48.58,
"step": 4270,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.769819913061478,
"grad_norm": 2.2593003701423755,
"learning_rate": 1.1867662753468517e-06,
"loss": 0.61837912,
"memory(GiB)": 48.58,
"step": 4275,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.7718898778720762,
"grad_norm": 2.4545294029220797,
"learning_rate": 1.176093916755603e-06,
"loss": 0.61774817,
"memory(GiB)": 48.58,
"step": 4280,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.7739598426826744,
"grad_norm": 1.9914997887032238,
"learning_rate": 1.1654215581643544e-06,
"loss": 0.58686681,
"memory(GiB)": 48.58,
"step": 4285,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.7760298074932725,
"grad_norm": 2.0999532307692896,
"learning_rate": 1.1547491995731057e-06,
"loss": 0.60145164,
"memory(GiB)": 48.58,
"step": 4290,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.7780997723038707,
"grad_norm": 2.480798626827661,
"learning_rate": 1.1440768409818571e-06,
"loss": 0.61948671,
"memory(GiB)": 48.58,
"step": 4295,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.780169737114469,
"grad_norm": 2.1964203822462527,
"learning_rate": 1.1334044823906084e-06,
"loss": 0.5903161,
"memory(GiB)": 48.58,
"step": 4300,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.7822397019250673,
"grad_norm": 2.2986740177948914,
"learning_rate": 1.1227321237993598e-06,
"loss": 0.60875359,
"memory(GiB)": 48.58,
"step": 4305,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.7843096667356655,
"grad_norm": 1.994450780247921,
"learning_rate": 1.112059765208111e-06,
"loss": 0.61587105,
"memory(GiB)": 48.58,
"step": 4310,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.7863796315462637,
"grad_norm": 2.096310617308162,
"learning_rate": 1.1013874066168623e-06,
"loss": 0.58633337,
"memory(GiB)": 48.58,
"step": 4315,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.788449596356862,
"grad_norm": 1.9466363269467861,
"learning_rate": 1.0907150480256138e-06,
"loss": 0.60186481,
"memory(GiB)": 48.58,
"step": 4320,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.7905195611674602,
"grad_norm": 2.308498919041305,
"learning_rate": 1.080042689434365e-06,
"loss": 0.59321814,
"memory(GiB)": 48.58,
"step": 4325,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.7925895259780584,
"grad_norm": 2.6206625429667874,
"learning_rate": 1.0693703308431163e-06,
"loss": 0.61461964,
"memory(GiB)": 48.58,
"step": 4330,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.7946594907886566,
"grad_norm": 2.4886244412597094,
"learning_rate": 1.0586979722518678e-06,
"loss": 0.61646094,
"memory(GiB)": 48.58,
"step": 4335,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.7967294555992548,
"grad_norm": 1.9987554252944004,
"learning_rate": 1.0480256136606192e-06,
"loss": 0.60513401,
"memory(GiB)": 48.58,
"step": 4340,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.798799420409853,
"grad_norm": 2.909678280995542,
"learning_rate": 1.0373532550693705e-06,
"loss": 0.58900928,
"memory(GiB)": 48.58,
"step": 4345,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.8008693852204511,
"grad_norm": 2.232043207476952,
"learning_rate": 1.0266808964781217e-06,
"loss": 0.61423898,
"memory(GiB)": 48.58,
"step": 4350,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.8029393500310493,
"grad_norm": 1.9536032593654695,
"learning_rate": 1.0160085378868732e-06,
"loss": 0.61663337,
"memory(GiB)": 48.58,
"step": 4355,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.8050093148416477,
"grad_norm": 2.5224030653638048,
"learning_rate": 1.0053361792956244e-06,
"loss": 0.61187458,
"memory(GiB)": 48.58,
"step": 4360,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.807079279652246,
"grad_norm": 1.7925531365832896,
"learning_rate": 9.946638207043757e-07,
"loss": 0.59958668,
"memory(GiB)": 48.58,
"step": 4365,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.809149244462844,
"grad_norm": 2.000073758007796,
"learning_rate": 9.839914621131271e-07,
"loss": 0.62268171,
"memory(GiB)": 48.58,
"step": 4370,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.8112192092734425,
"grad_norm": 1.8927383998351053,
"learning_rate": 9.733191035218784e-07,
"loss": 0.61844292,
"memory(GiB)": 48.58,
"step": 4375,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.8132891740840407,
"grad_norm": 2.2603010222528708,
"learning_rate": 9.626467449306296e-07,
"loss": 0.60280704,
"memory(GiB)": 48.58,
"step": 4380,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.8153591388946388,
"grad_norm": 2.0743626326418494,
"learning_rate": 9.519743863393811e-07,
"loss": 0.59184837,
"memory(GiB)": 48.58,
"step": 4385,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.817429103705237,
"grad_norm": 2.5328803417607273,
"learning_rate": 9.413020277481325e-07,
"loss": 0.60617228,
"memory(GiB)": 48.58,
"step": 4390,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.8194990685158352,
"grad_norm": 2.157648955819526,
"learning_rate": 9.306296691568837e-07,
"loss": 0.57980437,
"memory(GiB)": 48.58,
"step": 4395,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.8215690333264334,
"grad_norm": 2.238654145184739,
"learning_rate": 9.199573105656351e-07,
"loss": 0.59073811,
"memory(GiB)": 48.58,
"step": 4400,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.8236389981370316,
"grad_norm": 2.1560196555189677,
"learning_rate": 9.092849519743864e-07,
"loss": 0.60031624,
"memory(GiB)": 48.58,
"step": 4405,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.8257089629476297,
"grad_norm": 1.9778064514427567,
"learning_rate": 8.986125933831377e-07,
"loss": 0.59476948,
"memory(GiB)": 48.58,
"step": 4410,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.8277789277582281,
"grad_norm": 1.9062652252413357,
"learning_rate": 8.87940234791889e-07,
"loss": 0.60492353,
"memory(GiB)": 48.58,
"step": 4415,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.8298488925688263,
"grad_norm": 2.20251758356039,
"learning_rate": 8.772678762006404e-07,
"loss": 0.57913284,
"memory(GiB)": 48.58,
"step": 4420,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.8319188573794245,
"grad_norm": 2.374670539942745,
"learning_rate": 8.665955176093919e-07,
"loss": 0.61436529,
"memory(GiB)": 48.58,
"step": 4425,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.833988822190023,
"grad_norm": 1.8014659618328237,
"learning_rate": 8.559231590181431e-07,
"loss": 0.58954339,
"memory(GiB)": 48.58,
"step": 4430,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.836058787000621,
"grad_norm": 2.5328263785548706,
"learning_rate": 8.452508004268945e-07,
"loss": 0.62326274,
"memory(GiB)": 48.58,
"step": 4435,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.8381287518112193,
"grad_norm": 2.201541673623922,
"learning_rate": 8.345784418356458e-07,
"loss": 0.61668777,
"memory(GiB)": 48.58,
"step": 4440,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.8401987166218174,
"grad_norm": 1.9916726147288757,
"learning_rate": 8.239060832443971e-07,
"loss": 0.60296612,
"memory(GiB)": 48.58,
"step": 4445,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.8422686814324156,
"grad_norm": 2.2611511913260167,
"learning_rate": 8.132337246531484e-07,
"loss": 0.59926748,
"memory(GiB)": 48.58,
"step": 4450,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.8443386462430138,
"grad_norm": 1.8225104361277575,
"learning_rate": 8.025613660618997e-07,
"loss": 0.58590517,
"memory(GiB)": 48.58,
"step": 4455,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.846408611053612,
"grad_norm": 2.350449897326385,
"learning_rate": 7.91889007470651e-07,
"loss": 0.60878654,
"memory(GiB)": 48.58,
"step": 4460,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.8484785758642102,
"grad_norm": 2.4294512189314075,
"learning_rate": 7.812166488794024e-07,
"loss": 0.60740719,
"memory(GiB)": 48.58,
"step": 4465,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.8505485406748086,
"grad_norm": 2.0078565564513413,
"learning_rate": 7.705442902881538e-07,
"loss": 0.61569843,
"memory(GiB)": 48.58,
"step": 4470,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.8526185054854067,
"grad_norm": 2.416347350001525,
"learning_rate": 7.598719316969051e-07,
"loss": 0.59907522,
"memory(GiB)": 48.58,
"step": 4475,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.854688470296005,
"grad_norm": 2.5546773910049523,
"learning_rate": 7.491995731056565e-07,
"loss": 0.58857327,
"memory(GiB)": 48.58,
"step": 4480,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.8567584351066033,
"grad_norm": 2.1560857096799775,
"learning_rate": 7.385272145144078e-07,
"loss": 0.59972405,
"memory(GiB)": 48.58,
"step": 4485,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.8588283999172015,
"grad_norm": 2.8045450674344155,
"learning_rate": 7.278548559231591e-07,
"loss": 0.62519865,
"memory(GiB)": 48.58,
"step": 4490,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.8608983647277997,
"grad_norm": 1.88470841271018,
"learning_rate": 7.171824973319104e-07,
"loss": 0.59394321,
"memory(GiB)": 48.58,
"step": 4495,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.8629683295383979,
"grad_norm": 1.900323108929672,
"learning_rate": 7.065101387406617e-07,
"loss": 0.59171925,
"memory(GiB)": 48.58,
"step": 4500,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.8629683295383979,
"eval_loss": 0.7995001077651978,
"eval_runtime": 333.5022,
"eval_samples_per_second": 18.729,
"eval_steps_per_second": 1.172,
"step": 4500
},
{
"epoch": 1.865038294348996,
"grad_norm": 2.3418380686910854,
"learning_rate": 6.95837780149413e-07,
"loss": 0.62550597,
"memory(GiB)": 48.58,
"step": 4505,
"train_speed(iter/s)": 0.020009
},
{
"epoch": 1.8671082591595942,
"grad_norm": 2.2011802954910826,
"learning_rate": 6.851654215581644e-07,
"loss": 0.60017891,
"memory(GiB)": 48.58,
"step": 4510,
"train_speed(iter/s)": 0.02001
},
{
"epoch": 1.8691782239701924,
"grad_norm": 2.346777967941492,
"learning_rate": 6.744930629669158e-07,
"loss": 0.62277446,
"memory(GiB)": 48.58,
"step": 4515,
"train_speed(iter/s)": 0.020011
},
{
"epoch": 1.8712481887807906,
"grad_norm": 2.3039363321887385,
"learning_rate": 6.638207043756671e-07,
"loss": 0.63189201,
"memory(GiB)": 48.58,
"step": 4520,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.873318153591389,
"grad_norm": 2.4254296862394753,
"learning_rate": 6.531483457844184e-07,
"loss": 0.59847736,
"memory(GiB)": 48.58,
"step": 4525,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.8753881184019872,
"grad_norm": 2.520645231015452,
"learning_rate": 6.424759871931698e-07,
"loss": 0.60772429,
"memory(GiB)": 48.58,
"step": 4530,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.8774580832125853,
"grad_norm": 2.14246492812266,
"learning_rate": 6.31803628601921e-07,
"loss": 0.61204777,
"memory(GiB)": 48.58,
"step": 4535,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.8795280480231837,
"grad_norm": 1.9822250514442827,
"learning_rate": 6.211312700106724e-07,
"loss": 0.55947261,
"memory(GiB)": 48.58,
"step": 4540,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.881598012833782,
"grad_norm": 2.052778758307509,
"learning_rate": 6.104589114194238e-07,
"loss": 0.60927758,
"memory(GiB)": 48.58,
"step": 4545,
"train_speed(iter/s)": 0.020015
},
{
"epoch": 1.88366797764438,
"grad_norm": 2.3908992672472222,
"learning_rate": 5.997865528281751e-07,
"loss": 0.61153603,
"memory(GiB)": 48.58,
"step": 4550,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.8857379424549783,
"grad_norm": 1.9114509697538704,
"learning_rate": 5.891141942369264e-07,
"loss": 0.6099647,
"memory(GiB)": 48.58,
"step": 4555,
"train_speed(iter/s)": 0.020016
},
{
"epoch": 1.8878079072655765,
"grad_norm": 2.063913038194293,
"learning_rate": 5.784418356456777e-07,
"loss": 0.61249609,
"memory(GiB)": 48.58,
"step": 4560,
"train_speed(iter/s)": 0.020017
},
{
"epoch": 1.8898778720761746,
"grad_norm": 2.3274588748356404,
"learning_rate": 5.677694770544291e-07,
"loss": 0.60335112,
"memory(GiB)": 48.58,
"step": 4565,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.8919478368867728,
"grad_norm": 2.5803919299014404,
"learning_rate": 5.570971184631804e-07,
"loss": 0.59715414,
"memory(GiB)": 48.58,
"step": 4570,
"train_speed(iter/s)": 0.020018
},
{
"epoch": 1.894017801697371,
"grad_norm": 2.0982349697996727,
"learning_rate": 5.464247598719318e-07,
"loss": 0.59020104,
"memory(GiB)": 48.58,
"step": 4575,
"train_speed(iter/s)": 0.020019
},
{
"epoch": 1.8960877665079694,
"grad_norm": 2.4156310349424492,
"learning_rate": 5.35752401280683e-07,
"loss": 0.58418913,
"memory(GiB)": 48.58,
"step": 4580,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.8981577313185676,
"grad_norm": 2.1192384982329977,
"learning_rate": 5.250800426894344e-07,
"loss": 0.57976351,
"memory(GiB)": 48.58,
"step": 4585,
"train_speed(iter/s)": 0.02002
},
{
"epoch": 1.9002276961291658,
"grad_norm": 2.363144438078418,
"learning_rate": 5.144076840981858e-07,
"loss": 0.59181528,
"memory(GiB)": 48.58,
"step": 4590,
"train_speed(iter/s)": 0.020021
},
{
"epoch": 1.9022976609397642,
"grad_norm": 2.014916601780685,
"learning_rate": 5.037353255069371e-07,
"loss": 0.57993307,
"memory(GiB)": 48.58,
"step": 4595,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.9043676257503623,
"grad_norm": 2.1798732452842615,
"learning_rate": 4.930629669156884e-07,
"loss": 0.61651163,
"memory(GiB)": 48.58,
"step": 4600,
"train_speed(iter/s)": 0.020022
},
{
"epoch": 1.9064375905609605,
"grad_norm": 2.4067029242954847,
"learning_rate": 4.823906083244397e-07,
"loss": 0.60284195,
"memory(GiB)": 48.58,
"step": 4605,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.9085075553715587,
"grad_norm": 2.1464043153264587,
"learning_rate": 4.7171824973319113e-07,
"loss": 0.58217282,
"memory(GiB)": 48.58,
"step": 4610,
"train_speed(iter/s)": 0.020023
},
{
"epoch": 1.9105775201821569,
"grad_norm": 2.786031828765745,
"learning_rate": 4.6104589114194243e-07,
"loss": 0.59245019,
"memory(GiB)": 48.58,
"step": 4615,
"train_speed(iter/s)": 0.020024
},
{
"epoch": 1.912647484992755,
"grad_norm": 1.971002889872401,
"learning_rate": 4.5037353255069374e-07,
"loss": 0.55735373,
"memory(GiB)": 48.58,
"step": 4620,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.9147174498033532,
"grad_norm": 2.6133112222197097,
"learning_rate": 4.3970117395944504e-07,
"loss": 0.58715906,
"memory(GiB)": 48.58,
"step": 4625,
"train_speed(iter/s)": 0.020025
},
{
"epoch": 1.9167874146139514,
"grad_norm": 2.2493774425876496,
"learning_rate": 4.290288153681964e-07,
"loss": 0.61737943,
"memory(GiB)": 48.58,
"step": 4630,
"train_speed(iter/s)": 0.020026
},
{
"epoch": 1.9188573794245498,
"grad_norm": 2.066157101262181,
"learning_rate": 4.1835645677694775e-07,
"loss": 0.58552856,
"memory(GiB)": 48.58,
"step": 4635,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.920927344235148,
"grad_norm": 2.2279403099524164,
"learning_rate": 4.076840981856991e-07,
"loss": 0.60430613,
"memory(GiB)": 48.58,
"step": 4640,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.9229973090457462,
"grad_norm": 2.21706254471347,
"learning_rate": 3.970117395944504e-07,
"loss": 0.60209589,
"memory(GiB)": 48.58,
"step": 4645,
"train_speed(iter/s)": 0.020027
},
{
"epoch": 1.9250672738563446,
"grad_norm": 1.8175472634449323,
"learning_rate": 3.863393810032017e-07,
"loss": 0.57723808,
"memory(GiB)": 48.58,
"step": 4650,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.9271372386669428,
"grad_norm": 2.013736958312471,
"learning_rate": 3.756670224119531e-07,
"loss": 0.59187717,
"memory(GiB)": 48.58,
"step": 4655,
"train_speed(iter/s)": 0.020028
},
{
"epoch": 1.929207203477541,
"grad_norm": 2.9147581447233883,
"learning_rate": 3.649946638207044e-07,
"loss": 0.62887087,
"memory(GiB)": 48.58,
"step": 4660,
"train_speed(iter/s)": 0.020029
},
{
"epoch": 1.9312771682881391,
"grad_norm": 2.1113909828567605,
"learning_rate": 3.5432230522945573e-07,
"loss": 0.60369582,
"memory(GiB)": 48.58,
"step": 4665,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.9333471330987373,
"grad_norm": 2.42442448807154,
"learning_rate": 3.4364994663820703e-07,
"loss": 0.62186384,
"memory(GiB)": 48.58,
"step": 4670,
"train_speed(iter/s)": 0.02003
},
{
"epoch": 1.9354170979093355,
"grad_norm": 2.380814562965916,
"learning_rate": 3.329775880469584e-07,
"loss": 0.59275131,
"memory(GiB)": 48.58,
"step": 4675,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.9374870627199337,
"grad_norm": 2.6584440377132363,
"learning_rate": 3.2230522945570974e-07,
"loss": 0.60647793,
"memory(GiB)": 48.58,
"step": 4680,
"train_speed(iter/s)": 0.020031
},
{
"epoch": 1.9395570275305318,
"grad_norm": 1.9567852417003078,
"learning_rate": 3.116328708644611e-07,
"loss": 0.60262537,
"memory(GiB)": 48.58,
"step": 4685,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.9416269923411302,
"grad_norm": 2.3501175824898266,
"learning_rate": 3.009605122732124e-07,
"loss": 0.58797078,
"memory(GiB)": 48.58,
"step": 4690,
"train_speed(iter/s)": 0.020032
},
{
"epoch": 1.9436969571517284,
"grad_norm": 2.237666059037871,
"learning_rate": 2.9028815368196376e-07,
"loss": 0.60981102,
"memory(GiB)": 48.58,
"step": 4695,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.9457669219623266,
"grad_norm": 2.4233069808853993,
"learning_rate": 2.7961579509071506e-07,
"loss": 0.60726156,
"memory(GiB)": 48.58,
"step": 4700,
"train_speed(iter/s)": 0.020033
},
{
"epoch": 1.947836886772925,
"grad_norm": 2.260763775980905,
"learning_rate": 2.689434364994664e-07,
"loss": 0.60588284,
"memory(GiB)": 48.58,
"step": 4705,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.9499068515835232,
"grad_norm": 2.5081221538985625,
"learning_rate": 2.582710779082177e-07,
"loss": 0.59843764,
"memory(GiB)": 48.58,
"step": 4710,
"train_speed(iter/s)": 0.020034
},
{
"epoch": 1.9519768163941214,
"grad_norm": 2.3881113527972304,
"learning_rate": 2.475987193169691e-07,
"loss": 0.58006935,
"memory(GiB)": 48.58,
"step": 4715,
"train_speed(iter/s)": 0.020035
},
{
"epoch": 1.9540467812047195,
"grad_norm": 2.0678535427683564,
"learning_rate": 2.369263607257204e-07,
"loss": 0.60157442,
"memory(GiB)": 48.58,
"step": 4720,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.9561167460153177,
"grad_norm": 2.438506236548372,
"learning_rate": 2.2625400213447176e-07,
"loss": 0.61028309,
"memory(GiB)": 48.58,
"step": 4725,
"train_speed(iter/s)": 0.020036
},
{
"epoch": 1.9581867108259159,
"grad_norm": 2.0110599859655482,
"learning_rate": 2.1558164354322307e-07,
"loss": 0.59143724,
"memory(GiB)": 48.58,
"step": 4730,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.960256675636514,
"grad_norm": 2.0688837840068124,
"learning_rate": 2.049092849519744e-07,
"loss": 0.60122604,
"memory(GiB)": 48.58,
"step": 4735,
"train_speed(iter/s)": 0.020037
},
{
"epoch": 1.9623266404471122,
"grad_norm": 2.1607611011042964,
"learning_rate": 1.9423692636072575e-07,
"loss": 0.59857554,
"memory(GiB)": 48.58,
"step": 4740,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.9643966052577106,
"grad_norm": 1.9355873743150782,
"learning_rate": 1.8356456776947706e-07,
"loss": 0.58816404,
"memory(GiB)": 48.58,
"step": 4745,
"train_speed(iter/s)": 0.020038
},
{
"epoch": 1.9664665700683088,
"grad_norm": 2.1028560489657915,
"learning_rate": 1.728922091782284e-07,
"loss": 0.59075899,
"memory(GiB)": 48.58,
"step": 4750,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.968536534878907,
"grad_norm": 1.9716163219927025,
"learning_rate": 1.6221985058697972e-07,
"loss": 0.58181,
"memory(GiB)": 48.58,
"step": 4755,
"train_speed(iter/s)": 0.020039
},
{
"epoch": 1.9706064996895054,
"grad_norm": 2.243337925247892,
"learning_rate": 1.5154749199573107e-07,
"loss": 0.58783703,
"memory(GiB)": 48.58,
"step": 4760,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.9726764645001036,
"grad_norm": 2.4859440712948166,
"learning_rate": 1.408751334044824e-07,
"loss": 0.60655708,
"memory(GiB)": 48.58,
"step": 4765,
"train_speed(iter/s)": 0.02004
},
{
"epoch": 1.9747464293107018,
"grad_norm": 1.9831273691126385,
"learning_rate": 1.3020277481323373e-07,
"loss": 0.61188507,
"memory(GiB)": 48.58,
"step": 4770,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.9768163941213,
"grad_norm": 2.2314049587449807,
"learning_rate": 1.1953041622198506e-07,
"loss": 0.58718634,
"memory(GiB)": 48.58,
"step": 4775,
"train_speed(iter/s)": 0.020041
},
{
"epoch": 1.9788863589318981,
"grad_norm": 1.8150041254780722,
"learning_rate": 1.088580576307364e-07,
"loss": 0.58888893,
"memory(GiB)": 48.58,
"step": 4780,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.9809563237424963,
"grad_norm": 2.247396093692947,
"learning_rate": 9.818569903948773e-08,
"loss": 0.58732767,
"memory(GiB)": 48.58,
"step": 4785,
"train_speed(iter/s)": 0.020042
},
{
"epoch": 1.9830262885530945,
"grad_norm": 2.2098523979831644,
"learning_rate": 8.751334044823908e-08,
"loss": 0.56731772,
"memory(GiB)": 48.58,
"step": 4790,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.9850962533636927,
"grad_norm": 2.097191818370972,
"learning_rate": 7.68409818569904e-08,
"loss": 0.60409393,
"memory(GiB)": 48.58,
"step": 4795,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.987166218174291,
"grad_norm": 2.228865758940409,
"learning_rate": 6.616862326574174e-08,
"loss": 0.58583736,
"memory(GiB)": 48.58,
"step": 4800,
"train_speed(iter/s)": 0.020043
},
{
"epoch": 1.987166218174291,
"eval_loss": 0.7990086674690247,
"eval_runtime": 335.524,
"eval_samples_per_second": 18.616,
"eval_steps_per_second": 1.165,
"step": 4800
},
{
"epoch": 1.9892361829848892,
"grad_norm": 2.144126117644109,
"learning_rate": 5.5496264674493065e-08,
"loss": 0.59295273,
"memory(GiB)": 48.58,
"step": 4805,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.9913061477954874,
"grad_norm": 2.020692421672203,
"learning_rate": 4.48239060832444e-08,
"loss": 0.58215327,
"memory(GiB)": 48.58,
"step": 4810,
"train_speed(iter/s)": 0.020012
},
{
"epoch": 1.9933761126060858,
"grad_norm": 2.560575104653987,
"learning_rate": 3.415154749199574e-08,
"loss": 0.59236603,
"memory(GiB)": 48.58,
"step": 4815,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.995446077416684,
"grad_norm": 2.242799455811118,
"learning_rate": 2.347918890074707e-08,
"loss": 0.59284697,
"memory(GiB)": 48.58,
"step": 4820,
"train_speed(iter/s)": 0.020013
},
{
"epoch": 1.9975160422272822,
"grad_norm": 2.496931570288639,
"learning_rate": 1.28068303094984e-08,
"loss": 0.58884125,
"memory(GiB)": 48.58,
"step": 4825,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.9995860070378804,
"grad_norm": 2.1504541545847133,
"learning_rate": 2.1344717182497336e-09,
"loss": 0.57651815,
"memory(GiB)": 48.58,
"step": 4830,
"train_speed(iter/s)": 0.020014
},
{
"epoch": 1.9995860070378804,
"eval_loss": 0.7990483641624451,
"eval_runtime": 333.2574,
"eval_samples_per_second": 18.742,
"eval_steps_per_second": 1.173,
"step": 4830
}
],
"logging_steps": 5,
"max_steps": 4830,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 300,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0363792423256064e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}