NanoTranslator-S / trainer_state.json
Mxode's picture
update
d74e240
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 9676,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0010334849111202976,
"grad_norm": 3.4666478633880615,
"learning_rate": 0.0006200909466721785,
"loss": 7.9994,
"step": 10
},
{
"epoch": 0.002066969822240595,
"grad_norm": 1.0015512704849243,
"learning_rate": 0.001240181893344357,
"loss": 7.0943,
"step": 20
},
{
"epoch": 0.0031004547333608927,
"grad_norm": 0.8696095943450928,
"learning_rate": 0.0018602728400165356,
"loss": 6.3072,
"step": 30
},
{
"epoch": 0.00413393964448119,
"grad_norm": 0.32349398732185364,
"learning_rate": 0.002480363786688714,
"loss": 5.9694,
"step": 40
},
{
"epoch": 0.005167424555601488,
"grad_norm": 1.6740758419036865,
"learning_rate": 0.0031004547333608927,
"loss": 5.8536,
"step": 50
},
{
"epoch": 0.0062009094667217855,
"grad_norm": 1.6459004878997803,
"learning_rate": 0.003720545680033071,
"loss": 5.7039,
"step": 60
},
{
"epoch": 0.007234394377842084,
"grad_norm": 0.48763737082481384,
"learning_rate": 0.0043406366267052495,
"loss": 5.4789,
"step": 70
},
{
"epoch": 0.00826787928896238,
"grad_norm": 0.4977681338787079,
"learning_rate": 0.004960727573377428,
"loss": 5.269,
"step": 80
},
{
"epoch": 0.009301364200082679,
"grad_norm": 0.5395390391349792,
"learning_rate": 0.005580818520049607,
"loss": 5.0768,
"step": 90
},
{
"epoch": 0.010334849111202976,
"grad_norm": 0.4573260247707367,
"learning_rate": 0.006,
"loss": 4.9493,
"step": 100
},
{
"epoch": 0.011368334022323274,
"grad_norm": 0.5551472902297974,
"learning_rate": 0.006,
"loss": 4.7982,
"step": 110
},
{
"epoch": 0.012401818933443571,
"grad_norm": 0.4980376958847046,
"learning_rate": 0.006,
"loss": 4.6285,
"step": 120
},
{
"epoch": 0.01343530384456387,
"grad_norm": 0.4951136112213135,
"learning_rate": 0.006,
"loss": 4.5176,
"step": 130
},
{
"epoch": 0.014468788755684168,
"grad_norm": 0.5360985398292542,
"learning_rate": 0.006,
"loss": 4.3572,
"step": 140
},
{
"epoch": 0.015502273666804465,
"grad_norm": 0.4469200670719147,
"learning_rate": 0.006,
"loss": 4.214,
"step": 150
},
{
"epoch": 0.01653575857792476,
"grad_norm": 0.6157914400100708,
"learning_rate": 0.006,
"loss": 4.0854,
"step": 160
},
{
"epoch": 0.01756924348904506,
"grad_norm": 0.4493155777454376,
"learning_rate": 0.006,
"loss": 3.9583,
"step": 170
},
{
"epoch": 0.018602728400165358,
"grad_norm": 0.6523879766464233,
"learning_rate": 0.006,
"loss": 3.8434,
"step": 180
},
{
"epoch": 0.019636213311285657,
"grad_norm": 0.7103180885314941,
"learning_rate": 0.006,
"loss": 3.7414,
"step": 190
},
{
"epoch": 0.02066969822240595,
"grad_norm": 0.44575124979019165,
"learning_rate": 0.006,
"loss": 3.6829,
"step": 200
},
{
"epoch": 0.02170318313352625,
"grad_norm": 0.5272648334503174,
"learning_rate": 0.006,
"loss": 3.6056,
"step": 210
},
{
"epoch": 0.02273666804464655,
"grad_norm": 0.4514292776584625,
"learning_rate": 0.006,
"loss": 3.5279,
"step": 220
},
{
"epoch": 0.023770152955766847,
"grad_norm": 0.39017120003700256,
"learning_rate": 0.006,
"loss": 3.4585,
"step": 230
},
{
"epoch": 0.024803637866887142,
"grad_norm": 0.4593961536884308,
"learning_rate": 0.006,
"loss": 3.4018,
"step": 240
},
{
"epoch": 0.02583712277800744,
"grad_norm": 0.48860013484954834,
"learning_rate": 0.006,
"loss": 3.3523,
"step": 250
},
{
"epoch": 0.02687060768912774,
"grad_norm": 0.36506471037864685,
"learning_rate": 0.006,
"loss": 3.3124,
"step": 260
},
{
"epoch": 0.027904092600248037,
"grad_norm": 0.3506007194519043,
"learning_rate": 0.006,
"loss": 3.2678,
"step": 270
},
{
"epoch": 0.028937577511368336,
"grad_norm": 0.3703238070011139,
"learning_rate": 0.006,
"loss": 3.2231,
"step": 280
},
{
"epoch": 0.02997106242248863,
"grad_norm": 0.45603325963020325,
"learning_rate": 0.006,
"loss": 3.1847,
"step": 290
},
{
"epoch": 0.03100454733360893,
"grad_norm": 0.37536856532096863,
"learning_rate": 0.006,
"loss": 3.1585,
"step": 300
},
{
"epoch": 0.032038032244729224,
"grad_norm": 0.3810868561267853,
"learning_rate": 0.006,
"loss": 3.1063,
"step": 310
},
{
"epoch": 0.03307151715584952,
"grad_norm": 0.4568467438220978,
"learning_rate": 0.006,
"loss": 3.0649,
"step": 320
},
{
"epoch": 0.03410500206696982,
"grad_norm": 0.36088162660598755,
"learning_rate": 0.006,
"loss": 3.0495,
"step": 330
},
{
"epoch": 0.03513848697809012,
"grad_norm": 0.3663162887096405,
"learning_rate": 0.006,
"loss": 3.0188,
"step": 340
},
{
"epoch": 0.03617197188921042,
"grad_norm": 0.5975127220153809,
"learning_rate": 0.006,
"loss": 2.994,
"step": 350
},
{
"epoch": 0.037205456800330716,
"grad_norm": 0.3454865515232086,
"learning_rate": 0.006,
"loss": 2.9814,
"step": 360
},
{
"epoch": 0.038238941711451015,
"grad_norm": 0.4206918179988861,
"learning_rate": 0.006,
"loss": 2.9306,
"step": 370
},
{
"epoch": 0.03927242662257131,
"grad_norm": 0.3579295575618744,
"learning_rate": 0.006,
"loss": 2.9165,
"step": 380
},
{
"epoch": 0.04030591153369161,
"grad_norm": 0.3658403158187866,
"learning_rate": 0.006,
"loss": 2.9103,
"step": 390
},
{
"epoch": 0.0413393964448119,
"grad_norm": 0.3459131717681885,
"learning_rate": 0.006,
"loss": 2.8628,
"step": 400
},
{
"epoch": 0.0423728813559322,
"grad_norm": 0.4726301431655884,
"learning_rate": 0.006,
"loss": 2.8334,
"step": 410
},
{
"epoch": 0.0434063662670525,
"grad_norm": 0.3717091381549835,
"learning_rate": 0.006,
"loss": 2.8188,
"step": 420
},
{
"epoch": 0.0444398511781728,
"grad_norm": 0.3234303295612335,
"learning_rate": 0.006,
"loss": 2.8076,
"step": 430
},
{
"epoch": 0.0454733360892931,
"grad_norm": 0.4717100262641907,
"learning_rate": 0.006,
"loss": 2.7871,
"step": 440
},
{
"epoch": 0.046506821000413395,
"grad_norm": 0.35768577456474304,
"learning_rate": 0.006,
"loss": 2.7601,
"step": 450
},
{
"epoch": 0.047540305911533694,
"grad_norm": 0.3499695062637329,
"learning_rate": 0.006,
"loss": 2.73,
"step": 460
},
{
"epoch": 0.04857379082265399,
"grad_norm": 0.4336409866809845,
"learning_rate": 0.006,
"loss": 2.7235,
"step": 470
},
{
"epoch": 0.049607275733774284,
"grad_norm": 0.39916467666625977,
"learning_rate": 0.006,
"loss": 2.7007,
"step": 480
},
{
"epoch": 0.05064076064489458,
"grad_norm": 0.3249908983707428,
"learning_rate": 0.006,
"loss": 2.6705,
"step": 490
},
{
"epoch": 0.05167424555601488,
"grad_norm": 0.36379745602607727,
"learning_rate": 0.006,
"loss": 2.6393,
"step": 500
},
{
"epoch": 0.05270773046713518,
"grad_norm": 0.3417607843875885,
"learning_rate": 0.006,
"loss": 2.6767,
"step": 510
},
{
"epoch": 0.05374121537825548,
"grad_norm": 0.2848694920539856,
"learning_rate": 0.006,
"loss": 2.6244,
"step": 520
},
{
"epoch": 0.054774700289375776,
"grad_norm": 0.2725778818130493,
"learning_rate": 0.006,
"loss": 2.5704,
"step": 530
},
{
"epoch": 0.055808185200496074,
"grad_norm": 0.2887585461139679,
"learning_rate": 0.006,
"loss": 2.5508,
"step": 540
},
{
"epoch": 0.05684167011161637,
"grad_norm": 0.422413170337677,
"learning_rate": 0.006,
"loss": 2.5433,
"step": 550
},
{
"epoch": 0.05787515502273667,
"grad_norm": 0.5642419457435608,
"learning_rate": 0.006,
"loss": 2.5267,
"step": 560
},
{
"epoch": 0.05890863993385696,
"grad_norm": 0.33949220180511475,
"learning_rate": 0.006,
"loss": 2.4942,
"step": 570
},
{
"epoch": 0.05994212484497726,
"grad_norm": 0.36603158712387085,
"learning_rate": 0.006,
"loss": 2.4755,
"step": 580
},
{
"epoch": 0.06097560975609756,
"grad_norm": 0.3936167061328888,
"learning_rate": 0.006,
"loss": 2.4358,
"step": 590
},
{
"epoch": 0.06200909466721786,
"grad_norm": 0.46108072996139526,
"learning_rate": 0.006,
"loss": 2.4312,
"step": 600
},
{
"epoch": 0.06304257957833816,
"grad_norm": 0.3659123480319977,
"learning_rate": 0.006,
"loss": 2.4023,
"step": 610
},
{
"epoch": 0.06407606448945845,
"grad_norm": 0.4888301193714142,
"learning_rate": 0.006,
"loss": 2.3887,
"step": 620
},
{
"epoch": 0.06510954940057875,
"grad_norm": 0.33769381046295166,
"learning_rate": 0.006,
"loss": 2.3645,
"step": 630
},
{
"epoch": 0.06614303431169904,
"grad_norm": 0.4479112923145294,
"learning_rate": 0.006,
"loss": 2.3123,
"step": 640
},
{
"epoch": 0.06717651922281935,
"grad_norm": 0.3715905249118805,
"learning_rate": 0.006,
"loss": 2.3009,
"step": 650
},
{
"epoch": 0.06821000413393964,
"grad_norm": 0.43582767248153687,
"learning_rate": 0.006,
"loss": 2.3256,
"step": 660
},
{
"epoch": 0.06924348904505995,
"grad_norm": 0.295462042093277,
"learning_rate": 0.006,
"loss": 2.281,
"step": 670
},
{
"epoch": 0.07027697395618024,
"grad_norm": 0.2555060386657715,
"learning_rate": 0.006,
"loss": 2.2711,
"step": 680
},
{
"epoch": 0.07131045886730054,
"grad_norm": 0.2812485992908478,
"learning_rate": 0.006,
"loss": 2.2392,
"step": 690
},
{
"epoch": 0.07234394377842084,
"grad_norm": 0.2797698974609375,
"learning_rate": 0.006,
"loss": 2.2316,
"step": 700
},
{
"epoch": 0.07337742868954113,
"grad_norm": 0.4201503396034241,
"learning_rate": 0.006,
"loss": 2.1894,
"step": 710
},
{
"epoch": 0.07441091360066143,
"grad_norm": 0.48455363512039185,
"learning_rate": 0.006,
"loss": 2.1682,
"step": 720
},
{
"epoch": 0.07544439851178172,
"grad_norm": 0.4906412959098816,
"learning_rate": 0.006,
"loss": 2.1594,
"step": 730
},
{
"epoch": 0.07647788342290203,
"grad_norm": 0.3227013051509857,
"learning_rate": 0.006,
"loss": 2.148,
"step": 740
},
{
"epoch": 0.07751136833402232,
"grad_norm": 0.24952304363250732,
"learning_rate": 0.006,
"loss": 2.1323,
"step": 750
},
{
"epoch": 0.07854485324514263,
"grad_norm": 0.4688262641429901,
"learning_rate": 0.006,
"loss": 2.0781,
"step": 760
},
{
"epoch": 0.07957833815626292,
"grad_norm": 0.3051510453224182,
"learning_rate": 0.006,
"loss": 2.1005,
"step": 770
},
{
"epoch": 0.08061182306738322,
"grad_norm": 0.303608775138855,
"learning_rate": 0.006,
"loss": 2.0619,
"step": 780
},
{
"epoch": 0.08164530797850351,
"grad_norm": 0.41049811244010925,
"learning_rate": 0.006,
"loss": 2.069,
"step": 790
},
{
"epoch": 0.0826787928896238,
"grad_norm": 0.25236430764198303,
"learning_rate": 0.006,
"loss": 2.0536,
"step": 800
},
{
"epoch": 0.08371227780074411,
"grad_norm": 0.2919696569442749,
"learning_rate": 0.006,
"loss": 2.0221,
"step": 810
},
{
"epoch": 0.0847457627118644,
"grad_norm": 0.3965113162994385,
"learning_rate": 0.006,
"loss": 2.0008,
"step": 820
},
{
"epoch": 0.08577924762298471,
"grad_norm": 0.3784347176551819,
"learning_rate": 0.006,
"loss": 2.0041,
"step": 830
},
{
"epoch": 0.086812732534105,
"grad_norm": 0.32382646203041077,
"learning_rate": 0.006,
"loss": 1.9987,
"step": 840
},
{
"epoch": 0.0878462174452253,
"grad_norm": 0.28474798798561096,
"learning_rate": 0.006,
"loss": 1.9859,
"step": 850
},
{
"epoch": 0.0888797023563456,
"grad_norm": 0.35125505924224854,
"learning_rate": 0.006,
"loss": 1.9719,
"step": 860
},
{
"epoch": 0.08991318726746589,
"grad_norm": 0.2959253787994385,
"learning_rate": 0.006,
"loss": 1.9691,
"step": 870
},
{
"epoch": 0.0909466721785862,
"grad_norm": 0.2681068181991577,
"learning_rate": 0.006,
"loss": 1.946,
"step": 880
},
{
"epoch": 0.09198015708970649,
"grad_norm": 0.2882836163043976,
"learning_rate": 0.006,
"loss": 1.9227,
"step": 890
},
{
"epoch": 0.09301364200082679,
"grad_norm": 0.2792555093765259,
"learning_rate": 0.006,
"loss": 1.9067,
"step": 900
},
{
"epoch": 0.09404712691194708,
"grad_norm": 0.31801220774650574,
"learning_rate": 0.006,
"loss": 1.8997,
"step": 910
},
{
"epoch": 0.09508061182306739,
"grad_norm": 0.3104709982872009,
"learning_rate": 0.006,
"loss": 1.909,
"step": 920
},
{
"epoch": 0.09611409673418768,
"grad_norm": 0.4227132499217987,
"learning_rate": 0.006,
"loss": 1.8898,
"step": 930
},
{
"epoch": 0.09714758164530798,
"grad_norm": 0.2799486815929413,
"learning_rate": 0.006,
"loss": 1.8799,
"step": 940
},
{
"epoch": 0.09818106655642828,
"grad_norm": 0.25304150581359863,
"learning_rate": 0.006,
"loss": 1.8542,
"step": 950
},
{
"epoch": 0.09921455146754857,
"grad_norm": 0.40398165583610535,
"learning_rate": 0.006,
"loss": 1.8495,
"step": 960
},
{
"epoch": 0.10024803637866887,
"grad_norm": 0.24255605041980743,
"learning_rate": 0.006,
"loss": 1.8279,
"step": 970
},
{
"epoch": 0.10128152128978916,
"grad_norm": 0.251412957906723,
"learning_rate": 0.006,
"loss": 1.8448,
"step": 980
},
{
"epoch": 0.10231500620090947,
"grad_norm": 0.33690279722213745,
"learning_rate": 0.006,
"loss": 1.8345,
"step": 990
},
{
"epoch": 0.10334849111202976,
"grad_norm": 0.3986241817474365,
"learning_rate": 0.006,
"loss": 1.8272,
"step": 1000
},
{
"epoch": 0.10438197602315007,
"grad_norm": 0.275290310382843,
"learning_rate": 0.006,
"loss": 1.8177,
"step": 1010
},
{
"epoch": 0.10541546093427036,
"grad_norm": 0.24022069573402405,
"learning_rate": 0.006,
"loss": 1.8,
"step": 1020
},
{
"epoch": 0.10644894584539066,
"grad_norm": 0.22507286071777344,
"learning_rate": 0.006,
"loss": 1.7825,
"step": 1030
},
{
"epoch": 0.10748243075651095,
"grad_norm": 0.2965947985649109,
"learning_rate": 0.006,
"loss": 1.778,
"step": 1040
},
{
"epoch": 0.10851591566763125,
"grad_norm": 0.3291703164577484,
"learning_rate": 0.006,
"loss": 1.7864,
"step": 1050
},
{
"epoch": 0.10954940057875155,
"grad_norm": 0.2484293133020401,
"learning_rate": 0.006,
"loss": 1.7649,
"step": 1060
},
{
"epoch": 0.11058288548987184,
"grad_norm": 0.30289486050605774,
"learning_rate": 0.006,
"loss": 1.7756,
"step": 1070
},
{
"epoch": 0.11161637040099215,
"grad_norm": 0.2765989601612091,
"learning_rate": 0.006,
"loss": 1.7458,
"step": 1080
},
{
"epoch": 0.11264985531211244,
"grad_norm": 0.24408799409866333,
"learning_rate": 0.006,
"loss": 1.7598,
"step": 1090
},
{
"epoch": 0.11368334022323275,
"grad_norm": 0.2399667352437973,
"learning_rate": 0.006,
"loss": 1.754,
"step": 1100
},
{
"epoch": 0.11471682513435304,
"grad_norm": 0.2846289873123169,
"learning_rate": 0.006,
"loss": 1.7406,
"step": 1110
},
{
"epoch": 0.11575031004547334,
"grad_norm": 0.26227328181266785,
"learning_rate": 0.006,
"loss": 1.7336,
"step": 1120
},
{
"epoch": 0.11678379495659363,
"grad_norm": 0.2673957347869873,
"learning_rate": 0.006,
"loss": 1.7281,
"step": 1130
},
{
"epoch": 0.11781727986771393,
"grad_norm": 0.3353629410266876,
"learning_rate": 0.006,
"loss": 1.7169,
"step": 1140
},
{
"epoch": 0.11885076477883423,
"grad_norm": 0.21159133315086365,
"learning_rate": 0.006,
"loss": 1.7016,
"step": 1150
},
{
"epoch": 0.11988424968995452,
"grad_norm": 0.24120958149433136,
"learning_rate": 0.006,
"loss": 1.7033,
"step": 1160
},
{
"epoch": 0.12091773460107483,
"grad_norm": 0.26428887248039246,
"learning_rate": 0.006,
"loss": 1.6867,
"step": 1170
},
{
"epoch": 0.12195121951219512,
"grad_norm": 0.27754953503608704,
"learning_rate": 0.006,
"loss": 1.6901,
"step": 1180
},
{
"epoch": 0.12298470442331542,
"grad_norm": 0.36125853657722473,
"learning_rate": 0.006,
"loss": 1.6808,
"step": 1190
},
{
"epoch": 0.12401818933443572,
"grad_norm": 0.23145093023777008,
"learning_rate": 0.006,
"loss": 1.6946,
"step": 1200
},
{
"epoch": 0.12505167424555602,
"grad_norm": 0.22397373616695404,
"learning_rate": 0.006,
"loss": 1.6619,
"step": 1210
},
{
"epoch": 0.1260851591566763,
"grad_norm": 0.2219894528388977,
"learning_rate": 0.006,
"loss": 1.6483,
"step": 1220
},
{
"epoch": 0.1271186440677966,
"grad_norm": 0.2552240788936615,
"learning_rate": 0.006,
"loss": 1.6574,
"step": 1230
},
{
"epoch": 0.1281521289789169,
"grad_norm": 0.4382542073726654,
"learning_rate": 0.006,
"loss": 1.663,
"step": 1240
},
{
"epoch": 0.12918561389003722,
"grad_norm": 0.2551495134830475,
"learning_rate": 0.006,
"loss": 1.6543,
"step": 1250
},
{
"epoch": 0.1302190988011575,
"grad_norm": 0.21785561740398407,
"learning_rate": 0.006,
"loss": 1.6479,
"step": 1260
},
{
"epoch": 0.1312525837122778,
"grad_norm": 0.25551608204841614,
"learning_rate": 0.006,
"loss": 1.6303,
"step": 1270
},
{
"epoch": 0.1322860686233981,
"grad_norm": 0.30908089876174927,
"learning_rate": 0.006,
"loss": 1.6388,
"step": 1280
},
{
"epoch": 0.1333195535345184,
"grad_norm": 0.39932873845100403,
"learning_rate": 0.006,
"loss": 1.6192,
"step": 1290
},
{
"epoch": 0.1343530384456387,
"grad_norm": 0.2735249102115631,
"learning_rate": 0.006,
"loss": 1.6327,
"step": 1300
},
{
"epoch": 0.135386523356759,
"grad_norm": 0.23229070007801056,
"learning_rate": 0.006,
"loss": 1.6278,
"step": 1310
},
{
"epoch": 0.13642000826787928,
"grad_norm": 0.29770010709762573,
"learning_rate": 0.006,
"loss": 1.6137,
"step": 1320
},
{
"epoch": 0.13745349317899958,
"grad_norm": 0.2619315981864929,
"learning_rate": 0.006,
"loss": 1.6071,
"step": 1330
},
{
"epoch": 0.1384869780901199,
"grad_norm": 0.2108081877231598,
"learning_rate": 0.006,
"loss": 1.6189,
"step": 1340
},
{
"epoch": 0.13952046300124019,
"grad_norm": 0.2384573072195053,
"learning_rate": 0.006,
"loss": 1.6035,
"step": 1350
},
{
"epoch": 0.14055394791236048,
"grad_norm": 0.2051675170660019,
"learning_rate": 0.006,
"loss": 1.5866,
"step": 1360
},
{
"epoch": 0.14158743282348077,
"grad_norm": 0.3068487346172333,
"learning_rate": 0.006,
"loss": 1.5727,
"step": 1370
},
{
"epoch": 0.1426209177346011,
"grad_norm": 0.24623431265354156,
"learning_rate": 0.006,
"loss": 1.5786,
"step": 1380
},
{
"epoch": 0.14365440264572138,
"grad_norm": 0.3030165731906891,
"learning_rate": 0.006,
"loss": 1.5793,
"step": 1390
},
{
"epoch": 0.14468788755684167,
"grad_norm": 0.236822709441185,
"learning_rate": 0.006,
"loss": 1.5737,
"step": 1400
},
{
"epoch": 0.14572137246796196,
"grad_norm": 0.4476490318775177,
"learning_rate": 0.006,
"loss": 1.5743,
"step": 1410
},
{
"epoch": 0.14675485737908225,
"grad_norm": 0.3274894952774048,
"learning_rate": 0.006,
"loss": 1.5579,
"step": 1420
},
{
"epoch": 0.14778834229020257,
"grad_norm": 0.2623043358325958,
"learning_rate": 0.006,
"loss": 1.5666,
"step": 1430
},
{
"epoch": 0.14882182720132286,
"grad_norm": 0.23737861216068268,
"learning_rate": 0.006,
"loss": 1.553,
"step": 1440
},
{
"epoch": 0.14985531211244316,
"grad_norm": 0.23686368763446808,
"learning_rate": 0.006,
"loss": 1.5598,
"step": 1450
},
{
"epoch": 0.15088879702356345,
"grad_norm": 0.2711767256259918,
"learning_rate": 0.006,
"loss": 1.5508,
"step": 1460
},
{
"epoch": 0.15192228193468377,
"grad_norm": 0.1919260323047638,
"learning_rate": 0.006,
"loss": 1.5364,
"step": 1470
},
{
"epoch": 0.15295576684580406,
"grad_norm": 0.24973885715007782,
"learning_rate": 0.006,
"loss": 1.5444,
"step": 1480
},
{
"epoch": 0.15398925175692435,
"grad_norm": 0.2096351981163025,
"learning_rate": 0.006,
"loss": 1.5422,
"step": 1490
},
{
"epoch": 0.15502273666804464,
"grad_norm": 0.28911715745925903,
"learning_rate": 0.006,
"loss": 1.5562,
"step": 1500
},
{
"epoch": 0.15605622157916493,
"grad_norm": 0.2608921229839325,
"learning_rate": 0.006,
"loss": 1.5153,
"step": 1510
},
{
"epoch": 0.15708970649028525,
"grad_norm": 0.2605489194393158,
"learning_rate": 0.006,
"loss": 1.5262,
"step": 1520
},
{
"epoch": 0.15812319140140554,
"grad_norm": 0.2948496341705322,
"learning_rate": 0.006,
"loss": 1.5318,
"step": 1530
},
{
"epoch": 0.15915667631252584,
"grad_norm": 0.33066225051879883,
"learning_rate": 0.006,
"loss": 1.5138,
"step": 1540
},
{
"epoch": 0.16019016122364613,
"grad_norm": 0.2134758085012436,
"learning_rate": 0.006,
"loss": 1.5199,
"step": 1550
},
{
"epoch": 0.16122364613476645,
"grad_norm": 0.22920167446136475,
"learning_rate": 0.006,
"loss": 1.5084,
"step": 1560
},
{
"epoch": 0.16225713104588674,
"grad_norm": 0.22044587135314941,
"learning_rate": 0.006,
"loss": 1.5254,
"step": 1570
},
{
"epoch": 0.16329061595700703,
"grad_norm": 0.28658127784729004,
"learning_rate": 0.006,
"loss": 1.5149,
"step": 1580
},
{
"epoch": 0.16432410086812732,
"grad_norm": 0.1932067573070526,
"learning_rate": 0.006,
"loss": 1.4943,
"step": 1590
},
{
"epoch": 0.1653575857792476,
"grad_norm": 0.19694995880126953,
"learning_rate": 0.006,
"loss": 1.4871,
"step": 1600
},
{
"epoch": 0.16639107069036793,
"grad_norm": 0.2092551738023758,
"learning_rate": 0.006,
"loss": 1.481,
"step": 1610
},
{
"epoch": 0.16742455560148822,
"grad_norm": 0.20461459457874298,
"learning_rate": 0.006,
"loss": 1.4958,
"step": 1620
},
{
"epoch": 0.16845804051260851,
"grad_norm": 0.2160821110010147,
"learning_rate": 0.006,
"loss": 1.5013,
"step": 1630
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.23302793502807617,
"learning_rate": 0.006,
"loss": 1.4991,
"step": 1640
},
{
"epoch": 0.1705250103348491,
"grad_norm": 0.20844736695289612,
"learning_rate": 0.006,
"loss": 1.4765,
"step": 1650
},
{
"epoch": 0.17155849524596942,
"grad_norm": 0.18730531632900238,
"learning_rate": 0.006,
"loss": 1.4732,
"step": 1660
},
{
"epoch": 0.1725919801570897,
"grad_norm": 0.2798612415790558,
"learning_rate": 0.006,
"loss": 1.4809,
"step": 1670
},
{
"epoch": 0.17362546506821,
"grad_norm": 0.17376913130283356,
"learning_rate": 0.006,
"loss": 1.4847,
"step": 1680
},
{
"epoch": 0.1746589499793303,
"grad_norm": 0.1984642595052719,
"learning_rate": 0.006,
"loss": 1.4677,
"step": 1690
},
{
"epoch": 0.1756924348904506,
"grad_norm": 0.18918636441230774,
"learning_rate": 0.006,
"loss": 1.4623,
"step": 1700
},
{
"epoch": 0.1767259198015709,
"grad_norm": 0.22821608185768127,
"learning_rate": 0.006,
"loss": 1.4836,
"step": 1710
},
{
"epoch": 0.1777594047126912,
"grad_norm": 0.2023961991071701,
"learning_rate": 0.006,
"loss": 1.4665,
"step": 1720
},
{
"epoch": 0.17879288962381149,
"grad_norm": 0.21143494546413422,
"learning_rate": 0.006,
"loss": 1.4637,
"step": 1730
},
{
"epoch": 0.17982637453493178,
"grad_norm": 0.18367299437522888,
"learning_rate": 0.006,
"loss": 1.4564,
"step": 1740
},
{
"epoch": 0.1808598594460521,
"grad_norm": 0.18500660359859467,
"learning_rate": 0.006,
"loss": 1.4485,
"step": 1750
},
{
"epoch": 0.1818933443571724,
"grad_norm": 0.208240807056427,
"learning_rate": 0.006,
"loss": 1.449,
"step": 1760
},
{
"epoch": 0.18292682926829268,
"grad_norm": 0.28708386421203613,
"learning_rate": 0.006,
"loss": 1.4514,
"step": 1770
},
{
"epoch": 0.18396031417941297,
"grad_norm": 0.23758843541145325,
"learning_rate": 0.006,
"loss": 1.4376,
"step": 1780
},
{
"epoch": 0.1849937990905333,
"grad_norm": 0.23937605321407318,
"learning_rate": 0.006,
"loss": 1.4406,
"step": 1790
},
{
"epoch": 0.18602728400165358,
"grad_norm": 0.44142553210258484,
"learning_rate": 0.006,
"loss": 1.4331,
"step": 1800
},
{
"epoch": 0.18706076891277387,
"grad_norm": 0.1951988786458969,
"learning_rate": 0.006,
"loss": 1.4443,
"step": 1810
},
{
"epoch": 0.18809425382389416,
"grad_norm": 0.22376669943332672,
"learning_rate": 0.006,
"loss": 1.4345,
"step": 1820
},
{
"epoch": 0.18912773873501446,
"grad_norm": 0.24400733411312103,
"learning_rate": 0.006,
"loss": 1.4405,
"step": 1830
},
{
"epoch": 0.19016122364613477,
"grad_norm": 0.20870748162269592,
"learning_rate": 0.006,
"loss": 1.4291,
"step": 1840
},
{
"epoch": 0.19119470855725507,
"grad_norm": 0.2239631861448288,
"learning_rate": 0.006,
"loss": 1.4269,
"step": 1850
},
{
"epoch": 0.19222819346837536,
"grad_norm": 0.24455626308918,
"learning_rate": 0.006,
"loss": 1.4146,
"step": 1860
},
{
"epoch": 0.19326167837949565,
"grad_norm": 0.25358930230140686,
"learning_rate": 0.006,
"loss": 1.4431,
"step": 1870
},
{
"epoch": 0.19429516329061597,
"grad_norm": 0.1898830235004425,
"learning_rate": 0.006,
"loss": 1.4222,
"step": 1880
},
{
"epoch": 0.19532864820173626,
"grad_norm": 0.21167783439159393,
"learning_rate": 0.006,
"loss": 1.4251,
"step": 1890
},
{
"epoch": 0.19636213311285655,
"grad_norm": 0.2603727877140045,
"learning_rate": 0.006,
"loss": 1.426,
"step": 1900
},
{
"epoch": 0.19739561802397684,
"grad_norm": 0.193504199385643,
"learning_rate": 0.006,
"loss": 1.4041,
"step": 1910
},
{
"epoch": 0.19842910293509713,
"grad_norm": 0.22818297147750854,
"learning_rate": 0.006,
"loss": 1.4204,
"step": 1920
},
{
"epoch": 0.19946258784621745,
"grad_norm": 0.20980095863342285,
"learning_rate": 0.006,
"loss": 1.405,
"step": 1930
},
{
"epoch": 0.20049607275733775,
"grad_norm": 0.17794106900691986,
"learning_rate": 0.006,
"loss": 1.3871,
"step": 1940
},
{
"epoch": 0.20152955766845804,
"grad_norm": 0.2572697103023529,
"learning_rate": 0.006,
"loss": 1.4001,
"step": 1950
},
{
"epoch": 0.20256304257957833,
"grad_norm": 0.2605222463607788,
"learning_rate": 0.006,
"loss": 1.4029,
"step": 1960
},
{
"epoch": 0.20359652749069865,
"grad_norm": 0.27022120356559753,
"learning_rate": 0.006,
"loss": 1.4009,
"step": 1970
},
{
"epoch": 0.20463001240181894,
"grad_norm": 0.20254716277122498,
"learning_rate": 0.006,
"loss": 1.3946,
"step": 1980
},
{
"epoch": 0.20566349731293923,
"grad_norm": 0.19156356155872345,
"learning_rate": 0.006,
"loss": 1.3848,
"step": 1990
},
{
"epoch": 0.20669698222405952,
"grad_norm": 0.182766854763031,
"learning_rate": 0.006,
"loss": 1.3964,
"step": 2000
},
{
"epoch": 0.2077304671351798,
"grad_norm": 0.1684897243976593,
"learning_rate": 0.006,
"loss": 1.3872,
"step": 2010
},
{
"epoch": 0.20876395204630013,
"grad_norm": 0.19054663181304932,
"learning_rate": 0.006,
"loss": 1.3817,
"step": 2020
},
{
"epoch": 0.20979743695742042,
"grad_norm": 0.17960067093372345,
"learning_rate": 0.006,
"loss": 1.389,
"step": 2030
},
{
"epoch": 0.21083092186854072,
"grad_norm": 0.23440410196781158,
"learning_rate": 0.006,
"loss": 1.3707,
"step": 2040
},
{
"epoch": 0.211864406779661,
"grad_norm": 0.21306714415550232,
"learning_rate": 0.006,
"loss": 1.3789,
"step": 2050
},
{
"epoch": 0.21289789169078133,
"grad_norm": 0.24979044497013092,
"learning_rate": 0.006,
"loss": 1.3755,
"step": 2060
},
{
"epoch": 0.21393137660190162,
"grad_norm": 0.2036939263343811,
"learning_rate": 0.006,
"loss": 1.3825,
"step": 2070
},
{
"epoch": 0.2149648615130219,
"grad_norm": 0.21026568114757538,
"learning_rate": 0.006,
"loss": 1.3701,
"step": 2080
},
{
"epoch": 0.2159983464241422,
"grad_norm": 0.18213844299316406,
"learning_rate": 0.006,
"loss": 1.3835,
"step": 2090
},
{
"epoch": 0.2170318313352625,
"grad_norm": 0.23943543434143066,
"learning_rate": 0.006,
"loss": 1.3681,
"step": 2100
},
{
"epoch": 0.2180653162463828,
"grad_norm": 0.27890509366989136,
"learning_rate": 0.006,
"loss": 1.3746,
"step": 2110
},
{
"epoch": 0.2190988011575031,
"grad_norm": 0.1875576227903366,
"learning_rate": 0.006,
"loss": 1.3833,
"step": 2120
},
{
"epoch": 0.2201322860686234,
"grad_norm": 0.21590693295001984,
"learning_rate": 0.006,
"loss": 1.3679,
"step": 2130
},
{
"epoch": 0.2211657709797437,
"grad_norm": 0.16515956819057465,
"learning_rate": 0.006,
"loss": 1.3628,
"step": 2140
},
{
"epoch": 0.222199255890864,
"grad_norm": 0.3273829221725464,
"learning_rate": 0.006,
"loss": 1.3605,
"step": 2150
},
{
"epoch": 0.2232327408019843,
"grad_norm": 0.19835293292999268,
"learning_rate": 0.006,
"loss": 1.3599,
"step": 2160
},
{
"epoch": 0.2242662257131046,
"grad_norm": 0.20204949378967285,
"learning_rate": 0.006,
"loss": 1.3548,
"step": 2170
},
{
"epoch": 0.22529971062422488,
"grad_norm": 0.20704525709152222,
"learning_rate": 0.006,
"loss": 1.3625,
"step": 2180
},
{
"epoch": 0.22633319553534517,
"grad_norm": 0.16702768206596375,
"learning_rate": 0.006,
"loss": 1.3475,
"step": 2190
},
{
"epoch": 0.2273666804464655,
"grad_norm": 0.2254660427570343,
"learning_rate": 0.006,
"loss": 1.3508,
"step": 2200
},
{
"epoch": 0.22840016535758578,
"grad_norm": 0.17011414468288422,
"learning_rate": 0.006,
"loss": 1.3416,
"step": 2210
},
{
"epoch": 0.22943365026870607,
"grad_norm": 0.24636393785476685,
"learning_rate": 0.006,
"loss": 1.3487,
"step": 2220
},
{
"epoch": 0.23046713517982637,
"grad_norm": 0.2143375277519226,
"learning_rate": 0.006,
"loss": 1.345,
"step": 2230
},
{
"epoch": 0.23150062009094668,
"grad_norm": 0.1868714541196823,
"learning_rate": 0.006,
"loss": 1.3359,
"step": 2240
},
{
"epoch": 0.23253410500206698,
"grad_norm": 0.1683160364627838,
"learning_rate": 0.006,
"loss": 1.3399,
"step": 2250
},
{
"epoch": 0.23356758991318727,
"grad_norm": 0.199097141623497,
"learning_rate": 0.006,
"loss": 1.3568,
"step": 2260
},
{
"epoch": 0.23460107482430756,
"grad_norm": 0.18377502262592316,
"learning_rate": 0.006,
"loss": 1.3488,
"step": 2270
},
{
"epoch": 0.23563455973542785,
"grad_norm": 0.20199266076087952,
"learning_rate": 0.006,
"loss": 1.344,
"step": 2280
},
{
"epoch": 0.23666804464654817,
"grad_norm": 0.271793931722641,
"learning_rate": 0.006,
"loss": 1.3169,
"step": 2290
},
{
"epoch": 0.23770152955766846,
"grad_norm": 0.17339195311069489,
"learning_rate": 0.006,
"loss": 1.3328,
"step": 2300
},
{
"epoch": 0.23873501446878875,
"grad_norm": 0.17549367249011993,
"learning_rate": 0.006,
"loss": 1.3354,
"step": 2310
},
{
"epoch": 0.23976849937990904,
"grad_norm": 0.1800873875617981,
"learning_rate": 0.006,
"loss": 1.322,
"step": 2320
},
{
"epoch": 0.24080198429102936,
"grad_norm": 0.22731319069862366,
"learning_rate": 0.006,
"loss": 1.3235,
"step": 2330
},
{
"epoch": 0.24183546920214966,
"grad_norm": 0.22910168766975403,
"learning_rate": 0.006,
"loss": 1.3243,
"step": 2340
},
{
"epoch": 0.24286895411326995,
"grad_norm": 0.1611885130405426,
"learning_rate": 0.006,
"loss": 1.3352,
"step": 2350
},
{
"epoch": 0.24390243902439024,
"grad_norm": 0.19426490366458893,
"learning_rate": 0.006,
"loss": 1.3091,
"step": 2360
},
{
"epoch": 0.24493592393551053,
"grad_norm": 0.19100475311279297,
"learning_rate": 0.006,
"loss": 1.3198,
"step": 2370
},
{
"epoch": 0.24596940884663085,
"grad_norm": 0.2597349286079407,
"learning_rate": 0.006,
"loss": 1.302,
"step": 2380
},
{
"epoch": 0.24700289375775114,
"grad_norm": 0.16532281041145325,
"learning_rate": 0.006,
"loss": 1.3182,
"step": 2390
},
{
"epoch": 0.24803637866887143,
"grad_norm": 0.2065669149160385,
"learning_rate": 0.006,
"loss": 1.3182,
"step": 2400
},
{
"epoch": 0.24906986357999172,
"grad_norm": 0.1802499145269394,
"learning_rate": 0.006,
"loss": 1.3206,
"step": 2410
},
{
"epoch": 0.25010334849111204,
"grad_norm": 0.2926013469696045,
"learning_rate": 0.006,
"loss": 1.3089,
"step": 2420
},
{
"epoch": 0.2511368334022323,
"grad_norm": 0.28276434540748596,
"learning_rate": 0.006,
"loss": 1.307,
"step": 2430
},
{
"epoch": 0.2521703183133526,
"grad_norm": 0.18156251311302185,
"learning_rate": 0.006,
"loss": 1.3227,
"step": 2440
},
{
"epoch": 0.25320380322447295,
"grad_norm": 0.18208877742290497,
"learning_rate": 0.006,
"loss": 1.3065,
"step": 2450
},
{
"epoch": 0.2542372881355932,
"grad_norm": 0.18455766141414642,
"learning_rate": 0.006,
"loss": 1.3019,
"step": 2460
},
{
"epoch": 0.25527077304671353,
"grad_norm": 0.23398499190807343,
"learning_rate": 0.006,
"loss": 1.3126,
"step": 2470
},
{
"epoch": 0.2563042579578338,
"grad_norm": 0.1919931322336197,
"learning_rate": 0.006,
"loss": 1.2982,
"step": 2480
},
{
"epoch": 0.2573377428689541,
"grad_norm": 0.24925625324249268,
"learning_rate": 0.006,
"loss": 1.32,
"step": 2490
},
{
"epoch": 0.25837122778007443,
"grad_norm": 0.2440209984779358,
"learning_rate": 0.006,
"loss": 1.2904,
"step": 2500
},
{
"epoch": 0.2594047126911947,
"grad_norm": 0.16576959192752838,
"learning_rate": 0.006,
"loss": 1.3065,
"step": 2510
},
{
"epoch": 0.260438197602315,
"grad_norm": 0.18232128024101257,
"learning_rate": 0.006,
"loss": 1.3048,
"step": 2520
},
{
"epoch": 0.2614716825134353,
"grad_norm": 0.23745986819267273,
"learning_rate": 0.006,
"loss": 1.292,
"step": 2530
},
{
"epoch": 0.2625051674245556,
"grad_norm": 0.22818538546562195,
"learning_rate": 0.006,
"loss": 1.2964,
"step": 2540
},
{
"epoch": 0.2635386523356759,
"grad_norm": 0.1811428666114807,
"learning_rate": 0.006,
"loss": 1.2968,
"step": 2550
},
{
"epoch": 0.2645721372467962,
"grad_norm": 0.2562517821788788,
"learning_rate": 0.006,
"loss": 1.2985,
"step": 2560
},
{
"epoch": 0.2656056221579165,
"grad_norm": 0.16540707647800446,
"learning_rate": 0.006,
"loss": 1.2962,
"step": 2570
},
{
"epoch": 0.2666391070690368,
"grad_norm": 0.17198024690151215,
"learning_rate": 0.006,
"loss": 1.2835,
"step": 2580
},
{
"epoch": 0.2676725919801571,
"grad_norm": 0.15511544048786163,
"learning_rate": 0.006,
"loss": 1.2689,
"step": 2590
},
{
"epoch": 0.2687060768912774,
"grad_norm": 0.1916809380054474,
"learning_rate": 0.006,
"loss": 1.2822,
"step": 2600
},
{
"epoch": 0.26973956180239766,
"grad_norm": 0.17753975093364716,
"learning_rate": 0.006,
"loss": 1.2843,
"step": 2610
},
{
"epoch": 0.270773046713518,
"grad_norm": 0.18664464354515076,
"learning_rate": 0.006,
"loss": 1.2876,
"step": 2620
},
{
"epoch": 0.2718065316246383,
"grad_norm": 0.283917635679245,
"learning_rate": 0.006,
"loss": 1.2745,
"step": 2630
},
{
"epoch": 0.27284001653575857,
"grad_norm": 0.1494530886411667,
"learning_rate": 0.006,
"loss": 1.2848,
"step": 2640
},
{
"epoch": 0.2738735014468789,
"grad_norm": 0.2093777358531952,
"learning_rate": 0.006,
"loss": 1.2786,
"step": 2650
},
{
"epoch": 0.27490698635799915,
"grad_norm": 0.19870586693286896,
"learning_rate": 0.006,
"loss": 1.2895,
"step": 2660
},
{
"epoch": 0.27594047126911947,
"grad_norm": 0.18807923793792725,
"learning_rate": 0.006,
"loss": 1.2784,
"step": 2670
},
{
"epoch": 0.2769739561802398,
"grad_norm": 0.16479447484016418,
"learning_rate": 0.006,
"loss": 1.2815,
"step": 2680
},
{
"epoch": 0.27800744109136005,
"grad_norm": 0.19209875166416168,
"learning_rate": 0.006,
"loss": 1.2791,
"step": 2690
},
{
"epoch": 0.27904092600248037,
"grad_norm": 0.20269426703453064,
"learning_rate": 0.006,
"loss": 1.2828,
"step": 2700
},
{
"epoch": 0.28007441091360064,
"grad_norm": 0.1880909502506256,
"learning_rate": 0.006,
"loss": 1.2786,
"step": 2710
},
{
"epoch": 0.28110789582472095,
"grad_norm": 0.16020233929157257,
"learning_rate": 0.006,
"loss": 1.2637,
"step": 2720
},
{
"epoch": 0.2821413807358413,
"grad_norm": 0.17531338334083557,
"learning_rate": 0.006,
"loss": 1.264,
"step": 2730
},
{
"epoch": 0.28317486564696154,
"grad_norm": 0.16038616001605988,
"learning_rate": 0.006,
"loss": 1.2692,
"step": 2740
},
{
"epoch": 0.28420835055808186,
"grad_norm": 0.19676966965198517,
"learning_rate": 0.006,
"loss": 1.271,
"step": 2750
},
{
"epoch": 0.2852418354692022,
"grad_norm": 0.20371529459953308,
"learning_rate": 0.006,
"loss": 1.2637,
"step": 2760
},
{
"epoch": 0.28627532038032244,
"grad_norm": 0.18698766827583313,
"learning_rate": 0.006,
"loss": 1.2776,
"step": 2770
},
{
"epoch": 0.28730880529144276,
"grad_norm": 0.20747795701026917,
"learning_rate": 0.006,
"loss": 1.2573,
"step": 2780
},
{
"epoch": 0.288342290202563,
"grad_norm": 0.2328963726758957,
"learning_rate": 0.006,
"loss": 1.2568,
"step": 2790
},
{
"epoch": 0.28937577511368334,
"grad_norm": 0.1655196100473404,
"learning_rate": 0.006,
"loss": 1.2614,
"step": 2800
},
{
"epoch": 0.29040926002480366,
"grad_norm": 0.1615329086780548,
"learning_rate": 0.006,
"loss": 1.2712,
"step": 2810
},
{
"epoch": 0.2914427449359239,
"grad_norm": 0.15835103392601013,
"learning_rate": 0.006,
"loss": 1.2524,
"step": 2820
},
{
"epoch": 0.29247622984704424,
"grad_norm": 0.20250938832759857,
"learning_rate": 0.006,
"loss": 1.264,
"step": 2830
},
{
"epoch": 0.2935097147581645,
"grad_norm": 0.17166636884212494,
"learning_rate": 0.006,
"loss": 1.2563,
"step": 2840
},
{
"epoch": 0.2945431996692848,
"grad_norm": 0.2327214777469635,
"learning_rate": 0.006,
"loss": 1.2546,
"step": 2850
},
{
"epoch": 0.29557668458040515,
"grad_norm": 0.19193734228610992,
"learning_rate": 0.006,
"loss": 1.2647,
"step": 2860
},
{
"epoch": 0.2966101694915254,
"grad_norm": 0.21454569697380066,
"learning_rate": 0.006,
"loss": 1.2573,
"step": 2870
},
{
"epoch": 0.29764365440264573,
"grad_norm": 0.18797141313552856,
"learning_rate": 0.006,
"loss": 1.2661,
"step": 2880
},
{
"epoch": 0.298677139313766,
"grad_norm": 0.16613776981830597,
"learning_rate": 0.006,
"loss": 1.2593,
"step": 2890
},
{
"epoch": 0.2997106242248863,
"grad_norm": 0.15257228910923004,
"learning_rate": 0.006,
"loss": 1.2534,
"step": 2900
},
{
"epoch": 0.30074410913600663,
"grad_norm": 0.1718466728925705,
"learning_rate": 0.006,
"loss": 1.2526,
"step": 2910
},
{
"epoch": 0.3017775940471269,
"grad_norm": 0.17519408464431763,
"learning_rate": 0.006,
"loss": 1.236,
"step": 2920
},
{
"epoch": 0.3028110789582472,
"grad_norm": 0.1739313006401062,
"learning_rate": 0.006,
"loss": 1.2416,
"step": 2930
},
{
"epoch": 0.30384456386936753,
"grad_norm": 0.2000485360622406,
"learning_rate": 0.006,
"loss": 1.2525,
"step": 2940
},
{
"epoch": 0.3048780487804878,
"grad_norm": 0.21297138929367065,
"learning_rate": 0.006,
"loss": 1.2312,
"step": 2950
},
{
"epoch": 0.3059115336916081,
"grad_norm": 0.1601925939321518,
"learning_rate": 0.006,
"loss": 1.2575,
"step": 2960
},
{
"epoch": 0.3069450186027284,
"grad_norm": 0.22686167061328888,
"learning_rate": 0.006,
"loss": 1.2421,
"step": 2970
},
{
"epoch": 0.3079785035138487,
"grad_norm": 0.2552824318408966,
"learning_rate": 0.006,
"loss": 1.255,
"step": 2980
},
{
"epoch": 0.309011988424969,
"grad_norm": 0.1520908921957016,
"learning_rate": 0.006,
"loss": 1.2335,
"step": 2990
},
{
"epoch": 0.3100454733360893,
"grad_norm": 0.18067720532417297,
"learning_rate": 0.006,
"loss": 1.2278,
"step": 3000
},
{
"epoch": 0.3110789582472096,
"grad_norm": 0.2269776612520218,
"learning_rate": 0.006,
"loss": 1.2401,
"step": 3010
},
{
"epoch": 0.31211244315832987,
"grad_norm": 0.17168264091014862,
"learning_rate": 0.006,
"loss": 1.245,
"step": 3020
},
{
"epoch": 0.3131459280694502,
"grad_norm": 0.20345386862754822,
"learning_rate": 0.006,
"loss": 1.2286,
"step": 3030
},
{
"epoch": 0.3141794129805705,
"grad_norm": 0.17593909800052643,
"learning_rate": 0.006,
"loss": 1.2439,
"step": 3040
},
{
"epoch": 0.31521289789169077,
"grad_norm": 0.1948511153459549,
"learning_rate": 0.006,
"loss": 1.2513,
"step": 3050
},
{
"epoch": 0.3162463828028111,
"grad_norm": 0.1758977472782135,
"learning_rate": 0.006,
"loss": 1.2428,
"step": 3060
},
{
"epoch": 0.31727986771393135,
"grad_norm": 0.14619344472885132,
"learning_rate": 0.006,
"loss": 1.2295,
"step": 3070
},
{
"epoch": 0.31831335262505167,
"grad_norm": 0.15746112167835236,
"learning_rate": 0.006,
"loss": 1.2133,
"step": 3080
},
{
"epoch": 0.319346837536172,
"grad_norm": 0.14768971502780914,
"learning_rate": 0.006,
"loss": 1.2234,
"step": 3090
},
{
"epoch": 0.32038032244729225,
"grad_norm": 0.15294450521469116,
"learning_rate": 0.006,
"loss": 1.2255,
"step": 3100
},
{
"epoch": 0.3214138073584126,
"grad_norm": 0.18775102496147156,
"learning_rate": 0.006,
"loss": 1.2378,
"step": 3110
},
{
"epoch": 0.3224472922695329,
"grad_norm": 0.15432560443878174,
"learning_rate": 0.006,
"loss": 1.2359,
"step": 3120
},
{
"epoch": 0.32348077718065316,
"grad_norm": 0.17162322998046875,
"learning_rate": 0.006,
"loss": 1.2257,
"step": 3130
},
{
"epoch": 0.3245142620917735,
"grad_norm": 0.21141092479228973,
"learning_rate": 0.006,
"loss": 1.2301,
"step": 3140
},
{
"epoch": 0.32554774700289374,
"grad_norm": 0.1810036152601242,
"learning_rate": 0.006,
"loss": 1.2204,
"step": 3150
},
{
"epoch": 0.32658123191401406,
"grad_norm": 0.20127083361148834,
"learning_rate": 0.006,
"loss": 1.2263,
"step": 3160
},
{
"epoch": 0.3276147168251344,
"grad_norm": 0.1795072853565216,
"learning_rate": 0.006,
"loss": 1.2224,
"step": 3170
},
{
"epoch": 0.32864820173625464,
"grad_norm": 0.26130935549736023,
"learning_rate": 0.006,
"loss": 1.2236,
"step": 3180
},
{
"epoch": 0.32968168664737496,
"grad_norm": 0.17055651545524597,
"learning_rate": 0.006,
"loss": 1.2229,
"step": 3190
},
{
"epoch": 0.3307151715584952,
"grad_norm": 0.24257224798202515,
"learning_rate": 0.006,
"loss": 1.218,
"step": 3200
},
{
"epoch": 0.33174865646961554,
"grad_norm": 0.1574934422969818,
"learning_rate": 0.006,
"loss": 1.2248,
"step": 3210
},
{
"epoch": 0.33278214138073586,
"grad_norm": 0.18338268995285034,
"learning_rate": 0.006,
"loss": 1.211,
"step": 3220
},
{
"epoch": 0.3338156262918561,
"grad_norm": 0.23949076235294342,
"learning_rate": 0.006,
"loss": 1.2107,
"step": 3230
},
{
"epoch": 0.33484911120297645,
"grad_norm": 0.23181311786174774,
"learning_rate": 0.006,
"loss": 1.2157,
"step": 3240
},
{
"epoch": 0.3358825961140967,
"grad_norm": 0.1727987825870514,
"learning_rate": 0.006,
"loss": 1.2113,
"step": 3250
},
{
"epoch": 0.33691608102521703,
"grad_norm": 0.17050251364707947,
"learning_rate": 0.006,
"loss": 1.2196,
"step": 3260
},
{
"epoch": 0.33794956593633735,
"grad_norm": 0.1836351752281189,
"learning_rate": 0.006,
"loss": 1.2164,
"step": 3270
},
{
"epoch": 0.3389830508474576,
"grad_norm": 0.16653425991535187,
"learning_rate": 0.006,
"loss": 1.2108,
"step": 3280
},
{
"epoch": 0.34001653575857793,
"grad_norm": 0.19799716770648956,
"learning_rate": 0.006,
"loss": 1.2072,
"step": 3290
},
{
"epoch": 0.3410500206696982,
"grad_norm": 0.15341606736183167,
"learning_rate": 0.006,
"loss": 1.2157,
"step": 3300
},
{
"epoch": 0.3420835055808185,
"grad_norm": 0.15945138037204742,
"learning_rate": 0.006,
"loss": 1.2065,
"step": 3310
},
{
"epoch": 0.34311699049193883,
"grad_norm": 0.17721377313137054,
"learning_rate": 0.006,
"loss": 1.2056,
"step": 3320
},
{
"epoch": 0.3441504754030591,
"grad_norm": 0.1358070969581604,
"learning_rate": 0.006,
"loss": 1.2171,
"step": 3330
},
{
"epoch": 0.3451839603141794,
"grad_norm": 0.17187251150608063,
"learning_rate": 0.006,
"loss": 1.2046,
"step": 3340
},
{
"epoch": 0.34621744522529974,
"grad_norm": 0.14732858538627625,
"learning_rate": 0.006,
"loss": 1.2025,
"step": 3350
},
{
"epoch": 0.34725093013642,
"grad_norm": 0.224739208817482,
"learning_rate": 0.006,
"loss": 1.2213,
"step": 3360
},
{
"epoch": 0.3482844150475403,
"grad_norm": 0.20575636625289917,
"learning_rate": 0.006,
"loss": 1.1922,
"step": 3370
},
{
"epoch": 0.3493178999586606,
"grad_norm": 0.15750053524971008,
"learning_rate": 0.006,
"loss": 1.209,
"step": 3380
},
{
"epoch": 0.3503513848697809,
"grad_norm": 0.27751949429512024,
"learning_rate": 0.006,
"loss": 1.2079,
"step": 3390
},
{
"epoch": 0.3513848697809012,
"grad_norm": 0.14315839111804962,
"learning_rate": 0.006,
"loss": 1.1934,
"step": 3400
},
{
"epoch": 0.3524183546920215,
"grad_norm": 0.15051007270812988,
"learning_rate": 0.006,
"loss": 1.2067,
"step": 3410
},
{
"epoch": 0.3534518396031418,
"grad_norm": 0.15940986573696136,
"learning_rate": 0.006,
"loss": 1.2097,
"step": 3420
},
{
"epoch": 0.35448532451426207,
"grad_norm": 0.1412818878889084,
"learning_rate": 0.006,
"loss": 1.1959,
"step": 3430
},
{
"epoch": 0.3555188094253824,
"grad_norm": 0.1425037980079651,
"learning_rate": 0.006,
"loss": 1.1901,
"step": 3440
},
{
"epoch": 0.3565522943365027,
"grad_norm": 0.15428400039672852,
"learning_rate": 0.006,
"loss": 1.2037,
"step": 3450
},
{
"epoch": 0.35758577924762297,
"grad_norm": 0.1565396934747696,
"learning_rate": 0.006,
"loss": 1.2066,
"step": 3460
},
{
"epoch": 0.3586192641587433,
"grad_norm": 0.1757258176803589,
"learning_rate": 0.006,
"loss": 1.1947,
"step": 3470
},
{
"epoch": 0.35965274906986355,
"grad_norm": 0.23932603001594543,
"learning_rate": 0.006,
"loss": 1.1993,
"step": 3480
},
{
"epoch": 0.3606862339809839,
"grad_norm": 0.17783917486667633,
"learning_rate": 0.006,
"loss": 1.1881,
"step": 3490
},
{
"epoch": 0.3617197188921042,
"grad_norm": 0.13933712244033813,
"learning_rate": 0.006,
"loss": 1.1969,
"step": 3500
},
{
"epoch": 0.36275320380322446,
"grad_norm": 0.13872799277305603,
"learning_rate": 0.006,
"loss": 1.1958,
"step": 3510
},
{
"epoch": 0.3637866887143448,
"grad_norm": 0.18496429920196533,
"learning_rate": 0.006,
"loss": 1.1978,
"step": 3520
},
{
"epoch": 0.3648201736254651,
"grad_norm": 0.2182753086090088,
"learning_rate": 0.006,
"loss": 1.1957,
"step": 3530
},
{
"epoch": 0.36585365853658536,
"grad_norm": 0.15641410648822784,
"learning_rate": 0.006,
"loss": 1.1828,
"step": 3540
},
{
"epoch": 0.3668871434477057,
"grad_norm": 0.15196652710437775,
"learning_rate": 0.006,
"loss": 1.1901,
"step": 3550
},
{
"epoch": 0.36792062835882594,
"grad_norm": 0.15420852601528168,
"learning_rate": 0.006,
"loss": 1.1966,
"step": 3560
},
{
"epoch": 0.36895411326994626,
"grad_norm": 0.1419685333967209,
"learning_rate": 0.006,
"loss": 1.1835,
"step": 3570
},
{
"epoch": 0.3699875981810666,
"grad_norm": 0.20714856684207916,
"learning_rate": 0.006,
"loss": 1.1969,
"step": 3580
},
{
"epoch": 0.37102108309218684,
"grad_norm": 0.14852945506572723,
"learning_rate": 0.006,
"loss": 1.1979,
"step": 3590
},
{
"epoch": 0.37205456800330716,
"grad_norm": 0.1534539759159088,
"learning_rate": 0.006,
"loss": 1.1814,
"step": 3600
},
{
"epoch": 0.3730880529144274,
"grad_norm": 0.1536993384361267,
"learning_rate": 0.006,
"loss": 1.1935,
"step": 3610
},
{
"epoch": 0.37412153782554775,
"grad_norm": 0.14678268134593964,
"learning_rate": 0.006,
"loss": 1.1879,
"step": 3620
},
{
"epoch": 0.37515502273666806,
"grad_norm": 0.15194503962993622,
"learning_rate": 0.006,
"loss": 1.1851,
"step": 3630
},
{
"epoch": 0.37618850764778833,
"grad_norm": 0.1853533834218979,
"learning_rate": 0.006,
"loss": 1.177,
"step": 3640
},
{
"epoch": 0.37722199255890865,
"grad_norm": 0.19894924759864807,
"learning_rate": 0.006,
"loss": 1.1792,
"step": 3650
},
{
"epoch": 0.3782554774700289,
"grad_norm": 0.14289256930351257,
"learning_rate": 0.006,
"loss": 1.1916,
"step": 3660
},
{
"epoch": 0.37928896238114923,
"grad_norm": 0.15364837646484375,
"learning_rate": 0.006,
"loss": 1.1773,
"step": 3670
},
{
"epoch": 0.38032244729226955,
"grad_norm": 0.26057279109954834,
"learning_rate": 0.006,
"loss": 1.1772,
"step": 3680
},
{
"epoch": 0.3813559322033898,
"grad_norm": 0.14714686572551727,
"learning_rate": 0.006,
"loss": 1.183,
"step": 3690
},
{
"epoch": 0.38238941711451013,
"grad_norm": 0.13475820422172546,
"learning_rate": 0.006,
"loss": 1.1734,
"step": 3700
},
{
"epoch": 0.38342290202563045,
"grad_norm": 0.169892817735672,
"learning_rate": 0.006,
"loss": 1.1833,
"step": 3710
},
{
"epoch": 0.3844563869367507,
"grad_norm": 0.15819117426872253,
"learning_rate": 0.006,
"loss": 1.1693,
"step": 3720
},
{
"epoch": 0.38548987184787104,
"grad_norm": 0.2074589729309082,
"learning_rate": 0.006,
"loss": 1.1773,
"step": 3730
},
{
"epoch": 0.3865233567589913,
"grad_norm": 0.1474793255329132,
"learning_rate": 0.006,
"loss": 1.1746,
"step": 3740
},
{
"epoch": 0.3875568416701116,
"grad_norm": 0.20919297635555267,
"learning_rate": 0.006,
"loss": 1.1644,
"step": 3750
},
{
"epoch": 0.38859032658123194,
"grad_norm": 0.1802784502506256,
"learning_rate": 0.006,
"loss": 1.1785,
"step": 3760
},
{
"epoch": 0.3896238114923522,
"grad_norm": 0.13777178525924683,
"learning_rate": 0.006,
"loss": 1.1762,
"step": 3770
},
{
"epoch": 0.3906572964034725,
"grad_norm": 0.19703027606010437,
"learning_rate": 0.006,
"loss": 1.1715,
"step": 3780
},
{
"epoch": 0.3916907813145928,
"grad_norm": 0.13620315492153168,
"learning_rate": 0.006,
"loss": 1.1729,
"step": 3790
},
{
"epoch": 0.3927242662257131,
"grad_norm": 0.19151124358177185,
"learning_rate": 0.006,
"loss": 1.165,
"step": 3800
},
{
"epoch": 0.3937577511368334,
"grad_norm": 0.14211703836917877,
"learning_rate": 0.006,
"loss": 1.1705,
"step": 3810
},
{
"epoch": 0.3947912360479537,
"grad_norm": 0.13506080210208893,
"learning_rate": 0.006,
"loss": 1.1656,
"step": 3820
},
{
"epoch": 0.395824720959074,
"grad_norm": 0.15365293622016907,
"learning_rate": 0.006,
"loss": 1.1642,
"step": 3830
},
{
"epoch": 0.39685820587019427,
"grad_norm": 0.1530192494392395,
"learning_rate": 0.006,
"loss": 1.1621,
"step": 3840
},
{
"epoch": 0.3978916907813146,
"grad_norm": 0.1344013661146164,
"learning_rate": 0.006,
"loss": 1.1694,
"step": 3850
},
{
"epoch": 0.3989251756924349,
"grad_norm": 0.1469976007938385,
"learning_rate": 0.006,
"loss": 1.1706,
"step": 3860
},
{
"epoch": 0.39995866060355517,
"grad_norm": 0.14535875618457794,
"learning_rate": 0.006,
"loss": 1.1641,
"step": 3870
},
{
"epoch": 0.4009921455146755,
"grad_norm": 0.22596335411071777,
"learning_rate": 0.006,
"loss": 1.1835,
"step": 3880
},
{
"epoch": 0.4020256304257958,
"grad_norm": 0.214060440659523,
"learning_rate": 0.006,
"loss": 1.1688,
"step": 3890
},
{
"epoch": 0.4030591153369161,
"grad_norm": 0.14078359305858612,
"learning_rate": 0.006,
"loss": 1.158,
"step": 3900
},
{
"epoch": 0.4040926002480364,
"grad_norm": 0.15023833513259888,
"learning_rate": 0.006,
"loss": 1.1599,
"step": 3910
},
{
"epoch": 0.40512608515915666,
"grad_norm": 0.13650792837142944,
"learning_rate": 0.006,
"loss": 1.1512,
"step": 3920
},
{
"epoch": 0.406159570070277,
"grad_norm": 0.1389831155538559,
"learning_rate": 0.006,
"loss": 1.1624,
"step": 3930
},
{
"epoch": 0.4071930549813973,
"grad_norm": 0.1548847109079361,
"learning_rate": 0.006,
"loss": 1.1822,
"step": 3940
},
{
"epoch": 0.40822653989251756,
"grad_norm": 0.14826175570487976,
"learning_rate": 0.006,
"loss": 1.1539,
"step": 3950
},
{
"epoch": 0.4092600248036379,
"grad_norm": 0.27681705355644226,
"learning_rate": 0.006,
"loss": 1.1543,
"step": 3960
},
{
"epoch": 0.41029350971475814,
"grad_norm": 0.17979800701141357,
"learning_rate": 0.006,
"loss": 1.1583,
"step": 3970
},
{
"epoch": 0.41132699462587846,
"grad_norm": 0.14845605194568634,
"learning_rate": 0.006,
"loss": 1.1559,
"step": 3980
},
{
"epoch": 0.4123604795369988,
"grad_norm": 0.18015886843204498,
"learning_rate": 0.006,
"loss": 1.1663,
"step": 3990
},
{
"epoch": 0.41339396444811904,
"grad_norm": 0.2010691910982132,
"learning_rate": 0.006,
"loss": 1.1572,
"step": 4000
},
{
"epoch": 0.41442744935923936,
"grad_norm": 0.14041297137737274,
"learning_rate": 0.006,
"loss": 1.1621,
"step": 4010
},
{
"epoch": 0.4154609342703596,
"grad_norm": 0.13786040246486664,
"learning_rate": 0.006,
"loss": 1.157,
"step": 4020
},
{
"epoch": 0.41649441918147995,
"grad_norm": 0.17166124284267426,
"learning_rate": 0.006,
"loss": 1.178,
"step": 4030
},
{
"epoch": 0.41752790409260027,
"grad_norm": 0.13718323409557343,
"learning_rate": 0.006,
"loss": 1.1514,
"step": 4040
},
{
"epoch": 0.41856138900372053,
"grad_norm": 0.15878678858280182,
"learning_rate": 0.006,
"loss": 1.1718,
"step": 4050
},
{
"epoch": 0.41959487391484085,
"grad_norm": 0.14852339029312134,
"learning_rate": 0.006,
"loss": 1.1559,
"step": 4060
},
{
"epoch": 0.42062835882596117,
"grad_norm": 0.17088137567043304,
"learning_rate": 0.006,
"loss": 1.1527,
"step": 4070
},
{
"epoch": 0.42166184373708143,
"grad_norm": 0.15074481070041656,
"learning_rate": 0.006,
"loss": 1.1519,
"step": 4080
},
{
"epoch": 0.42269532864820175,
"grad_norm": 0.13103771209716797,
"learning_rate": 0.006,
"loss": 1.1475,
"step": 4090
},
{
"epoch": 0.423728813559322,
"grad_norm": 0.1428011655807495,
"learning_rate": 0.006,
"loss": 1.1528,
"step": 4100
},
{
"epoch": 0.42476229847044233,
"grad_norm": 0.14280404150485992,
"learning_rate": 0.006,
"loss": 1.1542,
"step": 4110
},
{
"epoch": 0.42579578338156265,
"grad_norm": 0.14426454901695251,
"learning_rate": 0.006,
"loss": 1.1632,
"step": 4120
},
{
"epoch": 0.4268292682926829,
"grad_norm": 0.13187375664710999,
"learning_rate": 0.006,
"loss": 1.1564,
"step": 4130
},
{
"epoch": 0.42786275320380324,
"grad_norm": 0.1555539071559906,
"learning_rate": 0.006,
"loss": 1.1497,
"step": 4140
},
{
"epoch": 0.4288962381149235,
"grad_norm": 0.14076194167137146,
"learning_rate": 0.006,
"loss": 1.1431,
"step": 4150
},
{
"epoch": 0.4299297230260438,
"grad_norm": 0.14553508162498474,
"learning_rate": 0.006,
"loss": 1.1444,
"step": 4160
},
{
"epoch": 0.43096320793716414,
"grad_norm": 0.14131340384483337,
"learning_rate": 0.006,
"loss": 1.1504,
"step": 4170
},
{
"epoch": 0.4319966928482844,
"grad_norm": 0.1349923461675644,
"learning_rate": 0.006,
"loss": 1.1499,
"step": 4180
},
{
"epoch": 0.4330301777594047,
"grad_norm": 0.1801183521747589,
"learning_rate": 0.006,
"loss": 1.1458,
"step": 4190
},
{
"epoch": 0.434063662670525,
"grad_norm": 0.16793182492256165,
"learning_rate": 0.006,
"loss": 1.1527,
"step": 4200
},
{
"epoch": 0.4350971475816453,
"grad_norm": 0.19141018390655518,
"learning_rate": 0.006,
"loss": 1.1451,
"step": 4210
},
{
"epoch": 0.4361306324927656,
"grad_norm": 0.1311262547969818,
"learning_rate": 0.006,
"loss": 1.1433,
"step": 4220
},
{
"epoch": 0.4371641174038859,
"grad_norm": 0.1293700635433197,
"learning_rate": 0.006,
"loss": 1.1251,
"step": 4230
},
{
"epoch": 0.4381976023150062,
"grad_norm": 0.13121871650218964,
"learning_rate": 0.006,
"loss": 1.1392,
"step": 4240
},
{
"epoch": 0.43923108722612647,
"grad_norm": 0.13999488949775696,
"learning_rate": 0.006,
"loss": 1.135,
"step": 4250
},
{
"epoch": 0.4402645721372468,
"grad_norm": 0.14969255030155182,
"learning_rate": 0.006,
"loss": 1.1399,
"step": 4260
},
{
"epoch": 0.4412980570483671,
"grad_norm": 0.14516979455947876,
"learning_rate": 0.006,
"loss": 1.1385,
"step": 4270
},
{
"epoch": 0.4423315419594874,
"grad_norm": 0.13377705216407776,
"learning_rate": 0.006,
"loss": 1.1409,
"step": 4280
},
{
"epoch": 0.4433650268706077,
"grad_norm": 0.1788671314716339,
"learning_rate": 0.006,
"loss": 1.1368,
"step": 4290
},
{
"epoch": 0.444398511781728,
"grad_norm": 0.1459660530090332,
"learning_rate": 0.006,
"loss": 1.1399,
"step": 4300
},
{
"epoch": 0.4454319966928483,
"grad_norm": 0.14903658628463745,
"learning_rate": 0.006,
"loss": 1.1347,
"step": 4310
},
{
"epoch": 0.4464654816039686,
"grad_norm": 0.16043029725551605,
"learning_rate": 0.006,
"loss": 1.1395,
"step": 4320
},
{
"epoch": 0.44749896651508886,
"grad_norm": 0.1458473652601242,
"learning_rate": 0.006,
"loss": 1.1442,
"step": 4330
},
{
"epoch": 0.4485324514262092,
"grad_norm": 0.1839747577905655,
"learning_rate": 0.006,
"loss": 1.1278,
"step": 4340
},
{
"epoch": 0.4495659363373295,
"grad_norm": 0.14207588136196136,
"learning_rate": 0.006,
"loss": 1.1355,
"step": 4350
},
{
"epoch": 0.45059942124844976,
"grad_norm": 0.1426732838153839,
"learning_rate": 0.006,
"loss": 1.1345,
"step": 4360
},
{
"epoch": 0.4516329061595701,
"grad_norm": 0.19898781180381775,
"learning_rate": 0.006,
"loss": 1.128,
"step": 4370
},
{
"epoch": 0.45266639107069034,
"grad_norm": 0.1501811444759369,
"learning_rate": 0.006,
"loss": 1.1294,
"step": 4380
},
{
"epoch": 0.45369987598181066,
"grad_norm": 0.13546693325042725,
"learning_rate": 0.006,
"loss": 1.1459,
"step": 4390
},
{
"epoch": 0.454733360892931,
"grad_norm": 0.1429995894432068,
"learning_rate": 0.006,
"loss": 1.1401,
"step": 4400
},
{
"epoch": 0.45576684580405125,
"grad_norm": 0.16272728145122528,
"learning_rate": 0.006,
"loss": 1.1388,
"step": 4410
},
{
"epoch": 0.45680033071517157,
"grad_norm": 0.13995374739170074,
"learning_rate": 0.006,
"loss": 1.1348,
"step": 4420
},
{
"epoch": 0.45783381562629183,
"grad_norm": 0.13635413348674774,
"learning_rate": 0.006,
"loss": 1.13,
"step": 4430
},
{
"epoch": 0.45886730053741215,
"grad_norm": 0.12863661348819733,
"learning_rate": 0.006,
"loss": 1.1289,
"step": 4440
},
{
"epoch": 0.45990078544853247,
"grad_norm": 0.14227396249771118,
"learning_rate": 0.006,
"loss": 1.1304,
"step": 4450
},
{
"epoch": 0.46093427035965273,
"grad_norm": 0.21000246703624725,
"learning_rate": 0.006,
"loss": 1.132,
"step": 4460
},
{
"epoch": 0.46196775527077305,
"grad_norm": 0.15686342120170593,
"learning_rate": 0.006,
"loss": 1.1456,
"step": 4470
},
{
"epoch": 0.46300124018189337,
"grad_norm": 0.13623683154582977,
"learning_rate": 0.006,
"loss": 1.1229,
"step": 4480
},
{
"epoch": 0.46403472509301363,
"grad_norm": 0.13092273473739624,
"learning_rate": 0.006,
"loss": 1.1199,
"step": 4490
},
{
"epoch": 0.46506821000413395,
"grad_norm": 0.13690875470638275,
"learning_rate": 0.006,
"loss": 1.1095,
"step": 4500
},
{
"epoch": 0.4661016949152542,
"grad_norm": 0.12770096957683563,
"learning_rate": 0.006,
"loss": 1.1251,
"step": 4510
},
{
"epoch": 0.46713517982637454,
"grad_norm": 0.12008727341890335,
"learning_rate": 0.006,
"loss": 1.1151,
"step": 4520
},
{
"epoch": 0.46816866473749486,
"grad_norm": 0.13290061056613922,
"learning_rate": 0.006,
"loss": 1.1278,
"step": 4530
},
{
"epoch": 0.4692021496486151,
"grad_norm": 0.16319867968559265,
"learning_rate": 0.006,
"loss": 1.1261,
"step": 4540
},
{
"epoch": 0.47023563455973544,
"grad_norm": 0.13294318318367004,
"learning_rate": 0.006,
"loss": 1.1314,
"step": 4550
},
{
"epoch": 0.4712691194708557,
"grad_norm": 0.15630632638931274,
"learning_rate": 0.006,
"loss": 1.1423,
"step": 4560
},
{
"epoch": 0.472302604381976,
"grad_norm": 0.16916899383068085,
"learning_rate": 0.006,
"loss": 1.1385,
"step": 4570
},
{
"epoch": 0.47333608929309634,
"grad_norm": 0.2240544855594635,
"learning_rate": 0.006,
"loss": 1.1307,
"step": 4580
},
{
"epoch": 0.4743695742042166,
"grad_norm": 0.12742292881011963,
"learning_rate": 0.006,
"loss": 1.1278,
"step": 4590
},
{
"epoch": 0.4754030591153369,
"grad_norm": 0.13861480355262756,
"learning_rate": 0.006,
"loss": 1.1248,
"step": 4600
},
{
"epoch": 0.4764365440264572,
"grad_norm": 0.16158758103847504,
"learning_rate": 0.006,
"loss": 1.1437,
"step": 4610
},
{
"epoch": 0.4774700289375775,
"grad_norm": 0.13849808275699615,
"learning_rate": 0.006,
"loss": 1.1345,
"step": 4620
},
{
"epoch": 0.4785035138486978,
"grad_norm": 0.1381804645061493,
"learning_rate": 0.006,
"loss": 1.124,
"step": 4630
},
{
"epoch": 0.4795369987598181,
"grad_norm": 0.1629112809896469,
"learning_rate": 0.006,
"loss": 1.1198,
"step": 4640
},
{
"epoch": 0.4805704836709384,
"grad_norm": 0.1835983693599701,
"learning_rate": 0.006,
"loss": 1.1198,
"step": 4650
},
{
"epoch": 0.48160396858205873,
"grad_norm": 0.13495475053787231,
"learning_rate": 0.006,
"loss": 1.1266,
"step": 4660
},
{
"epoch": 0.482637453493179,
"grad_norm": 0.15339604020118713,
"learning_rate": 0.006,
"loss": 1.1111,
"step": 4670
},
{
"epoch": 0.4836709384042993,
"grad_norm": 0.14230936765670776,
"learning_rate": 0.006,
"loss": 1.1002,
"step": 4680
},
{
"epoch": 0.4847044233154196,
"grad_norm": 0.13988140225410461,
"learning_rate": 0.006,
"loss": 1.1132,
"step": 4690
},
{
"epoch": 0.4857379082265399,
"grad_norm": 0.17499862611293793,
"learning_rate": 0.006,
"loss": 1.1223,
"step": 4700
},
{
"epoch": 0.4867713931376602,
"grad_norm": 0.15850752592086792,
"learning_rate": 0.006,
"loss": 1.1199,
"step": 4710
},
{
"epoch": 0.4878048780487805,
"grad_norm": 0.12476572394371033,
"learning_rate": 0.006,
"loss": 1.1252,
"step": 4720
},
{
"epoch": 0.4888383629599008,
"grad_norm": 0.14426189661026,
"learning_rate": 0.006,
"loss": 1.1132,
"step": 4730
},
{
"epoch": 0.48987184787102106,
"grad_norm": 0.14036618173122406,
"learning_rate": 0.006,
"loss": 1.1226,
"step": 4740
},
{
"epoch": 0.4909053327821414,
"grad_norm": 0.1434011608362198,
"learning_rate": 0.006,
"loss": 1.1123,
"step": 4750
},
{
"epoch": 0.4919388176932617,
"grad_norm": 0.19470463693141937,
"learning_rate": 0.006,
"loss": 1.1321,
"step": 4760
},
{
"epoch": 0.49297230260438196,
"grad_norm": 0.12864422798156738,
"learning_rate": 0.006,
"loss": 1.1176,
"step": 4770
},
{
"epoch": 0.4940057875155023,
"grad_norm": 0.16909734904766083,
"learning_rate": 0.006,
"loss": 1.1164,
"step": 4780
},
{
"epoch": 0.49503927242662255,
"grad_norm": 0.14817242324352264,
"learning_rate": 0.006,
"loss": 1.1086,
"step": 4790
},
{
"epoch": 0.49607275733774286,
"grad_norm": 0.16644009947776794,
"learning_rate": 0.006,
"loss": 1.1142,
"step": 4800
},
{
"epoch": 0.4971062422488632,
"grad_norm": 0.16777795553207397,
"learning_rate": 0.006,
"loss": 1.1077,
"step": 4810
},
{
"epoch": 0.49813972715998345,
"grad_norm": 0.1416415572166443,
"learning_rate": 0.006,
"loss": 1.122,
"step": 4820
},
{
"epoch": 0.49917321207110377,
"grad_norm": 0.1440482884645462,
"learning_rate": 0.006,
"loss": 1.1129,
"step": 4830
},
{
"epoch": 0.5002066969822241,
"grad_norm": 0.1519850343465805,
"learning_rate": 0.006,
"loss": 1.1144,
"step": 4840
},
{
"epoch": 0.5012401818933444,
"grad_norm": 0.24164853990077972,
"learning_rate": 0.006,
"loss": 1.1235,
"step": 4850
},
{
"epoch": 0.5022736668044646,
"grad_norm": 0.16202867031097412,
"learning_rate": 0.006,
"loss": 1.1054,
"step": 4860
},
{
"epoch": 0.5033071517155849,
"grad_norm": 0.13450877368450165,
"learning_rate": 0.006,
"loss": 1.1112,
"step": 4870
},
{
"epoch": 0.5043406366267053,
"grad_norm": 0.1484173834323883,
"learning_rate": 0.006,
"loss": 1.103,
"step": 4880
},
{
"epoch": 0.5053741215378256,
"grad_norm": 0.14877624809741974,
"learning_rate": 0.006,
"loss": 1.1353,
"step": 4890
},
{
"epoch": 0.5064076064489459,
"grad_norm": 0.14426067471504211,
"learning_rate": 0.006,
"loss": 1.1172,
"step": 4900
},
{
"epoch": 0.5074410913600661,
"grad_norm": 0.14011584222316742,
"learning_rate": 0.006,
"loss": 1.1171,
"step": 4910
},
{
"epoch": 0.5084745762711864,
"grad_norm": 0.20109255611896515,
"learning_rate": 0.006,
"loss": 1.1097,
"step": 4920
},
{
"epoch": 0.5095080611823067,
"grad_norm": 0.1302032619714737,
"learning_rate": 0.006,
"loss": 1.1092,
"step": 4930
},
{
"epoch": 0.5105415460934271,
"grad_norm": 0.13182514905929565,
"learning_rate": 0.006,
"loss": 1.1043,
"step": 4940
},
{
"epoch": 0.5115750310045474,
"grad_norm": 0.13346487283706665,
"learning_rate": 0.006,
"loss": 1.1133,
"step": 4950
},
{
"epoch": 0.5126085159156676,
"grad_norm": 0.13681770861148834,
"learning_rate": 0.006,
"loss": 1.1153,
"step": 4960
},
{
"epoch": 0.5136420008267879,
"grad_norm": 0.14637672901153564,
"learning_rate": 0.006,
"loss": 1.0984,
"step": 4970
},
{
"epoch": 0.5146754857379082,
"grad_norm": 0.14188942313194275,
"learning_rate": 0.006,
"loss": 1.1033,
"step": 4980
},
{
"epoch": 0.5157089706490285,
"grad_norm": 0.1467074155807495,
"learning_rate": 0.006,
"loss": 1.095,
"step": 4990
},
{
"epoch": 0.5167424555601489,
"grad_norm": 0.15010304749011993,
"learning_rate": 0.006,
"loss": 1.1029,
"step": 5000
},
{
"epoch": 0.5177759404712691,
"grad_norm": 0.13308066129684448,
"learning_rate": 0.006,
"loss": 1.1004,
"step": 5010
},
{
"epoch": 0.5188094253823894,
"grad_norm": 0.13878673315048218,
"learning_rate": 0.006,
"loss": 1.1055,
"step": 5020
},
{
"epoch": 0.5198429102935097,
"grad_norm": 0.14125706255435944,
"learning_rate": 0.006,
"loss": 1.0927,
"step": 5030
},
{
"epoch": 0.52087639520463,
"grad_norm": 0.1262088418006897,
"learning_rate": 0.006,
"loss": 1.1112,
"step": 5040
},
{
"epoch": 0.5219098801157503,
"grad_norm": 0.20926526188850403,
"learning_rate": 0.006,
"loss": 1.1169,
"step": 5050
},
{
"epoch": 0.5229433650268706,
"grad_norm": 0.2341676652431488,
"learning_rate": 0.006,
"loss": 1.1198,
"step": 5060
},
{
"epoch": 0.5239768499379909,
"grad_norm": 0.15894795954227448,
"learning_rate": 0.006,
"loss": 1.1017,
"step": 5070
},
{
"epoch": 0.5250103348491112,
"grad_norm": 0.14738234877586365,
"learning_rate": 0.006,
"loss": 1.1021,
"step": 5080
},
{
"epoch": 0.5260438197602315,
"grad_norm": 0.14371953904628754,
"learning_rate": 0.006,
"loss": 1.0978,
"step": 5090
},
{
"epoch": 0.5270773046713518,
"grad_norm": 0.13194578886032104,
"learning_rate": 0.006,
"loss": 1.1073,
"step": 5100
},
{
"epoch": 0.5281107895824722,
"grad_norm": 0.1411438137292862,
"learning_rate": 0.006,
"loss": 1.0931,
"step": 5110
},
{
"epoch": 0.5291442744935924,
"grad_norm": 0.1347801685333252,
"learning_rate": 0.006,
"loss": 1.0896,
"step": 5120
},
{
"epoch": 0.5301777594047127,
"grad_norm": 0.12243802100419998,
"learning_rate": 0.006,
"loss": 1.0912,
"step": 5130
},
{
"epoch": 0.531211244315833,
"grad_norm": 0.15618611872196198,
"learning_rate": 0.006,
"loss": 1.0925,
"step": 5140
},
{
"epoch": 0.5322447292269533,
"grad_norm": 0.133057102560997,
"learning_rate": 0.006,
"loss": 1.0937,
"step": 5150
},
{
"epoch": 0.5332782141380736,
"grad_norm": 0.14488716423511505,
"learning_rate": 0.006,
"loss": 1.1057,
"step": 5160
},
{
"epoch": 0.5343116990491938,
"grad_norm": 0.15518608689308167,
"learning_rate": 0.006,
"loss": 1.1073,
"step": 5170
},
{
"epoch": 0.5353451839603142,
"grad_norm": 0.14580237865447998,
"learning_rate": 0.006,
"loss": 1.1064,
"step": 5180
},
{
"epoch": 0.5363786688714345,
"grad_norm": 0.12775352597236633,
"learning_rate": 0.006,
"loss": 1.1043,
"step": 5190
},
{
"epoch": 0.5374121537825548,
"grad_norm": 0.15589575469493866,
"learning_rate": 0.006,
"loss": 1.0995,
"step": 5200
},
{
"epoch": 0.5384456386936751,
"grad_norm": 0.1348804533481598,
"learning_rate": 0.006,
"loss": 1.0959,
"step": 5210
},
{
"epoch": 0.5394791236047953,
"grad_norm": 0.1302793323993683,
"learning_rate": 0.006,
"loss": 1.0833,
"step": 5220
},
{
"epoch": 0.5405126085159156,
"grad_norm": 0.1520976573228836,
"learning_rate": 0.006,
"loss": 1.0991,
"step": 5230
},
{
"epoch": 0.541546093427036,
"grad_norm": 0.13820461928844452,
"learning_rate": 0.006,
"loss": 1.1039,
"step": 5240
},
{
"epoch": 0.5425795783381563,
"grad_norm": 0.13032610714435577,
"learning_rate": 0.006,
"loss": 1.0947,
"step": 5250
},
{
"epoch": 0.5436130632492766,
"grad_norm": 0.1633882224559784,
"learning_rate": 0.006,
"loss": 1.1009,
"step": 5260
},
{
"epoch": 0.5446465481603968,
"grad_norm": 0.2063656449317932,
"learning_rate": 0.006,
"loss": 1.0947,
"step": 5270
},
{
"epoch": 0.5456800330715171,
"grad_norm": 0.13242821395397186,
"learning_rate": 0.006,
"loss": 1.0904,
"step": 5280
},
{
"epoch": 0.5467135179826375,
"grad_norm": 0.15254896879196167,
"learning_rate": 0.006,
"loss": 1.0955,
"step": 5290
},
{
"epoch": 0.5477470028937578,
"grad_norm": 0.12627582252025604,
"learning_rate": 0.006,
"loss": 1.1101,
"step": 5300
},
{
"epoch": 0.5487804878048781,
"grad_norm": 0.132407084107399,
"learning_rate": 0.006,
"loss": 1.0947,
"step": 5310
},
{
"epoch": 0.5498139727159983,
"grad_norm": 0.1298685371875763,
"learning_rate": 0.006,
"loss": 1.0853,
"step": 5320
},
{
"epoch": 0.5508474576271186,
"grad_norm": 0.25235456228256226,
"learning_rate": 0.006,
"loss": 1.1048,
"step": 5330
},
{
"epoch": 0.5518809425382389,
"grad_norm": 0.15979771316051483,
"learning_rate": 0.006,
"loss": 1.0945,
"step": 5340
},
{
"epoch": 0.5529144274493593,
"grad_norm": 0.12209047377109528,
"learning_rate": 0.006,
"loss": 1.0918,
"step": 5350
},
{
"epoch": 0.5539479123604796,
"grad_norm": 0.14376087486743927,
"learning_rate": 0.006,
"loss": 1.1055,
"step": 5360
},
{
"epoch": 0.5549813972715998,
"grad_norm": 0.12132777273654938,
"learning_rate": 0.006,
"loss": 1.1009,
"step": 5370
},
{
"epoch": 0.5560148821827201,
"grad_norm": 0.12317829579114914,
"learning_rate": 0.006,
"loss": 1.0873,
"step": 5380
},
{
"epoch": 0.5570483670938404,
"grad_norm": 0.15228794515132904,
"learning_rate": 0.006,
"loss": 1.0883,
"step": 5390
},
{
"epoch": 0.5580818520049607,
"grad_norm": 0.11975687742233276,
"learning_rate": 0.006,
"loss": 1.1075,
"step": 5400
},
{
"epoch": 0.5591153369160811,
"grad_norm": 0.15219765901565552,
"learning_rate": 0.006,
"loss": 1.087,
"step": 5410
},
{
"epoch": 0.5601488218272013,
"grad_norm": 0.12764790654182434,
"learning_rate": 0.006,
"loss": 1.0928,
"step": 5420
},
{
"epoch": 0.5611823067383216,
"grad_norm": 0.18155109882354736,
"learning_rate": 0.006,
"loss": 1.0924,
"step": 5430
},
{
"epoch": 0.5622157916494419,
"grad_norm": 0.16299748420715332,
"learning_rate": 0.006,
"loss": 1.0923,
"step": 5440
},
{
"epoch": 0.5632492765605622,
"grad_norm": 0.1478397250175476,
"learning_rate": 0.006,
"loss": 1.0838,
"step": 5450
},
{
"epoch": 0.5642827614716825,
"grad_norm": 0.14866343140602112,
"learning_rate": 0.006,
"loss": 1.0861,
"step": 5460
},
{
"epoch": 0.5653162463828028,
"grad_norm": 0.15744392573833466,
"learning_rate": 0.006,
"loss": 1.0981,
"step": 5470
},
{
"epoch": 0.5663497312939231,
"grad_norm": 0.1575409471988678,
"learning_rate": 0.006,
"loss": 1.0972,
"step": 5480
},
{
"epoch": 0.5673832162050434,
"grad_norm": 0.2215876430273056,
"learning_rate": 0.006,
"loss": 1.0926,
"step": 5490
},
{
"epoch": 0.5684167011161637,
"grad_norm": 0.2060500532388687,
"learning_rate": 0.006,
"loss": 1.0885,
"step": 5500
},
{
"epoch": 0.569450186027284,
"grad_norm": 0.19035674631595612,
"learning_rate": 0.006,
"loss": 1.0852,
"step": 5510
},
{
"epoch": 0.5704836709384044,
"grad_norm": 0.15188254415988922,
"learning_rate": 0.006,
"loss": 1.0883,
"step": 5520
},
{
"epoch": 0.5715171558495246,
"grad_norm": 0.1428184062242508,
"learning_rate": 0.006,
"loss": 1.0777,
"step": 5530
},
{
"epoch": 0.5725506407606449,
"grad_norm": 0.12388137727975845,
"learning_rate": 0.006,
"loss": 1.0884,
"step": 5540
},
{
"epoch": 0.5735841256717652,
"grad_norm": 0.1246667355298996,
"learning_rate": 0.006,
"loss": 1.0807,
"step": 5550
},
{
"epoch": 0.5746176105828855,
"grad_norm": 0.15170446038246155,
"learning_rate": 0.006,
"loss": 1.0908,
"step": 5560
},
{
"epoch": 0.5756510954940058,
"grad_norm": 0.12245498597621918,
"learning_rate": 0.006,
"loss": 1.0915,
"step": 5570
},
{
"epoch": 0.576684580405126,
"grad_norm": 0.12406784296035767,
"learning_rate": 0.006,
"loss": 1.0821,
"step": 5580
},
{
"epoch": 0.5777180653162464,
"grad_norm": 0.173844575881958,
"learning_rate": 0.006,
"loss": 1.0879,
"step": 5590
},
{
"epoch": 0.5787515502273667,
"grad_norm": 0.1362677961587906,
"learning_rate": 0.006,
"loss": 1.0929,
"step": 5600
},
{
"epoch": 0.579785035138487,
"grad_norm": 0.13046661019325256,
"learning_rate": 0.006,
"loss": 1.0891,
"step": 5610
},
{
"epoch": 0.5808185200496073,
"grad_norm": 0.13165681064128876,
"learning_rate": 0.006,
"loss": 1.0842,
"step": 5620
},
{
"epoch": 0.5818520049607275,
"grad_norm": 0.1174607053399086,
"learning_rate": 0.006,
"loss": 1.0827,
"step": 5630
},
{
"epoch": 0.5828854898718479,
"grad_norm": 0.12023981660604477,
"learning_rate": 0.006,
"loss": 1.0725,
"step": 5640
},
{
"epoch": 0.5839189747829682,
"grad_norm": 0.16525237262248993,
"learning_rate": 0.006,
"loss": 1.0812,
"step": 5650
},
{
"epoch": 0.5849524596940885,
"grad_norm": 0.14079253375530243,
"learning_rate": 0.006,
"loss": 1.0859,
"step": 5660
},
{
"epoch": 0.5859859446052088,
"grad_norm": 0.1366300880908966,
"learning_rate": 0.006,
"loss": 1.0847,
"step": 5670
},
{
"epoch": 0.587019429516329,
"grad_norm": 0.12389519810676575,
"learning_rate": 0.006,
"loss": 1.0897,
"step": 5680
},
{
"epoch": 0.5880529144274493,
"grad_norm": 0.1193692535161972,
"learning_rate": 0.006,
"loss": 1.0727,
"step": 5690
},
{
"epoch": 0.5890863993385697,
"grad_norm": 0.14679639041423798,
"learning_rate": 0.006,
"loss": 1.0849,
"step": 5700
},
{
"epoch": 0.59011988424969,
"grad_norm": 0.13050967454910278,
"learning_rate": 0.006,
"loss": 1.0753,
"step": 5710
},
{
"epoch": 0.5911533691608103,
"grad_norm": 0.12082899361848831,
"learning_rate": 0.006,
"loss": 1.078,
"step": 5720
},
{
"epoch": 0.5921868540719305,
"grad_norm": 0.1468031257390976,
"learning_rate": 0.006,
"loss": 1.0866,
"step": 5730
},
{
"epoch": 0.5932203389830508,
"grad_norm": 0.12028497457504272,
"learning_rate": 0.006,
"loss": 1.0726,
"step": 5740
},
{
"epoch": 0.5942538238941711,
"grad_norm": 0.14379620552062988,
"learning_rate": 0.006,
"loss": 1.088,
"step": 5750
},
{
"epoch": 0.5952873088052915,
"grad_norm": 0.13077853620052338,
"learning_rate": 0.006,
"loss": 1.0743,
"step": 5760
},
{
"epoch": 0.5963207937164118,
"grad_norm": 0.14271710813045502,
"learning_rate": 0.006,
"loss": 1.078,
"step": 5770
},
{
"epoch": 0.597354278627532,
"grad_norm": 0.1357925534248352,
"learning_rate": 0.006,
"loss": 1.0777,
"step": 5780
},
{
"epoch": 0.5983877635386523,
"grad_norm": 0.12296663224697113,
"learning_rate": 0.006,
"loss": 1.0561,
"step": 5790
},
{
"epoch": 0.5994212484497726,
"grad_norm": 0.1592247039079666,
"learning_rate": 0.006,
"loss": 1.0731,
"step": 5800
},
{
"epoch": 0.600454733360893,
"grad_norm": 0.24744771420955658,
"learning_rate": 0.006,
"loss": 1.0703,
"step": 5810
},
{
"epoch": 0.6014882182720133,
"grad_norm": 0.18649180233478546,
"learning_rate": 0.006,
"loss": 1.0663,
"step": 5820
},
{
"epoch": 0.6025217031831335,
"grad_norm": 0.15970060229301453,
"learning_rate": 0.006,
"loss": 1.0814,
"step": 5830
},
{
"epoch": 0.6035551880942538,
"grad_norm": 0.15505217015743256,
"learning_rate": 0.006,
"loss": 1.0819,
"step": 5840
},
{
"epoch": 0.6045886730053741,
"grad_norm": 0.1281612068414688,
"learning_rate": 0.006,
"loss": 1.073,
"step": 5850
},
{
"epoch": 0.6056221579164944,
"grad_norm": 0.13322196900844574,
"learning_rate": 0.006,
"loss": 1.0851,
"step": 5860
},
{
"epoch": 0.6066556428276147,
"grad_norm": 0.1385001540184021,
"learning_rate": 0.006,
"loss": 1.0714,
"step": 5870
},
{
"epoch": 0.6076891277387351,
"grad_norm": 0.12118349224328995,
"learning_rate": 0.006,
"loss": 1.0744,
"step": 5880
},
{
"epoch": 0.6087226126498553,
"grad_norm": 0.12691213190555573,
"learning_rate": 0.006,
"loss": 1.0664,
"step": 5890
},
{
"epoch": 0.6097560975609756,
"grad_norm": 0.14859186112880707,
"learning_rate": 0.006,
"loss": 1.0821,
"step": 5900
},
{
"epoch": 0.6107895824720959,
"grad_norm": 0.14851497113704681,
"learning_rate": 0.006,
"loss": 1.08,
"step": 5910
},
{
"epoch": 0.6118230673832162,
"grad_norm": 0.1358513981103897,
"learning_rate": 0.006,
"loss": 1.0747,
"step": 5920
},
{
"epoch": 0.6128565522943366,
"grad_norm": 0.14147254824638367,
"learning_rate": 0.006,
"loss": 1.0805,
"step": 5930
},
{
"epoch": 0.6138900372054568,
"grad_norm": 0.13193708658218384,
"learning_rate": 0.006,
"loss": 1.079,
"step": 5940
},
{
"epoch": 0.6149235221165771,
"grad_norm": 0.1255359649658203,
"learning_rate": 0.006,
"loss": 1.0785,
"step": 5950
},
{
"epoch": 0.6159570070276974,
"grad_norm": 0.19004510343074799,
"learning_rate": 0.006,
"loss": 1.0728,
"step": 5960
},
{
"epoch": 0.6169904919388177,
"grad_norm": 0.2386738657951355,
"learning_rate": 0.006,
"loss": 1.0836,
"step": 5970
},
{
"epoch": 0.618023976849938,
"grad_norm": 0.14890602231025696,
"learning_rate": 0.006,
"loss": 1.0791,
"step": 5980
},
{
"epoch": 0.6190574617610582,
"grad_norm": 0.15145020186901093,
"learning_rate": 0.006,
"loss": 1.0717,
"step": 5990
},
{
"epoch": 0.6200909466721786,
"grad_norm": 0.12245503813028336,
"learning_rate": 0.006,
"loss": 1.064,
"step": 6000
},
{
"epoch": 0.6211244315832989,
"grad_norm": 0.13569168746471405,
"learning_rate": 0.006,
"loss": 1.0717,
"step": 6010
},
{
"epoch": 0.6221579164944192,
"grad_norm": 0.14238642156124115,
"learning_rate": 0.006,
"loss": 1.072,
"step": 6020
},
{
"epoch": 0.6231914014055395,
"grad_norm": 0.13332600891590118,
"learning_rate": 0.006,
"loss": 1.0821,
"step": 6030
},
{
"epoch": 0.6242248863166597,
"grad_norm": 0.12775711715221405,
"learning_rate": 0.006,
"loss": 1.0665,
"step": 6040
},
{
"epoch": 0.62525837122778,
"grad_norm": 0.12378199398517609,
"learning_rate": 0.006,
"loss": 1.0594,
"step": 6050
},
{
"epoch": 0.6262918561389004,
"grad_norm": 0.127638041973114,
"learning_rate": 0.006,
"loss": 1.064,
"step": 6060
},
{
"epoch": 0.6273253410500207,
"grad_norm": 0.1450446993112564,
"learning_rate": 0.006,
"loss": 1.0718,
"step": 6070
},
{
"epoch": 0.628358825961141,
"grad_norm": 0.14722521603107452,
"learning_rate": 0.006,
"loss": 1.0786,
"step": 6080
},
{
"epoch": 0.6293923108722612,
"grad_norm": 0.11357062309980392,
"learning_rate": 0.006,
"loss": 1.0713,
"step": 6090
},
{
"epoch": 0.6304257957833815,
"grad_norm": 0.1290910542011261,
"learning_rate": 0.006,
"loss": 1.0692,
"step": 6100
},
{
"epoch": 0.6314592806945019,
"grad_norm": 0.12275319546461105,
"learning_rate": 0.006,
"loss": 1.0599,
"step": 6110
},
{
"epoch": 0.6324927656056222,
"grad_norm": 0.14953821897506714,
"learning_rate": 0.006,
"loss": 1.0735,
"step": 6120
},
{
"epoch": 0.6335262505167425,
"grad_norm": 0.15243026614189148,
"learning_rate": 0.006,
"loss": 1.0641,
"step": 6130
},
{
"epoch": 0.6345597354278627,
"grad_norm": 0.17082957923412323,
"learning_rate": 0.006,
"loss": 1.0669,
"step": 6140
},
{
"epoch": 0.635593220338983,
"grad_norm": 0.14030581712722778,
"learning_rate": 0.006,
"loss": 1.0659,
"step": 6150
},
{
"epoch": 0.6366267052501033,
"grad_norm": 0.1736230105161667,
"learning_rate": 0.006,
"loss": 1.0698,
"step": 6160
},
{
"epoch": 0.6376601901612237,
"grad_norm": 0.12367543578147888,
"learning_rate": 0.006,
"loss": 1.0657,
"step": 6170
},
{
"epoch": 0.638693675072344,
"grad_norm": 0.14306339621543884,
"learning_rate": 0.006,
"loss": 1.0731,
"step": 6180
},
{
"epoch": 0.6397271599834642,
"grad_norm": 0.1417074203491211,
"learning_rate": 0.006,
"loss": 1.0704,
"step": 6190
},
{
"epoch": 0.6407606448945845,
"grad_norm": 0.12904278934001923,
"learning_rate": 0.006,
"loss": 1.0619,
"step": 6200
},
{
"epoch": 0.6417941298057048,
"grad_norm": 0.13394799828529358,
"learning_rate": 0.006,
"loss": 1.054,
"step": 6210
},
{
"epoch": 0.6428276147168251,
"grad_norm": 0.13006740808486938,
"learning_rate": 0.006,
"loss": 1.058,
"step": 6220
},
{
"epoch": 0.6438610996279455,
"grad_norm": 0.1411658525466919,
"learning_rate": 0.006,
"loss": 1.0688,
"step": 6230
},
{
"epoch": 0.6448945845390658,
"grad_norm": 0.12172463536262512,
"learning_rate": 0.006,
"loss": 1.0599,
"step": 6240
},
{
"epoch": 0.645928069450186,
"grad_norm": 0.1132531389594078,
"learning_rate": 0.006,
"loss": 1.0493,
"step": 6250
},
{
"epoch": 0.6469615543613063,
"grad_norm": 0.15614502131938934,
"learning_rate": 0.006,
"loss": 1.0637,
"step": 6260
},
{
"epoch": 0.6479950392724266,
"grad_norm": 0.14655426144599915,
"learning_rate": 0.006,
"loss": 1.0537,
"step": 6270
},
{
"epoch": 0.649028524183547,
"grad_norm": 0.1248323991894722,
"learning_rate": 0.006,
"loss": 1.0735,
"step": 6280
},
{
"epoch": 0.6500620090946673,
"grad_norm": 0.13113942742347717,
"learning_rate": 0.006,
"loss": 1.0561,
"step": 6290
},
{
"epoch": 0.6510954940057875,
"grad_norm": 0.11488605290651321,
"learning_rate": 0.006,
"loss": 1.0626,
"step": 6300
},
{
"epoch": 0.6521289789169078,
"grad_norm": 0.13470590114593506,
"learning_rate": 0.006,
"loss": 1.0629,
"step": 6310
},
{
"epoch": 0.6531624638280281,
"grad_norm": 0.13424867391586304,
"learning_rate": 0.006,
"loss": 1.0634,
"step": 6320
},
{
"epoch": 0.6541959487391484,
"grad_norm": 0.13190287351608276,
"learning_rate": 0.006,
"loss": 1.0697,
"step": 6330
},
{
"epoch": 0.6552294336502688,
"grad_norm": 0.12076553702354431,
"learning_rate": 0.006,
"loss": 1.0559,
"step": 6340
},
{
"epoch": 0.656262918561389,
"grad_norm": 0.12332470715045929,
"learning_rate": 0.006,
"loss": 1.0511,
"step": 6350
},
{
"epoch": 0.6572964034725093,
"grad_norm": 0.13250276446342468,
"learning_rate": 0.006,
"loss": 1.0772,
"step": 6360
},
{
"epoch": 0.6583298883836296,
"grad_norm": 0.12431011348962784,
"learning_rate": 0.006,
"loss": 1.0574,
"step": 6370
},
{
"epoch": 0.6593633732947499,
"grad_norm": 0.11310591548681259,
"learning_rate": 0.006,
"loss": 1.0548,
"step": 6380
},
{
"epoch": 0.6603968582058702,
"grad_norm": 0.12900187075138092,
"learning_rate": 0.006,
"loss": 1.0624,
"step": 6390
},
{
"epoch": 0.6614303431169904,
"grad_norm": 0.21367403864860535,
"learning_rate": 0.006,
"loss": 1.0681,
"step": 6400
},
{
"epoch": 0.6624638280281108,
"grad_norm": 0.17459499835968018,
"learning_rate": 0.006,
"loss": 1.0547,
"step": 6410
},
{
"epoch": 0.6634973129392311,
"grad_norm": 0.12943004071712494,
"learning_rate": 0.006,
"loss": 1.0659,
"step": 6420
},
{
"epoch": 0.6645307978503514,
"grad_norm": 0.12108403444290161,
"learning_rate": 0.006,
"loss": 1.0721,
"step": 6430
},
{
"epoch": 0.6655642827614717,
"grad_norm": 0.14485575258731842,
"learning_rate": 0.006,
"loss": 1.0502,
"step": 6440
},
{
"epoch": 0.6665977676725919,
"grad_norm": 0.12145304679870605,
"learning_rate": 0.006,
"loss": 1.0489,
"step": 6450
},
{
"epoch": 0.6676312525837123,
"grad_norm": 0.12879317998886108,
"learning_rate": 0.006,
"loss": 1.0603,
"step": 6460
},
{
"epoch": 0.6686647374948326,
"grad_norm": 0.13428542017936707,
"learning_rate": 0.006,
"loss": 1.0538,
"step": 6470
},
{
"epoch": 0.6696982224059529,
"grad_norm": 0.13670314848423004,
"learning_rate": 0.006,
"loss": 1.053,
"step": 6480
},
{
"epoch": 0.6707317073170732,
"grad_norm": 0.12726275622844696,
"learning_rate": 0.006,
"loss": 1.0555,
"step": 6490
},
{
"epoch": 0.6717651922281934,
"grad_norm": 0.1293756067752838,
"learning_rate": 0.006,
"loss": 1.0579,
"step": 6500
},
{
"epoch": 0.6727986771393137,
"grad_norm": 0.15011049807071686,
"learning_rate": 0.006,
"loss": 1.0528,
"step": 6510
},
{
"epoch": 0.6738321620504341,
"grad_norm": 0.1232418343424797,
"learning_rate": 0.006,
"loss": 1.0474,
"step": 6520
},
{
"epoch": 0.6748656469615544,
"grad_norm": 0.13466301560401917,
"learning_rate": 0.006,
"loss": 1.0561,
"step": 6530
},
{
"epoch": 0.6758991318726747,
"grad_norm": 0.18297189474105835,
"learning_rate": 0.006,
"loss": 1.0569,
"step": 6540
},
{
"epoch": 0.6769326167837949,
"grad_norm": 0.13041065633296967,
"learning_rate": 0.006,
"loss": 1.0614,
"step": 6550
},
{
"epoch": 0.6779661016949152,
"grad_norm": 0.15428952872753143,
"learning_rate": 0.006,
"loss": 1.0526,
"step": 6560
},
{
"epoch": 0.6789995866060355,
"grad_norm": 0.1226574033498764,
"learning_rate": 0.006,
"loss": 1.0476,
"step": 6570
},
{
"epoch": 0.6800330715171559,
"grad_norm": 0.1592113971710205,
"learning_rate": 0.006,
"loss": 1.057,
"step": 6580
},
{
"epoch": 0.6810665564282762,
"grad_norm": 0.131056547164917,
"learning_rate": 0.006,
"loss": 1.0643,
"step": 6590
},
{
"epoch": 0.6821000413393964,
"grad_norm": 0.17470133304595947,
"learning_rate": 0.006,
"loss": 1.0545,
"step": 6600
},
{
"epoch": 0.6831335262505167,
"grad_norm": 0.14145781099796295,
"learning_rate": 0.006,
"loss": 1.0459,
"step": 6610
},
{
"epoch": 0.684167011161637,
"grad_norm": 0.11432395130395889,
"learning_rate": 0.006,
"loss": 1.0454,
"step": 6620
},
{
"epoch": 0.6852004960727573,
"grad_norm": 0.12126456946134567,
"learning_rate": 0.006,
"loss": 1.0467,
"step": 6630
},
{
"epoch": 0.6862339809838777,
"grad_norm": 0.12609295547008514,
"learning_rate": 0.006,
"loss": 1.0734,
"step": 6640
},
{
"epoch": 0.687267465894998,
"grad_norm": 0.15360839664936066,
"learning_rate": 0.006,
"loss": 1.0424,
"step": 6650
},
{
"epoch": 0.6883009508061182,
"grad_norm": 0.13050951063632965,
"learning_rate": 0.006,
"loss": 1.0532,
"step": 6660
},
{
"epoch": 0.6893344357172385,
"grad_norm": 0.11345363408327103,
"learning_rate": 0.006,
"loss": 1.0616,
"step": 6670
},
{
"epoch": 0.6903679206283588,
"grad_norm": 0.12678933143615723,
"learning_rate": 0.006,
"loss": 1.0472,
"step": 6680
},
{
"epoch": 0.6914014055394792,
"grad_norm": 0.12220112234354019,
"learning_rate": 0.006,
"loss": 1.048,
"step": 6690
},
{
"epoch": 0.6924348904505995,
"grad_norm": 0.12634249031543732,
"learning_rate": 0.006,
"loss": 1.0489,
"step": 6700
},
{
"epoch": 0.6934683753617197,
"grad_norm": 0.14149808883666992,
"learning_rate": 0.006,
"loss": 1.0548,
"step": 6710
},
{
"epoch": 0.69450186027284,
"grad_norm": 0.12240725010633469,
"learning_rate": 0.006,
"loss": 1.0466,
"step": 6720
},
{
"epoch": 0.6955353451839603,
"grad_norm": 0.1353718340396881,
"learning_rate": 0.006,
"loss": 1.0423,
"step": 6730
},
{
"epoch": 0.6965688300950806,
"grad_norm": 0.18347500264644623,
"learning_rate": 0.006,
"loss": 1.0485,
"step": 6740
},
{
"epoch": 0.697602315006201,
"grad_norm": 0.12348928302526474,
"learning_rate": 0.006,
"loss": 1.0573,
"step": 6750
},
{
"epoch": 0.6986357999173212,
"grad_norm": 0.1428414285182953,
"learning_rate": 0.006,
"loss": 1.0412,
"step": 6760
},
{
"epoch": 0.6996692848284415,
"grad_norm": 0.1410830169916153,
"learning_rate": 0.006,
"loss": 1.0569,
"step": 6770
},
{
"epoch": 0.7007027697395618,
"grad_norm": 0.14577074348926544,
"learning_rate": 0.005935620400122873,
"loss": 1.0546,
"step": 6780
},
{
"epoch": 0.7017362546506821,
"grad_norm": 0.14377078413963318,
"learning_rate": 0.005842197295969981,
"loss": 1.045,
"step": 6790
},
{
"epoch": 0.7027697395618024,
"grad_norm": 0.12980251014232635,
"learning_rate": 0.005750244615429315,
"loss": 1.0595,
"step": 6800
},
{
"epoch": 0.7038032244729227,
"grad_norm": 0.13310031592845917,
"learning_rate": 0.005659739214915371,
"loss": 1.0492,
"step": 6810
},
{
"epoch": 0.704836709384043,
"grad_norm": 0.12173794209957123,
"learning_rate": 0.005570658315108789,
"loss": 1.0409,
"step": 6820
},
{
"epoch": 0.7058701942951633,
"grad_norm": 0.17054332792758942,
"learning_rate": 0.005482979495223033,
"loss": 1.0505,
"step": 6830
},
{
"epoch": 0.7069036792062836,
"grad_norm": 0.18403039872646332,
"learning_rate": 0.005396680687361299,
"loss": 1.0462,
"step": 6840
},
{
"epoch": 0.7079371641174039,
"grad_norm": 0.13148552179336548,
"learning_rate": 0.005311740170962234,
"loss": 1.049,
"step": 6850
},
{
"epoch": 0.7089706490285241,
"grad_norm": 0.13557471334934235,
"learning_rate": 0.005228136567333095,
"loss": 1.0348,
"step": 6860
},
{
"epoch": 0.7100041339396445,
"grad_norm": 0.11231327056884766,
"learning_rate": 0.005145848834268932,
"loss": 1.0309,
"step": 6870
},
{
"epoch": 0.7110376188507648,
"grad_norm": 0.14308036863803864,
"learning_rate": 0.005064856260756483,
"loss": 1.024,
"step": 6880
},
{
"epoch": 0.7120711037618851,
"grad_norm": 0.11384549736976624,
"learning_rate": 0.0049851384617614035,
"loss": 1.0255,
"step": 6890
},
{
"epoch": 0.7131045886730054,
"grad_norm": 0.1373758465051651,
"learning_rate": 0.004906675373097566,
"loss": 1.0248,
"step": 6900
},
{
"epoch": 0.7141380735841256,
"grad_norm": 0.13972637057304382,
"learning_rate": 0.00482944724637709,
"loss": 1.0197,
"step": 6910
},
{
"epoch": 0.7151715584952459,
"grad_norm": 0.11777855455875397,
"learning_rate": 0.004753434644039878,
"loss": 1.0197,
"step": 6920
},
{
"epoch": 0.7162050434063663,
"grad_norm": 0.10751146823167801,
"learning_rate": 0.004678618434461362,
"loss": 1.0272,
"step": 6930
},
{
"epoch": 0.7172385283174866,
"grad_norm": 0.11851785331964493,
"learning_rate": 0.004604979787137271,
"loss": 1.0109,
"step": 6940
},
{
"epoch": 0.7182720132286069,
"grad_norm": 0.1341877430677414,
"learning_rate": 0.004532500167944173,
"loss": 1.0106,
"step": 6950
},
{
"epoch": 0.7193054981397271,
"grad_norm": 0.11008955538272858,
"learning_rate": 0.004461161334474619,
"loss": 1.0152,
"step": 6960
},
{
"epoch": 0.7203389830508474,
"grad_norm": 0.16443827748298645,
"learning_rate": 0.0043909453314457106,
"loss": 1.0118,
"step": 6970
},
{
"epoch": 0.7213724679619677,
"grad_norm": 0.11731505393981934,
"learning_rate": 0.004321834486179928,
"loss": 1.0081,
"step": 6980
},
{
"epoch": 0.7224059528730881,
"grad_norm": 0.11518187075853348,
"learning_rate": 0.004253811404157096,
"loss": 1.004,
"step": 6990
},
{
"epoch": 0.7234394377842084,
"grad_norm": 0.10781227797269821,
"learning_rate": 0.004186858964636351,
"loss": 0.9879,
"step": 7000
},
{
"epoch": 0.7244729226953287,
"grad_norm": 0.12845131754875183,
"learning_rate": 0.004120960316347017,
"loss": 1.0148,
"step": 7010
},
{
"epoch": 0.7255064076064489,
"grad_norm": 0.11902609467506409,
"learning_rate": 0.004056098873247311,
"loss": 1.0012,
"step": 7020
},
{
"epoch": 0.7265398925175692,
"grad_norm": 0.11711709946393967,
"learning_rate": 0.003992258310349796,
"loss": 1.0134,
"step": 7030
},
{
"epoch": 0.7275733774286895,
"grad_norm": 0.12396284192800522,
"learning_rate": 0.003929422559612546,
"loss": 1.0023,
"step": 7040
},
{
"epoch": 0.7286068623398099,
"grad_norm": 0.10830879211425781,
"learning_rate": 0.0038675758058949712,
"loss": 1.0126,
"step": 7050
},
{
"epoch": 0.7296403472509302,
"grad_norm": 0.11904972046613693,
"learning_rate": 0.003806702482977311,
"loss": 0.9993,
"step": 7060
},
{
"epoch": 0.7306738321620504,
"grad_norm": 0.14758719503879547,
"learning_rate": 0.0037467872696427614,
"loss": 0.9919,
"step": 7070
},
{
"epoch": 0.7317073170731707,
"grad_norm": 0.10925702750682831,
"learning_rate": 0.0036878150858212825,
"loss": 0.9991,
"step": 7080
},
{
"epoch": 0.732740801984291,
"grad_norm": 0.1237645223736763,
"learning_rate": 0.0036297710887940875,
"loss": 1.004,
"step": 7090
},
{
"epoch": 0.7337742868954114,
"grad_norm": 0.12318721413612366,
"learning_rate": 0.003572640669457879,
"loss": 0.9833,
"step": 7100
},
{
"epoch": 0.7348077718065317,
"grad_norm": 0.10309872031211853,
"learning_rate": 0.003516409448647884,
"loss": 0.9919,
"step": 7110
},
{
"epoch": 0.7358412567176519,
"grad_norm": 0.10170971602201462,
"learning_rate": 0.003461063273518752,
"loss": 0.9908,
"step": 7120
},
{
"epoch": 0.7368747416287722,
"grad_norm": 0.13282230496406555,
"learning_rate": 0.0034065882139824315,
"loss": 0.9849,
"step": 7130
},
{
"epoch": 0.7379082265398925,
"grad_norm": 0.10541505366563797,
"learning_rate": 0.0033529705592020968,
"loss": 0.9893,
"step": 7140
},
{
"epoch": 0.7389417114510128,
"grad_norm": 0.10669828951358795,
"learning_rate": 0.0033001968141412714,
"loss": 0.9841,
"step": 7150
},
{
"epoch": 0.7399751963621332,
"grad_norm": 0.11185015738010406,
"learning_rate": 0.0032482536961672545,
"loss": 0.9888,
"step": 7160
},
{
"epoch": 0.7410086812732534,
"grad_norm": 0.11848310381174088,
"learning_rate": 0.003197128131708016,
"loss": 0.9753,
"step": 7170
},
{
"epoch": 0.7420421661843737,
"grad_norm": 0.12162330746650696,
"learning_rate": 0.0031468072529617073,
"loss": 0.9896,
"step": 7180
},
{
"epoch": 0.743075651095494,
"grad_norm": 0.10726076364517212,
"learning_rate": 0.0030972783946579593,
"loss": 0.967,
"step": 7190
},
{
"epoch": 0.7441091360066143,
"grad_norm": 0.11569386720657349,
"learning_rate": 0.003048529090870162,
"loss": 0.9691,
"step": 7200
},
{
"epoch": 0.7451426209177346,
"grad_norm": 0.11116993427276611,
"learning_rate": 0.0030005470718779117,
"loss": 0.9783,
"step": 7210
},
{
"epoch": 0.7461761058288549,
"grad_norm": 0.10824961215257645,
"learning_rate": 0.002953320261078841,
"loss": 0.9589,
"step": 7220
},
{
"epoch": 0.7472095907399752,
"grad_norm": 0.1097089946269989,
"learning_rate": 0.00290683677194906,
"loss": 0.9707,
"step": 7230
},
{
"epoch": 0.7482430756510955,
"grad_norm": 0.13413402438163757,
"learning_rate": 0.002861084905051434,
"loss": 0.9583,
"step": 7240
},
{
"epoch": 0.7492765605622158,
"grad_norm": 0.10966306924819946,
"learning_rate": 0.0028160531450909495,
"loss": 0.9656,
"step": 7250
},
{
"epoch": 0.7503100454733361,
"grad_norm": 0.10541026294231415,
"learning_rate": 0.0027717301580164277,
"loss": 0.9568,
"step": 7260
},
{
"epoch": 0.7513435303844563,
"grad_norm": 0.10642149299383163,
"learning_rate": 0.002728104788167857,
"loss": 0.973,
"step": 7270
},
{
"epoch": 0.7523770152955767,
"grad_norm": 0.10307744145393372,
"learning_rate": 0.0026851660554686207,
"loss": 0.973,
"step": 7280
},
{
"epoch": 0.753410500206697,
"grad_norm": 0.10917676985263824,
"learning_rate": 0.002642903152661921,
"loss": 0.9739,
"step": 7290
},
{
"epoch": 0.7544439851178173,
"grad_norm": 0.12075357139110565,
"learning_rate": 0.002601305442590699,
"loss": 0.9668,
"step": 7300
},
{
"epoch": 0.7554774700289376,
"grad_norm": 0.10885104537010193,
"learning_rate": 0.0025603624555203654,
"loss": 0.9732,
"step": 7310
},
{
"epoch": 0.7565109549400578,
"grad_norm": 0.11787907034158707,
"learning_rate": 0.0025200638865036733,
"loss": 0.9597,
"step": 7320
},
{
"epoch": 0.7575444398511781,
"grad_norm": 0.11566518247127533,
"learning_rate": 0.0024803995927870632,
"loss": 0.9595,
"step": 7330
},
{
"epoch": 0.7585779247622985,
"grad_norm": 0.09844059497117996,
"learning_rate": 0.002441359591257831,
"loss": 0.9561,
"step": 7340
},
{
"epoch": 0.7596114096734188,
"grad_norm": 0.1304013878107071,
"learning_rate": 0.002402934055931478,
"loss": 0.9454,
"step": 7350
},
{
"epoch": 0.7606448945845391,
"grad_norm": 0.1041470617055893,
"learning_rate": 0.002365113315478606,
"loss": 0.9453,
"step": 7360
},
{
"epoch": 0.7616783794956593,
"grad_norm": 0.10921091586351395,
"learning_rate": 0.00232788785079074,
"loss": 0.9644,
"step": 7370
},
{
"epoch": 0.7627118644067796,
"grad_norm": 0.12734107673168182,
"learning_rate": 0.0022912482925844615,
"loss": 0.9491,
"step": 7380
},
{
"epoch": 0.7637453493179,
"grad_norm": 0.11103381216526031,
"learning_rate": 0.0022551854190432517,
"loss": 0.9379,
"step": 7390
},
{
"epoch": 0.7647788342290203,
"grad_norm": 0.10525085031986237,
"learning_rate": 0.0022196901534964524,
"loss": 0.956,
"step": 7400
},
{
"epoch": 0.7658123191401406,
"grad_norm": 0.10325931012630463,
"learning_rate": 0.0021847535621347543,
"loss": 0.9478,
"step": 7410
},
{
"epoch": 0.7668458040512609,
"grad_norm": 0.10702349990606308,
"learning_rate": 0.0021503668517616494,
"loss": 0.9408,
"step": 7420
},
{
"epoch": 0.7678792889623811,
"grad_norm": 0.10343906283378601,
"learning_rate": 0.002116521367580266,
"loss": 0.9505,
"step": 7430
},
{
"epoch": 0.7689127738735014,
"grad_norm": 0.11631717532873154,
"learning_rate": 0.0020832085910150425,
"loss": 0.9483,
"step": 7440
},
{
"epoch": 0.7699462587846218,
"grad_norm": 0.114792600274086,
"learning_rate": 0.0020504201375676876,
"loss": 0.9425,
"step": 7450
},
{
"epoch": 0.7709797436957421,
"grad_norm": 0.10078386217355728,
"learning_rate": 0.0020181477547068817,
"loss": 0.955,
"step": 7460
},
{
"epoch": 0.7720132286068624,
"grad_norm": 0.10332197695970535,
"learning_rate": 0.0019863833197911983,
"loss": 0.9488,
"step": 7470
},
{
"epoch": 0.7730467135179826,
"grad_norm": 0.1078406572341919,
"learning_rate": 0.001955118838024712,
"loss": 0.9536,
"step": 7480
},
{
"epoch": 0.7740801984291029,
"grad_norm": 0.1059098169207573,
"learning_rate": 0.0019243464404447924,
"loss": 0.9396,
"step": 7490
},
{
"epoch": 0.7751136833402232,
"grad_norm": 0.10873686522245407,
"learning_rate": 0.0018940583819415564,
"loss": 0.9412,
"step": 7500
},
{
"epoch": 0.7761471682513436,
"grad_norm": 0.10201577097177505,
"learning_rate": 0.0018642470393085062,
"loss": 0.9406,
"step": 7510
},
{
"epoch": 0.7771806531624639,
"grad_norm": 0.10363386571407318,
"learning_rate": 0.0018349049093238407,
"loss": 0.9349,
"step": 7520
},
{
"epoch": 0.7782141380735841,
"grad_norm": 0.10290543735027313,
"learning_rate": 0.001806024606861967,
"loss": 0.9433,
"step": 7530
},
{
"epoch": 0.7792476229847044,
"grad_norm": 0.09915654361248016,
"learning_rate": 0.0017775988630347405,
"loss": 0.9333,
"step": 7540
},
{
"epoch": 0.7802811078958247,
"grad_norm": 0.10667795687913895,
"learning_rate": 0.0017496205233619542,
"loss": 0.9282,
"step": 7550
},
{
"epoch": 0.781314592806945,
"grad_norm": 0.0981353297829628,
"learning_rate": 0.001722082545970627,
"loss": 0.9443,
"step": 7560
},
{
"epoch": 0.7823480777180654,
"grad_norm": 0.09341371059417725,
"learning_rate": 0.0016949779998226353,
"loss": 0.9255,
"step": 7570
},
{
"epoch": 0.7833815626291856,
"grad_norm": 0.09437743574380875,
"learning_rate": 0.0016683000629702357,
"loss": 0.9374,
"step": 7580
},
{
"epoch": 0.7844150475403059,
"grad_norm": 0.10137518495321274,
"learning_rate": 0.001642042020839051,
"loss": 0.9434,
"step": 7590
},
{
"epoch": 0.7854485324514262,
"grad_norm": 0.10156513750553131,
"learning_rate": 0.001616197264538076,
"loss": 0.9282,
"step": 7600
},
{
"epoch": 0.7864820173625465,
"grad_norm": 0.10588687658309937,
"learning_rate": 0.0015907592891962843,
"loss": 0.9329,
"step": 7610
},
{
"epoch": 0.7875155022736668,
"grad_norm": 0.10155037045478821,
"learning_rate": 0.0015657216923254177,
"loss": 0.9268,
"step": 7620
},
{
"epoch": 0.788548987184787,
"grad_norm": 0.10379229485988617,
"learning_rate": 0.0015410781722085428,
"loss": 0.9437,
"step": 7630
},
{
"epoch": 0.7895824720959074,
"grad_norm": 0.10199972987174988,
"learning_rate": 0.0015168225263139693,
"loss": 0.9425,
"step": 7640
},
{
"epoch": 0.7906159570070277,
"grad_norm": 0.10469642281532288,
"learning_rate": 0.0014929486497341341,
"loss": 0.924,
"step": 7650
},
{
"epoch": 0.791649441918148,
"grad_norm": 0.09616528451442719,
"learning_rate": 0.0014694505336490583,
"loss": 0.931,
"step": 7660
},
{
"epoch": 0.7926829268292683,
"grad_norm": 0.10645470768213272,
"learning_rate": 0.001446322263813983,
"loss": 0.9266,
"step": 7670
},
{
"epoch": 0.7937164117403885,
"grad_norm": 0.10266025364398956,
"learning_rate": 0.0014235580190708143,
"loss": 0.9251,
"step": 7680
},
{
"epoch": 0.7947498966515089,
"grad_norm": 0.10014299303293228,
"learning_rate": 0.0014011520698829947,
"loss": 0.9268,
"step": 7690
},
{
"epoch": 0.7957833815626292,
"grad_norm": 0.10017547011375427,
"learning_rate": 0.0013790987768934346,
"loss": 0.9345,
"step": 7700
},
{
"epoch": 0.7968168664737495,
"grad_norm": 0.09650063514709473,
"learning_rate": 0.0013573925895051418,
"loss": 0.919,
"step": 7710
},
{
"epoch": 0.7978503513848698,
"grad_norm": 0.11645365506410599,
"learning_rate": 0.00133602804448419,
"loss": 0.9256,
"step": 7720
},
{
"epoch": 0.79888383629599,
"grad_norm": 0.09640100598335266,
"learning_rate": 0.0013149997645846788,
"loss": 0.9252,
"step": 7730
},
{
"epoch": 0.7999173212071103,
"grad_norm": 0.09893805533647537,
"learning_rate": 0.0012943024571953315,
"loss": 0.9209,
"step": 7740
},
{
"epoch": 0.8009508061182307,
"grad_norm": 0.10975757986307144,
"learning_rate": 0.001273930913007398,
"loss": 0.9259,
"step": 7750
},
{
"epoch": 0.801984291029351,
"grad_norm": 0.09872964024543762,
"learning_rate": 0.0012538800047035218,
"loss": 0.9139,
"step": 7760
},
{
"epoch": 0.8030177759404713,
"grad_norm": 0.10294859856367111,
"learning_rate": 0.0012341446856672469,
"loss": 0.9219,
"step": 7770
},
{
"epoch": 0.8040512608515916,
"grad_norm": 0.09699690341949463,
"learning_rate": 0.0012147199887128318,
"loss": 0.9171,
"step": 7780
},
{
"epoch": 0.8050847457627118,
"grad_norm": 0.10195200890302658,
"learning_rate": 0.0011956010248350593,
"loss": 0.9181,
"step": 7790
},
{
"epoch": 0.8061182306738321,
"grad_norm": 0.10153238475322723,
"learning_rate": 0.0011767829819787205,
"loss": 0.9122,
"step": 7800
},
{
"epoch": 0.8071517155849525,
"grad_norm": 0.09719067811965942,
"learning_rate": 0.0011582611238274692,
"loss": 0.9261,
"step": 7810
},
{
"epoch": 0.8081852004960728,
"grad_norm": 0.09960606694221497,
"learning_rate": 0.0011400307886117367,
"loss": 0.9159,
"step": 7820
},
{
"epoch": 0.8092186854071931,
"grad_norm": 0.10018911957740784,
"learning_rate": 0.001122087387935411,
"loss": 0.9043,
"step": 7830
},
{
"epoch": 0.8102521703183133,
"grad_norm": 0.09987188875675201,
"learning_rate": 0.0011044264056209812,
"loss": 0.923,
"step": 7840
},
{
"epoch": 0.8112856552294336,
"grad_norm": 0.10447081923484802,
"learning_rate": 0.0010870433965728626,
"loss": 0.912,
"step": 7850
},
{
"epoch": 0.812319140140554,
"grad_norm": 0.09891889244318008,
"learning_rate": 0.0010699339856586065,
"loss": 0.9231,
"step": 7860
},
{
"epoch": 0.8133526250516743,
"grad_norm": 0.11543388664722443,
"learning_rate": 0.0010530938666077263,
"loss": 0.9189,
"step": 7870
},
{
"epoch": 0.8143861099627946,
"grad_norm": 0.10272625833749771,
"learning_rate": 0.0010365188009278474,
"loss": 0.9218,
"step": 7880
},
{
"epoch": 0.8154195948739148,
"grad_norm": 0.11380596458911896,
"learning_rate": 0.0010202046168379232,
"loss": 0.9089,
"step": 7890
},
{
"epoch": 0.8164530797850351,
"grad_norm": 0.10323087871074677,
"learning_rate": 0.0010041472082182382,
"loss": 0.9053,
"step": 7900
},
{
"epoch": 0.8174865646961554,
"grad_norm": 0.10377652198076248,
"learning_rate": 0.0009883425335769375,
"loss": 0.9167,
"step": 7910
},
{
"epoch": 0.8185200496072758,
"grad_norm": 0.09777701646089554,
"learning_rate": 0.0009727866150328237,
"loss": 0.9134,
"step": 7920
},
{
"epoch": 0.8195535345183961,
"grad_norm": 0.10352311283349991,
"learning_rate": 0.0009574755373141626,
"loss": 0.9203,
"step": 7930
},
{
"epoch": 0.8205870194295163,
"grad_norm": 0.1033698320388794,
"learning_rate": 0.0009424054467732486,
"loss": 0.916,
"step": 7940
},
{
"epoch": 0.8216205043406366,
"grad_norm": 0.10148289799690247,
"learning_rate": 0.0009275725504164789,
"loss": 0.9055,
"step": 7950
},
{
"epoch": 0.8226539892517569,
"grad_norm": 0.1043614000082016,
"learning_rate": 0.0009129731149496945,
"loss": 0.9103,
"step": 7960
},
{
"epoch": 0.8236874741628772,
"grad_norm": 0.10638295859098434,
"learning_rate": 0.0008986034658385467,
"loss": 0.9095,
"step": 7970
},
{
"epoch": 0.8247209590739976,
"grad_norm": 0.09755829721689224,
"learning_rate": 0.0008844599863836532,
"loss": 0.911,
"step": 7980
},
{
"epoch": 0.8257544439851178,
"grad_norm": 0.09516286849975586,
"learning_rate": 0.0008705391168103099,
"loss": 0.8951,
"step": 7990
},
{
"epoch": 0.8267879288962381,
"grad_norm": 0.09857280552387238,
"learning_rate": 0.0008568373533725312,
"loss": 0.9129,
"step": 8000
},
{
"epoch": 0.8278214138073584,
"grad_norm": 0.09658730775117874,
"learning_rate": 0.0008433512474711912,
"loss": 0.9123,
"step": 8010
},
{
"epoch": 0.8288548987184787,
"grad_norm": 0.09653860330581665,
"learning_rate": 0.0008300774047860455,
"loss": 0.8981,
"step": 8020
},
{
"epoch": 0.829888383629599,
"grad_norm": 0.10416875779628754,
"learning_rate": 0.0008170124844214137,
"loss": 0.9173,
"step": 8030
},
{
"epoch": 0.8309218685407193,
"grad_norm": 0.09936217218637466,
"learning_rate": 0.0008041531980653093,
"loss": 0.9129,
"step": 8040
},
{
"epoch": 0.8319553534518396,
"grad_norm": 0.10089723765850067,
"learning_rate": 0.0007914963091618035,
"loss": 0.9138,
"step": 8050
},
{
"epoch": 0.8329888383629599,
"grad_norm": 0.09607098996639252,
"learning_rate": 0.0007790386320964167,
"loss": 0.8986,
"step": 8060
},
{
"epoch": 0.8340223232740802,
"grad_norm": 0.09406198561191559,
"learning_rate": 0.0007667770313943293,
"loss": 0.9062,
"step": 8070
},
{
"epoch": 0.8350558081852005,
"grad_norm": 0.1027892455458641,
"learning_rate": 0.0007547084209312148,
"loss": 0.9082,
"step": 8080
},
{
"epoch": 0.8360892930963207,
"grad_norm": 0.09503350406885147,
"learning_rate": 0.0007428297631564922,
"loss": 0.915,
"step": 8090
},
{
"epoch": 0.8371227780074411,
"grad_norm": 0.11514163017272949,
"learning_rate": 0.000731138068328804,
"loss": 0.8963,
"step": 8100
},
{
"epoch": 0.8381562629185614,
"grad_norm": 0.09855664521455765,
"learning_rate": 0.0007196303937635279,
"loss": 0.8902,
"step": 8110
},
{
"epoch": 0.8391897478296817,
"grad_norm": 0.0985339879989624,
"learning_rate": 0.000708303843092133,
"loss": 0.9087,
"step": 8120
},
{
"epoch": 0.840223232740802,
"grad_norm": 0.10324376076459885,
"learning_rate": 0.0006971555655331906,
"loss": 0.9129,
"step": 8130
},
{
"epoch": 0.8412567176519223,
"grad_norm": 0.10233034193515778,
"learning_rate": 0.0006861827551748618,
"loss": 0.8974,
"step": 8140
},
{
"epoch": 0.8422902025630425,
"grad_norm": 0.09102931618690491,
"learning_rate": 0.0006753826502686767,
"loss": 0.8945,
"step": 8150
},
{
"epoch": 0.8433236874741629,
"grad_norm": 0.09884581714868546,
"learning_rate": 0.0006647525325344291,
"loss": 0.9076,
"step": 8160
},
{
"epoch": 0.8443571723852832,
"grad_norm": 0.10044873505830765,
"learning_rate": 0.0006542897264760131,
"loss": 0.897,
"step": 8170
},
{
"epoch": 0.8453906572964035,
"grad_norm": 0.10729029029607773,
"learning_rate": 0.000643991598708026,
"loss": 0.9061,
"step": 8180
},
{
"epoch": 0.8464241422075238,
"grad_norm": 0.09865026921033859,
"learning_rate": 0.0006338555572929715,
"loss": 0.9198,
"step": 8190
},
{
"epoch": 0.847457627118644,
"grad_norm": 0.0979701355099678,
"learning_rate": 0.0006238790510888947,
"loss": 0.9027,
"step": 8200
},
{
"epoch": 0.8484911120297643,
"grad_norm": 0.10607071965932846,
"learning_rate": 0.0006140595691072842,
"loss": 0.9005,
"step": 8210
},
{
"epoch": 0.8495245969408847,
"grad_norm": 0.09750446677207947,
"learning_rate": 0.0006043946398810819,
"loss": 0.8955,
"step": 8220
},
{
"epoch": 0.850558081852005,
"grad_norm": 0.10503144562244415,
"learning_rate": 0.0005948818308426382,
"loss": 0.8987,
"step": 8230
},
{
"epoch": 0.8515915667631253,
"grad_norm": 0.09300017356872559,
"learning_rate": 0.0005855187477114587,
"loss": 0.8943,
"step": 8240
},
{
"epoch": 0.8526250516742455,
"grad_norm": 0.09706971049308777,
"learning_rate": 0.0005763030338915879,
"loss": 0.8942,
"step": 8250
},
{
"epoch": 0.8536585365853658,
"grad_norm": 0.11037931591272354,
"learning_rate": 0.000567232369878477,
"loss": 0.8958,
"step": 8260
},
{
"epoch": 0.8546920214964862,
"grad_norm": 0.09634577482938766,
"learning_rate": 0.0005583044726751868,
"loss": 0.8976,
"step": 8270
},
{
"epoch": 0.8557255064076065,
"grad_norm": 0.10242386907339096,
"learning_rate": 0.0005495170952177806,
"loss": 0.8902,
"step": 8280
},
{
"epoch": 0.8567589913187268,
"grad_norm": 0.10059966146945953,
"learning_rate": 0.00054086802580976,
"loss": 0.8896,
"step": 8290
},
{
"epoch": 0.857792476229847,
"grad_norm": 0.09923160821199417,
"learning_rate": 0.0005323550875654025,
"loss": 0.9073,
"step": 8300
},
{
"epoch": 0.8588259611409673,
"grad_norm": 0.09592054039239883,
"learning_rate": 0.000523976137861861,
"loss": 0.896,
"step": 8310
},
{
"epoch": 0.8598594460520876,
"grad_norm": 0.09789422899484634,
"learning_rate": 0.0005157290677998864,
"loss": 0.8878,
"step": 8320
},
{
"epoch": 0.860892930963208,
"grad_norm": 0.09961704909801483,
"learning_rate": 0.0005076118016730388,
"loss": 0.9077,
"step": 8330
},
{
"epoch": 0.8619264158743283,
"grad_norm": 0.10076375305652618,
"learning_rate": 0.0004996222964452523,
"loss": 0.8979,
"step": 8340
},
{
"epoch": 0.8629599007854485,
"grad_norm": 0.0950293093919754,
"learning_rate": 0.000491758541236623,
"loss": 0.8966,
"step": 8350
},
{
"epoch": 0.8639933856965688,
"grad_norm": 0.10502886027097702,
"learning_rate": 0.0004840185568172905,
"loss": 0.9094,
"step": 8360
},
{
"epoch": 0.8650268706076891,
"grad_norm": 0.09783428907394409,
"learning_rate": 0.0004764003951092846,
"loss": 0.904,
"step": 8370
},
{
"epoch": 0.8660603555188094,
"grad_norm": 0.10145466774702072,
"learning_rate": 0.00046890213869621406,
"loss": 0.8912,
"step": 8380
},
{
"epoch": 0.8670938404299298,
"grad_norm": 0.09995938092470169,
"learning_rate": 0.000461521900340671,
"loss": 0.91,
"step": 8390
},
{
"epoch": 0.86812732534105,
"grad_norm": 0.09882347285747528,
"learning_rate": 0.0004542578225092324,
"loss": 0.8935,
"step": 8400
},
{
"epoch": 0.8691608102521703,
"grad_norm": 0.1028551384806633,
"learning_rate": 0.00044710807690493667,
"loss": 0.9005,
"step": 8410
},
{
"epoch": 0.8701942951632906,
"grad_norm": 0.09993860870599747,
"learning_rate": 0.00044007086400711984,
"loss": 0.8883,
"step": 8420
},
{
"epoch": 0.8712277800744109,
"grad_norm": 0.09587843716144562,
"learning_rate": 0.0004331444126184934,
"loss": 0.896,
"step": 8430
},
{
"epoch": 0.8722612649855312,
"grad_norm": 0.10601107031106949,
"learning_rate": 0.0004263269794193516,
"loss": 0.9027,
"step": 8440
},
{
"epoch": 0.8732947498966515,
"grad_norm": 0.09365765005350113,
"learning_rate": 0.00041961684852879507,
"loss": 0.8928,
"step": 8450
},
{
"epoch": 0.8743282348077718,
"grad_norm": 0.10556492209434509,
"learning_rate": 0.0004130123310728602,
"loss": 0.8872,
"step": 8460
},
{
"epoch": 0.8753617197188921,
"grad_norm": 0.09618480503559113,
"learning_rate": 0.0004065117647594465,
"loss": 0.8917,
"step": 8470
},
{
"epoch": 0.8763952046300124,
"grad_norm": 0.100033700466156,
"learning_rate": 0.0004001135134599342,
"loss": 0.8949,
"step": 8480
},
{
"epoch": 0.8774286895411327,
"grad_norm": 0.09465377777814865,
"learning_rate": 0.00039381596679738596,
"loss": 0.8931,
"step": 8490
},
{
"epoch": 0.8784621744522529,
"grad_norm": 0.09475496411323547,
"learning_rate": 0.00038761753974123146,
"loss": 0.8856,
"step": 8500
},
{
"epoch": 0.8794956593633733,
"grad_norm": 0.09431330114603043,
"learning_rate": 0.0003815166722083307,
"loss": 0.8921,
"step": 8510
},
{
"epoch": 0.8805291442744936,
"grad_norm": 0.11336002498865128,
"learning_rate": 0.00037551182867031637,
"loss": 0.8944,
"step": 8520
},
{
"epoch": 0.8815626291856139,
"grad_norm": 0.09444057941436768,
"learning_rate": 0.00036960149776711637,
"loss": 0.8907,
"step": 8530
},
{
"epoch": 0.8825961140967342,
"grad_norm": 0.09592239558696747,
"learning_rate": 0.0003637841919265597,
"loss": 0.8993,
"step": 8540
},
{
"epoch": 0.8836295990078545,
"grad_norm": 0.09695941209793091,
"learning_rate": 0.0003580584469899685,
"loss": 0.8948,
"step": 8550
},
{
"epoch": 0.8846630839189747,
"grad_norm": 0.10035042464733124,
"learning_rate": 0.0003524228218436444,
"loss": 0.8897,
"step": 8560
},
{
"epoch": 0.8856965688300951,
"grad_norm": 0.11594710499048233,
"learning_rate": 0.0003468758980561539,
"loss": 0.896,
"step": 8570
},
{
"epoch": 0.8867300537412154,
"grad_norm": 0.10346604883670807,
"learning_rate": 0.00034141627952132355,
"loss": 0.8885,
"step": 8580
},
{
"epoch": 0.8877635386523357,
"grad_norm": 0.09101969003677368,
"learning_rate": 0.0003360425921068532,
"loss": 0.8867,
"step": 8590
},
{
"epoch": 0.888797023563456,
"grad_norm": 0.09395504742860794,
"learning_rate": 0.00033075348330846094,
"loss": 0.8871,
"step": 8600
},
{
"epoch": 0.8898305084745762,
"grad_norm": 0.09610182791948318,
"learning_rate": 0.00032554762190947073,
"loss": 0.8913,
"step": 8610
},
{
"epoch": 0.8908639933856966,
"grad_norm": 0.11420192569494247,
"learning_rate": 0.0003204236976457585,
"loss": 0.8897,
"step": 8620
},
{
"epoch": 0.8918974782968169,
"grad_norm": 0.10842256993055344,
"learning_rate": 0.00031538042087597126,
"loss": 0.8887,
"step": 8630
},
{
"epoch": 0.8929309632079372,
"grad_norm": 0.10306430608034134,
"learning_rate": 0.00031041652225693745,
"loss": 0.8882,
"step": 8640
},
{
"epoch": 0.8939644481190575,
"grad_norm": 0.09563726931810379,
"learning_rate": 0.00030553075242418533,
"loss": 0.8884,
"step": 8650
},
{
"epoch": 0.8949979330301777,
"grad_norm": 0.09344979375600815,
"learning_rate": 0.0003007218816774904,
"loss": 0.8853,
"step": 8660
},
{
"epoch": 0.896031417941298,
"grad_norm": 0.09084004163742065,
"learning_rate": 0.0002959886996713722,
"loss": 0.8844,
"step": 8670
},
{
"epoch": 0.8970649028524184,
"grad_norm": 0.0981561467051506,
"learning_rate": 0.0002913300151104616,
"loss": 0.8987,
"step": 8680
},
{
"epoch": 0.8980983877635387,
"grad_norm": 0.09380181133747101,
"learning_rate": 0.00028674465544966445,
"loss": 0.8881,
"step": 8690
},
{
"epoch": 0.899131872674659,
"grad_norm": 0.10047593712806702,
"learning_rate": 0.00028223146659904236,
"loss": 0.8949,
"step": 8700
},
{
"epoch": 0.9001653575857792,
"grad_norm": 0.10739541053771973,
"learning_rate": 0.0002777893126333406,
"loss": 0.8969,
"step": 8710
},
{
"epoch": 0.9011988424968995,
"grad_norm": 0.09939514100551605,
"learning_rate": 0.0002734170755060864,
"loss": 0.8808,
"step": 8720
},
{
"epoch": 0.9022323274080198,
"grad_norm": 0.09798012673854828,
"learning_rate": 0.0002691136547681875,
"loss": 0.8898,
"step": 8730
},
{
"epoch": 0.9032658123191402,
"grad_norm": 0.09667760878801346,
"learning_rate": 0.00026487796729096044,
"loss": 0.8996,
"step": 8740
},
{
"epoch": 0.9042992972302605,
"grad_norm": 0.09401620179414749,
"learning_rate": 0.0002607089469935173,
"loss": 0.8797,
"step": 8750
},
{
"epoch": 0.9053327821413807,
"grad_norm": 0.09637662023305893,
"learning_rate": 0.0002566055445744437,
"loss": 0.885,
"step": 8760
},
{
"epoch": 0.906366267052501,
"grad_norm": 0.0979810431599617,
"learning_rate": 0.00025256672724770034,
"loss": 0.8765,
"step": 8770
},
{
"epoch": 0.9073997519636213,
"grad_norm": 0.09577486664056778,
"learning_rate": 0.0002485914784826802,
"loss": 0.878,
"step": 8780
},
{
"epoch": 0.9084332368747416,
"grad_norm": 0.09690549224615097,
"learning_rate": 0.0002446787977483582,
"loss": 0.8951,
"step": 8790
},
{
"epoch": 0.909466721785862,
"grad_norm": 0.09132607281208038,
"learning_rate": 0.00024082770026146767,
"loss": 0.8922,
"step": 8800
},
{
"epoch": 0.9105002066969822,
"grad_norm": 0.10059286653995514,
"learning_rate": 0.00023703721673863946,
"loss": 0.8961,
"step": 8810
},
{
"epoch": 0.9115336916081025,
"grad_norm": 0.09476503729820251,
"learning_rate": 0.00023330639315244326,
"loss": 0.8891,
"step": 8820
},
{
"epoch": 0.9125671765192228,
"grad_norm": 0.097599096596241,
"learning_rate": 0.00022963429049126816,
"loss": 0.8877,
"step": 8830
},
{
"epoch": 0.9136006614303431,
"grad_norm": 0.09217467904090881,
"learning_rate": 0.00022601998452298264,
"loss": 0.893,
"step": 8840
},
{
"epoch": 0.9146341463414634,
"grad_norm": 0.11072806268930435,
"learning_rate": 0.00022246256556231452,
"loss": 0.8943,
"step": 8850
},
{
"epoch": 0.9156676312525837,
"grad_norm": 0.10245074331760406,
"learning_rate": 0.00021896113824189194,
"loss": 0.9008,
"step": 8860
},
{
"epoch": 0.916701116163704,
"grad_norm": 0.0995858684182167,
"learning_rate": 0.00021551482128688842,
"loss": 0.8831,
"step": 8870
},
{
"epoch": 0.9177346010748243,
"grad_norm": 0.09560049325227737,
"learning_rate": 0.0002121227472932145,
"loss": 0.8978,
"step": 8880
},
{
"epoch": 0.9187680859859446,
"grad_norm": 0.09800171852111816,
"learning_rate": 0.00020878406250920078,
"loss": 0.8993,
"step": 8890
},
{
"epoch": 0.9198015708970649,
"grad_norm": 0.09587955474853516,
"learning_rate": 0.00020549792662071688,
"loss": 0.8947,
"step": 8900
},
{
"epoch": 0.9208350558081853,
"grad_norm": 0.09761986881494522,
"learning_rate": 0.00020226351253967265,
"loss": 0.8968,
"step": 8910
},
{
"epoch": 0.9218685407193055,
"grad_norm": 0.10232323408126831,
"learning_rate": 0.00019908000619584844,
"loss": 0.8743,
"step": 8920
},
{
"epoch": 0.9229020256304258,
"grad_norm": 0.09598308801651001,
"learning_rate": 0.00019594660633200127,
"loss": 0.8886,
"step": 8930
},
{
"epoch": 0.9239355105415461,
"grad_norm": 0.09552961587905884,
"learning_rate": 0.0001928625243021967,
"loss": 0.898,
"step": 8940
},
{
"epoch": 0.9249689954526664,
"grad_norm": 0.09763569384813309,
"learning_rate": 0.000189826983873314,
"loss": 0.8872,
"step": 8950
},
{
"epoch": 0.9260024803637867,
"grad_norm": 0.10132598876953125,
"learning_rate": 0.00018683922102967604,
"loss": 0.8844,
"step": 8960
},
{
"epoch": 0.927035965274907,
"grad_norm": 0.10523135215044022,
"learning_rate": 0.00018389848378075436,
"loss": 0.8886,
"step": 8970
},
{
"epoch": 0.9280694501860273,
"grad_norm": 0.10074016451835632,
"learning_rate": 0.0001810040319719,
"loss": 0.8868,
"step": 8980
},
{
"epoch": 0.9291029350971476,
"grad_norm": 0.09313800185918808,
"learning_rate": 0.00017815513709805425,
"loss": 0.8944,
"step": 8990
},
{
"epoch": 0.9301364200082679,
"grad_norm": 0.09687028080224991,
"learning_rate": 0.00017535108212039094,
"loss": 0.8764,
"step": 9000
},
{
"epoch": 0.9311699049193882,
"grad_norm": 0.09959317743778229,
"learning_rate": 0.00017259116128584488,
"loss": 0.8879,
"step": 9010
},
{
"epoch": 0.9322033898305084,
"grad_norm": 0.09753499180078506,
"learning_rate": 0.0001698746799494807,
"loss": 0.8878,
"step": 9020
},
{
"epoch": 0.9332368747416288,
"grad_norm": 0.10160617530345917,
"learning_rate": 0.00016720095439965764,
"loss": 0.8878,
"step": 9030
},
{
"epoch": 0.9342703596527491,
"grad_norm": 0.09427917003631592,
"learning_rate": 0.00016456931168594587,
"loss": 0.8917,
"step": 9040
},
{
"epoch": 0.9353038445638694,
"grad_norm": 0.09126557409763336,
"learning_rate": 0.0001619790894497517,
"loss": 0.8756,
"step": 9050
},
{
"epoch": 0.9363373294749897,
"grad_norm": 0.10263457894325256,
"learning_rate": 0.0001594296357576083,
"loss": 0.8968,
"step": 9060
},
{
"epoch": 0.9373708143861099,
"grad_norm": 0.09004776179790497,
"learning_rate": 0.00015692030893709055,
"loss": 0.8747,
"step": 9070
},
{
"epoch": 0.9384042992972302,
"grad_norm": 0.09706980735063553,
"learning_rate": 0.0001544504774153122,
"loss": 0.8782,
"step": 9080
},
{
"epoch": 0.9394377842083506,
"grad_norm": 0.11443266272544861,
"learning_rate": 0.00015201951955996554,
"loss": 0.8954,
"step": 9090
},
{
"epoch": 0.9404712691194709,
"grad_norm": 0.09983173757791519,
"learning_rate": 0.00014962682352286227,
"loss": 0.8839,
"step": 9100
},
{
"epoch": 0.9415047540305912,
"grad_norm": 0.09485773742198944,
"learning_rate": 0.00014727178708593758,
"loss": 0.8873,
"step": 9110
},
{
"epoch": 0.9425382389417114,
"grad_norm": 0.09533528983592987,
"learning_rate": 0.00014495381750967773,
"loss": 0.8803,
"step": 9120
},
{
"epoch": 0.9435717238528317,
"grad_norm": 0.09723920375108719,
"learning_rate": 0.00014267233138393328,
"loss": 0.894,
"step": 9130
},
{
"epoch": 0.944605208763952,
"grad_norm": 0.10340782254934311,
"learning_rate": 0.00014042675448108058,
"loss": 0.8876,
"step": 9140
},
{
"epoch": 0.9456386936750724,
"grad_norm": 0.10088169574737549,
"learning_rate": 0.00013821652161149427,
"loss": 0.8797,
"step": 9150
},
{
"epoch": 0.9466721785861927,
"grad_norm": 0.09791452437639236,
"learning_rate": 0.0001360410764812946,
"loss": 0.8897,
"step": 9160
},
{
"epoch": 0.9477056634973129,
"grad_norm": 0.0989762470126152,
"learning_rate": 0.00013389987155233373,
"loss": 0.885,
"step": 9170
},
{
"epoch": 0.9487391484084332,
"grad_norm": 0.09686373174190521,
"learning_rate": 0.0001317923679043859,
"loss": 0.88,
"step": 9180
},
{
"epoch": 0.9497726333195535,
"grad_norm": 0.094807930290699,
"learning_rate": 0.00012971803509950622,
"loss": 0.8877,
"step": 9190
},
{
"epoch": 0.9508061182306738,
"grad_norm": 0.0951213538646698,
"learning_rate": 0.000127676351048525,
"loss": 0.8858,
"step": 9200
},
{
"epoch": 0.9518396031417942,
"grad_norm": 0.09922289848327637,
"learning_rate": 0.00012566680187964274,
"loss": 0.8826,
"step": 9210
},
{
"epoch": 0.9528730880529144,
"grad_norm": 0.09398606419563293,
"learning_rate": 0.0001236888818090938,
"loss": 0.8817,
"step": 9220
},
{
"epoch": 0.9539065729640347,
"grad_norm": 0.1025305688381195,
"learning_rate": 0.0001217420930138456,
"loss": 0.8809,
"step": 9230
},
{
"epoch": 0.954940057875155,
"grad_norm": 0.0998004898428917,
"learning_rate": 0.00011982594550630144,
"loss": 0.8878,
"step": 9240
},
{
"epoch": 0.9559735427862753,
"grad_norm": 0.09735240042209625,
"learning_rate": 0.00011793995701097545,
"loss": 0.8713,
"step": 9250
},
{
"epoch": 0.9570070276973957,
"grad_norm": 0.09462594240903854,
"learning_rate": 0.00011608365284310853,
"loss": 0.8932,
"step": 9260
},
{
"epoch": 0.958040512608516,
"grad_norm": 0.0939006358385086,
"learning_rate": 0.00011425656578919491,
"loss": 0.8865,
"step": 9270
},
{
"epoch": 0.9590739975196362,
"grad_norm": 0.10327833890914917,
"learning_rate": 0.00011245823598938918,
"loss": 0.885,
"step": 9280
},
{
"epoch": 0.9601074824307565,
"grad_norm": 0.10130874067544937,
"learning_rate": 0.00011068821082176394,
"loss": 0.8737,
"step": 9290
},
{
"epoch": 0.9611409673418768,
"grad_norm": 0.0958571583032608,
"learning_rate": 0.00010894604478838941,
"loss": 0.8836,
"step": 9300
},
{
"epoch": 0.9621744522529971,
"grad_norm": 0.10654882341623306,
"learning_rate": 0.00010723129940320598,
"loss": 0.8808,
"step": 9310
},
{
"epoch": 0.9632079371641175,
"grad_norm": 0.10343246906995773,
"learning_rate": 0.00010554354308166149,
"loss": 0.8838,
"step": 9320
},
{
"epoch": 0.9642414220752377,
"grad_norm": 0.09667105227708817,
"learning_rate": 0.00010388235103208581,
"loss": 0.8944,
"step": 9330
},
{
"epoch": 0.965274906986358,
"grad_norm": 0.09733343869447708,
"learning_rate": 0.00010224730514877476,
"loss": 0.8807,
"step": 9340
},
{
"epoch": 0.9663083918974783,
"grad_norm": 0.09809593111276627,
"learning_rate": 0.0001006379939067572,
"loss": 0.8824,
"step": 9350
},
{
"epoch": 0.9673418768085986,
"grad_norm": 0.09950044006109238,
"learning_rate": 9.905401225821806e-05,
"loss": 0.8895,
"step": 9360
},
{
"epoch": 0.9683753617197189,
"grad_norm": 0.09585009515285492,
"learning_rate": 9.749496153055196e-05,
"loss": 0.871,
"step": 9370
},
{
"epoch": 0.9694088466308391,
"grad_norm": 0.09470534324645996,
"learning_rate": 9.596044932602107e-05,
"loss": 0.8805,
"step": 9380
},
{
"epoch": 0.9704423315419595,
"grad_norm": 0.09828654676675797,
"learning_rate": 9.445008942299262e-05,
"loss": 0.8796,
"step": 9390
},
{
"epoch": 0.9714758164530798,
"grad_norm": 0.09540653228759766,
"learning_rate": 9.296350167873062e-05,
"loss": 0.8769,
"step": 9400
},
{
"epoch": 0.9725093013642001,
"grad_norm": 0.09175027906894684,
"learning_rate": 9.150031193371763e-05,
"loss": 0.8823,
"step": 9410
},
{
"epoch": 0.9735427862753204,
"grad_norm": 0.09416258335113525,
"learning_rate": 9.00601519174826e-05,
"loss": 0.8826,
"step": 9420
},
{
"epoch": 0.9745762711864406,
"grad_norm": 0.102226123213768,
"learning_rate": 8.864265915591077e-05,
"loss": 0.8813,
"step": 9430
},
{
"epoch": 0.975609756097561,
"grad_norm": 0.1030995100736618,
"learning_rate": 8.724747688001238e-05,
"loss": 0.89,
"step": 9440
},
{
"epoch": 0.9766432410086813,
"grad_norm": 0.09412838518619537,
"learning_rate": 8.587425393612774e-05,
"loss": 0.8869,
"step": 9450
},
{
"epoch": 0.9776767259198016,
"grad_norm": 0.09783417731523514,
"learning_rate": 8.45226446975449e-05,
"loss": 0.8745,
"step": 9460
},
{
"epoch": 0.9787102108309219,
"grad_norm": 0.09457999467849731,
"learning_rate": 8.31923089775091e-05,
"loss": 0.8854,
"step": 9470
},
{
"epoch": 0.9797436957420421,
"grad_norm": 0.0977085754275322,
"learning_rate": 8.188291194360097e-05,
"loss": 0.8743,
"step": 9480
},
{
"epoch": 0.9807771806531624,
"grad_norm": 0.09691467881202698,
"learning_rate": 8.059412403346259e-05,
"loss": 0.8798,
"step": 9490
},
{
"epoch": 0.9818106655642828,
"grad_norm": 0.09877024590969086,
"learning_rate": 7.932562087184977e-05,
"loss": 0.8785,
"step": 9500
},
{
"epoch": 0.9828441504754031,
"grad_norm": 0.09752603620290756,
"learning_rate": 7.807708318899015e-05,
"loss": 0.8787,
"step": 9510
},
{
"epoch": 0.9838776353865234,
"grad_norm": 0.10295706242322922,
"learning_rate": 7.684819674022599e-05,
"loss": 0.8782,
"step": 9520
},
{
"epoch": 0.9849111202976436,
"grad_norm": 0.09535824507474899,
"learning_rate": 7.563865222692196e-05,
"loss": 0.8861,
"step": 9530
},
{
"epoch": 0.9859446052087639,
"grad_norm": 0.09367913007736206,
"learning_rate": 7.444814521861768e-05,
"loss": 0.8791,
"step": 9540
},
{
"epoch": 0.9869780901198842,
"grad_norm": 0.10222353786230087,
"learning_rate": 7.327637607640558e-05,
"loss": 0.884,
"step": 9550
},
{
"epoch": 0.9880115750310046,
"grad_norm": 0.1052263155579567,
"learning_rate": 7.212304987751476e-05,
"loss": 0.8766,
"step": 9560
},
{
"epoch": 0.9890450599421249,
"grad_norm": 0.09064412862062454,
"learning_rate": 7.098787634108176e-05,
"loss": 0.8846,
"step": 9570
},
{
"epoch": 0.9900785448532451,
"grad_norm": 0.09217038005590439,
"learning_rate": 6.98705697550898e-05,
"loss": 0.878,
"step": 9580
},
{
"epoch": 0.9911120297643654,
"grad_norm": 0.08925709873437881,
"learning_rate": 6.87708489044578e-05,
"loss": 0.8659,
"step": 9590
},
{
"epoch": 0.9921455146754857,
"grad_norm": 0.09898435324430466,
"learning_rate": 6.768843700026141e-05,
"loss": 0.8891,
"step": 9600
},
{
"epoch": 0.993178999586606,
"grad_norm": 0.09614672511816025,
"learning_rate": 6.662306161006783e-05,
"loss": 0.8828,
"step": 9610
},
{
"epoch": 0.9942124844977264,
"grad_norm": 0.09297246485948563,
"learning_rate": 6.557445458936731e-05,
"loss": 0.8905,
"step": 9620
},
{
"epoch": 0.9952459694088466,
"grad_norm": 0.10573872923851013,
"learning_rate": 6.45423520140839e-05,
"loss": 0.8876,
"step": 9630
},
{
"epoch": 0.9962794543199669,
"grad_norm": 0.09352509677410126,
"learning_rate": 6.352649411414819e-05,
"loss": 0.8856,
"step": 9640
},
{
"epoch": 0.9973129392310872,
"grad_norm": 0.0931333377957344,
"learning_rate": 6.252662520811583e-05,
"loss": 0.8848,
"step": 9650
},
{
"epoch": 0.9983464241422075,
"grad_norm": 0.10176722705364227,
"learning_rate": 6.154249363881505e-05,
"loss": 0.8792,
"step": 9660
},
{
"epoch": 0.9993799090533279,
"grad_norm": 0.09999672323465347,
"learning_rate": 6.057385171000694e-05,
"loss": 0.8789,
"step": 9670
}
],
"logging_steps": 10,
"max_steps": 9676,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.24620217896713e+16,
"train_batch_size": 1024,
"trial_name": null,
"trial_params": null
}