muhtasham's picture
Model save
81eb4d0 verified
raw
history blame
113 kB
{
"best_metric": 0.22935771942138672,
"best_model_checkpoint": "output/output__lora/checkpoint-600",
"epoch": 1.7010935601458081,
"eval_steps": 100,
"global_step": 700,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002430133657351154,
"grad_norm": 2.0297858715057373,
"learning_rate": 0.0,
"loss": 0.3878,
"step": 1
},
{
"epoch": 0.004860267314702308,
"grad_norm": 2.0333988666534424,
"learning_rate": 8.859191006777897e-06,
"loss": 0.4074,
"step": 2
},
{
"epoch": 0.007290400972053463,
"grad_norm": 2.1134026050567627,
"learning_rate": 1.4041485532469073e-05,
"loss": 0.3097,
"step": 3
},
{
"epoch": 0.009720534629404616,
"grad_norm": 2.1784684658050537,
"learning_rate": 1.7718382013555794e-05,
"loss": 0.4578,
"step": 4
},
{
"epoch": 0.012150668286755772,
"grad_norm": 1.0290217399597168,
"learning_rate": 2.0570404496611053e-05,
"loss": 0.3389,
"step": 5
},
{
"epoch": 0.014580801944106925,
"grad_norm": 3.486976385116577,
"learning_rate": 2.2900676539246968e-05,
"loss": 0.2617,
"step": 6
},
{
"epoch": 0.01701093560145808,
"grad_norm": 1.7177977561950684,
"learning_rate": 2.4870893478326387e-05,
"loss": 0.5574,
"step": 7
},
{
"epoch": 0.019441069258809233,
"grad_norm": 0.8118329048156738,
"learning_rate": 2.6577573020333684e-05,
"loss": 0.4099,
"step": 8
},
{
"epoch": 0.02187120291616039,
"grad_norm": 2.2652881145477295,
"learning_rate": 2.8082971064938146e-05,
"loss": 0.3207,
"step": 9
},
{
"epoch": 0.024301336573511544,
"grad_norm": 1.2344919443130493,
"learning_rate": 2.9429595503388953e-05,
"loss": 0.2961,
"step": 10
},
{
"epoch": 0.026731470230862697,
"grad_norm": 1.2169722318649292,
"learning_rate": 3.064776548439465e-05,
"loss": 0.3689,
"step": 11
},
{
"epoch": 0.02916160388821385,
"grad_norm": 1.162546992301941,
"learning_rate": 3.1759867546024865e-05,
"loss": 0.351,
"step": 12
},
{
"epoch": 0.031591737545565005,
"grad_norm": 2.5353000164031982,
"learning_rate": 3.2782902272079295e-05,
"loss": 0.2424,
"step": 13
},
{
"epoch": 0.03402187120291616,
"grad_norm": 1.6772927045822144,
"learning_rate": 3.373008448510428e-05,
"loss": 0.2366,
"step": 14
},
{
"epoch": 0.03645200486026731,
"grad_norm": 1.5499287843704224,
"learning_rate": 3.4611890029080124e-05,
"loss": 0.3358,
"step": 15
},
{
"epoch": 0.038882138517618466,
"grad_norm": 1.383988618850708,
"learning_rate": 3.543676402711159e-05,
"loss": 0.3495,
"step": 16
},
{
"epoch": 0.041312272174969626,
"grad_norm": 1.4687221050262451,
"learning_rate": 3.621161404374383e-05,
"loss": 0.3973,
"step": 17
},
{
"epoch": 0.04374240583232078,
"grad_norm": 1.1256321668624878,
"learning_rate": 3.694216207171603e-05,
"loss": 0.2556,
"step": 18
},
{
"epoch": 0.046172539489671933,
"grad_norm": 0.9001079797744751,
"learning_rate": 3.76332012245438e-05,
"loss": 0.2283,
"step": 19
},
{
"epoch": 0.04860267314702309,
"grad_norm": 2.164158582687378,
"learning_rate": 3.8288786510166846e-05,
"loss": 0.2065,
"step": 20
},
{
"epoch": 0.05103280680437424,
"grad_norm": 0.6704838871955872,
"learning_rate": 3.8912379010795455e-05,
"loss": 0.3622,
"step": 21
},
{
"epoch": 0.053462940461725394,
"grad_norm": 1.380202293395996,
"learning_rate": 3.9506956491172545e-05,
"loss": 0.2949,
"step": 22
},
{
"epoch": 0.05589307411907655,
"grad_norm": 1.281702995300293,
"learning_rate": 4.007509939970292e-05,
"loss": 0.1766,
"step": 23
},
{
"epoch": 0.0583232077764277,
"grad_norm": 1.581787347793579,
"learning_rate": 4.061905855280276e-05,
"loss": 0.2676,
"step": 24
},
{
"epoch": 0.060753341433778855,
"grad_norm": 1.8878554105758667,
"learning_rate": 4.1140808993222106e-05,
"loss": 0.2572,
"step": 25
},
{
"epoch": 0.06318347509113001,
"grad_norm": 0.6384598612785339,
"learning_rate": 4.164209327885719e-05,
"loss": 0.231,
"step": 26
},
{
"epoch": 0.06561360874848117,
"grad_norm": 0.9673507809638977,
"learning_rate": 4.2124456597407214e-05,
"loss": 0.1772,
"step": 27
},
{
"epoch": 0.06804374240583232,
"grad_norm": 1.0678174495697021,
"learning_rate": 4.258927549188218e-05,
"loss": 0.2025,
"step": 28
},
{
"epoch": 0.07047387606318348,
"grad_norm": 0.9189561009407043,
"learning_rate": 4.303778154313212e-05,
"loss": 0.2345,
"step": 29
},
{
"epoch": 0.07290400972053462,
"grad_norm": 1.3609440326690674,
"learning_rate": 4.347108103585803e-05,
"loss": 0.2201,
"step": 30
},
{
"epoch": 0.07533414337788578,
"grad_norm": 1.3813296556472778,
"learning_rate": 4.389017139879164e-05,
"loss": 0.2438,
"step": 31
},
{
"epoch": 0.07776427703523693,
"grad_norm": 1.1185611486434937,
"learning_rate": 4.429595503388948e-05,
"loss": 0.1972,
"step": 32
},
{
"epoch": 0.08019441069258809,
"grad_norm": 0.876745343208313,
"learning_rate": 4.468925101686371e-05,
"loss": 0.2754,
"step": 33
},
{
"epoch": 0.08262454434993925,
"grad_norm": 0.7372068762779236,
"learning_rate": 4.507080505052173e-05,
"loss": 0.2719,
"step": 34
},
{
"epoch": 0.0850546780072904,
"grad_norm": 1.4229792356491089,
"learning_rate": 4.544129797493744e-05,
"loss": 0.3219,
"step": 35
},
{
"epoch": 0.08748481166464156,
"grad_norm": 1.4566582441329956,
"learning_rate": 4.5801353078493936e-05,
"loss": 0.1257,
"step": 36
},
{
"epoch": 0.0899149453219927,
"grad_norm": 1.217944860458374,
"learning_rate": 4.615154240700883e-05,
"loss": 0.3138,
"step": 37
},
{
"epoch": 0.09234507897934387,
"grad_norm": 1.882925033569336,
"learning_rate": 4.6492392231321696e-05,
"loss": 0.219,
"step": 38
},
{
"epoch": 0.09477521263669501,
"grad_norm": 0.8372556567192078,
"learning_rate": 4.682438780454837e-05,
"loss": 0.27,
"step": 39
},
{
"epoch": 0.09720534629404617,
"grad_norm": 0.9709153771400452,
"learning_rate": 4.714797751694474e-05,
"loss": 0.2759,
"step": 40
},
{
"epoch": 0.09963547995139732,
"grad_norm": 1.1802537441253662,
"learning_rate": 4.7463576537657414e-05,
"loss": 0.2411,
"step": 41
},
{
"epoch": 0.10206561360874848,
"grad_norm": 0.7289960980415344,
"learning_rate": 4.777157001757336e-05,
"loss": 0.2755,
"step": 42
},
{
"epoch": 0.10449574726609964,
"grad_norm": 0.8493523001670837,
"learning_rate": 4.8072315915252694e-05,
"loss": 0.3177,
"step": 43
},
{
"epoch": 0.10692588092345079,
"grad_norm": 1.0613563060760498,
"learning_rate": 4.8366147497950435e-05,
"loss": 0.2056,
"step": 44
},
{
"epoch": 0.10935601458080195,
"grad_norm": 1.1137558221817017,
"learning_rate": 4.8653375561549195e-05,
"loss": 0.293,
"step": 45
},
{
"epoch": 0.1117861482381531,
"grad_norm": 0.6552504301071167,
"learning_rate": 4.8934290406480814e-05,
"loss": 0.2478,
"step": 46
},
{
"epoch": 0.11421628189550426,
"grad_norm": 0.9905387759208679,
"learning_rate": 4.920916360113129e-05,
"loss": 0.2488,
"step": 47
},
{
"epoch": 0.1166464155528554,
"grad_norm": 1.101691484451294,
"learning_rate": 4.947824955958066e-05,
"loss": 0.1583,
"step": 48
},
{
"epoch": 0.11907654921020656,
"grad_norm": 1.3675787448883057,
"learning_rate": 4.9741786956652774e-05,
"loss": 0.3629,
"step": 49
},
{
"epoch": 0.12150668286755771,
"grad_norm": 2.6356711387634277,
"learning_rate": 5e-05,
"loss": 0.285,
"step": 50
},
{
"epoch": 0.12393681652490887,
"grad_norm": 1.5550183057785034,
"learning_rate": 5e-05,
"loss": 0.1688,
"step": 51
},
{
"epoch": 0.12636695018226002,
"grad_norm": 1.7116848230361938,
"learning_rate": 5e-05,
"loss": 0.2495,
"step": 52
},
{
"epoch": 0.12879708383961117,
"grad_norm": 1.3964793682098389,
"learning_rate": 5e-05,
"loss": 0.276,
"step": 53
},
{
"epoch": 0.13122721749696234,
"grad_norm": 0.8001569509506226,
"learning_rate": 5e-05,
"loss": 0.2976,
"step": 54
},
{
"epoch": 0.1336573511543135,
"grad_norm": 1.331112027168274,
"learning_rate": 5e-05,
"loss": 0.2725,
"step": 55
},
{
"epoch": 0.13608748481166463,
"grad_norm": 1.0783005952835083,
"learning_rate": 5e-05,
"loss": 0.188,
"step": 56
},
{
"epoch": 0.1385176184690158,
"grad_norm": 0.9507465958595276,
"learning_rate": 5e-05,
"loss": 0.2917,
"step": 57
},
{
"epoch": 0.14094775212636695,
"grad_norm": 0.8285686373710632,
"learning_rate": 5e-05,
"loss": 0.3322,
"step": 58
},
{
"epoch": 0.1433778857837181,
"grad_norm": 0.8201387524604797,
"learning_rate": 5e-05,
"loss": 0.2296,
"step": 59
},
{
"epoch": 0.14580801944106925,
"grad_norm": 0.743626594543457,
"learning_rate": 5e-05,
"loss": 0.1785,
"step": 60
},
{
"epoch": 0.14823815309842042,
"grad_norm": 0.9876251816749573,
"learning_rate": 5e-05,
"loss": 0.3416,
"step": 61
},
{
"epoch": 0.15066828675577157,
"grad_norm": 1.6088169813156128,
"learning_rate": 5e-05,
"loss": 0.2513,
"step": 62
},
{
"epoch": 0.15309842041312272,
"grad_norm": 1.2406659126281738,
"learning_rate": 5e-05,
"loss": 0.2856,
"step": 63
},
{
"epoch": 0.15552855407047386,
"grad_norm": 1.0257083177566528,
"learning_rate": 5e-05,
"loss": 0.1392,
"step": 64
},
{
"epoch": 0.15795868772782504,
"grad_norm": 0.7829582691192627,
"learning_rate": 5e-05,
"loss": 0.248,
"step": 65
},
{
"epoch": 0.16038882138517618,
"grad_norm": 0.7451558113098145,
"learning_rate": 5e-05,
"loss": 0.2826,
"step": 66
},
{
"epoch": 0.16281895504252733,
"grad_norm": 0.8115066885948181,
"learning_rate": 5e-05,
"loss": 0.3322,
"step": 67
},
{
"epoch": 0.1652490886998785,
"grad_norm": 0.8748694658279419,
"learning_rate": 5e-05,
"loss": 0.2057,
"step": 68
},
{
"epoch": 0.16767922235722965,
"grad_norm": 0.6844836473464966,
"learning_rate": 5e-05,
"loss": 0.2062,
"step": 69
},
{
"epoch": 0.1701093560145808,
"grad_norm": 0.9419387578964233,
"learning_rate": 5e-05,
"loss": 0.169,
"step": 70
},
{
"epoch": 0.17253948967193194,
"grad_norm": 0.8529036045074463,
"learning_rate": 5e-05,
"loss": 0.28,
"step": 71
},
{
"epoch": 0.17496962332928312,
"grad_norm": 0.8096457719802856,
"learning_rate": 5e-05,
"loss": 0.1699,
"step": 72
},
{
"epoch": 0.17739975698663427,
"grad_norm": 0.7595834136009216,
"learning_rate": 5e-05,
"loss": 0.2321,
"step": 73
},
{
"epoch": 0.1798298906439854,
"grad_norm": 0.6918485164642334,
"learning_rate": 5e-05,
"loss": 0.12,
"step": 74
},
{
"epoch": 0.1822600243013366,
"grad_norm": 0.8465882539749146,
"learning_rate": 5e-05,
"loss": 0.2344,
"step": 75
},
{
"epoch": 0.18469015795868773,
"grad_norm": 1.27568519115448,
"learning_rate": 5e-05,
"loss": 0.3018,
"step": 76
},
{
"epoch": 0.18712029161603888,
"grad_norm": 0.907692015171051,
"learning_rate": 5e-05,
"loss": 0.2154,
"step": 77
},
{
"epoch": 0.18955042527339003,
"grad_norm": 0.9362420439720154,
"learning_rate": 5e-05,
"loss": 0.2024,
"step": 78
},
{
"epoch": 0.1919805589307412,
"grad_norm": 1.1268069744110107,
"learning_rate": 5e-05,
"loss": 0.1509,
"step": 79
},
{
"epoch": 0.19441069258809235,
"grad_norm": 2.944066047668457,
"learning_rate": 5e-05,
"loss": 0.1669,
"step": 80
},
{
"epoch": 0.1968408262454435,
"grad_norm": 0.8719034790992737,
"learning_rate": 5e-05,
"loss": 0.2229,
"step": 81
},
{
"epoch": 0.19927095990279464,
"grad_norm": 0.7963365912437439,
"learning_rate": 5e-05,
"loss": 0.1485,
"step": 82
},
{
"epoch": 0.20170109356014582,
"grad_norm": 1.1861976385116577,
"learning_rate": 5e-05,
"loss": 0.1774,
"step": 83
},
{
"epoch": 0.20413122721749696,
"grad_norm": 1.029300570487976,
"learning_rate": 5e-05,
"loss": 0.2762,
"step": 84
},
{
"epoch": 0.2065613608748481,
"grad_norm": 0.7859634757041931,
"learning_rate": 5e-05,
"loss": 0.263,
"step": 85
},
{
"epoch": 0.20899149453219928,
"grad_norm": 1.2088333368301392,
"learning_rate": 5e-05,
"loss": 0.2248,
"step": 86
},
{
"epoch": 0.21142162818955043,
"grad_norm": 0.9491986036300659,
"learning_rate": 5e-05,
"loss": 0.1388,
"step": 87
},
{
"epoch": 0.21385176184690158,
"grad_norm": 1.3867732286453247,
"learning_rate": 5e-05,
"loss": 0.1932,
"step": 88
},
{
"epoch": 0.21628189550425272,
"grad_norm": 1.2418168783187866,
"learning_rate": 5e-05,
"loss": 0.2703,
"step": 89
},
{
"epoch": 0.2187120291616039,
"grad_norm": 1.514169692993164,
"learning_rate": 5e-05,
"loss": 0.3437,
"step": 90
},
{
"epoch": 0.22114216281895505,
"grad_norm": 1.1384562253952026,
"learning_rate": 5e-05,
"loss": 0.3157,
"step": 91
},
{
"epoch": 0.2235722964763062,
"grad_norm": 1.0910252332687378,
"learning_rate": 5e-05,
"loss": 0.3296,
"step": 92
},
{
"epoch": 0.22600243013365734,
"grad_norm": 0.8576286435127258,
"learning_rate": 5e-05,
"loss": 0.2572,
"step": 93
},
{
"epoch": 0.2284325637910085,
"grad_norm": 1.0798954963684082,
"learning_rate": 5e-05,
"loss": 0.2669,
"step": 94
},
{
"epoch": 0.23086269744835966,
"grad_norm": 1.0142358541488647,
"learning_rate": 5e-05,
"loss": 0.2607,
"step": 95
},
{
"epoch": 0.2332928311057108,
"grad_norm": 0.8707209825515747,
"learning_rate": 5e-05,
"loss": 0.2206,
"step": 96
},
{
"epoch": 0.23572296476306198,
"grad_norm": 1.038823127746582,
"learning_rate": 5e-05,
"loss": 0.1648,
"step": 97
},
{
"epoch": 0.23815309842041313,
"grad_norm": 1.3246694803237915,
"learning_rate": 5e-05,
"loss": 0.317,
"step": 98
},
{
"epoch": 0.24058323207776428,
"grad_norm": 1.0180795192718506,
"learning_rate": 5e-05,
"loss": 0.2653,
"step": 99
},
{
"epoch": 0.24301336573511542,
"grad_norm": 1.803346037864685,
"learning_rate": 5e-05,
"loss": 0.1221,
"step": 100
},
{
"epoch": 0.24301336573511542,
"eval_loss": 0.2808912396430969,
"eval_runtime": 508.6685,
"eval_samples_per_second": 5.402,
"eval_steps_per_second": 0.676,
"step": 100
},
{
"epoch": 0.2454434993924666,
"grad_norm": 1.4270009994506836,
"learning_rate": 5e-05,
"loss": 0.1513,
"step": 101
},
{
"epoch": 0.24787363304981774,
"grad_norm": 1.5230388641357422,
"learning_rate": 5e-05,
"loss": 0.2563,
"step": 102
},
{
"epoch": 0.2503037667071689,
"grad_norm": 1.0142149925231934,
"learning_rate": 5e-05,
"loss": 0.2021,
"step": 103
},
{
"epoch": 0.25273390036452004,
"grad_norm": 1.224928617477417,
"learning_rate": 5e-05,
"loss": 0.2227,
"step": 104
},
{
"epoch": 0.2551640340218712,
"grad_norm": 0.8966079950332642,
"learning_rate": 5e-05,
"loss": 0.1416,
"step": 105
},
{
"epoch": 0.25759416767922233,
"grad_norm": 4.392244815826416,
"learning_rate": 5e-05,
"loss": 0.1297,
"step": 106
},
{
"epoch": 0.2600243013365735,
"grad_norm": 0.7403039932250977,
"learning_rate": 5e-05,
"loss": 0.1383,
"step": 107
},
{
"epoch": 0.2624544349939247,
"grad_norm": 0.7826500535011292,
"learning_rate": 5e-05,
"loss": 0.1933,
"step": 108
},
{
"epoch": 0.2648845686512758,
"grad_norm": 1.2112810611724854,
"learning_rate": 5e-05,
"loss": 0.2118,
"step": 109
},
{
"epoch": 0.267314702308627,
"grad_norm": 1.2826062440872192,
"learning_rate": 5e-05,
"loss": 0.2028,
"step": 110
},
{
"epoch": 0.26974483596597815,
"grad_norm": 0.9675102829933167,
"learning_rate": 5e-05,
"loss": 0.1481,
"step": 111
},
{
"epoch": 0.27217496962332927,
"grad_norm": 1.2551299333572388,
"learning_rate": 5e-05,
"loss": 0.2172,
"step": 112
},
{
"epoch": 0.27460510328068044,
"grad_norm": 1.2526607513427734,
"learning_rate": 5e-05,
"loss": 0.213,
"step": 113
},
{
"epoch": 0.2770352369380316,
"grad_norm": 1.102669358253479,
"learning_rate": 5e-05,
"loss": 0.2387,
"step": 114
},
{
"epoch": 0.27946537059538273,
"grad_norm": 1.1452562808990479,
"learning_rate": 5e-05,
"loss": 0.2224,
"step": 115
},
{
"epoch": 0.2818955042527339,
"grad_norm": 1.4839582443237305,
"learning_rate": 5e-05,
"loss": 0.3572,
"step": 116
},
{
"epoch": 0.284325637910085,
"grad_norm": 0.7019845247268677,
"learning_rate": 5e-05,
"loss": 0.1531,
"step": 117
},
{
"epoch": 0.2867557715674362,
"grad_norm": 1.1344879865646362,
"learning_rate": 5e-05,
"loss": 0.1234,
"step": 118
},
{
"epoch": 0.2891859052247874,
"grad_norm": 0.7753707766532898,
"learning_rate": 5e-05,
"loss": 0.3054,
"step": 119
},
{
"epoch": 0.2916160388821385,
"grad_norm": 1.2467390298843384,
"learning_rate": 5e-05,
"loss": 0.0979,
"step": 120
},
{
"epoch": 0.29404617253948967,
"grad_norm": 0.8829241991043091,
"learning_rate": 5e-05,
"loss": 0.1251,
"step": 121
},
{
"epoch": 0.29647630619684084,
"grad_norm": 0.9316434860229492,
"learning_rate": 5e-05,
"loss": 0.1723,
"step": 122
},
{
"epoch": 0.29890643985419196,
"grad_norm": 0.9735895395278931,
"learning_rate": 5e-05,
"loss": 0.1992,
"step": 123
},
{
"epoch": 0.30133657351154314,
"grad_norm": 1.0356009006500244,
"learning_rate": 5e-05,
"loss": 0.2912,
"step": 124
},
{
"epoch": 0.3037667071688943,
"grad_norm": 0.8671007752418518,
"learning_rate": 5e-05,
"loss": 0.2204,
"step": 125
},
{
"epoch": 0.30619684082624543,
"grad_norm": 0.9986138939857483,
"learning_rate": 5e-05,
"loss": 0.1615,
"step": 126
},
{
"epoch": 0.3086269744835966,
"grad_norm": 0.8567970991134644,
"learning_rate": 5e-05,
"loss": 0.3127,
"step": 127
},
{
"epoch": 0.3110571081409477,
"grad_norm": 1.497078537940979,
"learning_rate": 5e-05,
"loss": 0.1461,
"step": 128
},
{
"epoch": 0.3134872417982989,
"grad_norm": 1.9854991436004639,
"learning_rate": 5e-05,
"loss": 0.1588,
"step": 129
},
{
"epoch": 0.3159173754556501,
"grad_norm": 0.9955502152442932,
"learning_rate": 5e-05,
"loss": 0.2703,
"step": 130
},
{
"epoch": 0.3183475091130012,
"grad_norm": 0.8443557620048523,
"learning_rate": 5e-05,
"loss": 0.1432,
"step": 131
},
{
"epoch": 0.32077764277035237,
"grad_norm": 0.8821367025375366,
"learning_rate": 5e-05,
"loss": 0.1113,
"step": 132
},
{
"epoch": 0.32320777642770354,
"grad_norm": 0.997931957244873,
"learning_rate": 5e-05,
"loss": 0.1363,
"step": 133
},
{
"epoch": 0.32563791008505466,
"grad_norm": 1.2628774642944336,
"learning_rate": 5e-05,
"loss": 0.134,
"step": 134
},
{
"epoch": 0.32806804374240583,
"grad_norm": 0.7694957852363586,
"learning_rate": 5e-05,
"loss": 0.1425,
"step": 135
},
{
"epoch": 0.330498177399757,
"grad_norm": 1.5881726741790771,
"learning_rate": 5e-05,
"loss": 0.17,
"step": 136
},
{
"epoch": 0.33292831105710813,
"grad_norm": 1.2904351949691772,
"learning_rate": 5e-05,
"loss": 0.1735,
"step": 137
},
{
"epoch": 0.3353584447144593,
"grad_norm": 1.3187663555145264,
"learning_rate": 5e-05,
"loss": 0.2035,
"step": 138
},
{
"epoch": 0.3377885783718105,
"grad_norm": 1.1753528118133545,
"learning_rate": 5e-05,
"loss": 0.2147,
"step": 139
},
{
"epoch": 0.3402187120291616,
"grad_norm": 1.6974895000457764,
"learning_rate": 5e-05,
"loss": 0.1303,
"step": 140
},
{
"epoch": 0.34264884568651277,
"grad_norm": 0.9945745468139648,
"learning_rate": 5e-05,
"loss": 0.1169,
"step": 141
},
{
"epoch": 0.3450789793438639,
"grad_norm": 1.0099682807922363,
"learning_rate": 5e-05,
"loss": 0.1446,
"step": 142
},
{
"epoch": 0.34750911300121506,
"grad_norm": 1.223264217376709,
"learning_rate": 5e-05,
"loss": 0.2118,
"step": 143
},
{
"epoch": 0.34993924665856624,
"grad_norm": 0.8575140833854675,
"learning_rate": 5e-05,
"loss": 0.1109,
"step": 144
},
{
"epoch": 0.35236938031591736,
"grad_norm": 1.0145286321640015,
"learning_rate": 5e-05,
"loss": 0.1664,
"step": 145
},
{
"epoch": 0.35479951397326853,
"grad_norm": 0.670891284942627,
"learning_rate": 5e-05,
"loss": 0.1166,
"step": 146
},
{
"epoch": 0.3572296476306197,
"grad_norm": 1.4332047700881958,
"learning_rate": 5e-05,
"loss": 0.1381,
"step": 147
},
{
"epoch": 0.3596597812879708,
"grad_norm": 1.0511322021484375,
"learning_rate": 5e-05,
"loss": 0.1279,
"step": 148
},
{
"epoch": 0.362089914945322,
"grad_norm": 0.751565158367157,
"learning_rate": 5e-05,
"loss": 0.1533,
"step": 149
},
{
"epoch": 0.3645200486026732,
"grad_norm": 1.612855076789856,
"learning_rate": 5e-05,
"loss": 0.1633,
"step": 150
},
{
"epoch": 0.3669501822600243,
"grad_norm": 1.2449802160263062,
"learning_rate": 5e-05,
"loss": 0.2037,
"step": 151
},
{
"epoch": 0.36938031591737547,
"grad_norm": 1.0060147047042847,
"learning_rate": 5e-05,
"loss": 0.0902,
"step": 152
},
{
"epoch": 0.3718104495747266,
"grad_norm": 0.9183961153030396,
"learning_rate": 5e-05,
"loss": 0.1748,
"step": 153
},
{
"epoch": 0.37424058323207776,
"grad_norm": 1.2824058532714844,
"learning_rate": 5e-05,
"loss": 0.1903,
"step": 154
},
{
"epoch": 0.37667071688942894,
"grad_norm": 0.524626612663269,
"learning_rate": 5e-05,
"loss": 0.1014,
"step": 155
},
{
"epoch": 0.37910085054678005,
"grad_norm": 0.946535587310791,
"learning_rate": 5e-05,
"loss": 0.2512,
"step": 156
},
{
"epoch": 0.38153098420413123,
"grad_norm": 0.8985329270362854,
"learning_rate": 5e-05,
"loss": 0.1247,
"step": 157
},
{
"epoch": 0.3839611178614824,
"grad_norm": 0.6025483012199402,
"learning_rate": 5e-05,
"loss": 0.0677,
"step": 158
},
{
"epoch": 0.3863912515188335,
"grad_norm": 0.9210860729217529,
"learning_rate": 5e-05,
"loss": 0.1361,
"step": 159
},
{
"epoch": 0.3888213851761847,
"grad_norm": 1.7067134380340576,
"learning_rate": 5e-05,
"loss": 0.2008,
"step": 160
},
{
"epoch": 0.39125151883353587,
"grad_norm": 0.8162620663642883,
"learning_rate": 5e-05,
"loss": 0.1108,
"step": 161
},
{
"epoch": 0.393681652490887,
"grad_norm": 0.535029947757721,
"learning_rate": 5e-05,
"loss": 0.1578,
"step": 162
},
{
"epoch": 0.39611178614823817,
"grad_norm": 0.7219657897949219,
"learning_rate": 5e-05,
"loss": 0.1012,
"step": 163
},
{
"epoch": 0.3985419198055893,
"grad_norm": 1.0636670589447021,
"learning_rate": 5e-05,
"loss": 0.1648,
"step": 164
},
{
"epoch": 0.40097205346294046,
"grad_norm": 1.1116470098495483,
"learning_rate": 5e-05,
"loss": 0.1191,
"step": 165
},
{
"epoch": 0.40340218712029163,
"grad_norm": 0.7233144044876099,
"learning_rate": 5e-05,
"loss": 0.1206,
"step": 166
},
{
"epoch": 0.40583232077764275,
"grad_norm": 1.7054411172866821,
"learning_rate": 5e-05,
"loss": 0.1466,
"step": 167
},
{
"epoch": 0.4082624544349939,
"grad_norm": 0.9031124114990234,
"learning_rate": 5e-05,
"loss": 0.1107,
"step": 168
},
{
"epoch": 0.4106925880923451,
"grad_norm": 0.8140673637390137,
"learning_rate": 5e-05,
"loss": 0.1503,
"step": 169
},
{
"epoch": 0.4131227217496962,
"grad_norm": 1.2189490795135498,
"learning_rate": 5e-05,
"loss": 0.1081,
"step": 170
},
{
"epoch": 0.4155528554070474,
"grad_norm": 0.8403185606002808,
"learning_rate": 5e-05,
"loss": 0.098,
"step": 171
},
{
"epoch": 0.41798298906439857,
"grad_norm": 0.8939706683158875,
"learning_rate": 5e-05,
"loss": 0.0961,
"step": 172
},
{
"epoch": 0.4204131227217497,
"grad_norm": 0.8737717270851135,
"learning_rate": 5e-05,
"loss": 0.1165,
"step": 173
},
{
"epoch": 0.42284325637910086,
"grad_norm": 0.9247780442237854,
"learning_rate": 5e-05,
"loss": 0.1787,
"step": 174
},
{
"epoch": 0.425273390036452,
"grad_norm": 0.9532423615455627,
"learning_rate": 5e-05,
"loss": 0.1676,
"step": 175
},
{
"epoch": 0.42770352369380316,
"grad_norm": 0.5811410546302795,
"learning_rate": 5e-05,
"loss": 0.1221,
"step": 176
},
{
"epoch": 0.43013365735115433,
"grad_norm": 0.6600822806358337,
"learning_rate": 5e-05,
"loss": 0.1809,
"step": 177
},
{
"epoch": 0.43256379100850545,
"grad_norm": 0.5638179183006287,
"learning_rate": 5e-05,
"loss": 0.0668,
"step": 178
},
{
"epoch": 0.4349939246658566,
"grad_norm": 0.5698412656784058,
"learning_rate": 5e-05,
"loss": 0.164,
"step": 179
},
{
"epoch": 0.4374240583232078,
"grad_norm": 1.2823748588562012,
"learning_rate": 5e-05,
"loss": 0.2111,
"step": 180
},
{
"epoch": 0.4398541919805589,
"grad_norm": 1.1044129133224487,
"learning_rate": 5e-05,
"loss": 0.1188,
"step": 181
},
{
"epoch": 0.4422843256379101,
"grad_norm": 0.9815372824668884,
"learning_rate": 5e-05,
"loss": 0.186,
"step": 182
},
{
"epoch": 0.44471445929526127,
"grad_norm": 0.8543856739997864,
"learning_rate": 5e-05,
"loss": 0.1237,
"step": 183
},
{
"epoch": 0.4471445929526124,
"grad_norm": 1.0749317407608032,
"learning_rate": 5e-05,
"loss": 0.1385,
"step": 184
},
{
"epoch": 0.44957472660996356,
"grad_norm": 1.4067716598510742,
"learning_rate": 5e-05,
"loss": 0.1269,
"step": 185
},
{
"epoch": 0.4520048602673147,
"grad_norm": 0.933139443397522,
"learning_rate": 5e-05,
"loss": 0.0476,
"step": 186
},
{
"epoch": 0.45443499392466585,
"grad_norm": 1.2160463333129883,
"learning_rate": 5e-05,
"loss": 0.0823,
"step": 187
},
{
"epoch": 0.456865127582017,
"grad_norm": 0.8305476903915405,
"learning_rate": 5e-05,
"loss": 0.2482,
"step": 188
},
{
"epoch": 0.45929526123936815,
"grad_norm": 0.8701503872871399,
"learning_rate": 5e-05,
"loss": 0.2007,
"step": 189
},
{
"epoch": 0.4617253948967193,
"grad_norm": 1.1040858030319214,
"learning_rate": 5e-05,
"loss": 0.1031,
"step": 190
},
{
"epoch": 0.4641555285540705,
"grad_norm": 1.149579644203186,
"learning_rate": 5e-05,
"loss": 0.1384,
"step": 191
},
{
"epoch": 0.4665856622114216,
"grad_norm": 1.0616861581802368,
"learning_rate": 5e-05,
"loss": 0.1961,
"step": 192
},
{
"epoch": 0.4690157958687728,
"grad_norm": 1.3416138887405396,
"learning_rate": 5e-05,
"loss": 0.2223,
"step": 193
},
{
"epoch": 0.47144592952612396,
"grad_norm": 1.2578601837158203,
"learning_rate": 5e-05,
"loss": 0.0833,
"step": 194
},
{
"epoch": 0.4738760631834751,
"grad_norm": 0.9381983876228333,
"learning_rate": 5e-05,
"loss": 0.135,
"step": 195
},
{
"epoch": 0.47630619684082626,
"grad_norm": 1.0591320991516113,
"learning_rate": 5e-05,
"loss": 0.1658,
"step": 196
},
{
"epoch": 0.4787363304981774,
"grad_norm": 0.9011105895042419,
"learning_rate": 5e-05,
"loss": 0.189,
"step": 197
},
{
"epoch": 0.48116646415552855,
"grad_norm": 0.9011105895042419,
"learning_rate": 5e-05,
"loss": 0.1729,
"step": 198
},
{
"epoch": 0.4835965978128797,
"grad_norm": 1.0387153625488281,
"learning_rate": 5e-05,
"loss": 0.1534,
"step": 199
},
{
"epoch": 0.48602673147023084,
"grad_norm": 1.7026396989822388,
"learning_rate": 5e-05,
"loss": 0.1534,
"step": 200
},
{
"epoch": 0.48602673147023084,
"eval_loss": 0.2694580554962158,
"eval_runtime": 510.8356,
"eval_samples_per_second": 5.379,
"eval_steps_per_second": 0.673,
"step": 200
},
{
"epoch": 0.488456865127582,
"grad_norm": 0.7090218663215637,
"learning_rate": 5e-05,
"loss": 0.1022,
"step": 201
},
{
"epoch": 0.4908869987849332,
"grad_norm": 0.6188836693763733,
"learning_rate": 5e-05,
"loss": 0.1839,
"step": 202
},
{
"epoch": 0.4933171324422843,
"grad_norm": 0.8001598715782166,
"learning_rate": 5e-05,
"loss": 0.1525,
"step": 203
},
{
"epoch": 0.4957472660996355,
"grad_norm": 1.006425142288208,
"learning_rate": 5e-05,
"loss": 0.1417,
"step": 204
},
{
"epoch": 0.49817739975698666,
"grad_norm": 0.9688583612442017,
"learning_rate": 5e-05,
"loss": 0.1504,
"step": 205
},
{
"epoch": 0.5006075334143378,
"grad_norm": 0.9707381129264832,
"learning_rate": 5e-05,
"loss": 0.0538,
"step": 206
},
{
"epoch": 0.503037667071689,
"grad_norm": 0.9181280136108398,
"learning_rate": 5e-05,
"loss": 0.2319,
"step": 207
},
{
"epoch": 0.5054678007290401,
"grad_norm": 1.059377670288086,
"learning_rate": 5e-05,
"loss": 0.2679,
"step": 208
},
{
"epoch": 0.5078979343863913,
"grad_norm": 0.6514914035797119,
"learning_rate": 5e-05,
"loss": 0.0623,
"step": 209
},
{
"epoch": 0.5103280680437424,
"grad_norm": 0.7265043258666992,
"learning_rate": 5e-05,
"loss": 0.1034,
"step": 210
},
{
"epoch": 0.5127582017010935,
"grad_norm": 1.5795567035675049,
"learning_rate": 5e-05,
"loss": 0.1236,
"step": 211
},
{
"epoch": 0.5151883353584447,
"grad_norm": 0.8533775806427002,
"learning_rate": 5e-05,
"loss": 0.0702,
"step": 212
},
{
"epoch": 0.5176184690157959,
"grad_norm": 0.7225982546806335,
"learning_rate": 5e-05,
"loss": 0.2005,
"step": 213
},
{
"epoch": 0.520048602673147,
"grad_norm": 0.9016891121864319,
"learning_rate": 5e-05,
"loss": 0.1329,
"step": 214
},
{
"epoch": 0.5224787363304981,
"grad_norm": 1.1623356342315674,
"learning_rate": 5e-05,
"loss": 0.1728,
"step": 215
},
{
"epoch": 0.5249088699878494,
"grad_norm": 1.1869186162948608,
"learning_rate": 5e-05,
"loss": 0.2517,
"step": 216
},
{
"epoch": 0.5273390036452005,
"grad_norm": 1.2271298170089722,
"learning_rate": 5e-05,
"loss": 0.1402,
"step": 217
},
{
"epoch": 0.5297691373025516,
"grad_norm": 1.182732105255127,
"learning_rate": 5e-05,
"loss": 0.1341,
"step": 218
},
{
"epoch": 0.5321992709599028,
"grad_norm": 0.8633968234062195,
"learning_rate": 5e-05,
"loss": 0.1469,
"step": 219
},
{
"epoch": 0.534629404617254,
"grad_norm": 0.9293995499610901,
"learning_rate": 5e-05,
"loss": 0.1121,
"step": 220
},
{
"epoch": 0.5370595382746051,
"grad_norm": 1.3438748121261597,
"learning_rate": 5e-05,
"loss": 0.1718,
"step": 221
},
{
"epoch": 0.5394896719319563,
"grad_norm": 1.2411259412765503,
"learning_rate": 5e-05,
"loss": 0.1304,
"step": 222
},
{
"epoch": 0.5419198055893074,
"grad_norm": 0.8153496384620667,
"learning_rate": 5e-05,
"loss": 0.1334,
"step": 223
},
{
"epoch": 0.5443499392466585,
"grad_norm": 0.7595478892326355,
"learning_rate": 5e-05,
"loss": 0.0994,
"step": 224
},
{
"epoch": 0.5467800729040098,
"grad_norm": 1.0710370540618896,
"learning_rate": 5e-05,
"loss": 0.0351,
"step": 225
},
{
"epoch": 0.5492102065613609,
"grad_norm": 1.2209466695785522,
"learning_rate": 5e-05,
"loss": 0.1049,
"step": 226
},
{
"epoch": 0.551640340218712,
"grad_norm": 0.5865522623062134,
"learning_rate": 5e-05,
"loss": 0.0948,
"step": 227
},
{
"epoch": 0.5540704738760632,
"grad_norm": 0.8379694223403931,
"learning_rate": 5e-05,
"loss": 0.1269,
"step": 228
},
{
"epoch": 0.5565006075334143,
"grad_norm": 0.7321844100952148,
"learning_rate": 5e-05,
"loss": 0.0912,
"step": 229
},
{
"epoch": 0.5589307411907655,
"grad_norm": 1.0253808498382568,
"learning_rate": 5e-05,
"loss": 0.0811,
"step": 230
},
{
"epoch": 0.5613608748481167,
"grad_norm": 0.9015201330184937,
"learning_rate": 5e-05,
"loss": 0.0725,
"step": 231
},
{
"epoch": 0.5637910085054678,
"grad_norm": 0.8760166168212891,
"learning_rate": 5e-05,
"loss": 0.1727,
"step": 232
},
{
"epoch": 0.5662211421628189,
"grad_norm": 0.892677903175354,
"learning_rate": 5e-05,
"loss": 0.084,
"step": 233
},
{
"epoch": 0.56865127582017,
"grad_norm": 0.6722489595413208,
"learning_rate": 5e-05,
"loss": 0.0845,
"step": 234
},
{
"epoch": 0.5710814094775213,
"grad_norm": 0.6187300682067871,
"learning_rate": 5e-05,
"loss": 0.1269,
"step": 235
},
{
"epoch": 0.5735115431348724,
"grad_norm": 0.8674693703651428,
"learning_rate": 5e-05,
"loss": 0.09,
"step": 236
},
{
"epoch": 0.5759416767922235,
"grad_norm": 0.9376964569091797,
"learning_rate": 5e-05,
"loss": 0.1916,
"step": 237
},
{
"epoch": 0.5783718104495748,
"grad_norm": 0.6777618527412415,
"learning_rate": 5e-05,
"loss": 0.1321,
"step": 238
},
{
"epoch": 0.5808019441069259,
"grad_norm": 1.1223134994506836,
"learning_rate": 5e-05,
"loss": 0.1811,
"step": 239
},
{
"epoch": 0.583232077764277,
"grad_norm": 0.8348031044006348,
"learning_rate": 5e-05,
"loss": 0.1508,
"step": 240
},
{
"epoch": 0.5856622114216282,
"grad_norm": 0.7387763857841492,
"learning_rate": 5e-05,
"loss": 0.058,
"step": 241
},
{
"epoch": 0.5880923450789793,
"grad_norm": 0.6327199339866638,
"learning_rate": 5e-05,
"loss": 0.1096,
"step": 242
},
{
"epoch": 0.5905224787363305,
"grad_norm": 0.7040870785713196,
"learning_rate": 5e-05,
"loss": 0.1281,
"step": 243
},
{
"epoch": 0.5929526123936817,
"grad_norm": 1.2110347747802734,
"learning_rate": 5e-05,
"loss": 0.1959,
"step": 244
},
{
"epoch": 0.5953827460510328,
"grad_norm": 0.8480271100997925,
"learning_rate": 5e-05,
"loss": 0.1234,
"step": 245
},
{
"epoch": 0.5978128797083839,
"grad_norm": 1.0009665489196777,
"learning_rate": 5e-05,
"loss": 0.1223,
"step": 246
},
{
"epoch": 0.6002430133657352,
"grad_norm": 0.7057176828384399,
"learning_rate": 5e-05,
"loss": 0.1747,
"step": 247
},
{
"epoch": 0.6026731470230863,
"grad_norm": 0.9945976138114929,
"learning_rate": 5e-05,
"loss": 0.1766,
"step": 248
},
{
"epoch": 0.6051032806804374,
"grad_norm": 0.7461872696876526,
"learning_rate": 5e-05,
"loss": 0.1224,
"step": 249
},
{
"epoch": 0.6075334143377886,
"grad_norm": 0.5109842419624329,
"learning_rate": 5e-05,
"loss": 0.1105,
"step": 250
},
{
"epoch": 0.6099635479951397,
"grad_norm": 0.7398750185966492,
"learning_rate": 5e-05,
"loss": 0.139,
"step": 251
},
{
"epoch": 0.6123936816524909,
"grad_norm": 0.7645841836929321,
"learning_rate": 5e-05,
"loss": 0.11,
"step": 252
},
{
"epoch": 0.6148238153098421,
"grad_norm": 0.8627939224243164,
"learning_rate": 5e-05,
"loss": 0.1277,
"step": 253
},
{
"epoch": 0.6172539489671932,
"grad_norm": 0.6933022737503052,
"learning_rate": 5e-05,
"loss": 0.1283,
"step": 254
},
{
"epoch": 0.6196840826245443,
"grad_norm": 0.8254914283752441,
"learning_rate": 5e-05,
"loss": 0.1087,
"step": 255
},
{
"epoch": 0.6221142162818954,
"grad_norm": 0.9561761617660522,
"learning_rate": 5e-05,
"loss": 0.1019,
"step": 256
},
{
"epoch": 0.6245443499392467,
"grad_norm": 1.0569480657577515,
"learning_rate": 5e-05,
"loss": 0.1235,
"step": 257
},
{
"epoch": 0.6269744835965978,
"grad_norm": 1.0708096027374268,
"learning_rate": 5e-05,
"loss": 0.1797,
"step": 258
},
{
"epoch": 0.6294046172539489,
"grad_norm": 0.5817265510559082,
"learning_rate": 5e-05,
"loss": 0.0787,
"step": 259
},
{
"epoch": 0.6318347509113001,
"grad_norm": 0.8255194425582886,
"learning_rate": 5e-05,
"loss": 0.1754,
"step": 260
},
{
"epoch": 0.6342648845686513,
"grad_norm": 0.8279481530189514,
"learning_rate": 5e-05,
"loss": 0.1668,
"step": 261
},
{
"epoch": 0.6366950182260024,
"grad_norm": 0.6393010020256042,
"learning_rate": 5e-05,
"loss": 0.0756,
"step": 262
},
{
"epoch": 0.6391251518833536,
"grad_norm": 0.49387434124946594,
"learning_rate": 5e-05,
"loss": 0.0757,
"step": 263
},
{
"epoch": 0.6415552855407047,
"grad_norm": 0.5705033540725708,
"learning_rate": 5e-05,
"loss": 0.0949,
"step": 264
},
{
"epoch": 0.6439854191980559,
"grad_norm": 0.9022333025932312,
"learning_rate": 5e-05,
"loss": 0.1716,
"step": 265
},
{
"epoch": 0.6464155528554071,
"grad_norm": 0.5299041271209717,
"learning_rate": 5e-05,
"loss": 0.0799,
"step": 266
},
{
"epoch": 0.6488456865127582,
"grad_norm": 0.7919138669967651,
"learning_rate": 5e-05,
"loss": 0.1476,
"step": 267
},
{
"epoch": 0.6512758201701093,
"grad_norm": 1.394972324371338,
"learning_rate": 5e-05,
"loss": 0.112,
"step": 268
},
{
"epoch": 0.6537059538274606,
"grad_norm": 0.9046385884284973,
"learning_rate": 5e-05,
"loss": 0.1683,
"step": 269
},
{
"epoch": 0.6561360874848117,
"grad_norm": 0.7190728783607483,
"learning_rate": 5e-05,
"loss": 0.1434,
"step": 270
},
{
"epoch": 0.6585662211421628,
"grad_norm": 1.0634692907333374,
"learning_rate": 5e-05,
"loss": 0.1244,
"step": 271
},
{
"epoch": 0.660996354799514,
"grad_norm": 0.569398045539856,
"learning_rate": 5e-05,
"loss": 0.0728,
"step": 272
},
{
"epoch": 0.6634264884568651,
"grad_norm": 0.9083871841430664,
"learning_rate": 5e-05,
"loss": 0.1467,
"step": 273
},
{
"epoch": 0.6658566221142163,
"grad_norm": 0.7927613854408264,
"learning_rate": 5e-05,
"loss": 0.1832,
"step": 274
},
{
"epoch": 0.6682867557715675,
"grad_norm": 1.0088517665863037,
"learning_rate": 5e-05,
"loss": 0.1422,
"step": 275
},
{
"epoch": 0.6707168894289186,
"grad_norm": 0.5374720692634583,
"learning_rate": 5e-05,
"loss": 0.0971,
"step": 276
},
{
"epoch": 0.6731470230862697,
"grad_norm": 0.7200155854225159,
"learning_rate": 5e-05,
"loss": 0.0722,
"step": 277
},
{
"epoch": 0.675577156743621,
"grad_norm": 0.7518504858016968,
"learning_rate": 5e-05,
"loss": 0.126,
"step": 278
},
{
"epoch": 0.6780072904009721,
"grad_norm": 0.8544089794158936,
"learning_rate": 5e-05,
"loss": 0.113,
"step": 279
},
{
"epoch": 0.6804374240583232,
"grad_norm": 0.7537454962730408,
"learning_rate": 5e-05,
"loss": 0.0615,
"step": 280
},
{
"epoch": 0.6828675577156743,
"grad_norm": 0.8440731763839722,
"learning_rate": 5e-05,
"loss": 0.0933,
"step": 281
},
{
"epoch": 0.6852976913730255,
"grad_norm": 1.7643004655838013,
"learning_rate": 5e-05,
"loss": 0.1807,
"step": 282
},
{
"epoch": 0.6877278250303767,
"grad_norm": 0.6502506732940674,
"learning_rate": 5e-05,
"loss": 0.1189,
"step": 283
},
{
"epoch": 0.6901579586877278,
"grad_norm": 0.8678569197654724,
"learning_rate": 5e-05,
"loss": 0.094,
"step": 284
},
{
"epoch": 0.692588092345079,
"grad_norm": 0.8592473864555359,
"learning_rate": 5e-05,
"loss": 0.1197,
"step": 285
},
{
"epoch": 0.6950182260024301,
"grad_norm": 0.8664891719818115,
"learning_rate": 5e-05,
"loss": 0.0575,
"step": 286
},
{
"epoch": 0.6974483596597812,
"grad_norm": 0.5715927481651306,
"learning_rate": 5e-05,
"loss": 0.0668,
"step": 287
},
{
"epoch": 0.6998784933171325,
"grad_norm": 1.0751949548721313,
"learning_rate": 5e-05,
"loss": 0.1457,
"step": 288
},
{
"epoch": 0.7023086269744836,
"grad_norm": 0.8039306402206421,
"learning_rate": 5e-05,
"loss": 0.1206,
"step": 289
},
{
"epoch": 0.7047387606318347,
"grad_norm": 1.5916086435317993,
"learning_rate": 5e-05,
"loss": 0.1471,
"step": 290
},
{
"epoch": 0.707168894289186,
"grad_norm": 0.7829633355140686,
"learning_rate": 5e-05,
"loss": 0.106,
"step": 291
},
{
"epoch": 0.7095990279465371,
"grad_norm": 0.8277871608734131,
"learning_rate": 5e-05,
"loss": 0.0347,
"step": 292
},
{
"epoch": 0.7120291616038882,
"grad_norm": 0.9798584580421448,
"learning_rate": 5e-05,
"loss": 0.1231,
"step": 293
},
{
"epoch": 0.7144592952612394,
"grad_norm": 0.6273385286331177,
"learning_rate": 5e-05,
"loss": 0.0838,
"step": 294
},
{
"epoch": 0.7168894289185905,
"grad_norm": 1.2111278772354126,
"learning_rate": 5e-05,
"loss": 0.1971,
"step": 295
},
{
"epoch": 0.7193195625759417,
"grad_norm": 0.5927327275276184,
"learning_rate": 5e-05,
"loss": 0.1198,
"step": 296
},
{
"epoch": 0.7217496962332929,
"grad_norm": 0.672064483165741,
"learning_rate": 5e-05,
"loss": 0.1295,
"step": 297
},
{
"epoch": 0.724179829890644,
"grad_norm": 0.6398962736129761,
"learning_rate": 5e-05,
"loss": 0.0933,
"step": 298
},
{
"epoch": 0.7266099635479951,
"grad_norm": 0.7458367943763733,
"learning_rate": 5e-05,
"loss": 0.1407,
"step": 299
},
{
"epoch": 0.7290400972053463,
"grad_norm": 0.6593539118766785,
"learning_rate": 5e-05,
"loss": 0.1055,
"step": 300
},
{
"epoch": 0.7290400972053463,
"eval_loss": 0.24557192623615265,
"eval_runtime": 513.4098,
"eval_samples_per_second": 5.352,
"eval_steps_per_second": 0.67,
"step": 300
},
{
"epoch": 0.7314702308626975,
"grad_norm": 0.5555610656738281,
"learning_rate": 5e-05,
"loss": 0.1641,
"step": 301
},
{
"epoch": 0.7339003645200486,
"grad_norm": 0.49084270000457764,
"learning_rate": 5e-05,
"loss": 0.0786,
"step": 302
},
{
"epoch": 0.7363304981773997,
"grad_norm": 0.9267096519470215,
"learning_rate": 5e-05,
"loss": 0.2053,
"step": 303
},
{
"epoch": 0.7387606318347509,
"grad_norm": 0.440594345331192,
"learning_rate": 5e-05,
"loss": 0.1224,
"step": 304
},
{
"epoch": 0.741190765492102,
"grad_norm": 1.0237218141555786,
"learning_rate": 5e-05,
"loss": 0.1263,
"step": 305
},
{
"epoch": 0.7436208991494532,
"grad_norm": 1.0087871551513672,
"learning_rate": 5e-05,
"loss": 0.1218,
"step": 306
},
{
"epoch": 0.7460510328068044,
"grad_norm": 0.8280074000358582,
"learning_rate": 5e-05,
"loss": 0.1851,
"step": 307
},
{
"epoch": 0.7484811664641555,
"grad_norm": 0.681713342666626,
"learning_rate": 5e-05,
"loss": 0.1757,
"step": 308
},
{
"epoch": 0.7509113001215066,
"grad_norm": 4.372740268707275,
"learning_rate": 5e-05,
"loss": 0.104,
"step": 309
},
{
"epoch": 0.7533414337788579,
"grad_norm": 0.8002400398254395,
"learning_rate": 5e-05,
"loss": 0.0922,
"step": 310
},
{
"epoch": 0.755771567436209,
"grad_norm": 0.7685486674308777,
"learning_rate": 5e-05,
"loss": 0.1391,
"step": 311
},
{
"epoch": 0.7582017010935601,
"grad_norm": 1.282363772392273,
"learning_rate": 5e-05,
"loss": 0.1141,
"step": 312
},
{
"epoch": 0.7606318347509113,
"grad_norm": 0.9099876284599304,
"learning_rate": 5e-05,
"loss": 0.1396,
"step": 313
},
{
"epoch": 0.7630619684082625,
"grad_norm": 0.5911856293678284,
"learning_rate": 5e-05,
"loss": 0.0631,
"step": 314
},
{
"epoch": 0.7654921020656136,
"grad_norm": 0.7148822546005249,
"learning_rate": 5e-05,
"loss": 0.1714,
"step": 315
},
{
"epoch": 0.7679222357229648,
"grad_norm": 0.6756410002708435,
"learning_rate": 5e-05,
"loss": 0.0894,
"step": 316
},
{
"epoch": 0.7703523693803159,
"grad_norm": 0.8938581347465515,
"learning_rate": 5e-05,
"loss": 0.1538,
"step": 317
},
{
"epoch": 0.772782503037667,
"grad_norm": 1.061012625694275,
"learning_rate": 5e-05,
"loss": 0.1069,
"step": 318
},
{
"epoch": 0.7752126366950183,
"grad_norm": 0.8262349963188171,
"learning_rate": 5e-05,
"loss": 0.0745,
"step": 319
},
{
"epoch": 0.7776427703523694,
"grad_norm": 1.3609825372695923,
"learning_rate": 5e-05,
"loss": 0.0689,
"step": 320
},
{
"epoch": 0.7800729040097205,
"grad_norm": 1.2151660919189453,
"learning_rate": 5e-05,
"loss": 0.0863,
"step": 321
},
{
"epoch": 0.7825030376670717,
"grad_norm": 0.7165310978889465,
"learning_rate": 5e-05,
"loss": 0.084,
"step": 322
},
{
"epoch": 0.7849331713244229,
"grad_norm": 1.2151451110839844,
"learning_rate": 5e-05,
"loss": 0.1229,
"step": 323
},
{
"epoch": 0.787363304981774,
"grad_norm": 0.9600043296813965,
"learning_rate": 5e-05,
"loss": 0.0524,
"step": 324
},
{
"epoch": 0.7897934386391251,
"grad_norm": 1.1785893440246582,
"learning_rate": 5e-05,
"loss": 0.1635,
"step": 325
},
{
"epoch": 0.7922235722964763,
"grad_norm": 0.6502353549003601,
"learning_rate": 5e-05,
"loss": 0.103,
"step": 326
},
{
"epoch": 0.7946537059538274,
"grad_norm": 0.6894869208335876,
"learning_rate": 5e-05,
"loss": 0.0824,
"step": 327
},
{
"epoch": 0.7970838396111786,
"grad_norm": 0.8827742338180542,
"learning_rate": 5e-05,
"loss": 0.0899,
"step": 328
},
{
"epoch": 0.7995139732685298,
"grad_norm": 0.82513827085495,
"learning_rate": 5e-05,
"loss": 0.0642,
"step": 329
},
{
"epoch": 0.8019441069258809,
"grad_norm": 0.7896872758865356,
"learning_rate": 5e-05,
"loss": 0.1869,
"step": 330
},
{
"epoch": 0.804374240583232,
"grad_norm": 1.421380639076233,
"learning_rate": 5e-05,
"loss": 0.0668,
"step": 331
},
{
"epoch": 0.8068043742405833,
"grad_norm": 0.826987087726593,
"learning_rate": 5e-05,
"loss": 0.1543,
"step": 332
},
{
"epoch": 0.8092345078979344,
"grad_norm": 0.7504023909568787,
"learning_rate": 5e-05,
"loss": 0.1167,
"step": 333
},
{
"epoch": 0.8116646415552855,
"grad_norm": 0.9217634797096252,
"learning_rate": 5e-05,
"loss": 0.0801,
"step": 334
},
{
"epoch": 0.8140947752126367,
"grad_norm": 0.6423681974411011,
"learning_rate": 5e-05,
"loss": 0.0445,
"step": 335
},
{
"epoch": 0.8165249088699879,
"grad_norm": 0.3945442736148834,
"learning_rate": 5e-05,
"loss": 0.0737,
"step": 336
},
{
"epoch": 0.818955042527339,
"grad_norm": 0.6396763324737549,
"learning_rate": 5e-05,
"loss": 0.1118,
"step": 337
},
{
"epoch": 0.8213851761846902,
"grad_norm": 1.31553053855896,
"learning_rate": 5e-05,
"loss": 0.0989,
"step": 338
},
{
"epoch": 0.8238153098420413,
"grad_norm": 0.962189793586731,
"learning_rate": 5e-05,
"loss": 0.1043,
"step": 339
},
{
"epoch": 0.8262454434993924,
"grad_norm": 0.9305171370506287,
"learning_rate": 5e-05,
"loss": 0.1333,
"step": 340
},
{
"epoch": 0.8286755771567437,
"grad_norm": 0.6102299690246582,
"learning_rate": 5e-05,
"loss": 0.0743,
"step": 341
},
{
"epoch": 0.8311057108140948,
"grad_norm": 0.9117159247398376,
"learning_rate": 5e-05,
"loss": 0.0399,
"step": 342
},
{
"epoch": 0.8335358444714459,
"grad_norm": 0.614378035068512,
"learning_rate": 5e-05,
"loss": 0.0767,
"step": 343
},
{
"epoch": 0.8359659781287971,
"grad_norm": 0.934684157371521,
"learning_rate": 5e-05,
"loss": 0.0814,
"step": 344
},
{
"epoch": 0.8383961117861483,
"grad_norm": 0.8567538261413574,
"learning_rate": 5e-05,
"loss": 0.0866,
"step": 345
},
{
"epoch": 0.8408262454434994,
"grad_norm": 1.083308219909668,
"learning_rate": 5e-05,
"loss": 0.0534,
"step": 346
},
{
"epoch": 0.8432563791008505,
"grad_norm": 0.8664528727531433,
"learning_rate": 5e-05,
"loss": 0.1348,
"step": 347
},
{
"epoch": 0.8456865127582017,
"grad_norm": 0.7151762843132019,
"learning_rate": 5e-05,
"loss": 0.0615,
"step": 348
},
{
"epoch": 0.8481166464155528,
"grad_norm": 0.784515917301178,
"learning_rate": 5e-05,
"loss": 0.1014,
"step": 349
},
{
"epoch": 0.850546780072904,
"grad_norm": 0.7713536620140076,
"learning_rate": 5e-05,
"loss": 0.0684,
"step": 350
},
{
"epoch": 0.8529769137302552,
"grad_norm": 0.8839780688285828,
"learning_rate": 5e-05,
"loss": 0.2673,
"step": 351
},
{
"epoch": 0.8554070473876063,
"grad_norm": 0.6095230579376221,
"learning_rate": 5e-05,
"loss": 0.1667,
"step": 352
},
{
"epoch": 0.8578371810449574,
"grad_norm": 1.1133002042770386,
"learning_rate": 5e-05,
"loss": 0.1746,
"step": 353
},
{
"epoch": 0.8602673147023087,
"grad_norm": 0.5206813812255859,
"learning_rate": 5e-05,
"loss": 0.1507,
"step": 354
},
{
"epoch": 0.8626974483596598,
"grad_norm": 0.6527222990989685,
"learning_rate": 5e-05,
"loss": 0.1045,
"step": 355
},
{
"epoch": 0.8651275820170109,
"grad_norm": 0.6029911637306213,
"learning_rate": 5e-05,
"loss": 0.1157,
"step": 356
},
{
"epoch": 0.8675577156743621,
"grad_norm": 0.7916355729103088,
"learning_rate": 5e-05,
"loss": 0.0817,
"step": 357
},
{
"epoch": 0.8699878493317132,
"grad_norm": 0.6410179734230042,
"learning_rate": 5e-05,
"loss": 0.0707,
"step": 358
},
{
"epoch": 0.8724179829890644,
"grad_norm": 1.0566637516021729,
"learning_rate": 5e-05,
"loss": 0.0774,
"step": 359
},
{
"epoch": 0.8748481166464156,
"grad_norm": 0.6340796947479248,
"learning_rate": 5e-05,
"loss": 0.1003,
"step": 360
},
{
"epoch": 0.8772782503037667,
"grad_norm": 0.6086530089378357,
"learning_rate": 5e-05,
"loss": 0.0441,
"step": 361
},
{
"epoch": 0.8797083839611178,
"grad_norm": 0.8955773115158081,
"learning_rate": 5e-05,
"loss": 0.1389,
"step": 362
},
{
"epoch": 0.8821385176184691,
"grad_norm": 0.4679882228374481,
"learning_rate": 5e-05,
"loss": 0.0708,
"step": 363
},
{
"epoch": 0.8845686512758202,
"grad_norm": 0.6539987921714783,
"learning_rate": 5e-05,
"loss": 0.1078,
"step": 364
},
{
"epoch": 0.8869987849331713,
"grad_norm": 0.8322189450263977,
"learning_rate": 5e-05,
"loss": 0.0477,
"step": 365
},
{
"epoch": 0.8894289185905225,
"grad_norm": 0.5329148769378662,
"learning_rate": 5e-05,
"loss": 0.0638,
"step": 366
},
{
"epoch": 0.8918590522478737,
"grad_norm": 0.7909948825836182,
"learning_rate": 5e-05,
"loss": 0.0941,
"step": 367
},
{
"epoch": 0.8942891859052248,
"grad_norm": 0.6703280806541443,
"learning_rate": 5e-05,
"loss": 0.0989,
"step": 368
},
{
"epoch": 0.8967193195625759,
"grad_norm": 0.6218695640563965,
"learning_rate": 5e-05,
"loss": 0.0712,
"step": 369
},
{
"epoch": 0.8991494532199271,
"grad_norm": 1.6482758522033691,
"learning_rate": 5e-05,
"loss": 0.1537,
"step": 370
},
{
"epoch": 0.9015795868772782,
"grad_norm": 0.7616936564445496,
"learning_rate": 5e-05,
"loss": 0.1234,
"step": 371
},
{
"epoch": 0.9040097205346294,
"grad_norm": 0.7617467045783997,
"learning_rate": 5e-05,
"loss": 0.0895,
"step": 372
},
{
"epoch": 0.9064398541919806,
"grad_norm": 0.7406112551689148,
"learning_rate": 5e-05,
"loss": 0.0559,
"step": 373
},
{
"epoch": 0.9088699878493317,
"grad_norm": 1.1627577543258667,
"learning_rate": 5e-05,
"loss": 0.0807,
"step": 374
},
{
"epoch": 0.9113001215066828,
"grad_norm": 0.6091077923774719,
"learning_rate": 5e-05,
"loss": 0.0649,
"step": 375
},
{
"epoch": 0.913730255164034,
"grad_norm": 0.8935027122497559,
"learning_rate": 5e-05,
"loss": 0.1563,
"step": 376
},
{
"epoch": 0.9161603888213852,
"grad_norm": 0.8287168741226196,
"learning_rate": 5e-05,
"loss": 0.0725,
"step": 377
},
{
"epoch": 0.9185905224787363,
"grad_norm": 0.5430625677108765,
"learning_rate": 5e-05,
"loss": 0.0984,
"step": 378
},
{
"epoch": 0.9210206561360875,
"grad_norm": 0.7538096308708191,
"learning_rate": 5e-05,
"loss": 0.1231,
"step": 379
},
{
"epoch": 0.9234507897934386,
"grad_norm": 0.9483656287193298,
"learning_rate": 5e-05,
"loss": 0.1341,
"step": 380
},
{
"epoch": 0.9258809234507898,
"grad_norm": 0.5689654350280762,
"learning_rate": 5e-05,
"loss": 0.1116,
"step": 381
},
{
"epoch": 0.928311057108141,
"grad_norm": 0.987369179725647,
"learning_rate": 5e-05,
"loss": 0.1815,
"step": 382
},
{
"epoch": 0.9307411907654921,
"grad_norm": 0.872460126876831,
"learning_rate": 5e-05,
"loss": 0.0994,
"step": 383
},
{
"epoch": 0.9331713244228432,
"grad_norm": 0.785040020942688,
"learning_rate": 5e-05,
"loss": 0.1006,
"step": 384
},
{
"epoch": 0.9356014580801945,
"grad_norm": 0.6228112578392029,
"learning_rate": 5e-05,
"loss": 0.0805,
"step": 385
},
{
"epoch": 0.9380315917375456,
"grad_norm": 0.746306836605072,
"learning_rate": 5e-05,
"loss": 0.0672,
"step": 386
},
{
"epoch": 0.9404617253948967,
"grad_norm": 0.44906482100486755,
"learning_rate": 5e-05,
"loss": 0.0831,
"step": 387
},
{
"epoch": 0.9428918590522479,
"grad_norm": 0.7880069613456726,
"learning_rate": 5e-05,
"loss": 0.0871,
"step": 388
},
{
"epoch": 0.945321992709599,
"grad_norm": 0.7579517960548401,
"learning_rate": 5e-05,
"loss": 0.1306,
"step": 389
},
{
"epoch": 0.9477521263669502,
"grad_norm": 0.5668421387672424,
"learning_rate": 5e-05,
"loss": 0.0738,
"step": 390
},
{
"epoch": 0.9501822600243013,
"grad_norm": 0.8249059319496155,
"learning_rate": 5e-05,
"loss": 0.069,
"step": 391
},
{
"epoch": 0.9526123936816525,
"grad_norm": 0.7063718438148499,
"learning_rate": 5e-05,
"loss": 0.0488,
"step": 392
},
{
"epoch": 0.9550425273390036,
"grad_norm": 0.9006249904632568,
"learning_rate": 5e-05,
"loss": 0.029,
"step": 393
},
{
"epoch": 0.9574726609963548,
"grad_norm": 0.5778633952140808,
"learning_rate": 5e-05,
"loss": 0.0859,
"step": 394
},
{
"epoch": 0.959902794653706,
"grad_norm": 0.9541341066360474,
"learning_rate": 5e-05,
"loss": 0.0698,
"step": 395
},
{
"epoch": 0.9623329283110571,
"grad_norm": 0.7825549244880676,
"learning_rate": 5e-05,
"loss": 0.0714,
"step": 396
},
{
"epoch": 0.9647630619684082,
"grad_norm": 0.729721188545227,
"learning_rate": 5e-05,
"loss": 0.0651,
"step": 397
},
{
"epoch": 0.9671931956257594,
"grad_norm": 0.7052625417709351,
"learning_rate": 5e-05,
"loss": 0.1518,
"step": 398
},
{
"epoch": 0.9696233292831106,
"grad_norm": 2.3800599575042725,
"learning_rate": 5e-05,
"loss": 0.0919,
"step": 399
},
{
"epoch": 0.9720534629404617,
"grad_norm": 0.5317090749740601,
"learning_rate": 5e-05,
"loss": 0.0828,
"step": 400
},
{
"epoch": 0.9720534629404617,
"eval_loss": 0.2335556447505951,
"eval_runtime": 513.3338,
"eval_samples_per_second": 5.353,
"eval_steps_per_second": 0.67,
"step": 400
},
{
"epoch": 0.9744835965978129,
"grad_norm": 0.6768686771392822,
"learning_rate": 5e-05,
"loss": 0.1675,
"step": 401
},
{
"epoch": 0.976913730255164,
"grad_norm": 0.5037873983383179,
"learning_rate": 5e-05,
"loss": 0.0199,
"step": 402
},
{
"epoch": 0.9793438639125152,
"grad_norm": 1.7766000032424927,
"learning_rate": 5e-05,
"loss": 0.0846,
"step": 403
},
{
"epoch": 0.9817739975698664,
"grad_norm": 0.978345513343811,
"learning_rate": 5e-05,
"loss": 0.0593,
"step": 404
},
{
"epoch": 0.9842041312272175,
"grad_norm": 1.4286214113235474,
"learning_rate": 5e-05,
"loss": 0.1963,
"step": 405
},
{
"epoch": 0.9866342648845686,
"grad_norm": 1.1499220132827759,
"learning_rate": 5e-05,
"loss": 0.0786,
"step": 406
},
{
"epoch": 0.9890643985419199,
"grad_norm": 1.4914474487304688,
"learning_rate": 5e-05,
"loss": 0.1152,
"step": 407
},
{
"epoch": 0.991494532199271,
"grad_norm": 0.7232803106307983,
"learning_rate": 5e-05,
"loss": 0.1338,
"step": 408
},
{
"epoch": 0.9939246658566221,
"grad_norm": 0.8456845283508301,
"learning_rate": 5e-05,
"loss": 0.0631,
"step": 409
},
{
"epoch": 0.9963547995139733,
"grad_norm": 0.5165804028511047,
"learning_rate": 5e-05,
"loss": 0.0725,
"step": 410
},
{
"epoch": 0.9987849331713244,
"grad_norm": 0.8364010453224182,
"learning_rate": 5e-05,
"loss": 0.1215,
"step": 411
},
{
"epoch": 1.0012150668286757,
"grad_norm": 0.6961699724197388,
"learning_rate": 5e-05,
"loss": 0.0327,
"step": 412
},
{
"epoch": 1.0036452004860268,
"grad_norm": 0.9358980655670166,
"learning_rate": 5e-05,
"loss": 0.0724,
"step": 413
},
{
"epoch": 1.006075334143378,
"grad_norm": 0.641076922416687,
"learning_rate": 5e-05,
"loss": 0.0824,
"step": 414
},
{
"epoch": 1.008505467800729,
"grad_norm": 0.6918608546257019,
"learning_rate": 5e-05,
"loss": 0.083,
"step": 415
},
{
"epoch": 1.0109356014580801,
"grad_norm": 0.7481106519699097,
"learning_rate": 5e-05,
"loss": 0.0887,
"step": 416
},
{
"epoch": 1.0133657351154313,
"grad_norm": 0.7577409148216248,
"learning_rate": 5e-05,
"loss": 0.0684,
"step": 417
},
{
"epoch": 1.0157958687727826,
"grad_norm": 1.0815515518188477,
"learning_rate": 5e-05,
"loss": 0.1167,
"step": 418
},
{
"epoch": 1.0182260024301337,
"grad_norm": 1.248225450515747,
"learning_rate": 5e-05,
"loss": 0.0644,
"step": 419
},
{
"epoch": 1.0206561360874848,
"grad_norm": 0.7970739006996155,
"learning_rate": 5e-05,
"loss": 0.1712,
"step": 420
},
{
"epoch": 1.023086269744836,
"grad_norm": 1.343601942062378,
"learning_rate": 5e-05,
"loss": 0.1359,
"step": 421
},
{
"epoch": 1.025516403402187,
"grad_norm": 0.705987274646759,
"learning_rate": 5e-05,
"loss": 0.0961,
"step": 422
},
{
"epoch": 1.0279465370595382,
"grad_norm": 0.5520442724227905,
"learning_rate": 5e-05,
"loss": 0.0839,
"step": 423
},
{
"epoch": 1.0303766707168893,
"grad_norm": 0.5799084305763245,
"learning_rate": 5e-05,
"loss": 0.0759,
"step": 424
},
{
"epoch": 1.0328068043742407,
"grad_norm": 0.610801637172699,
"learning_rate": 5e-05,
"loss": 0.0593,
"step": 425
},
{
"epoch": 1.0352369380315918,
"grad_norm": 1.035436987876892,
"learning_rate": 5e-05,
"loss": 0.194,
"step": 426
},
{
"epoch": 1.037667071688943,
"grad_norm": 0.9973961114883423,
"learning_rate": 5e-05,
"loss": 0.0954,
"step": 427
},
{
"epoch": 1.040097205346294,
"grad_norm": 0.8617413640022278,
"learning_rate": 5e-05,
"loss": 0.0917,
"step": 428
},
{
"epoch": 1.0425273390036451,
"grad_norm": 0.9682510495185852,
"learning_rate": 5e-05,
"loss": 0.1391,
"step": 429
},
{
"epoch": 1.0449574726609963,
"grad_norm": 0.3044584393501282,
"learning_rate": 5e-05,
"loss": 0.025,
"step": 430
},
{
"epoch": 1.0473876063183476,
"grad_norm": 0.6183155179023743,
"learning_rate": 5e-05,
"loss": 0.0347,
"step": 431
},
{
"epoch": 1.0498177399756987,
"grad_norm": 0.37442171573638916,
"learning_rate": 5e-05,
"loss": 0.0452,
"step": 432
},
{
"epoch": 1.0522478736330498,
"grad_norm": 0.565359890460968,
"learning_rate": 5e-05,
"loss": 0.0943,
"step": 433
},
{
"epoch": 1.054678007290401,
"grad_norm": 0.6165122389793396,
"learning_rate": 5e-05,
"loss": 0.0851,
"step": 434
},
{
"epoch": 1.057108140947752,
"grad_norm": 0.7150765061378479,
"learning_rate": 5e-05,
"loss": 0.0704,
"step": 435
},
{
"epoch": 1.0595382746051032,
"grad_norm": 0.7983779311180115,
"learning_rate": 5e-05,
"loss": 0.121,
"step": 436
},
{
"epoch": 1.0619684082624545,
"grad_norm": 0.5900336503982544,
"learning_rate": 5e-05,
"loss": 0.0938,
"step": 437
},
{
"epoch": 1.0643985419198057,
"grad_norm": 0.6962295174598694,
"learning_rate": 5e-05,
"loss": 0.1227,
"step": 438
},
{
"epoch": 1.0668286755771568,
"grad_norm": 1.0119210481643677,
"learning_rate": 5e-05,
"loss": 0.0737,
"step": 439
},
{
"epoch": 1.069258809234508,
"grad_norm": 0.4835350811481476,
"learning_rate": 5e-05,
"loss": 0.0644,
"step": 440
},
{
"epoch": 1.071688942891859,
"grad_norm": 0.6343538761138916,
"learning_rate": 5e-05,
"loss": 0.0796,
"step": 441
},
{
"epoch": 1.0741190765492101,
"grad_norm": 0.41634368896484375,
"learning_rate": 5e-05,
"loss": 0.087,
"step": 442
},
{
"epoch": 1.0765492102065615,
"grad_norm": 0.7903019189834595,
"learning_rate": 5e-05,
"loss": 0.0564,
"step": 443
},
{
"epoch": 1.0789793438639126,
"grad_norm": 0.45615118741989136,
"learning_rate": 5e-05,
"loss": 0.0928,
"step": 444
},
{
"epoch": 1.0814094775212637,
"grad_norm": 0.5224602222442627,
"learning_rate": 5e-05,
"loss": 0.1134,
"step": 445
},
{
"epoch": 1.0838396111786148,
"grad_norm": 0.7974500060081482,
"learning_rate": 5e-05,
"loss": 0.1592,
"step": 446
},
{
"epoch": 1.086269744835966,
"grad_norm": 0.545687735080719,
"learning_rate": 5e-05,
"loss": 0.0892,
"step": 447
},
{
"epoch": 1.088699878493317,
"grad_norm": 0.88418048620224,
"learning_rate": 5e-05,
"loss": 0.0406,
"step": 448
},
{
"epoch": 1.0911300121506682,
"grad_norm": 0.6175111532211304,
"learning_rate": 5e-05,
"loss": 0.0536,
"step": 449
},
{
"epoch": 1.0935601458080195,
"grad_norm": 0.5780444741249084,
"learning_rate": 5e-05,
"loss": 0.0806,
"step": 450
},
{
"epoch": 1.0959902794653706,
"grad_norm": 2.7306559085845947,
"learning_rate": 5e-05,
"loss": 0.102,
"step": 451
},
{
"epoch": 1.0984204131227218,
"grad_norm": 0.8727555871009827,
"learning_rate": 5e-05,
"loss": 0.0723,
"step": 452
},
{
"epoch": 1.1008505467800729,
"grad_norm": 0.49243301153182983,
"learning_rate": 5e-05,
"loss": 0.0589,
"step": 453
},
{
"epoch": 1.103280680437424,
"grad_norm": 1.0444567203521729,
"learning_rate": 5e-05,
"loss": 0.1075,
"step": 454
},
{
"epoch": 1.1057108140947751,
"grad_norm": 0.5982443690299988,
"learning_rate": 5e-05,
"loss": 0.0646,
"step": 455
},
{
"epoch": 1.1081409477521265,
"grad_norm": 0.4428943395614624,
"learning_rate": 5e-05,
"loss": 0.0421,
"step": 456
},
{
"epoch": 1.1105710814094776,
"grad_norm": 0.612250566482544,
"learning_rate": 5e-05,
"loss": 0.0694,
"step": 457
},
{
"epoch": 1.1130012150668287,
"grad_norm": 0.8445409536361694,
"learning_rate": 5e-05,
"loss": 0.1089,
"step": 458
},
{
"epoch": 1.1154313487241798,
"grad_norm": 0.5859739184379578,
"learning_rate": 5e-05,
"loss": 0.1143,
"step": 459
},
{
"epoch": 1.117861482381531,
"grad_norm": 0.38835424184799194,
"learning_rate": 5e-05,
"loss": 0.0866,
"step": 460
},
{
"epoch": 1.120291616038882,
"grad_norm": 0.7463281750679016,
"learning_rate": 5e-05,
"loss": 0.0815,
"step": 461
},
{
"epoch": 1.1227217496962334,
"grad_norm": 0.6839048266410828,
"learning_rate": 5e-05,
"loss": 0.1349,
"step": 462
},
{
"epoch": 1.1251518833535845,
"grad_norm": 1.4645401239395142,
"learning_rate": 5e-05,
"loss": 0.0799,
"step": 463
},
{
"epoch": 1.1275820170109356,
"grad_norm": 0.7507435083389282,
"learning_rate": 5e-05,
"loss": 0.0598,
"step": 464
},
{
"epoch": 1.1300121506682868,
"grad_norm": 0.7818620800971985,
"learning_rate": 5e-05,
"loss": 0.0938,
"step": 465
},
{
"epoch": 1.1324422843256379,
"grad_norm": 0.6816830635070801,
"learning_rate": 5e-05,
"loss": 0.0787,
"step": 466
},
{
"epoch": 1.134872417982989,
"grad_norm": 0.575626015663147,
"learning_rate": 5e-05,
"loss": 0.1461,
"step": 467
},
{
"epoch": 1.13730255164034,
"grad_norm": 0.73714679479599,
"learning_rate": 5e-05,
"loss": 0.0761,
"step": 468
},
{
"epoch": 1.1397326852976915,
"grad_norm": 0.7288960814476013,
"learning_rate": 5e-05,
"loss": 0.074,
"step": 469
},
{
"epoch": 1.1421628189550426,
"grad_norm": 0.5384907126426697,
"learning_rate": 5e-05,
"loss": 0.0732,
"step": 470
},
{
"epoch": 1.1445929526123937,
"grad_norm": 0.6071632504463196,
"learning_rate": 5e-05,
"loss": 0.03,
"step": 471
},
{
"epoch": 1.1470230862697448,
"grad_norm": 0.5744608044624329,
"learning_rate": 5e-05,
"loss": 0.0912,
"step": 472
},
{
"epoch": 1.149453219927096,
"grad_norm": 0.9392772912979126,
"learning_rate": 5e-05,
"loss": 0.1024,
"step": 473
},
{
"epoch": 1.1518833535844473,
"grad_norm": 1.170997142791748,
"learning_rate": 5e-05,
"loss": 0.1027,
"step": 474
},
{
"epoch": 1.1543134872417984,
"grad_norm": 0.791221559047699,
"learning_rate": 5e-05,
"loss": 0.1305,
"step": 475
},
{
"epoch": 1.1567436208991495,
"grad_norm": 0.5443055033683777,
"learning_rate": 5e-05,
"loss": 0.0749,
"step": 476
},
{
"epoch": 1.1591737545565006,
"grad_norm": 0.8006246089935303,
"learning_rate": 5e-05,
"loss": 0.0976,
"step": 477
},
{
"epoch": 1.1616038882138517,
"grad_norm": 0.47016945481300354,
"learning_rate": 5e-05,
"loss": 0.0386,
"step": 478
},
{
"epoch": 1.1640340218712029,
"grad_norm": 0.7536148428916931,
"learning_rate": 5e-05,
"loss": 0.1385,
"step": 479
},
{
"epoch": 1.166464155528554,
"grad_norm": 0.6575655341148376,
"learning_rate": 5e-05,
"loss": 0.0851,
"step": 480
},
{
"epoch": 1.1688942891859053,
"grad_norm": 0.484667032957077,
"learning_rate": 5e-05,
"loss": 0.0301,
"step": 481
},
{
"epoch": 1.1713244228432564,
"grad_norm": 0.8194088339805603,
"learning_rate": 5e-05,
"loss": 0.129,
"step": 482
},
{
"epoch": 1.1737545565006076,
"grad_norm": 0.6559669375419617,
"learning_rate": 5e-05,
"loss": 0.0667,
"step": 483
},
{
"epoch": 1.1761846901579587,
"grad_norm": 0.8752015829086304,
"learning_rate": 5e-05,
"loss": 0.0637,
"step": 484
},
{
"epoch": 1.1786148238153098,
"grad_norm": 0.6029524207115173,
"learning_rate": 5e-05,
"loss": 0.0557,
"step": 485
},
{
"epoch": 1.181044957472661,
"grad_norm": 0.5431787371635437,
"learning_rate": 5e-05,
"loss": 0.0338,
"step": 486
},
{
"epoch": 1.183475091130012,
"grad_norm": 0.5199941992759705,
"learning_rate": 5e-05,
"loss": 0.0603,
"step": 487
},
{
"epoch": 1.1859052247873634,
"grad_norm": 0.520455539226532,
"learning_rate": 5e-05,
"loss": 0.0221,
"step": 488
},
{
"epoch": 1.1883353584447145,
"grad_norm": 0.5124850273132324,
"learning_rate": 5e-05,
"loss": 0.0306,
"step": 489
},
{
"epoch": 1.1907654921020656,
"grad_norm": 0.6515238881111145,
"learning_rate": 5e-05,
"loss": 0.0612,
"step": 490
},
{
"epoch": 1.1931956257594167,
"grad_norm": 0.816235363483429,
"learning_rate": 5e-05,
"loss": 0.1585,
"step": 491
},
{
"epoch": 1.1956257594167679,
"grad_norm": 0.6307040452957153,
"learning_rate": 5e-05,
"loss": 0.0551,
"step": 492
},
{
"epoch": 1.1980558930741192,
"grad_norm": 0.46716415882110596,
"learning_rate": 5e-05,
"loss": 0.0457,
"step": 493
},
{
"epoch": 1.2004860267314703,
"grad_norm": 0.5497251749038696,
"learning_rate": 5e-05,
"loss": 0.0874,
"step": 494
},
{
"epoch": 1.2029161603888214,
"grad_norm": 1.0956586599349976,
"learning_rate": 5e-05,
"loss": 0.0796,
"step": 495
},
{
"epoch": 1.2053462940461726,
"grad_norm": 0.6908737421035767,
"learning_rate": 5e-05,
"loss": 0.0406,
"step": 496
},
{
"epoch": 1.2077764277035237,
"grad_norm": 0.8598071932792664,
"learning_rate": 5e-05,
"loss": 0.1157,
"step": 497
},
{
"epoch": 1.2102065613608748,
"grad_norm": 0.6752752065658569,
"learning_rate": 5e-05,
"loss": 0.068,
"step": 498
},
{
"epoch": 1.212636695018226,
"grad_norm": 0.5425324440002441,
"learning_rate": 5e-05,
"loss": 0.0976,
"step": 499
},
{
"epoch": 1.2150668286755772,
"grad_norm": 0.7367557883262634,
"learning_rate": 5e-05,
"loss": 0.1028,
"step": 500
},
{
"epoch": 1.2150668286755772,
"eval_loss": 0.23548956215381622,
"eval_runtime": 512.2983,
"eval_samples_per_second": 5.364,
"eval_steps_per_second": 0.671,
"step": 500
},
{
"epoch": 1.2174969623329284,
"grad_norm": 0.8351901769638062,
"learning_rate": 5e-05,
"loss": 0.0927,
"step": 501
},
{
"epoch": 1.2199270959902795,
"grad_norm": 0.5388638377189636,
"learning_rate": 5e-05,
"loss": 0.0956,
"step": 502
},
{
"epoch": 1.2223572296476306,
"grad_norm": 0.9194818139076233,
"learning_rate": 5e-05,
"loss": 0.0533,
"step": 503
},
{
"epoch": 1.2247873633049817,
"grad_norm": 0.7008382081985474,
"learning_rate": 5e-05,
"loss": 0.0554,
"step": 504
},
{
"epoch": 1.2272174969623328,
"grad_norm": 0.5537183284759521,
"learning_rate": 5e-05,
"loss": 0.1277,
"step": 505
},
{
"epoch": 1.229647630619684,
"grad_norm": 0.570819079875946,
"learning_rate": 5e-05,
"loss": 0.1037,
"step": 506
},
{
"epoch": 1.2320777642770353,
"grad_norm": 0.5867807269096375,
"learning_rate": 5e-05,
"loss": 0.0578,
"step": 507
},
{
"epoch": 1.2345078979343864,
"grad_norm": 0.7660566568374634,
"learning_rate": 5e-05,
"loss": 0.0856,
"step": 508
},
{
"epoch": 1.2369380315917375,
"grad_norm": 0.7493465542793274,
"learning_rate": 5e-05,
"loss": 0.0567,
"step": 509
},
{
"epoch": 1.2393681652490887,
"grad_norm": 0.6718665957450867,
"learning_rate": 5e-05,
"loss": 0.1113,
"step": 510
},
{
"epoch": 1.2417982989064398,
"grad_norm": 0.7018153667449951,
"learning_rate": 5e-05,
"loss": 0.0966,
"step": 511
},
{
"epoch": 1.2442284325637911,
"grad_norm": 1.3940482139587402,
"learning_rate": 5e-05,
"loss": 0.1352,
"step": 512
},
{
"epoch": 1.2466585662211422,
"grad_norm": 0.663314938545227,
"learning_rate": 5e-05,
"loss": 0.0556,
"step": 513
},
{
"epoch": 1.2490886998784934,
"grad_norm": 1.0090134143829346,
"learning_rate": 5e-05,
"loss": 0.0352,
"step": 514
},
{
"epoch": 1.2515188335358445,
"grad_norm": 0.5742064714431763,
"learning_rate": 5e-05,
"loss": 0.0669,
"step": 515
},
{
"epoch": 1.2539489671931956,
"grad_norm": 0.726700484752655,
"learning_rate": 5e-05,
"loss": 0.0533,
"step": 516
},
{
"epoch": 1.2563791008505467,
"grad_norm": 0.6788687109947205,
"learning_rate": 5e-05,
"loss": 0.0567,
"step": 517
},
{
"epoch": 1.2588092345078978,
"grad_norm": 0.48673132061958313,
"learning_rate": 5e-05,
"loss": 0.0406,
"step": 518
},
{
"epoch": 1.2612393681652492,
"grad_norm": 1.1016216278076172,
"learning_rate": 5e-05,
"loss": 0.0668,
"step": 519
},
{
"epoch": 1.2636695018226003,
"grad_norm": 1.2345398664474487,
"learning_rate": 5e-05,
"loss": 0.029,
"step": 520
},
{
"epoch": 1.2660996354799514,
"grad_norm": 0.7507830858230591,
"learning_rate": 5e-05,
"loss": 0.0991,
"step": 521
},
{
"epoch": 1.2685297691373025,
"grad_norm": 0.819266140460968,
"learning_rate": 5e-05,
"loss": 0.0723,
"step": 522
},
{
"epoch": 1.2709599027946537,
"grad_norm": 0.9876227974891663,
"learning_rate": 5e-05,
"loss": 0.1555,
"step": 523
},
{
"epoch": 1.273390036452005,
"grad_norm": 0.7902403473854065,
"learning_rate": 5e-05,
"loss": 0.156,
"step": 524
},
{
"epoch": 1.2758201701093559,
"grad_norm": 0.6342129111289978,
"learning_rate": 5e-05,
"loss": 0.1161,
"step": 525
},
{
"epoch": 1.2782503037667072,
"grad_norm": 0.9039121866226196,
"learning_rate": 5e-05,
"loss": 0.1078,
"step": 526
},
{
"epoch": 1.2806804374240583,
"grad_norm": 0.9820278882980347,
"learning_rate": 5e-05,
"loss": 0.1182,
"step": 527
},
{
"epoch": 1.2831105710814095,
"grad_norm": 0.45999497175216675,
"learning_rate": 5e-05,
"loss": 0.0469,
"step": 528
},
{
"epoch": 1.2855407047387606,
"grad_norm": 0.6754646301269531,
"learning_rate": 5e-05,
"loss": 0.0875,
"step": 529
},
{
"epoch": 1.2879708383961117,
"grad_norm": 0.7348721027374268,
"learning_rate": 5e-05,
"loss": 0.0639,
"step": 530
},
{
"epoch": 1.290400972053463,
"grad_norm": 0.7348721027374268,
"learning_rate": 5e-05,
"loss": 0.0764,
"step": 531
},
{
"epoch": 1.2928311057108142,
"grad_norm": 0.4331410229206085,
"learning_rate": 5e-05,
"loss": 0.0278,
"step": 532
},
{
"epoch": 1.2952612393681653,
"grad_norm": 0.7986178994178772,
"learning_rate": 5e-05,
"loss": 0.0893,
"step": 533
},
{
"epoch": 1.2976913730255164,
"grad_norm": 0.7255426049232483,
"learning_rate": 5e-05,
"loss": 0.0802,
"step": 534
},
{
"epoch": 1.3001215066828675,
"grad_norm": 0.590578556060791,
"learning_rate": 5e-05,
"loss": 0.0853,
"step": 535
},
{
"epoch": 1.3025516403402186,
"grad_norm": 0.6147257685661316,
"learning_rate": 5e-05,
"loss": 0.0462,
"step": 536
},
{
"epoch": 1.3049817739975698,
"grad_norm": 0.6174589991569519,
"learning_rate": 5e-05,
"loss": 0.0758,
"step": 537
},
{
"epoch": 1.307411907654921,
"grad_norm": 0.5428439378738403,
"learning_rate": 5e-05,
"loss": 0.1452,
"step": 538
},
{
"epoch": 1.3098420413122722,
"grad_norm": 0.554800271987915,
"learning_rate": 5e-05,
"loss": 0.0266,
"step": 539
},
{
"epoch": 1.3122721749696233,
"grad_norm": 0.627275824546814,
"learning_rate": 5e-05,
"loss": 0.0604,
"step": 540
},
{
"epoch": 1.3147023086269745,
"grad_norm": 0.5268173813819885,
"learning_rate": 5e-05,
"loss": 0.036,
"step": 541
},
{
"epoch": 1.3171324422843256,
"grad_norm": 0.4525201916694641,
"learning_rate": 5e-05,
"loss": 0.0746,
"step": 542
},
{
"epoch": 1.319562575941677,
"grad_norm": 1.299209475517273,
"learning_rate": 5e-05,
"loss": 0.0573,
"step": 543
},
{
"epoch": 1.3219927095990278,
"grad_norm": 0.5899752378463745,
"learning_rate": 5e-05,
"loss": 0.0373,
"step": 544
},
{
"epoch": 1.3244228432563792,
"grad_norm": 0.5044320225715637,
"learning_rate": 5e-05,
"loss": 0.0415,
"step": 545
},
{
"epoch": 1.3268529769137303,
"grad_norm": 0.4812248945236206,
"learning_rate": 5e-05,
"loss": 0.0898,
"step": 546
},
{
"epoch": 1.3292831105710814,
"grad_norm": 0.47461068630218506,
"learning_rate": 5e-05,
"loss": 0.0248,
"step": 547
},
{
"epoch": 1.3317132442284325,
"grad_norm": 0.5620236396789551,
"learning_rate": 5e-05,
"loss": 0.0418,
"step": 548
},
{
"epoch": 1.3341433778857836,
"grad_norm": 0.8658341765403748,
"learning_rate": 5e-05,
"loss": 0.0208,
"step": 549
},
{
"epoch": 1.336573511543135,
"grad_norm": 0.6640876531600952,
"learning_rate": 5e-05,
"loss": 0.0652,
"step": 550
},
{
"epoch": 1.339003645200486,
"grad_norm": 0.3873005509376526,
"learning_rate": 5e-05,
"loss": 0.018,
"step": 551
},
{
"epoch": 1.3414337788578372,
"grad_norm": 0.4546399414539337,
"learning_rate": 5e-05,
"loss": 0.0823,
"step": 552
},
{
"epoch": 1.3438639125151883,
"grad_norm": 0.5188842415809631,
"learning_rate": 5e-05,
"loss": 0.0604,
"step": 553
},
{
"epoch": 1.3462940461725394,
"grad_norm": 0.41030353307724,
"learning_rate": 5e-05,
"loss": 0.0112,
"step": 554
},
{
"epoch": 1.3487241798298906,
"grad_norm": 0.7154520750045776,
"learning_rate": 5e-05,
"loss": 0.0687,
"step": 555
},
{
"epoch": 1.3511543134872417,
"grad_norm": 0.6743410229682922,
"learning_rate": 5e-05,
"loss": 0.1262,
"step": 556
},
{
"epoch": 1.353584447144593,
"grad_norm": 0.38846057653427124,
"learning_rate": 5e-05,
"loss": 0.0371,
"step": 557
},
{
"epoch": 1.3560145808019441,
"grad_norm": 0.5022516846656799,
"learning_rate": 5e-05,
"loss": 0.0159,
"step": 558
},
{
"epoch": 1.3584447144592953,
"grad_norm": 0.5445972084999084,
"learning_rate": 5e-05,
"loss": 0.0652,
"step": 559
},
{
"epoch": 1.3608748481166464,
"grad_norm": 0.6346146464347839,
"learning_rate": 5e-05,
"loss": 0.1128,
"step": 560
},
{
"epoch": 1.3633049817739975,
"grad_norm": 0.7129392623901367,
"learning_rate": 5e-05,
"loss": 0.059,
"step": 561
},
{
"epoch": 1.3657351154313488,
"grad_norm": 0.4905102252960205,
"learning_rate": 5e-05,
"loss": 0.1114,
"step": 562
},
{
"epoch": 1.3681652490886997,
"grad_norm": 0.7290562391281128,
"learning_rate": 5e-05,
"loss": 0.1029,
"step": 563
},
{
"epoch": 1.370595382746051,
"grad_norm": 0.5576648116111755,
"learning_rate": 5e-05,
"loss": 0.0383,
"step": 564
},
{
"epoch": 1.3730255164034022,
"grad_norm": 0.595310389995575,
"learning_rate": 5e-05,
"loss": 0.0906,
"step": 565
},
{
"epoch": 1.3754556500607533,
"grad_norm": 0.7409107089042664,
"learning_rate": 5e-05,
"loss": 0.0239,
"step": 566
},
{
"epoch": 1.3778857837181044,
"grad_norm": 1.0220900774002075,
"learning_rate": 5e-05,
"loss": 0.0537,
"step": 567
},
{
"epoch": 1.3803159173754556,
"grad_norm": 1.0970425605773926,
"learning_rate": 5e-05,
"loss": 0.0615,
"step": 568
},
{
"epoch": 1.382746051032807,
"grad_norm": 0.45453736186027527,
"learning_rate": 5e-05,
"loss": 0.1,
"step": 569
},
{
"epoch": 1.385176184690158,
"grad_norm": 0.61972576379776,
"learning_rate": 5e-05,
"loss": 0.1029,
"step": 570
},
{
"epoch": 1.3876063183475091,
"grad_norm": 0.47734493017196655,
"learning_rate": 5e-05,
"loss": 0.0683,
"step": 571
},
{
"epoch": 1.3900364520048603,
"grad_norm": 0.3732990026473999,
"learning_rate": 5e-05,
"loss": 0.0706,
"step": 572
},
{
"epoch": 1.3924665856622114,
"grad_norm": 0.49241572618484497,
"learning_rate": 5e-05,
"loss": 0.0667,
"step": 573
},
{
"epoch": 1.3948967193195625,
"grad_norm": 0.6907554268836975,
"learning_rate": 5e-05,
"loss": 0.0602,
"step": 574
},
{
"epoch": 1.3973268529769136,
"grad_norm": 0.8935093879699707,
"learning_rate": 5e-05,
"loss": 0.2114,
"step": 575
},
{
"epoch": 1.399756986634265,
"grad_norm": 0.2839941680431366,
"learning_rate": 5e-05,
"loss": 0.0257,
"step": 576
},
{
"epoch": 1.402187120291616,
"grad_norm": 0.7202839255332947,
"learning_rate": 5e-05,
"loss": 0.0878,
"step": 577
},
{
"epoch": 1.4046172539489672,
"grad_norm": 0.6520135402679443,
"learning_rate": 5e-05,
"loss": 0.0295,
"step": 578
},
{
"epoch": 1.4070473876063183,
"grad_norm": 0.7360886335372925,
"learning_rate": 5e-05,
"loss": 0.065,
"step": 579
},
{
"epoch": 1.4094775212636694,
"grad_norm": 0.5855164527893066,
"learning_rate": 5e-05,
"loss": 0.0752,
"step": 580
},
{
"epoch": 1.4119076549210208,
"grad_norm": 0.38474756479263306,
"learning_rate": 5e-05,
"loss": 0.0541,
"step": 581
},
{
"epoch": 1.414337788578372,
"grad_norm": 0.5105370283126831,
"learning_rate": 5e-05,
"loss": 0.0749,
"step": 582
},
{
"epoch": 1.416767922235723,
"grad_norm": 0.5076673030853271,
"learning_rate": 5e-05,
"loss": 0.0528,
"step": 583
},
{
"epoch": 1.4191980558930741,
"grad_norm": 0.5034826397895813,
"learning_rate": 5e-05,
"loss": 0.0505,
"step": 584
},
{
"epoch": 1.4216281895504252,
"grad_norm": 0.6378771066665649,
"learning_rate": 5e-05,
"loss": 0.0676,
"step": 585
},
{
"epoch": 1.4240583232077764,
"grad_norm": 0.5122814774513245,
"learning_rate": 5e-05,
"loss": 0.0515,
"step": 586
},
{
"epoch": 1.4264884568651275,
"grad_norm": 0.5807358026504517,
"learning_rate": 5e-05,
"loss": 0.0507,
"step": 587
},
{
"epoch": 1.4289185905224788,
"grad_norm": 0.5156981348991394,
"learning_rate": 5e-05,
"loss": 0.0503,
"step": 588
},
{
"epoch": 1.43134872417983,
"grad_norm": 0.9181635975837708,
"learning_rate": 5e-05,
"loss": 0.0964,
"step": 589
},
{
"epoch": 1.433778857837181,
"grad_norm": 0.38104763627052307,
"learning_rate": 5e-05,
"loss": 0.055,
"step": 590
},
{
"epoch": 1.4362089914945322,
"grad_norm": 0.4987882673740387,
"learning_rate": 5e-05,
"loss": 0.0347,
"step": 591
},
{
"epoch": 1.4386391251518833,
"grad_norm": 0.5201525688171387,
"learning_rate": 5e-05,
"loss": 0.0904,
"step": 592
},
{
"epoch": 1.4410692588092346,
"grad_norm": 0.5238099694252014,
"learning_rate": 5e-05,
"loss": 0.077,
"step": 593
},
{
"epoch": 1.4434993924665855,
"grad_norm": 0.924893319606781,
"learning_rate": 5e-05,
"loss": 0.1248,
"step": 594
},
{
"epoch": 1.4459295261239369,
"grad_norm": 0.6318195462226868,
"learning_rate": 5e-05,
"loss": 0.0267,
"step": 595
},
{
"epoch": 1.448359659781288,
"grad_norm": 0.9913945198059082,
"learning_rate": 5e-05,
"loss": 0.0917,
"step": 596
},
{
"epoch": 1.4507897934386391,
"grad_norm": 0.4909703731536865,
"learning_rate": 5e-05,
"loss": 0.036,
"step": 597
},
{
"epoch": 1.4532199270959902,
"grad_norm": 0.5913828015327454,
"learning_rate": 5e-05,
"loss": 0.0353,
"step": 598
},
{
"epoch": 1.4556500607533414,
"grad_norm": 0.3615020215511322,
"learning_rate": 5e-05,
"loss": 0.0517,
"step": 599
},
{
"epoch": 1.4580801944106927,
"grad_norm": 0.5106006264686584,
"learning_rate": 5e-05,
"loss": 0.0444,
"step": 600
},
{
"epoch": 1.4580801944106927,
"eval_loss": 0.22935771942138672,
"eval_runtime": 507.4275,
"eval_samples_per_second": 5.416,
"eval_steps_per_second": 0.678,
"step": 600
},
{
"epoch": 1.4605103280680438,
"grad_norm": 0.7134807705879211,
"learning_rate": 5e-05,
"loss": 0.056,
"step": 601
},
{
"epoch": 1.462940461725395,
"grad_norm": 0.2919784188270569,
"learning_rate": 5e-05,
"loss": 0.0146,
"step": 602
},
{
"epoch": 1.465370595382746,
"grad_norm": 0.6526296734809875,
"learning_rate": 5e-05,
"loss": 0.1207,
"step": 603
},
{
"epoch": 1.4678007290400972,
"grad_norm": 0.5833859443664551,
"learning_rate": 5e-05,
"loss": 0.0657,
"step": 604
},
{
"epoch": 1.4702308626974483,
"grad_norm": 0.6107626557350159,
"learning_rate": 5e-05,
"loss": 0.0655,
"step": 605
},
{
"epoch": 1.4726609963547994,
"grad_norm": 0.5227603912353516,
"learning_rate": 5e-05,
"loss": 0.0504,
"step": 606
},
{
"epoch": 1.4750911300121508,
"grad_norm": 0.6677300333976746,
"learning_rate": 5e-05,
"loss": 0.0465,
"step": 607
},
{
"epoch": 1.4775212636695019,
"grad_norm": 0.7898095846176147,
"learning_rate": 5e-05,
"loss": 0.0932,
"step": 608
},
{
"epoch": 1.479951397326853,
"grad_norm": 0.9366335868835449,
"learning_rate": 5e-05,
"loss": 0.0819,
"step": 609
},
{
"epoch": 1.482381530984204,
"grad_norm": 0.6056090593338013,
"learning_rate": 5e-05,
"loss": 0.0638,
"step": 610
},
{
"epoch": 1.4848116646415552,
"grad_norm": 0.7640931606292725,
"learning_rate": 5e-05,
"loss": 0.1152,
"step": 611
},
{
"epoch": 1.4872417982989066,
"grad_norm": 0.7496662735939026,
"learning_rate": 5e-05,
"loss": 0.0461,
"step": 612
},
{
"epoch": 1.4896719319562575,
"grad_norm": 0.6870647072792053,
"learning_rate": 5e-05,
"loss": 0.0812,
"step": 613
},
{
"epoch": 1.4921020656136088,
"grad_norm": 0.7589767575263977,
"learning_rate": 5e-05,
"loss": 0.0846,
"step": 614
},
{
"epoch": 1.49453219927096,
"grad_norm": 0.681777834892273,
"learning_rate": 5e-05,
"loss": 0.0407,
"step": 615
},
{
"epoch": 1.496962332928311,
"grad_norm": 0.7369627952575684,
"learning_rate": 5e-05,
"loss": 0.0955,
"step": 616
},
{
"epoch": 1.4993924665856622,
"grad_norm": 0.7186546921730042,
"learning_rate": 5e-05,
"loss": 0.0745,
"step": 617
},
{
"epoch": 1.5018226002430133,
"grad_norm": 0.6271875500679016,
"learning_rate": 5e-05,
"loss": 0.0533,
"step": 618
},
{
"epoch": 1.5042527339003646,
"grad_norm": 0.5378838777542114,
"learning_rate": 5e-05,
"loss": 0.0688,
"step": 619
},
{
"epoch": 1.5066828675577155,
"grad_norm": 0.7019214630126953,
"learning_rate": 5e-05,
"loss": 0.0369,
"step": 620
},
{
"epoch": 1.5091130012150669,
"grad_norm": 0.9618707299232483,
"learning_rate": 5e-05,
"loss": 0.0469,
"step": 621
},
{
"epoch": 1.511543134872418,
"grad_norm": 0.4820462167263031,
"learning_rate": 5e-05,
"loss": 0.074,
"step": 622
},
{
"epoch": 1.513973268529769,
"grad_norm": 0.411123126745224,
"learning_rate": 5e-05,
"loss": 0.0307,
"step": 623
},
{
"epoch": 1.5164034021871204,
"grad_norm": 0.5718781352043152,
"learning_rate": 5e-05,
"loss": 0.0276,
"step": 624
},
{
"epoch": 1.5188335358444713,
"grad_norm": 0.4341799020767212,
"learning_rate": 5e-05,
"loss": 0.0651,
"step": 625
},
{
"epoch": 1.5212636695018227,
"grad_norm": 0.515855073928833,
"learning_rate": 5e-05,
"loss": 0.0413,
"step": 626
},
{
"epoch": 1.5236938031591738,
"grad_norm": 0.6174566149711609,
"learning_rate": 5e-05,
"loss": 0.0307,
"step": 627
},
{
"epoch": 1.526123936816525,
"grad_norm": 0.5931458473205566,
"learning_rate": 5e-05,
"loss": 0.0852,
"step": 628
},
{
"epoch": 1.528554070473876,
"grad_norm": 0.46777036786079407,
"learning_rate": 5e-05,
"loss": 0.0438,
"step": 629
},
{
"epoch": 1.5309842041312272,
"grad_norm": 0.4254392087459564,
"learning_rate": 5e-05,
"loss": 0.0606,
"step": 630
},
{
"epoch": 1.5334143377885785,
"grad_norm": 0.5670008063316345,
"learning_rate": 5e-05,
"loss": 0.0514,
"step": 631
},
{
"epoch": 1.5358444714459294,
"grad_norm": 1.1566509008407593,
"learning_rate": 5e-05,
"loss": 0.1133,
"step": 632
},
{
"epoch": 1.5382746051032807,
"grad_norm": 0.5660587549209595,
"learning_rate": 5e-05,
"loss": 0.0544,
"step": 633
},
{
"epoch": 1.5407047387606319,
"grad_norm": 0.5406026840209961,
"learning_rate": 5e-05,
"loss": 0.046,
"step": 634
},
{
"epoch": 1.543134872417983,
"grad_norm": 0.49723613262176514,
"learning_rate": 5e-05,
"loss": 0.0177,
"step": 635
},
{
"epoch": 1.545565006075334,
"grad_norm": 1.263848900794983,
"learning_rate": 5e-05,
"loss": 0.0659,
"step": 636
},
{
"epoch": 1.5479951397326852,
"grad_norm": 0.6971784830093384,
"learning_rate": 5e-05,
"loss": 0.0747,
"step": 637
},
{
"epoch": 1.5504252733900366,
"grad_norm": 0.5171895623207092,
"learning_rate": 5e-05,
"loss": 0.0512,
"step": 638
},
{
"epoch": 1.5528554070473874,
"grad_norm": 0.7318109273910522,
"learning_rate": 5e-05,
"loss": 0.0655,
"step": 639
},
{
"epoch": 1.5552855407047388,
"grad_norm": 0.46023720502853394,
"learning_rate": 5e-05,
"loss": 0.0568,
"step": 640
},
{
"epoch": 1.55771567436209,
"grad_norm": 0.7200696468353271,
"learning_rate": 5e-05,
"loss": 0.0511,
"step": 641
},
{
"epoch": 1.560145808019441,
"grad_norm": 1.0338624715805054,
"learning_rate": 5e-05,
"loss": 0.0822,
"step": 642
},
{
"epoch": 1.5625759416767924,
"grad_norm": 0.7976377010345459,
"learning_rate": 5e-05,
"loss": 0.0623,
"step": 643
},
{
"epoch": 1.5650060753341433,
"grad_norm": 0.6757969260215759,
"learning_rate": 5e-05,
"loss": 0.0737,
"step": 644
},
{
"epoch": 1.5674362089914946,
"grad_norm": 1.4083917140960693,
"learning_rate": 5e-05,
"loss": 0.0374,
"step": 645
},
{
"epoch": 1.5698663426488457,
"grad_norm": 0.348634272813797,
"learning_rate": 5e-05,
"loss": 0.0249,
"step": 646
},
{
"epoch": 1.5722964763061968,
"grad_norm": 0.7224584817886353,
"learning_rate": 5e-05,
"loss": 0.1013,
"step": 647
},
{
"epoch": 1.574726609963548,
"grad_norm": 0.919998288154602,
"learning_rate": 5e-05,
"loss": 0.1093,
"step": 648
},
{
"epoch": 1.577156743620899,
"grad_norm": 0.7341464161872864,
"learning_rate": 5e-05,
"loss": 0.1234,
"step": 649
},
{
"epoch": 1.5795868772782504,
"grad_norm": 0.8548154234886169,
"learning_rate": 5e-05,
"loss": 0.1628,
"step": 650
},
{
"epoch": 1.5820170109356013,
"grad_norm": 0.7931060194969177,
"learning_rate": 5e-05,
"loss": 0.1439,
"step": 651
},
{
"epoch": 1.5844471445929527,
"grad_norm": 0.9627839922904968,
"learning_rate": 5e-05,
"loss": 0.0327,
"step": 652
},
{
"epoch": 1.5868772782503038,
"grad_norm": 0.8406124114990234,
"learning_rate": 5e-05,
"loss": 0.0574,
"step": 653
},
{
"epoch": 1.589307411907655,
"grad_norm": 0.7667601108551025,
"learning_rate": 5e-05,
"loss": 0.0551,
"step": 654
},
{
"epoch": 1.5917375455650062,
"grad_norm": 0.4747026264667511,
"learning_rate": 5e-05,
"loss": 0.052,
"step": 655
},
{
"epoch": 1.5941676792223571,
"grad_norm": 0.632870078086853,
"learning_rate": 5e-05,
"loss": 0.0587,
"step": 656
},
{
"epoch": 1.5965978128797085,
"grad_norm": 0.7700726985931396,
"learning_rate": 5e-05,
"loss": 0.0986,
"step": 657
},
{
"epoch": 1.5990279465370594,
"grad_norm": 1.121201992034912,
"learning_rate": 5e-05,
"loss": 0.0318,
"step": 658
},
{
"epoch": 1.6014580801944107,
"grad_norm": 0.6213607788085938,
"learning_rate": 5e-05,
"loss": 0.0932,
"step": 659
},
{
"epoch": 1.6038882138517618,
"grad_norm": 0.30946460366249084,
"learning_rate": 5e-05,
"loss": 0.0333,
"step": 660
},
{
"epoch": 1.606318347509113,
"grad_norm": 0.5171956419944763,
"learning_rate": 5e-05,
"loss": 0.0617,
"step": 661
},
{
"epoch": 1.6087484811664643,
"grad_norm": 0.6288155317306519,
"learning_rate": 5e-05,
"loss": 0.0506,
"step": 662
},
{
"epoch": 1.6111786148238152,
"grad_norm": 0.6943643689155579,
"learning_rate": 5e-05,
"loss": 0.0285,
"step": 663
},
{
"epoch": 1.6136087484811665,
"grad_norm": 0.8144752383232117,
"learning_rate": 5e-05,
"loss": 0.0484,
"step": 664
},
{
"epoch": 1.6160388821385177,
"grad_norm": 0.4773041307926178,
"learning_rate": 5e-05,
"loss": 0.0993,
"step": 665
},
{
"epoch": 1.6184690157958688,
"grad_norm": 0.7703022956848145,
"learning_rate": 5e-05,
"loss": 0.0735,
"step": 666
},
{
"epoch": 1.62089914945322,
"grad_norm": 0.6951743960380554,
"learning_rate": 5e-05,
"loss": 0.0446,
"step": 667
},
{
"epoch": 1.623329283110571,
"grad_norm": 0.4781995415687561,
"learning_rate": 5e-05,
"loss": 0.0178,
"step": 668
},
{
"epoch": 1.6257594167679223,
"grad_norm": 0.5066447854042053,
"learning_rate": 5e-05,
"loss": 0.0969,
"step": 669
},
{
"epoch": 1.6281895504252732,
"grad_norm": 0.4288381040096283,
"learning_rate": 5e-05,
"loss": 0.0317,
"step": 670
},
{
"epoch": 1.6306196840826246,
"grad_norm": 0.5544601678848267,
"learning_rate": 5e-05,
"loss": 0.035,
"step": 671
},
{
"epoch": 1.6330498177399757,
"grad_norm": 0.48424339294433594,
"learning_rate": 5e-05,
"loss": 0.0231,
"step": 672
},
{
"epoch": 1.6354799513973268,
"grad_norm": 0.770346462726593,
"learning_rate": 5e-05,
"loss": 0.0767,
"step": 673
},
{
"epoch": 1.6379100850546782,
"grad_norm": 0.7101414203643799,
"learning_rate": 5e-05,
"loss": 0.0247,
"step": 674
},
{
"epoch": 1.640340218712029,
"grad_norm": 0.9238234758377075,
"learning_rate": 5e-05,
"loss": 0.0528,
"step": 675
},
{
"epoch": 1.6427703523693804,
"grad_norm": 0.7192332148551941,
"learning_rate": 5e-05,
"loss": 0.0738,
"step": 676
},
{
"epoch": 1.6452004860267315,
"grad_norm": 0.5393381118774414,
"learning_rate": 5e-05,
"loss": 0.0337,
"step": 677
},
{
"epoch": 1.6476306196840826,
"grad_norm": 0.2713688015937805,
"learning_rate": 5e-05,
"loss": 0.0159,
"step": 678
},
{
"epoch": 1.6500607533414338,
"grad_norm": 0.3296976089477539,
"learning_rate": 5e-05,
"loss": 0.0608,
"step": 679
},
{
"epoch": 1.6524908869987849,
"grad_norm": 0.4995911717414856,
"learning_rate": 5e-05,
"loss": 0.0119,
"step": 680
},
{
"epoch": 1.6549210206561362,
"grad_norm": 0.5983487367630005,
"learning_rate": 5e-05,
"loss": 0.0188,
"step": 681
},
{
"epoch": 1.6573511543134871,
"grad_norm": 0.4184721112251282,
"learning_rate": 5e-05,
"loss": 0.0207,
"step": 682
},
{
"epoch": 1.6597812879708385,
"grad_norm": 0.6056463122367859,
"learning_rate": 5e-05,
"loss": 0.0321,
"step": 683
},
{
"epoch": 1.6622114216281896,
"grad_norm": 0.8600037693977356,
"learning_rate": 5e-05,
"loss": 0.0716,
"step": 684
},
{
"epoch": 1.6646415552855407,
"grad_norm": 0.7863733768463135,
"learning_rate": 5e-05,
"loss": 0.1251,
"step": 685
},
{
"epoch": 1.6670716889428918,
"grad_norm": 0.4304426908493042,
"learning_rate": 5e-05,
"loss": 0.0824,
"step": 686
},
{
"epoch": 1.669501822600243,
"grad_norm": 0.6984259486198425,
"learning_rate": 5e-05,
"loss": 0.0332,
"step": 687
},
{
"epoch": 1.6719319562575943,
"grad_norm": 0.400698184967041,
"learning_rate": 5e-05,
"loss": 0.0444,
"step": 688
},
{
"epoch": 1.6743620899149452,
"grad_norm": 0.5730535387992859,
"learning_rate": 5e-05,
"loss": 0.0488,
"step": 689
},
{
"epoch": 1.6767922235722965,
"grad_norm": 0.4522067606449127,
"learning_rate": 5e-05,
"loss": 0.0596,
"step": 690
},
{
"epoch": 1.6792223572296476,
"grad_norm": 0.44992417097091675,
"learning_rate": 5e-05,
"loss": 0.0311,
"step": 691
},
{
"epoch": 1.6816524908869988,
"grad_norm": 0.8305267095565796,
"learning_rate": 5e-05,
"loss": 0.0933,
"step": 692
},
{
"epoch": 1.68408262454435,
"grad_norm": 0.7188434600830078,
"learning_rate": 5e-05,
"loss": 0.0443,
"step": 693
},
{
"epoch": 1.686512758201701,
"grad_norm": 0.5052474141120911,
"learning_rate": 5e-05,
"loss": 0.1067,
"step": 694
},
{
"epoch": 1.6889428918590523,
"grad_norm": 0.686128556728363,
"learning_rate": 5e-05,
"loss": 0.055,
"step": 695
},
{
"epoch": 1.6913730255164034,
"grad_norm": 0.4116796851158142,
"learning_rate": 5e-05,
"loss": 0.0728,
"step": 696
},
{
"epoch": 1.6938031591737546,
"grad_norm": 0.44450122117996216,
"learning_rate": 5e-05,
"loss": 0.0359,
"step": 697
},
{
"epoch": 1.6962332928311057,
"grad_norm": 0.3920440375804901,
"learning_rate": 5e-05,
"loss": 0.0654,
"step": 698
},
{
"epoch": 1.6986634264884568,
"grad_norm": 0.7238947749137878,
"learning_rate": 5e-05,
"loss": 0.0593,
"step": 699
},
{
"epoch": 1.7010935601458081,
"grad_norm": 0.3644774854183197,
"learning_rate": 5e-05,
"loss": 0.054,
"step": 700
},
{
"epoch": 1.7010935601458081,
"eval_loss": 0.23474913835525513,
"eval_runtime": 505.0402,
"eval_samples_per_second": 5.441,
"eval_steps_per_second": 0.681,
"step": 700
},
{
"epoch": 1.7010935601458081,
"step": 700,
"total_flos": 2.1333404857820774e+18,
"train_loss": 0.12394342508699213,
"train_runtime": 14191.792,
"train_samples_per_second": 2.255,
"train_steps_per_second": 0.07
}
],
"logging_steps": 1.0,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"total_flos": 2.1333404857820774e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}