zephyr-7b-kto-iter0 / trainer_state.json
DatPySci's picture
End of training
6576f06 verified
raw
history blame
125 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 100,
"global_step": 1563,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 42.15456008911133,
"kl": 0.017818570137023926,
"learning_rate": 1.592356687898089e-08,
"logps/chosen": -285.75128173828125,
"logps/rejected": -254.7062530517578,
"loss": 0.4999,
"rewards/chosen": 0.004669209010899067,
"rewards/margins": 0.0025329389609396458,
"rewards/rejected": 0.0021362705156207085,
"step": 5
},
{
"epoch": 0.01,
"grad_norm": 44.65619659423828,
"kl": 0.05461766570806503,
"learning_rate": 3.184713375796178e-08,
"logps/chosen": -286.11944580078125,
"logps/rejected": -276.1832580566406,
"loss": 0.496,
"rewards/chosen": 0.03117586299777031,
"rewards/margins": 0.033454541116952896,
"rewards/rejected": -0.00227867579087615,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 39.61723327636719,
"kl": 0.44255906343460083,
"learning_rate": 4.777070063694268e-08,
"logps/chosen": -269.203125,
"logps/rejected": -260.71966552734375,
"loss": 0.4849,
"rewards/chosen": 0.13460782170295715,
"rewards/margins": 0.11511580646038055,
"rewards/rejected": 0.0194920115172863,
"step": 15
},
{
"epoch": 0.01,
"grad_norm": 36.42387771606445,
"kl": 0.871076226234436,
"learning_rate": 6.369426751592356e-08,
"logps/chosen": -244.55203247070312,
"logps/rejected": -259.15496826171875,
"loss": 0.4573,
"rewards/chosen": 0.3324822187423706,
"rewards/margins": 0.3418889045715332,
"rewards/rejected": -0.00940666627138853,
"step": 20
},
{
"epoch": 0.02,
"grad_norm": 35.794044494628906,
"kl": 0.3753414750099182,
"learning_rate": 7.961783439490445e-08,
"logps/chosen": -261.01800537109375,
"logps/rejected": -271.74786376953125,
"loss": 0.4164,
"rewards/chosen": 0.48356980085372925,
"rewards/margins": 0.7098164558410645,
"rewards/rejected": -0.22624659538269043,
"step": 25
},
{
"epoch": 0.02,
"grad_norm": 34.50333786010742,
"kl": 0.0,
"learning_rate": 9.554140127388536e-08,
"logps/chosen": -306.0127868652344,
"logps/rejected": -244.1356658935547,
"loss": 0.3649,
"rewards/chosen": 0.746610164642334,
"rewards/margins": 1.2027219533920288,
"rewards/rejected": -0.4561118483543396,
"step": 30
},
{
"epoch": 0.02,
"grad_norm": 30.50051498413086,
"kl": 0.0,
"learning_rate": 1.1146496815286624e-07,
"logps/chosen": -280.65692138671875,
"logps/rejected": -262.2502136230469,
"loss": 0.3171,
"rewards/chosen": 0.774597704410553,
"rewards/margins": 1.728833794593811,
"rewards/rejected": -0.9542360305786133,
"step": 35
},
{
"epoch": 0.03,
"grad_norm": 29.806577682495117,
"kl": 0.0,
"learning_rate": 1.2738853503184713e-07,
"logps/chosen": -276.9497375488281,
"logps/rejected": -259.6419982910156,
"loss": 0.2884,
"rewards/chosen": 0.6780191659927368,
"rewards/margins": 2.165379047393799,
"rewards/rejected": -1.4873597621917725,
"step": 40
},
{
"epoch": 0.03,
"grad_norm": 22.508939743041992,
"kl": 0.0,
"learning_rate": 1.43312101910828e-07,
"logps/chosen": -253.7576141357422,
"logps/rejected": -268.8052062988281,
"loss": 0.2596,
"rewards/chosen": 1.205094575881958,
"rewards/margins": 2.685385227203369,
"rewards/rejected": -1.4802907705307007,
"step": 45
},
{
"epoch": 0.03,
"grad_norm": 29.29058074951172,
"kl": 0.0,
"learning_rate": 1.592356687898089e-07,
"logps/chosen": -284.4584045410156,
"logps/rejected": -254.226806640625,
"loss": 0.2614,
"rewards/chosen": 1.0118796825408936,
"rewards/margins": 2.8269615173339844,
"rewards/rejected": -1.8150818347930908,
"step": 50
},
{
"epoch": 0.04,
"grad_norm": 29.25771141052246,
"kl": 0.0,
"learning_rate": 1.7515923566878978e-07,
"logps/chosen": -266.92254638671875,
"logps/rejected": -260.49261474609375,
"loss": 0.233,
"rewards/chosen": 1.16305673122406,
"rewards/margins": 3.4210219383239746,
"rewards/rejected": -2.257965326309204,
"step": 55
},
{
"epoch": 0.04,
"grad_norm": 20.642614364624023,
"kl": 0.0,
"learning_rate": 1.9108280254777072e-07,
"logps/chosen": -231.07168579101562,
"logps/rejected": -258.5835876464844,
"loss": 0.241,
"rewards/chosen": 1.3277983665466309,
"rewards/margins": 3.532343626022339,
"rewards/rejected": -2.204545497894287,
"step": 60
},
{
"epoch": 0.04,
"grad_norm": 23.786853790283203,
"kl": 0.0,
"learning_rate": 2.070063694267516e-07,
"logps/chosen": -251.5561981201172,
"logps/rejected": -267.85986328125,
"loss": 0.2376,
"rewards/chosen": 1.2313438653945923,
"rewards/margins": 3.837125062942505,
"rewards/rejected": -2.605781316757202,
"step": 65
},
{
"epoch": 0.04,
"grad_norm": 24.48412322998047,
"kl": 0.0,
"learning_rate": 2.2292993630573247e-07,
"logps/chosen": -244.3961639404297,
"logps/rejected": -248.7190704345703,
"loss": 0.2304,
"rewards/chosen": 1.338683843612671,
"rewards/margins": 3.6262855529785156,
"rewards/rejected": -2.287601947784424,
"step": 70
},
{
"epoch": 0.05,
"grad_norm": 16.654287338256836,
"kl": 0.0,
"learning_rate": 2.388535031847134e-07,
"logps/chosen": -261.95819091796875,
"logps/rejected": -291.40020751953125,
"loss": 0.2003,
"rewards/chosen": 1.3882195949554443,
"rewards/margins": 4.626662254333496,
"rewards/rejected": -3.238443374633789,
"step": 75
},
{
"epoch": 0.05,
"grad_norm": 30.89764976501465,
"kl": 0.0,
"learning_rate": 2.5477707006369425e-07,
"logps/chosen": -267.4027099609375,
"logps/rejected": -271.3486328125,
"loss": 0.1985,
"rewards/chosen": 1.4253151416778564,
"rewards/margins": 4.816943645477295,
"rewards/rejected": -3.3916287422180176,
"step": 80
},
{
"epoch": 0.05,
"grad_norm": 55.57185363769531,
"kl": 0.0,
"learning_rate": 2.7070063694267513e-07,
"logps/chosen": -289.416259765625,
"logps/rejected": -282.303955078125,
"loss": 0.2075,
"rewards/chosen": 1.308215856552124,
"rewards/margins": 5.0450119972229,
"rewards/rejected": -3.7367959022521973,
"step": 85
},
{
"epoch": 0.06,
"grad_norm": 34.31930923461914,
"kl": 0.0,
"learning_rate": 2.86624203821656e-07,
"logps/chosen": -255.4053192138672,
"logps/rejected": -293.8216247558594,
"loss": 0.1877,
"rewards/chosen": 1.560572862625122,
"rewards/margins": 5.238338470458984,
"rewards/rejected": -3.6777656078338623,
"step": 90
},
{
"epoch": 0.06,
"grad_norm": 22.21489143371582,
"kl": 0.0,
"learning_rate": 3.0254777070063694e-07,
"logps/chosen": -286.2869873046875,
"logps/rejected": -271.2530212402344,
"loss": 0.1985,
"rewards/chosen": 1.478058099746704,
"rewards/margins": 5.051484107971191,
"rewards/rejected": -3.5734260082244873,
"step": 95
},
{
"epoch": 0.06,
"grad_norm": 29.9346981048584,
"kl": 0.0,
"learning_rate": 3.184713375796178e-07,
"logps/chosen": -245.080810546875,
"logps/rejected": -287.15435791015625,
"loss": 0.1656,
"rewards/chosen": 1.4543142318725586,
"rewards/margins": 5.874560356140137,
"rewards/rejected": -4.420246124267578,
"step": 100
},
{
"epoch": 0.07,
"grad_norm": 22.431930541992188,
"kl": 0.0,
"learning_rate": 3.343949044585987e-07,
"logps/chosen": -259.4337463378906,
"logps/rejected": -309.25164794921875,
"loss": 0.1767,
"rewards/chosen": 1.4657680988311768,
"rewards/margins": 5.859877109527588,
"rewards/rejected": -4.39410924911499,
"step": 105
},
{
"epoch": 0.07,
"grad_norm": 36.517799377441406,
"kl": 0.0,
"learning_rate": 3.5031847133757957e-07,
"logps/chosen": -264.08306884765625,
"logps/rejected": -295.64788818359375,
"loss": 0.1599,
"rewards/chosen": 1.291338562965393,
"rewards/margins": 5.8188042640686035,
"rewards/rejected": -4.5274658203125,
"step": 110
},
{
"epoch": 0.07,
"grad_norm": 25.18623924255371,
"kl": 0.0,
"learning_rate": 3.6624203821656045e-07,
"logps/chosen": -268.96527099609375,
"logps/rejected": -278.4827575683594,
"loss": 0.1735,
"rewards/chosen": 1.4270073175430298,
"rewards/margins": 5.516595363616943,
"rewards/rejected": -4.0895891189575195,
"step": 115
},
{
"epoch": 0.08,
"grad_norm": 24.899450302124023,
"kl": 0.0,
"learning_rate": 3.8216560509554143e-07,
"logps/chosen": -257.46807861328125,
"logps/rejected": -306.6875305175781,
"loss": 0.1675,
"rewards/chosen": 1.363029956817627,
"rewards/margins": 5.9437384605407715,
"rewards/rejected": -4.580708980560303,
"step": 120
},
{
"epoch": 0.08,
"grad_norm": 18.1636962890625,
"kl": 0.0,
"learning_rate": 3.980891719745223e-07,
"logps/chosen": -274.8599548339844,
"logps/rejected": -316.78057861328125,
"loss": 0.1509,
"rewards/chosen": 1.4044160842895508,
"rewards/margins": 6.668715000152588,
"rewards/rejected": -5.264299392700195,
"step": 125
},
{
"epoch": 0.08,
"grad_norm": 53.83689498901367,
"kl": 0.0,
"learning_rate": 4.140127388535032e-07,
"logps/chosen": -316.65460205078125,
"logps/rejected": -318.2517395019531,
"loss": 0.1667,
"rewards/chosen": 1.331627368927002,
"rewards/margins": 6.6092848777771,
"rewards/rejected": -5.277657508850098,
"step": 130
},
{
"epoch": 0.09,
"grad_norm": 16.44365692138672,
"kl": 0.0,
"learning_rate": 4.2993630573248406e-07,
"logps/chosen": -242.02651977539062,
"logps/rejected": -283.90106201171875,
"loss": 0.1762,
"rewards/chosen": 1.6127967834472656,
"rewards/margins": 6.091832160949707,
"rewards/rejected": -4.479035377502441,
"step": 135
},
{
"epoch": 0.09,
"grad_norm": 19.764270782470703,
"kl": 0.0,
"learning_rate": 4.4585987261146494e-07,
"logps/chosen": -252.82180786132812,
"logps/rejected": -319.73199462890625,
"loss": 0.1468,
"rewards/chosen": 1.5330384969711304,
"rewards/margins": 6.702307224273682,
"rewards/rejected": -5.169268608093262,
"step": 140
},
{
"epoch": 0.09,
"grad_norm": 21.560876846313477,
"kl": 0.0,
"learning_rate": 4.6178343949044587e-07,
"logps/chosen": -247.28427124023438,
"logps/rejected": -282.38848876953125,
"loss": 0.1868,
"rewards/chosen": 1.350294828414917,
"rewards/margins": 5.6418867111206055,
"rewards/rejected": -4.291592597961426,
"step": 145
},
{
"epoch": 0.1,
"grad_norm": 21.553871154785156,
"kl": 0.0,
"learning_rate": 4.777070063694267e-07,
"logps/chosen": -286.14312744140625,
"logps/rejected": -281.7044372558594,
"loss": 0.168,
"rewards/chosen": 1.5479779243469238,
"rewards/margins": 6.192603588104248,
"rewards/rejected": -4.644625663757324,
"step": 150
},
{
"epoch": 0.1,
"grad_norm": 24.488561630249023,
"kl": 0.0,
"learning_rate": 4.936305732484076e-07,
"logps/chosen": -272.4622497558594,
"logps/rejected": -323.91082763671875,
"loss": 0.1648,
"rewards/chosen": 1.3805065155029297,
"rewards/margins": 6.456129550933838,
"rewards/rejected": -5.07562255859375,
"step": 155
},
{
"epoch": 0.1,
"grad_norm": 21.614229202270508,
"kl": 0.0,
"learning_rate": 4.989331436699858e-07,
"logps/chosen": -207.8739776611328,
"logps/rejected": -324.73565673828125,
"loss": 0.1435,
"rewards/chosen": 1.465820074081421,
"rewards/margins": 6.953179836273193,
"rewards/rejected": -5.487359046936035,
"step": 160
},
{
"epoch": 0.11,
"grad_norm": 20.425729751586914,
"kl": 0.0,
"learning_rate": 4.971550497866287e-07,
"logps/chosen": -287.088623046875,
"logps/rejected": -310.1844787597656,
"loss": 0.1568,
"rewards/chosen": 1.5799640417099,
"rewards/margins": 6.89548397064209,
"rewards/rejected": -5.315520286560059,
"step": 165
},
{
"epoch": 0.11,
"grad_norm": 21.86062240600586,
"kl": 0.0,
"learning_rate": 4.953769559032717e-07,
"logps/chosen": -282.91741943359375,
"logps/rejected": -315.76470947265625,
"loss": 0.1516,
"rewards/chosen": 1.442077875137329,
"rewards/margins": 7.074441432952881,
"rewards/rejected": -5.632363796234131,
"step": 170
},
{
"epoch": 0.11,
"grad_norm": 20.31084632873535,
"kl": 0.0,
"learning_rate": 4.935988620199146e-07,
"logps/chosen": -274.0014953613281,
"logps/rejected": -310.8323059082031,
"loss": 0.1477,
"rewards/chosen": 1.4551314115524292,
"rewards/margins": 6.99865198135376,
"rewards/rejected": -5.543520450592041,
"step": 175
},
{
"epoch": 0.12,
"grad_norm": 20.067346572875977,
"kl": 0.0,
"learning_rate": 4.918207681365576e-07,
"logps/chosen": -274.91705322265625,
"logps/rejected": -330.29315185546875,
"loss": 0.1491,
"rewards/chosen": 1.5545661449432373,
"rewards/margins": 7.674098014831543,
"rewards/rejected": -6.119531631469727,
"step": 180
},
{
"epoch": 0.12,
"grad_norm": 26.703292846679688,
"kl": 0.0,
"learning_rate": 4.900426742532006e-07,
"logps/chosen": -267.2247619628906,
"logps/rejected": -308.6265563964844,
"loss": 0.142,
"rewards/chosen": 1.4495770931243896,
"rewards/margins": 7.4614386558532715,
"rewards/rejected": -6.011861324310303,
"step": 185
},
{
"epoch": 0.12,
"grad_norm": 22.81442642211914,
"kl": 0.0,
"learning_rate": 4.882645803698435e-07,
"logps/chosen": -243.8881072998047,
"logps/rejected": -318.4960632324219,
"loss": 0.1679,
"rewards/chosen": 1.2335823774337769,
"rewards/margins": 7.564291954040527,
"rewards/rejected": -6.330709934234619,
"step": 190
},
{
"epoch": 0.12,
"grad_norm": 17.875883102416992,
"kl": 0.0,
"learning_rate": 4.864864864864865e-07,
"logps/chosen": -276.0986328125,
"logps/rejected": -317.36163330078125,
"loss": 0.1442,
"rewards/chosen": 1.5666391849517822,
"rewards/margins": 7.757128715515137,
"rewards/rejected": -6.190489768981934,
"step": 195
},
{
"epoch": 0.13,
"grad_norm": 22.31620216369629,
"kl": 0.0,
"learning_rate": 4.847083926031294e-07,
"logps/chosen": -273.843994140625,
"logps/rejected": -329.43475341796875,
"loss": 0.154,
"rewards/chosen": 1.6012461185455322,
"rewards/margins": 7.519808292388916,
"rewards/rejected": -5.918562412261963,
"step": 200
},
{
"epoch": 0.13,
"grad_norm": 19.976451873779297,
"kl": 0.0,
"learning_rate": 4.829302987197724e-07,
"logps/chosen": -284.374755859375,
"logps/rejected": -326.24371337890625,
"loss": 0.1378,
"rewards/chosen": 1.5799314975738525,
"rewards/margins": 7.78595495223999,
"rewards/rejected": -6.206023216247559,
"step": 205
},
{
"epoch": 0.13,
"grad_norm": 23.919170379638672,
"kl": 0.0,
"learning_rate": 4.811522048364154e-07,
"logps/chosen": -236.8705291748047,
"logps/rejected": -316.30413818359375,
"loss": 0.1523,
"rewards/chosen": 1.5169861316680908,
"rewards/margins": 7.513286590576172,
"rewards/rejected": -5.99630069732666,
"step": 210
},
{
"epoch": 0.14,
"grad_norm": 25.041824340820312,
"kl": 0.0,
"learning_rate": 4.793741109530583e-07,
"logps/chosen": -219.86117553710938,
"logps/rejected": -331.60418701171875,
"loss": 0.1516,
"rewards/chosen": 1.5909864902496338,
"rewards/margins": 7.9639410972595215,
"rewards/rejected": -6.372954845428467,
"step": 215
},
{
"epoch": 0.14,
"grad_norm": 25.45568084716797,
"kl": 0.0,
"learning_rate": 4.775960170697012e-07,
"logps/chosen": -271.017333984375,
"logps/rejected": -333.2148132324219,
"loss": 0.1628,
"rewards/chosen": 1.5948388576507568,
"rewards/margins": 7.8658246994018555,
"rewards/rejected": -6.2709856033325195,
"step": 220
},
{
"epoch": 0.14,
"grad_norm": 22.259382247924805,
"kl": 0.0,
"learning_rate": 4.7581792318634425e-07,
"logps/chosen": -253.55789184570312,
"logps/rejected": -298.8450622558594,
"loss": 0.1452,
"rewards/chosen": 1.640355110168457,
"rewards/margins": 8.043792724609375,
"rewards/rejected": -6.403438568115234,
"step": 225
},
{
"epoch": 0.15,
"grad_norm": 26.135997772216797,
"kl": 0.0,
"learning_rate": 4.7403982930298717e-07,
"logps/chosen": -235.11349487304688,
"logps/rejected": -344.72369384765625,
"loss": 0.1472,
"rewards/chosen": 1.559660792350769,
"rewards/margins": 7.8217010498046875,
"rewards/rejected": -6.262040615081787,
"step": 230
},
{
"epoch": 0.15,
"grad_norm": 20.143089294433594,
"kl": 0.0,
"learning_rate": 4.7226173541963014e-07,
"logps/chosen": -249.42196655273438,
"logps/rejected": -277.12554931640625,
"loss": 0.1529,
"rewards/chosen": 1.6460959911346436,
"rewards/margins": 7.348568916320801,
"rewards/rejected": -5.702473163604736,
"step": 235
},
{
"epoch": 0.15,
"grad_norm": 23.44059181213379,
"kl": 0.0,
"learning_rate": 4.7048364153627306e-07,
"logps/chosen": -281.3484802246094,
"logps/rejected": -307.25457763671875,
"loss": 0.1387,
"rewards/chosen": 1.7489140033721924,
"rewards/margins": 7.860198974609375,
"rewards/rejected": -6.111284255981445,
"step": 240
},
{
"epoch": 0.16,
"grad_norm": 23.201915740966797,
"kl": 0.0,
"learning_rate": 4.6870554765291604e-07,
"logps/chosen": -272.64031982421875,
"logps/rejected": -310.50909423828125,
"loss": 0.1356,
"rewards/chosen": 1.6725397109985352,
"rewards/margins": 8.012718200683594,
"rewards/rejected": -6.340178489685059,
"step": 245
},
{
"epoch": 0.16,
"grad_norm": 19.45538902282715,
"kl": 0.0,
"learning_rate": 4.66927453769559e-07,
"logps/chosen": -272.3523254394531,
"logps/rejected": -338.20867919921875,
"loss": 0.1246,
"rewards/chosen": 1.727900743484497,
"rewards/margins": 8.646397590637207,
"rewards/rejected": -6.918497562408447,
"step": 250
},
{
"epoch": 0.16,
"grad_norm": 24.896251678466797,
"kl": 0.0,
"learning_rate": 4.65149359886202e-07,
"logps/chosen": -269.20794677734375,
"logps/rejected": -328.73486328125,
"loss": 0.1281,
"rewards/chosen": 1.6262487173080444,
"rewards/margins": 9.223333358764648,
"rewards/rejected": -7.597084999084473,
"step": 255
},
{
"epoch": 0.17,
"grad_norm": 20.800025939941406,
"kl": 0.0,
"learning_rate": 4.633712660028449e-07,
"logps/chosen": -270.3649597167969,
"logps/rejected": -330.872314453125,
"loss": 0.1279,
"rewards/chosen": 1.5803369283676147,
"rewards/margins": 9.708114624023438,
"rewards/rejected": -8.127778053283691,
"step": 260
},
{
"epoch": 0.17,
"grad_norm": 20.315540313720703,
"kl": 0.0,
"learning_rate": 4.615931721194879e-07,
"logps/chosen": -277.21807861328125,
"logps/rejected": -313.4653625488281,
"loss": 0.1461,
"rewards/chosen": 1.6411361694335938,
"rewards/margins": 7.905195713043213,
"rewards/rejected": -6.264059543609619,
"step": 265
},
{
"epoch": 0.17,
"grad_norm": 19.68859100341797,
"kl": 0.0,
"learning_rate": 4.5981507823613085e-07,
"logps/chosen": -275.69696044921875,
"logps/rejected": -295.79400634765625,
"loss": 0.1547,
"rewards/chosen": 1.6596572399139404,
"rewards/margins": 7.9982404708862305,
"rewards/rejected": -6.338583469390869,
"step": 270
},
{
"epoch": 0.18,
"grad_norm": 27.070371627807617,
"kl": 0.0,
"learning_rate": 4.580369843527738e-07,
"logps/chosen": -272.72735595703125,
"logps/rejected": -310.302734375,
"loss": 0.1673,
"rewards/chosen": 1.2246049642562866,
"rewards/margins": 8.44649600982666,
"rewards/rejected": -7.221890449523926,
"step": 275
},
{
"epoch": 0.18,
"grad_norm": 15.166617393493652,
"kl": 0.0,
"learning_rate": 4.562588904694168e-07,
"logps/chosen": -267.9317932128906,
"logps/rejected": -316.60479736328125,
"loss": 0.1304,
"rewards/chosen": 1.7944562435150146,
"rewards/margins": 8.752721786499023,
"rewards/rejected": -6.958265781402588,
"step": 280
},
{
"epoch": 0.18,
"grad_norm": 27.143291473388672,
"kl": 0.0,
"learning_rate": 4.544807965860597e-07,
"logps/chosen": -257.7388916015625,
"logps/rejected": -308.5174560546875,
"loss": 0.1421,
"rewards/chosen": 1.6366589069366455,
"rewards/margins": 8.678224563598633,
"rewards/rejected": -7.04156494140625,
"step": 285
},
{
"epoch": 0.19,
"grad_norm": 22.1202392578125,
"kl": 0.0,
"learning_rate": 4.5270270270270264e-07,
"logps/chosen": -266.8901062011719,
"logps/rejected": -322.6683654785156,
"loss": 0.1447,
"rewards/chosen": 1.787255883216858,
"rewards/margins": 9.206243515014648,
"rewards/rejected": -7.418987274169922,
"step": 290
},
{
"epoch": 0.19,
"grad_norm": 24.7782039642334,
"kl": 0.0,
"learning_rate": 4.509246088193456e-07,
"logps/chosen": -274.30438232421875,
"logps/rejected": -342.98541259765625,
"loss": 0.1421,
"rewards/chosen": 1.776602029800415,
"rewards/margins": 9.053030967712402,
"rewards/rejected": -7.276429176330566,
"step": 295
},
{
"epoch": 0.19,
"grad_norm": 23.54313087463379,
"kl": 0.0,
"learning_rate": 4.491465149359886e-07,
"logps/chosen": -215.9270477294922,
"logps/rejected": -303.5273742675781,
"loss": 0.1438,
"rewards/chosen": 1.868506669998169,
"rewards/margins": 9.129117965698242,
"rewards/rejected": -7.260611534118652,
"step": 300
},
{
"epoch": 0.2,
"grad_norm": 17.70969009399414,
"kl": 0.0,
"learning_rate": 4.4736842105263156e-07,
"logps/chosen": -240.4123992919922,
"logps/rejected": -336.1444396972656,
"loss": 0.1403,
"rewards/chosen": 1.7039588689804077,
"rewards/margins": 9.00100326538086,
"rewards/rejected": -7.297044277191162,
"step": 305
},
{
"epoch": 0.2,
"grad_norm": 19.908315658569336,
"kl": 0.0,
"learning_rate": 4.4559032716927454e-07,
"logps/chosen": -212.1704864501953,
"logps/rejected": -340.3092041015625,
"loss": 0.1338,
"rewards/chosen": 1.6462271213531494,
"rewards/margins": 8.726274490356445,
"rewards/rejected": -7.0800461769104,
"step": 310
},
{
"epoch": 0.2,
"grad_norm": 22.928499221801758,
"kl": 0.0,
"learning_rate": 4.438122332859175e-07,
"logps/chosen": -243.7815704345703,
"logps/rejected": -324.60772705078125,
"loss": 0.1323,
"rewards/chosen": 1.6998401880264282,
"rewards/margins": 8.809240341186523,
"rewards/rejected": -7.109400749206543,
"step": 315
},
{
"epoch": 0.2,
"grad_norm": 20.08190155029297,
"kl": 0.0,
"learning_rate": 4.420341394025605e-07,
"logps/chosen": -263.37725830078125,
"logps/rejected": -317.4496154785156,
"loss": 0.1443,
"rewards/chosen": 1.706053376197815,
"rewards/margins": 8.708757400512695,
"rewards/rejected": -7.002703666687012,
"step": 320
},
{
"epoch": 0.21,
"grad_norm": 25.728233337402344,
"kl": 0.0,
"learning_rate": 4.4025604551920335e-07,
"logps/chosen": -282.44989013671875,
"logps/rejected": -325.9697265625,
"loss": 0.1351,
"rewards/chosen": 1.6976518630981445,
"rewards/margins": 9.016552925109863,
"rewards/rejected": -7.318901062011719,
"step": 325
},
{
"epoch": 0.21,
"grad_norm": 22.311450958251953,
"kl": 0.0,
"learning_rate": 4.384779516358463e-07,
"logps/chosen": -260.39447021484375,
"logps/rejected": -309.620849609375,
"loss": 0.1311,
"rewards/chosen": 1.8077850341796875,
"rewards/margins": 9.406209945678711,
"rewards/rejected": -7.598425388336182,
"step": 330
},
{
"epoch": 0.21,
"grad_norm": 23.20633316040039,
"kl": 0.0,
"learning_rate": 4.366998577524893e-07,
"logps/chosen": -302.47509765625,
"logps/rejected": -318.767578125,
"loss": 0.128,
"rewards/chosen": 1.7027689218521118,
"rewards/margins": 10.0870943069458,
"rewards/rejected": -8.38432502746582,
"step": 335
},
{
"epoch": 0.22,
"grad_norm": 25.405933380126953,
"kl": 0.0,
"learning_rate": 4.3492176386913227e-07,
"logps/chosen": -238.94827270507812,
"logps/rejected": -326.12786865234375,
"loss": 0.1505,
"rewards/chosen": 1.7603483200073242,
"rewards/margins": 9.71554183959961,
"rewards/rejected": -7.955193519592285,
"step": 340
},
{
"epoch": 0.22,
"grad_norm": 22.41493034362793,
"kl": 0.0,
"learning_rate": 4.3314366998577524e-07,
"logps/chosen": -253.826416015625,
"logps/rejected": -330.7270812988281,
"loss": 0.1413,
"rewards/chosen": 1.8580402135849,
"rewards/margins": 9.722744941711426,
"rewards/rejected": -7.8647050857543945,
"step": 345
},
{
"epoch": 0.22,
"grad_norm": 26.315628051757812,
"kl": 0.0,
"learning_rate": 4.313655761024182e-07,
"logps/chosen": -286.4155578613281,
"logps/rejected": -342.62652587890625,
"loss": 0.1422,
"rewards/chosen": 1.7754218578338623,
"rewards/margins": 10.7272367477417,
"rewards/rejected": -8.951814651489258,
"step": 350
},
{
"epoch": 0.23,
"grad_norm": 22.07290267944336,
"kl": 0.0,
"learning_rate": 4.2958748221906114e-07,
"logps/chosen": -239.4450225830078,
"logps/rejected": -310.7589111328125,
"loss": 0.1108,
"rewards/chosen": 1.9980005025863647,
"rewards/margins": 10.0792818069458,
"rewards/rejected": -8.081281661987305,
"step": 355
},
{
"epoch": 0.23,
"grad_norm": 22.56456756591797,
"kl": 0.0,
"learning_rate": 4.278093883357041e-07,
"logps/chosen": -269.43560791015625,
"logps/rejected": -328.5326843261719,
"loss": 0.1524,
"rewards/chosen": 1.747078537940979,
"rewards/margins": 9.907144546508789,
"rewards/rejected": -8.160065650939941,
"step": 360
},
{
"epoch": 0.23,
"grad_norm": 24.159225463867188,
"kl": 0.0,
"learning_rate": 4.260312944523471e-07,
"logps/chosen": -241.72286987304688,
"logps/rejected": -353.99493408203125,
"loss": 0.1185,
"rewards/chosen": 2.0929551124572754,
"rewards/margins": 10.861230850219727,
"rewards/rejected": -8.768275260925293,
"step": 365
},
{
"epoch": 0.24,
"grad_norm": 21.119632720947266,
"kl": 0.0,
"learning_rate": 4.2425320056899e-07,
"logps/chosen": -263.73370361328125,
"logps/rejected": -343.69866943359375,
"loss": 0.1164,
"rewards/chosen": 1.9714298248291016,
"rewards/margins": 10.647806167602539,
"rewards/rejected": -8.676377296447754,
"step": 370
},
{
"epoch": 0.24,
"grad_norm": 23.59329605102539,
"kl": 0.0,
"learning_rate": 4.22475106685633e-07,
"logps/chosen": -244.0771026611328,
"logps/rejected": -348.35693359375,
"loss": 0.1218,
"rewards/chosen": 1.8124616146087646,
"rewards/margins": 10.569868087768555,
"rewards/rejected": -8.757406234741211,
"step": 375
},
{
"epoch": 0.24,
"grad_norm": 23.84940528869629,
"kl": 0.0,
"learning_rate": 4.2069701280227595e-07,
"logps/chosen": -241.4117889404297,
"logps/rejected": -336.4566345214844,
"loss": 0.1294,
"rewards/chosen": 1.8767503499984741,
"rewards/margins": 10.489054679870605,
"rewards/rejected": -8.612303733825684,
"step": 380
},
{
"epoch": 0.25,
"grad_norm": 16.005657196044922,
"kl": 0.0,
"learning_rate": 4.189189189189189e-07,
"logps/chosen": -307.21600341796875,
"logps/rejected": -341.0569152832031,
"loss": 0.1154,
"rewards/chosen": 1.901063323020935,
"rewards/margins": 10.400434494018555,
"rewards/rejected": -8.499369621276855,
"step": 385
},
{
"epoch": 0.25,
"grad_norm": 16.274534225463867,
"kl": 0.0,
"learning_rate": 4.1714082503556185e-07,
"logps/chosen": -253.8114471435547,
"logps/rejected": -333.0570983886719,
"loss": 0.1327,
"rewards/chosen": 1.955413818359375,
"rewards/margins": 10.895428657531738,
"rewards/rejected": -8.940014839172363,
"step": 390
},
{
"epoch": 0.25,
"grad_norm": 20.72968101501465,
"kl": 0.0,
"learning_rate": 4.153627311522048e-07,
"logps/chosen": -261.703857421875,
"logps/rejected": -339.4162902832031,
"loss": 0.1308,
"rewards/chosen": 1.7927815914154053,
"rewards/margins": 10.807384490966797,
"rewards/rejected": -9.014602661132812,
"step": 395
},
{
"epoch": 0.26,
"grad_norm": 21.173845291137695,
"kl": 0.0,
"learning_rate": 4.135846372688478e-07,
"logps/chosen": -260.5010681152344,
"logps/rejected": -359.41851806640625,
"loss": 0.1131,
"rewards/chosen": 1.8654544353485107,
"rewards/margins": 11.833639144897461,
"rewards/rejected": -9.968184471130371,
"step": 400
},
{
"epoch": 0.26,
"grad_norm": 17.008028030395508,
"kl": 0.0,
"learning_rate": 4.1180654338549077e-07,
"logps/chosen": -234.20315551757812,
"logps/rejected": -342.5408020019531,
"loss": 0.1224,
"rewards/chosen": 2.086550235748291,
"rewards/margins": 10.627596855163574,
"rewards/rejected": -8.541048049926758,
"step": 405
},
{
"epoch": 0.26,
"grad_norm": 19.24855613708496,
"kl": 0.0,
"learning_rate": 4.100284495021337e-07,
"logps/chosen": -282.03106689453125,
"logps/rejected": -347.0184631347656,
"loss": 0.1212,
"rewards/chosen": 1.9099693298339844,
"rewards/margins": 10.274763107299805,
"rewards/rejected": -8.36479377746582,
"step": 410
},
{
"epoch": 0.27,
"grad_norm": 19.914691925048828,
"kl": 0.0,
"learning_rate": 4.082503556187766e-07,
"logps/chosen": -278.811767578125,
"logps/rejected": -348.43096923828125,
"loss": 0.1209,
"rewards/chosen": 1.7562357187271118,
"rewards/margins": 11.304061889648438,
"rewards/rejected": -9.547826766967773,
"step": 415
},
{
"epoch": 0.27,
"grad_norm": 32.35190963745117,
"kl": 0.0,
"learning_rate": 4.064722617354196e-07,
"logps/chosen": -272.9192199707031,
"logps/rejected": -339.4073181152344,
"loss": 0.1005,
"rewards/chosen": 2.1069161891937256,
"rewards/margins": 11.689797401428223,
"rewards/rejected": -9.582880973815918,
"step": 420
},
{
"epoch": 0.27,
"grad_norm": 16.27385902404785,
"kl": 0.0,
"learning_rate": 4.0469416785206256e-07,
"logps/chosen": -233.26083374023438,
"logps/rejected": -329.8385314941406,
"loss": 0.1135,
"rewards/chosen": 2.1503446102142334,
"rewards/margins": 10.886509895324707,
"rewards/rejected": -8.736165046691895,
"step": 425
},
{
"epoch": 0.28,
"grad_norm": 22.10724449157715,
"kl": 0.0,
"learning_rate": 4.0291607396870553e-07,
"logps/chosen": -267.9533996582031,
"logps/rejected": -359.1152038574219,
"loss": 0.1197,
"rewards/chosen": 1.8767788410186768,
"rewards/margins": 11.441519737243652,
"rewards/rejected": -9.564741134643555,
"step": 430
},
{
"epoch": 0.28,
"grad_norm": 29.721330642700195,
"kl": 0.0,
"learning_rate": 4.011379800853485e-07,
"logps/chosen": -268.9942932128906,
"logps/rejected": -377.97308349609375,
"loss": 0.1061,
"rewards/chosen": 2.016845703125,
"rewards/margins": 11.474628448486328,
"rewards/rejected": -9.457781791687012,
"step": 435
},
{
"epoch": 0.28,
"grad_norm": 24.95069122314453,
"kl": 0.0,
"learning_rate": 3.993598862019915e-07,
"logps/chosen": -279.73944091796875,
"logps/rejected": -370.55474853515625,
"loss": 0.1045,
"rewards/chosen": 2.340444803237915,
"rewards/margins": 11.733491897583008,
"rewards/rejected": -9.393046379089355,
"step": 440
},
{
"epoch": 0.28,
"grad_norm": 26.99854278564453,
"kl": 0.0,
"learning_rate": 3.975817923186344e-07,
"logps/chosen": -245.24417114257812,
"logps/rejected": -336.7696533203125,
"loss": 0.1234,
"rewards/chosen": 2.2276217937469482,
"rewards/margins": 10.63569164276123,
"rewards/rejected": -8.40807056427002,
"step": 445
},
{
"epoch": 0.29,
"grad_norm": 22.343507766723633,
"kl": 0.0,
"learning_rate": 3.9580369843527737e-07,
"logps/chosen": -250.68197631835938,
"logps/rejected": -364.1970520019531,
"loss": 0.1342,
"rewards/chosen": 2.1794769763946533,
"rewards/margins": 10.827537536621094,
"rewards/rejected": -8.64806079864502,
"step": 450
},
{
"epoch": 0.29,
"grad_norm": 16.532583236694336,
"kl": 0.0,
"learning_rate": 3.940256045519203e-07,
"logps/chosen": -246.8504638671875,
"logps/rejected": -336.2493896484375,
"loss": 0.1151,
"rewards/chosen": 1.9471546411514282,
"rewards/margins": 11.059728622436523,
"rewards/rejected": -9.112573623657227,
"step": 455
},
{
"epoch": 0.29,
"grad_norm": 19.48154067993164,
"kl": 0.0,
"learning_rate": 3.9224751066856327e-07,
"logps/chosen": -246.47671508789062,
"logps/rejected": -328.21038818359375,
"loss": 0.1156,
"rewards/chosen": 2.022249221801758,
"rewards/margins": 11.125974655151367,
"rewards/rejected": -9.103724479675293,
"step": 460
},
{
"epoch": 0.3,
"grad_norm": 17.954296112060547,
"kl": 0.0,
"learning_rate": 3.9046941678520624e-07,
"logps/chosen": -254.69808959960938,
"logps/rejected": -347.45758056640625,
"loss": 0.1254,
"rewards/chosen": 2.0911641120910645,
"rewards/margins": 11.680787086486816,
"rewards/rejected": -9.58962345123291,
"step": 465
},
{
"epoch": 0.3,
"grad_norm": 21.16987419128418,
"kl": 0.0,
"learning_rate": 3.886913229018492e-07,
"logps/chosen": -226.58358764648438,
"logps/rejected": -334.48138427734375,
"loss": 0.1437,
"rewards/chosen": 1.8157398700714111,
"rewards/margins": 10.399962425231934,
"rewards/rejected": -8.584221839904785,
"step": 470
},
{
"epoch": 0.3,
"grad_norm": 18.404508590698242,
"kl": 0.0,
"learning_rate": 3.8691322901849213e-07,
"logps/chosen": -263.36395263671875,
"logps/rejected": -337.2552795410156,
"loss": 0.1078,
"rewards/chosen": 2.1132290363311768,
"rewards/margins": 11.090356826782227,
"rewards/rejected": -8.977127075195312,
"step": 475
},
{
"epoch": 0.31,
"grad_norm": 18.355857849121094,
"kl": 0.0,
"learning_rate": 3.851351351351351e-07,
"logps/chosen": -256.04412841796875,
"logps/rejected": -356.82073974609375,
"loss": 0.1145,
"rewards/chosen": 1.9654195308685303,
"rewards/margins": 11.159137725830078,
"rewards/rejected": -9.193717956542969,
"step": 480
},
{
"epoch": 0.31,
"grad_norm": 16.297401428222656,
"kl": 0.0,
"learning_rate": 3.833570412517781e-07,
"logps/chosen": -330.457275390625,
"logps/rejected": -367.7776794433594,
"loss": 0.113,
"rewards/chosen": 1.8794314861297607,
"rewards/margins": 11.851155281066895,
"rewards/rejected": -9.971722602844238,
"step": 485
},
{
"epoch": 0.31,
"grad_norm": 22.856266021728516,
"kl": 0.0,
"learning_rate": 3.8157894736842105e-07,
"logps/chosen": -254.3659210205078,
"logps/rejected": -343.2192687988281,
"loss": 0.1138,
"rewards/chosen": 2.034006118774414,
"rewards/margins": 11.903759002685547,
"rewards/rejected": -9.869752883911133,
"step": 490
},
{
"epoch": 0.32,
"grad_norm": 21.443723678588867,
"kl": 0.0,
"learning_rate": 3.7980085348506403e-07,
"logps/chosen": -278.9993591308594,
"logps/rejected": -333.3291320800781,
"loss": 0.1287,
"rewards/chosen": 1.906224012374878,
"rewards/margins": 10.766765594482422,
"rewards/rejected": -8.860541343688965,
"step": 495
},
{
"epoch": 0.32,
"grad_norm": 24.97509002685547,
"kl": 0.0,
"learning_rate": 3.7802275960170695e-07,
"logps/chosen": -238.44290161132812,
"logps/rejected": -338.4499206542969,
"loss": 0.1142,
"rewards/chosen": 2.103523015975952,
"rewards/margins": 11.162189483642578,
"rewards/rejected": -9.058666229248047,
"step": 500
},
{
"epoch": 0.32,
"grad_norm": 22.563114166259766,
"kl": 0.0,
"learning_rate": 3.7624466571834987e-07,
"logps/chosen": -238.7085723876953,
"logps/rejected": -352.23663330078125,
"loss": 0.1396,
"rewards/chosen": 1.8689464330673218,
"rewards/margins": 10.945247650146484,
"rewards/rejected": -9.076301574707031,
"step": 505
},
{
"epoch": 0.33,
"grad_norm": 17.017030715942383,
"kl": 0.0,
"learning_rate": 3.7446657183499284e-07,
"logps/chosen": -241.0718231201172,
"logps/rejected": -340.98968505859375,
"loss": 0.1197,
"rewards/chosen": 1.9388542175292969,
"rewards/margins": 11.241914749145508,
"rewards/rejected": -9.303060531616211,
"step": 510
},
{
"epoch": 0.33,
"grad_norm": 16.066030502319336,
"kl": 0.0,
"learning_rate": 3.726884779516358e-07,
"logps/chosen": -224.85086059570312,
"logps/rejected": -349.51800537109375,
"loss": 0.1094,
"rewards/chosen": 2.1027920246124268,
"rewards/margins": 11.039840698242188,
"rewards/rejected": -8.937047958374023,
"step": 515
},
{
"epoch": 0.33,
"grad_norm": 17.983535766601562,
"kl": 0.0,
"learning_rate": 3.709103840682788e-07,
"logps/chosen": -269.41693115234375,
"logps/rejected": -347.45477294921875,
"loss": 0.1234,
"rewards/chosen": 2.2909629344940186,
"rewards/margins": 11.061802864074707,
"rewards/rejected": -8.770838737487793,
"step": 520
},
{
"epoch": 0.34,
"grad_norm": 32.42300796508789,
"kl": 0.0,
"learning_rate": 3.6913229018492176e-07,
"logps/chosen": -274.3470764160156,
"logps/rejected": -335.10345458984375,
"loss": 0.1186,
"rewards/chosen": 2.1239845752716064,
"rewards/margins": 11.3440523147583,
"rewards/rejected": -9.220067977905273,
"step": 525
},
{
"epoch": 0.34,
"grad_norm": 16.601402282714844,
"kl": 0.0,
"learning_rate": 3.6735419630156474e-07,
"logps/chosen": -232.43392944335938,
"logps/rejected": -354.2206726074219,
"loss": 0.1218,
"rewards/chosen": 2.117833137512207,
"rewards/margins": 11.192573547363281,
"rewards/rejected": -9.07474136352539,
"step": 530
},
{
"epoch": 0.34,
"grad_norm": 24.04266357421875,
"kl": 0.0,
"learning_rate": 3.655761024182077e-07,
"logps/chosen": -277.41241455078125,
"logps/rejected": -334.04815673828125,
"loss": 0.1025,
"rewards/chosen": 2.3604273796081543,
"rewards/margins": 11.987722396850586,
"rewards/rejected": -9.627294540405273,
"step": 535
},
{
"epoch": 0.35,
"grad_norm": 15.607158660888672,
"kl": 0.0,
"learning_rate": 3.637980085348506e-07,
"logps/chosen": -286.7027282714844,
"logps/rejected": -319.68896484375,
"loss": 0.1124,
"rewards/chosen": 2.4556756019592285,
"rewards/margins": 11.582249641418457,
"rewards/rejected": -9.126574516296387,
"step": 540
},
{
"epoch": 0.35,
"grad_norm": 22.26024055480957,
"kl": 0.0,
"learning_rate": 3.6201991465149355e-07,
"logps/chosen": -239.4503173828125,
"logps/rejected": -345.2344970703125,
"loss": 0.112,
"rewards/chosen": 1.9915698766708374,
"rewards/margins": 11.853796005249023,
"rewards/rejected": -9.862226486206055,
"step": 545
},
{
"epoch": 0.35,
"grad_norm": 29.981088638305664,
"kl": 0.0,
"learning_rate": 3.602418207681365e-07,
"logps/chosen": -241.8583984375,
"logps/rejected": -365.6163024902344,
"loss": 0.1207,
"rewards/chosen": 1.9029747247695923,
"rewards/margins": 11.538789749145508,
"rewards/rejected": -9.635814666748047,
"step": 550
},
{
"epoch": 0.36,
"grad_norm": 17.248044967651367,
"kl": 0.0,
"learning_rate": 3.584637268847795e-07,
"logps/chosen": -255.35635375976562,
"logps/rejected": -348.6636962890625,
"loss": 0.1267,
"rewards/chosen": 2.0199761390686035,
"rewards/margins": 11.458740234375,
"rewards/rejected": -9.438763618469238,
"step": 555
},
{
"epoch": 0.36,
"grad_norm": 20.46364974975586,
"kl": 0.0,
"learning_rate": 3.5668563300142247e-07,
"logps/chosen": -271.98846435546875,
"logps/rejected": -319.4902038574219,
"loss": 0.1209,
"rewards/chosen": 2.140854597091675,
"rewards/margins": 11.506881713867188,
"rewards/rejected": -9.36602783203125,
"step": 560
},
{
"epoch": 0.36,
"grad_norm": 13.751622200012207,
"kl": 0.0,
"learning_rate": 3.5490753911806545e-07,
"logps/chosen": -249.31838989257812,
"logps/rejected": -363.65313720703125,
"loss": 0.1066,
"rewards/chosen": 2.302562713623047,
"rewards/margins": 12.324169158935547,
"rewards/rejected": -10.0216064453125,
"step": 565
},
{
"epoch": 0.36,
"grad_norm": 23.35749053955078,
"kl": 0.0,
"learning_rate": 3.5312944523470837e-07,
"logps/chosen": -242.62533569335938,
"logps/rejected": -326.2989807128906,
"loss": 0.1213,
"rewards/chosen": 2.164386749267578,
"rewards/margins": 11.125585556030273,
"rewards/rejected": -8.961198806762695,
"step": 570
},
{
"epoch": 0.37,
"grad_norm": 22.92085075378418,
"kl": 0.0,
"learning_rate": 3.5135135135135134e-07,
"logps/chosen": -215.1080322265625,
"logps/rejected": -343.92315673828125,
"loss": 0.1096,
"rewards/chosen": 2.084261417388916,
"rewards/margins": 12.25261402130127,
"rewards/rejected": -10.168352127075195,
"step": 575
},
{
"epoch": 0.37,
"grad_norm": 19.450267791748047,
"kl": 0.0,
"learning_rate": 3.495732574679943e-07,
"logps/chosen": -217.7822723388672,
"logps/rejected": -329.91375732421875,
"loss": 0.1294,
"rewards/chosen": 2.3254265785217285,
"rewards/margins": 11.483414649963379,
"rewards/rejected": -9.157987594604492,
"step": 580
},
{
"epoch": 0.37,
"grad_norm": 18.078920364379883,
"kl": 0.0,
"learning_rate": 3.4779516358463724e-07,
"logps/chosen": -253.4072723388672,
"logps/rejected": -344.491455078125,
"loss": 0.0927,
"rewards/chosen": 2.3925743103027344,
"rewards/margins": 12.241676330566406,
"rewards/rejected": -9.849101066589355,
"step": 585
},
{
"epoch": 0.38,
"grad_norm": 17.867704391479492,
"kl": 0.0,
"learning_rate": 3.460170697012802e-07,
"logps/chosen": -253.86178588867188,
"logps/rejected": -367.8619689941406,
"loss": 0.1152,
"rewards/chosen": 2.281512975692749,
"rewards/margins": 12.078498840332031,
"rewards/rejected": -9.79698657989502,
"step": 590
},
{
"epoch": 0.38,
"grad_norm": 19.863170623779297,
"kl": 0.0,
"learning_rate": 3.442389758179232e-07,
"logps/chosen": -289.83990478515625,
"logps/rejected": -346.80987548828125,
"loss": 0.1098,
"rewards/chosen": 2.1612110137939453,
"rewards/margins": 12.505953788757324,
"rewards/rejected": -10.344742774963379,
"step": 595
},
{
"epoch": 0.38,
"grad_norm": 14.594950675964355,
"kl": 0.0,
"learning_rate": 3.424608819345661e-07,
"logps/chosen": -199.16708374023438,
"logps/rejected": -352.14447021484375,
"loss": 0.0929,
"rewards/chosen": 1.9940685033798218,
"rewards/margins": 12.916508674621582,
"rewards/rejected": -10.922439575195312,
"step": 600
},
{
"epoch": 0.39,
"grad_norm": 20.205915451049805,
"kl": 0.0,
"learning_rate": 3.406827880512091e-07,
"logps/chosen": -251.85546875,
"logps/rejected": -347.35113525390625,
"loss": 0.1137,
"rewards/chosen": 2.3019704818725586,
"rewards/margins": 12.349814414978027,
"rewards/rejected": -10.047843933105469,
"step": 605
},
{
"epoch": 0.39,
"grad_norm": 21.62274742126465,
"kl": 0.0,
"learning_rate": 3.3890469416785205e-07,
"logps/chosen": -241.4209442138672,
"logps/rejected": -366.311767578125,
"loss": 0.104,
"rewards/chosen": 2.3162879943847656,
"rewards/margins": 11.884176254272461,
"rewards/rejected": -9.567889213562012,
"step": 610
},
{
"epoch": 0.39,
"grad_norm": 18.622377395629883,
"kl": 0.0,
"learning_rate": 3.37126600284495e-07,
"logps/chosen": -270.2699279785156,
"logps/rejected": -338.7982482910156,
"loss": 0.1127,
"rewards/chosen": 2.2165682315826416,
"rewards/margins": 11.826861381530762,
"rewards/rejected": -9.6102933883667,
"step": 615
},
{
"epoch": 0.4,
"grad_norm": 19.930374145507812,
"kl": 0.0,
"learning_rate": 3.35348506401138e-07,
"logps/chosen": -265.74432373046875,
"logps/rejected": -366.4137268066406,
"loss": 0.1235,
"rewards/chosen": 2.140533924102783,
"rewards/margins": 12.077839851379395,
"rewards/rejected": -9.937305450439453,
"step": 620
},
{
"epoch": 0.4,
"grad_norm": 20.78235626220703,
"kl": 0.0,
"learning_rate": 3.335704125177809e-07,
"logps/chosen": -251.783935546875,
"logps/rejected": -321.7081298828125,
"loss": 0.1251,
"rewards/chosen": 2.4168410301208496,
"rewards/margins": 11.567142486572266,
"rewards/rejected": -9.150300979614258,
"step": 625
},
{
"epoch": 0.4,
"grad_norm": 18.487886428833008,
"kl": 0.0,
"learning_rate": 3.3179231863442384e-07,
"logps/chosen": -244.60250854492188,
"logps/rejected": -386.168212890625,
"loss": 0.0922,
"rewards/chosen": 2.371992588043213,
"rewards/margins": 13.357122421264648,
"rewards/rejected": -10.985128402709961,
"step": 630
},
{
"epoch": 0.41,
"grad_norm": 16.354122161865234,
"kl": 0.0,
"learning_rate": 3.300142247510668e-07,
"logps/chosen": -252.8701171875,
"logps/rejected": -334.31744384765625,
"loss": 0.1003,
"rewards/chosen": 2.335937023162842,
"rewards/margins": 12.5576753616333,
"rewards/rejected": -10.2217378616333,
"step": 635
},
{
"epoch": 0.41,
"grad_norm": 16.126724243164062,
"kl": 0.0,
"learning_rate": 3.282361308677098e-07,
"logps/chosen": -240.2421417236328,
"logps/rejected": -346.31414794921875,
"loss": 0.1069,
"rewards/chosen": 2.2422432899475098,
"rewards/margins": 11.99687671661377,
"rewards/rejected": -9.754633903503418,
"step": 640
},
{
"epoch": 0.41,
"grad_norm": 21.066267013549805,
"kl": 0.0,
"learning_rate": 3.2645803698435276e-07,
"logps/chosen": -264.81964111328125,
"logps/rejected": -338.95379638671875,
"loss": 0.1126,
"rewards/chosen": 2.3289685249328613,
"rewards/margins": 11.437704086303711,
"rewards/rejected": -9.108736038208008,
"step": 645
},
{
"epoch": 0.42,
"grad_norm": 14.892672538757324,
"kl": 0.0,
"learning_rate": 3.2467994310099573e-07,
"logps/chosen": -240.17236328125,
"logps/rejected": -355.543212890625,
"loss": 0.0884,
"rewards/chosen": 2.1047472953796387,
"rewards/margins": 12.319875717163086,
"rewards/rejected": -10.215127944946289,
"step": 650
},
{
"epoch": 0.42,
"grad_norm": 22.722187042236328,
"kl": 0.0,
"learning_rate": 3.229018492176387e-07,
"logps/chosen": -235.84341430664062,
"logps/rejected": -363.74041748046875,
"loss": 0.0958,
"rewards/chosen": 2.262721300125122,
"rewards/margins": 12.947868347167969,
"rewards/rejected": -10.685147285461426,
"step": 655
},
{
"epoch": 0.42,
"grad_norm": 16.061914443969727,
"kl": 0.0,
"learning_rate": 3.211237553342817e-07,
"logps/chosen": -232.480224609375,
"logps/rejected": -363.93792724609375,
"loss": 0.0947,
"rewards/chosen": 2.245720863342285,
"rewards/margins": 13.019895553588867,
"rewards/rejected": -10.774174690246582,
"step": 660
},
{
"epoch": 0.43,
"grad_norm": 17.95708656311035,
"kl": 0.0,
"learning_rate": 3.193456614509246e-07,
"logps/chosen": -266.2756652832031,
"logps/rejected": -349.5681457519531,
"loss": 0.0981,
"rewards/chosen": 2.0552451610565186,
"rewards/margins": 12.47476863861084,
"rewards/rejected": -10.419523239135742,
"step": 665
},
{
"epoch": 0.43,
"grad_norm": 22.32909393310547,
"kl": 0.0,
"learning_rate": 3.175675675675675e-07,
"logps/chosen": -262.5782165527344,
"logps/rejected": -350.5921936035156,
"loss": 0.1155,
"rewards/chosen": 2.324476718902588,
"rewards/margins": 12.327436447143555,
"rewards/rejected": -10.002958297729492,
"step": 670
},
{
"epoch": 0.43,
"grad_norm": 14.884044647216797,
"kl": 0.0,
"learning_rate": 3.157894736842105e-07,
"logps/chosen": -257.66094970703125,
"logps/rejected": -344.09869384765625,
"loss": 0.096,
"rewards/chosen": 2.2759206295013428,
"rewards/margins": 13.264117240905762,
"rewards/rejected": -10.988197326660156,
"step": 675
},
{
"epoch": 0.44,
"grad_norm": 18.041053771972656,
"kl": 0.0,
"learning_rate": 3.1401137980085347e-07,
"logps/chosen": -270.93646240234375,
"logps/rejected": -360.5249938964844,
"loss": 0.0958,
"rewards/chosen": 2.366556167602539,
"rewards/margins": 13.1105318069458,
"rewards/rejected": -10.743974685668945,
"step": 680
},
{
"epoch": 0.44,
"grad_norm": 17.0706729888916,
"kl": 0.0,
"learning_rate": 3.1223328591749644e-07,
"logps/chosen": -214.73690795898438,
"logps/rejected": -342.7330017089844,
"loss": 0.1107,
"rewards/chosen": 2.152832508087158,
"rewards/margins": 12.126806259155273,
"rewards/rejected": -9.973973274230957,
"step": 685
},
{
"epoch": 0.44,
"grad_norm": 25.26626968383789,
"kl": 0.0,
"learning_rate": 3.104551920341394e-07,
"logps/chosen": -293.3183898925781,
"logps/rejected": -356.9164123535156,
"loss": 0.0973,
"rewards/chosen": 2.3354926109313965,
"rewards/margins": 13.199078559875488,
"rewards/rejected": -10.863585472106934,
"step": 690
},
{
"epoch": 0.44,
"grad_norm": 28.42453384399414,
"kl": 0.0,
"learning_rate": 3.0867709815078234e-07,
"logps/chosen": -285.2272644042969,
"logps/rejected": -367.3414001464844,
"loss": 0.1091,
"rewards/chosen": 2.12477707862854,
"rewards/margins": 13.724832534790039,
"rewards/rejected": -11.600054740905762,
"step": 695
},
{
"epoch": 0.45,
"grad_norm": 17.899433135986328,
"kl": 0.0,
"learning_rate": 3.068990042674253e-07,
"logps/chosen": -262.5505065917969,
"logps/rejected": -371.45025634765625,
"loss": 0.1083,
"rewards/chosen": 2.060147762298584,
"rewards/margins": 12.760570526123047,
"rewards/rejected": -10.700422286987305,
"step": 700
},
{
"epoch": 0.45,
"grad_norm": 19.646583557128906,
"kl": 0.0,
"learning_rate": 3.051209103840683e-07,
"logps/chosen": -274.9856262207031,
"logps/rejected": -360.3489685058594,
"loss": 0.1105,
"rewards/chosen": 2.4847919940948486,
"rewards/margins": 13.168169975280762,
"rewards/rejected": -10.683378219604492,
"step": 705
},
{
"epoch": 0.45,
"grad_norm": 14.362700462341309,
"kl": 0.0,
"learning_rate": 3.033428165007112e-07,
"logps/chosen": -253.5523681640625,
"logps/rejected": -336.803466796875,
"loss": 0.1024,
"rewards/chosen": 2.3810598850250244,
"rewards/margins": 12.508206367492676,
"rewards/rejected": -10.127145767211914,
"step": 710
},
{
"epoch": 0.46,
"grad_norm": 22.93763542175293,
"kl": 0.0,
"learning_rate": 3.015647226173542e-07,
"logps/chosen": -229.1265869140625,
"logps/rejected": -354.0566101074219,
"loss": 0.1135,
"rewards/chosen": 2.1574866771698,
"rewards/margins": 12.193730354309082,
"rewards/rejected": -10.036243438720703,
"step": 715
},
{
"epoch": 0.46,
"grad_norm": 22.314695358276367,
"kl": 0.0,
"learning_rate": 2.9978662873399715e-07,
"logps/chosen": -274.2774353027344,
"logps/rejected": -405.2040100097656,
"loss": 0.0953,
"rewards/chosen": 2.191157817840576,
"rewards/margins": 13.628583908081055,
"rewards/rejected": -11.43742561340332,
"step": 720
},
{
"epoch": 0.46,
"grad_norm": 20.03668975830078,
"kl": 0.0,
"learning_rate": 2.9800853485064007e-07,
"logps/chosen": -250.0018768310547,
"logps/rejected": -340.919677734375,
"loss": 0.1022,
"rewards/chosen": 2.4375367164611816,
"rewards/margins": 12.8548583984375,
"rewards/rejected": -10.417322158813477,
"step": 725
},
{
"epoch": 0.47,
"grad_norm": 20.21332359313965,
"kl": 0.0,
"learning_rate": 2.9623044096728305e-07,
"logps/chosen": -262.5707092285156,
"logps/rejected": -341.51556396484375,
"loss": 0.113,
"rewards/chosen": 2.2380969524383545,
"rewards/margins": 12.677546501159668,
"rewards/rejected": -10.439449310302734,
"step": 730
},
{
"epoch": 0.47,
"grad_norm": 18.686241149902344,
"kl": 0.0,
"learning_rate": 2.94452347083926e-07,
"logps/chosen": -259.5292053222656,
"logps/rejected": -340.3607482910156,
"loss": 0.1188,
"rewards/chosen": 2.1912286281585693,
"rewards/margins": 12.15953540802002,
"rewards/rejected": -9.968307495117188,
"step": 735
},
{
"epoch": 0.47,
"grad_norm": 19.728899002075195,
"kl": 0.0,
"learning_rate": 2.92674253200569e-07,
"logps/chosen": -249.8451385498047,
"logps/rejected": -381.3360290527344,
"loss": 0.0967,
"rewards/chosen": 2.2508890628814697,
"rewards/margins": 13.840009689331055,
"rewards/rejected": -11.589120864868164,
"step": 740
},
{
"epoch": 0.48,
"grad_norm": 17.338115692138672,
"kl": 0.0,
"learning_rate": 2.9089615931721197e-07,
"logps/chosen": -268.9515380859375,
"logps/rejected": -368.5078125,
"loss": 0.1015,
"rewards/chosen": 2.28879976272583,
"rewards/margins": 13.396432876586914,
"rewards/rejected": -11.107633590698242,
"step": 745
},
{
"epoch": 0.48,
"grad_norm": 19.87563133239746,
"kl": 0.0,
"learning_rate": 2.8911806543385494e-07,
"logps/chosen": -293.7141418457031,
"logps/rejected": -355.55535888671875,
"loss": 0.1071,
"rewards/chosen": 2.4517719745635986,
"rewards/margins": 13.449313163757324,
"rewards/rejected": -10.997541427612305,
"step": 750
},
{
"epoch": 0.48,
"grad_norm": 19.807565689086914,
"kl": 0.0,
"learning_rate": 2.873399715504978e-07,
"logps/chosen": -248.6322784423828,
"logps/rejected": -353.8228759765625,
"loss": 0.1088,
"rewards/chosen": 2.2985501289367676,
"rewards/margins": 12.722993850708008,
"rewards/rejected": -10.424444198608398,
"step": 755
},
{
"epoch": 0.49,
"grad_norm": 14.098715782165527,
"kl": 0.0,
"learning_rate": 2.855618776671408e-07,
"logps/chosen": -240.1453857421875,
"logps/rejected": -330.6163635253906,
"loss": 0.0891,
"rewards/chosen": 2.550762414932251,
"rewards/margins": 12.813929557800293,
"rewards/rejected": -10.263166427612305,
"step": 760
},
{
"epoch": 0.49,
"grad_norm": 16.266538619995117,
"kl": 0.0,
"learning_rate": 2.8378378378378376e-07,
"logps/chosen": -236.1558380126953,
"logps/rejected": -360.686767578125,
"loss": 0.0973,
"rewards/chosen": 2.2286760807037354,
"rewards/margins": 13.117490768432617,
"rewards/rejected": -10.888814926147461,
"step": 765
},
{
"epoch": 0.49,
"grad_norm": 16.414714813232422,
"kl": 0.0,
"learning_rate": 2.8200568990042673e-07,
"logps/chosen": -283.36944580078125,
"logps/rejected": -368.7194519042969,
"loss": 0.0967,
"rewards/chosen": 2.4076812267303467,
"rewards/margins": 12.956690788269043,
"rewards/rejected": -10.549009323120117,
"step": 770
},
{
"epoch": 0.5,
"grad_norm": 18.088735580444336,
"kl": 0.0,
"learning_rate": 2.802275960170697e-07,
"logps/chosen": -302.5070495605469,
"logps/rejected": -348.53289794921875,
"loss": 0.1047,
"rewards/chosen": 2.4432990550994873,
"rewards/margins": 12.935566902160645,
"rewards/rejected": -10.492268562316895,
"step": 775
},
{
"epoch": 0.5,
"grad_norm": 16.778833389282227,
"kl": 0.0,
"learning_rate": 2.784495021337127e-07,
"logps/chosen": -243.6858367919922,
"logps/rejected": -366.0279235839844,
"loss": 0.0852,
"rewards/chosen": 2.5511088371276855,
"rewards/margins": 13.161443710327148,
"rewards/rejected": -10.610334396362305,
"step": 780
},
{
"epoch": 0.5,
"grad_norm": 17.379335403442383,
"kl": 0.0,
"learning_rate": 2.766714082503556e-07,
"logps/chosen": -279.8948669433594,
"logps/rejected": -352.0287170410156,
"loss": 0.0957,
"rewards/chosen": 2.464888334274292,
"rewards/margins": 13.103485107421875,
"rewards/rejected": -10.638595581054688,
"step": 785
},
{
"epoch": 0.51,
"grad_norm": 21.737380981445312,
"kl": 0.0,
"learning_rate": 2.7489331436699857e-07,
"logps/chosen": -258.93865966796875,
"logps/rejected": -345.5055236816406,
"loss": 0.1177,
"rewards/chosen": 2.079983711242676,
"rewards/margins": 12.90088176727295,
"rewards/rejected": -10.820898056030273,
"step": 790
},
{
"epoch": 0.51,
"grad_norm": 15.227828025817871,
"kl": 0.0,
"learning_rate": 2.7311522048364154e-07,
"logps/chosen": -233.96194458007812,
"logps/rejected": -370.0096130371094,
"loss": 0.1054,
"rewards/chosen": 2.696063280105591,
"rewards/margins": 13.426877975463867,
"rewards/rejected": -10.730813980102539,
"step": 795
},
{
"epoch": 0.51,
"grad_norm": 13.938480377197266,
"kl": 0.0,
"learning_rate": 2.7133712660028446e-07,
"logps/chosen": -243.730224609375,
"logps/rejected": -350.44183349609375,
"loss": 0.0976,
"rewards/chosen": 2.3282103538513184,
"rewards/margins": 13.082319259643555,
"rewards/rejected": -10.754108428955078,
"step": 800
},
{
"epoch": 0.52,
"grad_norm": 29.015178680419922,
"kl": 0.0,
"learning_rate": 2.6955903271692744e-07,
"logps/chosen": -216.66488647460938,
"logps/rejected": -368.9279479980469,
"loss": 0.1045,
"rewards/chosen": 2.2219786643981934,
"rewards/margins": 13.255975723266602,
"rewards/rejected": -11.033994674682617,
"step": 805
},
{
"epoch": 0.52,
"grad_norm": 12.759780883789062,
"kl": 0.0,
"learning_rate": 2.677809388335704e-07,
"logps/chosen": -232.0436248779297,
"logps/rejected": -358.2416076660156,
"loss": 0.0943,
"rewards/chosen": 2.235215425491333,
"rewards/margins": 12.901985168457031,
"rewards/rejected": -10.666769027709961,
"step": 810
},
{
"epoch": 0.52,
"grad_norm": 19.0715389251709,
"kl": 0.0,
"learning_rate": 2.6600284495021333e-07,
"logps/chosen": -245.0729217529297,
"logps/rejected": -379.9324645996094,
"loss": 0.1173,
"rewards/chosen": 2.423140287399292,
"rewards/margins": 13.03075885772705,
"rewards/rejected": -10.60761833190918,
"step": 815
},
{
"epoch": 0.52,
"grad_norm": 15.53042984008789,
"kl": 0.0,
"learning_rate": 2.642247510668563e-07,
"logps/chosen": -261.9339294433594,
"logps/rejected": -327.780517578125,
"loss": 0.0903,
"rewards/chosen": 2.4564433097839355,
"rewards/margins": 13.401847839355469,
"rewards/rejected": -10.945404052734375,
"step": 820
},
{
"epoch": 0.53,
"grad_norm": 26.67053985595703,
"kl": 0.0,
"learning_rate": 2.624466571834993e-07,
"logps/chosen": -248.34487915039062,
"logps/rejected": -345.79083251953125,
"loss": 0.1071,
"rewards/chosen": 2.359476089477539,
"rewards/margins": 13.113537788391113,
"rewards/rejected": -10.754061698913574,
"step": 825
},
{
"epoch": 0.53,
"grad_norm": 22.24759292602539,
"kl": 0.0,
"learning_rate": 2.6066856330014225e-07,
"logps/chosen": -262.4079895019531,
"logps/rejected": -364.2686462402344,
"loss": 0.1076,
"rewards/chosen": 2.37235689163208,
"rewards/margins": 13.127764701843262,
"rewards/rejected": -10.75540828704834,
"step": 830
},
{
"epoch": 0.53,
"grad_norm": 17.638994216918945,
"kl": 0.0,
"learning_rate": 2.5889046941678523e-07,
"logps/chosen": -273.74432373046875,
"logps/rejected": -383.0593566894531,
"loss": 0.0984,
"rewards/chosen": 2.275177478790283,
"rewards/margins": 14.033803939819336,
"rewards/rejected": -11.758626937866211,
"step": 835
},
{
"epoch": 0.54,
"grad_norm": 15.510226249694824,
"kl": 0.0,
"learning_rate": 2.5711237553342815e-07,
"logps/chosen": -268.3088073730469,
"logps/rejected": -372.16876220703125,
"loss": 0.075,
"rewards/chosen": 2.7448737621307373,
"rewards/margins": 14.005487442016602,
"rewards/rejected": -11.260614395141602,
"step": 840
},
{
"epoch": 0.54,
"grad_norm": 18.002506256103516,
"kl": 0.0,
"learning_rate": 2.5533428165007107e-07,
"logps/chosen": -214.6236572265625,
"logps/rejected": -379.697265625,
"loss": 0.1156,
"rewards/chosen": 2.090531349182129,
"rewards/margins": 14.037800788879395,
"rewards/rejected": -11.947270393371582,
"step": 845
},
{
"epoch": 0.54,
"grad_norm": 18.8648681640625,
"kl": 0.0,
"learning_rate": 2.5355618776671404e-07,
"logps/chosen": -230.46810913085938,
"logps/rejected": -345.8827209472656,
"loss": 0.1028,
"rewards/chosen": 2.3623595237731934,
"rewards/margins": 13.34724235534668,
"rewards/rejected": -10.984883308410645,
"step": 850
},
{
"epoch": 0.55,
"grad_norm": 15.514137268066406,
"kl": 0.0,
"learning_rate": 2.51778093883357e-07,
"logps/chosen": -224.1614227294922,
"logps/rejected": -365.08477783203125,
"loss": 0.0985,
"rewards/chosen": 2.2946102619171143,
"rewards/margins": 14.019018173217773,
"rewards/rejected": -11.724408149719238,
"step": 855
},
{
"epoch": 0.55,
"grad_norm": 17.679288864135742,
"kl": 0.0,
"learning_rate": 2.5e-07,
"logps/chosen": -231.95565795898438,
"logps/rejected": -349.056640625,
"loss": 0.0966,
"rewards/chosen": 2.3342556953430176,
"rewards/margins": 13.233372688293457,
"rewards/rejected": -10.899115562438965,
"step": 860
},
{
"epoch": 0.55,
"grad_norm": 19.04916000366211,
"kl": 0.0,
"learning_rate": 2.4822190611664296e-07,
"logps/chosen": -233.5281982421875,
"logps/rejected": -374.4831848144531,
"loss": 0.0922,
"rewards/chosen": 2.58280873298645,
"rewards/margins": 13.661170959472656,
"rewards/rejected": -11.078360557556152,
"step": 865
},
{
"epoch": 0.56,
"grad_norm": 16.834497451782227,
"kl": 0.0,
"learning_rate": 2.4644381223328594e-07,
"logps/chosen": -236.75106811523438,
"logps/rejected": -353.3365478515625,
"loss": 0.096,
"rewards/chosen": 2.7401375770568848,
"rewards/margins": 13.776086807250977,
"rewards/rejected": -11.035948753356934,
"step": 870
},
{
"epoch": 0.56,
"grad_norm": 24.789772033691406,
"kl": 0.0,
"learning_rate": 2.4466571834992886e-07,
"logps/chosen": -267.3653869628906,
"logps/rejected": -370.1556091308594,
"loss": 0.0949,
"rewards/chosen": 2.2669379711151123,
"rewards/margins": 13.626495361328125,
"rewards/rejected": -11.35955810546875,
"step": 875
},
{
"epoch": 0.56,
"grad_norm": 12.911144256591797,
"kl": 0.0,
"learning_rate": 2.4288762446657183e-07,
"logps/chosen": -250.23556518554688,
"logps/rejected": -363.0490417480469,
"loss": 0.0717,
"rewards/chosen": 2.6297707557678223,
"rewards/margins": 15.319549560546875,
"rewards/rejected": -12.689778327941895,
"step": 880
},
{
"epoch": 0.57,
"grad_norm": 18.159881591796875,
"kl": 0.0,
"learning_rate": 2.411095305832148e-07,
"logps/chosen": -276.918701171875,
"logps/rejected": -351.54132080078125,
"loss": 0.1062,
"rewards/chosen": 2.5637714862823486,
"rewards/margins": 13.2720365524292,
"rewards/rejected": -10.708267211914062,
"step": 885
},
{
"epoch": 0.57,
"grad_norm": 16.24278450012207,
"kl": 0.0,
"learning_rate": 2.393314366998578e-07,
"logps/chosen": -264.356689453125,
"logps/rejected": -345.72064208984375,
"loss": 0.1061,
"rewards/chosen": 2.57415509223938,
"rewards/margins": 13.096124649047852,
"rewards/rejected": -10.521968841552734,
"step": 890
},
{
"epoch": 0.57,
"grad_norm": 15.353395462036133,
"kl": 0.0,
"learning_rate": 2.375533428165007e-07,
"logps/chosen": -245.558349609375,
"logps/rejected": -346.11151123046875,
"loss": 0.1098,
"rewards/chosen": 2.0979220867156982,
"rewards/margins": 13.061151504516602,
"rewards/rejected": -10.963228225708008,
"step": 895
},
{
"epoch": 0.58,
"grad_norm": 15.291166305541992,
"kl": 0.0,
"learning_rate": 2.3577524893314365e-07,
"logps/chosen": -268.0149841308594,
"logps/rejected": -359.99676513671875,
"loss": 0.0935,
"rewards/chosen": 2.660017251968384,
"rewards/margins": 13.785112380981445,
"rewards/rejected": -11.125094413757324,
"step": 900
},
{
"epoch": 0.58,
"grad_norm": 27.338308334350586,
"kl": 0.0,
"learning_rate": 2.3399715504978662e-07,
"logps/chosen": -223.98062133789062,
"logps/rejected": -382.53924560546875,
"loss": 0.0733,
"rewards/chosen": 2.573464870452881,
"rewards/margins": 14.070528030395508,
"rewards/rejected": -11.497062683105469,
"step": 905
},
{
"epoch": 0.58,
"grad_norm": 28.126638412475586,
"kl": 0.0,
"learning_rate": 2.322190611664296e-07,
"logps/chosen": -225.90432739257812,
"logps/rejected": -369.765380859375,
"loss": 0.1068,
"rewards/chosen": 2.4907584190368652,
"rewards/margins": 13.410183906555176,
"rewards/rejected": -10.919425964355469,
"step": 910
},
{
"epoch": 0.59,
"grad_norm": 23.488555908203125,
"kl": 0.0,
"learning_rate": 2.304409672830725e-07,
"logps/chosen": -264.053466796875,
"logps/rejected": -347.4461364746094,
"loss": 0.0978,
"rewards/chosen": 2.3625271320343018,
"rewards/margins": 13.296697616577148,
"rewards/rejected": -10.934170722961426,
"step": 915
},
{
"epoch": 0.59,
"grad_norm": 16.538496017456055,
"kl": 0.0,
"learning_rate": 2.2866287339971549e-07,
"logps/chosen": -305.2149353027344,
"logps/rejected": -387.3183898925781,
"loss": 0.0895,
"rewards/chosen": 2.51884126663208,
"rewards/margins": 14.519546508789062,
"rewards/rejected": -12.000704765319824,
"step": 920
},
{
"epoch": 0.59,
"grad_norm": 14.570239067077637,
"kl": 0.0,
"learning_rate": 2.2688477951635846e-07,
"logps/chosen": -288.55859375,
"logps/rejected": -360.0328063964844,
"loss": 0.0891,
"rewards/chosen": 2.6060962677001953,
"rewards/margins": 13.63697338104248,
"rewards/rejected": -11.030878067016602,
"step": 925
},
{
"epoch": 0.6,
"grad_norm": 14.252922058105469,
"kl": 0.0,
"learning_rate": 2.251066856330014e-07,
"logps/chosen": -237.12985229492188,
"logps/rejected": -375.7847900390625,
"loss": 0.0968,
"rewards/chosen": 2.277179718017578,
"rewards/margins": 13.618906021118164,
"rewards/rejected": -11.341727256774902,
"step": 930
},
{
"epoch": 0.6,
"grad_norm": 14.224340438842773,
"kl": 0.0,
"learning_rate": 2.2332859174964438e-07,
"logps/chosen": -247.994384765625,
"logps/rejected": -384.1309509277344,
"loss": 0.0921,
"rewards/chosen": 2.5065815448760986,
"rewards/margins": 14.67822551727295,
"rewards/rejected": -12.17164421081543,
"step": 935
},
{
"epoch": 0.6,
"grad_norm": 15.533319473266602,
"kl": 0.0,
"learning_rate": 2.2155049786628733e-07,
"logps/chosen": -217.72982788085938,
"logps/rejected": -364.0068054199219,
"loss": 0.1044,
"rewards/chosen": 2.4342358112335205,
"rewards/margins": 13.825854301452637,
"rewards/rejected": -11.391618728637695,
"step": 940
},
{
"epoch": 0.6,
"grad_norm": 12.951379776000977,
"kl": 0.0,
"learning_rate": 2.1977240398293027e-07,
"logps/chosen": -216.9298858642578,
"logps/rejected": -391.88916015625,
"loss": 0.0959,
"rewards/chosen": 2.2248167991638184,
"rewards/margins": 14.502099990844727,
"rewards/rejected": -12.27728271484375,
"step": 945
},
{
"epoch": 0.61,
"grad_norm": 12.726017951965332,
"kl": 0.0,
"learning_rate": 2.1799431009957325e-07,
"logps/chosen": -250.4014129638672,
"logps/rejected": -367.2666931152344,
"loss": 0.099,
"rewards/chosen": 2.6326889991760254,
"rewards/margins": 14.203544616699219,
"rewards/rejected": -11.570856094360352,
"step": 950
},
{
"epoch": 0.61,
"grad_norm": 24.116321563720703,
"kl": 0.0,
"learning_rate": 2.1621621621621622e-07,
"logps/chosen": -256.0219421386719,
"logps/rejected": -359.89410400390625,
"loss": 0.1114,
"rewards/chosen": 2.291504383087158,
"rewards/margins": 13.204824447631836,
"rewards/rejected": -10.913320541381836,
"step": 955
},
{
"epoch": 0.61,
"grad_norm": 21.19695472717285,
"kl": 0.0,
"learning_rate": 2.1443812233285914e-07,
"logps/chosen": -253.9010009765625,
"logps/rejected": -348.592041015625,
"loss": 0.0887,
"rewards/chosen": 2.41105580329895,
"rewards/margins": 13.58563232421875,
"rewards/rejected": -11.174577713012695,
"step": 960
},
{
"epoch": 0.62,
"grad_norm": 19.317626953125,
"kl": 0.0,
"learning_rate": 2.1266002844950212e-07,
"logps/chosen": -245.3134765625,
"logps/rejected": -380.17327880859375,
"loss": 0.0998,
"rewards/chosen": 2.4610061645507812,
"rewards/margins": 13.953335762023926,
"rewards/rejected": -11.492330551147461,
"step": 965
},
{
"epoch": 0.62,
"grad_norm": 22.688295364379883,
"kl": 0.0,
"learning_rate": 2.108819345661451e-07,
"logps/chosen": -278.2052917480469,
"logps/rejected": -357.1881408691406,
"loss": 0.1063,
"rewards/chosen": 2.404897928237915,
"rewards/margins": 13.51880931854248,
"rewards/rejected": -11.113912582397461,
"step": 970
},
{
"epoch": 0.62,
"grad_norm": 14.956331253051758,
"kl": 0.0,
"learning_rate": 2.0910384068278806e-07,
"logps/chosen": -274.7865295410156,
"logps/rejected": -348.69024658203125,
"loss": 0.1178,
"rewards/chosen": 2.560181140899658,
"rewards/margins": 13.09942626953125,
"rewards/rejected": -10.539244651794434,
"step": 975
},
{
"epoch": 0.63,
"grad_norm": 14.130335807800293,
"kl": 0.0,
"learning_rate": 2.0732574679943098e-07,
"logps/chosen": -241.42257690429688,
"logps/rejected": -374.1187438964844,
"loss": 0.0728,
"rewards/chosen": 2.970536470413208,
"rewards/margins": 14.709304809570312,
"rewards/rejected": -11.738768577575684,
"step": 980
},
{
"epoch": 0.63,
"grad_norm": 20.794334411621094,
"kl": 0.0,
"learning_rate": 2.0554765291607396e-07,
"logps/chosen": -230.2430419921875,
"logps/rejected": -380.54608154296875,
"loss": 0.09,
"rewards/chosen": 2.673881769180298,
"rewards/margins": 14.4403076171875,
"rewards/rejected": -11.766425132751465,
"step": 985
},
{
"epoch": 0.63,
"grad_norm": 16.297340393066406,
"kl": 0.0,
"learning_rate": 2.0376955903271693e-07,
"logps/chosen": -264.10137939453125,
"logps/rejected": -372.11346435546875,
"loss": 0.0896,
"rewards/chosen": 2.7595813274383545,
"rewards/margins": 13.790555953979492,
"rewards/rejected": -11.030974388122559,
"step": 990
},
{
"epoch": 0.64,
"grad_norm": 20.304162979125977,
"kl": 0.0,
"learning_rate": 2.0199146514935988e-07,
"logps/chosen": -243.2149200439453,
"logps/rejected": -342.080322265625,
"loss": 0.1091,
"rewards/chosen": 2.3006443977355957,
"rewards/margins": 13.069559097290039,
"rewards/rejected": -10.768914222717285,
"step": 995
},
{
"epoch": 0.64,
"grad_norm": 19.80646514892578,
"kl": 0.0,
"learning_rate": 2.0021337126600283e-07,
"logps/chosen": -251.78524780273438,
"logps/rejected": -360.4002380371094,
"loss": 0.088,
"rewards/chosen": 2.7110652923583984,
"rewards/margins": 13.96354866027832,
"rewards/rejected": -11.252483367919922,
"step": 1000
},
{
"epoch": 0.64,
"grad_norm": 19.04472541809082,
"kl": 0.0,
"learning_rate": 1.984352773826458e-07,
"logps/chosen": -227.35873413085938,
"logps/rejected": -363.24713134765625,
"loss": 0.0952,
"rewards/chosen": 2.7294869422912598,
"rewards/margins": 14.287317276000977,
"rewards/rejected": -11.557830810546875,
"step": 1005
},
{
"epoch": 0.65,
"grad_norm": 14.0511474609375,
"kl": 0.0,
"learning_rate": 1.9665718349928875e-07,
"logps/chosen": -248.2740478515625,
"logps/rejected": -359.29534912109375,
"loss": 0.0783,
"rewards/chosen": 2.7295162677764893,
"rewards/margins": 14.418438911437988,
"rewards/rejected": -11.688921928405762,
"step": 1010
},
{
"epoch": 0.65,
"grad_norm": 21.88315773010254,
"kl": 0.0,
"learning_rate": 1.9487908961593172e-07,
"logps/chosen": -228.99447631835938,
"logps/rejected": -387.5201721191406,
"loss": 0.0771,
"rewards/chosen": 2.3790504932403564,
"rewards/margins": 13.84051513671875,
"rewards/rejected": -11.461464881896973,
"step": 1015
},
{
"epoch": 0.65,
"grad_norm": 15.932575225830078,
"kl": 0.0,
"learning_rate": 1.931009957325747e-07,
"logps/chosen": -278.6766357421875,
"logps/rejected": -372.94793701171875,
"loss": 0.0826,
"rewards/chosen": 2.363306760787964,
"rewards/margins": 14.281936645507812,
"rewards/rejected": -11.918628692626953,
"step": 1020
},
{
"epoch": 0.66,
"grad_norm": 11.954800605773926,
"kl": 0.0,
"learning_rate": 1.9132290184921761e-07,
"logps/chosen": -223.6442108154297,
"logps/rejected": -371.5616149902344,
"loss": 0.0961,
"rewards/chosen": 2.332761526107788,
"rewards/margins": 13.71965217590332,
"rewards/rejected": -11.386890411376953,
"step": 1025
},
{
"epoch": 0.66,
"grad_norm": 14.154836654663086,
"kl": 0.0,
"learning_rate": 1.895448079658606e-07,
"logps/chosen": -223.6517791748047,
"logps/rejected": -386.1749267578125,
"loss": 0.0908,
"rewards/chosen": 2.7226128578186035,
"rewards/margins": 14.438260078430176,
"rewards/rejected": -11.71564769744873,
"step": 1030
},
{
"epoch": 0.66,
"grad_norm": 18.134370803833008,
"kl": 0.0,
"learning_rate": 1.8776671408250356e-07,
"logps/chosen": -260.0550842285156,
"logps/rejected": -346.24127197265625,
"loss": 0.1007,
"rewards/chosen": 2.5339343547821045,
"rewards/margins": 13.723528861999512,
"rewards/rejected": -11.189595222473145,
"step": 1035
},
{
"epoch": 0.67,
"grad_norm": 15.416353225708008,
"kl": 0.0,
"learning_rate": 1.859886201991465e-07,
"logps/chosen": -247.7535400390625,
"logps/rejected": -371.42034912109375,
"loss": 0.0912,
"rewards/chosen": 2.3352103233337402,
"rewards/margins": 13.56823444366455,
"rewards/rejected": -11.233022689819336,
"step": 1040
},
{
"epoch": 0.67,
"grad_norm": 18.23054313659668,
"kl": 0.0,
"learning_rate": 1.8421052631578946e-07,
"logps/chosen": -213.8344268798828,
"logps/rejected": -369.88702392578125,
"loss": 0.0758,
"rewards/chosen": 2.6615304946899414,
"rewards/margins": 14.028742790222168,
"rewards/rejected": -11.367212295532227,
"step": 1045
},
{
"epoch": 0.67,
"grad_norm": 18.42432975769043,
"kl": 0.0,
"learning_rate": 1.8243243243243243e-07,
"logps/chosen": -238.5988006591797,
"logps/rejected": -392.305419921875,
"loss": 0.0965,
"rewards/chosen": 2.5309016704559326,
"rewards/margins": 14.56842041015625,
"rewards/rejected": -12.037518501281738,
"step": 1050
},
{
"epoch": 0.67,
"grad_norm": 15.395295143127441,
"kl": 0.0,
"learning_rate": 1.8065433854907538e-07,
"logps/chosen": -242.9210968017578,
"logps/rejected": -375.423828125,
"loss": 0.0985,
"rewards/chosen": 2.644166946411133,
"rewards/margins": 13.905471801757812,
"rewards/rejected": -11.26130485534668,
"step": 1055
},
{
"epoch": 0.68,
"grad_norm": 16.343006134033203,
"kl": 0.0,
"learning_rate": 1.7887624466571835e-07,
"logps/chosen": -198.0528106689453,
"logps/rejected": -354.1863098144531,
"loss": 0.092,
"rewards/chosen": 2.4538371562957764,
"rewards/margins": 12.678072929382324,
"rewards/rejected": -10.224235534667969,
"step": 1060
},
{
"epoch": 0.68,
"grad_norm": 19.673669815063477,
"kl": 0.0,
"learning_rate": 1.770981507823613e-07,
"logps/chosen": -241.1757049560547,
"logps/rejected": -339.1280822753906,
"loss": 0.088,
"rewards/chosen": 2.550891876220703,
"rewards/margins": 13.831698417663574,
"rewards/rejected": -11.280807495117188,
"step": 1065
},
{
"epoch": 0.68,
"grad_norm": 15.624777793884277,
"kl": 0.0,
"learning_rate": 1.7532005689900424e-07,
"logps/chosen": -247.8914031982422,
"logps/rejected": -364.5996398925781,
"loss": 0.1038,
"rewards/chosen": 2.5451152324676514,
"rewards/margins": 13.162847518920898,
"rewards/rejected": -10.617732048034668,
"step": 1070
},
{
"epoch": 0.69,
"grad_norm": 16.0270938873291,
"kl": 0.0,
"learning_rate": 1.7354196301564722e-07,
"logps/chosen": -271.2415466308594,
"logps/rejected": -348.6362609863281,
"loss": 0.0832,
"rewards/chosen": 2.6901183128356934,
"rewards/margins": 13.692281723022461,
"rewards/rejected": -11.002164840698242,
"step": 1075
},
{
"epoch": 0.69,
"grad_norm": 16.198070526123047,
"kl": 0.0,
"learning_rate": 1.717638691322902e-07,
"logps/chosen": -228.4770965576172,
"logps/rejected": -371.384521484375,
"loss": 0.0879,
"rewards/chosen": 2.317864179611206,
"rewards/margins": 14.147501945495605,
"rewards/rejected": -11.82963752746582,
"step": 1080
},
{
"epoch": 0.69,
"grad_norm": 24.463367462158203,
"kl": 0.0,
"learning_rate": 1.6998577524893314e-07,
"logps/chosen": -271.0051574707031,
"logps/rejected": -403.8445129394531,
"loss": 0.0817,
"rewards/chosen": 2.82243275642395,
"rewards/margins": 14.68195629119873,
"rewards/rejected": -11.859524726867676,
"step": 1085
},
{
"epoch": 0.7,
"grad_norm": 21.81173324584961,
"kl": 0.0,
"learning_rate": 1.6820768136557609e-07,
"logps/chosen": -209.64572143554688,
"logps/rejected": -360.5612487792969,
"loss": 0.0849,
"rewards/chosen": 2.4915404319763184,
"rewards/margins": 13.439886093139648,
"rewards/rejected": -10.948348045349121,
"step": 1090
},
{
"epoch": 0.7,
"grad_norm": 20.325061798095703,
"kl": 0.0,
"learning_rate": 1.6642958748221906e-07,
"logps/chosen": -244.2580108642578,
"logps/rejected": -348.1650390625,
"loss": 0.0804,
"rewards/chosen": 2.744297742843628,
"rewards/margins": 13.572134017944336,
"rewards/rejected": -10.827836990356445,
"step": 1095
},
{
"epoch": 0.7,
"grad_norm": 24.380229949951172,
"kl": 0.0,
"learning_rate": 1.64651493598862e-07,
"logps/chosen": -253.1858673095703,
"logps/rejected": -348.04144287109375,
"loss": 0.0846,
"rewards/chosen": 2.656752824783325,
"rewards/margins": 13.392297744750977,
"rewards/rejected": -10.735544204711914,
"step": 1100
},
{
"epoch": 0.71,
"grad_norm": 12.965998649597168,
"kl": 0.0,
"learning_rate": 1.6287339971550498e-07,
"logps/chosen": -262.8606262207031,
"logps/rejected": -356.92242431640625,
"loss": 0.0783,
"rewards/chosen": 2.8926401138305664,
"rewards/margins": 14.1354341506958,
"rewards/rejected": -11.242793083190918,
"step": 1105
},
{
"epoch": 0.71,
"grad_norm": 16.383338928222656,
"kl": 0.0,
"learning_rate": 1.6109530583214793e-07,
"logps/chosen": -300.94384765625,
"logps/rejected": -358.71588134765625,
"loss": 0.0951,
"rewards/chosen": 2.539355993270874,
"rewards/margins": 13.336338996887207,
"rewards/rejected": -10.79698371887207,
"step": 1110
},
{
"epoch": 0.71,
"grad_norm": 15.756513595581055,
"kl": 0.0,
"learning_rate": 1.5931721194879087e-07,
"logps/chosen": -261.84637451171875,
"logps/rejected": -374.3198547363281,
"loss": 0.0791,
"rewards/chosen": 2.755915880203247,
"rewards/margins": 14.211552619934082,
"rewards/rejected": -11.455635070800781,
"step": 1115
},
{
"epoch": 0.72,
"grad_norm": 18.087120056152344,
"kl": 0.0,
"learning_rate": 1.5753911806543385e-07,
"logps/chosen": -274.73370361328125,
"logps/rejected": -326.76123046875,
"loss": 0.0935,
"rewards/chosen": 2.786874294281006,
"rewards/margins": 12.87488079071045,
"rewards/rejected": -10.088006973266602,
"step": 1120
},
{
"epoch": 0.72,
"grad_norm": 15.65069580078125,
"kl": 0.0,
"learning_rate": 1.5576102418207682e-07,
"logps/chosen": -235.80062866210938,
"logps/rejected": -340.6771545410156,
"loss": 0.1005,
"rewards/chosen": 2.691729784011841,
"rewards/margins": 13.200857162475586,
"rewards/rejected": -10.509127616882324,
"step": 1125
},
{
"epoch": 0.72,
"grad_norm": 25.325843811035156,
"kl": 0.0,
"learning_rate": 1.5398293029871974e-07,
"logps/chosen": -244.3973388671875,
"logps/rejected": -343.04412841796875,
"loss": 0.0883,
"rewards/chosen": 2.485719680786133,
"rewards/margins": 13.161664962768555,
"rewards/rejected": -10.675946235656738,
"step": 1130
},
{
"epoch": 0.73,
"grad_norm": 17.044322967529297,
"kl": 0.0,
"learning_rate": 1.5220483641536272e-07,
"logps/chosen": -244.70559692382812,
"logps/rejected": -358.86920166015625,
"loss": 0.0987,
"rewards/chosen": 2.5949511528015137,
"rewards/margins": 13.485272407531738,
"rewards/rejected": -10.890320777893066,
"step": 1135
},
{
"epoch": 0.73,
"grad_norm": 27.13024139404297,
"kl": 0.0,
"learning_rate": 1.504267425320057e-07,
"logps/chosen": -247.91641235351562,
"logps/rejected": -343.4930419921875,
"loss": 0.0899,
"rewards/chosen": 2.4175989627838135,
"rewards/margins": 13.832735061645508,
"rewards/rejected": -11.415135383605957,
"step": 1140
},
{
"epoch": 0.73,
"grad_norm": 14.320551872253418,
"kl": 0.0,
"learning_rate": 1.4864864864864866e-07,
"logps/chosen": -246.96670532226562,
"logps/rejected": -370.0877990722656,
"loss": 0.0758,
"rewards/chosen": 2.691598653793335,
"rewards/margins": 13.812780380249023,
"rewards/rejected": -11.121182441711426,
"step": 1145
},
{
"epoch": 0.74,
"grad_norm": 16.582712173461914,
"kl": 0.0,
"learning_rate": 1.4687055476529158e-07,
"logps/chosen": -224.07785034179688,
"logps/rejected": -349.31927490234375,
"loss": 0.082,
"rewards/chosen": 2.6934258937835693,
"rewards/margins": 13.236696243286133,
"rewards/rejected": -10.543269157409668,
"step": 1150
},
{
"epoch": 0.74,
"grad_norm": 17.989919662475586,
"kl": 0.0,
"learning_rate": 1.4509246088193456e-07,
"logps/chosen": -243.8923797607422,
"logps/rejected": -367.2593688964844,
"loss": 0.0846,
"rewards/chosen": 2.7338271141052246,
"rewards/margins": 14.203822135925293,
"rewards/rejected": -11.469995498657227,
"step": 1155
},
{
"epoch": 0.74,
"grad_norm": 17.582975387573242,
"kl": 0.0,
"learning_rate": 1.4331436699857753e-07,
"logps/chosen": -227.1583709716797,
"logps/rejected": -378.5196228027344,
"loss": 0.1011,
"rewards/chosen": 2.531038284301758,
"rewards/margins": 13.578967094421387,
"rewards/rejected": -11.047929763793945,
"step": 1160
},
{
"epoch": 0.75,
"grad_norm": 8.821680068969727,
"kl": 0.0,
"learning_rate": 1.4153627311522048e-07,
"logps/chosen": -277.82684326171875,
"logps/rejected": -364.6512145996094,
"loss": 0.0742,
"rewards/chosen": 2.9555764198303223,
"rewards/margins": 14.14686393737793,
"rewards/rejected": -11.19128704071045,
"step": 1165
},
{
"epoch": 0.75,
"grad_norm": 15.817394256591797,
"kl": 0.0,
"learning_rate": 1.3975817923186345e-07,
"logps/chosen": -271.718017578125,
"logps/rejected": -352.4165954589844,
"loss": 0.0848,
"rewards/chosen": 2.5184245109558105,
"rewards/margins": 13.12096881866455,
"rewards/rejected": -10.602544784545898,
"step": 1170
},
{
"epoch": 0.75,
"grad_norm": 18.2441463470459,
"kl": 0.0,
"learning_rate": 1.379800853485064e-07,
"logps/chosen": -232.19467163085938,
"logps/rejected": -379.3485412597656,
"loss": 0.0909,
"rewards/chosen": 2.5335030555725098,
"rewards/margins": 14.489652633666992,
"rewards/rejected": -11.956149101257324,
"step": 1175
},
{
"epoch": 0.75,
"grad_norm": 30.594867706298828,
"kl": 0.0,
"learning_rate": 1.3620199146514935e-07,
"logps/chosen": -265.6886291503906,
"logps/rejected": -361.285400390625,
"loss": 0.0964,
"rewards/chosen": 2.450854778289795,
"rewards/margins": 13.442533493041992,
"rewards/rejected": -10.991681098937988,
"step": 1180
},
{
"epoch": 0.76,
"grad_norm": 12.528332710266113,
"kl": 0.0,
"learning_rate": 1.3442389758179232e-07,
"logps/chosen": -288.8167724609375,
"logps/rejected": -352.4346618652344,
"loss": 0.0741,
"rewards/chosen": 2.7889742851257324,
"rewards/margins": 14.034700393676758,
"rewards/rejected": -11.245725631713867,
"step": 1185
},
{
"epoch": 0.76,
"grad_norm": 13.565973281860352,
"kl": 0.0,
"learning_rate": 1.326458036984353e-07,
"logps/chosen": -263.8364562988281,
"logps/rejected": -346.26019287109375,
"loss": 0.0843,
"rewards/chosen": 2.638777256011963,
"rewards/margins": 13.351663589477539,
"rewards/rejected": -10.712886810302734,
"step": 1190
},
{
"epoch": 0.76,
"grad_norm": 21.124759674072266,
"kl": 0.0,
"learning_rate": 1.3086770981507821e-07,
"logps/chosen": -245.15576171875,
"logps/rejected": -372.0450134277344,
"loss": 0.0898,
"rewards/chosen": 2.657459259033203,
"rewards/margins": 14.276753425598145,
"rewards/rejected": -11.619293212890625,
"step": 1195
},
{
"epoch": 0.77,
"grad_norm": 15.396940231323242,
"kl": 0.0,
"learning_rate": 1.290896159317212e-07,
"logps/chosen": -247.9556121826172,
"logps/rejected": -371.3470764160156,
"loss": 0.0828,
"rewards/chosen": 2.4249672889709473,
"rewards/margins": 14.25316333770752,
"rewards/rejected": -11.828195571899414,
"step": 1200
},
{
"epoch": 0.77,
"grad_norm": 14.409557342529297,
"kl": 0.0,
"learning_rate": 1.2731152204836416e-07,
"logps/chosen": -252.07241821289062,
"logps/rejected": -361.1800842285156,
"loss": 0.085,
"rewards/chosen": 2.680985689163208,
"rewards/margins": 14.30175495147705,
"rewards/rejected": -11.620769500732422,
"step": 1205
},
{
"epoch": 0.77,
"grad_norm": 13.202485084533691,
"kl": 0.0,
"learning_rate": 1.255334281650071e-07,
"logps/chosen": -237.88565063476562,
"logps/rejected": -350.7021179199219,
"loss": 0.0874,
"rewards/chosen": 2.6545567512512207,
"rewards/margins": 13.95142936706543,
"rewards/rejected": -11.29687213897705,
"step": 1210
},
{
"epoch": 0.78,
"grad_norm": 28.25654411315918,
"kl": 0.0,
"learning_rate": 1.2375533428165005e-07,
"logps/chosen": -282.26031494140625,
"logps/rejected": -377.5342712402344,
"loss": 0.0771,
"rewards/chosen": 2.6503818035125732,
"rewards/margins": 14.911274909973145,
"rewards/rejected": -12.260892868041992,
"step": 1215
},
{
"epoch": 0.78,
"grad_norm": 20.115049362182617,
"kl": 0.0,
"learning_rate": 1.2197724039829303e-07,
"logps/chosen": -234.886962890625,
"logps/rejected": -347.6160583496094,
"loss": 0.0879,
"rewards/chosen": 2.644916296005249,
"rewards/margins": 14.264989852905273,
"rewards/rejected": -11.620074272155762,
"step": 1220
},
{
"epoch": 0.78,
"grad_norm": 17.797800064086914,
"kl": 0.0,
"learning_rate": 1.2019914651493598e-07,
"logps/chosen": -246.7455291748047,
"logps/rejected": -392.910400390625,
"loss": 0.0791,
"rewards/chosen": 2.706545352935791,
"rewards/margins": 14.843210220336914,
"rewards/rejected": -12.136663436889648,
"step": 1225
},
{
"epoch": 0.79,
"grad_norm": 12.662943840026855,
"kl": 0.0,
"learning_rate": 1.1842105263157894e-07,
"logps/chosen": -225.7235870361328,
"logps/rejected": -402.0716247558594,
"loss": 0.0925,
"rewards/chosen": 2.4518020153045654,
"rewards/margins": 14.207118034362793,
"rewards/rejected": -11.755315780639648,
"step": 1230
},
{
"epoch": 0.79,
"grad_norm": 16.056779861450195,
"kl": 0.0,
"learning_rate": 1.166429587482219e-07,
"logps/chosen": -215.66708374023438,
"logps/rejected": -375.41717529296875,
"loss": 0.0844,
"rewards/chosen": 2.730950355529785,
"rewards/margins": 14.23332691192627,
"rewards/rejected": -11.502375602722168,
"step": 1235
},
{
"epoch": 0.79,
"grad_norm": 16.644628524780273,
"kl": 0.0,
"learning_rate": 1.1486486486486487e-07,
"logps/chosen": -266.43878173828125,
"logps/rejected": -358.8226318359375,
"loss": 0.0835,
"rewards/chosen": 2.870779514312744,
"rewards/margins": 14.435707092285156,
"rewards/rejected": -11.56492805480957,
"step": 1240
},
{
"epoch": 0.8,
"grad_norm": 8.560157775878906,
"kl": 0.0,
"learning_rate": 1.1308677098150782e-07,
"logps/chosen": -238.55319213867188,
"logps/rejected": -390.98358154296875,
"loss": 0.0804,
"rewards/chosen": 2.759446620941162,
"rewards/margins": 14.843562126159668,
"rewards/rejected": -12.084115028381348,
"step": 1245
},
{
"epoch": 0.8,
"grad_norm": 16.91299819946289,
"kl": 0.0,
"learning_rate": 1.1130867709815078e-07,
"logps/chosen": -234.5216064453125,
"logps/rejected": -338.272705078125,
"loss": 0.093,
"rewards/chosen": 2.591742992401123,
"rewards/margins": 14.04884147644043,
"rewards/rejected": -11.457098007202148,
"step": 1250
},
{
"epoch": 0.8,
"grad_norm": 13.1578369140625,
"kl": 0.0,
"learning_rate": 1.0953058321479374e-07,
"logps/chosen": -275.4759826660156,
"logps/rejected": -379.9882507324219,
"loss": 0.0692,
"rewards/chosen": 2.984293222427368,
"rewards/margins": 15.574743270874023,
"rewards/rejected": -12.59045124053955,
"step": 1255
},
{
"epoch": 0.81,
"grad_norm": 17.815027236938477,
"kl": 0.0,
"learning_rate": 1.077524893314367e-07,
"logps/chosen": -267.3926696777344,
"logps/rejected": -353.8580017089844,
"loss": 0.0725,
"rewards/chosen": 2.5867929458618164,
"rewards/margins": 14.201835632324219,
"rewards/rejected": -11.615041732788086,
"step": 1260
},
{
"epoch": 0.81,
"grad_norm": 11.758225440979004,
"kl": 0.0,
"learning_rate": 1.0597439544807964e-07,
"logps/chosen": -249.6155242919922,
"logps/rejected": -364.9371337890625,
"loss": 0.0769,
"rewards/chosen": 2.7065796852111816,
"rewards/margins": 14.05157470703125,
"rewards/rejected": -11.344995498657227,
"step": 1265
},
{
"epoch": 0.81,
"grad_norm": 16.8466739654541,
"kl": 0.0,
"learning_rate": 1.0419630156472262e-07,
"logps/chosen": -238.05673217773438,
"logps/rejected": -373.03546142578125,
"loss": 0.0748,
"rewards/chosen": 2.9012789726257324,
"rewards/margins": 14.46104621887207,
"rewards/rejected": -11.55976676940918,
"step": 1270
},
{
"epoch": 0.82,
"grad_norm": 14.22836685180664,
"kl": 0.0,
"learning_rate": 1.0241820768136557e-07,
"logps/chosen": -224.74887084960938,
"logps/rejected": -332.44415283203125,
"loss": 0.084,
"rewards/chosen": 2.65814208984375,
"rewards/margins": 13.293182373046875,
"rewards/rejected": -10.635040283203125,
"step": 1275
},
{
"epoch": 0.82,
"grad_norm": 11.591891288757324,
"kl": 0.0,
"learning_rate": 1.0064011379800854e-07,
"logps/chosen": -280.44427490234375,
"logps/rejected": -365.83392333984375,
"loss": 0.0801,
"rewards/chosen": 2.5924744606018066,
"rewards/margins": 14.31025505065918,
"rewards/rejected": -11.717779159545898,
"step": 1280
},
{
"epoch": 0.82,
"grad_norm": 14.504942893981934,
"kl": 0.0,
"learning_rate": 9.886201991465149e-08,
"logps/chosen": -277.6451416015625,
"logps/rejected": -350.4718322753906,
"loss": 0.0661,
"rewards/chosen": 2.9364330768585205,
"rewards/margins": 14.681096076965332,
"rewards/rejected": -11.744662284851074,
"step": 1285
},
{
"epoch": 0.83,
"grad_norm": 16.065649032592773,
"kl": 0.0,
"learning_rate": 9.708392603129445e-08,
"logps/chosen": -262.8613586425781,
"logps/rejected": -391.71649169921875,
"loss": 0.0723,
"rewards/chosen": 2.7925260066986084,
"rewards/margins": 15.433810234069824,
"rewards/rejected": -12.641283988952637,
"step": 1290
},
{
"epoch": 0.83,
"grad_norm": 21.038543701171875,
"kl": 0.0,
"learning_rate": 9.530583214793741e-08,
"logps/chosen": -262.80584716796875,
"logps/rejected": -380.09625244140625,
"loss": 0.0801,
"rewards/chosen": 2.761460781097412,
"rewards/margins": 15.412538528442383,
"rewards/rejected": -12.651077270507812,
"step": 1295
},
{
"epoch": 0.83,
"grad_norm": 18.872102737426758,
"kl": 0.0,
"learning_rate": 9.352773826458037e-08,
"logps/chosen": -285.86712646484375,
"logps/rejected": -342.8421325683594,
"loss": 0.0819,
"rewards/chosen": 2.6993701457977295,
"rewards/margins": 14.52336311340332,
"rewards/rejected": -11.823991775512695,
"step": 1300
},
{
"epoch": 0.83,
"grad_norm": 15.029848098754883,
"kl": 0.0,
"learning_rate": 9.174964438122331e-08,
"logps/chosen": -245.3412322998047,
"logps/rejected": -356.58172607421875,
"loss": 0.094,
"rewards/chosen": 2.818472385406494,
"rewards/margins": 14.51873779296875,
"rewards/rejected": -11.700265884399414,
"step": 1305
},
{
"epoch": 0.84,
"grad_norm": 17.082958221435547,
"kl": 0.0,
"learning_rate": 8.997155049786629e-08,
"logps/chosen": -286.61383056640625,
"logps/rejected": -395.6304626464844,
"loss": 0.0716,
"rewards/chosen": 2.8008289337158203,
"rewards/margins": 15.597787857055664,
"rewards/rejected": -12.796960830688477,
"step": 1310
},
{
"epoch": 0.84,
"grad_norm": 22.768911361694336,
"kl": 0.0,
"learning_rate": 8.819345661450925e-08,
"logps/chosen": -250.6887664794922,
"logps/rejected": -358.8474426269531,
"loss": 0.0929,
"rewards/chosen": 2.7452361583709717,
"rewards/margins": 14.326945304870605,
"rewards/rejected": -11.581708908081055,
"step": 1315
},
{
"epoch": 0.84,
"grad_norm": 21.478580474853516,
"kl": 0.0,
"learning_rate": 8.64153627311522e-08,
"logps/chosen": -267.79974365234375,
"logps/rejected": -369.04815673828125,
"loss": 0.0838,
"rewards/chosen": 2.744828224182129,
"rewards/margins": 14.584070205688477,
"rewards/rejected": -11.839241027832031,
"step": 1320
},
{
"epoch": 0.85,
"grad_norm": 12.477672576904297,
"kl": 0.0,
"learning_rate": 8.463726884779517e-08,
"logps/chosen": -239.75399780273438,
"logps/rejected": -383.63812255859375,
"loss": 0.0833,
"rewards/chosen": 2.6762642860412598,
"rewards/margins": 14.272786140441895,
"rewards/rejected": -11.596521377563477,
"step": 1325
},
{
"epoch": 0.85,
"grad_norm": 19.415245056152344,
"kl": 0.0,
"learning_rate": 8.285917496443812e-08,
"logps/chosen": -258.97406005859375,
"logps/rejected": -372.96173095703125,
"loss": 0.0874,
"rewards/chosen": 2.7613649368286133,
"rewards/margins": 14.08574390411377,
"rewards/rejected": -11.324378967285156,
"step": 1330
},
{
"epoch": 0.85,
"grad_norm": 13.463936805725098,
"kl": 0.0,
"learning_rate": 8.108108108108108e-08,
"logps/chosen": -228.6225128173828,
"logps/rejected": -344.24688720703125,
"loss": 0.0777,
"rewards/chosen": 2.942438840866089,
"rewards/margins": 14.587237358093262,
"rewards/rejected": -11.644798278808594,
"step": 1335
},
{
"epoch": 0.86,
"grad_norm": 13.836956977844238,
"kl": 0.0,
"learning_rate": 7.930298719772404e-08,
"logps/chosen": -247.45742797851562,
"logps/rejected": -371.2969055175781,
"loss": 0.0721,
"rewards/chosen": 3.017321825027466,
"rewards/margins": 14.967196464538574,
"rewards/rejected": -11.949874877929688,
"step": 1340
},
{
"epoch": 0.86,
"grad_norm": 16.551321029663086,
"kl": 0.0,
"learning_rate": 7.7524893314367e-08,
"logps/chosen": -230.99227905273438,
"logps/rejected": -384.6443786621094,
"loss": 0.0871,
"rewards/chosen": 2.6670827865600586,
"rewards/margins": 14.841870307922363,
"rewards/rejected": -12.174787521362305,
"step": 1345
},
{
"epoch": 0.86,
"grad_norm": 15.87748908996582,
"kl": 0.0,
"learning_rate": 7.574679943100994e-08,
"logps/chosen": -266.7972717285156,
"logps/rejected": -335.5400390625,
"loss": 0.0976,
"rewards/chosen": 2.702658176422119,
"rewards/margins": 13.561399459838867,
"rewards/rejected": -10.858741760253906,
"step": 1350
},
{
"epoch": 0.87,
"grad_norm": 17.662395477294922,
"kl": 0.0,
"learning_rate": 7.396870554765292e-08,
"logps/chosen": -269.71026611328125,
"logps/rejected": -383.00628662109375,
"loss": 0.0921,
"rewards/chosen": 2.5814826488494873,
"rewards/margins": 14.34886646270752,
"rewards/rejected": -11.76738452911377,
"step": 1355
},
{
"epoch": 0.87,
"grad_norm": 16.09382438659668,
"kl": 0.0,
"learning_rate": 7.219061166429587e-08,
"logps/chosen": -269.3471984863281,
"logps/rejected": -375.32208251953125,
"loss": 0.0886,
"rewards/chosen": 2.8077282905578613,
"rewards/margins": 14.510396003723145,
"rewards/rejected": -11.702667236328125,
"step": 1360
},
{
"epoch": 0.87,
"grad_norm": 13.27104377746582,
"kl": 0.0,
"learning_rate": 7.041251778093883e-08,
"logps/chosen": -273.42950439453125,
"logps/rejected": -371.14837646484375,
"loss": 0.0685,
"rewards/chosen": 2.944091558456421,
"rewards/margins": 15.448992729187012,
"rewards/rejected": -12.504900932312012,
"step": 1365
},
{
"epoch": 0.88,
"grad_norm": 12.18476390838623,
"kl": 0.0,
"learning_rate": 6.863442389758179e-08,
"logps/chosen": -256.47161865234375,
"logps/rejected": -384.1866760253906,
"loss": 0.0742,
"rewards/chosen": 2.7861623764038086,
"rewards/margins": 15.34550952911377,
"rewards/rejected": -12.559347152709961,
"step": 1370
},
{
"epoch": 0.88,
"grad_norm": 23.00941276550293,
"kl": 0.0,
"learning_rate": 6.685633001422475e-08,
"logps/chosen": -242.3636474609375,
"logps/rejected": -404.00848388671875,
"loss": 0.0747,
"rewards/chosen": 2.5074126720428467,
"rewards/margins": 14.919418334960938,
"rewards/rejected": -12.412006378173828,
"step": 1375
},
{
"epoch": 0.88,
"grad_norm": 10.992572784423828,
"kl": 0.0,
"learning_rate": 6.507823613086771e-08,
"logps/chosen": -248.75967407226562,
"logps/rejected": -368.59063720703125,
"loss": 0.0867,
"rewards/chosen": 2.401717185974121,
"rewards/margins": 14.726252555847168,
"rewards/rejected": -12.32453441619873,
"step": 1380
},
{
"epoch": 0.89,
"grad_norm": 11.035951614379883,
"kl": 0.0,
"learning_rate": 6.330014224751067e-08,
"logps/chosen": -240.3590850830078,
"logps/rejected": -347.6131896972656,
"loss": 0.0921,
"rewards/chosen": 2.680974245071411,
"rewards/margins": 14.051698684692383,
"rewards/rejected": -11.370722770690918,
"step": 1385
},
{
"epoch": 0.89,
"grad_norm": 21.619415283203125,
"kl": 0.0,
"learning_rate": 6.152204836415363e-08,
"logps/chosen": -240.7654266357422,
"logps/rejected": -381.09307861328125,
"loss": 0.087,
"rewards/chosen": 2.646531105041504,
"rewards/margins": 14.633687019348145,
"rewards/rejected": -11.987154960632324,
"step": 1390
},
{
"epoch": 0.89,
"grad_norm": 20.996952056884766,
"kl": 0.0,
"learning_rate": 5.974395448079659e-08,
"logps/chosen": -228.12826538085938,
"logps/rejected": -373.8050231933594,
"loss": 0.0903,
"rewards/chosen": 2.5932085514068604,
"rewards/margins": 14.422918319702148,
"rewards/rejected": -11.829710960388184,
"step": 1395
},
{
"epoch": 0.9,
"grad_norm": 12.941699981689453,
"kl": 0.0,
"learning_rate": 5.796586059743954e-08,
"logps/chosen": -266.34759521484375,
"logps/rejected": -377.67572021484375,
"loss": 0.0733,
"rewards/chosen": 2.6969380378723145,
"rewards/margins": 14.922724723815918,
"rewards/rejected": -12.225786209106445,
"step": 1400
},
{
"epoch": 0.9,
"grad_norm": 18.563539505004883,
"kl": 0.0,
"learning_rate": 5.61877667140825e-08,
"logps/chosen": -233.05337524414062,
"logps/rejected": -351.65478515625,
"loss": 0.0819,
"rewards/chosen": 2.4380977153778076,
"rewards/margins": 14.38970947265625,
"rewards/rejected": -11.951611518859863,
"step": 1405
},
{
"epoch": 0.9,
"grad_norm": 14.77774715423584,
"kl": 0.0,
"learning_rate": 5.4409672830725456e-08,
"logps/chosen": -272.6517028808594,
"logps/rejected": -359.89013671875,
"loss": 0.0762,
"rewards/chosen": 2.888676404953003,
"rewards/margins": 14.460576057434082,
"rewards/rejected": -11.571900367736816,
"step": 1410
},
{
"epoch": 0.91,
"grad_norm": 15.590229988098145,
"kl": 0.0,
"learning_rate": 5.2631578947368416e-08,
"logps/chosen": -256.3451843261719,
"logps/rejected": -375.3144226074219,
"loss": 0.0762,
"rewards/chosen": 2.667102336883545,
"rewards/margins": 15.10230827331543,
"rewards/rejected": -12.435206413269043,
"step": 1415
},
{
"epoch": 0.91,
"grad_norm": 10.8274507522583,
"kl": 0.0,
"learning_rate": 5.0853485064011376e-08,
"logps/chosen": -268.83636474609375,
"logps/rejected": -372.27838134765625,
"loss": 0.0648,
"rewards/chosen": 2.9176650047302246,
"rewards/margins": 14.840046882629395,
"rewards/rejected": -11.922381401062012,
"step": 1420
},
{
"epoch": 0.91,
"grad_norm": 15.774340629577637,
"kl": 0.0,
"learning_rate": 4.9075391180654337e-08,
"logps/chosen": -260.6099853515625,
"logps/rejected": -358.04241943359375,
"loss": 0.0585,
"rewards/chosen": 2.9233345985412598,
"rewards/margins": 14.653867721557617,
"rewards/rejected": -11.730535507202148,
"step": 1425
},
{
"epoch": 0.91,
"grad_norm": 15.554561614990234,
"kl": 0.0,
"learning_rate": 4.72972972972973e-08,
"logps/chosen": -238.23739624023438,
"logps/rejected": -386.2373962402344,
"loss": 0.0833,
"rewards/chosen": 2.658268451690674,
"rewards/margins": 14.15376091003418,
"rewards/rejected": -11.49549388885498,
"step": 1430
},
{
"epoch": 0.92,
"grad_norm": 17.29914093017578,
"kl": 0.0,
"learning_rate": 4.551920341394026e-08,
"logps/chosen": -248.63204956054688,
"logps/rejected": -371.215087890625,
"loss": 0.0872,
"rewards/chosen": 2.9267373085021973,
"rewards/margins": 14.192481994628906,
"rewards/rejected": -11.265745162963867,
"step": 1435
},
{
"epoch": 0.92,
"grad_norm": 14.301885604858398,
"kl": 0.0,
"learning_rate": 4.374110953058322e-08,
"logps/chosen": -240.0107421875,
"logps/rejected": -355.2042541503906,
"loss": 0.0846,
"rewards/chosen": 2.4140007495880127,
"rewards/margins": 13.864062309265137,
"rewards/rejected": -11.450060844421387,
"step": 1440
},
{
"epoch": 0.92,
"grad_norm": 13.32541275024414,
"kl": 0.0,
"learning_rate": 4.196301564722617e-08,
"logps/chosen": -247.22152709960938,
"logps/rejected": -360.07110595703125,
"loss": 0.0659,
"rewards/chosen": 2.7749335765838623,
"rewards/margins": 14.867749214172363,
"rewards/rejected": -12.092815399169922,
"step": 1445
},
{
"epoch": 0.93,
"grad_norm": 12.356550216674805,
"kl": 0.0,
"learning_rate": 4.018492176386913e-08,
"logps/chosen": -256.69683837890625,
"logps/rejected": -363.08489990234375,
"loss": 0.0799,
"rewards/chosen": 2.8861594200134277,
"rewards/margins": 14.505203247070312,
"rewards/rejected": -11.619044303894043,
"step": 1450
},
{
"epoch": 0.93,
"grad_norm": 14.969185829162598,
"kl": 0.0,
"learning_rate": 3.840682788051209e-08,
"logps/chosen": -254.7665252685547,
"logps/rejected": -354.30841064453125,
"loss": 0.0757,
"rewards/chosen": 2.8421378135681152,
"rewards/margins": 14.373845100402832,
"rewards/rejected": -11.531707763671875,
"step": 1455
},
{
"epoch": 0.93,
"grad_norm": 17.030691146850586,
"kl": 0.0,
"learning_rate": 3.6628733997155046e-08,
"logps/chosen": -263.8084411621094,
"logps/rejected": -395.4825439453125,
"loss": 0.0759,
"rewards/chosen": 2.9121432304382324,
"rewards/margins": 15.505640983581543,
"rewards/rejected": -12.593496322631836,
"step": 1460
},
{
"epoch": 0.94,
"grad_norm": 12.211559295654297,
"kl": 0.0,
"learning_rate": 3.4850640113798006e-08,
"logps/chosen": -236.26431274414062,
"logps/rejected": -357.1573791503906,
"loss": 0.0755,
"rewards/chosen": 2.843881130218506,
"rewards/margins": 14.507527351379395,
"rewards/rejected": -11.66364574432373,
"step": 1465
},
{
"epoch": 0.94,
"grad_norm": 17.574806213378906,
"kl": 0.0,
"learning_rate": 3.3072546230440967e-08,
"logps/chosen": -285.59942626953125,
"logps/rejected": -368.08245849609375,
"loss": 0.0676,
"rewards/chosen": 2.877531051635742,
"rewards/margins": 14.671765327453613,
"rewards/rejected": -11.794233322143555,
"step": 1470
},
{
"epoch": 0.94,
"grad_norm": 13.167614936828613,
"kl": 0.0,
"learning_rate": 3.129445234708392e-08,
"logps/chosen": -240.7215576171875,
"logps/rejected": -382.0995788574219,
"loss": 0.0651,
"rewards/chosen": 2.778978109359741,
"rewards/margins": 14.833264350891113,
"rewards/rejected": -12.054286003112793,
"step": 1475
},
{
"epoch": 0.95,
"grad_norm": 19.462112426757812,
"kl": 0.0,
"learning_rate": 2.9516358463726884e-08,
"logps/chosen": -243.0684356689453,
"logps/rejected": -366.4331359863281,
"loss": 0.0802,
"rewards/chosen": 2.8256452083587646,
"rewards/margins": 15.236946105957031,
"rewards/rejected": -12.411300659179688,
"step": 1480
},
{
"epoch": 0.95,
"grad_norm": 20.093833923339844,
"kl": 0.0,
"learning_rate": 2.7738264580369844e-08,
"logps/chosen": -209.17770385742188,
"logps/rejected": -359.9656677246094,
"loss": 0.0839,
"rewards/chosen": 2.901711940765381,
"rewards/margins": 14.679159164428711,
"rewards/rejected": -11.777449607849121,
"step": 1485
},
{
"epoch": 0.95,
"grad_norm": 13.270203590393066,
"kl": 0.0,
"learning_rate": 2.59601706970128e-08,
"logps/chosen": -272.44610595703125,
"logps/rejected": -357.05975341796875,
"loss": 0.0868,
"rewards/chosen": 2.6972391605377197,
"rewards/margins": 15.144004821777344,
"rewards/rejected": -12.446764945983887,
"step": 1490
},
{
"epoch": 0.96,
"grad_norm": 14.922286987304688,
"kl": 0.0,
"learning_rate": 2.418207681365576e-08,
"logps/chosen": -238.18191528320312,
"logps/rejected": -367.4979553222656,
"loss": 0.0824,
"rewards/chosen": 2.5710983276367188,
"rewards/margins": 14.130340576171875,
"rewards/rejected": -11.559242248535156,
"step": 1495
},
{
"epoch": 0.96,
"grad_norm": 21.186229705810547,
"kl": 0.0,
"learning_rate": 2.240398293029872e-08,
"logps/chosen": -240.6565399169922,
"logps/rejected": -387.65936279296875,
"loss": 0.0783,
"rewards/chosen": 2.6365768909454346,
"rewards/margins": 15.131543159484863,
"rewards/rejected": -12.494966506958008,
"step": 1500
},
{
"epoch": 0.96,
"grad_norm": 14.05397891998291,
"kl": 0.0,
"learning_rate": 2.0625889046941676e-08,
"logps/chosen": -307.04583740234375,
"logps/rejected": -361.1058654785156,
"loss": 0.0903,
"rewards/chosen": 2.804368257522583,
"rewards/margins": 14.683168411254883,
"rewards/rejected": -11.878799438476562,
"step": 1505
},
{
"epoch": 0.97,
"grad_norm": 10.290091514587402,
"kl": 0.0,
"learning_rate": 1.8847795163584636e-08,
"logps/chosen": -280.72406005859375,
"logps/rejected": -393.7200622558594,
"loss": 0.0643,
"rewards/chosen": 2.76891827583313,
"rewards/margins": 15.010998725891113,
"rewards/rejected": -12.24207878112793,
"step": 1510
},
{
"epoch": 0.97,
"grad_norm": 13.187542915344238,
"kl": 0.0,
"learning_rate": 1.7069701280227596e-08,
"logps/chosen": -250.7726287841797,
"logps/rejected": -378.99993896484375,
"loss": 0.0697,
"rewards/chosen": 2.8748562335968018,
"rewards/margins": 15.209932327270508,
"rewards/rejected": -12.335077285766602,
"step": 1515
},
{
"epoch": 0.97,
"grad_norm": 13.64420223236084,
"kl": 0.0,
"learning_rate": 1.5291607396870554e-08,
"logps/chosen": -253.7127227783203,
"logps/rejected": -367.41558837890625,
"loss": 0.0824,
"rewards/chosen": 3.031802177429199,
"rewards/margins": 14.762832641601562,
"rewards/rejected": -11.73102855682373,
"step": 1520
},
{
"epoch": 0.98,
"grad_norm": 10.866116523742676,
"kl": 0.0,
"learning_rate": 1.3513513513513514e-08,
"logps/chosen": -250.94400024414062,
"logps/rejected": -381.5174865722656,
"loss": 0.0793,
"rewards/chosen": 2.8337435722351074,
"rewards/margins": 15.124834060668945,
"rewards/rejected": -12.29109001159668,
"step": 1525
},
{
"epoch": 0.98,
"grad_norm": 11.604872703552246,
"kl": 0.0,
"learning_rate": 1.1735419630156473e-08,
"logps/chosen": -233.5682830810547,
"logps/rejected": -377.261962890625,
"loss": 0.0687,
"rewards/chosen": 3.119915008544922,
"rewards/margins": 15.134057998657227,
"rewards/rejected": -12.014142990112305,
"step": 1530
},
{
"epoch": 0.98,
"grad_norm": 20.006731033325195,
"kl": 0.0,
"learning_rate": 9.95732574679943e-09,
"logps/chosen": -257.07525634765625,
"logps/rejected": -346.5194396972656,
"loss": 0.0907,
"rewards/chosen": 2.676340103149414,
"rewards/margins": 14.137725830078125,
"rewards/rejected": -11.461384773254395,
"step": 1535
},
{
"epoch": 0.99,
"grad_norm": 17.203874588012695,
"kl": 0.0,
"learning_rate": 8.179231863442388e-09,
"logps/chosen": -226.61843872070312,
"logps/rejected": -381.11029052734375,
"loss": 0.0683,
"rewards/chosen": 2.940342903137207,
"rewards/margins": 15.23707103729248,
"rewards/rejected": -12.296728134155273,
"step": 1540
},
{
"epoch": 0.99,
"grad_norm": 14.477093696594238,
"kl": 0.0,
"learning_rate": 6.401137980085348e-09,
"logps/chosen": -240.14157104492188,
"logps/rejected": -388.9137268066406,
"loss": 0.0658,
"rewards/chosen": 2.875831127166748,
"rewards/margins": 15.294588088989258,
"rewards/rejected": -12.418756484985352,
"step": 1545
},
{
"epoch": 0.99,
"grad_norm": 10.444808959960938,
"kl": 0.0,
"learning_rate": 4.623044096728307e-09,
"logps/chosen": -273.6938781738281,
"logps/rejected": -361.58782958984375,
"loss": 0.0592,
"rewards/chosen": 3.0903258323669434,
"rewards/margins": 15.089597702026367,
"rewards/rejected": -11.999273300170898,
"step": 1550
},
{
"epoch": 0.99,
"grad_norm": 13.01125717163086,
"kl": 0.0,
"learning_rate": 2.844950213371266e-09,
"logps/chosen": -263.2748107910156,
"logps/rejected": -382.49481201171875,
"loss": 0.0724,
"rewards/chosen": 2.801457166671753,
"rewards/margins": 14.89887809753418,
"rewards/rejected": -12.097421646118164,
"step": 1555
},
{
"epoch": 1.0,
"grad_norm": 13.397442817687988,
"kl": 0.0,
"learning_rate": 1.0668563300142248e-09,
"logps/chosen": -225.07894897460938,
"logps/rejected": -353.08856201171875,
"loss": 0.0824,
"rewards/chosen": 2.47590970993042,
"rewards/margins": 14.8174409866333,
"rewards/rejected": -12.341531753540039,
"step": 1560
},
{
"epoch": 1.0,
"step": 1563,
"total_flos": 0.0,
"train_loss": 0.11715243643320149,
"train_runtime": 11089.6574,
"train_samples_per_second": 9.017,
"train_steps_per_second": 0.141
}
],
"logging_steps": 5,
"max_steps": 1563,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}