heuristic_snyder / checkpoint-753 /trainer_state.json
tomekkorbak's picture
Training in progress, step 753
ee02659
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.31879762912785775,
"global_step": 753,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"learning_rate": 4.1666666666666665e-05,
"loss": 3.0643,
"theoretical_loss": 3.321567680436603,
"tokens_seen": 2990538752
},
{
"epoch": 0.0,
"learning_rate": 8.333333333333333e-05,
"loss": 3.0798,
"theoretical_loss": 3.3215564803546,
"tokens_seen": 2990669824
},
{
"epoch": 0.0,
"learning_rate": 0.000125,
"loss": 2.8688,
"theoretical_loss": 3.321545280900887,
"tokens_seen": 2990800896
},
{
"epoch": 0.0,
"learning_rate": 0.00016666666666666666,
"loss": 2.7194,
"theoretical_loss": 3.3215340820754022,
"tokens_seen": 2990931968
},
{
"epoch": 0.0,
"learning_rate": 0.00020833333333333335,
"loss": 2.6193,
"theoretical_loss": 3.3215228838780817,
"tokens_seen": 2991063040
},
{
"epoch": 0.0,
"learning_rate": 0.00025,
"loss": 2.8571,
"theoretical_loss": 3.3215116863088636,
"tokens_seen": 2991194112
},
{
"epoch": 0.0,
"learning_rate": 0.0002916666666666667,
"loss": 2.7571,
"theoretical_loss": 3.3215004893676854,
"tokens_seen": 2991325184
},
{
"epoch": 0.0,
"learning_rate": 0.0003333333333333333,
"loss": 2.8877,
"theoretical_loss": 3.321489293054483,
"tokens_seen": 2991456256
},
{
"epoch": 0.0,
"learning_rate": 0.000375,
"loss": 2.714,
"theoretical_loss": 3.321478097369195,
"tokens_seen": 2991587328
},
{
"epoch": 0.0,
"learning_rate": 0.0004166666666666667,
"loss": 2.6564,
"theoretical_loss": 3.321466902311758,
"tokens_seen": 2991718400
},
{
"epoch": 0.0,
"learning_rate": 0.0004583333333333333,
"loss": 2.5638,
"theoretical_loss": 3.3214557078821096,
"tokens_seen": 2991849472
},
{
"epoch": 0.01,
"learning_rate": 0.0005,
"loss": 2.4354,
"theoretical_loss": 3.321444514080187,
"tokens_seen": 2991980544
},
{
"epoch": 0.01,
"objective/train/docs_used": 1640856,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5749809741973877,
"objective/train/theoretical_loss": 3.321438917414603,
"objective/train/tokens_used": 22097376,
"theoretical_loss": 3.321438917414603,
"tokens_seen": 2992046080
},
{
"epoch": 0.01,
"learning_rate": 0.0005416666666666666,
"loss": 2.5713,
"theoretical_loss": 3.321433320905927,
"tokens_seen": 2992111616
},
{
"epoch": 0.01,
"learning_rate": 0.0005833333333333334,
"loss": 2.4812,
"theoretical_loss": 3.3214221283592678,
"tokens_seen": 2992242688
},
{
"epoch": 0.01,
"learning_rate": 0.000625,
"loss": 2.7622,
"theoretical_loss": 3.321410936440146,
"tokens_seen": 2992373760
},
{
"epoch": 0.01,
"learning_rate": 0.0006666666666666666,
"loss": 2.609,
"theoretical_loss": 3.3213997451485,
"tokens_seen": 2992504832
},
{
"epoch": 0.01,
"learning_rate": 0.0007083333333333334,
"loss": 2.4537,
"theoretical_loss": 3.3213885544842654,
"tokens_seen": 2992635904
},
{
"epoch": 0.01,
"learning_rate": 0.00075,
"loss": 2.4831,
"theoretical_loss": 3.321377364447381,
"tokens_seen": 2992766976
},
{
"epoch": 0.01,
"learning_rate": 0.0007916666666666666,
"loss": 2.5607,
"theoretical_loss": 3.3213661750377836,
"tokens_seen": 2992898048
},
{
"epoch": 0.01,
"learning_rate": 0.0008333333333333334,
"loss": 2.5759,
"theoretical_loss": 3.3213549862554106,
"tokens_seen": 2993029120
},
{
"epoch": 0.01,
"learning_rate": 0.000875,
"loss": 2.3106,
"theoretical_loss": 3.3213437981001994,
"tokens_seen": 2993160192
},
{
"epoch": 0.01,
"learning_rate": 0.0009166666666666666,
"loss": 2.5471,
"theoretical_loss": 3.3213326105720875,
"tokens_seen": 2993291264
},
{
"epoch": 0.01,
"learning_rate": 0.0009583333333333334,
"loss": 2.602,
"theoretical_loss": 3.3213214236710122,
"tokens_seen": 2993422336
},
{
"epoch": 0.01,
"learning_rate": 0.001,
"loss": 2.5877,
"theoretical_loss": 3.321310237396911,
"tokens_seen": 2993553408
},
{
"epoch": 0.01,
"objective/train/docs_used": 1641461,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.557373523712158,
"objective/train/theoretical_loss": 3.3212990517497207,
"objective/train/tokens_used": 23735776,
"theoretical_loss": 3.3212990517497207,
"tokens_seen": 2993684480
},
{
"epoch": 0.01,
"learning_rate": 0.0009995722840034217,
"loss": 2.7042,
"theoretical_loss": 3.3212990517497207,
"tokens_seen": 2993684480
},
{
"epoch": 0.01,
"learning_rate": 0.0009991445680068436,
"loss": 2.5234,
"theoretical_loss": 3.3212878667293797,
"tokens_seen": 2993815552
},
{
"epoch": 0.01,
"learning_rate": 0.0009987168520102653,
"loss": 2.5502,
"theoretical_loss": 3.321276682335825,
"tokens_seen": 2993946624
},
{
"epoch": 0.01,
"learning_rate": 0.000998289136013687,
"loss": 2.6849,
"theoretical_loss": 3.3212654985689936,
"tokens_seen": 2994077696
},
{
"epoch": 0.01,
"learning_rate": 0.0009978614200171086,
"loss": 2.6348,
"theoretical_loss": 3.3212543154288237,
"tokens_seen": 2994208768
},
{
"epoch": 0.01,
"learning_rate": 0.0009974337040205303,
"loss": 2.6793,
"theoretical_loss": 3.3212431329152525,
"tokens_seen": 2994339840
},
{
"epoch": 0.01,
"learning_rate": 0.0009970059880239522,
"loss": 2.6212,
"theoretical_loss": 3.321231951028217,
"tokens_seen": 2994470912
},
{
"epoch": 0.01,
"learning_rate": 0.0009965782720273739,
"loss": 2.629,
"theoretical_loss": 3.3212207697676552,
"tokens_seen": 2994601984
},
{
"epoch": 0.01,
"learning_rate": 0.0009961505560307955,
"loss": 2.5865,
"theoretical_loss": 3.3212095891335043,
"tokens_seen": 2994733056
},
{
"epoch": 0.01,
"learning_rate": 0.0009957228400342174,
"loss": 2.667,
"theoretical_loss": 3.321198409125702,
"tokens_seen": 2994864128
},
{
"epoch": 0.01,
"learning_rate": 0.000995295124037639,
"loss": 2.55,
"theoretical_loss": 3.321187229744186,
"tokens_seen": 2994995200
},
{
"epoch": 0.02,
"learning_rate": 0.0009948674080410608,
"loss": 2.6324,
"theoretical_loss": 3.321176050988893,
"tokens_seen": 2995126272
},
{
"epoch": 0.02,
"learning_rate": 0.0009944396920444824,
"loss": 2.8406,
"theoretical_loss": 3.3211648728597614,
"tokens_seen": 2995257344
},
{
"epoch": 0.02,
"objective/train/docs_used": 1642666,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.8021185398101807,
"objective/train/theoretical_loss": 3.3211592840299864,
"objective/train/tokens_used": 25374176,
"theoretical_loss": 3.3211592840299864,
"tokens_seen": 2995322880
},
{
"epoch": 0.02,
"learning_rate": 0.0009940119760479041,
"loss": 2.5438,
"theoretical_loss": 3.3211536953567284,
"tokens_seen": 2995388416
},
{
"epoch": 0.02,
"learning_rate": 0.000993584260051326,
"loss": 2.7618,
"theoretical_loss": 3.321142518479731,
"tokens_seen": 2995519488
},
{
"epoch": 0.02,
"learning_rate": 0.0009931565440547477,
"loss": 2.6638,
"theoretical_loss": 3.321131342228708,
"tokens_seen": 2995650560
},
{
"epoch": 0.02,
"learning_rate": 0.0009927288280581694,
"loss": 2.7287,
"theoretical_loss": 3.321120166603596,
"tokens_seen": 2995781632
},
{
"epoch": 0.02,
"learning_rate": 0.000992301112061591,
"loss": 2.5832,
"theoretical_loss": 3.3211089916043326,
"tokens_seen": 2995912704
},
{
"epoch": 0.02,
"learning_rate": 0.0009918733960650127,
"loss": 2.6355,
"theoretical_loss": 3.3210978172308554,
"tokens_seen": 2996043776
},
{
"epoch": 0.02,
"learning_rate": 0.0009914456800684346,
"loss": 2.5139,
"theoretical_loss": 3.3210866434831026,
"tokens_seen": 2996174848
},
{
"epoch": 0.02,
"learning_rate": 0.0009910179640718563,
"loss": 2.5879,
"theoretical_loss": 3.3210754703610106,
"tokens_seen": 2996305920
},
{
"epoch": 0.02,
"learning_rate": 0.0009905902480752782,
"loss": 2.5935,
"theoretical_loss": 3.321064297864518,
"tokens_seen": 2996436992
},
{
"epoch": 0.02,
"learning_rate": 0.0009901625320786998,
"loss": 2.6176,
"theoretical_loss": 3.3210531259935627,
"tokens_seen": 2996568064
},
{
"epoch": 0.02,
"learning_rate": 0.0009897348160821215,
"loss": 2.6405,
"theoretical_loss": 3.321041954748081,
"tokens_seen": 2996699136
},
{
"epoch": 0.02,
"learning_rate": 0.0009893071000855432,
"loss": 2.7274,
"theoretical_loss": 3.321030784128012,
"tokens_seen": 2996830208
},
{
"epoch": 0.02,
"objective/train/docs_used": 1643300,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 3.023165464401245,
"objective/train/theoretical_loss": 3.321019614133292,
"objective/train/tokens_used": 27012576,
"theoretical_loss": 3.321019614133292,
"tokens_seen": 2996961280
},
{
"epoch": 0.02,
"learning_rate": 0.0009888793840889649,
"loss": 2.7003,
"theoretical_loss": 3.321019614133292,
"tokens_seen": 2996961280
},
{
"epoch": 0.02,
"learning_rate": 0.0009884516680923865,
"loss": 2.7436,
"theoretical_loss": 3.3210084447638595,
"tokens_seen": 2997092352
},
{
"epoch": 0.02,
"learning_rate": 0.0009880239520958084,
"loss": 2.5873,
"theoretical_loss": 3.320997276019652,
"tokens_seen": 2997223424
},
{
"epoch": 0.02,
"learning_rate": 0.00098759623609923,
"loss": 2.5974,
"theoretical_loss": 3.3209861079006067,
"tokens_seen": 2997354496
},
{
"epoch": 0.02,
"learning_rate": 0.000987168520102652,
"loss": 2.5806,
"theoretical_loss": 3.320974940406662,
"tokens_seen": 2997485568
},
{
"epoch": 0.02,
"learning_rate": 0.0009867408041060737,
"loss": 2.6771,
"theoretical_loss": 3.320963773537755,
"tokens_seen": 2997616640
},
{
"epoch": 0.02,
"learning_rate": 0.0009863130881094953,
"loss": 2.7313,
"theoretical_loss": 3.320952607293824,
"tokens_seen": 2997747712
},
{
"epoch": 0.02,
"learning_rate": 0.000985885372112917,
"loss": 2.7302,
"theoretical_loss": 3.320941441674806,
"tokens_seen": 2997878784
},
{
"epoch": 0.02,
"learning_rate": 0.0009854576561163387,
"loss": 2.893,
"theoretical_loss": 3.320930276680639,
"tokens_seen": 2998009856
},
{
"epoch": 0.02,
"learning_rate": 0.0009850299401197606,
"loss": 2.6886,
"theoretical_loss": 3.3209191123112607,
"tokens_seen": 2998140928
},
{
"epoch": 0.03,
"learning_rate": 0.0009846022241231823,
"loss": 2.7801,
"theoretical_loss": 3.320907948566609,
"tokens_seen": 2998272000
},
{
"epoch": 0.03,
"learning_rate": 0.000984174508126604,
"loss": 2.6538,
"theoretical_loss": 3.3208967854466214,
"tokens_seen": 2998403072
},
{
"epoch": 0.03,
"learning_rate": 0.0009837467921300258,
"loss": 2.5705,
"theoretical_loss": 3.3208856229512356,
"tokens_seen": 2998534144
},
{
"epoch": 0.03,
"objective/train/docs_used": 1644380,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.542499542236328,
"objective/train/theoretical_loss": 3.320880041937749,
"objective/train/tokens_used": 28650976,
"theoretical_loss": 3.320880041937749,
"tokens_seen": 2998599680
},
{
"epoch": 0.03,
"learning_rate": 0.0009833190761334475,
"loss": 2.5074,
"theoretical_loss": 3.3208744610803898,
"tokens_seen": 2998665216
},
{
"epoch": 0.03,
"learning_rate": 0.0009828913601368692,
"loss": 2.6151,
"theoretical_loss": 3.320863299834021,
"tokens_seen": 2998796288
},
{
"epoch": 0.03,
"learning_rate": 0.0009824636441402908,
"loss": 2.6994,
"theoretical_loss": 3.320852139212068,
"tokens_seen": 2998927360
},
{
"epoch": 0.03,
"learning_rate": 0.0009820359281437125,
"loss": 2.568,
"theoretical_loss": 3.3208409792144677,
"tokens_seen": 2999058432
},
{
"epoch": 0.03,
"learning_rate": 0.0009816082121471344,
"loss": 2.5552,
"theoretical_loss": 3.320829819841158,
"tokens_seen": 2999189504
},
{
"epoch": 0.03,
"learning_rate": 0.000981180496150556,
"loss": 2.6719,
"theoretical_loss": 3.320818661092077,
"tokens_seen": 2999320576
},
{
"epoch": 0.03,
"learning_rate": 0.0009807527801539778,
"loss": 2.5567,
"theoretical_loss": 3.3208075029671624,
"tokens_seen": 2999451648
},
{
"epoch": 0.03,
"learning_rate": 0.0009803250641573994,
"loss": 2.5511,
"theoretical_loss": 3.320796345466352,
"tokens_seen": 2999582720
},
{
"epoch": 0.03,
"learning_rate": 0.0009798973481608211,
"loss": 2.6608,
"theoretical_loss": 3.320785188589584,
"tokens_seen": 2999713792
},
{
"epoch": 0.03,
"learning_rate": 0.000979469632164243,
"loss": 2.5947,
"theoretical_loss": 3.3207740323367956,
"tokens_seen": 2999844864
},
{
"epoch": 0.03,
"learning_rate": 0.0009790419161676647,
"loss": 2.6511,
"theoretical_loss": 3.3207628767079242,
"tokens_seen": 2999975936
},
{
"epoch": 0.03,
"learning_rate": 0.0009786142001710863,
"loss": 2.5573,
"theoretical_loss": 3.3207517217029094,
"tokens_seen": 3000107008
},
{
"epoch": 0.03,
"objective/train/docs_used": 1645056,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4763801097869873,
"objective/train/theoretical_loss": 3.3207405673216877,
"objective/train/tokens_used": 30289376,
"theoretical_loss": 3.3207405673216877,
"tokens_seen": 3000238080
},
{
"epoch": 0.03,
"learning_rate": 0.0009781864841745082,
"loss": 2.6745,
"theoretical_loss": 3.3207405673216877,
"tokens_seen": 3000238080
},
{
"epoch": 0.03,
"learning_rate": 0.00097775876817793,
"loss": 2.697,
"theoretical_loss": 3.320729413564197,
"tokens_seen": 3000369152
},
{
"epoch": 0.03,
"learning_rate": 0.0009773310521813516,
"loss": 2.6853,
"theoretical_loss": 3.3207182604303753,
"tokens_seen": 3000500224
},
{
"epoch": 0.03,
"learning_rate": 0.0009769033361847733,
"loss": 2.3445,
"theoretical_loss": 3.320707107920161,
"tokens_seen": 3000631296
},
{
"epoch": 0.03,
"learning_rate": 0.000976475620188195,
"loss": 2.6763,
"theoretical_loss": 3.3206959560334917,
"tokens_seen": 3000762368
},
{
"epoch": 0.03,
"learning_rate": 0.0009760479041916168,
"loss": 2.5198,
"theoretical_loss": 3.320684804770305,
"tokens_seen": 3000893440
},
{
"epoch": 0.03,
"learning_rate": 0.0009756201881950385,
"loss": 2.7,
"theoretical_loss": 3.3206736541305393,
"tokens_seen": 3001024512
},
{
"epoch": 0.03,
"learning_rate": 0.0009751924721984602,
"loss": 2.6958,
"theoretical_loss": 3.3206625041141318,
"tokens_seen": 3001155584
},
{
"epoch": 0.04,
"learning_rate": 0.000974764756201882,
"loss": 2.6457,
"theoretical_loss": 3.3206513547210212,
"tokens_seen": 3001286656
},
{
"epoch": 0.04,
"learning_rate": 0.0009743370402053036,
"loss": 2.7946,
"theoretical_loss": 3.320640205951145,
"tokens_seen": 3001417728
},
{
"epoch": 0.04,
"learning_rate": 0.0009739093242087254,
"loss": 2.6682,
"theoretical_loss": 3.3206290578044415,
"tokens_seen": 3001548800
},
{
"epoch": 0.04,
"learning_rate": 0.0009734816082121472,
"loss": 2.5484,
"theoretical_loss": 3.3206179102808484,
"tokens_seen": 3001679872
},
{
"epoch": 0.04,
"learning_rate": 0.0009730538922155689,
"loss": 2.6724,
"theoretical_loss": 3.3206067633803036,
"tokens_seen": 3001810944
},
{
"epoch": 0.04,
"objective/train/docs_used": 1646327,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.547891855239868,
"objective/train/theoretical_loss": 3.320601190163655,
"objective/train/tokens_used": 31927776,
"theoretical_loss": 3.320601190163655,
"tokens_seen": 3001876480
},
{
"epoch": 0.04,
"learning_rate": 0.0009726261762189907,
"loss": 2.5267,
"theoretical_loss": 3.320595617102745,
"tokens_seen": 3001942016
},
{
"epoch": 0.04,
"learning_rate": 0.0009721984602224123,
"loss": 2.6548,
"theoretical_loss": 3.320584471448111,
"tokens_seen": 3002073088
},
{
"epoch": 0.04,
"learning_rate": 0.0009717707442258341,
"loss": 2.5147,
"theoretical_loss": 3.3205733264163393,
"tokens_seen": 3002204160
},
{
"epoch": 0.04,
"learning_rate": 0.0009713430282292558,
"loss": 2.4505,
"theoretical_loss": 3.320562182007368,
"tokens_seen": 3002335232
},
{
"epoch": 0.04,
"learning_rate": 0.0009709153122326775,
"loss": 2.6305,
"theoretical_loss": 3.320551038221135,
"tokens_seen": 3002466304
},
{
"epoch": 0.04,
"learning_rate": 0.0009704875962360993,
"loss": 2.5482,
"theoretical_loss": 3.3205398950575784,
"tokens_seen": 3002597376
},
{
"epoch": 0.04,
"learning_rate": 0.0009700598802395209,
"loss": 2.7266,
"theoretical_loss": 3.320528752516636,
"tokens_seen": 3002728448
},
{
"epoch": 0.04,
"learning_rate": 0.0009696321642429427,
"loss": 2.5155,
"theoretical_loss": 3.3205176105982463,
"tokens_seen": 3002859520
},
{
"epoch": 0.04,
"learning_rate": 0.0009692044482463645,
"loss": 2.7628,
"theoretical_loss": 3.320506469302347,
"tokens_seen": 3002990592
},
{
"epoch": 0.04,
"learning_rate": 0.0009687767322497862,
"loss": 2.6802,
"theoretical_loss": 3.3204953286288763,
"tokens_seen": 3003121664
},
{
"epoch": 0.04,
"learning_rate": 0.000968349016253208,
"loss": 2.7921,
"theoretical_loss": 3.3204841885777725,
"tokens_seen": 3003252736
},
{
"epoch": 0.04,
"learning_rate": 0.0009679213002566296,
"loss": 2.6088,
"theoretical_loss": 3.3204730491489727,
"tokens_seen": 3003383808
},
{
"epoch": 0.04,
"objective/train/docs_used": 1647543,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5402848720550537,
"objective/train/theoretical_loss": 3.3204619103424164,
"objective/train/tokens_used": 33566176,
"theoretical_loss": 3.3204619103424164,
"tokens_seen": 3003514880
},
{
"epoch": 0.04,
"learning_rate": 0.0009674935842600513,
"loss": 2.6446,
"theoretical_loss": 3.3204619103424164,
"tokens_seen": 3003514880
},
{
"epoch": 0.04,
"learning_rate": 0.0009670658682634731,
"loss": 2.5202,
"theoretical_loss": 3.3204507721580403,
"tokens_seen": 3003645952
},
{
"epoch": 0.04,
"learning_rate": 0.0009666381522668948,
"loss": 2.4832,
"theoretical_loss": 3.3204396345957834,
"tokens_seen": 3003777024
},
{
"epoch": 0.04,
"learning_rate": 0.0009662104362703165,
"loss": 2.6544,
"theoretical_loss": 3.320428497655584,
"tokens_seen": 3003908096
},
{
"epoch": 0.04,
"learning_rate": 0.0009657827202737382,
"loss": 2.7507,
"theoretical_loss": 3.320417361337379,
"tokens_seen": 3004039168
},
{
"epoch": 0.04,
"learning_rate": 0.00096535500427716,
"loss": 2.5786,
"theoretical_loss": 3.3204062256411078,
"tokens_seen": 3004170240
},
{
"epoch": 0.04,
"learning_rate": 0.0009649272882805818,
"loss": 2.6649,
"theoretical_loss": 3.320395090566708,
"tokens_seen": 3004301312
},
{
"epoch": 0.05,
"learning_rate": 0.0009644995722840035,
"loss": 2.6338,
"theoretical_loss": 3.3203839561141173,
"tokens_seen": 3004432384
},
{
"epoch": 0.05,
"learning_rate": 0.0009640718562874252,
"loss": 2.664,
"theoretical_loss": 3.320372822283275,
"tokens_seen": 3004563456
},
{
"epoch": 0.05,
"learning_rate": 0.0009636441402908469,
"loss": 2.6909,
"theoretical_loss": 3.3203616890741183,
"tokens_seen": 3004694528
},
{
"epoch": 0.05,
"learning_rate": 0.0009632164242942686,
"loss": 2.4478,
"theoretical_loss": 3.3203505564865856,
"tokens_seen": 3004825600
},
{
"epoch": 0.05,
"learning_rate": 0.0009627887082976904,
"loss": 2.6265,
"theoretical_loss": 3.3203394245206153,
"tokens_seen": 3004956672
},
{
"epoch": 0.05,
"learning_rate": 0.000962360992301112,
"loss": 2.5755,
"theoretical_loss": 3.320328293176145,
"tokens_seen": 3005087744
},
{
"epoch": 0.05,
"objective/train/docs_used": 1648109,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5592408180236816,
"objective/train/theoretical_loss": 3.3203227277369534,
"objective/train/tokens_used": 35204576,
"theoretical_loss": 3.3203227277369534,
"tokens_seen": 3005153280
},
{
"epoch": 0.05,
"learning_rate": 0.0009619332763045337,
"loss": 2.567,
"theoretical_loss": 3.320317162453114,
"tokens_seen": 3005218816
},
{
"epoch": 0.05,
"learning_rate": 0.0009615055603079555,
"loss": 2.6035,
"theoretical_loss": 3.3203060323514593,
"tokens_seen": 3005349888
},
{
"epoch": 0.05,
"learning_rate": 0.0009610778443113773,
"loss": 2.6209,
"theoretical_loss": 3.3202949028711197,
"tokens_seen": 3005480960
},
{
"epoch": 0.05,
"learning_rate": 0.0009606501283147991,
"loss": 2.522,
"theoretical_loss": 3.3202837740120335,
"tokens_seen": 3005612032
},
{
"epoch": 0.05,
"learning_rate": 0.0009602224123182207,
"loss": 2.5764,
"theoretical_loss": 3.3202726457741387,
"tokens_seen": 3005743104
},
{
"epoch": 0.05,
"learning_rate": 0.0009597946963216424,
"loss": 2.6805,
"theoretical_loss": 3.320261518157374,
"tokens_seen": 3005874176
},
{
"epoch": 0.05,
"learning_rate": 0.0009593669803250642,
"loss": 2.4835,
"theoretical_loss": 3.3202503911616765,
"tokens_seen": 3006005248
},
{
"epoch": 0.05,
"learning_rate": 0.0009589392643284859,
"loss": 2.5907,
"theoretical_loss": 3.320239264786986,
"tokens_seen": 3006136320
},
{
"epoch": 0.05,
"learning_rate": 0.0009585115483319077,
"loss": 2.617,
"theoretical_loss": 3.3202281390332393,
"tokens_seen": 3006267392
},
{
"epoch": 0.05,
"learning_rate": 0.0009580838323353293,
"loss": 2.5027,
"theoretical_loss": 3.320217013900376,
"tokens_seen": 3006398464
},
{
"epoch": 0.05,
"learning_rate": 0.000957656116338751,
"loss": 2.6857,
"theoretical_loss": 3.3202058893883333,
"tokens_seen": 3006529536
},
{
"epoch": 0.05,
"learning_rate": 0.0009572284003421729,
"loss": 2.6411,
"theoretical_loss": 3.3201947654970505,
"tokens_seen": 3006660608
},
{
"epoch": 0.05,
"objective/train/docs_used": 1649212,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.7101857662200928,
"objective/train/theoretical_loss": 3.320183642226465,
"objective/train/tokens_used": 36842976,
"theoretical_loss": 3.320183642226465,
"tokens_seen": 3006791680
},
{
"epoch": 0.05,
"learning_rate": 0.0009568006843455946,
"loss": 2.5369,
"theoretical_loss": 3.320183642226465,
"tokens_seen": 3006791680
},
{
"epoch": 0.05,
"learning_rate": 0.0009563729683490164,
"loss": 2.5836,
"theoretical_loss": 3.3201725195765155,
"tokens_seen": 3006922752
},
{
"epoch": 0.05,
"learning_rate": 0.000955945252352438,
"loss": 2.5358,
"theoretical_loss": 3.3201613975471402,
"tokens_seen": 3007053824
},
{
"epoch": 0.05,
"learning_rate": 0.0009555175363558597,
"loss": 2.6035,
"theoretical_loss": 3.3201502761382775,
"tokens_seen": 3007184896
},
{
"epoch": 0.05,
"learning_rate": 0.0009550898203592815,
"loss": 2.4594,
"theoretical_loss": 3.320139155349866,
"tokens_seen": 3007315968
},
{
"epoch": 0.06,
"learning_rate": 0.0009546621043627032,
"loss": 2.6376,
"theoretical_loss": 3.3201280351818436,
"tokens_seen": 3007447040
},
{
"epoch": 0.06,
"learning_rate": 0.0009542343883661248,
"loss": 2.5904,
"theoretical_loss": 3.320116915634149,
"tokens_seen": 3007578112
},
{
"epoch": 0.06,
"learning_rate": 0.0009538066723695466,
"loss": 2.7616,
"theoretical_loss": 3.3201057967067205,
"tokens_seen": 3007709184
},
{
"epoch": 0.06,
"learning_rate": 0.0009533789563729683,
"loss": 2.6076,
"theoretical_loss": 3.3200946783994962,
"tokens_seen": 3007840256
},
{
"epoch": 0.06,
"learning_rate": 0.0009529512403763902,
"loss": 2.6121,
"theoretical_loss": 3.3200835607124146,
"tokens_seen": 3007971328
},
{
"epoch": 0.06,
"learning_rate": 0.0009525235243798119,
"loss": 2.6083,
"theoretical_loss": 3.3200724436454143,
"tokens_seen": 3008102400
},
{
"epoch": 0.06,
"learning_rate": 0.0009520958083832335,
"loss": 2.6831,
"theoretical_loss": 3.3200613271984336,
"tokens_seen": 3008233472
},
{
"epoch": 0.06,
"learning_rate": 0.0009516680923866553,
"loss": 2.6156,
"theoretical_loss": 3.3200502113714108,
"tokens_seen": 3008364544
},
{
"epoch": 0.06,
"objective/train/docs_used": 1649940,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.545426368713379,
"objective/train/theoretical_loss": 3.3200446536903643,
"objective/train/tokens_used": 38481376,
"theoretical_loss": 3.3200446536903643,
"tokens_seen": 3008430080
},
{
"epoch": 0.06,
"learning_rate": 0.000951240376390077,
"loss": 2.6849,
"theoretical_loss": 3.3200390961642845,
"tokens_seen": 3008495616
},
{
"epoch": 0.06,
"learning_rate": 0.0009508126603934988,
"loss": 2.654,
"theoretical_loss": 3.3200279815769926,
"tokens_seen": 3008626688
},
{
"epoch": 0.06,
"learning_rate": 0.0009503849443969204,
"loss": 2.6215,
"theoretical_loss": 3.3200168676094743,
"tokens_seen": 3008757760
},
{
"epoch": 0.06,
"learning_rate": 0.0009499572284003421,
"loss": 2.4387,
"theoretical_loss": 3.320005754261668,
"tokens_seen": 3008888832
},
{
"epoch": 0.06,
"learning_rate": 0.0009495295124037639,
"loss": 2.5621,
"theoretical_loss": 3.319994641533511,
"tokens_seen": 3009019904
},
{
"epoch": 0.06,
"learning_rate": 0.0009491017964071857,
"loss": 2.731,
"theoretical_loss": 3.319983529424943,
"tokens_seen": 3009150976
},
{
"epoch": 0.06,
"learning_rate": 0.0009486740804106075,
"loss": 2.5847,
"theoretical_loss": 3.3199724179359027,
"tokens_seen": 3009282048
},
{
"epoch": 0.06,
"learning_rate": 0.0009482463644140291,
"loss": 2.5836,
"theoretical_loss": 3.319961307066327,
"tokens_seen": 3009413120
},
{
"epoch": 0.06,
"learning_rate": 0.0009478186484174508,
"loss": 2.6488,
"theoretical_loss": 3.3199501968161558,
"tokens_seen": 3009544192
},
{
"epoch": 0.06,
"learning_rate": 0.0009473909324208726,
"loss": 2.6244,
"theoretical_loss": 3.319939087185327,
"tokens_seen": 3009675264
},
{
"epoch": 0.06,
"learning_rate": 0.0009469632164242943,
"loss": 2.6873,
"theoretical_loss": 3.3199279781737796,
"tokens_seen": 3009806336
},
{
"epoch": 0.06,
"learning_rate": 0.000946535500427716,
"loss": 2.5921,
"theoretical_loss": 3.3199168697814514,
"tokens_seen": 3009937408
},
{
"epoch": 0.06,
"objective/train/docs_used": 1651249,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.582075834274292,
"objective/train/theoretical_loss": 3.3199057620082812,
"objective/train/tokens_used": 40119776,
"theoretical_loss": 3.3199057620082812,
"tokens_seen": 3010068480
},
{
"epoch": 0.06,
"learning_rate": 0.0009461077844311377,
"loss": 2.5527,
"theoretical_loss": 3.3199057620082812,
"tokens_seen": 3010068480
},
{
"epoch": 0.06,
"learning_rate": 0.0009456800684345594,
"loss": 2.7818,
"theoretical_loss": 3.319894654854208,
"tokens_seen": 3010199552
},
{
"epoch": 0.06,
"learning_rate": 0.0009452523524379812,
"loss": 2.628,
"theoretical_loss": 3.3198835483191695,
"tokens_seen": 3010330624
},
{
"epoch": 0.06,
"learning_rate": 0.000944824636441403,
"loss": 2.4649,
"theoretical_loss": 3.319872442403105,
"tokens_seen": 3010461696
},
{
"epoch": 0.07,
"learning_rate": 0.0009443969204448247,
"loss": 2.7571,
"theoretical_loss": 3.3198613371059524,
"tokens_seen": 3010592768
},
{
"epoch": 0.07,
"learning_rate": 0.0009439692044482464,
"loss": 2.5878,
"theoretical_loss": 3.319850232427651,
"tokens_seen": 3010723840
},
{
"epoch": 0.07,
"learning_rate": 0.0009435414884516681,
"loss": 2.6015,
"theoretical_loss": 3.3198391283681383,
"tokens_seen": 3010854912
},
{
"epoch": 0.07,
"learning_rate": 0.0009431137724550899,
"loss": 2.5823,
"theoretical_loss": 3.3198280249273546,
"tokens_seen": 3010985984
},
{
"epoch": 0.07,
"learning_rate": 0.0009426860564585116,
"loss": 2.7222,
"theoretical_loss": 3.319816922105237,
"tokens_seen": 3011117056
},
{
"epoch": 0.07,
"learning_rate": 0.0009422583404619332,
"loss": 2.6364,
"theoretical_loss": 3.319805819901724,
"tokens_seen": 3011248128
},
{
"epoch": 0.07,
"learning_rate": 0.000941830624465355,
"loss": 2.6811,
"theoretical_loss": 3.3197947183167553,
"tokens_seen": 3011379200
},
{
"epoch": 0.07,
"learning_rate": 0.0009414029084687767,
"loss": 2.8016,
"theoretical_loss": 3.319783617350269,
"tokens_seen": 3011510272
},
{
"epoch": 0.07,
"learning_rate": 0.0009409751924721985,
"loss": 2.5657,
"theoretical_loss": 3.319772517002204,
"tokens_seen": 3011641344
},
{
"epoch": 0.07,
"objective/train/docs_used": 1651905,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 3.0052103996276855,
"objective/train/theoretical_loss": 3.31976696706006,
"objective/train/tokens_used": 41758176,
"theoretical_loss": 3.31976696706006,
"tokens_seen": 3011706880
},
{
"epoch": 0.07,
"learning_rate": 0.0009405474764756203,
"loss": 2.6357,
"theoretical_loss": 3.319761417272498,
"tokens_seen": 3011772416
},
{
"epoch": 0.07,
"learning_rate": 0.0009401197604790419,
"loss": 2.5734,
"theoretical_loss": 3.319750318161091,
"tokens_seen": 3011903488
},
{
"epoch": 0.07,
"learning_rate": 0.0009396920444824637,
"loss": 2.424,
"theoretical_loss": 3.3197392196679205,
"tokens_seen": 3012034560
},
{
"epoch": 0.07,
"learning_rate": 0.0009392643284858854,
"loss": 2.5344,
"theoretical_loss": 3.3197281217929255,
"tokens_seen": 3012165632
},
{
"epoch": 0.07,
"learning_rate": 0.0009388366124893071,
"loss": 2.5689,
"theoretical_loss": 3.319717024536045,
"tokens_seen": 3012296704
},
{
"epoch": 0.07,
"learning_rate": 0.0009384088964927289,
"loss": 2.4989,
"theoretical_loss": 3.3197059278972176,
"tokens_seen": 3012427776
},
{
"epoch": 0.07,
"learning_rate": 0.0009379811804961505,
"loss": 2.6272,
"theoretical_loss": 3.3196948318763817,
"tokens_seen": 3012558848
},
{
"epoch": 0.07,
"learning_rate": 0.0009375534644995723,
"loss": 2.5959,
"theoretical_loss": 3.319683736473476,
"tokens_seen": 3012689920
},
{
"epoch": 0.07,
"learning_rate": 0.000937125748502994,
"loss": 2.5228,
"theoretical_loss": 3.3196726416884395,
"tokens_seen": 3012820992
},
{
"epoch": 0.07,
"learning_rate": 0.0009366980325064158,
"loss": 2.7357,
"theoretical_loss": 3.3196615475212106,
"tokens_seen": 3012952064
},
{
"epoch": 0.07,
"learning_rate": 0.0009362703165098376,
"loss": 2.6946,
"theoretical_loss": 3.3196504539717284,
"tokens_seen": 3013083136
},
{
"epoch": 0.07,
"learning_rate": 0.0009358426005132592,
"loss": 2.7344,
"theoretical_loss": 3.3196393610399317,
"tokens_seen": 3013214208
},
{
"epoch": 0.07,
"objective/train/docs_used": 1652881,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5316507816314697,
"objective/train/theoretical_loss": 3.3196282687257583,
"objective/train/tokens_used": 43396576,
"theoretical_loss": 3.3196282687257583,
"tokens_seen": 3013345280
},
{
"epoch": 0.07,
"learning_rate": 0.000935414884516681,
"loss": 2.7259,
"theoretical_loss": 3.3196282687257583,
"tokens_seen": 3013345280
},
{
"epoch": 0.07,
"learning_rate": 0.0009349871685201027,
"loss": 2.4882,
"theoretical_loss": 3.3196171770291483,
"tokens_seen": 3013476352
},
{
"epoch": 0.07,
"learning_rate": 0.0009345594525235244,
"loss": 2.6169,
"theoretical_loss": 3.3196060859500394,
"tokens_seen": 3013607424
},
{
"epoch": 0.08,
"learning_rate": 0.0009341317365269461,
"loss": 2.5293,
"theoretical_loss": 3.319594995488371,
"tokens_seen": 3013738496
},
{
"epoch": 0.08,
"learning_rate": 0.0009337040205303678,
"loss": 2.7782,
"theoretical_loss": 3.3195839056440812,
"tokens_seen": 3013869568
},
{
"epoch": 0.08,
"learning_rate": 0.0009332763045337895,
"loss": 2.6719,
"theoretical_loss": 3.3195728164171094,
"tokens_seen": 3014000640
},
{
"epoch": 0.08,
"learning_rate": 0.0009328485885372114,
"loss": 2.7268,
"theoretical_loss": 3.319561727807394,
"tokens_seen": 3014131712
},
{
"epoch": 0.08,
"learning_rate": 0.0009324208725406331,
"loss": 2.5997,
"theoretical_loss": 3.3195506398148744,
"tokens_seen": 3014262784
},
{
"epoch": 0.08,
"learning_rate": 0.0009319931565440548,
"loss": 2.7602,
"theoretical_loss": 3.319539552439489,
"tokens_seen": 3014393856
},
{
"epoch": 0.08,
"learning_rate": 0.0009315654405474765,
"loss": 2.5845,
"theoretical_loss": 3.3195284656811763,
"tokens_seen": 3014524928
},
{
"epoch": 0.08,
"learning_rate": 0.0009311377245508982,
"loss": 2.7053,
"theoretical_loss": 3.319517379539876,
"tokens_seen": 3014656000
},
{
"epoch": 0.08,
"learning_rate": 0.00093071000855432,
"loss": 2.6045,
"theoretical_loss": 3.3195062940155258,
"tokens_seen": 3014787072
},
{
"epoch": 0.08,
"learning_rate": 0.0009302822925577416,
"loss": 2.6324,
"theoretical_loss": 3.3194952091080654,
"tokens_seen": 3014918144
},
{
"epoch": 0.08,
"objective/train/docs_used": 1653310,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4821465015411377,
"objective/train/theoretical_loss": 3.3194896668856497,
"objective/train/tokens_used": 45034976,
"theoretical_loss": 3.3194896668856497,
"tokens_seen": 3014983680
},
{
"epoch": 0.08,
"learning_rate": 0.0009298545765611634,
"loss": 2.532,
"theoretical_loss": 3.3194841248174334,
"tokens_seen": 3015049216
},
{
"epoch": 0.08,
"learning_rate": 0.0009294268605645851,
"loss": 2.544,
"theoretical_loss": 3.3194730411435684,
"tokens_seen": 3015180288
},
{
"epoch": 0.08,
"learning_rate": 0.0009289991445680068,
"loss": 2.826,
"theoretical_loss": 3.3194619580864098,
"tokens_seen": 3015311360
},
{
"epoch": 0.08,
"learning_rate": 0.0009285714285714287,
"loss": 2.7561,
"theoretical_loss": 3.3194508756458965,
"tokens_seen": 3015442432
},
{
"epoch": 0.08,
"learning_rate": 0.0009281437125748503,
"loss": 2.5692,
"theoretical_loss": 3.319439793821967,
"tokens_seen": 3015573504
},
{
"epoch": 0.08,
"learning_rate": 0.000927715996578272,
"loss": 2.6322,
"theoretical_loss": 3.3194287126145596,
"tokens_seen": 3015704576
},
{
"epoch": 0.08,
"learning_rate": 0.0009272882805816938,
"loss": 2.6346,
"theoretical_loss": 3.3194176320236144,
"tokens_seen": 3015835648
},
{
"epoch": 0.08,
"learning_rate": 0.0009268605645851155,
"loss": 2.7908,
"theoretical_loss": 3.31940655204907,
"tokens_seen": 3015966720
},
{
"epoch": 0.08,
"learning_rate": 0.0009264328485885373,
"loss": 2.6439,
"theoretical_loss": 3.319395472690865,
"tokens_seen": 3016097792
},
{
"epoch": 0.08,
"learning_rate": 0.0009260051325919589,
"loss": 2.5589,
"theoretical_loss": 3.3193843939489382,
"tokens_seen": 3016228864
},
{
"epoch": 0.08,
"learning_rate": 0.0009255774165953806,
"loss": 2.6926,
"theoretical_loss": 3.319373315823229,
"tokens_seen": 3016359936
},
{
"epoch": 0.08,
"learning_rate": 0.0009251497005988024,
"loss": 2.7646,
"theoretical_loss": 3.3193622383136763,
"tokens_seen": 3016491008
},
{
"epoch": 0.08,
"objective/train/docs_used": 1654644,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.3390886783599854,
"objective/train/theoretical_loss": 3.3193511614202187,
"objective/train/tokens_used": 46673376,
"theoretical_loss": 3.3193511614202187,
"tokens_seen": 3016622080
},
{
"epoch": 0.08,
"learning_rate": 0.0009247219846022242,
"loss": 2.5287,
"theoretical_loss": 3.3193511614202187,
"tokens_seen": 3016622080
},
{
"epoch": 0.09,
"learning_rate": 0.000924294268605646,
"loss": 2.7607,
"theoretical_loss": 3.319340085142796,
"tokens_seen": 3016753152
},
{
"epoch": 0.09,
"learning_rate": 0.0009238665526090676,
"loss": 2.6289,
"theoretical_loss": 3.319329009481346,
"tokens_seen": 3016884224
},
{
"epoch": 0.09,
"learning_rate": 0.0009234388366124893,
"loss": 2.7648,
"theoretical_loss": 3.3193179344358086,
"tokens_seen": 3017015296
},
{
"epoch": 0.09,
"learning_rate": 0.0009230111206159111,
"loss": 2.497,
"theoretical_loss": 3.319306860006122,
"tokens_seen": 3017146368
},
{
"epoch": 0.09,
"learning_rate": 0.0009225834046193328,
"loss": 2.4963,
"theoretical_loss": 3.319295786192226,
"tokens_seen": 3017277440
},
{
"epoch": 0.09,
"learning_rate": 0.0009221556886227545,
"loss": 2.5823,
"theoretical_loss": 3.319284712994059,
"tokens_seen": 3017408512
},
{
"epoch": 0.09,
"learning_rate": 0.0009217279726261762,
"loss": 2.6555,
"theoretical_loss": 3.3192736404115606,
"tokens_seen": 3017539584
},
{
"epoch": 0.09,
"learning_rate": 0.0009213002566295979,
"loss": 2.5264,
"theoretical_loss": 3.3192625684446693,
"tokens_seen": 3017670656
},
{
"epoch": 0.09,
"learning_rate": 0.0009208725406330197,
"loss": 2.397,
"theoretical_loss": 3.3192514970933242,
"tokens_seen": 3017801728
},
{
"epoch": 0.09,
"learning_rate": 0.0009204448246364415,
"loss": 2.6273,
"theoretical_loss": 3.319240426357465,
"tokens_seen": 3017932800
},
{
"epoch": 0.09,
"learning_rate": 0.0009200171086398631,
"loss": 2.6414,
"theoretical_loss": 3.31922935623703,
"tokens_seen": 3018063872
},
{
"epoch": 0.09,
"learning_rate": 0.0009195893926432849,
"loss": 2.5706,
"theoretical_loss": 3.3192182867319584,
"tokens_seen": 3018194944
},
{
"epoch": 0.09,
"objective/train/docs_used": 1655335,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.3051180839538574,
"objective/train/theoretical_loss": 3.319212752210165,
"objective/train/tokens_used": 48311776,
"theoretical_loss": 3.319212752210165,
"tokens_seen": 3018260480
},
{
"epoch": 0.09,
"learning_rate": 0.0009191616766467066,
"loss": 2.5393,
"theoretical_loss": 3.3192072178421896,
"tokens_seen": 3018326016
},
{
"epoch": 0.09,
"learning_rate": 0.0009187339606501284,
"loss": 2.6545,
"theoretical_loss": 3.319196149567662,
"tokens_seen": 3018457088
},
{
"epoch": 0.09,
"learning_rate": 0.00091830624465355,
"loss": 2.5623,
"theoretical_loss": 3.3191850819083157,
"tokens_seen": 3018588160
},
{
"epoch": 0.09,
"learning_rate": 0.0009178785286569717,
"loss": 2.6804,
"theoretical_loss": 3.319174014864089,
"tokens_seen": 3018719232
},
{
"epoch": 0.09,
"learning_rate": 0.0009174508126603935,
"loss": 2.8051,
"theoretical_loss": 3.319162948434921,
"tokens_seen": 3018850304
},
{
"epoch": 0.09,
"learning_rate": 0.0009170230966638152,
"loss": 2.7044,
"theoretical_loss": 3.319151882620752,
"tokens_seen": 3018981376
},
{
"epoch": 0.09,
"learning_rate": 0.0009165953806672371,
"loss": 2.5421,
"theoretical_loss": 3.3191408174215193,
"tokens_seen": 3019112448
},
{
"epoch": 0.09,
"learning_rate": 0.0009161676646706587,
"loss": 2.7474,
"theoretical_loss": 3.3191297528371635,
"tokens_seen": 3019243520
},
{
"epoch": 0.09,
"learning_rate": 0.0009157399486740804,
"loss": 2.5804,
"theoretical_loss": 3.319118688867623,
"tokens_seen": 3019374592
},
{
"epoch": 0.09,
"learning_rate": 0.0009153122326775022,
"loss": 2.5145,
"theoretical_loss": 3.319107625512837,
"tokens_seen": 3019505664
},
{
"epoch": 0.09,
"learning_rate": 0.0009148845166809239,
"loss": 2.746,
"theoretical_loss": 3.3190965627727445,
"tokens_seen": 3019636736
},
{
"epoch": 0.09,
"learning_rate": 0.0009144568006843457,
"loss": 2.5949,
"theoretical_loss": 3.3190855006472857,
"tokens_seen": 3019767808
},
{
"epoch": 0.09,
"objective/train/docs_used": 1656670,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.1859257221221924,
"objective/train/theoretical_loss": 3.3190744391363984,
"objective/train/tokens_used": 49950176,
"theoretical_loss": 3.3190744391363984,
"tokens_seen": 3019898880
},
{
"epoch": 0.1,
"learning_rate": 0.0009140290846877673,
"loss": 2.6334,
"theoretical_loss": 3.3190744391363984,
"tokens_seen": 3019898880
},
{
"epoch": 0.1,
"learning_rate": 0.000913601368691189,
"loss": 2.6976,
"theoretical_loss": 3.3190633782400223,
"tokens_seen": 3020029952
},
{
"epoch": 0.1,
"learning_rate": 0.0009131736526946108,
"loss": 2.5658,
"theoretical_loss": 3.3190523179580973,
"tokens_seen": 3020161024
},
{
"epoch": 0.1,
"learning_rate": 0.0009127459366980325,
"loss": 2.7495,
"theoretical_loss": 3.3190412582905617,
"tokens_seen": 3020292096
},
{
"epoch": 0.1,
"learning_rate": 0.0009123182207014543,
"loss": 2.7093,
"theoretical_loss": 3.319030199237355,
"tokens_seen": 3020423168
},
{
"epoch": 0.1,
"learning_rate": 0.000911890504704876,
"loss": 2.5295,
"theoretical_loss": 3.3190191407984164,
"tokens_seen": 3020554240
},
{
"epoch": 0.1,
"learning_rate": 0.0009114627887082977,
"loss": 2.6081,
"theoretical_loss": 3.3190080829736854,
"tokens_seen": 3020685312
},
{
"epoch": 0.1,
"learning_rate": 0.0009110350727117195,
"loss": 2.6166,
"theoretical_loss": 3.318997025763101,
"tokens_seen": 3020816384
},
{
"epoch": 0.1,
"learning_rate": 0.0009106073567151412,
"loss": 2.7691,
"theoretical_loss": 3.318985969166602,
"tokens_seen": 3020947456
},
{
"epoch": 0.1,
"learning_rate": 0.0009101796407185628,
"loss": 2.6465,
"theoretical_loss": 3.3189749131841286,
"tokens_seen": 3021078528
},
{
"epoch": 0.1,
"learning_rate": 0.0009097519247219846,
"loss": 2.5314,
"theoretical_loss": 3.3189638578156195,
"tokens_seen": 3021209600
},
{
"epoch": 0.1,
"learning_rate": 0.0009093242087254063,
"loss": 2.8726,
"theoretical_loss": 3.3189528030610136,
"tokens_seen": 3021340672
},
{
"epoch": 0.1,
"learning_rate": 0.0009088964927288281,
"loss": 2.6762,
"theoretical_loss": 3.318941748920251,
"tokens_seen": 3021471744
},
{
"epoch": 0.1,
"objective/train/docs_used": 1657192,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.9018635749816895,
"objective/train/theoretical_loss": 3.318936222080042,
"objective/train/tokens_used": 51588576,
"theoretical_loss": 3.318936222080042,
"tokens_seen": 3021537280
},
{
"epoch": 0.1,
"learning_rate": 0.0009084687767322499,
"loss": 2.643,
"theoretical_loss": 3.318930695393271,
"tokens_seen": 3021602816
},
{
"epoch": 0.1,
"learning_rate": 0.0009080410607356715,
"loss": 2.6665,
"theoretical_loss": 3.3189196424800116,
"tokens_seen": 3021733888
},
{
"epoch": 0.1,
"learning_rate": 0.0009076133447390933,
"loss": 2.6265,
"theoretical_loss": 3.3189085901804134,
"tokens_seen": 3021864960
},
{
"epoch": 0.1,
"learning_rate": 0.000907185628742515,
"loss": 2.5788,
"theoretical_loss": 3.3188975384944155,
"tokens_seen": 3021996032
},
{
"epoch": 0.1,
"learning_rate": 0.0009067579127459367,
"loss": 2.6531,
"theoretical_loss": 3.318886487421957,
"tokens_seen": 3022127104
},
{
"epoch": 0.1,
"learning_rate": 0.0009063301967493585,
"loss": 2.7177,
"theoretical_loss": 3.318875436962977,
"tokens_seen": 3022258176
},
{
"epoch": 0.1,
"learning_rate": 0.0009059024807527801,
"loss": 2.4989,
"theoretical_loss": 3.3188643871174155,
"tokens_seen": 3022389248
},
{
"epoch": 0.1,
"learning_rate": 0.0009054747647562019,
"loss": 2.5249,
"theoretical_loss": 3.318853337885211,
"tokens_seen": 3022520320
},
{
"epoch": 0.1,
"learning_rate": 0.0009050470487596236,
"loss": 2.6031,
"theoretical_loss": 3.318842289266304,
"tokens_seen": 3022651392
},
{
"epoch": 0.1,
"learning_rate": 0.0009046193327630453,
"loss": 2.5702,
"theoretical_loss": 3.3188312412606327,
"tokens_seen": 3022782464
},
{
"epoch": 0.1,
"learning_rate": 0.0009041916167664672,
"loss": 2.5348,
"theoretical_loss": 3.3188201938681368,
"tokens_seen": 3022913536
},
{
"epoch": 0.11,
"learning_rate": 0.0009037639007698888,
"loss": 2.5876,
"theoretical_loss": 3.318809147088756,
"tokens_seen": 3023044608
},
{
"epoch": 0.11,
"objective/train/docs_used": 1658380,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.8608322143554688,
"objective/train/theoretical_loss": 3.3187981009224297,
"objective/train/tokens_used": 53226976,
"theoretical_loss": 3.3187981009224297,
"tokens_seen": 3023175680
},
{
"epoch": 0.11,
"learning_rate": 0.0009033361847733106,
"loss": 2.6644,
"theoretical_loss": 3.3187981009224297,
"tokens_seen": 3023175680
},
{
"epoch": 0.11,
"learning_rate": 0.0009029084687767323,
"loss": 2.5824,
"theoretical_loss": 3.3187870553690972,
"tokens_seen": 3023306752
},
{
"epoch": 0.11,
"learning_rate": 0.000902480752780154,
"loss": 2.7377,
"theoretical_loss": 3.3187760104286976,
"tokens_seen": 3023437824
},
{
"epoch": 0.11,
"learning_rate": 0.0009020530367835757,
"loss": 2.6572,
"theoretical_loss": 3.3187649661011704,
"tokens_seen": 3023568896
},
{
"epoch": 0.11,
"learning_rate": 0.0009016253207869974,
"loss": 2.5447,
"theoretical_loss": 3.3187539223864557,
"tokens_seen": 3023699968
},
{
"epoch": 0.11,
"learning_rate": 0.0009011976047904192,
"loss": 2.6011,
"theoretical_loss": 3.318742879284492,
"tokens_seen": 3023831040
},
{
"epoch": 0.11,
"learning_rate": 0.0009007698887938409,
"loss": 2.4863,
"theoretical_loss": 3.3187318367952194,
"tokens_seen": 3023962112
},
{
"epoch": 0.11,
"learning_rate": 0.0009003421727972627,
"loss": 2.4994,
"theoretical_loss": 3.318720794918577,
"tokens_seen": 3024093184
},
{
"epoch": 0.11,
"learning_rate": 0.0008999144568006844,
"loss": 2.5443,
"theoretical_loss": 3.3187097536545047,
"tokens_seen": 3024224256
},
{
"epoch": 0.11,
"learning_rate": 0.0008994867408041061,
"loss": 2.6808,
"theoretical_loss": 3.3186987130029415,
"tokens_seen": 3024355328
},
{
"epoch": 0.11,
"learning_rate": 0.0008990590248075278,
"loss": 2.7511,
"theoretical_loss": 3.3186876729638266,
"tokens_seen": 3024486400
},
{
"epoch": 0.11,
"learning_rate": 0.0008986313088109496,
"loss": 2.612,
"theoretical_loss": 3.3186766335371005,
"tokens_seen": 3024617472
},
{
"epoch": 0.11,
"learning_rate": 0.0008982035928143712,
"loss": 2.686,
"theoretical_loss": 3.318665594722702,
"tokens_seen": 3024748544
},
{
"epoch": 0.11,
"objective/train/docs_used": 1658975,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.8614556789398193,
"objective/train/theoretical_loss": 3.3186600755451066,
"objective/train/tokens_used": 54865376,
"theoretical_loss": 3.3186600755451066,
"tokens_seen": 3024814080
},
{
"epoch": 0.11,
"learning_rate": 0.000897775876817793,
"loss": 2.654,
"theoretical_loss": 3.3186545565205705,
"tokens_seen": 3024879616
},
{
"epoch": 0.11,
"learning_rate": 0.0008973481608212147,
"loss": 2.6409,
"theoretical_loss": 3.318643518930646,
"tokens_seen": 3025010688
},
{
"epoch": 0.11,
"learning_rate": 0.0008969204448246364,
"loss": 2.5854,
"theoretical_loss": 3.3186324819528674,
"tokens_seen": 3025141760
},
{
"epoch": 0.11,
"learning_rate": 0.0008964927288280582,
"loss": 2.6063,
"theoretical_loss": 3.318621445587175,
"tokens_seen": 3025272832
},
{
"epoch": 0.11,
"learning_rate": 0.0008960650128314799,
"loss": 2.537,
"theoretical_loss": 3.3186104098335076,
"tokens_seen": 3025403904
},
{
"epoch": 0.11,
"learning_rate": 0.0008956372968349017,
"loss": 2.5978,
"theoretical_loss": 3.318599374691805,
"tokens_seen": 3025534976
},
{
"epoch": 0.11,
"learning_rate": 0.0008952095808383234,
"loss": 2.6201,
"theoretical_loss": 3.318588340162007,
"tokens_seen": 3025666048
},
{
"epoch": 0.11,
"learning_rate": 0.0008947818648417451,
"loss": 2.7451,
"theoretical_loss": 3.3185773062440527,
"tokens_seen": 3025797120
},
{
"epoch": 0.11,
"learning_rate": 0.0008943541488451669,
"loss": 2.5093,
"theoretical_loss": 3.3185662729378826,
"tokens_seen": 3025928192
},
{
"epoch": 0.12,
"learning_rate": 0.0008939264328485885,
"loss": 2.517,
"theoretical_loss": 3.318555240243435,
"tokens_seen": 3026059264
},
{
"epoch": 0.12,
"learning_rate": 0.0008934987168520102,
"loss": 2.515,
"theoretical_loss": 3.3185442081606507,
"tokens_seen": 3026190336
},
{
"epoch": 0.12,
"learning_rate": 0.000893071000855432,
"loss": 2.6143,
"theoretical_loss": 3.3185331766894683,
"tokens_seen": 3026321408
},
{
"epoch": 0.12,
"objective/train/docs_used": 1660130,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4469985961914062,
"objective/train/theoretical_loss": 3.318522145829828,
"objective/train/tokens_used": 56503776,
"theoretical_loss": 3.318522145829828,
"tokens_seen": 3026452480
},
{
"epoch": 0.12,
"learning_rate": 0.0008926432848588537,
"loss": 2.4787,
"theoretical_loss": 3.318522145829828,
"tokens_seen": 3026452480
},
{
"epoch": 0.12,
"learning_rate": 0.0008922155688622756,
"loss": 2.6368,
"theoretical_loss": 3.318511115581669,
"tokens_seen": 3026583552
},
{
"epoch": 0.12,
"learning_rate": 0.0008917878528656972,
"loss": 2.6092,
"theoretical_loss": 3.3185000859449314,
"tokens_seen": 3026714624
},
{
"epoch": 0.12,
"learning_rate": 0.0008913601368691189,
"loss": 2.5731,
"theoretical_loss": 3.3184890569195544,
"tokens_seen": 3026845696
},
{
"epoch": 0.12,
"learning_rate": 0.0008909324208725407,
"loss": 2.5214,
"theoretical_loss": 3.3184780285054782,
"tokens_seen": 3026976768
},
{
"epoch": 0.12,
"learning_rate": 0.0008905047048759624,
"loss": 2.5803,
"theoretical_loss": 3.318467000702642,
"tokens_seen": 3027107840
},
{
"epoch": 0.12,
"learning_rate": 0.0008900769888793841,
"loss": 2.6223,
"theoretical_loss": 3.3184559735109853,
"tokens_seen": 3027238912
},
{
"epoch": 0.12,
"learning_rate": 0.0008896492728828058,
"loss": 2.5938,
"theoretical_loss": 3.3184449469304482,
"tokens_seen": 3027369984
},
{
"epoch": 0.12,
"learning_rate": 0.0008892215568862275,
"loss": 2.6878,
"theoretical_loss": 3.3184339209609703,
"tokens_seen": 3027501056
},
{
"epoch": 0.12,
"learning_rate": 0.0008887938408896493,
"loss": 2.4932,
"theoretical_loss": 3.318422895602491,
"tokens_seen": 3027632128
},
{
"epoch": 0.12,
"learning_rate": 0.000888366124893071,
"loss": 2.4597,
"theoretical_loss": 3.31841187085495,
"tokens_seen": 3027763200
},
{
"epoch": 0.12,
"learning_rate": 0.0008879384088964928,
"loss": 2.4485,
"theoretical_loss": 3.318400846718288,
"tokens_seen": 3027894272
},
{
"epoch": 0.12,
"learning_rate": 0.0008875106928999145,
"loss": 2.6036,
"theoretical_loss": 3.318389823192443,
"tokens_seen": 3028025344
},
{
"epoch": 0.12,
"objective/train/docs_used": 1661412,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.3641762733459473,
"objective/train/theoretical_loss": 3.3183843116585585,
"objective/train/tokens_used": 58142176,
"theoretical_loss": 3.3183843116585585,
"tokens_seen": 3028090880
},
{
"epoch": 0.12,
"learning_rate": 0.0008870829769033362,
"loss": 2.6277,
"theoretical_loss": 3.318378800277356,
"tokens_seen": 3028156416
},
{
"epoch": 0.12,
"learning_rate": 0.000886655260906758,
"loss": 2.5149,
"theoretical_loss": 3.3183677779729663,
"tokens_seen": 3028287488
},
{
"epoch": 0.12,
"learning_rate": 0.0008862275449101797,
"loss": 2.6112,
"theoretical_loss": 3.3183567562792136,
"tokens_seen": 3028418560
},
{
"epoch": 0.12,
"learning_rate": 0.0008857998289136013,
"loss": 2.6564,
"theoretical_loss": 3.3183457351960377,
"tokens_seen": 3028549632
},
{
"epoch": 0.12,
"learning_rate": 0.0008853721129170231,
"loss": 2.6973,
"theoretical_loss": 3.3183347147233784,
"tokens_seen": 3028680704
},
{
"epoch": 0.12,
"learning_rate": 0.0008849443969204448,
"loss": 2.4547,
"theoretical_loss": 3.3183236948611756,
"tokens_seen": 3028811776
},
{
"epoch": 0.12,
"learning_rate": 0.0008845166809238666,
"loss": 2.3115,
"theoretical_loss": 3.3183126756093686,
"tokens_seen": 3028942848
},
{
"epoch": 0.12,
"learning_rate": 0.0008840889649272883,
"loss": 2.5859,
"theoretical_loss": 3.318301656967898,
"tokens_seen": 3029073920
},
{
"epoch": 0.13,
"learning_rate": 0.00088366124893071,
"loss": 2.4681,
"theoretical_loss": 3.3182906389367024,
"tokens_seen": 3029204992
},
{
"epoch": 0.13,
"learning_rate": 0.0008832335329341318,
"loss": 2.4459,
"theoretical_loss": 3.3182796215157224,
"tokens_seen": 3029336064
},
{
"epoch": 0.13,
"learning_rate": 0.0008828058169375535,
"loss": 2.5877,
"theoretical_loss": 3.318268604704898,
"tokens_seen": 3029467136
},
{
"epoch": 0.13,
"learning_rate": 0.0008823781009409753,
"loss": 2.5306,
"theoretical_loss": 3.318257588504168,
"tokens_seen": 3029598208
},
{
"epoch": 0.13,
"objective/train/docs_used": 1662079,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.3965446949005127,
"objective/train/theoretical_loss": 3.318246572913474,
"objective/train/tokens_used": 59780576,
"theoretical_loss": 3.318246572913474,
"tokens_seen": 3029729280
},
{
"epoch": 0.13,
"learning_rate": 0.0008819503849443969,
"loss": 2.6399,
"theoretical_loss": 3.318246572913474,
"tokens_seen": 3029729280
},
{
"epoch": 0.13,
"learning_rate": 0.0008815226689478186,
"loss": 2.6156,
"theoretical_loss": 3.318235557932754,
"tokens_seen": 3029860352
},
{
"epoch": 0.13,
"learning_rate": 0.0008810949529512404,
"loss": 2.4887,
"theoretical_loss": 3.318224543561948,
"tokens_seen": 3029991424
},
{
"epoch": 0.13,
"learning_rate": 0.0008806672369546621,
"loss": 2.7212,
"theoretical_loss": 3.3182135298009974,
"tokens_seen": 3030122496
},
{
"epoch": 0.13,
"learning_rate": 0.0008802395209580839,
"loss": 2.6808,
"theoretical_loss": 3.3182025166498406,
"tokens_seen": 3030253568
},
{
"epoch": 0.13,
"learning_rate": 0.0008798118049615056,
"loss": 2.4704,
"theoretical_loss": 3.3181915041084182,
"tokens_seen": 3030384640
},
{
"epoch": 0.13,
"learning_rate": 0.0008793840889649273,
"loss": 2.4549,
"theoretical_loss": 3.3181804921766695,
"tokens_seen": 3030515712
},
{
"epoch": 0.13,
"learning_rate": 0.0008789563729683491,
"loss": 2.8214,
"theoretical_loss": 3.318169480854535,
"tokens_seen": 3030646784
},
{
"epoch": 0.13,
"learning_rate": 0.0008785286569717708,
"loss": 2.5934,
"theoretical_loss": 3.318158470141954,
"tokens_seen": 3030777856
},
{
"epoch": 0.13,
"learning_rate": 0.0008781009409751924,
"loss": 2.6187,
"theoretical_loss": 3.3181474600388667,
"tokens_seen": 3030908928
},
{
"epoch": 0.13,
"learning_rate": 0.0008776732249786142,
"loss": 2.3858,
"theoretical_loss": 3.318136450545213,
"tokens_seen": 3031040000
},
{
"epoch": 0.13,
"learning_rate": 0.0008772455089820359,
"loss": 2.5896,
"theoretical_loss": 3.318125441660933,
"tokens_seen": 3031171072
},
{
"epoch": 0.13,
"learning_rate": 0.0008768177929854577,
"loss": 2.388,
"theoretical_loss": 3.318114433385966,
"tokens_seen": 3031302144
},
{
"epoch": 0.13,
"objective/train/docs_used": 1662642,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5412111282348633,
"objective/train/theoretical_loss": 3.3181089294769563,
"objective/train/tokens_used": 61418976,
"theoretical_loss": 3.3181089294769563,
"tokens_seen": 3031367680
},
{
"epoch": 0.13,
"learning_rate": 0.0008763900769888794,
"loss": 2.6377,
"theoretical_loss": 3.3181034257202526,
"tokens_seen": 3031433216
},
{
"epoch": 0.13,
"learning_rate": 0.0008759623609923011,
"loss": 2.55,
"theoretical_loss": 3.318092418663732,
"tokens_seen": 3031564288
},
{
"epoch": 0.13,
"learning_rate": 0.0008755346449957229,
"loss": 2.6605,
"theoretical_loss": 3.3180814122163453,
"tokens_seen": 3031695360
},
{
"epoch": 0.13,
"learning_rate": 0.0008751069289991446,
"loss": 2.6276,
"theoretical_loss": 3.3180704063780313,
"tokens_seen": 3031826432
},
{
"epoch": 0.13,
"learning_rate": 0.0008746792130025664,
"loss": 2.6145,
"theoretical_loss": 3.318059401148731,
"tokens_seen": 3031957504
},
{
"epoch": 0.13,
"learning_rate": 0.0008742514970059881,
"loss": 2.7501,
"theoretical_loss": 3.3180483965283836,
"tokens_seen": 3032088576
},
{
"epoch": 0.14,
"learning_rate": 0.0008738237810094097,
"loss": 2.4986,
"theoretical_loss": 3.318037392516929,
"tokens_seen": 3032219648
},
{
"epoch": 0.14,
"learning_rate": 0.0008733960650128315,
"loss": 2.5994,
"theoretical_loss": 3.318026389114308,
"tokens_seen": 3032350720
},
{
"epoch": 0.14,
"learning_rate": 0.0008729683490162532,
"loss": 2.4541,
"theoretical_loss": 3.3180153863204596,
"tokens_seen": 3032481792
},
{
"epoch": 0.14,
"learning_rate": 0.0008725406330196749,
"loss": 2.6821,
"theoretical_loss": 3.3180043841353246,
"tokens_seen": 3032612864
},
{
"epoch": 0.14,
"learning_rate": 0.0008721129170230966,
"loss": 2.5988,
"theoretical_loss": 3.317993382558843,
"tokens_seen": 3032743936
},
{
"epoch": 0.14,
"learning_rate": 0.0008716852010265184,
"loss": 2.6567,
"theoretical_loss": 3.317982381590954,
"tokens_seen": 3032875008
},
{
"epoch": 0.14,
"objective/train/docs_used": 1663221,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.975888252258301,
"objective/train/theoretical_loss": 3.3179713812315983,
"objective/train/tokens_used": 63057376,
"theoretical_loss": 3.3179713812315983,
"tokens_seen": 3033006080
},
{
"epoch": 0.14,
"learning_rate": 0.0008712574850299402,
"loss": 2.6197,
"theoretical_loss": 3.3179713812315983,
"tokens_seen": 3033006080
},
{
"epoch": 0.14,
"learning_rate": 0.0008708297690333619,
"loss": 2.7088,
"theoretical_loss": 3.317960381480716,
"tokens_seen": 3033137152
},
{
"epoch": 0.14,
"learning_rate": 0.0008704020530367836,
"loss": 2.582,
"theoretical_loss": 3.317949382338247,
"tokens_seen": 3033268224
},
{
"epoch": 0.14,
"learning_rate": 0.0008699743370402053,
"loss": 2.5966,
"theoretical_loss": 3.3179383838041314,
"tokens_seen": 3033399296
},
{
"epoch": 0.14,
"learning_rate": 0.000869546621043627,
"loss": 2.8364,
"theoretical_loss": 3.317927385878309,
"tokens_seen": 3033530368
},
{
"epoch": 0.14,
"learning_rate": 0.0008691189050470488,
"loss": 2.6666,
"theoretical_loss": 3.31791638856072,
"tokens_seen": 3033661440
},
{
"epoch": 0.14,
"learning_rate": 0.0008686911890504705,
"loss": 2.6976,
"theoretical_loss": 3.317905391851305,
"tokens_seen": 3033792512
},
{
"epoch": 0.14,
"learning_rate": 0.0008682634730538922,
"loss": 2.662,
"theoretical_loss": 3.3178943957500033,
"tokens_seen": 3033923584
},
{
"epoch": 0.14,
"learning_rate": 0.000867835757057314,
"loss": 2.7223,
"theoretical_loss": 3.317883400256756,
"tokens_seen": 3034054656
},
{
"epoch": 0.14,
"learning_rate": 0.0008674080410607357,
"loss": 2.754,
"theoretical_loss": 3.3178724053715016,
"tokens_seen": 3034185728
},
{
"epoch": 0.14,
"learning_rate": 0.0008669803250641575,
"loss": 2.6831,
"theoretical_loss": 3.3178614110941815,
"tokens_seen": 3034316800
},
{
"epoch": 0.14,
"learning_rate": 0.0008665526090675792,
"loss": 2.8167,
"theoretical_loss": 3.3178504174247356,
"tokens_seen": 3034447872
},
{
"epoch": 0.14,
"learning_rate": 0.0008661248930710008,
"loss": 2.761,
"theoretical_loss": 3.317839424363104,
"tokens_seen": 3034578944
},
{
"epoch": 0.14,
"objective/train/docs_used": 1664363,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.568387746810913,
"objective/train/theoretical_loss": 3.3178339280602,
"objective/train/tokens_used": 64695776,
"theoretical_loss": 3.3178339280602,
"tokens_seen": 3034644480
},
{
"epoch": 0.14,
"learning_rate": 0.0008656971770744226,
"loss": 2.5995,
"theoretical_loss": 3.317828431909227,
"tokens_seen": 3034710016
},
{
"epoch": 0.14,
"learning_rate": 0.0008652694610778443,
"loss": 2.5506,
"theoretical_loss": 3.3178174400630445,
"tokens_seen": 3034841088
},
{
"epoch": 0.14,
"learning_rate": 0.000864841745081266,
"loss": 2.8138,
"theoretical_loss": 3.3178064488244967,
"tokens_seen": 3034972160
},
{
"epoch": 0.14,
"learning_rate": 0.0008644140290846878,
"loss": 2.5913,
"theoretical_loss": 3.3177954581935234,
"tokens_seen": 3035103232
},
{
"epoch": 0.14,
"learning_rate": 0.0008639863130881094,
"loss": 2.5688,
"theoretical_loss": 3.317784468170066,
"tokens_seen": 3035234304
},
{
"epoch": 0.15,
"learning_rate": 0.0008635585970915313,
"loss": 2.5422,
"theoretical_loss": 3.317773478754063,
"tokens_seen": 3035365376
},
{
"epoch": 0.15,
"learning_rate": 0.000863130881094953,
"loss": 2.6193,
"theoretical_loss": 3.317762489945456,
"tokens_seen": 3035496448
},
{
"epoch": 0.15,
"learning_rate": 0.0008627031650983747,
"loss": 2.5738,
"theoretical_loss": 3.3177515017441843,
"tokens_seen": 3035627520
},
{
"epoch": 0.15,
"learning_rate": 0.0008622754491017965,
"loss": 2.715,
"theoretical_loss": 3.3177405141501883,
"tokens_seen": 3035758592
},
{
"epoch": 0.15,
"learning_rate": 0.0008618477331052181,
"loss": 2.643,
"theoretical_loss": 3.317729527163409,
"tokens_seen": 3035889664
},
{
"epoch": 0.15,
"learning_rate": 0.0008614200171086399,
"loss": 2.987,
"theoretical_loss": 3.3177185407837855,
"tokens_seen": 3036020736
},
{
"epoch": 0.15,
"learning_rate": 0.0008609923011120616,
"loss": 2.594,
"theoretical_loss": 3.3177075550112587,
"tokens_seen": 3036151808
},
{
"epoch": 0.15,
"objective/train/docs_used": 1664891,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.584660291671753,
"objective/train/theoretical_loss": 3.3176965698457686,
"objective/train/tokens_used": 66334176,
"theoretical_loss": 3.3176965698457686,
"tokens_seen": 3036282880
},
{
"epoch": 0.15,
"learning_rate": 0.0008605645851154833,
"loss": 2.6228,
"theoretical_loss": 3.3176965698457686,
"tokens_seen": 3036282880
},
{
"epoch": 0.15,
"learning_rate": 0.000860136869118905,
"loss": 2.6502,
"theoretical_loss": 3.3176855852872555,
"tokens_seen": 3036413952
},
{
"epoch": 0.15,
"learning_rate": 0.0008597091531223268,
"loss": 2.6091,
"theoretical_loss": 3.3176746013356597,
"tokens_seen": 3036545024
},
{
"epoch": 0.15,
"learning_rate": 0.0008592814371257485,
"loss": 2.5742,
"theoretical_loss": 3.317663617990922,
"tokens_seen": 3036676096
},
{
"epoch": 0.15,
"learning_rate": 0.0008588537211291703,
"loss": 2.573,
"theoretical_loss": 3.3176526352529816,
"tokens_seen": 3036807168
},
{
"epoch": 0.15,
"learning_rate": 0.000858426005132592,
"loss": 2.6126,
"theoretical_loss": 3.31764165312178,
"tokens_seen": 3036938240
},
{
"epoch": 0.15,
"learning_rate": 0.0008579982891360137,
"loss": 2.6855,
"theoretical_loss": 3.3176306715972563,
"tokens_seen": 3037069312
},
{
"epoch": 0.15,
"learning_rate": 0.0008575705731394354,
"loss": 2.7764,
"theoretical_loss": 3.3176196906793516,
"tokens_seen": 3037200384
},
{
"epoch": 0.15,
"learning_rate": 0.0008571428571428571,
"loss": 2.625,
"theoretical_loss": 3.3176087103680056,
"tokens_seen": 3037331456
},
{
"epoch": 0.15,
"learning_rate": 0.0008567151411462789,
"loss": 2.6693,
"theoretical_loss": 3.3175977306631594,
"tokens_seen": 3037462528
},
{
"epoch": 0.15,
"learning_rate": 0.0008562874251497006,
"loss": 2.6201,
"theoretical_loss": 3.3175867515647526,
"tokens_seen": 3037593600
},
{
"epoch": 0.15,
"learning_rate": 0.0008558597091531223,
"loss": 2.5103,
"theoretical_loss": 3.317575773072726,
"tokens_seen": 3037724672
},
{
"epoch": 0.15,
"learning_rate": 0.0008554319931565441,
"loss": 2.4253,
"theoretical_loss": 3.3175647951870197,
"tokens_seen": 3037855744
},
{
"epoch": 0.15,
"objective/train/docs_used": 1665791,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.2302393913269043,
"objective/train/theoretical_loss": 3.317559306471518,
"objective/train/tokens_used": 67972576,
"theoretical_loss": 3.317559306471518,
"tokens_seen": 3037921280
},
{
"epoch": 0.15,
"learning_rate": 0.0008550042771599658,
"loss": 2.6287,
"theoretical_loss": 3.3175538179075743,
"tokens_seen": 3037986816
},
{
"epoch": 0.15,
"learning_rate": 0.0008545765611633876,
"loss": 2.5721,
"theoretical_loss": 3.31754284123433,
"tokens_seen": 3038117888
},
{
"epoch": 0.15,
"learning_rate": 0.0008541488451668093,
"loss": 2.5295,
"theoretical_loss": 3.3175318651672274,
"tokens_seen": 3038248960
},
{
"epoch": 0.15,
"learning_rate": 0.000853721129170231,
"loss": 2.6531,
"theoretical_loss": 3.3175208897062065,
"tokens_seen": 3038380032
},
{
"epoch": 0.16,
"learning_rate": 0.0008532934131736527,
"loss": 2.7652,
"theoretical_loss": 3.317509914851208,
"tokens_seen": 3038511104
},
{
"epoch": 0.16,
"learning_rate": 0.0008528656971770744,
"loss": 2.4813,
"theoretical_loss": 3.3174989406021718,
"tokens_seen": 3038642176
},
{
"epoch": 0.16,
"learning_rate": 0.0008524379811804962,
"loss": 2.7802,
"theoretical_loss": 3.317487966959039,
"tokens_seen": 3038773248
},
{
"epoch": 0.16,
"learning_rate": 0.0008520102651839178,
"loss": 2.6504,
"theoretical_loss": 3.3174769939217494,
"tokens_seen": 3038904320
},
{
"epoch": 0.16,
"learning_rate": 0.0008515825491873395,
"loss": 2.6791,
"theoretical_loss": 3.317466021490244,
"tokens_seen": 3039035392
},
{
"epoch": 0.16,
"learning_rate": 0.0008511548331907614,
"loss": 2.8812,
"theoretical_loss": 3.3174550496644626,
"tokens_seen": 3039166464
},
{
"epoch": 0.16,
"learning_rate": 0.0008507271171941831,
"loss": 2.6779,
"theoretical_loss": 3.317444078444346,
"tokens_seen": 3039297536
},
{
"epoch": 0.16,
"learning_rate": 0.0008502994011976049,
"loss": 2.6431,
"theoretical_loss": 3.317433107829835,
"tokens_seen": 3039428608
},
{
"epoch": 0.16,
"objective/train/docs_used": 1666368,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.657705307006836,
"objective/train/theoretical_loss": 3.31742213782087,
"objective/train/tokens_used": 69610976,
"theoretical_loss": 3.31742213782087,
"tokens_seen": 3039559680
},
{
"epoch": 0.16,
"learning_rate": 0.0008498716852010265,
"loss": 2.72,
"theoretical_loss": 3.31742213782087,
"tokens_seen": 3039559680
},
{
"epoch": 0.16,
"learning_rate": 0.0008494439692044482,
"loss": 2.589,
"theoretical_loss": 3.3174111684173906,
"tokens_seen": 3039690752
},
{
"epoch": 0.16,
"learning_rate": 0.00084901625320787,
"loss": 2.7118,
"theoretical_loss": 3.317400199619338,
"tokens_seen": 3039821824
},
{
"epoch": 0.16,
"learning_rate": 0.0008485885372112917,
"loss": 2.6093,
"theoretical_loss": 3.3173892314266524,
"tokens_seen": 3039952896
},
{
"epoch": 0.16,
"learning_rate": 0.0008481608212147135,
"loss": 2.724,
"theoretical_loss": 3.3173782638392746,
"tokens_seen": 3040083968
},
{
"epoch": 0.16,
"learning_rate": 0.0008477331052181351,
"loss": 2.6337,
"theoretical_loss": 3.3173672968571446,
"tokens_seen": 3040215040
},
{
"epoch": 0.16,
"learning_rate": 0.0008473053892215569,
"loss": 2.7281,
"theoretical_loss": 3.3173563304802034,
"tokens_seen": 3040346112
},
{
"epoch": 0.16,
"learning_rate": 0.0008468776732249787,
"loss": 2.6731,
"theoretical_loss": 3.3173453647083915,
"tokens_seen": 3040477184
},
{
"epoch": 0.16,
"learning_rate": 0.0008464499572284004,
"loss": 2.6926,
"theoretical_loss": 3.317334399541649,
"tokens_seen": 3040608256
},
{
"epoch": 0.16,
"learning_rate": 0.0008460222412318222,
"loss": 2.8351,
"theoretical_loss": 3.3173234349799166,
"tokens_seen": 3040739328
},
{
"epoch": 0.16,
"learning_rate": 0.0008455945252352438,
"loss": 2.6007,
"theoretical_loss": 3.317312471023135,
"tokens_seen": 3040870400
},
{
"epoch": 0.16,
"learning_rate": 0.0008451668092386655,
"loss": 2.7641,
"theoretical_loss": 3.3173015076712447,
"tokens_seen": 3041001472
},
{
"epoch": 0.16,
"learning_rate": 0.0008447390932420873,
"loss": 2.7725,
"theoretical_loss": 3.3172905449241865,
"tokens_seen": 3041132544
},
{
"epoch": 0.16,
"objective/train/docs_used": 1667402,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.6917831897735596,
"objective/train/theoretical_loss": 3.317285063777451,
"objective/train/tokens_used": 71249376,
"theoretical_loss": 3.317285063777451,
"tokens_seen": 3041198080
},
{
"epoch": 0.16,
"learning_rate": 0.000844311377245509,
"loss": 2.6852,
"theoretical_loss": 3.3172795827819,
"tokens_seen": 3041263616
},
{
"epoch": 0.16,
"learning_rate": 0.0008438836612489306,
"loss": 2.6591,
"theoretical_loss": 3.3172686212443274,
"tokens_seen": 3041394688
},
{
"epoch": 0.17,
"learning_rate": 0.0008434559452523525,
"loss": 2.74,
"theoretical_loss": 3.317257660311408,
"tokens_seen": 3041525760
},
{
"epoch": 0.17,
"learning_rate": 0.0008430282292557742,
"loss": 2.83,
"theoretical_loss": 3.3172466999830825,
"tokens_seen": 3041656832
},
{
"epoch": 0.17,
"learning_rate": 0.000842600513259196,
"loss": 2.6488,
"theoretical_loss": 3.317235740259292,
"tokens_seen": 3041787904
},
{
"epoch": 0.17,
"learning_rate": 0.0008421727972626177,
"loss": 2.5988,
"theoretical_loss": 3.3172247811399767,
"tokens_seen": 3041918976
},
{
"epoch": 0.17,
"learning_rate": 0.0008417450812660393,
"loss": 2.6984,
"theoretical_loss": 3.317213822625077,
"tokens_seen": 3042050048
},
{
"epoch": 0.17,
"learning_rate": 0.0008413173652694611,
"loss": 2.6926,
"theoretical_loss": 3.3172028647145346,
"tokens_seen": 3042181120
},
{
"epoch": 0.17,
"learning_rate": 0.0008408896492728828,
"loss": 2.6579,
"theoretical_loss": 3.317191907408289,
"tokens_seen": 3042312192
},
{
"epoch": 0.17,
"learning_rate": 0.0008404619332763046,
"loss": 2.6408,
"theoretical_loss": 3.3171809507062817,
"tokens_seen": 3042443264
},
{
"epoch": 0.17,
"learning_rate": 0.0008400342172797262,
"loss": 2.7066,
"theoretical_loss": 3.3171699946084523,
"tokens_seen": 3042574336
},
{
"epoch": 0.17,
"learning_rate": 0.0008396065012831479,
"loss": 2.7149,
"theoretical_loss": 3.3171590391147427,
"tokens_seen": 3042705408
},
{
"epoch": 0.17,
"objective/train/docs_used": 1668521,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.528013229370117,
"objective/train/theoretical_loss": 3.3171480842250927,
"objective/train/tokens_used": 72887776,
"theoretical_loss": 3.3171480842250927,
"tokens_seen": 3042836480
},
{
"epoch": 0.17,
"learning_rate": 0.0008391787852865698,
"loss": 2.6473,
"theoretical_loss": 3.3171480842250927,
"tokens_seen": 3042836480
},
{
"epoch": 0.17,
"learning_rate": 0.0008387510692899915,
"loss": 2.6857,
"theoretical_loss": 3.317137129939443,
"tokens_seen": 3042967552
},
{
"epoch": 0.17,
"learning_rate": 0.0008383233532934132,
"loss": 2.585,
"theoretical_loss": 3.3171261762577346,
"tokens_seen": 3043098624
},
{
"epoch": 0.17,
"learning_rate": 0.000837895637296835,
"loss": 2.6799,
"theoretical_loss": 3.3171152231799086,
"tokens_seen": 3043229696
},
{
"epoch": 0.17,
"learning_rate": 0.0008374679213002566,
"loss": 2.6625,
"theoretical_loss": 3.317104270705905,
"tokens_seen": 3043360768
},
{
"epoch": 0.17,
"learning_rate": 0.0008370402053036784,
"loss": 2.6606,
"theoretical_loss": 3.3170933188356644,
"tokens_seen": 3043491840
},
{
"epoch": 0.17,
"learning_rate": 0.0008366124893071001,
"loss": 2.7666,
"theoretical_loss": 3.3170823675691277,
"tokens_seen": 3043622912
},
{
"epoch": 0.17,
"learning_rate": 0.0008361847733105218,
"loss": 2.7028,
"theoretical_loss": 3.317071416906236,
"tokens_seen": 3043753984
},
{
"epoch": 0.17,
"learning_rate": 0.0008357570573139435,
"loss": 2.664,
"theoretical_loss": 3.3170604668469297,
"tokens_seen": 3043885056
},
{
"epoch": 0.17,
"learning_rate": 0.0008353293413173653,
"loss": 2.7291,
"theoretical_loss": 3.31704951739115,
"tokens_seen": 3044016128
},
{
"epoch": 0.17,
"learning_rate": 0.0008349016253207871,
"loss": 2.6537,
"theoretical_loss": 3.317038568538837,
"tokens_seen": 3044147200
},
{
"epoch": 0.17,
"learning_rate": 0.0008344739093242088,
"loss": 2.5645,
"theoretical_loss": 3.317027620289932,
"tokens_seen": 3044278272
},
{
"epoch": 0.17,
"learning_rate": 0.0008340461933276305,
"loss": 2.7629,
"theoretical_loss": 3.317016672644375,
"tokens_seen": 3044409344
},
{
"epoch": 0.17,
"objective/train/docs_used": 1668980,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.177739143371582,
"objective/train/theoretical_loss": 3.3170111990478337,
"objective/train/tokens_used": 74526176,
"theoretical_loss": 3.3170111990478337,
"tokens_seen": 3044474880
},
{
"epoch": 0.17,
"learning_rate": 0.0008336184773310522,
"loss": 2.5701,
"theoretical_loss": 3.3170057256021077,
"tokens_seen": 3044540416
},
{
"epoch": 0.18,
"learning_rate": 0.0008331907613344739,
"loss": 2.6897,
"theoretical_loss": 3.3169947791630703,
"tokens_seen": 3044671488
},
{
"epoch": 0.18,
"learning_rate": 0.0008327630453378957,
"loss": 2.7863,
"theoretical_loss": 3.3169838333272037,
"tokens_seen": 3044802560
},
{
"epoch": 0.18,
"learning_rate": 0.0008323353293413174,
"loss": 2.6692,
"theoretical_loss": 3.316972888094449,
"tokens_seen": 3044933632
},
{
"epoch": 0.18,
"learning_rate": 0.000831907613344739,
"loss": 2.6618,
"theoretical_loss": 3.3169619434647464,
"tokens_seen": 3045064704
},
{
"epoch": 0.18,
"learning_rate": 0.0008314798973481608,
"loss": 2.5543,
"theoretical_loss": 3.3169509994380375,
"tokens_seen": 3045195776
},
{
"epoch": 0.18,
"learning_rate": 0.0008310521813515826,
"loss": 2.7893,
"theoretical_loss": 3.3169400560142623,
"tokens_seen": 3045326848
},
{
"epoch": 0.18,
"learning_rate": 0.0008306244653550043,
"loss": 2.6742,
"theoretical_loss": 3.3169291131933623,
"tokens_seen": 3045457920
},
{
"epoch": 0.18,
"learning_rate": 0.0008301967493584261,
"loss": 2.6488,
"theoretical_loss": 3.316918170975278,
"tokens_seen": 3045588992
},
{
"epoch": 0.18,
"learning_rate": 0.0008297690333618477,
"loss": 2.5798,
"theoretical_loss": 3.31690722935995,
"tokens_seen": 3045720064
},
{
"epoch": 0.18,
"learning_rate": 0.0008293413173652695,
"loss": 2.7802,
"theoretical_loss": 3.3168962883473205,
"tokens_seen": 3045851136
},
{
"epoch": 0.18,
"learning_rate": 0.0008289136013686912,
"loss": 2.7032,
"theoretical_loss": 3.316885347937329,
"tokens_seen": 3045982208
},
{
"epoch": 0.18,
"objective/train/docs_used": 1670028,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.408548355102539,
"objective/train/theoretical_loss": 3.316874408129916,
"objective/train/tokens_used": 76164576,
"theoretical_loss": 3.316874408129916,
"tokens_seen": 3046113280
},
{
"epoch": 0.18,
"learning_rate": 0.0008284858853721129,
"loss": 2.7009,
"theoretical_loss": 3.316874408129916,
"tokens_seen": 3046113280
},
{
"epoch": 0.18,
"learning_rate": 0.0008280581693755347,
"loss": 2.6379,
"theoretical_loss": 3.316863468925024,
"tokens_seen": 3046244352
},
{
"epoch": 0.18,
"learning_rate": 0.0008276304533789563,
"loss": 2.7495,
"theoretical_loss": 3.3168525303225924,
"tokens_seen": 3046375424
},
{
"epoch": 0.18,
"learning_rate": 0.0008272027373823782,
"loss": 2.8538,
"theoretical_loss": 3.316841592322563,
"tokens_seen": 3046506496
},
{
"epoch": 0.18,
"learning_rate": 0.0008267750213857999,
"loss": 2.8283,
"theoretical_loss": 3.3168306549248765,
"tokens_seen": 3046637568
},
{
"epoch": 0.18,
"learning_rate": 0.0008263473053892216,
"loss": 2.6814,
"theoretical_loss": 3.316819718129474,
"tokens_seen": 3046768640
},
{
"epoch": 0.18,
"learning_rate": 0.0008259195893926434,
"loss": 2.6987,
"theoretical_loss": 3.3168087819362957,
"tokens_seen": 3046899712
},
{
"epoch": 0.18,
"learning_rate": 0.000825491873396065,
"loss": 2.6017,
"theoretical_loss": 3.316797846345283,
"tokens_seen": 3047030784
},
{
"epoch": 0.18,
"learning_rate": 0.0008250641573994867,
"loss": 2.7868,
"theoretical_loss": 3.316786911356377,
"tokens_seen": 3047161856
},
{
"epoch": 0.18,
"learning_rate": 0.0008246364414029085,
"loss": 2.5386,
"theoretical_loss": 3.316775976969519,
"tokens_seen": 3047292928
},
{
"epoch": 0.18,
"learning_rate": 0.0008242087254063302,
"loss": 2.806,
"theoretical_loss": 3.316765043184649,
"tokens_seen": 3047424000
},
{
"epoch": 0.18,
"learning_rate": 0.0008237810094097519,
"loss": 2.7629,
"theoretical_loss": 3.316754110001708,
"tokens_seen": 3047555072
},
{
"epoch": 0.19,
"learning_rate": 0.0008233532934131736,
"loss": 2.7679,
"theoretical_loss": 3.3167431774206384,
"tokens_seen": 3047686144
},
{
"epoch": 0.19,
"objective/train/docs_used": 1670628,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.13972806930542,
"objective/train/theoretical_loss": 3.316737711355786,
"objective/train/tokens_used": 77802976,
"theoretical_loss": 3.316737711355786,
"tokens_seen": 3047751680
},
{
"epoch": 0.19,
"learning_rate": 0.0008229255774165954,
"loss": 2.5883,
"theoretical_loss": 3.3167322454413792,
"tokens_seen": 3047817216
},
{
"epoch": 0.19,
"learning_rate": 0.0008224978614200172,
"loss": 2.7089,
"theoretical_loss": 3.316721314063873,
"tokens_seen": 3047948288
},
{
"epoch": 0.19,
"learning_rate": 0.0008220701454234389,
"loss": 2.7955,
"theoretical_loss": 3.3167103832880604,
"tokens_seen": 3048079360
},
{
"epoch": 0.19,
"learning_rate": 0.0008216424294268606,
"loss": 2.7126,
"theoretical_loss": 3.316699453113882,
"tokens_seen": 3048210432
},
{
"epoch": 0.19,
"learning_rate": 0.0008212147134302823,
"loss": 2.6486,
"theoretical_loss": 3.3166885235412784,
"tokens_seen": 3048341504
},
{
"epoch": 0.19,
"learning_rate": 0.000820786997433704,
"loss": 2.6465,
"theoretical_loss": 3.316677594570192,
"tokens_seen": 3048472576
},
{
"epoch": 0.19,
"learning_rate": 0.0008203592814371258,
"loss": 2.6168,
"theoretical_loss": 3.316666666200563,
"tokens_seen": 3048603648
},
{
"epoch": 0.19,
"learning_rate": 0.0008199315654405474,
"loss": 2.6939,
"theoretical_loss": 3.316655738432332,
"tokens_seen": 3048734720
},
{
"epoch": 0.19,
"learning_rate": 0.0008195038494439692,
"loss": 2.6707,
"theoretical_loss": 3.3166448112654408,
"tokens_seen": 3048865792
},
{
"epoch": 0.19,
"learning_rate": 0.000819076133447391,
"loss": 2.6281,
"theoretical_loss": 3.3166338846998302,
"tokens_seen": 3048996864
},
{
"epoch": 0.19,
"learning_rate": 0.0008186484174508127,
"loss": 2.7623,
"theoretical_loss": 3.316622958735442,
"tokens_seen": 3049127936
},
{
"epoch": 0.19,
"learning_rate": 0.0008182207014542345,
"loss": 2.7285,
"theoretical_loss": 3.3166120333722158,
"tokens_seen": 3049259008
},
{
"epoch": 0.19,
"objective/train/docs_used": 1671752,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.2737972736358643,
"objective/train/theoretical_loss": 3.3166011086100937,
"objective/train/tokens_used": 79441376,
"theoretical_loss": 3.3166011086100937,
"tokens_seen": 3049390080
},
{
"epoch": 0.19,
"learning_rate": 0.0008177929854576561,
"loss": 2.7943,
"theoretical_loss": 3.3166011086100937,
"tokens_seen": 3049390080
},
{
"epoch": 0.19,
"learning_rate": 0.0008173652694610778,
"loss": 2.6782,
"theoretical_loss": 3.3165901844490167,
"tokens_seen": 3049521152
},
{
"epoch": 0.19,
"learning_rate": 0.0008169375534644996,
"loss": 2.6773,
"theoretical_loss": 3.3165792608889255,
"tokens_seen": 3049652224
},
{
"epoch": 0.19,
"learning_rate": 0.0008165098374679213,
"loss": 2.7108,
"theoretical_loss": 3.3165683379297612,
"tokens_seen": 3049783296
},
{
"epoch": 0.19,
"learning_rate": 0.0008160821214713431,
"loss": 2.644,
"theoretical_loss": 3.3165574155714657,
"tokens_seen": 3049914368
},
{
"epoch": 0.19,
"learning_rate": 0.0008156544054747647,
"loss": 2.7035,
"theoretical_loss": 3.3165464938139797,
"tokens_seen": 3050045440
},
{
"epoch": 0.19,
"learning_rate": 0.0008152266894781864,
"loss": 2.6701,
"theoretical_loss": 3.3165355726572434,
"tokens_seen": 3050176512
},
{
"epoch": 0.19,
"learning_rate": 0.0008147989734816083,
"loss": 2.6013,
"theoretical_loss": 3.3165246521011995,
"tokens_seen": 3050307584
},
{
"epoch": 0.19,
"learning_rate": 0.00081437125748503,
"loss": 2.5665,
"theoretical_loss": 3.3165137321457885,
"tokens_seen": 3050438656
},
{
"epoch": 0.19,
"learning_rate": 0.0008139435414884518,
"loss": 2.5955,
"theoretical_loss": 3.3165028127909513,
"tokens_seen": 3050569728
},
{
"epoch": 0.19,
"learning_rate": 0.0008135158254918734,
"loss": 2.7426,
"theoretical_loss": 3.3164918940366293,
"tokens_seen": 3050700800
},
{
"epoch": 0.2,
"learning_rate": 0.0008130881094952951,
"loss": 2.6797,
"theoretical_loss": 3.3164809758827634,
"tokens_seen": 3050831872
},
{
"epoch": 0.2,
"learning_rate": 0.0008126603934987169,
"loss": 2.7291,
"theoretical_loss": 3.3164700583292954,
"tokens_seen": 3050962944
},
{
"epoch": 0.2,
"objective/train/docs_used": 1672176,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4254088401794434,
"objective/train/theoretical_loss": 3.316464599777692,
"objective/train/tokens_used": 81079776,
"theoretical_loss": 3.316464599777692,
"tokens_seen": 3051028480
},
{
"epoch": 0.2,
"learning_rate": 0.0008122326775021386,
"loss": 2.6885,
"theoretical_loss": 3.316459141376166,
"tokens_seen": 3051094016
},
{
"epoch": 0.2,
"learning_rate": 0.0008118049615055603,
"loss": 2.6706,
"theoretical_loss": 3.3164482250233163,
"tokens_seen": 3051225088
},
{
"epoch": 0.2,
"learning_rate": 0.000811377245508982,
"loss": 2.6934,
"theoretical_loss": 3.316437309270688,
"tokens_seen": 3051356160
},
{
"epoch": 0.2,
"learning_rate": 0.0008109495295124037,
"loss": 2.6764,
"theoretical_loss": 3.316426394118222,
"tokens_seen": 3051487232
},
{
"epoch": 0.2,
"learning_rate": 0.0008105218135158256,
"loss": 2.6293,
"theoretical_loss": 3.316415479565859,
"tokens_seen": 3051618304
},
{
"epoch": 0.2,
"learning_rate": 0.0008100940975192473,
"loss": 2.7261,
"theoretical_loss": 3.3164045656135417,
"tokens_seen": 3051749376
},
{
"epoch": 0.2,
"learning_rate": 0.0008096663815226689,
"loss": 2.6812,
"theoretical_loss": 3.3163936522612096,
"tokens_seen": 3051880448
},
{
"epoch": 0.2,
"learning_rate": 0.0008092386655260907,
"loss": 2.5557,
"theoretical_loss": 3.3163827395088052,
"tokens_seen": 3052011520
},
{
"epoch": 0.2,
"learning_rate": 0.0008088109495295124,
"loss": 2.6232,
"theoretical_loss": 3.3163718273562695,
"tokens_seen": 3052142592
},
{
"epoch": 0.2,
"learning_rate": 0.0008083832335329342,
"loss": 2.5755,
"theoretical_loss": 3.3163609158035436,
"tokens_seen": 3052273664
},
{
"epoch": 0.2,
"learning_rate": 0.0008079555175363559,
"loss": 2.5787,
"theoretical_loss": 3.3163500048505687,
"tokens_seen": 3052404736
},
{
"epoch": 0.2,
"learning_rate": 0.0008075278015397775,
"loss": 2.6212,
"theoretical_loss": 3.3163390944972857,
"tokens_seen": 3052535808
},
{
"epoch": 0.2,
"objective/train/docs_used": 1673386,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.2504236698150635,
"objective/train/theoretical_loss": 3.316328184743637,
"objective/train/tokens_used": 82718176,
"theoretical_loss": 3.316328184743637,
"tokens_seen": 3052666880
},
{
"epoch": 0.2,
"learning_rate": 0.0008071000855431993,
"loss": 2.5091,
"theoretical_loss": 3.316328184743637,
"tokens_seen": 3052666880
},
{
"epoch": 0.2,
"learning_rate": 0.0008066723695466211,
"loss": 2.7788,
"theoretical_loss": 3.3163172755895634,
"tokens_seen": 3052797952
},
{
"epoch": 0.2,
"learning_rate": 0.0008062446535500429,
"loss": 2.6996,
"theoretical_loss": 3.3163063670350055,
"tokens_seen": 3052929024
},
{
"epoch": 0.2,
"learning_rate": 0.0008058169375534645,
"loss": 2.639,
"theoretical_loss": 3.3162954590799054,
"tokens_seen": 3053060096
},
{
"epoch": 0.2,
"learning_rate": 0.0008053892215568862,
"loss": 2.7192,
"theoretical_loss": 3.316284551724204,
"tokens_seen": 3053191168
},
{
"epoch": 0.2,
"learning_rate": 0.000804961505560308,
"loss": 2.5435,
"theoretical_loss": 3.3162736449678434,
"tokens_seen": 3053322240
},
{
"epoch": 0.2,
"learning_rate": 0.0008045337895637297,
"loss": 2.7468,
"theoretical_loss": 3.3162627388107637,
"tokens_seen": 3053453312
},
{
"epoch": 0.2,
"learning_rate": 0.0008041060735671514,
"loss": 2.6621,
"theoretical_loss": 3.316251833252908,
"tokens_seen": 3053584384
},
{
"epoch": 0.2,
"learning_rate": 0.0008036783575705731,
"loss": 2.6691,
"theoretical_loss": 3.3162409282942154,
"tokens_seen": 3053715456
},
{
"epoch": 0.2,
"learning_rate": 0.0008032506415739948,
"loss": 2.6699,
"theoretical_loss": 3.316230023934629,
"tokens_seen": 3053846528
},
{
"epoch": 0.21,
"learning_rate": 0.0008028229255774167,
"loss": 2.711,
"theoretical_loss": 3.3162191201740896,
"tokens_seen": 3053977600
},
{
"epoch": 0.21,
"learning_rate": 0.0008023952095808384,
"loss": 2.6899,
"theoretical_loss": 3.3162082170125387,
"tokens_seen": 3054108672
},
{
"epoch": 0.21,
"learning_rate": 0.00080196749358426,
"loss": 2.4566,
"theoretical_loss": 3.3161973144499175,
"tokens_seen": 3054239744
},
{
"epoch": 0.21,
"objective/train/docs_used": 1673816,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.680955171585083,
"objective/train/theoretical_loss": 3.316191863393187,
"objective/train/tokens_used": 84356576,
"theoretical_loss": 3.316191863393187,
"tokens_seen": 3054305280
},
{
"epoch": 0.21,
"learning_rate": 0.0008015397775876818,
"loss": 2.5931,
"theoretical_loss": 3.3161864124861675,
"tokens_seen": 3054370816
},
{
"epoch": 0.21,
"learning_rate": 0.0008011120615911035,
"loss": 2.6795,
"theoretical_loss": 3.31617551112123,
"tokens_seen": 3054501888
},
{
"epoch": 0.21,
"learning_rate": 0.0008006843455945253,
"loss": 2.7542,
"theoretical_loss": 3.316164610355047,
"tokens_seen": 3054632960
},
{
"epoch": 0.21,
"learning_rate": 0.000800256629597947,
"loss": 2.6446,
"theoretical_loss": 3.316153710187559,
"tokens_seen": 3054764032
},
{
"epoch": 0.21,
"learning_rate": 0.0007998289136013686,
"loss": 2.7551,
"theoretical_loss": 3.316142810618708,
"tokens_seen": 3054895104
},
{
"epoch": 0.21,
"learning_rate": 0.0007994011976047904,
"loss": 2.7207,
"theoretical_loss": 3.3161319116484353,
"tokens_seen": 3055026176
},
{
"epoch": 0.21,
"learning_rate": 0.0007989734816082121,
"loss": 2.8592,
"theoretical_loss": 3.3161210132766823,
"tokens_seen": 3055157248
},
{
"epoch": 0.21,
"learning_rate": 0.000798545765611634,
"loss": 2.8129,
"theoretical_loss": 3.316110115503391,
"tokens_seen": 3055288320
},
{
"epoch": 0.21,
"learning_rate": 0.0007981180496150557,
"loss": 2.6611,
"theoretical_loss": 3.316099218328502,
"tokens_seen": 3055419392
},
{
"epoch": 0.21,
"learning_rate": 0.0007976903336184773,
"loss": 2.7847,
"theoretical_loss": 3.3160883217519572,
"tokens_seen": 3055550464
},
{
"epoch": 0.21,
"learning_rate": 0.0007972626176218991,
"loss": 2.5896,
"theoretical_loss": 3.316077425773698,
"tokens_seen": 3055681536
},
{
"epoch": 0.21,
"learning_rate": 0.0007968349016253208,
"loss": 2.699,
"theoretical_loss": 3.316066530393666,
"tokens_seen": 3055812608
},
{
"epoch": 0.21,
"objective/train/docs_used": 1674780,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 3.0617194175720215,
"objective/train/theoretical_loss": 3.3160556356118027,
"objective/train/tokens_used": 85994976,
"theoretical_loss": 3.3160556356118027,
"tokens_seen": 3055943680
},
{
"epoch": 0.21,
"learning_rate": 0.0007964071856287425,
"loss": 2.7631,
"theoretical_loss": 3.3160556356118027,
"tokens_seen": 3055943680
},
{
"epoch": 0.21,
"learning_rate": 0.0007959794696321643,
"loss": 2.8357,
"theoretical_loss": 3.3160447414280494,
"tokens_seen": 3056074752
},
{
"epoch": 0.21,
"learning_rate": 0.0007955517536355859,
"loss": 2.6724,
"theoretical_loss": 3.316033847842348,
"tokens_seen": 3056205824
},
{
"epoch": 0.21,
"learning_rate": 0.0007951240376390077,
"loss": 2.6585,
"theoretical_loss": 3.3160229548546396,
"tokens_seen": 3056336896
},
{
"epoch": 0.21,
"learning_rate": 0.0007946963216424295,
"loss": 2.6504,
"theoretical_loss": 3.316012062464866,
"tokens_seen": 3056467968
},
{
"epoch": 0.21,
"learning_rate": 0.0007942686056458512,
"loss": 2.7904,
"theoretical_loss": 3.316001170672968,
"tokens_seen": 3056599040
},
{
"epoch": 0.21,
"learning_rate": 0.000793840889649273,
"loss": 2.6297,
"theoretical_loss": 3.3159902794788887,
"tokens_seen": 3056730112
},
{
"epoch": 0.21,
"learning_rate": 0.0007934131736526946,
"loss": 2.6602,
"theoretical_loss": 3.3159793888825684,
"tokens_seen": 3056861184
},
{
"epoch": 0.22,
"learning_rate": 0.0007929854576561164,
"loss": 2.7146,
"theoretical_loss": 3.315968498883949,
"tokens_seen": 3056992256
},
{
"epoch": 0.22,
"learning_rate": 0.0007925577416595381,
"loss": 2.681,
"theoretical_loss": 3.3159576094829726,
"tokens_seen": 3057123328
},
{
"epoch": 0.22,
"learning_rate": 0.0007921300256629598,
"loss": 2.8578,
"theoretical_loss": 3.3159467206795794,
"tokens_seen": 3057254400
},
{
"epoch": 0.22,
"learning_rate": 0.0007917023096663815,
"loss": 2.6611,
"theoretical_loss": 3.3159358324737123,
"tokens_seen": 3057385472
},
{
"epoch": 0.22,
"learning_rate": 0.0007912745936698032,
"loss": 2.5705,
"theoretical_loss": 3.3159249448653125,
"tokens_seen": 3057516544
},
{
"epoch": 0.22,
"objective/train/docs_used": 1675938,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.6079812049865723,
"objective/train/theoretical_loss": 3.3159195012851446,
"objective/train/tokens_used": 87633376,
"theoretical_loss": 3.3159195012851446,
"tokens_seen": 3057582080
},
{
"epoch": 0.22,
"learning_rate": 0.0007908468776732249,
"loss": 2.6312,
"theoretical_loss": 3.3159140578543216,
"tokens_seen": 3057647616
},
{
"epoch": 0.22,
"learning_rate": 0.0007904191616766468,
"loss": 2.6494,
"theoretical_loss": 3.315903171440681,
"tokens_seen": 3057778688
},
{
"epoch": 0.22,
"learning_rate": 0.0007899914456800685,
"loss": 2.5289,
"theoretical_loss": 3.3158922856243325,
"tokens_seen": 3057909760
},
{
"epoch": 0.22,
"learning_rate": 0.0007895637296834902,
"loss": 2.6143,
"theoretical_loss": 3.3158814004052175,
"tokens_seen": 3058040832
},
{
"epoch": 0.22,
"learning_rate": 0.0007891360136869119,
"loss": 2.6765,
"theoretical_loss": 3.3158705157832786,
"tokens_seen": 3058171904
},
{
"epoch": 0.22,
"learning_rate": 0.0007887082976903336,
"loss": 2.6434,
"theoretical_loss": 3.315859631758456,
"tokens_seen": 3058302976
},
{
"epoch": 0.22,
"learning_rate": 0.0007882805816937554,
"loss": 2.7441,
"theoretical_loss": 3.3158487483306924,
"tokens_seen": 3058434048
},
{
"epoch": 0.22,
"learning_rate": 0.000787852865697177,
"loss": 2.5689,
"theoretical_loss": 3.3158378654999288,
"tokens_seen": 3058565120
},
{
"epoch": 0.22,
"learning_rate": 0.0007874251497005988,
"loss": 2.67,
"theoretical_loss": 3.315826983266107,
"tokens_seen": 3058696192
},
{
"epoch": 0.22,
"learning_rate": 0.0007869974337040205,
"loss": 2.746,
"theoretical_loss": 3.31581610162917,
"tokens_seen": 3058827264
},
{
"epoch": 0.22,
"learning_rate": 0.0007865697177074422,
"loss": 2.6029,
"theoretical_loss": 3.3158052205890574,
"tokens_seen": 3058958336
},
{
"epoch": 0.22,
"learning_rate": 0.0007861420017108641,
"loss": 2.5993,
"theoretical_loss": 3.315794340145712,
"tokens_seen": 3059089408
},
{
"epoch": 0.22,
"objective/train/docs_used": 1676633,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 3.0798161029815674,
"objective/train/theoretical_loss": 3.3157834602990754,
"objective/train/tokens_used": 89271776,
"theoretical_loss": 3.3157834602990754,
"tokens_seen": 3059220480
},
{
"epoch": 0.22,
"learning_rate": 0.0007857142857142857,
"loss": 2.6821,
"theoretical_loss": 3.3157834602990754,
"tokens_seen": 3059220480
},
{
"epoch": 0.22,
"learning_rate": 0.0007852865697177075,
"loss": 2.6591,
"theoretical_loss": 3.3157725810490892,
"tokens_seen": 3059351552
},
{
"epoch": 0.22,
"learning_rate": 0.0007848588537211292,
"loss": 2.5473,
"theoretical_loss": 3.3157617023956956,
"tokens_seen": 3059482624
},
{
"epoch": 0.22,
"learning_rate": 0.0007844311377245509,
"loss": 2.737,
"theoretical_loss": 3.3157508243388354,
"tokens_seen": 3059613696
},
{
"epoch": 0.22,
"learning_rate": 0.0007840034217279727,
"loss": 2.5373,
"theoretical_loss": 3.315739946878451,
"tokens_seen": 3059744768
},
{
"epoch": 0.22,
"learning_rate": 0.0007835757057313943,
"loss": 2.7326,
"theoretical_loss": 3.3157290700144837,
"tokens_seen": 3059875840
},
{
"epoch": 0.22,
"learning_rate": 0.000783147989734816,
"loss": 2.8193,
"theoretical_loss": 3.315718193746876,
"tokens_seen": 3060006912
},
{
"epoch": 0.23,
"learning_rate": 0.0007827202737382378,
"loss": 2.8582,
"theoretical_loss": 3.315707318075569,
"tokens_seen": 3060137984
},
{
"epoch": 0.23,
"learning_rate": 0.0007822925577416596,
"loss": 2.6654,
"theoretical_loss": 3.3156964430005047,
"tokens_seen": 3060269056
},
{
"epoch": 0.23,
"learning_rate": 0.0007818648417450814,
"loss": 2.7077,
"theoretical_loss": 3.315685568521625,
"tokens_seen": 3060400128
},
{
"epoch": 0.23,
"learning_rate": 0.000781437125748503,
"loss": 2.761,
"theoretical_loss": 3.3156746946388713,
"tokens_seen": 3060531200
},
{
"epoch": 0.23,
"learning_rate": 0.0007810094097519247,
"loss": 2.6981,
"theoretical_loss": 3.315663821352186,
"tokens_seen": 3060662272
},
{
"epoch": 0.23,
"learning_rate": 0.0007805816937553465,
"loss": 2.8834,
"theoretical_loss": 3.3156529486615103,
"tokens_seen": 3060793344
},
{
"epoch": 0.23,
"objective/train/docs_used": 1677094,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.535158634185791,
"objective/train/theoretical_loss": 3.315647512539658,
"objective/train/tokens_used": 90910176,
"theoretical_loss": 3.315647512539658,
"tokens_seen": 3060858880
},
{
"epoch": 0.23,
"learning_rate": 0.0007801539777587682,
"loss": 2.7827,
"theoretical_loss": 3.3156420765667862,
"tokens_seen": 3060924416
},
{
"epoch": 0.23,
"learning_rate": 0.00077972626176219,
"loss": 2.7269,
"theoretical_loss": 3.3156312050679553,
"tokens_seen": 3061055488
},
{
"epoch": 0.23,
"learning_rate": 0.0007792985457656116,
"loss": 2.751,
"theoretical_loss": 3.31562033416496,
"tokens_seen": 3061186560
},
{
"epoch": 0.23,
"learning_rate": 0.0007788708297690333,
"loss": 2.7144,
"theoretical_loss": 3.315609463857742,
"tokens_seen": 3061317632
},
{
"epoch": 0.23,
"learning_rate": 0.0007784431137724552,
"loss": 2.6884,
"theoretical_loss": 3.3155985941462425,
"tokens_seen": 3061448704
},
{
"epoch": 0.23,
"learning_rate": 0.0007780153977758769,
"loss": 2.6604,
"theoretical_loss": 3.315587725030404,
"tokens_seen": 3061579776
},
{
"epoch": 0.23,
"learning_rate": 0.0007775876817792986,
"loss": 2.7039,
"theoretical_loss": 3.315576856510168,
"tokens_seen": 3061710848
},
{
"epoch": 0.23,
"learning_rate": 0.0007771599657827203,
"loss": 2.6965,
"theoretical_loss": 3.315565988585477,
"tokens_seen": 3061841920
},
{
"epoch": 0.23,
"learning_rate": 0.000776732249786142,
"loss": 2.7133,
"theoretical_loss": 3.3155551212562724,
"tokens_seen": 3061972992
},
{
"epoch": 0.23,
"learning_rate": 0.0007763045337895638,
"loss": 2.8031,
"theoretical_loss": 3.3155442545224956,
"tokens_seen": 3062104064
},
{
"epoch": 0.23,
"learning_rate": 0.0007758768177929855,
"loss": 2.6672,
"theoretical_loss": 3.315533388384089,
"tokens_seen": 3062235136
},
{
"epoch": 0.23,
"learning_rate": 0.0007754491017964071,
"loss": 2.7976,
"theoretical_loss": 3.315522522840995,
"tokens_seen": 3062366208
},
{
"epoch": 0.23,
"objective/train/docs_used": 1678209,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4061381816864014,
"objective/train/theoretical_loss": 3.3155116578931545,
"objective/train/tokens_used": 92548576,
"theoretical_loss": 3.3155116578931545,
"tokens_seen": 3062497280
},
{
"epoch": 0.23,
"learning_rate": 0.0007750213857998289,
"loss": 2.4624,
"theoretical_loss": 3.3155116578931545,
"tokens_seen": 3062497280
},
{
"epoch": 0.23,
"learning_rate": 0.0007745936698032506,
"loss": 2.7517,
"theoretical_loss": 3.31550079354051,
"tokens_seen": 3062628352
},
{
"epoch": 0.23,
"learning_rate": 0.0007741659538066725,
"loss": 2.7074,
"theoretical_loss": 3.315489929783004,
"tokens_seen": 3062759424
},
{
"epoch": 0.23,
"learning_rate": 0.0007737382378100942,
"loss": 2.7777,
"theoretical_loss": 3.315479066620577,
"tokens_seen": 3062890496
},
{
"epoch": 0.23,
"learning_rate": 0.0007733105218135158,
"loss": 2.7602,
"theoretical_loss": 3.3154682040531718,
"tokens_seen": 3063021568
},
{
"epoch": 0.23,
"learning_rate": 0.0007728828058169376,
"loss": 2.6515,
"theoretical_loss": 3.3154573420807303,
"tokens_seen": 3063152640
},
{
"epoch": 0.24,
"learning_rate": 0.0007724550898203593,
"loss": 2.7385,
"theoretical_loss": 3.3154464807031943,
"tokens_seen": 3063283712
},
{
"epoch": 0.24,
"learning_rate": 0.0007720273738237811,
"loss": 2.6795,
"theoretical_loss": 3.315435619920506,
"tokens_seen": 3063414784
},
{
"epoch": 0.24,
"learning_rate": 0.0007715996578272027,
"loss": 2.6428,
"theoretical_loss": 3.315424759732607,
"tokens_seen": 3063545856
},
{
"epoch": 0.24,
"learning_rate": 0.0007711719418306244,
"loss": 2.6636,
"theoretical_loss": 3.31541390013944,
"tokens_seen": 3063676928
},
{
"epoch": 0.24,
"learning_rate": 0.0007707442258340462,
"loss": 2.6116,
"theoretical_loss": 3.3154030411409465,
"tokens_seen": 3063808000
},
{
"epoch": 0.24,
"learning_rate": 0.0007703165098374679,
"loss": 2.6947,
"theoretical_loss": 3.3153921827370683,
"tokens_seen": 3063939072
},
{
"epoch": 0.24,
"learning_rate": 0.0007698887938408897,
"loss": 2.6784,
"theoretical_loss": 3.3153813249277473,
"tokens_seen": 3064070144
},
{
"epoch": 0.24,
"objective/train/docs_used": 1678783,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.995398759841919,
"objective/train/theoretical_loss": 3.315375896246028,
"objective/train/tokens_used": 94186976,
"theoretical_loss": 3.315375896246028,
"tokens_seen": 3064135680
},
{
"epoch": 0.24,
"learning_rate": 0.0007694610778443114,
"loss": 2.712,
"theoretical_loss": 3.3153704677129263,
"tokens_seen": 3064201216
},
{
"epoch": 0.24,
"learning_rate": 0.0007690333618477331,
"loss": 2.6632,
"theoretical_loss": 3.3153596110925467,
"tokens_seen": 3064332288
},
{
"epoch": 0.24,
"learning_rate": 0.0007686056458511549,
"loss": 2.6325,
"theoretical_loss": 3.315348755066551,
"tokens_seen": 3064463360
},
{
"epoch": 0.24,
"learning_rate": 0.0007681779298545766,
"loss": 2.611,
"theoretical_loss": 3.31533789963488,
"tokens_seen": 3064594432
},
{
"epoch": 0.24,
"learning_rate": 0.0007677502138579982,
"loss": 2.5939,
"theoretical_loss": 3.3153270447974776,
"tokens_seen": 3064725504
},
{
"epoch": 0.24,
"learning_rate": 0.00076732249786142,
"loss": 2.7007,
"theoretical_loss": 3.3153161905542845,
"tokens_seen": 3064856576
},
{
"epoch": 0.24,
"learning_rate": 0.0007668947818648417,
"loss": 2.6422,
"theoretical_loss": 3.315305336905243,
"tokens_seen": 3064987648
},
{
"epoch": 0.24,
"learning_rate": 0.0007664670658682635,
"loss": 2.6363,
"theoretical_loss": 3.3152944838502956,
"tokens_seen": 3065118720
},
{
"epoch": 0.24,
"learning_rate": 0.0007660393498716853,
"loss": 2.7748,
"theoretical_loss": 3.3152836313893843,
"tokens_seen": 3065249792
},
{
"epoch": 0.24,
"learning_rate": 0.0007656116338751069,
"loss": 2.8551,
"theoretical_loss": 3.3152727795224504,
"tokens_seen": 3065380864
},
{
"epoch": 0.24,
"learning_rate": 0.0007651839178785287,
"loss": 2.6915,
"theoretical_loss": 3.3152619282494373,
"tokens_seen": 3065511936
},
{
"epoch": 0.24,
"learning_rate": 0.0007647562018819504,
"loss": 2.4881,
"theoretical_loss": 3.315251077570286,
"tokens_seen": 3065643008
},
{
"epoch": 0.24,
"objective/train/docs_used": 1679875,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.7984094619750977,
"objective/train/theoretical_loss": 3.3152402274849395,
"objective/train/tokens_used": 95825376,
"theoretical_loss": 3.3152402274849395,
"tokens_seen": 3065774080
},
{
"epoch": 0.24,
"learning_rate": 0.0007643284858853722,
"loss": 2.7687,
"theoretical_loss": 3.3152402274849395,
"tokens_seen": 3065774080
},
{
"epoch": 0.24,
"learning_rate": 0.0007639007698887939,
"loss": 2.5306,
"theoretical_loss": 3.315229377993339,
"tokens_seen": 3065905152
},
{
"epoch": 0.24,
"learning_rate": 0.0007634730538922155,
"loss": 2.5974,
"theoretical_loss": 3.315218529095427,
"tokens_seen": 3066036224
},
{
"epoch": 0.24,
"learning_rate": 0.0007630453378956373,
"loss": 2.6238,
"theoretical_loss": 3.315207680791146,
"tokens_seen": 3066167296
},
{
"epoch": 0.25,
"learning_rate": 0.000762617621899059,
"loss": 2.6609,
"theoretical_loss": 3.3151968330804378,
"tokens_seen": 3066298368
},
{
"epoch": 0.25,
"learning_rate": 0.0007621899059024807,
"loss": 2.6483,
"theoretical_loss": 3.3151859859632444,
"tokens_seen": 3066429440
},
{
"epoch": 0.25,
"learning_rate": 0.0007617621899059026,
"loss": 2.5367,
"theoretical_loss": 3.3151751394395084,
"tokens_seen": 3066560512
},
{
"epoch": 0.25,
"learning_rate": 0.0007613344739093242,
"loss": 2.8444,
"theoretical_loss": 3.3151642935091714,
"tokens_seen": 3066691584
},
{
"epoch": 0.25,
"learning_rate": 0.000760906757912746,
"loss": 2.5866,
"theoretical_loss": 3.3151534481721763,
"tokens_seen": 3066822656
},
{
"epoch": 0.25,
"learning_rate": 0.0007604790419161677,
"loss": 2.7557,
"theoretical_loss": 3.3151426034284643,
"tokens_seen": 3066953728
},
{
"epoch": 0.25,
"learning_rate": 0.0007600513259195894,
"loss": 2.8116,
"theoretical_loss": 3.3151317592779788,
"tokens_seen": 3067084800
},
{
"epoch": 0.25,
"learning_rate": 0.0007596236099230111,
"loss": 2.7216,
"theoretical_loss": 3.315120915720661,
"tokens_seen": 3067215872
},
{
"epoch": 0.25,
"learning_rate": 0.0007591958939264328,
"loss": 2.6847,
"theoretical_loss": 3.315110072756454,
"tokens_seen": 3067346944
},
{
"epoch": 0.25,
"objective/train/docs_used": 1680552,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.6995866298675537,
"objective/train/theoretical_loss": 3.3151046514967484,
"objective/train/tokens_used": 97463776,
"theoretical_loss": 3.3151046514967484,
"tokens_seen": 3067412480
},
{
"epoch": 0.25,
"learning_rate": 0.0007587681779298546,
"loss": 2.5473,
"theoretical_loss": 3.315099230385299,
"tokens_seen": 3067478016
},
{
"epoch": 0.25,
"learning_rate": 0.0007583404619332763,
"loss": 2.6589,
"theoretical_loss": 3.3150883886071387,
"tokens_seen": 3067609088
},
{
"epoch": 0.25,
"learning_rate": 0.0007579127459366981,
"loss": 2.5557,
"theoretical_loss": 3.3150775474219154,
"tokens_seen": 3067740160
},
{
"epoch": 0.25,
"learning_rate": 0.0007574850299401198,
"loss": 2.6966,
"theoretical_loss": 3.3150667068295716,
"tokens_seen": 3067871232
},
{
"epoch": 0.25,
"learning_rate": 0.0007570573139435415,
"loss": 2.6355,
"theoretical_loss": 3.3150558668300487,
"tokens_seen": 3068002304
},
{
"epoch": 0.25,
"learning_rate": 0.0007566295979469632,
"loss": 2.6891,
"theoretical_loss": 3.31504502742329,
"tokens_seen": 3068133376
},
{
"epoch": 0.25,
"learning_rate": 0.000756201881950385,
"loss": 2.6008,
"theoretical_loss": 3.3150341886092374,
"tokens_seen": 3068264448
},
{
"epoch": 0.25,
"learning_rate": 0.0007557741659538067,
"loss": 2.6525,
"theoretical_loss": 3.3150233503878326,
"tokens_seen": 3068395520
},
{
"epoch": 0.25,
"learning_rate": 0.0007553464499572284,
"loss": 2.7266,
"theoretical_loss": 3.3150125127590186,
"tokens_seen": 3068526592
},
{
"epoch": 0.25,
"learning_rate": 0.0007549187339606501,
"loss": 2.6693,
"theoretical_loss": 3.3150016757227374,
"tokens_seen": 3068657664
},
{
"epoch": 0.25,
"learning_rate": 0.0007544910179640718,
"loss": 2.5191,
"theoretical_loss": 3.314990839278931,
"tokens_seen": 3068788736
},
{
"epoch": 0.25,
"learning_rate": 0.0007540633019674937,
"loss": 2.5744,
"theoretical_loss": 3.314980003427542,
"tokens_seen": 3068919808
},
{
"epoch": 0.25,
"objective/train/docs_used": 1681818,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.981553554534912,
"objective/train/theoretical_loss": 3.3149691681685134,
"objective/train/tokens_used": 99102176,
"theoretical_loss": 3.3149691681685134,
"tokens_seen": 3069050880
},
{
"epoch": 0.25,
"learning_rate": 0.0007536355859709153,
"loss": 2.6833,
"theoretical_loss": 3.3149691681685134,
"tokens_seen": 3069050880
},
{
"epoch": 0.25,
"learning_rate": 0.0007532078699743371,
"loss": 2.7529,
"theoretical_loss": 3.314958333501786,
"tokens_seen": 3069181952
},
{
"epoch": 0.25,
"learning_rate": 0.0007527801539777588,
"loss": 2.5656,
"theoretical_loss": 3.3149474994273036,
"tokens_seen": 3069313024
},
{
"epoch": 0.26,
"learning_rate": 0.0007523524379811805,
"loss": 2.7531,
"theoretical_loss": 3.314936665945008,
"tokens_seen": 3069444096
},
{
"epoch": 0.26,
"learning_rate": 0.0007519247219846023,
"loss": 2.6828,
"theoretical_loss": 3.314925833054841,
"tokens_seen": 3069575168
},
{
"epoch": 0.26,
"learning_rate": 0.0007514970059880239,
"loss": 2.6544,
"theoretical_loss": 3.3149150007567454,
"tokens_seen": 3069706240
},
{
"epoch": 0.26,
"learning_rate": 0.0007510692899914457,
"loss": 2.6764,
"theoretical_loss": 3.314904169050664,
"tokens_seen": 3069837312
},
{
"epoch": 0.26,
"learning_rate": 0.0007506415739948674,
"loss": 2.6181,
"theoretical_loss": 3.3148933379365384,
"tokens_seen": 3069968384
},
{
"epoch": 0.26,
"learning_rate": 0.0007502138579982891,
"loss": 2.5999,
"theoretical_loss": 3.3148825074143113,
"tokens_seen": 3070099456
},
{
"epoch": 0.26,
"learning_rate": 0.000749786142001711,
"loss": 2.5656,
"theoretical_loss": 3.314871677483925,
"tokens_seen": 3070230528
},
{
"epoch": 0.26,
"learning_rate": 0.0007493584260051326,
"loss": 2.6492,
"theoretical_loss": 3.3148608481453223,
"tokens_seen": 3070361600
},
{
"epoch": 0.26,
"learning_rate": 0.0007489307100085543,
"loss": 2.6454,
"theoretical_loss": 3.3148500193984454,
"tokens_seen": 3070492672
},
{
"epoch": 0.26,
"learning_rate": 0.0007485029940119761,
"loss": 2.6889,
"theoretical_loss": 3.3148391912432364,
"tokens_seen": 3070623744
},
{
"epoch": 0.26,
"objective/train/docs_used": 1682472,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.804523468017578,
"objective/train/theoretical_loss": 3.3148337773874896,
"objective/train/tokens_used": 100740576,
"theoretical_loss": 3.3148337773874896,
"tokens_seen": 3070689280
},
{
"epoch": 0.26,
"learning_rate": 0.0007480752780153978,
"loss": 2.6604,
"theoretical_loss": 3.3148283636796383,
"tokens_seen": 3070754816
},
{
"epoch": 0.26,
"learning_rate": 0.0007476475620188196,
"loss": 2.6386,
"theoretical_loss": 3.3148175367075927,
"tokens_seen": 3070885888
},
{
"epoch": 0.26,
"learning_rate": 0.0007472198460222412,
"loss": 2.6414,
"theoretical_loss": 3.314806710327043,
"tokens_seen": 3071016960
},
{
"epoch": 0.26,
"learning_rate": 0.0007467921300256629,
"loss": 2.8612,
"theoretical_loss": 3.3147958845379306,
"tokens_seen": 3071148032
},
{
"epoch": 0.26,
"learning_rate": 0.0007463644140290847,
"loss": 2.5636,
"theoretical_loss": 3.3147850593401986,
"tokens_seen": 3071279104
},
{
"epoch": 0.26,
"learning_rate": 0.0007459366980325064,
"loss": 2.7602,
"theoretical_loss": 3.3147742347337896,
"tokens_seen": 3071410176
},
{
"epoch": 0.26,
"learning_rate": 0.0007455089820359282,
"loss": 2.8901,
"theoretical_loss": 3.3147634107186454,
"tokens_seen": 3071541248
},
{
"epoch": 0.26,
"learning_rate": 0.0007450812660393499,
"loss": 2.6674,
"theoretical_loss": 3.3147525872947092,
"tokens_seen": 3071672320
},
{
"epoch": 0.26,
"learning_rate": 0.0007446535500427716,
"loss": 2.5025,
"theoretical_loss": 3.3147417644619233,
"tokens_seen": 3071803392
},
{
"epoch": 0.26,
"learning_rate": 0.0007442258340461934,
"loss": 2.6975,
"theoretical_loss": 3.31473094222023,
"tokens_seen": 3071934464
},
{
"epoch": 0.26,
"learning_rate": 0.0007437981180496151,
"loss": 2.8009,
"theoretical_loss": 3.3147201205695715,
"tokens_seen": 3072065536
},
{
"epoch": 0.26,
"learning_rate": 0.0007433704020530368,
"loss": 2.5947,
"theoretical_loss": 3.3147092995098912,
"tokens_seen": 3072196608
},
{
"epoch": 0.26,
"objective/train/docs_used": 1683394,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.7559075355529785,
"objective/train/theoretical_loss": 3.3146984790411307,
"objective/train/tokens_used": 102378976,
"theoretical_loss": 3.3146984790411307,
"tokens_seen": 3072327680
},
{
"epoch": 0.26,
"learning_rate": 0.0007429426860564585,
"loss": 2.6029,
"theoretical_loss": 3.3146984790411307,
"tokens_seen": 3072327680
},
{
"epoch": 0.27,
"learning_rate": 0.0007425149700598802,
"loss": 2.7892,
"theoretical_loss": 3.314687659163233,
"tokens_seen": 3072458752
},
{
"epoch": 0.27,
"learning_rate": 0.000742087254063302,
"loss": 2.4519,
"theoretical_loss": 3.3146768398761406,
"tokens_seen": 3072589824
},
{
"epoch": 0.27,
"learning_rate": 0.0007416595380667238,
"loss": 2.7062,
"theoretical_loss": 3.3146660211797956,
"tokens_seen": 3072720896
},
{
"epoch": 0.27,
"learning_rate": 0.0007412318220701454,
"loss": 2.6081,
"theoretical_loss": 3.3146552030741416,
"tokens_seen": 3072851968
},
{
"epoch": 0.27,
"learning_rate": 0.0007408041060735672,
"loss": 2.6677,
"theoretical_loss": 3.31464438555912,
"tokens_seen": 3072983040
},
{
"epoch": 0.27,
"learning_rate": 0.0007403763900769889,
"loss": 2.5899,
"theoretical_loss": 3.314633568634674,
"tokens_seen": 3073114112
},
{
"epoch": 0.27,
"learning_rate": 0.0007399486740804107,
"loss": 2.7511,
"theoretical_loss": 3.314622752300746,
"tokens_seen": 3073245184
},
{
"epoch": 0.27,
"learning_rate": 0.0007395209580838323,
"loss": 2.6366,
"theoretical_loss": 3.3146119365572786,
"tokens_seen": 3073376256
},
{
"epoch": 0.27,
"learning_rate": 0.000739093242087254,
"loss": 2.5128,
"theoretical_loss": 3.3146011214042144,
"tokens_seen": 3073507328
},
{
"epoch": 0.27,
"learning_rate": 0.0007386655260906758,
"loss": 2.6395,
"theoretical_loss": 3.314590306841496,
"tokens_seen": 3073638400
},
{
"epoch": 0.27,
"learning_rate": 0.0007382378100940975,
"loss": 2.808,
"theoretical_loss": 3.3145794928690657,
"tokens_seen": 3073769472
},
{
"epoch": 0.27,
"learning_rate": 0.0007378100940975194,
"loss": 2.7223,
"theoretical_loss": 3.3145686794868667,
"tokens_seen": 3073900544
},
{
"epoch": 0.27,
"objective/train/docs_used": 1684169,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.4948744773864746,
"objective/train/theoretical_loss": 3.314563273017086,
"objective/train/tokens_used": 104017376,
"theoretical_loss": 3.314563273017086,
"tokens_seen": 3073966080
},
{
"epoch": 0.27,
"learning_rate": 0.000737382378100941,
"loss": 2.7059,
"theoretical_loss": 3.314557866694841,
"tokens_seen": 3074031616
},
{
"epoch": 0.27,
"learning_rate": 0.0007369546621043627,
"loss": 2.8122,
"theoretical_loss": 3.314547054492932,
"tokens_seen": 3074162688
},
{
"epoch": 0.27,
"learning_rate": 0.0007365269461077845,
"loss": 2.7304,
"theoretical_loss": 3.314536242881082,
"tokens_seen": 3074293760
},
{
"epoch": 0.27,
"learning_rate": 0.0007360992301112062,
"loss": 2.7804,
"theoretical_loss": 3.3145254318592325,
"tokens_seen": 3074424832
},
{
"epoch": 0.27,
"learning_rate": 0.0007356715141146278,
"loss": 2.6001,
"theoretical_loss": 3.3145146214273282,
"tokens_seen": 3074555904
},
{
"epoch": 0.27,
"learning_rate": 0.0007352437981180496,
"loss": 2.6449,
"theoretical_loss": 3.3145038115853103,
"tokens_seen": 3074686976
},
{
"epoch": 0.27,
"learning_rate": 0.0007348160821214713,
"loss": 2.6983,
"theoretical_loss": 3.314493002333122,
"tokens_seen": 3074818048
},
{
"epoch": 0.27,
"learning_rate": 0.0007343883661248931,
"loss": 2.7784,
"theoretical_loss": 3.3144821936707056,
"tokens_seen": 3074949120
},
{
"epoch": 0.27,
"learning_rate": 0.0007339606501283148,
"loss": 2.6079,
"theoretical_loss": 3.3144713855980044,
"tokens_seen": 3075080192
},
{
"epoch": 0.27,
"learning_rate": 0.0007335329341317365,
"loss": 2.6211,
"theoretical_loss": 3.3144605781149608,
"tokens_seen": 3075211264
},
{
"epoch": 0.27,
"learning_rate": 0.0007331052181351583,
"loss": 2.6204,
"theoretical_loss": 3.314449771221517,
"tokens_seen": 3075342336
},
{
"epoch": 0.27,
"learning_rate": 0.00073267750213858,
"loss": 2.7189,
"theoretical_loss": 3.3144389649176165,
"tokens_seen": 3075473408
},
{
"epoch": 0.27,
"objective/train/docs_used": 1684867,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.818236827850342,
"objective/train/theoretical_loss": 3.3144281592032017,
"objective/train/tokens_used": 105655776,
"theoretical_loss": 3.3144281592032017,
"tokens_seen": 3075604480
},
{
"epoch": 0.28,
"learning_rate": 0.0007322497861420018,
"loss": 2.7713,
"theoretical_loss": 3.3144281592032017,
"tokens_seen": 3075604480
},
{
"epoch": 0.28,
"learning_rate": 0.0007318220701454235,
"loss": 2.5885,
"theoretical_loss": 3.3144173540782154,
"tokens_seen": 3075735552
},
{
"epoch": 0.28,
"learning_rate": 0.0007313943541488451,
"loss": 2.6047,
"theoretical_loss": 3.3144065495426,
"tokens_seen": 3075866624
},
{
"epoch": 0.28,
"learning_rate": 0.0007309666381522669,
"loss": 2.6418,
"theoretical_loss": 3.3143957455962982,
"tokens_seen": 3075997696
},
{
"epoch": 0.28,
"learning_rate": 0.0007305389221556886,
"loss": 2.7568,
"theoretical_loss": 3.3143849422392533,
"tokens_seen": 3076128768
},
{
"epoch": 0.28,
"learning_rate": 0.0007301112061591104,
"loss": 2.6375,
"theoretical_loss": 3.3143741394714077,
"tokens_seen": 3076259840
},
{
"epoch": 0.28,
"learning_rate": 0.0007296834901625322,
"loss": 2.761,
"theoretical_loss": 3.314363337292704,
"tokens_seen": 3076390912
},
{
"epoch": 0.28,
"learning_rate": 0.0007292557741659538,
"loss": 2.731,
"theoretical_loss": 3.314352535703086,
"tokens_seen": 3076521984
},
{
"epoch": 0.28,
"learning_rate": 0.0007288280581693756,
"loss": 2.7631,
"theoretical_loss": 3.314341734702495,
"tokens_seen": 3076653056
},
{
"epoch": 0.28,
"learning_rate": 0.0007284003421727973,
"loss": 2.6186,
"theoretical_loss": 3.3143309342908744,
"tokens_seen": 3076784128
},
{
"epoch": 0.28,
"learning_rate": 0.000727972626176219,
"loss": 2.6961,
"theoretical_loss": 3.314320134468167,
"tokens_seen": 3076915200
},
{
"epoch": 0.28,
"learning_rate": 0.0007275449101796407,
"loss": 2.7757,
"theoretical_loss": 3.3143093352343165,
"tokens_seen": 3077046272
},
{
"epoch": 0.28,
"learning_rate": 0.0007271171941830624,
"loss": 2.6504,
"theoretical_loss": 3.314298536589264,
"tokens_seen": 3077177344
},
{
"epoch": 0.28,
"objective/train/docs_used": 1686119,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.809108018875122,
"objective/train/theoretical_loss": 3.3142931374875197,
"objective/train/tokens_used": 107294176,
"theoretical_loss": 3.3142931374875197,
"tokens_seen": 3077242880
},
{
"epoch": 0.28,
"learning_rate": 0.0007266894781864842,
"loss": 2.6963,
"theoretical_loss": 3.3142877385329537,
"tokens_seen": 3077308416
},
{
"epoch": 0.28,
"learning_rate": 0.0007262617621899059,
"loss": 2.6956,
"theoretical_loss": 3.314276941065328,
"tokens_seen": 3077439488
},
{
"epoch": 0.28,
"learning_rate": 0.0007258340461933276,
"loss": 2.7936,
"theoretical_loss": 3.3142661441863295,
"tokens_seen": 3077570560
},
{
"epoch": 0.28,
"learning_rate": 0.0007254063301967494,
"loss": 2.6945,
"theoretical_loss": 3.3142553478959007,
"tokens_seen": 3077701632
},
{
"epoch": 0.28,
"learning_rate": 0.0007249786142001711,
"loss": 2.634,
"theoretical_loss": 3.3142445521939856,
"tokens_seen": 3077832704
},
{
"epoch": 0.28,
"learning_rate": 0.0007245508982035929,
"loss": 2.7157,
"theoretical_loss": 3.314233757080526,
"tokens_seen": 3077963776
},
{
"epoch": 0.28,
"learning_rate": 0.0007241231822070146,
"loss": 2.6782,
"theoretical_loss": 3.3142229625554656,
"tokens_seen": 3078094848
},
{
"epoch": 0.28,
"learning_rate": 0.0007236954662104363,
"loss": 2.6866,
"theoretical_loss": 3.3142121686187465,
"tokens_seen": 3078225920
},
{
"epoch": 0.28,
"learning_rate": 0.000723267750213858,
"loss": 2.7431,
"theoretical_loss": 3.3142013752703123,
"tokens_seen": 3078356992
},
{
"epoch": 0.28,
"learning_rate": 0.0007228400342172797,
"loss": 2.6258,
"theoretical_loss": 3.314190582510105,
"tokens_seen": 3078488064
},
{
"epoch": 0.28,
"learning_rate": 0.0007224123182207014,
"loss": 2.6916,
"theoretical_loss": 3.3141797903380685,
"tokens_seen": 3078619136
},
{
"epoch": 0.29,
"learning_rate": 0.0007219846022241232,
"loss": 2.6934,
"theoretical_loss": 3.314168998754145,
"tokens_seen": 3078750208
},
{
"epoch": 0.29,
"objective/train/docs_used": 1686853,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.6352412700653076,
"objective/train/theoretical_loss": 3.314158207758278,
"objective/train/tokens_used": 108932576,
"theoretical_loss": 3.314158207758278,
"tokens_seen": 3078881280
},
{
"epoch": 0.29,
"learning_rate": 0.0007215568862275448,
"loss": 2.8914,
"theoretical_loss": 3.314158207758278,
"tokens_seen": 3078881280
},
{
"epoch": 0.29,
"learning_rate": 0.0007211291702309667,
"loss": 2.6968,
"theoretical_loss": 3.3141474173504095,
"tokens_seen": 3079012352
},
{
"epoch": 0.29,
"learning_rate": 0.0007207014542343884,
"loss": 2.7933,
"theoretical_loss": 3.3141366275304835,
"tokens_seen": 3079143424
},
{
"epoch": 0.29,
"learning_rate": 0.0007202737382378101,
"loss": 2.7595,
"theoretical_loss": 3.3141258382984424,
"tokens_seen": 3079274496
},
{
"epoch": 0.29,
"learning_rate": 0.0007198460222412319,
"loss": 2.8446,
"theoretical_loss": 3.3141150496542293,
"tokens_seen": 3079405568
},
{
"epoch": 0.29,
"learning_rate": 0.0007194183062446535,
"loss": 2.814,
"theoretical_loss": 3.3141042615977865,
"tokens_seen": 3079536640
},
{
"epoch": 0.29,
"learning_rate": 0.0007189905902480753,
"loss": 2.6338,
"theoretical_loss": 3.314093474129058,
"tokens_seen": 3079667712
},
{
"epoch": 0.29,
"learning_rate": 0.000718562874251497,
"loss": 2.7369,
"theoretical_loss": 3.314082687247986,
"tokens_seen": 3079798784
},
{
"epoch": 0.29,
"learning_rate": 0.0007181351582549187,
"loss": 2.5899,
"theoretical_loss": 3.314071900954514,
"tokens_seen": 3079929856
},
{
"epoch": 0.29,
"learning_rate": 0.0007177074422583405,
"loss": 2.7351,
"theoretical_loss": 3.3140611152485846,
"tokens_seen": 3080060928
},
{
"epoch": 0.29,
"learning_rate": 0.0007172797262617622,
"loss": 2.7682,
"theoretical_loss": 3.314050330130141,
"tokens_seen": 3080192000
},
{
"epoch": 0.29,
"learning_rate": 0.000716852010265184,
"loss": 2.6207,
"theoretical_loss": 3.314039545599126,
"tokens_seen": 3080323072
},
{
"epoch": 0.29,
"learning_rate": 0.0007164242942686057,
"loss": 2.6523,
"theoretical_loss": 3.314028761655483,
"tokens_seen": 3080454144
},
{
"epoch": 0.29,
"objective/train/docs_used": 1688263,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5861594676971436,
"objective/train/theoretical_loss": 3.314023369903908,
"objective/train/tokens_used": 110570976,
"theoretical_loss": 3.314023369903908,
"tokens_seen": 3080519680
},
{
"epoch": 0.29,
"learning_rate": 0.0007159965782720274,
"loss": 2.6073,
"theoretical_loss": 3.3140179782991552,
"tokens_seen": 3080585216
},
{
"epoch": 0.29,
"learning_rate": 0.0007155688622754492,
"loss": 2.7999,
"theoretical_loss": 3.3140071955300847,
"tokens_seen": 3080716288
},
{
"epoch": 0.29,
"learning_rate": 0.0007151411462788708,
"loss": 2.6396,
"theoretical_loss": 3.3139964133482147,
"tokens_seen": 3080847360
},
{
"epoch": 0.29,
"learning_rate": 0.0007147134302822925,
"loss": 2.6213,
"theoretical_loss": 3.3139856317534893,
"tokens_seen": 3080978432
},
{
"epoch": 0.29,
"learning_rate": 0.0007142857142857143,
"loss": 2.7394,
"theoretical_loss": 3.31397485074585,
"tokens_seen": 3081109504
},
{
"epoch": 0.29,
"learning_rate": 0.000713857998289136,
"loss": 2.6453,
"theoretical_loss": 3.3139640703252415,
"tokens_seen": 3081240576
},
{
"epoch": 0.29,
"learning_rate": 0.0007134302822925579,
"loss": 2.6476,
"theoretical_loss": 3.3139532904916056,
"tokens_seen": 3081371648
},
{
"epoch": 0.29,
"learning_rate": 0.0007130025662959795,
"loss": 2.7387,
"theoretical_loss": 3.3139425112448864,
"tokens_seen": 3081502720
},
{
"epoch": 0.29,
"learning_rate": 0.0007125748502994012,
"loss": 2.6949,
"theoretical_loss": 3.3139317325850257,
"tokens_seen": 3081633792
},
{
"epoch": 0.3,
"learning_rate": 0.000712147134302823,
"loss": 2.6522,
"theoretical_loss": 3.3139209545119677,
"tokens_seen": 3081764864
},
{
"epoch": 0.3,
"learning_rate": 0.0007117194183062447,
"loss": 2.5179,
"theoretical_loss": 3.313910177025655,
"tokens_seen": 3081895936
},
{
"epoch": 0.3,
"learning_rate": 0.0007112917023096664,
"loss": 2.5659,
"theoretical_loss": 3.3138994001260307,
"tokens_seen": 3082027008
},
{
"epoch": 0.3,
"objective/train/docs_used": 1688735,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.452139377593994,
"objective/train/theoretical_loss": 3.313888623813038,
"objective/train/tokens_used": 112209376,
"theoretical_loss": 3.313888623813038,
"tokens_seen": 3082158080
},
{
"epoch": 0.3,
"learning_rate": 0.0007108639863130881,
"loss": 2.6064,
"theoretical_loss": 3.313888623813038,
"tokens_seen": 3082158080
},
{
"epoch": 0.3,
"learning_rate": 0.0007104362703165098,
"loss": 2.6823,
"theoretical_loss": 3.3138778480866202,
"tokens_seen": 3082289152
},
{
"epoch": 0.3,
"learning_rate": 0.0007100085543199316,
"loss": 2.6733,
"theoretical_loss": 3.3138670729467203,
"tokens_seen": 3082420224
},
{
"epoch": 0.3,
"learning_rate": 0.0007095808383233532,
"loss": 2.7262,
"theoretical_loss": 3.3138562983932816,
"tokens_seen": 3082551296
},
{
"epoch": 0.3,
"learning_rate": 0.0007091531223267751,
"loss": 2.6266,
"theoretical_loss": 3.313845524426247,
"tokens_seen": 3082682368
},
{
"epoch": 0.3,
"learning_rate": 0.0007087254063301968,
"loss": 2.7868,
"theoretical_loss": 3.3138347510455595,
"tokens_seen": 3082813440
},
{
"epoch": 0.3,
"learning_rate": 0.0007082976903336185,
"loss": 2.6533,
"theoretical_loss": 3.3138239782511625,
"tokens_seen": 3082944512
},
{
"epoch": 0.3,
"learning_rate": 0.0007078699743370403,
"loss": 2.8224,
"theoretical_loss": 3.3138132060429992,
"tokens_seen": 3083075584
},
{
"epoch": 0.3,
"learning_rate": 0.0007074422583404619,
"loss": 2.6842,
"theoretical_loss": 3.313802434421013,
"tokens_seen": 3083206656
},
{
"epoch": 0.3,
"learning_rate": 0.0007070145423438836,
"loss": 2.5755,
"theoretical_loss": 3.313791663385146,
"tokens_seen": 3083337728
},
{
"epoch": 0.3,
"learning_rate": 0.0007065868263473054,
"loss": 2.7376,
"theoretical_loss": 3.313780892935343,
"tokens_seen": 3083468800
},
{
"epoch": 0.3,
"learning_rate": 0.0007061591103507271,
"loss": 2.7195,
"theoretical_loss": 3.313770123071546,
"tokens_seen": 3083599872
},
{
"epoch": 0.3,
"learning_rate": 0.0007057313943541489,
"loss": 2.7146,
"theoretical_loss": 3.313759353793699,
"tokens_seen": 3083730944
},
{
"epoch": 0.3,
"objective/train/docs_used": 1689720,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.289289951324463,
"objective/train/theoretical_loss": 3.313753969374489,
"objective/train/tokens_used": 113847776,
"theoretical_loss": 3.313753969374489,
"tokens_seen": 3083796480
},
{
"epoch": 0.3,
"learning_rate": 0.0007053036783575705,
"loss": 2.6381,
"theoretical_loss": 3.3137485851017447,
"tokens_seen": 3083862016
},
{
"epoch": 0.3,
"learning_rate": 0.0007048759623609923,
"loss": 2.7474,
"theoretical_loss": 3.3137378169956264,
"tokens_seen": 3083993088
},
{
"epoch": 0.3,
"learning_rate": 0.0007044482463644141,
"loss": 2.752,
"theoretical_loss": 3.313727049475287,
"tokens_seen": 3084124160
},
{
"epoch": 0.3,
"learning_rate": 0.0007040205303678358,
"loss": 2.4917,
"theoretical_loss": 3.3137162825406707,
"tokens_seen": 3084255232
},
{
"epoch": 0.3,
"learning_rate": 0.0007035928143712576,
"loss": 2.5718,
"theoretical_loss": 3.31370551619172,
"tokens_seen": 3084386304
},
{
"epoch": 0.3,
"learning_rate": 0.0007031650983746792,
"loss": 2.8449,
"theoretical_loss": 3.313694750428378,
"tokens_seen": 3084517376
},
{
"epoch": 0.3,
"learning_rate": 0.0007027373823781009,
"loss": 2.5374,
"theoretical_loss": 3.313683985250589,
"tokens_seen": 3084648448
},
{
"epoch": 0.3,
"learning_rate": 0.0007023096663815227,
"loss": 2.7122,
"theoretical_loss": 3.313673220658295,
"tokens_seen": 3084779520
},
{
"epoch": 0.31,
"learning_rate": 0.0007018819503849444,
"loss": 2.5981,
"theoretical_loss": 3.31366245665144,
"tokens_seen": 3084910592
},
{
"epoch": 0.31,
"learning_rate": 0.000701454234388366,
"loss": 2.5364,
"theoretical_loss": 3.3136516932299673,
"tokens_seen": 3085041664
},
{
"epoch": 0.31,
"learning_rate": 0.0007010265183917879,
"loss": 2.7192,
"theoretical_loss": 3.3136409303938197,
"tokens_seen": 3085172736
},
{
"epoch": 0.31,
"learning_rate": 0.0007005988023952096,
"loss": 2.683,
"theoretical_loss": 3.313630168142941,
"tokens_seen": 3085303808
},
{
"epoch": 0.31,
"objective/train/docs_used": 1690994,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.9632725715637207,
"objective/train/theoretical_loss": 3.3136194064772746,
"objective/train/tokens_used": 115486176,
"theoretical_loss": 3.3136194064772746,
"tokens_seen": 3085434880
},
{
"epoch": 0.31,
"learning_rate": 0.0007001710863986314,
"loss": 2.8662,
"theoretical_loss": 3.3136194064772746,
"tokens_seen": 3085434880
},
{
"epoch": 0.31,
"learning_rate": 0.0006997433704020531,
"loss": 2.6794,
"theoretical_loss": 3.3136086453967635,
"tokens_seen": 3085565952
},
{
"epoch": 0.31,
"learning_rate": 0.0006993156544054747,
"loss": 2.6724,
"theoretical_loss": 3.313597884901351,
"tokens_seen": 3085697024
},
{
"epoch": 0.31,
"learning_rate": 0.0006988879384088965,
"loss": 2.5711,
"theoretical_loss": 3.3135871249909803,
"tokens_seen": 3085828096
},
{
"epoch": 0.31,
"learning_rate": 0.0006984602224123182,
"loss": 2.5722,
"theoretical_loss": 3.3135763656655954,
"tokens_seen": 3085959168
},
{
"epoch": 0.31,
"learning_rate": 0.00069803250641574,
"loss": 2.5725,
"theoretical_loss": 3.313565606925139,
"tokens_seen": 3086090240
},
{
"epoch": 0.31,
"learning_rate": 0.0006976047904191617,
"loss": 2.4772,
"theoretical_loss": 3.313554848769555,
"tokens_seen": 3086221312
},
{
"epoch": 0.31,
"learning_rate": 0.0006971770744225833,
"loss": 2.6553,
"theoretical_loss": 3.313544091198786,
"tokens_seen": 3086352384
},
{
"epoch": 0.31,
"learning_rate": 0.0006967493584260052,
"loss": 2.4844,
"theoretical_loss": 3.313533334212776,
"tokens_seen": 3086483456
},
{
"epoch": 0.31,
"learning_rate": 0.0006963216424294269,
"loss": 2.7982,
"theoretical_loss": 3.3135225778114683,
"tokens_seen": 3086614528
},
{
"epoch": 0.31,
"learning_rate": 0.0006958939264328487,
"loss": 2.6032,
"theoretical_loss": 3.313511821994806,
"tokens_seen": 3086745600
},
{
"epoch": 0.31,
"learning_rate": 0.0006954662104362704,
"loss": 2.6065,
"theoretical_loss": 3.313501066762733,
"tokens_seen": 3086876672
},
{
"epoch": 0.31,
"learning_rate": 0.000695038494439692,
"loss": 2.7262,
"theoretical_loss": 3.3134903121151926,
"tokens_seen": 3087007744
},
{
"epoch": 0.31,
"objective/train/docs_used": 1691628,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.5217933654785156,
"objective/train/theoretical_loss": 3.313484935010604,
"objective/train/tokens_used": 117124576,
"theoretical_loss": 3.313484935010604,
"tokens_seen": 3087073280
},
{
"epoch": 0.31,
"learning_rate": 0.0006946107784431138,
"loss": 2.6959,
"theoretical_loss": 3.3134795580521277,
"tokens_seen": 3087138816
},
{
"epoch": 0.31,
"learning_rate": 0.0006941830624465355,
"loss": 2.8117,
"theoretical_loss": 3.313468804573482,
"tokens_seen": 3087269888
},
{
"epoch": 0.31,
"learning_rate": 0.0006937553464499572,
"loss": 2.5203,
"theoretical_loss": 3.313458051679199,
"tokens_seen": 3087400960
},
{
"epoch": 0.31,
"learning_rate": 0.0006933276304533789,
"loss": 2.7896,
"theoretical_loss": 3.3134472993692223,
"tokens_seen": 3087532032
},
{
"epoch": 0.31,
"learning_rate": 0.0006928999144568007,
"loss": 2.6886,
"theoretical_loss": 3.3134365476434953,
"tokens_seen": 3087663104
},
{
"epoch": 0.31,
"learning_rate": 0.0006924721984602225,
"loss": 2.69,
"theoretical_loss": 3.313425796501961,
"tokens_seen": 3087794176
},
{
"epoch": 0.31,
"learning_rate": 0.0006920444824636442,
"loss": 2.7939,
"theoretical_loss": 3.3134150459445633,
"tokens_seen": 3087925248
},
{
"epoch": 0.32,
"learning_rate": 0.0006916167664670659,
"loss": 2.8524,
"theoretical_loss": 3.313404295971245,
"tokens_seen": 3088056320
},
{
"epoch": 0.32,
"learning_rate": 0.0006911890504704876,
"loss": 2.6725,
"theoretical_loss": 3.313393546581951,
"tokens_seen": 3088187392
},
{
"epoch": 0.32,
"learning_rate": 0.0006907613344739093,
"loss": 2.5903,
"theoretical_loss": 3.3133827977766237,
"tokens_seen": 3088318464
},
{
"epoch": 0.32,
"learning_rate": 0.0006903336184773311,
"loss": 2.741,
"theoretical_loss": 3.3133720495552064,
"tokens_seen": 3088449536
},
{
"epoch": 0.32,
"learning_rate": 0.0006899059024807528,
"loss": 2.6592,
"theoretical_loss": 3.3133613019176433,
"tokens_seen": 3088580608
},
{
"epoch": 0.32,
"objective/train/docs_used": 1692791,
"objective/train/instantaneous_batch_size": 16,
"objective/train/instantaneous_microbatch_size": 16384,
"objective/train/original_loss": 2.493699073791504,
"objective/train/theoretical_loss": 3.3133505548638778,
"objective/train/tokens_used": 118762976,
"theoretical_loss": 3.3133505548638778,
"tokens_seen": 3088711680
},
{
"epoch": 0.32,
"learning_rate": 0.0006894781864841744,
"loss": 2.69,
"theoretical_loss": 3.3133505548638778,
"tokens_seen": 3088711680
},
{
"epoch": 0.32,
"learning_rate": 0.0006890504704875963,
"loss": 2.7953,
"theoretical_loss": 3.3133398083938532,
"tokens_seen": 3088842752
},
{
"epoch": 0.32,
"learning_rate": 0.000688622754491018,
"loss": 2.7766,
"theoretical_loss": 3.313329062507513,
"tokens_seen": 3088973824
},
{
"epoch": 0.32,
"learning_rate": 0.0006881950384944397,
"loss": 2.6679,
"theoretical_loss": 3.313318317204801,
"tokens_seen": 3089104896
}
],
"max_steps": 2362,
"num_train_epochs": 9223372036854775807,
"total_flos": 5.0368742424576e+16,
"trial_name": null,
"trial_params": null
}