{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.21253175275190517, "global_step": 502, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 4.1666666666666665e-05, "loss": 3.0643, "theoretical_loss": 3.321567680436603, "tokens_seen": 2990538752 }, { "epoch": 0.0, "learning_rate": 8.333333333333333e-05, "loss": 3.0798, "theoretical_loss": 3.3215564803546, "tokens_seen": 2990669824 }, { "epoch": 0.0, "learning_rate": 0.000125, "loss": 2.8688, "theoretical_loss": 3.321545280900887, "tokens_seen": 2990800896 }, { "epoch": 0.0, "learning_rate": 0.00016666666666666666, "loss": 2.7194, "theoretical_loss": 3.3215340820754022, "tokens_seen": 2990931968 }, { "epoch": 0.0, "learning_rate": 0.00020833333333333335, "loss": 2.6193, "theoretical_loss": 3.3215228838780817, "tokens_seen": 2991063040 }, { "epoch": 0.0, "learning_rate": 0.00025, "loss": 2.8571, "theoretical_loss": 3.3215116863088636, "tokens_seen": 2991194112 }, { "epoch": 0.0, "learning_rate": 0.0002916666666666667, "loss": 2.7571, "theoretical_loss": 3.3215004893676854, "tokens_seen": 2991325184 }, { "epoch": 0.0, "learning_rate": 0.0003333333333333333, "loss": 2.8877, "theoretical_loss": 3.321489293054483, "tokens_seen": 2991456256 }, { "epoch": 0.0, "learning_rate": 0.000375, "loss": 2.714, "theoretical_loss": 3.321478097369195, "tokens_seen": 2991587328 }, { "epoch": 0.0, "learning_rate": 0.0004166666666666667, "loss": 2.6564, "theoretical_loss": 3.321466902311758, "tokens_seen": 2991718400 }, { "epoch": 0.0, "learning_rate": 0.0004583333333333333, "loss": 2.5638, "theoretical_loss": 3.3214557078821096, "tokens_seen": 2991849472 }, { "epoch": 0.01, "learning_rate": 0.0005, "loss": 2.4354, "theoretical_loss": 3.321444514080187, "tokens_seen": 2991980544 }, { "epoch": 0.01, "objective/train/docs_used": 1640856, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5749809741973877, "objective/train/theoretical_loss": 3.321438917414603, "objective/train/tokens_used": 22097376, "theoretical_loss": 3.321438917414603, "tokens_seen": 2992046080 }, { "epoch": 0.01, "learning_rate": 0.0005416666666666666, "loss": 2.5713, "theoretical_loss": 3.321433320905927, "tokens_seen": 2992111616 }, { "epoch": 0.01, "learning_rate": 0.0005833333333333334, "loss": 2.4812, "theoretical_loss": 3.3214221283592678, "tokens_seen": 2992242688 }, { "epoch": 0.01, "learning_rate": 0.000625, "loss": 2.7622, "theoretical_loss": 3.321410936440146, "tokens_seen": 2992373760 }, { "epoch": 0.01, "learning_rate": 0.0006666666666666666, "loss": 2.609, "theoretical_loss": 3.3213997451485, "tokens_seen": 2992504832 }, { "epoch": 0.01, "learning_rate": 0.0007083333333333334, "loss": 2.4537, "theoretical_loss": 3.3213885544842654, "tokens_seen": 2992635904 }, { "epoch": 0.01, "learning_rate": 0.00075, "loss": 2.4831, "theoretical_loss": 3.321377364447381, "tokens_seen": 2992766976 }, { "epoch": 0.01, "learning_rate": 0.0007916666666666666, "loss": 2.5607, "theoretical_loss": 3.3213661750377836, "tokens_seen": 2992898048 }, { "epoch": 0.01, "learning_rate": 0.0008333333333333334, "loss": 2.5759, "theoretical_loss": 3.3213549862554106, "tokens_seen": 2993029120 }, { "epoch": 0.01, "learning_rate": 0.000875, "loss": 2.3106, "theoretical_loss": 3.3213437981001994, "tokens_seen": 2993160192 }, { "epoch": 0.01, "learning_rate": 0.0009166666666666666, "loss": 2.5471, "theoretical_loss": 3.3213326105720875, "tokens_seen": 2993291264 }, { "epoch": 0.01, "learning_rate": 0.0009583333333333334, "loss": 2.602, "theoretical_loss": 3.3213214236710122, "tokens_seen": 2993422336 }, { "epoch": 0.01, "learning_rate": 0.001, "loss": 2.5877, "theoretical_loss": 3.321310237396911, "tokens_seen": 2993553408 }, { "epoch": 0.01, "objective/train/docs_used": 1641461, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.557373523712158, "objective/train/theoretical_loss": 3.3212990517497207, "objective/train/tokens_used": 23735776, "theoretical_loss": 3.3212990517497207, "tokens_seen": 2993684480 }, { "epoch": 0.01, "learning_rate": 0.0009995722840034217, "loss": 2.7042, "theoretical_loss": 3.3212990517497207, "tokens_seen": 2993684480 }, { "epoch": 0.01, "learning_rate": 0.0009991445680068436, "loss": 2.5234, "theoretical_loss": 3.3212878667293797, "tokens_seen": 2993815552 }, { "epoch": 0.01, "learning_rate": 0.0009987168520102653, "loss": 2.5502, "theoretical_loss": 3.321276682335825, "tokens_seen": 2993946624 }, { "epoch": 0.01, "learning_rate": 0.000998289136013687, "loss": 2.6849, "theoretical_loss": 3.3212654985689936, "tokens_seen": 2994077696 }, { "epoch": 0.01, "learning_rate": 0.0009978614200171086, "loss": 2.6348, "theoretical_loss": 3.3212543154288237, "tokens_seen": 2994208768 }, { "epoch": 0.01, "learning_rate": 0.0009974337040205303, "loss": 2.6793, "theoretical_loss": 3.3212431329152525, "tokens_seen": 2994339840 }, { "epoch": 0.01, "learning_rate": 0.0009970059880239522, "loss": 2.6212, "theoretical_loss": 3.321231951028217, "tokens_seen": 2994470912 }, { "epoch": 0.01, "learning_rate": 0.0009965782720273739, "loss": 2.629, "theoretical_loss": 3.3212207697676552, "tokens_seen": 2994601984 }, { "epoch": 0.01, "learning_rate": 0.0009961505560307955, "loss": 2.5865, "theoretical_loss": 3.3212095891335043, "tokens_seen": 2994733056 }, { "epoch": 0.01, "learning_rate": 0.0009957228400342174, "loss": 2.667, "theoretical_loss": 3.321198409125702, "tokens_seen": 2994864128 }, { "epoch": 0.01, "learning_rate": 0.000995295124037639, "loss": 2.55, "theoretical_loss": 3.321187229744186, "tokens_seen": 2994995200 }, { "epoch": 0.02, "learning_rate": 0.0009948674080410608, "loss": 2.6324, "theoretical_loss": 3.321176050988893, "tokens_seen": 2995126272 }, { "epoch": 0.02, "learning_rate": 0.0009944396920444824, "loss": 2.8406, "theoretical_loss": 3.3211648728597614, "tokens_seen": 2995257344 }, { "epoch": 0.02, "objective/train/docs_used": 1642666, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8021185398101807, "objective/train/theoretical_loss": 3.3211592840299864, "objective/train/tokens_used": 25374176, "theoretical_loss": 3.3211592840299864, "tokens_seen": 2995322880 }, { "epoch": 0.02, "learning_rate": 0.0009940119760479041, "loss": 2.5438, "theoretical_loss": 3.3211536953567284, "tokens_seen": 2995388416 }, { "epoch": 0.02, "learning_rate": 0.000993584260051326, "loss": 2.7618, "theoretical_loss": 3.321142518479731, "tokens_seen": 2995519488 }, { "epoch": 0.02, "learning_rate": 0.0009931565440547477, "loss": 2.6638, "theoretical_loss": 3.321131342228708, "tokens_seen": 2995650560 }, { "epoch": 0.02, "learning_rate": 0.0009927288280581694, "loss": 2.7287, "theoretical_loss": 3.321120166603596, "tokens_seen": 2995781632 }, { "epoch": 0.02, "learning_rate": 0.000992301112061591, "loss": 2.5832, "theoretical_loss": 3.3211089916043326, "tokens_seen": 2995912704 }, { "epoch": 0.02, "learning_rate": 0.0009918733960650127, "loss": 2.6355, "theoretical_loss": 3.3210978172308554, "tokens_seen": 2996043776 }, { "epoch": 0.02, "learning_rate": 0.0009914456800684346, "loss": 2.5139, "theoretical_loss": 3.3210866434831026, "tokens_seen": 2996174848 }, { "epoch": 0.02, "learning_rate": 0.0009910179640718563, "loss": 2.5879, "theoretical_loss": 3.3210754703610106, "tokens_seen": 2996305920 }, { "epoch": 0.02, "learning_rate": 0.0009905902480752782, "loss": 2.5935, "theoretical_loss": 3.321064297864518, "tokens_seen": 2996436992 }, { "epoch": 0.02, "learning_rate": 0.0009901625320786998, "loss": 2.6176, "theoretical_loss": 3.3210531259935627, "tokens_seen": 2996568064 }, { "epoch": 0.02, "learning_rate": 0.0009897348160821215, "loss": 2.6405, "theoretical_loss": 3.321041954748081, "tokens_seen": 2996699136 }, { "epoch": 0.02, "learning_rate": 0.0009893071000855432, "loss": 2.7274, "theoretical_loss": 3.321030784128012, "tokens_seen": 2996830208 }, { "epoch": 0.02, "objective/train/docs_used": 1643300, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.023165464401245, "objective/train/theoretical_loss": 3.321019614133292, "objective/train/tokens_used": 27012576, "theoretical_loss": 3.321019614133292, "tokens_seen": 2996961280 }, { "epoch": 0.02, "learning_rate": 0.0009888793840889649, "loss": 2.7003, "theoretical_loss": 3.321019614133292, "tokens_seen": 2996961280 }, { "epoch": 0.02, "learning_rate": 0.0009884516680923865, "loss": 2.7436, "theoretical_loss": 3.3210084447638595, "tokens_seen": 2997092352 }, { "epoch": 0.02, "learning_rate": 0.0009880239520958084, "loss": 2.5873, "theoretical_loss": 3.320997276019652, "tokens_seen": 2997223424 }, { "epoch": 0.02, "learning_rate": 0.00098759623609923, "loss": 2.5974, "theoretical_loss": 3.3209861079006067, "tokens_seen": 2997354496 }, { "epoch": 0.02, "learning_rate": 0.000987168520102652, "loss": 2.5806, "theoretical_loss": 3.320974940406662, "tokens_seen": 2997485568 }, { "epoch": 0.02, "learning_rate": 0.0009867408041060737, "loss": 2.6771, "theoretical_loss": 3.320963773537755, "tokens_seen": 2997616640 }, { "epoch": 0.02, "learning_rate": 0.0009863130881094953, "loss": 2.7313, "theoretical_loss": 3.320952607293824, "tokens_seen": 2997747712 }, { "epoch": 0.02, "learning_rate": 0.000985885372112917, "loss": 2.7302, "theoretical_loss": 3.320941441674806, "tokens_seen": 2997878784 }, { "epoch": 0.02, "learning_rate": 0.0009854576561163387, "loss": 2.893, "theoretical_loss": 3.320930276680639, "tokens_seen": 2998009856 }, { "epoch": 0.02, "learning_rate": 0.0009850299401197606, "loss": 2.6886, "theoretical_loss": 3.3209191123112607, "tokens_seen": 2998140928 }, { "epoch": 0.03, "learning_rate": 0.0009846022241231823, "loss": 2.7801, "theoretical_loss": 3.320907948566609, "tokens_seen": 2998272000 }, { "epoch": 0.03, "learning_rate": 0.000984174508126604, "loss": 2.6538, "theoretical_loss": 3.3208967854466214, "tokens_seen": 2998403072 }, { "epoch": 0.03, "learning_rate": 0.0009837467921300258, "loss": 2.5705, "theoretical_loss": 3.3208856229512356, "tokens_seen": 2998534144 }, { "epoch": 0.03, "objective/train/docs_used": 1644380, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.542499542236328, "objective/train/theoretical_loss": 3.320880041937749, "objective/train/tokens_used": 28650976, "theoretical_loss": 3.320880041937749, "tokens_seen": 2998599680 }, { "epoch": 0.03, "learning_rate": 0.0009833190761334475, "loss": 2.5074, "theoretical_loss": 3.3208744610803898, "tokens_seen": 2998665216 }, { "epoch": 0.03, "learning_rate": 0.0009828913601368692, "loss": 2.6151, "theoretical_loss": 3.320863299834021, "tokens_seen": 2998796288 }, { "epoch": 0.03, "learning_rate": 0.0009824636441402908, "loss": 2.6994, "theoretical_loss": 3.320852139212068, "tokens_seen": 2998927360 }, { "epoch": 0.03, "learning_rate": 0.0009820359281437125, "loss": 2.568, "theoretical_loss": 3.3208409792144677, "tokens_seen": 2999058432 }, { "epoch": 0.03, "learning_rate": 0.0009816082121471344, "loss": 2.5552, "theoretical_loss": 3.320829819841158, "tokens_seen": 2999189504 }, { "epoch": 0.03, "learning_rate": 0.000981180496150556, "loss": 2.6719, "theoretical_loss": 3.320818661092077, "tokens_seen": 2999320576 }, { "epoch": 0.03, "learning_rate": 0.0009807527801539778, "loss": 2.5567, "theoretical_loss": 3.3208075029671624, "tokens_seen": 2999451648 }, { "epoch": 0.03, "learning_rate": 0.0009803250641573994, "loss": 2.5511, "theoretical_loss": 3.320796345466352, "tokens_seen": 2999582720 }, { "epoch": 0.03, "learning_rate": 0.0009798973481608211, "loss": 2.6608, "theoretical_loss": 3.320785188589584, "tokens_seen": 2999713792 }, { "epoch": 0.03, "learning_rate": 0.000979469632164243, "loss": 2.5947, "theoretical_loss": 3.3207740323367956, "tokens_seen": 2999844864 }, { "epoch": 0.03, "learning_rate": 0.0009790419161676647, "loss": 2.6511, "theoretical_loss": 3.3207628767079242, "tokens_seen": 2999975936 }, { "epoch": 0.03, "learning_rate": 0.0009786142001710863, "loss": 2.5573, "theoretical_loss": 3.3207517217029094, "tokens_seen": 3000107008 }, { "epoch": 0.03, "objective/train/docs_used": 1645056, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4763801097869873, "objective/train/theoretical_loss": 3.3207405673216877, "objective/train/tokens_used": 30289376, "theoretical_loss": 3.3207405673216877, "tokens_seen": 3000238080 }, { "epoch": 0.03, "learning_rate": 0.0009781864841745082, "loss": 2.6745, "theoretical_loss": 3.3207405673216877, "tokens_seen": 3000238080 }, { "epoch": 0.03, "learning_rate": 0.00097775876817793, "loss": 2.697, "theoretical_loss": 3.320729413564197, "tokens_seen": 3000369152 }, { "epoch": 0.03, "learning_rate": 0.0009773310521813516, "loss": 2.6853, "theoretical_loss": 3.3207182604303753, "tokens_seen": 3000500224 }, { "epoch": 0.03, "learning_rate": 0.0009769033361847733, "loss": 2.3445, "theoretical_loss": 3.320707107920161, "tokens_seen": 3000631296 }, { "epoch": 0.03, "learning_rate": 0.000976475620188195, "loss": 2.6763, "theoretical_loss": 3.3206959560334917, "tokens_seen": 3000762368 }, { "epoch": 0.03, "learning_rate": 0.0009760479041916168, "loss": 2.5198, "theoretical_loss": 3.320684804770305, "tokens_seen": 3000893440 }, { "epoch": 0.03, "learning_rate": 0.0009756201881950385, "loss": 2.7, "theoretical_loss": 3.3206736541305393, "tokens_seen": 3001024512 }, { "epoch": 0.03, "learning_rate": 0.0009751924721984602, "loss": 2.6958, "theoretical_loss": 3.3206625041141318, "tokens_seen": 3001155584 }, { "epoch": 0.04, "learning_rate": 0.000974764756201882, "loss": 2.6457, "theoretical_loss": 3.3206513547210212, "tokens_seen": 3001286656 }, { "epoch": 0.04, "learning_rate": 0.0009743370402053036, "loss": 2.7946, "theoretical_loss": 3.320640205951145, "tokens_seen": 3001417728 }, { "epoch": 0.04, "learning_rate": 0.0009739093242087254, "loss": 2.6682, "theoretical_loss": 3.3206290578044415, "tokens_seen": 3001548800 }, { "epoch": 0.04, "learning_rate": 0.0009734816082121472, "loss": 2.5484, "theoretical_loss": 3.3206179102808484, "tokens_seen": 3001679872 }, { "epoch": 0.04, "learning_rate": 0.0009730538922155689, "loss": 2.6724, "theoretical_loss": 3.3206067633803036, "tokens_seen": 3001810944 }, { "epoch": 0.04, "objective/train/docs_used": 1646327, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.547891855239868, "objective/train/theoretical_loss": 3.320601190163655, "objective/train/tokens_used": 31927776, "theoretical_loss": 3.320601190163655, "tokens_seen": 3001876480 }, { "epoch": 0.04, "learning_rate": 0.0009726261762189907, "loss": 2.5267, "theoretical_loss": 3.320595617102745, "tokens_seen": 3001942016 }, { "epoch": 0.04, "learning_rate": 0.0009721984602224123, "loss": 2.6548, "theoretical_loss": 3.320584471448111, "tokens_seen": 3002073088 }, { "epoch": 0.04, "learning_rate": 0.0009717707442258341, "loss": 2.5147, "theoretical_loss": 3.3205733264163393, "tokens_seen": 3002204160 }, { "epoch": 0.04, "learning_rate": 0.0009713430282292558, "loss": 2.4505, "theoretical_loss": 3.320562182007368, "tokens_seen": 3002335232 }, { "epoch": 0.04, "learning_rate": 0.0009709153122326775, "loss": 2.6305, "theoretical_loss": 3.320551038221135, "tokens_seen": 3002466304 }, { "epoch": 0.04, "learning_rate": 0.0009704875962360993, "loss": 2.5482, "theoretical_loss": 3.3205398950575784, "tokens_seen": 3002597376 }, { "epoch": 0.04, "learning_rate": 0.0009700598802395209, "loss": 2.7266, "theoretical_loss": 3.320528752516636, "tokens_seen": 3002728448 }, { "epoch": 0.04, "learning_rate": 0.0009696321642429427, "loss": 2.5155, "theoretical_loss": 3.3205176105982463, "tokens_seen": 3002859520 }, { "epoch": 0.04, "learning_rate": 0.0009692044482463645, "loss": 2.7628, "theoretical_loss": 3.320506469302347, "tokens_seen": 3002990592 }, { "epoch": 0.04, "learning_rate": 0.0009687767322497862, "loss": 2.6802, "theoretical_loss": 3.3204953286288763, "tokens_seen": 3003121664 }, { "epoch": 0.04, "learning_rate": 0.000968349016253208, "loss": 2.7921, "theoretical_loss": 3.3204841885777725, "tokens_seen": 3003252736 }, { "epoch": 0.04, "learning_rate": 0.0009679213002566296, "loss": 2.6088, "theoretical_loss": 3.3204730491489727, "tokens_seen": 3003383808 }, { "epoch": 0.04, "objective/train/docs_used": 1647543, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5402848720550537, "objective/train/theoretical_loss": 3.3204619103424164, "objective/train/tokens_used": 33566176, "theoretical_loss": 3.3204619103424164, "tokens_seen": 3003514880 }, { "epoch": 0.04, "learning_rate": 0.0009674935842600513, "loss": 2.6446, "theoretical_loss": 3.3204619103424164, "tokens_seen": 3003514880 }, { "epoch": 0.04, "learning_rate": 0.0009670658682634731, "loss": 2.5202, "theoretical_loss": 3.3204507721580403, "tokens_seen": 3003645952 }, { "epoch": 0.04, "learning_rate": 0.0009666381522668948, "loss": 2.4832, "theoretical_loss": 3.3204396345957834, "tokens_seen": 3003777024 }, { "epoch": 0.04, "learning_rate": 0.0009662104362703165, "loss": 2.6544, "theoretical_loss": 3.320428497655584, "tokens_seen": 3003908096 }, { "epoch": 0.04, "learning_rate": 0.0009657827202737382, "loss": 2.7507, "theoretical_loss": 3.320417361337379, "tokens_seen": 3004039168 }, { "epoch": 0.04, "learning_rate": 0.00096535500427716, "loss": 2.5786, "theoretical_loss": 3.3204062256411078, "tokens_seen": 3004170240 }, { "epoch": 0.04, "learning_rate": 0.0009649272882805818, "loss": 2.6649, "theoretical_loss": 3.320395090566708, "tokens_seen": 3004301312 }, { "epoch": 0.05, "learning_rate": 0.0009644995722840035, "loss": 2.6338, "theoretical_loss": 3.3203839561141173, "tokens_seen": 3004432384 }, { "epoch": 0.05, "learning_rate": 0.0009640718562874252, "loss": 2.664, "theoretical_loss": 3.320372822283275, "tokens_seen": 3004563456 }, { "epoch": 0.05, "learning_rate": 0.0009636441402908469, "loss": 2.6909, "theoretical_loss": 3.3203616890741183, "tokens_seen": 3004694528 }, { "epoch": 0.05, "learning_rate": 0.0009632164242942686, "loss": 2.4478, "theoretical_loss": 3.3203505564865856, "tokens_seen": 3004825600 }, { "epoch": 0.05, "learning_rate": 0.0009627887082976904, "loss": 2.6265, "theoretical_loss": 3.3203394245206153, "tokens_seen": 3004956672 }, { "epoch": 0.05, "learning_rate": 0.000962360992301112, "loss": 2.5755, "theoretical_loss": 3.320328293176145, "tokens_seen": 3005087744 }, { "epoch": 0.05, "objective/train/docs_used": 1648109, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5592408180236816, "objective/train/theoretical_loss": 3.3203227277369534, "objective/train/tokens_used": 35204576, "theoretical_loss": 3.3203227277369534, "tokens_seen": 3005153280 }, { "epoch": 0.05, "learning_rate": 0.0009619332763045337, "loss": 2.567, "theoretical_loss": 3.320317162453114, "tokens_seen": 3005218816 }, { "epoch": 0.05, "learning_rate": 0.0009615055603079555, "loss": 2.6035, "theoretical_loss": 3.3203060323514593, "tokens_seen": 3005349888 }, { "epoch": 0.05, "learning_rate": 0.0009610778443113773, "loss": 2.6209, "theoretical_loss": 3.3202949028711197, "tokens_seen": 3005480960 }, { "epoch": 0.05, "learning_rate": 0.0009606501283147991, "loss": 2.522, "theoretical_loss": 3.3202837740120335, "tokens_seen": 3005612032 }, { "epoch": 0.05, "learning_rate": 0.0009602224123182207, "loss": 2.5764, "theoretical_loss": 3.3202726457741387, "tokens_seen": 3005743104 }, { "epoch": 0.05, "learning_rate": 0.0009597946963216424, "loss": 2.6805, "theoretical_loss": 3.320261518157374, "tokens_seen": 3005874176 }, { "epoch": 0.05, "learning_rate": 0.0009593669803250642, "loss": 2.4835, "theoretical_loss": 3.3202503911616765, "tokens_seen": 3006005248 }, { "epoch": 0.05, "learning_rate": 0.0009589392643284859, "loss": 2.5907, "theoretical_loss": 3.320239264786986, "tokens_seen": 3006136320 }, { "epoch": 0.05, "learning_rate": 0.0009585115483319077, "loss": 2.617, "theoretical_loss": 3.3202281390332393, "tokens_seen": 3006267392 }, { "epoch": 0.05, "learning_rate": 0.0009580838323353293, "loss": 2.5027, "theoretical_loss": 3.320217013900376, "tokens_seen": 3006398464 }, { "epoch": 0.05, "learning_rate": 0.000957656116338751, "loss": 2.6857, "theoretical_loss": 3.3202058893883333, "tokens_seen": 3006529536 }, { "epoch": 0.05, "learning_rate": 0.0009572284003421729, "loss": 2.6411, "theoretical_loss": 3.3201947654970505, "tokens_seen": 3006660608 }, { "epoch": 0.05, "objective/train/docs_used": 1649212, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.7101857662200928, "objective/train/theoretical_loss": 3.320183642226465, "objective/train/tokens_used": 36842976, "theoretical_loss": 3.320183642226465, "tokens_seen": 3006791680 }, { "epoch": 0.05, "learning_rate": 0.0009568006843455946, "loss": 2.5369, "theoretical_loss": 3.320183642226465, "tokens_seen": 3006791680 }, { "epoch": 0.05, "learning_rate": 0.0009563729683490164, "loss": 2.5836, "theoretical_loss": 3.3201725195765155, "tokens_seen": 3006922752 }, { "epoch": 0.05, "learning_rate": 0.000955945252352438, "loss": 2.5358, "theoretical_loss": 3.3201613975471402, "tokens_seen": 3007053824 }, { "epoch": 0.05, "learning_rate": 0.0009555175363558597, "loss": 2.6035, "theoretical_loss": 3.3201502761382775, "tokens_seen": 3007184896 }, { "epoch": 0.05, "learning_rate": 0.0009550898203592815, "loss": 2.4594, "theoretical_loss": 3.320139155349866, "tokens_seen": 3007315968 }, { "epoch": 0.06, "learning_rate": 0.0009546621043627032, "loss": 2.6376, "theoretical_loss": 3.3201280351818436, "tokens_seen": 3007447040 }, { "epoch": 0.06, "learning_rate": 0.0009542343883661248, "loss": 2.5904, "theoretical_loss": 3.320116915634149, "tokens_seen": 3007578112 }, { "epoch": 0.06, "learning_rate": 0.0009538066723695466, "loss": 2.7616, "theoretical_loss": 3.3201057967067205, "tokens_seen": 3007709184 }, { "epoch": 0.06, "learning_rate": 0.0009533789563729683, "loss": 2.6076, "theoretical_loss": 3.3200946783994962, "tokens_seen": 3007840256 }, { "epoch": 0.06, "learning_rate": 0.0009529512403763902, "loss": 2.6121, "theoretical_loss": 3.3200835607124146, "tokens_seen": 3007971328 }, { "epoch": 0.06, "learning_rate": 0.0009525235243798119, "loss": 2.6083, "theoretical_loss": 3.3200724436454143, "tokens_seen": 3008102400 }, { "epoch": 0.06, "learning_rate": 0.0009520958083832335, "loss": 2.6831, "theoretical_loss": 3.3200613271984336, "tokens_seen": 3008233472 }, { "epoch": 0.06, "learning_rate": 0.0009516680923866553, "loss": 2.6156, "theoretical_loss": 3.3200502113714108, "tokens_seen": 3008364544 }, { "epoch": 0.06, "objective/train/docs_used": 1649940, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.545426368713379, "objective/train/theoretical_loss": 3.3200446536903643, "objective/train/tokens_used": 38481376, "theoretical_loss": 3.3200446536903643, "tokens_seen": 3008430080 }, { "epoch": 0.06, "learning_rate": 0.000951240376390077, "loss": 2.6849, "theoretical_loss": 3.3200390961642845, "tokens_seen": 3008495616 }, { "epoch": 0.06, "learning_rate": 0.0009508126603934988, "loss": 2.654, "theoretical_loss": 3.3200279815769926, "tokens_seen": 3008626688 }, { "epoch": 0.06, "learning_rate": 0.0009503849443969204, "loss": 2.6215, "theoretical_loss": 3.3200168676094743, "tokens_seen": 3008757760 }, { "epoch": 0.06, "learning_rate": 0.0009499572284003421, "loss": 2.4387, "theoretical_loss": 3.320005754261668, "tokens_seen": 3008888832 }, { "epoch": 0.06, "learning_rate": 0.0009495295124037639, "loss": 2.5621, "theoretical_loss": 3.319994641533511, "tokens_seen": 3009019904 }, { "epoch": 0.06, "learning_rate": 0.0009491017964071857, "loss": 2.731, "theoretical_loss": 3.319983529424943, "tokens_seen": 3009150976 }, { "epoch": 0.06, "learning_rate": 0.0009486740804106075, "loss": 2.5847, "theoretical_loss": 3.3199724179359027, "tokens_seen": 3009282048 }, { "epoch": 0.06, "learning_rate": 0.0009482463644140291, "loss": 2.5836, "theoretical_loss": 3.319961307066327, "tokens_seen": 3009413120 }, { "epoch": 0.06, "learning_rate": 0.0009478186484174508, "loss": 2.6488, "theoretical_loss": 3.3199501968161558, "tokens_seen": 3009544192 }, { "epoch": 0.06, "learning_rate": 0.0009473909324208726, "loss": 2.6244, "theoretical_loss": 3.319939087185327, "tokens_seen": 3009675264 }, { "epoch": 0.06, "learning_rate": 0.0009469632164242943, "loss": 2.6873, "theoretical_loss": 3.3199279781737796, "tokens_seen": 3009806336 }, { "epoch": 0.06, "learning_rate": 0.000946535500427716, "loss": 2.5921, "theoretical_loss": 3.3199168697814514, "tokens_seen": 3009937408 }, { "epoch": 0.06, "objective/train/docs_used": 1651249, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.582075834274292, "objective/train/theoretical_loss": 3.3199057620082812, "objective/train/tokens_used": 40119776, "theoretical_loss": 3.3199057620082812, "tokens_seen": 3010068480 }, { "epoch": 0.06, "learning_rate": 0.0009461077844311377, "loss": 2.5527, "theoretical_loss": 3.3199057620082812, "tokens_seen": 3010068480 }, { "epoch": 0.06, "learning_rate": 0.0009456800684345594, "loss": 2.7818, "theoretical_loss": 3.319894654854208, "tokens_seen": 3010199552 }, { "epoch": 0.06, "learning_rate": 0.0009452523524379812, "loss": 2.628, "theoretical_loss": 3.3198835483191695, "tokens_seen": 3010330624 }, { "epoch": 0.06, "learning_rate": 0.000944824636441403, "loss": 2.4649, "theoretical_loss": 3.319872442403105, "tokens_seen": 3010461696 }, { "epoch": 0.07, "learning_rate": 0.0009443969204448247, "loss": 2.7571, "theoretical_loss": 3.3198613371059524, "tokens_seen": 3010592768 }, { "epoch": 0.07, "learning_rate": 0.0009439692044482464, "loss": 2.5878, "theoretical_loss": 3.319850232427651, "tokens_seen": 3010723840 }, { "epoch": 0.07, "learning_rate": 0.0009435414884516681, "loss": 2.6015, "theoretical_loss": 3.3198391283681383, "tokens_seen": 3010854912 }, { "epoch": 0.07, "learning_rate": 0.0009431137724550899, "loss": 2.5823, "theoretical_loss": 3.3198280249273546, "tokens_seen": 3010985984 }, { "epoch": 0.07, "learning_rate": 0.0009426860564585116, "loss": 2.7222, "theoretical_loss": 3.319816922105237, "tokens_seen": 3011117056 }, { "epoch": 0.07, "learning_rate": 0.0009422583404619332, "loss": 2.6364, "theoretical_loss": 3.319805819901724, "tokens_seen": 3011248128 }, { "epoch": 0.07, "learning_rate": 0.000941830624465355, "loss": 2.6811, "theoretical_loss": 3.3197947183167553, "tokens_seen": 3011379200 }, { "epoch": 0.07, "learning_rate": 0.0009414029084687767, "loss": 2.8016, "theoretical_loss": 3.319783617350269, "tokens_seen": 3011510272 }, { "epoch": 0.07, "learning_rate": 0.0009409751924721985, "loss": 2.5657, "theoretical_loss": 3.319772517002204, "tokens_seen": 3011641344 }, { "epoch": 0.07, "objective/train/docs_used": 1651905, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.0052103996276855, "objective/train/theoretical_loss": 3.31976696706006, "objective/train/tokens_used": 41758176, "theoretical_loss": 3.31976696706006, "tokens_seen": 3011706880 }, { "epoch": 0.07, "learning_rate": 0.0009405474764756203, "loss": 2.6357, "theoretical_loss": 3.319761417272498, "tokens_seen": 3011772416 }, { "epoch": 0.07, "learning_rate": 0.0009401197604790419, "loss": 2.5734, "theoretical_loss": 3.319750318161091, "tokens_seen": 3011903488 }, { "epoch": 0.07, "learning_rate": 0.0009396920444824637, "loss": 2.424, "theoretical_loss": 3.3197392196679205, "tokens_seen": 3012034560 }, { "epoch": 0.07, "learning_rate": 0.0009392643284858854, "loss": 2.5344, "theoretical_loss": 3.3197281217929255, "tokens_seen": 3012165632 }, { "epoch": 0.07, "learning_rate": 0.0009388366124893071, "loss": 2.5689, "theoretical_loss": 3.319717024536045, "tokens_seen": 3012296704 }, { "epoch": 0.07, "learning_rate": 0.0009384088964927289, "loss": 2.4989, "theoretical_loss": 3.3197059278972176, "tokens_seen": 3012427776 }, { "epoch": 0.07, "learning_rate": 0.0009379811804961505, "loss": 2.6272, "theoretical_loss": 3.3196948318763817, "tokens_seen": 3012558848 }, { "epoch": 0.07, "learning_rate": 0.0009375534644995723, "loss": 2.5959, "theoretical_loss": 3.319683736473476, "tokens_seen": 3012689920 }, { "epoch": 0.07, "learning_rate": 0.000937125748502994, "loss": 2.5228, "theoretical_loss": 3.3196726416884395, "tokens_seen": 3012820992 }, { "epoch": 0.07, "learning_rate": 0.0009366980325064158, "loss": 2.7357, "theoretical_loss": 3.3196615475212106, "tokens_seen": 3012952064 }, { "epoch": 0.07, "learning_rate": 0.0009362703165098376, "loss": 2.6946, "theoretical_loss": 3.3196504539717284, "tokens_seen": 3013083136 }, { "epoch": 0.07, "learning_rate": 0.0009358426005132592, "loss": 2.7344, "theoretical_loss": 3.3196393610399317, "tokens_seen": 3013214208 }, { "epoch": 0.07, "objective/train/docs_used": 1652881, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5316507816314697, "objective/train/theoretical_loss": 3.3196282687257583, "objective/train/tokens_used": 43396576, "theoretical_loss": 3.3196282687257583, "tokens_seen": 3013345280 }, { "epoch": 0.07, "learning_rate": 0.000935414884516681, "loss": 2.7259, "theoretical_loss": 3.3196282687257583, "tokens_seen": 3013345280 }, { "epoch": 0.07, "learning_rate": 0.0009349871685201027, "loss": 2.4882, "theoretical_loss": 3.3196171770291483, "tokens_seen": 3013476352 }, { "epoch": 0.07, "learning_rate": 0.0009345594525235244, "loss": 2.6169, "theoretical_loss": 3.3196060859500394, "tokens_seen": 3013607424 }, { "epoch": 0.08, "learning_rate": 0.0009341317365269461, "loss": 2.5293, "theoretical_loss": 3.319594995488371, "tokens_seen": 3013738496 }, { "epoch": 0.08, "learning_rate": 0.0009337040205303678, "loss": 2.7782, "theoretical_loss": 3.3195839056440812, "tokens_seen": 3013869568 }, { "epoch": 0.08, "learning_rate": 0.0009332763045337895, "loss": 2.6719, "theoretical_loss": 3.3195728164171094, "tokens_seen": 3014000640 }, { "epoch": 0.08, "learning_rate": 0.0009328485885372114, "loss": 2.7268, "theoretical_loss": 3.319561727807394, "tokens_seen": 3014131712 }, { "epoch": 0.08, "learning_rate": 0.0009324208725406331, "loss": 2.5997, "theoretical_loss": 3.3195506398148744, "tokens_seen": 3014262784 }, { "epoch": 0.08, "learning_rate": 0.0009319931565440548, "loss": 2.7602, "theoretical_loss": 3.319539552439489, "tokens_seen": 3014393856 }, { "epoch": 0.08, "learning_rate": 0.0009315654405474765, "loss": 2.5845, "theoretical_loss": 3.3195284656811763, "tokens_seen": 3014524928 }, { "epoch": 0.08, "learning_rate": 0.0009311377245508982, "loss": 2.7053, "theoretical_loss": 3.319517379539876, "tokens_seen": 3014656000 }, { "epoch": 0.08, "learning_rate": 0.00093071000855432, "loss": 2.6045, "theoretical_loss": 3.3195062940155258, "tokens_seen": 3014787072 }, { "epoch": 0.08, "learning_rate": 0.0009302822925577416, "loss": 2.6324, "theoretical_loss": 3.3194952091080654, "tokens_seen": 3014918144 }, { "epoch": 0.08, "objective/train/docs_used": 1653310, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4821465015411377, "objective/train/theoretical_loss": 3.3194896668856497, "objective/train/tokens_used": 45034976, "theoretical_loss": 3.3194896668856497, "tokens_seen": 3014983680 }, { "epoch": 0.08, "learning_rate": 0.0009298545765611634, "loss": 2.532, "theoretical_loss": 3.3194841248174334, "tokens_seen": 3015049216 }, { "epoch": 0.08, "learning_rate": 0.0009294268605645851, "loss": 2.544, "theoretical_loss": 3.3194730411435684, "tokens_seen": 3015180288 }, { "epoch": 0.08, "learning_rate": 0.0009289991445680068, "loss": 2.826, "theoretical_loss": 3.3194619580864098, "tokens_seen": 3015311360 }, { "epoch": 0.08, "learning_rate": 0.0009285714285714287, "loss": 2.7561, "theoretical_loss": 3.3194508756458965, "tokens_seen": 3015442432 }, { "epoch": 0.08, "learning_rate": 0.0009281437125748503, "loss": 2.5692, "theoretical_loss": 3.319439793821967, "tokens_seen": 3015573504 }, { "epoch": 0.08, "learning_rate": 0.000927715996578272, "loss": 2.6322, "theoretical_loss": 3.3194287126145596, "tokens_seen": 3015704576 }, { "epoch": 0.08, "learning_rate": 0.0009272882805816938, "loss": 2.6346, "theoretical_loss": 3.3194176320236144, "tokens_seen": 3015835648 }, { "epoch": 0.08, "learning_rate": 0.0009268605645851155, "loss": 2.7908, "theoretical_loss": 3.31940655204907, "tokens_seen": 3015966720 }, { "epoch": 0.08, "learning_rate": 0.0009264328485885373, "loss": 2.6439, "theoretical_loss": 3.319395472690865, "tokens_seen": 3016097792 }, { "epoch": 0.08, "learning_rate": 0.0009260051325919589, "loss": 2.5589, "theoretical_loss": 3.3193843939489382, "tokens_seen": 3016228864 }, { "epoch": 0.08, "learning_rate": 0.0009255774165953806, "loss": 2.6926, "theoretical_loss": 3.319373315823229, "tokens_seen": 3016359936 }, { "epoch": 0.08, "learning_rate": 0.0009251497005988024, "loss": 2.7646, "theoretical_loss": 3.3193622383136763, "tokens_seen": 3016491008 }, { "epoch": 0.08, "objective/train/docs_used": 1654644, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3390886783599854, "objective/train/theoretical_loss": 3.3193511614202187, "objective/train/tokens_used": 46673376, "theoretical_loss": 3.3193511614202187, "tokens_seen": 3016622080 }, { "epoch": 0.08, "learning_rate": 0.0009247219846022242, "loss": 2.5287, "theoretical_loss": 3.3193511614202187, "tokens_seen": 3016622080 }, { "epoch": 0.09, "learning_rate": 0.000924294268605646, "loss": 2.7607, "theoretical_loss": 3.319340085142796, "tokens_seen": 3016753152 }, { "epoch": 0.09, "learning_rate": 0.0009238665526090676, "loss": 2.6289, "theoretical_loss": 3.319329009481346, "tokens_seen": 3016884224 }, { "epoch": 0.09, "learning_rate": 0.0009234388366124893, "loss": 2.7648, "theoretical_loss": 3.3193179344358086, "tokens_seen": 3017015296 }, { "epoch": 0.09, "learning_rate": 0.0009230111206159111, "loss": 2.497, "theoretical_loss": 3.319306860006122, "tokens_seen": 3017146368 }, { "epoch": 0.09, "learning_rate": 0.0009225834046193328, "loss": 2.4963, "theoretical_loss": 3.319295786192226, "tokens_seen": 3017277440 }, { "epoch": 0.09, "learning_rate": 0.0009221556886227545, "loss": 2.5823, "theoretical_loss": 3.319284712994059, "tokens_seen": 3017408512 }, { "epoch": 0.09, "learning_rate": 0.0009217279726261762, "loss": 2.6555, "theoretical_loss": 3.3192736404115606, "tokens_seen": 3017539584 }, { "epoch": 0.09, "learning_rate": 0.0009213002566295979, "loss": 2.5264, "theoretical_loss": 3.3192625684446693, "tokens_seen": 3017670656 }, { "epoch": 0.09, "learning_rate": 0.0009208725406330197, "loss": 2.397, "theoretical_loss": 3.3192514970933242, "tokens_seen": 3017801728 }, { "epoch": 0.09, "learning_rate": 0.0009204448246364415, "loss": 2.6273, "theoretical_loss": 3.319240426357465, "tokens_seen": 3017932800 }, { "epoch": 0.09, "learning_rate": 0.0009200171086398631, "loss": 2.6414, "theoretical_loss": 3.31922935623703, "tokens_seen": 3018063872 }, { "epoch": 0.09, "learning_rate": 0.0009195893926432849, "loss": 2.5706, "theoretical_loss": 3.3192182867319584, "tokens_seen": 3018194944 }, { "epoch": 0.09, "objective/train/docs_used": 1655335, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3051180839538574, "objective/train/theoretical_loss": 3.319212752210165, "objective/train/tokens_used": 48311776, "theoretical_loss": 3.319212752210165, "tokens_seen": 3018260480 }, { "epoch": 0.09, "learning_rate": 0.0009191616766467066, "loss": 2.5393, "theoretical_loss": 3.3192072178421896, "tokens_seen": 3018326016 }, { "epoch": 0.09, "learning_rate": 0.0009187339606501284, "loss": 2.6545, "theoretical_loss": 3.319196149567662, "tokens_seen": 3018457088 }, { "epoch": 0.09, "learning_rate": 0.00091830624465355, "loss": 2.5623, "theoretical_loss": 3.3191850819083157, "tokens_seen": 3018588160 }, { "epoch": 0.09, "learning_rate": 0.0009178785286569717, "loss": 2.6804, "theoretical_loss": 3.319174014864089, "tokens_seen": 3018719232 }, { "epoch": 0.09, "learning_rate": 0.0009174508126603935, "loss": 2.8051, "theoretical_loss": 3.319162948434921, "tokens_seen": 3018850304 }, { "epoch": 0.09, "learning_rate": 0.0009170230966638152, "loss": 2.7044, "theoretical_loss": 3.319151882620752, "tokens_seen": 3018981376 }, { "epoch": 0.09, "learning_rate": 0.0009165953806672371, "loss": 2.5421, "theoretical_loss": 3.3191408174215193, "tokens_seen": 3019112448 }, { "epoch": 0.09, "learning_rate": 0.0009161676646706587, "loss": 2.7474, "theoretical_loss": 3.3191297528371635, "tokens_seen": 3019243520 }, { "epoch": 0.09, "learning_rate": 0.0009157399486740804, "loss": 2.5804, "theoretical_loss": 3.319118688867623, "tokens_seen": 3019374592 }, { "epoch": 0.09, "learning_rate": 0.0009153122326775022, "loss": 2.5145, "theoretical_loss": 3.319107625512837, "tokens_seen": 3019505664 }, { "epoch": 0.09, "learning_rate": 0.0009148845166809239, "loss": 2.746, "theoretical_loss": 3.3190965627727445, "tokens_seen": 3019636736 }, { "epoch": 0.09, "learning_rate": 0.0009144568006843457, "loss": 2.5949, "theoretical_loss": 3.3190855006472857, "tokens_seen": 3019767808 }, { "epoch": 0.09, "objective/train/docs_used": 1656670, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.1859257221221924, "objective/train/theoretical_loss": 3.3190744391363984, "objective/train/tokens_used": 49950176, "theoretical_loss": 3.3190744391363984, "tokens_seen": 3019898880 }, { "epoch": 0.1, "learning_rate": 0.0009140290846877673, "loss": 2.6334, "theoretical_loss": 3.3190744391363984, "tokens_seen": 3019898880 }, { "epoch": 0.1, "learning_rate": 0.000913601368691189, "loss": 2.6976, "theoretical_loss": 3.3190633782400223, "tokens_seen": 3020029952 }, { "epoch": 0.1, "learning_rate": 0.0009131736526946108, "loss": 2.5658, "theoretical_loss": 3.3190523179580973, "tokens_seen": 3020161024 }, { "epoch": 0.1, "learning_rate": 0.0009127459366980325, "loss": 2.7495, "theoretical_loss": 3.3190412582905617, "tokens_seen": 3020292096 }, { "epoch": 0.1, "learning_rate": 0.0009123182207014543, "loss": 2.7093, "theoretical_loss": 3.319030199237355, "tokens_seen": 3020423168 }, { "epoch": 0.1, "learning_rate": 0.000911890504704876, "loss": 2.5295, "theoretical_loss": 3.3190191407984164, "tokens_seen": 3020554240 }, { "epoch": 0.1, "learning_rate": 0.0009114627887082977, "loss": 2.6081, "theoretical_loss": 3.3190080829736854, "tokens_seen": 3020685312 }, { "epoch": 0.1, "learning_rate": 0.0009110350727117195, "loss": 2.6166, "theoretical_loss": 3.318997025763101, "tokens_seen": 3020816384 }, { "epoch": 0.1, "learning_rate": 0.0009106073567151412, "loss": 2.7691, "theoretical_loss": 3.318985969166602, "tokens_seen": 3020947456 }, { "epoch": 0.1, "learning_rate": 0.0009101796407185628, "loss": 2.6465, "theoretical_loss": 3.3189749131841286, "tokens_seen": 3021078528 }, { "epoch": 0.1, "learning_rate": 0.0009097519247219846, "loss": 2.5314, "theoretical_loss": 3.3189638578156195, "tokens_seen": 3021209600 }, { "epoch": 0.1, "learning_rate": 0.0009093242087254063, "loss": 2.8726, "theoretical_loss": 3.3189528030610136, "tokens_seen": 3021340672 }, { "epoch": 0.1, "learning_rate": 0.0009088964927288281, "loss": 2.6762, "theoretical_loss": 3.318941748920251, "tokens_seen": 3021471744 }, { "epoch": 0.1, "objective/train/docs_used": 1657192, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.9018635749816895, "objective/train/theoretical_loss": 3.318936222080042, "objective/train/tokens_used": 51588576, "theoretical_loss": 3.318936222080042, "tokens_seen": 3021537280 }, { "epoch": 0.1, "learning_rate": 0.0009084687767322499, "loss": 2.643, "theoretical_loss": 3.318930695393271, "tokens_seen": 3021602816 }, { "epoch": 0.1, "learning_rate": 0.0009080410607356715, "loss": 2.6665, "theoretical_loss": 3.3189196424800116, "tokens_seen": 3021733888 }, { "epoch": 0.1, "learning_rate": 0.0009076133447390933, "loss": 2.6265, "theoretical_loss": 3.3189085901804134, "tokens_seen": 3021864960 }, { "epoch": 0.1, "learning_rate": 0.000907185628742515, "loss": 2.5788, "theoretical_loss": 3.3188975384944155, "tokens_seen": 3021996032 }, { "epoch": 0.1, "learning_rate": 0.0009067579127459367, "loss": 2.6531, "theoretical_loss": 3.318886487421957, "tokens_seen": 3022127104 }, { "epoch": 0.1, "learning_rate": 0.0009063301967493585, "loss": 2.7177, "theoretical_loss": 3.318875436962977, "tokens_seen": 3022258176 }, { "epoch": 0.1, "learning_rate": 0.0009059024807527801, "loss": 2.4989, "theoretical_loss": 3.3188643871174155, "tokens_seen": 3022389248 }, { "epoch": 0.1, "learning_rate": 0.0009054747647562019, "loss": 2.5249, "theoretical_loss": 3.318853337885211, "tokens_seen": 3022520320 }, { "epoch": 0.1, "learning_rate": 0.0009050470487596236, "loss": 2.6031, "theoretical_loss": 3.318842289266304, "tokens_seen": 3022651392 }, { "epoch": 0.1, "learning_rate": 0.0009046193327630453, "loss": 2.5702, "theoretical_loss": 3.3188312412606327, "tokens_seen": 3022782464 }, { "epoch": 0.1, "learning_rate": 0.0009041916167664672, "loss": 2.5348, "theoretical_loss": 3.3188201938681368, "tokens_seen": 3022913536 }, { "epoch": 0.11, "learning_rate": 0.0009037639007698888, "loss": 2.5876, "theoretical_loss": 3.318809147088756, "tokens_seen": 3023044608 }, { "epoch": 0.11, "objective/train/docs_used": 1658380, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8608322143554688, "objective/train/theoretical_loss": 3.3187981009224297, "objective/train/tokens_used": 53226976, "theoretical_loss": 3.3187981009224297, "tokens_seen": 3023175680 }, { "epoch": 0.11, "learning_rate": 0.0009033361847733106, "loss": 2.6644, "theoretical_loss": 3.3187981009224297, "tokens_seen": 3023175680 }, { "epoch": 0.11, "learning_rate": 0.0009029084687767323, "loss": 2.5824, "theoretical_loss": 3.3187870553690972, "tokens_seen": 3023306752 }, { "epoch": 0.11, "learning_rate": 0.000902480752780154, "loss": 2.7377, "theoretical_loss": 3.3187760104286976, "tokens_seen": 3023437824 }, { "epoch": 0.11, "learning_rate": 0.0009020530367835757, "loss": 2.6572, "theoretical_loss": 3.3187649661011704, "tokens_seen": 3023568896 }, { "epoch": 0.11, "learning_rate": 0.0009016253207869974, "loss": 2.5447, "theoretical_loss": 3.3187539223864557, "tokens_seen": 3023699968 }, { "epoch": 0.11, "learning_rate": 0.0009011976047904192, "loss": 2.6011, "theoretical_loss": 3.318742879284492, "tokens_seen": 3023831040 }, { "epoch": 0.11, "learning_rate": 0.0009007698887938409, "loss": 2.4863, "theoretical_loss": 3.3187318367952194, "tokens_seen": 3023962112 }, { "epoch": 0.11, "learning_rate": 0.0009003421727972627, "loss": 2.4994, "theoretical_loss": 3.318720794918577, "tokens_seen": 3024093184 }, { "epoch": 0.11, "learning_rate": 0.0008999144568006844, "loss": 2.5443, "theoretical_loss": 3.3187097536545047, "tokens_seen": 3024224256 }, { "epoch": 0.11, "learning_rate": 0.0008994867408041061, "loss": 2.6808, "theoretical_loss": 3.3186987130029415, "tokens_seen": 3024355328 }, { "epoch": 0.11, "learning_rate": 0.0008990590248075278, "loss": 2.7511, "theoretical_loss": 3.3186876729638266, "tokens_seen": 3024486400 }, { "epoch": 0.11, "learning_rate": 0.0008986313088109496, "loss": 2.612, "theoretical_loss": 3.3186766335371005, "tokens_seen": 3024617472 }, { "epoch": 0.11, "learning_rate": 0.0008982035928143712, "loss": 2.686, "theoretical_loss": 3.318665594722702, "tokens_seen": 3024748544 }, { "epoch": 0.11, "objective/train/docs_used": 1658975, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.8614556789398193, "objective/train/theoretical_loss": 3.3186600755451066, "objective/train/tokens_used": 54865376, "theoretical_loss": 3.3186600755451066, "tokens_seen": 3024814080 }, { "epoch": 0.11, "learning_rate": 0.000897775876817793, "loss": 2.654, "theoretical_loss": 3.3186545565205705, "tokens_seen": 3024879616 }, { "epoch": 0.11, "learning_rate": 0.0008973481608212147, "loss": 2.6409, "theoretical_loss": 3.318643518930646, "tokens_seen": 3025010688 }, { "epoch": 0.11, "learning_rate": 0.0008969204448246364, "loss": 2.5854, "theoretical_loss": 3.3186324819528674, "tokens_seen": 3025141760 }, { "epoch": 0.11, "learning_rate": 0.0008964927288280582, "loss": 2.6063, "theoretical_loss": 3.318621445587175, "tokens_seen": 3025272832 }, { "epoch": 0.11, "learning_rate": 0.0008960650128314799, "loss": 2.537, "theoretical_loss": 3.3186104098335076, "tokens_seen": 3025403904 }, { "epoch": 0.11, "learning_rate": 0.0008956372968349017, "loss": 2.5978, "theoretical_loss": 3.318599374691805, "tokens_seen": 3025534976 }, { "epoch": 0.11, "learning_rate": 0.0008952095808383234, "loss": 2.6201, "theoretical_loss": 3.318588340162007, "tokens_seen": 3025666048 }, { "epoch": 0.11, "learning_rate": 0.0008947818648417451, "loss": 2.7451, "theoretical_loss": 3.3185773062440527, "tokens_seen": 3025797120 }, { "epoch": 0.11, "learning_rate": 0.0008943541488451669, "loss": 2.5093, "theoretical_loss": 3.3185662729378826, "tokens_seen": 3025928192 }, { "epoch": 0.12, "learning_rate": 0.0008939264328485885, "loss": 2.517, "theoretical_loss": 3.318555240243435, "tokens_seen": 3026059264 }, { "epoch": 0.12, "learning_rate": 0.0008934987168520102, "loss": 2.515, "theoretical_loss": 3.3185442081606507, "tokens_seen": 3026190336 }, { "epoch": 0.12, "learning_rate": 0.000893071000855432, "loss": 2.6143, "theoretical_loss": 3.3185331766894683, "tokens_seen": 3026321408 }, { "epoch": 0.12, "objective/train/docs_used": 1660130, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4469985961914062, "objective/train/theoretical_loss": 3.318522145829828, "objective/train/tokens_used": 56503776, "theoretical_loss": 3.318522145829828, "tokens_seen": 3026452480 }, { "epoch": 0.12, "learning_rate": 0.0008926432848588537, "loss": 2.4787, "theoretical_loss": 3.318522145829828, "tokens_seen": 3026452480 }, { "epoch": 0.12, "learning_rate": 0.0008922155688622756, "loss": 2.6368, "theoretical_loss": 3.318511115581669, "tokens_seen": 3026583552 }, { "epoch": 0.12, "learning_rate": 0.0008917878528656972, "loss": 2.6092, "theoretical_loss": 3.3185000859449314, "tokens_seen": 3026714624 }, { "epoch": 0.12, "learning_rate": 0.0008913601368691189, "loss": 2.5731, "theoretical_loss": 3.3184890569195544, "tokens_seen": 3026845696 }, { "epoch": 0.12, "learning_rate": 0.0008909324208725407, "loss": 2.5214, "theoretical_loss": 3.3184780285054782, "tokens_seen": 3026976768 }, { "epoch": 0.12, "learning_rate": 0.0008905047048759624, "loss": 2.5803, "theoretical_loss": 3.318467000702642, "tokens_seen": 3027107840 }, { "epoch": 0.12, "learning_rate": 0.0008900769888793841, "loss": 2.6223, "theoretical_loss": 3.3184559735109853, "tokens_seen": 3027238912 }, { "epoch": 0.12, "learning_rate": 0.0008896492728828058, "loss": 2.5938, "theoretical_loss": 3.3184449469304482, "tokens_seen": 3027369984 }, { "epoch": 0.12, "learning_rate": 0.0008892215568862275, "loss": 2.6878, "theoretical_loss": 3.3184339209609703, "tokens_seen": 3027501056 }, { "epoch": 0.12, "learning_rate": 0.0008887938408896493, "loss": 2.4932, "theoretical_loss": 3.318422895602491, "tokens_seen": 3027632128 }, { "epoch": 0.12, "learning_rate": 0.000888366124893071, "loss": 2.4597, "theoretical_loss": 3.31841187085495, "tokens_seen": 3027763200 }, { "epoch": 0.12, "learning_rate": 0.0008879384088964928, "loss": 2.4485, "theoretical_loss": 3.318400846718288, "tokens_seen": 3027894272 }, { "epoch": 0.12, "learning_rate": 0.0008875106928999145, "loss": 2.6036, "theoretical_loss": 3.318389823192443, "tokens_seen": 3028025344 }, { "epoch": 0.12, "objective/train/docs_used": 1661412, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3641762733459473, "objective/train/theoretical_loss": 3.3183843116585585, "objective/train/tokens_used": 58142176, "theoretical_loss": 3.3183843116585585, "tokens_seen": 3028090880 }, { "epoch": 0.12, "learning_rate": 0.0008870829769033362, "loss": 2.6277, "theoretical_loss": 3.318378800277356, "tokens_seen": 3028156416 }, { "epoch": 0.12, "learning_rate": 0.000886655260906758, "loss": 2.5149, "theoretical_loss": 3.3183677779729663, "tokens_seen": 3028287488 }, { "epoch": 0.12, "learning_rate": 0.0008862275449101797, "loss": 2.6112, "theoretical_loss": 3.3183567562792136, "tokens_seen": 3028418560 }, { "epoch": 0.12, "learning_rate": 0.0008857998289136013, "loss": 2.6564, "theoretical_loss": 3.3183457351960377, "tokens_seen": 3028549632 }, { "epoch": 0.12, "learning_rate": 0.0008853721129170231, "loss": 2.6973, "theoretical_loss": 3.3183347147233784, "tokens_seen": 3028680704 }, { "epoch": 0.12, "learning_rate": 0.0008849443969204448, "loss": 2.4547, "theoretical_loss": 3.3183236948611756, "tokens_seen": 3028811776 }, { "epoch": 0.12, "learning_rate": 0.0008845166809238666, "loss": 2.3115, "theoretical_loss": 3.3183126756093686, "tokens_seen": 3028942848 }, { "epoch": 0.12, "learning_rate": 0.0008840889649272883, "loss": 2.5859, "theoretical_loss": 3.318301656967898, "tokens_seen": 3029073920 }, { "epoch": 0.13, "learning_rate": 0.00088366124893071, "loss": 2.4681, "theoretical_loss": 3.3182906389367024, "tokens_seen": 3029204992 }, { "epoch": 0.13, "learning_rate": 0.0008832335329341318, "loss": 2.4459, "theoretical_loss": 3.3182796215157224, "tokens_seen": 3029336064 }, { "epoch": 0.13, "learning_rate": 0.0008828058169375535, "loss": 2.5877, "theoretical_loss": 3.318268604704898, "tokens_seen": 3029467136 }, { "epoch": 0.13, "learning_rate": 0.0008823781009409753, "loss": 2.5306, "theoretical_loss": 3.318257588504168, "tokens_seen": 3029598208 }, { "epoch": 0.13, "objective/train/docs_used": 1662079, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.3965446949005127, "objective/train/theoretical_loss": 3.318246572913474, "objective/train/tokens_used": 59780576, "theoretical_loss": 3.318246572913474, "tokens_seen": 3029729280 }, { "epoch": 0.13, "learning_rate": 0.0008819503849443969, "loss": 2.6399, "theoretical_loss": 3.318246572913474, "tokens_seen": 3029729280 }, { "epoch": 0.13, "learning_rate": 0.0008815226689478186, "loss": 2.6156, "theoretical_loss": 3.318235557932754, "tokens_seen": 3029860352 }, { "epoch": 0.13, "learning_rate": 0.0008810949529512404, "loss": 2.4887, "theoretical_loss": 3.318224543561948, "tokens_seen": 3029991424 }, { "epoch": 0.13, "learning_rate": 0.0008806672369546621, "loss": 2.7212, "theoretical_loss": 3.3182135298009974, "tokens_seen": 3030122496 }, { "epoch": 0.13, "learning_rate": 0.0008802395209580839, "loss": 2.6808, "theoretical_loss": 3.3182025166498406, "tokens_seen": 3030253568 }, { "epoch": 0.13, "learning_rate": 0.0008798118049615056, "loss": 2.4704, "theoretical_loss": 3.3181915041084182, "tokens_seen": 3030384640 }, { "epoch": 0.13, "learning_rate": 0.0008793840889649273, "loss": 2.4549, "theoretical_loss": 3.3181804921766695, "tokens_seen": 3030515712 }, { "epoch": 0.13, "learning_rate": 0.0008789563729683491, "loss": 2.8214, "theoretical_loss": 3.318169480854535, "tokens_seen": 3030646784 }, { "epoch": 0.13, "learning_rate": 0.0008785286569717708, "loss": 2.5934, "theoretical_loss": 3.318158470141954, "tokens_seen": 3030777856 }, { "epoch": 0.13, "learning_rate": 0.0008781009409751924, "loss": 2.6187, "theoretical_loss": 3.3181474600388667, "tokens_seen": 3030908928 }, { "epoch": 0.13, "learning_rate": 0.0008776732249786142, "loss": 2.3858, "theoretical_loss": 3.318136450545213, "tokens_seen": 3031040000 }, { "epoch": 0.13, "learning_rate": 0.0008772455089820359, "loss": 2.5896, "theoretical_loss": 3.318125441660933, "tokens_seen": 3031171072 }, { "epoch": 0.13, "learning_rate": 0.0008768177929854577, "loss": 2.388, "theoretical_loss": 3.318114433385966, "tokens_seen": 3031302144 }, { "epoch": 0.13, "objective/train/docs_used": 1662642, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.5412111282348633, "objective/train/theoretical_loss": 3.3181089294769563, "objective/train/tokens_used": 61418976, "theoretical_loss": 3.3181089294769563, "tokens_seen": 3031367680 }, { "epoch": 0.13, "learning_rate": 0.0008763900769888794, "loss": 2.6377, "theoretical_loss": 3.3181034257202526, "tokens_seen": 3031433216 }, { "epoch": 0.13, "learning_rate": 0.0008759623609923011, "loss": 2.55, "theoretical_loss": 3.318092418663732, "tokens_seen": 3031564288 }, { "epoch": 0.13, "learning_rate": 0.0008755346449957229, "loss": 2.6605, "theoretical_loss": 3.3180814122163453, "tokens_seen": 3031695360 }, { "epoch": 0.13, "learning_rate": 0.0008751069289991446, "loss": 2.6276, "theoretical_loss": 3.3180704063780313, "tokens_seen": 3031826432 }, { "epoch": 0.13, "learning_rate": 0.0008746792130025664, "loss": 2.6145, "theoretical_loss": 3.318059401148731, "tokens_seen": 3031957504 }, { "epoch": 0.13, "learning_rate": 0.0008742514970059881, "loss": 2.7501, "theoretical_loss": 3.3180483965283836, "tokens_seen": 3032088576 }, { "epoch": 0.14, "learning_rate": 0.0008738237810094097, "loss": 2.4986, "theoretical_loss": 3.318037392516929, "tokens_seen": 3032219648 }, { "epoch": 0.14, "learning_rate": 0.0008733960650128315, "loss": 2.5994, "theoretical_loss": 3.318026389114308, "tokens_seen": 3032350720 }, { "epoch": 0.14, "learning_rate": 0.0008729683490162532, "loss": 2.4541, "theoretical_loss": 3.3180153863204596, "tokens_seen": 3032481792 }, { "epoch": 0.14, "learning_rate": 0.0008725406330196749, "loss": 2.6821, "theoretical_loss": 3.3180043841353246, "tokens_seen": 3032612864 }, { "epoch": 0.14, "learning_rate": 0.0008721129170230966, "loss": 2.5988, "theoretical_loss": 3.317993382558843, "tokens_seen": 3032743936 }, { "epoch": 0.14, "learning_rate": 0.0008716852010265184, "loss": 2.6567, "theoretical_loss": 3.317982381590954, "tokens_seen": 3032875008 }, { "epoch": 0.14, "objective/train/docs_used": 1663221, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.975888252258301, "objective/train/theoretical_loss": 3.3179713812315983, "objective/train/tokens_used": 63057376, "theoretical_loss": 3.3179713812315983, "tokens_seen": 3033006080 }, { "epoch": 0.14, "learning_rate": 0.0008712574850299402, "loss": 2.6197, "theoretical_loss": 3.3179713812315983, "tokens_seen": 3033006080 }, { "epoch": 0.14, "learning_rate": 0.0008708297690333619, "loss": 2.7088, "theoretical_loss": 3.317960381480716, "tokens_seen": 3033137152 }, { "epoch": 0.14, "learning_rate": 0.0008704020530367836, "loss": 2.582, "theoretical_loss": 3.317949382338247, "tokens_seen": 3033268224 }, { "epoch": 0.14, "learning_rate": 0.0008699743370402053, "loss": 2.5966, "theoretical_loss": 3.3179383838041314, "tokens_seen": 3033399296 }, { "epoch": 0.14, "learning_rate": 0.000869546621043627, "loss": 2.8364, "theoretical_loss": 3.317927385878309, "tokens_seen": 3033530368 }, { "epoch": 0.14, "learning_rate": 0.0008691189050470488, "loss": 2.6666, "theoretical_loss": 3.31791638856072, "tokens_seen": 3033661440 }, { "epoch": 0.14, "learning_rate": 0.0008686911890504705, "loss": 2.6976, "theoretical_loss": 3.317905391851305, "tokens_seen": 3033792512 }, { "epoch": 0.14, "learning_rate": 0.0008682634730538922, "loss": 2.662, "theoretical_loss": 3.3178943957500033, "tokens_seen": 3033923584 }, { "epoch": 0.14, "learning_rate": 0.000867835757057314, "loss": 2.7223, "theoretical_loss": 3.317883400256756, "tokens_seen": 3034054656 }, { "epoch": 0.14, "learning_rate": 0.0008674080410607357, "loss": 2.754, "theoretical_loss": 3.3178724053715016, "tokens_seen": 3034185728 }, { "epoch": 0.14, "learning_rate": 0.0008669803250641575, "loss": 2.6831, "theoretical_loss": 3.3178614110941815, "tokens_seen": 3034316800 }, { "epoch": 0.14, "learning_rate": 0.0008665526090675792, "loss": 2.8167, "theoretical_loss": 3.3178504174247356, "tokens_seen": 3034447872 }, { "epoch": 0.14, "learning_rate": 0.0008661248930710008, "loss": 2.761, "theoretical_loss": 3.317839424363104, "tokens_seen": 3034578944 }, { "epoch": 0.14, "objective/train/docs_used": 1664363, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.568387746810913, "objective/train/theoretical_loss": 3.3178339280602, "objective/train/tokens_used": 64695776, "theoretical_loss": 3.3178339280602, "tokens_seen": 3034644480 }, { "epoch": 0.14, "learning_rate": 0.0008656971770744226, "loss": 2.5995, "theoretical_loss": 3.317828431909227, "tokens_seen": 3034710016 }, { "epoch": 0.14, "learning_rate": 0.0008652694610778443, "loss": 2.5506, "theoretical_loss": 3.3178174400630445, "tokens_seen": 3034841088 }, { "epoch": 0.14, "learning_rate": 0.000864841745081266, "loss": 2.8138, "theoretical_loss": 3.3178064488244967, "tokens_seen": 3034972160 }, { "epoch": 0.14, "learning_rate": 0.0008644140290846878, "loss": 2.5913, "theoretical_loss": 3.3177954581935234, "tokens_seen": 3035103232 }, { "epoch": 0.14, "learning_rate": 0.0008639863130881094, "loss": 2.5688, "theoretical_loss": 3.317784468170066, "tokens_seen": 3035234304 }, { "epoch": 0.15, "learning_rate": 0.0008635585970915313, "loss": 2.5422, "theoretical_loss": 3.317773478754063, "tokens_seen": 3035365376 }, { "epoch": 0.15, "learning_rate": 0.000863130881094953, "loss": 2.6193, "theoretical_loss": 3.317762489945456, "tokens_seen": 3035496448 }, { "epoch": 0.15, "learning_rate": 0.0008627031650983747, "loss": 2.5738, "theoretical_loss": 3.3177515017441843, "tokens_seen": 3035627520 }, { "epoch": 0.15, "learning_rate": 0.0008622754491017965, "loss": 2.715, "theoretical_loss": 3.3177405141501883, "tokens_seen": 3035758592 }, { "epoch": 0.15, "learning_rate": 0.0008618477331052181, "loss": 2.643, "theoretical_loss": 3.317729527163409, "tokens_seen": 3035889664 }, { "epoch": 0.15, "learning_rate": 0.0008614200171086399, "loss": 2.987, "theoretical_loss": 3.3177185407837855, "tokens_seen": 3036020736 }, { "epoch": 0.15, "learning_rate": 0.0008609923011120616, "loss": 2.594, "theoretical_loss": 3.3177075550112587, "tokens_seen": 3036151808 }, { "epoch": 0.15, "objective/train/docs_used": 1664891, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.584660291671753, "objective/train/theoretical_loss": 3.3176965698457686, "objective/train/tokens_used": 66334176, "theoretical_loss": 3.3176965698457686, "tokens_seen": 3036282880 }, { "epoch": 0.15, "learning_rate": 0.0008605645851154833, "loss": 2.6228, "theoretical_loss": 3.3176965698457686, "tokens_seen": 3036282880 }, { "epoch": 0.15, "learning_rate": 0.000860136869118905, "loss": 2.6502, "theoretical_loss": 3.3176855852872555, "tokens_seen": 3036413952 }, { "epoch": 0.15, "learning_rate": 0.0008597091531223268, "loss": 2.6091, "theoretical_loss": 3.3176746013356597, "tokens_seen": 3036545024 }, { "epoch": 0.15, "learning_rate": 0.0008592814371257485, "loss": 2.5742, "theoretical_loss": 3.317663617990922, "tokens_seen": 3036676096 }, { "epoch": 0.15, "learning_rate": 0.0008588537211291703, "loss": 2.573, "theoretical_loss": 3.3176526352529816, "tokens_seen": 3036807168 }, { "epoch": 0.15, "learning_rate": 0.000858426005132592, "loss": 2.6126, "theoretical_loss": 3.31764165312178, "tokens_seen": 3036938240 }, { "epoch": 0.15, "learning_rate": 0.0008579982891360137, "loss": 2.6855, "theoretical_loss": 3.3176306715972563, "tokens_seen": 3037069312 }, { "epoch": 0.15, "learning_rate": 0.0008575705731394354, "loss": 2.7764, "theoretical_loss": 3.3176196906793516, "tokens_seen": 3037200384 }, { "epoch": 0.15, "learning_rate": 0.0008571428571428571, "loss": 2.625, "theoretical_loss": 3.3176087103680056, "tokens_seen": 3037331456 }, { "epoch": 0.15, "learning_rate": 0.0008567151411462789, "loss": 2.6693, "theoretical_loss": 3.3175977306631594, "tokens_seen": 3037462528 }, { "epoch": 0.15, "learning_rate": 0.0008562874251497006, "loss": 2.6201, "theoretical_loss": 3.3175867515647526, "tokens_seen": 3037593600 }, { "epoch": 0.15, "learning_rate": 0.0008558597091531223, "loss": 2.5103, "theoretical_loss": 3.317575773072726, "tokens_seen": 3037724672 }, { "epoch": 0.15, "learning_rate": 0.0008554319931565441, "loss": 2.4253, "theoretical_loss": 3.3175647951870197, "tokens_seen": 3037855744 }, { "epoch": 0.15, "objective/train/docs_used": 1665791, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2302393913269043, "objective/train/theoretical_loss": 3.317559306471518, "objective/train/tokens_used": 67972576, "theoretical_loss": 3.317559306471518, "tokens_seen": 3037921280 }, { "epoch": 0.15, "learning_rate": 0.0008550042771599658, "loss": 2.6287, "theoretical_loss": 3.3175538179075743, "tokens_seen": 3037986816 }, { "epoch": 0.15, "learning_rate": 0.0008545765611633876, "loss": 2.5721, "theoretical_loss": 3.31754284123433, "tokens_seen": 3038117888 }, { "epoch": 0.15, "learning_rate": 0.0008541488451668093, "loss": 2.5295, "theoretical_loss": 3.3175318651672274, "tokens_seen": 3038248960 }, { "epoch": 0.15, "learning_rate": 0.000853721129170231, "loss": 2.6531, "theoretical_loss": 3.3175208897062065, "tokens_seen": 3038380032 }, { "epoch": 0.16, "learning_rate": 0.0008532934131736527, "loss": 2.7652, "theoretical_loss": 3.317509914851208, "tokens_seen": 3038511104 }, { "epoch": 0.16, "learning_rate": 0.0008528656971770744, "loss": 2.4813, "theoretical_loss": 3.3174989406021718, "tokens_seen": 3038642176 }, { "epoch": 0.16, "learning_rate": 0.0008524379811804962, "loss": 2.7802, "theoretical_loss": 3.317487966959039, "tokens_seen": 3038773248 }, { "epoch": 0.16, "learning_rate": 0.0008520102651839178, "loss": 2.6504, "theoretical_loss": 3.3174769939217494, "tokens_seen": 3038904320 }, { "epoch": 0.16, "learning_rate": 0.0008515825491873395, "loss": 2.6791, "theoretical_loss": 3.317466021490244, "tokens_seen": 3039035392 }, { "epoch": 0.16, "learning_rate": 0.0008511548331907614, "loss": 2.8812, "theoretical_loss": 3.3174550496644626, "tokens_seen": 3039166464 }, { "epoch": 0.16, "learning_rate": 0.0008507271171941831, "loss": 2.6779, "theoretical_loss": 3.317444078444346, "tokens_seen": 3039297536 }, { "epoch": 0.16, "learning_rate": 0.0008502994011976049, "loss": 2.6431, "theoretical_loss": 3.317433107829835, "tokens_seen": 3039428608 }, { "epoch": 0.16, "objective/train/docs_used": 1666368, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.657705307006836, "objective/train/theoretical_loss": 3.31742213782087, "objective/train/tokens_used": 69610976, "theoretical_loss": 3.31742213782087, "tokens_seen": 3039559680 }, { "epoch": 0.16, "learning_rate": 0.0008498716852010265, "loss": 2.72, "theoretical_loss": 3.31742213782087, "tokens_seen": 3039559680 }, { "epoch": 0.16, "learning_rate": 0.0008494439692044482, "loss": 2.589, "theoretical_loss": 3.3174111684173906, "tokens_seen": 3039690752 }, { "epoch": 0.16, "learning_rate": 0.00084901625320787, "loss": 2.7118, "theoretical_loss": 3.317400199619338, "tokens_seen": 3039821824 }, { "epoch": 0.16, "learning_rate": 0.0008485885372112917, "loss": 2.6093, "theoretical_loss": 3.3173892314266524, "tokens_seen": 3039952896 }, { "epoch": 0.16, "learning_rate": 0.0008481608212147135, "loss": 2.724, "theoretical_loss": 3.3173782638392746, "tokens_seen": 3040083968 }, { "epoch": 0.16, "learning_rate": 0.0008477331052181351, "loss": 2.6337, "theoretical_loss": 3.3173672968571446, "tokens_seen": 3040215040 }, { "epoch": 0.16, "learning_rate": 0.0008473053892215569, "loss": 2.7281, "theoretical_loss": 3.3173563304802034, "tokens_seen": 3040346112 }, { "epoch": 0.16, "learning_rate": 0.0008468776732249787, "loss": 2.6731, "theoretical_loss": 3.3173453647083915, "tokens_seen": 3040477184 }, { "epoch": 0.16, "learning_rate": 0.0008464499572284004, "loss": 2.6926, "theoretical_loss": 3.317334399541649, "tokens_seen": 3040608256 }, { "epoch": 0.16, "learning_rate": 0.0008460222412318222, "loss": 2.8351, "theoretical_loss": 3.3173234349799166, "tokens_seen": 3040739328 }, { "epoch": 0.16, "learning_rate": 0.0008455945252352438, "loss": 2.6007, "theoretical_loss": 3.317312471023135, "tokens_seen": 3040870400 }, { "epoch": 0.16, "learning_rate": 0.0008451668092386655, "loss": 2.7641, "theoretical_loss": 3.3173015076712447, "tokens_seen": 3041001472 }, { "epoch": 0.16, "learning_rate": 0.0008447390932420873, "loss": 2.7725, "theoretical_loss": 3.3172905449241865, "tokens_seen": 3041132544 }, { "epoch": 0.16, "objective/train/docs_used": 1667402, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.6917831897735596, "objective/train/theoretical_loss": 3.317285063777451, "objective/train/tokens_used": 71249376, "theoretical_loss": 3.317285063777451, "tokens_seen": 3041198080 }, { "epoch": 0.16, "learning_rate": 0.000844311377245509, "loss": 2.6852, "theoretical_loss": 3.3172795827819, "tokens_seen": 3041263616 }, { "epoch": 0.16, "learning_rate": 0.0008438836612489306, "loss": 2.6591, "theoretical_loss": 3.3172686212443274, "tokens_seen": 3041394688 }, { "epoch": 0.17, "learning_rate": 0.0008434559452523525, "loss": 2.74, "theoretical_loss": 3.317257660311408, "tokens_seen": 3041525760 }, { "epoch": 0.17, "learning_rate": 0.0008430282292557742, "loss": 2.83, "theoretical_loss": 3.3172466999830825, "tokens_seen": 3041656832 }, { "epoch": 0.17, "learning_rate": 0.000842600513259196, "loss": 2.6488, "theoretical_loss": 3.317235740259292, "tokens_seen": 3041787904 }, { "epoch": 0.17, "learning_rate": 0.0008421727972626177, "loss": 2.5988, "theoretical_loss": 3.3172247811399767, "tokens_seen": 3041918976 }, { "epoch": 0.17, "learning_rate": 0.0008417450812660393, "loss": 2.6984, "theoretical_loss": 3.317213822625077, "tokens_seen": 3042050048 }, { "epoch": 0.17, "learning_rate": 0.0008413173652694611, "loss": 2.6926, "theoretical_loss": 3.3172028647145346, "tokens_seen": 3042181120 }, { "epoch": 0.17, "learning_rate": 0.0008408896492728828, "loss": 2.6579, "theoretical_loss": 3.317191907408289, "tokens_seen": 3042312192 }, { "epoch": 0.17, "learning_rate": 0.0008404619332763046, "loss": 2.6408, "theoretical_loss": 3.3171809507062817, "tokens_seen": 3042443264 }, { "epoch": 0.17, "learning_rate": 0.0008400342172797262, "loss": 2.7066, "theoretical_loss": 3.3171699946084523, "tokens_seen": 3042574336 }, { "epoch": 0.17, "learning_rate": 0.0008396065012831479, "loss": 2.7149, "theoretical_loss": 3.3171590391147427, "tokens_seen": 3042705408 }, { "epoch": 0.17, "objective/train/docs_used": 1668521, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.528013229370117, "objective/train/theoretical_loss": 3.3171480842250927, "objective/train/tokens_used": 72887776, "theoretical_loss": 3.3171480842250927, "tokens_seen": 3042836480 }, { "epoch": 0.17, "learning_rate": 0.0008391787852865698, "loss": 2.6473, "theoretical_loss": 3.3171480842250927, "tokens_seen": 3042836480 }, { "epoch": 0.17, "learning_rate": 0.0008387510692899915, "loss": 2.6857, "theoretical_loss": 3.317137129939443, "tokens_seen": 3042967552 }, { "epoch": 0.17, "learning_rate": 0.0008383233532934132, "loss": 2.585, "theoretical_loss": 3.3171261762577346, "tokens_seen": 3043098624 }, { "epoch": 0.17, "learning_rate": 0.000837895637296835, "loss": 2.6799, "theoretical_loss": 3.3171152231799086, "tokens_seen": 3043229696 }, { "epoch": 0.17, "learning_rate": 0.0008374679213002566, "loss": 2.6625, "theoretical_loss": 3.317104270705905, "tokens_seen": 3043360768 }, { "epoch": 0.17, "learning_rate": 0.0008370402053036784, "loss": 2.6606, "theoretical_loss": 3.3170933188356644, "tokens_seen": 3043491840 }, { "epoch": 0.17, "learning_rate": 0.0008366124893071001, "loss": 2.7666, "theoretical_loss": 3.3170823675691277, "tokens_seen": 3043622912 }, { "epoch": 0.17, "learning_rate": 0.0008361847733105218, "loss": 2.7028, "theoretical_loss": 3.317071416906236, "tokens_seen": 3043753984 }, { "epoch": 0.17, "learning_rate": 0.0008357570573139435, "loss": 2.664, "theoretical_loss": 3.3170604668469297, "tokens_seen": 3043885056 }, { "epoch": 0.17, "learning_rate": 0.0008353293413173653, "loss": 2.7291, "theoretical_loss": 3.31704951739115, "tokens_seen": 3044016128 }, { "epoch": 0.17, "learning_rate": 0.0008349016253207871, "loss": 2.6537, "theoretical_loss": 3.317038568538837, "tokens_seen": 3044147200 }, { "epoch": 0.17, "learning_rate": 0.0008344739093242088, "loss": 2.5645, "theoretical_loss": 3.317027620289932, "tokens_seen": 3044278272 }, { "epoch": 0.17, "learning_rate": 0.0008340461933276305, "loss": 2.7629, "theoretical_loss": 3.317016672644375, "tokens_seen": 3044409344 }, { "epoch": 0.17, "objective/train/docs_used": 1668980, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.177739143371582, "objective/train/theoretical_loss": 3.3170111990478337, "objective/train/tokens_used": 74526176, "theoretical_loss": 3.3170111990478337, "tokens_seen": 3044474880 }, { "epoch": 0.17, "learning_rate": 0.0008336184773310522, "loss": 2.5701, "theoretical_loss": 3.3170057256021077, "tokens_seen": 3044540416 }, { "epoch": 0.18, "learning_rate": 0.0008331907613344739, "loss": 2.6897, "theoretical_loss": 3.3169947791630703, "tokens_seen": 3044671488 }, { "epoch": 0.18, "learning_rate": 0.0008327630453378957, "loss": 2.7863, "theoretical_loss": 3.3169838333272037, "tokens_seen": 3044802560 }, { "epoch": 0.18, "learning_rate": 0.0008323353293413174, "loss": 2.6692, "theoretical_loss": 3.316972888094449, "tokens_seen": 3044933632 }, { "epoch": 0.18, "learning_rate": 0.000831907613344739, "loss": 2.6618, "theoretical_loss": 3.3169619434647464, "tokens_seen": 3045064704 }, { "epoch": 0.18, "learning_rate": 0.0008314798973481608, "loss": 2.5543, "theoretical_loss": 3.3169509994380375, "tokens_seen": 3045195776 }, { "epoch": 0.18, "learning_rate": 0.0008310521813515826, "loss": 2.7893, "theoretical_loss": 3.3169400560142623, "tokens_seen": 3045326848 }, { "epoch": 0.18, "learning_rate": 0.0008306244653550043, "loss": 2.6742, "theoretical_loss": 3.3169291131933623, "tokens_seen": 3045457920 }, { "epoch": 0.18, "learning_rate": 0.0008301967493584261, "loss": 2.6488, "theoretical_loss": 3.316918170975278, "tokens_seen": 3045588992 }, { "epoch": 0.18, "learning_rate": 0.0008297690333618477, "loss": 2.5798, "theoretical_loss": 3.31690722935995, "tokens_seen": 3045720064 }, { "epoch": 0.18, "learning_rate": 0.0008293413173652695, "loss": 2.7802, "theoretical_loss": 3.3168962883473205, "tokens_seen": 3045851136 }, { "epoch": 0.18, "learning_rate": 0.0008289136013686912, "loss": 2.7032, "theoretical_loss": 3.316885347937329, "tokens_seen": 3045982208 }, { "epoch": 0.18, "objective/train/docs_used": 1670028, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.408548355102539, "objective/train/theoretical_loss": 3.316874408129916, "objective/train/tokens_used": 76164576, "theoretical_loss": 3.316874408129916, "tokens_seen": 3046113280 }, { "epoch": 0.18, "learning_rate": 0.0008284858853721129, "loss": 2.7009, "theoretical_loss": 3.316874408129916, "tokens_seen": 3046113280 }, { "epoch": 0.18, "learning_rate": 0.0008280581693755347, "loss": 2.6379, "theoretical_loss": 3.316863468925024, "tokens_seen": 3046244352 }, { "epoch": 0.18, "learning_rate": 0.0008276304533789563, "loss": 2.7495, "theoretical_loss": 3.3168525303225924, "tokens_seen": 3046375424 }, { "epoch": 0.18, "learning_rate": 0.0008272027373823782, "loss": 2.8538, "theoretical_loss": 3.316841592322563, "tokens_seen": 3046506496 }, { "epoch": 0.18, "learning_rate": 0.0008267750213857999, "loss": 2.8283, "theoretical_loss": 3.3168306549248765, "tokens_seen": 3046637568 }, { "epoch": 0.18, "learning_rate": 0.0008263473053892216, "loss": 2.6814, "theoretical_loss": 3.316819718129474, "tokens_seen": 3046768640 }, { "epoch": 0.18, "learning_rate": 0.0008259195893926434, "loss": 2.6987, "theoretical_loss": 3.3168087819362957, "tokens_seen": 3046899712 }, { "epoch": 0.18, "learning_rate": 0.000825491873396065, "loss": 2.6017, "theoretical_loss": 3.316797846345283, "tokens_seen": 3047030784 }, { "epoch": 0.18, "learning_rate": 0.0008250641573994867, "loss": 2.7868, "theoretical_loss": 3.316786911356377, "tokens_seen": 3047161856 }, { "epoch": 0.18, "learning_rate": 0.0008246364414029085, "loss": 2.5386, "theoretical_loss": 3.316775976969519, "tokens_seen": 3047292928 }, { "epoch": 0.18, "learning_rate": 0.0008242087254063302, "loss": 2.806, "theoretical_loss": 3.316765043184649, "tokens_seen": 3047424000 }, { "epoch": 0.18, "learning_rate": 0.0008237810094097519, "loss": 2.7629, "theoretical_loss": 3.316754110001708, "tokens_seen": 3047555072 }, { "epoch": 0.19, "learning_rate": 0.0008233532934131736, "loss": 2.7679, "theoretical_loss": 3.3167431774206384, "tokens_seen": 3047686144 }, { "epoch": 0.19, "objective/train/docs_used": 1670628, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.13972806930542, "objective/train/theoretical_loss": 3.316737711355786, "objective/train/tokens_used": 77802976, "theoretical_loss": 3.316737711355786, "tokens_seen": 3047751680 }, { "epoch": 0.19, "learning_rate": 0.0008229255774165954, "loss": 2.5883, "theoretical_loss": 3.3167322454413792, "tokens_seen": 3047817216 }, { "epoch": 0.19, "learning_rate": 0.0008224978614200172, "loss": 2.7089, "theoretical_loss": 3.316721314063873, "tokens_seen": 3047948288 }, { "epoch": 0.19, "learning_rate": 0.0008220701454234389, "loss": 2.7955, "theoretical_loss": 3.3167103832880604, "tokens_seen": 3048079360 }, { "epoch": 0.19, "learning_rate": 0.0008216424294268606, "loss": 2.7126, "theoretical_loss": 3.316699453113882, "tokens_seen": 3048210432 }, { "epoch": 0.19, "learning_rate": 0.0008212147134302823, "loss": 2.6486, "theoretical_loss": 3.3166885235412784, "tokens_seen": 3048341504 }, { "epoch": 0.19, "learning_rate": 0.000820786997433704, "loss": 2.6465, "theoretical_loss": 3.316677594570192, "tokens_seen": 3048472576 }, { "epoch": 0.19, "learning_rate": 0.0008203592814371258, "loss": 2.6168, "theoretical_loss": 3.316666666200563, "tokens_seen": 3048603648 }, { "epoch": 0.19, "learning_rate": 0.0008199315654405474, "loss": 2.6939, "theoretical_loss": 3.316655738432332, "tokens_seen": 3048734720 }, { "epoch": 0.19, "learning_rate": 0.0008195038494439692, "loss": 2.6707, "theoretical_loss": 3.3166448112654408, "tokens_seen": 3048865792 }, { "epoch": 0.19, "learning_rate": 0.000819076133447391, "loss": 2.6281, "theoretical_loss": 3.3166338846998302, "tokens_seen": 3048996864 }, { "epoch": 0.19, "learning_rate": 0.0008186484174508127, "loss": 2.7623, "theoretical_loss": 3.316622958735442, "tokens_seen": 3049127936 }, { "epoch": 0.19, "learning_rate": 0.0008182207014542345, "loss": 2.7285, "theoretical_loss": 3.3166120333722158, "tokens_seen": 3049259008 }, { "epoch": 0.19, "objective/train/docs_used": 1671752, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2737972736358643, "objective/train/theoretical_loss": 3.3166011086100937, "objective/train/tokens_used": 79441376, "theoretical_loss": 3.3166011086100937, "tokens_seen": 3049390080 }, { "epoch": 0.19, "learning_rate": 0.0008177929854576561, "loss": 2.7943, "theoretical_loss": 3.3166011086100937, "tokens_seen": 3049390080 }, { "epoch": 0.19, "learning_rate": 0.0008173652694610778, "loss": 2.6782, "theoretical_loss": 3.3165901844490167, "tokens_seen": 3049521152 }, { "epoch": 0.19, "learning_rate": 0.0008169375534644996, "loss": 2.6773, "theoretical_loss": 3.3165792608889255, "tokens_seen": 3049652224 }, { "epoch": 0.19, "learning_rate": 0.0008165098374679213, "loss": 2.7108, "theoretical_loss": 3.3165683379297612, "tokens_seen": 3049783296 }, { "epoch": 0.19, "learning_rate": 0.0008160821214713431, "loss": 2.644, "theoretical_loss": 3.3165574155714657, "tokens_seen": 3049914368 }, { "epoch": 0.19, "learning_rate": 0.0008156544054747647, "loss": 2.7035, "theoretical_loss": 3.3165464938139797, "tokens_seen": 3050045440 }, { "epoch": 0.19, "learning_rate": 0.0008152266894781864, "loss": 2.6701, "theoretical_loss": 3.3165355726572434, "tokens_seen": 3050176512 }, { "epoch": 0.19, "learning_rate": 0.0008147989734816083, "loss": 2.6013, "theoretical_loss": 3.3165246521011995, "tokens_seen": 3050307584 }, { "epoch": 0.19, "learning_rate": 0.00081437125748503, "loss": 2.5665, "theoretical_loss": 3.3165137321457885, "tokens_seen": 3050438656 }, { "epoch": 0.19, "learning_rate": 0.0008139435414884518, "loss": 2.5955, "theoretical_loss": 3.3165028127909513, "tokens_seen": 3050569728 }, { "epoch": 0.19, "learning_rate": 0.0008135158254918734, "loss": 2.7426, "theoretical_loss": 3.3164918940366293, "tokens_seen": 3050700800 }, { "epoch": 0.2, "learning_rate": 0.0008130881094952951, "loss": 2.6797, "theoretical_loss": 3.3164809758827634, "tokens_seen": 3050831872 }, { "epoch": 0.2, "learning_rate": 0.0008126603934987169, "loss": 2.7291, "theoretical_loss": 3.3164700583292954, "tokens_seen": 3050962944 }, { "epoch": 0.2, "objective/train/docs_used": 1672176, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.4254088401794434, "objective/train/theoretical_loss": 3.316464599777692, "objective/train/tokens_used": 81079776, "theoretical_loss": 3.316464599777692, "tokens_seen": 3051028480 }, { "epoch": 0.2, "learning_rate": 0.0008122326775021386, "loss": 2.6885, "theoretical_loss": 3.316459141376166, "tokens_seen": 3051094016 }, { "epoch": 0.2, "learning_rate": 0.0008118049615055603, "loss": 2.6706, "theoretical_loss": 3.3164482250233163, "tokens_seen": 3051225088 }, { "epoch": 0.2, "learning_rate": 0.000811377245508982, "loss": 2.6934, "theoretical_loss": 3.316437309270688, "tokens_seen": 3051356160 }, { "epoch": 0.2, "learning_rate": 0.0008109495295124037, "loss": 2.6764, "theoretical_loss": 3.316426394118222, "tokens_seen": 3051487232 }, { "epoch": 0.2, "learning_rate": 0.0008105218135158256, "loss": 2.6293, "theoretical_loss": 3.316415479565859, "tokens_seen": 3051618304 }, { "epoch": 0.2, "learning_rate": 0.0008100940975192473, "loss": 2.7261, "theoretical_loss": 3.3164045656135417, "tokens_seen": 3051749376 }, { "epoch": 0.2, "learning_rate": 0.0008096663815226689, "loss": 2.6812, "theoretical_loss": 3.3163936522612096, "tokens_seen": 3051880448 }, { "epoch": 0.2, "learning_rate": 0.0008092386655260907, "loss": 2.5557, "theoretical_loss": 3.3163827395088052, "tokens_seen": 3052011520 }, { "epoch": 0.2, "learning_rate": 0.0008088109495295124, "loss": 2.6232, "theoretical_loss": 3.3163718273562695, "tokens_seen": 3052142592 }, { "epoch": 0.2, "learning_rate": 0.0008083832335329342, "loss": 2.5755, "theoretical_loss": 3.3163609158035436, "tokens_seen": 3052273664 }, { "epoch": 0.2, "learning_rate": 0.0008079555175363559, "loss": 2.5787, "theoretical_loss": 3.3163500048505687, "tokens_seen": 3052404736 }, { "epoch": 0.2, "learning_rate": 0.0008075278015397775, "loss": 2.6212, "theoretical_loss": 3.3163390944972857, "tokens_seen": 3052535808 }, { "epoch": 0.2, "objective/train/docs_used": 1673386, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.2504236698150635, "objective/train/theoretical_loss": 3.316328184743637, "objective/train/tokens_used": 82718176, "theoretical_loss": 3.316328184743637, "tokens_seen": 3052666880 }, { "epoch": 0.2, "learning_rate": 0.0008071000855431993, "loss": 2.5091, "theoretical_loss": 3.316328184743637, "tokens_seen": 3052666880 }, { "epoch": 0.2, "learning_rate": 0.0008066723695466211, "loss": 2.7788, "theoretical_loss": 3.3163172755895634, "tokens_seen": 3052797952 }, { "epoch": 0.2, "learning_rate": 0.0008062446535500429, "loss": 2.6996, "theoretical_loss": 3.3163063670350055, "tokens_seen": 3052929024 }, { "epoch": 0.2, "learning_rate": 0.0008058169375534645, "loss": 2.639, "theoretical_loss": 3.3162954590799054, "tokens_seen": 3053060096 }, { "epoch": 0.2, "learning_rate": 0.0008053892215568862, "loss": 2.7192, "theoretical_loss": 3.316284551724204, "tokens_seen": 3053191168 }, { "epoch": 0.2, "learning_rate": 0.000804961505560308, "loss": 2.5435, "theoretical_loss": 3.3162736449678434, "tokens_seen": 3053322240 }, { "epoch": 0.2, "learning_rate": 0.0008045337895637297, "loss": 2.7468, "theoretical_loss": 3.3162627388107637, "tokens_seen": 3053453312 }, { "epoch": 0.2, "learning_rate": 0.0008041060735671514, "loss": 2.6621, "theoretical_loss": 3.316251833252908, "tokens_seen": 3053584384 }, { "epoch": 0.2, "learning_rate": 0.0008036783575705731, "loss": 2.6691, "theoretical_loss": 3.3162409282942154, "tokens_seen": 3053715456 }, { "epoch": 0.2, "learning_rate": 0.0008032506415739948, "loss": 2.6699, "theoretical_loss": 3.316230023934629, "tokens_seen": 3053846528 }, { "epoch": 0.21, "learning_rate": 0.0008028229255774167, "loss": 2.711, "theoretical_loss": 3.3162191201740896, "tokens_seen": 3053977600 }, { "epoch": 0.21, "learning_rate": 0.0008023952095808384, "loss": 2.6899, "theoretical_loss": 3.3162082170125387, "tokens_seen": 3054108672 }, { "epoch": 0.21, "learning_rate": 0.00080196749358426, "loss": 2.4566, "theoretical_loss": 3.3161973144499175, "tokens_seen": 3054239744 }, { "epoch": 0.21, "objective/train/docs_used": 1673816, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 2.680955171585083, "objective/train/theoretical_loss": 3.316191863393187, "objective/train/tokens_used": 84356576, "theoretical_loss": 3.316191863393187, "tokens_seen": 3054305280 }, { "epoch": 0.21, "learning_rate": 0.0008015397775876818, "loss": 2.5931, "theoretical_loss": 3.3161864124861675, "tokens_seen": 3054370816 }, { "epoch": 0.21, "learning_rate": 0.0008011120615911035, "loss": 2.6795, "theoretical_loss": 3.31617551112123, "tokens_seen": 3054501888 }, { "epoch": 0.21, "learning_rate": 0.0008006843455945253, "loss": 2.7542, "theoretical_loss": 3.316164610355047, "tokens_seen": 3054632960 }, { "epoch": 0.21, "learning_rate": 0.000800256629597947, "loss": 2.6446, "theoretical_loss": 3.316153710187559, "tokens_seen": 3054764032 }, { "epoch": 0.21, "learning_rate": 0.0007998289136013686, "loss": 2.7551, "theoretical_loss": 3.316142810618708, "tokens_seen": 3054895104 }, { "epoch": 0.21, "learning_rate": 0.0007994011976047904, "loss": 2.7207, "theoretical_loss": 3.3161319116484353, "tokens_seen": 3055026176 }, { "epoch": 0.21, "learning_rate": 0.0007989734816082121, "loss": 2.8592, "theoretical_loss": 3.3161210132766823, "tokens_seen": 3055157248 }, { "epoch": 0.21, "learning_rate": 0.000798545765611634, "loss": 2.8129, "theoretical_loss": 3.316110115503391, "tokens_seen": 3055288320 }, { "epoch": 0.21, "learning_rate": 0.0007981180496150557, "loss": 2.6611, "theoretical_loss": 3.316099218328502, "tokens_seen": 3055419392 }, { "epoch": 0.21, "learning_rate": 0.0007976903336184773, "loss": 2.7847, "theoretical_loss": 3.3160883217519572, "tokens_seen": 3055550464 }, { "epoch": 0.21, "learning_rate": 0.0007972626176218991, "loss": 2.5896, "theoretical_loss": 3.316077425773698, "tokens_seen": 3055681536 }, { "epoch": 0.21, "learning_rate": 0.0007968349016253208, "loss": 2.699, "theoretical_loss": 3.316066530393666, "tokens_seen": 3055812608 }, { "epoch": 0.21, "objective/train/docs_used": 1674780, "objective/train/instantaneous_batch_size": 16, "objective/train/instantaneous_microbatch_size": 16384, "objective/train/original_loss": 3.0617194175720215, "objective/train/theoretical_loss": 3.3160556356118027, "objective/train/tokens_used": 85994976, "theoretical_loss": 3.3160556356118027, "tokens_seen": 3055943680 }, { "epoch": 0.21, "learning_rate": 0.0007964071856287425, "loss": 2.7631, "theoretical_loss": 3.3160556356118027, "tokens_seen": 3055943680 }, { "epoch": 0.21, "learning_rate": 0.0007959794696321643, "loss": 2.8357, "theoretical_loss": 3.3160447414280494, "tokens_seen": 3056074752 }, { "epoch": 0.21, "learning_rate": 0.0007955517536355859, "loss": 2.6724, "theoretical_loss": 3.316033847842348, "tokens_seen": 3056205824 } ], "max_steps": 2362, "num_train_epochs": 9223372036854775807, "total_flos": 3.3579161616384e+16, "trial_name": null, "trial_params": null }