{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 6411, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.2, "grad_norm": 6479046246400.0, "learning_rate": 3.3255451713395644e-05, "loss": 31699221108.3091, "step": 427 }, { "epoch": 0.4, "grad_norm": 8700652158976.0, "learning_rate": 4.8162593170393484e-05, "loss": 14378762930.6604, "step": 854 }, { "epoch": 0.6, "grad_norm": 16.28143310546875, "learning_rate": 4.446177847113885e-05, "loss": 3635725784.4309, "step": 1281 }, { "epoch": 0.8, "grad_norm": 6.706409454345703, "learning_rate": 4.076096377188421e-05, "loss": 1017393142.4075, "step": 1708 }, { "epoch": 1.0, "grad_norm": 9.295268058776855, "learning_rate": 3.7060149072629574e-05, "loss": 1230655574.3326, "step": 2135 }, { "epoch": 1.2, "grad_norm": 7.9164958000183105, "learning_rate": 3.335933437337493e-05, "loss": 387735187.4848, "step": 2562 }, { "epoch": 1.4, "grad_norm": 18.9678897857666, "learning_rate": 2.96585196741203e-05, "loss": 610998305.5738, "step": 2989 }, { "epoch": 1.6, "grad_norm": 22.521636962890625, "learning_rate": 2.595770497486566e-05, "loss": 286763645.9016, "step": 3416 }, { "epoch": 1.8, "grad_norm": 6.2225751876831055, "learning_rate": 2.2256890275611026e-05, "loss": 146726173.377, "step": 3843 }, { "epoch": 2.0, "grad_norm": 4.1397271156311035, "learning_rate": 1.8556075576356388e-05, "loss": 153057112.1311, "step": 4270 }, { "epoch": 2.2, "grad_norm": 20.3540096282959, "learning_rate": 1.4855260877101752e-05, "loss": 167597749.0585, "step": 4697 }, { "epoch": 2.4, "grad_norm": 2.9089903831481934, "learning_rate": 1.1154446177847114e-05, "loss": 31491796.2342, "step": 5124 }, { "epoch": 2.6, "grad_norm": 11.952803611755371, "learning_rate": 7.453631478592478e-06, "loss": 228096436.459, "step": 5551 }, { "epoch": 2.8, "grad_norm": 8.380680084228516, "learning_rate": 3.7528167793378402e-06, "loss": 170669039.2131, "step": 5978 }, { "epoch": 3.0, "grad_norm": 11.055083274841309, "learning_rate": 5.2002080083203335e-08, "loss": 77525903.2881, "step": 6405 } ], "logging_steps": 427, "max_steps": 6411, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "total_flos": 1.089435891661996e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }