{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9830508474576272, "eval_steps": 8, "global_step": 29, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03389830508474576, "grad_norm": 0.38113964825930874, "learning_rate": 1.0000000000000002e-06, "loss": 1.1589, "step": 1 }, { "epoch": 0.03389830508474576, "eval_loss": 1.130394697189331, "eval_runtime": 17.254, "eval_samples_per_second": 5.738, "eval_steps_per_second": 0.406, "step": 1 }, { "epoch": 0.06779661016949153, "grad_norm": 0.3402704961945935, "learning_rate": 2.0000000000000003e-06, "loss": 1.2375, "step": 2 }, { "epoch": 0.1016949152542373, "grad_norm": 0.3059822893552262, "learning_rate": 3e-06, "loss": 1.1753, "step": 3 }, { "epoch": 0.13559322033898305, "grad_norm": 0.2967260032476386, "learning_rate": 4.000000000000001e-06, "loss": 1.2375, "step": 4 }, { "epoch": 0.1694915254237288, "grad_norm": 0.34744521968923137, "learning_rate": 5e-06, "loss": 1.2547, "step": 5 }, { "epoch": 0.2033898305084746, "grad_norm": 0.35347423132256417, "learning_rate": 6e-06, "loss": 1.1697, "step": 6 }, { "epoch": 0.23728813559322035, "grad_norm": 0.3384189211692584, "learning_rate": 7e-06, "loss": 1.1887, "step": 7 }, { "epoch": 0.2711864406779661, "grad_norm": 0.30721189309671465, "learning_rate": 8.000000000000001e-06, "loss": 1.1859, "step": 8 }, { "epoch": 0.2711864406779661, "eval_loss": 1.128434181213379, "eval_runtime": 15.9228, "eval_samples_per_second": 6.218, "eval_steps_per_second": 0.44, "step": 8 }, { "epoch": 0.3050847457627119, "grad_norm": 0.4453319451582458, "learning_rate": 9e-06, "loss": 1.1759, "step": 9 }, { "epoch": 0.3389830508474576, "grad_norm": 0.3351797755126229, "learning_rate": 1e-05, "loss": 1.2088, "step": 10 }, { "epoch": 0.3728813559322034, "grad_norm": 0.3941569965942398, "learning_rate": 9.931806517013612e-06, "loss": 1.207, "step": 11 }, { "epoch": 0.4067796610169492, "grad_norm": 0.3441878205899112, "learning_rate": 9.729086208503174e-06, "loss": 1.1378, "step": 12 }, { "epoch": 0.4406779661016949, "grad_norm": 0.41959508213449587, "learning_rate": 9.397368756032445e-06, "loss": 1.1803, "step": 13 }, { "epoch": 0.4745762711864407, "grad_norm": 0.3756216639700884, "learning_rate": 8.94570254698197e-06, "loss": 1.1101, "step": 14 }, { "epoch": 0.5084745762711864, "grad_norm": 0.43480517409238956, "learning_rate": 8.386407858128707e-06, "loss": 1.217, "step": 15 }, { "epoch": 0.5423728813559322, "grad_norm": 0.4491620141825034, "learning_rate": 7.734740790612137e-06, "loss": 1.0445, "step": 16 }, { "epoch": 0.5423728813559322, "eval_loss": 1.1105434894561768, "eval_runtime": 15.8633, "eval_samples_per_second": 6.241, "eval_steps_per_second": 0.441, "step": 16 }, { "epoch": 0.576271186440678, "grad_norm": 0.4666669843609912, "learning_rate": 7.008477123264849e-06, "loss": 1.2094, "step": 17 }, { "epoch": 0.6101694915254238, "grad_norm": 0.3831948927984007, "learning_rate": 6.227427435703997e-06, "loss": 1.1375, "step": 18 }, { "epoch": 0.6440677966101694, "grad_norm": 0.4141983253111484, "learning_rate": 5.412896727361663e-06, "loss": 1.1568, "step": 19 }, { "epoch": 0.6779661016949152, "grad_norm": 0.4528810988014663, "learning_rate": 4.587103272638339e-06, "loss": 1.0813, "step": 20 }, { "epoch": 0.711864406779661, "grad_norm": 0.463559725998681, "learning_rate": 3.7725725642960047e-06, "loss": 1.0759, "step": 21 }, { "epoch": 0.7457627118644068, "grad_norm": 0.5481602130859397, "learning_rate": 2.991522876735154e-06, "loss": 1.1317, "step": 22 }, { "epoch": 0.7796610169491526, "grad_norm": 0.4227477837808019, "learning_rate": 2.265259209387867e-06, "loss": 1.1669, "step": 23 }, { "epoch": 0.8135593220338984, "grad_norm": 0.42412143974912303, "learning_rate": 1.6135921418712959e-06, "loss": 1.1946, "step": 24 }, { "epoch": 0.8135593220338984, "eval_loss": 1.0844968557357788, "eval_runtime": 15.8637, "eval_samples_per_second": 6.241, "eval_steps_per_second": 0.441, "step": 24 }, { "epoch": 0.847457627118644, "grad_norm": 0.42758121975218416, "learning_rate": 1.0542974530180327e-06, "loss": 1.174, "step": 25 }, { "epoch": 0.8813559322033898, "grad_norm": 0.5245615771378881, "learning_rate": 6.026312439675553e-07, "loss": 1.1217, "step": 26 }, { "epoch": 0.9152542372881356, "grad_norm": 0.36119220018810566, "learning_rate": 2.7091379149682683e-07, "loss": 1.0711, "step": 27 }, { "epoch": 0.9491525423728814, "grad_norm": 0.42475966514744157, "learning_rate": 6.819348298638839e-08, "loss": 0.9551, "step": 28 }, { "epoch": 0.9830508474576272, "grad_norm": 0.4061658497077543, "learning_rate": 0.0, "loss": 1.1796, "step": 29 } ], "logging_steps": 1, "max_steps": 29, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 281581411368960.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }