llama-3-86-lora-pretrain_v2 / all_results.json
ytcheng's picture
End of training
9653e8f verified
raw
history blame
399 Bytes
{
"epoch": 2.9964796996010326,
"eval_loss": 2.235288619995117,
"eval_runtime": 115.517,
"eval_samples_per_second": 8.198,
"eval_steps_per_second": 4.103,
"perplexity": 9.349179823068619,
"total_flos": 1.1780062420402176e+18,
"train_loss": 2.428264606566656,
"train_runtime": 11538.2228,
"train_samples_per_second": 2.216,
"train_steps_per_second": 0.138
}