learning_rate: 5.0e-06 nb_training_epochs: 5 dataloader_nb_workers: 0 batch_size: 4 seed: 0 gradient_accumulation_steps: 1 do_metrics_in_training: false metric_to_track_lr: loss log_step_interval: 100