|
cutoff_len: 1024 |
|
dataset_dir: data |
|
do_predict: true |
|
eval_dataset: truth_dev |
|
finetuning_type: full |
|
flash_attn: auto |
|
max_new_tokens: 512 |
|
max_samples: 100000 |
|
model_name_or_path: saves/LLaMA3.1-8B-Chat/full/train_2024-07-29-16-36-04_llama3.1_reeval |
|
output_dir: saves/LLaMA3.1-8B-Chat/full/eval_2024-07-29-16-36-04_llama3.1_reeval |
|
per_device_eval_batch_size: 2 |
|
predict_with_generate: true |
|
preprocessing_num_workers: 16 |
|
quantization_method: bitsandbytes |
|
stage: sft |
|
temperature: 0.95 |
|
template: llama3 |
|
top_p: 0.7 |
|
|