model: | |
_component_: torchtune.models.llama2.qlora_llama2_7b | |
lora_attn_modules: | |
- q_proj | |
- v_proj | |
- k_proj | |
- output_proj | |
apply_lora_to_mlp: true | |
apply_lora_to_output: false | |
lora_rank: 8 | |
lora_alpha: 16 | |
tokenizer: | |
_component_: torchtune.models.llama2.llama2_tokenizer | |
path: ./checkpoint/tokenizer.model | |
checkpointer: | |
_component_: torchtune.utils.FullModelHFCheckpointer | |
checkpoint_dir: ./checkpoint | |
checkpoint_files: | |
- pytorch_model-00001-of-00002.bin | |
- pytorch_model-00002-of-00002.bin | |
adapter_checkpoint: null | |
recipe_checkpoint: null | |
output_dir: ./checkpoint | |
model_type: LLAMA2 | |
resume_from_checkpoint: false | |
dataset: | |
_component_: torchtune.datasets.alpaca_cleaned_dataset | |
train_on_input: true | |
seed: null | |
shuffle: true | |
batch_size: 2 | |
optimizer: | |
_component_: torch.optim.AdamW | |
weight_decay: 0.01 | |
lr: 0.0003 | |
lr_scheduler: | |
_component_: torchtune.modules.get_cosine_schedule_with_warmup | |
num_warmup_steps: 100 | |
loss: | |
_component_: torch.nn.CrossEntropyLoss | |
epochs: 1 | |
max_steps_per_epoch: null | |
gradient_accumulation_steps: 16 | |
compile: false | |
output_dir: /tmp/qlora_finetune_output/ | |
metric_logger: | |
_component_: torchtune.utils.metric_logging.WandBLogger | |
log_dir: ${output_dir} | |
log_every_n_steps: 1 | |
device: cuda | |
dtype: bf16 | |
enable_activation_checkpointing: true | |
profiler: | |
_component_: torchtune.utils.profiler | |
enabled: false | |
output_dir: ${output_dir}/torchtune_perf_tracing.json | |