metadata
license: apache-2.0
library_name: peft
tags:
- trl
- sft
- generated_from_trainer
base_model: mistralai/Mistral-7B-v0.1
model-index:
- name: lc_full
results: []
lc_full
This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 1.8715
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 1
- eval_batch_size: 1
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- num_epochs: 50
Training results
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.7424 | 1.0 | 486 | 1.6914 |
1.301 | 2.0 | 972 | 1.6780 |
1.5718 | 3.0 | 1458 | 1.6743 |
1.6632 | 4.0 | 1944 | 1.6793 |
1.8588 | 5.0 | 2430 | 1.6794 |
1.5308 | 6.0 | 2916 | 1.6894 |
1.5776 | 7.0 | 3402 | 1.6985 |
1.6394 | 8.0 | 3888 | 1.7073 |
1.4696 | 9.0 | 4374 | 1.7187 |
1.4191 | 10.0 | 4860 | 1.7298 |
1.4776 | 11.0 | 5346 | 1.7414 |
1.4767 | 12.0 | 5832 | 1.7512 |
1.3546 | 13.0 | 6318 | 1.7731 |
1.542 | 14.0 | 6804 | 1.7610 |
1.3709 | 15.0 | 7290 | 1.7679 |
1.3167 | 16.0 | 7776 | 1.7936 |
1.3563 | 17.0 | 8262 | 1.8007 |
1.4615 | 18.0 | 8748 | 1.8008 |
1.511 | 19.0 | 9234 | 1.8068 |
1.3145 | 20.0 | 9720 | 1.8232 |
1.1285 | 21.0 | 10206 | 1.8204 |
1.5045 | 22.0 | 10692 | 1.8204 |
1.2697 | 23.0 | 11178 | 1.8453 |
1.302 | 24.0 | 11664 | 1.8386 |
1.4892 | 25.0 | 12150 | 1.8434 |
1.5042 | 26.0 | 12636 | 1.8471 |
1.1989 | 27.0 | 13122 | 1.8472 |
1.2353 | 28.0 | 13608 | 1.8545 |
1.145 | 29.0 | 14094 | 1.8560 |
1.4146 | 30.0 | 14580 | 1.8612 |
1.3598 | 31.0 | 15066 | 1.8611 |
1.2659 | 32.0 | 15552 | 1.8695 |
1.2085 | 33.0 | 16038 | 1.8631 |
1.0623 | 34.0 | 16524 | 1.8679 |
1.4594 | 35.0 | 17010 | 1.8694 |
1.3038 | 36.0 | 17496 | 1.8685 |
1.5902 | 37.0 | 17982 | 1.8695 |
1.2771 | 38.0 | 18468 | 1.8709 |
1.2738 | 39.0 | 18954 | 1.8698 |
1.3209 | 40.0 | 19440 | 1.8707 |
1.2578 | 41.0 | 19926 | 1.8709 |
1.1108 | 42.0 | 20412 | 1.8717 |
1.3264 | 43.0 | 20898 | 1.8711 |
1.3152 | 44.0 | 21384 | 1.8709 |
1.4287 | 45.0 | 21870 | 1.8709 |
1.299 | 46.0 | 22356 | 1.8709 |
1.2863 | 47.0 | 22842 | 1.8710 |
1.1795 | 48.0 | 23328 | 1.8716 |
1.27 | 49.0 | 23814 | 1.8719 |
1.3156 | 50.0 | 24300 | 1.8715 |
Framework versions
- PEFT 0.11.1
- Transformers 4.41.2
- Pytorch 2.1.0+cu118
- Datasets 2.19.2
- Tokenizers 0.19.1