Edit model card

lc_full

This model is a fine-tuned version of mistralai/Mistral-7B-v0.1 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.8715

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 1
  • eval_batch_size: 1
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • num_epochs: 50

Training results

Training Loss Epoch Step Validation Loss
1.7424 1.0 486 1.6914
1.301 2.0 972 1.6780
1.5718 3.0 1458 1.6743
1.6632 4.0 1944 1.6793
1.8588 5.0 2430 1.6794
1.5308 6.0 2916 1.6894
1.5776 7.0 3402 1.6985
1.6394 8.0 3888 1.7073
1.4696 9.0 4374 1.7187
1.4191 10.0 4860 1.7298
1.4776 11.0 5346 1.7414
1.4767 12.0 5832 1.7512
1.3546 13.0 6318 1.7731
1.542 14.0 6804 1.7610
1.3709 15.0 7290 1.7679
1.3167 16.0 7776 1.7936
1.3563 17.0 8262 1.8007
1.4615 18.0 8748 1.8008
1.511 19.0 9234 1.8068
1.3145 20.0 9720 1.8232
1.1285 21.0 10206 1.8204
1.5045 22.0 10692 1.8204
1.2697 23.0 11178 1.8453
1.302 24.0 11664 1.8386
1.4892 25.0 12150 1.8434
1.5042 26.0 12636 1.8471
1.1989 27.0 13122 1.8472
1.2353 28.0 13608 1.8545
1.145 29.0 14094 1.8560
1.4146 30.0 14580 1.8612
1.3598 31.0 15066 1.8611
1.2659 32.0 15552 1.8695
1.2085 33.0 16038 1.8631
1.0623 34.0 16524 1.8679
1.4594 35.0 17010 1.8694
1.3038 36.0 17496 1.8685
1.5902 37.0 17982 1.8695
1.2771 38.0 18468 1.8709
1.2738 39.0 18954 1.8698
1.3209 40.0 19440 1.8707
1.2578 41.0 19926 1.8709
1.1108 42.0 20412 1.8717
1.3264 43.0 20898 1.8711
1.3152 44.0 21384 1.8709
1.4287 45.0 21870 1.8709
1.299 46.0 22356 1.8709
1.2863 47.0 22842 1.8710
1.1795 48.0 23328 1.8716
1.27 49.0 23814 1.8719
1.3156 50.0 24300 1.8715

Framework versions

  • PEFT 0.11.1
  • Transformers 4.41.2
  • Pytorch 2.1.0+cu118
  • Datasets 2.19.2
  • Tokenizers 0.19.1
Downloads last month
0
Safetensors
Model size
7.24B params
Tensor type
BF16
·
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for nerottt/lc_full

Adapter
(1171)
this model

Space using nerottt/lc_full 1