Model save
Browse files
README.md
ADDED
@@ -0,0 +1,59 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: mistralai/Mistral-7B-Instruct-v0.2
|
3 |
+
library_name: peft
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- trl
|
7 |
+
- orpo
|
8 |
+
- generated_from_trainer
|
9 |
+
model-index:
|
10 |
+
- name: Mistral-7B-Instruct-v0.2-mergedORPO-r8
|
11 |
+
results: []
|
12 |
+
---
|
13 |
+
|
14 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
15 |
+
should probably proofread and complete it, then remove this comment. -->
|
16 |
+
|
17 |
+
# Mistral-7B-Instruct-v0.2-mergedORPO-r8
|
18 |
+
|
19 |
+
This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
|
20 |
+
|
21 |
+
## Model description
|
22 |
+
|
23 |
+
More information needed
|
24 |
+
|
25 |
+
## Intended uses & limitations
|
26 |
+
|
27 |
+
More information needed
|
28 |
+
|
29 |
+
## Training and evaluation data
|
30 |
+
|
31 |
+
More information needed
|
32 |
+
|
33 |
+
## Training procedure
|
34 |
+
|
35 |
+
### Training hyperparameters
|
36 |
+
|
37 |
+
The following hyperparameters were used during training:
|
38 |
+
- learning_rate: 0.0001
|
39 |
+
- train_batch_size: 1
|
40 |
+
- eval_batch_size: 1
|
41 |
+
- seed: 42
|
42 |
+
- gradient_accumulation_steps: 64
|
43 |
+
- total_train_batch_size: 64
|
44 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
45 |
+
- lr_scheduler_type: linear
|
46 |
+
- lr_scheduler_warmup_steps: 10
|
47 |
+
- num_epochs: 1
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
|
52 |
+
|
53 |
+
### Framework versions
|
54 |
+
|
55 |
+
- PEFT 0.12.0
|
56 |
+
- Transformers 4.44.2
|
57 |
+
- Pytorch 2.4.0+cu121
|
58 |
+
- Datasets 2.21.0
|
59 |
+
- Tokenizers 0.19.1
|
last-checkpoint/adapter_config.json
CHANGED
@@ -20,13 +20,13 @@
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
-
"
|
24 |
"q_proj",
|
25 |
"o_proj",
|
|
|
26 |
"gate_proj",
|
27 |
"k_proj",
|
28 |
-
"
|
29 |
-
"up_proj"
|
30 |
],
|
31 |
"task_type": "CAUSAL_LM",
|
32 |
"use_dora": false,
|
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
+
"up_proj",
|
24 |
"q_proj",
|
25 |
"o_proj",
|
26 |
+
"v_proj",
|
27 |
"gate_proj",
|
28 |
"k_proj",
|
29 |
+
"down_proj"
|
|
|
30 |
],
|
31 |
"task_type": "CAUSAL_LM",
|
32 |
"use_dora": false,
|
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 83945296
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:349c41609a51e6cfc6896a2a5d1a67153c68f9e2b7976427d560cec1ef2695d0
|
3 |
size 83945296
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 43706258
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66a5843d74545eb236796350e0ece8c200e91256593c92b31b5d53bd3686a8f4
|
3 |
size 43706258
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5f6daa0fed03c921127a64c255f462e66919bc76a6e6ba1e9093cf71d1068c6f
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
-
"epoch": 0.
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -799,96 +799,6 @@
|
|
799 |
"rewards/margins": 0.0255814790725708,
|
800 |
"rewards/rejected": -0.08916367590427399,
|
801 |
"step": 44
|
802 |
-
},
|
803 |
-
{
|
804 |
-
"epoch": 0.9174896463841988,
|
805 |
-
"grad_norm": 1.5401713848114014,
|
806 |
-
"learning_rate": 1.0256410256410256e-05,
|
807 |
-
"log_odds_chosen": 0.5436590313911438,
|
808 |
-
"log_odds_ratio": -0.49072104692459106,
|
809 |
-
"logits/chosen": -2.357189655303955,
|
810 |
-
"logits/rejected": -2.368823528289795,
|
811 |
-
"logps/chosen": -0.6493522524833679,
|
812 |
-
"logps/rejected": -0.9346315860748291,
|
813 |
-
"loss": 0.9199,
|
814 |
-
"nll_loss": 0.8708474040031433,
|
815 |
-
"rewards/accuracies": 0.859375,
|
816 |
-
"rewards/chosen": -0.06493522971868515,
|
817 |
-
"rewards/margins": 0.02852793037891388,
|
818 |
-
"rewards/rejected": -0.09346316009759903,
|
819 |
-
"step": 45
|
820 |
-
},
|
821 |
-
{
|
822 |
-
"epoch": 0.9378783051927365,
|
823 |
-
"grad_norm": 1.42889404296875,
|
824 |
-
"learning_rate": 7.692307692307694e-06,
|
825 |
-
"log_odds_chosen": 0.49409711360931396,
|
826 |
-
"log_odds_ratio": -0.506333589553833,
|
827 |
-
"logits/chosen": -2.390031576156616,
|
828 |
-
"logits/rejected": -2.3763091564178467,
|
829 |
-
"logps/chosen": -0.6777428984642029,
|
830 |
-
"logps/rejected": -0.9434182643890381,
|
831 |
-
"loss": 0.9246,
|
832 |
-
"nll_loss": 0.8739202618598938,
|
833 |
-
"rewards/accuracies": 0.828125,
|
834 |
-
"rewards/chosen": -0.06777428835630417,
|
835 |
-
"rewards/margins": 0.02656753547489643,
|
836 |
-
"rewards/rejected": -0.09434183686971664,
|
837 |
-
"step": 46
|
838 |
-
},
|
839 |
-
{
|
840 |
-
"epoch": 0.9582669640012743,
|
841 |
-
"grad_norm": 1.4630545377731323,
|
842 |
-
"learning_rate": 5.128205128205128e-06,
|
843 |
-
"log_odds_chosen": 0.5089167356491089,
|
844 |
-
"log_odds_ratio": -0.49921250343322754,
|
845 |
-
"logits/chosen": -2.377633810043335,
|
846 |
-
"logits/rejected": -2.36307954788208,
|
847 |
-
"logps/chosen": -0.6796414256095886,
|
848 |
-
"logps/rejected": -0.9439955353736877,
|
849 |
-
"loss": 0.969,
|
850 |
-
"nll_loss": 0.9190601110458374,
|
851 |
-
"rewards/accuracies": 0.828125,
|
852 |
-
"rewards/chosen": -0.06796413660049438,
|
853 |
-
"rewards/margins": 0.02643541246652603,
|
854 |
-
"rewards/rejected": -0.09439954906702042,
|
855 |
-
"step": 47
|
856 |
-
},
|
857 |
-
{
|
858 |
-
"epoch": 0.978655622809812,
|
859 |
-
"grad_norm": 1.4280872344970703,
|
860 |
-
"learning_rate": 2.564102564102564e-06,
|
861 |
-
"log_odds_chosen": 0.5206915736198425,
|
862 |
-
"log_odds_ratio": -0.4926137626171112,
|
863 |
-
"logits/chosen": -2.381681442260742,
|
864 |
-
"logits/rejected": -2.3803200721740723,
|
865 |
-
"logps/chosen": -0.6376308798789978,
|
866 |
-
"logps/rejected": -0.9052294492721558,
|
867 |
-
"loss": 0.894,
|
868 |
-
"nll_loss": 0.8447284698486328,
|
869 |
-
"rewards/accuracies": 0.890625,
|
870 |
-
"rewards/chosen": -0.0637630894780159,
|
871 |
-
"rewards/margins": 0.026759855449199677,
|
872 |
-
"rewards/rejected": -0.09052293747663498,
|
873 |
-
"step": 48
|
874 |
-
},
|
875 |
-
{
|
876 |
-
"epoch": 0.9990442816183498,
|
877 |
-
"grad_norm": 1.549211859703064,
|
878 |
-
"learning_rate": 0.0,
|
879 |
-
"log_odds_chosen": 0.5898391008377075,
|
880 |
-
"log_odds_ratio": -0.47122782468795776,
|
881 |
-
"logits/chosen": -2.377497911453247,
|
882 |
-
"logits/rejected": -2.376384735107422,
|
883 |
-
"logps/chosen": -0.6647161245346069,
|
884 |
-
"logps/rejected": -0.974918007850647,
|
885 |
-
"loss": 0.9571,
|
886 |
-
"nll_loss": 0.9099375605583191,
|
887 |
-
"rewards/accuracies": 0.875,
|
888 |
-
"rewards/chosen": -0.06647161394357681,
|
889 |
-
"rewards/margins": 0.031020179390907288,
|
890 |
-
"rewards/rejected": -0.0974918082356453,
|
891 |
-
"step": 49
|
892 |
}
|
893 |
],
|
894 |
"logging_steps": 1,
|
@@ -903,7 +813,7 @@
|
|
903 |
"should_evaluate": false,
|
904 |
"should_log": false,
|
905 |
"should_save": true,
|
906 |
-
"should_training_stop":
|
907 |
},
|
908 |
"attributes": {}
|
909 |
}
|
|
|
1 |
{
|
2 |
"best_metric": null,
|
3 |
"best_model_checkpoint": null,
|
4 |
+
"epoch": 0.897100987575661,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 44,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
799 |
"rewards/margins": 0.0255814790725708,
|
800 |
"rewards/rejected": -0.08916367590427399,
|
801 |
"step": 44
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
802 |
}
|
803 |
],
|
804 |
"logging_steps": 1,
|
|
|
813 |
"should_evaluate": false,
|
814 |
"should_log": false,
|
815 |
"should_save": true,
|
816 |
+
"should_training_stop": false
|
817 |
},
|
818 |
"attributes": {}
|
819 |
}
|
last-checkpoint/training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5560
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c6bcc3f11131b07d173fb79ad8c31d1a50ed3843eadae9ec2e4213021d99bed1
|
3 |
size 5560
|