k-r-l commited on
Commit
74a27bb
1 Parent(s): 4cb3a30

Model save

Browse files
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
3
+ library_name: peft
4
+ license: apache-2.0
5
+ tags:
6
+ - trl
7
+ - orpo
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: Mistral-7B-Instruct-v0.2-mergedORPO-r8
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # Mistral-7B-Instruct-v0.2-mergedORPO-r8
18
+
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 0.0001
39
+ - train_batch_size: 1
40
+ - eval_batch_size: 1
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 64
43
+ - total_train_batch_size: 64
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_steps: 10
47
+ - num_epochs: 1
48
+
49
+ ### Training results
50
+
51
+
52
+
53
+ ### Framework versions
54
+
55
+ - PEFT 0.12.0
56
+ - Transformers 4.44.2
57
+ - Pytorch 2.4.0+cu121
58
+ - Datasets 2.21.0
59
+ - Tokenizers 0.19.1
last-checkpoint/adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "down_proj",
24
  "q_proj",
25
  "o_proj",
 
26
  "gate_proj",
27
  "k_proj",
28
- "v_proj",
29
- "up_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "up_proj",
24
  "q_proj",
25
  "o_proj",
26
+ "v_proj",
27
  "gate_proj",
28
  "k_proj",
29
+ "down_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a1e4c3b90d664cf744230b1635c69595b3e6feb863b87b843e43e3d3a7ae1d5
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:349c41609a51e6cfc6896a2a5d1a67153c68f9e2b7976427d560cec1ef2695d0
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2a154beda3c570368f5acddceb38cab4dc0cd4ce9afacb1a4b5ede7235e6a54
3
  size 43706258
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66a5843d74545eb236796350e0ece8c200e91256593c92b31b5d53bd3686a8f4
3
  size 43706258
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:25704ee5478cd4521ee5db3e9e14050e9c72edebe1088acf05e49ba2bb8ed7fb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f6daa0fed03c921127a64c255f462e66919bc76a6e6ba1e9093cf71d1068c6f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.9990442816183498,
5
  "eval_steps": 500,
6
- "global_step": 49,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -799,96 +799,6 @@
799
  "rewards/margins": 0.0255814790725708,
800
  "rewards/rejected": -0.08916367590427399,
801
  "step": 44
802
- },
803
- {
804
- "epoch": 0.9174896463841988,
805
- "grad_norm": 1.5401713848114014,
806
- "learning_rate": 1.0256410256410256e-05,
807
- "log_odds_chosen": 0.5436590313911438,
808
- "log_odds_ratio": -0.49072104692459106,
809
- "logits/chosen": -2.357189655303955,
810
- "logits/rejected": -2.368823528289795,
811
- "logps/chosen": -0.6493522524833679,
812
- "logps/rejected": -0.9346315860748291,
813
- "loss": 0.9199,
814
- "nll_loss": 0.8708474040031433,
815
- "rewards/accuracies": 0.859375,
816
- "rewards/chosen": -0.06493522971868515,
817
- "rewards/margins": 0.02852793037891388,
818
- "rewards/rejected": -0.09346316009759903,
819
- "step": 45
820
- },
821
- {
822
- "epoch": 0.9378783051927365,
823
- "grad_norm": 1.42889404296875,
824
- "learning_rate": 7.692307692307694e-06,
825
- "log_odds_chosen": 0.49409711360931396,
826
- "log_odds_ratio": -0.506333589553833,
827
- "logits/chosen": -2.390031576156616,
828
- "logits/rejected": -2.3763091564178467,
829
- "logps/chosen": -0.6777428984642029,
830
- "logps/rejected": -0.9434182643890381,
831
- "loss": 0.9246,
832
- "nll_loss": 0.8739202618598938,
833
- "rewards/accuracies": 0.828125,
834
- "rewards/chosen": -0.06777428835630417,
835
- "rewards/margins": 0.02656753547489643,
836
- "rewards/rejected": -0.09434183686971664,
837
- "step": 46
838
- },
839
- {
840
- "epoch": 0.9582669640012743,
841
- "grad_norm": 1.4630545377731323,
842
- "learning_rate": 5.128205128205128e-06,
843
- "log_odds_chosen": 0.5089167356491089,
844
- "log_odds_ratio": -0.49921250343322754,
845
- "logits/chosen": -2.377633810043335,
846
- "logits/rejected": -2.36307954788208,
847
- "logps/chosen": -0.6796414256095886,
848
- "logps/rejected": -0.9439955353736877,
849
- "loss": 0.969,
850
- "nll_loss": 0.9190601110458374,
851
- "rewards/accuracies": 0.828125,
852
- "rewards/chosen": -0.06796413660049438,
853
- "rewards/margins": 0.02643541246652603,
854
- "rewards/rejected": -0.09439954906702042,
855
- "step": 47
856
- },
857
- {
858
- "epoch": 0.978655622809812,
859
- "grad_norm": 1.4280872344970703,
860
- "learning_rate": 2.564102564102564e-06,
861
- "log_odds_chosen": 0.5206915736198425,
862
- "log_odds_ratio": -0.4926137626171112,
863
- "logits/chosen": -2.381681442260742,
864
- "logits/rejected": -2.3803200721740723,
865
- "logps/chosen": -0.6376308798789978,
866
- "logps/rejected": -0.9052294492721558,
867
- "loss": 0.894,
868
- "nll_loss": 0.8447284698486328,
869
- "rewards/accuracies": 0.890625,
870
- "rewards/chosen": -0.0637630894780159,
871
- "rewards/margins": 0.026759855449199677,
872
- "rewards/rejected": -0.09052293747663498,
873
- "step": 48
874
- },
875
- {
876
- "epoch": 0.9990442816183498,
877
- "grad_norm": 1.549211859703064,
878
- "learning_rate": 0.0,
879
- "log_odds_chosen": 0.5898391008377075,
880
- "log_odds_ratio": -0.47122782468795776,
881
- "logits/chosen": -2.377497911453247,
882
- "logits/rejected": -2.376384735107422,
883
- "logps/chosen": -0.6647161245346069,
884
- "logps/rejected": -0.974918007850647,
885
- "loss": 0.9571,
886
- "nll_loss": 0.9099375605583191,
887
- "rewards/accuracies": 0.875,
888
- "rewards/chosen": -0.06647161394357681,
889
- "rewards/margins": 0.031020179390907288,
890
- "rewards/rejected": -0.0974918082356453,
891
- "step": 49
892
  }
893
  ],
894
  "logging_steps": 1,
@@ -903,7 +813,7 @@
903
  "should_evaluate": false,
904
  "should_log": false,
905
  "should_save": true,
906
- "should_training_stop": true
907
  },
908
  "attributes": {}
909
  }
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.897100987575661,
5
  "eval_steps": 500,
6
+ "global_step": 44,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
799
  "rewards/margins": 0.0255814790725708,
800
  "rewards/rejected": -0.08916367590427399,
801
  "step": 44
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
802
  }
803
  ],
804
  "logging_steps": 1,
 
813
  "should_evaluate": false,
814
  "should_log": false,
815
  "should_save": true,
816
+ "should_training_stop": false
817
  },
818
  "attributes": {}
819
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:128f0cf3da8b1abd86530282a4bc926272323030313ebb6ed3dbec18e05701bf
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6bcc3f11131b07d173fb79ad8c31d1a50ed3843eadae9ec2e4213021d99bed1
3
  size 5560