liuylhf commited on
Commit
01b640c
1 Parent(s): 6ea7144

Model save

Browse files
README.md CHANGED
@@ -2,7 +2,6 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
- - axolotl
6
  - generated_from_trainer
7
  base_model: mistralai/Mistral-7B-Instruct-v0.2
8
  model-index:
@@ -18,94 +17,74 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  axolotl version: `0.4.0`
20
  ```yaml
 
 
 
21
  base_model: mistralai/Mistral-7B-Instruct-v0.2
22
- model_type: AutoModelForCausalLM
23
- tokenizer_type: LlamaTokenizer
24
- is_mistral_derived_model: true
25
- load_in_8bit: false
26
- load_in_4bit: false
27
- strict: false
28
  chat_template: inst
29
-
30
- datasets:
31
- - path: ./data/raw_format/tool_used_training.jsonl
32
- type: sharegpt
33
- - path: ./data/raw_format/tool_not_used_training.jsonl
34
- type: sharegpt
35
- - path: ./data/raw_format/no_tools_training.jsonl
36
- type: sharegpt
37
-
38
  dataset_prepared_path: last_run_prepared
39
- val_set_size: 0.1
40
- output_dir: ../../text-generation-webui/loras/mistral-instruct-raw-format-v2-more-positive-inst
41
-
42
- adapter: lora
43
- lora_model_dir:
44
-
45
- sequence_len: 4096
46
- sample_packing: true
47
- pad_to_sequence_len: true
48
-
49
- lora_r: 16
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  lora_alpha: 16
51
  lora_dropout: 0.1
 
 
 
52
  lora_target_linear: true
53
- lora_fan_in_fan_out:
54
-
55
- hub_model_id: liuylhf/mistral-lora
56
-
57
- wandb_project: function-call
58
- wandb_name: mixtral-instruct-qlora-v1
59
- wandb_log_model: end
60
-
61
- gradient_accumulation_steps: 4
62
  micro_batch_size: 2
 
63
  num_epochs: 2
64
  optimizer: paged_adamw_8bit
65
- lr_scheduler: cosine
66
- learning_rate: 0.001
67
- adam_beta2: 0.95
68
- adam_epsilon: 0.00001
69
- max_grad_norm: 1.0
70
-
71
- train_on_inputs: false
72
- group_by_length: false
73
- bf16: auto
74
- fp16:
75
  tf32: false
76
-
77
- gradient_checkpointing: true
78
- early_stopping_patience:
79
- resume_from_checkpoint:
80
- local_rank:
81
- logging_steps: 1
82
- xformers_attention:
83
- flash_attention: true
84
-
85
- # loss_watchdog_threshold: 5.0
86
- # loss_watchdog_patience: 3
87
-
88
  warmup_steps: 10
89
- # evals_per_epoch: 20
90
- eval_steps: 0.2
91
- save_steps: 0.2
92
- eval_table_size:
93
- eval_max_new_tokens: 256
94
- # saves_per_epoch: 1
95
- debug:
96
- deepspeed:
97
  weight_decay: 1.0
98
- fsdp:
99
- fsdp_config:
100
  ```
101
 
102
  </details><br>
103
 
104
  # mistral-lora
105
 
106
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the None dataset.
107
- It achieves the following results on the evaluation set:
108
- - Loss: 0.1520
109
 
110
  ## Model description
111
 
@@ -138,21 +117,10 @@ The following hyperparameters were used during training:
138
  - lr_scheduler_warmup_steps: 10
139
  - num_epochs: 2
140
 
141
- ### Training results
142
-
143
- | Training Loss | Epoch | Step | Validation Loss |
144
- |:-------------:|:-----:|:----:|:---------------:|
145
- | 1.3787 | 0.0 | 1 | 1.4156 |
146
- | 0.1825 | 0.4 | 122 | 0.1627 |
147
- | 0.1398 | 0.8 | 244 | 0.1575 |
148
- | 0.0862 | 1.19 | 366 | 0.1562 |
149
- | 0.0868 | 1.59 | 488 | 0.1520 |
150
-
151
-
152
  ### Framework versions
153
 
154
  - PEFT 0.8.2
155
- - Transformers 4.38.0.dev0
156
  - Pytorch 2.2.0+cu121
157
  - Datasets 2.17.1
158
  - Tokenizers 0.15.0
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - generated_from_trainer
6
  base_model: mistralai/Mistral-7B-Instruct-v0.2
7
  model-index:
 
17
 
18
  axolotl version: `0.4.0`
19
  ```yaml
20
+ adam_beta2: 0.95
21
+ adam_epsilon: 1.0e-05
22
+ adapter: lora
23
  base_model: mistralai/Mistral-7B-Instruct-v0.2
24
+ bf16: auto
 
 
 
 
 
25
  chat_template: inst
 
 
 
 
 
 
 
 
 
26
  dataset_prepared_path: last_run_prepared
27
+ datasets:
28
+ - conversation: mistral
29
+ path: 4e9501d816a24795b7d619faea6fe0b7/./data/raw_format/tool_used_training_small.jsonl
30
+ type: sharegpt
31
+ debug: null
32
+ deepspeed: null
33
+ early_stopping_patience: null
34
+ eval_max_new_tokens: 256
35
+ eval_steps: 0.2
36
+ eval_table_size: null
37
+ flash_attention: true
38
+ fp16: null
39
+ fsdp: null
40
+ fsdp_config: null
41
+ gradient_accumulation_steps: 4
42
+ gradient_checkpointing: true
43
+ group_by_length: false
44
+ hub_model_id: liuylhf/mistral-lora
45
+ is_mistral_derived_model: true
46
+ learning_rate: 0.001
47
+ load_in_4bit: false
48
+ load_in_8bit: false
49
+ local_rank: null
50
+ logging_steps: 1
51
  lora_alpha: 16
52
  lora_dropout: 0.1
53
+ lora_fan_in_fan_out: null
54
+ lora_model_dir: null
55
+ lora_r: 16
56
  lora_target_linear: true
57
+ lr_scheduler: cosine
58
+ max_grad_norm: 1.0
 
 
 
 
 
 
 
59
  micro_batch_size: 2
60
+ model_type: AutoModelForCausalLM
61
  num_epochs: 2
62
  optimizer: paged_adamw_8bit
63
+ output_dir: ../../text-generation-webui/loras/mistral-instruct-raw-format-v2-more-positive-inst
64
+ pad_to_sequence_len: true
65
+ resume_from_checkpoint: null
66
+ sample_packing: true
67
+ save_steps: 0.2
68
+ sequence_len: 4096
69
+ strict: false
 
 
 
70
  tf32: false
71
+ tokenizer_type: LlamaTokenizer
72
+ train_on_inputs: false
73
+ val_set_size: 0.1
74
+ wandb_log_model: end
75
+ wandb_name: mixtral-instruct-qlora-v1
76
+ wandb_project: function-call
 
 
 
 
 
 
77
  warmup_steps: 10
 
 
 
 
 
 
 
 
78
  weight_decay: 1.0
79
+ xformers_attention: null
80
+
81
  ```
82
 
83
  </details><br>
84
 
85
  # mistral-lora
86
 
87
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
 
 
88
 
89
  ## Model description
90
 
 
117
  - lr_scheduler_warmup_steps: 10
118
  - num_epochs: 2
119
 
 
 
 
 
 
 
 
 
 
 
 
120
  ### Framework versions
121
 
122
  - PEFT 0.8.2
123
+ - Transformers 4.39.0.dev0
124
  - Pytorch 2.2.0+cu121
125
  - Datasets 2.17.1
126
  - Tokenizers 0.15.0
adapter_config.json CHANGED
@@ -19,13 +19,13 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
  "down_proj",
24
- "up_proj",
25
- "v_proj",
26
  "q_proj",
27
- "gate_proj",
28
- "k_proj"
 
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "gate_proj",
23
  "down_proj",
24
+ "o_proj",
 
25
  "q_proj",
26
+ "k_proj",
27
+ "v_proj",
28
+ "up_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "use_rslora": false
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6b79d285afc334ab5611226c2e84e5281b514dc161f24757d0161c5ca6cec0f
3
  size 84047946
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc4c49f4231047910e5b89dd68dce81cbac2f8438273133924b43ce4d7b32bd4
3
  size 84047946
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1e322c9e264875578ffd3376c0b176458cdf2f9a545ed4ef00e6f01df4450a7
3
  size 83946192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5e53c2bbc490a6f3bf85f23c969433fb56992e73dcd5ac2b9489a7ffb199196
3
  size 83946192
config.json CHANGED
@@ -20,7 +20,7 @@
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
- "transformers_version": "4.38.0.dev0",
24
  "use_cache": false,
25
  "vocab_size": 32000
26
  }
 
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.0.dev0",
24
  "use_cache": false,
25
  "vocab_size": 32000
26
  }
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -38,7 +39,6 @@
38
  "sp_model_kwargs": {},
39
  "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
41
- "trust_remote_code": false,
42
  "unk_token": "<unk>",
43
  "use_default_system_prompt": false,
44
  "use_fast": true
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
41
  "tokenizer_class": "LlamaTokenizer",
 
42
  "unk_token": "<unk>",
43
  "use_default_system_prompt": false,
44
  "use_fast": true
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2de5a83572cd0d10a4f700ce4fc93c48a03a750f8117aacdd75dff8cef72784
3
- size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4132915575eea232b6b96893db1644b5e5e781de46620bde99e6ca920cfad17e
3
+ size 5752