liuylhf commited on
Commit
1d72239
1 Parent(s): ceef875

Model save

Browse files
Files changed (2) hide show
  1. README.md +152 -0
  2. adapter_model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,152 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ library_name: peft
4
+ tags:
5
+ - generated_from_trainer
6
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
7
+ model-index:
8
+ - name: mistral-lora
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
16
+ <details><summary>See axolotl config</summary>
17
+
18
+ axolotl version: `0.4.0`
19
+ ```yaml
20
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
21
+ model_type: AutoModelForCausalLM
22
+ tokenizer_type: LlamaTokenizer
23
+ is_mistral_derived_model: true
24
+ load_in_8bit: false
25
+ load_in_4bit: false
26
+ strict: false
27
+ chat_template: inst
28
+
29
+ datasets:
30
+ - path: ./data/raw_format/tool_used_training.jsonl
31
+ type: sharegpt
32
+ - path: ./data/raw_format/tool_not_used_training.jsonl
33
+ type: sharegpt
34
+ - path: ./data/raw_format/no_tools_training.jsonl
35
+ type: sharegpt
36
+
37
+ dataset_prepared_path: last_run_prepared
38
+ val_set_size: 0.1
39
+ output_dir: ../../text-generation-webui/loras/mistral-instruct-raw-format-v2-more-positive-new-tokens-inst
40
+
41
+ adapter: lora
42
+ lora_model_dir:
43
+
44
+ sequence_len: 4096
45
+ sample_packing: true
46
+ pad_to_sequence_len: true
47
+
48
+ lora_r: 16
49
+ lora_alpha: 16
50
+ lora_dropout: 0.1
51
+ lora_target_linear: true
52
+ lora_fan_in_fan_out:
53
+
54
+ lora_modules_to_save:
55
+ - embed_tokens
56
+ - lm_head
57
+
58
+ tokens:
59
+ function_token: "[f]"
60
+ conversation_token: "[c]"
61
+
62
+ hub_model_id: liuylhf/mistral-lora
63
+
64
+ wandb_project: function-call
65
+ wandb_name: mixtral-instruct-qlora-v1
66
+ wandb_log_model: end
67
+
68
+ gradient_accumulation_steps: 4
69
+ micro_batch_size: 2
70
+ num_epochs: 1
71
+ optimizer: paged_adamw_8bit
72
+ lr_scheduler: cosine
73
+ learning_rate: 0.001
74
+ adam_beta2: 0.95
75
+ adam_epsilon: 0.00001
76
+ max_grad_norm: 1.0
77
+
78
+ train_on_inputs: false
79
+ group_by_length: false
80
+ bf16: auto
81
+ fp16:
82
+ tf32: false
83
+
84
+ gradient_checkpointing: true
85
+ early_stopping_patience:
86
+ resume_from_checkpoint:
87
+ local_rank:
88
+ logging_steps: 1
89
+ xformers_attention:
90
+ flash_attention: true
91
+
92
+ # loss_watchdog_threshold: 5.0
93
+ # loss_watchdog_patience: 3
94
+
95
+ warmup_steps: 10
96
+ # evals_per_epoch: 20
97
+ # eval_steps: 0.1
98
+ save_steps: 0.1
99
+ eval_table_size:
100
+ eval_max_new_tokens: 256
101
+ # saves_per_epoch: 1
102
+ debug:
103
+ deepspeed:
104
+ weight_decay: 1.0
105
+ fsdp:
106
+ fsdp_config:
107
+ ```
108
+
109
+ </details><br>
110
+
111
+ # mistral-lora
112
+
113
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
114
+
115
+ ## Model description
116
+
117
+ More information needed
118
+
119
+ ## Intended uses & limitations
120
+
121
+ More information needed
122
+
123
+ ## Training and evaluation data
124
+
125
+ More information needed
126
+
127
+ ## Training procedure
128
+
129
+ ### Training hyperparameters
130
+
131
+ The following hyperparameters were used during training:
132
+ - learning_rate: 0.001
133
+ - train_batch_size: 2
134
+ - eval_batch_size: 2
135
+ - seed: 42
136
+ - distributed_type: multi-GPU
137
+ - num_devices: 2
138
+ - gradient_accumulation_steps: 4
139
+ - total_train_batch_size: 16
140
+ - total_eval_batch_size: 4
141
+ - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-05
142
+ - lr_scheduler_type: cosine
143
+ - lr_scheduler_warmup_steps: 10
144
+ - num_epochs: 1
145
+
146
+ ### Framework versions
147
+
148
+ - PEFT 0.8.2
149
+ - Transformers 4.38.0.dev0
150
+ - Pytorch 2.2.0+cu121
151
+ - Datasets 2.17.1
152
+ - Tokenizers 0.15.0
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3452cb5d829a101799dea60b0ea266e833df23c067c2b92561543d5f07344e4
3
  size 1656911144
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:316878b4a49d1c171a8d228fdc437744ff5d02b881a48e5415e1b417e2a18a30
3
  size 1656911144