liuylhf commited on
Commit
52a2824
1 Parent(s): 3f692ea

Model save

Browse files
Files changed (1) hide show
  1. README.md +6 -24
README.md CHANGED
@@ -2,7 +2,6 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
- - axolotl
6
  - generated_from_trainer
7
  base_model: mistralai/Mistral-7B-Instruct-v0.2
8
  model-index:
@@ -37,7 +36,7 @@ datasets:
37
 
38
  dataset_prepared_path: last_run_prepared
39
  val_set_size: 0.1
40
- output_dir: ../../text-generation-webui/loras/mistral-instruct-raw-format-v2-more-positive-new-tokens-inst
41
 
42
  adapter: lora
43
  lora_model_dir:
@@ -52,14 +51,6 @@ lora_dropout: 0.1
52
  lora_target_linear: true
53
  lora_fan_in_fan_out:
54
 
55
- lora_modules_to_save:
56
- - embed_tokens
57
- - lm_head
58
-
59
- tokens:
60
- function_token: "[f]"
61
- conversation_token: "[c]"
62
-
63
  hub_model_id: liuylhf/mistral-lora
64
 
65
  wandb_project: function-call
@@ -68,7 +59,7 @@ wandb_log_model: end
68
 
69
  gradient_accumulation_steps: 4
70
  micro_batch_size: 2
71
- num_epochs: 1
72
  optimizer: paged_adamw_8bit
73
  lr_scheduler: cosine
74
  learning_rate: 0.001
@@ -95,8 +86,8 @@ flash_attention: true
95
 
96
  warmup_steps: 10
97
  # evals_per_epoch: 20
98
- # eval_steps: 0.1
99
- save_steps: 0.1
100
  eval_table_size:
101
  eval_max_new_tokens: 256
102
  # saves_per_epoch: 1
@@ -111,9 +102,7 @@ fsdp_config:
111
 
112
  # mistral-lora
113
 
114
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the None dataset.
115
- It achieves the following results on the evaluation set:
116
- - Loss: 0.2163
117
 
118
  ## Model description
119
 
@@ -144,14 +133,7 @@ The following hyperparameters were used during training:
144
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-05
145
  - lr_scheduler_type: cosine
146
  - lr_scheduler_warmup_steps: 10
147
- - num_epochs: 1
148
-
149
- ### Training results
150
-
151
- | Training Loss | Epoch | Step | Validation Loss |
152
- |:-------------:|:-----:|:----:|:---------------:|
153
- | 0.149 | 1.0 | 304 | 0.2163 |
154
-
155
 
156
  ### Framework versions
157
 
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - generated_from_trainer
6
  base_model: mistralai/Mistral-7B-Instruct-v0.2
7
  model-index:
 
36
 
37
  dataset_prepared_path: last_run_prepared
38
  val_set_size: 0.1
39
+ output_dir: ../../text-generation-webui/loras/mistral-instruct-raw-format-v2-more-positive-inst
40
 
41
  adapter: lora
42
  lora_model_dir:
 
51
  lora_target_linear: true
52
  lora_fan_in_fan_out:
53
 
 
 
 
 
 
 
 
 
54
  hub_model_id: liuylhf/mistral-lora
55
 
56
  wandb_project: function-call
 
59
 
60
  gradient_accumulation_steps: 4
61
  micro_batch_size: 2
62
+ num_epochs: 0.5
63
  optimizer: paged_adamw_8bit
64
  lr_scheduler: cosine
65
  learning_rate: 0.001
 
86
 
87
  warmup_steps: 10
88
  # evals_per_epoch: 20
89
+ eval_steps: 0.2
90
+ save_steps: 0.2
91
  eval_table_size:
92
  eval_max_new_tokens: 256
93
  # saves_per_epoch: 1
 
102
 
103
  # mistral-lora
104
 
105
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
 
 
106
 
107
  ## Model description
108
 
 
133
  - optimizer: Adam with betas=(0.9,0.95) and epsilon=1e-05
134
  - lr_scheduler_type: cosine
135
  - lr_scheduler_warmup_steps: 10
136
+ - num_epochs: 0.5
 
 
 
 
 
 
 
137
 
138
  ### Framework versions
139