liuylhf commited on
Commit
bb0480f
1 Parent(s): 7a6c7d4

Model save

Browse files
README.md CHANGED
@@ -2,7 +2,6 @@
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
5
- - axolotl
6
  - generated_from_trainer
7
  base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
8
  model-index:
@@ -27,10 +26,10 @@ chat_template: inst
27
  dataset_prepared_path: last_run_prepared
28
  datasets:
29
  - conversation: mistral
30
- path: 14050e2bd5b9487a9afae30b4c86d554/./data/with_function_response/function_not_used.jsonl
31
  type: sharegpt
32
  - conversation: mistral
33
- path: 14050e2bd5b9487a9afae30b4c86d554/./data/with_function_response/function_used.jsonl
34
  type: sharegpt
35
  debug: null
36
  deepspeed: null
@@ -69,7 +68,7 @@ model_config:
69
  model_type: AutoModelForCausalLM
70
  num_epochs: 1
71
  optimizer: paged_adamw_8bit
72
- output_dir: 14050e2bd5b9487a9afae30b4c86d554/model
73
  pad_to_sequence_len: true
74
  resume_from_checkpoint: null
75
  sample_packing: true
@@ -93,9 +92,7 @@ xformers_attention: null
93
 
94
  # mixtral-remove-negative-data
95
 
96
- This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the None dataset.
97
- It achieves the following results on the evaluation set:
98
- - Loss: 0.0936
99
 
100
  ## Model description
101
 
@@ -128,17 +125,6 @@ The following hyperparameters were used during training:
128
  - lr_scheduler_warmup_steps: 10
129
  - num_epochs: 1
130
 
131
- ### Training results
132
-
133
- | Training Loss | Epoch | Step | Validation Loss |
134
- |:-------------:|:-----:|:----:|:---------------:|
135
- | 4.0791 | 0.0 | 1 | 4.0657 |
136
- | 0.1507 | 0.2 | 56 | 0.1291 |
137
- | 0.1178 | 0.4 | 112 | 0.1052 |
138
- | 0.1046 | 0.61 | 168 | 0.0977 |
139
- | 0.102 | 0.81 | 224 | 0.0936 |
140
-
141
-
142
  ### Framework versions
143
 
144
  - PEFT 0.8.2
 
2
  license: apache-2.0
3
  library_name: peft
4
  tags:
 
5
  - generated_from_trainer
6
  base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
7
  model-index:
 
26
  dataset_prepared_path: last_run_prepared
27
  datasets:
28
  - conversation: mistral
29
+ path: 497b7e7687f94b09b0e633b679727a04/./data/with_function_response/function_not_used_training.jsonl
30
  type: sharegpt
31
  - conversation: mistral
32
+ path: 497b7e7687f94b09b0e633b679727a04/./data/with_function_response/function_used_training.jsonl
33
  type: sharegpt
34
  debug: null
35
  deepspeed: null
 
68
  model_type: AutoModelForCausalLM
69
  num_epochs: 1
70
  optimizer: paged_adamw_8bit
71
+ output_dir: 497b7e7687f94b09b0e633b679727a04/model
72
  pad_to_sequence_len: true
73
  resume_from_checkpoint: null
74
  sample_packing: true
 
92
 
93
  # mixtral-remove-negative-data
94
 
95
+ This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on an unknown dataset.
 
 
96
 
97
  ## Model description
98
 
 
125
  - lr_scheduler_warmup_steps: 10
126
  - num_epochs: 1
127
 
 
 
 
 
 
 
 
 
 
 
 
128
  ### Framework versions
129
 
130
  - PEFT 0.8.2
adapter_config.json CHANGED
@@ -20,9 +20,9 @@
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
23
- "q_proj",
24
  "o_proj",
25
- "k_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
 
20
  "revision": null,
21
  "target_modules": [
22
  "v_proj",
23
+ "k_proj",
24
  "o_proj",
25
+ "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_rslora": false
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:baa33ea1c35f01d785b890afec3242a7156fda827d1569188ae1abf34c177d13
3
  size 109086416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e38812772ac55f93a5c705bc028f6bda448c0ea78a32f7b9be1d647caaba5224
3
  size 109086416
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ffdef099be9ab3284c1e5a2c92dfe97b9bbcd4e8428fcac33d5afa90f34ccc4
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d41dbd958da6f59f7b0c8f21d54fd6a4ca27c71d3c6d51e8942b99a2c60ebf9
3
  size 5624