End of training
Browse files- README.md +30 -1
- adapter_model.bin +3 -0
README.md
CHANGED
@@ -2,6 +2,7 @@
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
|
|
5 |
- generated_from_trainer
|
6 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
7 |
model-index:
|
@@ -86,7 +87,9 @@ weight_decay: 0.0
|
|
86 |
|
87 |
# empower-functions-clean-data-one-more-functions
|
88 |
|
89 |
-
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on
|
|
|
|
|
90 |
|
91 |
## Model description
|
92 |
|
@@ -119,6 +122,32 @@ The following hyperparameters were used during training:
|
|
119 |
- lr_scheduler_warmup_steps: 10
|
120 |
- num_epochs: 1
|
121 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
122 |
### Framework versions
|
123 |
|
124 |
- PEFT 0.9.0
|
|
|
2 |
license: apache-2.0
|
3 |
library_name: peft
|
4 |
tags:
|
5 |
+
- axolotl
|
6 |
- generated_from_trainer
|
7 |
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
8 |
model-index:
|
|
|
87 |
|
88 |
# empower-functions-clean-data-one-more-functions
|
89 |
|
90 |
+
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the None dataset.
|
91 |
+
It achieves the following results on the evaluation set:
|
92 |
+
- Loss: 0.0863
|
93 |
|
94 |
## Model description
|
95 |
|
|
|
122 |
- lr_scheduler_warmup_steps: 10
|
123 |
- num_epochs: 1
|
124 |
|
125 |
+
### Training results
|
126 |
+
|
127 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
128 |
+
|:-------------:|:-----:|:----:|:---------------:|
|
129 |
+
| 2.0157 | 0.0 | 1 | 2.1200 |
|
130 |
+
| 0.153 | 0.05 | 23 | 0.1454 |
|
131 |
+
| 0.1236 | 0.1 | 46 | 0.1160 |
|
132 |
+
| 0.1043 | 0.15 | 69 | 0.1073 |
|
133 |
+
| 0.1163 | 0.2 | 92 | 0.1035 |
|
134 |
+
| 0.1072 | 0.25 | 115 | 0.0996 |
|
135 |
+
| 0.0988 | 0.31 | 138 | 0.0978 |
|
136 |
+
| 0.0962 | 0.36 | 161 | 0.0963 |
|
137 |
+
| 0.0823 | 0.41 | 184 | 0.0939 |
|
138 |
+
| 0.0785 | 0.46 | 207 | 0.0938 |
|
139 |
+
| 0.0941 | 0.51 | 230 | 0.0918 |
|
140 |
+
| 0.0968 | 0.56 | 253 | 0.0905 |
|
141 |
+
| 0.0856 | 0.61 | 276 | 0.0899 |
|
142 |
+
| 0.0965 | 0.66 | 299 | 0.0895 |
|
143 |
+
| 0.0894 | 0.71 | 322 | 0.0881 |
|
144 |
+
| 0.086 | 0.76 | 345 | 0.0872 |
|
145 |
+
| 0.0941 | 0.82 | 368 | 0.0869 |
|
146 |
+
| 0.0894 | 0.87 | 391 | 0.0867 |
|
147 |
+
| 0.0782 | 0.92 | 414 | 0.0864 |
|
148 |
+
| 0.0815 | 0.97 | 437 | 0.0863 |
|
149 |
+
|
150 |
+
|
151 |
### Framework versions
|
152 |
|
153 |
- PEFT 0.9.0
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4c08c92b59732c88a22bc40bb57b9557f018107acd9ba6016a9bd1d9833ebffc
|
3 |
+
size 109144714
|