Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
language:
|
3 |
- en
|
4 |
license: apache-2.0
|
5 |
-
base_model: mistralai/Mixtral-8x7B-v0.1
|
6 |
inference:
|
7 |
parameters:
|
8 |
temperature: 0.5
|
@@ -12,7 +12,7 @@ widget:
|
|
12 |
content: What is your favorite condiment?
|
13 |
---
|
14 |
|
15 |
-
This model is compressed from the Mixtral-8x7B. Using Low-Rank Approximation, I removed 10 billion parameters from the MLP experts' matrices, enough to run the model on a single A100 80GB GPU using half precision.
|
16 |
|
17 |
|
18 |
Without being retrained or fine-tuned, the model still retains its core performance:
|
|
|
2 |
language:
|
3 |
- en
|
4 |
license: apache-2.0
|
5 |
+
base_model: mistralai/Mixtral-8x7B-v0.1-Instruct
|
6 |
inference:
|
7 |
parameters:
|
8 |
temperature: 0.5
|
|
|
12 |
content: What is your favorite condiment?
|
13 |
---
|
14 |
|
15 |
+
This model is compressed from the Mixtral-8x7B-Instruct. Using Low-Rank Approximation, I removed 10 billion parameters from the MLP experts' matrices, enough to run the model on a single A100 80GB GPU using half precision.
|
16 |
|
17 |
|
18 |
Without being retrained or fine-tuned, the model still retains its core performance:
|