Knightcodin
commited on
Commit
•
37d4d58
1
Parent(s):
952ed4f
Update README.md
Browse files
README.md
CHANGED
@@ -1,20 +1,20 @@
|
|
1 |
---
|
|
|
|
|
2 |
language:
|
3 |
- en
|
4 |
-
|
5 |
pipeline_tag: text-generation
|
6 |
-
tags:
|
7 |
-
- facebook
|
8 |
-
- meta
|
9 |
-
- pytorch
|
10 |
-
- llama
|
11 |
-
- llama-3
|
12 |
-
- moe
|
13 |
-
- code
|
14 |
-
license: other
|
15 |
-
license_name: llama3
|
16 |
-
license_link: LICENSE
|
17 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
|
20 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/657eb5b256c9c67605a6e8b5/8JXktjAyUPCWQGnRExiVI.png)
|
@@ -23,10 +23,6 @@ license_link: LICENSE
|
|
23 |
|
24 |
Aplite-Instruct-4x8B-Llama-3 is a experimental MoE (Mixture of Experts) model based on the Llama-3 architecture using Mergekit.
|
25 |
|
26 |
-
**GGUF Quant [Here](https://huggingface.co/raincandy-u/Llama-3-Aplite-Instruct-4x8B-GGUF)**
|
27 |
-
|
28 |
-
**<|eot_id|> bug has been fixed.**
|
29 |
-
|
30 |
## Disclaimer
|
31 |
|
32 |
This model is a research experiment and may generate incorrect or harmful content. The model's outputs should not be taken as factual or representative of the views of the model's creator or any other individual.
|
|
|
1 |
---
|
2 |
+
---
|
3 |
+
license: cc-by-nc-4.0
|
4 |
language:
|
5 |
- en
|
6 |
+
quantized_by: KnightCodin
|
7 |
pipeline_tag: text-generation
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8 |
---
|
9 |
+
## Exllama v2 Quantizations of L3-Solana-8B-v1
|
10 |
+
|
11 |
+
Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.19">turboderp's ExLlamaV2 v0.0.19</a> for quantization.
|
12 |
+
|
13 |
+
<b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
|
14 |
+
|
15 |
+
Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
|
16 |
+
|
17 |
+
Original model: https://huggingface.co/raincandy-u/Llama-3-Aplite-Instruct-4x8B
|
18 |
|
19 |
|
20 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/657eb5b256c9c67605a6e8b5/8JXktjAyUPCWQGnRExiVI.png)
|
|
|
23 |
|
24 |
Aplite-Instruct-4x8B-Llama-3 is a experimental MoE (Mixture of Experts) model based on the Llama-3 architecture using Mergekit.
|
25 |
|
|
|
|
|
|
|
|
|
26 |
## Disclaimer
|
27 |
|
28 |
This model is a research experiment and may generate incorrect or harmful content. The model's outputs should not be taken as factual or representative of the views of the model's creator or any other individual.
|