File size: 5,163 Bytes
c3ae0fc 3e7bf68 c3ae0fc bd75fba c3ae0fc 5cf01ba c3ae0fc 2797ceb 4d9a318 2797ceb 4d9a318 990ba86 4d9a318 29376dc 26136a8 2797ceb 4d9a318 2797ceb 4d9a318 26136a8 1cfe03a c3ae0fc 1cfe03a c3ae0fc 7ef75e6 58c1afc 9996285 58c1afc 2788441 58c1afc 7ef75e6 63a54dd 7ef75e6 c3ae0fc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 |
---
base_model:
- nothingiisreal/L3.1-8B-Celeste-V1.5
- Sao10K/Llama-3.1-8B-Stheno-v3.4
- Sao10K/L3.1-8B-Niitama-v1.1
- arcee-ai/Llama-3.1-SuperNova-Lite
- akjindal53244/Llama-3.1-Storm-8B
- arcee-ai/Llama-Spark
- grimjim/Llama-3-Instruct-abliteration-LoRA-8B
- crestf411/sunfall-peft
tags:
- llama
- merge
- llama3
- mixtral
library_name: transformers
---
> [!WARNING]
> **Content:**<br>
> This models output's can be a bit unhinged.
# Llama-3.1-Celestial-Stone-2x8B (BF16)
* *Mixture of Experts (14B).*
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/lBrXRa3sVRinE3cabs-oQ.png)
Both experts are used in tandem when generating a token.
------------------------------------------------------------------------------
* *Llama.CPP - GGUF.*
# Thank you mradermacher for the quants!
----> [GGUF iMatrix](https://huggingface.co/mradermacher/L3.1-Celestial-Stone-2x8B-i1-GGUF)
----> [GGUF static](https://huggingface.co/mradermacher/L3.1-Celestial-Stone-2x8B-GGUF)
# Thank you QuantFactory for the quants!:
----> [GGUF static](https://huggingface.co/QuantFactory/L3.1-Celestial-Stone-2x8B-GGUF)
Other alternative quants:
----> [Q8_0 GGUF by dasChronos1](https://huggingface.co/dasChronos1/L3.1-Celestial-Stone-2x8B-Q8_0-GGUF)
----> [Q6_K GGUF](https://huggingface.co/v000000/L3.1-Celestial-Stone-2x8B-Q6_K-GGUF)
----> [Q4_K_M GGUF by aashish1904](https://huggingface.co/aashish1904/L3.1-Celestial-Stone-2x8B-Q4_K_M-GGUF)
----> [Q2_K GGUF by aashish1904](https://huggingface.co/aashish1904/L3.1-Celestial-Stone-2x8B-Q2_K-GGUF)
------------------------------------------------------------------------------
*The first expert* is Instruct 405B distillation/RP vector merge <b>(Supernova-Lite, Niitama1.1, Storm)</b>
*The second expert* is ERP/Reddit data merge <b>(Celeste1.5, Stheno3.4, Storm)</b>
-------------------------------------------------------------------------------
*The base model* is <b>Sao10k/L3.1-Stheno-3.4</b> with the <b>Sunfall LoRa 0.6.1</b> to make it understand SillyTavern prompts and storywriting better.
-------------------------------------------------------------------------------
# Prompt Template:
```bash
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
{output}<|eot_id|>
```
* *Other Details:*
*The model has 131072 context length, and is on Llama-3.1 and Mixtral architecture.*
*I did not abliterate the base model at all, so it will refuse zero-shot unethical questions. I recommend avoiding keywords like 'assistant, helpful, kind'*
# Recipe (I'm sorry...):
```yaml
slices:
- sources:
- model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
layer_range: [0, 32]
- model: akjindal53244/Llama-3.1-Storm-8B
layer_range: [0, 32]
merge_method: nearswap
base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
parameters:
t:
- value: 0.0001
dtype: bfloat16
out_type: float16
slices:
- sources:
- model: v000000/Llama-3.1-8B-Stheno-v3.4-abliterated
layer_range: [0, 32]
- model: akjindal53244/Llama-3.1-Storm-8B
layer_range: [0, 32]
merge_method: slerp
base_model: v000000/Llama-3.1-8B-Stheno-v3.4-abliterated
parameters:
t:
- filter: self_attn
value: [0.1, 0.6, 0.3, 0.8, 0.5]
- filter: mlp
value: [0.9, 0.4, 0.7, 0.2, 0.5]
- value: 0.5
dtype: float32
models:
- model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
weight: 1.0
- model: v000000/L3.1-Niitorm-8B-t0.0001
parameters:
weight: 0.4
merge_method: task_arithmetic
base_model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
normalize: false
dtype: float16
models:
- model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
weight: 0.0
- model: v000000/L3.1-Niitorm-8B-t0.0001
parameters:
weight: 1.25
merge_method: task_arithmetic
base_model: arcee-ai/Llama-3.1-SuperNova-Lite
parameters:
normalize: false
dtype: float16
models:
- model: v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
merge_method: slerp
base_model: v000000/L3.1-8B-RP-Test-002-Task_Arithmetic+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
parameters:
t:
- value: [0, 0, 0.3, 0.4, 0.5, 0.6, 0.5, 0.4, 0.3, 0, 0]
dtype: float16
base_model: nothingiisreal/L3.1-8B-Celeste-V1.5+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
dtype: bfloat16
merge_method: task_arithmetic
parameters:
normalize: false
slices:
- sources:
- layer_range: [0, 32]
model: nothingiisreal/L3.1-8B-Celeste-V1.5+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
parameters:
weight: 0.7
- layer_range: [0, 32]
model: v000000/L3.1-Sthenorm-8B
parameters:
weight: 0.2
- layer_range: [0, 32]
model: nothingiisreal/L3.1-8B-Celeste-V1.5
parameters:
weight: 0.2
base_model: crestf411/L3.1-8B-sunfall-stheno-v0.6.1
experts_per_token: 2
local_experts: 2
gate_mode: random
dtype: bfloat16
experts:
- source_model: v000000/L3.1-Storniitova-8B
- source_model: x0000001/l3.1-part_aaa
``` |