File size: 1,053 Bytes
d1dcaad
 
11c74be
 
 
 
 
77e2bca
d1dcaad
233d575
 
 
7f9a963
11c74be
233d575
 
11c74be
 
d877497
 
ced6501
7f9a963
 
01d79eb
11c74be
 
206dd29
 
 
 
11c74be
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
license: cc-by-nc-4.0
tags:
- not-for-all-audiences
- mixtral
- conversational
- gguf
- iMat
---
<img src="https://i.imgur.com/P68dXux.png" width="400"/>


# CybersurferNyandroidLexicat-8x7B-iMat-GGUF



[CybersurferNyandroidLexicat](https://huggingface.co/Envoid/CybersurferNyandroidLexicat-8x7B) quantized from fp16 with love.

Uses the same imat calculation method as the later batch of maid-yuzu-v8-alter-iMat-GGUF. 

<b>Legacy quants (i.e. Q5_K_M, Q6_K, etc) in this repo have all been enhanced with importance matrix calculation. These quants show improved KL-Divergence over their static counterparts.</b>

All files included here for your convenience. No need to clone the entire repo, just pick the quant that's right for you.

For more information on latest iMatrix quants see this PR - https://github.com/ggerganov/llama.cpp/pull/5747

<b>Tip:</b> The letter at the end of the quant name indicates its size. Larger sizes have better quality, smaller sizes are faster.
* IQ3_XS - XS (Extra Small)
* IQ3_S - S (Small)
* IQ3_M - M (medium)