|
--- |
|
license: cc-by-nc-4.0 |
|
tags: |
|
- not-for-all-audiences |
|
- mixtral |
|
- conversational |
|
- gguf |
|
- iMat |
|
--- |
|
<img src="https://i.imgur.com/P68dXux.png" width="400"/> |
|
|
|
|
|
# CybersurferNyandroidLexicat-8x7B-iMat-GGUF |
|
|
|
|
|
|
|
[CybersurferNyandroidLexicat](https://huggingface.co/Envoid/CybersurferNyandroidLexicat-8x7B) quantized from fp16 with love. |
|
|
|
Uses the same imat calculation method as the later batch of maid-yuzu-v8-alter-iMat-GGUF. |
|
|
|
<b>Legacy quants (i.e. Q5_K_M, Q6_K, etc) in this repo have all been enhanced with importance matrix calculation. These quants show improved KL-Divergence over their static counterparts.</b> |
|
|
|
All files included here for your convenience. No need to clone the entire repo, just pick the quant that's right for you. |
|
|
|
For more information on latest iMatrix quants see this PR - https://github.com/ggerganov/llama.cpp/pull/5747 |
|
|
|
<b>Tip:</b> The letter at the end of the quant name indicates its size. Larger sizes have better quality, smaller sizes are faster. |
|
* IQ3_XS - XS (Extra Small) |
|
* IQ3_S - S (Small) |
|
* IQ3_M - M (medium) |
|
|
|
|