What 2bit quantization approach are you using?
1
#13 opened 8 months ago
by
ingridstevens
Not able to use with ctransformers
1
#12 opened 9 months ago
by
aravindsr
NousCapybara 34b
#11 opened 9 months ago
by
Alastar-Smith
code-llama-70b
#10 opened 9 months ago
by
eramax
What -ctx and -chunks parameters did you use to make the iMatrix of the Lllama 2 70b?
1
#9 opened 10 months ago
by
Nexesenex
Quantize these amazing models
#8 opened 10 months ago
by
Tanvir1337
mixtral-instruct-8x7b for Q2KS as well
#7 opened 10 months ago
by
shing3232
Would love a deepseekcode 2bit quant. I bet others would love it too :)
2
#6 opened 10 months ago
by
subiculumforge
[Model request] Saily 100b, Saily 220b
1
#5 opened 10 months ago
by
Perpetuity7
Could We combine AWQ and Importance Matrix calculation together to further improve perplexity.
3
#4 opened 10 months ago
by
shing3232
[Model Request] cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser
#3 opened 10 months ago
by
Joseph717171
Magic Issues with nous-hermes-2-34b-2.16bpw.gguf (Log Attached...)
#2 opened 10 months ago
by
Joseph717171
Please 3b model rocket 2bit?
17
#1 opened 10 months ago
by
Shqmil