DavidAU commited on
Commit
649c245
1 Parent(s): 4218fce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -84,6 +84,7 @@ One version is not stronger than the other, they are different and result in dif
84
 
85
  This chart shows the order in terms of "BPW" for each quant with "IQ1_S" with the least, and "Q8_0" with the most:
86
 
 
87
  IQ1_S | IQ1_M
88
 
89
  IQ2_XXS | IQ2_XS | Q2_K_S | IQ2_S | Q2_K | IQ2_M
@@ -97,9 +98,11 @@ Q5_K_S | Q5_K_M
97
  Q6_K
98
 
99
  Q8_0
 
100
 
101
- More BPW mean better quality, but higher VRAM requirements (and larger file size) and lower Tokens per second.
102
  The larger the model in terms of parameters the lower the size of quant you can run with less quality losses.
 
103
 
104
  (not all quants may be at this repo)
105
 
 
84
 
85
  This chart shows the order in terms of "BPW" for each quant with "IQ1_S" with the least, and "Q8_0" with the most:
86
 
87
+ <PRE>
88
  IQ1_S | IQ1_M
89
 
90
  IQ2_XXS | IQ2_XS | Q2_K_S | IQ2_S | Q2_K | IQ2_M
 
98
  Q6_K
99
 
100
  Q8_0
101
+ </pre>
102
 
103
+ More BPW mean better quality, but higher VRAM requirements (and larger file size) and lower tokens per second.
104
  The larger the model in terms of parameters the lower the size of quant you can run with less quality losses.
105
+ Note that "quality losses" refers to both instruction following and output quality.
106
 
107
  (not all quants may be at this repo)
108