OpenSourceRonin commited on
Commit
6d2cd96
1 Parent(s): 7a0c1ba

Update index.html

Browse files
Files changed (1) hide show
  1. index.html +3 -1
index.html CHANGED
@@ -27,7 +27,9 @@
27
  <p>
28
  <b>VPTQ (Vector Post-Training Quantization)</b> is an advanced compression technique that dramatically reduces the size of large language models such as the 70B and 405B Llama models. VPTQ efficiently compresses these models to 1-2 bits within just a few hours, enabling them to run effectively on GPUs with limited memory.
29
  For more information, visit the following links:
30
- <p>The current demo runs on a free, shared A100 provided by HUGGINGFACE, which may lead to long load times for model loading and acquiring an available GPU. This demo is intended to showcase the quality of the quantized model, not inference speed.</p>
 
 
31
  <ul>
32
  <li>
33
  <a href="https://arxiv.org/abs/2409.17066" target="_blank" class="link-styled">
 
27
  <p>
28
  <b>VPTQ (Vector Post-Training Quantization)</b> is an advanced compression technique that dramatically reduces the size of large language models such as the 70B and 405B Llama models. VPTQ efficiently compresses these models to 1-2 bits within just a few hours, enabling them to run effectively on GPUs with limited memory.
29
  For more information, visit the following links:
30
+ <p style="font-weight: bold; font-size: larger;">
31
+ The current demo runs on a free, shared A100 provided by HUGGINGFACE, which may lead to long load times for model loading and acquiring an available GPU. This demo is intended to showcase the quality of the quantized model, not inference speed.
32
+ </p>
33
  <ul>
34
  <li>
35
  <a href="https://arxiv.org/abs/2409.17066" target="_blank" class="link-styled">