Spaces:
Running
Running
OpenSourceRonin
commited on
Commit
•
6d2cd96
1
Parent(s):
7a0c1ba
Update index.html
Browse files- index.html +3 -1
index.html
CHANGED
@@ -27,7 +27,9 @@
|
|
27 |
<p>
|
28 |
<b>VPTQ (Vector Post-Training Quantization)</b> is an advanced compression technique that dramatically reduces the size of large language models such as the 70B and 405B Llama models. VPTQ efficiently compresses these models to 1-2 bits within just a few hours, enabling them to run effectively on GPUs with limited memory.
|
29 |
For more information, visit the following links:
|
30 |
-
<p
|
|
|
|
|
31 |
<ul>
|
32 |
<li>
|
33 |
<a href="https://arxiv.org/abs/2409.17066" target="_blank" class="link-styled">
|
|
|
27 |
<p>
|
28 |
<b>VPTQ (Vector Post-Training Quantization)</b> is an advanced compression technique that dramatically reduces the size of large language models such as the 70B and 405B Llama models. VPTQ efficiently compresses these models to 1-2 bits within just a few hours, enabling them to run effectively on GPUs with limited memory.
|
29 |
For more information, visit the following links:
|
30 |
+
<p style="font-weight: bold; font-size: larger;">
|
31 |
+
The current demo runs on a free, shared A100 provided by HUGGINGFACE, which may lead to long load times for model loading and acquiring an available GPU. This demo is intended to showcase the quality of the quantized model, not inference speed.
|
32 |
+
</p>
|
33 |
<ul>
|
34 |
<li>
|
35 |
<a href="https://arxiv.org/abs/2409.17066" target="_blank" class="link-styled">
|