OneBit: Towards Extremely Low-bit Large Language Models
Abstract
Model quantification uses low bit-width values to represent the weight matrices of models, which is a promising approach to reduce both storage and computational overheads of deploying highly anticipated LLMs. However, existing quantization methods suffer severe performance degradation when the bit-width is extremely reduced, and thus focus on utilizing 4-bit or 8-bit values to quantize models. This paper boldly quantizes the weight matrices of LLMs to 1-bit, paving the way for the extremely low bit-width deployment of LLMs. For this target, we introduce a 1-bit quantization-aware training (QAT) framework named OneBit, including a novel 1-bit parameter representation method to better quantize LLMs as well as an effective parameter initialization method based on matrix decomposition to improve the convergence speed of the QAT framework. Sufficient experimental results indicate that OneBit achieves good performance (at least 83% of the non-quantized performance) with robust training processes when only using 1-bit weight matrices.
Community
Looks like their results are better than BiLLM. They only go up to 13B in the paper, perhaps 70B would be more favorable.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- BitDistiller: Unleashing the Potential of Sub-4-Bit LLMs via Self-Distillation (2024)
- EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge (2024)
- LQER: Low-Rank Quantization Error Reconstruction for LLMs (2024)
- ApiQ: Finetuning of 2-Bit Quantized Large Language Model (2024)
- L4Q: Parameter Efficient Quantization-Aware Training on Large Language Models via LoRA-wise LSQ (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Code and checkpoints are available at https://github.com/xuyuzhuang11/OneBit
Welcome to join us to explore the 1-bit LLM world! π€π€π€
Revolutionizing Large Language Models: OneBit's 1-Bit Quantization Breakthrough
Links π:
π Subscribe: https://www.youtube.com/@Arxflix
π Twitter: https://x.com/arxflix
π LMNT (Partner): https://lmnt.com/
π€π€π€Thanks for your attention!
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper