# QServe benchmarks This huggingface repository contains configurations and tokenizer files for all models benchmarked in our [QServe](https://github.com/mit-han-lab/qserve) project: - Llama-3-8B - Llama-2-7B - Llama-2-13B - Llama-2-70B - Llama-30B - Mistral-7B - Yi-34B - Qwen1.5-72B Please clone this repository if you wish to run our QServe benchmark code without cloning full models. Please consider citing our paper if it is helpful: ``` @article{lin2024qserve, title={QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving}, author={Lin*, Yujun and Tang*, Haotian and Yang*, Shang and Zhang, Zhekai and Xiao, Guangxuan and Gan, Chuang and Han, Song}, year={2024} } ```