t5-v1_1-xxl / README.md
RyanJDick's picture
Add bnb_llm_int8 quantization method.
4376554
---
license: apache-2.0
---
# T5-XXL Encoder
This repo contains copies of the T5-XXL encoder in various quantization formats. The models in this repo are intended for use in [InvokeAI](https://github.com/invoke-ai/InvokeAI).
Contents:
- `bfloat16/` - T5-XXL encoder cast to bfloat16. Copied from [here](https://huggingface.co/black-forest-labs/FLUX.1-schnell/tree/741f7c3ce8b383c54771c7003378a50191e9efe9/text_encoder_2).
- `bnb_llm_int8/` - T5-XXL encoder quantized using bitsandbytes LLM.int8() quantization.
- `optimum_quanto_qfloat8/` - T5-XXL encoder quantized using [optimum-quanto](https://github.com/huggingface/optimum-quanto) qfloat8 quantization.