text-generation-webui / docs /What Works.md
zhengr's picture
update
33473a0
|
raw
history blame
1.61 kB

What Works

Loader Loading 1 LoRA Loading 2 or more LoRAs Training LoRAs Multimodal extension Perplexity evaluation
Transformers βœ… βœ…** βœ…* βœ… βœ…
llama.cpp ❌ ❌ ❌ ❌ use llamacpp_HF
llamacpp_HF ❌ ❌ ❌ ❌ βœ…
ExLlamav2_HF βœ… βœ… ❌ ❌ βœ…
ExLlamav2 βœ… βœ… ❌ ❌ use ExLlamav2_HF
AutoGPTQ βœ… ❌ ❌ βœ… βœ…
AutoAWQ ? ❌ ? ? βœ…
HQQ ? ? ? ? βœ…

❌ = not implemented

βœ… = implemented

* Training LoRAs with GPTQ models also works with the Transformers loader. Make sure to check "auto-devices" and "disable_exllama" before loading the model.

** Multi-LoRA in PEFT is tricky and the current implementation does not work reliably in all cases.