--- pipeline_tag: conversational tags: - vicuna - llama - text-generation-inference --- Converted for use with [llama.cpp](https://github.com/ggerganov/llama.cpp) --- - 4-bit quantized - Needs ~10GB of CPU RAM - Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp) --- Smaller 7B version can be found here: https://huggingface.co/eachadea/ggml-vicuna-7b-4bit --- tags: - vicuna - llama - text-generation-inference ---