eachadea's picture
Update README.md
6e50f34
|
raw
history blame
455 Bytes
metadata
pipeline_tag: conversational
tags:
  - vicuna
  - llama
  - text-generation-inference

Converted for use with llama.cpp

  • 4-bit quantized
  • Needs ~10GB of CPU RAM
  • Won't work with alpaca.cpp or old llama.cpp (new ggml format requires latest llama.cpp)

Smaller 7B version can be found here: https://huggingface.co/eachadea/ggml-vicuna-7b-4bit


tags: - vicuna - llama - text-generation-inference