--- library_name: transformers pipeline_tag: text-generation tags: - code - llama - llama-2 - text-generation-inference base_model: codellama/CodeLlama-7b-Instruct-hf inference: false --- # Mistral-7B-v0.1-GGUF - Quantized version of [CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) - Created using llama.cpp ## Available Quants * Q2_K * Q3_K_L * Q3_K_M * Q3_K_S * Q4_0 * Q4_K_M * Q4_K_S * Q5_0 * Q5_K_M * Q5_K_S * Q6_K * Q8_0