metadata
license: apache-2.0
inference: false
bling-tiny-llama-ov
bling-tiny-llama-ov is an OpenVino int4 quantized version of BLING Tiny-Llama 1B, providing a very fast, very small inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
bling-tiny-llama is a fact-based question-answering model, optimized for complex business documents.
Get started right away with OpenVino
Looking for AI PC solutions and demos, contact us at llmware
Model Description
- Developed by: llmware
- Model type: tinyllama
- Parameters: 1.1 billion
- Model Parent: llmware/bling-tiny-llama-v0
- Language(s) (NLP): English
- License: Apache 2.0
- Uses: Fact-based question-answering
- RAG Benchmark Accuracy Score: 86.5
- Quantization: int4