|
--- |
|
language: |
|
- en |
|
tags: |
|
- openvino |
|
--- |
|
|
|
# baichuan-inc/Baichuan2-13B-Chat |
|
|
|
This is the [baichuan-inc/Baichuan2-13B-Chat](https://huggingface.co/baichuan-inc/Baichuan2-13B-Chat) model converted to [OpenVINO](https://openvino.ai) with INT8 weights compression for accelerated inference. |
|
|
|
An example of how to do inference on this model: |
|
```python |
|
from optimum.intel import OVModelForCausalLM |
|
from transformers import AutoTokenizer, pipeline |
|
|
|
# model_id should be set to either a local directory or a model available on the HuggingFace hub. |
|
model_id = "helenai/baichuan-inc-Baichuan2-13B-Chat-ov" |
|
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True) |
|
model = OVModelForCausalLM.from_pretrained(model_id) |
|
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) |
|
result = pipe("hello world") |
|
print(result) |
|
``` |
|
|
|
|