|
--- |
|
language: |
|
- en |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
--- |
|
# GenZ 13B v2 GGML |
|
|
|
The instruction finetuned model with 4K input length. The model is finetuned on top of pretrained LLaMa2 |
|
|
|
|
|
## Inference |
|
|
|
```python |
|
import ctransformers |
|
from ctransformers import AutoModelForCausalLM |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained('budecosystem/genz-13b-v2-ggml', model_type="llama") |
|
|
|
prompt = """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. |
|
USER: who are you? ASSISTANT: """ |
|
print(model(prompt)) |
|
``` |
|
Support [LM Studio](https://lmstudio.ai/) for Mac & Windows users |
|
|
|
Use following prompt template |
|
|
|
``` |
|
A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. USER: Hi, how are you? ASSISTANT: |
|
``` |
|
|
|
Check the GitHub for the code -> [GenZ](https://github.com/BudEcosystem/GenZ) |