|
--- |
|
library_name: transformers |
|
datasets: |
|
- jondurbin/truthy-dpo-v0.1 |
|
--- |
|
|
|
# MBX-7B-v3-DPO |
|
|
|
This model is a finetune of [flemmingmiguel/MBX-7B-v3](https://huggingface.co/flemmingmiguel/MBX-7B-v3) using jondurbin/truthy-dpo-v0.1 |
|
|
|
![MBX-v3-orca](MBX-v3-orca.png) |
|
|
|
## Code Example |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
import torch |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("macadeliccc/MBX-7B-v3-DPO") |
|
model = AutoModelForCausalLM.from_pretrained("macadeliccc/MBX-7B-v3-DPO") |
|
|
|
messages = [ |
|
{"role": "system", "content": "You are a helpful assistant."}, |
|
{"role": "user", "content": "Can you write me a creative haiku?"} |
|
] |
|
gen_input = tokenizer.apply_chat_template(messages, return_tensors="pt") |
|
|
|
response = model.generate(**gen_input) |
|
decoded_response = tokenizer.decode(response[0], skip_special_tokens=True) |
|
|
|
print(decoded_response) |
|
``` |
|
## GGUF |
|
|
|
TODO |
|
|
|
## Evaluations |
|
|
|
TODO |