MBX-7B-v3-DPO / README.md
macadeliccc's picture
Update README.md
88575bb verified
|
raw
history blame
914 Bytes
---
library_name: transformers
datasets:
- jondurbin/truthy-dpo-v0.1
---
# MBX-7B-v3-DPO
This model is a finetune of [flemmingmiguel/MBX-7B-v3](https://huggingface.co/flemmingmiguel/MBX-7B-v3) using jondurbin/truthy-dpo-v0.1
![MBX-v3-orca](MBX-v3-orca.png)
## Code Example
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
tokenizer = AutoTokenizer.from_pretrained("macadeliccc/MBX-7B-v3-DPO")
model = AutoModelForCausalLM.from_pretrained("macadeliccc/MBX-7B-v3-DPO")
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Can you write me a creative haiku?"}
]
gen_input = tokenizer.apply_chat_template(messages, return_tensors="pt")
response = model.generate(**gen_input)
decoded_response = tokenizer.decode(response[0], skip_special_tokens=True)
print(decoded_response)
```
## GGUF
TODO
## Evaluations
TODO