Update README.md
Browse files
README.md
CHANGED
@@ -23,6 +23,12 @@ quantized_by: dfurman
|
|
23 |
|
24 |
This model is a finetune of `MaziyarPanahi/calme-2.4-rys-78b` on 1.5k rows of the `mlabonne/orpo-dpo-mix-40k` dataset.
|
25 |
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/62afc20ca5bd7cef3e1ab3f4/NG5WGL0ljzLsNhSBRVqnD.png)
|
27 |
|
28 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/62afc20ca5bd7cef3e1ab3f4/Zhk5Bpr1I2NrzX98Bhtp8.png)
|
@@ -57,7 +63,7 @@ else:
|
|
57 |
attn_implementation = "eager"
|
58 |
torch_dtype = torch.float16
|
59 |
|
60 |
-
# quantize if necessary
|
61 |
# bnb_config = BitsAndBytesConfig(
|
62 |
# load_in_4bit=True,
|
63 |
# bnb_4bit_quant_type="nf4",
|
@@ -84,6 +90,30 @@ pipeline = transformers.pipeline(
|
|
84 |
|
85 |
### Example 1
|
86 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
87 |
```python
|
88 |
question = """The bakers at the Beverly Hills Bakery baked 200 loaves of bread on Monday morning.
|
89 |
They sold 93 loaves in the morning and 39 loaves in the afternoon.
|
@@ -114,7 +144,7 @@ print(outputs[0]["generated_text"][len(prompt):])
|
|
114 |
|3|Adjust for returns|Add returned loaves to remaining|74|
|
115 |
```
|
116 |
|
117 |
-
### Example
|
118 |
|
119 |
```python
|
120 |
question = "What's a good recipe for a spicy margarita?"
|
|
|
23 |
|
24 |
This model is a finetune of `MaziyarPanahi/calme-2.4-rys-78b` on 1.5k rows of the `mlabonne/orpo-dpo-mix-40k` dataset.
|
25 |
|
26 |
+
It was trained as a generalist language model supporting a variety of text generation use cases, including agentic capabilities, roleplaying, reasoning, multi-turn conversation, long context coherence, and more.
|
27 |
+
|
28 |
+
## 🚅 Training
|
29 |
+
|
30 |
+
Here are a few visualizations of the finetune run:
|
31 |
+
|
32 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/62afc20ca5bd7cef3e1ab3f4/NG5WGL0ljzLsNhSBRVqnD.png)
|
33 |
|
34 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/62afc20ca5bd7cef3e1ab3f4/Zhk5Bpr1I2NrzX98Bhtp8.png)
|
|
|
63 |
attn_implementation = "eager"
|
64 |
torch_dtype = torch.float16
|
65 |
|
66 |
+
# # quantize if necessary
|
67 |
# bnb_config = BitsAndBytesConfig(
|
68 |
# load_in_4bit=True,
|
69 |
# bnb_4bit_quant_type="nf4",
|
|
|
90 |
|
91 |
### Example 1
|
92 |
|
93 |
+
```python
|
94 |
+
question = "Is the number 9.11 larger than 9.9?"
|
95 |
+
|
96 |
+
messages = [
|
97 |
+
{"role": "system", "content": "You are a helpful assistant that thinks step by step."},
|
98 |
+
{"role": "user", "content": question},
|
99 |
+
]
|
100 |
+
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
101 |
+
# print("***Prompt:\n", prompt)
|
102 |
+
|
103 |
+
outputs = pipeline(
|
104 |
+
prompt, max_new_tokens=1000, do_sample=True, temperature=0.01, top_k=50, top_p=0.95
|
105 |
+
)
|
106 |
+
print("***Generation:")
|
107 |
+
print(outputs[0]["generated_text"][len(prompt) :])
|
108 |
+
```
|
109 |
+
|
110 |
+
```
|
111 |
+
***Generation:
|
112 |
+
To compare these two numbers, it's important to look at their decimal places after the whole number part, which is 9 in both cases. Comparing the tenths place, 9.11 has a '1' and 9.9 has a '9'. Since '9' is greater than '1', 9.9 is larger than 9.11.
|
113 |
+
```
|
114 |
+
|
115 |
+
### Example 2
|
116 |
+
|
117 |
```python
|
118 |
question = """The bakers at the Beverly Hills Bakery baked 200 loaves of bread on Monday morning.
|
119 |
They sold 93 loaves in the morning and 39 loaves in the afternoon.
|
|
|
144 |
|3|Adjust for returns|Add returned loaves to remaining|74|
|
145 |
```
|
146 |
|
147 |
+
### Example 3
|
148 |
|
149 |
```python
|
150 |
question = "What's a good recipe for a spicy margarita?"
|