AlexWortega
commited on
Commit
•
81d5541
1
Parent(s):
5db0bb9
Update README.md
Browse files
README.md
CHANGED
@@ -11,8 +11,8 @@ language:
|
|
11 |
|
12 |
Added a lot more data to sft, now json and multiturn work more stable on long context and hard prompts
|
13 |
|
14 |
-
- [Google Colab](https://colab.research.google.com/drive/
|
15 |
-
- [GGUF](https://huggingface.co/Vikhrmodels/
|
16 |
|
17 |
```python
|
18 |
|
@@ -24,7 +24,7 @@ model = AutoModelForCausalLM.from_pretrained("Vikhrmodels/it-5.2-fp16-cp",
|
|
24 |
attn_implementation="flash_attention_2",
|
25 |
torch_dtype=torch.bfloat16)
|
26 |
|
27 |
-
tokenizer = AutoTokenizer.from_pretrained("Vikhrmodels/
|
28 |
from transformers import AutoTokenizer, pipeline
|
29 |
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
30 |
prompts = [
|
|
|
11 |
|
12 |
Added a lot more data to sft, now json and multiturn work more stable on long context and hard prompts
|
13 |
|
14 |
+
- [Google Colab](https://colab.research.google.com/drive/1-_BWsJycBm3rEyjpBx2_ejshpemQYHbe?usp=sharing)
|
15 |
+
- [GGUF](https://huggingface.co/Vikhrmodels/it-5.2-fp16-cp-GGUF)
|
16 |
|
17 |
```python
|
18 |
|
|
|
24 |
attn_implementation="flash_attention_2",
|
25 |
torch_dtype=torch.bfloat16)
|
26 |
|
27 |
+
tokenizer = AutoTokenizer.from_pretrained("Vikhrmodels/Vikhr-7B-instruct_0.4")
|
28 |
from transformers import AutoTokenizer, pipeline
|
29 |
pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
30 |
prompts = [
|