File size: 1,492 Bytes
013cc95
2a760f6
 
 
013cc95
 
 
 
 
 
 
 
4042167
 
ef26f36
 
 
 
 
 
 
 
 
 
013cc95
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
---
license: other
license_name: yi-license
license_link: LICENSE
datasets:
- totally-not-an-llm/EverythingLM-data-V3
pipeline_tag: text-generation
---
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63732ebbbd81fae2b3aaf3fb/4OQkvAa1zOK4Devv-aUdL.png)

In this case the tokenizer is the yi_tokenizer, loading it requires trust_remote_code=True

Have some fun with this fellow. 

It can eat a lot of vram, depends on settings to make it useable on two 24 gb vram gpus:

Without fused attention, it's 27 gb vram, will need some if if yo do stuff. 

![image.png](https://cdn-uploads.huggingface.co/production/uploads/63732ebbbd81fae2b3aaf3fb/1cbqKp55WhN4BQD337E-n.png)

You can also let if have fused attention and just reduce the max_seq_length to something way smaller yet still useful

![image.png](https://cdn-uploads.huggingface.co/production/uploads/63732ebbbd81fae2b3aaf3fb/JRi4sakPziGpmOFCBfcJS.png)



License
The Yi series models are fully open for academic research and free commercial usage with permission via applications. All usage must adhere to the Model License Agreement 2.0. To apply for the official commercial license, please contact us ([email protected]).



Prompt Example:
```
### System:

You are an AI assistant. User will you give you a task. Your goal is to complete the task as faithfully as you can. While performing the task think step-by-step and justify your steps.

### Instruction: 

How do you fine tune a large language model? 

### Response:
```