wenge-research
commited on
Commit
•
321524b
1
Parent(s):
dc16eed
Update README.md
Browse files
README.md
CHANGED
@@ -27,13 +27,48 @@ YAYI 2 is a collection of open-source large language models launched by Wenge Te
|
|
27 |
For more details about the YAYI 2, please refer to our GitHub repository. Stay tuned for more technical details in our upcoming technical report! 🔥
|
28 |
|
29 |
|
30 |
-
##
|
31 |
|
32 |
| Model Name | Context Length | 🤗 HF Model Name |
|
33 |
|:----------|:----------:|:----------:|
|
34 |
| YAYI2-30B | 4096 | wenge-research/yayi2-30b|
|
35 |
|
36 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
## 评测结果/Evaluation
|
38 |
|
39 |
我们在多个基准数据集上进行了评测,包括 C-Eval、MMLU、 CMMLU、AGIEval、GAOKAO-Bench、GSM8K、MATH、BBH、HumanEval 以及 MBPP。我们考察了模型在语言理解、学科知识、数学推理、逻辑推理以及代码生成方面的表现。YAYI 2 模型在与其规模相近的开源模型中展现出了显著的性能提升。
|
@@ -200,27 +235,6 @@ We evaluate our model on standard benchmarks, including C-Eval, MMLU, CMMLU, AGI
|
|
200 |
We evaluate our model using the source code from the [OpenCompass Github repository](https://github.com/open-compass/opencompass). If available, we report results for comparative models assessed by OpenCompass with the evaluation reference date set to Dec. 15th, 2013. For MPT, Falfon, and Llama, which have not been evaluated by OpenCompass, we use the results reported in the [LLaMA 2](https://arxiv.org/abs/2307.09288) paper.
|
201 |
|
202 |
|
203 |
-
## 快速开始/Quick Start
|
204 |
-
|
205 |
-
```python
|
206 |
-
>>> from transformers import AutoModelForCausalLM, AutoTokenizer
|
207 |
-
>>> tokenizer = AutoTokenizer.from_pretrained("wenge-research/yayi2-30b", trust_remote_code=True)
|
208 |
-
>>> model = AutoModelForCausalLM.from_pretrained("wenge-research/yayi2-30b", device_map="auto", trust_remote_code=True)
|
209 |
-
>>> inputs = tokenizer('The winter in Beijing is', return_tensors='pt')
|
210 |
-
>>> inputs = inputs.to('cuda')
|
211 |
-
>>> pred = model.generate(
|
212 |
-
**inputs,
|
213 |
-
max_new_tokens=256,
|
214 |
-
eos_token_id=tokenizer.eos_token_id,
|
215 |
-
do_sample=True,
|
216 |
-
repetition_penalty=1.2,
|
217 |
-
temperature=0.4,
|
218 |
-
top_k=100,
|
219 |
-
top_p=0.8
|
220 |
-
)
|
221 |
-
>>> print(tokenizer.decode(pred.cpu()[0], skip_special_tokens=True))
|
222 |
-
```
|
223 |
-
|
224 |
|
225 |
## 协议/Liencese
|
226 |
|
|
|
27 |
For more details about the YAYI 2, please refer to our GitHub repository. Stay tuned for more technical details in our upcoming technical report! 🔥
|
28 |
|
29 |
|
30 |
+
## 模型细节/Model
|
31 |
|
32 |
| Model Name | Context Length | 🤗 HF Model Name |
|
33 |
|:----------|:----------:|:----------:|
|
34 |
| YAYI2-30B | 4096 | wenge-research/yayi2-30b|
|
35 |
|
36 |
|
37 |
+
|
38 |
+
|
39 |
+
## 要求 Requirements)
|
40 |
+
|
41 |
+
python 3.8及以上版本
|
42 |
+
pytorch 1.12及以上版本,推荐2.0及以上版本
|
43 |
+
建议使用CUDA 11.4及以上(GPU用户、flash-attention用户等需考虑此选项)
|
44 |
+
运行BF16或FP16模型需要多卡至少144GB显存(例如2xA100-80G或5xV100-32G);运行Int4模型至少需要48GB显存(例如1xA100-80G或2xV100-32G)。
|
45 |
+
python 3.8 and above
|
46 |
+
pytorch 1.12 and above, 2.0 and above are recommended
|
47 |
+
CUDA 11.4 and above are recommended (this is for GPU users, flash-attention users, etc.) To run Qwen-72B-Chat in bf16/fp16, at least 144GB GPU memory is required (e.g., 2xA100-80G or 5xV100-32G). To run it in int4, at least 48GB GPU memory is requred (e.g., 1xA100-80G or 2xV100-32G).
|
48 |
+
|
49 |
+
|
50 |
+
## 快速开始/Quick Start
|
51 |
+
|
52 |
+
```python
|
53 |
+
>>> from transformers import AutoModelForCausalLM, AutoTokenizer
|
54 |
+
>>> tokenizer = AutoTokenizer.from_pretrained("wenge-research/yayi2-30b", trust_remote_code=True)
|
55 |
+
>>> model = AutoModelForCausalLM.from_pretrained("wenge-research/yayi2-30b", device_map="auto", trust_remote_code=True)
|
56 |
+
>>> inputs = tokenizer('The winter in Beijing is', return_tensors='pt')
|
57 |
+
>>> inputs = inputs.to('cuda')
|
58 |
+
>>> pred = model.generate(
|
59 |
+
**inputs,
|
60 |
+
max_new_tokens=256,
|
61 |
+
eos_token_id=tokenizer.eos_token_id,
|
62 |
+
do_sample=True,
|
63 |
+
repetition_penalty=1.2,
|
64 |
+
temperature=0.4,
|
65 |
+
top_k=100,
|
66 |
+
top_p=0.8
|
67 |
+
)
|
68 |
+
>>> print(tokenizer.decode(pred.cpu()[0], skip_special_tokens=True))
|
69 |
+
```
|
70 |
+
|
71 |
+
|
72 |
## 评测结果/Evaluation
|
73 |
|
74 |
我们在多个基准数据集上进行了评测,包括 C-Eval、MMLU、 CMMLU、AGIEval、GAOKAO-Bench、GSM8K、MATH、BBH、HumanEval 以及 MBPP。我们考察了模型在语言理解、学科知识、数学推理、逻辑推理以及代码生成方面的表现。YAYI 2 模型在与其规模相近的开源模型中展现出了显著的性能提升。
|
|
|
235 |
We evaluate our model using the source code from the [OpenCompass Github repository](https://github.com/open-compass/opencompass). If available, we report results for comparative models assessed by OpenCompass with the evaluation reference date set to Dec. 15th, 2013. For MPT, Falfon, and Llama, which have not been evaluated by OpenCompass, we use the results reported in the [LLaMA 2](https://arxiv.org/abs/2307.09288) paper.
|
236 |
|
237 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
238 |
|
239 |
## 协议/Liencese
|
240 |
|