gmchuggingface commited on
Commit
677d4fe
1 Parent(s): 541ddd4

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - BelleGroup/train_0.5M_CN
5
+ language:
6
+ - en
7
+ - zh
8
+ tags:
9
+ - text-generation-inference
10
+ widget:
11
+ - text: |-
12
+ <|im_start|>user
13
+ 请以『春天的北京』为题写一首诗歌
14
+ <|im_end|>
15
+ <|im_start|>assistant
16
+ example_title: generation zh
17
+ ---
18
+ # Baichuan 7B ChatML
19
+
20
+ ## 介绍 Introduction
21
+
22
+ `baichuan-7B-chatml` 是支持多轮对话兼容于 ChatML 的模型。
23
+ 模型基于 [baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B) 微调而成。
24
+ `baichuan-7B-chatml` 模型支持商用。但按照baichuan-7B的要求,如果将baichuan-7B衍生品用作商业用途,需要联系[baichuan-7B 的许可方](https://github.com/baichuan-inc/baichuan-7B#%E5%8D%8F%E8%AE%AE)。
25
+
26
+ __需要注意:在面对事实性知识任务时,模型可能会生成不正确的信息或者产生不稳定的输出(有时可以返回正确答案,有时不能)。__
27
+
28
+ `baichuan-7B-chatml` is a model that supports multi-turn dialog and is compatible with ChatML.
29
+ The model is fine-tuned based on [baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B).
30
+ `baichuan-7B-chatml` model supports commercial use. However, according to the requirements of baichuan-7B, if baichuan-7B derivatives are used for commercial purposes, you need to contact [baichuan-7B](https://github.com/baichuan-inc/baichuan-7B#%E5%8D%8F%E8%AE%AE)。
31
+
32
+ __Note: When dealing with factual knowledge tasks, it may generate incorrect information or unstable output (sometimes it can return the correct answer, sometimes not).__
33
+
34
+ ## 代码示例 Examples
35
+
36
+ 模型在百川的基础上提供了对轮对话的函数供调用。
37
+
38
+ The model provides a function for multi-turn dialogs.
39
+
40
+ ```ipython
41
+ >>> from transformers import AutoTokenizer, AutoModelForCausalLM
42
+ >>> tokenizer = AutoTokenizer.from_pretrained("tibok/baichuan-7B-chatml", trust_remote_code=True)
43
+ >>> model = AutoModelForCausalLM.from_pretrained("tibok/baichuan-7B-chatml", device_map="auto", trust_remote_code=True)
44
+ >>> response, history = model.chat(tokenizer, "请以『春天的北京』为题写一首诗歌", history=[])
45
+ 春天的北京,
46
+ 花开万丈,
47
+ 春意盎然,
48
+ 清风送暖。
49
+ <|im_end|>
50
+ >>> response, history = model.chat(tokenizer, "能不能再写一首关于香山的?", history=history)
51
+ >>> print(response)
52
+ 香山之巅,
53
+ 芳草连天。
54
+ 清泉潺潺,
55
+ 山峦绵绵。
56
+ <|im_end|>
57
+ ```
58
+
59
+ ## 更多细节 Details
60
+
61
+ - Dataset: [BelleGroup/train_0.5M_CN](https://huggingface.co/datasets/BelleGroup/train_0.5M_CN)
62
+ - steps: 13800
63
+ - batch_size: 8
64
+ - seq_len: 2048