suolyer commited on
Commit
c85423a
1 Parent(s): 0867a87

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - zh
4
+ license: apache-2.0
5
+
6
+ inference: false
7
+
8
+ ---
9
+ # Randeng-770M model (Chinese),one model of [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM).
10
+ The 770 million parameter Randeng-770M large model, using 280G Chinese data, 16 A100 training for 14 days,which is a standard transformer structure.
11
+
12
+
13
+ ## Usage
14
+ There is no structure of Randeng-770M in [Transformers](https://github.com/huggingface/transformers), you can run follow code to get structure of Randeng-770M from [Fengshenbang-LM](https://github.com/IDEA-CCNL/Fengshenbang-LM)
15
+
16
+ ```shell
17
+ git clone https://github.com/IDEA-CCNL/Fengshenbang-LM.git
18
+ ```
19
+
20
+ ## Usage
21
+ ```python
22
+ from model.megatron_t5.modeling_megatron_t5 import T5EncoderModel
23
+ from model.megatron_t5.configuration_magetron_t5 import T5Config
24
+ from model.megatron_t5.tokenization_megatron_t5 import T5Tokenizer
25
+
26
+ tokenizer = T5Tokenizer.from_pretrained('IDEA-CCNL/Randeng-770M')
27
+ config = T5Config.from_pretrained('IDEA-CCNL/Randeng-770M')
28
+ model = T5EncoderModel.from_pretrained('IDEA-CCNL/Randeng-770M')
29
+
30
+ ```
31
+
32
+ ## Citation
33
+ If you find the resource is useful, please cite the following website in your paper.
34
+ ```
35
+ @misc{Fengshenbang-LM,
36
+ title={Fengshenbang-LM},
37
+ author={IDEA-CCNL},
38
+ year={2021},
39
+ howpublished={\url{https://github.com/IDEA-CCNL/Fengshenbang-LM}},
40
+ }
41
+ ```