Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,9 @@ widget:
|
|
9 |
|
10 |
## Model description
|
11 |
|
12 |
-
This is the set of 5 Chinese RoBERTa-Base classification models fine-tuned by [UER-py](https://arxiv.org/abs/1909.05658).
|
|
|
|
|
13 |
|
14 |
| Dataset | Link |
|
15 |
| :-----------: | :-------------------------------------------------------: |
|
@@ -34,7 +36,7 @@ You can use this model directly with a pipeline for text classification (take th
|
|
34 |
|
35 |
## Training data
|
36 |
|
37 |
-
5 Chinese text classification datasets are used. JD full, JD binary, and Dianping datasets consist of user reviews of different sentiment polarities. Ifeng and Chinanews consist of first paragraphs of news articles of different topic classes. They are collected by [Glyph](https://github.com/zhangxiangxiao/glyph) project and more details are discussed in corresponding [paper](https://arxiv.org/abs/1708.02657).
|
38 |
|
39 |
## Training procedure
|
40 |
|
@@ -62,13 +64,6 @@ python3 scripts/convert_bert_text_classification_from_uer_to_huggingface.py --in
|
|
62 |
### BibTeX entry and citation info
|
63 |
|
64 |
```
|
65 |
-
@article{devlin2018bert,
|
66 |
-
title={BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding},
|
67 |
-
author={Devlin, Jacob and Chang, Ming-Wei and Lee, Kenton and Toutanova, Kristina},
|
68 |
-
journal={arXiv preprint arXiv:1810.04805},
|
69 |
-
year={2018}
|
70 |
-
}
|
71 |
-
|
72 |
@article{liu2019roberta,
|
73 |
title={Roberta: A robustly optimized bert pretraining approach},
|
74 |
author={Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
|
@@ -90,6 +85,13 @@ python3 scripts/convert_bert_text_classification_from_uer_to_huggingface.py --in
|
|
90 |
pages={241},
|
91 |
year={2019}
|
92 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
93 |
```
|
94 |
|
95 |
[jd_full]:https://huggingface.co/uer/roberta-base-finetuned-jd-full-chinese
|
|
|
9 |
|
10 |
## Model description
|
11 |
|
12 |
+
This is the set of 5 Chinese RoBERTa-Base classification models fine-tuned by [UER-py](https://github.com/dbiir/UER-py/), which is introduced in [this paper](https://arxiv.org/abs/1909.05658). Besides, the models could also be fine-tuned by [TencentPretrain](https://github.com/Tencent/TencentPretrain) introduced in [this paper](https://arxiv.org/abs/2212.06385), which inherits UER-py to support models with parameters above one billion, and extends it to a multimodal pre-training framework.
|
13 |
+
|
14 |
+
You can download the 5 Chinese RoBERTa-Base classification models either from the [UER-py Modelzoo page](https://github.com/dbiir/UER-py/wiki/Modelzoo), or via HuggingFace from the links below:
|
15 |
|
16 |
| Dataset | Link |
|
17 |
| :-----------: | :-------------------------------------------------------: |
|
|
|
36 |
|
37 |
## Training data
|
38 |
|
39 |
+
5 Chinese text classification datasets are used. JD full, JD binary, and Dianping datasets consist of user reviews of different sentiment polarities. Ifeng and Chinanews consist of first paragraphs of news articles of different topic classes. They are collected by [Glyph](https://github.com/zhangxiangxiao/glyph) project and more details are discussed in the corresponding [paper](https://arxiv.org/abs/1708.02657).
|
40 |
|
41 |
## Training procedure
|
42 |
|
|
|
64 |
### BibTeX entry and citation info
|
65 |
|
66 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
67 |
@article{liu2019roberta,
|
68 |
title={Roberta: A robustly optimized bert pretraining approach},
|
69 |
author={Liu, Yinhan and Ott, Myle and Goyal, Naman and Du, Jingfei and Joshi, Mandar and Chen, Danqi and Levy, Omer and Lewis, Mike and Zettlemoyer, Luke and Stoyanov, Veselin},
|
|
|
85 |
pages={241},
|
86 |
year={2019}
|
87 |
}
|
88 |
+
|
89 |
+
@article{zhao2023tencentpretrain,
|
90 |
+
title={TencentPretrain: A Scalable and Flexible Toolkit for Pre-training Models of Different Modalities},
|
91 |
+
author={Zhao, Zhe and Li, Yudong and Hou, Cheng and Zhao, Jing and others},
|
92 |
+
journal={ACL 2023},
|
93 |
+
pages={217},
|
94 |
+
year={2023}
|
95 |
```
|
96 |
|
97 |
[jd_full]:https://huggingface.co/uer/roberta-base-finetuned-jd-full-chinese
|