ai-forever vmkhlv commited on
Commit
08569b5
1 Parent(s): 7d0ab19

Update README.md (#4)

Browse files

- Update README.md (fd04d6e1a984345cb5d62d389b22eb489ea19c4c)


Co-authored-by: Vladislav Mikhailov <[email protected]>

Files changed (1) hide show
  1. README.md +16 -2
README.md CHANGED
@@ -7,7 +7,9 @@ tags:
7
  thumbnail: "https://github.com/sberbank-ai/model-zoo"
8
  ---
9
  # ruT5-base
10
- Model was trained by [SberDevices](https://sberdevices.ru/).
 
 
11
  * Task: `text2text generation`
12
  * Type: `encoder-decoder`
13
  * Tokenizer: `bpe`
@@ -17,4 +19,16 @@ Model was trained by [SberDevices](https://sberdevices.ru/).
17
 
18
  # Authors
19
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
20
- + Dmitry Zmitrovich
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  thumbnail: "https://github.com/sberbank-ai/model-zoo"
8
  ---
9
  # ruT5-base
10
+ The model architecture design, pretraining, and evaluation are documented in our preprint: [**A Family of Pretrained Transformer Language Models for Russian**](https://arxiv.org/abs/2309.10931).
11
+
12
+ The model was trained by the [SberDevices](https://sberdevices.ru/).
13
  * Task: `text2text generation`
14
  * Type: `encoder-decoder`
15
  * Tokenizer: `bpe`
 
19
 
20
  # Authors
21
  + NLP core team RnD [Telegram channel](https://t.me/nlpcoreteam):
22
+ + Dmitry Zmitrovich
23
+
24
+ # Cite us
25
+ ```
26
+ @misc{zmitrovich2023family,
27
+ title={A Family of Pretrained Transformer Language Models for Russian},
28
+ author={Dmitry Zmitrovich and Alexander Abramov and Andrey Kalmykov and Maria Tikhonova and Ekaterina Taktasheva and Danil Astafurov and Mark Baushenko and Artem Snegirev and Tatiana Shavrina and Sergey Markov and Vladislav Mikhailov and Alena Fenogenova},
29
+ year={2023},
30
+ eprint={2309.10931},
31
+ archivePrefix={arXiv},
32
+ primaryClass={cs.CL}
33
+ }
34
+ ```