sbmaruf commited on
Commit
8664527
1 Parent(s): b53f232

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -0
README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # bengali-t5-large
2
+
3
+ **bengali-t5-large** is a model trained on the Bengali portion of MT5 dataset.
4
+
5
+ [Flax/Jax Community Week](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104), organized by [HuggingFace](https://huggingface.co/) and TPU usage sponsored by Google.
6
+
7
+ The model is trained on around ~11B tokens (64 size batch, 512 tokens, 350k steps).
8
+
9
+ # Proposal
10
+ - [Project Proposal](https://discuss.huggingface.co/t/pretrain-t5-from-scratch-in-bengali/7121)
11
+
12
+ # Participants
13
+ - [Ibraheem Muhammad Moosa](https://huggingface.co/ibraheemmoosa)
14
+ - [Tasnim Mohiuddin](https://huggingface.co/tasnim)
15
+ - [M Saiful Bari](https://huggingface.co/sbmaruf)
16
+
17
+ ## Useful links
18
+ - [Community Week timeline](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104#summary-timeline-calendar-6)
19
+ - [Community Week README](https://github.com/huggingface/transformers/blob/master/examples/research_projects/jax-projects/README.md)
20
+ - [Masked Language Modelling example scripts](https://github.com/huggingface/transformers/tree/master/examples/flax/language-modeling)
21
+ - [Model Repository](https://huggingface.co/flax-community/roberta-base-als-demo)