bengali-t5-base / README.md
sbmaruf's picture
Update README.md
899ac4b
|
raw
history blame
2 kB

bengali-t5-base

bengali-t5-base is a model trained on the Bengali portion of MT5 dataset. We used the T5-base model for this model.

Flax/Jax Community Week, organized by HuggingFace and TPU usage sponsored by Google.

The model is trained on around ~11B tokens (64 size batch, 512 tokens, 350k steps).

load tokenizer

>>> tokenizer = transformers.AutoTokenizer.from_pretrained("flax-community/bengali-t5-large")
>>> tokenizer.encode("আমি বাংলার গান গাই")
>>> tokenizer.decode([93, 1912, 814, 5995, 3, 1])
[93, 1912, 814, 5995, 3, 1]
'আমি বাংলার গান গাই </s>'

load model

config  = T5Config.from_pretrained("flax-community/bengali-t5-base")
model = FlaxT5ForConditionalGeneration.from_pretrained("flax-community/bengali-t5-base", config=config)

Please note that we haven't finetuned the model in any downstream task. If you are finetuning the model in any downstream task, please let us know about it. Shoot us an email (sbmaruf at gmail dot com)

Proposal

Participants

Useful links