|
--- |
|
license: cc-by-4.0 |
|
language: mr |
|
datasets: |
|
- L3Cube-MahaCorpus |
|
--- |
|
|
|
## MahaBERT |
|
|
|
New version of this model is available here: https://huggingface.co/l3cube-pune/marathi-bert-v2 |
|
|
|
MahaBERT is a Marathi BERT model. It is a multilingual BERT (bert-base-multilingual-cased) model fine-tuned on L3Cube-MahaCorpus and other publicly available Marathi monolingual datasets. |
|
[dataset link] (https://github.com/l3cube-pune/MarathiNLP) |
|
|
|
More details on the dataset, models, and baseline results can be found in our [paper] (https://arxiv.org/abs/2202.01159) |
|
|
|
``` |
|
@InProceedings{joshi:2022:WILDRE6, |
|
author = {Joshi, Raviraj}, |
|
title = {L3Cube-MahaCorpus and MahaBERT: Marathi Monolingual Corpus, Marathi BERT Language Models, and Resources}, |
|
booktitle = {Proceedings of The WILDRE-6 Workshop within the 13th Language Resources and Evaluation Conference}, |
|
month = {June}, |
|
year = {2022}, |
|
address = {Marseille, France}, |
|
publisher = {European Language Resources Association}, |
|
pages = {97--101} |
|
} |
|
``` |