Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

RoBERTurk is pretrained on Oscar Turkish Split (27GB) and a small chunk of C4 Turkish Split (1GB) with sentencepiece BPE tokenizer that is trained on randomly selected 30M sentences from the training data, which is composed of 90M sentences. The training data in total contains 5.3B tokens and the vocabulary size is 50K. The learning rate is warmed up to the peak value of 1e-5 for the first 10K updates and linearly decayed at $0.01$ rate. The model is pretrained for maximum 600K updates only with sequences of at most T=256 length.

Tokenizer

Load the pretrained tokenizer as follows:

from transformers import AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("Nuri-Tas/roberturk-base")

Model

Get the pretrained model with:

from transformers import RobertaModel
model = RobertaModel.from_pretrained("Nuri-Tas/roberturk-base")

Caveats

There is a slight mismatch between our tokenizer and the default tokenizer used by RobertaTokenizer, which results in some underperformance. I'm working on the issue and will update the tokenizer/model accordingly. Additional TODOs are (although some of them can take some time and I may include them on different repositories):

  • Using Zemberek as an alternative tokenizer
  • Adjusting masking algorithm to be able to mask morphologies besides only complete words
  • Preferably pretraining BPE on the whole training data
  • Pretraining with 512 max sequence length + more data
Downloads last month
5
Safetensors
Model size
125M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.