--- language: - it pipeline_tag: fill-mask library_name: transformers widget: - text: "ciao" example_title: "ciao" --- This model (based on [bert-tiny](https://huggingface.co/google/bert_uncased_L-2_H-128_A-2)) has been trained from scratch on 20k italian Wikipedia articles and on a wide collection of italian words and dictionary definitions. The project is still a work in progress, new versions will come with time. # Training - epochs: 200 - lr: 1e-5 - optim: AdamW - weight_decay: 1e-3 # Eval - perplexity: 50 (it's a 12MB model, don't expect this to be ChatGPT anytime soon :)