File size: 591 Bytes
f28d01d
 
 
7df81a3
 
5111699
 
 
c65f616
9e433c4
339c7f5
9e433c4
8933331
9e433c4
 
 
 
 
 
 
 
 
 
963c004
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
---
language:
- it
pipeline_tag: fill-mask
library_name: transformers
widget:
- text: "ciao"
  example_title: "ciao"
---

This model (based on [bert-tiny](https://huggingface.co/google/bert_uncased_L-2_H-128_A-2)) has been trained from scratch on 20k italian Wikipedia articles and on a wide collection of italian words and dictionary definitions.

The project is still a work in progress, new versions will come with time.

# Training

- epochs: 200
- lr: 1e-5
- optim: AdamW
- weight_decay: 1e-3

# Eval

- perplexity: 50 (it's a 12MB model, don't expect this to be ChatGPT anytime soon :)