system HF staff commited on
Commit
afd5192
1 Parent(s): f69c84c

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "architectures": [
3
- "RobertaForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
@@ -25,7 +25,7 @@
25
  "length_penalty": 1.0,
26
  "max_length": 20,
27
  "max_position_embeddings": 258,
28
- "model_type": "bert",
29
  "num_attention_heads": 16,
30
  "num_beams": 1,
31
  "num_hidden_layers": 24,
 
1
  {
2
  "architectures": [
3
+ "PhobertForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
 
25
  "length_penalty": 1.0,
26
  "max_length": 20,
27
  "max_position_embeddings": 258,
28
+ "model_type": "phobert",
29
  "num_attention_heads": 16,
30
  "num_beams": 1,
31
  "num_hidden_layers": 24,