Joemgu commited on
Commit
26d5962
1 Parent(s): e7060da

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +14 -2
config.json CHANGED
@@ -8,7 +8,7 @@
8
  "d_model": 1024,
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "gelu_new",
11
- "dropout_rate": 0.1,
12
  "encoder_attention_type": "transient-global",
13
  "eos_token_id": 1,
14
  "feed_forward_proj": "gated-gelu",
@@ -28,8 +28,20 @@
28
  "relative_attention_max_distance": 128,
29
  "relative_attention_num_buckets": 32,
30
  "tie_word_embeddings": false,
 
31
  "torch_dtype": "float32",
32
  "transformers_version": "4.30.2",
33
- "use_cache": false,
 
 
 
 
 
 
 
 
 
 
 
34
  "vocab_size": 256384
35
  }
 
8
  "d_model": 1024,
9
  "decoder_start_token_id": 0,
10
  "dense_act_fn": "gelu_new",
11
+ "dropout_rate": 0.0,
12
  "encoder_attention_type": "transient-global",
13
  "eos_token_id": 1,
14
  "feed_forward_proj": "gated-gelu",
 
28
  "relative_attention_max_distance": 128,
29
  "relative_attention_num_buckets": 32,
30
  "tie_word_embeddings": false,
31
+ "tie_weights": false,
32
  "torch_dtype": "float32",
33
  "transformers_version": "4.30.2",
34
+ "use_cache": true,
35
+ "task_specific_params": {
36
+ "summarization": {
37
+ "use_cache": true,
38
+ "top_k": 3,
39
+ "penalty_alpha": 0.4,
40
+ "renormalize_logits": true,
41
+ "do_sample": false,
42
+ "num_beams": 1,
43
+ "no_repeat_ngram_size": 5,
44
+ }
45
+ },
46
  "vocab_size": 256384
47
  }