abuelnasr commited on
Commit
6342222
1 Parent(s): b0980e1

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -1
config.json CHANGED
@@ -13,7 +13,7 @@
13
  "layer_norm_epsilon": 1e-05,
14
  "dropout": 0.0,
15
  "max_sequence_length": 131072,
16
- "training_sequence_length": 4096,
17
  "rope_max_wavelength": 10000.0,
18
  "rope_scaling_type": "su",
19
  "rope_scaling_short_factor": [
 
13
  "layer_norm_epsilon": 1e-05,
14
  "dropout": 0.0,
15
  "max_sequence_length": 131072,
16
+ "pretraining_sequence_length": 4096,
17
  "rope_max_wavelength": 10000.0,
18
  "rope_scaling_type": "su",
19
  "rope_scaling_short_factor": [