|
{ |
|
"add_ffn_bias": false, |
|
"add_lm_head_bias": true, |
|
"add_qkv_bias": true, |
|
"architectures": [ |
|
"OpenBAForConditionalGeneration" |
|
], |
|
"attention_dropout": 0.0, |
|
"auto_map": { |
|
"AutoConfig": "configuration_openba.OpenBAConfig", |
|
"AutoModel": "modeling_openba.OpenBAForConditionalGeneration", |
|
"AutoModelForCausalLM": "modeling_openba.OpenBAForConditionalGeneration", |
|
"AutoModelForSeq2SeqLM": "modeling_openba.OpenBAForConditionalGeneration" |
|
}, |
|
"decoder_max_seq_length": 256, |
|
"decoder_start_token_id": 0, |
|
"eos_token_id": 1, |
|
"ffn_hidden_size": 6912, |
|
"hidden_dropout": 0.0, |
|
"hidden_size": 2560, |
|
"initializer_factor": 1.0, |
|
"is_encoder_decoder": true, |
|
"kv_channels": 128, |
|
"max_seq_length": 1024, |
|
"model_type": "openba", |
|
"num_decoder_layers": 24, |
|
"num_heads": 20, |
|
"num_layers": 8, |
|
"pad_token_id": 0, |
|
"tie_word_embeddings": false, |
|
"tokenizer_class": "OpenBATokenizer", |
|
"transformers_version": "4.32.0", |
|
"use_cache": true, |
|
"vocab_size": 250240 |
|
} |
|
|