fan_large_24_p16_224 / config.json
ksmcg's picture
Upload config
c76e5f4
raw
history blame
538 Bytes
{
"depth": 24,
"embed_dim": 480,
"eta": 1.0,
"img_size": [
224,
224
],
"in_chans": 3,
"initializer_range": 1.0,
"model_type": "fully_attentional_networks",
"num_classes": 1000,
"num_heads": 10,
"patch_size": 16,
"se_mlp": true,
"sharpen_attn": false,
"sr_ratio": [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
],
"tokens_norm": true,
"transformers_version": "4.22.0.dev0"
}