fan_large_24_p16_224 / config.json
ksmcg's picture
Upload config
0eb2ac2
raw
history blame
1.06 kB
{
"act_layer": null,
"attn_drop_rate": 0.0,
"backbone": null,
"channel_dims": null,
"channels": 256,
"cls_attn_layers": 2,
"decoder_dropout": 0.1,
"decoder_hidden_size": 768,
"depths": null,
"drop_path_rate": 0.0,
"drop_rate": 0.0,
"embed_dim": 480,
"eta": 1.0,
"feat_downsample": false,
"feature_strides": [
4,
8,
16,
32
],
"hybrid_patch_size": 2,
"img_size": [
224,
224
],
"in_chans": 3,
"in_index": [
0,
1,
2,
3
],
"initializer_range": 1.0,
"mlp_ratio": 4.0,
"model_type": "fan",
"norm_layer": null,
"num_attention_heads": 10,
"num_classes": 1000,
"num_hidden_layers": 24,
"out_index": -1,
"patch_size": 16,
"qkv_bias": true,
"reshape_last_stage": false,
"rounding_mode": "floor",
"se_mlp": false,
"segmentation_in_channels": [
128,
256,
480,
480
],
"semantic_loss_ignore_index": -100,
"sharpen_attn": false,
"tokens_norm": true,
"transformers_version": "4.24.0.dev0",
"use_checkpoint": false,
"use_pos_embed": true
}