File size: 432 Bytes
8775b52 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
{
"depth": 24,
"embed_dim": 480,
"eta": 1.0,
"model_type": "fully_attentional_networks",
"num_heads": 10,
"patch_size": 16,
"se_mlp": true,
"sharpen_attn": false,
"sr_ratio": [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
],
"tokens_norm": true,
"transformers_version": "4.22.0.dev0"
}
|