Upload config
Browse files- config.json +3 -5
config.json
CHANGED
@@ -32,7 +32,6 @@
|
|
32 |
"cls_attn_layers": 2,
|
33 |
"decoder_dropout": 0.1,
|
34 |
"decoder_hidden_size": 768,
|
35 |
-
"depth": 23,
|
36 |
"depths": [
|
37 |
3,
|
38 |
7
|
@@ -76,9 +75,7 @@
|
|
76 |
"mlp_ratio": 4.0,
|
77 |
"model_type": "fan",
|
78 |
"norm_layer": null,
|
79 |
-
"num_attention_heads":
|
80 |
-
"num_classes": 1000,
|
81 |
-
"num_heads": [
|
82 |
11,
|
83 |
11,
|
84 |
11,
|
@@ -103,7 +100,8 @@
|
|
103 |
16,
|
104 |
16
|
105 |
],
|
106 |
-
"
|
|
|
107 |
"out_index": -1,
|
108 |
"patch_size": 16,
|
109 |
"qkv_bias": true,
|
|
|
32 |
"cls_attn_layers": 2,
|
33 |
"decoder_dropout": 0.1,
|
34 |
"decoder_hidden_size": 768,
|
|
|
35 |
"depths": [
|
36 |
3,
|
37 |
7
|
|
|
75 |
"mlp_ratio": 4.0,
|
76 |
"model_type": "fan",
|
77 |
"norm_layer": null,
|
78 |
+
"num_attention_heads": [
|
|
|
|
|
79 |
11,
|
80 |
11,
|
81 |
11,
|
|
|
100 |
16,
|
101 |
16
|
102 |
],
|
103 |
+
"num_classes": 1000,
|
104 |
+
"num_hidden_layers": 23,
|
105 |
"out_index": -1,
|
106 |
"patch_size": 16,
|
107 |
"qkv_bias": true,
|