RaushanTurganbay HF staff commited on
Commit
a55e1f6
1 Parent(s): fa5aeee

Upload LlavaNextForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +197 -0
  2. generation_config.json +6 -0
  3. model.safetensors +3 -0
config.json ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/raid/raushan/si-500",
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "ignore_index": -100,
7
+ "image_grid_pinpoints": [
8
+ [
9
+ 384,
10
+ 384
11
+ ],
12
+ [
13
+ 384,
14
+ 768
15
+ ],
16
+ [
17
+ 384,
18
+ 1152
19
+ ],
20
+ [
21
+ 384,
22
+ 1536
23
+ ],
24
+ [
25
+ 384,
26
+ 1920
27
+ ],
28
+ [
29
+ 384,
30
+ 2304
31
+ ],
32
+ [
33
+ 768,
34
+ 384
35
+ ],
36
+ [
37
+ 768,
38
+ 768
39
+ ],
40
+ [
41
+ 768,
42
+ 1152
43
+ ],
44
+ [
45
+ 768,
46
+ 1536
47
+ ],
48
+ [
49
+ 768,
50
+ 1920
51
+ ],
52
+ [
53
+ 768,
54
+ 2304
55
+ ],
56
+ [
57
+ 1152,
58
+ 384
59
+ ],
60
+ [
61
+ 1152,
62
+ 768
63
+ ],
64
+ [
65
+ 1152,
66
+ 1152
67
+ ],
68
+ [
69
+ 1152,
70
+ 1536
71
+ ],
72
+ [
73
+ 1152,
74
+ 1920
75
+ ],
76
+ [
77
+ 1152,
78
+ 2304
79
+ ],
80
+ [
81
+ 1536,
82
+ 384
83
+ ],
84
+ [
85
+ 1536,
86
+ 768
87
+ ],
88
+ [
89
+ 1536,
90
+ 1152
91
+ ],
92
+ [
93
+ 1536,
94
+ 1536
95
+ ],
96
+ [
97
+ 1536,
98
+ 1920
99
+ ],
100
+ [
101
+ 1536,
102
+ 2304
103
+ ],
104
+ [
105
+ 1920,
106
+ 384
107
+ ],
108
+ [
109
+ 1920,
110
+ 768
111
+ ],
112
+ [
113
+ 1920,
114
+ 1152
115
+ ],
116
+ [
117
+ 1920,
118
+ 1536
119
+ ],
120
+ [
121
+ 1920,
122
+ 1920
123
+ ],
124
+ [
125
+ 1920,
126
+ 2304
127
+ ],
128
+ [
129
+ 2304,
130
+ 384
131
+ ],
132
+ [
133
+ 2304,
134
+ 768
135
+ ],
136
+ [
137
+ 2304,
138
+ 1152
139
+ ],
140
+ [
141
+ 2304,
142
+ 1536
143
+ ],
144
+ [
145
+ 2304,
146
+ 1920
147
+ ],
148
+ [
149
+ 2304,
150
+ 2304
151
+ ]
152
+ ],
153
+ "image_token_index": 151646,
154
+ "model_type": "llava_next",
155
+ "projector_hidden_act": "gelu",
156
+ "text_config": {
157
+ "_name_or_path": "Qwen/Qwen2-0.5B-Instruct",
158
+ "architectures": [
159
+ "Qwen2ForCausalLM"
160
+ ],
161
+ "bos_token_id": 151643,
162
+ "eos_token_id": 151645,
163
+ "hidden_size": 896,
164
+ "intermediate_size": 4864,
165
+ "max_position_embeddings": 32768,
166
+ "max_window_layers": 24,
167
+ "model_type": "qwen2",
168
+ "num_attention_heads": 14,
169
+ "num_hidden_layers": 24,
170
+ "num_key_value_heads": 2,
171
+ "rope_theta": 1000000.0,
172
+ "sliding_window": null,
173
+ "tie_word_embeddings": true,
174
+ "torch_dtype": "bfloat16",
175
+ "use_sliding_window": false,
176
+ "vocab_size": 152000
177
+ },
178
+ "tie_word_embeddings": false,
179
+ "torch_dtype": "float16",
180
+ "transformers_version": "4.45.0.dev0",
181
+ "use_image_newline_parameter": true,
182
+ "vision_aspect_ratio": "anyres_max_9",
183
+ "vision_config": {
184
+ "hidden_act": "gelu_pytorch_tanh",
185
+ "hidden_size": 1152,
186
+ "image_size": 384,
187
+ "intermediate_size": 4304,
188
+ "layer_norm_eps": 1e-06,
189
+ "model_type": "siglip_vision_model",
190
+ "num_attention_heads": 16,
191
+ "num_hidden_layers": 26,
192
+ "patch_size": 14,
193
+ "vision_use_head": false
194
+ },
195
+ "vision_feature_layer": -1,
196
+ "vision_feature_select_strategy": "full"
197
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 151643,
4
+ "eos_token_id": 151645,
5
+ "transformers_version": "4.45.0.dev0"
6
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70af414eb9a64bc78f1b5aa9a42fc6ff6b374bf53c561642169080b1ca615fce
3
+ size 1787445680