可亲
commited on
Commit
•
d1eab90
1
Parent(s):
2a3f54c
fix(pad zero) pad intermediate_size to 29696 to make sure quantized model can use 8 tensor-parallel in vllm
Browse files- config.json +10 -16
- model-00002-of-00021.safetensors +2 -2
- model-00003-of-00021.safetensors +2 -2
- model-00004-of-00021.safetensors +2 -2
- model-00005-of-00021.safetensors +2 -2
- model-00006-of-00021.safetensors +2 -2
- model-00007-of-00021.safetensors +2 -2
- model-00008-of-00021.safetensors +2 -2
- model-00009-of-00021.safetensors +2 -2
- model-00010-of-00021.safetensors +2 -2
- model-00011-of-00021.safetensors +2 -2
- model-00012-of-00021.safetensors +2 -2
- model-00013-of-00021.safetensors +2 -2
- model-00014-of-00021.safetensors +2 -2
- model-00015-of-00021.safetensors +2 -2
- model-00016-of-00021.safetensors +2 -2
- model-00017-of-00021.safetensors +2 -2
- model-00018-of-00021.safetensors +2 -2
- model-00019-of-00021.safetensors +2 -2
- model-00020-of-00021.safetensors +2 -2
- model.safetensors.index.json +61 -61
config.json
CHANGED
@@ -1,20 +1,16 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "Qwen/Qwen2-VL-72B-Instruct-
|
3 |
"architectures": [
|
4 |
"Qwen2VLForConditionalGeneration"
|
5 |
],
|
6 |
"attention_dropout": 0.0,
|
7 |
"bos_token_id": 151643,
|
8 |
"eos_token_id": 151645,
|
9 |
-
"vision_start_token_id": 151652,
|
10 |
-
"vision_end_token_id": 151653,
|
11 |
-
"vision_token_id": 151654,
|
12 |
-
"image_token_id": 151655,
|
13 |
-
"video_token_id": 151656,
|
14 |
"hidden_act": "silu",
|
15 |
"hidden_size": 8192,
|
|
|
16 |
"initializer_range": 0.02,
|
17 |
-
"intermediate_size":
|
18 |
"max_position_embeddings": 32768,
|
19 |
"max_window_layers": 80,
|
20 |
"model_type": "qwen2_vl",
|
@@ -48,17 +44,15 @@
|
|
48 |
"transformers_version": "4.45.0.dev0",
|
49 |
"use_cache": true,
|
50 |
"use_sliding_window": false,
|
|
|
51 |
"vision_config": {
|
52 |
-
"depth": 32,
|
53 |
-
"embed_dim": 1280,
|
54 |
-
"mlp_ratio": 4,
|
55 |
-
"num_heads": 16,
|
56 |
-
"in_chans": 3,
|
57 |
"hidden_size": 8192,
|
58 |
-
"
|
59 |
-
"
|
60 |
-
"spatial_patch_size": 14
|
61 |
-
"temporal_patch_size": 2
|
62 |
},
|
|
|
|
|
|
|
63 |
"vocab_size": 152064
|
64 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2-VL-72B-Instruct-315-fix-dim-pad",
|
3 |
"architectures": [
|
4 |
"Qwen2VLForConditionalGeneration"
|
5 |
],
|
6 |
"attention_dropout": 0.0,
|
7 |
"bos_token_id": 151643,
|
8 |
"eos_token_id": 151645,
|
|
|
|
|
|
|
|
|
|
|
9 |
"hidden_act": "silu",
|
10 |
"hidden_size": 8192,
|
11 |
+
"image_token_id": 151655,
|
12 |
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 29696,
|
14 |
"max_position_embeddings": 32768,
|
15 |
"max_window_layers": 80,
|
16 |
"model_type": "qwen2_vl",
|
|
|
44 |
"transformers_version": "4.45.0.dev0",
|
45 |
"use_cache": true,
|
46 |
"use_sliding_window": false,
|
47 |
+
"video_token_id": 151656,
|
48 |
"vision_config": {
|
|
|
|
|
|
|
|
|
|
|
49 |
"hidden_size": 8192,
|
50 |
+
"in_chans": 3,
|
51 |
+
"model_type": "qwen2_vl",
|
52 |
+
"spatial_patch_size": 14
|
|
|
53 |
},
|
54 |
+
"vision_end_token_id": 151653,
|
55 |
+
"vision_start_token_id": 151652,
|
56 |
+
"vision_token_id": 151654,
|
57 |
"vocab_size": 152064
|
58 |
}
|
model-00002-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22b21f6727ea22cc69edecb38bfb2939b7f5501b617a36eda68355f2e0b2ee50
|
3 |
+
size 3933722616
|
model-00003-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:493b6a9b2df51300de7c77c59db0a4a1c0e1d56e562229f763a333be82a9500f
|
3 |
+
size 3856284568
|
model-00004-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc1b2bcc5ccc001827901e93d244641b7fc466107f8a41afbc7efa2cb97e33f2
|
3 |
+
size 3933653192
|
model-00005-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:110c935df7e411c8f74f2967aaca9e6a4de82025e75257d23c5c322c47ba4593
|
3 |
+
size 3933722760
|
model-00006-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:51dcd557a8c049e4a71651c8ed636dd28aa04fc7cd9ae1bb0089894fb9fe2e7a
|
3 |
+
size 3856284704
|
model-00007-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:63a2ebb51123b1726a9bd68b0773cd1c3dc31cf8be36551b74624767ef716968
|
3 |
+
size 3933653240
|
model-00008-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ffcc710199307f40b8e0854233c57561da443fe33300b00fcc401a3d7a94420c
|
3 |
+
size 3933722760
|
model-00009-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d8dca5803cb00b40867bef50951ab8c15f9c10cd7487f164ae5dffbbd3893914
|
3 |
+
size 3856284704
|
model-00010-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:081e6a9f411358b7a91b3444f4ae49534d884232e05e6d9b420d6339e3d25662
|
3 |
+
size 3933653240
|
model-00011-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f1893bc6b77673e185c8bcfcb10aacc9f1c63785522cafb6c090250919ff1ee
|
3 |
+
size 3933722760
|
model-00012-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ebcfd57658001265ac2e935bf6c323f2d850fe61d6589abe80a5a327733befe
|
3 |
+
size 3856284704
|
model-00013-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81bd0841831608340f38cb807e72c5bc624f8ac2e991c12dbc84b74c1bb2e4cd
|
3 |
+
size 3933653240
|
model-00014-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a19deab70a4fba0534f53ae393720b47e8a22ae9c3b87509ec4c936438b86de
|
3 |
+
size 3933722760
|
model-00015-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02cf34da9b4ed556bde70032e3afe4c3d8caa3a7145c9ea8fef2977624383ee6
|
3 |
+
size 3856284704
|
model-00016-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:07ba4cad0397cdeef6c2c7d3c37e7923913c778a9efdfeecb84bc8f25c6ea625
|
3 |
+
size 3933653240
|
model-00017-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a3257b539618933c2dbaa57d0b133815bd167415fedf06f0ce76f7aebc3b9bd9
|
3 |
+
size 3933722760
|
model-00018-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d32113a8a028c7ee42a5695cd0ca766ef8615f1b7732e4e28f14dd22f061887c
|
3 |
+
size 3856284704
|
model-00019-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f663fddc7364084668196c5fb56c1a635822257e8340b79eed4eb1dd6c4932be
|
3 |
+
size 3933653240
|
model-00020-of-00021.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:680110343579090179e68f0fc5133fe79cd7589525dee02f6a63dbd52957a700
|
3 |
+
size 1726321024
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.weight": "model-00021-of-00021.safetensors",
|
@@ -193,16 +193,16 @@
|
|
193 |
"model.layers.13.self_attn.o_proj.qweight": "model-00004-of-00021.safetensors",
|
194 |
"model.layers.13.self_attn.o_proj.qzeros": "model-00004-of-00021.safetensors",
|
195 |
"model.layers.13.self_attn.o_proj.scales": "model-00004-of-00021.safetensors",
|
196 |
-
"model.layers.13.self_attn.q_proj.bias": "model-
|
197 |
-
"model.layers.13.self_attn.q_proj.g_idx": "model-
|
198 |
-
"model.layers.13.self_attn.q_proj.qweight": "model-
|
199 |
-
"model.layers.13.self_attn.q_proj.qzeros": "model-
|
200 |
-
"model.layers.13.self_attn.q_proj.scales": "model-
|
201 |
-
"model.layers.13.self_attn.v_proj.bias": "model-
|
202 |
-
"model.layers.13.self_attn.v_proj.g_idx": "model-
|
203 |
-
"model.layers.13.self_attn.v_proj.qweight": "model-
|
204 |
-
"model.layers.13.self_attn.v_proj.qzeros": "model-
|
205 |
-
"model.layers.13.self_attn.v_proj.scales": "model-
|
206 |
"model.layers.14.input_layernorm.weight": "model-00005-of-00021.safetensors",
|
207 |
"model.layers.14.mlp.down_proj.g_idx": "model-00005-of-00021.safetensors",
|
208 |
"model.layers.14.mlp.down_proj.qweight": "model-00005-of-00021.safetensors",
|
@@ -655,16 +655,16 @@
|
|
655 |
"model.layers.26.self_attn.o_proj.qweight": "model-00007-of-00021.safetensors",
|
656 |
"model.layers.26.self_attn.o_proj.qzeros": "model-00007-of-00021.safetensors",
|
657 |
"model.layers.26.self_attn.o_proj.scales": "model-00007-of-00021.safetensors",
|
658 |
-
"model.layers.26.self_attn.q_proj.bias": "model-
|
659 |
-
"model.layers.26.self_attn.q_proj.g_idx": "model-
|
660 |
-
"model.layers.26.self_attn.q_proj.qweight": "model-
|
661 |
-
"model.layers.26.self_attn.q_proj.qzeros": "model-
|
662 |
-
"model.layers.26.self_attn.q_proj.scales": "model-
|
663 |
-
"model.layers.26.self_attn.v_proj.bias": "model-
|
664 |
-
"model.layers.26.self_attn.v_proj.g_idx": "model-
|
665 |
-
"model.layers.26.self_attn.v_proj.qweight": "model-
|
666 |
-
"model.layers.26.self_attn.v_proj.qzeros": "model-
|
667 |
-
"model.layers.26.self_attn.v_proj.scales": "model-
|
668 |
"model.layers.27.input_layernorm.weight": "model-00008-of-00021.safetensors",
|
669 |
"model.layers.27.mlp.down_proj.g_idx": "model-00008-of-00021.safetensors",
|
670 |
"model.layers.27.mlp.down_proj.qweight": "model-00008-of-00021.safetensors",
|
@@ -1117,16 +1117,16 @@
|
|
1117 |
"model.layers.39.self_attn.o_proj.qweight": "model-00010-of-00021.safetensors",
|
1118 |
"model.layers.39.self_attn.o_proj.qzeros": "model-00010-of-00021.safetensors",
|
1119 |
"model.layers.39.self_attn.o_proj.scales": "model-00010-of-00021.safetensors",
|
1120 |
-
"model.layers.39.self_attn.q_proj.bias": "model-
|
1121 |
-
"model.layers.39.self_attn.q_proj.g_idx": "model-
|
1122 |
-
"model.layers.39.self_attn.q_proj.qweight": "model-
|
1123 |
-
"model.layers.39.self_attn.q_proj.qzeros": "model-
|
1124 |
-
"model.layers.39.self_attn.q_proj.scales": "model-
|
1125 |
-
"model.layers.39.self_attn.v_proj.bias": "model-
|
1126 |
-
"model.layers.39.self_attn.v_proj.g_idx": "model-
|
1127 |
-
"model.layers.39.self_attn.v_proj.qweight": "model-
|
1128 |
-
"model.layers.39.self_attn.v_proj.qzeros": "model-
|
1129 |
-
"model.layers.39.self_attn.v_proj.scales": "model-
|
1130 |
"model.layers.4.input_layernorm.weight": "model-00003-of-00021.safetensors",
|
1131 |
"model.layers.4.mlp.down_proj.g_idx": "model-00002-of-00021.safetensors",
|
1132 |
"model.layers.4.mlp.down_proj.qweight": "model-00002-of-00021.safetensors",
|
@@ -1612,16 +1612,16 @@
|
|
1612 |
"model.layers.52.self_attn.o_proj.qweight": "model-00013-of-00021.safetensors",
|
1613 |
"model.layers.52.self_attn.o_proj.qzeros": "model-00013-of-00021.safetensors",
|
1614 |
"model.layers.52.self_attn.o_proj.scales": "model-00013-of-00021.safetensors",
|
1615 |
-
"model.layers.52.self_attn.q_proj.bias": "model-
|
1616 |
-
"model.layers.52.self_attn.q_proj.g_idx": "model-
|
1617 |
-
"model.layers.52.self_attn.q_proj.qweight": "model-
|
1618 |
-
"model.layers.52.self_attn.q_proj.qzeros": "model-
|
1619 |
-
"model.layers.52.self_attn.q_proj.scales": "model-
|
1620 |
-
"model.layers.52.self_attn.v_proj.bias": "model-
|
1621 |
-
"model.layers.52.self_attn.v_proj.g_idx": "model-
|
1622 |
-
"model.layers.52.self_attn.v_proj.qweight": "model-
|
1623 |
-
"model.layers.52.self_attn.v_proj.qzeros": "model-
|
1624 |
-
"model.layers.52.self_attn.v_proj.scales": "model-
|
1625 |
"model.layers.53.input_layernorm.weight": "model-00014-of-00021.safetensors",
|
1626 |
"model.layers.53.mlp.down_proj.g_idx": "model-00014-of-00021.safetensors",
|
1627 |
"model.layers.53.mlp.down_proj.qweight": "model-00014-of-00021.safetensors",
|
@@ -2074,16 +2074,16 @@
|
|
2074 |
"model.layers.65.self_attn.o_proj.qweight": "model-00016-of-00021.safetensors",
|
2075 |
"model.layers.65.self_attn.o_proj.qzeros": "model-00016-of-00021.safetensors",
|
2076 |
"model.layers.65.self_attn.o_proj.scales": "model-00016-of-00021.safetensors",
|
2077 |
-
"model.layers.65.self_attn.q_proj.bias": "model-
|
2078 |
-
"model.layers.65.self_attn.q_proj.g_idx": "model-
|
2079 |
-
"model.layers.65.self_attn.q_proj.qweight": "model-
|
2080 |
-
"model.layers.65.self_attn.q_proj.qzeros": "model-
|
2081 |
-
"model.layers.65.self_attn.q_proj.scales": "model-
|
2082 |
-
"model.layers.65.self_attn.v_proj.bias": "model-
|
2083 |
-
"model.layers.65.self_attn.v_proj.g_idx": "model-
|
2084 |
-
"model.layers.65.self_attn.v_proj.qweight": "model-
|
2085 |
-
"model.layers.65.self_attn.v_proj.qzeros": "model-
|
2086 |
-
"model.layers.65.self_attn.v_proj.scales": "model-
|
2087 |
"model.layers.66.input_layernorm.weight": "model-00017-of-00021.safetensors",
|
2088 |
"model.layers.66.mlp.down_proj.g_idx": "model-00017-of-00021.safetensors",
|
2089 |
"model.layers.66.mlp.down_proj.qweight": "model-00017-of-00021.safetensors",
|
@@ -2536,16 +2536,16 @@
|
|
2536 |
"model.layers.78.self_attn.o_proj.qweight": "model-00019-of-00021.safetensors",
|
2537 |
"model.layers.78.self_attn.o_proj.qzeros": "model-00019-of-00021.safetensors",
|
2538 |
"model.layers.78.self_attn.o_proj.scales": "model-00019-of-00021.safetensors",
|
2539 |
-
"model.layers.78.self_attn.q_proj.bias": "model-
|
2540 |
-
"model.layers.78.self_attn.q_proj.g_idx": "model-
|
2541 |
-
"model.layers.78.self_attn.q_proj.qweight": "model-
|
2542 |
-
"model.layers.78.self_attn.q_proj.qzeros": "model-
|
2543 |
-
"model.layers.78.self_attn.q_proj.scales": "model-
|
2544 |
-
"model.layers.78.self_attn.v_proj.bias": "model-
|
2545 |
-
"model.layers.78.self_attn.v_proj.g_idx": "model-
|
2546 |
-
"model.layers.78.self_attn.v_proj.qweight": "model-
|
2547 |
-
"model.layers.78.self_attn.v_proj.qzeros": "model-
|
2548 |
-
"model.layers.78.self_attn.v_proj.scales": "model-
|
2549 |
"model.layers.79.input_layernorm.weight": "model-00020-of-00021.safetensors",
|
2550 |
"model.layers.79.mlp.down_proj.g_idx": "model-00020-of-00021.safetensors",
|
2551 |
"model.layers.79.mlp.down_proj.qweight": "model-00020-of-00021.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 78526868480
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.weight": "model-00021-of-00021.safetensors",
|
|
|
193 |
"model.layers.13.self_attn.o_proj.qweight": "model-00004-of-00021.safetensors",
|
194 |
"model.layers.13.self_attn.o_proj.qzeros": "model-00004-of-00021.safetensors",
|
195 |
"model.layers.13.self_attn.o_proj.scales": "model-00004-of-00021.safetensors",
|
196 |
+
"model.layers.13.self_attn.q_proj.bias": "model-00005-of-00021.safetensors",
|
197 |
+
"model.layers.13.self_attn.q_proj.g_idx": "model-00005-of-00021.safetensors",
|
198 |
+
"model.layers.13.self_attn.q_proj.qweight": "model-00005-of-00021.safetensors",
|
199 |
+
"model.layers.13.self_attn.q_proj.qzeros": "model-00005-of-00021.safetensors",
|
200 |
+
"model.layers.13.self_attn.q_proj.scales": "model-00005-of-00021.safetensors",
|
201 |
+
"model.layers.13.self_attn.v_proj.bias": "model-00005-of-00021.safetensors",
|
202 |
+
"model.layers.13.self_attn.v_proj.g_idx": "model-00005-of-00021.safetensors",
|
203 |
+
"model.layers.13.self_attn.v_proj.qweight": "model-00005-of-00021.safetensors",
|
204 |
+
"model.layers.13.self_attn.v_proj.qzeros": "model-00005-of-00021.safetensors",
|
205 |
+
"model.layers.13.self_attn.v_proj.scales": "model-00005-of-00021.safetensors",
|
206 |
"model.layers.14.input_layernorm.weight": "model-00005-of-00021.safetensors",
|
207 |
"model.layers.14.mlp.down_proj.g_idx": "model-00005-of-00021.safetensors",
|
208 |
"model.layers.14.mlp.down_proj.qweight": "model-00005-of-00021.safetensors",
|
|
|
655 |
"model.layers.26.self_attn.o_proj.qweight": "model-00007-of-00021.safetensors",
|
656 |
"model.layers.26.self_attn.o_proj.qzeros": "model-00007-of-00021.safetensors",
|
657 |
"model.layers.26.self_attn.o_proj.scales": "model-00007-of-00021.safetensors",
|
658 |
+
"model.layers.26.self_attn.q_proj.bias": "model-00008-of-00021.safetensors",
|
659 |
+
"model.layers.26.self_attn.q_proj.g_idx": "model-00008-of-00021.safetensors",
|
660 |
+
"model.layers.26.self_attn.q_proj.qweight": "model-00008-of-00021.safetensors",
|
661 |
+
"model.layers.26.self_attn.q_proj.qzeros": "model-00008-of-00021.safetensors",
|
662 |
+
"model.layers.26.self_attn.q_proj.scales": "model-00008-of-00021.safetensors",
|
663 |
+
"model.layers.26.self_attn.v_proj.bias": "model-00008-of-00021.safetensors",
|
664 |
+
"model.layers.26.self_attn.v_proj.g_idx": "model-00008-of-00021.safetensors",
|
665 |
+
"model.layers.26.self_attn.v_proj.qweight": "model-00008-of-00021.safetensors",
|
666 |
+
"model.layers.26.self_attn.v_proj.qzeros": "model-00008-of-00021.safetensors",
|
667 |
+
"model.layers.26.self_attn.v_proj.scales": "model-00008-of-00021.safetensors",
|
668 |
"model.layers.27.input_layernorm.weight": "model-00008-of-00021.safetensors",
|
669 |
"model.layers.27.mlp.down_proj.g_idx": "model-00008-of-00021.safetensors",
|
670 |
"model.layers.27.mlp.down_proj.qweight": "model-00008-of-00021.safetensors",
|
|
|
1117 |
"model.layers.39.self_attn.o_proj.qweight": "model-00010-of-00021.safetensors",
|
1118 |
"model.layers.39.self_attn.o_proj.qzeros": "model-00010-of-00021.safetensors",
|
1119 |
"model.layers.39.self_attn.o_proj.scales": "model-00010-of-00021.safetensors",
|
1120 |
+
"model.layers.39.self_attn.q_proj.bias": "model-00011-of-00021.safetensors",
|
1121 |
+
"model.layers.39.self_attn.q_proj.g_idx": "model-00011-of-00021.safetensors",
|
1122 |
+
"model.layers.39.self_attn.q_proj.qweight": "model-00011-of-00021.safetensors",
|
1123 |
+
"model.layers.39.self_attn.q_proj.qzeros": "model-00011-of-00021.safetensors",
|
1124 |
+
"model.layers.39.self_attn.q_proj.scales": "model-00011-of-00021.safetensors",
|
1125 |
+
"model.layers.39.self_attn.v_proj.bias": "model-00011-of-00021.safetensors",
|
1126 |
+
"model.layers.39.self_attn.v_proj.g_idx": "model-00011-of-00021.safetensors",
|
1127 |
+
"model.layers.39.self_attn.v_proj.qweight": "model-00011-of-00021.safetensors",
|
1128 |
+
"model.layers.39.self_attn.v_proj.qzeros": "model-00011-of-00021.safetensors",
|
1129 |
+
"model.layers.39.self_attn.v_proj.scales": "model-00011-of-00021.safetensors",
|
1130 |
"model.layers.4.input_layernorm.weight": "model-00003-of-00021.safetensors",
|
1131 |
"model.layers.4.mlp.down_proj.g_idx": "model-00002-of-00021.safetensors",
|
1132 |
"model.layers.4.mlp.down_proj.qweight": "model-00002-of-00021.safetensors",
|
|
|
1612 |
"model.layers.52.self_attn.o_proj.qweight": "model-00013-of-00021.safetensors",
|
1613 |
"model.layers.52.self_attn.o_proj.qzeros": "model-00013-of-00021.safetensors",
|
1614 |
"model.layers.52.self_attn.o_proj.scales": "model-00013-of-00021.safetensors",
|
1615 |
+
"model.layers.52.self_attn.q_proj.bias": "model-00014-of-00021.safetensors",
|
1616 |
+
"model.layers.52.self_attn.q_proj.g_idx": "model-00014-of-00021.safetensors",
|
1617 |
+
"model.layers.52.self_attn.q_proj.qweight": "model-00014-of-00021.safetensors",
|
1618 |
+
"model.layers.52.self_attn.q_proj.qzeros": "model-00014-of-00021.safetensors",
|
1619 |
+
"model.layers.52.self_attn.q_proj.scales": "model-00014-of-00021.safetensors",
|
1620 |
+
"model.layers.52.self_attn.v_proj.bias": "model-00014-of-00021.safetensors",
|
1621 |
+
"model.layers.52.self_attn.v_proj.g_idx": "model-00014-of-00021.safetensors",
|
1622 |
+
"model.layers.52.self_attn.v_proj.qweight": "model-00014-of-00021.safetensors",
|
1623 |
+
"model.layers.52.self_attn.v_proj.qzeros": "model-00014-of-00021.safetensors",
|
1624 |
+
"model.layers.52.self_attn.v_proj.scales": "model-00014-of-00021.safetensors",
|
1625 |
"model.layers.53.input_layernorm.weight": "model-00014-of-00021.safetensors",
|
1626 |
"model.layers.53.mlp.down_proj.g_idx": "model-00014-of-00021.safetensors",
|
1627 |
"model.layers.53.mlp.down_proj.qweight": "model-00014-of-00021.safetensors",
|
|
|
2074 |
"model.layers.65.self_attn.o_proj.qweight": "model-00016-of-00021.safetensors",
|
2075 |
"model.layers.65.self_attn.o_proj.qzeros": "model-00016-of-00021.safetensors",
|
2076 |
"model.layers.65.self_attn.o_proj.scales": "model-00016-of-00021.safetensors",
|
2077 |
+
"model.layers.65.self_attn.q_proj.bias": "model-00017-of-00021.safetensors",
|
2078 |
+
"model.layers.65.self_attn.q_proj.g_idx": "model-00017-of-00021.safetensors",
|
2079 |
+
"model.layers.65.self_attn.q_proj.qweight": "model-00017-of-00021.safetensors",
|
2080 |
+
"model.layers.65.self_attn.q_proj.qzeros": "model-00017-of-00021.safetensors",
|
2081 |
+
"model.layers.65.self_attn.q_proj.scales": "model-00017-of-00021.safetensors",
|
2082 |
+
"model.layers.65.self_attn.v_proj.bias": "model-00017-of-00021.safetensors",
|
2083 |
+
"model.layers.65.self_attn.v_proj.g_idx": "model-00017-of-00021.safetensors",
|
2084 |
+
"model.layers.65.self_attn.v_proj.qweight": "model-00017-of-00021.safetensors",
|
2085 |
+
"model.layers.65.self_attn.v_proj.qzeros": "model-00017-of-00021.safetensors",
|
2086 |
+
"model.layers.65.self_attn.v_proj.scales": "model-00017-of-00021.safetensors",
|
2087 |
"model.layers.66.input_layernorm.weight": "model-00017-of-00021.safetensors",
|
2088 |
"model.layers.66.mlp.down_proj.g_idx": "model-00017-of-00021.safetensors",
|
2089 |
"model.layers.66.mlp.down_proj.qweight": "model-00017-of-00021.safetensors",
|
|
|
2536 |
"model.layers.78.self_attn.o_proj.qweight": "model-00019-of-00021.safetensors",
|
2537 |
"model.layers.78.self_attn.o_proj.qzeros": "model-00019-of-00021.safetensors",
|
2538 |
"model.layers.78.self_attn.o_proj.scales": "model-00019-of-00021.safetensors",
|
2539 |
+
"model.layers.78.self_attn.q_proj.bias": "model-00020-of-00021.safetensors",
|
2540 |
+
"model.layers.78.self_attn.q_proj.g_idx": "model-00020-of-00021.safetensors",
|
2541 |
+
"model.layers.78.self_attn.q_proj.qweight": "model-00020-of-00021.safetensors",
|
2542 |
+
"model.layers.78.self_attn.q_proj.qzeros": "model-00020-of-00021.safetensors",
|
2543 |
+
"model.layers.78.self_attn.q_proj.scales": "model-00020-of-00021.safetensors",
|
2544 |
+
"model.layers.78.self_attn.v_proj.bias": "model-00020-of-00021.safetensors",
|
2545 |
+
"model.layers.78.self_attn.v_proj.g_idx": "model-00020-of-00021.safetensors",
|
2546 |
+
"model.layers.78.self_attn.v_proj.qweight": "model-00020-of-00021.safetensors",
|
2547 |
+
"model.layers.78.self_attn.v_proj.qzeros": "model-00020-of-00021.safetensors",
|
2548 |
+
"model.layers.78.self_attn.v_proj.scales": "model-00020-of-00021.safetensors",
|
2549 |
"model.layers.79.input_layernorm.weight": "model-00020-of-00021.safetensors",
|
2550 |
"model.layers.79.mlp.down_proj.g_idx": "model-00020-of-00021.safetensors",
|
2551 |
"model.layers.79.mlp.down_proj.qweight": "model-00020-of-00021.safetensors",
|