可亲 commited on
Commit
d1eab90
1 Parent(s): 2a3f54c

fix(pad zero) pad intermediate_size to 29696 to make sure quantized model can use 8 tensor-parallel in vllm

Browse files
config.json CHANGED
@@ -1,20 +1,16 @@
1
  {
2
- "_name_or_path": "Qwen/Qwen2-VL-72B-Instruct-GPTQ-Int8",
3
  "architectures": [
4
  "Qwen2VLForConditionalGeneration"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
  "eos_token_id": 151645,
9
- "vision_start_token_id": 151652,
10
- "vision_end_token_id": 151653,
11
- "vision_token_id": 151654,
12
- "image_token_id": 151655,
13
- "video_token_id": 151656,
14
  "hidden_act": "silu",
15
  "hidden_size": 8192,
 
16
  "initializer_range": 0.02,
17
- "intermediate_size": 29568,
18
  "max_position_embeddings": 32768,
19
  "max_window_layers": 80,
20
  "model_type": "qwen2_vl",
@@ -48,17 +44,15 @@
48
  "transformers_version": "4.45.0.dev0",
49
  "use_cache": true,
50
  "use_sliding_window": false,
 
51
  "vision_config": {
52
- "depth": 32,
53
- "embed_dim": 1280,
54
- "mlp_ratio": 4,
55
- "num_heads": 16,
56
- "in_chans": 3,
57
  "hidden_size": 8192,
58
- "patch_size": 14,
59
- "spatial_merge_size": 2,
60
- "spatial_patch_size": 14,
61
- "temporal_patch_size": 2
62
  },
 
 
 
63
  "vocab_size": 152064
64
  }
 
1
  {
2
+ "_name_or_path": "Qwen/Qwen2-VL-72B-Instruct-315-fix-dim-pad",
3
  "architectures": [
4
  "Qwen2VLForConditionalGeneration"
5
  ],
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 151643,
8
  "eos_token_id": 151645,
 
 
 
 
 
9
  "hidden_act": "silu",
10
  "hidden_size": 8192,
11
+ "image_token_id": 151655,
12
  "initializer_range": 0.02,
13
+ "intermediate_size": 29696,
14
  "max_position_embeddings": 32768,
15
  "max_window_layers": 80,
16
  "model_type": "qwen2_vl",
 
44
  "transformers_version": "4.45.0.dev0",
45
  "use_cache": true,
46
  "use_sliding_window": false,
47
+ "video_token_id": 151656,
48
  "vision_config": {
 
 
 
 
 
49
  "hidden_size": 8192,
50
+ "in_chans": 3,
51
+ "model_type": "qwen2_vl",
52
+ "spatial_patch_size": 14
 
53
  },
54
+ "vision_end_token_id": 151653,
55
+ "vision_start_token_id": 151652,
56
+ "vision_token_id": 151654,
57
  "vocab_size": 152064
58
  }
model-00002-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08bb28260e24d622e27668e6a03d8af143d0f5623151f23d452aa41073e1bc0d
3
- size 3919769080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22b21f6727ea22cc69edecb38bfb2939b7f5501b617a36eda68355f2e0b2ee50
3
+ size 3933722616
model-00003-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:beab19aede9e0ae43e22f33ddea2a687f10e88a7ab3111d1438125514f7d35b5
3
- size 3842331544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:493b6a9b2df51300de7c77c59db0a4a1c0e1d56e562229f763a333be82a9500f
3
+ size 3856284568
model-00004-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:591982fc0dda4f43769267a265d7e435e650fcfb479c2041fca4ef6fc7ac4bea
3
- size 3997052216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc1b2bcc5ccc001827901e93d244641b7fc466107f8a41afbc7efa2cb97e33f2
3
+ size 3933653192
model-00005-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1b44a2b18f45435952333819aefdefe11245302757280ace2fe9aa8b389635b
3
- size 3842417184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:110c935df7e411c8f74f2967aaca9e6a4de82025e75257d23c5c322c47ba4593
3
+ size 3933722760
model-00006-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d6488e4420bc01faef1cccfc3e756dbfe1680a152a12caa7c0fae3e45bd8b93
3
- size 3842331680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51dcd557a8c049e4a71651c8ed636dd28aa04fc7cd9ae1bb0089894fb9fe2e7a
3
+ size 3856284704
model-00007-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec85660e46ecbaba6ed872889e810a7aa0bc089d9c4e4353fdc425b7fec9c18e
3
- size 3997052264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63a2ebb51123b1726a9bd68b0773cd1c3dc31cf8be36551b74624767ef716968
3
+ size 3933653240
model-00008-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f4b5dbda7c3fe2e807e04af67e3ee65a9c1e0cd37b5b1087841d06589e4da919
3
- size 3842417184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffcc710199307f40b8e0854233c57561da443fe33300b00fcc401a3d7a94420c
3
+ size 3933722760
model-00009-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:236b33745d06773bccf078a38ec563dbc4037ab7ec755c0fa6e53e2aa4a78d6a
3
- size 3842331680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8dca5803cb00b40867bef50951ab8c15f9c10cd7487f164ae5dffbbd3893914
3
+ size 3856284704
model-00010-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fff317816b3185918f59740dcdc5df80b4f7c5f353972810d5e84ce383443b0
3
- size 3997052264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:081e6a9f411358b7a91b3444f4ae49534d884232e05e6d9b420d6339e3d25662
3
+ size 3933653240
model-00011-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d09b7131a4c6a25b375b04bccdaf35957a473d89ffe6d83677e7b91c75ed0b78
3
- size 3842417184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f1893bc6b77673e185c8bcfcb10aacc9f1c63785522cafb6c090250919ff1ee
3
+ size 3933722760
model-00012-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fea53411a91d9e46b0a9ff2554037aef559a67f659af7cb098c7ce33c5ed5fe7
3
- size 3842331680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ebcfd57658001265ac2e935bf6c323f2d850fe61d6589abe80a5a327733befe
3
+ size 3856284704
model-00013-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d19f003f3c9f479d4c2b080af61ff23249c7272702ffa84c6242a02cf98ab924
3
- size 3997052264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81bd0841831608340f38cb807e72c5bc624f8ac2e991c12dbc84b74c1bb2e4cd
3
+ size 3933653240
model-00014-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:823bd58a92d46ba9fa089713df5221ee29f426d54a81ad9067d95a96313ba149
3
- size 3842417184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a19deab70a4fba0534f53ae393720b47e8a22ae9c3b87509ec4c936438b86de
3
+ size 3933722760
model-00015-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31197c33385fe298b2f0b173ca0e6480a2e416831ccd9a5360a2e766f5460142
3
- size 3842331680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02cf34da9b4ed556bde70032e3afe4c3d8caa3a7145c9ea8fef2977624383ee6
3
+ size 3856284704
model-00016-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d436c46d4337b997699d408321d9d82d91d6db340b558e7c341711a04f8b0af4
3
- size 3997052264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07ba4cad0397cdeef6c2c7d3c37e7923913c778a9efdfeecb84bc8f25c6ea625
3
+ size 3933653240
model-00017-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b12be6e85fee44720a86d20c856787b7310b8e3af59ff76d22fc2af9a880df3
3
- size 3842417184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3257b539618933c2dbaa57d0b133815bd167415fedf06f0ce76f7aebc3b9bd9
3
+ size 3933722760
model-00018-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11369c3cb0a914e500ac4a329d585da6f6bbe9645cdd95c1253269b37e0f8637
3
- size 3842331680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d32113a8a028c7ee42a5695cd0ca766ef8615f1b7732e4e28f14dd22f061887c
3
+ size 3856284704
model-00019-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d4c2a4025a94114b8edfddeab86bac8c1910da695a585a6474d8cda0ad6fd66
3
- size 3997052264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f663fddc7364084668196c5fb56c1a635822257e8340b79eed4eb1dd6c4932be
3
+ size 3933653240
model-00020-of-00021.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c19bda47e855ee77e8262790cfb532ab285a7b98cbe0a22cf6248630bd69dc98
3
- size 1642529048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:680110343579090179e68f0fc5133fe79cd7589525dee02f6a63dbd52957a700
3
+ size 1726321024
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 78269271040
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00021-of-00021.safetensors",
@@ -193,16 +193,16 @@
193
  "model.layers.13.self_attn.o_proj.qweight": "model-00004-of-00021.safetensors",
194
  "model.layers.13.self_attn.o_proj.qzeros": "model-00004-of-00021.safetensors",
195
  "model.layers.13.self_attn.o_proj.scales": "model-00004-of-00021.safetensors",
196
- "model.layers.13.self_attn.q_proj.bias": "model-00004-of-00021.safetensors",
197
- "model.layers.13.self_attn.q_proj.g_idx": "model-00004-of-00021.safetensors",
198
- "model.layers.13.self_attn.q_proj.qweight": "model-00004-of-00021.safetensors",
199
- "model.layers.13.self_attn.q_proj.qzeros": "model-00004-of-00021.safetensors",
200
- "model.layers.13.self_attn.q_proj.scales": "model-00004-of-00021.safetensors",
201
- "model.layers.13.self_attn.v_proj.bias": "model-00004-of-00021.safetensors",
202
- "model.layers.13.self_attn.v_proj.g_idx": "model-00004-of-00021.safetensors",
203
- "model.layers.13.self_attn.v_proj.qweight": "model-00004-of-00021.safetensors",
204
- "model.layers.13.self_attn.v_proj.qzeros": "model-00004-of-00021.safetensors",
205
- "model.layers.13.self_attn.v_proj.scales": "model-00004-of-00021.safetensors",
206
  "model.layers.14.input_layernorm.weight": "model-00005-of-00021.safetensors",
207
  "model.layers.14.mlp.down_proj.g_idx": "model-00005-of-00021.safetensors",
208
  "model.layers.14.mlp.down_proj.qweight": "model-00005-of-00021.safetensors",
@@ -655,16 +655,16 @@
655
  "model.layers.26.self_attn.o_proj.qweight": "model-00007-of-00021.safetensors",
656
  "model.layers.26.self_attn.o_proj.qzeros": "model-00007-of-00021.safetensors",
657
  "model.layers.26.self_attn.o_proj.scales": "model-00007-of-00021.safetensors",
658
- "model.layers.26.self_attn.q_proj.bias": "model-00007-of-00021.safetensors",
659
- "model.layers.26.self_attn.q_proj.g_idx": "model-00007-of-00021.safetensors",
660
- "model.layers.26.self_attn.q_proj.qweight": "model-00007-of-00021.safetensors",
661
- "model.layers.26.self_attn.q_proj.qzeros": "model-00007-of-00021.safetensors",
662
- "model.layers.26.self_attn.q_proj.scales": "model-00007-of-00021.safetensors",
663
- "model.layers.26.self_attn.v_proj.bias": "model-00007-of-00021.safetensors",
664
- "model.layers.26.self_attn.v_proj.g_idx": "model-00007-of-00021.safetensors",
665
- "model.layers.26.self_attn.v_proj.qweight": "model-00007-of-00021.safetensors",
666
- "model.layers.26.self_attn.v_proj.qzeros": "model-00007-of-00021.safetensors",
667
- "model.layers.26.self_attn.v_proj.scales": "model-00007-of-00021.safetensors",
668
  "model.layers.27.input_layernorm.weight": "model-00008-of-00021.safetensors",
669
  "model.layers.27.mlp.down_proj.g_idx": "model-00008-of-00021.safetensors",
670
  "model.layers.27.mlp.down_proj.qweight": "model-00008-of-00021.safetensors",
@@ -1117,16 +1117,16 @@
1117
  "model.layers.39.self_attn.o_proj.qweight": "model-00010-of-00021.safetensors",
1118
  "model.layers.39.self_attn.o_proj.qzeros": "model-00010-of-00021.safetensors",
1119
  "model.layers.39.self_attn.o_proj.scales": "model-00010-of-00021.safetensors",
1120
- "model.layers.39.self_attn.q_proj.bias": "model-00010-of-00021.safetensors",
1121
- "model.layers.39.self_attn.q_proj.g_idx": "model-00010-of-00021.safetensors",
1122
- "model.layers.39.self_attn.q_proj.qweight": "model-00010-of-00021.safetensors",
1123
- "model.layers.39.self_attn.q_proj.qzeros": "model-00010-of-00021.safetensors",
1124
- "model.layers.39.self_attn.q_proj.scales": "model-00010-of-00021.safetensors",
1125
- "model.layers.39.self_attn.v_proj.bias": "model-00010-of-00021.safetensors",
1126
- "model.layers.39.self_attn.v_proj.g_idx": "model-00010-of-00021.safetensors",
1127
- "model.layers.39.self_attn.v_proj.qweight": "model-00010-of-00021.safetensors",
1128
- "model.layers.39.self_attn.v_proj.qzeros": "model-00010-of-00021.safetensors",
1129
- "model.layers.39.self_attn.v_proj.scales": "model-00010-of-00021.safetensors",
1130
  "model.layers.4.input_layernorm.weight": "model-00003-of-00021.safetensors",
1131
  "model.layers.4.mlp.down_proj.g_idx": "model-00002-of-00021.safetensors",
1132
  "model.layers.4.mlp.down_proj.qweight": "model-00002-of-00021.safetensors",
@@ -1612,16 +1612,16 @@
1612
  "model.layers.52.self_attn.o_proj.qweight": "model-00013-of-00021.safetensors",
1613
  "model.layers.52.self_attn.o_proj.qzeros": "model-00013-of-00021.safetensors",
1614
  "model.layers.52.self_attn.o_proj.scales": "model-00013-of-00021.safetensors",
1615
- "model.layers.52.self_attn.q_proj.bias": "model-00013-of-00021.safetensors",
1616
- "model.layers.52.self_attn.q_proj.g_idx": "model-00013-of-00021.safetensors",
1617
- "model.layers.52.self_attn.q_proj.qweight": "model-00013-of-00021.safetensors",
1618
- "model.layers.52.self_attn.q_proj.qzeros": "model-00013-of-00021.safetensors",
1619
- "model.layers.52.self_attn.q_proj.scales": "model-00013-of-00021.safetensors",
1620
- "model.layers.52.self_attn.v_proj.bias": "model-00013-of-00021.safetensors",
1621
- "model.layers.52.self_attn.v_proj.g_idx": "model-00013-of-00021.safetensors",
1622
- "model.layers.52.self_attn.v_proj.qweight": "model-00013-of-00021.safetensors",
1623
- "model.layers.52.self_attn.v_proj.qzeros": "model-00013-of-00021.safetensors",
1624
- "model.layers.52.self_attn.v_proj.scales": "model-00013-of-00021.safetensors",
1625
  "model.layers.53.input_layernorm.weight": "model-00014-of-00021.safetensors",
1626
  "model.layers.53.mlp.down_proj.g_idx": "model-00014-of-00021.safetensors",
1627
  "model.layers.53.mlp.down_proj.qweight": "model-00014-of-00021.safetensors",
@@ -2074,16 +2074,16 @@
2074
  "model.layers.65.self_attn.o_proj.qweight": "model-00016-of-00021.safetensors",
2075
  "model.layers.65.self_attn.o_proj.qzeros": "model-00016-of-00021.safetensors",
2076
  "model.layers.65.self_attn.o_proj.scales": "model-00016-of-00021.safetensors",
2077
- "model.layers.65.self_attn.q_proj.bias": "model-00016-of-00021.safetensors",
2078
- "model.layers.65.self_attn.q_proj.g_idx": "model-00016-of-00021.safetensors",
2079
- "model.layers.65.self_attn.q_proj.qweight": "model-00016-of-00021.safetensors",
2080
- "model.layers.65.self_attn.q_proj.qzeros": "model-00016-of-00021.safetensors",
2081
- "model.layers.65.self_attn.q_proj.scales": "model-00016-of-00021.safetensors",
2082
- "model.layers.65.self_attn.v_proj.bias": "model-00016-of-00021.safetensors",
2083
- "model.layers.65.self_attn.v_proj.g_idx": "model-00016-of-00021.safetensors",
2084
- "model.layers.65.self_attn.v_proj.qweight": "model-00016-of-00021.safetensors",
2085
- "model.layers.65.self_attn.v_proj.qzeros": "model-00016-of-00021.safetensors",
2086
- "model.layers.65.self_attn.v_proj.scales": "model-00016-of-00021.safetensors",
2087
  "model.layers.66.input_layernorm.weight": "model-00017-of-00021.safetensors",
2088
  "model.layers.66.mlp.down_proj.g_idx": "model-00017-of-00021.safetensors",
2089
  "model.layers.66.mlp.down_proj.qweight": "model-00017-of-00021.safetensors",
@@ -2536,16 +2536,16 @@
2536
  "model.layers.78.self_attn.o_proj.qweight": "model-00019-of-00021.safetensors",
2537
  "model.layers.78.self_attn.o_proj.qzeros": "model-00019-of-00021.safetensors",
2538
  "model.layers.78.self_attn.o_proj.scales": "model-00019-of-00021.safetensors",
2539
- "model.layers.78.self_attn.q_proj.bias": "model-00019-of-00021.safetensors",
2540
- "model.layers.78.self_attn.q_proj.g_idx": "model-00019-of-00021.safetensors",
2541
- "model.layers.78.self_attn.q_proj.qweight": "model-00019-of-00021.safetensors",
2542
- "model.layers.78.self_attn.q_proj.qzeros": "model-00019-of-00021.safetensors",
2543
- "model.layers.78.self_attn.q_proj.scales": "model-00019-of-00021.safetensors",
2544
- "model.layers.78.self_attn.v_proj.bias": "model-00019-of-00021.safetensors",
2545
- "model.layers.78.self_attn.v_proj.g_idx": "model-00019-of-00021.safetensors",
2546
- "model.layers.78.self_attn.v_proj.qweight": "model-00019-of-00021.safetensors",
2547
- "model.layers.78.self_attn.v_proj.qzeros": "model-00019-of-00021.safetensors",
2548
- "model.layers.78.self_attn.v_proj.scales": "model-00019-of-00021.safetensors",
2549
  "model.layers.79.input_layernorm.weight": "model-00020-of-00021.safetensors",
2550
  "model.layers.79.mlp.down_proj.g_idx": "model-00020-of-00021.safetensors",
2551
  "model.layers.79.mlp.down_proj.qweight": "model-00020-of-00021.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 78526868480
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00021-of-00021.safetensors",
 
193
  "model.layers.13.self_attn.o_proj.qweight": "model-00004-of-00021.safetensors",
194
  "model.layers.13.self_attn.o_proj.qzeros": "model-00004-of-00021.safetensors",
195
  "model.layers.13.self_attn.o_proj.scales": "model-00004-of-00021.safetensors",
196
+ "model.layers.13.self_attn.q_proj.bias": "model-00005-of-00021.safetensors",
197
+ "model.layers.13.self_attn.q_proj.g_idx": "model-00005-of-00021.safetensors",
198
+ "model.layers.13.self_attn.q_proj.qweight": "model-00005-of-00021.safetensors",
199
+ "model.layers.13.self_attn.q_proj.qzeros": "model-00005-of-00021.safetensors",
200
+ "model.layers.13.self_attn.q_proj.scales": "model-00005-of-00021.safetensors",
201
+ "model.layers.13.self_attn.v_proj.bias": "model-00005-of-00021.safetensors",
202
+ "model.layers.13.self_attn.v_proj.g_idx": "model-00005-of-00021.safetensors",
203
+ "model.layers.13.self_attn.v_proj.qweight": "model-00005-of-00021.safetensors",
204
+ "model.layers.13.self_attn.v_proj.qzeros": "model-00005-of-00021.safetensors",
205
+ "model.layers.13.self_attn.v_proj.scales": "model-00005-of-00021.safetensors",
206
  "model.layers.14.input_layernorm.weight": "model-00005-of-00021.safetensors",
207
  "model.layers.14.mlp.down_proj.g_idx": "model-00005-of-00021.safetensors",
208
  "model.layers.14.mlp.down_proj.qweight": "model-00005-of-00021.safetensors",
 
655
  "model.layers.26.self_attn.o_proj.qweight": "model-00007-of-00021.safetensors",
656
  "model.layers.26.self_attn.o_proj.qzeros": "model-00007-of-00021.safetensors",
657
  "model.layers.26.self_attn.o_proj.scales": "model-00007-of-00021.safetensors",
658
+ "model.layers.26.self_attn.q_proj.bias": "model-00008-of-00021.safetensors",
659
+ "model.layers.26.self_attn.q_proj.g_idx": "model-00008-of-00021.safetensors",
660
+ "model.layers.26.self_attn.q_proj.qweight": "model-00008-of-00021.safetensors",
661
+ "model.layers.26.self_attn.q_proj.qzeros": "model-00008-of-00021.safetensors",
662
+ "model.layers.26.self_attn.q_proj.scales": "model-00008-of-00021.safetensors",
663
+ "model.layers.26.self_attn.v_proj.bias": "model-00008-of-00021.safetensors",
664
+ "model.layers.26.self_attn.v_proj.g_idx": "model-00008-of-00021.safetensors",
665
+ "model.layers.26.self_attn.v_proj.qweight": "model-00008-of-00021.safetensors",
666
+ "model.layers.26.self_attn.v_proj.qzeros": "model-00008-of-00021.safetensors",
667
+ "model.layers.26.self_attn.v_proj.scales": "model-00008-of-00021.safetensors",
668
  "model.layers.27.input_layernorm.weight": "model-00008-of-00021.safetensors",
669
  "model.layers.27.mlp.down_proj.g_idx": "model-00008-of-00021.safetensors",
670
  "model.layers.27.mlp.down_proj.qweight": "model-00008-of-00021.safetensors",
 
1117
  "model.layers.39.self_attn.o_proj.qweight": "model-00010-of-00021.safetensors",
1118
  "model.layers.39.self_attn.o_proj.qzeros": "model-00010-of-00021.safetensors",
1119
  "model.layers.39.self_attn.o_proj.scales": "model-00010-of-00021.safetensors",
1120
+ "model.layers.39.self_attn.q_proj.bias": "model-00011-of-00021.safetensors",
1121
+ "model.layers.39.self_attn.q_proj.g_idx": "model-00011-of-00021.safetensors",
1122
+ "model.layers.39.self_attn.q_proj.qweight": "model-00011-of-00021.safetensors",
1123
+ "model.layers.39.self_attn.q_proj.qzeros": "model-00011-of-00021.safetensors",
1124
+ "model.layers.39.self_attn.q_proj.scales": "model-00011-of-00021.safetensors",
1125
+ "model.layers.39.self_attn.v_proj.bias": "model-00011-of-00021.safetensors",
1126
+ "model.layers.39.self_attn.v_proj.g_idx": "model-00011-of-00021.safetensors",
1127
+ "model.layers.39.self_attn.v_proj.qweight": "model-00011-of-00021.safetensors",
1128
+ "model.layers.39.self_attn.v_proj.qzeros": "model-00011-of-00021.safetensors",
1129
+ "model.layers.39.self_attn.v_proj.scales": "model-00011-of-00021.safetensors",
1130
  "model.layers.4.input_layernorm.weight": "model-00003-of-00021.safetensors",
1131
  "model.layers.4.mlp.down_proj.g_idx": "model-00002-of-00021.safetensors",
1132
  "model.layers.4.mlp.down_proj.qweight": "model-00002-of-00021.safetensors",
 
1612
  "model.layers.52.self_attn.o_proj.qweight": "model-00013-of-00021.safetensors",
1613
  "model.layers.52.self_attn.o_proj.qzeros": "model-00013-of-00021.safetensors",
1614
  "model.layers.52.self_attn.o_proj.scales": "model-00013-of-00021.safetensors",
1615
+ "model.layers.52.self_attn.q_proj.bias": "model-00014-of-00021.safetensors",
1616
+ "model.layers.52.self_attn.q_proj.g_idx": "model-00014-of-00021.safetensors",
1617
+ "model.layers.52.self_attn.q_proj.qweight": "model-00014-of-00021.safetensors",
1618
+ "model.layers.52.self_attn.q_proj.qzeros": "model-00014-of-00021.safetensors",
1619
+ "model.layers.52.self_attn.q_proj.scales": "model-00014-of-00021.safetensors",
1620
+ "model.layers.52.self_attn.v_proj.bias": "model-00014-of-00021.safetensors",
1621
+ "model.layers.52.self_attn.v_proj.g_idx": "model-00014-of-00021.safetensors",
1622
+ "model.layers.52.self_attn.v_proj.qweight": "model-00014-of-00021.safetensors",
1623
+ "model.layers.52.self_attn.v_proj.qzeros": "model-00014-of-00021.safetensors",
1624
+ "model.layers.52.self_attn.v_proj.scales": "model-00014-of-00021.safetensors",
1625
  "model.layers.53.input_layernorm.weight": "model-00014-of-00021.safetensors",
1626
  "model.layers.53.mlp.down_proj.g_idx": "model-00014-of-00021.safetensors",
1627
  "model.layers.53.mlp.down_proj.qweight": "model-00014-of-00021.safetensors",
 
2074
  "model.layers.65.self_attn.o_proj.qweight": "model-00016-of-00021.safetensors",
2075
  "model.layers.65.self_attn.o_proj.qzeros": "model-00016-of-00021.safetensors",
2076
  "model.layers.65.self_attn.o_proj.scales": "model-00016-of-00021.safetensors",
2077
+ "model.layers.65.self_attn.q_proj.bias": "model-00017-of-00021.safetensors",
2078
+ "model.layers.65.self_attn.q_proj.g_idx": "model-00017-of-00021.safetensors",
2079
+ "model.layers.65.self_attn.q_proj.qweight": "model-00017-of-00021.safetensors",
2080
+ "model.layers.65.self_attn.q_proj.qzeros": "model-00017-of-00021.safetensors",
2081
+ "model.layers.65.self_attn.q_proj.scales": "model-00017-of-00021.safetensors",
2082
+ "model.layers.65.self_attn.v_proj.bias": "model-00017-of-00021.safetensors",
2083
+ "model.layers.65.self_attn.v_proj.g_idx": "model-00017-of-00021.safetensors",
2084
+ "model.layers.65.self_attn.v_proj.qweight": "model-00017-of-00021.safetensors",
2085
+ "model.layers.65.self_attn.v_proj.qzeros": "model-00017-of-00021.safetensors",
2086
+ "model.layers.65.self_attn.v_proj.scales": "model-00017-of-00021.safetensors",
2087
  "model.layers.66.input_layernorm.weight": "model-00017-of-00021.safetensors",
2088
  "model.layers.66.mlp.down_proj.g_idx": "model-00017-of-00021.safetensors",
2089
  "model.layers.66.mlp.down_proj.qweight": "model-00017-of-00021.safetensors",
 
2536
  "model.layers.78.self_attn.o_proj.qweight": "model-00019-of-00021.safetensors",
2537
  "model.layers.78.self_attn.o_proj.qzeros": "model-00019-of-00021.safetensors",
2538
  "model.layers.78.self_attn.o_proj.scales": "model-00019-of-00021.safetensors",
2539
+ "model.layers.78.self_attn.q_proj.bias": "model-00020-of-00021.safetensors",
2540
+ "model.layers.78.self_attn.q_proj.g_idx": "model-00020-of-00021.safetensors",
2541
+ "model.layers.78.self_attn.q_proj.qweight": "model-00020-of-00021.safetensors",
2542
+ "model.layers.78.self_attn.q_proj.qzeros": "model-00020-of-00021.safetensors",
2543
+ "model.layers.78.self_attn.q_proj.scales": "model-00020-of-00021.safetensors",
2544
+ "model.layers.78.self_attn.v_proj.bias": "model-00020-of-00021.safetensors",
2545
+ "model.layers.78.self_attn.v_proj.g_idx": "model-00020-of-00021.safetensors",
2546
+ "model.layers.78.self_attn.v_proj.qweight": "model-00020-of-00021.safetensors",
2547
+ "model.layers.78.self_attn.v_proj.qzeros": "model-00020-of-00021.safetensors",
2548
+ "model.layers.78.self_attn.v_proj.scales": "model-00020-of-00021.safetensors",
2549
  "model.layers.79.input_layernorm.weight": "model-00020-of-00021.safetensors",
2550
  "model.layers.79.mlp.down_proj.g_idx": "model-00020-of-00021.safetensors",
2551
  "model.layers.79.mlp.down_proj.qweight": "model-00020-of-00021.safetensors",