diff --git a/config.json b/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..b11b2f749cea89cd67e8517619f0dbb7e5231171
--- /dev/null
+++ b/config.json
@@ -0,0 +1,75 @@
+{
+ "_name_or_path": "converted",
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": 128000,
+ "compression_config": {
+ "config_groups": {
+ "group_0": {
+ "input_activations": {
+ "block_structure": null,
+ "dynamic": true,
+ "group_size": null,
+ "num_bits": 8,
+ "observer": "memoryless",
+ "observer_kwargs": {},
+ "strategy": "token",
+ "symmetric": true,
+ "type": "float"
+ },
+ "output_activations": null,
+ "targets": [
+ "Linear"
+ ],
+ "weights": {
+ "block_structure": null,
+ "dynamic": false,
+ "group_size": null,
+ "num_bits": 8,
+ "observer": "minmax",
+ "observer_kwargs": {},
+ "strategy": "channel",
+ "symmetric": true,
+ "type": "float"
+ }
+ }
+ },
+ "format": "float-quantized",
+ "global_compression_ratio": 1.240844678218891,
+ "ignore": [
+ "lm_head"
+ ],
+ "kv_cache_scheme": null,
+ "quant_method": "compressed-tensors",
+ "quantization_status": "frozen"
+ },
+ "eos_token_id": 128039,
+ "hidden_act": "silu",
+ "hidden_size": 16384,
+ "initializer_range": 0.02,
+ "intermediate_size": 53248,
+ "max_position_embeddings": 131072,
+ "mlp_bias": false,
+ "model_type": "llama",
+ "num_attention_heads": 128,
+ "num_hidden_layers": 126,
+ "num_key_value_heads": 8,
+ "pretraining_tp": 1,
+ "rms_norm_eps": 1e-05,
+ "rope_scaling": {
+ "factor": 8.0,
+ "high_freq_factor": 4.0,
+ "low_freq_factor": 1.0,
+ "original_max_position_embeddings": 8192,
+ "rope_type": "llama3"
+ },
+ "rope_theta": 500000.0,
+ "tie_word_embeddings": false,
+ "torch_dtype": "float16",
+ "transformers_version": "4.44.0.dev0",
+ "use_cache": false,
+ "vocab_size": 128256
+}
\ No newline at end of file
diff --git a/generation_config.json b/generation_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..8433be282f84fd89b2ebb60092daba281e371036
--- /dev/null
+++ b/generation_config.json
@@ -0,0 +1,7 @@
+{
+ "_from_model_config": true,
+ "bos_token_id": 128000,
+ "eos_token_id": 128039,
+ "transformers_version": "4.44.0.dev0",
+ "use_cache": false
+}
diff --git a/model-00001-of-00086.safetensors b/model-00001-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a7a9183c6aceabfccfd2b8c5912d9ebb72193069
--- /dev/null
+++ b/model-00001-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:054c70a66e8192fe6220b1be537b1fb7b28d4c7e772edd533fd6d9a61436420a
+size 4773188688
diff --git a/model-00002-of-00086.safetensors b/model-00002-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a36ff3f05d6e5dd5cd180388ecbed0662d49c1c9
--- /dev/null
+++ b/model-00002-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b3a46132d9aebc0bfd102b2375907e5360bece99e03418183805a8e72c4e5d51
+size 4933097728
diff --git a/model-00003-of-00086.safetensors b/model-00003-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3d71159d93233ce3dda5ab96d4b657ec863a0bef
--- /dev/null
+++ b/model-00003-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d6d20f477d22e7ce28c29222471dc95770c9883733ad15fc7351eafa51946a2e
+size 4631063712
diff --git a/model-00004-of-00086.safetensors b/model-00004-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7ff9fbec648114a78d953cd388bac2833e3b75c4
--- /dev/null
+++ b/model-00004-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:523a21eab5707cd2f3e36ac70d8c92d1eb3309b7daeca9cfb2a33ae88ff8bef2
+size 4933097728
diff --git a/model-00005-of-00086.safetensors b/model-00005-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..36181668750296d21cf31417a450585c46a35ef2
--- /dev/null
+++ b/model-00005-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:50f4fcd12894200314d73f0bf50630ad524adc0958ad5c10e1701e457009c32f
+size 4631063712
diff --git a/model-00006-of-00086.safetensors b/model-00006-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4e83b244942ef565b7429b6d9b4e709669173e10
--- /dev/null
+++ b/model-00006-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9a05128a2578fea96a0710c4fa6983982874cd3f937617c56b48952e09aaafe6
+size 4933097728
diff --git a/model-00007-of-00086.safetensors b/model-00007-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..210c16bd0bc03b625febdb8e1aebe004a1e6635e
--- /dev/null
+++ b/model-00007-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:361f2f3c42fa34f77c4188293039a1972ceeb88a36807ce59ad854ffe623ea94
+size 4631063712
diff --git a/model-00008-of-00086.safetensors b/model-00008-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a17c9af2ba0b5a3536ac26442377652fedba7fdc
--- /dev/null
+++ b/model-00008-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:40a97526ea98d257b3851370141aab15b084929733a793fb8e76f00cbb06d96a
+size 4933097744
diff --git a/model-00009-of-00086.safetensors b/model-00009-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..2248eba9cc801e6161df202ea15e7b47a90871bf
--- /dev/null
+++ b/model-00009-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:76054b1c2961dbedaa49c1a038f69ac28e8b2bc629cc79448c777e2a9e661d86
+size 4631063736
diff --git a/model-00010-of-00086.safetensors b/model-00010-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..32cea7401b3684336a27660e63c1616551f1a3f6
--- /dev/null
+++ b/model-00010-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d4fff42469386a5c03cb4324193336118fc4405ef9ed053e76b026d30cc5bed0
+size 4933097752
diff --git a/model-00011-of-00086.safetensors b/model-00011-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d166872fe41aa8863c329dbe0afd09766a0b66a2
--- /dev/null
+++ b/model-00011-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f24f40b1b500d55d0a53a7e4f410ce0e89f6176df899528d7187e87278818433
+size 4631063736
diff --git a/model-00012-of-00086.safetensors b/model-00012-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..302a17815bb13b73d3b39f4ecd70899a4a71406c
--- /dev/null
+++ b/model-00012-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:df06d654f6874a19d9c2faf618ad5bc5b22a6cd140b8f5a1a327b55c39e876af
+size 4933097752
diff --git a/model-00013-of-00086.safetensors b/model-00013-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9c6708b7472f27344a9bb514af0a99bcf6299508
--- /dev/null
+++ b/model-00013-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:7c87355df220611cf6903aed181128eb8ba1151ff28d72986dde0c24fc5508ac
+size 4631063736
diff --git a/model-00014-of-00086.safetensors b/model-00014-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6aa0ddb7baae38b2291dbbd9d8b6a0841fd8b86c
--- /dev/null
+++ b/model-00014-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e07badb07b16988ce69b1f71e52457b83b87b491908355c2f853c7f3ea520358
+size 4933097752
diff --git a/model-00015-of-00086.safetensors b/model-00015-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..94f740093e78ff9d78137c90895232da3b6ed070
--- /dev/null
+++ b/model-00015-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:74b3fd80b4811e7ffa1ecced878be34ff56b688f9ce743b4149c19e8defc8c6b
+size 4631063736
diff --git a/model-00016-of-00086.safetensors b/model-00016-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..fb41b26211122c7156e65d8ed08001be8bfcbd41
--- /dev/null
+++ b/model-00016-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b47dbea81b4d4e88fdf49d5060dd62b2c3b9fb3a111263725a2533c64d2b1448
+size 4933097752
diff --git a/model-00017-of-00086.safetensors b/model-00017-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..50aa9bd324a14e51f2d9cbce4f9a6ae2f0803f3b
--- /dev/null
+++ b/model-00017-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1c7d3b46edd24e829a7c406c38630c55a8f3162d6bba5e968e9bd94826d713f7
+size 4631063736
diff --git a/model-00018-of-00086.safetensors b/model-00018-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7adcbf7dcb03ed8059c4b5d892d594fd9a9dc76b
--- /dev/null
+++ b/model-00018-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4d61124a6ef7200594577717921390ceab63dd12510e7c3953c96c7b995583bd
+size 4933097752
diff --git a/model-00019-of-00086.safetensors b/model-00019-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..74ff78dfe6691570d3d55adef0e6bc6e436ac696
--- /dev/null
+++ b/model-00019-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:44c65cccc511fc28a24f7e2a5ef9aa3341a625421702986f68e4e339c41958ed
+size 4631063736
diff --git a/model-00020-of-00086.safetensors b/model-00020-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f7083f8975ac92161fc92021b08d9650beac4789
--- /dev/null
+++ b/model-00020-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d473412191dd48f31b672def9b42fb66913e2d1944e40bea50e382dc612c9be4
+size 4933097752
diff --git a/model-00021-of-00086.safetensors b/model-00021-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3c6218bc02959f7d23d1eb32e08fb1d814f54dcd
--- /dev/null
+++ b/model-00021-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5bec8f8d1eb0c8a1ad077d96056084123637ca99f381a4de2955ade19987628c
+size 4631063736
diff --git a/model-00022-of-00086.safetensors b/model-00022-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4b7341885711bce46e7c0e8469c4cb9aeec500bc
--- /dev/null
+++ b/model-00022-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:98ec3444c030a09eddad91250f4f60773b6665463d96b223d3b4a8448108e159
+size 4933097752
diff --git a/model-00023-of-00086.safetensors b/model-00023-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..99c95c5d755e5d5b34bc81231e13dd55d2d692eb
--- /dev/null
+++ b/model-00023-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:49b4ace14cf348d17c331a333d4cd411427286d2b89677307d080e6b1aeae9d7
+size 4631063736
diff --git a/model-00024-of-00086.safetensors b/model-00024-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b798817f422a37f08cdbb86342f514581dbd8507
--- /dev/null
+++ b/model-00024-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:65b55cb6fb2a4131b34eff9a7a601227c9c46c7294b7662763b09cbfb26555f2
+size 4933097752
diff --git a/model-00025-of-00086.safetensors b/model-00025-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..3185e968b28a208bcba36afee51c9775f38364cf
--- /dev/null
+++ b/model-00025-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e4179f91f6049039212439e26dc9f6ffb7c802971264f0199a9a92d227bc2dc5
+size 4631063736
diff --git a/model-00026-of-00086.safetensors b/model-00026-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a27397b90cbb12c7547e0ac6510d461b0832c3fa
--- /dev/null
+++ b/model-00026-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:953b79fadda926a2d9513f50120d08231e0648a1cb3fef6e57f03826606fc5ff
+size 4933097752
diff --git a/model-00027-of-00086.safetensors b/model-00027-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7aef05a3d2fd69122db596fb997e5ee25b64a763
--- /dev/null
+++ b/model-00027-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:05f8dad29541ed1b4a52405d36b6d85c2b931d6a62c8bff16145b2d602408a88
+size 4631063736
diff --git a/model-00028-of-00086.safetensors b/model-00028-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f2598d498807a4c21ad170f360cb17bd470469a7
--- /dev/null
+++ b/model-00028-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fd15baba42998452d32ec8abd853e5a19ad823f24ede44f3abb14f88966cdf4a
+size 4933097752
diff --git a/model-00029-of-00086.safetensors b/model-00029-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ba0819ae6c9363b3872d607f58b3ecf6ec298e75
--- /dev/null
+++ b/model-00029-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d409c76878b76a24ccb603db51816dbfcac79cffe8680bf85cac6de1465a8810
+size 4631063736
diff --git a/model-00030-of-00086.safetensors b/model-00030-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6d2680df39d002362ee459862d3b5b6984e7acdd
--- /dev/null
+++ b/model-00030-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:174e4e1b53195a079d3027551d0d1ccc4fffcc427e774b363c5824e933c0fc5c
+size 4933097752
diff --git a/model-00031-of-00086.safetensors b/model-00031-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..28273da8f13c3f02da4186a9aaf46bb00f2e2d5a
--- /dev/null
+++ b/model-00031-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:56c3dbb5b991088a6677a2d9784a696ab50baf0ac44001aeda002a32ba5b773c
+size 4631063736
diff --git a/model-00032-of-00086.safetensors b/model-00032-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..988d031c5e445e00234ab14ad6656cbc9c8ac063
--- /dev/null
+++ b/model-00032-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bdacf7d13d450c9e2178fff233263267907d68545761fd89da0e55860e6e440e
+size 4933097752
diff --git a/model-00033-of-00086.safetensors b/model-00033-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..16bc184904dc62811d18f18360ab8fa6e636b5a5
--- /dev/null
+++ b/model-00033-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1d04d3fe5b4f181913690de0ceb51e51109d015e9de7292698c2656c44d38280
+size 4631063736
diff --git a/model-00034-of-00086.safetensors b/model-00034-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ddc8b30fe9803373571cfe5fd584cc0b2eaf71b0
--- /dev/null
+++ b/model-00034-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:065ed293c2d3946ed3b3b68c770287be77da656d19e89c9c6fccb17dffc08739
+size 4933097752
diff --git a/model-00035-of-00086.safetensors b/model-00035-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e3e4ae169f89014b366b40a43d36657fabfce4c7
--- /dev/null
+++ b/model-00035-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1c51359aabec1abb505c731e22b5569d27570f0dfe457bae5e9eeb5f6ef4f5fc
+size 4631063736
diff --git a/model-00036-of-00086.safetensors b/model-00036-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..26c5b96c7f073a64e83ab36f1c946f0c1602c730
--- /dev/null
+++ b/model-00036-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:cf7da8746b23d84c74872b006461196987c4440a8a3a30ad6dfb81dc5fc56629
+size 4933097752
diff --git a/model-00037-of-00086.safetensors b/model-00037-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..410f58db5ae4e5d411bb871eaa50e3f396d28d0b
--- /dev/null
+++ b/model-00037-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a29b2447dc134aaf805a0d2851ac5b0aa4dd31e6f954132d77ae46729bfe1d1f
+size 4631063736
diff --git a/model-00038-of-00086.safetensors b/model-00038-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a7e684bad9b57c806975f42de37d337f010fc100
--- /dev/null
+++ b/model-00038-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4a95dea482ec0381ac8741cdefab1f5176d0df2a42c993048fea89c90d0e0de4
+size 4933097752
diff --git a/model-00039-of-00086.safetensors b/model-00039-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..ddbc6c17add40107fe4a6ee3b5d642f5377f8e77
--- /dev/null
+++ b/model-00039-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5b4d782358e8423b27c1bcd4db4d418354e01f85f12337acb7f9e3d44ca06be9
+size 4631063736
diff --git a/model-00040-of-00086.safetensors b/model-00040-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8ef79dd56d63cbb2cd447f91ed56b5b33802770f
--- /dev/null
+++ b/model-00040-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fe4d14146857e892daf49ac27526662da3d2456f42c7871c7df370b9f642d586
+size 4933097752
diff --git a/model-00041-of-00086.safetensors b/model-00041-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..dac490044bc962a9dfd1f86b78f29eee9696cabd
--- /dev/null
+++ b/model-00041-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:196efd5cd073b208a485c8e6ca55404c9babb5a5566a5c7f214f0b077f6bad14
+size 4631063736
diff --git a/model-00042-of-00086.safetensors b/model-00042-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..862e172cb0cc412eb558400f664394d0a5599f59
--- /dev/null
+++ b/model-00042-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e5e4b029cf898cf5deb1cebee16e3feeed32743363d0a6ce9433c70199b19526
+size 4933097752
diff --git a/model-00043-of-00086.safetensors b/model-00043-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..48a8cb5130a86fbf14a2376e2f4e8380828f6743
--- /dev/null
+++ b/model-00043-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:14b713aa6d24a06034e882832140173a0481375aa5579aa68c5f3e44ff2dc93f
+size 4631063736
diff --git a/model-00044-of-00086.safetensors b/model-00044-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c4662a913f252d59bf58437a0ba93c9b94ab0f23
--- /dev/null
+++ b/model-00044-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d9aa1f68b4bf4d1adc7ec1f1cda4b837ce13cf7a22b55882931c58814e94c889
+size 4933097752
diff --git a/model-00045-of-00086.safetensors b/model-00045-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..7e8f33337a6862c6cabbe8d6a1698c4b89451f6a
--- /dev/null
+++ b/model-00045-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fe0e00ab458d15edc9c6129dc828146a4be994839808abf430821505b29c467e
+size 4631063736
diff --git a/model-00046-of-00086.safetensors b/model-00046-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..08a6a844b7a6691bcbe05859919faf728ad75529
--- /dev/null
+++ b/model-00046-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1c1d6a13c700f19bdf7b19681bb5ea4e591cc8ded3d53d515e84799d46d70206
+size 4933097752
diff --git a/model-00047-of-00086.safetensors b/model-00047-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..1235d6d81021cbb539e51d367d21826e589b7a4a
--- /dev/null
+++ b/model-00047-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0a87f542f5ded70b766bedc55ad0a3fa39c10e5f6da5708ae0e5dd87809b8640
+size 4631063736
diff --git a/model-00048-of-00086.safetensors b/model-00048-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..124b0a0d4ff5c9c9a7983f85bc2e8ffc5540d031
--- /dev/null
+++ b/model-00048-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4077fbfc99c9bb418f6735a8e69dcf10824959eab63181f9ca32f2fa56b02b86
+size 4933097752
diff --git a/model-00049-of-00086.safetensors b/model-00049-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d0a12c95a43cb6b3e051624a22646e0248b9e820
--- /dev/null
+++ b/model-00049-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8b0d4cf700a71bb42be0c5d408ab5437a8a79622e0786baad89748d31b36a186
+size 4631063736
diff --git a/model-00050-of-00086.safetensors b/model-00050-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..099b4cb7db8670cfc846ebf74c49003e9071f0b9
--- /dev/null
+++ b/model-00050-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:646d1b28f0e376dce06392fc193154afb097c1c9e453d2975650a93aa971e234
+size 4933097752
diff --git a/model-00051-of-00086.safetensors b/model-00051-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c0f94a984fb421e7ec64146c386a7dd6b3f3fd93
--- /dev/null
+++ b/model-00051-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:940bc442d582130347509589138eeed49c1d6d08dd56761f682fb206bde8f679
+size 4631063736
diff --git a/model-00052-of-00086.safetensors b/model-00052-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f7610a00058bdd6b14c7ff9817448a30294f8eeb
--- /dev/null
+++ b/model-00052-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2238147f3fada4d97316302cf3a17e7e01b0a7a3eec4648b0e2eb70c3d1a4b61
+size 4933097752
diff --git a/model-00053-of-00086.safetensors b/model-00053-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..8463d8c22e9a1e3e7927a6cc5687cec299d54171
--- /dev/null
+++ b/model-00053-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ec6cd4a9bec8d5b5451a3592e56e987617a637835be7d9dd69e24a7991abb28f
+size 4631063736
diff --git a/model-00054-of-00086.safetensors b/model-00054-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..9cdba5867833e100e0dfaf10122689e0d0587e17
--- /dev/null
+++ b/model-00054-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b2f58759b1a9936d9cbaee30d5d2daf03f1a915d95e2f826774f2ade362c0260
+size 4933097752
diff --git a/model-00055-of-00086.safetensors b/model-00055-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..41427673b4590adb608b168c8fa3f8ffc192081b
--- /dev/null
+++ b/model-00055-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6a6e5082d081f821dfec9b795a0e190d184ed4e5b878fb833e099a019bb38839
+size 4631063736
diff --git a/model-00056-of-00086.safetensors b/model-00056-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..27167e5b921e71fa99f105713299fe46162d2f02
--- /dev/null
+++ b/model-00056-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f9d97e59da1e8f84b94e0b9feb63328364b00b3bdf5c7c08916d9a0203cb4a10
+size 4933097752
diff --git a/model-00057-of-00086.safetensors b/model-00057-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..fb9679c1becd956b6eb0c9ab8075a56bab85ec6f
--- /dev/null
+++ b/model-00057-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4acbd90361b025d0749648013af93c1850c6941b19fda1a9b4bb83654c8434ef
+size 4631063736
diff --git a/model-00058-of-00086.safetensors b/model-00058-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..fdab6c704b52f433a8a2839eb83395a14ee00e19
--- /dev/null
+++ b/model-00058-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ba3ae1366e8eabb4d0e3d904c24875b112a59d2a65618533d59a6ddc6cc656d7
+size 4933097752
diff --git a/model-00059-of-00086.safetensors b/model-00059-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..d3bc086588af8f768025ec5ad463926991cc4a64
--- /dev/null
+++ b/model-00059-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ae94898a9e40fb6b01d01dc71ce2b73f53c83bdd2eec8474281d265fe867a959
+size 4631063736
diff --git a/model-00060-of-00086.safetensors b/model-00060-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6f73bc871a4b43b5a9ddcd167c39d1fe2bfca99e
--- /dev/null
+++ b/model-00060-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a5ae72a7bba4fedd980742ce6da5cfb3e171407a91086d73a0ba7400a6c4a158
+size 4933097752
diff --git a/model-00061-of-00086.safetensors b/model-00061-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..dd72de3e5acf959c0baa423b3fa7d865107e9f5a
--- /dev/null
+++ b/model-00061-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:471354f8febf78eb7d5efaf3104e4e425af84dd9733a7b034e42c9acd765abd2
+size 4631063736
diff --git a/model-00062-of-00086.safetensors b/model-00062-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e76bf325a7d4e7f1806cafa846ddca4297b3f752
--- /dev/null
+++ b/model-00062-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:58485c51c87eaf18aea1651cdf54924651da7bea80038103a7f52a0f98da784e
+size 4933097752
diff --git a/model-00063-of-00086.safetensors b/model-00063-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..99ff0a533fc6ae93eca8f24d0706e8408aaf053e
--- /dev/null
+++ b/model-00063-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b8a146f4ad1192cf9c30180b20c0a2eb63dbbfda4e08d97a182e8537e7569e5b
+size 4631063736
diff --git a/model-00064-of-00086.safetensors b/model-00064-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..be1622a78beaec02eb2ff51e8341d97f9f5b015b
--- /dev/null
+++ b/model-00064-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b88aea3e2782393062029f8f58aae5fbc4a7d46276205f56f200c540e01c2267
+size 4933097752
diff --git a/model-00065-of-00086.safetensors b/model-00065-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5e6e2a0b99f2f845b890637e161fe921f04542b6
--- /dev/null
+++ b/model-00065-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:8973c3a6f0641b469409a9f6bf0bced21436d2c6eaa22a6f77e6a292af90970e
+size 4631063736
diff --git a/model-00066-of-00086.safetensors b/model-00066-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..53997f099ce76dde6452aa31df2af5ced37a2b51
--- /dev/null
+++ b/model-00066-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fe6595656e4c2040e7cf9aacc7feae7547c75af5cc0d53d96b4645dceaecb83d
+size 4933097752
diff --git a/model-00067-of-00086.safetensors b/model-00067-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c6eff1f632d3b2d5608eba05543674d81f2e6119
--- /dev/null
+++ b/model-00067-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:47c1ec025092d0f3e36c8be50a9f81760bde0022924c5f6b4e5eadc4be2af8ed
+size 4631063736
diff --git a/model-00068-of-00086.safetensors b/model-00068-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..440ec22e8ae7ef79aaa0f2e8ef4f283844a4c2c7
--- /dev/null
+++ b/model-00068-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:902430cbe51e2bddd65d61817d17e27137eca7d1041e1d5bec25ecfc224b3137
+size 4933097768
diff --git a/model-00069-of-00086.safetensors b/model-00069-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..eeb74563fcfd41f86b1c1a8f10e09d7f7900c5ed
--- /dev/null
+++ b/model-00069-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:dd5c0828f7b79d6df1e8a2f4922441cfa405324618b8caee0dc7d1fa9b7bf07c
+size 4631063768
diff --git a/model-00070-of-00086.safetensors b/model-00070-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..2c4a090bd96e9f4b3c27e4f14c121e7769db6df2
--- /dev/null
+++ b/model-00070-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6355989941bffa38efad0397aaa808d0bf881e5892c01fb9dc87e2d78b24d31a
+size 4933097768
diff --git a/model-00071-of-00086.safetensors b/model-00071-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..854f16291d54eaf480c26c7f3e20b7823692391e
--- /dev/null
+++ b/model-00071-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:bf614a96a0c3021785b3bb11a4a5c1792013ca0abab51a2215b2d037d7d5b8cb
+size 4631063768
diff --git a/model-00072-of-00086.safetensors b/model-00072-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5eb4ee43df7a1a5d80934e3e6161bf175bab6a2b
--- /dev/null
+++ b/model-00072-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0fcba875964647df7ddc18050d4daeb170c12feefa59d1c617c9b17f28d6e9ce
+size 4933097768
diff --git a/model-00073-of-00086.safetensors b/model-00073-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0678cac11b605a050baf31947a8ec5c4dd933417
--- /dev/null
+++ b/model-00073-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:772798c30a73dbe9d6d4f372f4dc2abfba591a3764f73f453c27d598cfcb3981
+size 4631063768
diff --git a/model-00074-of-00086.safetensors b/model-00074-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..44bc4b403599b355afdd013d9e630414c7b3437a
--- /dev/null
+++ b/model-00074-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:efecc6c4795984d62e504ccc5846744f78dcd1f71a578e1af343e6bd98c173db
+size 4933097768
diff --git a/model-00075-of-00086.safetensors b/model-00075-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..c3b11592994ed27cfe2a2e4fe0a72cee6952ec0a
--- /dev/null
+++ b/model-00075-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ac094fce6c1e52357345439970b2ccfcda90dd882a1932c590c10b99e3932c89
+size 4631063768
diff --git a/model-00076-of-00086.safetensors b/model-00076-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..e319048329f3238ced91aef8dfdf5934c7b97266
--- /dev/null
+++ b/model-00076-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:963c485a77b2257fec720b86952de7003790f222af5b446375ff80ad4f66ac5a
+size 4933097768
diff --git a/model-00077-of-00086.safetensors b/model-00077-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..35bef991efd47d8a757f5ed6c64fe0ee971be153
--- /dev/null
+++ b/model-00077-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3d0bf9886cffe03b93374cf14bd335cb5b5c0e084702e098f7fd4b8fa7f08a7a
+size 4631063768
diff --git a/model-00078-of-00086.safetensors b/model-00078-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..de5200e5cdfae24225f6e581213ad86000ae385b
--- /dev/null
+++ b/model-00078-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c854132122a0878827d5213e79530203bb7b130d014bbafcccbca65bc075300b
+size 4933097768
diff --git a/model-00079-of-00086.safetensors b/model-00079-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..0e36b4e10f98add2c8a7285f98abfdb4faf54dd0
--- /dev/null
+++ b/model-00079-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:2c1acf1b65b800a8b9d285b9506f667d1ad2a4425c39065e3bde0a2f6b0f9f49
+size 4631063768
diff --git a/model-00080-of-00086.safetensors b/model-00080-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..521e850fb5d43187e908daee374e8bfe34a4cadf
--- /dev/null
+++ b/model-00080-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:4f723b75728db0575e7649a26bf3a07c792b87c46292f00e6fc889305098f363
+size 4933097768
diff --git a/model-00081-of-00086.safetensors b/model-00081-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a807a31deed8a2b375d4543c8e3795c8e29fe251
--- /dev/null
+++ b/model-00081-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ef633508d2a513d4c859b4b7f8fe1642d2fd58f094d46c0be71cfe397d319940
+size 4631063768
diff --git a/model-00082-of-00086.safetensors b/model-00082-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..a5acf503deebe90a5636a7e696d5cfed4c04d13e
--- /dev/null
+++ b/model-00082-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e44d7653980d5a7d0843fecbbe7444f5164817a41dfa6ae4f64e22a3a6007948
+size 4933097768
diff --git a/model-00083-of-00086.safetensors b/model-00083-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..6e7e5f81d23deb5815e803d9a6acda0a2c6e4f3f
--- /dev/null
+++ b/model-00083-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9c15cfe3f6b9037156bf7ce8ced6d351c7295d1f82be709c0d691d016e861883
+size 4631063768
diff --git a/model-00084-of-00086.safetensors b/model-00084-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..10c98ed835b092088dd07e8190ee5585d74d5487
--- /dev/null
+++ b/model-00084-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c0267577bf3b9f6b3c014bc27ca8ed8dd786cb9679017505e7f282413ae44162
+size 4933097768
diff --git a/model-00085-of-00086.safetensors b/model-00085-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4f3bde2d7a966122d79b4b58301cba2a66eec488
--- /dev/null
+++ b/model-00085-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f4acb781b39cc5e0bf41f9d9ff171b9166305511188f8b3c9858021b1af37c63
+size 4060600688
diff --git a/model-00086-of-00086.safetensors b/model-00086-of-00086.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..4321e3c72946a45e5f8f4356e4495231810365aa
--- /dev/null
+++ b/model-00086-of-00086.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:d210b97557f8416dbd3c0709f05c4bf120a8f1bf98f2f380ab73f2d5d6ec41c1
+size 4202692736
diff --git a/model.safetensors.index.json b/model.safetensors.index.json
new file mode 100644
index 0000000000000000000000000000000000000000..861a38388765478cc85524fef32cf0999e39c778
--- /dev/null
+++ b/model.safetensors.index.json
@@ -0,0 +1,2026 @@
+{
+ "metadata": {
+ "total_size": 410099965952
+ },
+ "weight_map": {
+ "lm_head.weight": "model-00086-of-00086.safetensors",
+ "model.embed_tokens.weight": "model-00001-of-00086.safetensors",
+ "model.layers.0.input_layernorm.weight": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.down_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.down_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.gate_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.up_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.0.mlp.up_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00086.safetensors",
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.k_proj.weight_scale": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.o_proj.weight_scale": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.q_proj.weight_scale": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00086.safetensors",
+ "model.layers.0.self_attn.v_proj.weight_scale": "model-00001-of-00086.safetensors",
+ "model.layers.1.input_layernorm.weight": "model-00003-of-00086.safetensors",
+ "model.layers.1.mlp.down_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.1.mlp.down_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.mlp.gate_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.1.mlp.up_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.mlp.up_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00086.safetensors",
+ "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.k_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.o_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.q_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00086.safetensors",
+ "model.layers.1.self_attn.v_proj.weight_scale": "model-00002-of-00086.safetensors",
+ "model.layers.10.input_layernorm.weight": "model-00009-of-00086.safetensors",
+ "model.layers.10.mlp.down_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.10.mlp.down_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.mlp.gate_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.10.mlp.up_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.mlp.up_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00086.safetensors",
+ "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.k_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.o_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.q_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.10.self_attn.v_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.100.input_layernorm.weight": "model-00069-of-00086.safetensors",
+ "model.layers.100.mlp.down_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.100.mlp.down_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.100.mlp.gate_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.mlp.gate_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.100.mlp.up_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.mlp.up_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.100.post_attention_layernorm.weight": "model-00069-of-00086.safetensors",
+ "model.layers.100.self_attn.k_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.k_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.o_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.o_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.q_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.q_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.v_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.100.self_attn.v_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.101.input_layernorm.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.down_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.down_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.gate_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.gate_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.up_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.mlp.up_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.post_attention_layernorm.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.k_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.k_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.o_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.o_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.q_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.q_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.v_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.101.self_attn.v_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.102.input_layernorm.weight": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.down_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.down_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.gate_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.gate_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.up_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.102.mlp.up_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.102.post_attention_layernorm.weight": "model-00070-of-00086.safetensors",
+ "model.layers.102.self_attn.k_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.k_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.o_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.o_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.q_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.q_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.v_proj.weight": "model-00069-of-00086.safetensors",
+ "model.layers.102.self_attn.v_proj.weight_scale": "model-00069-of-00086.safetensors",
+ "model.layers.103.input_layernorm.weight": "model-00071-of-00086.safetensors",
+ "model.layers.103.mlp.down_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.103.mlp.down_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.103.mlp.gate_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.mlp.gate_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.103.mlp.up_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.mlp.up_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.103.post_attention_layernorm.weight": "model-00071-of-00086.safetensors",
+ "model.layers.103.self_attn.k_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.k_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.o_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.o_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.q_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.q_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.v_proj.weight": "model-00070-of-00086.safetensors",
+ "model.layers.103.self_attn.v_proj.weight_scale": "model-00070-of-00086.safetensors",
+ "model.layers.104.input_layernorm.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.down_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.down_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.gate_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.gate_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.up_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.mlp.up_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.post_attention_layernorm.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.k_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.k_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.o_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.o_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.q_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.q_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.v_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.104.self_attn.v_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.105.input_layernorm.weight": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.down_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.down_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.gate_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.gate_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.up_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.105.mlp.up_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.105.post_attention_layernorm.weight": "model-00072-of-00086.safetensors",
+ "model.layers.105.self_attn.k_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.k_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.o_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.o_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.q_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.q_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.v_proj.weight": "model-00071-of-00086.safetensors",
+ "model.layers.105.self_attn.v_proj.weight_scale": "model-00071-of-00086.safetensors",
+ "model.layers.106.input_layernorm.weight": "model-00073-of-00086.safetensors",
+ "model.layers.106.mlp.down_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.106.mlp.down_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.106.mlp.gate_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.mlp.gate_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.106.mlp.up_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.mlp.up_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.106.post_attention_layernorm.weight": "model-00073-of-00086.safetensors",
+ "model.layers.106.self_attn.k_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.k_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.o_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.o_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.q_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.q_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.v_proj.weight": "model-00072-of-00086.safetensors",
+ "model.layers.106.self_attn.v_proj.weight_scale": "model-00072-of-00086.safetensors",
+ "model.layers.107.input_layernorm.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.down_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.down_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.gate_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.gate_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.up_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.mlp.up_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.post_attention_layernorm.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.k_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.k_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.o_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.o_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.q_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.q_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.v_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.107.self_attn.v_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.108.input_layernorm.weight": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.down_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.down_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.gate_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.gate_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.up_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.108.mlp.up_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.108.post_attention_layernorm.weight": "model-00074-of-00086.safetensors",
+ "model.layers.108.self_attn.k_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.k_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.o_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.o_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.q_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.q_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.v_proj.weight": "model-00073-of-00086.safetensors",
+ "model.layers.108.self_attn.v_proj.weight_scale": "model-00073-of-00086.safetensors",
+ "model.layers.109.input_layernorm.weight": "model-00075-of-00086.safetensors",
+ "model.layers.109.mlp.down_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.109.mlp.down_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.109.mlp.gate_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.mlp.gate_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.109.mlp.up_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.mlp.up_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.109.post_attention_layernorm.weight": "model-00075-of-00086.safetensors",
+ "model.layers.109.self_attn.k_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.k_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.o_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.o_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.q_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.q_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.v_proj.weight": "model-00074-of-00086.safetensors",
+ "model.layers.109.self_attn.v_proj.weight_scale": "model-00074-of-00086.safetensors",
+ "model.layers.11.input_layernorm.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.down_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.down_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.gate_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.up_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.mlp.up_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.post_attention_layernorm.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.k_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.o_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.q_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.11.self_attn.v_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.110.input_layernorm.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.down_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.down_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.gate_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.gate_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.up_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.mlp.up_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.post_attention_layernorm.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.k_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.k_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.o_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.o_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.q_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.q_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.v_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.110.self_attn.v_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.111.input_layernorm.weight": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.down_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.down_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.gate_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.gate_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.up_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.111.mlp.up_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.111.post_attention_layernorm.weight": "model-00076-of-00086.safetensors",
+ "model.layers.111.self_attn.k_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.k_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.o_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.o_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.q_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.q_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.v_proj.weight": "model-00075-of-00086.safetensors",
+ "model.layers.111.self_attn.v_proj.weight_scale": "model-00075-of-00086.safetensors",
+ "model.layers.112.input_layernorm.weight": "model-00077-of-00086.safetensors",
+ "model.layers.112.mlp.down_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.112.mlp.down_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.112.mlp.gate_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.mlp.gate_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.112.mlp.up_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.mlp.up_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.112.post_attention_layernorm.weight": "model-00077-of-00086.safetensors",
+ "model.layers.112.self_attn.k_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.k_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.o_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.o_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.q_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.q_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.v_proj.weight": "model-00076-of-00086.safetensors",
+ "model.layers.112.self_attn.v_proj.weight_scale": "model-00076-of-00086.safetensors",
+ "model.layers.113.input_layernorm.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.down_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.down_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.gate_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.gate_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.up_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.mlp.up_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.post_attention_layernorm.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.k_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.k_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.o_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.o_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.q_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.q_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.v_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.113.self_attn.v_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.114.input_layernorm.weight": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.down_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.down_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.gate_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.gate_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.up_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.114.mlp.up_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.114.post_attention_layernorm.weight": "model-00078-of-00086.safetensors",
+ "model.layers.114.self_attn.k_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.k_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.o_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.o_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.q_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.q_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.v_proj.weight": "model-00077-of-00086.safetensors",
+ "model.layers.114.self_attn.v_proj.weight_scale": "model-00077-of-00086.safetensors",
+ "model.layers.115.input_layernorm.weight": "model-00079-of-00086.safetensors",
+ "model.layers.115.mlp.down_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.115.mlp.down_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.115.mlp.gate_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.mlp.gate_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.115.mlp.up_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.mlp.up_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.115.post_attention_layernorm.weight": "model-00079-of-00086.safetensors",
+ "model.layers.115.self_attn.k_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.k_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.o_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.o_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.q_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.q_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.v_proj.weight": "model-00078-of-00086.safetensors",
+ "model.layers.115.self_attn.v_proj.weight_scale": "model-00078-of-00086.safetensors",
+ "model.layers.116.input_layernorm.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.down_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.down_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.gate_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.gate_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.up_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.mlp.up_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.post_attention_layernorm.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.k_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.k_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.o_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.o_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.q_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.q_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.v_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.116.self_attn.v_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.117.input_layernorm.weight": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.down_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.down_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.gate_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.gate_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.up_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.117.mlp.up_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.117.post_attention_layernorm.weight": "model-00080-of-00086.safetensors",
+ "model.layers.117.self_attn.k_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.k_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.o_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.o_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.q_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.q_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.v_proj.weight": "model-00079-of-00086.safetensors",
+ "model.layers.117.self_attn.v_proj.weight_scale": "model-00079-of-00086.safetensors",
+ "model.layers.118.input_layernorm.weight": "model-00081-of-00086.safetensors",
+ "model.layers.118.mlp.down_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.118.mlp.down_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.118.mlp.gate_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.mlp.gate_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.118.mlp.up_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.mlp.up_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.118.post_attention_layernorm.weight": "model-00081-of-00086.safetensors",
+ "model.layers.118.self_attn.k_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.k_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.o_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.o_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.q_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.q_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.v_proj.weight": "model-00080-of-00086.safetensors",
+ "model.layers.118.self_attn.v_proj.weight_scale": "model-00080-of-00086.safetensors",
+ "model.layers.119.input_layernorm.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.down_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.down_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.gate_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.gate_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.up_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.mlp.up_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.post_attention_layernorm.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.k_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.k_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.o_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.o_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.q_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.q_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.v_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.119.self_attn.v_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.12.input_layernorm.weight": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.down_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.down_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.gate_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.up_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.12.mlp.up_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00086.safetensors",
+ "model.layers.12.self_attn.k_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.k_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.o_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.o_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.q_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.q_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.v_proj.weight": "model-00009-of-00086.safetensors",
+ "model.layers.12.self_attn.v_proj.weight_scale": "model-00009-of-00086.safetensors",
+ "model.layers.120.input_layernorm.weight": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.down_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.down_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.gate_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.gate_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.up_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.120.mlp.up_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.120.post_attention_layernorm.weight": "model-00082-of-00086.safetensors",
+ "model.layers.120.self_attn.k_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.k_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.o_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.o_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.q_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.q_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.v_proj.weight": "model-00081-of-00086.safetensors",
+ "model.layers.120.self_attn.v_proj.weight_scale": "model-00081-of-00086.safetensors",
+ "model.layers.121.input_layernorm.weight": "model-00083-of-00086.safetensors",
+ "model.layers.121.mlp.down_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.121.mlp.down_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.121.mlp.gate_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.mlp.gate_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.121.mlp.up_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.mlp.up_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.121.post_attention_layernorm.weight": "model-00083-of-00086.safetensors",
+ "model.layers.121.self_attn.k_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.k_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.o_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.o_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.q_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.q_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.v_proj.weight": "model-00082-of-00086.safetensors",
+ "model.layers.121.self_attn.v_proj.weight_scale": "model-00082-of-00086.safetensors",
+ "model.layers.122.input_layernorm.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.down_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.down_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.gate_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.gate_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.up_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.mlp.up_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.post_attention_layernorm.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.k_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.k_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.o_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.o_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.q_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.q_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.v_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.122.self_attn.v_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.123.input_layernorm.weight": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.down_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.down_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.gate_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.gate_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.up_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.123.mlp.up_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.123.post_attention_layernorm.weight": "model-00084-of-00086.safetensors",
+ "model.layers.123.self_attn.k_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.k_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.o_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.o_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.q_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.q_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.v_proj.weight": "model-00083-of-00086.safetensors",
+ "model.layers.123.self_attn.v_proj.weight_scale": "model-00083-of-00086.safetensors",
+ "model.layers.124.input_layernorm.weight": "model-00085-of-00086.safetensors",
+ "model.layers.124.mlp.down_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.124.mlp.down_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.124.mlp.gate_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.mlp.gate_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.124.mlp.up_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.mlp.up_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.124.post_attention_layernorm.weight": "model-00085-of-00086.safetensors",
+ "model.layers.124.self_attn.k_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.k_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.o_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.o_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.q_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.q_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.v_proj.weight": "model-00084-of-00086.safetensors",
+ "model.layers.124.self_attn.v_proj.weight_scale": "model-00084-of-00086.safetensors",
+ "model.layers.125.input_layernorm.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.down_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.down_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.gate_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.gate_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.up_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.mlp.up_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.post_attention_layernorm.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.k_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.k_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.o_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.o_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.q_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.q_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.v_proj.weight": "model-00085-of-00086.safetensors",
+ "model.layers.125.self_attn.v_proj.weight_scale": "model-00085-of-00086.safetensors",
+ "model.layers.13.input_layernorm.weight": "model-00011-of-00086.safetensors",
+ "model.layers.13.mlp.down_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.13.mlp.down_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.13.mlp.gate_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.mlp.gate_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.13.mlp.up_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.mlp.up_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00086.safetensors",
+ "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.k_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.o_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.q_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00086.safetensors",
+ "model.layers.13.self_attn.v_proj.weight_scale": "model-00010-of-00086.safetensors",
+ "model.layers.14.input_layernorm.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.down_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.down_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.gate_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.up_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.mlp.up_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.post_attention_layernorm.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.k_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.o_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.q_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.14.self_attn.v_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.15.input_layernorm.weight": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.down_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.down_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.gate_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.up_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.15.mlp.up_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.15.post_attention_layernorm.weight": "model-00012-of-00086.safetensors",
+ "model.layers.15.self_attn.k_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.k_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.o_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.o_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.q_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.q_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.v_proj.weight": "model-00011-of-00086.safetensors",
+ "model.layers.15.self_attn.v_proj.weight_scale": "model-00011-of-00086.safetensors",
+ "model.layers.16.input_layernorm.weight": "model-00013-of-00086.safetensors",
+ "model.layers.16.mlp.down_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.16.mlp.down_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.16.mlp.gate_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.mlp.gate_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.16.mlp.up_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.mlp.up_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00086.safetensors",
+ "model.layers.16.self_attn.k_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.k_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.o_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.o_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.q_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.q_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.v_proj.weight": "model-00012-of-00086.safetensors",
+ "model.layers.16.self_attn.v_proj.weight_scale": "model-00012-of-00086.safetensors",
+ "model.layers.17.input_layernorm.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.down_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.down_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.gate_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.gate_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.up_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.mlp.up_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.post_attention_layernorm.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.k_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.o_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.q_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.17.self_attn.v_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.18.input_layernorm.weight": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.down_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.down_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.gate_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.up_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.18.mlp.up_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.18.post_attention_layernorm.weight": "model-00014-of-00086.safetensors",
+ "model.layers.18.self_attn.k_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.k_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.o_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.o_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.q_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.q_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.v_proj.weight": "model-00013-of-00086.safetensors",
+ "model.layers.18.self_attn.v_proj.weight_scale": "model-00013-of-00086.safetensors",
+ "model.layers.19.input_layernorm.weight": "model-00015-of-00086.safetensors",
+ "model.layers.19.mlp.down_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.19.mlp.down_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.19.mlp.gate_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.mlp.gate_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.19.mlp.up_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.mlp.up_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.19.post_attention_layernorm.weight": "model-00015-of-00086.safetensors",
+ "model.layers.19.self_attn.k_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.k_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.o_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.o_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.q_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.q_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.v_proj.weight": "model-00014-of-00086.safetensors",
+ "model.layers.19.self_attn.v_proj.weight_scale": "model-00014-of-00086.safetensors",
+ "model.layers.2.input_layernorm.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.down_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.down_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.gate_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.up_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.mlp.up_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.k_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.o_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.q_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.2.self_attn.v_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.20.input_layernorm.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.down_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.down_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.gate_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.gate_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.up_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.mlp.up_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.post_attention_layernorm.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.k_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.k_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.o_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.o_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.q_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.q_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.v_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.20.self_attn.v_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.21.input_layernorm.weight": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.down_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.down_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.gate_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.gate_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.up_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.21.mlp.up_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.21.post_attention_layernorm.weight": "model-00016-of-00086.safetensors",
+ "model.layers.21.self_attn.k_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.k_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.o_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.o_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.q_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.q_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.v_proj.weight": "model-00015-of-00086.safetensors",
+ "model.layers.21.self_attn.v_proj.weight_scale": "model-00015-of-00086.safetensors",
+ "model.layers.22.input_layernorm.weight": "model-00017-of-00086.safetensors",
+ "model.layers.22.mlp.down_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.22.mlp.down_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.22.mlp.gate_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.mlp.gate_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.22.mlp.up_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.mlp.up_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.22.post_attention_layernorm.weight": "model-00017-of-00086.safetensors",
+ "model.layers.22.self_attn.k_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.k_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.o_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.o_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.q_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.q_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.v_proj.weight": "model-00016-of-00086.safetensors",
+ "model.layers.22.self_attn.v_proj.weight_scale": "model-00016-of-00086.safetensors",
+ "model.layers.23.input_layernorm.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.down_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.down_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.gate_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.gate_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.up_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.mlp.up_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.post_attention_layernorm.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.k_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.k_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.o_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.o_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.q_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.q_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.v_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.23.self_attn.v_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.24.input_layernorm.weight": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.down_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.down_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.gate_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.gate_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.up_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.24.mlp.up_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.24.post_attention_layernorm.weight": "model-00018-of-00086.safetensors",
+ "model.layers.24.self_attn.k_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.k_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.o_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.o_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.q_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.q_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.v_proj.weight": "model-00017-of-00086.safetensors",
+ "model.layers.24.self_attn.v_proj.weight_scale": "model-00017-of-00086.safetensors",
+ "model.layers.25.input_layernorm.weight": "model-00019-of-00086.safetensors",
+ "model.layers.25.mlp.down_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.25.mlp.down_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.25.mlp.gate_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.mlp.gate_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.25.mlp.up_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.mlp.up_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.25.post_attention_layernorm.weight": "model-00019-of-00086.safetensors",
+ "model.layers.25.self_attn.k_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.k_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.o_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.o_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.q_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.q_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.v_proj.weight": "model-00018-of-00086.safetensors",
+ "model.layers.25.self_attn.v_proj.weight_scale": "model-00018-of-00086.safetensors",
+ "model.layers.26.input_layernorm.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.down_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.down_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.gate_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.gate_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.up_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.mlp.up_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.post_attention_layernorm.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.k_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.k_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.o_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.o_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.q_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.q_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.v_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.26.self_attn.v_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.27.input_layernorm.weight": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.down_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.down_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.gate_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.gate_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.up_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.27.mlp.up_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.27.post_attention_layernorm.weight": "model-00020-of-00086.safetensors",
+ "model.layers.27.self_attn.k_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.k_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.o_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.o_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.q_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.q_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.v_proj.weight": "model-00019-of-00086.safetensors",
+ "model.layers.27.self_attn.v_proj.weight_scale": "model-00019-of-00086.safetensors",
+ "model.layers.28.input_layernorm.weight": "model-00021-of-00086.safetensors",
+ "model.layers.28.mlp.down_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.28.mlp.down_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.28.mlp.gate_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.mlp.gate_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.28.mlp.up_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.mlp.up_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.28.post_attention_layernorm.weight": "model-00021-of-00086.safetensors",
+ "model.layers.28.self_attn.k_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.k_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.o_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.o_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.q_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.q_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.v_proj.weight": "model-00020-of-00086.safetensors",
+ "model.layers.28.self_attn.v_proj.weight_scale": "model-00020-of-00086.safetensors",
+ "model.layers.29.input_layernorm.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.down_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.down_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.gate_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.gate_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.up_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.mlp.up_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.post_attention_layernorm.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.k_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.k_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.o_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.o_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.q_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.q_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.v_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.29.self_attn.v_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.3.input_layernorm.weight": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.down_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.down_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.gate_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.up_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.3.mlp.up_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00086.safetensors",
+ "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.k_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.o_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.q_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00086.safetensors",
+ "model.layers.3.self_attn.v_proj.weight_scale": "model-00003-of-00086.safetensors",
+ "model.layers.30.input_layernorm.weight": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.down_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.down_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.gate_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.gate_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.up_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.30.mlp.up_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.30.post_attention_layernorm.weight": "model-00022-of-00086.safetensors",
+ "model.layers.30.self_attn.k_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.k_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.o_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.o_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.q_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.q_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.v_proj.weight": "model-00021-of-00086.safetensors",
+ "model.layers.30.self_attn.v_proj.weight_scale": "model-00021-of-00086.safetensors",
+ "model.layers.31.input_layernorm.weight": "model-00023-of-00086.safetensors",
+ "model.layers.31.mlp.down_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.31.mlp.down_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.31.mlp.gate_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.mlp.gate_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.31.mlp.up_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.mlp.up_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.31.post_attention_layernorm.weight": "model-00023-of-00086.safetensors",
+ "model.layers.31.self_attn.k_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.k_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.o_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.o_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.q_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.q_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.v_proj.weight": "model-00022-of-00086.safetensors",
+ "model.layers.31.self_attn.v_proj.weight_scale": "model-00022-of-00086.safetensors",
+ "model.layers.32.input_layernorm.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.down_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.down_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.gate_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.gate_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.up_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.mlp.up_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.post_attention_layernorm.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.k_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.k_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.o_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.o_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.q_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.q_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.v_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.32.self_attn.v_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.33.input_layernorm.weight": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.down_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.down_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.gate_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.gate_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.up_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.33.mlp.up_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.33.post_attention_layernorm.weight": "model-00024-of-00086.safetensors",
+ "model.layers.33.self_attn.k_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.k_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.o_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.o_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.q_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.q_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.v_proj.weight": "model-00023-of-00086.safetensors",
+ "model.layers.33.self_attn.v_proj.weight_scale": "model-00023-of-00086.safetensors",
+ "model.layers.34.input_layernorm.weight": "model-00025-of-00086.safetensors",
+ "model.layers.34.mlp.down_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.34.mlp.down_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.34.mlp.gate_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.mlp.gate_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.34.mlp.up_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.mlp.up_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.34.post_attention_layernorm.weight": "model-00025-of-00086.safetensors",
+ "model.layers.34.self_attn.k_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.k_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.o_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.o_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.q_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.q_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.v_proj.weight": "model-00024-of-00086.safetensors",
+ "model.layers.34.self_attn.v_proj.weight_scale": "model-00024-of-00086.safetensors",
+ "model.layers.35.input_layernorm.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.down_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.down_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.gate_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.gate_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.up_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.mlp.up_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.post_attention_layernorm.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.k_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.k_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.o_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.o_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.q_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.q_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.v_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.35.self_attn.v_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.36.input_layernorm.weight": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.down_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.down_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.gate_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.gate_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.up_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.36.mlp.up_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.36.post_attention_layernorm.weight": "model-00026-of-00086.safetensors",
+ "model.layers.36.self_attn.k_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.k_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.o_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.o_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.q_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.q_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.v_proj.weight": "model-00025-of-00086.safetensors",
+ "model.layers.36.self_attn.v_proj.weight_scale": "model-00025-of-00086.safetensors",
+ "model.layers.37.input_layernorm.weight": "model-00027-of-00086.safetensors",
+ "model.layers.37.mlp.down_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.37.mlp.down_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.37.mlp.gate_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.mlp.gate_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.37.mlp.up_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.mlp.up_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.37.post_attention_layernorm.weight": "model-00027-of-00086.safetensors",
+ "model.layers.37.self_attn.k_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.k_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.o_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.o_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.q_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.q_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.v_proj.weight": "model-00026-of-00086.safetensors",
+ "model.layers.37.self_attn.v_proj.weight_scale": "model-00026-of-00086.safetensors",
+ "model.layers.38.input_layernorm.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.down_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.down_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.gate_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.gate_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.up_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.mlp.up_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.post_attention_layernorm.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.k_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.k_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.o_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.o_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.q_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.q_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.v_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.38.self_attn.v_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.39.input_layernorm.weight": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.down_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.down_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.gate_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.gate_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.up_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.39.mlp.up_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.39.post_attention_layernorm.weight": "model-00028-of-00086.safetensors",
+ "model.layers.39.self_attn.k_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.k_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.o_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.o_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.q_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.q_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.v_proj.weight": "model-00027-of-00086.safetensors",
+ "model.layers.39.self_attn.v_proj.weight_scale": "model-00027-of-00086.safetensors",
+ "model.layers.4.input_layernorm.weight": "model-00005-of-00086.safetensors",
+ "model.layers.4.mlp.down_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.4.mlp.down_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.mlp.gate_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.4.mlp.up_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.mlp.up_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.4.post_attention_layernorm.weight": "model-00005-of-00086.safetensors",
+ "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.k_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.o_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.q_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00086.safetensors",
+ "model.layers.4.self_attn.v_proj.weight_scale": "model-00004-of-00086.safetensors",
+ "model.layers.40.input_layernorm.weight": "model-00029-of-00086.safetensors",
+ "model.layers.40.mlp.down_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.40.mlp.down_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.40.mlp.gate_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.mlp.gate_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.40.mlp.up_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.mlp.up_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.40.post_attention_layernorm.weight": "model-00029-of-00086.safetensors",
+ "model.layers.40.self_attn.k_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.k_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.o_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.o_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.q_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.q_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.v_proj.weight": "model-00028-of-00086.safetensors",
+ "model.layers.40.self_attn.v_proj.weight_scale": "model-00028-of-00086.safetensors",
+ "model.layers.41.input_layernorm.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.down_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.down_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.gate_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.gate_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.up_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.mlp.up_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.post_attention_layernorm.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.k_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.k_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.o_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.o_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.q_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.q_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.v_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.41.self_attn.v_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.42.input_layernorm.weight": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.down_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.down_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.gate_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.gate_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.up_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.42.mlp.up_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.42.post_attention_layernorm.weight": "model-00030-of-00086.safetensors",
+ "model.layers.42.self_attn.k_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.k_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.o_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.o_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.q_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.q_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.v_proj.weight": "model-00029-of-00086.safetensors",
+ "model.layers.42.self_attn.v_proj.weight_scale": "model-00029-of-00086.safetensors",
+ "model.layers.43.input_layernorm.weight": "model-00031-of-00086.safetensors",
+ "model.layers.43.mlp.down_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.43.mlp.down_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.43.mlp.gate_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.mlp.gate_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.43.mlp.up_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.mlp.up_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.43.post_attention_layernorm.weight": "model-00031-of-00086.safetensors",
+ "model.layers.43.self_attn.k_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.k_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.o_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.o_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.q_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.q_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.v_proj.weight": "model-00030-of-00086.safetensors",
+ "model.layers.43.self_attn.v_proj.weight_scale": "model-00030-of-00086.safetensors",
+ "model.layers.44.input_layernorm.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.down_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.down_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.gate_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.gate_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.up_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.mlp.up_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.post_attention_layernorm.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.k_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.k_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.o_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.o_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.q_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.q_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.v_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.44.self_attn.v_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.45.input_layernorm.weight": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.down_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.down_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.gate_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.gate_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.up_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.45.mlp.up_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.45.post_attention_layernorm.weight": "model-00032-of-00086.safetensors",
+ "model.layers.45.self_attn.k_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.k_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.o_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.o_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.q_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.q_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.v_proj.weight": "model-00031-of-00086.safetensors",
+ "model.layers.45.self_attn.v_proj.weight_scale": "model-00031-of-00086.safetensors",
+ "model.layers.46.input_layernorm.weight": "model-00033-of-00086.safetensors",
+ "model.layers.46.mlp.down_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.46.mlp.down_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.46.mlp.gate_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.mlp.gate_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.46.mlp.up_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.mlp.up_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.46.post_attention_layernorm.weight": "model-00033-of-00086.safetensors",
+ "model.layers.46.self_attn.k_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.k_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.o_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.o_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.q_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.q_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.v_proj.weight": "model-00032-of-00086.safetensors",
+ "model.layers.46.self_attn.v_proj.weight_scale": "model-00032-of-00086.safetensors",
+ "model.layers.47.input_layernorm.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.down_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.down_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.gate_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.gate_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.up_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.mlp.up_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.post_attention_layernorm.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.k_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.k_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.o_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.o_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.q_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.q_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.v_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.47.self_attn.v_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.48.input_layernorm.weight": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.down_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.down_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.gate_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.gate_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.up_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.48.mlp.up_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.48.post_attention_layernorm.weight": "model-00034-of-00086.safetensors",
+ "model.layers.48.self_attn.k_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.k_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.o_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.o_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.q_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.q_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.v_proj.weight": "model-00033-of-00086.safetensors",
+ "model.layers.48.self_attn.v_proj.weight_scale": "model-00033-of-00086.safetensors",
+ "model.layers.49.input_layernorm.weight": "model-00035-of-00086.safetensors",
+ "model.layers.49.mlp.down_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.49.mlp.down_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.49.mlp.gate_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.mlp.gate_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.49.mlp.up_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.mlp.up_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.49.post_attention_layernorm.weight": "model-00035-of-00086.safetensors",
+ "model.layers.49.self_attn.k_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.k_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.o_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.o_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.q_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.q_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.v_proj.weight": "model-00034-of-00086.safetensors",
+ "model.layers.49.self_attn.v_proj.weight_scale": "model-00034-of-00086.safetensors",
+ "model.layers.5.input_layernorm.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.down_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.down_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.gate_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.up_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.mlp.up_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.k_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.k_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.o_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.o_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.q_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.q_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.v_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.5.self_attn.v_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.50.input_layernorm.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.down_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.down_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.gate_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.gate_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.up_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.mlp.up_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.post_attention_layernorm.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.k_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.k_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.o_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.o_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.q_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.q_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.v_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.50.self_attn.v_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.51.input_layernorm.weight": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.down_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.down_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.gate_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.gate_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.up_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.51.mlp.up_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.51.post_attention_layernorm.weight": "model-00036-of-00086.safetensors",
+ "model.layers.51.self_attn.k_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.k_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.o_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.o_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.q_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.q_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.v_proj.weight": "model-00035-of-00086.safetensors",
+ "model.layers.51.self_attn.v_proj.weight_scale": "model-00035-of-00086.safetensors",
+ "model.layers.52.input_layernorm.weight": "model-00037-of-00086.safetensors",
+ "model.layers.52.mlp.down_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.52.mlp.down_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.52.mlp.gate_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.mlp.gate_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.52.mlp.up_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.mlp.up_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.52.post_attention_layernorm.weight": "model-00037-of-00086.safetensors",
+ "model.layers.52.self_attn.k_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.k_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.o_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.o_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.q_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.q_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.v_proj.weight": "model-00036-of-00086.safetensors",
+ "model.layers.52.self_attn.v_proj.weight_scale": "model-00036-of-00086.safetensors",
+ "model.layers.53.input_layernorm.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.down_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.down_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.gate_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.gate_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.up_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.mlp.up_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.post_attention_layernorm.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.k_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.k_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.o_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.o_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.q_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.q_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.v_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.53.self_attn.v_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.54.input_layernorm.weight": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.down_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.down_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.gate_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.gate_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.up_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.54.mlp.up_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.54.post_attention_layernorm.weight": "model-00038-of-00086.safetensors",
+ "model.layers.54.self_attn.k_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.k_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.o_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.o_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.q_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.q_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.v_proj.weight": "model-00037-of-00086.safetensors",
+ "model.layers.54.self_attn.v_proj.weight_scale": "model-00037-of-00086.safetensors",
+ "model.layers.55.input_layernorm.weight": "model-00039-of-00086.safetensors",
+ "model.layers.55.mlp.down_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.55.mlp.down_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.55.mlp.gate_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.mlp.gate_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.55.mlp.up_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.mlp.up_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.55.post_attention_layernorm.weight": "model-00039-of-00086.safetensors",
+ "model.layers.55.self_attn.k_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.k_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.o_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.o_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.q_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.q_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.v_proj.weight": "model-00038-of-00086.safetensors",
+ "model.layers.55.self_attn.v_proj.weight_scale": "model-00038-of-00086.safetensors",
+ "model.layers.56.input_layernorm.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.down_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.down_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.gate_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.gate_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.up_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.mlp.up_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.post_attention_layernorm.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.k_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.k_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.o_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.o_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.q_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.q_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.v_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.56.self_attn.v_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.57.input_layernorm.weight": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.down_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.down_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.gate_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.gate_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.up_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.57.mlp.up_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.57.post_attention_layernorm.weight": "model-00040-of-00086.safetensors",
+ "model.layers.57.self_attn.k_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.k_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.o_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.o_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.q_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.q_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.v_proj.weight": "model-00039-of-00086.safetensors",
+ "model.layers.57.self_attn.v_proj.weight_scale": "model-00039-of-00086.safetensors",
+ "model.layers.58.input_layernorm.weight": "model-00041-of-00086.safetensors",
+ "model.layers.58.mlp.down_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.58.mlp.down_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.58.mlp.gate_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.mlp.gate_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.58.mlp.up_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.mlp.up_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.58.post_attention_layernorm.weight": "model-00041-of-00086.safetensors",
+ "model.layers.58.self_attn.k_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.k_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.o_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.o_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.q_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.q_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.v_proj.weight": "model-00040-of-00086.safetensors",
+ "model.layers.58.self_attn.v_proj.weight_scale": "model-00040-of-00086.safetensors",
+ "model.layers.59.input_layernorm.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.down_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.down_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.gate_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.gate_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.up_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.mlp.up_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.post_attention_layernorm.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.k_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.k_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.o_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.o_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.q_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.q_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.v_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.59.self_attn.v_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.6.input_layernorm.weight": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.down_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.down_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.gate_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.gate_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.up_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.6.mlp.up_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00086.safetensors",
+ "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.k_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.o_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.q_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00086.safetensors",
+ "model.layers.6.self_attn.v_proj.weight_scale": "model-00005-of-00086.safetensors",
+ "model.layers.60.input_layernorm.weight": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.down_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.down_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.gate_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.gate_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.up_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.60.mlp.up_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.60.post_attention_layernorm.weight": "model-00042-of-00086.safetensors",
+ "model.layers.60.self_attn.k_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.k_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.o_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.o_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.q_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.q_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.v_proj.weight": "model-00041-of-00086.safetensors",
+ "model.layers.60.self_attn.v_proj.weight_scale": "model-00041-of-00086.safetensors",
+ "model.layers.61.input_layernorm.weight": "model-00043-of-00086.safetensors",
+ "model.layers.61.mlp.down_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.61.mlp.down_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.61.mlp.gate_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.mlp.gate_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.61.mlp.up_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.mlp.up_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.61.post_attention_layernorm.weight": "model-00043-of-00086.safetensors",
+ "model.layers.61.self_attn.k_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.k_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.o_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.o_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.q_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.q_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.v_proj.weight": "model-00042-of-00086.safetensors",
+ "model.layers.61.self_attn.v_proj.weight_scale": "model-00042-of-00086.safetensors",
+ "model.layers.62.input_layernorm.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.down_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.down_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.gate_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.gate_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.up_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.mlp.up_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.post_attention_layernorm.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.k_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.k_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.o_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.o_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.q_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.q_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.v_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.62.self_attn.v_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.63.input_layernorm.weight": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.down_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.down_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.gate_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.gate_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.up_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.63.mlp.up_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.63.post_attention_layernorm.weight": "model-00044-of-00086.safetensors",
+ "model.layers.63.self_attn.k_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.k_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.o_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.o_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.q_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.q_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.v_proj.weight": "model-00043-of-00086.safetensors",
+ "model.layers.63.self_attn.v_proj.weight_scale": "model-00043-of-00086.safetensors",
+ "model.layers.64.input_layernorm.weight": "model-00045-of-00086.safetensors",
+ "model.layers.64.mlp.down_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.64.mlp.down_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.64.mlp.gate_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.mlp.gate_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.64.mlp.up_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.mlp.up_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.64.post_attention_layernorm.weight": "model-00045-of-00086.safetensors",
+ "model.layers.64.self_attn.k_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.k_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.o_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.o_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.q_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.q_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.v_proj.weight": "model-00044-of-00086.safetensors",
+ "model.layers.64.self_attn.v_proj.weight_scale": "model-00044-of-00086.safetensors",
+ "model.layers.65.input_layernorm.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.down_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.down_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.gate_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.gate_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.up_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.mlp.up_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.post_attention_layernorm.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.k_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.k_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.o_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.o_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.q_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.q_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.v_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.65.self_attn.v_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.66.input_layernorm.weight": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.down_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.down_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.gate_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.gate_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.up_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.66.mlp.up_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.66.post_attention_layernorm.weight": "model-00046-of-00086.safetensors",
+ "model.layers.66.self_attn.k_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.k_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.o_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.o_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.q_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.q_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.v_proj.weight": "model-00045-of-00086.safetensors",
+ "model.layers.66.self_attn.v_proj.weight_scale": "model-00045-of-00086.safetensors",
+ "model.layers.67.input_layernorm.weight": "model-00047-of-00086.safetensors",
+ "model.layers.67.mlp.down_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.67.mlp.down_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.67.mlp.gate_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.mlp.gate_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.67.mlp.up_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.mlp.up_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.67.post_attention_layernorm.weight": "model-00047-of-00086.safetensors",
+ "model.layers.67.self_attn.k_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.k_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.o_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.o_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.q_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.q_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.v_proj.weight": "model-00046-of-00086.safetensors",
+ "model.layers.67.self_attn.v_proj.weight_scale": "model-00046-of-00086.safetensors",
+ "model.layers.68.input_layernorm.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.down_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.down_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.gate_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.gate_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.up_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.mlp.up_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.post_attention_layernorm.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.k_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.k_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.o_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.o_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.q_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.q_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.v_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.68.self_attn.v_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.69.input_layernorm.weight": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.down_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.down_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.gate_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.gate_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.up_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.69.mlp.up_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.69.post_attention_layernorm.weight": "model-00048-of-00086.safetensors",
+ "model.layers.69.self_attn.k_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.k_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.o_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.o_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.q_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.q_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.v_proj.weight": "model-00047-of-00086.safetensors",
+ "model.layers.69.self_attn.v_proj.weight_scale": "model-00047-of-00086.safetensors",
+ "model.layers.7.input_layernorm.weight": "model-00007-of-00086.safetensors",
+ "model.layers.7.mlp.down_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.7.mlp.down_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.mlp.gate_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.7.mlp.up_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.mlp.up_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00086.safetensors",
+ "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.k_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.o_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.q_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00086.safetensors",
+ "model.layers.7.self_attn.v_proj.weight_scale": "model-00006-of-00086.safetensors",
+ "model.layers.70.input_layernorm.weight": "model-00049-of-00086.safetensors",
+ "model.layers.70.mlp.down_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.70.mlp.down_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.70.mlp.gate_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.mlp.gate_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.70.mlp.up_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.mlp.up_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.70.post_attention_layernorm.weight": "model-00049-of-00086.safetensors",
+ "model.layers.70.self_attn.k_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.k_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.o_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.o_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.q_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.q_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.v_proj.weight": "model-00048-of-00086.safetensors",
+ "model.layers.70.self_attn.v_proj.weight_scale": "model-00048-of-00086.safetensors",
+ "model.layers.71.input_layernorm.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.down_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.down_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.gate_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.gate_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.up_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.mlp.up_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.post_attention_layernorm.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.k_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.k_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.o_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.o_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.q_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.q_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.v_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.71.self_attn.v_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.72.input_layernorm.weight": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.down_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.down_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.gate_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.gate_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.up_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.72.mlp.up_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.72.post_attention_layernorm.weight": "model-00050-of-00086.safetensors",
+ "model.layers.72.self_attn.k_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.k_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.o_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.o_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.q_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.q_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.v_proj.weight": "model-00049-of-00086.safetensors",
+ "model.layers.72.self_attn.v_proj.weight_scale": "model-00049-of-00086.safetensors",
+ "model.layers.73.input_layernorm.weight": "model-00051-of-00086.safetensors",
+ "model.layers.73.mlp.down_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.73.mlp.down_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.73.mlp.gate_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.mlp.gate_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.73.mlp.up_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.mlp.up_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.73.post_attention_layernorm.weight": "model-00051-of-00086.safetensors",
+ "model.layers.73.self_attn.k_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.k_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.o_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.o_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.q_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.q_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.v_proj.weight": "model-00050-of-00086.safetensors",
+ "model.layers.73.self_attn.v_proj.weight_scale": "model-00050-of-00086.safetensors",
+ "model.layers.74.input_layernorm.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.down_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.down_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.gate_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.gate_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.up_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.mlp.up_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.post_attention_layernorm.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.k_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.k_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.o_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.o_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.q_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.q_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.v_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.74.self_attn.v_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.75.input_layernorm.weight": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.down_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.down_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.gate_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.gate_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.up_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.75.mlp.up_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.75.post_attention_layernorm.weight": "model-00052-of-00086.safetensors",
+ "model.layers.75.self_attn.k_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.k_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.o_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.o_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.q_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.q_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.v_proj.weight": "model-00051-of-00086.safetensors",
+ "model.layers.75.self_attn.v_proj.weight_scale": "model-00051-of-00086.safetensors",
+ "model.layers.76.input_layernorm.weight": "model-00053-of-00086.safetensors",
+ "model.layers.76.mlp.down_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.76.mlp.down_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.76.mlp.gate_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.mlp.gate_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.76.mlp.up_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.mlp.up_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.76.post_attention_layernorm.weight": "model-00053-of-00086.safetensors",
+ "model.layers.76.self_attn.k_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.k_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.o_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.o_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.q_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.q_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.v_proj.weight": "model-00052-of-00086.safetensors",
+ "model.layers.76.self_attn.v_proj.weight_scale": "model-00052-of-00086.safetensors",
+ "model.layers.77.input_layernorm.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.down_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.down_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.gate_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.gate_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.up_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.mlp.up_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.post_attention_layernorm.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.k_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.k_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.o_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.o_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.q_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.q_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.v_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.77.self_attn.v_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.78.input_layernorm.weight": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.down_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.down_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.gate_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.gate_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.up_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.78.mlp.up_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.78.post_attention_layernorm.weight": "model-00054-of-00086.safetensors",
+ "model.layers.78.self_attn.k_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.k_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.o_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.o_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.q_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.q_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.v_proj.weight": "model-00053-of-00086.safetensors",
+ "model.layers.78.self_attn.v_proj.weight_scale": "model-00053-of-00086.safetensors",
+ "model.layers.79.input_layernorm.weight": "model-00055-of-00086.safetensors",
+ "model.layers.79.mlp.down_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.79.mlp.down_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.79.mlp.gate_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.mlp.gate_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.79.mlp.up_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.mlp.up_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.79.post_attention_layernorm.weight": "model-00055-of-00086.safetensors",
+ "model.layers.79.self_attn.k_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.k_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.o_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.o_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.q_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.q_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.v_proj.weight": "model-00054-of-00086.safetensors",
+ "model.layers.79.self_attn.v_proj.weight_scale": "model-00054-of-00086.safetensors",
+ "model.layers.8.input_layernorm.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.down_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.down_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.gate_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.up_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.mlp.up_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.k_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.o_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.q_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.8.self_attn.v_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.80.input_layernorm.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.down_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.down_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.gate_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.gate_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.up_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.mlp.up_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.post_attention_layernorm.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.k_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.k_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.o_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.o_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.q_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.q_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.v_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.80.self_attn.v_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.81.input_layernorm.weight": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.down_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.down_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.gate_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.gate_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.up_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.81.mlp.up_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.81.post_attention_layernorm.weight": "model-00056-of-00086.safetensors",
+ "model.layers.81.self_attn.k_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.k_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.o_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.o_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.q_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.q_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.v_proj.weight": "model-00055-of-00086.safetensors",
+ "model.layers.81.self_attn.v_proj.weight_scale": "model-00055-of-00086.safetensors",
+ "model.layers.82.input_layernorm.weight": "model-00057-of-00086.safetensors",
+ "model.layers.82.mlp.down_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.82.mlp.down_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.82.mlp.gate_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.mlp.gate_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.82.mlp.up_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.mlp.up_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.82.post_attention_layernorm.weight": "model-00057-of-00086.safetensors",
+ "model.layers.82.self_attn.k_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.k_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.o_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.o_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.q_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.q_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.v_proj.weight": "model-00056-of-00086.safetensors",
+ "model.layers.82.self_attn.v_proj.weight_scale": "model-00056-of-00086.safetensors",
+ "model.layers.83.input_layernorm.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.down_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.down_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.gate_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.gate_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.up_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.mlp.up_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.post_attention_layernorm.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.k_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.k_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.o_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.o_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.q_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.q_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.v_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.83.self_attn.v_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.84.input_layernorm.weight": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.down_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.down_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.gate_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.gate_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.up_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.84.mlp.up_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.84.post_attention_layernorm.weight": "model-00058-of-00086.safetensors",
+ "model.layers.84.self_attn.k_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.k_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.o_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.o_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.q_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.q_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.v_proj.weight": "model-00057-of-00086.safetensors",
+ "model.layers.84.self_attn.v_proj.weight_scale": "model-00057-of-00086.safetensors",
+ "model.layers.85.input_layernorm.weight": "model-00059-of-00086.safetensors",
+ "model.layers.85.mlp.down_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.85.mlp.down_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.85.mlp.gate_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.mlp.gate_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.85.mlp.up_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.mlp.up_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.85.post_attention_layernorm.weight": "model-00059-of-00086.safetensors",
+ "model.layers.85.self_attn.k_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.k_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.o_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.o_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.q_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.q_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.v_proj.weight": "model-00058-of-00086.safetensors",
+ "model.layers.85.self_attn.v_proj.weight_scale": "model-00058-of-00086.safetensors",
+ "model.layers.86.input_layernorm.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.down_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.down_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.gate_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.gate_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.up_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.mlp.up_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.post_attention_layernorm.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.k_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.k_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.o_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.o_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.q_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.q_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.v_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.86.self_attn.v_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.87.input_layernorm.weight": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.down_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.down_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.gate_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.gate_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.up_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.87.mlp.up_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.87.post_attention_layernorm.weight": "model-00060-of-00086.safetensors",
+ "model.layers.87.self_attn.k_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.k_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.o_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.o_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.q_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.q_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.v_proj.weight": "model-00059-of-00086.safetensors",
+ "model.layers.87.self_attn.v_proj.weight_scale": "model-00059-of-00086.safetensors",
+ "model.layers.88.input_layernorm.weight": "model-00061-of-00086.safetensors",
+ "model.layers.88.mlp.down_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.88.mlp.down_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.88.mlp.gate_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.mlp.gate_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.88.mlp.up_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.mlp.up_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.88.post_attention_layernorm.weight": "model-00061-of-00086.safetensors",
+ "model.layers.88.self_attn.k_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.k_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.o_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.o_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.q_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.q_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.v_proj.weight": "model-00060-of-00086.safetensors",
+ "model.layers.88.self_attn.v_proj.weight_scale": "model-00060-of-00086.safetensors",
+ "model.layers.89.input_layernorm.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.down_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.down_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.gate_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.gate_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.up_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.mlp.up_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.post_attention_layernorm.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.k_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.k_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.o_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.o_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.q_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.q_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.v_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.89.self_attn.v_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.9.input_layernorm.weight": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.down_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.down_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.gate_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.up_proj.weight": "model-00008-of-00086.safetensors",
+ "model.layers.9.mlp.up_proj.weight_scale": "model-00008-of-00086.safetensors",
+ "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00086.safetensors",
+ "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.k_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.o_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.q_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00086.safetensors",
+ "model.layers.9.self_attn.v_proj.weight_scale": "model-00007-of-00086.safetensors",
+ "model.layers.90.input_layernorm.weight": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.down_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.down_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.gate_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.gate_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.up_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.90.mlp.up_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.90.post_attention_layernorm.weight": "model-00062-of-00086.safetensors",
+ "model.layers.90.self_attn.k_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.k_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.o_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.o_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.q_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.q_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.v_proj.weight": "model-00061-of-00086.safetensors",
+ "model.layers.90.self_attn.v_proj.weight_scale": "model-00061-of-00086.safetensors",
+ "model.layers.91.input_layernorm.weight": "model-00063-of-00086.safetensors",
+ "model.layers.91.mlp.down_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.91.mlp.down_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.91.mlp.gate_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.mlp.gate_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.91.mlp.up_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.mlp.up_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.91.post_attention_layernorm.weight": "model-00063-of-00086.safetensors",
+ "model.layers.91.self_attn.k_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.k_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.o_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.o_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.q_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.q_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.v_proj.weight": "model-00062-of-00086.safetensors",
+ "model.layers.91.self_attn.v_proj.weight_scale": "model-00062-of-00086.safetensors",
+ "model.layers.92.input_layernorm.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.down_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.down_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.gate_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.gate_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.up_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.mlp.up_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.post_attention_layernorm.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.k_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.k_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.o_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.o_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.q_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.q_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.v_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.92.self_attn.v_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.93.input_layernorm.weight": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.down_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.down_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.gate_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.gate_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.up_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.93.mlp.up_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.93.post_attention_layernorm.weight": "model-00064-of-00086.safetensors",
+ "model.layers.93.self_attn.k_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.k_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.o_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.o_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.q_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.q_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.v_proj.weight": "model-00063-of-00086.safetensors",
+ "model.layers.93.self_attn.v_proj.weight_scale": "model-00063-of-00086.safetensors",
+ "model.layers.94.input_layernorm.weight": "model-00065-of-00086.safetensors",
+ "model.layers.94.mlp.down_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.94.mlp.down_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.94.mlp.gate_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.mlp.gate_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.94.mlp.up_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.mlp.up_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.94.post_attention_layernorm.weight": "model-00065-of-00086.safetensors",
+ "model.layers.94.self_attn.k_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.k_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.o_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.o_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.q_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.q_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.v_proj.weight": "model-00064-of-00086.safetensors",
+ "model.layers.94.self_attn.v_proj.weight_scale": "model-00064-of-00086.safetensors",
+ "model.layers.95.input_layernorm.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.down_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.down_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.gate_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.gate_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.up_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.mlp.up_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.post_attention_layernorm.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.k_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.k_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.o_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.o_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.q_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.q_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.v_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.95.self_attn.v_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.96.input_layernorm.weight": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.down_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.down_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.gate_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.gate_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.up_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.96.mlp.up_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.96.post_attention_layernorm.weight": "model-00066-of-00086.safetensors",
+ "model.layers.96.self_attn.k_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.k_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.o_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.o_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.q_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.q_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.v_proj.weight": "model-00065-of-00086.safetensors",
+ "model.layers.96.self_attn.v_proj.weight_scale": "model-00065-of-00086.safetensors",
+ "model.layers.97.input_layernorm.weight": "model-00067-of-00086.safetensors",
+ "model.layers.97.mlp.down_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.97.mlp.down_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.97.mlp.gate_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.mlp.gate_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.97.mlp.up_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.mlp.up_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.97.post_attention_layernorm.weight": "model-00067-of-00086.safetensors",
+ "model.layers.97.self_attn.k_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.k_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.o_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.o_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.q_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.q_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.v_proj.weight": "model-00066-of-00086.safetensors",
+ "model.layers.97.self_attn.v_proj.weight_scale": "model-00066-of-00086.safetensors",
+ "model.layers.98.input_layernorm.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.down_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.down_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.gate_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.gate_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.up_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.mlp.up_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.post_attention_layernorm.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.k_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.k_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.o_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.o_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.q_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.q_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.v_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.98.self_attn.v_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.99.input_layernorm.weight": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.down_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.down_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.gate_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.gate_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.up_proj.weight": "model-00068-of-00086.safetensors",
+ "model.layers.99.mlp.up_proj.weight_scale": "model-00068-of-00086.safetensors",
+ "model.layers.99.post_attention_layernorm.weight": "model-00068-of-00086.safetensors",
+ "model.layers.99.self_attn.k_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.k_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.o_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.o_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.q_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.q_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.v_proj.weight": "model-00067-of-00086.safetensors",
+ "model.layers.99.self_attn.v_proj.weight_scale": "model-00067-of-00086.safetensors",
+ "model.norm.weight": "model-00085-of-00086.safetensors"
+ }
+}
diff --git a/recipe.yaml b/recipe.yaml
new file mode 100644
index 0000000000000000000000000000000000000000..9c7755fcffd47fc652a1078c28b5ad89671bc70f
--- /dev/null
+++ b/recipe.yaml
@@ -0,0 +1,9 @@
+quant_stage:
+ quant_modifiers:
+ QuantizationModifier:
+ ignore: [lm_head]
+ config_groups:
+ group_0:
+ weights: {num_bits: 8, type: float, strategy: channel, dynamic: false, symmetric: true}
+ input_activations: {num_bits: 8, type: float, strategy: token, dynamic: true, symmetric: true}
+ targets: [Linear]
diff --git a/special_tokens_map.json b/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..44e8cb8943364b2bfdcf16fc526008d9d0577c48
--- /dev/null
+++ b/special_tokens_map.json
@@ -0,0 +1,23 @@
+{
+ "bos_token": {
+ "content": "<|begin_of_text|>",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "<|im_end|>",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": {
+ "content": "<|end_of_text|>",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/tokenizer.json b/tokenizer.json
new file mode 100644
index 0000000000000000000000000000000000000000..cc11a80de9a55dc804fec9e3720775b37ae11a2b
--- /dev/null
+++ b/tokenizer.json
@@ -0,0 +1,410563 @@
+{
+ "version": "1.0",
+ "truncation": null,
+ "padding": null,
+ "added_tokens": [
+ {
+ "id": 128000,
+ "content": "<|begin_of_text|>",
+ "single_word": false,
+ "lstrip": false,
+ "rstrip": false,
+ "normalized": false,
+ "special": true
+ },
+ {
+ "id": 128001,
+ "content": "<|end_of_text|>",
+ "single_word": false,
+ "lstrip": false,
+ "rstrip": false,
+ "normalized": false,
+ "special": true
+ },
+ {
+ "id": 128002,
+ "content": "
&": 5909, + "CON": 5910, + "Ġrepl": 5911, + "Ġregular": 5912, + "Storage": 5913, + "ramework": 5914, + "Ġgoal": 5915, + "Ġtouch": 5916, + ".widget": 5917, + "Ġbuilt": 5918, + "des": 5919, + "Part": 5920, + "(re": 5921, + "Ġworth": 5922, + "hib": 5923, + "game": 5924, + "91": 5925, + "192": 5926, + "Ġв": 5927, + "acion": 5928, + "ĠWhite": 5929, + "(type": 5930, + "(`": 5931, + "81": 5932, + "Ġnatural": 5933, + "Ġinj": 5934, + "Ġcalcul": 5935, + "ĠApril": 5936, + ".List": 5937, + "Ġassociated": 5938, + "ĉSystem": 5939, + "~~": 5940, + "=[": 5941, + "Ġstorage": 5942, + "Ġbytes": 5943, + "Ġtravel": 5944, + "Ġsou": 5945, + "Ġpassed": 5946, + "!=": 5947, + "ascript": 5948, + ".open": 5949, + "Ġgrid": 5950, + "Ġbus": 5951, + "Ġrecogn": 5952, + "Ab": 5953, + "Ġhon": 5954, + "ĠCenter": 5955, + "Ġprec": 5956, + "build": 5957, + "73": 5958, + "HTML": 5959, + "ĠSan": 5960, + "Ġcountries": 5961, + "aled": 5962, + "token": 5963, + "kt": 5964, + "Ġqual": 5965, + "Last": 5966, + "adow": 5967, + "Ġmanufact": 5968, + "idad": 5969, + "jango": 5970, + "Next": 5971, + "xf": 5972, + ".a": 5973, + "Ġporno": 5974, + "ĠPM": 5975, + "erve": 5976, + "iting": 5977, + "_th": 5978, + "ci": 5979, + "=None": 5980, + "gs": 5981, + "Ġlogin": 5982, + "atives": 5983, + "']);Ċ": 5984, + "Äħ": 5985, + "Ġill": 5986, + "IA": 5987, + "children": 5988, + "DO": 5989, + "Ġlevels": 5990, + "Ġ{{": 5991, + "Ġlooks": 5992, + "Ġ\"#": 5993, + "ToString": 5994, + "Ġnecessary": 5995, + "ĠĠĠĊ": 5996, + "cell": 5997, + "Entry": 5998, + "Ġ'#": 5999, + "Ġextrem": 6000, + "Selector": 6001, + "Ġplaceholder": 6002, + "Load": 6003, + "Ġreleased": 6004, + "ORE": 6005, + "Enumer": 6006, + "ĠTV": 6007, + "SET": 6008, + "inq": 6009, + "Press": 6010, + "ĠDepartment": 6011, + "Ġproperties": 6012, + "Ġrespond": 6013, + "Search": 6014, + "ael": 6015, + "Ġrequ": 6016, + "ĠBook": 6017, + "/Ċ": 6018, + "(st": 6019, + "Ġfinancial": 6020, + "icket": 6021, + "_input": 6022, + "Ġthreat": 6023, + "(in": 6024, + "Strip": 6025, + "ìĿ": 6026, + "ção": 6027, + "71": 6028, + "Ġevidence": 6029, + "));": 6030, + "ĠBro": 6031, + "Ġ[];Ċ": 6032, + "Ġou": 6033, + "buf": 6034, + "Script": 6035, + "dat": 6036, + "Ġrule": 6037, + "#import": 6038, + "=\"/": 6039, + "Serial": 6040, + "Ġstarting": 6041, + "[index": 6042, + "ae": 6043, + "Ġcontrib": 6044, + "session": 6045, + "_new": 6046, + "utable": 6047, + "ober": 6048, + "Ġ\"./": 6049, + "Ġlogger": 6050, + "Ġrecently": 6051, + "Ġreturned": 6052, + "ččĊ": 6053, + ")))Ċ": 6054, + "itions": 6055, + "Ġseek": 6056, + "Ġcommunic": 6057, + "Ġ\".": 6058, + "Ġusername": 6059, + "ECT": 6060, + "DS": 6061, + "Ġotherwise": 6062, + "ĠGerman": 6063, + ".aw": 6064, + "Adapter": 6065, + "ixel": 6066, + "Ġsystems": 6067, + "Ġdrop": 6068, + "83": 6069, + "Ġstructure": 6070, + "Ġ$(\"#": 6071, + "encies": 6072, + "anning": 6073, + "ĠLink": 6074, + "ĠResponse": 6075, + "Ġstri": 6076, + "ż": 6077, + "ĠDB": 6078, + "æĹ": 6079, + "android": 6080, + "submit": 6081, + "otion": 6082, + "92": 6083, + "(@": 6084, + ".test": 6085, + "82": 6086, + "ĊĊĊĊĊĊĊĊ": 6087, + "];čĊ": 6088, + "Ġdirectly": 6089, + "Ġ\"%": 6090, + "ris": 6091, + "elta": 6092, + "AIL": 6093, + "){čĊ": 6094, + "mine": 6095, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 6096, + "(k": 6097, + "bon": 6098, + "asic": 6099, + "pite": 6100, + "___": 6101, + "Max": 6102, + "Ġerrors": 6103, + "ĠWhile": 6104, + "Ġarguments": 6105, + "Ġensure": 6106, + "Right": 6107, + "-based": 6108, + "Web": 6109, + "Ġ-=": 6110, + "Ġintrodu": 6111, + "ĠInst": 6112, + "ĠWash": 6113, + "ordin": 6114, + "join": 6115, + "Database": 6116, + "Ġgrad": 6117, + "Ġusually": 6118, + "ITE": 6119, + "Props": 6120, + "?>Ċ": 6121, + "ĠGo": 6122, + "@Override": 6123, + "REF": 6124, + "Ġip": 6125, + "ĠAustral": 6126, + "Ġist": 6127, + "ViewById": 6128, + "Ġserious": 6129, + "Ġcustomer": 6130, + ".prototype": 6131, + "odo": 6132, + "cor": 6133, + "Ġdoor": 6134, + "ĠWITHOUT": 6135, + "Ġplant": 6136, + "Ġbegan": 6137, + "Ġdistance": 6138, + "()).": 6139, + "Ġchance": 6140, + "Ġord": 6141, + "came": 6142, + "pragma": 6143, + "Ġprotect": 6144, + "ragment": 6145, + "ĠNode": 6146, + "ening": 6147, + "Ñĩ": 6148, + "Ġroute": 6149, + "ĠSchool": 6150, + "hi": 6151, + "Ġneighb": 6152, + "After": 6153, + "licit": 6154, + "Ġcontr": 6155, + "Ġprimary": 6156, + "AA": 6157, + ".WriteLine": 6158, + "utils": 6159, + "Ġbi": 6160, + "Red": 6161, + ".Linq": 6162, + ".object": 6163, + "Ġleaders": 6164, + "unities": 6165, + "Ġgun": 6166, + "onth": 6167, + "ĠDev": 6168, + "FILE": 6169, + "Ġcomments": 6170, + "_len": 6171, + "arrow": 6172, + "amount": 6173, + "Range": 6174, + "sert": 6175, + "GridView": 6176, + "Ġupdated": 6177, + "ĠMo": 6178, + "Ġinform": 6179, + "ociety": 6180, + "ala": 6181, + "Access": 6182, + "Ġhab": 6183, + "Ġcreat": 6184, + "_arg": 6185, + "ĠJanuary": 6186, + "ĠDay": 6187, + "\")čĊ": 6188, + "uple": 6189, + "document": 6190, + "gorith": 6191, + "menu": 6192, + "ĠOver": 6193, + "bb": 6194, + ".title": 6195, + "_out": 6196, + "Ġled": 6197, + "uri": 6198, + "Ġ?>": 6199, + "gl": 6200, + "Ġbank": 6201, + "ayment": 6202, + "ĉprintf": 6203, + "MD": 6204, + "Ġsample": 6205, + "Ġhands": 6206, + "ĠVersion": 6207, + "uario": 6208, + "Ġoffers": 6209, + "ityEngine": 6210, + "Ġshape": 6211, + "Ġsleep": 6212, + "_point": 6213, + "Settings": 6214, + "Ġachie": 6215, + "Ġsold": 6216, + "ota": 6217, + ".bind": 6218, + "Am": 6219, + "Ġsafe": 6220, + "Store": 6221, + "Ġshared": 6222, + "Ġpriv": 6223, + "_VAL": 6224, + "Ġsens": 6225, + "){": 6226, + "Ġremember": 6227, + "shared": 6228, + "element": 6229, + "Ġshoot": 6230, + "Vert": 6231, + "cout": 6232, + "Ġenv": 6233, + "_label": 6234, + "Ġ>Ċ": 6235, + "run": 6236, + "Ġscene": 6237, + "(array": 6238, + "device": 6239, + "_title": 6240, + "agon": 6241, + "]čĊ": 6242, + "aby": 6243, + "Ġbecame": 6244, + "boolean": 6245, + "Ġpark": 6246, + "ĠCode": 6247, + "upload": 6248, + "riday": 6249, + "ĠSeptember": 6250, + "Fe": 6251, + "Ġsen": 6252, + "cing": 6253, + "FL": 6254, + "Col": 6255, + "uts": 6256, + "_page": 6257, + "inn": 6258, + "Ġimplied": 6259, + "aling": 6260, + "Ġyourself": 6261, + ".Count": 6262, + "conf": 6263, + "Ġaud": 6264, + "_init": 6265, + ".)": 6266, + "Ġwrote": 6267, + "003": 6268, + "NG": 6269, + ".Error": 6270, + "ä»": 6271, + ".for": 6272, + "Ġequal": 6273, + "ĠRequest": 6274, + "Ġserial": 6275, + "Ġallows": 6276, + "XX": 6277, + "Ġmiddle": 6278, + "chor": 6279, + "195": 6280, + "94": 6281, + "ø": 6282, + "erval": 6283, + ".Column": 6284, + "reading": 6285, + "Ġescort": 6286, + "ĠAugust": 6287, + "Ġquickly": 6288, + "Ġweap": 6289, + "ĠCG": 6290, + "ropri": 6291, + "ho": 6292, + "Ġcop": 6293, + "(struct": 6294, + "ĠBig": 6295, + "Ġvs": 6296, + "Ġfrequ": 6297, + ".Value": 6298, + "Ġactions": 6299, + "Ġproper": 6300, + "Ġinn": 6301, + "Ġobjects": 6302, + "Ġmatrix": 6303, + "avascript": 6304, + "Ġones": 6305, + ".group": 6306, + "Ġgreen": 6307, + "Ġpaint": 6308, + "ools": 6309, + "ycl": 6310, + "encode": 6311, + "olt": 6312, + "comment": 6313, + ".api": 6314, + "Dir": 6315, + "Ġune": 6316, + "izont": 6317, + ".position": 6318, + "Ġdesigned": 6319, + "_val": 6320, + "avi": 6321, + "iring": 6322, + "tab": 6323, + "Ġlayer": 6324, + "Ġviews": 6325, + "Ġreve": 6326, + "rael": 6327, + "ĠON": 6328, + "rics": 6329, + "160": 6330, + "np": 6331, + "Ġcore": 6332, + "());čĊ": 6333, + "Main": 6334, + "Ġexpert": 6335, + "ĉĉčĊ": 6336, + "_en": 6337, + "Ġ/>": 6338, + "utter": 6339, + "IAL": 6340, + "ails": 6341, + "ĠKing": 6342, + "*/ĊĊ": 6343, + "ĠMet": 6344, + "_end": 6345, + "addr": 6346, + "ora": 6347, + "Ġir": 6348, + "Min": 6349, + "Ġsurpr": 6350, + "Ġrepe": 6351, + "Ġdirectory": 6352, + "PUT": 6353, + "-S": 6354, + "Ġelection": 6355, + "haps": 6356, + ".pre": 6357, + "cm": 6358, + "Values": 6359, + "Ġ\"Ċ": 6360, + "column": 6361, + "ivil": 6362, + "Login": 6363, + "inue": 6364, + "93": 6365, + "Ġbeautiful": 6366, + "Ġsecret": 6367, + "(event": 6368, + "Ġchat": 6369, + "ums": 6370, + "Ġorigin": 6371, + "Ġeffects": 6372, + "Ġmanagement": 6373, + "illa": 6374, + "tk": 6375, + "Ġsetting": 6376, + "ĠCour": 6377, + "Ġmassage": 6378, + "ĉend": 6379, + "Ġhappy": 6380, + "Ġfinish": 6381, + "Ġcamera": 6382, + "ĠVer": 6383, + "ĠDemocr": 6384, + "ĠHer": 6385, + "(Q": 6386, + "cons": 6387, + "ita": 6388, + "Ġ'.": 6389, + "{}": 6390, + "ĉC": 6391, + "Ġstuff": 6392, + "194": 6393, + "Ġ:Ċ": 6394, + "ĠAR": 6395, + "Task": 6396, + "hidden": 6397, + "eros": 6398, + "IGN": 6399, + "atio": 6400, + "ĠHealth": 6401, + "olute": 6402, + "Enter": 6403, + "'>": 6404, + "ĠTwitter": 6405, + "ĠCounty": 6406, + "scribe": 6407, + "Ġ=>Ċ": 6408, + "Ġhy": 6409, + "fit": 6410, + "Ġmilitary": 6411, + "Ġsale": 6412, + "required": 6413, + "non": 6414, + "bootstrap": 6415, + "hold": 6416, + "rim": 6417, + "-old": 6418, + "ĠDown": 6419, + "Ġmention": 6420, + "contact": 6421, + "_group": 6422, + "oday": 6423, + "Ġtown": 6424, + "Ġsolution": 6425, + "uate": 6426, + "elling": 6427, + "]->": 6428, + "otes": 6429, + "ental": 6430, + "omen": 6431, + "ospital": 6432, + "ĠSup": 6433, + "_EN": 6434, + "Ġslow": 6435, + "SESSION": 6436, + "Ġblue": 6437, + "ago": 6438, + "Ġlives": 6439, + "Ġ^": 6440, + ".un": 6441, + "inst": 6442, + "enge": 6443, + "Ġcustomers": 6444, + "Ġcast": 6445, + "udget": 6446, + "ï¼ģ": 6447, + "icens": 6448, + "Ġdetermin": 6449, + "Selected": 6450, + "_pl": 6451, + "ueue": 6452, + "Ġdark": 6453, + "//ĊĊ": 6454, + "si": 6455, + "thern": 6456, + "ĠJapan": 6457, + "/w": 6458, + "PU": 6459, + "ĠEast": 6460, + "ovie": 6461, + "Ġpackage": 6462, + "Ġnor": 6463, + "Ġapi": 6464, + "bot": 6465, + "\"];Ċ": 6466, + "_post": 6467, + "ulate": 6468, + "Ġclub": 6469, + "'));Ċ": 6470, + "Ġloop": 6471, + "PIO": 6472, + "ione": 6473, + "shot": 6474, + "Initial": 6475, + "Ġplayed": 6476, + "register": 6477, + "rought": 6478, + "_max": 6479, + "acement": 6480, + "match": 6481, + "raphics": 6482, + "AST": 6483, + "Ġexisting": 6484, + "Ġcomplex": 6485, + "DA": 6486, + ".Ch": 6487, + ".common": 6488, + "mo": 6489, + "Ġ'../../": 6490, + "ito": 6491, + "Ġanalysis": 6492, + "Ġdeliver": 6493, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 6494, + "idx": 6495, + "Ãł": 6496, + "ongo": 6497, + "ĠEnglish": 6498, + "Ċ": 10197, + "_default": 10198, + "ĠDatabase": 10199, + "rep": 10200, + "ESS": 10201, + "nergy": 10202, + ".Find": 10203, + "_mask": 10204, + "Ġrise": 10205, + "Ġkernel": 10206, + "::$": 10207, + ".Q": 10208, + "Ġoffering": 10209, + "decl": 10210, + "ĠCS": 10211, + "Ġlisted": 10212, + "Ġmostly": 10213, + "enger": 10214, + "Ġblocks": 10215, + "olo": 10216, + "Ġgoverning": 10217, + "\\F": 10218, + "Ġconcent": 10219, + ".getText": 10220, + "Ġmb": 10221, + "Ġoccurred": 10222, + "Ġchanging": 10223, + "Scene": 10224, + "_CODE": 10225, + "Beh": 10226, + "\"The": 10227, + "Ġtile": 10228, + "ĠAssociation": 10229, + "ĉP": 10230, + "alty": 10231, + "_ad": 10232, + "odies": 10233, + "iated": 10234, + "Ġprepared": 10235, + "possible": 10236, + "Ġmort": 10237, + "TEST": 10238, + "142": 10239, + "Ġignore": 10240, + "Ġcalc": 10241, + "Ġrs": 10242, + "ĠassertEquals": 10243, + "Ġsz": 10244, + "ĠTHIS": 10245, + ".\"Ċ": 10246, + "Ġcanvas": 10247, + "java": 10248, + "Ġdut": 10249, + "VALID": 10250, + ".sql": 10251, + ".input": 10252, + "Ġaux": 10253, + "Sup": 10254, + "Ġartist": 10255, + "Vec": 10256, + "_TIME": 10257, + ".stringify": 10258, + "etween": 10259, + "ĠCategory": 10260, + "Ġ[-": 10261, + "ĠDevExpress": 10262, + "ĠJul": 10263, + "Ġring": 10264, + ".ed": 10265, + "YY": 10266, + "Let": 10267, + "TextField": 10268, + "Ġflat": 10269, + "_print": 10270, + "ĠOTHER": 10271, + "adian": 10272, + "Ġchecked": 10273, + "ele": 10274, + "Align": 10275, + "standing": 10276, + "Ġ[],": 10277, + "Ġlab": 10278, + "ucky": 10279, + "ĠChristmas": 10280, + "(image": 10281, + ".module": 10282, + "Ġlots": 10283, + "Ġslightly": 10284, + "(final": 10285, + "erge": 10286, + "è¿": 10287, + "147": 10288, + "ĠPolice": 10289, + "143": 10290, + "ĠRight": 10291, + "Ġaward": 10292, + "ĠOS": 10293, + "Ġ{}ĊĊ": 10294, + "Ġptr": 10295, + "oves": 10296, + "icated": 10297, + "ем": 10298, + "Ġmanage": 10299, + "oliday": 10300, + "Amount": 10301, + "oolStrip": 10302, + "tbody": 10303, + "Nav": 10304, + "wrap": 10305, + "BB": 10306, + "Ġwatching": 10307, + "arios": 10308, + "Ġoptional": 10309, + "_K": 10310, + "ĠLicensed": 10311, + ".Map": 10312, + "Timer": 10313, + "ĠAP": 10314, + "ĠRev": 10315, + "(o": 10316, + ",c": 10317, + "umin": 10318, + "etailed": 10319, + "ĠHy": 10320, + "Ġblank": 10321, + "agger": 10322, + "ĠSelf": 10323, + "()[": 10324, + ".make": 10325, + "earn": 10326, + "channel": 10327, + ";Ċ": 10342, + "World": 10343, + "Ġpython": 10344, + "Ġlif": 10345, + "Ġtrav": 10346, + "Ġconven": 10347, + "company": 10348, + "ĠClub": 10349, + "138": 10350, + "Ver": 10351, + "Btn": 10352, + "Ġzone": 10353, + "products": 10354, + "ĠEduc": 10355, + "Ġverify": 10356, + "ĠMil": 10357, + "ono": 10358, + "]);ĊĊ": 10359, + "ENCE": 10360, + "Ġpacket": 10361, + "Ġcer": 10362, + "Ġenumer": 10363, + "Ġpars": 10364, + "formed": 10365, + "Ġoccup": 10366, + "tre": 10367, + "Ġexercise": 10368, + "Day": 10369, + "_sum": 10370, + "Ġasking": 10371, + "aption": 10372, + "Ġorders": 10373, + "Ġspending": 10374, + "ĠERR": 10375, + ".Dis": 10376, + "ĠUtil": 10377, + "âĢľI": 10378, + "\\'": 10379, + "?)": 10380, + "/>Ċ": 10381, + "Ġemot": 10382, + "Ġinfluence": 10383, + "ĠAfrica": 10384, + "atters": 10385, + "Ùħ": 10386, + ".session": 10387, + "Ġchief": 10388, + "ĉĉĉĉĉĉĉĉĉĉĉ": 10389, + "Ġtom": 10390, + "cluded": 10391, + "serial": 10392, + "_handler": 10393, + ".Type": 10394, + "aped": 10395, + "Ġpolicies": 10396, + "-ex": 10397, + "-tr": 10398, + "blank": 10399, + "merce": 10400, + "Ġcoverage": 10401, + "Ġrc": 10402, + "_matrix": 10403, + "_box": 10404, + "Ġcharges": 10405, + "ĠBoston": 10406, + "Pe": 10407, + "Ġcircum": 10408, + "Ġfilled": 10409, + "148": 10410, + "Ġnorth": 10411, + "ictureBox": 10412, + "ĉres": 10413, + "è®": 10414, + "Ġtermin": 10415, + "Ġ[âĢ¦": 10416, + "IRECT": 10417, + "Ġber": 10418, + "Ġ\"../../": 10419, + "retch": 10420, + ".code": 10421, + "_col": 10422, + "ĠGovernment": 10423, + "Ġargv": 10424, + "ĠLord": 10425, + "asi": 10426, + "Exec": 10427, + "ĉlet": 10428, + "vertis": 10429, + "Ġdiscussion": 10430, + "enance": 10431, + "outube": 10432, + "typeof": 10433, + "Ġserved": 10434, + "ĠPut": 10435, + "ĉx": 10436, + "Ġsweet": 10437, + "Before": 10438, + "ategy": 10439, + ".of": 10440, + "ĠMaterial": 10441, + "Sort": 10442, + "ONT": 10443, + "igital": 10444, + "Why": 10445, + "Ġsust": 10446, + "Ġç": 10447, + "abet": 10448, + "Ġsegment": 10449, + "Ġ[],Ċ": 10450, + "ĠMuslim": 10451, + "ĠfindViewById": 10452, + "cut": 10453, + "_TEXT": 10454, + "ĠMary": 10455, + "Ġloved": 10456, + "Ġlie": 10457, + "ĠJO": 10458, + "Ġisset": 10459, + "month": 10460, + "Ġprime": 10461, + "ti": 10462, + "ĠCarol": 10463, + "Use": 10464, + "146": 10465, + "ĠPop": 10466, + "ĠSave": 10467, + "Interval": 10468, + "execute": 10469, + "dy": 10470, + "ĠIran": 10471, + "_cont": 10472, + "ĉT": 10473, + "Ġphase": 10474, + "checkbox": 10475, + "week": 10476, + "Ġhide": 10477, + "Ġtil": 10478, + "Ġju": 10479, + "Custom": 10480, + "burg": 10481, + "/M": 10482, + "TON": 10483, + "Ġquant": 10484, + "Ġrub": 10485, + "ixels": 10486, + "Ġinstalled": 10487, + "Ġdump": 10488, + "Ġproperly": 10489, + "(List": 10490, + "Ġdecide": 10491, + "apply": 10492, + "Has": 10493, + "Ġkeeping": 10494, + "Ġcitizens": 10495, + "Ġjoint": 10496, + "pool": 10497, + "Socket": 10498, + "_op": 10499, + "Ġweapon": 10500, + "gnore": 10501, + "ĠExec": 10502, + "otten": 10503, + "ĠMS": 10504, + "Ġ(-": 10505, + "ĠReview": 10506, + "Ġexamples": 10507, + "Ġtight": 10508, + "!(": 10509, + "DP": 10510, + "ĠMessageBox": 10511, + "Ġphotograph": 10512, + "164": 10513, + "URI": 10514, + "ét": 10515, + "low": 10516, + "ĠGrand": 10517, + ".persistence": 10518, + "Ġmaintain": 10519, + "Ġnums": 10520, + "Ġzip": 10521, + "ials": 10522, + "ĠGets": 10523, + "peg": 10524, + "ĠBuffer": 10525, + "~~~~": 10526, + "rastructure": 10527, + "ĠPL": 10528, + "uen": 10529, + "obby": 10530, + "sizeof": 10531, + "Ġpic": 10532, + "Ġseed": 10533, + "Ġexperienced": 10534, + "Ġodd": 10535, + "Ġkick": 10536, + "Ġprocedure": 10537, + "avigator": 10538, + "-on": 10539, + ",j": 10540, + "ĠAlthough": 10541, + "ĠuserId": 10542, + "accept": 10543, + "Blue": 10544, + "IColor": 10545, + "layer": 10546, + "available": 10547, + "Ġends": 10548, + ".table": 10549, + "Ġdataset": 10550, + "bus": 10551, + "Ġexplain": 10552, + "(pro": 10553, + "ĠCommittee": 10554, + "Ġnoted": 10555, + "]:Ċ": 10556, + "Dim": 10557, + "stdio": 10558, + "154": 10559, + ".\",Ċ": 10560, + "_source": 10561, + "181": 10562, + "ĠWeek": 10563, + "ĠEdge": 10564, + "Ġoperating": 10565, + "Ġeste": 10566, + "ipl": 10567, + "330": 10568, + "agination": 10569, + "Ġproceed": 10570, + "Ġanimation": 10571, + ".Models": 10572, + "ĠWatch": 10573, + "iat": 10574, + "Ġoppon": 10575, + "/A": 10576, + "Report": 10577, + "Ġsounds": 10578, + "_buf": 10579, + "IELD": 10580, + "Ġbund": 10581, + "ĉget": 10582, + ".pr": 10583, + "(tmp": 10584, + "Ġkid": 10585, + ">ĊĊĊ": 10586, + "Ġyang": 10587, + "NotFound": 10588, + "ÑĨ": 10589, + "math": 10590, + "@gmail": 10591, + "ĠLIMIT": 10592, + "redients": 10593, + "Ġvent": 10594, + "avigate": 10595, + "Look": 10596, + "Ġreligious": 10597, + "Ġrand": 10598, + "rio": 10599, + "(GL": 10600, + "_ip": 10601, + "uan": 10602, + "iciency": 10603, + "ĠChange": 10604, + ">čĊčĊ": 10605, + "ĠEntity": 10606, + "Ġrencontre": 10607, + "ĠRet": 10608, + "plan": 10609, + "én": 10610, + "BOOL": 10611, + "uries": 10612, + "train": 10613, + "Definition": 10614, + "============": 10615, + "zz": 10616, + "450": 10617, + "Animation": 10618, + "ĠOK": 10619, + "_menu": 10620, + ".bl": 10621, + "_score": 10622, + "Ġacad": 10623, + "(System": 10624, + "Ġrefresh": 10625, + "'=>$": 10626, + ".Graphics": 10627, + "amento": 10628, + "pid": 10629, + "tc": 10630, + "Ġtips": 10631, + "Ġhomes": 10632, + "Ġfuel": 10633, + "âĸ": 10634, + "_helper": 10635, + "ĠĠčĊ": 10636, + "ĠRoom": 10637, + ".Close": 10638, + "_attr": 10639, + "ĠMount": 10640, + "ĠEv": 10641, + "arser": 10642, + "_top": 10643, + "eah": 10644, + "ĠDelete": 10645, + "ãĢį": 10646, + "uke": 10647, + "Ġusage": 10648, + "aria": 10649, + "_dev": 10650, + "Ġtexture": 10651, + "Ġconversation": 10652, + "eper": 10653, + "Bean": 10654, + "done": 10655, + "nonatomic": 10656, + "ĠSecond": 10657, + "Ġshooting": 10658, + "_pre": 10659, + "Components": 10660, + "Ġ]ĊĊ": 10661, + "__,": 10662, + "stitution": 10663, + ".Char": 10664, + ">();ĊĊ": 10665, + "Ġpresented": 10666, + "Ġwa": 10667, + "oker": 10668, + "-ĊĊ": 10669, + "iner": 10670, + "Ġbecoming": 10671, + "Ġincident": 10672, + "Att": 10673, + "162": 10674, + "Ġrevealed": 10675, + "forc": 10676, + "Ġboot": 10677, + ".page": 10678, + "Enumerator": 10679, + "165": 10680, + "_->": 10681, + "Photo": 10682, + "Ġspring": 10683, + ".\",": 10684, + "ĠDictionary": 10685, + "BJECT": 10686, + "Ġlocations": 10687, + "Ġsamples": 10688, + "InputStream": 10689, + "ĠBrown": 10690, + "Ġstats": 10691, + "quality": 10692, + "Ñħ": 10693, + "-dis": 10694, + "Ġhelping": 10695, + "Ġped": 10696, + "224": 10697, + "(se": 10698, + "ĠWho": 10699, + "alian": 10700, + "internal": 10701, + "Ġft": 10702, + ">().": 10703, + "->{": 10704, + "Ġmine": 10705, + "Ġsector": 10706, + "Ġgro": 10707, + "Ġopportunities": 10708, + "Ġü": 10709, + "Ġmp": 10710, + "Ġalleged": 10711, + "Ġdoubt": 10712, + "Mouse": 10713, + "About": 10714, + "_part": 10715, + "Ġchair": 10716, + "Ġstopped": 10717, + "161": 10718, + "loop": 10719, + "entities": 10720, + "Ġapps": 10721, + "ansion": 10722, + "Ġmental": 10723, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10724, + "FR": 10725, + "Ġdefend": 10726, + "care": 10727, + "Ġideal": 10728, + "/api": 10729, + "urface": 10730, + "011": 10731, + "Ġele": 10732, + "ulator": 10733, + "ĠRights": 10734, + "anguages": 10735, + "Ġfunds": 10736, + "Ġadapt": 10737, + "Attributes": 10738, + "Ġdeploy": 10739, + "opts": 10740, + "Ġvalidation": 10741, + "Ġconcerns": 10742, + "uce": 10743, + ".num": 10744, + "ulture": 10745, + "ila": 10746, + "Ġcup": 10747, + "Ġpure": 10748, + ".Fore": 10749, + "183": 10750, + "ĠHashMap": 10751, + ".valueOf": 10752, + "asm": 10753, + "MO": 10754, + "Ġcs": 10755, + "Ġstores": 10756, + "Ġ************************************************************************": 10757, + "Ġcommunication": 10758, + "mem": 10759, + ".EventHandler": 10760, + ".Status": 10761, + "_right": 10762, + ".setOn": 10763, + "Sheet": 10764, + "Ġidentify": 10765, + "enerated": 10766, + "ordered": 10767, + "Ġ\"[": 10768, + "Ġswe": 10769, + "Condition": 10770, + "ĠAccording": 10771, + "Ġprepare": 10772, + "Ġrob": 10773, + "Pool": 10774, + "Ġsport": 10775, + "rv": 10776, + "ĠRouter": 10777, + "Ġalternative": 10778, + "([]": 10779, + "ĠChicago": 10780, + "ipher": 10781, + "ische": 10782, + "ĠDirector": 10783, + "kl": 10784, + "ĠWil": 10785, + "keys": 10786, + "Ġmysql": 10787, + "Ġwelcome": 10788, + "king": 10789, + "ĠManager": 10790, + "Ġcaught": 10791, + ")}Ċ": 10792, + "Score": 10793, + "_PR": 10794, + "Ġsurvey": 10795, + "hab": 10796, + "Headers": 10797, + "ADER": 10798, + "Ġdecor": 10799, + "Ġturns": 10800, + "Ġradius": 10801, + "errupt": 10802, + "Cor": 10803, + "Ġmel": 10804, + "Ġintr": 10805, + "(q": 10806, + "ĠAC": 10807, + "amos": 10808, + "MAX": 10809, + "ĠGrid": 10810, + "ĠJesus": 10811, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 10812, + ".DE": 10813, + "Ġts": 10814, + "Ġlinked": 10815, + "free": 10816, + "ĠQt": 10817, + "Ġ/**čĊ": 10818, + "Ġfaster": 10819, + "ctr": 10820, + "_J": 10821, + "DT": 10822, + ".Check": 10823, + "Ġcombination": 10824, + "Ġintended": 10825, + "-the": 10826, + "-type": 10827, + "182": 10828, + "ectors": 10829, + "ami": 10830, + "uting": 10831, + "Ġuma": 10832, + "XML": 10833, + "UCT": 10834, + "Ap": 10835, + "ĠRandom": 10836, + "Ġran": 10837, + ".sort": 10838, + "Ġsorted": 10839, + ".Un": 10840, + "401": 10841, + "_PER": 10842, + "itory": 10843, + "Ġpriority": 10844, + "ĠGal": 10845, + "ĠOld": 10846, + "hot": 10847, + "ĠDisplay": 10848, + "(sub": 10849, + "_TH": 10850, + "_Y": 10851, + "ĠCare": 10852, + "loading": 10853, + "Kind": 10854, + "_handle": 10855, + ",,": 10856, + "rase": 10857, + "_replace": 10858, + ".addEventListener": 10859, + "ĠRT": 10860, + "172": 10861, + "Ġentered": 10862, + "gers": 10863, + "Ġich": 10864, + "(start": 10865, + "205": 10866, + "/app": 10867, + "Ġbrother": 10868, + "Memory": 10869, + "Outlet": 10870, + "Ġutf": 10871, + "prec": 10872, + "Ġnavigation": 10873, + "ORK": 10874, + "Ġdst": 10875, + "Detail": 10876, + "Ġaudience": 10877, + "Ġdur": 10878, + "Ġcluster": 10879, + "unched": 10880, + "Ġ],": 10881, + "Ġcomfortable": 10882, + ".values": 10883, + "ĠTotal": 10884, + "Ġsnap": 10885, + "Ġstandards": 10886, + "Ġperformed": 10887, + "hand": 10888, + "(\"@": 10889, + "åŃ": 10890, + "Ġphil": 10891, + "ibr": 10892, + "trim": 10893, + "Ġforget": 10894, + "157": 10895, + "Ġdoctor": 10896, + ".TextBox": 10897, + "377": 10898, + "icons": 10899, + ",s": 10900, + "ĠOp": 10901, + "Sm": 10902, + "Stop": 10903, + "ĉList": 10904, + "ĉu": 10905, + "Comment": 10906, + "_VERSION": 10907, + ".Xtra": 10908, + "Person": 10909, + "rb": 10910, + "LOB": 10911, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĊ": 10912, + "ĠCentral": 10913, + "270": 10914, + "ICK": 10915, + "raq": 10916, + "Ġputting": 10917, + "Ġmd": 10918, + "ĠLove": 10919, + "Program": 10920, + "Border": 10921, + "oor": 10922, + "Ġallowing": 10923, + "after": 10924, + "Ġentries": 10925, + "ĠMaybe": 10926, + "]).": 10927, + "ĠShort": 10928, + ")\\": 10929, + ".now": 10930, + "friend": 10931, + "Ġprefer": 10932, + "ĠGPIO": 10933, + "osis": 10934, + "ĠGameObject": 10935, + "Ġskip": 10936, + "Ġcompetition": 10937, + "_match": 10938, + "lications": 10939, + "_CONT": 10940, + ".groupBox": 10941, + "Ġals": 10942, + "666": 10943, + "\"We": 10944, + "_eq": 10945, + "lan": 10946, + "_search": 10947, + "ĠMusic": 10948, + "asis": 10949, + "Ġbind": 10950, + "ĠIsland": 10951, + "rum": 10952, + "(E": 10953, + "Ġseat": 10954, + "Video": 10955, + "Ġack": 10956, + "reek": 10957, + "={()": 10958, + "Ġrating": 10959, + "Ġrestaurant": 10960, + "456": 10961, + "DEX": 10962, + "(buf": 10963, + "pping": 10964, + "uality": 10965, + "Ġleague": 10966, + "176": 10967, + "Ġfocused": 10968, + "apon": 10969, + "$data": 10970, + "CLUD": 10971, + "CLUDING": 10972, + "Ġabsolute": 10973, + "(query": 10974, + "Ġtells": 10975, + "Ang": 10976, + "Ġcommunities": 10977, + "Ġhonest": 10978, + "oking": 10979, + "Ġapart": 10980, + "arity": 10981, + "/$": 10982, + "_module": 10983, + "ĠEnc": 10984, + ".an": 10985, + ".Config": 10986, + "Cre": 10987, + "Ġshock": 10988, + "ĠArab": 10989, + "IENT": 10990, + "/re": 10991, + "Ġretrie": 10992, + "ycler": 10993, + "isa": 10994, + "ĠOrgan": 10995, + ".graph": 10996, + "Ġí": 10997, + "ĠBAS": 10998, + "Enum": 10999, + "Ġpossibly": 11000, + "ÑĢаÐ": 11001, + "ĠJapanese": 11002, + "Ġcraft": 11003, + "ĠPlace": 11004, + "Ġtalent": 11005, + "Ġfunding": 11006, + "Ġconfirmed": 11007, + "Ġcycle": 11008, + "/x": 11009, + "GE": 11010, + "Ġhearing": 11011, + "Ġplants": 11012, + "Ġmouth": 11013, + "pages": 11014, + "oria": 11015, + "ĠRemove": 11016, + "_total": 11017, + "Ġod": 11018, + "ollapse": 11019, + "door": 11020, + "Ġbought": 11021, + "Ġaddr": 11022, + "ARCH": 11023, + "_dim": 11024, + "dden": 11025, + "Ġdecades": 11026, + "REQUEST": 11027, + "Ġversions": 11028, + "fire": 11029, + "006": 11030, + "Ġmoves": 11031, + "fb": 11032, + "Ġcoffee": 11033, + ".connect": 11034, + "ĠRow": 11035, + "Ġschema": 11036, + "Scope": 11037, + "-Type": 11038, + "Ġfighting": 11039, + "Ġretail": 11040, + "Ġmodified": 11041, + "TF": 11042, + "Files": 11043, + "nie": 11044, + "_command": 11045, + "stone": 11046, + "ĠÑĤ": 11047, + "_thread": 11048, + "Ġbond": 11049, + "ĠDevelopment": 11050, + "Ġpt": 11051, + "FORM": 11052, + "plet": 11053, + "Ġidentified": 11054, + "cpp": 11055, + "206": 11056, + "225": 11057, + "Ġcoding": 11058, + "oked": 11059, + "ĠMaster": 11060, + "IDTH": 11061, + "Ġresidents": 11062, + "redit": 11063, + "ĠPhoto": 11064, + "=-": 11065, + "unte": 11066, + "ateur": 11067, + "159": 11068, + "_STATE": 11069, + "ĠSing": 11070, + "Ġsheet": 11071, + ".val": 11072, + "orse": 11073, + "Ġhers": 11074, + "Ġdetermined": 11075, + "Common": 11076, + "Ġwed": 11077, + "_queue": 11078, + "PH": 11079, + "ĠAtl": 11080, + "cred": 11081, + "/LICENSE": 11082, + "Ġmes": 11083, + "Ġadvanced": 11084, + ".java": 11085, + ".Sh": 11086, + "Go": 11087, + "kill": 11088, + "fp": 11089, + "_settings": 11090, + "Ġpal": 11091, + "Ġtruck": 11092, + "Ġcombined": 11093, + "Ġ\"${": 11094, + "ĠCorpor": 11095, + "Ġjoined": 11096, + "ĠJose": 11097, + "ĠCup": 11098, + "uns": 11099, + "estival": 11100, + "levision": 11101, + "Ġbroken": 11102, + "Ġmarriage": 11103, + "ĠWestern": 11104, + "Ġrepresents": 11105, + "ĠTitle": 11106, + "Ġss": 11107, + ".Ass": 11108, + "ongoose": 11109, + "iento": 11110, + "<>();Ċ": 11111, + "Ġabsolutely": 11112, + "Ġsmooth": 11113, + "TERN": 11114, + "ĠUnless": 11115, + "Word": 11116, + "Ġmerge": 11117, + "igan": 11118, + "ĠVol": 11119, + "Ġnn": 11120, + ".getId": 11121, + "Ġз": 11122, + "171": 11123, + "Ġsexy": 11124, + "Ġseeking": 11125, + "Single": 11126, + ".this": 11127, + "179": 11128, + "Ġkom": 11129, + "bound": 11130, + ";\"": 11131, + "ĠfontSize": 11132, + "_df": 11133, + "Ġinjury": 11134, + "(H": 11135, + "Ġissued": 11136, + "_END": 11137, + ":self": 11138, + "020": 11139, + "Ġpatch": 11140, + "Ġleaves": 11141, + "Ġadopt": 11142, + "FileName": 11143, + "ãĢIJ": 11144, + "Ġexecutive": 11145, + "ĠByte": 11146, + "]))Ċ": 11147, + "Ġnu": 11148, + "outing": 11149, + "cluding": 11150, + "-R": 11151, + ".options": 11152, + "Ġsubstant": 11153, + "avax": 11154, + "ĠBUT": 11155, + "Ġtechnical": 11156, + "Ġtwice": 11157, + "Ġmás": 11158, + "Ġunivers": 11159, + "yr": 11160, + "Ġdrag": 11161, + "ĠDC": 11162, + "Ġsed": 11163, + "Ġbot": 11164, + "ĠPal": 11165, + "ĠHall": 11166, + "forcement": 11167, + "Ġauch": 11168, + ".mod": 11169, + "notation": 11170, + "_files": 11171, + ".line": 11172, + "_flag": 11173, + "[name": 11174, + "Ġresolution": 11175, + "Ġbott": 11176, + "(\"[": 11177, + "ende": 11178, + "(arr": 11179, + "Free": 11180, + "(@\"": 11181, + "ĠDistrict": 11182, + "PEC": 11183, + ":-": 11184, + "Picker": 11185, + "ĠJo": 11186, + "ĠĠĠĠĠĊ": 11187, + "ĠRiver": 11188, + "_rows": 11189, + "Ġhelpful": 11190, + "Ġmassive": 11191, + "---Ċ": 11192, + "Ġmeasures": 11193, + "007": 11194, + "ĠRuntime": 11195, + "Ġworry": 11196, + "ĠSpec": 11197, + "ĉD": 11198, + "ãĢij": 11199, + "Ġ){Ċ": 11200, + "Ġworse": 11201, + "(filename": 11202, + "Ġlay": 11203, + "Ġmagic": 11204, + "ĠTheir": 11205, + "oul": 11206, + "stroy": 11207, + "ĠWhere": 11208, + "280": 11209, + "Ġsudden": 11210, + "Ġdefe": 11211, + "Ġbinding": 11212, + "Ġflight": 11213, + "ĠOnInit": 11214, + "ĠWomen": 11215, + "ĠPolicy": 11216, + "Ġdrugs": 11217, + "ishing": 11218, + "('../": 11219, + "ĠMel": 11220, + "peat": 11221, + "tor": 11222, + "Ġproposed": 11223, + "Ġstated": 11224, + "_RES": 11225, + "Ġeast": 11226, + "212": 11227, + "ĠCONDITION": 11228, + "_desc": 11229, + "Ġwinning": 11230, + "folio": 11231, + "Mapper": 11232, + "ĠPan": 11233, + "ĠAnge": 11234, + ".servlet": 11235, + "Ġcopies": 11236, + "LM": 11237, + "Ġvm": 11238, + "åį": 11239, + "Ġdictionary": 11240, + "Seg": 11241, + "177": 11242, + "elines": 11243, + "ĠSend": 11244, + "Ġiron": 11245, + "ĠFort": 11246, + "166": 11247, + ".domain": 11248, + "Ġdebate": 11249, + "NotNull": 11250, + "eq": 11251, + "acher": 11252, + "lf": 11253, + "ĉfmt": 11254, + "Ġlawy": 11255, + "178": 11256, + "ÄŁ": 11257, + "ĠMen": 11258, + "Ġtrim": 11259, + "(NULL": 11260, + "Ġ!!": 11261, + "Ġpad": 11262, + "Ġfollows": 11263, + "\"][\"": 11264, + "requ": 11265, + "ĠEp": 11266, + ".github": 11267, + "(img": 11268, + "eto": 11269, + "('\\": 11270, + "Services": 11271, + "umbnail": 11272, + "_main": 11273, + "pleted": 11274, + "fortunately": 11275, + "Ġwindows": 11276, + "Ġplane": 11277, + "ĠConnection": 11278, + ".local": 11279, + "uard": 11280, + "}\\": 11281, + "==\"": 11282, + "andon": 11283, + "ĠRoy": 11284, + "west": 11285, + "158": 11286, + "iginal": 11287, + "emies": 11288, + "itz": 11289, + "'):Ċ": 11290, + "ĠPeter": 11291, + "Ġtough": 11292, + "Ġreduced": 11293, + "Ġcalculate": 11294, + "Ġrapid": 11295, + "customer": 11296, + "Ġefficient": 11297, + "Ġmedium": 11298, + "Ġfell": 11299, + ".ref": 11300, + "ĠCas": 11301, + "Ġfeedback": 11302, + "Speed": 11303, + "(output": 11304, + "aje": 11305, + "Ġcategories": 11306, + "Ġfee": 11307, + "};": 11308, + "Ġdeleted": 11309, + "reh": 11310, + "Ġproof": 11311, + "Desc": 11312, + "Build": 11313, + "Ġsides": 11314, + ".ArrayList": 11315, + "-%": 11316, + "ĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠĠ": 11317, + "ر": 11318, + ".match": 11319, + "ли": 11320, + "Ġfeels": 11321, + "Ġachieve": 11322, + "Ġclim": 11323, + "_ON": 11324, + "ĠCD": 11325, + "Ġteacher": 11326, + "_current": 11327, + "bn": 11328, + "_PL": 11329, + "isting": 11330, + "Enable": 11331, + "GEN": 11332, + "Ġtv": 11333, + "Ġsock": 11334, + "Ġplays": 11335, + "Ġdiscount": 11336, + "ĠKE": 11337, + "ĠDebug": 11338, + "Fore": 11339, + "ĠIraq": 11340, + "Ġappearance": 11341, + "Mon": 11342, + "Ġstyled": 11343, + "ĠHuman": 11344, + "iot": 11345, + "ĠHistory": 11346, + "Ġsac": 11347, + "ĠCollection": 11348, + "Ġrecommended": 11349, + ".Selected": 11350, + "Ġorganizations": 11351, + "Ġdiscovered": 11352, + "cohol": 11353, + "adas": 11354, + "ĠThomas": 11355, + "May": 11356, + "Ġconserv": 11357, + "Ġdomin": 11358, + "ĠFollow": 11359, + "ĠSection": 11360, + "ĠThanks": 11361, + "Username": 11362, + "Ġrecipe": 11363, + "Ġwonderful": 11364, + ".sleep": 11365, + "_if": 11366, + "ĉĊĉĊ": 11367, + "orno": 11368, + "Ġru": 11369, + "_target": 11370, + ".\"\"": 11371, + "à¦": 11372, + "EventArgs": 11373, + "Ġinputs": 11374, + "Ġfif": 11375, + "Ġvision": 11376, + "cy": 11377, + "ĠSeries": 11378, + ")(((": 11379, + "Ġtrading": 11380, + "Ġmarker": 11381, + "Begin": 11382, + "Ġtypically": 11383, + "Ġcauses": 11384, + "dropdown": 11385, + "_DEBUG": 11386, + "260": 11387, + "Ġdetect": 11388, + "country": 11389, + "!\");Ċ": 11390, + "ĉR": 11391, + "appy": 11392, + "Ġcref": 11393, + "('<": 11394, + "\"=>": 11395, + "ĠLE": 11396, + "reader": 11397, + "Ġadministr": 11398, + "õ": 11399, + "ucket": 11400, + "Ġfashion": 11401, + ".char": 11402, + "izar": 11403, + "Ġdisable": 11404, + "Ġsuc": 11405, + "ĠLive": 11406, + "issue": 11407, + "Ġmetadata": 11408, + "flags": 11409, + "ĠðŁ": 11410, + "Ġcommitted": 11411, + "Ġva": 11412, + "Ġrough": 11413, + "Ġ'''Ċ": 11414, + "Ġhighlight": 11415, + "_vars": 11416, + "VO": 11417, + "Ġencoding": 11418, + "-Z": 11419, + "_sign": 11420, + "$(\"#": 11421, + "Ġrain": 11422, + "reatest": 11423, + "ĠEND": 11424, + "Selection": 11425, + "Ġcandidates": 11426, + "Ġsav": 11427, + ".Empty": 11428, + "Ġdecisions": 11429, + "Ġcollabor": 11430, + "ridge": 11431, + "feed": 11432, + "ression": 11433, + "Ġpersons": 11434, + "VM": 11435, + "008": 11436, + "ega": 11437, + "_BIT": 11438, + "According": 11439, + "acked": 11440, + "Ġdollars": 11441, + "_loss": 11442, + "ĠCost": 11443, + "}\"Ċ": 11444, + "Notification": 11445, + "Ġprostit": 11446, + "Ġauthority": 11447, + ".rec": 11448, + "Ġspokes": 11449, + "ĠToday": 11450, + "istant": 11451, + "ĠHead": 11452, + "âĢĿ.": 11453, + "ertainment": 11454, + "cean": 11455, + "culate": 11456, + "Ġven": 11457, + "However": 11458, + "_arr": 11459, + "Ġtokens": 11460, + "Graph": 11461, + "ĠJud": 11462, + "ĠVirgin": 11463, + "ĠSerial": 11464, + "unning": 11465, + "Mutable": 11466, + "agers": 11467, + ".csv": 11468, + "Ġdeveloping": 11469, + "Ġinstructions": 11470, + "Ġpromise": 11471, + "Ġrequested": 11472, + "_encode": 11473, + "/\"": 11474, + "ĠIcon": 11475, + "uilt": 11476, + "-day": 11477, + "Ġintelligence": 11478, + ".IS": 11479, + "ĠObservable": 11480, + "ĠHard": 11481, + "Bool": 11482, + "211": 11483, + "idential": 11484, + ".Anchor": 11485, + "Ġselling": 11486, + "CI": 11487, + "AGES": 11488, + "tle": 11489, + "bur": 11490, + "UFFER": 11491, + "RY": 11492, + "Ġbigger": 11493, + "Ġrat": 11494, + "Ġfamous": 11495, + "Ġtypename": 11496, + "Ġexplained": 11497, + "}}Ċ": 11498, + "Ġnuclear": 11499, + "-N": 11500, + "Ġcrisis": 11501, + "ĠEnter": 11502, + "Ġanswers": 11503, + "/${": 11504, + "/pl": 11505, + "Ġsequ": 11506, + "_next": 11507, + "mask": 11508, + "Ġstanding": 11509, + "Ġplenty": 11510, + "ĠCross": 11511, + "ĉret": 11512, + "dro": 11513, + "ĠCast": 11514, + "167": 11515, + "=true": 11516, + "ĠChris": 11517, + "icio": 11518, + "ĠMike": 11519, + "Decimal": 11520, + "addComponent": 11521, + "Len": 11522, + "Ġcock": 11523, + "Ġ#{": 11524, + "URN": 11525, + "": 11657, + "Ġ*=": 11658, + "ĠPS": 11659, + "Ġdangerous": 11660, + "[p": 11661, + "OME": 11662, + "Other": 11663, + "ĠStringBuilder": 11664, + "Points": 11665, + "heading": 11666, + "Ġcurrency": 11667, + "Ġpercentage": 11668, + "_API": 11669, + "Ġclassic": 11670, + "thead": 11671, + "ĠMO": 11672, + "FE": 11673, + "Idx": 11674, + "await": 11675, + "Ġè": 11676, + "Ġaccident": 11677, + "Ġvariant": 11678, + "Ġmyst": 11679, + "ĠLand": 11680, + "ĠBre": 11681, + "Ġharm": 11682, + "ĠAcc": 11683, + "Ġcharged": 11684, + "iones": 11685, + "Visibility": 11686, + "arry": 11687, + "ĠLanguage": 11688, + "Ġwalking": 11689, + "\".ĊĊ": 11690, + "ifer": 11691, + "Ġleadership": 11692, + ".From": 11693, + "ynam": 11694, + "Ġtimestamp": 11695, + "ipt": 11696, + "ĠHas": 11697, + "REFER": 11698, + "ĠIts": 11699, + "Ġlistener": 11700, + "UTE": 11701, + "213": 11702, + "_description": 11703, + "Ġexperiences": 11704, + "Ġcreates": 11705, + "RS": 11706, + "cart": 11707, + "black": 11708, + "Ġchoices": 11709, + "war": 11710, + "750": 11711, + "Ġ'''": 11712, + "Ġordered": 11713, + "Ġevening": 11714, + "Ġpil": 11715, + "Ġtun": 11716, + "ĠBad": 11717, + "(app": 11718, + "random": 11719, + "Ġexplicit": 11720, + "Ġarrived": 11721, + "Ġfly": 11722, + "Ġeconom": 11723, + "-mail": 11724, + "Ġlists": 11725, + "Ġarchitect": 11726, + "234": 11727, + "ĠPay": 11728, + "Ġds": 11729, + "ĠSol": 11730, + "Ġvehicles": 11731, + "Hz": 11732, + "-com": 11733, + "Ġking": 11734, + "_equal": 11735, + "ĠHelp": 11736, + "Ġabuse": 11737, + "480": 11738, + "169": 11739, + "--;Ċ": 11740, + "Ġextr": 11741, + "Ġchemical": 11742, + "ä¿": 11743, + "Ġorient": 11744, + "Ġbreath": 11745, + "ĠSpace": 11746, + "(element": 11747, + "wait": 11748, + "DED": 11749, + "igma": 11750, + "Ġentr": 11751, + "Ġsob": 11752, + "-name": 11753, + "Ġaffected": 11754, + "ika": 11755, + "Ġcoal": 11756, + "_work": 11757, + "Ġhundreds": 11758, + "Ġpolitics": 11759, + "subject": 11760, + "Ġconsumer": 11761, + "ANGE": 11762, + "Ġrepeated": 11763, + "Send": 11764, + "Ġ#[": 11765, + "Ġprotocol": 11766, + "Ġleads": 11767, + "useum": 11768, + "Every": 11769, + "808": 11770, + "174": 11771, + "Import": 11772, + "(count": 11773, + "Ġchallenges": 11774, + "Ġnovel": 11775, + "Ġdepart": 11776, + "bits": 11777, + ".Current": 11778, + "Ġ`${": 11779, + "oting": 11780, + "(\\": 11781, + "Ġcreative": 11782, + "Ġbuff": 11783, + "Ġintroduced": 11784, + "usic": 11785, + "modules": 11786, + "Are": 11787, + "-doc": 11788, + "language": 11789, + "_cache": 11790, + "Ġtod": 11791, + "?>": 11792, + "omething": 11793, + "Ġhun": 11794, + "åº": 11795, + "aters": 11796, + "Intent": 11797, + "Ġimplemented": 11798, + "ĠCase": 11799, + "Children": 11800, + "Ġnotification": 11801, + "Renderer": 11802, + "Wrapper": 11803, + "Objects": 11804, + "tl": 11805, + ".Contains": 11806, + "Plugin": 11807, + ".row": 11808, + "Ġforg": 11809, + "Ġpermit": 11810, + "Ġtargets": 11811, + "ĠIF": 11812, + "Ġtip": 11813, + "sex": 11814, + "Ġsupports": 11815, + "Ġfold": 11816, + "photo": 11817, + "},čĊ": 11818, + "Ġgoogle": 11819, + "$('#": 11820, + "Ġsharing": 11821, + "Ġgoods": 11822, + "vs": 11823, + "ĠDan": 11824, + "Rate": 11825, + "ĠMartin": 11826, + "Ġmanner": 11827, + "lie": 11828, + ".The": 11829, + "Internal": 11830, + "ĠCONTR": 11831, + "Mock": 11832, + "RIGHT": 11833, + "Ġ'{": 11834, + "Ġcontrols": 11835, + "Mat": 11836, + "Ġmand": 11837, + "Ġextended": 11838, + "Ok": 11839, + "Ġembed": 11840, + "Ġplanet": 11841, + "ĠNon": 11842, + "-ch": 11843, + ")\",": 11844, + "epar": 11845, + "Ġbelieved": 11846, + "ĠEnvironment": 11847, + "ĠFriend": 11848, + "-res": 11849, + "Ġhandling": 11850, + "nic": 11851, + "-level": 11852, + "scri": 11853, + "Xml": 11854, + "BE": 11855, + "ungen": 11856, + "Ġalter": 11857, + "[idx": 11858, + "Pop": 11859, + "cam": 11860, + "Ġ(((": 11861, + "Ġshipping": 11862, + "Ġbattery": 11863, + "iddleware": 11864, + "MC": 11865, + "Ġimpl": 11866, + "otation": 11867, + "ĠLab": 11868, + "