jykim310 commited on
Commit
a51d99b
1 Parent(s): 0a7481c

Upload 35 files

Browse files
mlc-chat-config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "qwen2",
3
+ "quantization": "q4f16_2",
4
+ "model_config": {
5
+ "hidden_act": "silu",
6
+ "hidden_size": 1536,
7
+ "intermediate_size": 8960,
8
+ "num_attention_heads": 12,
9
+ "num_hidden_layers": 28,
10
+ "num_key_value_heads": 2,
11
+ "rms_norm_eps": 1e-06,
12
+ "rope_theta": 1000000.0,
13
+ "vocab_size": 151936,
14
+ "context_window_size": 768,
15
+ "prefill_chunk_size": 768,
16
+ "tensor_parallel_shards": 1,
17
+ "head_dim": 128,
18
+ "dtype": "float32"
19
+ },
20
+ "vocab_size": 151936,
21
+ "context_window_size": 768,
22
+ "sliding_window_size": -1,
23
+ "prefill_chunk_size": 768,
24
+ "attention_sink_size": -1,
25
+ "tensor_parallel_shards": 1,
26
+ "mean_gen_len": 128,
27
+ "max_gen_len": 512,
28
+ "shift_fill_factor": 0.3,
29
+ "temperature": 0.0,
30
+ "presence_penalty": 0.0,
31
+ "frequency_penalty": 0.0,
32
+ "repetition_penalty": 1.1,
33
+ "top_p": 0.9,
34
+ "conv_template": {
35
+ "name": "qwen2",
36
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
37
+ "system_message": "You are a helpful assistant.",
38
+ "add_role_after_system_message": true,
39
+ "roles": {
40
+ "user": "<|im_start|>user",
41
+ "assistant": "<|im_start|>assistant"
42
+ },
43
+ "role_templates": {
44
+ "user": "{user_message}",
45
+ "assistant": "{assistant_message}",
46
+ "tool": "{tool_message}"
47
+ },
48
+ "messages": [],
49
+ "seps": [
50
+ "<|im_end|>\n"
51
+ ],
52
+ "role_content_sep": "\n",
53
+ "role_empty_sep": "\n",
54
+ "stop_str": [
55
+ "<|endoftext|>",
56
+ "<|im_end|>"
57
+ ],
58
+ "stop_token_ids": [
59
+ 151643,
60
+ 151645
61
+ ],
62
+ "function_string": "",
63
+ "use_function_calling": false,
64
+ "image_token_index": -1
65
+ },
66
+ "pad_token_id": 151643,
67
+ "bos_token_id": 151643,
68
+ "eos_token_id": [
69
+ 151645,
70
+ 151643
71
+ ],
72
+ "tokenizer_files": [
73
+ "tokenizer.json",
74
+ "vocab.json",
75
+ "merges.txt",
76
+ "added_tokens.json",
77
+ "tokenizer_config.json"
78
+ ],
79
+ "version": "0.1.0"
80
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e52e67f3e01b3b2a2e781a630b8c210d65068019185bf2962837a0b963e3508d
3
+ size 466747392
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802ffd09ad5ca2e0af2db0c1f803c3c50ca1807b42b90b6f28d60648572076a2
3
+ size 29431808
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35c300b2ada08cd7a09f22b53dfc767e2864d152677c0cebfda535ce4fd21035
3
+ size 26331136
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f14fb97e9a10ffc817f74b620a3d92d56cb17ee941e762aafd8a26aaa868016
3
+ size 26331136
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a3a0689f043e86c3e44c5deff625eaacf634300fa710a44431f355e122d01cd
3
+ size 26331136
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eba7db6f25338e5a46dfe91b76eba322262bc494db33b9aae13e4b7e0a8f401
3
+ size 26331136
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb267743150bf31524efddf1b8ec6ddc3532dbee376b5ec1552c7925b247cb8d
3
+ size 26331136
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ebfbe000ac0866930409c88aa4d08ecd5342b8a8625a66d927bfa703a4cc6a
3
+ size 26331136
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12c167036a88334dfc463926bdc58978bed5cc97531a6f37fb3bf3b6bbb18fe9
3
+ size 26331136
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ed76521d738dc2f0df14c095a4fbffe2ff4866d0f4e4dd3d57bd5617510ec4
3
+ size 26331136
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51dc45e5b4d58683a54654dcb377be1b7f6cfef144895742ede6adaaa918882e
3
+ size 26331136
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6939c8e48a492970a249ea55035a98fd9a319e05ee3b64a0259a87f5ada3fbb
3
+ size 26331136
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94d4942ab92c2adb53c56d14399bd53ce71bc4f05154f16211b3fb4b43421b59
3
+ size 26331136
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b201518822a2d9dc647e0c7e7157aeef2e58016855c74bad0a0ade117a0f2f60
3
+ size 26331136
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:615c741bd2a90b45c38eb047ba5facc30ae941cefad434ae566e5661564a9fa9
3
+ size 26331136
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84b7eb8df5569065ceccb4dea6875a1d5e3dbe8359edacc9cbe316d84805f00d
3
+ size 26331136
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac35c343c87c6b2fd6f0e745a767741dd1170ec57e92bcdcb7cedf0252424a92
3
+ size 26331136
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e58137100e3074c42db11c26d10ff235d830deee15321a64a8fa5e34a533c669
3
+ size 26331136
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf44212d2cd2294594956e070ea8f23514faf86524a1a1d6fe1eb0316270fe2d
3
+ size 26331136
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcb568892d2861343692a9f4c3f2bdc18d8eaa405868367970145eacd563265b
3
+ size 26331136
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:506374cdf32d066fd3ea709bf86dd9dd9472ad0a698cbd40f04d1b0e9e0452ea
3
+ size 26331136
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e52e67f3e01b3b2a2e781a630b8c210d65068019185bf2962837a0b963e3508d
3
+ size 466747392
params_shard_29.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a8105db06715251b8225b4e9899c60a1666b82052d7ebf39799614e75ec5083
3
+ size 23233536
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b8e0e24335145e83d26c4734174c5b5c0f3f588ee98de24687891f8552b6ba3
3
+ size 26331136
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ff299a72de28709f465edd3dfbade51f01290c2f095ef03a26c3a35965af232
3
+ size 26331136
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443951ce5df26b11dd69db5850f4613c3b7b39fa1603326a2b9bf46c4c70307f
3
+ size 26331136
params_shard_6.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59d342734d53c540192e0845aa1c7bf8c860aa9123b85a69d992c5847a6ba225
3
+ size 26331136
params_shard_7.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9488a5b85ee1bfb6b284df299197b81f0a6532e8c61e64a064f12faa96c04118
3
+ size 26331136
params_shard_8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea8f9886b45d9aa5b014d8b1977a0ebb2954f0300df54f88445a1f65ef62a08b
3
+ size 26331136
params_shard_9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0276d0911c48f62ad70913f21b6f22d430ad3b5c3eb975046baf250d16789e7e
3
+ size 26331136
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff