jykim310 commited on
Commit
b4ad476
1 Parent(s): 8827570

Upload 36 files

Browse files
added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|endoftext|>": 151643,
3
+ "<|im_end|>": 151645,
4
+ "<|im_start|>": 151644
5
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
mlc-chat-config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "qwen2",
3
+ "quantization": "q4f16_2",
4
+ "model_config": {
5
+ "hidden_act": "silu",
6
+ "hidden_size": 1536,
7
+ "intermediate_size": 8960,
8
+ "num_attention_heads": 12,
9
+ "num_hidden_layers": 28,
10
+ "num_key_value_heads": 2,
11
+ "rms_norm_eps": 1e-06,
12
+ "rope_theta": 1000000.0,
13
+ "vocab_size": 151936,
14
+ "context_window_size": 768,
15
+ "prefill_chunk_size": 768,
16
+ "tensor_parallel_shards": 1,
17
+ "head_dim": 128,
18
+ "dtype": "float32"
19
+ },
20
+ "vocab_size": 151936,
21
+ "context_window_size": 768,
22
+ "sliding_window_size": -1,
23
+ "prefill_chunk_size": 768,
24
+ "attention_sink_size": -1,
25
+ "tensor_parallel_shards": 1,
26
+ "mean_gen_len": 128,
27
+ "max_gen_len": 512,
28
+ "shift_fill_factor": 0.3,
29
+ "temperature": 0.7,
30
+ "presence_penalty": 0.0,
31
+ "frequency_penalty": 0.0,
32
+ "repetition_penalty": 1.1,
33
+ "top_p": 0.8,
34
+ "conv_template": {
35
+ "name": "qwen2",
36
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
37
+ "system_message": "You are a helpful assistant.",
38
+ "add_role_after_system_message": true,
39
+ "roles": {
40
+ "user": "<|im_start|>user",
41
+ "assistant": "<|im_start|>assistant"
42
+ },
43
+ "role_templates": {
44
+ "user": "{user_message}",
45
+ "assistant": "{assistant_message}",
46
+ "tool": "{tool_message}"
47
+ },
48
+ "messages": [],
49
+ "seps": [
50
+ "<|im_end|>\n"
51
+ ],
52
+ "role_content_sep": "\n",
53
+ "role_empty_sep": "\n",
54
+ "stop_str": [
55
+ "<|endoftext|>",
56
+ "<|im_end|>"
57
+ ],
58
+ "stop_token_ids": [
59
+ 151643,
60
+ 151645
61
+ ],
62
+ "function_string": "",
63
+ "use_function_calling": false,
64
+ "image_token_index": -1
65
+ },
66
+ "pad_token_id": 151643,
67
+ "bos_token_id": 151643,
68
+ "eos_token_id": [
69
+ 151645,
70
+ 151643
71
+ ],
72
+ "tokenizer_files": [
73
+ "tokenizer.json",
74
+ "vocab.json",
75
+ "merges.txt",
76
+ "added_tokens.json",
77
+ "tokenizer_config.json"
78
+ ],
79
+ "version": "0.1.0"
80
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43256d0b8f0c812f4ea9e707c40695df8e357fda9eb562d6728a237a534b7f7d
3
+ size 466747392
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43256d0b8f0c812f4ea9e707c40695df8e357fda9eb562d6728a237a534b7f7d
3
+ size 466747392
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3af07a0b51b23bf8897d7d9f9aa1847a5cf4803957b963a20d4c6ffb61c67479
3
+ size 26331136
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59bc00e7e4c24c7eb422a47b122491119c93efff2f89cee6766500bdec274811
3
+ size 26331136
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccdeaff3423f6ea0f0f172323d02b1c46d7b3efe9151ca944bf16bc6520a2c43
3
+ size 26331136
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fc798908ce5d790808ee5bbc6a6c55295ad2cc0566e785a2fb7ceaa039733e4
3
+ size 26331136
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7733e32ce1c21e388e223c0233921b2314f0f3d54097fa624363606c7fb0e8f
3
+ size 26331136
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:342a61e8db52c4be2aa6dbb6b6f3f6e257c6f61d8609d39d3ffad7ef9778664b
3
+ size 32349184
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a318c1a648beae43f3424cf62db410563b65f1f7fca1b458e30abde635cd29d3
3
+ size 31155200
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:750542be451278278b5bc0c59f79b23216aa4b619c21beda53cd00455db28d37
3
+ size 33212416
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee9ff34c7c5cb6e91bc7bbeffeee6bae9ed49424bc77611bd3bc3f09c7f0abb7
3
+ size 27191296
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa4c200a81475a6db414562dcf03b8c0d7444329204587f49398e15b82b577c1
3
+ size 26331136
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc307829af50e77a068ca85e26145a634a5e57c1e2100c8b848b721f5de4c625
3
+ size 33215488
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f50ad297ecb1b00ad6f5560730953fa3ce7c7bd3aa76f356092f0684d751b0e3
3
+ size 26331136
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beb334bc61a0b088f17e31803c31241f0bfbc24a65e00870ca63ebff24d7f444
3
+ size 26331136
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443b39b3d2a960035ce9843d93149c99e942c57609fc33e9160721d52dd3e1d4
3
+ size 33218560
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37025c71b9011fb44d6509607454a2faaf2eda61dcf44812c98ee03d5703e575
3
+ size 27191296
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66cf4fac18abf708cad1f99af96e1fa0481458215380389128f5d9343a88dc90
3
+ size 26331136
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7642ee14a0adfa74bc4c679121b7dd68ca394869738b6c572de122493540262
3
+ size 26331136
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13a0ffa165c764765668debad57251d0cfbe22be13538bc36824c33aca116745
3
+ size 26331136
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9a7380d48792c3a54658e6876ba76176c241f07950fba6ff85c55aa7d8289b2
3
+ size 26331136
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4db3cbb08157d8ca510790c55a0c241d7e6245e28132aef7d730c23cf902b8
3
+ size 18589696
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02319c6431c7cd6aae9030921a96cef6b548b434deca046fb4b78886ff126bc1
3
+ size 27191296
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d3cc695d6a5a99c5ae1d07bdbfcae10425b7c11fbc2292708b7c8e9b03ef740
3
+ size 26331136
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5432b7a0338f74352ea7e84e7e439ca77f476ab83d81bbac9da76d38a2418def
3
+ size 26331136
params_shard_6.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:642ae58228088ebbc2521c4b53cf0c2ae43bef27b641b92631b2c6438aaf92e2
3
+ size 26331136
params_shard_7.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88c80fa53c282f275cfc2eb8467f31ad87c8703f74c22411ff475a9e621f4e59
3
+ size 26331136
params_shard_8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6e8a021b305ec1d73118c54deba798fca66f6efb497b3f0a18a968b8fcdafab
3
+ size 26331136
params_shard_9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed74f0802eda9e480f70e2314d6830cdb42bf932ae790f2e8c64935bc658a2b7
3
+ size 26331136
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": [
30
+ "<|im_start|>",
31
+ "<|im_end|>"
32
+ ],
33
+ "bos_token": null,
34
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
35
+ "clean_up_tokenization_spaces": false,
36
+ "eos_token": "<|im_end|>",
37
+ "errors": "replace",
38
+ "model_max_length": 32768,
39
+ "pad_token": "<|endoftext|>",
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "Qwen2Tokenizer",
42
+ "unk_token": null
43
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff