Corianas commited on
Commit
2fe22ff
1 Parent(s): f731f5e

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - Corianas/Microllama_Char_500k_step
7
+ - Corianas/Microllama_Char_300k_step
8
+ base_model:
9
+ - Corianas/Microllama_Char_500k_step
10
+ - Corianas/Microllama_Char_300k_step
11
+ ---
12
+
13
+ # Microllama-300.500kmerge
14
+
15
+ Microllama-300.500kmerge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
+ * [Corianas/Microllama_Char_500k_step](https://huggingface.co/Corianas/Microllama_Char_500k_step)
17
+ * [Corianas/Microllama_Char_300k_step](https://huggingface.co/Corianas/Microllama_Char_300k_step)
18
+
19
+ ## 🧩 Configuration
20
+
21
+ ```yaml
22
+ slices:
23
+ - sources:
24
+ - model: Corianas/Microllama_Char_500k_step
25
+ layer_range: [0, 12]
26
+ - model: Corianas/Microllama_Char_300k_step
27
+ layer_range: [0, 12]
28
+ merge_method: slerp
29
+ base_model: Corianas/Microllama_Char_300k_step
30
+ parameters:
31
+ t:
32
+ - filter: self_attn
33
+ value: [0, 0.5, 0.3, 0.7, 1]
34
+ - filter: mlp
35
+ value: [1, 0.5, 0.7, 0.3, 0]
36
+ - value: 0.5
37
+ dtype: bfloat16
38
+ ```
39
+
40
+ ## 💻 Usage
41
+
42
+ ```python
43
+ !pip install -qU transformers accelerate
44
+
45
+ from transformers import AutoTokenizer
46
+ import transformers
47
+ import torch
48
+
49
+ model = "Corianas/Microllama-300.500kmerge"
50
+ messages = [{"role": "user", "content": "What is a large language model?"}]
51
+
52
+ tokenizer = AutoTokenizer.from_pretrained(model)
53
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
54
+ pipeline = transformers.pipeline(
55
+ "text-generation",
56
+ model=model,
57
+ torch_dtype=torch.float16,
58
+ device_map="auto",
59
+ )
60
+
61
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
62
+ print(outputs[0]["generated_text"])
63
+ ```
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Corianas/Microllama_Char_300k_step",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 2048,
14
+ "max_position_embeddings": 2048,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 12,
18
+ "num_key_value_heads": 12,
19
+ "pad_token_id": 0,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
+ "tie_word_embeddings": true,
25
+ "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.40.0",
27
+ "use_cache": true,
28
+ "vocab_size": 341
29
+ }
mergekit_config.yml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ slices:
3
+ - sources:
4
+ - model: Corianas/Microllama_Char_500k_step
5
+ layer_range: [0, 12]
6
+ - model: Corianas/Microllama_Char_300k_step
7
+ layer_range: [0, 12]
8
+ merge_method: slerp
9
+ base_model: Corianas/Microllama_Char_300k_step
10
+ parameters:
11
+ t:
12
+ - filter: self_attn
13
+ value: [0, 0.5, 0.3, 0.7, 1]
14
+ - filter: mlp
15
+ value: [1, 0.5, 0.7, 0.3, 0]
16
+ - value: 0.5
17
+ dtype: bfloat16
model-00001-of-00001.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e14f4aed5c90705a0c6c12f9c1b3ede8b175667a31b8ee899c8d4eaadb870bc
3
+ size 170967552
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 170955264}, "weight_map": {"lm_head.weight": "model-00001-of-00001.safetensors", "model.embed_tokens.weight": "model-00001-of-00001.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.input_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00001.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00001.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00001.safetensors", "model.norm.weight": "model-00001-of-00001.safetensors"}}
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
@@ -0,0 +1,483 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "<unk>",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "<s>",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "</s>",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ }
33
+ ],
34
+ "normalizer": {
35
+ "type": "Sequence",
36
+ "normalizers": [
37
+ {
38
+ "type": "Prepend",
39
+ "prepend": "▁"
40
+ },
41
+ {
42
+ "type": "Replace",
43
+ "pattern": {
44
+ "String": " "
45
+ },
46
+ "content": "▁"
47
+ }
48
+ ]
49
+ },
50
+ "pre_tokenizer": null,
51
+ "post_processor": {
52
+ "type": "TemplateProcessing",
53
+ "single": [
54
+ {
55
+ "SpecialToken": {
56
+ "id": "<s>",
57
+ "type_id": 0
58
+ }
59
+ },
60
+ {
61
+ "Sequence": {
62
+ "id": "A",
63
+ "type_id": 0
64
+ }
65
+ }
66
+ ],
67
+ "pair": [
68
+ {
69
+ "SpecialToken": {
70
+ "id": "<s>",
71
+ "type_id": 0
72
+ }
73
+ },
74
+ {
75
+ "Sequence": {
76
+ "id": "A",
77
+ "type_id": 0
78
+ }
79
+ },
80
+ {
81
+ "SpecialToken": {
82
+ "id": "<s>",
83
+ "type_id": 1
84
+ }
85
+ },
86
+ {
87
+ "Sequence": {
88
+ "id": "B",
89
+ "type_id": 1
90
+ }
91
+ }
92
+ ],
93
+ "special_tokens": {
94
+ "<s>": {
95
+ "id": "<s>",
96
+ "ids": [
97
+ 1
98
+ ],
99
+ "tokens": [
100
+ "<s>"
101
+ ]
102
+ }
103
+ }
104
+ },
105
+ "decoder": {
106
+ "type": "Sequence",
107
+ "decoders": [
108
+ {
109
+ "type": "Replace",
110
+ "pattern": {
111
+ "String": "▁"
112
+ },
113
+ "content": " "
114
+ },
115
+ {
116
+ "type": "ByteFallback"
117
+ },
118
+ {
119
+ "type": "Fuse"
120
+ },
121
+ {
122
+ "type": "Strip",
123
+ "content": " ",
124
+ "start": 1,
125
+ "stop": 0
126
+ }
127
+ ]
128
+ },
129
+ "model": {
130
+ "type": "BPE",
131
+ "dropout": null,
132
+ "unk_token": "<unk>",
133
+ "continuing_subword_prefix": null,
134
+ "end_of_word_suffix": null,
135
+ "fuse_unk": true,
136
+ "byte_fallback": true,
137
+ "ignore_merges": false,
138
+ "vocab": {
139
+ "<unk>": 0,
140
+ "<s>": 1,
141
+ "</s>": 2,
142
+ "<0x00>": 3,
143
+ "<0x01>": 4,
144
+ "<0x02>": 5,
145
+ "<0x03>": 6,
146
+ "<0x04>": 7,
147
+ "<0x05>": 8,
148
+ "<0x06>": 9,
149
+ "<0x07>": 10,
150
+ "<0x08>": 11,
151
+ "<0x09>": 12,
152
+ "<0x0A>": 13,
153
+ "<0x0B>": 14,
154
+ "<0x0C>": 15,
155
+ "<0x0D>": 16,
156
+ "<0x0E>": 17,
157
+ "<0x0F>": 18,
158
+ "<0x10>": 19,
159
+ "<0x11>": 20,
160
+ "<0x12>": 21,
161
+ "<0x13>": 22,
162
+ "<0x14>": 23,
163
+ "<0x15>": 24,
164
+ "<0x16>": 25,
165
+ "<0x17>": 26,
166
+ "<0x18>": 27,
167
+ "<0x19>": 28,
168
+ "<0x1A>": 29,
169
+ "<0x1B>": 30,
170
+ "<0x1C>": 31,
171
+ "<0x1D>": 32,
172
+ "<0x1E>": 33,
173
+ "<0x1F>": 34,
174
+ "<0x20>": 35,
175
+ "<0x21>": 36,
176
+ "<0x22>": 37,
177
+ "<0x23>": 38,
178
+ "<0x24>": 39,
179
+ "<0x25>": 40,
180
+ "<0x26>": 41,
181
+ "<0x27>": 42,
182
+ "<0x28>": 43,
183
+ "<0x29>": 44,
184
+ "<0x2A>": 45,
185
+ "<0x2B>": 46,
186
+ "<0x2C>": 47,
187
+ "<0x2D>": 48,
188
+ "<0x2E>": 49,
189
+ "<0x2F>": 50,
190
+ "<0x30>": 51,
191
+ "<0x31>": 52,
192
+ "<0x32>": 53,
193
+ "<0x33>": 54,
194
+ "<0x34>": 55,
195
+ "<0x35>": 56,
196
+ "<0x36>": 57,
197
+ "<0x37>": 58,
198
+ "<0x38>": 59,
199
+ "<0x39>": 60,
200
+ "<0x3A>": 61,
201
+ "<0x3B>": 62,
202
+ "<0x3C>": 63,
203
+ "<0x3D>": 64,
204
+ "<0x3E>": 65,
205
+ "<0x3F>": 66,
206
+ "<0x40>": 67,
207
+ "<0x41>": 68,
208
+ "<0x42>": 69,
209
+ "<0x43>": 70,
210
+ "<0x44>": 71,
211
+ "<0x45>": 72,
212
+ "<0x46>": 73,
213
+ "<0x47>": 74,
214
+ "<0x48>": 75,
215
+ "<0x49>": 76,
216
+ "<0x4A>": 77,
217
+ "<0x4B>": 78,
218
+ "<0x4C>": 79,
219
+ "<0x4D>": 80,
220
+ "<0x4E>": 81,
221
+ "<0x4F>": 82,
222
+ "<0x50>": 83,
223
+ "<0x51>": 84,
224
+ "<0x52>": 85,
225
+ "<0x53>": 86,
226
+ "<0x54>": 87,
227
+ "<0x55>": 88,
228
+ "<0x56>": 89,
229
+ "<0x57>": 90,
230
+ "<0x58>": 91,
231
+ "<0x59>": 92,
232
+ "<0x5A>": 93,
233
+ "<0x5B>": 94,
234
+ "<0x5C>": 95,
235
+ "<0x5D>": 96,
236
+ "<0x5E>": 97,
237
+ "<0x5F>": 98,
238
+ "<0x60>": 99,
239
+ "<0x61>": 100,
240
+ "<0x62>": 101,
241
+ "<0x63>": 102,
242
+ "<0x64>": 103,
243
+ "<0x65>": 104,
244
+ "<0x66>": 105,
245
+ "<0x67>": 106,
246
+ "<0x68>": 107,
247
+ "<0x69>": 108,
248
+ "<0x6A>": 109,
249
+ "<0x6B>": 110,
250
+ "<0x6C>": 111,
251
+ "<0x6D>": 112,
252
+ "<0x6E>": 113,
253
+ "<0x6F>": 114,
254
+ "<0x70>": 115,
255
+ "<0x71>": 116,
256
+ "<0x72>": 117,
257
+ "<0x73>": 118,
258
+ "<0x74>": 119,
259
+ "<0x75>": 120,
260
+ "<0x76>": 121,
261
+ "<0x77>": 122,
262
+ "<0x78>": 123,
263
+ "<0x79>": 124,
264
+ "<0x7A>": 125,
265
+ "<0x7B>": 126,
266
+ "<0x7C>": 127,
267
+ "<0x7D>": 128,
268
+ "<0x7E>": 129,
269
+ "<0x7F>": 130,
270
+ "<0x80>": 131,
271
+ "<0x81>": 132,
272
+ "<0x82>": 133,
273
+ "<0x83>": 134,
274
+ "<0x84>": 135,
275
+ "<0x85>": 136,
276
+ "<0x86>": 137,
277
+ "<0x87>": 138,
278
+ "<0x88>": 139,
279
+ "<0x89>": 140,
280
+ "<0x8A>": 141,
281
+ "<0x8B>": 142,
282
+ "<0x8C>": 143,
283
+ "<0x8D>": 144,
284
+ "<0x8E>": 145,
285
+ "<0x8F>": 146,
286
+ "<0x90>": 147,
287
+ "<0x91>": 148,
288
+ "<0x92>": 149,
289
+ "<0x93>": 150,
290
+ "<0x94>": 151,
291
+ "<0x95>": 152,
292
+ "<0x96>": 153,
293
+ "<0x97>": 154,
294
+ "<0x98>": 155,
295
+ "<0x99>": 156,
296
+ "<0x9A>": 157,
297
+ "<0x9B>": 158,
298
+ "<0x9C>": 159,
299
+ "<0x9D>": 160,
300
+ "<0x9E>": 161,
301
+ "<0x9F>": 162,
302
+ "<0xA0>": 163,
303
+ "<0xA1>": 164,
304
+ "<0xA2>": 165,
305
+ "<0xA3>": 166,
306
+ "<0xA4>": 167,
307
+ "<0xA5>": 168,
308
+ "<0xA6>": 169,
309
+ "<0xA7>": 170,
310
+ "<0xA8>": 171,
311
+ "<0xA9>": 172,
312
+ "<0xAA>": 173,
313
+ "<0xAB>": 174,
314
+ "<0xAC>": 175,
315
+ "<0xAD>": 176,
316
+ "<0xAE>": 177,
317
+ "<0xAF>": 178,
318
+ "<0xB0>": 179,
319
+ "<0xB1>": 180,
320
+ "<0xB2>": 181,
321
+ "<0xB3>": 182,
322
+ "<0xB4>": 183,
323
+ "<0xB5>": 184,
324
+ "<0xB6>": 185,
325
+ "<0xB7>": 186,
326
+ "<0xB8>": 187,
327
+ "<0xB9>": 188,
328
+ "<0xBA>": 189,
329
+ "<0xBB>": 190,
330
+ "<0xBC>": 191,
331
+ "<0xBD>": 192,
332
+ "<0xBE>": 193,
333
+ "<0xBF>": 194,
334
+ "<0xC0>": 195,
335
+ "<0xC1>": 196,
336
+ "<0xC2>": 197,
337
+ "<0xC3>": 198,
338
+ "<0xC4>": 199,
339
+ "<0xC5>": 200,
340
+ "<0xC6>": 201,
341
+ "<0xC7>": 202,
342
+ "<0xC8>": 203,
343
+ "<0xC9>": 204,
344
+ "<0xCA>": 205,
345
+ "<0xCB>": 206,
346
+ "<0xCC>": 207,
347
+ "<0xCD>": 208,
348
+ "<0xCE>": 209,
349
+ "<0xCF>": 210,
350
+ "<0xD0>": 211,
351
+ "<0xD1>": 212,
352
+ "<0xD2>": 213,
353
+ "<0xD3>": 214,
354
+ "<0xD4>": 215,
355
+ "<0xD5>": 216,
356
+ "<0xD6>": 217,
357
+ "<0xD7>": 218,
358
+ "<0xD8>": 219,
359
+ "<0xD9>": 220,
360
+ "<0xDA>": 221,
361
+ "<0xDB>": 222,
362
+ "<0xDC>": 223,
363
+ "<0xDD>": 224,
364
+ "<0xDE>": 225,
365
+ "<0xDF>": 226,
366
+ "<0xE0>": 227,
367
+ "<0xE1>": 228,
368
+ "<0xE2>": 229,
369
+ "<0xE3>": 230,
370
+ "<0xE4>": 231,
371
+ "<0xE5>": 232,
372
+ "<0xE6>": 233,
373
+ "<0xE7>": 234,
374
+ "<0xE8>": 235,
375
+ "<0xE9>": 236,
376
+ "<0xEA>": 237,
377
+ "<0xEB>": 238,
378
+ "<0xEC>": 239,
379
+ "<0xED>": 240,
380
+ "<0xEE>": 241,
381
+ "<0xEF>": 242,
382
+ "<0xF0>": 243,
383
+ "<0xF1>": 244,
384
+ "<0xF2>": 245,
385
+ "<0xF3>": 246,
386
+ "<0xF4>": 247,
387
+ "<0xF5>": 248,
388
+ "<0xF6>": 249,
389
+ "<0xF7>": 250,
390
+ "<0xF8>": 251,
391
+ "<0xF9>": 252,
392
+ "<0xFA>": 253,
393
+ "<0xFB>": 254,
394
+ "<0xFC>": 255,
395
+ "<0xFD>": 256,
396
+ "<0xFE>": 257,
397
+ "<0xFF>": 258,
398
+ "▁": 259,
399
+ "e": 260,
400
+ "t": 261,
401
+ "a": 262,
402
+ "o": 263,
403
+ "h": 264,
404
+ "n": 265,
405
+ "s": 266,
406
+ "i": 267,
407
+ "r": 268,
408
+ "d": 269,
409
+ "l": 270,
410
+ "u": 271,
411
+ "w": 272,
412
+ "m": 273,
413
+ "↨": 274,
414
+ "g": 275,
415
+ "c": 276,
416
+ "f": 277,
417
+ "y": 278,
418
+ ".": 279,
419
+ "p": 280,
420
+ ",": 281,
421
+ "b": 282,
422
+ "\r": 283,
423
+ "k": 284,
424
+ "v": 285,
425
+ "\"": 286,
426
+ "'": 287,
427
+ "j": 288,
428
+ "x": 289,
429
+ "z": 290,
430
+ "q": 291,
431
+ "§": 292,
432
+ "?": 293,
433
+ "-": 294,
434
+ "!": 295,
435
+ ":": 296,
436
+ "1": 297,
437
+ "0": 298,
438
+ "2": 299,
439
+ "5": 300,
440
+ "3": 301,
441
+ ";": 302,
442
+ "4": 303,
443
+ "9": 304,
444
+ "8": 305,
445
+ "$": 306,
446
+ "6": 307,
447
+ "7": 308,
448
+ "+": 309,
449
+ "=": 310,
450
+ ")": 311,
451
+ "(": 312,
452
+ "&": 313,
453
+ "/": 314,
454
+ "%": 315,
455
+ "^": 316,
456
+ "*": 317,
457
+ "`": 318,
458
+ "_": 319,
459
+ "]": 320,
460
+ "<": 321,
461
+ "[": 322,
462
+ "\\": 323,
463
+ ">": 324,
464
+ "{": 325,
465
+ "}": 326,
466
+ "|": 327,
467
+ "¼": 328,
468
+ "½": 329,
469
+ "π": 330,
470
+ "←": 331,
471
+ "↑": 332,
472
+ "→": 333,
473
+ "↓": 334,
474
+ "▲": 335,
475
+ "►": 336,
476
+ "▼": 337,
477
+ "◄": 338,
478
+ "♪": 339,
479
+ "♫": 340
480
+ },
481
+ "merges": []
482
+ }
483
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a9ecce7fdafd04b12010f1b0b21feb89a2c0495b6ff899fac1ba9a285f4abc5
3
+ size 5543
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }