AswanthCManoj commited on
Commit
0c354ef
1 Parent(s): a6df6cf

Training in progress, step 75

Browse files
adapter_config.json CHANGED
@@ -1,17 +1,26 @@
1
  {
 
2
  "auto_mapping": null,
3
- "base_model_name_or_path": "deepseek-ai/deepseek-coder-1.3b-instruct",
4
- "encoder_dropout": 0.0,
5
- "encoder_hidden_size": 250,
6
- "encoder_num_layers": 2,
7
- "encoder_reparameterization_type": "MLP",
8
  "inference_mode": true,
9
- "num_attention_heads": 16,
10
- "num_layers": 24,
11
- "num_transformer_submodules": 1,
12
- "num_virtual_tokens": 50,
13
- "peft_type": "P_TUNING",
 
 
 
 
 
 
14
  "revision": null,
15
- "task_type": "CAUSAL_LM",
16
- "token_dim": 2048
 
 
 
 
17
  }
 
1
  {
2
+ "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "teknium/OpenHermes-2.5-Mistral-7B",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
 
 
7
  "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layers_pattern": null,
10
+ "layers_to_transform": null,
11
+ "lora_alpha": 32,
12
+ "lora_dropout": 0.1,
13
+ "modules_to_save": [
14
+ "embed_tokens"
15
+ ],
16
+ "peft_type": "LORA",
17
+ "r": 16,
18
+ "rank_pattern": {},
19
  "revision": null,
20
+ "target_modules": [
21
+ "gate_proj",
22
+ "up_proj",
23
+ "down_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79acdcc1b8e57db310c2b04aab07db81dd60615a67b42d7a105b2b40f171d1e8
3
- size 409720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab69c31afaffb1acf340f47e56d2f4deaa97626231f62effadcb10ae6d5c88b2
3
+ size 637592768
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 32000,
3
+ "<|im_start|>": 32001
4
+ }
special_tokens_map.json CHANGED
@@ -1,17 +1,6 @@
1
  {
2
- "bos_token": {
3
- "content": "<|begin▁of▁sentence|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|EOT|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": "Ġsq"
17
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "pad_token": "</s>",
5
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
6
  }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json CHANGED
@@ -2,192 +2,60 @@
2
  "add_bos_token": true,
3
  "add_eos_token": true,
4
  "added_tokens_decoder": {
5
- "32000": {
6
- "content": "õ",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": false
12
- },
13
- "32001": {
14
- "content": "÷",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": false
20
- },
21
- "32002": {
22
- "content": "Á",
23
- "lstrip": false,
24
- "normalized": true,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": false
28
- },
29
- "32003": {
30
- "content": "ý",
31
- "lstrip": false,
32
- "normalized": true,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": false
36
- },
37
- "32004": {
38
- "content": "À",
39
- "lstrip": false,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": false
44
- },
45
- "32005": {
46
- "content": "ÿ",
47
- "lstrip": false,
48
- "normalized": true,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": false
52
- },
53
- "32006": {
54
- "content": "ø",
55
- "lstrip": false,
56
- "normalized": true,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": false
60
- },
61
- "32007": {
62
- "content": "ú",
63
- "lstrip": false,
64
- "normalized": true,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": false
68
- },
69
- "32008": {
70
- "content": "þ",
71
- "lstrip": false,
72
- "normalized": true,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": false
76
- },
77
- "32009": {
78
- "content": "ü",
79
- "lstrip": false,
80
- "normalized": true,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": false
84
- },
85
- "32010": {
86
- "content": "ù",
87
  "lstrip": false,
88
- "normalized": true,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": false
92
- },
93
- "32011": {
94
- "content": "ö",
95
- "lstrip": false,
96
- "normalized": true,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": false
100
- },
101
- "32012": {
102
- "content": "û",
103
- "lstrip": false,
104
- "normalized": true,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": false
108
- },
109
- "32013": {
110
- "content": "<|begin▁of▁sentence|>",
111
- "lstrip": false,
112
- "normalized": true,
113
  "rstrip": false,
114
  "single_word": false,
115
  "special": true
116
  },
117
- "32014": {
118
- "content": "<|end▁of▁sentence|>",
119
  "lstrip": false,
120
- "normalized": true,
121
  "rstrip": false,
122
  "single_word": false,
123
  "special": true
124
  },
125
- "32015": {
126
- "content": "<|fim▁hole|>",
127
  "lstrip": false,
128
- "normalized": true,
129
  "rstrip": false,
130
  "single_word": false,
131
- "special": false
132
- },
133
- "32016": {
134
- "content": "<|fim▁begin|>",
135
- "lstrip": false,
136
- "normalized": true,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "32017": {
142
- "content": "<|fim▁end|>",
143
- "lstrip": false,
144
- "normalized": true,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "32018": {
150
- "content": "<pad>",
151
- "lstrip": false,
152
- "normalized": true,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "32019": {
158
- "content": "<|User|>",
159
- "lstrip": false,
160
- "normalized": true,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
  },
165
- "32020": {
166
- "content": "<|Assistant|>",
167
  "lstrip": false,
168
- "normalized": true,
169
  "rstrip": false,
170
  "single_word": false,
171
- "special": false
172
  },
173
- "32021": {
174
- "content": "<|EOT|>",
175
  "lstrip": false,
176
- "normalized": true,
177
  "rstrip": false,
178
  "single_word": false,
179
  "special": true
180
  }
181
  },
182
- "bos_token": "<|begin▁of▁sentence|>",
183
- "chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{bos_token}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
 
184
  "clean_up_tokenization_spaces": false,
185
- "eos_token": "<|EOT|>",
186
  "legacy": true,
187
- "model_max_length": 16384,
188
- "pad_token": "Ġsq",
189
  "sp_model_kwargs": {},
 
190
  "tokenizer_class": "LlamaTokenizer",
191
- "unk_token": null,
192
- "use_default_system_prompt": false
 
 
193
  }
 
2
  "add_bos_token": true,
3
  "add_eos_token": true,
4
  "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  "lstrip": false,
8
+ "normalized": false,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
12
  },
13
+ "1": {
14
+ "content": "<s>",
15
  "lstrip": false,
16
+ "normalized": false,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
20
  },
21
+ "2": {
22
+ "content": "</s>",
23
  "lstrip": false,
24
+ "normalized": false,
25
  "rstrip": false,
26
  "single_word": false,
27
+ "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  },
29
+ "32000": {
30
+ "content": "<|im_end|>",
31
  "lstrip": false,
32
+ "normalized": false,
33
  "rstrip": false,
34
  "single_word": false,
35
+ "special": true
36
  },
37
+ "32001": {
38
+ "content": "<|im_start|>",
39
  "lstrip": false,
40
+ "normalized": false,
41
  "rstrip": false,
42
  "single_word": false,
43
  "special": true
44
  }
45
  },
46
+ "additional_special_tokens": [],
47
+ "bos_token": "<s>",
48
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
49
  "clean_up_tokenization_spaces": false,
50
+ "eos_token": "</s>",
51
  "legacy": true,
52
+ "model_max_length": 1000000000000000019884624838656,
53
+ "pad_token": "</s>",
54
  "sp_model_kwargs": {},
55
+ "spaces_between_special_tokens": false,
56
  "tokenizer_class": "LlamaTokenizer",
57
+ "trust_remote_code": false,
58
+ "unk_token": "<unk>",
59
+ "use_default_system_prompt": true,
60
+ "use_fast": true
61
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:759434762d7bfb55a21f68b18a6e6125b5035fbd73274ea67425ff0b6cc062ad
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b239c79424204a0c8ffc8edf94952bccae3e9a354372d5dc4f94a81aad08e77e
3
  size 4664