Zeb commited on
Commit
ff34e0d
1 Parent(s): 2a2cdda

Upload model

Browse files
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 3,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 3,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_embd": 768,
14
+ "n_head": 12,
15
+ "n_inner": 3072,
16
+ "n_layer": 12,
17
+ "n_positions": 256,
18
+ "reorder_and_upcast_attn": false,
19
+ "resid_pdrop": 0.1,
20
+ "scale_attn_by_inverse_layer_idx": false,
21
+ "scale_attn_weights": true,
22
+ "summary_activation": null,
23
+ "summary_first_dropout": 0.1,
24
+ "summary_proj_to_labels": true,
25
+ "summary_type": "cls_index",
26
+ "summary_use_proj": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.43.3",
29
+ "use_cache": true,
30
+ "vocab_size": 51
31
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452fe745e87696095ab7b0b2bda9d3a8fdf41527e9ea659af9d1d7e269232182
3
+ size 341182080
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "UTT_BOUNDARY",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "UTT_BOUNDARY",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "PAD",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "UNK",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "UNK",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "PAD",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 3,
26
+ "content": "UTT_BOUNDARY",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ }
33
+ ],
34
+ "normalizer": {
35
+ "type": "Sequence",
36
+ "normalizers": [
37
+ {
38
+ "type": "Replace",
39
+ "pattern": {
40
+ "String": " WORD_BOUNDARY"
41
+ },
42
+ "content": ""
43
+ },
44
+ {
45
+ "type": "Strip",
46
+ "strip_left": true,
47
+ "strip_right": true
48
+ }
49
+ ]
50
+ },
51
+ "pre_tokenizer": {
52
+ "type": "Whitespace"
53
+ },
54
+ "post_processor": {
55
+ "type": "TemplateProcessing",
56
+ "single": [
57
+ {
58
+ "SpecialToken": {
59
+ "id": "UTT_BOUNDARY",
60
+ "type_id": 0
61
+ }
62
+ },
63
+ {
64
+ "Sequence": {
65
+ "id": "A",
66
+ "type_id": 0
67
+ }
68
+ }
69
+ ],
70
+ "pair": [
71
+ {
72
+ "SpecialToken": {
73
+ "id": "UTT_BOUNDARY",
74
+ "type_id": 0
75
+ }
76
+ },
77
+ {
78
+ "Sequence": {
79
+ "id": "A",
80
+ "type_id": 0
81
+ }
82
+ },
83
+ {
84
+ "SpecialToken": {
85
+ "id": "UTT_BOUNDARY",
86
+ "type_id": 0
87
+ }
88
+ },
89
+ {
90
+ "Sequence": {
91
+ "id": "B",
92
+ "type_id": 1
93
+ }
94
+ }
95
+ ],
96
+ "special_tokens": {
97
+ "UTT_BOUNDARY": {
98
+ "id": "UTT_BOUNDARY",
99
+ "ids": [
100
+ 3
101
+ ],
102
+ "tokens": [
103
+ "UTT_BOUNDARY"
104
+ ]
105
+ }
106
+ }
107
+ },
108
+ "decoder": null,
109
+ "model": {
110
+ "type": "WordLevel",
111
+ "vocab": {
112
+ "UNK": 0,
113
+ "PAD": 1,
114
+ "WORD_BOUNDARY": 2,
115
+ "UTT_BOUNDARY": 3,
116
+ "j": 4,
117
+ "ɛ": 5,
118
+ "h": 6,
119
+ "k": 7,
120
+ "ɑ": 8,
121
+ "m": 9,
122
+ "p": 10,
123
+ "aʊ": 11,
124
+ "n": 12,
125
+ "d": 13,
126
+ "z": 14,
127
+ "θ": 15,
128
+ "ɪ": 16,
129
+ "ŋ": 17,
130
+ "l": 18,
131
+ "aɪ": 19,
132
+ "s": 20,
133
+ "ɜː": 21,
134
+ "t": 22,
135
+ "w": 23,
136
+ "v": 24,
137
+ "ð": 25,
138
+ "æ": 26,
139
+ "ɔ": 27,
140
+ "ɹ": 28,
141
+ "ʌ": 29,
142
+ "f": 30,
143
+ "ə": 31,
144
+ "b": 32,
145
+ "iː": 33,
146
+ "eɪ": 34,
147
+ "oʊ": 35,
148
+ "d̠ʒ": 36,
149
+ "i": 37,
150
+ "uː": 38,
151
+ "iə": 39,
152
+ "ʊ": 40,
153
+ "ɡ": 41,
154
+ "t̠ʃ": 42,
155
+ "ɔɪ": 43,
156
+ "ʃ": 44,
157
+ "ʒ": 45,
158
+ "r": 46,
159
+ "x": 47,
160
+ "ɬ": 48,
161
+ "ɑ̃": 49,
162
+ "nʲ": 50
163
+ },
164
+ "unk_token": "UNK"
165
+ }
166
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "UNK",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "PAD",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "3": {
21
+ "content": "UTT_BOUNDARY",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "bos_token": "UTT_BOUNDARY",
30
+ "clean_up_tokenization_spaces": true,
31
+ "eos_token": "UTT_BOUNDARY",
32
+ "model_max_length": 1000000000000000019884624838656,
33
+ "pad_token": "PAD",
34
+ "tokenizer_class": "GPT2Tokenizer",
35
+ "unk_token": "UNK"
36
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07b857f0c7f5ff42d26627cc5c529e93d9d5e7281f236267235c06475167c2c6
3
+ size 5368
vocab.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"UNK":0,"PAD":1,"WORD_BOUNDARY":2,"UTT_BOUNDARY":3,"j":4,"ɛ":5,"h":6,"k":7,"ɑ":8,"m":9,"p":10,"aʊ":11,"n":12,"d":13,"z":14,"θ":15,"ɪ":16,"ŋ":17,"l":18,"aɪ":19,"s":20,"ɜː":21,"t":22,"w":23,"v":24,"ð":25,"æ":26,"ɔ":27,"ɹ":28,"ʌ":29,"f":30,"ə":31,"b":32,"iː":33,"eɪ":34,"oʊ":35,"d̠ʒ":36,"i":37,"uː":38,"iə":39,"ʊ":40,"ɡ":41,"t̠ʃ":42,"ɔɪ":43,"ʃ":44,"ʒ":45,"r":46,"x":47,"ɬ":48,"ɑ̃":49,"nʲ":50}