samabos commited on
Commit
5b81c09
1 Parent(s): 137de7b

Training in progress, epoch 1

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
config.json ADDED
@@ -0,0 +1,354 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-small",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "id2label": {
11
+ "0": 105.13,
12
+ "1": 105.94,
13
+ "2": 106.0,
14
+ "3": 106.12,
15
+ "4": 106.2,
16
+ "5": 106.31,
17
+ "6": 205.0,
18
+ "7": 301.93,
19
+ "8": 302.84,
20
+ "9": 305.54,
21
+ "10": 305.72,
22
+ "11": 601.1,
23
+ "12": 603.19,
24
+ "13": 603.9,
25
+ "14": 710.21,
26
+ "15": 710.22,
27
+ "16": 710.8,
28
+ "17": 713.1,
29
+ "18": 713.3,
30
+ "19": 713.39,
31
+ "20": 801.11,
32
+ "21": 802.9,
33
+ "22": 811.2,
34
+ "23": 811.9,
35
+ "24": 907.2,
36
+ "25": 908.12,
37
+ "26": 910.12,
38
+ "27": 1209.24,
39
+ "28": 1522.0,
40
+ "29": 1601.0,
41
+ "30": 1702.6,
42
+ "31": 1901.1,
43
+ "32": 2006.0,
44
+ "33": 2101.12,
45
+ "34": 2101.2,
46
+ "35": 2101.3,
47
+ "36": 2207.2,
48
+ "37": 2208.4,
49
+ "38": 2305.0,
50
+ "39": 2309.1,
51
+ "40": 2515.12,
52
+ "41": 2516.12,
53
+ "42": 2521.0,
54
+ "43": 2803.0,
55
+ "44": 2918.91,
56
+ "45": 2931.35,
57
+ "46": 3002.13,
58
+ "47": 3002.14,
59
+ "48": 3002.15,
60
+ "49": 3207.2,
61
+ "50": 3405.2,
62
+ "51": 3914.0,
63
+ "52": 3917.31,
64
+ "53": 3922.2,
65
+ "54": 4003.0,
66
+ "55": 4013.1,
67
+ "56": 4017.0,
68
+ "57": 4101.9,
69
+ "58": 4115.1,
70
+ "59": 4301.9,
71
+ "60": 4302.2,
72
+ "61": 4413.0,
73
+ "62": 4415.1,
74
+ "63": 4504.1,
75
+ "64": 4707.2,
76
+ "65": 4817.2,
77
+ "66": 4819.1,
78
+ "67": 4819.2,
79
+ "68": 4819.3,
80
+ "69": 4820.4,
81
+ "70": 4903.0,
82
+ "71": 4909.0,
83
+ "72": 5104.0,
84
+ "73": 5702.5,
85
+ "74": 5801.21,
86
+ "75": 5801.23,
87
+ "76": 5801.27,
88
+ "77": 5801.31,
89
+ "78": 5801.33,
90
+ "79": 5801.37,
91
+ "80": 5806.1,
92
+ "81": 6001.1,
93
+ "82": 6502.0,
94
+ "83": 6507.0,
95
+ "84": 6602.0,
96
+ "85": 6603.2,
97
+ "86": 7001.0,
98
+ "87": 7003.12,
99
+ "88": 7004.2,
100
+ "89": 7005.21,
101
+ "90": 7009.1,
102
+ "91": 7016.1,
103
+ "92": 7107.0,
104
+ "93": 7109.0,
105
+ "94": 7111.0,
106
+ "95": 7204.41,
107
+ "96": 7208.1,
108
+ "97": 7208.4,
109
+ "98": 7211.13,
110
+ "99": 7302.3,
111
+ "100": 7323.1,
112
+ "101": 7403.21,
113
+ "102": 7407.21,
114
+ "103": 7408.21,
115
+ "104": 7408.22,
116
+ "105": 7409.4,
117
+ "106": 7411.21,
118
+ "107": 7411.22,
119
+ "108": 7609.0,
120
+ "109": 8102.95,
121
+ "110": 8205.3,
122
+ "111": 8209.0,
123
+ "112": 8306.3,
124
+ "113": 8430.1,
125
+ "114": 8441.2,
126
+ "115": 8443.91,
127
+ "116": 8452.9,
128
+ "117": 8483.1,
129
+ "118": 8483.3,
130
+ "119": 8483.6,
131
+ "120": 8483.9,
132
+ "121": 8507.1,
133
+ "122": 8517.11,
134
+ "123": 8519.2,
135
+ "124": 8525.8,
136
+ "125": 8528.52,
137
+ "126": 8540.2,
138
+ "127": 8543.7,
139
+ "128": 8545.9,
140
+ "129": 8607.3,
141
+ "130": 8703.21,
142
+ "131": 8703.22,
143
+ "132": 8703.23,
144
+ "133": 8703.24,
145
+ "134": 8703.31,
146
+ "135": 8703.32,
147
+ "136": 8703.33,
148
+ "137": 8703.4,
149
+ "138": 8708.21,
150
+ "139": 8708.5,
151
+ "140": 8714.91,
152
+ "141": 8714.96,
153
+ "142": 8907.0,
154
+ "143": 9002.11,
155
+ "144": 9018.41,
156
+ "145": 9021.4,
157
+ "146": 9033.0,
158
+ "147": 9111.2,
159
+ "148": 9301.2,
160
+ "149": 9401.3,
161
+ "150": 9603.5,
162
+ "151": 9606.1,
163
+ "152": 9607.11,
164
+ "153": 9616.1,
165
+ "154": 9616.2,
166
+ "155": 9617.0,
167
+ "156": 9618.0
168
+ },
169
+ "initializer_range": 0.02,
170
+ "intermediate_size": 3072,
171
+ "label2id": {
172
+ "105.13": 0,
173
+ "105.94": 1,
174
+ "106.0": 2,
175
+ "106.12": 3,
176
+ "106.2": 4,
177
+ "106.31": 5,
178
+ "205.0": 6,
179
+ "301.93": 7,
180
+ "302.84": 8,
181
+ "305.54": 9,
182
+ "305.72": 10,
183
+ "601.1": 11,
184
+ "603.19": 12,
185
+ "603.9": 13,
186
+ "710.21": 14,
187
+ "710.22": 15,
188
+ "710.8": 16,
189
+ "713.1": 17,
190
+ "713.3": 18,
191
+ "713.39": 19,
192
+ "801.11": 20,
193
+ "802.9": 21,
194
+ "811.2": 22,
195
+ "811.9": 23,
196
+ "907.2": 24,
197
+ "908.12": 25,
198
+ "910.12": 26,
199
+ "1209.24": 27,
200
+ "1522.0": 28,
201
+ "1601.0": 29,
202
+ "1702.6": 30,
203
+ "1901.1": 31,
204
+ "2006.0": 32,
205
+ "2101.12": 33,
206
+ "2101.2": 34,
207
+ "2101.3": 35,
208
+ "2207.2": 36,
209
+ "2208.4": 37,
210
+ "2305.0": 38,
211
+ "2309.1": 39,
212
+ "2515.12": 40,
213
+ "2516.12": 41,
214
+ "2521.0": 42,
215
+ "2803.0": 43,
216
+ "2918.91": 44,
217
+ "2931.35": 45,
218
+ "3002.13": 46,
219
+ "3002.14": 47,
220
+ "3002.15": 48,
221
+ "3207.2": 49,
222
+ "3405.2": 50,
223
+ "3914.0": 51,
224
+ "3917.31": 52,
225
+ "3922.2": 53,
226
+ "4003.0": 54,
227
+ "4013.1": 55,
228
+ "4017.0": 56,
229
+ "4101.9": 57,
230
+ "4115.1": 58,
231
+ "4301.9": 59,
232
+ "4302.2": 60,
233
+ "4413.0": 61,
234
+ "4415.1": 62,
235
+ "4504.1": 63,
236
+ "4707.2": 64,
237
+ "4817.2": 65,
238
+ "4819.1": 66,
239
+ "4819.2": 67,
240
+ "4819.3": 68,
241
+ "4820.4": 69,
242
+ "4903.0": 70,
243
+ "4909.0": 71,
244
+ "5104.0": 72,
245
+ "5702.5": 73,
246
+ "5801.21": 74,
247
+ "5801.23": 75,
248
+ "5801.27": 76,
249
+ "5801.31": 77,
250
+ "5801.33": 78,
251
+ "5801.37": 79,
252
+ "5806.1": 80,
253
+ "6001.1": 81,
254
+ "6502.0": 82,
255
+ "6507.0": 83,
256
+ "6602.0": 84,
257
+ "6603.2": 85,
258
+ "7001.0": 86,
259
+ "7003.12": 87,
260
+ "7004.2": 88,
261
+ "7005.21": 89,
262
+ "7009.1": 90,
263
+ "7016.1": 91,
264
+ "7107.0": 92,
265
+ "7109.0": 93,
266
+ "7111.0": 94,
267
+ "7204.41": 95,
268
+ "7208.1": 96,
269
+ "7208.4": 97,
270
+ "7211.13": 98,
271
+ "7302.3": 99,
272
+ "7323.1": 100,
273
+ "7403.21": 101,
274
+ "7407.21": 102,
275
+ "7408.21": 103,
276
+ "7408.22": 104,
277
+ "7409.4": 105,
278
+ "7411.21": 106,
279
+ "7411.22": 107,
280
+ "7609.0": 108,
281
+ "8102.95": 109,
282
+ "8205.3": 110,
283
+ "8209.0": 111,
284
+ "8306.3": 112,
285
+ "8430.1": 113,
286
+ "8441.2": 114,
287
+ "8443.91": 115,
288
+ "8452.9": 116,
289
+ "8483.1": 117,
290
+ "8483.3": 118,
291
+ "8483.6": 119,
292
+ "8483.9": 120,
293
+ "8507.1": 121,
294
+ "8517.11": 122,
295
+ "8519.2": 123,
296
+ "8525.8": 124,
297
+ "8528.52": 125,
298
+ "8540.2": 126,
299
+ "8543.7": 127,
300
+ "8545.9": 128,
301
+ "8607.3": 129,
302
+ "8703.21": 130,
303
+ "8703.22": 131,
304
+ "8703.23": 132,
305
+ "8703.24": 133,
306
+ "8703.31": 134,
307
+ "8703.32": 135,
308
+ "8703.33": 136,
309
+ "8703.4": 137,
310
+ "8708.21": 138,
311
+ "8708.5": 139,
312
+ "8714.91": 140,
313
+ "8714.96": 141,
314
+ "8907.0": 142,
315
+ "9002.11": 143,
316
+ "9018.41": 144,
317
+ "9021.4": 145,
318
+ "9033.0": 146,
319
+ "9111.2": 147,
320
+ "9301.2": 148,
321
+ "9401.3": 149,
322
+ "9603.5": 150,
323
+ "9606.1": 151,
324
+ "9607.11": 152,
325
+ "9616.1": 153,
326
+ "9616.2": 154,
327
+ "9617.0": 155,
328
+ "9618.0": 156
329
+ },
330
+ "layer_norm_eps": 1e-07,
331
+ "max_position_embeddings": 512,
332
+ "max_relative_positions": -1,
333
+ "model_type": "deberta-v2",
334
+ "norm_rel_ebd": "layer_norm",
335
+ "num_attention_heads": 12,
336
+ "num_hidden_layers": 6,
337
+ "pad_token_id": 0,
338
+ "pooler_dropout": 0,
339
+ "pooler_hidden_act": "gelu",
340
+ "pooler_hidden_size": 768,
341
+ "pos_att_type": [
342
+ "p2c",
343
+ "c2p"
344
+ ],
345
+ "position_biased_input": false,
346
+ "position_buckets": 256,
347
+ "problem_type": "multi_label_classification",
348
+ "relative_attention": true,
349
+ "share_att_key": true,
350
+ "torch_dtype": "float32",
351
+ "transformers_version": "4.44.2",
352
+ "type_vocab_size": 0,
353
+ "vocab_size": 128100
354
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb0ed97f4044575e3aff4932f2458a7c6540407f1a5a02547ff580c35b6c08ef
3
+ size 568075348
runs/Sep22_13-50-23_965508b51b08/events.out.tfevents.1727013025.965508b51b08.347.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73da1d3fdfada313bd1b9953271197688af400c21288f54d19d9d1e46cda6f4f
3
+ size 12814
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe3c4d25eba88cec974544385398b7efad99a48e88121e63ccc52338f1fa18a4
3
+ size 5240