clincolnoz
commited on
Commit
•
231a361
1
Parent(s):
f46b89f
first commit, epoch 10
Browse files- README.md +9 -0
- added_tokens.json +126 -0
- config.json +26 -0
- optimizer.pt +3 -0
- pytorch_model.bin +3 -0
- rng_state.pth +3 -0
- scaler.pt +3 -0
- scheduler.pt +3 -0
- special_tokens_map.json +7 -0
- tokenizer.json +0 -0
- tokenizer_config.json +13 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
README.md
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
metrics:
|
5 |
+
- accuracy
|
6 |
+
pipeline_tag: fill-mask
|
7 |
+
tags:
|
8 |
+
- not-for-all-audiences
|
9 |
+
---
|
added_tokens.json
ADDED
@@ -0,0 +1,126 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"#gabfam": 30571,
|
3 |
+
"#maga": 30626,
|
4 |
+
"#speakfreely": 30610,
|
5 |
+
"&": 30549,
|
6 |
+
"​": 30531,
|
7 |
+
">": 30533,
|
8 |
+
">i": 30639,
|
9 |
+
">the": 30530,
|
10 |
+
"100%": 30548,
|
11 |
+
"[[url]]": 30568,
|
12 |
+
"[url]": 30556,
|
13 |
+
"[user]": 30552,
|
14 |
+
"ain't": 30623,
|
15 |
+
"alot": 30603,
|
16 |
+
"and/or": 30563,
|
17 |
+
"anyways": 30529,
|
18 |
+
"aren’t": 30601,
|
19 |
+
"assholes": 30606,
|
20 |
+
"autistic": 30591,
|
21 |
+
"blackpill": 30587,
|
22 |
+
"bruh": 30523,
|
23 |
+
"btw": 30589,
|
24 |
+
"cant": 30569,
|
25 |
+
"can’t": 30644,
|
26 |
+
"chads": 30611,
|
27 |
+
"congrats": 30534,
|
28 |
+
"couldn’t": 30580,
|
29 |
+
"covid": 30621,
|
30 |
+
"cringe": 30555,
|
31 |
+
"cuz": 30637,
|
32 |
+
"delusional": 30613,
|
33 |
+
"didnt": 30604,
|
34 |
+
"didn’t": 30619,
|
35 |
+
"doesnt": 30609,
|
36 |
+
"doesn’t": 30595,
|
37 |
+
"dont": 30628,
|
38 |
+
"don’t": 30636,
|
39 |
+
"downvoted": 30542,
|
40 |
+
"dudes": 30583,
|
41 |
+
"dunno": 30541,
|
42 |
+
"foids": 30600,
|
43 |
+
"fucks": 30586,
|
44 |
+
"gab": 30645,
|
45 |
+
"haha": 30643,
|
46 |
+
"hahaha": 30575,
|
47 |
+
"hateful": 30592,
|
48 |
+
"haven’t": 30550,
|
49 |
+
"he’s": 30631,
|
50 |
+
"hobbies": 30522,
|
51 |
+
"idk": 30525,
|
52 |
+
"imo": 30578,
|
53 |
+
"incel": 30620,
|
54 |
+
"incels": 30624,
|
55 |
+
"insecure": 30630,
|
56 |
+
"irl": 30559,
|
57 |
+
"isnt": 30527,
|
58 |
+
"isn’t": 30537,
|
59 |
+
"it'll": 30545,
|
60 |
+
"it’s": 30588,
|
61 |
+
"ive": 30539,
|
62 |
+
"i’d": 30618,
|
63 |
+
"i’ll": 30558,
|
64 |
+
"i’m": 30560,
|
65 |
+
"i’ve": 30582,
|
66 |
+
"jfl": 30570,
|
67 |
+
"leftists": 30625,
|
68 |
+
"legit": 30554,
|
69 |
+
"lmao": 30599,
|
70 |
+
"lol": 30538,
|
71 |
+
"ltr": 30638,
|
72 |
+
"meme": 30564,
|
73 |
+
"memes": 30543,
|
74 |
+
"mentality": 30579,
|
75 |
+
"mgtow": 30585,
|
76 |
+
"mindset": 30590,
|
77 |
+
"mods": 30547,
|
78 |
+
"moron": 30605,
|
79 |
+
"morons": 30576,
|
80 |
+
"normie": 30635,
|
81 |
+
"normies": 30546,
|
82 |
+
"omg": 30562,
|
83 |
+
"people's": 30577,
|
84 |
+
"pilled": 30532,
|
85 |
+
"ppl": 30567,
|
86 |
+
"pua": 30557,
|
87 |
+
"reddit": 30584,
|
88 |
+
"redpill": 30536,
|
89 |
+
"retard": 30607,
|
90 |
+
"retarded": 30566,
|
91 |
+
"retards": 30597,
|
92 |
+
"she’s": 30622,
|
93 |
+
"shouldn’t": 30596,
|
94 |
+
"sidebar": 30528,
|
95 |
+
"simp": 30574,
|
96 |
+
"simps": 30615,
|
97 |
+
"sjw": 30641,
|
98 |
+
"sjws": 30614,
|
99 |
+
"smv": 30565,
|
100 |
+
"someone's": 30602,
|
101 |
+
"subreddit": 30612,
|
102 |
+
"subs": 30642,
|
103 |
+
"tbh": 30544,
|
104 |
+
"thats": 30633,
|
105 |
+
"that’s": 30629,
|
106 |
+
"theres": 30526,
|
107 |
+
"there’s": 30573,
|
108 |
+
"they’re": 30640,
|
109 |
+
"tinder": 30540,
|
110 |
+
"trp": 30616,
|
111 |
+
"upvote": 30561,
|
112 |
+
"wasn’t": 30598,
|
113 |
+
"we’re": 30581,
|
114 |
+
"whats": 30608,
|
115 |
+
"what’s": 30593,
|
116 |
+
"wont": 30551,
|
117 |
+
"won’t": 30524,
|
118 |
+
"wouldn’t": 30572,
|
119 |
+
"wtf": 30553,
|
120 |
+
"yea": 30627,
|
121 |
+
"youre": 30634,
|
122 |
+
"you’ll": 30535,
|
123 |
+
"you’re": 30617,
|
124 |
+
"you’ve": 30632,
|
125 |
+
"yup": 30594
|
126 |
+
}
|
config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "bert-base-uncased",
|
3 |
+
"architectures": [
|
4 |
+
"BertForPreTraining"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
+
"gradient_checkpointing": false,
|
9 |
+
"hidden_act": "gelu",
|
10 |
+
"hidden_dropout_prob": 0.1,
|
11 |
+
"hidden_size": 768,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 3072,
|
14 |
+
"layer_norm_eps": 1e-12,
|
15 |
+
"max_position_embeddings": 512,
|
16 |
+
"model_type": "bert",
|
17 |
+
"num_attention_heads": 12,
|
18 |
+
"num_hidden_layers": 12,
|
19 |
+
"pad_token_id": 0,
|
20 |
+
"position_embedding_type": "absolute",
|
21 |
+
"torch_dtype": "float32",
|
22 |
+
"transformers_version": "4.27.0.dev0",
|
23 |
+
"type_vocab_size": 2,
|
24 |
+
"use_cache": true,
|
25 |
+
"vocab_size": 30646
|
26 |
+
}
|
optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8403339132f725bce1a6674a1594fced6aca457e9cd7fadb8b1a30fcd6ed1033
|
3 |
+
size 881735429
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b071798c884d9ee4c472e62ca99e26cc31f1f01253bf622cd875e27850340fb
|
3 |
+
size 440881865
|
rng_state.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:23fdf9bc81503ab406c254ff0089dd1ef401196d8db34f8000b5a8893c7b1cfa
|
3 |
+
size 14575
|
scaler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a55b4470acc74a4765783c6241b601e0b370613e97ce3af429581ef93689074
|
3 |
+
size 557
|
scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f18605fa5face0401dcf97615d4dba2ebe08db5c350bff8bbd5f1df829df2599
|
3 |
+
size 627
|
special_tokens_map.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"mask_token": "[MASK]",
|
4 |
+
"pad_token": "[PAD]",
|
5 |
+
"sep_token": "[SEP]",
|
6 |
+
"unk_token": "[UNK]"
|
7 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"cls_token": "[CLS]",
|
3 |
+
"do_lower_case": true,
|
4 |
+
"mask_token": "[MASK]",
|
5 |
+
"model_max_length": 512,
|
6 |
+
"pad_token": "[PAD]",
|
7 |
+
"sep_token": "[SEP]",
|
8 |
+
"special_tokens_map_file": null,
|
9 |
+
"strip_accents": null,
|
10 |
+
"tokenize_chinese_chars": true,
|
11 |
+
"tokenizer_class": "BertTokenizer",
|
12 |
+
"unk_token": "[UNK]"
|
13 |
+
}
|
trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:122a872cb609b5fc05adf84a8d6d9565266277c61206274491cd92c2a443a99f
|
3 |
+
size 3515
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|