vukosi commited on
Commit
54689dc
1 Parent(s): f959343

Upload 5 files

Browse files
Files changed (5) hide show
  1. config.json +65 -0
  2. eval_results.txt +4 -0
  3. pytorch_model.bin +3 -0
  4. tokenizer.json +0 -0
  5. vocab.json +0 -0
config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../models/PuoBERTa-resume-v4/",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "X",
15
+ "1": "ADJ",
16
+ "10": "PART",
17
+ "11": "PRON",
18
+ "12": "PROPN",
19
+ "13": "PUNCT",
20
+ "14": "SCONJ",
21
+ "15": "SYM",
22
+ "16": "VERB",
23
+ "2": "ADP",
24
+ "3": "ADV",
25
+ "4": "AUX",
26
+ "5": "CCONJ",
27
+ "6": "DET",
28
+ "7": "INTJ",
29
+ "8": "NOUN",
30
+ "9": "NUM"
31
+ },
32
+ "initializer_range": 0.02,
33
+ "intermediate_size": 3072,
34
+ "label2id": {
35
+ "ADJ": 1,
36
+ "ADP": 2,
37
+ "ADV": 3,
38
+ "AUX": 4,
39
+ "CCONJ": 5,
40
+ "DET": 6,
41
+ "INTJ": 7,
42
+ "NOUN": 8,
43
+ "NUM": 9,
44
+ "PART": 10,
45
+ "PRON": 11,
46
+ "PROPN": 12,
47
+ "PUNCT": 13,
48
+ "SCONJ": 14,
49
+ "SYM": 15,
50
+ "VERB": 16,
51
+ "X": 0
52
+ },
53
+ "layer_norm_eps": 1e-12,
54
+ "max_position_embeddings": 514,
55
+ "model_type": "roberta",
56
+ "num_attention_heads": 12,
57
+ "num_hidden_layers": 6,
58
+ "pad_token_id": 1,
59
+ "position_embedding_type": "absolute",
60
+ "torch_dtype": "float32",
61
+ "transformers_version": "4.32.1",
62
+ "type_vocab_size": 1,
63
+ "use_cache": true,
64
+ "vocab_size": 52000
65
+ }
eval_results.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ f1 = 0.8610216934919525
2
+ loss = 0.5192539895835676
3
+ precision = 0.8613833660039205
4
+ recall = 0.8606603245663123
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1ace8795af73685f02bea5165d91f0bc9ba3b3ec8cdb518ebe60af1d9f7957b
3
+ size 331528489
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff