lafias commited on
Commit
a5ead5a
1 Parent(s): 24e841d

Adding config

Browse files
Files changed (1) hide show
  1. config.cfg +142 -0
config.cfg ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [paths]
2
+ train = "../ner/train.spacy"
3
+ dev = "../ner/dev.spacy"
4
+ vectors = null
5
+ init_tok2vec = null
6
+
7
+ [system]
8
+ gpu_allocator = "pytorch"
9
+ seed = 0
10
+
11
+ [nlp]
12
+ lang = "en"
13
+ pipeline = ["transformer","ner"]
14
+ batch_size = 128
15
+ disabled = []
16
+ before_creation = null
17
+ after_creation = null
18
+ after_pipeline_creation = null
19
+ tokenizer = {"@tokenizers":"spacy.Tokenizer.v1"}
20
+
21
+ [components]
22
+
23
+ [components.ner]
24
+ factory = "ner"
25
+ incorrect_spans_key = null
26
+ moves = null
27
+ update_with_oracle_cut_size = 100
28
+
29
+ [components.ner.model]
30
+ @architectures = "spacy.TransitionBasedParser.v2"
31
+ state_type = "ner"
32
+ extra_state_tokens = false
33
+ hidden_width = 64
34
+ maxout_pieces = 2
35
+ use_upper = false
36
+ nO = null
37
+
38
+ [components.ner.model.tok2vec]
39
+ @architectures = "spacy-transformers.TransformerListener.v1"
40
+ grad_factor = 1.0
41
+ pooling = {"@layers":"reduce_mean.v1"}
42
+ upstream = "*"
43
+
44
+ [components.transformer]
45
+ factory = "transformer"
46
+ max_batch_items = 4096
47
+ set_extra_annotations = {"@annotation_setters":"spacy-transformers.null_annotation_setter.v1"}
48
+
49
+ [components.transformer.model]
50
+ @architectures = "spacy-transformers.TransformerModel.v1"
51
+ name = "roberta-base"
52
+
53
+ [components.transformer.model.get_spans]
54
+ @span_getters = "spacy-transformers.strided_spans.v1"
55
+ window = 128
56
+ stride = 96
57
+
58
+ [components.transformer.model.tokenizer_config]
59
+ use_fast = true
60
+
61
+ [corpora]
62
+
63
+ [corpora.dev]
64
+ @readers = "spacy.Corpus.v1"
65
+ path = ${paths.dev}
66
+ max_length = 0
67
+ gold_preproc = false
68
+ limit = 0
69
+ augmenter = null
70
+
71
+ [corpora.train]
72
+ @readers = "spacy.Corpus.v1"
73
+ path = ${paths.train}
74
+ max_length = 500
75
+ gold_preproc = false
76
+ limit = 0
77
+ augmenter = null
78
+
79
+ [training]
80
+ accumulate_gradient = 3
81
+ dev_corpus = "corpora.dev"
82
+ train_corpus = "corpora.train"
83
+ seed = ${system.seed}
84
+ gpu_allocator = ${system.gpu_allocator}
85
+ dropout = 0.5958824689
86
+ patience = 1600
87
+ max_epochs = 0
88
+ max_steps = 20000
89
+ eval_frequency = 200
90
+ frozen_components = []
91
+ annotating_components = []
92
+ before_to_disk = null
93
+
94
+ [training.batcher]
95
+ @batchers = "spacy.batch_by_padded.v1"
96
+ discard_oversize = true
97
+ size = 2000
98
+ buffer = 256
99
+ get_length = null
100
+
101
+ [training.logger]
102
+ @loggers = "spacy.WandbLogger.v2"
103
+ project_name = "bib-tuning"
104
+ remove_config_values = []
105
+ model_log_interval = null
106
+ log_dataset_dir = null
107
+
108
+ [training.optimizer]
109
+ @optimizers = "Adam.v1"
110
+ beta1 = 0.9
111
+ beta2 = 0.999
112
+ L2_is_weight_decay = true
113
+ L2 = 0.01
114
+ grad_clip = 1.0
115
+ use_averages = false
116
+ eps = 0.00000001
117
+
118
+ [training.optimizer.learn_rate]
119
+ @schedules = "warmup_linear.v1"
120
+ warmup_steps = 250
121
+ total_steps = 20000
122
+ initial_rate = 0.00005
123
+
124
+ [training.score_weights]
125
+ ents_f = 1.0
126
+ ents_p = 0.0
127
+ ents_r = 0.0
128
+ ents_per_type = null
129
+
130
+ [pretraining]
131
+
132
+ [initialize]
133
+ vectors = null
134
+ init_tok2vec = ${paths.init_tok2vec}
135
+ vocab_data = null
136
+ lookups = null
137
+ before_init = null
138
+ after_init = null
139
+
140
+ [initialize.components]
141
+
142
+ [initialize.tokenizer]