Safetensors
xlm-roberta
affiliations
ner
science
nicolauduran45 commited on
Commit
3c7601a
1 Parent(s): 85f670c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +261 -3
README.md CHANGED
@@ -1,3 +1,261 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - af
5
+ - am
6
+ - ar
7
+ - as
8
+ - az
9
+ - be
10
+ - bg
11
+ - bn
12
+ - br
13
+ - bs
14
+ - ca
15
+ - cs
16
+ - cy
17
+ - da
18
+ - de
19
+ - el
20
+ - en
21
+ - eo
22
+ - es
23
+ - et
24
+ - eu
25
+ - fa
26
+ - fi
27
+ - fr
28
+ - fy
29
+ - ga
30
+ - gd
31
+ - gl
32
+ - gu
33
+ - ha
34
+ - he
35
+ - hi
36
+ - hr
37
+ - hu
38
+ - hy
39
+ - id
40
+ - is
41
+ - it
42
+ - ja
43
+ - jv
44
+ - ka
45
+ - kk
46
+ - km
47
+ - kn
48
+ - ko
49
+ - ku
50
+ - ky
51
+ - la
52
+ - lo
53
+ - lt
54
+ - lv
55
+ - mg
56
+ - mk
57
+ - ml
58
+ - mn
59
+ - mr
60
+ - ms
61
+ - my
62
+ - ne
63
+ - nl
64
+ - 'no'
65
+ - om
66
+ - or
67
+ - pa
68
+ - pl
69
+ - ps
70
+ - pt
71
+ - ro
72
+ - ru
73
+ - sa
74
+ - sd
75
+ - si
76
+ - sk
77
+ - sl
78
+ - so
79
+ - sq
80
+ - sr
81
+ - su
82
+ - sv
83
+ - sw
84
+ - ta
85
+ - te
86
+ - th
87
+ - tl
88
+ - tr
89
+ - ug
90
+ - uk
91
+ - ur
92
+ - uz
93
+ - vi
94
+ - xh
95
+ - yi
96
+ - zh
97
+ base_model:
98
+ - SIRIS-Lab/affilgood-affilxlm
99
+ tags:
100
+ - affiliations
101
+ - ner
102
+ - science
103
+ ---
104
+
105
+ # AffilGood-SPAN
106
+
107
+ ## Overview
108
+
109
+ <details>
110
+ <summary>Click to expand</summary>
111
+
112
+ - **Model type:** Language Model
113
+ - **Architecture:** XLM-RoBERTa-base
114
+ - **Language:** Multilingual
115
+ - **License:** Apache 2.0
116
+ - **Task:** Named Entity Recognition
117
+ - **Data:** AffilGood-SPAN
118
+ - **Additional Resources:**
119
+ - [Paper](https://https://aclanthology.org/2024.sdp-1.13/)
120
+ - [GitHub](https://github.com/sirisacademic/affilgood)
121
+ </details>
122
+
123
+ ## Model description
124
+
125
+ The multilingual version of **affilgood-span-multilingual** is a Named Entity Recognition (NER) model for identifying span of single affiliation strings in raw affiliation strings from scientific papers and projects, fine-tuned from the [AffilRoberta](https://huggingface.co/SIRIS-Lab/affilgood-affilroberta) model, a [RoBERTa](https://arxiv.org/abs/1907.11692) base model futher pre-trained for MLM task on a medium-size corpus of raw affiliation stirngs collected from OpenAlex.
126
+
127
+ Raw affiliation span identification task is aimed at extracting and cleaning affiliation strings when there is noise and/or when there are multiple affiliation strings in the same signature.
128
+ Typically, multiple institutions have been considered to be separated by semicolons. However, other punctuation marks, spaces or *and* connectors are frequently used to separate affiliations.
129
+
130
+ We have annotated a dataset containing 2,072 raw affiliation strings obtained from OpenAlex to identify spans containing relevant affiliation data within them.
131
+ The annotated instances were selected by a stratified random sampling by country, focusing on ensuring diversity in affiliation languages and origins.
132
+ Additional manually-chosen instances with noisy sequences were included in the annotated data so we could train our model to filter out non-affiliation strings.
133
+ It is frequent that affiliation data automatically extracted from PDF files contain texts that should have been discarded (e.g. email, acknowledgements or part of the contents of the publication). These data can introduce errors in the subsequent steps of the pipeline.
134
+
135
+ We have fine-tuned the models for 25 epochs, using 80% of the dataset for training, 10% for validation and 10% for testing. Best performing model is available here.
136
+
137
+ ## Intended Usage
138
+
139
+ This model is intended to be used for multilingual raw affiliation strings, because this model is pre-trained on XLM-RoBERTa, NER and large further pre-training corpora are both multilingual.
140
+
141
+ ## How to use
142
+
143
+ ```python
144
+ from transformers import pipeline
145
+ affilgood_span_pipeline("ner", model=model, tokenizer=tokenizer, aggregation_strategy="simple")
146
+ sentence = "Deanery of Biomedical Sciences, University of Edinburgh, and TW2Informatics Ltd, Gothenburg, 42166, Sweden / Received in revised form February 22, 2016; ENaC, epithelial sodium channel"
147
+ output = affilgood_span_pipeline(sentence)
148
+ print(output)
149
+ ```
150
+
151
+ ## Limitations and bias
152
+
153
+ No measures have been taken to estimate the bias and toxicity embedded in the model.
154
+
155
+ We have annotated a dataset containing 2,072 raw affiliation strings obtained from OpenAlex to identify spans containing relevant affiliation data within them.
156
+ The annotated instances were selected by a stratified random sampling by country, focusing on ensuring diversity in affiliation languages and origins.
157
+ Additional manually-chosen instances with noisy sequences were included in the annotated data so we could train our model to filter out non-affiliation strings.
158
+ It is frequent that affiliation data automatically extracted from PDF files contain texts that should have been discarded (e.g. email, acknowledgements or part of the contents of the publication). These data can introduce errors in the subsequent steps of the pipeline.
159
+
160
+ ## Training
161
+
162
+ We used the [AffilGood-SPAN dataset](link) for training and evaluation.
163
+
164
+ We fine-tuned the adapted and base models for token classification with the IOB annotation schema.
165
+ We trained the models for 25 epochs, using 80% of the dataset for training, 10% for validation and 10% for testing.
166
+
167
+ Hyperparameters used for training are described here:
168
+ - Learning Rate: 2e-5
169
+ - Learning Rate Decay: Linear
170
+ - Weight Decay: 0.01
171
+ - Warmup Portion: 0.06
172
+ - Batch Size: 128
173
+ - Number of Steps: 25k steps
174
+ - Adam ε: 1e-6
175
+ - Adam β<sub>1</sub>: 0.9
176
+ - Adam β<sub>2</sub>: 0.999
177
+
178
+ The **best performing epoch (considering macro-averaged F1 with *strict* matching criteria) was used to select the model**.
179
+
180
+ ### Evaluation
181
+
182
+ The model's performance was evaluated on a 10% of the dataset.
183
+
184
+ | **Model** | **Exact F1** | **Partial F1** |
185
+ |--------------------------|--------------|----------------|
186
+ | Semicolon split (baseline) | .793 | .907 |
187
+ | RoBERTa | .929 | .981 |
188
+ | XLM | .931 | .978 |
189
+ | AffilGood-RoBERTa | **.938** | **.981** |
190
+ | **AffilGood-XLM (this model)**| .927 | .979 |
191
+
192
+ All the numbers reported above represent F1-score with *strict* and *partial* match, when both the boundaries and types of the entities match.
193
+
194
+ ## Additional information
195
+
196
+ ### Authors
197
+
198
+ - SIRIS Lab, Research Division of SIRIS Academic, Barcelona, Spain
199
+ - LaSTUS Lab, TALN Group, Universitat Pompeu Fabra, Barcelona, Spain
200
+ - Institute of Computer Science, Polish Academy of Sciences, Warsaw, Poland
201
+
202
+ ### Contact
203
+
204
+ For further information, send an email to either <[email protected]> or <[email protected]>.
205
+
206
+ ### License
207
+
208
+ This work is distributed under a [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).
209
+
210
+ ### Funding
211
+
212
+ This work was partially funded and supporter by:
213
+ - Industrial Doctorates Plan of the Department of Research and Universities of the Generalitat de Catalunya, by Departament de Recerca i Universitats de la Generalitat de Catalunya (ajuts SGR-Cat 2021),
214
+ - Maria de Maeztu Units of Excellence Programme CEX2021-001195-M, funded by MCIN/AEI /10.13039/501100011033
215
+ - EU HORIZON SciLake (Grant Agreement 101058573)
216
+ - EU HORIZON ERINIA (Grant Agreement 101060930)
217
+
218
+ ### Citation
219
+
220
+ ```bibtex
221
+ @inproceedings{duran-silva-etal-2024-affilgood,
222
+ title = "{A}ffil{G}ood: Building reliable institution name disambiguation tools to improve scientific literature analysis",
223
+ author = "Duran-Silva, Nicolau and
224
+ Accuosto, Pablo and
225
+ Przyby{\l}a, Piotr and
226
+ Saggion, Horacio",
227
+ editor = "Ghosal, Tirthankar and
228
+ Singh, Amanpreet and
229
+ Waard, Anita and
230
+ Mayr, Philipp and
231
+ Naik, Aakanksha and
232
+ Weller, Orion and
233
+ Lee, Yoonjoo and
234
+ Shen, Shannon and
235
+ Qin, Yanxia",
236
+ booktitle = "Proceedings of the Fourth Workshop on Scholarly Document Processing (SDP 2024)",
237
+ month = aug,
238
+ year = "2024",
239
+ address = "Bangkok, Thailand",
240
+ publisher = "Association for Computational Linguistics",
241
+ url = "https://aclanthology.org/2024.sdp-1.13",
242
+ pages = "135--144",
243
+ }
244
+ ```
245
+
246
+ ### Disclaimer
247
+
248
+ <details>
249
+ <summary>Click to expand</summary>
250
+
251
+ The model published in this repository is intended for a generalist purpose
252
+ and is made available to third parties under a Apache v2.0 License.
253
+
254
+ Please keep in mind that the model may have bias and/or any other undesirable distortions.
255
+ When third parties deploy or provide systems and/or services to other parties using this model
256
+ (or a system based on it) or become users of the model itself, they should note that it is under
257
+ their responsibility to mitigate the risks arising from its use and, in any event, to comply with
258
+ applicable regulations, including regulations regarding the use of Artificial Intelligence.
259
+
260
+ In no event shall the owners and creators of the model be liable for any results arising from the use made by third parties.
261
+ </details>