title
stringlengths 5
342
| author
stringlengths 3
2.17k
| year
int64 1.95k
2.02k
| abstract
stringlengths 0
12.7k
| pages
stringlengths 1
702
| queryID
stringlengths 4
40
| query
stringlengths 1
300
| paperID
stringlengths 0
40
| include
int64 0
1
|
---|---|---|---|---|---|---|---|---|
Improving Semantic Parsing via Answer Type Inference | Yavuz, Semih and
Gur, Izzeddin and
Su, Yu and
Srivatsa, Mudhakar and
Yan, Xifeng | 2,016 | nan | 149--159 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f3594f9d60c98cac88f9033c69c2b666713ed6d6 | 1 |
Automatic Extraction of Implicit Interpretations from Modal Constructions | Sanders, Jordan and
Blanco, Eduardo | 2,016 | nan | 1098--1107 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 813779418a613d1faecd7b1deb9b4456121a9b7e | 0 |
Factorizing Complex Models: A Case Study in Mention Detection | Florian, Radu and
Jing, Hongyan and
Kambhatla, Nanda and
Zitouni, Imed | 2,006 | nan | 473--480 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 3221a2c32488439c61bcfad832c50917e1ef3bdf | 1 |
Translation Context Sensitive {WSD} | Specia, Lucia and
das Gra{\c{c}}as Volpe Nunes, Maria and
Stevenson, Mark | 2,006 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | cffc77dc35c179166bb37124a759585cbcfd5d8a | 0 |
End-to-End Trainable Attentive Decoder for Hierarchical Entity Classification | Karn, Sanjeev and
Waltinger, Ulli and
Sch{\"u}tze, Hinrich | 2,017 | We address fine-grained entity classification and propose a novel attention-based recurrent neural network (RNN) encoder-decoder that generates paths in the type hierarchy and can be trained end-to-end. We show that our model performs better on fine-grained entity classification than prior work that relies on flat or local classifiers that do not directly model hierarchical structure. | 752--758 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 7615ed4f35f84cc086b9ae8e421891f3d33c68a6 | 1 |
Using Gaze Data to Predict Multiword Expressions | Rohanian, Omid and
Taslimipoor, Shiva and
Yaneva, Victoria and
Ha, Le An | 2,017 | In recent years gaze data has been increasingly used to improve and evaluate NLP models due to the fact that it carries information about the cognitive processing of linguistic phenomena. In this paper we conduct a preliminary study towards the automatic identification of multiword expressions based on gaze features from native and non-native speakers of English. We report comparisons between a part-of-speech (POS) and frequency baseline to: i) a prediction model based solely on gaze data and ii) a combined model of gaze data, POS and frequency. In spite of the challenging nature of the task, best performance was achieved by the latter. Furthermore, we explore how the type of gaze data (from native versus non-native speakers) affects the prediction, showing that data from the two groups is discriminative to an equal degree for the task. Finally, we show that late processing measures are more predictive than early ones, which is in line with previous research on idioms and other formulaic structures. | 601--609 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 6cbb05739c8d033eb027222c1a395391d877bc61 | 0 |
A Maximum Entropy Approach to Natural Language Processing | Berger, Adam L. and
Della Pietra, Stephen A. and
Della Pietra, Vincent J. | 1,996 | nan | 39--71 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | fb486e03369a64de2d5b0df86ec0a7b55d3907db | 1 |
Extracting Topics from Texts Based on Situations | Ma, Zhiyi and
Zhan, Xuegong and
Yao, Tianshun | 1,996 | nan | 357--362 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f8c4d4bd5f119e1d29e7fa38e81de6316817bda3 | 0 |
Attributed and Predictive Entity Embedding for Fine-Grained Entity Typing in Knowledge Bases | Jin, Hailong and
Hou, Lei and
Li, Juanzi and
Dong, Tiansi | 2,018 | Fine-grained entity typing aims at identifying the semantic type of an entity in KB. Type information is very important in knowledge bases, but are unfortunately incomplete even in some large knowledge bases. Limitations of existing methods are either ignoring the structure and type information in KB or requiring large scale annotated corpus. To address these issues, we propose an attributed and predictive entity embedding method, which can fully utilize various kinds of information comprehensively. Extensive experiments on two real DBpedia datasets show that our proposed method significantly outperforms 8 state-of-the-art methods, with 4.0{\%} and 5.2{\%} improvement in Mi-F1 and Ma-F1, respectively. | 282--292 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | fc958f9f9876689acbe99ad80c508ed7ad0e40bf | 1 |
{GKR}: the Graphical Knowledge Representation for semantic parsing | Kalouli, Aikaterini-Lida and
Crouch, Richard | 2,018 | This paper describes the first version of an open-source semantic parser that creates graphical representations of sentences to be used for further semantic processing, e.g. for natural language inference, reasoning and semantic similarity. The Graphical Knowledge Representation which is output by the parser is inspired by the Abstract Knowledge Representation, which separates out conceptual and contextual levels of representation that deal respectively with the subject matter of a sentence and its existential commitments. Our representation is a layered graph with each sub-graph holding different kinds of information, including one sub-graph for concepts and one for contexts. Our first evaluation of the system shows an F-score of 85{\%} in accurately representing sentences as semantic graphs. | 27--37 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0e195eb08cb94a2aedadae06442ee2d4e0cc1016 | 0 |
Imposing Label-Relational Inductive Bias for Extremely Fine-Grained Entity Typing | Xiong, Wenhan and
Wu, Jiawei and
Lei, Deren and
Yu, Mo and
Chang, Shiyu and
Guo, Xiaoxiao and
Wang, William Yang | 2,019 | Existing entity typing systems usually exploit the type hierarchy provided by knowledge base (KB) schema to model label correlations and thus improve the overall performance. Such techniques, however, are not directly applicable to more open and practical scenarios where the type set is not restricted by KB schema and includes a vast number of free-form types. To model the underlying label correlations without access to manually annotated label structures, we introduce a novel label-relational inductive bias, represented by a graph propagation layer that effectively encodes both global label co-occurrence statistics and word-level similarities. On a large dataset with over 10,000 free-form types, the graph-enhanced model equipped with an attention-based matching module is able to achieve a much higher recall score while maintaining a high-level precision. Specifically, it achieves a 15.3{\%} relative F1 improvement and also less inconsistency in the outputs. We further show that a simple modification of our proposed graph layer can also improve the performance on a conventional and widely-tested dataset that only includes KB-schema types. | 773--784 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a0713d945b2e5c2bdeeba68399c8ac6ea84e0ca6 | 1 |
Sample Size in {A}rabic Authorship Verification | Ahmed, Hossam | 2,019 | nan | 84--91 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f5f6cdf8deb778abc8469baed6824aeaf28289be | 0 |
The Life and Death of Discourse Entities: Identifying Singleton Mentions | Recasens, Marta and
de Marneffe, Marie-Catherine and
Potts, Christopher | 2,013 | nan | 627--633 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0fdf52226bde594ca069b2768249a4b10da9255d | 1 |
Annotating Legitimate Disagreement in Corpus Construction | Wong, Billy T.M. and
Lee, Sophia Y.M. | 2,013 | nan | 51--57 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0e575e0c10cf734b491b7a11b2fd47f5dcd7c33e | 0 |
Coreference in {W}ikipedia: Main Concept Resolution | Ghaddar, Abbas and
Langlais, Phillippe | 2,016 | nan | 229--238 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | da07e8d3951aeda87846f4b7db321576b48a2c60 | 1 |
Comparing Named-Entity Recognizers in a Targeted Domain: Handcrafted Rules vs Machine Learning | Partalas, Ioannis and
Lopez, C{\'e}dric and
Segond, Fr{\'e}d{\'e}rique | 2,016 | Comparing Named-Entity Recognizers in a Targeted Domain : Handcrafted Rules vs. Machine Learning Named-Entity Recognition concerns the classification of textual objects in a predefined set of categories such as persons, organizations, and localizations. While Named-Entity Recognition is well studied since 20 years, the application to specialized domains still poses challenges for current systems. We developed a rule-based system and two machine learning approaches to tackle the same task : recognition of product names, brand names, etc., in the domain of Cosmetics, for French. Our systems can thus be compared under ideal conditions. In this paper, we introduce both systems and we compare them. | 389--395 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f64f17a5be7e9a5e05e5a9fce1c01058ba024d1d | 0 |
Question Answering on Knowledge Bases and Text using Universal Schema and Memory Networks | Das, Rajarshi and
Zaheer, Manzil and
Reddy, Siva and
McCallum, Andrew | 2,017 | Existing question answering methods infer answers either from a knowledge base or from raw text. While knowledge base (KB) methods are good at answering compositional questions, their performance is often affected by the incompleteness of the KB. Au contraire, web text contains millions of facts that are absent in the KB, however in an unstructured form. Universal schema can support reasoning on the union of both structured KBs and unstructured text by aligning them in a common embedded space. In this paper we extend universal schema to natural language question answering, employing Memory networks to attend to the large body of facts in the combination of text and KB. Our models can be trained in an end-to-end fashion on question-answer pairs. Evaluation results on Spades fill-in-the-blank question answering dataset show that exploiting universal schema for question answering is better than using either a KB or text alone. This model also outperforms the current state-of-the-art by 8.5 F1 points. | 358--365 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 2b2090eab4abe27e6e5e4ca94afaf82e511b63bd | 1 |
Applying {BLAST} to Text Reuse Detection in {F}innish Newspapers and Journals, 1771-1910 | Vesanto, Aleksi and
Nivala, Asko and
Rantala, Heli and
Salakoski, Tapio and
Salmi, Hannu and
Ginter, Filip | 2,017 | nan | 54--58 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f69a2cf20e8a80cc9da747e2273f230b39847515 | 0 |
Embeddings of Label Components for Sequence Labeling: A Case Study of Fine-grained Named Entity Recognition | Kato, Takuma and
Abe, Kaori and
Ouchi, Hiroki and
Miyawaki, Shumpei and
Suzuki, Jun and
Inui, Kentaro | 2,020 | In general, the labels used in sequence labeling consist of different types of elements. For example, IOB-format entity labels, such as B-Person and I-Person, can be decomposed into span (B and I) and type information (Person). However, while most sequence labeling models do not consider such label components, the shared components across labels, such as Person, can be beneficial for label prediction. In this work, we propose to integrate label component information as embeddings into models. Through experiments on English and Japanese fine-grained named entity recognition, we demonstrate that the proposed method improves performance, especially for instances with low-frequency labels. | 222--229 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 926588efc881ba9f4322ddfaae14555d058bbb46 | 1 |
Autoencoding Keyword Correlation Graph for Document Clustering | Chiu, Billy and
Sahu, Sunil Kumar and
Thomas, Derek and
Sengupta, Neha and
Mahdy, Mohammady | 2,020 | Document clustering requires a deep understanding of the complex structure of long-text; in particular, the intra-sentential (local) and inter-sentential features (global). Existing representation learning models do not fully capture these features. To address this, we present a novel graph-based representation for document clustering that builds a \textit{graph autoencoder} (GAE) on a Keyword Correlation Graph. The graph is constructed with topical keywords as nodes and multiple local and global features as edges. A GAE is employed to aggregate the two sets of features by learning a latent representation which can jointly reconstruct them. Clustering is then performed on the learned representations, using vector dimensions as features for inducing document classes. Extensive experiments on two datasets show that the features learned by our approach can achieve better clustering performance than other existing features, including term frequency-inverse document frequency and average embedding. | 3974--3981 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e78b1c6cf0fbe935f12adf3a5ce3cde629252316 | 0 |
A Joint Named Entity Recognition and Entity Linking System | Stern, Rosa and
Sagot, Beno{\^\i}t and
B{\'e}chet, Fr{\'e}d{\'e}ric | 2,012 | nan | 52--60 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 443e814ab3b87ea51a18d4a3925a0fadeca03a9a | 1 |
Regular polysemy: A distributional model | Boleda, Gemma and
Pad{\'o}, Sebastian and
Utt, Jason | 2,012 | nan | 151--160 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 969fdeafe142994f4bf41ffc35ffea235de2aa18 | 0 |
Mining Entity Types from Query Logs via User Intent Modeling | Pantel, Patrick and
Lin, Thomas and
Gamon, Michael | 2,012 | nan | 563--571 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | c19b48e088983bf3ab71751000d78409293f4cf0 | 1 |
Nouvelle approche pour le regroupement des locuteurs dans des {\'e}missions radiophoniques et t{\'e}l{\'e}visuelles (New approach for speaker clustering of broadcast news) [in {F}rench] | Rouvier, Mickael and
Meignier, Sylvain | 2,012 | nan | 97--104 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ba643220c830a2db0b0d0e0aa3d556bc30d40b2b | 0 |
Label Embedding for Zero-shot Fine-grained Named Entity Typing | Ma, Yukun and
Cambria, Erik and
Gao, Sa | 2,016 | Named entity typing is the task of detecting the types of a named entity in context. For instance, given {``}Eric is giving a presentation{''}, our goal is to infer that {`}Eric{'} is a speaker or a presenter and a person. Existing approaches to named entity typing cannot work with a growing type set and fails to recognize entity mentions of unseen types. In this paper, we present a label embedding method that incorporates prototypical and hierarchical information to learn pre-trained label embeddings. In addition, we adapt a zero-shot learning framework that can predict both seen and previously unseen entity types. We perform evaluation on three benchmark datasets with two settings: 1) few-shots recognition where all types are covered by the training set; and 2) zero-shot recognition where fine-grained types are assumed absent from training set. Results show that prior knowledge encoded using our label embedding methods can significantly boost the performance of classification for both cases. | 171--180 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0d12035f96d795fef0d6b4f70340934dd3dd98a1 | 1 |
{ELMD}: An Automatically Generated Entity Linking Gold Standard Dataset in the Music Domain | Oramas, Sergio and
Anke, Luis Espinosa and
Sordo, Mohamed and
Saggion, Horacio and
Serra, Xavier | 2,016 | In this paper we present a gold standard dataset for Entity Linking (EL) in the Music Domain. It contains thousands of musical named entities such as Artist, Song or Record Label, which have been automatically annotated on a set of artist biographies coming from the Music website and social network Last.fm. The annotation process relies on the analysis of the hyperlinks present in the source texts and in a voting-based algorithm for EL, which considers, for each entity mention in text, the degree of agreement across three state-of-the-art EL systems. Manual evaluation shows that EL Precision is at least 94{\%}, and due to its tunable nature, it is possible to derive annotations favouring higher Precision or Recall, at will. We make available the annotated dataset along with evaluation data and the code. | 3312--3317 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a529e05211d685bbc6f80f1081e4784c325ea8d0 | 0 |
Fine-Grained Entity Typing via Hierarchical Multi Graph Convolutional Networks | Jin, Hailong and
Hou, Lei and
Li, Juanzi and
Dong, Tiansi | 2,019 | This paper addresses the problem of inferring the fine-grained type of an entity from a knowledge base. We convert this problem into the task of graph-based semi-supervised classification, and propose Hierarchical Multi Graph Convolutional Network (HMGCN), a novel Deep Learning architecture to tackle this problem. We construct three kinds of connectivity matrices to capture different kinds of semantic correlations between entities. A recursive regularization is proposed to model the subClassOf relations between types in given type hierarchy. Extensive experiments with two large-scale public datasets show that our proposed method significantly outperforms four state-of-the-art methods. | 4969--4978 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 074e3497b03366caf2e17acd59fb1c52ccf8be55 | 1 |
Multi-grained Attention with Object-level Grounding for Visual Question Answering | Huang, Pingping and
Huang, Jianhui and
Guo, Yuqing and
Qiao, Min and
Zhu, Yong | 2,019 | Attention mechanisms are widely used in Visual Question Answering (VQA) to search for visual clues related to the question. Most approaches train attention models from a coarse-grained association between sentences and images, which tends to fail on small objects or uncommon concepts. To address this problem, this paper proposes a multi-grained attention method. It learns explicit word-object correspondence by two types of word-level attention complementary to the sentence-image association. Evaluated on the VQA benchmark, the multi-grained attention model achieves competitive performance with state-of-the-art models. And the visualized attention maps demonstrate that addition of object-level groundings leads to a better understanding of the images and locates the attended objects more precisely. | 3595--3600 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 144f4d5dcd0b13935ff0d0890c2ec37aa40039b1 | 0 |
Automatic Acquisition of Hyponyms from Large Text Corpora | Hearst, Marti A. | 1,992 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | dbfd191afbbc8317577cbc44afe7156df546e143 | 1 |
The Acquisition of Lexical Knowledge from Combined Machine-Readable Dictionary Sources | Sanfilippo, Antonio and
Poznatlski, Victor | 1,992 | nan | 80--87 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0616b0f5e6edce01f153081e53bd0152c8d0a4bd | 0 |
Evaluating Pretrained Transformer-based Models on the Task of Fine-Grained Named Entity Recognition | Lothritz, Cedric and
Allix, Kevin and
Veiber, Lisa and
Bissyand{\'e}, Tegawend{\'e} F. and
Klein, Jacques | 2,020 | Named Entity Recognition (NER) is a fundamental Natural Language Processing (NLP) task and has remained an active research field. In recent years, transformer models and more specifically the BERT model developed at Google revolutionised the field of NLP. While the performance of transformer-based approaches such as BERT has been studied for NER, there has not yet been a study for the fine-grained Named Entity Recognition (FG-NER) task. In this paper, we compare three transformer-based models (BERT, RoBERTa, and XLNet) to two non-transformer-based models (CRF and BiLSTM-CNN-CRF). Furthermore, we apply each model to a multitude of distinct domains. We find that transformer-based models incrementally outperform the studied non-transformer-based models in most domains with respect to the F1 score. Furthermore, we find that the choice of domains significantly influenced the performance regardless of the respective data size or the model chosen. | 3750--3760 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e25dc08340655401a034a90bf091c1a185c422b6 | 1 |
Cross-lingual Semantic Representation for {NLP} with {UCCA} | Abend, Omri and
Dvir, Dotan and
Hershcovich, Daniel and
Prange, Jakob and
Schneider, Nathan | 2,020 | This is an introductory tutorial to UCCA (Universal Conceptual Cognitive Annotation), a cross-linguistically applicable framework for semantic representation, with corpora annotated in English, German and French, and ongoing annotation in Russian and Hebrew. UCCA builds on extensive typological work and supports rapid annotation. The tutorial will provide a detailed introduction to the UCCA annotation guidelines, design philosophy and the available resources; and a comparison to other meaning representations. It will also survey the existing parsing work, including the findings of three recent shared tasks, in SemEval and CoNLL, that addressed UCCA parsing. Finally, the tutorial will present recent applications and extensions to the scheme, demonstrating its value for natural language processing in a range of languages and domains. | 1--9 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 4ddb820ad3dceeb777967d98fbae6711de5077eb | 0 |
Fine-Grained Entity Typing in Hyperbolic Space | L{\'o}pez, Federico and
Heinzerling, Benjamin and
Strube, Michael | 2,019 | How can we represent hierarchical information present in large type inventories for entity typing? We study the suitability of hyperbolic embeddings to capture hierarchical relations between mentions in context and their target types in a shared vector space. We evaluate on two datasets and propose two different techniques to extract hierarchical information from the type inventory: from an expert-generated ontology and by automatically mining the dataset. The hyperbolic model shows improvements in some but not all cases over its Euclidean counterpart. Our analysis suggests that the adequacy of this geometry depends on the granularity of the type inventory and the representation of its distribution. | 169--180 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | cded59d31ab4841baa1517ff0359f0f2f4b865f5 | 1 |
The Impact of Semantic Linguistic Features in Relation Extraction: A Logical Relational Learning Approach | Lima, Rinaldo and
Espinasse, Bernard and
Freitas, Frederico | 2,019 | Relation Extraction (RE) consists in detecting and classifying semantic relations between entities in a sentence. The vast majority of the state-of-the-art RE systems relies on morphosyntactic features and supervised machine learning algorithms. This paper tries to answer important questions concerning both the impact of semantic based features, and the integration of external linguistic knowledge resources on RE performance. For that, a RE system based on a logical and relational learning algorithm was used and evaluated on three reference datasets from two distinct domains. The yielded results confirm that the classifiers induced using the proposed richer feature set outperformed the classifiers built with morphosyntactic features in average 4{\%} (F1-measure). | 648--654 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 72f07997e910d48a7e6557441c893ddca107c6f8 | 0 |
Automatic construction of a hypernym-labeled noun hierarchy from text | Caraballo, Sharon A. | 1,999 | nan | 120--126 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | aab329ef59d21060c31afce413f6e447b1c0b8b7 | 1 |
Improved Alignment Models for Statistical Machine Translation | Och, Franz Josef and
Tillmann, Christoph and
Ney, Hermann | 1,999 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 8b0495331238da6c0e7be0bfdb9b5453b33c1f98 | 0 |
Hierarchical Losses and New Resources for Fine-grained Entity Typing and Linking | Murty, Shikhar and
Verga, Patrick and
Vilnis, Luke and
Radovanovic, Irena and
McCallum, Andrew | 2,018 | Extraction from raw text to a knowledge base of entities and fine-grained types is often cast as prediction into a flat set of entity and type labels, neglecting the rich hierarchies over types and entities contained in curated ontologies. Previous attempts to incorporate hierarchical structure have yielded little benefit and are restricted to shallow ontologies. This paper presents new methods using real and complex bilinear mappings for integrating hierarchical information, yielding substantial improvement over flat predictions in entity linking and fine-grained entity typing, and achieving new state-of-the-art results for end-to-end models on the benchmark FIGER dataset. We also present two new human-annotated datasets containing wide and deep hierarchies which we will release to the community to encourage further research in this direction: \textit{MedMentions}, a collection of PubMed abstracts in which 246k mentions have been mapped to the massive UMLS ontology; and \textit{TypeNet}, which aligns Freebase types with the WordNet hierarchy to obtain nearly 2k entity types. In experiments on all three datasets we show substantial gains from hierarchy-aware training. | 97--109 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 35112824817b78156a6b2bcd2a5622a26ee16600 | 1 |
Findings of the {E}2{E} {NLG} Challenge | Du{\v{s}}ek, Ond{\v{r}}ej and
Novikova, Jekaterina and
Rieser, Verena | 2,018 | This paper summarises the experimental setup and results of the first shared task on end-to-end (E2E) natural language generation (NLG) in spoken dialogue systems. Recent end-to-end generation systems are promising since they reduce the need for data annotation. However, they are currently limited to small, delexicalised datasets. The E2E NLG shared task aims to assess whether these novel approaches can generate better-quality output by learning from a dataset containing higher lexical richness, syntactic complexity and diverse discourse phenomena. We compare 62 systems submitted by 17 institutions, covering a wide range of approaches, including machine learning architectures {--} with the majority implementing sequence-to-sequence models (seq2seq) {--} as well as systems based on grammatical rules and templates. | 322--328 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 92cfd6d2eb957805aaf4786dacb484081a469e80 | 0 |
No Noun Phrase Left Behind: Detecting and Typing Unlinkable Entities | Lin, Thomas and
{Mausam} and
Etzioni, Oren | 2,012 | nan | 893--903 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | cffb556ee3d1e188f4688b71a8608bbe1883bc49 | 1 |
Language Richness of the Web | Majli{\v{s}}, Martin and
{\v{Z}}abokrtsk{\'y}, Zden{\v{e}}k | 2,012 | We have built a corpus containing texts in 106 languages from texts available on the Internet and on Wikipedia. The W2C Web Corpus contains 54.7{\textasciitilde}GB of text and the W2C Wiki Corpus contains 8.5{\textasciitilde}GB of text. The W2C Web Corpus contains more than 100{\textasciitilde}MB of text available for 75 languages. At least 10{\textasciitilde}MB of text is available for 100 languages. These corpora are a unique data source for linguists, since they outclass all published works both in the size of the material collected and the number of languages covered. This language data resource can be of use particularly to researchers specialized in multilingual technologies development. We also developed software that greatly simplifies the creation of a new text corpus for a given language, using text materials freely available on the Internet. Special attention was given to components for filtering and de-duplication that allow to keep the material quality very high. | 2927--2934 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 33e7d35324cd138730022db297c358fc66149160 | 0 |
Collective Entity Resolution with Multi-Focal Attention | Globerson, Amir and
Lazic, Nevena and
Chakrabarti, Soumen and
Subramanya, Amarnag and
Ringgaard, Michael and
Pereira, Fernando | 2,016 | nan | 621--631 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 4988a269e9f61c6fd1da502e34648b93dfd1a54d | 1 |
Results of the 4th edition of {B}io{ASQ} Challenge | Krithara, Anastasia and
Nentidis, Anastasios and
Paliouras, Georgios and
Kakadiaris, Ioannis | 2,016 | nan | 1--7 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 2d5108706bfd88506c27adceb87ce46e75f9cad2 | 0 |
The Automatic Content Extraction ({ACE}) Program {--} Tasks, Data, and Evaluation | Doddington, George and
Mitchell, Alexis and
Przybocki, Mark and
Ramshaw, Lance and
Strassel, Stephanie and
Weischedel, Ralph | 2,004 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0617dd6924df7a3491c299772b70e90507b195dc | 1 |
Generating Paired Transliterated-cognates Using Multiple Pronunciation Characteristics from Web corpora | Kuo, Jin-Shea and
Yang, Ying-Kuei | 2,004 | nan | 275--282 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | adfbf15a0059f68af95c2836b890577807d66550 | 0 |
Transforming {W}ikipedia into Named Entity Training Data | Nothman, Joel and
Curran, James R. and
Murphy, Tara | 2,008 | nan | 124--132 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 04ca48e573c0800fc572f2af1d475dd2645e840a | 1 |
Automatic Extraction of Briefing Templates | Das, Dipanjan and
Kumar, Mohit and
Rudnicky, Alexander I. | 2,008 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e1dfe8ad1f4cfb45484118d3faf0f13505e483e1 | 0 |
{AFET}: Automatic Fine-Grained Entity Typing by Hierarchical Partial-Label Embedding | Ren, Xiang and
He, Wenqi and
Qu, Meng and
Huang, Lifu and
Ji, Heng and
Han, Jiawei | 2,016 | nan | 1369--1378 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ee42c6c3c5db2f0eb40faacf6e3b80035a645287 | 1 |
Understanding Discourse on Work and Job-Related Well-Being in Public Social Media | Liu, Tong and
Homan, Christopher and
Ovesdotter Alm, Cecilia and
Lytle, Megan and
Marie White, Ann and
Kautz, Henry | 2,016 | nan | 1044--1053 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e32666501e824d1cfdc19534b0ed009c7268cd8a | 0 |
{J}-{NERD}: Joint Named Entity Recognition and Disambiguation with Rich Linguistic Features | Nguyen, Dat Ba and
Theobald, Martin and
Weikum, Gerhard | 2,016 | Methods for Named Entity Recognition and Disambiguation (NERD) perform NER and NED in two separate stages. Therefore, NED may be penalized with respect to precision by NER false positives, and suffers in recall from NER false negatives. Conversely, NED does not fully exploit information computed by NER such as types of mentions. This paper presents J-NERD, a new approach to perform NER and NED jointly, by means of a probabilistic graphical model that captures mention spans, mention types, and the mapping of mentions to entities in a knowledge base. We present experiments with different kinds of texts from the CoNLL{'}03, ACE{'}05, and ClueWeb{'}09-FACC1 corpora. J-NERD consistently outperforms state-of-the-art competitors in end-to-end NERD precision, recall, and F1. | 215--229 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 181e0f95a61d769e01f6d2c520d60d4df228c5c0 | 1 |
{F}rench Learners Audio Corpus of {G}erman Speech ({FLACGS}) | Wottawa, Jane and
Adda-Decker, Martine | 2,016 | The French Learners Audio Corpus of German Speech (FLACGS) was created to compare German speech production of German native speakers (GG) and French learners of German (FG) across three speech production tasks of increasing production complexity: repetition, reading and picture description. 40 speakers, 20 GG and 20 FG performed each of the three tasks, which in total leads to approximately 7h of speech. The corpus was manually transcribed and automatically aligned. Analysis that can be performed on this type of corpus are for instance segmental differences in the speech production of L2 learners compared to native speakers. We chose the realization of the velar nasal consonant engma. In spoken French, engma does not appear in a VCV context which leads to production difficulties in FG. With increasing speech production complexity (reading and picture description), engma is realized as engma + plosive by FG in over 50{\%} of the cases. The results of a two way ANOVA with unequal sample sizes on the durations of the different realizations of engma indicate that duration is a reliable factor to distinguish between engma and engma + plosive in FG productions compared to the engma productions in GG in a VCV context. The FLACGS corpus allows to study L2 production and perception. | 3215--3219 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e5a5dd2a6b9d3aa69f57a7180fd129b338d82866 | 0 |
Entity Linking via Joint Encoding of Types, Descriptions, and Context | Gupta, Nitish and
Singh, Sameer and
Roth, Dan | 2,017 | For accurate entity linking, we need to capture various information aspects of an entity, such as its description in a KB, contexts in which it is mentioned, and structured knowledge. Additionally, a linking system should work on texts from different domains without requiring domain-specific training data or hand-engineered features. In this work we present a neural, modular entity linking system that learns a unified dense representation for each entity using multiple sources of information, such as its description, contexts around its mentions, and its fine-grained types. We show that the resulting entity linking system is effective at combining these sources, and performs competitively, sometimes out-performing current state-of-the-art systems across datasets, without requiring any domain-specific training data or hand-engineered features. We also show that our model can effectively {``}embed{''} entities that are new to the KB, and is able to link its mentions accurately. | 2681--2690 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 2927dfc481446568fc9108795570eb4d416be021 | 1 |
Efficient Attention using a Fixed-Size Memory Representation | Britz, Denny and
Guan, Melody and
Luong, Minh-Thang | 2,017 | The standard content-based attention mechanism typically used in sequence-to-sequence models is computationally expensive as it requires the comparison of large encoder and decoder states at each time step. In this work, we propose an alternative attention mechanism based on a fixed size memory representation that is more efficient. Our technique predicts a compact set of K attention contexts during encoding and lets the decoder compute an efficient lookup that does not need to consult the memory. We show that our approach performs on-par with the standard attention mechanism while yielding inference speedups of 20{\%} for real-world translation tasks and more for tasks with longer sequences. By visualizing attention scores we demonstrate that our models learn distinct, meaningful alignments. | 392--400 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 87fc28cbb193a3bc100e13a4a57a8dc9ce7e31a3 | 0 |
Structured Generative Models for Unsupervised Named-Entity Clustering | Elsner, Micha and
Charniak, Eugene and
Johnson, Mark | 2,009 | nan | 164--172 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | dfedad21048cafdb7066cd2caeba13228e83d4eb | 1 |
Fertility-based Source-Language-biased Inversion Transduction Grammar for Word Alignment | Huang, Chung-Chi and
Chang, Jason S. | 2,009 | nan | 1--18 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 498b24e27e2a2cec829afbb1163cd456a01ba668 | 0 |
Instance-Based Ontology Population Exploiting Named-Entity Substitution | Giuliano, Claudio and
Gliozzo, Alfio | 2,008 | nan | 265--272 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | aa480756d7ecee36e500ed05e13d0eb3bfe0aa2d | 1 |
The {M}ove{O}n Motorcycle Speech Corpus | Winkler, Thomas and
Kostoulas, Theodoros and
Adderley, Richard and
Bonkowski, Christian and
Ganchev, Todor and
K{\"o}hler, Joachim and
Fakotakis, Nikos | 2,008 | A speech and noise corpus dealing with the extreme conditions of the motorcycle environment is developed within the MoveOn project. Speech utterances in British English are recorded and processed approaching the issue of command and control and template driven dialog systems on the motorcycle. The major part of the corpus comprises noisy speech and environmental noise recorded on a motorcycle, but several clean speech recordings in a silent environment are also available. The corpus development focuses on distortion free recordings and accurate descriptions of both recorded speech and noise. Not only speech segments are annotated but also annotation of environmental noise is performed. The corpus is a small-sized speech corpus with about 12 hours of clean and noisy speech utterances and about 30 hours of segments with environmental noise without speech. This paper addresses the motivation and development of the speech corpus and finally presents some statistics and results of the database creation. | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | beae598d40e025802505ef076d95e8d3f2ca2d3c | 0 |
Semantic Class Learning from the Web with Hyponym Pattern Linkage Graphs | Kozareva, Zornitsa and
Riloff, Ellen and
Hovy, Eduard | 2,008 | nan | 1048--1056 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 194587b9d80e29aa6e50e9b0c628b581b66ea364 | 1 |
Machine Translation for {I}ndonesian and {T}agalog | Laugher, Brianna and
MacLeod, Ben | 2,008 | Kataku is a hybrid MT system for Indonesian to English and English to Indonesian translation, available on Windows, Linux and web-based platforms. This paper briefly presents the technical background to Kataku, some of its use cases and extensions. Kataku is the flagship product of ToggleText, a language technology company based in Melbourne, Australia. | 397--401 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 5ab505f24d2b3f5391ed64b1dbf6796411f6f1fd | 0 |
{S}cience{E}xam{CER}: A High-Density Fine-Grained Science-Domain Corpus for Common Entity Recognition | Smith, Hannah and
Zhang, Zeyu and
Culnan, John and
Jansen, Peter | 2,020 | Named entity recognition identifies common classes of entities in text, but these entity labels are generally sparse, limiting utility to downstream tasks. In this work we present ScienceExamCER, a densely-labeled semantic classification corpus of 133k mentions in the science exam domain where nearly all (96{\%}) of content words have been annotated with one or more fine-grained semantic class labels including taxonomic groups, meronym groups, verb/action groups, properties and values, and synonyms. Semantic class labels are drawn from a manually-constructed fine-grained typology of 601 classes generated through a data-driven analysis of 4,239 science exam questions. We show an off-the-shelf BERT-based named entity recognition model modified for multi-label classification achieves an accuracy of 0.85 F1 on this task, suggesting strong utility for downstream tasks in science domain question answering requiring densely-labeled semantic classification. | 4529--4546 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e9a679c4215762478f1b849101b09102ed39c6b1 | 1 |
Exclusive Hierarchical Decoding for Deep Keyphrase Generation | Chen, Wang and
Chan, Hou Pong and
Li, Piji and
King, Irwin | 2,020 | Keyphrase generation (KG) aims to summarize the main ideas of a document into a set of keyphrases. A new setting is recently introduced into this problem, in which, given a document, the model needs to predict a set of keyphrases and simultaneously determine the appropriate number of keyphrases to produce. Previous work in this setting employs a sequential decoding process to generate keyphrases. However, such a decoding method ignores the intrinsic hierarchical compositionality existing in the keyphrase set of a document. Moreover, previous work tends to generate duplicated keyphrases, which wastes time and computing resources. To overcome these limitations, we propose an exclusive hierarchical decoding framework that includes a hierarchical decoding process and either a soft or a hard exclusion mechanism. The hierarchical decoding process is to explicitly model the hierarchical compositionality of a keyphrase set. Both the soft and the hard exclusion mechanisms keep track of previously-predicted keyphrases within a window size to enhance the diversity of the generated keyphrases. Extensive experiments on multiple KG benchmark datasets demonstrate the effectiveness of our method to generate less duplicated and more accurate keyphrases. | 1095--1105 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ba46ece6feba34c408d081a8dce66f0ecf4b7a60 | 0 |
Two/Too Simple Adaptations of {W}ord2{V}ec for Syntax Problems | Ling, Wang and
Dyer, Chris and
Black, Alan W. and
Trancoso, Isabel | 2,015 | nan | 1299--1304 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | b92513dac9d5b6a4683bcc625b94dd1ced98734e | 1 |
Gradiant-Analytics: Training Polarity Shifters with {CRF}s for Message Level Polarity Detection | Cerezo-Costas, H{\'e}ctor and
Celix-Salgado, Diego | 2,015 | nan | 539--544 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a109aa9ab559dbdbd57df26ba2a5a08fba7aa34b | 0 |
Knowledge Base Population: Successful Approaches and Challenges | Ji, Heng and
Grishman, Ralph | 2,011 | nan | 1148--1158 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 77d2698e8efadda698b0edb457cd8de75224bfa0 | 1 |
Using machine translation in computer-aided translation to suggest the target-side words to change | Espl{\`a}-Gomis, Miquel and
S{\'a}nchez-Mart{\'\i}nez, Felipe and
Forcada, Mikel L. | 2,011 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ffe80022365aa36087f736a57e536dffea78ab53 | 0 |
Fine-Grained Entity Type Classification by Jointly Learning Representations and Label Embeddings | Abhishek, Abhishek and
Anand, Ashish and
Awekar, Amit | 2,017 | Fine-grained entity type classification (FETC) is the task of classifying an entity mention to a broad set of types. Distant supervision paradigm is extensively used to generate training data for this task. However, generated training data assigns same set of labels to every mention of an entity without considering its local context. Existing FETC systems have two major drawbacks: assuming training data to be noise free and use of hand crafted features. Our work overcomes both drawbacks. We propose a neural network model that jointly learns entity mentions and their context representation to eliminate use of hand crafted features. Our model treats training data as noisy and uses non-parametric variant of hinge loss function. Experiments show that the proposed model outperforms previous state-of-the-art methods on two publicly available datasets, namely FIGER (GOLD) and BBN with an average relative improvement of 2.69{\%} in micro-F1 score. Knowledge learnt by our model on one dataset can be transferred to other datasets while using same model or other FETC systems. These approaches of transferring knowledge further improve the performance of respective models. | 797--807 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | f4283dbf7883b1ab1a7fe01b58ebd627bcfdf008 | 1 |
Grounding Language by Continuous Observation of Instruction Following | Han, Ting and
Schlangen, David | 2,017 | Grounded semantics is typically learnt from utterance-level meaning representations (e.g., successful database retrievals, denoted objects in images, moves in a game). We explore learning word and utterance meanings by continuous observation of the actions of an instruction follower (IF). While an instruction giver (IG) provided a verbal description of a configuration of objects, IF recreated it using a GUI. Aligning these GUI actions to sub-utterance chunks allows a simple maximum entropy model to associate them as chunk meaning better than just providing it with the utterance-final configuration. This shows that semantics useful for incremental (word-by-word) application, as required in natural dialogue, might also be better acquired from incremental settings. | 491--496 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 0ceab2555088ce7ca49336f472f5902191661ff1 | 0 |
{M}essage {U}nderstanding {C}onference- 6: A Brief History | Grishman, Ralph and
Sundheim, Beth | 1,996 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 6723dda58e5e09089ec78ba42827b65859f030e2 | 1 |
{C}hinese String Searching Using the {KMP} Algorithm | Luk, Robert W.P. | 1,996 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 5f0d19f9fbbb03b8fb0fad8e810734da0242a60c | 0 |
Collective Cross-Document Relation Extraction Without Labelled Data | Yao, Limin and
Riedel, Sebastian and
McCallum, Andrew | 2,010 | nan | 1013--1023 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a3fa819575c78be3cbcc8aa394fd21a182dce669 | 1 |
Employing Machine Translation in Glocalization Tasks {--} A Use Case Study | Sch{\"u}tz, J{\"o}rg and
Andr{\"a}, Sven Christian | 2,010 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 7dafa546ad602a551fda556e1db89c522e506a9d | 0 |
Class-Based \textit{n}-gram Models of Natural Language | Brown, Peter F. and
Della Pietra, Vincent J. and
deSouza, Peter V. and
Lai, Jenifer C. and
Mercer, Robert L. | 1,992 | nan | 467--480 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 3de5d40b60742e3dfa86b19e7f660962298492af | 1 |
{N}ew {Y}ork {U}niversity Description of the {PROTEUS} System as Used for {MUC}-4 | Grishman, Ralph and
Macleod, Catherine and
Sterling, John | 1,992 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 74384de9df7422e25c3e66749ec4674b474e13f8 | 0 |
Creating an Extended Named Entity Dictionary from {W}ikipedia | Higashinaka, Ryuichiro and
Sadamitsu, Kugatsu and
Saito, Kuniko and
Makino, Toshiro and
Matsuo, Yoshihiro | 2,012 | nan | 1163--1178 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 06d44cb242b2454527e6e5e0b020664ab65a3059 | 1 |
String Re-writing Kernel | Bu, Fan and
Li, Hang and
Zhu, Xiaoyan | 2,012 | nan | 449--458 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 32924bbd545430a09969fc700965f9e030c45e67 | 0 |
Constructing Datasets for Multi-hop Reading Comprehension Across Documents | Welbl, Johannes and
Stenetorp, Pontus and
Riedel, Sebastian | 2,018 | Most Reading Comprehension methods limit themselves to queries which can be answered using a single sentence, paragraph, or document. Enabling models to combine disjoint pieces of textual evidence would extend the scope of machine comprehension methods, but currently no resources exist to train and test this capability. We propose a novel task to encourage the development of models for text understanding across multiple documents and to investigate the limits of existing methods. In our task, a model learns to seek and combine evidence {---} effectively performing multihop, alias multi-step, inference. We devise a methodology to produce datasets for this task, given a collection of query-answer pairs and thematically linked documents. Two datasets from different domains are induced, and we identify potential pitfalls and devise circumvention strategies. We evaluate two previously proposed competitive models and find that one can integrate information across documents. However, both models struggle to select relevant information; and providing documents guaranteed to be relevant greatly improves their performance. While the models outperform several strong baselines, their best accuracy reaches 54.5{\%} on an annotated test set, compared to human performance at 85.0{\%}, leaving ample room for improvement. | 287--302 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 7d5cf22c70484fe217936c66741fb73b2a278bde | 1 |
{E}i{TAKA} at {S}em{E}val-2018 Task 1: An Ensemble of N-Channels {C}onv{N}et and {XG}boost Regressors for Emotion Analysis of Tweets | Jabreel, Mohammed and
Moreno, Antonio | 2,018 | This paper describes our system that has been used in Task1 Affect in Tweets. We combine two different approaches. The first one called N-Stream ConvNets, which is a deep learning approach where the second one is XGboost regressor based on a set of embedding and lexicons based features. Our system was evaluated on the testing sets of the tasks outperforming all other approaches for the Arabic version of valence intensity regression task and valence ordinal classification task. | 193--199 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 58ee3e694bcaa126d9f2438dc0326824a0de5584 | 0 |
Improving Fine-grained Entity Typing with Entity Linking | Dai, Hongliang and
Du, Donghong and
Li, Xin and
Song, Yangqiu | 2,019 | Fine-grained entity typing is a challenging problem since it usually involves a relatively large tag set and may require to understand the context of the entity mention. In this paper, we use entity linking to help with the fine-grained entity type classification process. We propose a deep neural model that makes predictions based on both the context and the information obtained from entity linking results. Experimental results on two commonly used datasets demonstrates the effectiveness of our approach. On both datasets, it achieves more than 5{\%} absolute strict accuracy improvement over the state of the art. | 6210--6215 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | b74b272c7fe881614f3eb8c2504b037439571eec | 1 |
Comparison of Diverse Decoding Methods from Conditional Language Models | Ippolito, Daphne and
Kriz, Reno and
Sedoc, Jo{\~a}o and
Kustikova, Maria and
Callison-Burch, Chris | 2,019 | While conditional language models have greatly improved in their ability to output high quality natural language, many NLP applications benefit from being able to generate a diverse set of candidate sequences. Diverse decoding strategies aim to, within a given-sized candidate list, cover as much of the space of high-quality outputs as possible, leading to improvements for tasks that rerank and combine candidate outputs. Standard decoding methods, such as beam search, optimize for generating high likelihood sequences rather than diverse ones, though recent work has focused on increasing diversity in these methods. In this work, we perform an extensive survey of decoding-time strategies for generating diverse outputs from a conditional language model. In addition, we present a novel method where we over-sample candidates, then use clustering to remove similar sequences, thus achieving high diversity without sacrificing quality. | 3752--3762 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | fd846869e6f25d9b1a524aef8b54a08b81a1b1fa | 0 |
Generating Fine-Grained Open Vocabulary Entity Type Descriptions | Bhowmik, Rajarshi and
de Melo, Gerard | 2,018 | While large-scale knowledge graphs provide vast amounts of structured facts about entities, a short textual description can often be useful to succinctly characterize an entity and its type. Unfortunately, many knowledge graphs entities lack such textual descriptions. In this paper, we introduce a dynamic memory-based network that generates a short open vocabulary description of an entity by jointly leveraging induced fact embeddings as well as the dynamic context of the generated sequence of words. We demonstrate the ability of our architecture to discern relevant information for more accurate generation of type description by pitting the system against several strong baselines. | 877--888 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e2468753d57300d06accc5a31479e6803c90e5d4 | 1 |
Polyglot Semantic Parsing in {API}s | Richardson, Kyle and
Berant, Jonathan and
Kuhn, Jonas | 2,018 | Traditional approaches to semantic parsing (SP) work by training individual models for each available parallel dataset of text-meaning pairs. In this paper, we explore the idea of polyglot semantic translation, or learning semantic parsing models that are trained on multiple datasets and natural languages. In particular, we focus on translating text to code signature representations using the software component datasets of Richardson and Kuhn (2017b,a). The advantage of such models is that they can be used for parsing a wide variety of input natural languages and output programming languages, or mixed input languages, using a single unified model. To facilitate modeling of this type, we develop a novel graph-based decoding framework that achieves state-of-the-art performance on the above datasets, and apply this method to two other benchmark SP tasks. | 720--730 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | cf8ec112520e53a3864f01a827b085c3869c55e8 | 0 |
Fine Grained Classification of Named Entities | Fleischman, Michael and
Hovy, Eduard | 2,002 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 198b711915429fa55162e749a0b964755b36a62e | 1 |
How to prevent adjoining in {TAG}s and its impact on the Average Case Complexity | Woch, Jens | 2,002 | nan | 102--107 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e0fb80262a15e457af9e0f3de46c1f35c8f8da82 | 0 |
Neural Architectures for Fine-grained Entity Type Classification | Shimaoka, Sonse and
Stenetorp, Pontus and
Inui, Kentaro and
Riedel, Sebastian | 2,017 | In this work, we investigate several neural network architectures for fine-grained entity type classification and make three key contributions. Despite being a natural comparison and addition, previous work on attentive neural architectures have not considered hand-crafted features and we combine these with learnt features and establish that they complement each other. Additionally, through quantitative analysis we establish that the attention mechanism learns to attend over syntactic heads and the phrase containing the mention, both of which are known to be strong hand-crafted features for our task. We introduce parameter sharing between labels through a hierarchical encoding method, that in low-dimensional projections show clear clusters for each type hierarchy. Lastly, despite using the same evaluation dataset, the literature frequently compare models trained using different data. We demonstrate that the choice of training data has a drastic impact on performance, which decreases by as much as 9.85{\%} loose micro F1 score for a previously proposed method. Despite this discrepancy, our best model achieves state-of-the-art results with 75.36{\%} loose micro F1 score on the well-established Figer (GOLD) dataset and we report the best results for models trained using publicly available data for the OntoNotes dataset with 64.93{\%} loose micro F1 score. | 1271--1280 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 800dd1672789fe97513b84e65e75e370b10d6c13 | 1 |
Question Difficulty {--} How to Estimate Without Norming, How to Use for Automated Grading | Pad{\'o}, Ulrike | 2,017 | Question difficulty estimates guide test creation, but are too costly for small-scale testing. We empirically verify that Bloom{'}s Taxonomy, a standard tool for difficulty estimation during question creation, reliably predicts question difficulty observed after testing in a short-answer corpus. We also find that difficulty is mirrored in the amount of variation in student answers, which can be computed before grading. We show that question difficulty and its approximations are useful for \textit{automated grading}, allowing us to identify the optimal feature set for grading each question even in an unseen-question setting. | 1--10 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 12f66979e1b1a7d8c2e1a4ff8a6219c19483138e | 0 |
Learning to Bootstrap for Entity Set Expansion | Yan, Lingyong and
Han, Xianpei and
Sun, Le and
He, Ben | 2,019 | Bootstrapping for Entity Set Expansion (ESE) aims at iteratively acquiring new instances of a specific target category. Traditional bootstrapping methods often suffer from two problems: 1) delayed feedback, i.e., the pattern evaluation relies on both its direct extraction quality and extraction quality in later iterations. 2) sparse supervision, i.e., only few seed entities are used as the supervision. To address the above two problems, we propose a novel bootstrapping method combining the Monte Carlo Tree Search (MCTS) algorithm with a deep similarity network, which can efficiently estimate delayed feedback for pattern evaluation and adaptively score entities given sparse supervision signals. Experimental results confirm the effectiveness of the proposed method. | 292--301 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | b2057b7ae7205c3fce709d349d575683a3dc40d1 | 1 |
Evidence Sentence Extraction for Machine Reading Comprehension | Wang, Hai and
Yu, Dian and
Sun, Kai and
Chen, Jianshu and
Yu, Dong and
McAllester, David and
Roth, Dan | 2,019 | Remarkable success has been achieved in the last few years on some limited machine reading comprehension (MRC) tasks. However, it is still difficult to interpret the predictions of existing MRC models. In this paper, we focus on extracting evidence sentences that can explain or support the answers of multiple-choice MRC tasks, where the majority of answer options cannot be directly extracted from reference documents. Due to the lack of ground truth evidence sentence labels in most cases, we apply distant supervision to generate imperfect labels and then use them to train an evidence sentence extractor. To denoise the noisy labels, we apply a recently proposed deep probabilistic logic learning framework to incorporate both sentence-level and cross-sentence linguistic indicators for indirect supervision. We feed the extracted evidence sentences into existing MRC models and evaluate the end-to-end performance on three challenging multiple-choice MRC datasets: MultiRC, RACE, and DREAM, achieving comparable or better performance than the same models that take as input the full reference document. To the best of our knowledge, this is the first work extracting evidence sentences for multiple-choice MRC. | 696--707 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | bb104dc51121a0f64a5327526fad449cb03dd1bb | 0 |
Neural Fine-Grained Entity Type Classification with Hierarchy-Aware Loss | Xu, Peng and
Barbosa, Denilson | 2,018 | The task of Fine-grained Entity Type Classification (FETC) consists of assigning types from a hierarchy to entity mentions in text. Existing methods rely on distant supervision and are thus susceptible to noisy labels that can be out-of-context or overly-specific for the training sentence. Previous methods that attempt to address these issues do so with heuristics or with the help of hand-crafted features. Instead, we propose an end-to-end solution with a neural network model that uses a variant of cross-entropy loss function to handle out-of-context labels, and hierarchical loss normalization to cope with overly-specific ones. Also, previous work solve FETC a multi-label classification followed by ad-hoc post-processing. In contrast, our solution is more elegant: we use public word embeddings to train a single-label that jointly learns representations for entity mentions and their context. We show experimentally that our approach is robust against noise and consistently outperforms the state-of-the-art on established benchmarks for the task. | 16--25 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 008405f7ee96677ac23cc38be360832af2d9f437 | 1 |
Cross-Domain Sentiment Classification with Target Domain Specific Information | Peng, Minlong and
Zhang, Qi and
Jiang, Yu-gang and
Huang, Xuanjing | 2,018 | The task of adopting a model with good performance to a target domain that is different from the source domain used for training has received considerable attention in sentiment analysis. Most existing approaches mainly focus on learning representations that are domain-invariant in both the source and target domains. Few of them pay attention to domain-specific information, which should also be informative. In this work, we propose a method to simultaneously extract domain specific and invariant representations and train a classifier on each of the representation, respectively. And we introduce a few target domain labeled data for learning domain-specific information. To effectively utilize the target domain labeled data, we train the domain invariant representation based classifier with both the source and target domain labeled data and train the domain-specific representation based classifier with only the target domain labeled data. These two classifiers then boost each other in a co-training style. Extensive sentiment analysis experiments demonstrated that the proposed method could achieve better performance than state-of-the-art methods. | 2505--2513 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a99122b0b96cc40982ac267ba7e99c72a1dcc2e9 | 0 |
An Attentive Neural Architecture for Fine-grained Entity Type Classification | Shimaoka, Sonse and
Stenetorp, Pontus and
Inui, Kentaro and
Riedel, Sebastian | 2,016 | nan | 69--74 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 5c22ff7fe5fc588e3648b5897255f151feb61fee | 1 |
Producing Monolingual and Parallel Web Corpora at the Same Time - {S}pider{L}ing and Bitextor{'}s Love Affair | Ljube{\v{s}}i{\'c}, Nikola and
Espl{\`a}-Gomis, Miquel and
Toral, Antonio and
Rojas, Sergio Ortiz and
Klubi{\v{c}}ka, Filip | 2,016 | This paper presents an approach for building large monolingual corpora and, at the same time, extracting parallel data by crawling the top-level domain of a given language of interest. For gathering linguistically relevant data from top-level domains we use the SpiderLing crawler, modified to crawl data written in multiple languages. The output of this process is then fed to Bitextor, a tool for harvesting parallel data from a collection of documents. We call the system combining these two tools Spidextor, a blend of the names of its two crucial parts. We evaluate the described approach intrinsically by measuring the accuracy of the extracted bitexts from the Croatian top-level domain {``}.hr{''} and the Slovene top-level domain {``}.si{''}, and extrinsically on the English-Croatian language pair by comparing an SMT system built from the crawled data with third-party systems. We finally present parallel datasets collected with our approach for the English-Croatian, English-Finnish, English-Serbian and English-Slovene language pairs. | 2949--2956 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | e5ce9182054fa811d3c65c7e98b30bf2d90af8a4 | 0 |
Assessing the Challenge of Fine-Grained Named Entity Recognition and Classification | Ekbal, Asif and
Sourjikova, Eva and
Frank, Anette and
Ponzetto, Simone Paolo | 2,010 | nan | 93--101 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 9538eed00e2ed9d077c88afe7492766f6855ba78 | 1 |
Workshop on Advanced Corpus Solutions | Johannessen, Janne Bondi | 2,010 | nan | 717--719 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | a33ace522fd2a4dd26cb02a4d62eec8436e8f6b3 | 0 |
An Attentive Fine-Grained Entity Typing Model with Latent Type Representation | Lin, Ying and
Ji, Heng | 2,019 | We propose a fine-grained entity typing model with a novel attention mechanism and a hybrid type classifier. We advance existing methods in two aspects: feature extraction and type prediction. To capture richer contextual information, we adopt contextualized word representations instead of fixed word embeddings used in previous work. In addition, we propose a two-step mention-aware attention mechanism to enable the model to focus on important words in mentions and contexts. We also present a hybrid classification method beyond binary relevance to exploit type inter-dependency with latent type representation. Instead of independently predicting each type, we predict a low-dimensional vector that encodes latent type features and reconstruct the type vector from this latent representation. Experiment results on multiple data sets show that our model significantly advances the state-of-the-art on fine-grained entity typing, obtaining up to 6.1{\%} and 5.5{\%} absolute gains in macro averaged F-score and micro averaged F-score respectively. | 6197--6202 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ed3a6ff80bd9892a5d8bf6490147fcd518ebc413 | 1 |
An adaptable task-oriented dialog system for stand-alone embedded devices | Duong, Long and
Hoang, Vu Cong Duy and
Pham, Tuyen Quang and
Hong, Yu-Heng and
Dovgalecs, Vladislavs and
Bashkansky, Guy and
Black, Jason and
Bleeker, Andrew and
Huitouze, Serge Le and
Johnson, Mark | 2,019 | This paper describes a spoken-language end-to-end task-oriented dialogue system for small embedded devices such as home appliances. While the current system implements a smart alarm clock with advanced calendar scheduling functionality, the system is designed to make it easy to port to other application domains (e.g., the dialogue component factors out domain-specific execution from domain-general actions such as requesting and updating slot values). The system does not require internet connectivity because all components, including speech recognition, natural language understanding, dialogue management, execution and text-to-speech, run locally on the embedded device (our demo uses a Raspberry Pi). This simplifies deployment, minimizes server costs and most importantly, eliminates user privacy risks. The demo video in alarm domain is here youtu.be/N3IBMGocvHU | 49--57 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | ea2ad7e330070aed3e909a2e263ae88e320663b3 | 0 |
Transforming {W}ikipedia into a Large-Scale Fine-Grained Entity Type Corpus | Ghaddar, Abbas and
Langlais, Philippe | 2,018 | nan | nan | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 8296e7f869172ac6fe34380574706f753328eda5 | 1 |
{B}in{L}in: A Simple Method of Dependency Tree Linearization | Puzikov, Yevgeniy and
Gurevych, Iryna | 2,018 | Surface Realization Shared Task 2018 is a workshop on generating sentences from lemmatized sets of dependency triples. This paper describes the results of our participation in the challenge. We develop a data-driven pipeline system which first orders the lemmas and then conjugates the words to finish the surface realization process. Our contribution is a novel sequential method of ordering lemmas, which, despite its simplicity, achieves promising results. We demonstrate the effectiveness of the proposed approach, describe its limitations and outline ways to improve it. | 13--28 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 933afbd7aea1671e6950d65511c23af7adf38de1 | 0 |
A Joint Model for Entity Analysis: Coreference, Typing, and Linking | Durrett, Greg and
Klein, Dan | 2,014 | We present a joint model of three core tasks in the entity analysis stack: coreference resolution (within-document clustering), named entity recognition (coarse semantic typing), and entity linking (matching to Wikipedia entities). Our model is formally a structured conditional random field. Unary factors encode local features from strong baselines for each task. We then add binary and ternary factors to capture cross-task interactions, such as the constraint that coreferent mentions have the same semantic type. On the ACE 2005 and OntoNotes datasets, we achieve state-of-the-art results for all three tasks. Moreover, joint modeling improves performance on each task over strong independent baselines. | 477--490 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 28eb033eee5f51c5e5389cbb6b777779203a6778 | 1 |
Transfer learning of feedback head expressions in {D}anish and {P}olish comparable multimodal corpora | Navarretta, Costanza and
Lis, Magdalena | 2,014 | The paper is an investigation of the reusability of the annotations of head movements in a corpus in a language to predict the feedback functions of head movements in a comparable corpus in another language. The two corpora consist of naturally occurring triadic conversations in Danish and Polish, which were annotated according to the same scheme. The intersection of common annotation features was used in the experiments. A Na{\"\i}ve Bayes classifier was trained on the annotations of a corpus and tested on the annotations of the other corpus. Training and test datasets were then reversed and the experiments repeated. The results show that the classifier identifies more feedback behaviours than the majority baseline in both cases and the improvements are significant. The performance of the classifier decreases significantly compared with the results obtained when training and test data belong to the same corpus. Annotating multimodal data is resource consuming, thus the results are promising. However, they also confirm preceding studies that have identified both similarities and differences in the use of feedback head movements in different languages. Since our datasets are small and only regard a communicative behaviour in two languages, the experiments should be tested on more data types. | 3597--3603 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | afd1491f8a07ce7a4388e0ab40bf4eb7c06333a8 | 0 |
{W}iki{C}oref: An {E}nglish Coreference-annotated Corpus of {W}ikipedia Articles | Ghaddar, Abbas and
Langlais, Phillippe | 2,016 | This paper presents WikiCoref, an English corpus annotated for anaphoric relations, where all documents are from the English version of Wikipedia. Our annotation scheme follows the one of OntoNotes with a few disparities. We annotated each markable with coreference type, mention type and the equivalent Freebase topic. Since most similar annotation efforts concentrate on very specific types of written text, mainly newswire, there is a lack of resources for otherwise over-used Wikipedia texts. The corpus described in this paper addresses this issue. We present a freely available resource we initially devised for improving coreference resolution algorithms dedicated to Wikipedia texts. Our corpus has no restriction on the topics of the documents being annotated, and documents of various sizes have been considered for annotation. | 136--142 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | fe5cffa25cb2ab412da9f19ea7e9656ac72d454c | 1 |
Improving Reliability of Word Similarity Evaluation by Redesigning Annotation Task and Performance Measure | Avraham, Oded and
Goldberg, Yoav | 2,016 | nan | 106--110 | 566bd3f672357b8e35343ab6bda4cc25a3071922 | Fine-Grained Entity Typing With a Type Taxonomy: A Systematic Review | 1c7174bd2b01831920217088e2b48cb151691110 | 0 |