ReLiK: Retrieve and LinK, Fast and Accurate Entity Linking and Relation Extraction on an Academic Budget
Abstract
Entity Linking (EL) and Relation Extraction (RE) are fundamental tasks in Natural Language Processing, serving as critical components in a wide range of applications. In this paper, we propose ReLiK, a Retriever-Reader architecture for both EL and RE, where, given an input text, the Retriever module undertakes the identification of candidate entities or relations that could potentially appear within the text. Subsequently, the Reader module is tasked to discern the pertinent retrieved entities or relations and establish their alignment with the corresponding textual spans. Notably, we put forward an innovative input representation that incorporates the candidate entities or relations alongside the text, making it possible to link entities or extract relations in a single forward pass and to fully leverage pre-trained language models contextualization capabilities, in contrast with previous Retriever-Reader-based methods, which require a forward pass for each candidate. Our formulation of EL and RE achieves state-of-the-art performance in both in-domain and out-of-domain benchmarks while using academic budget training and with up to 40x inference speed compared to competitors. Finally, we show how our architecture can be used seamlessly for Information Extraction (cIE), i.e. EL + RE, and setting a new state of the art by employing a shared Reader that simultaneously extracts entities and relations.
Community
The paper presents a Retriever-Reader architecture for Entity Linking and Relation Extraction. First, entity and relations are retrieved from a Knowledge Base using the Retriever, from which topk are selected to be given to the Reader. Unlike previous work, the Reader is capable of linking spans of text to their corresponding entity as well as predicting the relations between them in a single forward pass. This is achieved by leveraging descriptive modeling, i.e. the candidate entities and relation types are given to the reader as part of the input, and then some specialized layers link spans to candidate entities, and predict relations between them. This image should explain this process visually:
There is a collection of models released, a space demo, GitHub, pip package and Docker:
๐ค Check out our live demo on Huggingface: ReLiK Information Extraction
๐ Check out the model collection also on Huggingface: ReLiK Collection
๐ฅ Explore the code on GitHub: SapienzaNLP/ReLiK
๐ Python Package on PyPI
๐ณ Docker images on Docker Hub
๐ Read the full paper on Arxiv: ReLiK Paper
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Contrastive Entity Coreference and Disambiguation for Historical Texts (2024)
- On the Robustness of Document-Level Relation Extraction Models to Entity Name Variations (2024)
- mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval (2024)
- Relation Extraction with Fine-Tuned Large Language Models in Retrieval Augmented Generation Frameworks (2024)
- LLMAEL: Large Language Models are Good Context Augmenters for Entity Linking (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 23
Browse 23 models citing this paperDatasets citing this paper 0
No dataset linking this paper