Abstract
While recent language models have the ability to take long contexts as input, relatively little is known about how well the language models use longer context. We analyze language model performance on two tasks that require identifying relevant information within their input contexts: multi-document question answering and key-value retrieval. We find that performance is often highest when relevant information occurs at the beginning or end of the input context, and significantly degrades when models must access relevant information in the middle of long contexts. Furthermore, performance substantially decreases as the input context grows longer, even for explicitly long-context models. Our analysis provides a better understanding of how language models use their input context and provides new evaluation protocols for future long-context models.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Marathon: A Race Through the Realm of Long Context with Large Language Models (2023)
- "Paraphrasing The Original Text" Makes High Accuracy Long-Context QA (2023)
- Enhancing Large Language Model Performance To Answer Questions and Extract Information More Accurately (2024)
- LongAlign: A Recipe for Long Context Alignment of Large Language Models (2024)
- Structured Packing in LLM Training Improves Long Context Utilization (2023)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
How Current Language Models Struggle with Long Contexts: Key Insights
Links π:
π Subscribe: https://www.youtube.com/@Arxflix
π Twitter: https://x.com/arxflix
π LMNT (Partner): https://lmnt.com/