A Survey on Hallucination in Large Language Models: Principles, Taxonomy, Challenges, and Open Questions Paper • 2311.05232 • Published Nov 9, 2023
RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models Paper • 2401.00396 • Published Dec 31, 2023 • 3
MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents Paper • 2404.10774 • Published Apr 16 • 2
WildHallucinations: Evaluating Long-form Factuality in LLMs with Real-World Entity Queries Paper • 2407.17468 • Published Jul 24
HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild Paper • 2403.04307 • Published Mar 7
FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation Paper • 2305.14251 • Published May 23, 2023 • 2
HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language Models Paper • 2305.11747 • Published May 19, 2023
Foundational Autoraters: Taming Large Language Models for Better Automatic Evaluation Paper • 2407.10817 • Published Jul 15 • 13
G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment Paper • 2303.16634 • Published Mar 29, 2023 • 3