Awesome
Hallucination in Large Foundation Models
This repository will be updated to include all the contemporary papers related to hallucination in foundation models. We broadly categorize the papers into four major categories as follows.
Text
LLMs
- SELFCHECKGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models
- Beyond Factuality: A Comprehensive Evaluation of Large Language Models as Knowledge Generators
- HaluEval: A Large-Scale Hallucination Evaluation Benchmark for Large Language Models
- Self-contradictory Hallucinations of Large Language Models: Evaluation, Detection and Mitigation
- PURR: Efficiently Editing Language Model Hallucinations by Denoising Language Model Corruptions
- Mitigating Language Model Hallucination with Interactive Question-Knowledge Alignment
- How Language Model Hallucinations Can Snowball
- Check Your Facts and Try Again: Improving Large Language Models with External Knowledge and Automated Feedback
- The Internal State of an LLM Knows When its Lying
- Chain of Knowledge: A Framework for Grounding Large Language Models with Structured Knowledge Bases
- HALO: Estimation and Reduction of Hallucinations in Open-Source Weak Large Language Models
- A Stitch in Time Saves Nine: Detecting and Mitigating Hallucinations of LLMs by Validating Low-Confidence Generation
- Dehallucinating Large Language Models Using Formal Methods Guided Iterative Prompting
- Sources of Hallucination by Large Language Models on Inference Tasks
- Citation: A Key to Building Responsible and Accountable Large Language Models
- Zero-resource hallucination prevention for large language models
- RARR: Researching and Revising What Language Models Say, Using Language Models
Multilingual LLMs
Domain-specific LLMs
Image
- Evaluating Object Hallucination in Large Vision-Language Models
- Detecting and Preventing Hallucinations in Large Vision Language Models
- Plausible May Not Be Faithful: Probing Object Hallucination in Vision-Language Pre-training
- Hallucination Improves the Performance of Unsupervised Visual Representation Learning
Video
- Let’s Think Frame by Frame: Evaluating Video Chain of Thought with Video Infilling and Prediction
- Putting People in Their Place: Affordance-Aware Human Insertion into Scenes
- VideoChat : Chat-Centric Video Understanding