Home

Awesome

<p align="center"> <img src="pictures/logo.jpg" width="500"/> <p> <div align="center">

Awesome License: MIT

</div>

Uncritical trust in LLMs can give rise to a phenomenon Cognitive Mirage, leading to misguided decision-making and a cascade of unintended consequences. To effectively control the risk of hallucinations, we summarize recent progress in hallucination theories and solutions in this paper. We propose to organize relevant work by a comprehensive survey.

[:bell: News! :bell: ] We have released a new survey paper:"Cognitive Mirage: A Review of Hallucinations in Large Language Models" based on this repository, with a perspective of Hallucinations in LLMs! We are looking forward to any comments or discussions on this topic :)

🕵️ Introduction

As large language models continue to develop in the field of AI, text generation systems are susceptible to a worrisome phenomenon known as hallucination. In this study, we summarize recent compelling insights into hallucinations in LLMs. We present a novel taxonomy of hallucinations from various text generation tasks, thus provide theoretical insights, detection methods and improvement approaches. Based on this, future research directions are proposed. Our contribution are threefold: (1) We provide a detailed and complete taxonomy for hallucinations appearing in text generation tasks; (2) We provide theoretical analyses of hallucinations in LLMs and provide existing detection and improvement methods; (3) We propose several research directions that can be developed in the future. As hallucinations garner significant attention from the community, we will maintain updates on relevant research progress.

🏆 A Timeline of LLMs

LLM NameTitleAuthorsPublication Date
T5Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer.Colin Raffel, Noam Shazeer, Adam Roberts2019.10
GPT-3Language Models are Few-Shot Learners.Tom B. Brown, Benjamin Mann, Nick Ryder2020.12
mT5mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer.Linting Xue, Noah Constant, Adam Roberts2021.3
CodexEvaluating Large Language Models Trained on Code.Mark Chen, Jerry Tworek, Heewoo Jun2021.7
FLANFinetuned Language Models are Zero-Shot Learners.Jason Wei, Maarten Bosma, Vincent Y. Zhao2021.9
WebGPTWebGPT: Browser-assisted question-answering with human feedback.Reiichiro Nakano, Jacob Hilton, Suchir Balaji2021.12
InstructGPTTraining language models to follow instructions with human feedback.Long Ouyang, Jeffrey Wu, Xu Jiang2022.3
CodeGenCodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis.Erik Nijkamp, Bo Pang, Hiroaki Hayashi2022.3
ClaudeTraining a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback.Yuntao Bai, Andy Jones, Kamal Ndousse2022.4
PaLMPaLM: Scaling Language Modeling with Pathways.Aakanksha Chowdhery, Sharan Narang, Jacob Devlin2022.4
OPTOPT: Open Pre-trained Transformer Language Models.Susan Zhang, Stephen Roller, Naman Goyal2022.5
Super-NaturalInstructionsESuper-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks.Yizhong Wang, Swaroop Mishra, Pegah Alipoormolabashi, Yeganeh Kordi2022.9
GLMGLM-130B: An Open Bilingual Pre-trained Model.Aohan Zeng, Xiao Liu, Zhengxiao Du2022.10
BLOOMBLOOM: A 176B-Parameter Open-Access Multilingual Language Model.Teven Le Scao, Angela Fan, Christopher Akiki2022.11
LLaMALLaMA: Open and Efficient Foundation Language Models.Hugo Touvron, Thibaut Lavril, Gautier Izacard2023.2
AlpacaAlpaca: A Strong, Replicable Instruction-Following Model.Rohan Taori, Ishaan Gulrajani, Tianyi Zhang2023.3
GPT-4GPT-4 Technical Report.OpenAI2023.3
WizardLMWizardLM: Empowering Large Language Models to Follow Complex Instructions.Can Xu, Qingfeng Sun, Kai Zheng2023.4
VicunaVicuna: An Open-Source Chatbot Impressing GPT-4 with 90% ChatGPT Quality.The Vicuna Team2023.5
ChatGLMChatGLM.Wisdom and Clear Speech Team2023.6
Llama2Llama 2: Open Foundation and Fine-Tuned Chat Models.Hugo Touvron, Louis Martin, Kevin Stone2023.7

🏳‍🌈 Definition of Hallucination

🎉 Mechanism Analysis

✨Data Attribution

✨Knowledge Gap

✨Optimum Formulation

🪁Taxonomy of LLMs Hallucination in NLP tasks

✨Machine Translation

✨Question and Answer

✨Dialog System

✨Summarization System

✨Knowledge Graphs with LLMs

✨Cross-modal System

✨Others

🔮 Hallucination Detection

✨Inference Classifier

✨Uncertainty Measure

✨Self-Evaluation

✨Evidence Retrieval

🪄 Hallucination Correction

✨Parameter Enhancement

✨Post-hoc Attribution and Edit Technology

✨Utilizing Programming Languages

✨Leverage External Knowledge

✨Assessment Feedback

✨Mindset Society

🌟 TIPS

If you find this repository useful to your research or work, it is really appreciate to star this repository.