Home

Awesome

Awesome-Align-LLM-Human

A collection of papers and resources about aligning large language models (LLMs) with human.

Large Language Models (LLMs) trained on extensive textual corpora have emerged as leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite their notable performance, these models are prone to certain limitations such as misunderstanding human instructions, generating potentially biased content, or factually incorrect (hallucinated) information. Hence, aligning LLMs with human expectations has become an active area of interest within the research community. This survey presents a comprehensive overview of these alignment technologies, including the following aspects. (1) Data collection (2) Training methodologies (3) Model Evaluation. In conclusion, we collate and distill our findings, shedding light on several promising future research avenues in the field. This survey, therefore, serves as a valuable resource for anyone invested in understanding and advancing the alignment of LLMs to better suit human-oriented tasks and expectations.

We hope this repository can help researchers and practitioners to get a better understanding of this emerging field. If this repository is helpful for you, please help us by citing this paper:

@article{aligning_llm_human,
    title={Aligning Large Language Models with Human: A Survey},
    author={Yufei Wang and Wanjun Zhong and Liangyou Li and Fei Mi and Xingshan Zeng and Wenyong Huang and Lifeng Shang and Xin Jiang and Qun Liu},
    journal={arXiv preprint arXiv:2307.12966},
    year={2023}
}

News

🔭 This project is under development. You can hit the STAR and WATCH to follow the updates.

Table of Contents

Related Surveys

Alignment Data

Data From Human

NLP Benchmarks

Domain Knowledge

Hand-crafted Instructions

Human Preference Data

Data From Strong LLMs

General Instructions

Improving Input Quality
Improving Output Quality

Reasoning Instructions

General Reasoning
Code
Maths

Conversational Instructions

Multilingual Instructions

Instructions Management

Instruction Implications

Instruction Quantity

Alignment Training

Online Human Alignment

Offline Human Alignment

Rank-based Training

Language-based Training

Parameter-Efficient Training

Model Architecture Design

Alignment Evaluation

Evaluation Design Principles

Evaluation Benchmarks

Closed-set Benchmarks

General Knowledge
Reasoning

Open-set Benchmarks

General Chat
Safety
Long Context

Evaluation Paradigms

Human-based Evaluation

LLMs-based Evaluation

LLMs for Evaluation
LLMs bias in Evaluation
Evaluation-specific LLMs

Alignment Toolkits