Home

Awesome

Awesome LLMs Evaluation Papers :bookmark_tabs:

The papers are organized according to our survey:

<p align="center"><strong>Evaluating Large Language Models: A Comprehensive Survey</strong></p> <p align="center">Zishan Guo*, Renren Jin*, Chuang Liu*, Yufei Huang, Dan Shi, Supryadi, </p> <p align="center">Linhao Yu, Yan Liu, Jiaxuan Li, Bojian Xiong, Deyi Xiong†</p> <p align="center">Tianjin University</p> <p align="center">(*: Co-first authors, †: Corresponding author)</p> <div align=center> <img src="./imgs/Figure_1.png" style="zoom:30%"/> </div>

If you find our survey useful, please kindly cite our paper:

@article{guo2023evaluating,
  title={Evaluating Large Language Models: A Comprehensive Survey},
  author={Guo, Zishan and Jin, Renren and Liu, Chuang and Huang, Yufei and Shi, Dan and Yu, Linhao and Liu, Yan and Li, Jiaxuan and Xiong, Bojian and Xiong, Deyi and others},
  journal={arXiv preprint arXiv:2310.19736},
  year={2023}
}

Contributing to this paper list

Feel free to open an issue/PR or e-mail guozishan@tju.edu.cn, rrjin@tju.edu.cn, liuc_09@tju.edu.cn and dyxiong@tju.edu.cn if you find any missing areas, papers, or datasets. We will keep updating this list and survey.

Updates

Survey Introduction

Large language models (LLMs) have demonstrated remarkable capabilities across a broad spectrum of tasks. They have attracted significant attention and been deployed in numerous downstream applications. Nevertheless, akin to a double-edged sword, LLMs also present potential risks. They could suffer from private data leaks or yield inappropriate, harmful, or misleading content. Additionally, the rapid progress of LLMs raises concerns about the potential emergence of superintelligent systems without adequate safeguards. To effectively capitalize on LLM capacities as well as ensure their safe and beneficial development, it is critical to conduct a rigorous and comprehensive evaluation of LLMs.

This survey endeavors to offer a panoramic perspective on the evaluation of LLMs. We categorize the evaluation of LLMs into three major groups: knowledge and capability evaluation, alignment evaluation and safety evaluation. In addition to the comprehensive review on the evaluation methodologies and benchmarks on these three aspects, we collate a compendium of evaluations pertaining to LLMs' performance in specialized domains, and discuss the construction of comprehensive evaluation platforms that covers LLM evaluations on capabilities, alignment, safety, sand applicability.

We hope that this comprehensive overview will stimulate further research interests in the evaluation of LLMs, with the ultimate goal of making evaluation serve as a cornerstone in guiding the responsible development of LLMs. We envision that this will channel their evolution into a direction that maximizes societal benefit while minimizing potential risks.

Markups

The paper proposes a dataset that can be used for LLMs evaluation.

The paper proposes an evaluation method that can be used for LLMs.

The paper proposes a platform for LLMs evaluation.

The paper examines the performance of LLMs in a particular domain.

Table of Contents

Related Surveys for LLMs Evaluation

  1. "Through the Lens of Core Competency: Survey on Evaluation of Large Language Models".

    Ziyu Zhuang et al. arXiv 2023. [Paper] [GitHub]

  2. "A Survey on Evaluation of Large Language Models".

    Yupeng Chang and Xu Wang et al. arXiv 2023. [Paper] [GitHub]

Papers

:books:Knowledge and Capability Evaluation

Question Answering

  1. Squad: "Squad: 100, 000+ questions for machine comprehension of text".

    Pranav Rajpurkar et al. EMNLP 2016. [Paper] [Source]

  2. NarrativeQA: "The narrativeqa reading comprehension challenge".

    Tomás Kociský et al. arXiv 2017. [Paper] [Github]

  3. Hotpotqa: "Hotpotqa: A dataset for diverse, explainable multi-hop question answering".

    Zhilin Yang et al. EMNLP 2018. [Paper] [Github]

  4. CoQA: "Coqa: A conversational question answering challenge".

    Siva Reddy et al. NAACL 2019. [Paper] [Github]

  5. NQ: "Natural questions: a benchmark for question answering research".

    Tom Kwiatkowski et al. [Paper] [Github]

  6. DuReader: "Dureader_robust: A chinese dataset towards evaluating robustness and generalization of machine reading comprehension in real-world applications".

    Hongxuan Tang et al. NAACL-HLT 2019. [Paper] [Github]

  7. RAGAS: "RAGAS: Automated Evaluation of Retrieval Augmented Generation".

    Shahul Es et al. arXiv 2023. [Paper] [Github]

  8. "Why Does ChatGPT Fall Short in Providing Truthful Answers?".

    Shen Zheng and Jie Huang et al. arXiv 2023. [Paper]

Knowledge Completion

  1. LAMA: "Language Models as Knowledge Bases?".

    In Kentaro Inui et al. EMNLP-IJCNLP 2019. [Paper] [GitHub]

  2. Kola: "Kola: Carefully Benchmarking World Knowledge of Large Language models".

    JiaFang Yu et al. arXiv 2023. [Paper] [Source]

  3. WikiFact: "Assessing the Factual Accuracy of Generated Text".

    Ben Goodrich et al. KDD 2019. [Paper]

Reasoning

Commonsense Reasoning
  1. ARC: "Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge".

    Peter Clark et al. arXiv 2018. [Paper] [GitHub]

  2. QASC: "QASC: A Dataset for Question Answering via Sentence Composition".

    Tushar Khot et al. AAAI 2020. [Paper] [GitHub]

  3. MCTACO: ""Going on a vacation" takes longer than "Going for a walk": A Study of Temporal Commonsense Understanding".

    Ben Zhou et al. EMNLP 2019. [Paper] [Source]

  4. TRACIE: "Temporal Reasoning on Implicit Events from Distant Supervision".

    Ben Zhou et al. NAACL 2021. [Paper] [Source]

  5. TIMEDIAL: "TIMEDIAL: Temporal Commonsense Reasoning in Dialog".

    Lianhui Qin et al. ACL 2021. [Paper] [GitHub]

  6. HellaSWAG: "HellaSwag: Can a Machine Really Finish Your Sentence?".

    Rowan Zellers et al. ACL 2019. [Paper] [Source]

  7. PIQA: "PIQA: Reasoning about Physical Commonsense in Natural Language".

    Yonatan Bisk et al. AAAI 2020. [Paper] [Source]

  8. Pep-3k: "Modeling Semantic Plausibility by Injecting World Knowledge".

    Su Wang et al. NAACL-HLT 2018. [Paper] [GitHub]

  9. Social IQA: "Social IQa: Commonsense Reasoning about Social Interactions".

    Maarten Sap and Hannah Rashkin et al. EMNLP 2019. [Paper] [Source]

  10. CommonsenseQA: "CommonsenseQA: A Question Answering Challenge Targeting Commonsense Knowledge".

    Alon Talmor and Jonathan Herzig et al. NAACL 2019. [Paper] [GitHub]

  11. OpenBookQA: "Can a Suit of Armor Conduct Electricity? A New Dataset for Open Book Question Answering".

    Todor Mihaylov et al. EMNLP 2018. [Paper] [Source]

  12. "A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity".

    Yejin Bang et al. arXiv 2023. [Paper] [GitHub]

  13. "ChatGPT is a Knowledgeable but Inexperienced Solver: An Investigation of Commonsense Problem in Large Language Models".

    Ning Bian et al. arXiv 2023. [Paper]

Logical Reasoning
  1. SNLI: "A large annotated corpus for learning natural language inference".

    Samuel R. Bowman et al. EMNLP 2015. [Paper]

  2. MultiNLI: "A Broad-Coverage Challenge Corpus for Sentence Understanding through Inference".

    Adina Williams et al. NAACL-HLT 2018. [Paper] [GitHub]

  3. LogicNLI: "Diagnosing the First-Order Logical Reasoning Ability Through LogicNLI".

    Jidong Tian and Yitian Li et al. EMNLP 2021. [Paper]

  4. ConTRoL: "Natural Language Inference in Context — Investigating Contextual Reasoning over Long Texts".

    Hanmeng Liu et al. EMNLP 2015. [Paper] [GitHub]

  5. MED: "Can Neural Networks Understand Monotonicity Reasoning?".

    Hitomi Yanaka et al. ACL Workshop BlackboxNLP 2019. [Paper] [GitHub]

  6. HELP: "HELP: A Dataset for Identifying Shortcomings of Neural Models in Monotonicity Reasoning".

    Hitomi Yanaka et al. *SEM 2019. [Paper] [GitHub]

  7. ConjNLI: "ConjNLI: Natural Language Inference Over Conjunctive Sentences".

    Swarnadeep Saha et al. EMNLP 2020. [Paper] [GitHub]

  8. TaxiNLI: "TaxiNLI: Taking a Ride up the NLU Hill".

    Pratik Joshi, Somak Aditya and Aalok Sathe et al. CoNLL 2020. [Paper] [GitHub]

  9. ReClor: "ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning".

    Weihao Yu and Zihang Jiang et al. ICLR 2020. [Paper] [Source]

  10. LogiQA: "LogiQA: A Challenge Dataset for Machine Reading Comprehension with Logical Reasoning".

    Jian Liu et al. IJCAI 2020. [Paper] [GitHub]

  11. LogiQA 2.0: "LogiQA 2.0 — An Improved Dataset for Logical Reasoning in Natural Language Understanding".

    Hanmeng Liu et al. TASLP 2023. [Paper] [GitHub]

  12. LSAT: "From LSAT: The Progress and Challenges of Complex Reasoning".

    Siyuan Wang et al. TASLP 2021. [Paper]

  13. LogicInference: "LogicInference: A New Dataset for Teaching Logical Inference to seq2seq Models".

    Santiago Ontanon et al. ICLR OSC workshop 2022. [Paper] [GitHub]

  14. FOLIO: "FOLIO: Natural Language Reasoning with First-Order Logic".

    Simeng Han et al. arXiv 2022. [Paper] [GitHub]

  15. "Are Large Language Models Really Good Logical Reasoners? A Comprehensive Evaluation and Beyond".

    Fangzhi Xu and Qika Lin et al. arXiv 2023. [Paper] [GitHub]

  16. "A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity".

    Yejin Bang et al. arXiv 2023. [Paper] [GitHub]

  17. "Evaluating the Logical Reasoning Ability of ChatGPT and GPT-4".

    Hanmeng Liu et al. arXiv 2023. [Paper] [GitHub]

Multi-hop Reasoning
  1. HotpotQA: "HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering".

    Zhilin Yang, Peng Qi and Saizheng Zhang et al. EMNLP 2018. [Paper] [GitHub]

  2. HybridQA: "HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data".

    Wenhu Chen et al. EMNLP (Findings) 2020. [Paper] [GitHub]

  3. MultiRC: "Looking Beyond the Surface: A Challenge Set for Reading Comprehension over Multiple Sentences".

    Daniel Khashabi et al. NAACL 2018. [Paper] [Source]

  4. NarrativeQA: "The NarrativeQA Reading Comprehension Challenge".

    Tomas Kocisk et al. TACL 2018. [Paper] [Source]

  5. Wikihop, Medhop: "Constructing Datasets for Multi-hop Reading Comprehension Across Documents".

    Johannes Welbl et al. TACL 2018. [Paper] [Source]

  6. "A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity".

    Yejin Bang et al. arXiv 2023. [Paper] [GitHub]

  7. "How is ChatGPT's behavior changing over time?".

    Lingjiao Chen et al. arXiv 2023. [Paper] [GitHub]

Mathematical Reasoning
  1. MultiArith: "Solving General Arithmetic Word Problems".

    Subhro Roy and Dan Roth et al. EMNLP 2015. [Paper]

  2. AddSub: "Learning to Solve Arithmetic Word Problems with Verb Categorization".

    Mohammad Javad Hosseini et al. ACL 2014. [Paper]

  3. AQUA: "Program Induction by Rationale Generation: Learning to Solve and Explain Algebraic Word Problems".

    Wang Ling et al. ACL 2017. [Paper]

  4. SVAMP: "Are NLP Models Really Able to Solve Simple Math Word Problems".

    Arkil Patel et al. ACL 2021. [Paper] [GitHub]

  5. GSM8K: "Training Verifiers to Solve Math Word Problems".

    Karl Cobbe et al. arXiv 2021. [Paper] [GitHub]

  6. M3KE: "M3KE: A Massive Multi-level Multi-subject Knowledge Evaluation Benchmark for Chinese Large Language Models".

    Liu Chuang et al. arXiv 2023. [Paper] [GitHub]

  7. VNHSGE: "VNHSGE: Vietnamese High School Graduation Examination Dataset for Large Language Models".

    Xuan-Quy Dao et al. arXiv 2023. [Paper] [GitHub]

  8. MATH: "Measuring Mathematical Problem Solving with the MATH Dataset".

    Dan Hendrycks et al. NeurIPS 2021. [Paper] [GitHub]

  9. JEEBench: "Have LLMs Advanced Enough A Challenging Problem Solving Benchmark for Large Language Models".

    Daman Arora et al. EMNLP 2023. [Paper] [GitHub]

  10. MATH401: "How Well Do Large Language Models Perform in Arithmetic Tasks".

    Zheng Yuan et al. arXiv 2023. [Paper] [GitHub]

  11. CMATH: "CMATH: Can Your Language Model Pass Chinese Elementary School Math Test?".

    WeiTian Wen et al. arXiv 2023. [Paper]

  12. AUTOPROMPT: "Chain-of-Thought Prompting Elicits Reasoning in Large Language Models".

    Jason Wei et al. NeurIPS 2022. [Paper]

  13. "Evaluating Language Models for Mathematics Through Interactions".

    Katherine M. Collins et al. arXiv 2023. [Paper]

Tool Learning

  1. RestBench: "RestGPT: Connecting Large Language Models with Real-World RESTful APIs".

    Yifan Song et al. arXiv 2023. [Paper] [GitHub]

  2. SayCan: "Do As I Can, Not As I Say: Grounding Language in Robotic Affordances".

    Michael Ahn et al. arXiv 2023. [Paper] [GitHub]

  3. WebCPM: "WebCPM: Interactive Web Search for Chinese Long-form Question Answering".

    Yujia Qin et al. ACL 2023. [Paper] [GitHub]

  4. WebShop: "WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents".

    Shunyu Yao et al. NeurIPS 2022. [Paper] [GitHub] [Source]

  5. ToolAlpaca: "ToolAlpaca: Generalized Tool Learning for Language Models with 3000 Simulated Cases".

    Qiaoyu Tang et al. arXiv 2023. [Paper] [GitHub]

  6. "Tool Documentation Enables Zero-Shot Tool-Usage with Large Language Models".

    Cheng-Yu Hsieh et al. arXiv 2023. [Paper]

  7. ToolQA: "ToolQA: A Dataset for LLM Question Answering with External Tools".

    Yuchen Zhuang et al. arXiv 2023. [Paper] [GitHub]

  8. Toolformer: "Toolformer: Language Models Can Teach Themselves to Use Tools".

    Timo Schick et al. arXiv 2023. [Paper] [GitHub]

  9. ALFRED: "ALFRED: A Benchmark for Interpreting Grounded Instructions for Everyday Tasks".

    Mohit Shridhar et al. CVPR 2020. [Paper] [GitHub]

  10. ALFWorld: "ALFWorld: Aligning Text and Embodied Environments for Interactive Learning".

    Mohit Shridhar et al. ICLR 2021. [Paper] [GitHub]

  11. BEHAVIOR: "BEHAVIOR: Benchmark for Everyday Household Activities in Virtual, Interactive, and Ecological Environments".

    Sanjana Srivastava et al. PMLR 2021. [Paper] [GitHub]

  12. Inner Monologue: "Inner Monologue: Embodied Reasoning through Planning with Language Models".

    Wenlong Huang et al. PMLR 2023. [Paper] [GitHub]

  13. API-Bank: "API-Bank: A Comprehensive Benchmark for Tool-Augmented LLMs".

    Minghao Li et al. arXiv 2023. [Paper] [Source]

  14. "On the Tool Manipulation Capability of Open-source Large Language Models".

    Qiantong Xu et al. arXiv 2023. [Paper]

  15. "Tool Learning with Foundation Models".

    Yujia Qin et al. arXiv 2023. [Paper] [GitHub]

  16. ToolEval: "ToolLLM: Facilitating Large Language Models to Master 16000+ Real-world APIs".

    Yujia Qin et al. arXiv 2023. [Paper] [GitHub]

  17. LaMDA: "LaMDA: Language Models for Dialog Applications".

    Romal Thoppilan et al. arXiv 2022. [Paper] [GitHub]

  18. GeneGPT: "GeneGPT: Augmenting Large Language Models with Domain Tools for Improved Access to Biomedical Information".

    Qiao Jin et al. arXiv 2023. [Paper] [GitHub]

  19. Code as Policies: "Code as Policies: Language Model Programs for Embodied Control".

    Jacky Liang et al. ICRA 2023. [Paper] [GitHub]

  20. "Augmented Language Models: a Survey".

    Grégoire Mialon et al. arXiv 2023. [Paper]

Negation

  1. "Negated and Misprimed Probes for Pretrained Language Models: Birds Can Talk, But Cannot Fly".

    Nora Kassner et al. ACL 2020. [Paper][Github]

  2. "UnCommonSense: Informative Negative Knowledge about Everyday Concepts".

    Hiba Arnaout et al. CIKM 2022. [Paper][Source]

  3. "Beyond Positive Scaling: How Negation Impacts Scaling Trends of Language Models".

    Yuhui Zhang and Michihiro Yasunaga et al. ACL (Findings) 2023. [Paper][Github]

  4. "Say What You Mean! Large Language Models Speak Too Positively about Negative Commonsense Knowledge".

    Jiangjie Chen et al. ACL 2023. [Paper][Github]

  5. ScoNe: "ScoNe: Benchmarking Negation Reasoning in Language Models With Fine-Tuning and In-Context Learning".

    Jingyuan Selena She et al. ACL 2023. [Paper][Github]

  6. xNot360: "A negation detection assessment of GPTs: analysis with the xNot360 dataset".

    Ha Thanh Nguyen et al. arXiv 2023. [Paper][Source]

  7. "This is not a Dataset: A Large Negation Benchmark to Challenge Large Language Models".

    Iker García-Ferrero et al. EMNLP 2023. [Paper][Github][Source]

:triangular_ruler:Alignment Evaluation

Ethics and Morality

  1. "Classification of moral foundations in microblog political discourse".

    Kristen Johnson et al. ACL 2018. [Paper]

  2. Social chemistry 101: "Social chemistry 101: Learning to reason about social and moral norms".

    Maxwell Forbes et al. EMNLP 2020. [Paper] [Github]

  3. Moral Foundations Twitter Corpus: "Moral foundations twitter corpus: A collection of 35k tweets annotated for moral sentiment".

    Joe Hoover et al. [Paper]

  4. "Moral stories: Situated reasoning about norms, intents, actions, and their consequences".

    Denis Emelin et al. EMNLP 2021. [Paper] [Github]

  5. "Analysis of moral judgement on reddit".

    Nicholas Botzer et al. CoRR 2021. [Paper]

  6. MIC: "The moral integrity corpus: A benchmark for ethical dialogue systems".

    Caleb Ziems et al. ACL 2022. [Paper] [Github]

  7. “When to make exceptions:Exploring language models as accounts of human moral judgment”.

    Zhijing Jin et al. NeurIPS 2022. [Paper] [Github]

  8. "Prosocialdialog: A prosocial backbone for conversational agents".

    Hyunwoo Kim et al. NAACL-HLT 2021. [Paper] [Github]

  9. SCRUPLES: "SCRUPLES: A corpus of community ethical judgments on 32, 000 real-life anecdotes".

    Nicholas Lourie et al. AAAI 2021. [Paper] [Github]

  10. "Trustgpt:A benchmark for trustworthy and responsible large language models".

    Yue Huang et al. CoRL 2022. [Paper] [Github]

  11. "Aligning AI with shared human values".

    Dan Hendrycks et al. ICLR 2021. [Paper] [Github]

  12. "Evaluating the moral beliefs encoded in llms".

    Nino Scherrer et al. CORR 2023. [Paper] [Github]

Bias

  1. Winogender: "Gender Bias in Coreference Resolution".

    Rachel Rudinger et al. NAACL-HLT 2018. [Paper] [GitHub]

  2. WinoBias: "Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods".

    Jieyu Zhao et al. NAACL-HLT 2018. [Paper] [GitHub]

  3. GICOREF: "Toward Gender-Inclusive Coreference Resolution: An Analysis of Gender and Bias Throughout the Machine Learning Lifecycle".

    Yang Trista Cao et al. Comput. Linguistics 2021. [Paper]

  4. WinoMT: "Evaluating Gender Bias in Machine Translation".

    Gabriel Stanovsky et al. ACL 2019. [Paper] [GitHub]

  5. "Investigating Failures of Automatic Translationin the Case of Unambiguous Gender".

    Adithya Renduchintala et al. ACL 2022. [Paper]

  6. "Addressing Age-Related Bias in Sentiment Analysis".

    Díaz Mark et al. IJCAI 2019. [Paper] [Source]

  7. EEC: "Examining Gender and Race Bias in Two Hundred Sentiment Analysis Systems".

    Kiritchenko Svetlana et al. NAACL HLT 2018. [Paper] [Source]

  8. WikiGenderBias: "Towards Understanding Gender Bias in Relation Extraction".

    Gaut Andrew et al. ACL 2020. [Paper] [GitHub]

  9. "Measuring and Mitigating Unintended Bias in Text Classification".

    Lucas Dixon et al. AAAI 2018. [Paper] [GitHub]

  10. "Nuanced Metrics for Measuring Unintended Bias with Real Data for Text Classification".

    Daniel Borkan et al. WWW 2019. [Paper]

  11. "Social Bias Frames: Reasoning about Social and Power Implications of Language".

    Sap Maarten et al. ACL 2020. [Paper] [Source]

  12. "Finding Microaggressions in the Wild: A Case for Locating Elusive Phenomena in Social Media Posts".

    Breitfeller Luke et al. EMNLP-IJCNLP 2019. [Paper]

  13. Latent Hatred: "Latent Hatred: A Benchmark for Understanding Implicit Hate Speech".

    Mai ElSherief et al. EMNLP 2021. [Paper] [GitHub]

  14. DynaHate: "Learning from the Worst: Dynamically Generated Datasets to Improve Online Hate Detection".

    Vidgen Bertie et al. ACL/IJCNLP 2021. [Paper] [GitHub]

  15. TOXIGEN: "ToxiGen: A Large-Scale Machine-Generated Dataset for Adversarial and Implicit Hate Speech Detection".

    Thomas Hartvigsen et al. ACL 2022. [Paper] [GitHub] [Source]

  16. CDail-Bias: "Towards Identifying Social Bias in Dialog Systems: Frame, Datasets, and Benchmarks".

    Jingyan Zhou et al. EMNLP 2022. [Paper] [GitHub]

  17. CORGI-PM: "CORGI-PM: A Chinese Corpus For Gender Bias Probing and Mitigation".

    Ge Zhang et al. arXiv 2023. [Paper] [GitHub]

  18. HateCheck: "HateCheck: Functional Tests for Hate Speech Detection Models".

    Paul Röttger et al. ACL/IJCNLP 2021. [Paper] [GitHub]

  19. StereoSet: "StereoSet: Measuring stereotypical bias in pretrained language models".

    Moin Nadeem et al. ACL/IJCNLP 2021. [Paper] [GitHub] [Source]

  20. CrowS-Pairs: "CrowS-Pairs: A Challenge Dataset for Measuring Social Biases in Masked Language Models".

    Nikita Nangia et al. EMNLP 2020. [Paper] [GitHub] [Source]

  21. "Does gender matter? towards fairness in dialogue systems".

    Haochen Liu et al. COLING 2020. [Paper] [GitHub]

  22. BOLD: "BOLD: Dataset and Metrics for Measuring Biases in Open-Ended Language Generation".

    Jwala Dhamala et al. FAccT 2021. [Paper] [GitHub] [Source]

  23. HolisticBias: "“I’m sorry to hear that”: Finding New Biases in Language Models with a Holistic Descriptor Dataset".

    Eric Michael Smith et al. EMNLP 2022. [Paper] [GitHub]

  24. Multilingual Holistic Bias: "Multilingual Holistic Bias: Extending Descriptors and Patterns to Unveil Demographic Biases in Languages at Scale".

    Eric Michael Smith et al. arXiv 2023. [Paper]

  25. Unqover: "UNQOVERing Stereotyping Biases via Underspecified Questions".

    Tao Li et al. EMNLP 2020. [Paper] [GitHub]

  26. BBQ: "BBQ: A Hand-Built Bias Benchmark for Question Answering".

    Alicia Parrish et al. ACL 2022. [Paper] [GitHub]

  27. CBBQ: "CBBQ: A Chinese Bias Benchmark Dataset Curated with Human-AI Collaboration for Large Language Models".

    Yufei Huang et al. arXiv 2023. [Paper] [GitHub]

  28. "Gender Bias in Multilingual Embeddings and Cross-Lingual Transfer".

    Jieyu Zhao et al. ACL 2020. [Paper] [GitHub]

  29. FairLex: "FairLex: A Multilingual Benchmark for Evaluating Fairness in Legal Text Processing".

    Ilias Chalkidis et al. ACL 2022. [Paper] [GitHub]

  30. "Nuanced Metrics for Measuring Unintended Bias with Real Data for Text Classification".

    Daniel Borkan et al. WWW 2019. [Paper]

  31. "On measuring and mitigating biased inferences of word embeddings".

Sunipa Dev et al. AAAI 2020. [Paper]

  1. "An Empirical Study of Metrics to Measure Representational Harms in Pre-Trained Language Models".

    Saghar Hosseini et al. TrustNLP 2023. [Paper] [GitHub]

  2. "Revealing Persona Biases in Dialogue Systems".

    Emily Sheng et al. arXiv 2021. [Paper] [GitHub]

  3. "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? ".

    Emily M. Bender et al. FAccT 2021. [Paper]

  4. "A Survey on Hate Speech Detection using Natural Language Processing."

    Anna Schmidt et al. SocialNLP 2017. [Paper]

  5. "Red teaming ChatGPT via Jailbreaking: Bias, Robustness, Reliability and Toxicity".

    Terry Yue Zhuo et al. arXiv 2023. [Paper]

Toxicity

  1. OLID: "Predicting the Type and Target of Offensive Posts in Social Media".

    Marcos Zampiari et al. NAACL-HLT 2019. [Paper]

  2. SOLID: "The narrativeqa reading comprehension challenge".

    Sara Rosenthal et al. ACL/IJCNLP (Findings) 2021. [Paper] [Source]

  3. OLID-BR: "OLID‑BR: ofensive language identifcation dataset for Brazilian Portuguese".

    Douglas Trajano et al. LRE 2023. [Paper] [Github]

  4. KODOLI: ""Why do I feel offended?" - Korean Dataset for Offensive Language Identification".

    San-Hee Park et al. EACL (Findings) 2023. [Paper] [Github]

  5. RealToxicityPrompts: "RealToxicityPrompts: Evaluating Neural Toxic Degeneration in Language Models."

    Samuel Gehman et al. EMNLP (Findings) 2020 [Paper] [Source]

  6. HarmfulQ: "On Second Thought, Let’s Not Think Step by Step! Bias and Toxicity in Zero-Shot Reasoning".

    Omar Shaikh et al. ACL 2023. [Paper] [Github]

  7. "Toxicity in ChatGPT: Analyzing Persona-assigned Language Models".

    Ameet Deshpande et al. arXiv 2023 [Paper]

  8. "Red teaming ChatGPT via Jailbreaking: Bias, Robustness, Reliability and Toxicity".

    Terry Yue Zhuo et al. arXiv 2023. [Paper]

Truthfulness

  1. NewsQA: "NewsQA: A Machine Comprehension Dataset".

    Adam Trischler, Tong Wang, and Xingdi Yuan et al. Rep4NLP@ACL 2017. [Paper] [GitHub]

  2. SQuAD 2.0: "Know What You Don't Know: Unanswerable Questions for SQuAD".

    Pranav Rajpurkar and Robin Jia et al. ACL 2018. [Paper] [Source]

  3. BIG-bench: "Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models".

    Aarohi Srivastava et al. arXiv 2022. [Paper] [GitHub]

  4. SelfAware: "Do Large Language Models Know What They Don’t Know?".

    Zhangyue Yin et al. ACL (Findings) 2023. [Paper] [GitHub]

  5. TruthfulQA: "TruthfulQA: Measuring How Models Mimic Human Falsehoods".

    Stephanie Lin et al. ACL 2022. [Paper] [GitHub]

  6. HalluQA: "Evaluating Hallucinations in Chinese Large Language Models".

    Qinyuan Cheng et al. arXiv 2023. [Paper] [GitHub]

  7. DialFact: "DialFact: A Benchmark for Fact-Checking in Dialogue".

    Prakhar Gupta et al. ACL 2022. [Paper] [GitHub]

  8. "Q2: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering".

    Or Honovich et al. EMNLP 2021. [Paper] [GitHub]

  9. BEGIN: "Evaluating Attribution in Dialogue Systems: The BEGIN Benchmark".

    Nouha Dziri and Hannah Rashkin et al. TACL 2022. [Paper] [GitHub]

  10. ConsisTest: "What Was Your Name Again? Interrogating Generative Conversational Models For Factual Consistency Evaluation".

    Ehsan Lotfi et al. GEM 2022. [Paper] [GitHub]

  11. XSumFaith: "On Faithfulness and Factuality in Abstractive Summarization".

    Joshua Maynez and Shashi Narayan et al. ACL 2020. [Paper] [GitHub]

  12. FactCC: "Evaluating the Factual Consistency of Abstractive Text Summarization".

    Wojciech Kryściński et al. EMNLP 2020. [Paper] [GitHub]

  13. SummEval: "SummEval: Re-evaluating Summarization Evaluation".

    Alexander R. Fabbri and Wojciech Kryściński et al. TACL 2021. [Paper] [GitHub]

  14. FRANK: "Understanding Factuality in Abstractive Summarization with FRANK: A Benchmark for Factuality Metrics".

    Artidoro Pagnoni et al. NAACL 2021. [Paper] [GitHub]

  15. SummaC: "SummaC: Re-Visiting NLI-based Models for Inconsistency Detection in Summarization".

    Philippe Laban et al. TACL 2022. [Paper] [GitHub]

  16. "Asking and Answering Questions to Evaluate the Factual Consistency of Summaries".

    Alex Wang et al. ACL 2020. [Paper] [GitHub]

  17. "Annotating and Modeling Fine-grained Factuality in Summarization".

    Tanya Goyal et al. NAACL 2021. [Paper] [GitHub]

  18. "Hallucinated but Factual! Inspecting the Factuality of Hallucinations in Abstractive Summarization".

    Meng Cao et al. ACL 2022. [Paper] [GitHub]

  19. CLIFF: "CLIFF: Contrastive Learning for Improving Faithfulness and Factuality in Abstractive Summarization".

    Shuyang Cao et al. EMNLP 2021. [Paper] [GitHub]

  20. AggreFact: "Understanding Factual Errors in Summarization: Errors, Summarizers, Datasets, Error Detectors".

    Liyan Tang et al. ACL 2023. [Paper] [GitHub]

  21. PolyTope: "What Have We Achieved on Text Summarization?".

    Dandan Huang and Leyang Cui et al. EMNLP 2020. [Paper] [GitHub]

  22. FIB: "Evaluating the Factual Consistency of Large Language Models Through News Summarization".

    Derek Tam et al. ACL (Findings) 2023. [Paper] [GitHub]

  23. FacTool: "FacTool: Factuality Detection in Generative AI -- A Tool Augmented Framework for Multi-Task and Multi-Domain Scenarios".

    I-Chun Chern et al. arXiv 2023. [Paper] [GitHub]

  24. CONNER: "Beyond Factuality: A Comprehensive Evaluation of Large Language Models as Knowledge Generators".

    Liang Chen et al. EMNLP 2023. [Paper] [GitHub]

  25. FActScore: "FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation".

    Sewon Min et al. EMNLP 2023. [Paper] [GitHub]

  26. SelfCheckGPT: "SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Models".

    Potsawee Manakul et al. EMNLP 2023. [Paper] [GitHub]

  27. SAPLMA: "The Internal State of an LLM Knows When It's Lying".

    Amos Azaria et al. arXiv 2023. [Paper]

  28. "Teaching Models to Express Their Uncertainty in Words".

    Stephanie Lin et al. arXiv 2022. [Paper]

  29. "Language Models (Mostly) Know What They Know".

    Saurav Kadavath et al. arXiv 2022. [Paper]

  30. "Dialogue Natural Language Inference".

    Sean Welleck et al. ACL 2019. [Paper]

  31. "Ranking Generated Summaries by Correctness: An Interesting but Challenging Application for Natural Language Inference".

    Tobias Falke et al. ACL 2019. [Paper]

  32. "mFACE: Multilingual Summarization with Factual Consistency Evaluation".

    Roee Aharoni et al. arXiv 2022. [Paper]

  33. "Falsesum: Generating Document-level NLI Examples for Recognizing Factual Inconsistency in Summarization".

    Prasetya Ajie Utama et al. NAACL 2022. [Paper] [GitHub]

  34. "Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback".

    Paul Roit, Johan Ferret, and Lior Shani et al. ACL 2023. [Paper]

  35. FEQA: "FEQA: A Question Answering Evaluation Framework for Faithfulness Assessment in Abstractive Summarization".

    Esin Durmus et al. ACL 2020. [Paper] [GitHub]

  36. QuestEval: "QuestEval: Summarization Asks for Fact-based Evaluation".

    Thomas Scialom et al. EMNLP 2021. [Paper] [GitHub]

  37. QAFactEval: "QAFactEval: Improved QA-Based Factual Consistency Evaluation for Summarization".

    Alexander R. Fabbri et al. NAACL 2022. [Paper] [GitHub]

  38. FaithDial: "FaithDial: A Faithful Benchmark for Information-Seeking Dialogue".

    Nouha Dziri et al. TACL 2022. [Paper] [GitHub]

  39. "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions".

    Lorenzo Pacchiardi and Alex J. Chan et al. arXiv 2023. [Paper] [GitHub]

  40. "Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning".

    Fuxiao Liu et al. arXiv 2023. [Paper] [GitHub]

  41. "HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models".

    Fuxiao Liu and Tianrui Guan et al. arXiv 2023. [Paper] [GitHub]

  42. "Analyzing and Evaluating Faithfulness in Dialogue Summarization".

    Bin Wang et al. EMNLP 2022. [Paper] [GitHub]

General Alignment Evaluation

  1. "TrustGPT: A Benchmark for Trustworthy and Responsible Large Language Models".

    Yue Huang et al. arXiv 2023. [Paper] [GitHub]

  2. "Safety Assessment of Chinese Large Language Models".

    Hao Sun et al. arXiv 2023. [Paper] [GitHub] [Source]

  3. "FLASK: Fine-grained Language Model Evaluation Based on Alignment Skill Sets".

    Seonghyeon Ye et al. arXiv 2023. [Paper] [GitHub]

  4. "Judging LLM-as-a-judge with MT-Bench and Chatbot Arena".

    Lianmin Zheng et al. arXiv 2023. [Paper] [GitHub]

  5. "Helpful, Honest, & Harmless - a Pragmatic Alignment Evaluation".

    Amanda Askell et al. GitHub 2022. [GitHub]

  6. "A Critical Evaluation of Evaluations for Long-form Question Answering".

    Fangyuan Xu et al. arXiv 2023. [Paper] [GitHub]

  7. "AlpacaEval: An Automatic Evaluator of Instruction-following Models".

    Xuechen Li et al. Github 2023. [Github]

  8. "PandaLM: An Automatic Evaluation Benchmark for LLM Instruction Tuning Optimization".

    Yidong Wang et al. arXiv 2023. [Paper] [GitHub]

  9. "Large Language Models are not Fair Evaluators".

    Peiyi Wang et al. arXiv 2023. [Paper] [GitHub]

  10. "G-Eval: NLG Evaluation using GPT-4 with Better Human Alignment".

    Yang Liu et al. arXiv 2023. [Paper] [GitHub]

  11. "Benchmarking Foundation Models with Language-Model-as-an-Examiner".

    Yushi Bai et al. arXiv 2023. [Paper] [Source]

  12. "PRD: Peer Rank and Discussion Improve Large Language Model based Evaluations".

    Ruosen Li et al. arXiv 2023. [Paper] [Source]

  13. "SELF-INSTRUCT: Aligning Language Models with Self-Generated Instructions"

    Yizhong Wang et al. arXiv 2023. [Paper] [GitHub]

:closed_lock_with_key:Safety Evaluation

Robustness Evaluation

  1. PromptBench: "PromptBench: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts".

    Kaijie Zhu et al. arXiv 2023. [Paper] [Github]

  2. "On the Robustness of ChatGPT: An Adversarial and Out-of-distribution Perspective".

    Jindong Wang et al. ICLR 2023. [Paper] [Github]

  3. RobuT: "RobuT: A Systematic Study of Table QA Robustness Against Human-Annotated Adversarial Perturbations".

    Yilun Zhao et al. ACL 2023. [Paper] [Github]

  4. SynTextBench: "On Robustness-Accuracy Characterization of Large Language Models using Synthetic Datasets".

    Ching-Yun Ko et al. ICML 2023. [Paper]

  5. ReCode: "ReCode: Robustness Evaluation of Code Generation Models".

    Shiqi Wang et al. ACL 2023. [Paper] [Github]

  6. "Exploring the Robustness of Large Language Models for Solving Programming Problems".

    Atsushi Shirafuji et al. arXiv 2023. [Paper] [Github]

  7. "A Causal Framework to Quantify the Robustness of Mathematical Reasoning with Language Models".

    Alessandro Stolfo et al. ACL 2023. [Paper] [Github]

  8. DGSlow: "White-Box Multi-Objective Adversarial Attack on Dialogue Generation".

    Yufei Li et al. ACL 2023. [Paper] [Github]

  9. "Jailbreaking ChatGPT via Prompt Engineering: An Empirical Study".

    Yi Liu et al. arXiv 2023. [Paper]

  10. MasterKey: "MasterKey: Automated Jailbreak Across Multiple Large Language Model Chatbots".

    Gelei Deng et al. arXiv 2023. [Paper]

  11. JailBroken: "Jailbroken: How Does LLM Safety Training Fail?".

    Alexander Wei et al. NeurIPS 2023. [Paper]

  12. "Red teaming ChatGPT via Jailbreaking: Bias, Robustness, Reliability and Toxicity".

    Terry Yue Zhuo et al. arXiv 2023. [Paper]

  13. "On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex".

    Terry Yue Zhuo et al. EACL 2023. [Paper]

  14. "How Important are Good Method Names in Neural Code Generation? A Model Robustness Perspective".

    Guang Yang et al. TOSEM 2023. [Paper]

  15. "Ask Again, Then Fail: Large Language Models' Vacillations in Judgement".

    Qiming Xie and Zengzhi Wang et al. arXiv 2023. [Paper] [Github]

Risk Evaluation

  1. "Frontier AI Regulation: Managing Emerging Risks to Public Safety".

    Markus Anderljung et al. arXiv 2023. [Paper]

  2. "Model evaluation for extreme risks".

    Toby Shevlane et al. arXiv 2023. [Paper]

  3. "Is Power-Seeking AI an Existential Risk?".

    Joseph Carlsmith. arXiv 2023. [Paper]

Evaluating LLMs Behaviors
  1. "Discovering Language Model Behaviors with Model-Written Evaluations".

    Ethan Perez et al. ACL (Findings) 2023. [Paper]

  2. "Evaluating Superhuman Models with Consistency Checks".

    Lukas Fluri et al. arXiv 2023. [Paper]

  3. "Understanding Social Reasoning in Language Models with Language Models".

    Kanishk Gandhi et al. arXiv 2023. [Paper]

  4. "Towards the Scalable Evaluation of Cooperativeness in Language Models".

    Alan Chan et al. arXiv 2023. [Paper]

  5. "Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations".

    Yanda Chen et al. arXiv 2023. [Paper]

Evaluating LLMs as Agents
  1. "AgentBench: Evaluating LLMs as Agents".

    Xiao Liu et al. arXiv 2023. [Paper]

  2. "WebArena: A Realistic Web Environment for Building Autonomous Agents".

    Shuyan Zhou et al. arXiv 2023. [Paper]

  3. "Training Socially Aligned Language Models in Simulated Human Society".

    Ruibo Liu et al. arXiv 2023. [Paper]

  4. "AgentSims: An Open-Source Sandbox for Large Language Model Evaluation".

    Jiaju Lin et al. EMNLP 2023 demo track. [Paper]

  5. "Evaluating Language-Model Agents on Realistic Autonomous Tasks".

    Megan Kinniment et al. ARC Evals. [Paper]

  6. MINT: "MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback"

    Xingyao Wang et al. arXiv 2023. [Paper] [Github] [Source]

  7. "Leveraging Word Guessing Games to Assess the Intelligence of Large Language Models"

    Tian Liang et al. arXiv 2023. [Paper] [Github]

  8. "InterCode: Standardizing and Benchmarking Interactive Coding with Execution Feedback"

    John Yang et al. NeurIPS 2023 Datasets & Benchmarks track. [Paper] [Github] [Source]

:syringe::woman_judge::computer::moneybag:Specialized LLMs Evaluation

Biology and Medicine

  1. MultiMedQA: "Large Language Models Encode Clinical Knowledge".

    Karan Singhal, Shekoofeh Azizi and Tao Tu et al. arXiv 2022. [Paper]

  2. PubMedQA: "PubMedQA: A Dataset for Biomedical Research Question Answering".

    Qiao Jin et al. EMNLP 2019. [Paper] [GitHub]

  3. LiveQA: "Overview of the Medical Question Answering Task at TREC 2017 LiveQA".

    Asma Ben Abacha et al. TREC 2017. [Paper] [GitHub]

  4. CLUE: "Clinical language understanding evaluation (CLUE)".

    Travis R. Goodwin et al. arXiv 2022. [Paper]

  5. "Towards Expert-Level Medical Question Answering with Large Language Models".

    Karan Singhal, Tao Tu, Juraj Gottweis and Rory Sayres et al. arXiv 2023. [Paper]

  6. "Performance of ChatGPT on USMLE: Unlocking the Potential of Large Language Models for AI-Assisted Medical Education".

    Prabin Sharma et al. arXiv 2023. [Paper]

  7. "Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum".

    John W. Ayers et al. JAMA Internal Medicine 2023. [Paper]

  8. "Evaluating large language models on medical evidence summarization".

    Liyan Tang et al. npj Digital Medicine 2023. [Paper]

  9. "Can large language models reason about medical questions?".

    Valentin Liévin et al. arXiv 2023. [Paper] [GitHub]

  10. "Capabilities of GPT-4 on Medical Challenge Problems".

    Harsha Nori et al. arXiv 2023. [Paper]

  11. "Evaluating the performance of chatgpt in ophthalmology: An analysis of its successes and shortcomings".

    Fares Antaki et al. Ophthalmology Science 2023. [Paper]

  12. "Chatgpt goes to the operating room: evaluating gpt-4 performance and its potential in surgical education and training in the era of large language models".

    Namkee Oh et al. Annals of Surgical Treatment and Research 2023. [Paper]

Education

  1. "The AI teacher test: Measuring the pedagogical ability of blender and GPT-3 in educational dialogues".

    Anaïs Tack et al. arXiv 2022. [Paper] [GitHub]

  2. "Is ChatGPT a Good Teacher Coach? Measuring Zero-Shot Performance For Scoring and Providing Actionable Insights on Classroom Instruction".

    Rose Wang et al. BEA 2023. [Paper] [GitHub]

  3. "Learning gain differences between ChatGPT and human tutor generated algebra hints".

    Zachary A. Pardos et al. arXiv 2023. [Paper] [GitHub]

  4. "Can Large Language Models Provide Feedback to Students? A Case Study on ChatGPT".

    Wei Dai et al. ICALT 2023. [Paper]

Legislation

  1. "GPT-4 Passes the Bar Exam".

    Daniel Martin Katz et al. SSRN 2023. [Paper] [GitHub]

  2. L’ART: "How well do SOTA legal reasoning models support abductive reasoning?".

    Ha-Thanh Nguyen et al. ICLP 2023. [Paper]

  3. "GPT Takes the Bar Exam".

    Michael Bommarito II et al. arXiv 2022. [Paper] [GitHub]

  4. "ChatGPT Goes to Law School".

    Jonathan H. Choi et al. SSRN 2023. [Paper]

  5. "Explaining Legal Concepts with Augmented Large Language Models (GPT-4)".

    Jaromir Savelka et al. arXiv 2023. [Paper]

  6. "How Ready are Pre-trained Abstractive Models and LLMs for Legal Case Judgement Summarization?".

    Aniket Deroy et al. LegalAIIA 2023. [Paper]

  7. "Legal Prompting: Teaching a Language Model to Think Like a Lawyer".

    Fangyi Yu et al. arXiv 2022. [Paper]

  8. "Can GPT-3 Perform Statutory Reasoning?".

    AndrewBlair-Stanek et al. ICAIL 2023. [Paper] [GitHub]

  9. LawBench: "LawBench: Benchmarking Legal Knowledge of Large Language Models".

    Zhiwei Fei, Xiaoyu Shen and Dawei Zhu et al. arXiv 2023. [Paper] [GitHub]

  10. "A Comprehensive Evaluation of Large Language Models on Legal Judgment Prediction".

    Ruihao Shui et al. EMNLP Findings 2023. [Paper] [GitHub]

Computer Science

  1. "A Systematic Evaluation of Large Language Models of Code".

    Xu, Frank F et al. DL4C@ICLR 2022. [Paper] [Github]

  2. "Is your code generated by chatgpt really correct? rigorous evaluation of large language models for code generation".

    Liu J et al. arXiv 2023. [Paper] [Github]

  3. "Lost at C: A user study on the security implications of large language model code assistants".

    Sandoval G et al. arXiv 2023. [Paper]

  4. "SWE-bench: Can Language Models Resolve Real-World GitHub Issues?".

    Jimenez, Carlos E et al. arXiv 2023. [Paper] [Github] [Source]

  5. "InterCode: Standardizing and Benchmarking Interactive Coding with Execution Feedback".

    John Yang et al. NeurIPS 2023 Datasets & Benchmarks track. [Paper] [Github] [Source]

  6. "DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation".

    John Yang et al. arXiv 2023. [Paper] [Source]

Finance

  1. "Xuanyuan 2.0: A large chinese financial chat model with hundreds of billions parameters".

    Zhang X et al. CIKM 2023. [Paper]

  2. "FinBERT: A large language model for extracting information from financial text".

    Huang A H et al. Contemporary Accounting Research 2023. [Paper]

  3. "ChatGPT: Unlocking the future of NLP in finance".

    Zaremba A et al. SSRN 2023. [Paper]

  4. "GPT as a Financial Advisor".

    Niszczota P et al. SSRN 2023. [Paper]

:earth_americas:Evaluation Organization

Benchmarks for NLU and NLG

  1. GLUE: "GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding".

    Alex Wang et al. ICLR 2019. [Paper] [Source]

  2. SuperGLUE: "SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems".

    Alex Wang et al. NeurIPS 2019. [Paper] [Source]

  3. LongBench: "LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding".

    Yushi Bai et al. arXiv 2023. [Paper] [GitHub]

Benchmarks for Knowledge and Reasoning

  1. MMLU: "Measuring Massive Multitask Language Understanding".

    Dan Hendrycks et al. ICLR 2021. [Paper] [GitHub]

  2. MMCU: "Measuring Massive Multitask Chinese Understanding".

    Hui Zeng et al. arXiv 2023. [Paper] [GitHub]

  3. C-Eval: "C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models".

    Yuzhen Huang et al. arXiv 2023. [Paper] [Source]

  4. M3KE: "M3KE: A Massive Multi-Level Multi-Subject Knowledge Evaluation Benchmark for Chinese Large Language Models".

    Chuang Liu et al. arXiv 2023. [Paper] [GitHub]

  5. CMMLU: "CMMLU: Measuring massive multitask language understanding in Chinese".

    Haonan Li et al. arXiv 2023. [Paper] [GitHub]

  6. AGIEval: "AGIEval: A Human-Centric Benchmark for Evaluating Foundation Models".

    Wanjun Zhong et al. arXiv 2023. [Paper] [GitHub]

  7. M3Exam: "M3Exam: A Multilingual, Multimodal, Multilevel Benchmark for Examining Large Language Models".

    Wenxuan Zhang et al. arXiv 2023. [Paper] [GitHub]

  8. LucyEval: "Evaluating the Generation Capabilities of Large Chinese Language Models".

    Hui Zeng et al. arXiv 2023. [Paper] [Source] [GitHub]

Benchmark for Holistic Evaluation

  1. Big-bench: "Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models".

    Dan Hendrycks et al. ICLR 2021. [Paper] [GitHub]

  2. Evaluation Harness: "A framework for few-shot language model evaluation".

    Leo Gao et al. arXiv 2023. [GitHub]

  3. HELM: "Holistic Evaluation of Language Models".

    Yuzhen Huang et al. arXiv 2023. [Paper] [Source] [GitHub]

  4. OpenAI Evals [GitHub]

  5. GPT-Fathom: "GPT-Fathom: Benchmarking Large Language Models to Decipher the Evolutionary Path towards GPT-4 and Beyond".

    Shen Zheng and Yuyu Zhang et al. arXiv 2023. [Paper] [GitHub]

  6. "INSTRUCTEVAL: Towards Holistic Evaluation of Instruction-Tuned Large Language Models".

    Yew Ken Chia et al. arXiv 2023. [Paper] [Source] [GitHub]

  7. Huggingface Open LLM Leaderboard [Source]

  8. Chatbot Arena: "Judging LLM-as-a-judge with MT-Bench and Chatbot Arena".

    Lianmin Zheng et al. arXiv 2023. [Paper] [Source] [GitHub]

  9. FlagEval [Source] [GitHub]

  10. OpenCompass: "Evaluating the Generation Capabilities of Large Chinese Language Models".

    Yuan Liu et al. arXiv 2023. [Source] [GitHub]

  11. CLEVA: "CLEVA: Chinese Language Models EVAluation Platform".

    Yanyang Li et al. arXiv 2023. [Paper] [Source] [GitHub]

  12. OpenEval [Source]

LLM Leaderboards

PlatformAccessDomain
Chatbot Arena[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
CLEVA[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
FlagEval[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
HELM[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
Huggingface Open LLM Leaderboard[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
InstructEval[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
LLMonitor[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
OpenCompass[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
Open Ko-LLM[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
SuperCLUE[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
TheoremOne LLM Benchmarking Metrics[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
Toloka[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
Open Multilingual LLM Eval[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
OpenEval[Source]Evaluation Organization/ Benchmark for Holistic Evaluation
ANGO[Source]Evaluation Organization/ Benchmarks for Knowledge and Reasoning
C-Eval[Source]Evaluation Organization/ Benchmarks for Knowledge and Reasoning
LucyEval[Source]Evaluation Organization/ Benchmarks for Knowledge and Reasoning
MMLU[Source]Evaluation Organization/ Benchmarks for Knowledge and Reasoning
OpenKG LLM[Source]Evaluation Organization/ Benchmarks for Knowledge and Reasoning
SEED-Bench[Source]Evaluation Organization/ Benchmarks for NLU and NLG
SuperGLUE[Source]Evaluation Organization/ Benchmarks for NLU and NLG
Toolbench[Source]Knowledge and Capability Evaluation/ Tool Learning
Hallucination Leaderboard[Source]Alignment Evaluation/ Truthfulness
AlpacaEval[Source]Alignment Evaluation/ General Alignment Evaluation
AgentBench[Source]Safety Evaluation/ Evaluating LLMs as Agents
InterCode[Source]Safety Evaluation/ Evaluating LLMs as Agents
SafetyBench[Source]Safety Evaluation
Nucleotide Transformer[Source]Specialized LLMs Evaluation/ Biology and Medicine
LAiW[Source]Specialized LLMs Evaluation/ Legislation
Big Code Models Leaderboard[Source]Specialized LLMs Evaluation/ Computer Science
Huggingface LLM Perf Leaderboard[Source]the Performance of LLMs

Contributors

<a href="https://github.com/tjunlp-lab/Awesome-LLMs-Evaluation-Papers/graphs/contributors"> <img src="https://contrib.rocks/image?repo=tjunlp-lab/Awesome-LLMs-Evaluation-Papers" /> </a>