Home

Awesome

<div align="center"> <img src="figures/logo.png" width="700px"/> </div> <div align="center"> <h1>A Multi-evidence, Position-aware, and Scalable Benchmark for Evaluating Long-Context Large Language Models</h1> </div>

In this work, we propose a multi-evidence, position-aware, and scalable benchmark for evaluating long-context LLMs, named Counting-Stars, which evaluates long-context LLMs by using two tasks: multi-evidence acquisition and multi-evidence reasoning.

Based on the Counting-Stars test, we conduct experiments to evaluate long-context LLMs (i.e., GPT-4 Turbo, Gemini 1.5 Pro, Claude3 Opus, GLM-4, and Moonshot-v1). Experimental results demonstrate that Gemini 1.5 Pro achieves the best overall results, while the performance of GPT-4 Turbo is the most stable across various tasks. Furthermore, our analysis of these LLMs, which are extended to handle long-context scenarios, indicates that there is potential for improvement as the length of the input context and the intricacy of the tasks are increasing.

Please find more details of this work in the paper.

Note

We'd like to encourage you to test the Counting-Stars using

, the 128K English and Chinese versions of the Counting-Stars.

RankModelsClaimed LengthMe-Acq.(ZH)Me-Acq.(EN)Me-Rea.(ZH)Me-Rea.(EN)Avg.
1Gemini 1.5 Pro1M0.7750.8330.5750.3710.639
2GPT-4 Turbo (1106)128K0.6970.7180.4730.6510.635
3Claude3 Opus200K0.8070.7050.4880.3740.594
4GPT-4 Turbo (0125)128K0.6630.6620.3860.6100.580
5Moonshot-v1200K0.6060.5590.3440.4600.492
6GLM-4128K0.6820.3890.4750.1790.431
-Claude3 Sonnet200K0.788----
-Claude3 Haiku200K0.698----
-Baichuan3-Turbo128K0.7590.490---

Task Description

<p> <img src="figures/tasks.png" width = "100%" /> </p>

Evaluation Results

<p> <img src="figures/results.png" width = "100%" /> </p> <p> <img src="figures/acq.png" width = "100%" /> </p>

Visualization of the results on the Chinese version of the Counting-Stars-32-(Multi-evidence Acquisition).

<p> <img src="figures/rea.png" width = "100%" /> </p>

Visualization of the results on the Chinese version of the Counting-Stars-32-(Multi-evidence Reasoning).

Cite

If you use this benchmark, please cite this paper

@misc{song2024countingstars,
      title={Counting-Stars: A Multi-evidence, Position-aware, and Scalable Benchmark for Evaluating Long-Context Large Language Models}, 
      author={Mingyang Song and Mao Zheng and Xuan Luo},
      year={2024},
      eprint={2403.11802},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

CONTACT

For any questions, feel free to create an issue, and we will try our best to solve it.
If the problem is more urgent, you can email me simultaneously (I check email almost daily).

NAME: Mingyang Song
EMAIL: nickmysong@tencent.com

Our visualization code is built on the source code from NeedleInAHaystack. Thanks for their work.