nick7nlp / Counting-Stars

Counting-Stars (★)
https://arxiv.org/abs/2403.11802
MIT License
76 stars 2 forks source link
evaluation-metrics large-language-model long-context

A Multi-evidence, Position-aware, and Scalable Benchmark for Evaluating Long-Context Large Language Models

In this work, we propose a multi-evidence, position-aware, and scalable benchmark for evaluating long-context LLMs, named Counting-Stars, which evaluates long-context LLMs by using two tasks: multi-evidence acquisition and multi-evidence reasoning.

Based on the Counting-Stars test, we conduct experiments to evaluate long-context LLMs (i.e., GPT-4 Turbo, Gemini 1.5 Pro, Claude3 Opus, GLM-4, and Moonshot-v1). Experimental results demonstrate that Gemini 1.5 Pro achieves the best overall results, while the performance of GPT-4 Turbo is the most stable across various tasks. Furthermore, our analysis of these LLMs, which are extended to handle long-context scenarios, indicates that there is potential for improvement as the length of the input context and the intricacy of the tasks are increasing.

Please find more details of this work in the paper.

Note

We'd like to encourage you to test the Counting-Stars using

, the 128K English and Chinese versions of the Counting-Stars.

Rank Models Claimed Length Me-Acq.(ZH) Me-Acq.(EN) Me-Rea.(ZH) Me-Rea.(EN) Avg.
1 Gemini 1.5 Pro 1M 0.775 0.833 0.575 0.371 0.639
2 GPT-4 Turbo (1106) 128K 0.697 0.718 0.473 0.651 0.635
3 Claude3 Opus 200K 0.807 0.705 0.488 0.374 0.594
4 GPT-4 Turbo (0125) 128K 0.663 0.662 0.386 0.610 0.580
5 Moonshot-v1 200K 0.606 0.559 0.344 0.460 0.492
6 GLM-4 128K 0.682 0.389 0.475 0.179 0.431
- Claude3 Sonnet 200K 0.788 - - - -
- Claude3 Haiku 200K 0.698 - - - -
- Baichuan3-Turbo 128K 0.759 0.490 - - -

Task Description

Evaluation Results

Visualization of the results on the Chinese version of the Counting-Stars-32-(Multi-evidence Acquisition).

Visualization of the results on the Chinese version of the Counting-Stars-32-(Multi-evidence Reasoning).

Cite

If you use this benchmark, please cite this paper

@misc{song2024countingstars,
      title={Counting-Stars: A Multi-evidence, Position-aware, and Scalable Benchmark for Evaluating Long-Context Large Language Models}, 
      author={Mingyang Song and Mao Zheng and Xuan Luo},
      year={2024},
      eprint={2403.11802},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

CONTACT

For any questions, feel free to create an issue, and we will try our best to solve it. \ If the problem is more urgent, you can email me simultaneously (I check email almost daily).

NAME: Mingyang Song
EMAIL: nickmysong@tencent.com

Our visualization code is built on the source code from NeedleInAHaystack. Thanks for their work.