gomate-community / rageval

Evaluation tools for Retrieval-augmented Generation (RAG) methods.
Apache License 2.0
81 stars 9 forks source link
evalution llm rag

Rageval

Evaluation tools for Retrieval-augmented Generation (RAG) methods.

python workflow status codecov pydocstyle PEP8

Rageval is a tool that helps you evaluate RAG system. The evaluation consists of six sub-tasks, including query rewriting, document ranking, information compression, evidence verify, answer generating, and result validating.

Definition of tasks and metrics

1. The generate task

The generate task is to answer the question based on the contexts provided by retrieval modules in RAG. Typically, the context could be extracted/generated text snippets from the compressor, or relevant documents from the re-ranker. Here, we divide metrics used in the generate task into two categories, namely answer correctness and answer groundedness.

(1) Answer Correctness: this category of metrics is to evaluate the correctness by comparing the generated answer with the groundtruth answer. Here are some commonly used metrics:

(2) Answer Groundedness: this category of metrics is to evaluate the groundedness (also known as factual consistency) by comparing the generated answer with the provided contexts. Here are some commonly used metrics:

2. The rewrite task

The rewrite task is to reformulate user question into a set of queries, making them more friendly to the search module in RAG.

3. The search task

The search task is to retrieve relevant documents from the knowledge base.

(1) Context Adequacy: this category of metrics is to evaluate the adequacy by comparing the retrieved documents with the groundtruth contexts. Here are some commonly used metrics:

(2) Context Relevance: this category of metrics is to evaluate the relevance by comparing the retrieved documents with the groundtruth answers. Here are some commonly used metrics:

Benchmark Results

1. ASQA benchmark

ASQA dataset is a question-answering dataset that contains factoid questions and long-form answers. The benchmark evaluates the correctness of the answer in the dataset. All detailed results can be download from this repo. Besides, these results can be reproduced based on the script in this repo.

Model Retriever Metric
String EM Rouge L Disambig F1 D-R Score
gpt-3.5-turbo-instruct no-retrieval 33.8 30.2 30.7 30.5
mistral-7b no-retrieval 20.6 31.1 26.6 28.7
llama2-7b-chat no-retrieval 21.7 30.7 28.0 29.3
llama3-8b-base no-retrieval 25.7 31.0 28.4 29.7
llama3-8b-instruct no-retrieval 27.1 30.9 29.4 30.1
solar-10.7b-instruct no-retrieval 23.0 24.9 28.1 26.5

2. ALCE Benchmark

ALCE is a benchmark for Automatic LLMs' Citation Evaluation. ALCE contains three datasets: ASQA, QAMPARI, and ELI5. All detailed results can be download from this repo. Besides, these results can be reproduced based on the script in this repo.

For more evaluation results, please view the benchmark's README: ALCE-ASQA and ALCE-ELI5.

Dataset Model Method Metric
retriever prompt MAUVE EM Recall Claim Recall Citation Recall Citation Precision
ASQA llama2-7b-chat GTR vanilla(5-psg) - 33.3 - 55.9 80.0
DPR vanilla(5-psg) - 29.2 - 49.2 81.0
Oracle vanilla(5-psg) - 41.7 - 58.1 78.9
ELI5 llama2-7b-chat BM25 vanilla(5-psg) - - 11.5 26.6 74.5
Oracle vanilla(5-psg) - - 17.8 34.0 75.6

Installation

git clone https://github.com/gomate-community/rageval.git
cd rageval
python setup.py install

Usage

1. Metric

Take F1 as an example.

from datasets import Dataset
import rageval as rl

sample = {
    "answers": [
        "Democrat rick kriseman won the 2016 mayoral election, while re- publican former mayor rick baker did so in the 2017 mayoral election."
    ],
    "gt_answers": [
        [
            "Kriseman",
            "Rick Kriseman"
        ]
    ]
}
dataset = Dataset.from_dict(sample)
metric = rl.metrics.AnswerF1Correctness()
score, dataset = metric.compute(dataset)

2. Benchmark

Benchmarks can be run directly using scripts (Take ALCE-ELI5 as an example).

bash benchmarks/ALCE/ELI5/run.sh

Contribution

Please make sure to read the Contributing Guide before creating a pull request.

About

This project is currently at its preliminary stage.