SkillAgentSearch skills...

Rageval

Evaluation tools for Retrieval-augmented Generation (RAG) methods.

Install / Use

/learn @gomate-community/Rageval
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

Rageval

Evaluation tools for Retrieval-augmented Generation (RAG) methods.

python workflow status codecov pydocstyle PEP8

Rageval is a tool that helps you evaluate RAG system. The evaluation consists of six sub-tasks, including query rewriting, document ranking, information compression, evidence verify, answer generating, and result validating.

Definition of tasks and metrics

1. The generate task

The generate task is to answer the question based on the contexts provided by retrieval modules in RAG. Typically, the context could be extracted/generated text snippets from the compressor, or relevant documents from the re-ranker. Here, we divide metrics used in the generate task into two categories, namely answer correctness and answer groundedness.

(1) Answer Correctness: this category of metrics is to evaluate the correctness by comparing the generated answer with the groundtruth answer. Here are some commonly used metrics:

(2) Answer Groundedness: this category of metrics is to evaluate the groundedness (also known as factual consistency) by comparing the generated answer with the provided contexts. Here are some commonly used metrics:

2. The rewrite task

The rewrite task is to reformulate user question into a set of queries, making them more friendly to the search module in RAG.

3. The search task

The search task is to retrieve relevant documents from the knowledge base.

(1) Context Adequacy: this category of metrics is to evaluate the adequacy by comparing the retrieved documents with the groundtruth contexts. Here are some commonly used metrics:

(2) Context Relevance: this category of metrics is to evaluate the relevance by comparing the retrieved documents with the groundtruth answers. Here are some commonly used metrics:

Setup Evaluator LLMs

Some metrics evaluations rely on LLMs as evaluators. You can either directly call OpenAI's API or deploy an open-source model as a RESTful API in the OpenAI format for evaluation.

  • OpenAI
os.environ["OPENAI_API_KEY"] = "<your_openai_api_key>"
  • Open source LLMs

Please use vllm to setup the API server for open source LLMs. For example, use the following command to deploy a Llama-3-8B model hosted on HuggingFace:

python -m vllm.entrypoints.openai.api_server \
  --model meta-llama/Meta-Llama-3-8B-Instruct \
  --tensor-parallel-size 8 \
  --dtype auto \
  --api-key sk-123456789 \
  --gpu-memory-utilization 0.9 \
  --port 5000

Benchmark Results

1. ASQA benchmark

ASQA dataset is a question-answering dataset that contains factoid questions and long-form answers. The benchmark evaluates the correctness of the answer in the dataset. All detailed results can be download from this repo. Besides, these results can be reproduced based on the script in this repo.

<table> <col width=166> <col width=125> <col width=125 span=4> <tr> <td rowspan=2 align="center"><b>Model</b></td> <td rowspan=2 align="center"><b>Retriever</b></td> <td colspan=4 align="center"><b>Metric</b></td> </tr> <tr> <td align="center"><a href="rageval\metrics\_answer_exact_match.py">String EM</a></td> <td align="center"><a href="rageval\metrics\_answer_rouge_correctness.py">Rouge L</a></td> <td align="center"><a href="rageval\metrics\_answer_disambig_f1.py">Disambig F1</a></td> <td align="center"><a href="benchmarks\ASQA\asqa_benchmark.py">D-R Score</a></td> </tr> <tr> <td>gpt-3.5-turbo-instruct</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/gpt_3.5_turbo_instruct">no-retrieval</a></td> <td align="center">33.8</td> <td align="center">30.2</td> <td align="center">30.7</td> <td align="center">30.5</td> </tr> <tr> <td>mistral-7b</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/mistral_7b">no-retrieval</a></td> <td align="center">20.6</td> <td align="center">31.1</td> <td align="center">26.6</td> <td align="center">28.7</td> </tr> <tr> <td>llama2-7b-chat</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/llama2_7b_chat">no-retrieval</a></td> <td align="center">21.7</td> <td align="center">30.7</td> <td align="center">28.0</td> <td align="center">29.3</td> </tr> <tr> <td>llama3-8b-base</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/llama3_8b_base">no-retrieval</a></td> <td align="center">25.7</td> <td align="center">31.0</td> <td align="center">28.4</td> <td align="center">29.7</td> </tr> <tr> <td>llama3-8b-instruct</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/llama3_8b_instruct">no-retrieval</a></td> <td align="center">27.1</td> <td align="center">30.9</td> <td align="center">29.4</td> <td align="center">30.1</td> </tr> <tr> <td>solar-10.7b-instruct</td> <td><a href="https://huggingface.co/datasets/golaxy/rag-bench/viewer/asqa/solar_10.7b_instruct">no-retrieval</a></td> <td align="center">23.0</td> <td align="center">24.9</td> <td align="center">28.1</td> <td align="center">26.5</td> </tr> </table>

2. ALCE Benchmark

ALCE is a benchmark for Automatic LLMs' Citation Evaluation. ALCE contains three datasets: ASQA, QAMPARI, and ELI5. All detailed results can be download from this repo. Besides, these results can be reproduced based on the script in this repo.

For more evaluation results, please view the benchmark's README: ALCE-ASQA and ALCE-ELI5.

<table> <col width=75> <col width=125> <col width=85> <col width=145> <col width=125 span=5> <tr> <td rowspan=2 align="center"><b>Dataset</b></td> <td rowspan=2 align="center"><b>Model</b></td> <td colspan=2 align="center"><b>Method</b></td> <td colspan=5 align="center"><b>Metric</b></td> </tr> <tr> <td align="center">retriever</td> <td align="center">prompt</td> <td align="center">MAUVE</td> <td align="center"><a href="rageval\metrics\_answer_exact_match.py">EM Recall</a></td> <td align="center"><a href="rageval\metrics\_answer_claim_recall.py">Claim Recall</a></td> <td align="center"><a hr
View on GitHub
GitHub Stars170
CategoryDevelopment
Updated2mo ago
Forks10

Languages

Python

Security Score

100/100

Audited on Jan 31, 2026

No findings