SkillAgentSearch skills...

EvaLearn

EvaLearn is a pioneering benchmark designed to evaluate large language models (LLMs) on their learning capability and efficiency in challenging tasks.

Install / Use

/learn @ByteDance-Seed/EvaLearn
About this skill

Quality Score

0/100

Category

Design

Supported Platforms

Universal

README

<div align="center"> <img src="logo.png" alt="Bytedance-seed" width="300"/> </div> <div align="center"> <h2>EvaLearn: Quantifying the Learning Capability and Efficiency of LLMs via Sequential Problem Solving</h2>

Paper Code License Data License

</div>

📰 News

  • 📅 Sep 18, 2025: EvaLearn was accepted to the NeurIPS 2025 main track with a high score of 5/5/5/5! 🎉
  • 📅 Jul 15, 2025: We've released a new version! 🎉 Open-sourced complete Chinese rubrics, updated Chinese README documentation, and optimized evaluation scripts for improved efficiency and accuracy.
  • 📅 Jun 5, 2025: EvaLearn is officially open-sourced! 🚀 We released this innovative benchmark for evaluating the learning capability and efficiency of large language models.

📚 Overview

EvaLearn is a benchmark designed to evaluate large language models (LLMs) on their learning capability and efficiency. It contains 648 challenging problems across six task types, grouped into 182 sequences. Unlike traditional benchmarks that evaluate models in parallel, EvaLearn requires models to solve problems sequentially, allowing them to leverage experience from previous solutions.

🧩 Framework Components

The EvaLearn evaluation framework consists of:

  1. A streamlined sequential evaluation tool (Evaluate/evaluate.py) that processes sequences of questions
  2. A dataset of problem definitions (Dataset/EvaLearn_Problem.json)
  3. A dataset of sequence definitions (Dataset/EvaLearn_Sequence.json)
  4. A metrics evaluation tool (Evaluate/evaluate_metric.py) for analyzing results

🚀 Getting Started

Installation

git clone https://github.com/YOUR_USERNAME/EvaLearn.git
cd EvaLearn
pip install -r requirements.txt

🛠️ Usage

Command Line Interface

Run the evaluation:

python Evaluate/evaluate.py --input Dataset/EvaLearn_Problem.json \
                               --seq Dataset/EvaLearn_Sequence.json \
                               --output results.json \
                               --workers 4 \
                               --client-api-key YOUR_CLIENT_API_KEY \
                               --judge-api-key YOUR_JUDGE_API_KEY

Command Line Arguments

| Argument | Description | | ------------------------- | ---------------------------------------------------------------- | | --input | Path to the problem JSON file | | --seq | Path to the sequence JSON file | | --output | Path to save the evaluation results | | --workers | Number of worker threads for parallel processing | | --no-check-empty | Skip checking for empty responses | | --judge-api-key | API key for the judge model | | --client-api-key | API key for the client model | | --judge-model | Model to use for judging (default: "gpt-4o-2024-11-20") | | --client-model | Model to use for client responses (default: "gpt-4o-2024-11-20") | | --judge-api-base-url | Custom base URL for judge API calls | | --client-api-base-url | Custom base URL for client API calls |

Key Features

  • Checkpoint Recovery: Automatically resumes interrupted evaluations
  • API Compatibility: Support for custom API endpoints
  • Parallel Processing: Multi-threaded execution for faster processing

Library Usage

from Evaluate.evaluate import sequentialEval

sequentialEval(
    input_json_path="Dataset/EvaLearn_Problem.json",
    seq_json_path="Dataset/EvaLearn_Sequence.json",
    output_json_path="results.json",
    client_api_key="YOUR_CLIENT_API_KEY",
    judge_api_key="YOUR_JUDGE_API_KEY"
)

📈 Evaluation Metrics

Use Evaluate/evaluate_metric.py to compute learning metrics from your results:

python Evaluate/evaluate_metric.py --results results.json --output report.json

Metrics

  • Overall sequence accuracy
  • Position-wise Accuracy
  • Slope of fitted accuracy curve
  • Average position of first correct solution
  • Average number of consecutive correct solutions
  • Post-warmup Accuracy

For detailed metric descriptions, please refer to the Section 2.3 of the paper.

Usage

1. Prepare Your Results

Your results should be in a JSON file, where each item contains at least:sequence_id: Unique identifier for a sequence

  • position_in_sequence: Position (1-based) of the problem in the sequence
  • type: (Optional) Task type/category
  • gpt4judge: String containing a JSON with an answer_score field

2. Run the Evaluation

python Evaluate/evaluate_metric.py --results <results.json> [--problems 7] [--warmup 3] [--output <report.json>]
  • --results: Path to your results JSON file (required)
  • --problems: Number of problems per sequence (default: 7)
  • --warmup: Number of initial problems to exclude for post-warmup accuracy (default: 3)
  • --output: Path to save the report as JSON (default: report_<results.json>)

3. Output

  • Prints a summary of all metrics to the console, including:
    • Overall metrics
    • Position-wise accuracy
    • Metrics by task type
  • Saves a detailed report as a JSON file (if --output is specified).

4. Example

python Evaluate/evaluate_metric.py --results my_eval_results.json --problems 7 --warmup 3 --output my_report.json

Logging

  • Logs are saved to evaluation_metrics.log and also printed to the console.

📊 Data Format

Problem JSON Format

Each problem in Dataset/EvaLearn_Problem.json has the following structure:

{
  "id": 1,
  "type": "Logical Reasoning",
  "source": "LogicGame-crypto_puzzle",
  "level": 1,
  "prompt": ["The question text that will be presented to the model"],
  "rubric_zh": "用于判断模型回答质量的中文评分标准",
  "rubric_en": "English evaluation criteria used by the judge model",
  "canonical_answer": "The expected correct answer"
}

| Field | Description | | -------------------- | ----------------------------------------------------------------------------- | | id | Unique identifier for the problem | | type | Category of the problem (e.g., "Logical Reasoning", "Mathematical Reasoning") | | source | Origin of the problem | | level | Difficulty level | | prompt | The question text (can be a string or an array of strings) | | rubric_zh | Chinese evaluation criteria used by the judge model | | rubric_en | English evaluation criteria used by the judge model | | canonical_answer | The expected correct answer |

Note: The results in our paper use the Chinese rubric, which was carefully annotated by our annotation team and is of high quality. The English version was translated using a large language model to help understand the meaning of the rubric. Therefore, we strongly recommend that everyone use the Chinese rubric for evaluation. We will also update it with a high-quality English rubric in the future.

Sequence JSON Format

Each sequence in Dataset/EvaLearn_Sequence.json has the following structure:

{
  "sequence_id": 1,
  "type": "Extraction",
  "question_ids": [252, 258, 297, 263, 245, 273, 241]
}

| Field | Description | | ---------------- | ------------------------------------------------------------------ | | sequence_id | Unique identifier for the sequence | | type | Category of the sequence (e.g., "Extraction", "Logical Reasoning") | | question_ids | Ordered list of problem IDs that form the sequence |

🔑 Key Functions

Main Evaluation Function

sequentialEval

The main evaluation function that processes sequences of questions with checkpoint recovery and API flexibility.

sequentialEval(
    input_json_path,
    seq_json_path,
    output_json_path,
    worker_nums=None,
    check_empty=True,
    judge_api_key=None,
    client_api_key=None,
    judge_model="gpt-4o-2024-11-20",
    client_model="gpt-4o-2024-11-20",
    judge_api_base_url=None,
    client_api_base_url=None
)

Parameters:

  • input_json_path: Path to the problem JSON file
  • seq_json_path: Path to the sequence JSON file
  • output_json_path: Path to save evaluation results
  • worker_nums: Number of worker threads (default: 5)
  • check_empty: Whether to check and reprocess empty responses (default: True)
  • judge_api_key: API key for judge model
  • client_api_key: API key for client model
  • judge_model: Model name for judging (default: "gpt-4o-2024-11-20")
  • client_model: Model name for responses (default: "gpt-4o-2024-11-20")
  • judge_api_base_url: Custom base URL for judge API
  • client_api_base_url: Custom base URL for client API

Core Processing Functions

sequential_infer_and_judge

Processes a sequence of questions with inference and judging.

process_sequence_batch

Processes a batch of sequences in parall

Related Skills

View on GitHub
GitHub Stars432
CategoryDesign
Updated13d ago
Forks12

Languages

Python

Security Score

100/100

Audited on Mar 7, 2026

No findings