SkillAgentSearch skills...

SciDef

Repository for the paper "SciDef: Automating Definition Extraction from Academic Literature with Large Language Models"

Install / Use

/learn @Media-Bias-Group/SciDef
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

SciDef: Automated Definition Extraction from Scientific Literature

SciDef - Workflow

<p align="center"> <a href="https://arxiv.org/abs/2602.05413"><img src="https://img.shields.io/badge/arXiv-2602.05413-b31b1b" alt="arXiv:2602.05413"></a> <a href="https://sigir.org/"><img src="https://img.shields.io/badge/SIGIR%202026-under%20review-0054a6" alt="SIGIR 2026 under review"></a> <a href="https://huggingface.co/datasets/mediabiasgroup/DefExtra"><img src="https://img.shields.io/badge/HF%20Dataset-DefExtra-ff9d00" alt="HF Dataset DefExtra"></a> <a href="https://huggingface.co/datasets/mediabiasgroup/DefSim"><img src="https://img.shields.io/badge/HF%20Dataset-DefSim-ff9d00" alt="HF Dataset DefSim"></a> <a href="https://media-bias-group.github.io/SciDef-ProjectPage/"><img src="https://img.shields.io/badge/Project%20Page-SciDef-2e7d32" alt="SciDef Project Page"></a> <a href="https://doi.org/10.5281/zenodo.18501198"><img src="https://img.shields.io/badge/Zenodo-10.5281%2Fzenodo.18501198-1682D4?logo=zenodo" alt="Zenodo DOI: 10.5281/zenodo.18501198"></a> </p>

Overview

With the rapid growth of publications, identifying definitions relevant to a given keyword has become increasingly difficult. SciDef provides resources to support research on definition extraction and definition similarity from scientific literature.

This repository contains:

  • An LLM-based definition extraction pipeline
  • Scripts for running and evaluating definition extraction
  • DefExtra, a human-annotated dataset for definition extraction
  • DefSim, a human-annotated dataset for definition similarity
  • Evaluation scripts covering multiple models, prompting strategies, and similarity metrics
  • DSPy-optimized prompts in artifacts directory for various open-weight & proprietary models

The goal of SciDef is to provide resources for reproducible research on on definition extraction from scientific articles.

Datasets

To facilitate future research in Definition Extraction from Scientific articles we publish 2 human annotated datasets.

DefExtra: Definition Extraction Dataset

DefExtra is a human-annotated dataset for the evaluation of definition extraction.

Content:

  • 268 definitions from 75 papers
  • 60 media bias related and 15 non-media bias related papers

Important:

  • The public DefExtra release ships markers only (no excerpts). You must hydrate it from your own PDFs, then convert the hydrated CSV to SciDef's JSON ground-truth format. See docs/defextra_integration.md.

DefSim: Definition Similarity Dataset

DefSim is a human-annotated dataset for the evaluation of definition similarity.

Content:

  • 60 definition definition pairs
  • Similarity rating on a 1-5 scale

Scripts and Usage

To support user-friendly usage, we provide scripts for running the SciDef pipeline, evaluation methods and other utility functions in the scripts/ directory.

SciDef uses uv for package and environment management.

Documentation (AI-generated)

DISCLAIMER: The documentation in docs/ was auto-generated with AI assistance. Please verify commands and settings in your environment.

Example

uv run python scripts/benchmark_nli.py --datasets stsb sick --sample-size 100

Note on Contribution

We have recreated the repository for clean release and due to squashing of Git history, the commits do not reflect author's contribution.

Citation

If you use this resource, please cite:

@misc{kucera2026scidefautomatingdefinitionextraction,
      title={SciDef: Automating Definition Extraction from Academic Literature with Large Language Models},
      author={Filip Ku\v{c}era and Christoph Mandl and Isao Echizen and Radu Timofte and Timo Spinde},
      year={2026},
      eprint={2602.05413},
      archivePrefix={arXiv},
      primaryClass={cs.IR},
      url={https://arxiv.org/abs/2602.05413},
}
View on GitHub
GitHub Stars7
CategoryDevelopment
Updated1d ago
Forks0

Languages

Python

Security Score

85/100

Audited on Apr 7, 2026

No findings