SkillAgentSearch skills...

TARS

TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs

Install / Use

/learn @KejiaZhang-Robust/TARS
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

<div align="center"> <h2 style="font-size: 36px; font-weight: bold; color: #333;"> TARS: MinMax Token-Adaptive Preference Strategy for MLLM Hallucination Reduction </h2> <h4 style="font-size: 20px; color: #777; font-style: italic;"> A tribute to TARS from <i>Interstellar</i> — not piloting through wormholes, but steering MLLMs away from the gravity of hallucination. </h4> </div> <div align="center" style="margin-top: 20px;"> <!-- GitHub Badges --> <img alt="GitHub Repo stars" src="https://img.shields.io/github/stars/KejiaZhang-Robust/TARS?style=social" style="margin: 0 5px;"> <img alt="GitHub forks" src="https://img.shields.io/github/forks/KejiaZhang-Robust/TARS?style=social" style="margin: 0 5px;"> <a href="https://arxiv.org/abs/2507.21584"> <img src="https://img.shields.io/badge/arXiv-2507.21584-b31b1b?style=flat-square" alt="arXiv" style="margin: 0 0px;" /> </a> <a href="https://kejiazhang-robust.github.io/tars_web/"> <img src="https://img.shields.io/badge/Project Page-TARS-008080?style=flat-square" alt="Project Page" style="margin: 0 5px;"> </a> <img alt="GitHub License" src="https://img.shields.io/github/license/KejiaZhang-Robust/TARS?style=flat-square" style="margin: 0 5px;"> <img alt="Language" src="https://img.shields.io/github/languages/top/KejiaZhang-Robust/TARS?style=flat-square&color=9acd32" style="margin: 0 5px;"> </div> <div align="center" style="margin-top: 30px;"> <h3 style="font-size: 24px; font-weight: bold; color: #333;"> Kejia Zhang, Keda Tao, Zhiming Luo, Chang Liu, Jiasheng Tang, Huan Wang </h3> </div> <!-- LOGO --> <div align="center" style="margin-top: 20px;"> <img src="image/logo.png" height="100" alt="Logos" style="margin-right: 20px; display: inline-block;"> </div>

📖 Paper Teaser

<div align="center" style="margin-top: 20px;"> <img src="image/Teaser.png" alt="TARS Teaser" width="80%" style="border-radius: 8px; box-shadow: 0 4px 8px rgba(0, 0, 0, 0.1);"> </div> <div align="center" style="margin-top: 15px;"> <p style="font-size: 12px; font-weight: 500; color: #444;"> <b>Left:</b> We present <i>TARS</i>, a <u>t</u>oken-<u>a</u>daptive p<u>r</u>eference <u>s</u>trategy for mitigating hallucinations in MLLMs. TARS reformulates Direct Preference Optimization (DPO) as a min-max objective that (1) minimizes behavioral misalignment via preference feedback, and (2) maximizes adaptability through perturbations of visual-agnostic tokens. <br><br> <b>Right:</b> Evaluation on LLaVA-v1.5-13B and industrial MLLMs under the AMBER benchmark shows that TARS consistently outperforms standard DPO baselines and matches GPT-4o in hallucination suppression. </p> </div>

🚀 News

📢 [2025-07-28] TARS is now open-source! Check out the repo and get started. 🔥


🧪 Quick Start

📦 Environment Setup

conda create -n DPO python=3.10 -y
conda activate DPO
pip install -e .

🔧 Base Models

We conduct experiments based on the following pretrained models:

📊 Hallucination Benchmarks

We evaluate hallucination suppression performance on several widely-used benchmarks:

📁 DPO Dataset

We adopt the RLHF-V-Dataset and sampled a subset of 4.8k data for training.

🚀 Run TARS-DPO

To launch training with our TARS-DPO strategy, simply run:

bash scripts/TARS.sh

📌 Citation

If you find our work helpful, please consider citing our paper:

@article{zhang2025tars,
  title={TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs},
  author={Zhang, Kejia and Tao, Keda and Luo, Zhiming and Liu, Chang and Tang, Jiasheng and Wang, Huan},
  journal={arXiv preprint arXiv:2507.21584},
  year={2025}
}

Your citation helps support our research and further advances the field of reliable vision-language models. 🚀


View on GitHub
GitHub Stars24
CategoryDevelopment
Updated21d ago
Forks0

Languages

Python

Security Score

95/100

Audited on Mar 9, 2026

No findings