TARS
TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs
Install / Use
/learn @KejiaZhang-Robust/TARSREADME
<div align="center">
<h2 style="font-size: 36px; font-weight: bold; color: #333;">
TARS: MinMax Token-Adaptive Preference Strategy for MLLM Hallucination Reduction
</h2>
<h4 style="font-size: 20px; color: #777; font-style: italic;">
A tribute to TARS from <i>Interstellar</i> — not piloting through wormholes, but steering MLLMs away from the gravity of hallucination.
</h4>
</div>
<div align="center" style="margin-top: 20px;">
<!-- GitHub Badges -->
<img alt="GitHub Repo stars" src="https://img.shields.io/github/stars/KejiaZhang-Robust/TARS?style=social" style="margin: 0 5px;">
<img alt="GitHub forks" src="https://img.shields.io/github/forks/KejiaZhang-Robust/TARS?style=social" style="margin: 0 5px;">
<a href="https://arxiv.org/abs/2507.21584">
<img src="https://img.shields.io/badge/arXiv-2507.21584-b31b1b?style=flat-square" alt="arXiv" style="margin: 0 0px;" />
</a>
<a href="https://kejiazhang-robust.github.io/tars_web/">
<img src="https://img.shields.io/badge/Project Page-TARS-008080?style=flat-square" alt="Project Page" style="margin: 0 5px;">
</a>
<img alt="GitHub License" src="https://img.shields.io/github/license/KejiaZhang-Robust/TARS?style=flat-square" style="margin: 0 5px;">
<img alt="Language" src="https://img.shields.io/github/languages/top/KejiaZhang-Robust/TARS?style=flat-square&color=9acd32" style="margin: 0 5px;">
</div>
<div align="center" style="margin-top: 30px;">
<h3 style="font-size: 24px; font-weight: bold; color: #333;">
Kejia Zhang, Keda Tao, Zhiming Luo, Chang Liu, Jiasheng Tang, Huan Wang
</h3>
</div>
<!-- LOGO -->
<div align="center" style="margin-top: 20px;">
<img src="image/logo.png" height="100" alt="Logos" style="margin-right: 20px; display: inline-block;">
</div>
📖 Paper Teaser
<div align="center" style="margin-top: 20px;"> <img src="image/Teaser.png" alt="TARS Teaser" width="80%" style="border-radius: 8px; box-shadow: 0 4px 8px rgba(0, 0, 0, 0.1);"> </div> <div align="center" style="margin-top: 15px;"> <p style="font-size: 12px; font-weight: 500; color: #444;"> <b>Left:</b> We present <i>TARS</i>, a <u>t</u>oken-<u>a</u>daptive p<u>r</u>eference <u>s</u>trategy for mitigating hallucinations in MLLMs. TARS reformulates Direct Preference Optimization (DPO) as a min-max objective that (1) minimizes behavioral misalignment via preference feedback, and (2) maximizes adaptability through perturbations of visual-agnostic tokens. <br><br> <b>Right:</b> Evaluation on LLaVA-v1.5-13B and industrial MLLMs under the AMBER benchmark shows that TARS consistently outperforms standard DPO baselines and matches GPT-4o in hallucination suppression. </p> </div>🚀 News
📢 [2025-07-28] TARS is now open-source! Check out the repo and get started. 🔥
🧪 Quick Start
📦 Environment Setup
conda create -n DPO python=3.10 -y
conda activate DPO
pip install -e .
🔧 Base Models
We conduct experiments based on the following pretrained models:
📊 Hallucination Benchmarks
We evaluate hallucination suppression performance on several widely-used benchmarks:
📁 DPO Dataset
We adopt the RLHF-V-Dataset and sampled a subset of 4.8k data for training.
🚀 Run TARS-DPO
To launch training with our TARS-DPO strategy, simply run:
bash scripts/TARS.sh
📌 Citation
If you find our work helpful, please consider citing our paper:
@article{zhang2025tars,
title={TARS: MinMax Token-Adaptive Preference Strategy for Hallucination Reduction in MLLMs},
author={Zhang, Kejia and Tao, Keda and Luo, Zhiming and Liu, Chang and Tang, Jiasheng and Wang, Huan},
journal={arXiv preprint arXiv:2507.21584},
year={2025}
}
Your citation helps support our research and further advances the field of reliable vision-language models. 🚀
