AMD
[CVPR 2026] Code for "The Coherence Trap: When MLLM-Crafted Narratives Exploit Manipulated Visual Contexts"
Install / Use
/learn @YcZhangSing/AMDREADME
The Coherence Trap: When MLLM-Crafted Narratives Exploit Manipulated Visual Contexts
<font size=4><div align='center' > [📄 Paper]
</div></font> <!-- <font size=4><div align='center' > [\[📄 Paper\]](https://arxiv.org/abs/2505.17476) [\[🤗 Datasets\]](https://www.kaggle.com/datasets/yaxiongwang/mdsm-dataset-under-review/data) </div></font> -->🔥 Overview
We propose MLLM-Driven Synthetic Multimodal (MDSM), a large scale, semantic-aligned multi-modal benchmark with high-fidelity texts from MLLM, supporting fake news detecting and grounding tasks. <a href=""> <img src="assets/overview.png" alt="Logo" > </a>
And an Artifact-aware Manipulation Diagnosis framework (AMD) for the MDSM problem that synergizes artifact pre-perception encoding and manipulation-oriented reasoning to effectively adapt MLLMs for precise manipulation analysis is proposed.
Setup
conda create -n AMD python=3.10
conda activate AMD
pip install -r requirements.txt
If you encounter issues with flash-attn, you can fix it with the following command:
pip install -U flash-attn --no-build-isolation
Alternatively, you can visit the flash-attention releases page to find the version compatible with your environment and follow the installation instructions provided there.
Data
Our MDSM dataset is avaliable at at kaggle. And the DGM4 dataset is avaliable at DGM4
You can also prepare your training or inference data like:
images/:
image1.png
image2.png
test.json:
[
{
"id": 1556711,
"image": "image1.jpg",
"text": "your text1",
"fake_cls": "orig",
"fake_image_box": [],
"mtcnn_boxes": [
[
113,
27,
208,
165
],
[
436,
56,
491,
124
]
]
},
{
"id": 236385,
"image": "DGM4/origin/usa_today/0244/128.jpg",
"text": "your text1",
"fake_cls": "face_swap&text_swap",
"fake_image_box": [
64,
48,
145,
159
],
"mtcnn_boxes": [
[
64,
48,
145,
159
],
[
104,
180,
179,
285
]
]
},
]
Inference
Please run:
sh test.sh
Training
Please run:
sh train.sh
Please fill the MODEL_PATH, TRAIN_JS, and VAL_JS with your real checkpoint path and data path.
🗞️ News
2026-02-22: Our paper is accepted by CVPR 2026.2025-05-15: We release the AMD repository.
🤝 Acknowledgements
We sincerely thank projects DGM4 and Florence2 for providing their open-source resources.
