MIML
[CVPR2024] Towards Modern Image Manipulation Localization: A Large-Scale Dataset and Novel Methods
Install / Use
/learn @qcf-568/MIMLREADME
Towards Modern Image Manipulation Localization: A Large-Scale Dataset and Novel Methods
This is the official implementation of the paper Towards Modern Image Manipulation Localization: A Large-Scale Dataset and Novel Methods. paper<br/>
The Version 2 is also available.
The main contributions of this work are as follows:
- <font size=10>We propose to harness Constrained Image Manipulation Localization (CIML) models to automatically annotate the numerous unlabelled manually forged images from the web (e.g. those from PhotoshopBattles). Thereby addressing the severe scarcity of non-synthetic data for image manipulation localization.</font>
- <font size=10>We propose a novel and effective paradigm CAAA for constrained image manipulation localization, which significantly improves the accuracy of the automatic annotations. We believe that this is the best paradigm for CIML-based auto-annotation.</font>
- <font size=10>We propose a novel metric QES to automatically to filter out the possible bad annotations, which is crucial to ensure the quality of an automatically annotated dataset. This metric is quite effective in reflecting the quality of the predictions during the construction of the dataset, where the ground truth is not available.</font>
- <font size=10>Based on the above techniques, we construct a large-scale dataset, termed as MIML, with 123,150 manually forged images and pixel-level annotations denoting the forged regions. The MIML dataset can significantly improve the generalization of different forgery localization models, especially on modern-style images (as those in IMD20 dataset).</font> <br/> <br/> <br/>
<font size=10>We are confirmed that large-scale non-synthetic data is vital for deep image manipulation localization models. We sincerely hope that our methods and our dataset can shed light on the community and promote the real-world applications of deep image forensic models. </font><br/>
<font size=10>This work is an initial attempt of automatic annotation for IML, futher improvements could be made. We are glad to witnessed the development of this field together with the community. </font><br/>
The Modern Image Manipulation Localization (MIML) dataset is now publicly available at Kaggle and Baidu Drive.
<font size=10>Researchers are welcome 😃 to apply for this dataset by sending an email to 202221012612@mail.scut.edu.cn (with institution email address) and explaining:</font><br/>
- Who you are and your institution.
- Who is your supervisor/mentor.
Disclaimer
The original data of the dataset is sourced from public channels such as the Internet, and its copyright shall remain with the original providers. The collated and annotated dataset presented in this case is for non-commercial use only and is currently licensed to universities and research institutions. To apply for the use of this dataset, please fill in the corresponding application form in accordance with the requirements specified on the dataset’s official website. The applicant must be a full-time employee of a university or research institute and is required to sign the application form. For the convenience of review, it is recommended to affix an official seal (a seal of a secondary-level department is acceptable).
Environment package version
Python3.9
pip install torch==1.13.0 torchvision==0.14.0 mmcv=1.6.0 prettytable matplotlib timm einops numpy==1.26.4 albumentations==1.3.1
<font size=10>Any question about this work please contact 202221012612@mail.scut.edu.cn.</font><br/>
License
The project is under CC-BY-NC-4.0 license.
Citation
If you use the MIML dataset, CAAA method or the APSC-Net, please cite this paper.
@inproceedings{qu2024towards,
title={Towards Modern Image Manipulation Localization: A Large-Scale Dataset and Novel Methods},
author={Qu, Chenfan and Zhong, Yiwu and Liu, Chongyu and Xu, Guitao and Peng, Dezhi and Guo, Fengjun and Jin, Lianwen},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={10781--10790},
year={2024}
}
