ILCACM
[AAAI 2025 Oral] Implicit Location-Caption Alignment via Complementary Masking for Weakly-Supervised Dense Video Captioning
Install / Use
/learn @ShipingGe/ILCACMREADME
Repository for our AAAI'25 paper titled "Implicit Location-Caption Alignment via Complementary Masking for Weakly-Supervised Dense Video Captioning".
About
Weakly-Supervised Dense Video Captioning (WSDVC) aims to localize and describe all events of interest in a video without requiring annotations of event boundaries. This setting poses a great challenge in accurately locating the temporal location of event, as the relevant supervision is unavailable. Existing methods rely on explicit alignment constraints between event locations and captions, which involve complex event proposal procedures during both training and inference. To tackle this problem, we propose a novel implicit location-caption alignment paradigm by complementary masking, which simplifies the complex event proposal and localization process while maintaining effectiveness. Specifically, our model comprises two components: a dual-mode video captioning module and a mask generation module. The dual-mode video captioning module captures global event information and generates descriptive captions, while the mask generation module generates differentiable positive and negative masks for localizing the events. These masks enable the implicit alignment of event locations and captions by ensuring that captions generated from positively and negatively masked videos are complementary, thereby forming a complete video description. In this way, even under weak supervision, the event location and event caption can be aligned implicitly. Extensive experiments on the public datasets demonstrate that our method outperforms existing weakly-supervised methods and achieves competitive results compared to fully-supervised methods.
<div align="center"> <img src="img/framework.jpg" width="80%" alt="wsdvc" align="center"> </div>Preparation
- Download the dataset and the base model.
- Install the requirements.
Training & Inference
cd src
accelerate launch --config_file accelerate.yaml main_accelerate.py --cap-epochs 10 --total-epochs 20 --lr 1e-4
Citation
If you find our paper and code useful in your research, please consider giving a star ⭐ and citation 📝 :)
@misc{ge2024implicitlocationcaptionalignmentcomplementary,
title={Implicit Location-Caption Alignment via Complementary Masking for Weakly-Supervised Dense Video Captioning},
author={Shiping Ge and Qiang Chen and Zhiwei Jiang and Yafeng Yin and Liu Qin and Ziyao Chen and Qing Gu},
year={2024},
eprint={2412.12791},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2412.12791},
}
Related Skills
qqbot-channel
352.2kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.6k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
352.2kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
arscontexta
3.1kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.
