VisTR
[CVPR2021 Oral] End-to-End Video Instance Segmentation with Transformers
Install / Use
/learn @YuqingWang1029/VisTRREADME
VisTR: End-to-End Video Instance Segmentation with Transformers
This is the official implementation of the VisTR paper:
<p align="center"> <img src="https://user-images.githubusercontent.com/16319629/110786946-b99aa080-82a7-11eb-98e4-85478ca4eeac.png" width="600"> </p>Installation
We provide instructions how to install dependencies via conda. First, clone the repository locally:
git clone https://github.com/Epiphqny/vistr.git
Then, install PyTorch 1.6 and torchvision 0.7:
conda install pytorch==1.6.0 torchvision==0.7.0
Install pycocotools
conda install cython scipy
pip install -U 'git+https://github.com/cocodataset/cocoapi.git#subdirectory=PythonAPI'
pip install git+https://github.com/youtubevos/cocoapi.git#"egg=pycocotools&subdirectory=PythonAPI"
Compile DCN module(requires GCC>=5.3, cuda>=10.0)
cd models/dcn
python setup.py build_ext --inplace
Preparation
Download and extract 2019 version of YoutubeVIS train and val images with annotations from CodeLab or YoutubeVIS. We expect the directory structure to be the following:
VisTR
├── data
│ ├── train
│ ├── val
│ ├── annotations
│ │ ├── instances_train_sub.json
│ │ ├── instances_val_sub.json
├── models
...
Download the pretrained DETR models Google Drive BaiduYun(passcode:alge) on COCO and save it to the pretrained path.
Training
Training of the model requires at least 32g memory GPU, we performed the experiment on 32g V100 card. (As the training resolution is limited by the GPU memory, if you have a larger memory GPU and want to perform the experiment, please contact with me, thanks very much)
To train baseline VisTR on a single node with 8 gpus for 18 epochs, run:
python -m torch.distributed.launch --nproc_per_node=8 --use_env main.py --backbone resnet101/50 --ytvos_path /path/to/ytvos --masks --pretrained_weights /path/to/pretrained_path
Inference
python inference.py --masks --model_path /path/to/model_weights --save_path /path/to/results.json
Models
We provide baseline VisTR models, and plan to include more in future. AP is computed on YouTubeVIS dataset by submitting the result json file to the CodeLab system, and inference time is calculated by pure model inference time (without data-loading and post-processing).
<table> <thead> <tr style="text-align: right;"> <th></th> <th>name</th> <th>backbone</th> <th>FPS</th> <th>mask AP</th> <th>model</th> <th>result json zip</th> <th>detailed AP </th> </tr> </thead> <tbody> <tr> <th>0</th> <td>VisTR</td> <td>R50</td> <td>69.9</td> <td>36.2</td> <td><a href="https://drive.google.com/file/d/10lfe_QJSoZJzcJKxWoxt67QbJG35X55e/view?usp=sharing">vistr_r50.pth </a></td> <td><a href="https://drive.google.com/file/d/1vnZvxFR94EQ5TsrWixe368WMnqJ2KHIv/view?usp=sharing">vistr_r50.zip</a></td> <td><p align="center"> <img src="https://user-images.githubusercontent.com/16319629/115868905-b4686e00-a46f-11eb-9fe0-c9170026fca9.png" width="100"> </p></td> </tr> <tr> <th>1</th> <td>VisTR</td> <td>R101</td> <td>57.7</td> <td>40.1</td> <td><a href="https://drive.google.com/file/d/1WTkrpbITPsjVQESaetgI-nPyjQybRc2M/view?usp=sharing">vistr_r101.pth </a></td> <td><a href="https://drive.google.com/file/d/1bfZO3MNF9e0aO0W8vwnALsf_m84lCFfb/view?usp=sharing">vistr_r101.zip</a></td> <td> <p align="center"> <img src="https://user-images.githubusercontent.com/16319629/115869052-e974c080-a46f-11eb-92e3-1778aad83c71.png" width="100"> </p> </td> </tr> </table>License
VisTR is released under the Apache 2.0 license. Please see the LICENSE file for more information.
Acknowledgement
We would like to thank the DETR open-source project for its awesome work, part of the code are modified from its project.
Citation
Please consider citing our paper in your publications if the project helps your research. BibTeX reference is as follow.
@inproceedings{wang2020end,
title={End-to-End Video Instance Segmentation with Transformers},
author={Wang, Yuqing and Xu, Zhaoliang and Wang, Xinlong and Shen, Chunhua and Cheng, Baoshan and Shen, Hao and Xia, Huaxia},
booktitle = {Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR)},
year={2021}
}
Related Skills
docs-writer
98.6k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
328.7kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Design
Campus Second-Hand Trading Platform \- General Design Document (v5.0 \- React Architecture \- Complete Final Version)1\. System Overall Design 1.1. Project Overview This project aims t
arscontexta
2.8kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.
