SkillAgentSearch skills...

Mtmct

Design and Implementation of a Multi-Target Multi-Camera Tracking Solution

Install / Use

/learn @nolanzzz/Mtmct

README

English | 简体中文

MTMCT

This project demonstrates the design and implementation of a Multi-Target Multi-Camera Tracking (MTMCT) solution.

Pipeline of our solution: <img src="readme_files/pipeline_white.png" style="zoom:35%;" />

Tracking performance

Results and comparisons with FairMOT and wda_tracker trained and tested on a 6x2-minute MTA dataset

<table> <thead> <tr> <th rowspan=2>Method</th> <th colspan=5>Single-Camera</th> <th colspan=5>Multi-Camera</th> </tr> <tr> <th>MOTA</th> <th>IDF1</th> <th>IDs</th> <th>MT</th> <th>ML</th> <th>MOTA</th> <th>IDF1</th> <th>IDs</th> <th>MT</th> <th>ML</th> </tr> </thead> <tbody> <tr> <td>WDA</td> <td>58.2</td> <td>37.3</td> <td>534.2</td> <td>16.8%</td> <td>17.2</td> <td>46.6</td> <td>19.8</td> <td>563.8</td> <td>6.5%</td> <td>7.0%</td> </tr> <tr> <td>FairMOT</td> <td>64.1</td> <td><strong>48.0</strong></td> <td>588.2</td> <td>34.7%</td> <td>7.8%</td> <td>N/A</td> <td>N/A</td> <td>N/A</td> <td>N/A</td> <td>N/A</td> </tr> <tr> <td>Ours</td> <td><strong>70.8</strong></td> <td>47.8</td> <td><strong>470.2</strong></td> <td><strong>40.5%</strong></td> <td><strong>5.6%</strong></td> <td><strong>65.6</strong></td> <td><strong>31.5</strong></td> <td><strong>494.5</strong></td> <td><strong>31.2%</strong></td> <td><strong>1.1%</strong></td> </tr> </tbody> </table>

Demo on Multi Camera Track Auto (MTA) dataset

<img src="readme_files/cam_0_sample.png" width="400"/> <img src="readme_files/cam_1_sample.png" width="400"/> <img src="readme_files/cam_2_sample.png" width="400"/> <img src="readme_files/cam_3_sample.png" width="400"/> <img src="readme_files/cam_4_sample.png" width="400"/> <img src="readme_files/cam_5_sample.png" width="400"/>

Demo GIFs can be seen here

Full-length demo videos can be found at: https://youtu.be/lS9YvbrhOdo

Installation

conda create -n mtmct python=3.7.7 -y
conda activate mtmct
pip install -r requirements.txt

Install dependencies for FairMOT:

cd trackers/fair
conda install pytorch==1.7.0 torchvision==0.8.0 cudatoolkit=10.2 -c pytorch
pip install cython
pip install -r requirements.txt
cd DCNv2
./make.sh
conda install -c conda-forge ffmpeg

Download data

Go to https://github.com/schuar-iosb/mta-dataset to download the MTA data. Or use other datasets that match the same format.

Configurations

Modify config files under tracker_configs and clustering_configs for customization. Create a work_dirs and see more instructions at FairMOT and wda_tracker.

E.g. in configs/tracker_configs/fair_high_30e set the data -> source -> base_folder to your dataset location.

Tracking

Run single and the multi-camera tracking with one script:

sh start.sh fair_high_30e

Modify config files under tracker_configs and clustering_configs for customization. More instructions can be found at FairMOT and wda_tracker.

Acknowledgement

A large part of the code is borrowed from FairMOT and wda_tracker. The dataset used is MTA

Copyright

Ruizhe Zhang is the author of this repository and the corresponding report, the copyright belongs to Wireless System Research Group (WiSeR), McMaster University.

View on GitHub
GitHub Stars110
CategoryDesign
Updated1mo ago
Forks16

Languages

Python

Security Score

100/100

Audited on Feb 24, 2026

No findings