SkillAgentSearch skills...

COESOT

[Pattern Recognition 2025] A large-scale benchmark dataset for color-event based visual tracking

Install / Use

/learn @Event-AHU/COESOT

README

<div align="center"> <img src="https://github.com/Event-AHU/COESOT/blob/main/figures/COESOT.png" width="600">

A general and large-scale benchmark COESOT dataset for color-event based visual tracking


</div>

Revisiting Color-Event based Tracking: A Unified Network, Dataset, and Metric, Chuanming Tang, Xiao Wang, Ju Huang, Bo Jiang, Lin Zhu, Jianlin Zhang, Yaowei Wang, Yonghong Tian [Project]

Update Log

  • :fire: [2025.11.05] COESOT is accepted by the Journal Pattern Recognition!

  • :fire: [2024.03.12] A New Long-term RGB-Event based Visual Object Tracking Benchmark Dataset (termed FELT) is available at [Paper] [Code] [DemoVideo]

  • :fire: [2024.03.06] Tracking results of CEUTrack on VisEvent dataset is available at [ceutrack_visevent_dataset_tracking_results.zip]

  • :fire: [2023.09.27] A High Definition (HD) Event based Visual Object Tracking Benchmark Dataset (termed EventVOT) is available at [arXiv] [Github]

Demo Video:

Dataset Download:

Baidu Download link:https://pan.baidu.com/s/12XDlKABlz3lDkJJEDvsu9A     Passcode:AHUT 

The directory should have the below format:

├── COESOT dataset
    ├── Training Subset (827 videos, 160GB)
        ├── dvSave-2021_09_01_06_59_10
            ├── dvSave-2021_09_01_06_59_10_aps
            ├── dvSave-2021_09_01_06_59_10_dvs
            ├── dvSave-2021_09_01_06_59_10.aedat4
            ├── groundtruth.txt
            ├── absent.txt
            ├── start_end_index.txt
        ├── ... 
    ├── Testing Subset (528 videos, 105GB)
        ├── dvSave-2021_07_30_11_04_12
            ├── dvSave-2021_07_30_11_04_12_aps
            ├── dvSave-2021_07_30_11_04_12_dvs
            ├── dvSave-2021_07_30_11_04_12.aedat4
            ├── groundtruth.txt
            ├── absent.txt
            ├── start_end_index.txt
        ├── ... 
<p align="center"> <img width="85%" src="https://github.com/Event-AHU/COESOT/blob/main/figures/CODSOT_benchmarkSamples.jpg" alt="Framework"/> </p>

COESOT_eval_toolkit

  1. unzip the COESOT_eval_toolkit.zip, and open it with Matlab (over Matlab R2020).

  2. add your tracking results and baseline results (Passcode:siaw) in $/coesot_tracking_results/ and modify the name in $/utils/config_tracker.m. BTW, here we also provide the event-only baseline tracking methods results in [Event_only Results] Passcode:qblp

  3. run Evaluate_COESOT_benchmark_SP_PR_only.m for the overall performance evaluation, including SR, PR, NPR.

<p align="left"> <img width="100%" src="./figures/SRPRNPR.png" alt="SR_PR_NPR"/> </p>
  1. run plot_BOC.m for BOC score evaluation and figure plot.
  2. run plot_radar.m for attributes radar figrue plot.
<p align="center"> <img width="43%" src="./figures/radar1.png" alt="Radar"/><img width="57%" src="./figures/BOC_score.png" alt="Radar"/> </p>
  1. run Evaluate_COESOT_benchmark_attributes.m for attributes analysis and figure saved in $/res_fig/.

CEUTrack

A unified framework for color-event tracking.

[Models] Passcode:0uk0 [Raw Results] Passcode:yeow [Training logs] Passcode:hnim

<p align="center"> <img width="85%" src="./figures/frameworkV2.jpg" alt="Framework"/> </p>

Install env

conda create -n event python=3.7
conda activate event
bash install.sh

Run the following command to set paths for this project

python tracking/create_default_local_file.py --workspace_dir . --data_dir ./data --save_dir ./output

After running this command, you can also modify paths by editing these two files

lib/train/admin/local.py  # paths about training
lib/test/evaluation/local.py  # paths about testing

Then, put the tracking datasets COESOT in ./data.

Download pre-trained MAE ViT-Base weights and put it under $/pretrained_models

Download the model weights and put it on $/output/checkpoints/train/ceutrack

  • [Note] More useful scripts can be found from:
https://github.com/Event-AHU/COESOT/tree/main/CEUTrack/scripts

Train & Test & Evaluation

    # train
    export CUDA_VISIBLE_DEVICES=0
    python tracking/train.py --script ceutrack --config ceutrack_coesot  \
    --save_dir ./output --mode multiple --nproc_per_node 1 --use_wandb  0
    # test
    python tracking/test.py   ceutrack ceutrack_coesot --dataset coesot --threads 4 --num_gpus 1
    # eval
    python tracking/analysis_results.py --dataset coesot  --parameter_name ceutrack_coesot

Test FLOPs, and Speed

Note: The speeds reported in our paper were tested on a single RTX 3090 GPU.

# Profiling ceutrack_coesot
python tracking/profile_model.py --script ceutrack --config ceutrack_coesot

Activation Visualization

Use the script from: [show_CAM.py]

from .show_CAM import getCAM
getCAM(response, curr_image, self.idx)
<p align="center"> <img width="85%" src="./figures/responseMAPs.png" alt="responseMAPs"/> </p>

TODO List

  • [x] Paper (arXiv) release
  • [x] COESOT dataset release
  • [x] Evaluation Toolkit release
  • [x] Source Code release
  • [x] Tracking Models release

Acknowledgments

Citation:

@article{tang2022coesot,
  title={Revisiting Color-Event based Tracking: A Unified Network, Dataset, and Metric},
  author={Tang, Chuanming and Wang, Xiao and Huang, Ju and Jiang, Bo and Zhu, Lin and Zhang, Jianlin and Wang, Yaowei and Tian, Yonghong},
  journal={arXiv preprint arXiv:2211.11010},
  year={2022}
}

Star History

<a href="https://star-history.com/#Event-AHU/COESOT&Date"> <picture> <source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=Event-AHU/COESOT&type=Date&theme=dark" /> <source media="(prefers-color-scheme: light)" srcset="https://api.star-history.com/svg?repos=Event-AHU/COESOT&type=Date" /> <img alt="Star History Chart" src="https://api.star-history.com/svg?repos=Event-AHU/COESOT&type=Date" /> </picture> </a>
View on GitHub
GitHub Stars76
CategoryDevelopment
Updated5d ago
Forks4

Languages

Python

Security Score

85/100

Audited on Mar 30, 2026

No findings