SkillAgentSearch skills...

Wear

WEAR: An Outdoor Sports Dataset for Wearable and Egocentric Activity Recognition, Bock et al., published in IMWUT (Issue 4, Vol. 8, Article 175 November 2024)

Install / Use

/learn @mariusbock/Wear

README

WEAR: An Outdoor Sports Dataset for Wearable and Egocentric Activity Recognition

<img loop src="teaser.gif" width="100%"/>

arXiv CC BY-NC-SA 4.0 GitHub forks GitHub forks

Abstract

Research has shown the complementarity of camera- and inertial-based data for modeling human activities, yet datasets with both egocentric video and inertial-based sensor data remain scarce. In this paper, we introduce WEAR, an outdoor sports dataset for both vision- and inertial-based human activity recognition (HAR). Data from 22 participants performing a total of 18 different workout activities was collected with synchronized inertial (acceleration) and camera (egocentric video) data recorded at 11 different outside locations. WEAR provides a challenging prediction scenario in changing outdoor environments using a sensor placement, in line with recent trends in real-world applications. Benchmark results show that through our sensor placement, each modality interestingly offers complementary strengths and weaknesses in their prediction performance. Further, in light of the recent success of single-stage Temporal Action Localization (TAL) models, we demonstrate their versatility of not only being trained using visual data, but also using raw inertial data and being capable to fuse both modalities by means of simple concatenation. The dataset and code to reproduce experiments is publicly available via: mariusbock.github.io/wear/. An arXiv version of our paper is available here.

Changelog

  • 14/10/2024: WEAR accepted at IMWUT; updated code base, main paper and supplementary material
  • 14/06/2023: updated code base and arXiv available.
  • 18/04/2023: provided code to reproduce experiments.
  • 12/04/2023: initial commit and arXiv uploaded.

Supplementary Material

Additional results and figures can be found in the supplementary_material.pdf.

Installation

Please follow instructions mentioned in the INSTALL.md file.

Download

The full dataset can be downloaded here

The download folder is divided into 3 subdirectories

  • annotations (14MB): JSON-files containing annotations per-subject using the THUMOS14-style
  • processed (44GB): precomputed I3D, inertial and combined per-subject features
  • raw (164GB): Raw, per-subject video and inertial data

Reproduce Experiments

Once having installed requirements, one can rerun experiments by running the main.py script:

python main.py --config ./configs/60_frames_30_stride/actionformer_combined.yaml --seed 1 --eval_type split

Each config file represents one type of experiment. Each experiment was run three times using three different random seeds (i.e. 1, 2, 3). To rerun the experiments without changing anything about the config files, please place the complete dataset download into a folder called data/wear in the main directory of the repository.

Postprocessing

Please follow instructions mentioned in the README.md file in the postprocessing subfolder.

Logging using Neptune.ai

In order to log experiments to Neptune.ai please provide projectand api_token information in your local deployment (see lines 34-35 in main.py)

Record your own Data

Please follow instructions mentioned in the README.md file in the data creation subfolder.

License

WEAR is offered under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License. You are free to use, copy, and redistribute the material for non-commercial purposes provided you give appropriate credit, provide a link to the license, and indicate if changes were made. If you remix, transform, or build upon the material, you must distribute your contributions under the same license as the original. You may not use the material for commercial purposes.

Contact

Marius Bock (marius.bock@uni-siegen.de)

Cite as

@article{bock2024wear,
    author={Bock, Marius and Kuehne, Hilde and Van Laerhoven, Kristof and Moeller, Michael},
    title = {WEAR: An Outdoor Sports Dataset for Wearable and Egocentric Activity Recognition},
    year = {2024},
    volume = {8},
    number = {4},
    journal = {Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. (IMWUT)},
    numpages = {21},
    articleno = {175},
    doi = {10.1145/3699776},
    url={https://dl.acm.org/doi/10.1145/3699776}
}
View on GitHub
GitHub Stars24
CategoryDevelopment
Updated19d ago
Forks3

Languages

Python

Security Score

80/100

Audited on Mar 17, 2026

No findings