MRI
Data repo for mRI: Multi-modal 3D Human Pose Estimation Dataset using mmWave, RGB-D, and Inertial Sensors
Install / Use
/learn @SizheAn/MRIREADME
mRI:
Data repo for mRI: Multi-modal 3D Human Pose Estimation Dataset using mmWave, RGB-D, and Inertial Sensors
Demo in Project page: https://sizhean.github.io/mri
Dataset download link in google drive
New dataset download link with RGB videos included and a new readme file
Please note that we need to process the part of the data (camera related modalities) due to privacy-preserving protocol, which might delay the data release. The dataset (including camera realted modalities) will be fully open-sourced soon.
After unzip the dataset_release.zip, the folder structure should be like this:
${ROOT}
|-- raw_data
| |-- imu
| |-- eaf_file
| |-- radar
| |-- unixtime
| |-- videolabels
|-- aligned_data
| |-- imu
| |-- radar
| |-- pose_labels
|-- features
| |-- imu
| |-- radar
|-- model
| |-- imu
| | |-- results
| | |-- *.pkl
| |-- mmWave
| | |-- results
| | |-- *.pkl
Load cpl file The .cpl file is essentially pickle file, to read them, use:
import pickle
file = pickle.load(open('.../file_path/XXX.cpl', 'rb'))
raw_data folder contains all raw_data before synchronization. It includes imu raw data, radar raw data, eaf annotations, unix timestamp from camera, and videolabels generated from the eaf file.
aligned_data folder contains all data after temporal alignment. It includes imu data, radar data, and the pose_labels. pose_labels for each subject contain following information:
'2d_l_avail_frames': available frames for 2d human detection, left camera
'2d_r_avail_frames': available frames for 2d human detection, right camera
'camera_matrix': camera parameters
'gt_avail_frames': available frames for 3d human joints ground truth
'imu_avail_frames': available frames for imu-estimated keypoints
'imu_est_kps': imu-estimated keypoints
'naive_gt_kps': naive triangulation keypoints
'pose_2d_l': human 2d keypoints from left camera
'pose_2d_r': human 2d keypoints from right camera
'radar_avail_frames': available frames for radar-estimated keypoints
'radar_est_kps': radar-estimated keypoints
'refined_gt_kps': refined triangulation keypoints ground truth
'rgb_avail_frames': available frames for rgb-estimated keypoints
'rgb_est_kps': rgb-estimated keypoints
'video_label': video action labels
feature folder contains imu, radar features for deep learning models. The features are generated from the synced data.
Dimension of the radar feature is (frames, 14, 14, 5). The final 5 means x, y, z-axis coordinates, Doppler velocity, and intensity.
Dimension of the radar feature is (frames, 6, 12). 6 is the number of IMUs and 12 is flattened 3x3 rotation and 3 accelerations.
model folder contains the pretrained model .pkl files and and results.
Related Skills
node-connect
348.5kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
109.1kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
348.5kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
348.5kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
