DirectionNet
Wide-Baseline Relative Camera Pose Estimation with Directional Learning (CVPR 2021)
Install / Use
/learn @arthurchen0518/DirectionNetREADME
DirectionNet
This repository will contain the TensorFlow code for the model introduced in the CVPR 2020 paper:
Wide-Baseline Relative Camera Pose Estimation with Directional Learning
Kefan Chen, Noah Snavely, Ameesh Makadia
IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2020).
arXiv
Requirements
Required packages: tensorflow 1.15, tensorflow_graphics, tensorflow_addons, tensorflow_probability, tf_slim, pickle
Dataset and Pre-trained Models
MatterportA test data
MatterportB test data
-
dataset.generate_from_meta can create the datasets of images and ground truth from the Matterport3D dataset given the meta data files.
-
dataset.generate_random_views can be used to generate a large-scale wide stereo dataset with camera pose labels from a panoramic image dataset.
Train DirectionNet-R and DirectionNet-T
- train DirectionNet-R.
python train.py \
--checkpoint_dir <path_to_checkpoints_and_logs> \
--data_dir <path_to_training_set> \
--model 9D
- Run DirectionNet-R on the training and test sets, then save the estimated rotations as Python pickle dictionary in the data directories respectively.
- train DirectionNet-T.
python train.py \
--checkpoint_dir <path_to_checkpoints_and_logs> \
--data_dir <path_to_training_set> \
--model T
Evaluation
DirectionNet-R
python eval.py \
--checkpoint_dir <path_to_load_checkpoints> \
--eval_data_dir <path_to_test_set> \
--save_summary_dir <path_to_save_logs> \
--testset_size <testset_size> \
--batch <test_batch> \
--model 9D
DirectionNet-T
python eval.py \
--checkpoint_dir <path_to_load_checkpoints> \
--eval_data_dir <path_to_test_set> \
--save_summary_dir <path_to_save_logs> \
--testset_size <testset_size> \
--batch <test_batch> \
--model T
Citation
@InProceedings{Chen_2021_CVPR,
author = {Chen, Kefan and Snavely, Noah and Makadia, Ameesh},
title = {Wide-Baseline Relative Camera Pose Estimation With Directional Learning},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2021},
pages = {3258-3268}
}
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
research_rules
Research & Verification Rules Quote Verification Protocol Primary Task "Make sure that the quote is relevant to the chapter and so you we want to make sure that we want to have it identifie
groundhog
398Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
