Snca.pytorch
Improving Generalization via Scalable Neighborhood Component Analysis
Install / Use
/learn @microsoft/Snca.pytorchREADME
Improving Generalization via Scalable Neighborhood Component Analysis
This repo constains the pytorch implementation for the ECCV 2018 paper (paper). We use deep networks to learn feature representations optimized for nearest neighbor classifiers, which could generalize better for new object categories. This project is a re-investigation of Neighborhood Component Analysis (NCA) with recent technologies to make it scalable to deep networks and large-scale datasets.
Much of code is extended from the previous unsupervised learning project. Please refer to this repo for more details.
<img src="http://zhirongw.westus2.cloudapp.azure.com/figs/snca.png" width="800px"/>Pretrained Models
Currently, we provide three pretrained ResNet models. Each release contains the feature representation of all ImageNet training images (600 mb) and model weights (100-200mb). Models and their performance with nearest neighbor classifiers are as follows.
- ResNet 18 (top 1 accuracy 70.59%)
- ResNet 34 (top 1 accuracy 74.41%)
- ResNet 50 (top 1 accuracy 76.57%)
Code to reproduce the rest of the experiments are comming soon.
Nearest Neighbors
Please follow this link for a list of nearest neighbors on ImageNet. Results are visualized from our ResNet50 feature, compared with baseline ResNet50 feature, raw image features and previous unsupervised features. First column is the query image, followed by 20 retrievals ranked by the similarity.
<img src="http://zhirongw.westus2.cloudapp.azure.com/figs/nn.png" width="800px"/>Usage
Our code extends the pytorch implementation of imagenet classification in official pytorch release. Please refer to the official repo for details of data preparation and hardware configurations.
-
install python2 and pytorch>=0.4
-
clone this repo:
git clone https://github.com/Microsoft/snca.pytorch -
Training on ImageNet:
python main.py DATAPATH --arch resnet18 -j 32 --temperature 0.05 --low-dim 128 -b 256- During training, we monitor the supervised validation accuracy by K nearest neighbor with k=1, as it's faster, and gives a good estimation of the feature quality.
-
Testing on ImageNet:
python main.py DATAPATH --arch resnet18 --resume input_model.pth.tar -eruns testing with default K=30 neighbors. -
Memory Consumption and Computation Issues
Memory consumption is more of an issue than computation time. Currently, the implementation of nca module is not paralleled across multiple GPUs. Hence, the first GPU will consume much more memory than the others. For example, when training a ResNet18 network, GPU 0 will consume 11GB memory, while the others each takes 2.5GB. You will need to set the Caffe style "-b 128 --iter-size 2" for training deeper networks. Our released models are trained with V100 machines.
-
Training on CIFAR10:
python cifar.py --temperature 0.05 --lr 0.1
Citation
@inproceedings{wu2018improving,
title={Improving Generalization via Scalable Neighborhood Component Analysis},
author={Wu, Zhirong and Efros, Alexei A and Yu, Stella},
booktitle={European Conference on Computer Vision (ECCV) 2018},
year={2018}
}
Contact
For any questions, please feel free to reach
Zhirong Wu: xavibrowu@gmail.com
Contributing
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.microsoft.com.
When you submit a pull request, a CLA-bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., label, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
research_rules
Research & Verification Rules Quote Verification Protocol Primary Task "Make sure that the quote is relevant to the chapter and so you we want to make sure that we want to have it identifie
groundhog
398Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
