This repository contains the official PyTorch implementation of paper "Equivariance and Invariance Inductive Bias for Learning from Insufficient Data".
Equivariance and Invariance Inductive Bias for Learning from Insufficient Data
Tan Wang, Qianru Sun, Sugiri Pranata, Karlekar Jayashree, Hanwang Zhang
European Conference on Computer Vision (ECCV), 2022
[Paper] [Poster] [Slides]
From this project, you can:
- Try our algorithm for data efficient learning tasks, for example, VIPriors Challenge.
- Use the dataset or generate your own data with our script for evaluation.
- Improve our equivariance and/or invariance idea and apply in your own project.
If you find our codes helpful, please cite our paper:
@inproceedings{wang2022equivariance,
title={Equivariance and invariance inductive bias for learning from insufficient data},
author={Wang, Tan and Sun, Qianru and Pranata, Sugiri and Jayashree, Karlekar and Zhang, Hanwang},
booktitle={European Conference on Computer Vision},
year={2022}
}
- Python 3.7
- PyTorch 1.9.0
- tqdm
- randaugment
- opencv-python
Please download dataset in this link and put it into the data
folder.
Ps:
- We also provide the dataset generation script here and you can generate any efficient learning dataset that you want.
- Besides the
train
andval
set, we also provide thetestgt
set for test accuracy evaluation. This can be achieved since VIPriors Challenge use the part of ImageNet Val set for testing.
pretrain_path
: path of SSL pretrained modelsstage1_model
: the model type of the SSL training stage.num_shot
: number of samples for each class in the dataset.class_num
: number of classes in the dataset.activate_type
: the activation type added on the mask.inv_start
: when to add the invariance regularization.inv_weight
: the weight of the invariance regularization.opt_mask
: if optimize the mask.
CUDA_VISIBLE_DEVICES=0,1,2,3 python baseline.py -b 256 --name vipriors10_rn50 -j 8 --lr 0.1 data/imagenet_10
You can also try the built-in augmentation algorithms, such as Mixup
CUDA_VISIBLE_DEVICES=0,1,2,3 python baseline.py -b 256 --name vipriors10_rn50_mixup -j 8 --lr 0.1 data/imagenet_10 --mixup
CUDA_VISIBLE_DEVICES=4,5,6,7 python baseline_eq_ipirm.py -b 256 --name vipriors10_rn50_lr0.1_ipirm -j 8 --lr 0.1 data/imagenet_10 --pretrain_path phase1_ssl_methods/run_imagenet10/ipirm_imagenet10/model_ipirm.pth
For the SSL pretraining process, please follow the chapter below.
Step-1: SSL Pretraining (Equivariance Learning)
Please follow the original codebase. We list the code we used below:
Please put the pretrained models in phase1_ssl_methods
. You can also choose to directly use our SSL pretrained models (IP-IRM) here
Step-2/3: Downstream Fine-tuning (Invariance Learning)
Running Commands
CUDA_VISIBLE_DEVICES=4,5,6,7 python vipriors_eqinv.py -b 128 --name vipriors10_ipirm_mask_sigmoid_rex100._start10 -j 24 data/imagenet_10 --pretrain_path phase1_ssl_methods/run_imagenet10/ipirm_imagenet10/model_ipirm.pth --inv rex --inv_weight 100. --opt_mask --activat_type sigmoid --inv_start 10 --mlp --stage1_model ipirm --num_shot 10
You can also adopt Random Augmentation
to achieve better results:
CUDA_VISIBLE_DEVICES=4,5,6,7 python vipriors_eqinv.py -b 128 --name vipriors10_ipirm_mask_sigmoid_rex10._start10_randaug -j 24 data/imagenet_10 --pretrain_path phase1_ssl_methods/run_imagenet10/ipirm_imagenet10/model_ipirm.pth --inv rex --inv_weight 10. --opt_mask --activat_type sigmoid --inv_start 10 --mlp --stage1_model ipirm --num_shot 10 --random_aug
For other dataset, you can try:
CUDA_VISIBLE_DEVICES=0,1,2,3 python vipriors_eqinv.py -b 128 --name vipriors20_ipirm_mask_sigmoid_rex10._start10_randaug -j 24 data/imagenet_20 --pretrain_path phase1_ssl_methods/run_imagenet20/ipirm_imagenet20/model_ipirm.pth --inv rex --inv_weight 10. --opt_mask --activat_type sigmoid --inv_start 10 --mlp --stage1_model ipirm --num_shot 20 --random_aug
CUDA_VISIBLE_DEVICES=0,1,2,3 python vipriors_eqinv.py -b 128 --name vipriors50_ipirm_mask_sigmoid_rex10._start10_randaug -j 24 data/imagenet_50 --pretrain_path phase1_ssl_methods/run_imagenet50/ipirm_imagenet50/model_ipirm.pth --inv rex --inv_weight 10. --opt_mask --activat_type sigmoid --inv_start 10 --mlp --stage1_model ipirm --num_shot 50 --random_aug
If you have any questions, please feel free to email me (TAN317@ntu.edu.sg).