Bridging the Gap Between Anchor-based and Anchor-free Detection via Adaptive Training Sample Selection
By Shifeng Zhang, Cheng Chi, Yongqiang Yao, Zhen Lei, Stan Z. Li.
In this work, we first point out that the essential difference between anchor-based and anchor-free detection is actually how to define positive and negative training samples. Then we propose an Adaptive Training Sample Selection (ATSS) to automatically select positive and negative samples according to statistical characteristics of object, which significantly improves the performance of anchor-based and anchor-free detectors and bridges the gap between them. Finally, we demonstrate that tiling multiple anchors per location on the image to detect objects is a thankless operation under current situations. Extensive experiments conducted on MS COCO support our aforementioned analysis and conclusions. With the newly introduced ATSS, we improve state-of-the-art detectors by a large margin to 50.7% AP without introducing any overhead. For more details, please refer to our paper.
Note: The lite version of our ATSS has been merged to the official code of FCOS as the center sampling improvement, which improves its performance by ~0.8%. The full version of our ATSS can further improve the performance.
This ATSS implementation is based on FCOS and maskrcnn-benchmark and the installation is the same as them. Please check INSTALL.md for installation instructions.
Once the installation is done, you can download ATSS_R_50_FPN_1x.pth from Google or Baidu to run a quick demo.
# assume that you are under the root directory of this project,
# and you have activated your virtual environment if needed.
python demo/atss_demo.py
The inference command line on coco minival split:
python tools/test_net.py \
--config-file configs/atss/atss_R_50_FPN_1x.yaml \
MODEL.WEIGHT ATSS_R_50_FPN_1x.pth \
TEST.IMS_PER_BATCH 4
Please note that:
- If your model's name is different, please replace
ATSS_R_50_FPN_1x.pth
with your own. - If you enounter out-of-memory error, please try to reduce
TEST.IMS_PER_BATCH
to 1. - If you want to evaluate a different model, please change
--config-file
to its config file (in configs/atss) andMODEL.WEIGHT
to its weights file.
For your convenience, we provide the following trained models. All models are trained with 16 images in a mini-batch and frozen batch normalization (i.e., consistent with models in FCOS and maskrcnn_benchmark).*
Model | Multi-scale training | Testing time / im | AP (minival) | AP (test-dev) | Link |
---|---|---|---|---|---|
ATSS_R_50_FPN_1x | No | 44ms | 39.3 | 39.3 | Google/Baidu |
ATSS_dcnv2_R_50_FPN_1x | No | 54ms | 43.2 | 43.0 | Google/Baidu |
ATSS_R_101_FPN_2x | Yes | 57ms | 43.5 | 43.6 | Google/Baidu |
ATSS_dcnv2_R_101_FPN_2x | Yes | 73ms | 46.1 | 46.3 | Google/Baidu |
ATSS_X_101_32x8d_FPN_2x | Yes | 110ms | 44.8 | 45.1 | Google/Baidu |
ATSS_dcnv2_X_101_32x8d_FPN_2x | Yes | 143ms | 47.7 | 47.7 | Google/Baidu |
ATSS_X_101_64x4d_FPN_2x | Yes | 112ms | 45.5 | 45.6 | Google/Baidu |
ATSS_dcnv2_X_101_64x4d_FPN_2x | Yes | 144ms | 47.7 | 47.7 | Google/Baidu |
[1] The testing time is taken from FCOS, because our method only redefines positive and negative training samples without incurring any additional overhead.
[2] 1x and 2x mean the model is trained for 90K and 180K iterations, respectively.
[3] All results are obtained with a single model and without any test time data augmentation such as multi-scale, flipping and etc..
[4] dcnv2
denotes deformable convolutional networks v2. Note that for ResNet based models, we apply deformable convolutions from stage c3 to c5 in backbones. For ResNeXt based models, only stage c4 and c5 use deformable convolutions. All models use deformable convolutions in the last layer of detector towers.
[5] The model ATSS_dcnv2_X_101_64x4d_FPN_2x
with multi-scale testing achieves 50.7% in AP on COCO test-dev. Please use TEST.BBOX_AUG.ENABLED True
to enable multi-scale testing.
The following command line will train ATSS_R_50_FPN_1x on 8 GPUs with Synchronous Stochastic Gradient Descent (SGD):
python -m torch.distributed.launch \
--nproc_per_node=8 \
--master_port=$((RANDOM + 10000)) \
tools/train_net.py \
--config-file configs/atss/atss_R_50_FPN_1x.yaml \
DATALOADER.NUM_WORKERS 2 \
OUTPUT_DIR training_dir/atss_R_50_FPN_1x
Please note that:
- If you want to use fewer GPUs, please change
--nproc_per_node
to the number of GPUs. No other settings need to be changed. The total batch size does not depends onnproc_per_node
. If you want to change the total batch size, please changeSOLVER.IMS_PER_BATCH
in configs/atss/atss_R_50_FPN_1x.yaml. - The models will be saved into
OUTPUT_DIR
. - If you want to train ATSS with other backbones, please change
--config-file
.
Any pull requests or issues are welcome.
Please cite our paper in your publications if it helps your research:
@article{zhang2019bridging,
title = {Bridging the Gap Between Anchor-based and Anchor-free Detection via Adaptive Training Sample Selection},
author = {Zhang, Shifeng and Chi, Cheng and Yao, Yongqiang and Lei, Zhen and Li, Stan Z.},
journal = {arXiv preprint arXiv:1912.02424},
year = {2019}
}