RepPoints

Last updated on Feb 23, 2021

RepPoints (R-101-FPN, 2x, GN=Y, Anchor=none)

Memory (M) 5800.0
inference time (s/im) 0.07299
File Size 213.53 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture Group Normalization, RepPoints, ResNet, FPN
lr sched 2x
Memory (M) 5800.0
Backbone Layers 101
inference time (s/im) 0.07299
SHOW MORE
SHOW LESS
RepPoints (R-101-FPN-DCN, 2x, GN=Y, Anchor=none)

Memory (M) 5900.0
inference time (s/im) 0.08264
File Size 218.44 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture Group Normalization, Deformable Convolution, RepPoints, FPN, ResNet
lr sched 2x
Memory (M) 5900.0
Backbone Layers 101
inference time (s/im) 0.08264
SHOW MORE
SHOW LESS
RepPoints (R-50-FPN, 1x, GN=N, Anchor=none)

Memory (M) 3300.0
inference time (s/im) 0.05405
File Size 140.82 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture RepPoints, ResNet, FPN
lr sched 1x
Memory (M) 3300.0
Backbone Layers 50
inference time (s/im) 0.05405
SHOW MORE
SHOW LESS
RepPoints (R-50-FPN, 1x, GN=Y, Anchor=none)

Memory (M) 3900.0
inference time (s/im) 0.05714
File Size 140.83 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture Group Normalization, RepPoints, ResNet, FPN
lr sched 1x
Memory (M) 3900.0
Backbone Layers 50
inference time (s/im) 0.05714
SHOW MORE
SHOW LESS
RepPoints (R-50-FPN, 2x, GN=Y, Anchor=none)

Memory (M) 3900.0
Backbone Layers 50
File Size 140.83 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture Group Normalization, RepPoints, ResNet, FPN
lr sched 2x
Memory (M) 3900.0
Backbone Layers 50
SHOW MORE
SHOW LESS
RepPoints (X-101-FPN-DCN, 2x, GN=Y, Anchor=none)

Memory (M) 7100.0
inference time (s/im) 0.10753
File Size 222.06 MB
Training Data MS COCO
Training Resources 8x NVIDIA V100 GPUs
Training Time

Architecture ResNeXt, Group Normalization, Deformable Convolution, RepPoints, FPN
lr sched 2x
Memory (M) 7100.0
Backbone Layers 101
inference time (s/im) 0.10753
SHOW MORE
SHOW LESS
README.md

RepPoints: Point Set Representation for Object Detection

By Ze Yang, Shaohui Liu, and Han Hu.

We provide code support and configuration files to reproduce the results in the paper for "RepPoints: Point Set Representation for Object Detection" on COCO object detection.

Introduction

[ALGORITHM]

RepPoints, initially described in arXiv, is a new representation method for visual objects, on which visual understanding tasks are typically centered. Visual object representation, aiming at both geometric description and appearance feature extraction, is conventionally achieved by bounding box + RoIPool (RoIAlign). The bounding box representation is convenient to use; however, it provides only a rectangular localization of objects that lacks geometric precision and may consequently degrade feature quality. Our new representation, RepPoints, models objects by a point set instead of a bounding box, which learns to adaptively position themselves over an object in a manner that circumscribes the object’s spatial extent and enables semantically aligned feature extraction. This richer and more flexible representation maintains the convenience of bounding boxes while facilitating various visual understanding applications. This repo demonstrated the effectiveness of RepPoints for COCO object detection.

Another feature of this repo is the demonstration of an anchor-free detector, which can be as effective as state-of-the-art anchor-based detection methods. The anchor-free detector can utilize either bounding box or RepPoints as the basic object representation.

Learning RepPoints in Object Detection.

Citing RepPoints

@inproceedings{yang2019reppoints,
  title={RepPoints: Point Set Representation for Object Detection},
  author={Yang, Ze and Liu, Shaohui and Hu, Han and Wang, Liwei and Lin, Stephen},
  booktitle={The IEEE International Conference on Computer Vision (ICCV)},
  month={Oct},
  year={2019}
}

Results and models

The results on COCO 2017val are shown in the table below.

Method Backbone GN Anchor convert func Lr schd Mem (GB) Inf time (fps) box AP Config Download
BBox R-50-FPN Y single - 1x 3.9 15.9 36.4 config model | log
BBox R-50-FPN Y none - 1x 3.9 15.4 37.4 config model | log
RepPoints R-50-FPN N none moment 1x 3.3 18.5 37.0 config model | log
RepPoints R-50-FPN Y none moment 1x 3.9 17.5 38.1 config model | log
RepPoints R-50-FPN Y none moment 2x 3.9 - 38.6 config model | log
RepPoints R-101-FPN Y none moment 2x 5.8 13.7 40.5 config model | log
RepPoints R-101-FPN-DCN Y none moment 2x 5.9 12.1 42.9 config model | log
RepPoints X-101-FPN-DCN Y none moment 2x 7.1 9.3 44.2 config model | log

Notes:

  • R-xx, X-xx denote the ResNet and ResNeXt architectures, respectively.
  • DCN denotes replacing 3x3 conv with the 3x3 deformable convolution in c3-c5 stages of backbone.
  • none in the anchor column means 2-d center point (x,y) is used to represent the initial object hypothesis. single denotes one 4-d anchor box (x,y,w,h) with IoU based label assign criterion is adopted.
  • moment, partial MinMax, MinMax in the convert func column are three functions to convert a point set to a pseudo box.
  • Note the results here are slightly different from those reported in the paper, due to framework change. While the original paper uses an MXNet implementation, we re-implement the method in PyTorch based on mmdetection.

Results

Object Detection on COCO minival

Object Detection on COCO minival
MODEL BOX AP
RepPoints (X-101-FPN-DCN, 2x, GN=Y, Anchor=none) 44.2
RepPoints (R-101-FPN-DCN, 2x, GN=Y, Anchor=none) 42.9
RepPoints (R-101-FPN, 2x, GN=Y, Anchor=none) 40.5
RepPoints (R-50-FPN, 2x, GN=Y, Anchor=none) 38.6
RepPoints (R-50-FPN, 1x, GN=Y, Anchor=none) 38.1
RepPoints (R-50-FPN, 1x, GN=N, Anchor=none) 37.0