Gradient Harmonized Single-stage Detector

13 Nov 2018  ·  Buyu Li, Yu Liu, Xiaogang Wang ·

Despite the great success of two-stage detectors, single-stage detector is still a more elegant and efficient way, yet suffers from the two well-known disharmonies during training, i.e. the huge difference in quantity between positive and negative examples as well as between easy and hard examples. In this work, we first point out that the essential effect of the two disharmonies can be summarized in term of the gradient. Further, we propose a novel gradient harmonizing mechanism (GHM) to be a hedging for the disharmonies. The philosophy behind GHM can be easily embedded into both classification loss function like cross-entropy (CE) and regression loss function like smooth-$L_1$ ($SL_1$) loss. To this end, two novel loss functions called GHM-C and GHM-R are designed to balancing the gradient flow for anchor classification and bounding box refinement, respectively. Ablation study on MS COCO demonstrates that without laborious hyper-parameter tuning, both GHM-C and GHM-R can bring substantial improvement for single-stage detector. Without any whistles and bells, our model achieves 41.6 mAP on COCO test-dev set which surpasses the state-of-the-art method, Focal Loss (FL) + $SL_1$, by 0.8.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Object Detection COCO minival GHM-C + GHM-R (RetinaNet-FPN-ResNet-50, M=30) box AP 35.8 # 193
AP50 55.5 # 97
AP75 38.1 # 92
APS 19.6 # 76
APM 39.6 # 76
APL 46.7 # 83
Object Detection COCO test-dev GHM-C + GHM-R (RetinaNet-FPN-ResNeXt-101) box mAP 41.6 # 176
AP50 62.8 # 106
AP75 44.2 # 129
APS 22.3 # 119
APM 45.1 # 108
APL 55.3 # 98
Hardware Burden None # 1
Operations per network pass None # 1

Methods