Normalization

Local Response Normalization

Introduced by Krizhevsky et al. in ImageNet Classification with Deep Convolutional Neural Networks

Local Response Normalization is a normalization layer that implements the idea of lateral inhibition. Lateral inhibition is a concept in neurobiology that refers to the phenomenon of an excited neuron inhibiting its neighbours: this leads to a peak in the form of a local maximum, creating contrast in that area and increasing sensory perception. In practice, we can either normalize within the same channel or normalize across channels when we apply LRN to convolutional neural networks.

$$ b_{c} = a_{c}\left(k + \frac{\alpha}{n}\sum_{c'=\max(0, c-n/2)}^{\min(N-1,c+n/2)}a_{c'}^2\right)^{-\beta} $$

Where the size is the number of neighbouring channels used for normalization, $\alpha$ is multiplicative factor, $\beta$ an exponent and $k$ an additive factor

Source: ImageNet Classification with Deep Convolutional Neural Networks

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
General Classification 68 12.43%
Image Classification 52 9.51%
Quantization 45 8.23%
Classification 37 6.76%
Object Detection 34 6.22%
Object Recognition 24 4.39%
Model Compression 13 2.38%
Clustering 10 1.83%
Network Pruning 9 1.65%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories