Weight Modulation is an alternative to adaptive instance normalization for use in generative adversarial networks, specifically it is introduced in StyleGAN2. The purpose of instance normalization is to remove the effect of $s$ - the scales of the features maps - from the statistics of the convolution’s output feature maps. Weight modulation tries to achieve this goal more directly. Assuming that input activations are i.i.d. random variables with unit standard deviation. After modulation and convolution, the output activations have standard deviation of:
$$ \sigma_{j} = \sqrt{{\sum_{i,k}w_{ijk}'}^{2}} $$
i.e., the outputs are scaled by the $L_{2}$ norm of the corresponding weights. The subsequent normalization aims to restore the outputs back to unit standard deviation. This can be achieved if we scale (“demodulate”) each output feature map $j$ by $1/\sigma_{j}$ . Alternatively, we can again bake this into the convolution weights:
$$ w''_{ijk} = w'_{ijk} / \sqrt{{\sum_{i, k}w'_{ijk}}^{2} + \epsilon} $$
where $\epsilon$ is a small constant to avoid numerical issues.
Source: Analyzing and Improving the Image Quality of StyleGANPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Image Generation | 53 | 17.79% |
Face Generation | 13 | 4.36% |
Disentanglement | 13 | 4.36% |
Image Manipulation | 13 | 4.36% |
Diversity | 11 | 3.69% |
Face Recognition | 10 | 3.36% |
Translation | 7 | 2.35% |
Conditional Image Generation | 7 | 2.35% |
Face Swapping | 6 | 2.01% |