PIRL

Last updated on Feb 27, 2021

PIRL ResNet-50 (200 epochs, linear head)

Parameters 26 Million
FLOPs 4 Billion
File Size 825.09 MB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head linear
Epochs 200
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 1
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
PIRL ResNet-50 (200 epochs, MLP head)

Parameters 26 Million
FLOPs 4 Billion
File Size 889.13 MB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture Dense Connections, 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head MLP
Epochs 200
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 1
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
PIRL ResNet-50 (800 epochs, linear head)

Parameters 26 Million
FLOPs 4 Billion
File Size 825.09 MB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head linear
Epochs 800
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 1
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
PIRL ResNet-50 (800 epochs, MLP head)

Parameters 26 Million
FLOPs 4 Billion
File Size 889.13 MB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture Dense Connections, 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head MLP
Epochs 800
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 1
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
PIRL ResNet-50-w2 (400 epochs, linear head)

Parameters 94 Million
Layers 50
File Size 1.37 GB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head Linear
Epochs 400
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 2
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
PIRL ResNet-50-w2 (400 epochs, MLP head)

Parameters 94 Million
Layers 50
File Size 1.62 GB
Training Data ImageNet
Training Resources 8x NVIDIA V100 GPUs
Training Time

Training Techniques PIRL, Weight Decay, SGD with Momentum, Cosine Annealing
Architecture Dense Connections, 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
Head MLP
Epochs 400
Layers 50
Classes 1000
Momentum 0.9
Negatives 32000
Weight Decay 0.0001
Width Multiplier 2
NCE Loss Temperature 0.07
SHOW MORE
SHOW LESS
README.md

Summary

Pretext-Invariant Representation Learning (PIRL, pronounced as “pearl”) learns invariant representations based on pretext tasks. PIRL is used with a commonly used pretext task that involves solving jigsaw puzzles. Specifically, PIRL constructs image representations that are similar to the representation of transformed versions of the same image and different from the representations of other images.

How do I train this model?

Get started with VISSL by trying one of the Colab tutorial notebooks.

Citation

@inproceedings{misra2020pirl,
  title={Self-Supervised Learning of Pretext-Invariant Representations},
  author={Misra, Ishan and van der Maaten, Laurens},
  booktitle={CVPR},
  year={2020}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY
PIRL ResNet-50-w2 (400 epochs, MLP head) 70.9%
PIRL ResNet-50 (800 epochs, MLP head) 69.9%
PIRL ResNet-50-w2 (400 epochs, linear head) 69.3%
PIRL ResNet-50 (200 epochs, MLP head) 65.8%
PIRL ResNet-50 (800 epochs, linear head) 64.29%
PIRL ResNet-50 (200 epochs, linear head) 62.9%