Training Techniques | Weight Decay, SGD with Momentum |
---|---|
Architecture | 1x1 Convolution, Wide Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Connection, ReLU, Max Pooling, Softmax |
ID | wide_resnet101_2 |
SHOW MORE |
Training Techniques | Weight Decay, SGD with Momentum |
---|---|
Architecture | 1x1 Convolution, Wide Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Connection, ReLU, Max Pooling, Softmax |
ID | wide_resnet50_2 |
SHOW MORE |
Wide Residual Networks are a variant on ResNets where we decrease depth and increase the width of residual networks. This is achieved through the use of wide residual blocks.
To load a pretrained model:
import torchvision.models as models
wide_resnet50_2 = models.wide_resnet50_2(pretrained=True)
Replace the model name with the variant you want to use, e.g. wide_resnet50_2
. You can find
the IDs in the model summaries at the top of this page.
To evaluate the model, use the image classification recipes from the library.
python train.py --test-only --model='<model_name>'
You can follow the torchvision recipe on GitHub for training a new model afresh.
@article{DBLP:journals/corr/ZagoruykoK16,
author = {Sergey Zagoruyko and
Nikos Komodakis},
title = {Wide Residual Networks},
journal = {CoRR},
volume = {abs/1605.07146},
year = {2016},
url = {http://arxiv.org/abs/1605.07146},
archivePrefix = {arXiv},
eprint = {1605.07146},
timestamp = {Mon, 13 Aug 2018 16:46:42 +0200},
biburl = {https://dblp.org/rec/journals/corr/ZagoruykoK16.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
BENCHMARK | MODEL | METRIC NAME | METRIC VALUE | GLOBAL RANK |
---|---|---|---|---|
ImageNet | Wide ResNet-101-2 | Top 1 Accuracy | 78.84% | # 141 |
Top 5 Accuracy | 94.28% | # 141 | ||
ImageNet | Wide ResNet-50-2 | Top 1 Accuracy | 78.51% | # 151 |
Top 5 Accuracy | 94.09% | # 151 |