ECAResNet

Last updated on Feb 14, 2021

ecaresnet101d

Parameters 45 Million
FLOPs 10 Billion
File Size 170.53 MB
Training Data ImageNet
Training Resources 4x RTX 2080Ti GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Efficient Channel Attention, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Squeeze-and-Excitation Block
ID ecaresnet101d
LR 0.1
Epochs 100
Layers 101
Crop Pct 0.875
Batch Size 256
Image Size 224
Weight Decay 0.0001
Interpolation bicubic
SHOW MORE
SHOW LESS
ecaresnet101d_pruned

Parameters 25 Million
FLOPs 4 Billion
File Size 95.23 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Efficient Channel Attention, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Squeeze-and-Excitation Block
ID ecaresnet101d_pruned
Layers 101
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
ecaresnet50d

Parameters 26 Million
FLOPs 6 Billion
File Size 97.83 MB
Training Data ImageNet
Training Resources 4x RTX 2080Ti GPUs
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Efficient Channel Attention, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Squeeze-and-Excitation Block
ID ecaresnet50d
LR 0.1
Epochs 100
Layers 50
Crop Pct 0.875
Batch Size 256
Image Size 224
Weight Decay 0.0001
Interpolation bicubic
SHOW MORE
SHOW LESS
ecaresnet50d_pruned

Parameters 20 Million
FLOPs 3 Billion
File Size 76.28 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Efficient Channel Attention, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Squeeze-and-Excitation Block
ID ecaresnet50d_pruned
Layers 50
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
ecaresnetlight

Parameters 30 Million
FLOPs 5 Billion
File Size 115.35 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Efficient Channel Attention, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Squeeze-and-Excitation Block
ID ecaresnetlight
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
README.md

Summary

An ECA ResNet is a variant on a ResNet that utilises an Efficient Channel Attention module. Efficient Channel Attention is an architectural unit based on squeeze-and-excitation blocks that reduces model complexity without dimensionality reduction.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('ecaresnet50d', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. ecaresnet50d. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{wang2020ecanet,
      title={ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks}, 
      author={Qilong Wang and Banggu Wu and Pengfei Zhu and Peihua Li and Wangmeng Zuo and Qinghua Hu},
      year={2020},
      eprint={1910.03151},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
ecaresnet101d 82.18% 96.06%
ecaresnet101d_pruned 80.82% 95.64%
ecaresnet50d 80.61% 95.31%
ecaresnetlight 80.46% 95.25%
ecaresnet50d_pruned 79.71% 94.88%