TF EfficientNet

Last updated on Feb 14, 2021

tf_efficientnet_b0

Parameters 5 Million
FLOPs 489 Million
File Size 20.39 MB
Training Data ImageNet
Training Resources TPUv3 Cloud TPU
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b0
LR 0.256
Epochs 350
Crop Pct 0.875
Momentum 0.9
Batch Size 2048
Image Size 224
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b1

Parameters 8 Million
FLOPs 884 Million
File Size 30.05 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b1
LR 0.256
Epochs 350
Crop Pct 0.882
Momentum 0.9
Batch Size 2048
Image Size 240
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b2

Parameters 9 Million
FLOPs 1 Billion
File Size 35.09 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b2
LR 0.256
Epochs 350
Crop Pct 0.89
Momentum 0.9
Batch Size 2048
Image Size 260
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b3

Parameters 12 Million
FLOPs 2 Billion
File Size 47.09 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b3
LR 0.256
Epochs 350
Crop Pct 0.904
Momentum 0.9
Batch Size 2048
Image Size 300
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b4

Parameters 19 Million
FLOPs 6 Billion
File Size 74.38 MB
Training Data ImageNet
Training Resources TPUv3 Cloud TPU
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b4
LR 0.256
Epochs 350
Crop Pct 0.922
Momentum 0.9
Batch Size 2048
Image Size 380
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b5

Parameters 30 Million
FLOPs 13 Billion
File Size 116.73 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b5
LR 0.256
Epochs 350
Crop Pct 0.934
Momentum 0.9
Batch Size 2048
Image Size 456
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b6

Parameters 43 Million
FLOPs 24 Billion
File Size 165.21 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b6
LR 0.256
Epochs 350
Crop Pct 0.942
Momentum 0.9
Batch Size 2048
Image Size 528
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b7

Parameters 66 Million
FLOPs 48 Billion
File Size 254.49 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b7
LR 0.256
Epochs 350
Crop Pct 0.949
Momentum 0.9
Batch Size 2048
Image Size 600
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_b8

Parameters 87 Million
FLOPs 81 Billion
File Size 335.10 MB
Training Data ImageNet
Training Resources
Training Time

Training Techniques RMSProp, Weight Decay, Label Smoothing, AutoAugment, Stochastic Depth
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_b8
LR 0.256
Epochs 350
Crop Pct 0.954
Momentum 0.9
Batch Size 2048
Image Size 672
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
SHOW MORE
SHOW LESS
tf_efficientnet_el

Parameters 11 Million
FLOPs 9 Billion
File Size 40.82 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_el
Crop Pct 0.904
Image Size 300
Interpolation bicubic
SHOW MORE
SHOW LESS
tf_efficientnet_em

Parameters 7 Million
FLOPs 4 Billion
File Size 26.64 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_em
Crop Pct 0.882
Image Size 240
Interpolation bicubic
SHOW MORE
SHOW LESS
tf_efficientnet_es

Parameters 5 Million
FLOPs 2 Billion
File Size 20.99 MB
Training Data ImageNet
Training Resources
Training Time

Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_es
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS
tf_efficientnet_l2_ns_475

Parameters 480 Million
FLOPs 218 Billion
File Size 1.84 GB
Training Data JFT-300M, ImageNet
Training Resources TPUv3 Cloud TPU
Training Time

Training Techniques Noisy Student, FixRes, RMSProp, Weight Decay, Label Smoothing, AutoAugment, RandAugment
Architecture 1x1 Convolution, Average Pooling, Convolution, Dense Connections, Dropout, Inverted Residual Block, Batch Normalization, Squeeze-and-Excitation Block, Swish
ID tf_efficientnet_l2_ns_475
LR 0.128
Epochs 350
Dropout 0.5
Crop Pct 0.936
Momentum 0.9
Batch Size 2048
Image Size 475
Weight Decay 0.00001
Interpolation bicubic
RMSProp Decay 0.9
Label Smoothing 0.1
BatchNorm Momentum 0.99
Stochastic Depth Survival 0.8
SHOW MORE
SHOW LESS
README.md

Summary

EfficientNet is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a compound coefficient. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use $2^N$ times more computational resources, then we can simply increase the network depth by $\alpha ^ N$, width by $\beta ^ N$, and image size by $\gamma ^ N$, where $\alpha, \beta, \gamma$ are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient $\phi$ to uniformly scales network width, depth, and resolution in a principled way.

The compound scaling method is justified by the intuition that if the input image is bigger, then the network needs more layers to increase the receptive field and more channels to capture more fine-grained patterns on the bigger image.

The base EfficientNet-B0 network is based on the inverted bottleneck residual blocks of MobileNetV2, in addition to squeeze-and-excitation blocks.

How do I load this model?

To load a pretrained model:

import timm
m = timm.create_model('tf_efficientnet_b0', pretrained=True)
m.eval()

Replace the model name with the variant you want to use, e.g. tf_efficientnet_b0. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@misc{tan2020efficientnet,
      title={EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks}, 
      author={Mingxing Tan and Quoc V. Le},
      year={2020},
      eprint={1905.11946},
      archivePrefix={arXiv},
      primaryClass={cs.LG}
}

Results

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
tf_efficientnet_l2_ns_475 88.24% 98.55%
tf_efficientnet_b8 85.35% 97.39%
tf_efficientnet_b7 84.93% 97.2%
tf_efficientnet_b6 84.11% 96.89%
tf_efficientnet_b5 83.81% 96.75%
tf_efficientnet_b4 83.03% 96.3%
tf_efficientnet_b3 81.65% 95.72%
tf_efficientnet_el 80.45% 95.17%
tf_efficientnet_b2 80.07% 94.9%
tf_efficientnet_b1 78.84% 94.2%
tf_efficientnet_em 78.71% 94.33%
tf_efficientnet_es 77.28% 93.6%
tf_efficientnet_b0 76.85% 93.23%