​​Revisiting ResNets: Improved Training and Scaling Strategies



The authors of the paper (from Google Brain and UC Berkeley) have decided to analyze the effects of the model architecture, training, and scaling strategies separately and concluded that these strategies might have a higher impact on the score than the architecture.



They offer two new strategies:

- scale model depth if overfitting is possible, scale model width otherwise

- increase image resolution slower than recommended in previous papers



Based on these ideas, the new architecture ResNet-RS was developed. It is 2.1x–3.3x faster than EfficientNets on GPU while reaching similar accuracy on ImageNet.



In semi-supervised learning, ResNet-RS achieves 86.2% top-1 ImageNet accuracy while being 4.7x faster than EfficientNet-NoisyStudent.



Transfer learning on downstream tasks also has improved performance.



The authors suggest using these ResNet-RS as a baseline for further research.





Paper: https://arxiv.org/abs/2103.07579



Code and checkpoints are available in TensorFlow:

https://github.com/tensorflow/models/tree/master/official/vision/beta



https://github.com/tensorflow/tpu/tree/master/models/official/resnet/resnet_rs



A detailed unofficial overview of the paper: https://andlukyane.com/blog/paper-review-resnetsr





#deeplearning #computervision #sota