Papers with Code - ResNet (2024)

Add Metadata

Your model lacks metadata. Adding metadata gives context on how your model was trained.

Take the following JSON template, fill it in with your model's correct values:

{ "Parameters": 62000000 "FLOPs": 524000000 "Training Time": "24 hours", "Training Resources": "8 NVIDIA V100 GPUs", "Training Data": ["ImageNet, Instagram"], "Training Techniques": ["AdamW, CutMix"]}

[INSERT ADVICE HERE]

Papers with Code - ResNet (1) rwightman / pytorch-image-models

Last updated on Feb 14, 2021

resnet18

Parameters 12 Million

FLOPs 2 Billion

File Size 44.66 MB

Training Data ImageNet

Training Resources

Training Time

Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID resnet18
Crop Pct 0.875
Image Size 224
Interpolation bilinear
SHOW MORE
SHOW LESS

resnet26

Parameters 16 Million

FLOPs 3 Billion

File Size 61.16 MB

Training Data ImageNet

Training Resources

Training Time

Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID resnet26
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS

resnet34

Parameters 22 Million

FLOPs 5 Billion

File Size 83.25 MB

Training Data ImageNet

Training Resources

Training Time

Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID resnet34
Crop Pct 0.875
Image Size 224
Interpolation bilinear
SHOW MORE
SHOW LESS

resnet50

Parameters 26 Million

FLOPs 5 Billion

File Size 97.74 MB

Training Data ImageNet

Training Resources

Training Time

Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID resnet50
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS

resnetblur50

Parameters 26 Million

FLOPs 7 Billion

File Size 97.74 MB

Training Data ImageNet

Training Resources

Training Time

Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax, Blur Pooling
ID resnetblur50
Crop Pct 0.875
Image Size 224
Interpolation bicubic
SHOW MORE
SHOW LESS

tv_resnet101

Parameters 45 Million

FLOPs 10 Billion

File Size 170.45 MB

Training Data ImageNet

Training Resources

Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID tv_resnet101
LR 0.1
Epochs 90
Crop Pct 0.875
LR Gamma 0.1
Momentum 0.9
Batch Size 32
Image Size 224
LR Step Size 30
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS

tv_resnet152

Parameters 60 Million

FLOPs 15 Billion

File Size 230.34 MB

Training Data ImageNet

Training Resources

Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID tv_resnet152
LR 0.1
Epochs 90
Crop Pct 0.875
LR Gamma 0.1
Momentum 0.9
Batch Size 32
Image Size 224
LR Step Size 30
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS

tv_resnet34

Parameters 22 Million

FLOPs 5 Billion

File Size 83.26 MB

Training Data ImageNet

Training Resources

Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID tv_resnet34
LR 0.1
Epochs 90
Crop Pct 0.875
LR Gamma 0.1
Momentum 0.9
Batch Size 32
Image Size 224
LR Step Size 30
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS

tv_resnet50

Parameters 26 Million

FLOPs 5 Billion

File Size 97.75 MB

Training Data ImageNet

Training Resources

Training Time

Training Techniques SGD with Momentum, Weight Decay
Architecture 1x1 Convolution, Bottleneck Residual Block, Batch Normalization, Convolution, Global Average Pooling, Residual Block, Residual Connection, ReLU, Max Pooling, Softmax
ID tv_resnet50
LR 0.1
Epochs 90
Crop Pct 0.875
LR Gamma 0.1
Momentum 0.9
Batch Size 32
Image Size 224
LR Step Size 30
Weight Decay 0.0001
Interpolation bilinear
SHOW MORE
SHOW LESS

README.md

Residual Networks, or ResNets, learn residual functions with reference to the layer inputs, instead of learning unreferenced functions. Instead of hoping each few stacked layers directly fit a desired underlying mapping, residual nets let these layers fit a residual mapping. They stack residual blocks ontop of each other to form network: e.g. a ResNet-50 has fifty layers using these blocks.

How do I load this model?

To load a pretrained model:

import timmm = timm.create_model('resnet18', pretrained=True)m.eval()

Replace the model name with the variant you want to use, e.g. resnet18. You can find the IDs in the model summaries at the top of this page.

How do I train this model?

You can follow the timm recipe scripts for training a new model afresh.

Citation

@article{DBLP:journals/corr/HeZRS15, author = {Kaiming He and Xiangyu Zhang and Shaoqing Ren and Jian Sun}, title = {Deep Residual Learning for Image Recognition}, journal = {CoRR}, volume = {abs/1512.03385}, year = {2015}, url = {http://arxiv.org/abs/1512.03385}, archivePrefix = {arXiv}, eprint = {1512.03385}, timestamp = {Wed, 17 Apr 2019 17:23:45 +0200}, biburl = {https://dblp.org/rec/journals/corr/HeZRS15.bib}, bibsource = {dblp computer science bibliography, https://dblp.org}}

Image Classification on ImageNet

Image Classification on ImageNet
MODEL TOP 1 ACCURACY TOP 5 ACCURACY
resnetblur50 79.29% 94.64%
resnet50 79.04% 94.39%
tv_resnet152 78.32% 94.05%
tv_resnet101 77.37% 93.56%
tv_resnet50 76.16% 92.88%
resnet26 75.29% 92.57%
resnet34 75.11% 92.28%
tv_resnet34 73.3% 91.42%
resnet18 69.74% 89.09%
Papers with Code - ResNet (2024)

FAQs

How accurate is ResNet training? ›

curve basically stabilizes when the number of iterations is greater than 2000. The ResNet-50 model training accuracy rises to 97% from 65.6%, and the final model training accuracy reaches 99.61% when the number of iterations is between 200 and 450, which is shown in the ResNet-50 model in Figure 8.

What is the weakness of ResNet? ›

However, one disadvantage of using ResNet is the potential disappearance of gradients in very deep networks, which can make the gradient descent process slow. This can hinder the training of the network, especially as the number of layers increases.

How accurate is the ResNet algorithm? ›

The depth of ResNet for best accuracy is over four times deeper than previous deep networks. Achieved 3.57% top 5 error rate on the test set with 152 layer ResNet on ensemble model.

Why ResNet was able to train deep network select the most appropriate answer? ›

The residual block in ResNet combines the features of the input layer with the features of the adjacent output layer by adding a shortcut to achieve residual learning. Such processing enables the deep network to maintain the same performance as the shallow network in the worst case.

How much time does it take to train ResNet? ›

In a blog-post (http://torch.ch/blog/2016/02/04/resnets.html) linked from the original github, 0.4s / batch of 32 images is recorded for ResNet-50, and it's written that "training took from 3.5 days for the 18-layer model to 14 days for the 101-layer model".

What is a good learning rate for ResNet? ›

Authors follow ResNet to choose 0.1 as the initial learning rate for batch size 256, then when changing to a larger batch size b, the initial learning rate is increased to 0.1×b/256.

What is the problem with ResNet? ›

ResNet and Residual Blocks

The primary problem ResNet solved was the degradation problem in deep neural networks. As networks become deeper, their accuracy saturates and then degrades rapidly. This degradation is not caused by overfitting, but rather the difficulty of optimizing the training process.

Is ViT better than ResNet? ›

Computer vision research indicates that when pre-trained with a sufficient amount of data, ViT models are at least as robust as ResNet models.

Is ResNet better than CNN? ›

ResNet is superior to CNN because it introduces the concept of residual units, which allows deep layers to directly learn from shallow layers, reducing the difficulty of network convergence. This results in better learning ability and improved performance in image recognition tasks 1.

Which ResNet model is best? ›

Therefore, each of the 2-layer blocks in Resnet34 was replaced with a 3-layer bottleneck block, forming the Resnet-50 architecture. This has much higher accuracy than the 34-layer ResNet model. The 50-layer ResNet-50 achieves a performance of 3.8 bn FLOPS.

Is VGG better than ResNet? ›

Performance: Generally, ResNets perform better than VGG in terms of accuracy, especially on deeper models. They are also faster to train. Transfer Learning Use: ResNet, with its deeper and more complex architecture, is often used in transfer learning for more sophisticated tasks.

Why is ResNet so powerful? ›

And with increased network layers, the ResNet-152 has much lower complexity than other deeper models. ResNet has shown that we can effectively architect deeper and deeper networks by creating short paths from the early to later layers. There's no doubt that ResNet has proven powerful in a wide number of applications.

What are the limitations of ResNet? ›

ResNets have limitations in their performance compared to supervised learning on the ImageNet classification benchmark, which restricts their use in performance-critical scenarios. Additionally, conventional training methods for ResNets do not effectively learn iterative solutions, even though ResNets can express them.

Why Inception is better than ResNet? ›

While Inception focuses on computational cost, ResNet focuses on computational accuracy. Intuitively, deeper networks should not perform worse than the shallower networks, but in practice, the deeper networks performed worse than the shallower networks, caused not by overfitting, but by an optimization problem.

Who invented ResNet? ›

ResNet, short for Residual Network is a specific type of neural network that was introduced in 2015 by Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun in their paper “Deep Residual Learning for Image Recognition”.

How accurate is ImageNet ResNet? ›

A paper called “Resnet Strikes Back” demonstrated the significance of the right training for training ResNet50 on ImageNet. They boosted ResNet50 to a top-1 accuracy of 80.4% on ImageNet-1K. The original ResNet50 recipe reached 75.8% accuracy, so this improved.

How accurate is ResNet 34? ›

¶ We used a RESNET34 model to predict and classify fish images and achieved a test accuracy of 99.41%.

How accurate is machine learning training and validation? ›

Training is the process of adjusting the parameters of a neural network to minimize a cost function. Validation is the process of measuring the accuracy of a model on a subset of the data. Accuracy is a measure of how well a model is able to predict the correct output given the input.

How to improve ResNet50 accuracy? ›

The solution here would be to clean up your data and get more of it… That could improve the accuracy. The architecture you are using is not the best one for the job. Instead of a better resnet, you could try switching architectures and see if that improves things.

References

Top Articles
Latest Posts
Article information

Author: Nathanial Hackett

Last Updated:

Views: 6107

Rating: 4.1 / 5 (72 voted)

Reviews: 87% of readers found this page helpful

Author information

Name: Nathanial Hackett

Birthday: 1997-10-09

Address: Apt. 935 264 Abshire Canyon, South Nerissachester, NM 01800

Phone: +9752624861224

Job: Forward Technology Assistant

Hobby: Listening to music, Shopping, Vacation, Baton twirling, Flower arranging, Blacksmithing, Do it yourself

Introduction: My name is Nathanial Hackett, I am a lovely, curious, smiling, lively, thoughtful, courageous, lively person who loves writing and wants to share my knowledge and understanding with you.