| | # (Gluon) Inception v3 |
| |
|
| | **Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifier](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module). |
| |
|
| | The weights from this model were ported from [Gluon](https://cv.gluon.ai/model_zoo/classification.html). |
| |
|
| | ## How do I use this model on an image? |
| |
|
| | To load a pretrained model: |
| |
|
| | ```py |
| | >>> import timm |
| | >>> model = timm.create_model('gluon_inception_v3', pretrained=True) |
| | >>> model.eval() |
| | ``` |
| |
|
| | To load and preprocess the image: |
| |
|
| | ```py |
| | >>> import urllib |
| | >>> from PIL import Image |
| | >>> from timm.data import resolve_data_config |
| | >>> from timm.data.transforms_factory import create_transform |
| |
|
| | >>> config = resolve_data_config({}, model=model) |
| | >>> transform = create_transform(**config) |
| |
|
| | >>> url, filename = ("https://github.com/pytorch/hub/raw/master/images/dog.jpg", "dog.jpg") |
| | >>> urllib.request.urlretrieve(url, filename) |
| | >>> img = Image.open(filename).convert('RGB') |
| | >>> tensor = transform(img).unsqueeze(0) |
| | ``` |
| |
|
| | To get the model predictions: |
| |
|
| | ```py |
| | >>> import torch |
| | >>> with torch.no_grad(): |
| | ... out = model(tensor) |
| | >>> probabilities = torch.nn.functional.softmax(out[0], dim=0) |
| | >>> print(probabilities.shape) |
| | >>> |
| | ``` |
| |
|
| | To get the top-5 predictions class names: |
| |
|
| | ```py |
| | >>> |
| | >>> url, filename = ("https://raw.githubusercontent.com/pytorch/hub/master/imagenet_classes.txt", "imagenet_classes.txt") |
| | >>> urllib.request.urlretrieve(url, filename) |
| | >>> with open("imagenet_classes.txt", "r") as f: |
| | ... categories = [s.strip() for s in f.readlines()] |
| |
|
| | >>> |
| | >>> top5_prob, top5_catid = torch.topk(probabilities, 5) |
| | >>> for i in range(top5_prob.size(0)): |
| | ... print(categories[top5_catid[i]], top5_prob[i].item()) |
| | >>> |
| | >>> |
| | ``` |
| |
|
| | Replace the model name with the variant you want to use, e.g. `gluon_inception_v3`. You can find the IDs in the model summaries at the top of this page. |
| |
|
| | To extract image features with this model, follow the [timm feature extraction examples](../feature_extraction), just change the name of the model you want to use. |
| |
|
| | ## How do I finetune this model? |
| |
|
| | You can finetune any of the pre-trained models just by changing the classifier (the last layer). |
| |
|
| | ```py |
| | >>> model = timm.create_model('gluon_inception_v3', pretrained=True, num_classes=NUM_FINETUNE_CLASSES) |
| | ``` |
| | To finetune on your own dataset, you have to write a training loop or adapt [timm's training |
| | script](https://github.com/rwightman/pytorch-image-models/blob/master/train.py) to use your dataset. |
| |
|
| | ## How do I train this model? |
| |
|
| | You can follow the [timm recipe scripts](../training_script) for training a new model afresh. |
| |
|
| | ## Citation |
| |
|
| | ```BibTeX |
| | @article{DBLP:journals/corr/SzegedyVISW15, |
| | author = {Christian Szegedy and |
| | Vincent Vanhoucke and |
| | Sergey Ioffe and |
| | Jonathon Shlens and |
| | Zbigniew Wojna}, |
| | title = {Rethinking the Inception Architecture for Computer Vision}, |
| | journal = {CoRR}, |
| | volume = {abs/1512.00567}, |
| | year = {2015}, |
| | url = {http://arxiv.org/abs/1512.00567}, |
| | archivePrefix = {arXiv}, |
| | eprint = {1512.00567}, |
| | timestamp = {Mon, 13 Aug 2018 16:49:07 +0200}, |
| | biburl = {https://dblp.org/rec/journals/corr/SzegedyVISW15.bib}, |
| | bibsource = {dblp computer science bibliography, https://dblp.org} |
| | } |
| | ``` |
| |
|
| | <!-- |
| | Type: model-index |
| | Collections: |
| | - Name: Gloun Inception v3 |
| | Paper: |
| | Title: Rethinking the Inception Architecture for Computer Vision |
| | URL: https://paperswithcode.com/paper/rethinking-the-inception-architecture-for |
| | Models: |
| | - Name: gluon_inception_v3 |
| | In Collection: Gloun Inception v3 |
| | Metadata: |
| | FLOPs: 7352418880 |
| | Parameters: 23830000 |
| | File Size: 95567055 |
| | Architecture: |
| | - 1x1 Convolution |
| | - Auxiliary Classifier |
| | - Average Pooling |
| | - Average Pooling |
| | - Batch Normalization |
| | - Convolution |
| | - Dense Connections |
| | - Dropout |
| | - Inception-v3 Module |
| | - Max Pooling |
| | - ReLU |
| | - Softmax |
| | Tasks: |
| | - Image Classification |
| | Training Data: |
| | - ImageNet |
| | ID: gluon_inception_v3 |
| | Crop Pct: '0.875' |
| | Image Size: '299' |
| | Interpolation: bicubic |
| | Code: https://github.com/rwightman/pytorch-image-models/blob/d8e69206be253892b2956341fea09fdebfaae4e3/timm/models/inception_v3.py#L464 |
| | Weights: https://github.com/rwightman/pytorch-image-models/releases/download/v0.1-weights/gluon_inception_v3-9f746940.pth |
| | Results: |
| | - Task: Image Classification |
| | Dataset: ImageNet |
| | Metrics: |
| | Top 1 Accuracy: 78.8% |
| | Top 5 Accuracy: 94.38% |
| | --> |