Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

refactor: unify the words into one #338

Merged
merged 1 commit into from
Sep 26, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion facebookresearch_WSL-Images_resnext.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
print(torch.nn.functional.softmax(output[0], dim=0))
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
print(torch.nn.functional.softmax(output[0], dim=0))
Expand Down
2 changes: 1 addition & 1 deletion nvidia_deeplearningexamples_efficientnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -55,7 +55,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
print(f'Using {device} for inference')
```

Load the model pretrained on IMAGENET dataset.
Load the model pretrained on ImageNet dataset.

You can choose among the following models:

Expand Down
2 changes: 1 addition & 1 deletion nvidia_deeplearningexamples_gpunet.md
Original file line number Diff line number Diff line change
Expand Up @@ -54,7 +54,7 @@ print(f'Using {device} for inference')
```

### Load Pretrained model
Loads NVIDIA GPUNet-0 model by default pre-trained on IMAGENET dataset. You can switch the default pre-trained model loading from GPUNet-0 to one of the following models listed below.
Loads NVIDIA GPUNet-0 model by default pre-trained on ImageNet dataset. You can switch the default pre-trained model loading from GPUNet-0 to one of the following models listed below.

The model architecture is visible as output of the loaded model. For details architecture and latency info please refer to [architecture section](https://github.com/NVIDIA/DeepLearningExamples/tree/torchhub/PyTorch/Classification/GPUNet#model-architecture) in the original repo and Table#[3](https://arxiv.org/pdf/2205.00841.pdf) in the CVPR-2022 paper, respectively.

Expand Down
2 changes: 1 addition & 1 deletion nvidia_deeplearningexamples_resnet50.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
print(f'Using {device} for inference')
```

Load the model pretrained on IMAGENET dataset.
Load the model pretrained on ImageNet dataset.
```python
resnet50 = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_resnet50', pretrained=True)
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')
Expand Down
2 changes: 1 addition & 1 deletion nvidia_deeplearningexamples_resnext.md
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
print(f'Using {device} for inference')
```

Load the model pretrained on IMAGENET dataset.
Load the model pretrained on ImageNet dataset.
```python
resneXt = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_resneXt')
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')
Expand Down
2 changes: 1 addition & 1 deletion nvidia_deeplearningexamples_se-resnext.md
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cp
print(f'Using {device} for inference')
```

Load the model pretrained on IMAGENET dataset.
Load the model pretrained on ImageNet dataset.
```python
resneXt = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_se_resnext101_32x4d')
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')
Expand Down
6 changes: 3 additions & 3 deletions pytorch_vision_alexnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand All @@ -85,11 +85,11 @@ for i in range(top5_prob.size(0)):

AlexNet competed in the ImageNet Large Scale Visual Recognition Challenge on September 30, 2012. The network achieved a top-5 error of 15.3%, more than 10.8 percentage points lower than that of the runner up. The original paper's primary result was that the depth of the model was essential for its high performance, which was computationally expensive, but made feasible due to the utilization of graphics processing units (GPUs) during training.

The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
| alexnet | 43.45 | 20.91 |
| AlexNet | 43.45 | 20.91 |

### References

Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_densenet.md
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand All @@ -89,7 +89,7 @@ for i in range(top5_prob.size(0)):

Dense Convolutional Network (DenseNet), connects each layer to every other layer in a feed-forward fashion. Whereas traditional convolutional networks with L layers have L connections - one between each layer and its subsequent layer - our network has L(L+1)/2 direct connections. For each layer, the feature-maps of all preceding layers are used as inputs, and its own feature-maps are used as inputs into all subsequent layers. DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters.

The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_ghostnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_googlenet.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_hardnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down Expand Up @@ -95,7 +95,7 @@ were designed for comparing with MobileNet).

Here we have the 4 versions of hardnet models, which contains 39, 68, 85 layers
w/ or w/o Depthwise Separable Conv respectively.
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_ibnnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
6 changes: 3 additions & 3 deletions pytorch_vision_inception_v3.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ layout: hub_detail
background-class: hub-background
body-class: hub
title: Inception_v3
summary: Also called GoogleNetv3, a famous ConvNet trained on Imagenet from 2015
summary: Also called GoogleNetv3, a famous ConvNet trained on ImageNet from 2015
category: researchers
image: inception_v3.png
author: Pytorch Team
Expand Down Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand All @@ -85,7 +85,7 @@ for i in range(top5_prob.size(0)):

Inception v3: Based on the exploration of ways to scale up networks in ways that aim at utilizing the added computation as efficiently as possible by suitably factorized convolutions and aggressive regularization. We benchmark our methods on the ILSVRC 2012 classification challenge validation set demonstrate substantial gains over the state of the art: 21.2% top-1 and 5.6% top-5 error for single frame evaluation using a network with a computational cost of 5 billion multiply-adds per inference and with using less than 25 million parameters. With an ensemble of 4 models and multi-crop evaluation, we report 3.5% top-5 error on the validation set (3.6% error on the test set) and 17.3% top-1 error on the validation set.

The 1-crop error rates on the imagenet dataset with the pretrained model are listed below.
The 1-crop error rates on the ImageNet dataset with the pretrained model are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_meal_v2.md
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_mobilenet_v2.md
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_once_for_all.md
Original file line number Diff line number Diff line change
Expand Up @@ -105,7 +105,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_proxylessnas.md
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_resnest.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_resnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand All @@ -91,7 +91,7 @@ for i in range(top5_prob.size(0)):
Resnet models were proposed in "Deep Residual Learning for Image Recognition".
Here we have the 5 versions of resnet models, which contains 18, 34, 50, 101, 152 layers respectively.
Detailed model architectures can be found in Table 1.
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_resnext.md
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down Expand Up @@ -90,7 +90,7 @@ for i in range(top5_prob.size(0)):
Resnext models were proposed in [Aggregated Residual Transformations for Deep Neural Networks](https://arxiv.org/abs/1611.05431).
Here we have the 2 versions of resnet models, which contains 50, 101 layers repspectively.
A comparison in model archetechure between resnet50 and resnext50 can be found in Table 1.
Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.

| Model structure | Top-1 error | Top-5 error |
| ----------------- | ----------- | ----------- |
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_shufflenet_v2.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ layout: hub_detail
background-class: hub-background
body-class: hub
title: ShuffleNet v2
summary: An efficient ConvNet optimized for speed and memory, pre-trained on Imagenet
summary: An efficient ConvNet optimized for speed and memory, pre-trained on ImageNet
category: researchers
image: shufflenet_v2_1.png
author: Pytorch Team
Expand Down Expand Up @@ -59,7 +59,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_snnmlp.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
print(torch.nn.functional.softmax(output[0], dim=0))
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_squeezenet.md
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down Expand Up @@ -90,7 +90,7 @@ Model `squeezenet1_0` is from the [SqueezeNet: AlexNet-level accuracy with 50x f
Model `squeezenet1_1` is from the [official squeezenet repo](https://github.com/DeepScale/SqueezeNet/tree/master/SqueezeNet_v1.1).
It has 2.4x less computation and slightly fewer parameters than `squeezenet1_0`, without sacrificing accuracy.

Their 1-crop error rates on imagenet dataset with pretrained models are listed below.
Their 1-crop error rates on ImageNet dataset with pretrained models are listed below.

| Model structure | Top-1 error | Top-5 error |
| --------------- | ----------- | ----------- |
Expand Down
4 changes: 2 additions & 2 deletions pytorch_vision_vgg.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@ layout: hub_detail
background-class: hub-background
body-class: hub
title: vgg-nets
summary: Award winning ConvNets from 2014 Imagenet ILSVRC challenge
summary: Award winning ConvNets from 2014 ImageNet ILSVRC challenge
category: researchers
image: vgg.png
author: Pytorch Team
Expand Down Expand Up @@ -67,7 +67,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
2 changes: 1 addition & 1 deletion pytorch_vision_wide_resnet.md
Original file line number Diff line number Diff line change
Expand Up @@ -62,7 +62,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand Down
4 changes: 2 additions & 2 deletions simplenet.md
Original file line number Diff line number Diff line change
Expand Up @@ -67,7 +67,7 @@ if torch.cuda.is_available():

with torch.no_grad():
output = model(input_batch)
# Tensor of shape 1000, with confidence scores over Imagenet's 1000 classes
# Tensor of shape 1000, with confidence scores over ImageNet's 1000 classes
print(output[0])
# The output has unnormalized scores. To get probabilities, you can run a softmax on it.
probabilities = torch.nn.functional.softmax(output[0], dim=0)
Expand All @@ -94,7 +94,7 @@ for i in range(top5_prob.size(0)):
SimpleNet models were proposed in "Lets Keep it simple, Using simple architectures to outperform deeper and more complex architectures".
Here we have the 8 versions of simplenet models, which contains 1.5m, 3.2m, 5.7m and 9.5m parameters respectively.
Detailed model architectures can be found in Table 1 and Table 2.
Their 1-crop errors on imagenet dataset with pretrained models are listed below.
Their 1-crop errors on ImageNet dataset with pretrained models are listed below.

The m2 variants

Expand Down
Loading