Skip to content

Commit 8d81fdf

Browse files
JosuaRiederrwightman
authored andcommitted
Fix typos
1 parent 3677f67 commit 8d81fdf

33 files changed

+48
-48
lines changed

CONTRIBUTING.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Code linting and auto-format (black) are not currently in place but open to cons
1010

1111
A few specific differences from Google style (or black)
1212
1. Line length is 120 char. Going over is okay in some cases (e.g. I prefer not to break URL across lines).
13-
2. Hanging indents are always prefered, please avoid aligning arguments with closing brackets or braces.
13+
2. Hanging indents are always preferred, please avoid aligning arguments with closing brackets or braces.
1414

1515
Example, from Google guide, but this is a NO here:
1616
```

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -238,7 +238,7 @@ Add a set of new very well trained ResNet & ResNet-V2 18/34 (basic block) weight
238238
### May 14, 2024
239239
* Support loading PaliGemma jax weights into SigLIP ViT models with average pooling.
240240
* Add Hiera models from Meta (https://github.com/facebookresearch/hiera).
241-
* Add `normalize=` flag for transorms, return non-normalized torch.Tensor with original dytpe (for `chug`)
241+
* Add `normalize=` flag for transforms, return non-normalized torch.Tensor with original dytpe (for `chug`)
242242
* Version 1.0.3 release
243243

244244
### May 11, 2024

hfdocs/source/changes.mdx

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -93,7 +93,7 @@
9393
### May 14, 2024
9494
* Support loading PaliGemma jax weights into SigLIP ViT models with average pooling.
9595
* Add Hiera models from Meta (https://github.com/facebookresearch/hiera).
96-
* Add `normalize=` flag for transorms, return non-normalized torch.Tensor with original dytpe (for `chug`)
96+
* Add `normalize=` flag for transforms, return non-normalized torch.Tensor with original dytpe (for `chug`)
9797
* Version 1.0.3 release
9898

9999
### May 11, 2024
@@ -125,7 +125,7 @@
125125
### April 11, 2024
126126
* Prepping for a long overdue 1.0 release, things have been stable for a while now.
127127
* Significant feature that's been missing for a while, `features_only=True` support for ViT models with flat hidden states or non-std module layouts (so far covering `'vit_*', 'twins_*', 'deit*', 'beit*', 'mvitv2*', 'eva*', 'samvit_*', 'flexivit*'`)
128-
* Above feature support achieved through a new `forward_intermediates()` API that can be used with a feature wrapping module or direclty.
128+
* Above feature support achieved through a new `forward_intermediates()` API that can be used with a feature wrapping module or directly.
129129
```python
130130
model = timm.create_model('vit_base_patch16_224')
131131
final_feat, intermediates = model.forward_intermediates(input)
@@ -360,7 +360,7 @@ Datasets & transform refactoring
360360
* 0.8.15dev0
361361

362362
### Feb 20, 2023
363-
* Add 320x320 `convnext_large_mlp.clip_laion2b_ft_320` and `convnext_lage_mlp.clip_laion2b_ft_soup_320` CLIP image tower weights for features & fine-tune
363+
* Add 320x320 `convnext_large_mlp.clip_laion2b_ft_320` and `convnext_large_mlp.clip_laion2b_ft_soup_320` CLIP image tower weights for features & fine-tune
364364
* 0.8.13dev0 pypi release for latest changes w/ move to huggingface org
365365

366366
### Feb 16, 2023
@@ -745,7 +745,7 @@ More models, more fixes
745745
* Add 'group matching' API to all models to allow grouping model parameters for application of 'layer-wise' LR decay, lr scale added to LR scheduler
746746
* Gradient checkpointing support added to many models
747747
* `forward_head(x, pre_logits=False)` fn added to all models to allow separate calls of `forward_features` + `forward_head`
748-
* All vision transformer and vision MLP models update to return non-pooled / non-token selected features from `foward_features`, for consistency with CNN models, token selection or pooling now applied in `forward_head`
748+
* All vision transformer and vision MLP models update to return non-pooled / non-token selected features from `forward_features`, for consistency with CNN models, token selection or pooling now applied in `forward_head`
749749

750750
### Feb 2, 2022
751751
* [Chris Hughes](https://github.com/Chris-hughes10) posted an exhaustive run through of `timm` on his blog yesterday. Well worth a read. [Getting Started with PyTorch Image Models (timm): A Practitioner’s Guide](https://towardsdatascience.com/getting-started-with-pytorch-image-models-timm-a-practitioners-guide-4e77b4bf9055)
@@ -1058,7 +1058,7 @@ More models, more fixes
10581058
* Add 'group matching' API to all models to allow grouping model parameters for application of 'layer-wise' LR decay, lr scale added to LR scheduler
10591059
* Gradient checkpointing support added to many models
10601060
* `forward_head(x, pre_logits=False)` fn added to all models to allow separate calls of `forward_features` + `forward_head`
1061-
* All vision transformer and vision MLP models update to return non-pooled / non-token selected features from `foward_features`, for consistency with CNN models, token selection or pooling now applied in `forward_head`
1061+
* All vision transformer and vision MLP models update to return non-pooled / non-token selected features from `forward_features`, for consistency with CNN models, token selection or pooling now applied in `forward_head`
10621062

10631063
### Feb 2, 2022
10641064
* [Chris Hughes](https://github.com/Chris-hughes10) posted an exhaustive run through of `timm` on his blog yesterday. Well worth a read. [Getting Started with PyTorch Image Models (timm): A Practitioner’s Guide](https://towardsdatascience.com/getting-started-with-pytorch-image-models-timm-a-practitioners-guide-4e77b4bf9055)

hfdocs/source/models/adversarial-inception-v3.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Adversarial Inception v3
22

3-
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifer](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
3+
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifier](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
44

55
This particular model was trained for study of adversarial examples (adversarial training).
66

hfdocs/source/models/gloun-inception-v3.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# (Gluon) Inception v3
22

3-
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifer](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
3+
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifier](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
44

55
The weights from this model were ported from [Gluon](https://cv.gluon.ai/model_zoo/classification.html).
66

hfdocs/source/models/inception-v3.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Inception v3
22

3-
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifer](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
3+
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifier](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
44

55
## How do I use this model on an image?
66

hfdocs/source/models/tf-inception-v3.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# (Tensorflow) Inception v3
22

3-
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifer](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
3+
**Inception v3** is a convolutional neural network architecture from the Inception family that makes several improvements including using [Label Smoothing](https://paperswithcode.com/method/label-smoothing), Factorized 7 x 7 convolutions, and the use of an [auxiliary classifier](https://paperswithcode.com/method/auxiliary-classifier) to propagate label information lower down the network (along with the use of batch normalization for layers in the sidehead). The key building block is an [Inception Module](https://paperswithcode.com/method/inception-v3-module).
44

55
The weights from this model were ported from [Tensorflow/Models](https://github.com/tensorflow/models).
66

timm/data/auto_augment.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -954,7 +954,7 @@ def augment_and_mix_transform(config_str: str, hparams: Optional[Dict] = None):
954954
Args:
955955
config_str (str): String defining configuration of random augmentation. Consists of multiple sections separated
956956
by dashes ('-'). The first section defines the specific variant of rand augment (currently only 'rand').
957-
The remaining sections, not order sepecific determine
957+
The remaining sections, not order specific determine
958958
'm' - integer magnitude (severity) of augmentation mix (default: 3)
959959
'w' - integer width of augmentation chain (default: 3)
960960
'd' - integer depth of augmentation chain (-1 is random [1, 3], default: -1)

timm/data/imagenet_info.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ def __init__(self, subset: str = 'imagenet-1k'):
5252
subset = re.sub(r'[-_\s]', '', subset.lower())
5353
assert subset in _SUBSETS, f'Unknown imagenet subset {subset}.'
5454

55-
# WordNet synsets (part-of-speach + offset) are the unique class label names for ImageNet classifiers
55+
# WordNet synsets (part-of-speech + offset) are the unique class label names for ImageNet classifiers
5656
synset_file = _SUBSETS[subset]
5757
synset_data = pkgutil.get_data(__name__, os.path.join('_info', synset_file))
5858
self._synsets = synset_data.decode('utf-8').splitlines()

timm/data/readers/reader_hfids.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -80,7 +80,7 @@ def __init__(
8080
self.num_samples = split_info.num_examples
8181
else:
8282
raise ValueError(
83-
"Dataset length is unknown, please pass `num_samples` explicitely. "
83+
"Dataset length is unknown, please pass `num_samples` explicitly. "
8484
"The number of steps needs to be known in advance for the learning rate scheduler."
8585
)
8686

0 commit comments

Comments
 (0)