You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+18-32Lines changed: 18 additions & 32 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,15 @@
2
2
3
3
## What's New
4
4
5
+
### Feb 10, 2021
6
+
* More model archs, incl a flexible ByobNet backbone ('Bring-your-own-blocks')
7
+
* GPU-Efficient-Networks (https://github.com/idstcv/GPU-Efficient-Networks), impl in `byobnet.py`
8
+
* RepVGG (https://github.com/DingXiaoH/RepVGG), impl in `byobnet.py`
9
+
* classic VGG (from torchvision, impl in `vgg.py`)
10
+
* Refinements to normalizer layer arg handling and normalizer+act layer handling in some models
11
+
* Default AMP mode changed to native PyTorch AMP instead of APEX. Issues not being fixed with APEX. Native works with `--channels-last` and `--torchscript` model training, APEX does not.
12
+
* Fix a few bugs introduced since last pypi release
13
+
5
14
### Feb 8, 2021
6
15
* Add several ResNet weights with ECA attention. 26t & 50t trained @ 256, test @ 320. 269d train @ 256, fine-tune @320, test @ 352.
* Some import cleanup and classifier reset changes, all models will have classifier reset to nn.Identity on reset_classifer(0) call
119
128
* Prep for 0.1.28 pip release
120
129
121
-
### May 12, 2020
122
-
* Add ResNeSt models (code adapted from https://github.com/zhanghang1989/ResNeSt, paper https://arxiv.org/abs/2004.08955))
123
-
124
-
### May 3, 2020
125
-
* Pruned EfficientNet B1, B2, and B3 (https://arxiv.org/abs/2002.08258) contributed by [Yonathan Aflalo](https://github.com/yoniaflalo)
126
-
127
-
### May 1, 2020
128
-
* Merged a number of execellent contributions in the ResNet model family over the past month
129
-
* BlurPool2D and resnetblur models initiated by [Chris Ha](https://github.com/VRandme), I trained resnetblur50 to 79.3.
130
-
* TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by [mrT23](https://github.com/mrT23)
131
-
* ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (https://arxiv.org/abs/2002.08258) by [Yonathan Aflalo](https://github.com/yoniaflalo)
132
-
* 200 pretrained models in total now with updated results csv in results folder
133
-
134
-
### April 5, 2020
135
-
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
136
-
* 3.5M param MobileNet-V2 100 @ 73%
137
-
* 4.5M param MobileNet-V2 110d @ 75%
138
-
* 6.1M param MobileNet-V2 140 @ 76.5%
139
-
* 5.8M param MobileNet-V2 120d @ 77.3%
140
-
141
-
### March 18, 2020
142
-
* Add EfficientNet-Lite models w/ weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite)
143
-
* Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by [Andrew Lavin](https://github.com/andravin) (see Training section for hparams)
144
-
145
130
## Introduction
146
131
147
132
Py**T**orch **Im**age **M**odels (`timm`) is a collection of image models, layers, utilities, optimizers, schedulers, data-loaders / augmentations, and reference training / validation scripts that aim to pull together a wide variety of SOTA models with ability to reproduce ImageNet training results.
@@ -150,7 +135,7 @@ The work of many others is present here. I've tried to make sure all source mate
150
135
151
136
## Models
152
137
153
-
All model architecture families include variants with pretrained weights. The are variants without any weights. Help training new or better weights is always appreciated. Here are some example [training hparams](https://rwightman.github.io/pytorch-image-models/training_hparam_examples) to get you started.
138
+
All model architecture families include variants with pretrained weights. There are specific model variants without any weights, it is NOT a bug. Help training new or better weights is always appreciated. Here are some example [training hparams](https://rwightman.github.io/pytorch-image-models/training_hparam_examples) to get you started.
154
139
155
140
A full version of the list below with source links can be found in the [documentation](https://rwightman.github.io/pytorch-image-models/models/).
156
141
@@ -170,6 +155,7 @@ A full version of the list below with source links can be found in the [document
170
155
* MNASNet B1, A1 (Squeeze-Excite), and Small - https://arxiv.org/abs/1807.11626
171
156
* MobileNet-V2 - https://arxiv.org/abs/1801.04381
172
157
* Single-Path NAS - https://arxiv.org/abs/1904.02877
@@ -261,9 +248,10 @@ The root folder of the repository contains reference train, validation, and infe
261
248
262
249
One of the greatest assets of PyTorch is the community and their contributions. A few of my favourite resources that pair well with the models and componenets here are listed below.
Copy file name to clipboardExpand all lines: docs/archived_changes.md
+24Lines changed: 24 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,29 @@
1
1
# Archived Changes
2
2
3
+
### May 12, 2020
4
+
* Add ResNeSt models (code adapted from https://github.com/zhanghang1989/ResNeSt, paper https://arxiv.org/abs/2004.08955))
5
+
6
+
### May 3, 2020
7
+
* Pruned EfficientNet B1, B2, and B3 (https://arxiv.org/abs/2002.08258) contributed by [Yonathan Aflalo](https://github.com/yoniaflalo)
8
+
9
+
### May 1, 2020
10
+
* Merged a number of execellent contributions in the ResNet model family over the past month
11
+
* BlurPool2D and resnetblur models initiated by [Chris Ha](https://github.com/VRandme), I trained resnetblur50 to 79.3.
12
+
* TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by [mrT23](https://github.com/mrT23)
13
+
* ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (https://arxiv.org/abs/2002.08258) by [Yonathan Aflalo](https://github.com/yoniaflalo)
14
+
* 200 pretrained models in total now with updated results csv in results folder
15
+
16
+
### April 5, 2020
17
+
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
18
+
* 3.5M param MobileNet-V2 100 @ 73%
19
+
* 4.5M param MobileNet-V2 110d @ 75%
20
+
* 6.1M param MobileNet-V2 140 @ 76.5%
21
+
* 5.8M param MobileNet-V2 120d @ 77.3%
22
+
23
+
### March 18, 2020
24
+
* Add EfficientNet-Lite models w/ weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite)
25
+
* Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by [Andrew Lavin](https://github.com/andravin) (see Training section for hparams)
26
+
3
27
### April 5, 2020
4
28
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
Copy file name to clipboardExpand all lines: docs/changes.md
+50Lines changed: 50 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,55 @@
1
1
# Recent Changes
2
2
3
+
### Feb 10, 2021
4
+
* More model archs, incl a flexible ByobNet backbone ('Bring-your-own-blocks')
5
+
* GPU-Efficient-Networks (https://github.com/idstcv/GPU-Efficient-Networks), impl in `byobnet.py`
6
+
* RepVGG (https://github.com/DingXiaoH/RepVGG), impl in `byobnet.py`
7
+
* classic VGG (from torchvision, impl in `vgg`)
8
+
* Refinements to normalizer layer arg handling and normalizer+act layer handling in some models
9
+
* Default AMP mode changed to native PyTorch AMP instead of APEX. Issues not being fixed with APEX. Native works with `--channels-last` and `--torchscript` model training, APEX does not.
10
+
* Fix a few bugs introduced since last pypi release
11
+
12
+
### Feb 8, 2021
13
+
* Add several ResNet weights with ECA attention. 26t & 50t trained @ 256, test @ 320. 269d train @ 256, fine-tune @320, test @ 352.
* Remove separate tiered (`t`) vs tiered_narrow (`tn`) ResNet model defs, all `tn` changed to `t` and `t` models removed (`seresnext26t_32x4d` only model w/ weights that was removed).
18
+
* Support model default_cfgs with separate train vs test resolution `test_input_size` and remove extra `_320` suffix ResNet model defs that were just for test.
19
+
20
+
### Jan 30, 2021
21
+
* Add initial "Normalization Free" NF-RegNet-B* and NF-ResNet model definitions based on [paper](https://arxiv.org/abs/2101.08692)
22
+
23
+
### Jan 25, 2021
24
+
* Add ResNetV2 Big Transfer (BiT) models w/ ImageNet-1k and 21k weights from https://github.com/google-research/big_transfer
25
+
* Add official R50+ViT-B/16 hybrid models + weights from https://github.com/google-research/vision_transformer
26
+
* ImageNet-21k ViT weights are added w/ model defs and representation layer (pre logits) support
27
+
* NOTE: ImageNet-21k classifier heads were zero'd in original weights, they are only useful for transfer learning
28
+
* Add model defs and weights for DeiT Vision Transformer models from https://github.com/facebookresearch/deit
29
+
* Refactor dataset classes into ImageDataset/IterableImageDataset + dataset specific parser classes
30
+
* Add Tensorflow-Datasets (TFDS) wrapper to allow use of TFDS image classification sets with train script
Copy file name to clipboardExpand all lines: docs/scripts.md
+3-3Lines changed: 3 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,9 +10,9 @@ The variety of training args is large and not all combinations of options (or ev
10
10
11
11
To train an SE-ResNet34 on ImageNet, locally distributed, 4 GPUs, one process per GPU w/ cosine schedule, random-erasing prob of 50% and per-pixel random value:
NOTE: NVIDIA APEX should be installed to run in per-process distributed via DDP or to enable AMP mixed precision with the --amp flag
15
+
NOTE: It is recommended to use PyTorch 1.7+ w/ PyTorch native AMP and DDP instead of APEX AMP. `--amp` defaults to native AMP as of timm ver 0.4.3. `--apex-amp` will force use of APEX components if they are installed.
16
16
17
17
## Validation / Inference Scripts
18
18
@@ -24,4 +24,4 @@ To validate with the model's pretrained weights (if they exist):
0 commit comments