You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: CONTRIBUTING.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
*This guideline is very much a work-in-progress.*
2
2
3
-
Contriubtions to `timm` for code, documentation, tests are more than welcome!
3
+
Contributions to `timm` for code, documentation, tests are more than welcome!
4
4
5
5
There haven't been any formal guidelines to date so please bear with me, and feel free to add to this guide.
6
6
@@ -49,7 +49,7 @@ This is YES:
49
49
}
50
50
```
51
51
52
-
When there is descrepancy in a given source file (there are many origins for various bits of code and not all have been updated to what I consider current goal), please follow the style in a given file.
52
+
When there is discrepancy in a given source file (there are many origins for various bits of code and not all have been updated to what I consider current goal), please follow the style in a given file.
53
53
54
54
In general, if you add new code, formatting it with black using the following options should result in a style that is compatible with the rest of the code base:
Copy file name to clipboardExpand all lines: README.md
+30Lines changed: 30 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -26,6 +26,36 @@
26
26
* The Hugging Face Hub (https://huggingface.co/timm) is now the primary source for `timm` weights. Model cards include link to papers, original source, license.
27
27
* Previous 0.6.x can be cloned from [0.6.x](https://github.com/rwightman/pytorch-image-models/tree/0.6.x) branch or installed via pip with version.
* Fix Python 3.7 compat, will be dropping support for it soon
32
+
* Other misc fixes
33
+
* Release 0.9.12
34
+
35
+
### Nov 20, 2023
36
+
* Added significant flexibility for Hugging Face Hub based timm models via `model_args` config entry. `model_args` will be passed as kwargs through to models on creation.
37
+
* See example at https://huggingface.co/gaunernst/vit_base_patch16_1024_128.audiomae_as2m_ft_as20k/blob/main/config.json
Copy file name to clipboardExpand all lines: hfdocs/source/models/efficientnet-pruned.mdx
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# EfficientNet (Knapsack Pruned)
2
2
3
-
**EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use $2^N$ times more computational resources, then we can simply increase the network depth by $\alpha ^ N$, width by $\beta ^ N$, and image size by $\gamma ^ N$, where $\alpha, \beta, \gamma$ are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient $\phi$ to uniformly scales network width, depth, and resolution in a principled way.
3
+
**EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use \\( 2^N\\) times more computational resources, then we can simply increase the network depth by \\( \alpha ^ N\\), width by \\( \beta ^ N\\), and image size by \\( \gamma ^ N\\), where \\( \alpha, \beta, \gamma\\) are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient \\( \phi\\) to uniformly scales network width, depth, and resolution in a principled way.
4
4
5
5
The compound scaling method is justified by the intuition that if the input image is bigger, then the network needs more layers to increase the receptive field and more channels to capture more fine-grained patterns on the bigger image.
6
6
@@ -20,7 +20,7 @@ To load a pretrained model:
20
20
21
21
To load and preprocess the image:
22
22
23
-
```py
23
+
```py
24
24
>>>import urllib
25
25
>>>fromPILimport Image
26
26
>>>from timm.data import resolve_data_config
@@ -51,7 +51,7 @@ To get the top-5 predictions class names:
Copy file name to clipboardExpand all lines: hfdocs/source/models/efficientnet.mdx
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# EfficientNet
2
2
3
-
**EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use $2^N$ times more computational resources, then we can simply increase the network depth by $\alpha ^ N$, width by $\beta ^ N$, and image size by $\gamma ^ N$, where $\alpha, \beta, \gamma$ are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient $\phi$ to uniformly scales network width, depth, and resolution in a principled way.
3
+
**EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network width, depth, and resolution with a set of fixed scaling coefficients. For example, if we want to use \\( 2^N\\) times more computational resources, then we can simply increase the network depth by \\( \alpha ^ N\\), width by \\( \beta ^ N\\), and image size by \\( \gamma ^ N\\), where \\( \alpha, \beta, \gamma\\) are constant coefficients determined by a small grid search on the original small model. EfficientNet uses a compound coefficient \\( \phi\\) to uniformly scales network width, depth, and resolution in a principled way.
4
4
5
5
The compound scaling method is justified by the intuition that if the input image is bigger, then the network needs more layers to increase the receptive field and more channels to capture more fine-grained patterns on the bigger image.
6
6
@@ -18,7 +18,7 @@ To load a pretrained model:
18
18
19
19
To load and preprocess the image:
20
20
21
-
```py
21
+
```py
22
22
>>>import urllib
23
23
>>>fromPILimport Image
24
24
>>>from timm.data import resolve_data_config
@@ -49,7 +49,7 @@ To get the top-5 predictions class names:
Copy file name to clipboardExpand all lines: hfdocs/source/models/gloun-resnext.mdx
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,6 +1,6 @@
1
1
# (Gluon) ResNeXt
2
2
3
-
A **ResNeXt** repeats a [building block](https://paperswithcode.com/method/resnext-block) that aggregates a set of transformations with the same topology. Compared to a [ResNet](https://paperswithcode.com/method/resnet), it exposes a new dimension, *cardinality* (the size of the set of transformations) $C$, as an essential factor in addition to the dimensions of depth and width.
3
+
A **ResNeXt** repeats a [building block](https://paperswithcode.com/method/resnext-block) that aggregates a set of transformations with the same topology. Compared to a [ResNet](https://paperswithcode.com/method/resnet), it exposes a new dimension, *cardinality* (the size of the set of transformations) \\( C \\), as an essential factor in addition to the dimensions of depth and width.
4
4
5
5
The weights from this model were ported from [Gluon](https://cv.gluon.ai/model_zoo/classification.html).
6
6
@@ -16,7 +16,7 @@ To load a pretrained model:
16
16
17
17
To load and preprocess the image:
18
18
19
-
```py
19
+
```py
20
20
>>>import urllib
21
21
>>>fromPILimport Image
22
22
>>>from timm.data import resolve_data_config
@@ -47,7 +47,7 @@ To get the top-5 predictions class names:
0 commit comments