Skip to content

Commit 17f4dd2

Browse files
committed
Merge branch 'michalwols-docs'
2 parents 31cf125 + 022130d commit 17f4dd2

File tree

5 files changed

+411
-0
lines changed

5 files changed

+411
-0
lines changed

docs/changes.md

Lines changed: 62 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,62 @@
1+
### June 11, 2020
2+
Bunch of changes:
3+
4+
* DenseNet models updated with memory efficient addition from torchvision (fixed a bug), blur pooling and deep stem additions
5+
* VoVNet V1 and V2 models added, 39 V2 variant (ese_vovnet_39b) trained to 79.3 top-1
6+
* Activation factory added along with new activations:
7+
* select act at model creation time for more flexibility in using activations compatible with scripting or tracing (ONNX export)
8+
* hard_mish (experimental) added with memory-efficient grad, along with ME hard_swish
9+
* context mgr for setting exportable/scriptable/no_jit states
10+
* Norm + Activation combo layers added with initial trial support in DenseNet and VoVNet along with impl of EvoNorm and InplaceAbn wrapper that fit the interface
11+
* Torchscript works for all but two of the model types as long as using Pytorch 1.5+, tests added for this
12+
* Some import cleanup and classifier reset changes, all models will have classifier reset to nn.Identity on reset_classifer(0) call
13+
* Prep for 0.1.28 pip release
14+
15+
### May 12, 2020
16+
* Add ResNeSt models (code adapted from https://github.com/zhanghang1989/ResNeSt, paper https://arxiv.org/abs/2004.08955))
17+
18+
### May 3, 2020
19+
* Pruned EfficientNet B1, B2, and B3 (https://arxiv.org/abs/2002.08258) contributed by [Yonathan Aflalo](https://github.com/yoniaflalo)
20+
21+
### May 1, 2020
22+
* Merged a number of execellent contributions in the ResNet model family over the past month
23+
* BlurPool2D and resnetblur models initiated by [Chris Ha](https://github.com/VRandme), I trained resnetblur50 to 79.3.
24+
* TResNet models and SpaceToDepth, AntiAliasDownsampleLayer layers by [mrT23](https://github.com/mrT23)
25+
* ecaresnet (50d, 101d, light) models and two pruned variants using pruning as per (https://arxiv.org/abs/2002.08258) by [Yonathan Aflalo](https://github.com/yoniaflalo)
26+
* 200 pretrained models in total now with updated results csv in results folder
27+
28+
### April 5, 2020
29+
* Add some newly trained MobileNet-V2 models trained with latest h-params, rand augment. They compare quite favourably to EfficientNet-Lite
30+
* 3.5M param MobileNet-V2 100 @ 73%
31+
* 4.5M param MobileNet-V2 110d @ 75%
32+
* 6.1M param MobileNet-V2 140 @ 76.5%
33+
* 5.8M param MobileNet-V2 120d @ 77.3%
34+
35+
### March 18, 2020
36+
* Add EfficientNet-Lite models w/ weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet/lite)
37+
* Add RandAugment trained ResNeXt-50 32x4d weights with 79.8 top-1. Trained by [Andrew Lavin](https://github.com/andravin) (see Training section for hparams)
38+
39+
### Feb 29, 2020
40+
* New MobileNet-V3 Large weights trained from stratch with this code to 75.77% top-1
41+
* IMPORTANT CHANGE - default weight init changed for all MobilenetV3 / EfficientNet / related models
42+
* overall results similar to a bit better training from scratch on a few smaller models tried
43+
* performance early in training seems consistently improved but less difference by end
44+
* set `fix_group_fanout=False` in `_init_weight_goog` fn if you need to reproducte past behaviour
45+
* Experimental LR noise feature added applies a random perturbation to LR each epoch in specified range of training
46+
47+
### Feb 18, 2020
48+
* Big refactor of model layers and addition of several attention mechanisms. Several additions motivated by 'Compounding the Performance Improvements...' (https://arxiv.org/abs/2001.06268):
49+
* Move layer/module impl into `layers` subfolder/module of `models` and organize in a more granular fashion
50+
* ResNet downsample paths now properly support dilation (output stride != 32) for avg_pool ('D' variant) and 3x3 (SENets) networks
51+
* Add Selective Kernel Nets on top of ResNet base, pretrained weights
52+
* skresnet18 - 73% top-1
53+
* skresnet34 - 76.9% top-1
54+
* skresnext50_32x4d (equiv to SKNet50) - 80.2% top-1
55+
* ECA and CECA (circular padding) attention layer contributed by [Chris Ha](https://github.com/VRandme)
56+
* CBAM attention experiment (not the best results so far, may remove)
57+
* Attention factory to allow dynamically selecting one of SE, ECA, CBAM in the `.se` position for all ResNets
58+
* Add DropBlock and DropPath (formerly DropConnect for EfficientNet/MobileNetv3) support to all ResNet variants
59+
* Full dataset results updated that incl NoisyStudent weights and 2 of the 3 SK weights
60+
61+
### Feb 12, 2020
62+
* Add EfficientNet-L2 and B0-B7 NoisyStudent weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet)

docs/index.md

Lines changed: 38 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,38 @@
1+
# Getting Started
2+
3+
4+
## Install
5+
6+
The library can be installed with pip:
7+
8+
```
9+
pip install timm
10+
```
11+
12+
!!! info "Conda Environment"
13+
All development and testing has been done in Conda Python 3 environments
14+
on Linux x86-64 systems, specifically Python 3.6.x and 3.7.x.
15+
16+
To install `timm` in a conda environment:
17+
```
18+
conda create -n torch-env
19+
conda activate torch-env
20+
conda install -c pytorch pytorch torchvision cudatoolkit=10.1
21+
conda install pyyaml
22+
pip install timm
23+
```
24+
25+
26+
## Load Pretrained Model
27+
28+
Pretrained models can be loaded using `timm.create_model`
29+
30+
```python
31+
import timm
32+
33+
m = timm.create_model('mobilenetv3_100', pretrained=True)
34+
m.eval()
35+
```
36+
37+
To load a different model see [the list of pretrained weights](/models
38+
/#pretrained-imagenet-weights).

0 commit comments

Comments
 (0)