You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+32-46Lines changed: 32 additions & 46 deletions
Original file line number
Diff line number
Diff line change
@@ -21,12 +21,36 @@ And a big thanks to all GitHub sponsors who helped with some of my costs before
21
21
22
22
## What's New
23
23
24
-
### 🤗 Survey: Feedback Appreciated 🤗
25
-
26
-
For a few months now, `timm` has been part of the Hugging Face ecosystem. Yearly, we survey users of our tools to see what we could do better, what we need to continue doing, or what we need to stop doing.
27
-
28
-
If you have a couple of minutes and want to participate in shaping the future of the ecosystem, please share your thoughts:
* ❗Updates after Oct 10, 2022 are available in 0.8.x pre-releases (`pip install --pre timm`) or cloning main❗
25
+
* Stable releases are 0.6.x and available by normal pip install or clone from [0.6.x](https://github.com/rwightman/pytorch-image-models/tree/0.6.x) branch.
26
+
27
+
### Jan 11, 2023
28
+
* Update ConvNeXt ImageNet-12k pretrain series w/ two new fine-tuned weights (and pre FT `.in12k` tags)
* Cleanup some popular models to better support arg passthrough / merge with model configs, more to go.
38
+
39
+
### Jan 5, 2023
40
+
* ConvNeXt-V2 models and weights added to existing `convnext.py`
41
+
* Paper: [ConvNeXt V2: Co-designing and Scaling ConvNets with Masked Autoencoders](http://arxiv.org/abs/2301.00808)
42
+
* Reference impl: https://github.com/facebookresearch/ConvNeXt-V2 (NOTE: weights currently CC-BY-NC)
43
+
44
+
### Dec 23, 2022 🎄☃
45
+
* Add FlexiViT models and weights from https://github.com/google-research/big_vision (check out paper at https://arxiv.org/abs/2212.08013)
46
+
* NOTE currently resizing is static on model creation, on-the-fly dynamic / train patch size sampling is a WIP
47
+
* Many more models updated to multi-weight and downloadable via HF hub now (convnext, efficientnet, mobilenet, vision_transformer*, beit)
48
+
* More model pretrained tag and adjustments, some model names changed (working on deprecation translations, consider main branch DEV branch right now, use 0.6.x for stable use)
49
+
* More ImageNet-12k (subset of 22k) pretrain models popping up:
* Add 'EVA l' to `vision_transformer.py`, MAE style ViT-L/14 MIM pretrain w/ EVA-CLIP targets, FT on ImageNet-1k (w/ ImageNet-22k intermediate for some)
@@ -325,46 +349,6 @@ More models, more fixes
325
349
* TinyNet models added by [rsomani95](https://github.com/rsomani95)
326
350
* LCNet added via MobileNetV3 architecture
327
351
328
-
### Nov 22, 2021
329
-
* A number of updated weights anew new model defs
330
-
*`eca_halonext26ts` - 79.5 @ 256
331
-
*`resnet50_gn` (new) - 80.1 @ 224, 81.3 @ 288
332
-
*`resnet50` - 80.7 @ 224, 80.9 @ 288 (trained at 176, not replacing current a1 weights as default since these don't scale as well to higher res, [weights](https://github.com/rwightman/pytorch-image-models/releases/download/v0.1-rsb-weights/resnet50_a1h2_176-001a1197.pth))
* Groundwork in for FX feature extraction thanks to [Alexander Soare](https://github.com/alexander-soare)
345
-
* models updated for tracing compatibility (almost full support with some distlled transformer exceptions)
346
-
347
-
### Oct 19, 2021
348
-
* ResNet strikes back (https://arxiv.org/abs/2110.00476) weights added, plus any extra training components used. Model weights and some more details here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-rsb-weights)
349
-
* BCE loss and Repeated Augmentation support for RSB paper
350
-
* 4 series of ResNet based attention model experiments being added (implemented across byobnet.py/byoanet.py). These include all sorts of attention, from channel attn like SE, ECA to 2D QKV self-attention layers such as Halo, Bottlneck, Lambda. Details here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-attn-weights)
351
-
* Working implementations of the following 2D self-attention modules (likely to be differences from paper or eventual official impl):
* A RegNetZ series of models with some attention experiments (being added to). These do not follow the paper (https://arxiv.org/abs/2103.06877) in any way other than block architecture, details of official models are not available. See more here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-attn-weights)
* freeze/unfreeze helpers by [Alexander Soare](https://github.com/alexander-soare)
358
-
359
-
### Aug 18, 2021
360
-
* Optimizer bonanza!
361
-
* Add LAMB and LARS optimizers, incl trust ratio clipping options. Tweaked to work properly in PyTorch XLA (tested on TPUs w/ `timm bits`[branch](https://github.com/rwightman/pytorch-image-models/tree/bits_and_tpu/timm/bits))
362
-
* Add MADGRAD from FB research w/ a few tweaks (decoupled decay option, step handling that works with PyTorch XLA)
363
-
* Some cleanup on all optimizers and factory. No more `.data`, a bit more consistency, unit tests for all!
364
-
* SGDP and AdamP still won't work with PyTorch XLA but others should (have yet to test Adabelief, Adafactor, Adahessian myself).
365
-
* EfficientNet-V2 XL TF ported weights added, but they don't validate well in PyTorch (L is better). The pre-processing for the V2 TF training is a bit diff and the fine-tuned 21k -> 1k weights are very sensitive and less robust than the 1k weights.
366
-
* Added PyTorch trained EfficientNet-V2 'Tiny' w/ GlobalContext attn weights. Only .1-.2 top-1 better than the SE so more of a curiosity for those interested.
367
-
368
352
## Introduction
369
353
370
354
Py**T**orch **Im**age **M**odels (`timm`) is a collection of image models, layers, utilities, optimizers, schedulers, data-loaders / augmentations, and reference training / validation scripts that aim to pull together a wide variety of SOTA models with ability to reproduce ImageNet training results.
@@ -385,6 +369,7 @@ A full version of the list below with source links can be found in the [document
Copy file name to clipboardExpand all lines: docs/archived_changes.md
+40Lines changed: 40 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,45 @@
1
1
# Archived Changes
2
2
3
+
### Nov 22, 2021
4
+
* A number of updated weights anew new model defs
5
+
*`eca_halonext26ts` - 79.5 @ 256
6
+
*`resnet50_gn` (new) - 80.1 @ 224, 81.3 @ 288
7
+
*`resnet50` - 80.7 @ 224, 80.9 @ 288 (trained at 176, not replacing current a1 weights as default since these don't scale as well to higher res, [weights](https://github.com/rwightman/pytorch-image-models/releases/download/v0.1-rsb-weights/resnet50_a1h2_176-001a1197.pth))
* Groundwork in for FX feature extraction thanks to [Alexander Soare](https://github.com/alexander-soare)
20
+
* models updated for tracing compatibility (almost full support with some distlled transformer exceptions)
21
+
22
+
### Oct 19, 2021
23
+
* ResNet strikes back (https://arxiv.org/abs/2110.00476) weights added, plus any extra training components used. Model weights and some more details here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-rsb-weights)
24
+
* BCE loss and Repeated Augmentation support for RSB paper
25
+
* 4 series of ResNet based attention model experiments being added (implemented across byobnet.py/byoanet.py). These include all sorts of attention, from channel attn like SE, ECA to 2D QKV self-attention layers such as Halo, Bottlneck, Lambda. Details here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-attn-weights)
26
+
* Working implementations of the following 2D self-attention modules (likely to be differences from paper or eventual official impl):
* A RegNetZ series of models with some attention experiments (being added to). These do not follow the paper (https://arxiv.org/abs/2103.06877) in any way other than block architecture, details of official models are not available. See more here (https://github.com/rwightman/pytorch-image-models/releases/tag/v0.1-attn-weights)
* freeze/unfreeze helpers by [Alexander Soare](https://github.com/alexander-soare)
33
+
34
+
### Aug 18, 2021
35
+
* Optimizer bonanza!
36
+
* Add LAMB and LARS optimizers, incl trust ratio clipping options. Tweaked to work properly in PyTorch XLA (tested on TPUs w/ `timm bits`[branch](https://github.com/rwightman/pytorch-image-models/tree/bits_and_tpu/timm/bits))
37
+
* Add MADGRAD from FB research w/ a few tweaks (decoupled decay option, step handling that works with PyTorch XLA)
38
+
* Some cleanup on all optimizers and factory. No more `.data`, a bit more consistency, unit tests for all!
39
+
* SGDP and AdamP still won't work with PyTorch XLA but others should (have yet to test Adabelief, Adafactor, Adahessian myself).
40
+
* EfficientNet-V2 XL TF ported weights added, but they don't validate well in PyTorch (L is better). The pre-processing for the V2 TF training is a bit diff and the fine-tuned 21k -> 1k weights are very sensitive and less robust than the 1k weights.
41
+
* Added PyTorch trained EfficientNet-V2 'Tiny' w/ GlobalContext attn weights. Only .1-.2 top-1 better than the SE so more of a curiosity for those interested.
42
+
3
43
### July 12, 2021
4
44
* Add XCiT models from [official facebook impl](https://github.com/facebookresearch/xcit). Contributed by [Alexander Soare](https://github.com/alexander-soare)
0 commit comments