You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+21Lines changed: 21 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -2,6 +2,20 @@
2
2
3
3
## What's New
4
4
5
+
### Feb 18, 2020
6
+
* Big refactor of model layers and addition of several attention mechanisms. Several additions motivated by 'Compounding the Performance Improvements...' (https://arxiv.org/abs/2001.06268):
7
+
* Move layer/module impl into `layers` subfolder/module of `models` and organize in a more granular fashion
8
+
* ResNet downsample paths now properly support dilation (output stride != 32) for avg_pool ('D' variant) and 3x3 (SENets) networks
9
+
* Add Selective Kernel Nets on top of ResNet base, pretrained weights
10
+
* skresnet18 - 73% top-1
11
+
* skresnet34 - 76.9% top-1
12
+
* skresnext50_32x4d (equiv to SKNet50) - 80.2% top-1
13
+
* ECA and CECA (circular padding) attention layer contributed by [Chris Ha](https://github.com/VRandme)
14
+
* CBAM attention experiment (not the best results so far, may remove)
15
+
* Attention factory to allow dynamically selecting one of SE, ECA, CBAM in the `.se` position for all ResNets
16
+
* Add DropBlock and DropPath (formerly DropConnect for EfficientNet/MobileNetv3) support to all ResNet variants
17
+
* Full dataset results updated that incl NoisyStudent weights and 2 of the 3 SK weights
18
+
5
19
### Feb 12, 2020
6
20
* Add EfficientNet-L2 and B0-B7 NoisyStudent weights ported from [Tensorflow TPU](https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet)
7
21
@@ -86,6 +100,7 @@ Included models:
86
100
* 'Bag of Tricks' / Gluon C, D, E, S variations (https://arxiv.org/abs/1812.01187)
87
101
* Instagram trained / ImageNet tuned ResNeXt101-32x8d to 32x48d from from [facebookresearch](https://pytorch.org/hub/facebookresearch_WSL-Images_resnext/)
@@ -138,6 +153,8 @@ Several (less common) features that I often utilize in my projects are included.
138
153
* AutoAugment (https://arxiv.org/abs/1805.09501) and RandAugment (https://arxiv.org/abs/1909.13719) ImageNet configurations modeled after impl for EfficientNet training (https://github.com/tensorflow/tpu/blob/master/models/official/efficientnet/autoaugment.py)
139
154
* AugMix w/ JSD loss (https://arxiv.org/abs/1912.02781), JSD w/ clean + augmented mixing support works with AutoAugment and RandAugment as well
140
155
* SplitBachNorm - allows splitting batch norm layers between clean and augmented (auxiliary batch norm) data
0 commit comments