Skip to content

Commit 9de2ec5

Browse files
committed
Update README for AGC and bump version to 0.4.4
1 parent 01653db commit 9de2ec5

File tree

2 files changed

+9
-1
lines changed

2 files changed

+9
-1
lines changed

README.md

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,13 @@
22

33
## What's New
44

5+
### Feb 16, 2021
6+
* Add Adaptive Gradient Clipping (AGC) as per https://arxiv.org/abs/2102.06171. Integrated w/ PyTorch gradient clipping via mode arg that defaults to prev 'norm' mode. For backward arg compat, clip-grad arg must be specified to enable when using train.py.
7+
* AGC w/ default clipping factor `--clip-grad .01 --clip-mode agc`
8+
* PyTorch global norm of 1.0 (old behaviour, always norm), `--clip-grad 1.0`
9+
* PyTorch value clipping of 10, `--clip-grad 10. --clip-mode value`
10+
* AGC performance is definitely sensitive to the clipping factor. More experimentation needed to determine good values for smaller batch sizes and optimizers besides those in paper. So far I've found .001-.005 is necessary for stable RMSProp training.
11+
512
### Feb 12, 2021
613
* Update Normalization-Free nets to include new NFNet-F (https://arxiv.org/abs/2102.06171) model defs
714

@@ -238,6 +245,7 @@ Several (less common) features that I often utilize in my projects are included.
238245
* Efficient Channel Attention - ECA (https://arxiv.org/abs/1910.03151)
239246
* Blur Pooling (https://arxiv.org/abs/1904.11486)
240247
* Space-to-Depth by [mrT23](https://github.com/mrT23/TResNet/blob/master/src/models/tresnet/layers/space_to_depth.py) (https://arxiv.org/abs/1801.04590) -- original paper?
248+
* Adaptive Gradient Clipping (https://arxiv.org/abs/2102.06171, https://github.com/deepmind/deepmind-research/tree/master/nfnets)
241249

242250
## Results
243251

timm/version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
__version__ = '0.4.3'
1+
__version__ = '0.4.4'

0 commit comments

Comments
 (0)