Skip to content

Commit aa26892

Browse files
committed
Update default, add note about scheduler
1 parent 706ce52 commit aa26892

File tree

1 file changed

+4
-3
lines changed

1 file changed

+4
-3
lines changed

examples/imagenet.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -21,6 +21,7 @@
2121

2222
from utils import *
2323

24+
# install warmup_scheduler from https://github.com/ildoonet/pytorch-gradual-warmup-lr
2425
from warmup_scheduler import GradualWarmupScheduler
2526
import bnn.models.resnet as models
2627
from bnn.engine import BinaryChef
@@ -48,11 +49,11 @@
4849
help='mini-batch size (default: 256), this is the total '
4950
'batch size of all GPUs on the current node when '
5051
'using Data Parallel or Distributed Data Parallel')
51-
parser.add_argument('--lr', '--learning-rate', default=0.1, type=float,
52+
parser.add_argument('--lr', '--learning-rate', default=0.001, type=float,
5253
metavar='LR', help='initial learning rate', dest='lr')
5354
parser.add_argument('--momentum', default=0.9, type=float, metavar='M',
5455
help='momentum')
55-
parser.add_argument('--wd', '--weight-decay', default=1e-4, type=float,
56+
parser.add_argument('--wd', '--weight-decay', default=1e-2, type=float,
5657
metavar='W', help='weight decay (default: 1e-4)',
5758
dest='weight_decay')
5859
parser.add_argument('-p', '--print-freq', default=100, type=int,
@@ -190,7 +191,7 @@ def main_worker(gpu, ngpus_per_node, args):
190191

191192
parameters = model.parameters()
192193
if args.optimizer == 'adamw':
193-
wd = 0.01 if args.step == 0 else 0.01
194+
wd = args.weight_decay if args.step == 0 else 0
194195
optimizer = torch.optim.AdamW(parameters, args.lr, weight_decay=wd)
195196
elif args.optimizer == 'adam':
196197
optimizer = torch.optim.Adam(parameters, args.lr)

0 commit comments

Comments
 (0)