@@ -61,8 +61,10 @@ def create_optimizer(args, model, filter_bias_and_bn=True):
61
61
opt_split = opt_lower .split ('_' )
62
62
opt_lower = opt_split [- 1 ]
63
63
if opt_lower == 'sgd' or opt_lower == 'nesterov' :
64
+ del opt_args ['eps' ]
64
65
optimizer = optim .SGD (parameters , momentum = args .momentum , nesterov = True , ** opt_args )
65
66
elif opt_lower == 'momentum' :
67
+ del opt_args ['eps' ]
66
68
optimizer = optim .SGD (parameters , momentum = args .momentum , nesterov = False , ** opt_args )
67
69
elif opt_lower == 'adam' :
68
70
optimizer = optim .Adam (parameters , ** opt_args )
@@ -93,8 +95,10 @@ def create_optimizer(args, model, filter_bias_and_bn=True):
93
95
elif opt_lower == 'nvnovograd' :
94
96
optimizer = NvNovoGrad (parameters , ** opt_args )
95
97
elif opt_lower == 'fusedsgd' :
98
+ del opt_args ['eps' ]
96
99
optimizer = FusedSGD (parameters , momentum = args .momentum , nesterov = True , ** opt_args )
97
100
elif opt_lower == 'fusedmomentum' :
101
+ del opt_args ['eps' ]
98
102
optimizer = FusedSGD (parameters , momentum = args .momentum , nesterov = False , ** opt_args )
99
103
elif opt_lower == 'fusedadam' :
100
104
optimizer = FusedAdam (parameters , adam_w_mode = False , ** opt_args )
0 commit comments