@@ -121,11 +121,12 @@ class AdamW(Optimizer):
121121 >>> beta1 = paddle.to_tensor([0.9], dtype="float32")
122122 >>> beta2 = paddle.to_tensor([0.99], dtype="float32")
123123
124- >>> opt = paddle.optimizer.AdamW(learning_rate=0.1,
125- ... parameters=linear.parameters(),
126- ... beta1=beta1,
127- ... beta2=beta2,
128- ... weight_decay=0.01
124+ >>> opt = paddle.optimizer.AdamW(
125+ ... learning_rate=0.1,
126+ ... parameters=linear.parameters(),
127+ ... beta1=beta1,
128+ ... beta2=beta2,
129+ ... weight_decay=0.01
129130 ... )
130131 >>> loss.backward()
131132 >>> opt.step()
@@ -171,9 +172,9 @@ def __init__(
171172 beta1 : float | Tensor = 0.9 ,
172173 beta2 : float | Tensor = 0.999 ,
173174 epsilon : float | Tensor = 1e-8 ,
174- parameters : Sequence [ Tensor ]
175- | Sequence [_AdamParameterConfig ]
176- | None = None ,
175+ parameters : (
176+ Sequence [ Tensor ] | Sequence [_AdamParameterConfig ] | None
177+ ) = None ,
177178 weight_decay : float | Tensor = 0.01 ,
178179 lr_ratio : Callable [[Tensor ], float ] | None = None ,
179180 apply_decay_param_fun : Callable [[str ], bool ] | None = None ,
@@ -383,9 +384,11 @@ def _add_moments_pows(self, p):
383384 name = self ._beta1_pow_acc_str ,
384385 param = p ,
385386 dtype = acc_dtype ,
386- fill_value = 0.9
387- if isinstance (self ._beta1 , (Variable , Value ))
388- else self ._beta1 ,
387+ fill_value = (
388+ 0.9
389+ if isinstance (self ._beta1 , (Variable , Value ))
390+ else self ._beta1
391+ ),
389392 shape = [1 ],
390393 type = core .VarDesc .VarType .LOD_TENSOR ,
391394 device = 'cpu' ,
@@ -394,9 +397,11 @@ def _add_moments_pows(self, p):
394397 name = self ._beta2_pow_acc_str ,
395398 param = p ,
396399 dtype = acc_dtype ,
397- fill_value = 0.999
398- if isinstance (self ._beta2 , (Variable , Value ))
399- else self ._beta2 ,
400+ fill_value = (
401+ 0.999
402+ if isinstance (self ._beta2 , (Variable , Value ))
403+ else self ._beta2
404+ ),
400405 shape = [1 ],
401406 type = core .VarDesc .VarType .LOD_TENSOR ,
402407 device = 'cpu' ,
@@ -538,9 +543,11 @@ def _append_optimize_op(self, block, param_and_grad):
538543 "multi_precision" : find_master ,
539544 "with_decay" : with_decay ,
540545 "coeff" : self ._weight_decay ,
541- "lr_ratio" : 1.0
542- if self ._lr_ratio is None
543- else self ._lr_ratio (param_and_grad [0 ]),
546+ "lr_ratio" : (
547+ 1.0
548+ if self ._lr_ratio is None
549+ else self ._lr_ratio (param_and_grad [0 ])
550+ ),
544551 }
545552
546553 if isinstance (self ._beta1 , Variable ):
0 commit comments