DistributedAdam¶
- class paddle.fluid.incubate.fleet.parameter_server.pslib.optimizer_factory. DistributedAdam ( optimizer ) [source]
-
adam optimizer in distributed training
-
minimize
(
losses,
startup_program=None,
parameter_list=None,
no_grad_set=None
)
minimize¶
-
- Parameters
-
losses (Variable) – loss variable defined by user
startup_program (Program) – startup program that defined by user
parameter_list (str list) – parameter names defined by users
no_grad_set (set) – a set of variables that is defined by users so that these variables do not need gradient computation
-
minimize
(
losses,
startup_program=None,
parameter_list=None,
no_grad_set=None
)