torch.optim.NAdamOptions
export interface NAdamOptions {
/** Learning rate (default: 2e-3) */
lr?: number;
/** Coefficients for computing running averages of gradient and its square (default: [0.9, 0.999]) */
betas?: [number, number];
/** Term added to the denominator to improve numerical stability (default: 1e-8) */
eps?: number;
/** Weight decay (L2 penalty) (default: 0) */
weight_decay?: number;
/** Momentum decay (default: 4e-3) */
momentum_decay?: number;
/** Use decoupled weight decay (default: false) */
decoupled_weight_decay?: boolean;
/** Whether to use foreach implementation (default: false) */
foreach?: boolean;
/** Maximize the objective with respect to params (default: false) */
maximize?: boolean;
/** Whether to use a differentiable learning rate (default: false) */
differentiable?: boolean;
}lr(number)optional- – Learning rate (default: 2e-3)
betas([number, number])optional- – Coefficients for computing running averages of gradient and its square (default: [0.9, 0.999])
eps(number)optional- – Term added to the denominator to improve numerical stability (default: 1e-8)
weight_decay(number)optional- – Weight decay (L2 penalty) (default: 0)
momentum_decay(number)optional- – Momentum decay (default: 4e-3)
decoupled_weight_decay(boolean)optional- – Use decoupled weight decay (default: false)
foreach(boolean)optional- – Whether to use foreach implementation (default: false)
maximize(boolean)optional- – Maximize the objective with respect to params (default: false)
differentiable(boolean)optional- – Whether to use a differentiable learning rate (default: false)
Options for NAdam optimizer.