optimizer.schedule_free.AdamWScheduleFree
optimizer.schedule_free.AdamWScheduleFree(
params,
lr=0.0025,
betas=(0.9, 0.999),
eps=1e-08,
weight_decay=0,
warmup_steps=0,
r=0.0,
weight_lr_power=2.0,
)Schedule-Free AdamW in PyTorch.
As the name suggests, no scheduler is needed with this optimizer. To add warmup, rather than using a learning rate schedule you can just set the warmup_steps parameter.
This optimizer requires that .train() and .eval() be called on the optimizer before the beginning of training and evaluation respectively.
Reference: https://github.com/facebookresearch/schedule_free
Methods
| Name | Description |
|---|---|
| eval | Switch model parameters to evaluation mode (using averaged parameters ‘x’). |
| step | Performs a single optimization step. |
| train | Switch model parameters to training mode (using current iterate ‘y’). |
eval
optimizer.schedule_free.AdamWScheduleFree.eval()Switch model parameters to evaluation mode (using averaged parameters ‘x’).
step
optimizer.schedule_free.AdamWScheduleFree.step(closure=None)Performs a single optimization step.
train
optimizer.schedule_free.AdamWScheduleFree.train()Switch model parameters to training mode (using current iterate ‘y’).