2.6 KiB
2.6 KiB
Optimization
The .optimization module provides:
- an optimizer with weight decay fixed that can be used to fine-tuned models, and
- several schedules in the form of schedule objects that inherit from
_LRSchedule: - a gradient accumulation class to accumulate the gradients of multiple batches
AdaFactor
autodoc Adafactor
Schedules
SchedulerType
autodoc SchedulerType
get_scheduler
autodoc get_scheduler
get_constant_schedule
autodoc get_constant_schedule
get_constant_schedule_with_warmup
autodoc get_constant_schedule_with_warmup
get_cosine_schedule_with_warmup
autodoc get_cosine_schedule_with_warmup
get_cosine_with_hard_restarts_schedule_with_warmup
autodoc get_cosine_with_hard_restarts_schedule_with_warmup
get_cosine_with_min_lr_schedule_with_warmup
autodoc get_cosine_with_min_lr_schedule_with_warmup
get_cosine_with_min_lr_schedule_with_warmup_lr_rate
autodoc get_cosine_with_min_lr_schedule_with_warmup_lr_rate
get_linear_schedule_with_warmup
autodoc get_linear_schedule_with_warmup
get_polynomial_decay_schedule_with_warmup
autodoc get_polynomial_decay_schedule_with_warmup
get_inverse_sqrt_schedule
autodoc get_inverse_sqrt_schedule
get_reduce_on_plateau_schedule
autodoc get_reduce_on_plateau_schedule
get_wsd_schedule
autodoc get_wsd_schedule