pytorch-optimizer v3.1.0
Change Log
Feature
- Implement
AdaLomo
optimizer. (#258) - Support
Q-GaLore
optimizer. (#258)- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.
- you can use by
optimizer = load_optimizer('q_galore_adamw8bit')
- Support more bnb optimizers. (#258)
bnb_paged_adam8bit
,bnb_paged_adamw8bit
,bnb_*_*32bit
.
- Improve
power_iteration()
speed up to 40%. (#259) - Improve
reg_noise()
(E-MCMC) speed up to 120%. (#260) - Support
disable_lr_scheduler
parameter forRanger21
optimizer to disable built-in learning rate scheduler. (#261)
Refactor
- Refactor
AdamMini
optimizer. (#258) - Deprecate optional dependency,
bitsandbytes
. (#258) - Move
get_rms
,approximate_sq_grad
functions toBaseOptimizer
for reusability. (#258) - Refactor
shampoo_utils.py
. (#259) - Add
debias
,debias_adam
methods inBaseOptimizer
. (#261) - Refactor to use
BaseOptimizer
only, not inherit multiple classes. (#261)
Bug
- Fix several bugs in
AdamMini
optimizer. (#257)
Contributions
thanks to @sdbds