Releases: kozistr/pytorch_optimizer
Releases · kozistr/pytorch_optimizer
pytorch-optimizer v2.3.0
Change Log
Feature
- re-implement Shampoo Optimizer (#97, related to #93)
- layer-wise grafting (none, adagrad, sgd)
- block partitioner
- preconditioner
- remove casting to
fp16
orbf16
inside of thestep()
not to lose consistency with the other optimizers. #96 - change some ops to in-place operations to speed up. #96
Fix
- fix
exp_avg_var
whenamsgrad
is True. #96
Refactor
- change linter from
Pylint
toRuff
, #97
pytorch-optimizer v2.2.1
Change Log
Feature
- Support
max_grad_norm
(Adan optimizer) - Support
gradient averaging
(Lamb optimizer) - Support
dampening
,nesterov
parameters (Lars optimizer)
Refactor
- move
step
parameter fromstate
togroup
. (to reduce computation cost & memory) - load
betas
bygroup
, not a parameter. - change to in-place operations.
Fix
- fix when
momentum
is 0 (Lars optimizer)
pytorch-optimizer v2.2.0
Change Log
- Implement GSAM (Surrogate Gap Guided Sharpness-Aware Minimization) optimizer, ICLR 22
pytorch-optimizer v2.1.1
Change Log
Feature
- Support
gradient centralization
forAdai
optimizer - Support
AdamD debias
forAdaPNM
optimizer - Register custom exceptions (e.g. NoSparseGradientError, NoClosureError, ...)
Documentation
- Add API documentation
Bug
- Fix
SAM
optimizer
pytorch-optimizer v2.1.0
pytorch-optimizer v2.0.1
pytorch-optimizer v2.0.0
Chage Log
- Refactor the package depth
- 4 depths
pytorch_optimizer.lr_scheduler
: lr schedulerspytorch_optimizer.optimizer
: optimizerspytorch_optimizer.base
: base utilspytorch_optimizer.experimental
: any experimental features
pytorch_optimizer.adamp
->pytorch_optimizer.optimizer.adamp
- Still
from pytorch_optimizer import AdamP
is possible
- 4 depths
- Implement lr schedulers
-
CosineAnealingWarmupRestarts
-
- Implement (experimental) lr schedulers
-
DeBERTaV3-large
layer-wise lr scheduler
-
Other changes (bug fixes, small refactors)
- Fix
AGC
(to returning the parameter) - Make a room for
experimental features
(atpytorch_optimizer.experimental
) - base types
pytorch-optimizer v1.3.2
pytorch-optimizer v1.3.1
Change Log
-
raw
directive in RST format cannot be used due to the security issue e.g. code injection. #75