pytorch-optimizer v2.3.0
Change Log
Feature
- re-implement Shampoo Optimizer (#97, related to #93)
- layer-wise grafting (none, adagrad, sgd)
- block partitioner
- preconditioner
- remove casting to
fp16
orbf16
inside of thestep()
not to lose consistency with the other optimizers. #96 - change some ops to in-place operations to speed up. #96
Fix
- fix
exp_avg_var
whenamsgrad
is True. #96
Refactor
- change linter from
Pylint
toRuff
, #97