Skip to content

Latest commit

 

History

History
9 lines (5 loc) · 618 Bytes

qa.md

File metadata and controls

9 lines (5 loc) · 618 Bytes

Frequently asked questions

Q1) SophiaH, AdaHessian optimizers give RuntimeError: ~ tensors does not require grad and does not have a grad_fn in compute_hutchinson_hessian().

create_graph must be set True when calling backward(). here's an example.

Q2) Memory leak happens when using SophiaH, AdaHessian optimizers.

torch.autograd.grad with complex gradient flows sometimes leads memory leak issues, and you might encounter OOM issue. related issue