-
-
Notifications
You must be signed in to change notification settings - Fork 898
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
'AdamW' object has no attribute 'optim_bits' #2191
Comments
Could you try with the regular adamw_8bit optimizer please? |
OK I will try with that and get back to you |
@e-p-armstrong @winglian Looks like the issue is with accelerate. I find that downgrading accelerate to version 1.0.1 bypass this error for now. Will follow up on accelerate upstream |
This issue seems to only affect zero2. Zero3 works fine. |
@winglian Reproduced with a different optimizer and it happened even with DPO tuning.
@bursteratom Thanks for searching to find the problem! |
@e-p-armstrong In the meantime I recommend keeping the current accelerate version 1.2.1 while using zero3 instead of zero2 |
@e-p-armstrong @winglian started an upstream PR on accelerate to fix this: huggingface/accelerate#3305 |
Please check that this issue hasn't been reported before.
Expected Behavior
Full-parameter chatml finetuning of Llama 3.1 should work on the main:latest docker image on runpod on 6x A40s with deepspeed
Current behaviour
Training never gets a chance to start:
Stacktrace:
This issue has been around for about a week now? I first reported it on the Discord.
Steps to reproduce
Attempt to full finetune llama 3 using the settings provided (need to add some generic chatml dataset as I had to redact my data files)
Config yaml
The text was updated successfully, but these errors were encountered: