-
-
Notifications
You must be signed in to change notification settings - Fork 899
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
lora example not working with deepspeed zero3 #1481
Comments
I met the same problem. Have you tried any good solutions since then? |
I'm facing this issue now. It has been two months since this was reported. Did anybody find a solution? |
Also running into this |
Ran into this and setting load_in_8bit to false made it work. |
same here. |
Hi this is a known issue where fix is depending on transformers upstream. Currently Zero3 works with 4-bit qlora but not 8-bit lora. See #2068 |
Please check that this issue hasn't been reported before.
Expected Behavior
should be able to do training as usual
Current behaviour
crash with the following error message
Steps to reproduce
run the codellama-7b lora example with deepspeed zero3
Config yaml
Possible solution
No response
Which Operating Systems are you using?
Python Version
3.10
axolotl branch-commit
main/c2b64e4dcff59cfbd754626e5172688433cc13e1
Acknowledgements
The text was updated successfully, but these errors were encountered: