-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Compute compatibility? #6
Comments
The model should run on 4 GPUs with ~24GB of memory each. I will change the default batch size in If you haven't tried it yet, I also suggest using sparse attention, as you get some memory usage reduction and speed-up when training the model. |
Thank you so much! I'll give that a try. |
Don't want to keep prodding you, but I ran the provided Sparse Attention installation script:
And received this trace:
I tried installing llvm another way:
This worked, but the subsequent
|
Hmm not too sure what the issue is. Have you tried running For the |
The trace is pretty long, but I think it was this:
Maybe our system has some issue with gcc? I'm not too familiar with this system-level stuff. |
I believe that is essentially the same error that you mentioned above Have you tried looking at some of the github issues on the Deepspeed repo that might be relevant? Such as this one One other option is to try out the Dockerfile in the other VideoGPT related repo |
Would you happen to have a rough estimate of the kind of compute needed to run this model? Unfortunately, we are subject to a very limited compute scenario and I am getting memory allocation errors when trying to run under the default settings.
Thank you for any support.
The text was updated successfully, but these errors were encountered: