-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can we use multi GPU while exporting (diffusers ) onnx model? #96
Comments
@wxsma backbone.eval()
with torch.no_grad():
modelopt_export_sd(backbone, f"{str(args.onnx_dir)}", args.model, args.format) And also move other parts to cpu, like vae and clip. Please let me know if it works. |
Sadly it does not work. I manage to export the onnx model in A800 and complie in 4090. |
I'll take a look and get back to you, I barely tested on 4090. Just to confirm, can you export the FP16 SDXL on a 4090? |
Thank you, I will try it later. |
Has there been any progress on this issue? I encountered the same problem on an RTX 4090. Eventually, I performed the ONNX model conversion on an A800. Using nvidia-smi, I noticed that the ONNX conversion process requires around 30GB of VRAM model:SDXL-1.0 |
I'm building a SDXL model in float16 using 4090x2, therefore the GPU memory available is ~48GB.
however, the script in
diffusers/quantizatoin
does not looks like to able to use both of them, and raise OOM error while exporting onnx model.I tried to export the model using CPU, but it's too slow.
The text was updated successfully, but these errors were encountered: