-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can you share your create svdq-int4-flux.1-dev.safetensors script? #10
Comments
Gotcha, it seems that i should use deepcompressor to create w4a4 quantized model |
@chuck-ma was wondering if you managed to load the quantization results from deepcompressor to nunchaku, because I got scaler.pt, smooth.pt, ... as a result and not sure how to convert it to nunchaku format. |
|
@zzz11223345 Here is the response from the devs. mit-han-lab/deepcompressor#27 (comment) |
Hey @ZukaChilachava, I quantized my fine-tuned flux model, but my model.pt (22GB) is bigger than my fine-tuned model (12GB). Is your files just like that or am I missing something. I did the exactly same things in the README.md. Also @lmxyy can you give me your thougths about this? |
Hi, thanks for your work. But I'm wondering how to create an int4 model like svdq-int4-flux.1-dev.safetensors, because I have a model fine-tuned from the dev model.
The text was updated successfully, but these errors were encountered: