Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix dispatch for quantized model #1855

Merged
merged 1 commit into from
Aug 17, 2023

Conversation

SunMarc
Copy link
Member

@SunMarc SunMarc commented Aug 17, 2023

What does this do ?

This PR fixes the dispatch function for quantized model. Only bnb model needs to have hooks in a single gpu setup. For other quantization methods such as GPTQ, we don't need that.

Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Makes sense, thanks Marc!

@SunMarc SunMarc merged commit 21d1273 into huggingface:main Aug 17, 2023
24 checks passed
@SunMarc SunMarc deleted the fix_dispatch_quantized_model branch August 17, 2023 16:23
@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants