Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Feat (activation_calibration): speed-up by skipping quantization #1029

Merged
merged 8 commits into from
Oct 8, 2024

Conversation

Giuseppe5
Copy link
Collaborator

No description provided.

@Giuseppe5 Giuseppe5 requested review from nickfraser and removed request for nickfraser September 23, 2024 16:50
@Giuseppe5 Giuseppe5 requested review from nickfraser and removed request for nickfraser September 30, 2024 15:33
@Giuseppe5 Giuseppe5 added the next release PRs which should be merged for the next release label Oct 2, 2024
@nickfraser
Copy link
Collaborator

How much speedup does this give for the examples you have tried?

Copy link
Collaborator

@nickfraser nickfraser left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM - Thank you for the extra tests!

@Giuseppe5
Copy link
Collaborator Author

SDXL with minifloat quantization on 500 samples goes from 8h to 2h

@Giuseppe5 Giuseppe5 merged commit 746d97e into Xilinx:dev Oct 8, 2024
23 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
next release PRs which should be merged for the next release
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants