-
Notifications
You must be signed in to change notification settings - Fork 46
Issues: NVIDIA/TensorRT-Model-Optimizer
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[ONNX][PTQ] Quantization failed with --dq_only flag in ConvTranspose
#117
opened Dec 19, 2024 by
ry3s
What is difference of torch.quantization and onnx.quantization for speed and accuracy ?
#114
opened Dec 11, 2024 by
demuxin
Is there a plan to support more recent PTQ methods for INT8 ViT?
#113
opened Dec 10, 2024 by
dedoogong
FP16 and FP32 shows 30% lower accuracy compared to INT8 for the ViT Example in ONNX_PTQ
#106
opened Nov 13, 2024 by
chjej202
In cache_diffusion example, can we use dynamic image shape & batch size?
#101
opened Nov 4, 2024 by
wxsms
Previous Next
ProTip!
Follow long discussions with comments:>50.