-
Notifications
You must be signed in to change notification settings - Fork 145
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
don't check max_batch_size for cpu #298
Conversation
for cpu models, we don't expect users to execute params.try_use_cuda_graph_with_max_batch_size(1). I think we can add a check to conditionally pass try_use_cuda_graph...() in the model-qa.py |
if user called it, it should not crash. I think we can add a warning later instead of quit. |
It would still crash (throw an exception) since you're checking for cuda graph enabled in cpu ep, no? |
It will still crash if user set to enable cuda_graph for cpu |
How is that any different from what it was earlier? It appears that max_batch_size_ is only set when try_use_cuda_graph_with_max_batch_size is called. |
It will not crash with this PR if user call try_use_cuda_graph_with_max_batch_size for CPU device |
Sorry, I didn't check the definition of |
No description provided.