We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
https://github.com/NVIDIA/TensorRT-LLM/blob/main/tensorrt_llm/auto_parallel/cluster_info.py
here it handles a list of devices which doesn't include H200, can we add H200 (and potentially GB200) to that list as well?
The text was updated successfully, but these errors were encountered:
What is infer_cluster_key used for
Sorry, something went wrong.
Hi @dongluw thanks for suggestion. We'll add H200 into infer_cluster_key() menthod and this update will be avaiable in next release.
nv-guomingz
No branches or pull requests
https://github.com/NVIDIA/TensorRT-LLM/blob/main/tensorrt_llm/auto_parallel/cluster_info.py
here it handles a list of devices which doesn't include H200, can we add H200 (and potentially GB200) to that list as well?
The text was updated successfully, but these errors were encountered: