-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Qwen2-VL Model support #1
Comments
It maybe difficult now because trtllm do not support M-ROPE(NVIDIA/TensorRT-LLM#2183). I will follow up continuously. |
Thank you! |
qwen2-vl and m-rope have been supported in the latest tensorrtllm master. Will grps consider supporting it? refer to: |
Implementing the Qwen2-VL processor in C++ is a bit complicated, but I’m trying to support it. |
Master branch have support. Docker image is beta version at now and will update in the futrue. |
Thank you very much, but the image registry.cn-hangzhou.aliyuncs.com/opengrps/grps_gpu:grps1.1.0_cuda12.5_cudnn9.2_trtllm0.16.0_py3.12 cannot be pulled, and an error is reported:
|
Current use image beta version: registry.cn-hangzhou.aliyuncs.com/opengrps/grps_gpu:grps1.1.0_cuda12.5_cudnn9.2_trtllm0.16.0_py3.12_beta Docker image will be update in the future. |
During the execution of the command “grpst archive .”, an error keeps reporting:
Do you have any ideas to solve this problem? The Chinese mirror who replaced Cargo.toml in the tokenizers-cpp project rust folder also reported the same error. |
I haven't meet this. It seems that is your network issue when access crates.io. You can try set network proxy. |
0817b67 Docker image size have been reduced. |
Hello
Will it be possible to include support for Qwen2-VL model? Thank you
The text was updated successfully, but these errors were encountered: