Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Request for Ascend NPU support #6368

Closed
xuedinge233 opened this issue Jul 12, 2024 · 14 comments
Closed

[Feature]: Request for Ascend NPU support #6368

xuedinge233 opened this issue Jul 12, 2024 · 14 comments

Comments

@xuedinge233
Copy link

🚀 The feature, motivation and pitch

Background

Currently, the project supports various hardware accelerators such as GPUs, but there is no support for NPUs. Adding NPU support could significationly benefit users who have access to these devices, enabling faster and more efficient computations.

Reference Materials

Ascend is a full-stack AI computing infrastructure for industry applications and services based on Huawei Ascend processors and software. For more information about Ascend, see Ascend Community.

CANN (Compute Architecture of Neural Networks), developped by Huawei, is a heterogeneous computing architecture for AI.

Pytorch has officially announced support for Ascend NPU (through key PrivateUse1), please see the PrivateUse1 tutorial here.

Specific Request

we would like to request the addition of support for NPUs within the project. In order to achieve this goal, we will contributing code and providing feedback. This request may additional resources and effort on your part, but we hope to get your help if possible.

Alternatives

No response

Additional context

No response

@mgoin
Copy link
Member

mgoin commented Jul 12, 2024

Duplicate of #1606 and #6066

@xuedinge233
Copy link
Author

Duplicate of #1606 and #6066

Thinks for your reply, we are ready to develop Ascend ADAPTS for the vllm-project, after completion, we will provide you with a public Ascend prototype

@mgoin
Copy link
Member

mgoin commented Jul 15, 2024

That's great to hear @xuedinge233! Having hardware instances available for CI would be great for maintaining support

@xuedinge233
Copy link
Author

Yeah, we‘re working on it, and we will upload it as soon as there is progress in the development

@BrightXiaoHan
Copy link

Yeah, we‘re working on it, and we will upload it as soon as there is progress in the development

Any update about this?

@xuedinge233
Copy link
Author

xuedinge233 commented Aug 1, 2024

Yeah, we‘re working on it, and we will upload it as soon as there is progress in the development

Any update about this?

We are making steady progress and expect to deliver the first version of the code in a couple of weeks.

@seoibiubiu
Copy link

Is it planned to support NPU+CANN environments where the CPU is the aarch64 architecture?

@dogeeelin
Copy link

Looking forward to NPU support on vLLM!

@MengqingCao
Copy link
Contributor

Is it planned to support NPU+CANN environments where the CPU is the aarch64 architecture?

Yes, NPU+MindIE+CANN is in plan, and both aarch64 and x86_64 will be supported.

@ccly1996
Copy link

ccly1996 commented Aug 26, 2024

Yeah, we‘re working on it, and we will upload it as soon as there is progress in the development

Any update about this?

We are making steady progress and expect to deliver the first version of the code in a couple of weeks.

Excuse me, is there any progress?

@seoibiubiu
Copy link

Is it planned to support NPU+CANN environments where the CPU is the aarch64 architecture?

Yes, NPU+MindIE+CANN is in plan, and both aarch64 and x86_64 will be supported.

I mean NPU+VLLM+CANN, and based on aarch64 CPU architecture. NPU device is Atlas 300I Duo or others.

@shilei4260
Copy link

@xuedinge233
How long will it take to release the vllm version of npu

@xuedinge233
Copy link
Author

@xuedinge233 How long will it take to release the vllm version of npu

The latest updates on the progress can be found here #8054

Copy link

github-actions bot commented Dec 1, 2024

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

8 participants