Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[RFC]: Add support for custom ops #156

Open
Yikun opened this issue Feb 25, 2025 · 0 comments
Open

[RFC]: Add support for custom ops #156

Yikun opened this issue Feb 25, 2025 · 0 comments
Assignees

Comments

@Yikun
Copy link
Collaborator

Yikun commented Feb 25, 2025

Motivation.

Currently:

vLLM supports a variety of custom ops by.

vLLM Ascend current (v0.7.1rc1) supports torch native ops (with torch npu), the whole workflow like: vllm --> torch --> torch_npu --> atb ---> cann, but in this way:

  1. the devs should have to first implements the ops in atb
  2. then exposed to torch_npu
  3. upgrade torch_npu to latest version as dependency.
  4. finally, users can use the ops.

The lengthy version matching and upgrade process discourages developers from implementing the Ascend operator.

Proposed Change.

This RFC aims to smooth out the complicated process for ops development and make everything clear and simple. It can also help Ascend developers to create ops with better collaboration.

This RFC is going to start with exploring custom ops support via two ways:

  1. AscendCL (aclnn)
  2. AscendC

We propose to support custom ops via torch bindings to archive this goal.

Work items:

  • Custom Ops framework for vLLM Ascend
  • A real ops implements with CI passed
  • A turtorial to help users understand how to develop the custom ops

Feedback Period.

now - 2025.03.06

CC List.

cc @wangxiyuan
cc @ganyi1996ppo

Any Other Things.

Ready in 2025 Q1 (vLLM Ascend first release v1.7.3)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants