Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Added lora support proposal #216

Merged
merged 2 commits into from
Mar 4, 2025
Merged

Conversation

wangchen615
Copy link
Contributor

@wangchen615 wangchen615 commented Mar 3, 2025

This is a PR to start a proposal design for LoRA support.

Continue progress toward #205

BEFORE SUBMITTING, PLEASE READ THE CHECKLIST BELOW AND FILL IN THE DESCRIPTION ABOVE


  • Make sure the code changes pass the pre-commit checks.
  • Sign-off your commit by using -s when doing git commit
  • [Feat] LoRA management support.
Proposal to support LoRA Adapter Management for vLLM Production Stack on Kubernetes

[Feature] LoRA Adapter Management Support

  • [Feat] for new features in the cluster (e.g., autoscaling, disaggregated prefill, etc.).
  • [Router] for changes to the vllm_router (e.g., routing algorithm, router observability, etc.).

Note: If the PR spans more than one category, please include all relevant prefixes.

Code Quality

The PR need to meet the following code quality standards:

  • Pass all linter checks. Please use pre-commit to format your code. See README.md for installation.
  • The code need to be well-documented to ensure future contributors can easily understand the code.
  • Please include sufficient tests to ensure the change is stay correct and robust. This includes both unit tests and integration tests.

DCO and Signed-off-by

When contributing changes to this project, you must agree to the DCO. Commits must include a Signed-off-by: header which certifies agreement with the terms of the DCO.

Using -s with git commit will automatically add this header.

What to Expect for the Reviews

We aim to address all PRs in a timely manner. If no one reviews your PR within 5 days, please @-mention one of YuhanLiu11
, Shaoting-Feng or ApostaC.

Copy link
Collaborator

@ApostaC ApostaC left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @wangchen615 ! This proposal looks pretty good!

Got a few questions though:

  1. A clarification question: IIUC, the scheduler is on pod-level (which means different pods in the same deployment may get different LoRA adapters), right?
  2. What should the scheduler do if there are newly added pods (i.e., when autoscale happens)
  3. Would anything special happen when a vLLM pod dies?

Thanks!

@wangchen615
Copy link
Contributor Author

Thanks @wangchen615 ! This proposal looks pretty good!

Got a few questions though:

  1. A clarification question: IIUC, the scheduler is on pod-level (which means different pods in the same deployment may get different LoRA adapters), right?

Yes, the scheduler is supposed to have algorithms at pod level, yet the initial default algorithms would be just register all adapters belonging to the same base model to all replicas of that base model deployment.

  1. What should the scheduler do if there are newly added pods (i.e., when autoscale happens)
    It will trigger the controller to reconcile, namely redo the scheduling of adapters belonging to that base model. The default algorithm will just register all adapters belonging to that base models to the new replica.
  1. Would anything special happen when a vLLM pod dies?
    We need some special mechanism to handle request failures at the router for a short period when the adapter to pod mapping cache has not been updated. Will come up with more designs when this controller and scheduler algorithms are put into work.

Thanks!

@ApostaC
Copy link
Collaborator

ApostaC commented Mar 4, 2025

LGTM! Merging this proposal now, thanks!

@ApostaC ApostaC merged commit c564acc into vllm-project:main Mar 4, 2025
5 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants