Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open 6
[V1] Feedback Thread
#12568 opened Jan 30, 2025 by simon-mo
Open 49
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: ChatCompletionRequest rejects its own defaults bug Something isn't working
#14351 opened Mar 6, 2025 by schoennenbeck
1 task done
[Bug]: vllm cannot connect to an external ray cluster bug Something isn't working
#14349 opened Mar 6, 2025 by dotbalo
1 task done
[Feature]: Ovis2 VLM series feature request New feature or request
#14346 opened Mar 6, 2025 by Strand2013
1 task done
[Doc]: How can I set the date_string for the chat templates documentation Improvements or additions to documentation
#14344 opened Mar 6, 2025 by FelixNeutatzMainWebSolutions
1 task done
[Bug]: online-rl sampling is different from offline-sampling bug Something isn't working
#14341 opened Mar 6, 2025 by yyht
1 task done
[Feature]: eagle支持多模态模型 feature request New feature or request
#14337 opened Mar 6, 2025 by perdone
1 task done
[Feature]: reasoning_tokens in Chat Completion Response usage feature request New feature or request
#14335 opened Mar 6, 2025 by gaocegege
1 task done
[Bug]: GPU索引指定失败 bug Something isn't working
#14334 opened Mar 6, 2025 by hackerhaiJu
1 task done
[Bug]: vLLM returning 415 status code at high load bug Something isn't working
#14333 opened Mar 6, 2025 by chiragjn
1 task done
[Bug]: opentelemetry POC vLLM span cannot be concatenated with HTTP spans. bug Something isn't working
#14330 opened Mar 6, 2025 by RichardoMrMu
1 task done
[Usage]: How do I set the input image size when using qwen2-vl? usage How to use vllm
#14325 opened Mar 6, 2025 by glamourzc
1 task done
[Doc]: Why is max block_size on CUDA 32? documentation Improvements or additions to documentation
#14319 opened Mar 5, 2025 by ptarasiewiczNV
1 task done
Issue with Mistral Small and greek characters usage How to use vllm
#14307 opened Mar 5, 2025 by myrulezzz
1 task done
[New Model]: llava-onevision-qwen2-72b-ov-sft new model Requests to new models
#14290 opened Mar 5, 2025 by Song-WenPo
1 task done
[Feature]: Chat inputs to AsyncLLMEngine feature request New feature or request
#14289 opened Mar 5, 2025 by sfc-gh-mkrubinski
1 task done
ProTip! Updated in the last three days: updated:>2025-03-03.