Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open 3
vLLM's V1 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 11
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

Memory leak while using tensor_parallel_size>1 bug Something isn't working unstale
#694 opened Aug 8, 2023 by haiasd
Generate nothing from VLLM output bug Something isn't working
#1185 opened Sep 26, 2023 by FocusLiwen
API causes slowdown in batch request handling bug Something isn't working unstale
#1707 opened Nov 17, 2023 by jpeig
anyone can Qwen-14B-Chat-AWQ work with VLLM/TP ? ray anything related with ray unstale
#2419 opened Jan 11, 2024 by s-natsubori
Conda Forge Package keep-open
#3126 opened Feb 29, 2024 by iamthebot
ProTip! Updated in the last three days: updated:>2025-01-24.