Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q2 2024
#3861 opened Apr 4, 2024 by simon-mo
Open 21
v0.4.2 Release Tracker
#4505 by simon-mo was closed May 5, 2024
Closed 11
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Assignee
Filter by who’s assigned
Sort

Issues list

[Bug]: Unable to serve Llama3 using vLLM Docker container bug Something isn't working
#4725 opened May 10, 2024 by vecorro
[Bug]: Not able to do lora inference with phi-3 bug Something isn't working
#4715 opened May 9, 2024 by WeiXiaoSummer
[Doc]: OpenAI Server Command Line Args Broken documentation Improvements or additions to documentation
#4707 opened May 9, 2024 by noamgat
[Bug]: use thread after call multiple times. KeyError: request_id bug Something isn't working
#4706 opened May 9, 2024 by xubzhlin
[Performance]: benchmarking vllm copy kernel and pytorch index copy help wanted Extra attention is needed performance Performance-related issues
#4698 opened May 9, 2024 by youkaichao
2 tasks
[Bug]: batched prefill returning gibberish in some cases. bug Something isn't working
#4697 opened May 8, 2024 by fmmoret
[Bug]: VLLM + tritonserver bug Something isn't working
#4695 opened May 8, 2024 by dlopes78
Installation with CPU with errors installation Installation problems
#4692 opened May 8, 2024 by ming-ddtechcg
[Doc]: API reference for LLM class documentation Improvements or additions to documentation
#4684 opened May 8, 2024 by zplizzi
[Usage]: Get time statistics with each request usage How to use vllm
#4683 opened May 8, 2024 by arunpatala
[Bug]: assert parts[0] == "base_model" AssertionError bug Something isn't working
#4682 opened May 8, 2024 by Edisonwei54
[Usage]: Out of Memory w/ multiple models usage How to use vllm
#4678 opened May 8, 2024 by yudataguy
[Bug]: i cant't use vllm lora with 2gpus, but 1 gpu is ok bug Something isn't working
#4676 opened May 8, 2024 by kynow2
ProTip! Add no:assignee to see everything that’s not assigned.