Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open
vLLM's V1 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 9
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Usage]: Multi-Step Scheduling with Speculative Decoding usage How to use vllm
#11917 opened Jan 10, 2025 by ynwang007
1 task done
[Bug]: deepseek-v3-bf16 only generates a null char ""! bug Something isn't working
#11913 opened Jan 10, 2025 by janelu9
1 task done
[Bug]: LLAMA3.1 output not matching with HuggingFace when beam search is enabled. bug Something isn't working
#11911 opened Jan 10, 2025 by pratcooper
1 task done
[Bug]: python offline_inference_whisper.py example issue bug Something isn't working
#11909 opened Jan 10, 2025 by silvacarl2
1 task done
[Bug]: example/openai_chat_completion_client_with_tools.py not working bug Something isn't working
#11903 opened Jan 9, 2025 by Hurricane31337
1 task done
[Bug]: Problems with releasing memory after starting the vllm container bug Something isn't working
#11902 opened Jan 9, 2025 by JohnConnor123
1 task done
[Bug]: VLLM get stucks with Qwen VL 7B bug Something isn't working
#11899 opened Jan 9, 2025 by engleccma
1 task done
[Usage]: Use multiprocessing ? usage How to use vllm
#11897 opened Jan 9, 2025 by vinceRV
1 task done
[Performance]: Huge prompts impact other parallel generations performance Performance-related issues
#11893 opened Jan 9, 2025 by iamcristi
1 task done
[New Model]: Support Efficient-Large-Model/NVILA new model Requests to new models
#11887 opened Jan 9, 2025 by z1054136399
1 task done
[Doc]: Add GitHub Action to auto-sync Dockerfile dependency graph documentation Improvements or additions to documentation
#11880 opened Jan 9, 2025 by WineChord
1 task done
[Performance]: Distributed Inference and Serving performance Performance-related issues
#11874 opened Jan 9, 2025 by zlwzlwzlw
1 task done
[Bug]: Engine is gracefully shutting down bug Something isn't working
#11873 opened Jan 9, 2025 by Bryce1010
1 task done
[Usage]: 为什么CPU KV cache usage一直为0.0% usage How to use vllm
#11871 opened Jan 9, 2025 by WangJianQ-0118
1 task done
[Bug]: Mistral's Pixtral error for vllm>=0.6.5 on 4 T4's bug Something isn't working
#11865 opened Jan 8, 2025 by jgen1
1 task done
[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
36 tasks
ProTip! Exclude everything labeled bug with -label:bug.