-
-
Notifications
You must be signed in to change notification settings - Fork 5.1k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Usage]: Multi-Step Scheduling with Speculative Decoding
usage
How to use vllm
#11917
opened Jan 10, 2025 by
ynwang007
1 task done
[Usage]: Compilation and Execution Issues Across Different GPU Models After Modifying vLLM Source Code
usage
How to use vllm
#11914
opened Jan 10, 2025 by
Yang1032
1 task done
[Bug]: deepseek-v3-bf16 only generates a null char ""!
bug
Something isn't working
#11913
opened Jan 10, 2025 by
janelu9
1 task done
[Performance]: Performance regression for long prompt length since vLLM0.6.4.post1
performance
Performance-related issues
#11912
opened Jan 10, 2025 by
hustxiayang
1 task done
[Bug]: LLAMA3.1 output not matching with HuggingFace when beam search is enabled.
bug
Something isn't working
#11911
opened Jan 10, 2025 by
pratcooper
1 task done
[Bug]: python offline_inference_whisper.py example issue
bug
Something isn't working
#11909
opened Jan 10, 2025 by
silvacarl2
1 task done
[RFC]: Implement Structured Output support for V1 engine
RFC
#11908
opened Jan 9, 2025 by
russellb
1 task done
[Feature]: Support Multiple Tasks Per Model
feature request
#11905
opened Jan 9, 2025 by
FurtherAI
1 task done
[Bug]: example/openai_chat_completion_client_with_tools.py not working
bug
Something isn't working
#11903
opened Jan 9, 2025 by
Hurricane31337
1 task done
[Bug]: Problems with releasing memory after starting the vllm container
bug
Something isn't working
#11902
opened Jan 9, 2025 by
JohnConnor123
1 task done
[Bug]: VLLM get stucks with Qwen VL 7B
bug
Something isn't working
#11899
opened Jan 9, 2025 by
engleccma
1 task done
[Usage]: Use multiprocessing ?
usage
How to use vllm
#11897
opened Jan 9, 2025 by
vinceRV
1 task done
[Performance]: Huge prompts impact other parallel generations
performance
Performance-related issues
#11893
opened Jan 9, 2025 by
iamcristi
1 task done
[New Model]: Support Efficient-Large-Model/NVILA
new model
Requests to new models
#11887
opened Jan 9, 2025 by
z1054136399
1 task done
[Installation]: Could not find a version that satisfies the requirement xgrammar>=0.1.6; platform_machine == "x86_64" (from vllm) (from versions: none)
installation
Installation problems
#11886
opened Jan 9, 2025 by
Xingkangze
1 task done
[Feature]: Support sigmoid for classification models
feature request
#11881
opened Jan 9, 2025 by
tonyay163
1 task done
[Doc]: Add GitHub Action to auto-sync Dockerfile dependency graph
documentation
Improvements or additions to documentation
#11880
opened Jan 9, 2025 by
WineChord
1 task done
[Performance]: Distributed Inference and Serving
performance
Performance-related issues
#11874
opened Jan 9, 2025 by
zlwzlwzlw
1 task done
[Bug]: Engine is gracefully shutting down
bug
Something isn't working
#11873
opened Jan 9, 2025 by
Bryce1010
1 task done
[Usage]: Is there a more refined control method for VLLM caching strategy?
usage
How to use vllm
#11872
opened Jan 9, 2025 by
zhouyuustc
1 task done
[Usage]: 为什么CPU KV cache usage一直为0.0%
usage
How to use vllm
#11871
opened Jan 9, 2025 by
WangJianQ-0118
1 task done
[Feature]: Make
auto
load format handle bitsandbytes models
feature request
#11867
opened Jan 8, 2025 by
alugowski
1 task done
[Bug]: Mistral's Pixtral error for vllm>=0.6.5 on 4 T4's
bug
Something isn't working
#11865
opened Jan 8, 2025 by
jgen1
1 task done
[Usage]: How do I set default temperature for openai compatible server?
usage
How to use vllm
#11861
opened Jan 8, 2025 by
hdnh2006
1 task done
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.