-
Notifications
You must be signed in to change notification settings - Fork 66
Issues: vllm-project/production-stack
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
feature: introduce pyproject.toml and use uv
feature request
New feature or request
#184
opened Feb 25, 2025 by
bufferoverflow
feature: unify naming of production-stack, vllm-stack and vllm-router
discussion
feature request
New feature or request
#178
opened Feb 25, 2025 by
bufferoverflow
bug: Metrics do not update when no new queries are routed to the endpoint
bug
Something isn't working
#173
opened Feb 24, 2025 by
xqe2011
feature: Terraform Quickstart Tutorials for Google GKE
feature request
New feature or request
#172
opened Feb 23, 2025 by
falconlee236
feature: Terraform Quickstart Tutorials for Underlying Infrastructure
feature request
New feature or request
#167
opened Feb 21, 2025 by
0xThresh
Discussion - QPS routing when there are multiple router replicas
discussion
question
Further information is requested
#166
opened Feb 21, 2025 by
aishwaryaraimule21
Internal registry
feature request
New feature or request
#165
opened Feb 20, 2025 by
tylerlittlefield
bug: flaky test case Functionality test for helm chart / Multiple-Models
bug
Something isn't working
#152
opened Feb 19, 2025 by
gaocegege
bug: Model not found when enable vllm api key
bug
Something isn't working
#150
opened Feb 18, 2025 by
JustinDuy
Discussion: Unifying versions for helm and router
question
Further information is requested
#80
opened Feb 7, 2025 by
gaocegege
Feat: Router observability (Current QPS, router-side queueing delay, etc)
feature request
New feature or request
#78
opened Feb 7, 2025 by
sitloboi2012
feat: Distributed tracing for router
feature request
New feature or request
help wanted
Extra attention is needed
#77
opened Feb 7, 2025 by
gaocegege
feat: Allow remote backend configuration
feature request
New feature or request
#75
opened Feb 7, 2025 by
askulkarni2
Download model from Object Storage
feature request
New feature or request
#69
opened Feb 6, 2025 by
nitin302
Why Hugging Face Token?
question
Further information is requested
#67
opened Feb 6, 2025 by
nitin302
Create an Example Building Ingress for Router Service
documentation
Improvements or additions to documentation
#60
opened Feb 4, 2025 by
0xThresh
Helm Chart Lacks Clear Support for Multi-Node vLLM Deployment
help wanted
Extra attention is needed
#50
opened Jan 31, 2025 by
shohamyamin
feat: Offline batched inference based on OpenAI offline batching API
feature request
New feature or request
#47
opened Jan 31, 2025 by
gaocegege
Fix Installation Steps in vLLM Production Stack Tutorials
documentation
Improvements or additions to documentation
#37
opened Jan 29, 2025 by
dxu104
Pluggable Router Implementations
feature request
New feature or request
#36
opened Jan 29, 2025 by
danehans
[Roadmap] vLLM production stack roadmap for 2025 Q1
#26
opened Jan 27, 2025 by
ApostaC
7 of 15 tasks
ProTip!
Adding no:label will show everything without a label.