Skip to content

Issues: vllm-project/vllm

[Roadmap] vLLM Roadmap Q1 2025
#11862 opened Jan 8, 2025 by simon-mo
Open
vLLM's V1 Engine Architecture
#8779 opened Sep 24, 2024 by simon-mo
Open 10
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

[Usage]: what happens if served lora module is incomptaible with main model? usage How to use vllm
#12106 opened Jan 16, 2025 by yxchng
1 task done
[Performance]: Question about TTFT for ngram speculative decoding performance Performance-related issues
#12101 opened Jan 16, 2025 by ynwang007
1 task done
[Bug]: Corrupted responses for Llama-3.2-3B-Instruct with v0.6.6.post1 bug Something isn't working
#12096 opened Jan 15, 2025 by bsatzger
1 task done
[Feature]: Support Python 3.13 misc
#12083 opened Jan 15, 2025 by manueldeprada
1 task done
[Usage]: Automatic Prefix Cache life cycle usage How to use vllm
#12077 opened Jan 15, 2025 by hyuenmin-choi
1 task done
[Usage]: Will vLLM support LoRA for classification models? usage How to use vllm
#12075 opened Jan 15, 2025 by lullabies777
1 task done
[New Model]: support minimax-01 new model Requests to new models
#12073 opened Jan 15, 2025 by liyawei87
1 task done
[Bug]: Memory profiler does not consider CUDA context memory bug Something isn't working
#12059 opened Jan 14, 2025 by benchislett
1 task done
[Usage]: Running Tensor Parallel on TPUs on Ray Cluster ray anything related with ray usage How to use vllm
#12058 opened Jan 14, 2025 by BabyChouSr
1 task done
[Usage]: Issues related to model meta llama 3.1 70 b instruct usage How to use vllm
#12056 opened Jan 14, 2025 by karimhussain10
1 task done
[Bug]: Drop use of pickle where possible bug Something isn't working
#12055 opened Jan 14, 2025 by russellb
[New Model]: nomic-ai/nomic-embed-text-v1 new model Requests to new models
#12054 opened Jan 14, 2025 by Fmstrat
1 task done
[Usage]: who to run cluster withou docker usage How to use vllm
#12053 opened Jan 14, 2025 by Eutenacity
1 task done
[Bug]: PaliGemma2 not working with OpenAI Docker serve bug Something isn't working
#12052 opened Jan 14, 2025 by IngLP
1 task done
[Usage]: Failed to serve local model in distributed inference usage How to use vllm
#12035 opened Jan 14, 2025 by kerthcet
1 task done
[Bug]: Profiling on vLLM server hangs when --num-scheduler-steps > 1 bug Something isn't working
#12032 opened Jan 14, 2025 by Jacob0226
1 task done
[Bug]: another example of structured output xgrammar does not support bug Something isn't working
#12028 opened Jan 14, 2025 by hustxiayang
1 task done
[Bug]: server crash when glm4-9b-chat got an image request bug Something isn't working
#12024 opened Jan 14, 2025 by liuyanyi
1 task done
ProTip! Adding no:label will show everything without a label.