-
-
Notifications
You must be signed in to change notification settings - Fork 5.1k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[New Model]: Cosmos-1.0-Autoregressive (World Foundation Models)
new model
Requests to new models
#11968
opened Jan 12, 2025 by
Haoxiang-Wang
1 task done
[Doc]: Invalid JSON examples in Engine Args Document
documentation
Improvements or additions to documentation
#11965
opened Jan 12, 2025 by
ardapekis
1 task done
[Feature]: 【support tts llm like cosyvoice2.0】
feature request
#11964
opened Jan 12, 2025 by
wang-TJ-20
1 task done
[Bug]: failure when compiling httptools
bug
Something isn't working
#11961
opened Jan 11, 2025 by
gnusupport
1 task done
[Usage]: How to reach 100% GPU Compute Utilization ?
usage
How to use vllm
#11959
opened Jan 11, 2025 by
MohamedAliRashad
1 task done
[Bug]: How to run LanguageBind/Video-LLaVA-7B-hf
bug
Something isn't working
#11954
opened Jan 11, 2025 by
jianghuyihei
1 task done
[Bug]: The random seed behavior when loading a model in vLLM is confusing.
bug
Something isn't working
#11953
opened Jan 11, 2025 by
Aratako
1 task done
[Feature]: Support for serving two llms at the same time
feature request
#11950
opened Jan 11, 2025 by
czg1225
1 task done
[Feature] [Spec Decode]: Simplify the Use of Eagle Spec Decode
feature request
#11943
opened Jan 10, 2025 by
LiuXiaoxuanPKU
1 task done
[Bug]: Loading model from S3 using RunAI Model Streamer excludes too many files
bug
Something isn't working
#11929
opened Jan 10, 2025 by
svantesorberg
1 task done
[Feature]: Support out-of-tree register of customize quantization method
feature request
#11926
opened Jan 10, 2025 by
ice-tong
1 task done
[Usage]: Multi-Step Scheduling with Speculative Decoding
usage
How to use vllm
#11917
opened Jan 10, 2025 by
ynwang007
1 task done
[Usage]: Compilation and Execution Issues Across Different GPU Models After Modifying vLLM Source Code
usage
How to use vllm
#11914
opened Jan 10, 2025 by
Yang1032
1 task done
[Bug]: deepseek-v3-bf16 only generates a null char ""!
bug
Something isn't working
#11913
opened Jan 10, 2025 by
janelu9
1 task done
[Performance]: Performance regression for long prompt length since vLLM0.6.4.post1
performance
Performance-related issues
#11912
opened Jan 10, 2025 by
hustxiayang
1 task done
[Bug]: LLAMA3.1 output not matching with HuggingFace when beam search is enabled.
bug
Something isn't working
#11911
opened Jan 10, 2025 by
pratcooper
1 task done
[Bug]: python offline_inference_whisper.py example issue
bug
Something isn't working
#11909
opened Jan 10, 2025 by
silvacarl2
1 task done
[RFC]: Implement Structured Output support for V1 engine
RFC
#11908
opened Jan 9, 2025 by
russellb
1 task done
[Feature]: Support Multiple Tasks Per Model
feature request
#11905
opened Jan 9, 2025 by
FurtherAI
1 task done
[Bug]: example/openai_chat_completion_client_with_tools.py not working
bug
Something isn't working
#11903
opened Jan 9, 2025 by
Hurricane31337
1 task done
[Bug]: Problems with releasing memory after starting the vllm container
bug
Something isn't working
#11902
opened Jan 9, 2025 by
JohnConnor123
1 task done
[Bug]: VLLM get stucks with Qwen VL 7B
bug
Something isn't working
#11899
opened Jan 9, 2025 by
engleccma
1 task done
[Usage]: Use multiprocessing ?
usage
How to use vllm
#11897
opened Jan 9, 2025 by
vinceRV
1 task done
[Performance]: Huge prompts impact other parallel generations
performance
Performance-related issues
#11893
opened Jan 9, 2025 by
iamcristi
1 task done
Previous Next
ProTip!
Adding no:label will show everything without a label.