-
-
Notifications
You must be signed in to change notification settings - Fork 6.6k
Issues: vllm-project/vllm
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[RFC]: Data Parallel Attention and Expert Parallel MoEs
RFC
#16037
opened Apr 3, 2025 by
tlrmchlsmth
7 of 13 tasks
[Bug]: xgrammar missing file crashes the server
bug
Something isn't working
#16030
opened Apr 3, 2025 by
servient-ashwin
1 task done
[Feature]: Adding tool_choice: required for lm-format-enforcer
feature request
New feature or request
#16029
opened Apr 3, 2025 by
ItzAmirreza
1 task done
[Bug]: Two beginning of sequence tokens for Lllama-3.2-3B-Instruct
bug
Something isn't working
#16028
opened Apr 3, 2025 by
Naqu6
1 task done
[Bug]: Unable to run Phi4 with tensor-parallel-size 4 torch.compile compatiblity
bug
Something isn't working
#16021
opened Apr 3, 2025 by
roguetech
1 task done
[New Model]: support for fashion-clip
new model
Requests to new models
#16019
opened Apr 3, 2025 by
priyankaiiit14
1 task done
[RFC]: Cache Salting for Secure and Flexible Prefix Caching in vLLM
RFC
#16016
opened Apr 3, 2025 by
dr75
1 task done
[Bug]: Null response for Mistral3.1
bug
Something isn't working
#16014
opened Apr 3, 2025 by
hahmad2008
1 task done
[Bug]: Cannot use FlashAttention-2 backend because the vllm.vllm_flash_attn package is not found. Make sure that vllm_flash_attn was built and installed (on by default).
bug
Something isn't working
#16013
opened Apr 3, 2025 by
GGBond8488
1 task done
[RFC]: Is
huggingface-cli[hf_xet]
needed for vllm build?
RFC
#16012
opened Apr 3, 2025 by
Shafi-Hussain
1 task done
[Usage]: Performance Comparison: 1x8 (TP=8) vs 2x4 (TP=4) in vLLM - Why Does 1x8 Outperform 2x4 in Concurrency?
usage
How to use vllm
#16011
opened Apr 3, 2025 by
hwb96
1 task done
[Bug]: TypeError: __init__() missing 1 required positional argument: 'inner_exception'
bug
Something isn't working
torch.compile
#16009
opened Apr 3, 2025 by
Satonio1
1 task done
[Bug]: Tool call auto not working with Qwen models in v0.8.2
bug
Something isn't working
#16008
opened Apr 3, 2025 by
pivotal-marcela-campo
1 task done
[Bug]: crash during debug, works ok running cli
bug
Something isn't working
torch.compile
#16006
opened Apr 3, 2025 by
CharlesJu1
1 task done
[Bug] [Misc]: test_sharded_state_loader run failed
bug
Something isn't working
#16004
opened Apr 3, 2025 by
Accelerator1996
1 task done
[Usage]: Is it possible to run vLLM inside a Jupyter Notebook?
usage
How to use vllm
#16003
opened Apr 3, 2025 by
repodiac
1 task done
[Bug]: The reasoning_parser doesn't dynamically apply to parser.add_argument after I added a custom Reasoning Parser
bug
Something isn't working
#15999
opened Apr 3, 2025 by
amazingkmy
1 task done
[Bug]: creating helm chart with vllm image and downloaded model throws raise KeyboardInterrupt("terminated") KeyboardInterrupt: terminated error
bug
Something isn't working
#15997
opened Apr 3, 2025 by
Chennakesavulu5
1 task done
[Bug]: System gets stuck during pressure test
bug
Something isn't working
#15996
opened Apr 3, 2025 by
oldcpple
1 task done
[Bug]: Fail to load LLM either using offline inference or vllm serve
bug
Something isn't working
#15992
opened Apr 3, 2025 by
oximi123
1 task done
[Bug]: amd gpu Multiple graphics cards cannot function properly
bug
Something isn't working
#15991
opened Apr 3, 2025 by
taikai-zz
1 task
[Installation]: vllm 0.8.0 installation from source failed
installation
Installation problems
#15989
opened Apr 3, 2025 by
Sabiha1225
1 task done
[Bug]: EADDRINUSE (-98) error when setting up NCCL communicator
bug
Something isn't working
#15987
opened Apr 3, 2025 by
jglaser
1 task done
[Bug]: can not use uv run or uv run python on mac series
bug
Something isn't working
#15985
opened Apr 3, 2025 by
yihong0618
1 task done
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.