Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Add multi-LoRA support for Whisper models
#28528 opened Nov 12, 2025 by daje0601 Loading…
5 tasks
[Bugfix] Missing cached item in the MultiModalReceiverCache multi-modality Related to multi-modality (#4194)
#28525 opened Nov 12, 2025 by knlnguyen1802 Loading…
5 tasks
Fix Llama4 Pipeline Parallelism llama Related to Llama models
#28524 opened Nov 12, 2025 by River12 Loading…
debug tests/tool_use/test_tool_calls.py failure on AMD ci/build documentation Improvements or additions to documentation frontend performance Performance-related issues rocm Related to AMD ROCm tool-calling
#28523 opened Nov 12, 2025 by zhewenl Draft
5 tasks
[feat]: log number of preempted requests v1
#28522 opened Nov 12, 2025 by 610lyn Loading…
[XPU]Fix crash due to removed VLLM_USE_V1 attribute
#28520 opened Nov 12, 2025 by chaojun-zhang Loading…
5 tasks
[BugFix] Ensure EngineArgs.create_engine_config is idempotent bug Something isn't working ready ONLY add when PR is ready to merge/full CI is needed
#28515 opened Nov 12, 2025 by njhill Loading…
Work in Progress
#28513 opened Nov 12, 2025 by xujinheng Draft
5 tasks
[Benchmark] Fix client seed synchronization in multi-turn benchmark performance Performance-related issues
#28512 opened Nov 12, 2025 by ai-jz Loading…
[XPU] Support Triton path for LoRA operations on XPU ready ONLY add when PR is ready to merge/full CI is needed
#28511 opened Nov 12, 2025 by faaany Loading…
3 of 4 tasks
[quantization][config] enable override existing quant_config
#28510 opened Nov 12, 2025 by ILikeIneine Loading…
5 tasks
Enable PP on tpu_inference v1
#28506 opened Nov 12, 2025 by Chenyaaang Draft
5 tasks
clean up activation hardcoding to silu
#28503 opened Nov 12, 2025 by cyrusd98 Loading…
[ROCm] [Bugfix] Fix fused_qknorm_rope_kernel rocm compatibility ready ONLY add when PR is ready to merge/full CI is needed rocm Related to AMD ROCm
#28500 opened Nov 11, 2025 by tjtanaa Loading…
5 tasks
[Kernel] Unified attention kernel performance tuning
#28497 opened Nov 11, 2025 by cagrikymk Loading…
[Bugfix] Prevent premature multimodal preprocess cache eviction multi-modality Related to multi-modality (#4194)
#28496 opened Nov 11, 2025 by aykoppol Loading…
update to 2.9.1 PyTorch release ci/build nvidia rocm Related to AMD ROCm
#28495 opened Nov 11, 2025 by atalman Loading…
[Performance][Hopper] Avoid M dim padding to 4x for most cases (due to cuda graphs paddings) deepseek Related to DeepSeek models nvidia ready ONLY add when PR is ready to merge/full CI is needed
#28492 opened Nov 11, 2025 by alexm-redhat Loading…
ProTip! no:milestone will show everything without a milestone.