Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[LoRA] Add LoRA support for Qwen3OmniMoeThinkerForConditionalGeneration qwen Related to Qwen models
#37193 opened Mar 16, 2026 by pratapyash Loading…
4 tasks done
WIP: [Feature] KVCACHE NVFP4 v1
#37192 opened Mar 16, 2026 by JartX Loading…
[Performance] Enable Triton autotuning disk cache by default
#37188 opened Mar 16, 2026 by arpera Loading…
2 of 5 tasks
Add ability to replace oot ops when using lora
#37181 opened Mar 16, 2026 by kyuyeunk Loading…
5 tasks
Bugfix for offloading+prefetch for GLM-4.7-FP8 bug Something isn't working
#37178 opened Mar 16, 2026 by sfbemerk Loading…
Fix KV cache size estimation regression in v0.17+ v1
#37172 opened Mar 16, 2026 by xueliangyang-oeuler Loading…
5 tasks
[Bugfix] Fix "Already borrowed" tokenizer race in Hermes tool parser bug Something isn't working
#37169 opened Mar 16, 2026 by stonelazy Loading…
[Bugfix] Fix mock.patch resolution failure for standalone_compile.FakeTensorMode on Python <= 3.10 bug Something isn't working
#37158 opened Mar 16, 2026 by dbari Loading…
3 of 5 tasks
[openapi] remove redundant exception stack trace[4/N] frontend
#37157 opened Mar 16, 2026 by andyxning Loading…
5 tasks
Fix issue #37037 v1
#37156 opened Mar 16, 2026 by xueliangyang-oeuler Loading…
5 tasks
ProTip! Find all pull requests that aren't related to any open issues with -linked:issue.