Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Update serial_utils.py v1
#20379 opened Jul 2, 2025 by jue-cmd Loading…
4 tasks
[CI/Build] Fix torch nightly CI dependencies part 3 ci/build ready ONLY add when PR is ready to merge/full CI is needed
#20378 opened Jul 2, 2025 by zou3519 Loading…
3 of 4 tasks
[Docs] Update EAGLE example documentation Improvements or additions to documentation ready ONLY add when PR is ready to merge/full CI is needed
#20375 opened Jul 2, 2025 by NickLucche Loading…
[V1] feat:add engine v1 tracing v1
#20372 opened Jul 2, 2025 by RichardoMrMu Loading…
[Misc] add handler HF_TOKEN is emptry string ready ONLY add when PR is ready to merge/full CI is needed
#20369 opened Jul 2, 2025 by lengrongfu Loading…
1 of 4 tasks
[Structured Outputs][V1] Skipping with models doesn't contain tokenizers ready ONLY add when PR is ready to merge/full CI is needed structured-output v1
#20365 opened Jul 2, 2025 by aarnphm Loading…
[Bugfix] Fix flaky test_streaming_response test ready ONLY add when PR is ready to merge/full CI is needed
#20363 opened Jul 2, 2025 by NickLucche Loading…
[PP][V1]: Integrate Token Throttling into vLLM v1
#20359 opened Jul 2, 2025 by gty111 Loading…
4 tasks done
[WIP][RC] Update PyTorch to 2.8.0 ci/build rocm Related to AMD ROCm
#20358 opened Jul 2, 2025 by huydhn Draft
[Core] Move multimodal placeholder from chat utils to model definition deepseek Related to DeepSeek models documentation Improvements or additions to documentation frontend llama Related to Llama models qwen Related to Qwen models
#20355 opened Jul 2, 2025 by DarkLight1337 Loading…
2 of 4 tasks
[DP] Copy environment variables to Ray DPEngineCoreActors v1
#20344 opened Jul 2, 2025 by ruisearch42 Loading…
3 of 4 tasks
[TPU][Bugfix] fix the MoE OOM issue
#20339 opened Jul 1, 2025 by yaochengji Loading…
Add benchmark dataset for mlperf llama tasks llama Related to Llama models perf-benchmarks performance Performance-related issues
#20338 opened Jul 1, 2025 by mgoin Loading…
Change default model to Qwen3-0.6B qwen Related to Qwen models ready ONLY add when PR is ready to merge/full CI is needed
#20335 opened Jul 1, 2025 by tlrmchlsmth Loading…
[Perf] Optimize Vectorization Utils for Int 8 Quantization Kernels performance Performance-related issues ready ONLY add when PR is ready to merge/full CI is needed
#20331 opened Jul 1, 2025 by yewentao256 Loading…
[ROCm] warpSize is being made non constexpr in ROCm 7.0 rocm Related to AMD ROCm
#20330 opened Jul 1, 2025 by gshtras Loading…
Add MoE config files for Nvidia Pro 6000 Blackwell Workstation Edition
#20329 opened Jul 1, 2025 by Chen-zexi Loading…
1 of 4 tasks
[USAGE] Improve error handling for weight initialization in Unquantized… documentation Improvements or additions to documentation v1
#20321 opened Jul 1, 2025 by koiker Loading…
3 of 4 tasks
HF Hub LoRA Resolver ci/build documentation Improvements or additions to documentation
#20320 opened Jul 1, 2025 by alex-jw-brooks Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.