Skip to content

Pull requests: HabanaAI/vllm-fork

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

consolidate pd/dp scripts
#1346 opened May 31, 2025 by hlin99 Loading…
by default disable contiguous_pa on Gaudi2.
#1345 opened May 30, 2025 by ccrhx4 Loading…
Revise DeepSeek-R1 README and update start scripts
#1339 opened May 29, 2025 by taotod Loading…
fix requirements/hpu.txt for hpu extension
#1336 opened May 29, 2025 by ranzhejiang Loading…
Fix prefill warm up issue
#1335 opened May 29, 2025 by yeonsily Draft
Upgrade to HPU docker 1.21.0 and update run_cluster.sh
#1331 opened May 28, 2025 by tvoas Loading…
Fix vllm crash when running with lm-eval
#1321 opened May 27, 2025 by ccrhx4 Loading…
Add Flag to speed up Qwen3 fp8 warmup issue
#1319 opened May 27, 2025 by Yanli2190 Loading…
[Torch compile] Torch compilation on Sampler
#1314 opened May 26, 2025 by jczaja Loading…
Enabled MoE for both BF16 and INC based FP8.
#1309 opened May 23, 2025 by gyou2021 Loading…
parallel compile for fast warm up
#1304 opened May 22, 2025 by inkcherry Loading…
optimize transfer time. use mooncake put/get_unsafe.
#1297 opened May 22, 2025 by jikunshang Loading…
Qwen2.5 Omni
#1296 opened May 22, 2025 by wenbinc-Bin Loading…
Add torch.compile tests into test_config.yaml
#1289 opened May 21, 2025 by kzawora-intel Loading…
ProTip! Adding no:label will show everything without a label.