Skip to content

Pull requests: vllm-project/vllm-gaudi

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[upstream crash] fix spec decode due to upstream 24986
#265 opened Sep 26, 2025 by xuechendi Loading…
Convert padding itertools.islice to list
#264 opened Sep 25, 2025 by malsbat Loading…
Enable modification of prompt BS
#258 opened Sep 25, 2025 by ksmusz Loading…
Update long context README
#256 opened Sep 25, 2025 by iboiko-habana Loading…
Update LoRA tests
#255 opened Sep 25, 2025 by vivekgoe Loading…
enable p2d4
#253 opened Sep 24, 2025 by hsubramony Draft
Support DP for unified attention
#242 opened Sep 24, 2025 by wuxun-zhang Loading…
Add assert for empty buckets
#236 opened Sep 23, 2025 by iboiko-habana Loading…
Fix calculating used blocks in unified attn
#232 opened Sep 23, 2025 by madamczyk-intel Loading…
[Docs] README update - bucketing, warmup
#231 opened Sep 23, 2025 by adobrzyn Loading…
Enable modification of prompt BS
#224 opened Sep 22, 2025 by ksmusz Loading…
KV cache sharing
#223 opened Sep 22, 2025 by jakub-sochacki Draft
[DO NOT MERGE] Update README.md
#220 opened Sep 22, 2025 by kzawora-intel Loading…
Adding prompt context flags for linear warmup
#217 opened Sep 22, 2025 by iboiko-habana Loading…
ProTip! Adding no:label will show everything without a label.