Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[Bugfix] Fix issues for Pixtral-Large-Instruct-2411 ready ONLY add when PR is ready to merge/full CI is needed
#11393 opened Dec 21, 2024 by ywang96 Loading…
[Bugfix] Fix fully sharded LoRAs with Mixtral
#11390 opened Dec 21, 2024 by n1hility Loading…
Update Dockerfile.rocm ci/build
#11387 opened Dec 20, 2024 by yx-lamini Loading…
[Core] Support global prefix caching
#11385 opened Dec 20, 2024 by lyppg Loading…
[Misc] Adding API Key to the benchmark
#11384 opened Dec 20, 2024 by bjb19 Loading…
[V1] Make AsyncLLMEngine v1-v0 opaque frontend ready ONLY add when PR is ready to merge/full CI is needed
#11383 opened Dec 20, 2024 by rickyyx Loading…
[Bugfix] Last token measurement fix
#11376 opened Dec 20, 2024 by rajveerb Loading…
[Bugfix] Fix sliding window in cohere2 model
#11358 opened Dec 20, 2024 by CXIAAAAA Loading…
[VLM] Support caching in merged multi-modal processor documentation Improvements or additions to documentation
#11341 opened Dec 19, 2024 by DarkLight1337 Loading…
[platform] support pytorch custom op pluggable
#11328 opened Dec 19, 2024 by wangxiyuan Loading…
[platform] support custom torch.compile backend key
#11318 opened Dec 19, 2024 by wangxiyuan Loading…
[Core] Changes to support 0.2.0 flashinfer
#11314 opened Dec 19, 2024 by pavanimajety Loading…
ProTip! Mix and match filters to narrow down what you’re looking for.