Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Roadmap] vLLM Roadmap Q2 2024 #3861

Closed
36 of 65 tasks
simon-mo opened this issue Apr 4, 2024 · 38 comments
Closed
36 of 65 tasks

[Roadmap] vLLM Roadmap Q2 2024 #3861

simon-mo opened this issue Apr 4, 2024 · 38 comments

Comments

@simon-mo
Copy link
Collaborator

simon-mo commented Apr 4, 2024

This document includes the features in vLLM's roadmap for Q2 2024. Please feel free to discuss and contribute to the specific features at related RFC/Issues/PRs and add anything else you'd like to talk about in this issue.

You can see our historical roadmap at #2681, #244. This roadmap contains work committed by the vLLM team from UC Berkeley, as well as the broader vLLM contributor groups including but not limited to Anyscale, IBM, NeuralMagic, Roblox, Oracle Cloud. You can also find help wanted items in this roadmap as well! Additionally, this roadmap is shaped by you, our user community!

Themes.

We categorized our roadmap into 6 broad themes:

  • Broad model support: vLLM should support a wide range of transformer based models. It should be kept up to date as much as possible. This includes new auto-regressive decoder models, encoder-decoder models, hybrid architectures, and models supporting multi-modal inputs.
  • Excellent hardware coverage: vLLM should run on a wide range of accelerators for production AI workload. This includes GPUs, tensor accelerators, and CPUs. We will work closely with hardware vendors to ensure vLLM utilizes the greatest performance out of the chip.
  • Performance optimization:vLLM should be kept up to date with the latest performance optimization techniques. Users of vLLM can trust its performance to be competitive and strong.
  • Production level engine: vLLM should be the go-to choice for production level serving engine with a suite of features bridging the gaps from single forward pass to 24/7 service.
  • Strong OSS product: vLLM is and will be a true community project. We want it to be a healthy project with regular release cadence, good documentation, and adding new reviewers to the codebase.
  • Extensible architectures: For vLLM to grow at an even faster pace, it needs good abstractions to support a wide range of scheduling policies, hardware backends, and inference optimizations. We will work on refactoring the codebase to support that.

Broad Model Support

Help Wanted:

Excellent Hardware Coverage

  • AMD MI300x: enhancing fp8 performance [enable FP8 compute]
  • NVIDIA H100: enhancing fp8 performance
  • AWS Trainium and Inferentia
  • Google TPU
  • Intel CPU
  • Intel GPU
  • Intel Gaudi

Performance Optimization

  • Speculative decoding
    • Speculative decoding framework for top-1 proposals w/draft model
    • Proposer improvement: Prompt-lookup n-gram speculations
    • Scoring improvement: Make batch expansion optional
    • Scoring improvement: dynamic scoring length policy
  • Kernels:
  • Quantization:

Help Wanted:

  • Sparse kv cache (H2O, compression, FastDecode)
  • Speculative decoding
    • Proposer/scoring/verifier improvement: Top-k “tree attention” proposals for Eagle/Medusa/Draft model
    • Proposer improvement: RAG n-gram speculations
    • Proposer improvement: Eagle/Medusa top-1 proposals
    • Proposer improvement: Quantized draft models
    • Verifier improvement: Typical acceptance

Production Level Engine

Help Wanted:

  • Logging serving FLOPs for performance analysis
  • Dynamic LoRA adapter downloads from hub/S3

Strong OSS Product

  • Continuous benchmarks (resource needed!)
  • Commit to 2wk release cadence
  • Growing reviewer and committer base
  • Better docs
    • doc: memory and performance tuning guide
    • doc: apc documentation
    • doc: hardware support levels, feature matrix, and policies
    • doc: guide to horizontally scale up vLLM service
    • doc: developer guide for adding new draft based models or draft-less optimizations
  • Automatic CD of nightly wheels and docker images

Help Wanted:

  • ARM aarch-64 support for AWS Graviton based instances and GH200
  • Full correctness test with HuggingFace transformers. Resources needed.
  • Well tested support for lm-eval-harness (logprobs, get tokenizers)
  • Local development workflow without cuda

Extensible Architecture

  • Prototype pipeline parallelism
  • Extensible memory manager
  • Extensible scheduler
  • torch.compile investigations
    • use compile for quantization kernel fusion
    • use compile for future proofing graph mode
    • use compile for xpu or other accelerators
  • Architecture for queue management and request prioritization
  • Streaming LLM, prototype it on new block manager
  • Investigate Tensor + Pipeline parallelism (LIGER)
@simon-mo simon-mo added misc and removed misc labels Apr 4, 2024
@simon-mo simon-mo pinned this issue Apr 4, 2024
@Jeffwan
Copy link
Contributor

Jeffwan commented Apr 5, 2024

@simon-mo for prefill disaggregation. from the splitwise and distserve paper, they all build solution on top of vLLM for evaluation. Any contribution from these teams? is vLLM community open for public contribution for this feature?

@simon-mo
Copy link
Collaborator Author

simon-mo commented Apr 5, 2024

@Jeffwan yes! We are actively with the authors of both papers to integrate the work properly. We are also working with Sarathi's authors for chunked prefill as well.

@kanseaveg
Copy link

Any update for PEFT?

please consider support huggingface peft, thank you. #1129

Copy link
Collaborator Author

simon-mo commented Apr 5, 2024

Hi @kanseaveg, we do support LoRA and planning to add prefix tuning support, which should allow Hugging face PEFT model format. Which PEFT methods are you interested in?

@kanseaveg
Copy link

kanseaveg commented Apr 5, 2024

@simon-mo Thank you very much for your reply.There are three common types of tuning methods that I am currently concerned about:

  • prefix-tuning / p-tuning v2
  • adapter-tuning
  • lora-tuning (currently supported)
    I hope the vllm framework can support this, which is what I mentioned in Q3 last year and Q1 this year.
    Thank you very much for your reply.

@accupham
Copy link

accupham commented Apr 5, 2024

Maybe consider supporting QuaRot quantization scheme?

QuaRot: Outlier-Free 4-Bit Inference in Rotated LLMs
We introduce QuaRot, a new Quantization scheme based on Rotations, which is able to quantize LLMs end-to-end, including all weights, activations, and KV cache in 4 bits. QuaRot rotates LLMs in a way that removes outliers from the hidden state without changing the output, making quantization easier. This computational invariance is applied to the hidden state (residual) of the LLM, as well as to the activations of the feed-forward components, aspects of the attention mechanism and to the KV cache. The result is a quantized model where all matrix multiplications are performed in 4-bits, without any channels identified for retention in higher precision. Our quantized LLaMa2-70B model has losses of at most 0.29 WikiText-2 perplexity and retains 99% of the zero-shot performance. Code is available at: this https URL.

I think this would be huge for larger models like Command-R+ (104B) being able to fit into a single 80G A100 with negligible performance losses.

@zbloss
Copy link

zbloss commented Apr 5, 2024

Very excited to see both Embedding models and CPU support on the roadmap!

These being implemented would make vLLM my default model serving engine.

@sangstar
Copy link
Contributor

sangstar commented Apr 5, 2024

Very excited to see that the tensorizer PR is in this roadmap! Sorry about all the pings, I'm just passionate about getting this to vLLM users :D More than happy to be of any assistance in getting that feature implemented :)

@PenutChen
Copy link

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

@yukavio
Copy link

yukavio commented Apr 9, 2024

I'm very interested in implementing tree attention for speculative decoding. @simon-mo

@jeejeelee
Copy link
Collaborator

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

#4015 had done this

@qZhang88
Copy link

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

#4015 had done this

This is strange, serving lora finetune for Llama-3 (vocab size 12800) has the same problem, When using LoRA, vocab size must be 32000 >= vocab_size <= 33024, however same code finetune for Qwen1.5-7B-Chat, with vocab size 151643, has no such serving problem, why?

@jeejeelee
Copy link
Collaborator

jeejeelee commented Apr 19, 2024

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

#4015 had done this

This is strange, serving lora finetune for Llama-3 (vocab size 12800) has the same problem, When using LoRA, vocab size must be 32000 >= vocab_size <= 33024, however same code finetune for Qwen1.5-7B-Chat, with vocab size 151643, has no such serving problem, why?

the function create_lora_weights from LogitsProcessorWithLoRA throws this error.

Model using llama architecture designate lm_head as a target module for lora, and need instantiate LogitsProcessorWithLoRA,refer to: https://github.com/vllm-project/vllm/blob/main/vllm/lora/models.py#438

Models such as qwen-2 don't designate lm_head as a target module for lora,so,They don't instantiate LogitsProcessorWithLoRA

@qZhang88
Copy link

qZhang88 commented Apr 20, 2024

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

#4015 had done this

This is strange, serving lora finetune for Llama-3 (vocab size 12800) has the same problem, When using LoRA, vocab size must be 32000 >= vocab_size <= 33024, however same code finetune for Qwen1.5-7B-Chat, with vocab size 151643, has no such serving problem, why?

the function create_lora_weights from LogitsProcessorWithLoRA throws this error.

Model using llama architecture designate lm_head as a target module for lora, and need instantiate LogitsProcessorWithLoRA,refer to: https://github.com/vllm-project/vllm/blob/main/vllm/lora/models.py#438

Models such as qwen-2 don't designate lm_head as a target module for lora,so,They don't instantiate LogitsProcessorWithLoRA

I see, but lm_head is not finetuned during lora, so there is no need to replace logits_processor. In my adapter_config.json, target_modules does not contains lm_head

  "target_modules": [
    "gate_proj",
    "v_proj",
    "q_proj",
    "o_proj",
    "up_proj",
    "k_proj",
    "down_proj"
  ],

@jeejeelee
Copy link
Collaborator

Will larger vocabulary size for multi-lora be supported in Q2 2024? Related: #3000

#4015 had done this

This is strange, serving lora finetune for Llama-3 (vocab size 12800) has the same problem, When using LoRA, vocab size must be 32000 >= vocab_size <= 33024, however same code finetune for Qwen1.5-7B-Chat, with vocab size 151643, has no such serving problem, why?

the function create_lora_weights from LogitsProcessorWithLoRA throws this error.
Model using llama architecture designate lm_head as a target module for lora, and need instantiate LogitsProcessorWithLoRA,refer to: https://github.com/vllm-project/vllm/blob/main/vllm/lora/models.py#438
Models such as qwen-2 don't designate lm_head as a target module for lora,so,They don't instantiate LogitsProcessorWithLoRA

I see, but lm_head is not finetuned during lora, so there is no need to replace logits_processor. In my adapter_config.json, target_modules does not contains lm_head

  "target_modules": [
    "gate_proj",
    "v_proj",
    "q_proj",
    "o_proj",
    "up_proj",
    "k_proj",
    "down_proj"
  ],

vllm support multi-lora, whether to replace logits_processor is determined by the model's support_modules, not by the adapter_config.json.

@Vermeille
Copy link

would like to help with #620

@WangErXiao
Copy link

@Jeffwan yes! We are actively with the authors of both papers to integrate the work properly. We are also working with Sarathi's authors for chunked prefill as well.

Looking forward to the release of vllm support for Prefill-Decode Disaggregation feature

@colourful-tree
Copy link

@simon-mo Hi, How about https://arxiv.org/abs/2404.18057? It seems to have a significant advantage in long sequences, and it does not conflict with page-attention technology.

@kanseaveg
Copy link

kanseaveg commented May 10, 2024

@simon-mo Any thing update about the #3117 ? This issue was raised in February, and it has been nearly three months. We sincerely look forward to your updating in this regard, thank you.
image

@simon-mo
Copy link
Collaborator Author

@simon-mo Any thing update about the #3117 ? This issue was raised in February, and it has been nearly three months. We sincerely look forward to your updating in this regard, thank you.

Still in progress. @robertgshaw2-neuralmagic can help comment more.

@zxy-zzz
Copy link

zxy-zzz commented May 11, 2024

Do you have plans to incorporate RISC-V or ARM CPU backends into the vLLM project? Thank you.

@robertgshaw2-neuralmagic
Copy link
Collaborator

We should consider long-context optimizations for Q3.

@sumukshashidhar
Copy link

Hi - with smaller models being popular these days - I'm wondering, if for Q3, there are any plans for data parallelism support (loading the same model onto gpu's as copies)

If not - I can help with this

@johnsonwag03
Copy link

do you have plan to support nvidia device jetson with aarch64 ?

@robertgshaw2-neuralmagic
Copy link
Collaborator

Hi - with smaller models being popular these days - I'm wondering, if for Q3, there are any plans for data parallelism support (loading the same model onto gpu's as copies)

If not - I can help with this

Are you thinking this would be something handled internally by LLMEngine or a new front end that stands in front?

If handled internally, this will require significant changes to the core logic.

Also, if this is targeted at offline batch mode, perhaps we will see some gains, though I suspect not too much since we can saturate the GPU via batching even with TP

If this is targeted at online serving, I do not think we should be implementing a load balancer in vLLM. This should be handled by higher level orchestrators like kuberentes or ray

@sumukshashidhar
Copy link

Hi - with smaller models being popular these days - I'm wondering, if for Q3, there are any plans for data parallelism support (loading the same model onto gpu's as copies)
If not - I can help with this

Are you thinking this would be something handled internally by LLMEngine or a new front end that stands in front?

If handled internally, this will require significant changes to the core logic.

Also, if this is targeted at offline batch mode, perhaps we will see some gains, though I suspect not too much since we can saturate the GPU via batching even with TP

If this is targeted at online serving, I do not think we should be implementing a load balancer in vLLM. This should be handled by higher level orchestrators like kuberentes or ray

My particular use-case is automatic large offline batches, for which I have a hotfix - I spin up multiple OpenAI servers, and distribute the prompts among them. Curiously, I see large speedups when I do this, as opposed to TP.

Also, if this is targeted at offline batch mode, perhaps we will see some gains, though I suspect not too much since we can saturate the GPU via batching even with TP.

I'm not sure if this is a bug or something else, because I did indeed see large speedups with this, when I completely removed ray worker communication (some digging said that the overhead is not worth it). If this is not expected, I can try out some experiments and post them here. (This may be an artifact of me having a PCIE GPU cluster, not sped up by NVLINK)

@robertgshaw2-neuralmagic
Copy link
Collaborator

Hi - with smaller models being popular these days - I'm wondering, if for Q3, there are any plans for data parallelism support (loading the same model onto gpu's as copies)
If not - I can help with this

Are you thinking this would be something handled internally by LLMEngine or a new front end that stands in front?
If handled internally, this will require significant changes to the core logic.
Also, if this is targeted at offline batch mode, perhaps we will see some gains, though I suspect not too much since we can saturate the GPU via batching even with TP
If this is targeted at online serving, I do not think we should be implementing a load balancer in vLLM. This should be handled by higher level orchestrators like kuberentes or ray

My particular use-case is automatic large offline batches, for which I have a hotfix - I spin up multiple OpenAI servers, and distribute the prompts among them. Curiously, I see large speedups when I do this, as opposed to TP.

Also, if this is targeted at offline batch mode, perhaps we will see some gains, though I suspect not too much since we can saturate the GPU via batching even with TP.

I'm not sure if this is a bug or something else, because I did indeed see large speedups with this, when I completely removed ray worker communication (some digging said that the overhead is not worth it). If this is not expected, I can try out some experiments and post them here. (This may be an artifact of me having a PCIE GPU cluster, not sped up by NVLINK)

Okay great. We would welcome a contribution focused on the offline batch processing case.

Could you make an RFC issue to discuss a potential design? I think we should try hard to not modify LLMEngine and see if we can handle things in the LLM class

@fenggwsx
Copy link

Very excited to see function calling support in OpenAI-Compatible server is in this roadmap!
This is quite helpful when using LangChain.

@irasin
Copy link
Contributor

irasin commented Jun 3, 2024

@Jeffwan yes! We are actively with the authors of both papers to integrate the work properly. We are also working with Sarathi's authors for chunked prefill as well.

Hi @simon-mo. Is there any update about splitwise? It seems that the development of #2809 has stopped.

@K-Mistele
Copy link
Contributor

Would love to see updates to the docs on how to use supported vision models, embedding models, and the new support for tools with forced tool choice (auto tool choice is still WIP as I understand)

@cason0126
Copy link

Hi @simon-mo , is there any plan to support Huawei's NPU HardWare ?

@CSEEduanyu
Copy link

CSEEduanyu commented Jun 23, 2024

Hi @simon-mo , is there any plan to support Huawei's NPU HardWare ?
@simon-mo Some company with no moral bottom line, don't have anything to do with them。

@simon-mo
Copy link
Collaborator Author

Q3 published here #5805

@simon-mo simon-mo unpinned this issue Jun 25, 2024
@2-fly-4-ai
Copy link

Is function calling available yet?

@K-Mistele
Copy link
Contributor

Is function calling available yet?

Soon, for Hermes and mistral models in #5649

If there are other specific models you're interested in, let me know and I can add it in my follow up PR along with Llama 3.1

@githebs
Copy link

githebs commented Aug 19, 2024

Is function calling available yet?

Soon, for Hermes and mistral models in #5649

If there are other specific models you're interested in, let me know and I can add it in my follow up PR along with Llama 3.1

So initially Llama not included ? Thanks

@joshdevins
Copy link

If there are other specific models you're interested in, let me know and I can add it in my follow up PR along with Llama 3.1

@K-Mistele Is there a PR or issue I can follow for function calling support with Llama 3.1 (70B specifically)?

@K-Mistele
Copy link
Contributor

If there are other specific models you're interested in, let me know and I can add it in my follow up PR along with Llama 3.1

@K-Mistele Is there a PR or issue I can follow for function calling support with Llama 3.1 (70B specifically)?

There is a branch on my vLLM fork, but not a PR yet since #5649 needs to be merged before I open another PR based on it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests