-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
WIP: LoRA Adapters #304
base: main
Are you sure you want to change the base?
WIP: LoRA Adapters #304
Conversation
Nice drawing. This is very helpful! 🌻 |
Can you show an example that has the url field? I'm assuming the url field must be used to specify the base model? |
@samos123 I currently have all examples in the diagrams |
That's where I looked but none of them have the base model URL set? |
Model |
Note, it looks like vLLM supports loading adapters from huggingface: vllm-project/vllm#6234 |
Note, vLLM has an endpoint to support dynamic loading/unloading of adapters: vllm-project/vllm#6566 |
#url: hf://meta-llama/Llama-2-7b | ||
adapters: | ||
- id: test | ||
url: hf://jashing/tinyllama-colorist-lora |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
does vLLM support directly loading this adapter from HF or is it a hard requirement to download the lora adapter first?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
vLLM can load it from HF but not S3
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Addresses #132