From 04ad904c92086635497d8358ef7e0854d88108d8 Mon Sep 17 00:00:00 2001 From: Travis Addair Date: Thu, 21 Dec 2023 22:11:13 -0800 Subject: [PATCH] Update README.md --- README.md | 26 ++++++++++++++++++++++++-- 1 file changed, 24 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 9322681d8..1af5a00df 100644 --- a/README.md +++ b/README.md @@ -25,6 +25,7 @@ LoRAX (LoRA eXchange) is a framework that allows users to serve thousands of fin - [Launch LoRAX Server](#launch-lorax-server) - [Prompt via REST API](#prompt-via-rest-api) - [Prompt via Python Client](#prompt-via-python-client) + - [Next steps](#next-steps) - [πŸ™‡ Acknowledgements](#-acknowledgements) - [πŸ—ΊοΈ Roadmap](#️-roadmap) @@ -79,7 +80,12 @@ Prompt base LLM: ```shell curl 127.0.0.1:8080/generate \ -X POST \ - -d '{"inputs": "[INST] Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May? [/INST]", "parameters": {"max_new_tokens": 64}}' \ + -d '{ + "inputs": "[INST] Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May? [/INST]", + "parameters": { + "max_new_tokens": 64 + } + }' \ -H 'Content-Type: application/json' ``` @@ -88,7 +94,13 @@ Prompt a LoRA adapter: ```shell curl 127.0.0.1:8080/generate \ -X POST \ - -d '{"inputs": "[INST] Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May? [/INST]", "parameters": {"max_new_tokens": 64, "adapter_id": "vineetsharma/qlora-adapter-Mistral-7B-Instruct-v0.1-gsm8k"}}' \ + -d '{ + "inputs": "[INST] Natalia sold clips to 48 of her friends in April, and then she sold half as many clips in May. How many clips did Natalia sell altogether in April and May? [/INST]", + "parameters": { + "max_new_tokens": 64, + "adapter_id": "vineetsharma/qlora-adapter-Mistral-7B-Instruct-v0.1-gsm8k" + } + }' \ -H 'Content-Type: application/json' ``` @@ -122,6 +134,16 @@ See [Reference - Python Client](https://predibase.github.io/lorax/reference/pyth For other ways to run LoRAX, see [Getting Started - Kubernetes](https://predibase.github.io/lorax/getting_started/kubernetes), [Getting Started - SkyPilot](https://predibase.github.io/lorax/getting_started/skypilot), and [Getting Started - Local](https://predibase.github.io/lorax/getting_started/local). +### Next steps + +Here are some other interesting Mistral-7B fine-tuned models to try out: + +- [alignment-handbook/zephyr-7b-dpo-lora](https://huggingface.co/alignment-handbook/zephyr-7b-dpo-lora): Mistral-7b fine-tuned on Zephyr-7B dataset with DPO. +- [IlyaGusev/saiga_mistral_7b_lora](https://huggingface.co/IlyaGusev/saiga_mistral_7b_lora): Russian chatbot based on `Open-Orca/Mistral-7B-OpenOrca`. +- [Undi95/Mistral-7B-roleplay_alpaca-lora](https://huggingface.co/Undi95/Mistral-7B-roleplay_alpaca-lora): Fine-tuned using role-play prompts. + +You can find more LoRA adapters [here](https://huggingface.co/models?pipeline_tag=text-generation&sort=trending&search=-lora), or try fine-tuning your own with [PEFT](https://github.com/huggingface/peft) or [Ludwig](https://ludwig.ai). + ## πŸ™‡ Acknowledgements LoRAX is built on top of HuggingFace's [text-generation-inference](https://github.com/huggingface/text-generation-inference), forked from v0.9.4 (Apache 2.0).