diff --git a/docs/source/_gallery_original/index.rst b/docs/source/_gallery_original/index.rst index 7500b240f52..14cc9609b41 100644 --- a/docs/source/_gallery_original/index.rst +++ b/docs/source/_gallery_original/index.rst @@ -35,7 +35,7 @@ Contents :caption: LLM Models Mixtral (Mistral AI) - MiniCPM (openbmb) + MiniCPM (openbmb) Mistral 7B (Mistral AI) DBRX (Databricks) Llama-2 (Meta) diff --git a/llm/minicpm/README.md b/llm/minicpm/README.md index 35c0af3f723..1bd0d059408 100644 --- a/llm/minicpm/README.md +++ b/llm/minicpm/README.md @@ -1,6 +1,6 @@ -📰 **Update (26 April 2024) -** SkyPilot now also supports the [**MiniCPM-2B**](https://openbmb.vercel.app/?category=Chinese+Blog/) model! Use [serve-2b.yaml](https://github.com/skypilot-org/skypilot/blob/master/llm/minicpm/serve-2b.yaml) to serve the 110B model. +📰 **Update (26 April 2024) -** SkyPilot now also supports the [**MiniCPM-2B**](https://openbmb.vercel.app/?category=Chinese+Blog/) model! Use [serve-2b.yaml](https://github.com/skypilot-org/skypilot/blob/master/llm/minicpm/serve-2b.yaml) to serve the 2B model. 📰 **Update (6 Jun 2024) -** SkyPilot now also supports the [**MiniCPM-1B**](https://openbmb.vercel.app/?category=Chinese+Blog/) model! @@ -63,7 +63,7 @@ curl http://$IP:8000/v1/chat/completions \ ``` -## **Optional:** Accessing Qwen with Chat GUI +## **Optional:** Accessing Cpm with Chat GUI It is also possible to access the Qwen service with a GUI using [vLLM](https://github.com/vllm-project/vllm). @@ -76,4 +76,3 @@ sky launch -c cpm-gui ./gui.yaml --env MODEL_NAME='openbmb/MiniCPM-2B-sft-bf16' ``` | INFO | stdout | Running on public URL: https://6141e84201ce0bb4ed.gradio.live ``` -