-
Notifications
You must be signed in to change notification settings - Fork 5
Running LLM As Chatbot in your cloud
This repository is a fork of deep-diver/LLM-As-Chatbot.
The only difference from the original repo is the .dstack.yml
file, which allows you to run LLM-As-Chatbot in your cloud with a single dstack run
command, automatically provisioning cloud resources for you.
git clone https://github.com/dstackai/LLM-As-Chatbot.git
cd LLM-As-Chatbot
pip install "dstack[aws,gcp,azure,lambda]" -U
dstack start
Once the server is up, make sure to log in and create a project with your cloud credentials (AWS, GCP, or Azure). After creating the project, copy its dstack config
command and run it locally to configure the CLI to use this project.
Create a .dstack/profiles.yml
file that points to the created project and describes the resources.
Example:
profiles:
- name: gcp
project: gcp
resources:
memory: 48GB
gpu:
memory: 24GB
default: true
Run the dstack init
command:
dstack init
Use the dstack run
command:
dstack run .
This command will build the environment and run LLM-As-Chatbot in your cloud.
dstack
will automatically forward the port to your local machine, providing secure and convenient access.
For more details on how dstack
works, check out its documentation.