diff --git a/llm/vllm/README.md b/llm/vllm/README.md index 4d63b7acb71..5c407e877c3 100644 --- a/llm/vllm/README.md +++ b/llm/vllm/README.md @@ -54,9 +54,7 @@ sky launch -c vllm-llama2 serve-openai-api.yaml --gpus V100:1 ``` 2. Check the IP for the cluster with: ``` -sky status -a -# Or get the IP with Python API: -IP=$(python -c "import sky; print(sky.status('vllm-llama2')[0]['handle'].head_ip)") +IP=$(sky status --ip vllm-llama2) ``` 3. You can now use the OpenAI API to interact with the model. - Query the models hosted on the cluster: