Skip to content

Commit

Permalink
update docs
Browse files Browse the repository at this point in the history
  • Loading branch information
RunningLeon committed Oct 18, 2023
1 parent 2054d63 commit 68cd9f2
Show file tree
Hide file tree
Showing 12 changed files with 18 additions and 14 deletions.
3 changes: 2 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -99,7 +99,7 @@ And the request throughput of TurboMind is 30% higher than vLLM.
Install lmdeploy with pip ( python 3.8+) or [from source](./docs/en/build.md)

```shell
pip install lmdeploy
pip install lmdeploy[all]
```

### Deploy InternLM
Expand Down Expand Up @@ -181,6 +181,7 @@ bash workspace/service_docker_up.sh
Then, you can communicate with the inference server by command line,

```shell
python3 -m pip install tritonclient[grpc]
python3 -m lmdeploy.serve.client {server_ip_addresss}:33337
```

Expand Down
3 changes: 2 additions & 1 deletion README_zh-CN.md
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ TurboMind 的 output token throughput 超过 2000 token/s, 整体比 DeepSpeed
使用 pip ( python 3.8+) 安装 LMDeploy,或者[源码安装](./docs/zh_cn/build.md)

```shell
pip install lmdeploy
pip install lmdeploy[all]
```

### 部署 InternLM
Expand Down Expand Up @@ -181,6 +181,7 @@ bash workspace/service_docker_up.sh
你可以通过命令行方式与推理服务进行对话:

```shell
python3 -m pip install tritonclient[grpc]
python3 -m lmdeploy.serve.client {server_ip_addresss}:33337
```

Expand Down
4 changes: 2 additions & 2 deletions docs/en/faq.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ It may have been caused by the following reasons.
1. You haven't installed lmdeploy's precompiled package. `_turbomind` is the pybind package of c++ turbomind, which involves compilation. It is recommended that you install the precompiled one.

```shell
pip install lmdeploy
pip install lmdeploy[all]
```

2. If you have installed it and still encounter this issue, it is probably because you are executing turbomind-related command in the root directory of lmdeploy source code. Switching to another directory will fix it
Expand All @@ -26,7 +26,7 @@ pip install lmdeploy

### libnccl.so.2 not found

Make sure you have install lmdeploy (>=v0.0.5) through `pip install lmdeploy`.
Make sure you have install lmdeploy (>=v0.0.5) through `pip install lmdeploy[all]`.

If the issue still exists after lmdeploy installation, add the path of `libnccl.so.2` to environment variable LD_LIBRARY_PATH.

Expand Down
2 changes: 1 addition & 1 deletion docs/en/supported_models/codellama.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@ Based on the above table, download the model that meets your requirements. Execu

```shell
# install lmdeploy
python3 -m pip install lmdeploy
python3 -m pip install lmdeploy[all]

# convert weight layout
python3 -m lmdeploy.serve.turbomind.deploy codellama /the/path/of/codellama/model
Expand Down
2 changes: 1 addition & 1 deletion docs/en/w4a16.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ LMDeploy supports LLM model inference of 4-bit weight, with the minimum requirem
Before proceeding with the inference, please ensure that lmdeploy is installed.

```shell
pip install lmdeploy
pip install lmdeploy[all]
```

## 4-bit LLM model Inference
Expand Down
4 changes: 2 additions & 2 deletions docs/zh_cn/faq.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ pip install --upgrade mmengine
1. 您没有安装 lmdeploy 的预编译包。`_turbomind`是 turbomind c++ 的 pybind部分,涉及到编译。推荐您直接安装预编译包。

```
pip install lmdeploy
pip install lmdeploy[all]
```

2. 如果已经安装了,还是出现这个问题,请检查下执行目录。不要在 lmdeploy 的源码根目录下执行 python -m lmdeploy.turbomind.\*下的package,换到其他目录下执行。
Expand All @@ -26,7 +26,7 @@ pip install lmdeploy

### libnccl.so.2 not found

确保通过 `pip install lmdeploy` 安装了 lmdeploy (>=v0.0.5)。
确保通过 `pip install lmdeploy[all]` 安装了 lmdeploy (>=v0.0.5)。

如果安装之后,问题还存在,那么就把`libnccl.so.2`的路径加入到环境变量 LD_LIBRARY_PATH 中。

Expand Down
2 changes: 1 addition & 1 deletion docs/zh_cn/supported_models/codellama.md
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@

```shell
# 安装 lmdeploy
python3 -m pip install lmdeploy
python3 -m pip install lmdeploy[all]

# 转模型格式
python3 -m lmdeploy.serve.turbomind.deploy codellama /path/of/codellama/model
Expand Down
2 changes: 1 addition & 1 deletion docs/zh_cn/w4a16.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ LMDeploy 支持 4bit 权重模型的推理,**对 NVIDIA 显卡的最低要求
在推理之前,请确保安装了 lmdeploy

```shell
pip install lmdeploy
pip install lmdeploy[all]
```

## 4bit 权重模型推理
Expand Down
3 changes: 2 additions & 1 deletion requirements.txt
Original file line number Diff line number Diff line change
@@ -1,3 +1,4 @@
-r requirements/build.txt
-r requirements/runtime.txt
-r requirements/optional.txt
-r requirements/lite.txt
-r requirements/serve.txt
2 changes: 2 additions & 0 deletions requirements/lite.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,2 @@
accelerate
datasets
2 changes: 0 additions & 2 deletions requirements/optional.txt → requirements/serve.txt
Original file line number Diff line number Diff line change
@@ -1,5 +1,3 @@
accelerate
datasets
fastapi
shortuuid
uvicorn
3 changes: 2 additions & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -138,7 +138,8 @@ def gen_packages_items():
install_requires=parse_requirements('requirements/runtime.txt'),
extras_require={
'all': parse_requirements('requirements.txt'),
'optional': parse_requirements('requirements/optional.txt'),
'lite': parse_requirements('requirements/lite.txt'),
'serve': parse_requirements('requirements/serve.txt'),
},
has_ext_modules=check_ext_modules,
classifiers=[
Expand Down

0 comments on commit 68cd9f2

Please sign in to comment.