Skip to content

Commit 29d3456

Browse files
committed
fix lmdeploy 后端
1 parent dfa4676 commit 29d3456

File tree

1 file changed

+3
-7
lines changed

1 file changed

+3
-7
lines changed

gpt_server/model_backend/lmdeploy_backend.py

Lines changed: 3 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -29,13 +29,9 @@ def __init__(self, model_path) -> None:
2929
backend = backend_map[os.getenv("backend")]
3030
logger.info(f"后端 {backend}")
3131
if backend == "pytorch":
32-
backend_config = PytorchEngineConfig(
33-
model_name="", tp=int(os.getenv("num_gpus", "1")), thread_safe=False
34-
)
32+
backend_config = PytorchEngineConfig(tp=int(os.getenv("num_gpus", "1")))
3533
if backend == "turbomind":
36-
backend_config = TurbomindEngineConfig(
37-
model_name="", tp=int(os.getenv("num_gpus", "1")), thread_safe=True
38-
)
34+
backend_config = TurbomindEngineConfig(tp=int(os.getenv("num_gpus", "1")))
3935
pipeline_type, pipeline_class = get_task(model_path)
4036
logger.info(f"模型架构:{pipeline_type}")
4137
self.async_engine = pipeline_class(
@@ -45,7 +41,7 @@ def __init__(self, model_path) -> None:
4541
)
4642

4743
async def stream_chat(self, params: Dict[str, Any]) -> AsyncGenerator:
48-
prompt = params.get("prompt","")
44+
prompt = params.get("prompt", "")
4945
logger.info(prompt)
5046
messages = params["messages"]
5147
request_id = params.get("request_id", "0")

0 commit comments

Comments
 (0)