@@ -29,13 +29,9 @@ def __init__(self, model_path) -> None:
2929 backend = backend_map [os .getenv ("backend" )]
3030 logger .info (f"后端 { backend } " )
3131 if backend == "pytorch" :
32- backend_config = PytorchEngineConfig (
33- model_name = "" , tp = int (os .getenv ("num_gpus" , "1" )), thread_safe = False
34- )
32+ backend_config = PytorchEngineConfig (tp = int (os .getenv ("num_gpus" , "1" )))
3533 if backend == "turbomind" :
36- backend_config = TurbomindEngineConfig (
37- model_name = "" , tp = int (os .getenv ("num_gpus" , "1" )), thread_safe = True
38- )
34+ backend_config = TurbomindEngineConfig (tp = int (os .getenv ("num_gpus" , "1" )))
3935 pipeline_type , pipeline_class = get_task (model_path )
4036 logger .info (f"模型架构:{ pipeline_type } " )
4137 self .async_engine = pipeline_class (
@@ -45,7 +41,7 @@ def __init__(self, model_path) -> None:
4541 )
4642
4743 async def stream_chat (self , params : Dict [str , Any ]) -> AsyncGenerator :
48- prompt = params .get ("prompt" ,"" )
44+ prompt = params .get ("prompt" , "" )
4945 logger .info (prompt )
5046 messages = params ["messages" ]
5147 request_id = params .get ("request_id" , "0" )
0 commit comments