From 07c21dfe63e4bdbc2d5beb923f33c45e8766d06c Mon Sep 17 00:00:00 2001 From: Kalila Date: Sat, 25 Nov 2023 14:50:22 +0000 Subject: [PATCH] Lint + fix typo. --- fastchat/model/model_adapter.py | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/fastchat/model/model_adapter.py b/fastchat/model/model_adapter.py index ba54f7db4..d5da78992 100644 --- a/fastchat/model/model_adapter.py +++ b/fastchat/model/model_adapter.py @@ -88,7 +88,7 @@ def load_model( use_fast=False, revision=revision, trust_remote_code=True, - gptq_transformers_config=gptq_transformers_config + gptq_transformers_config=gptq_transformers_config, ) try: model = AutoModelForCausalLM.from_pretrained( @@ -96,7 +96,7 @@ def load_model( from_pretrained_kwargs=from_pretrained_kwargs, low_cpu_mem_usage=True, trust_remote_code=True, - gptq_transformers_config=gptq_transformers_config + gptq_transformers_config=gptq_transformers_config, ) except NameError: model = AutoModel.from_pretrained( @@ -104,7 +104,7 @@ def load_model( from_pretrained_kwargs=from_pretrained_kwargs, low_cpu_mem_usage=True, trust_remote_code=True, - gptq_transformers_config=gptq_transformers_config + gptq_transformers_config=gptq_transformers_config, ) return model, tokenizer @@ -868,7 +868,7 @@ def load_model( model_path, from_pretrained_kwargs=from_pretrained_kwargs, trust_remote_code=True, - add_special_tokens=add_special_tokens + add_special_tokens=add_special_tokens, gptq_transformers_config=gptq_transformers_config, ) return model, tokenizer @@ -1937,7 +1937,7 @@ def load_model( trust_remote_code=True, low_cpu_mem_usage=True, ) - + return model, tokenizer def get_default_conv_template(self, model_path: str) -> Conversation: