From ea660512be4a686901aaf9f38fff78907cd4639f Mon Sep 17 00:00:00 2001 From: "pre-commit-ci[bot]" <66853113+pre-commit-ci[bot]@users.noreply.github.com> Date: Mon, 25 Mar 2024 07:19:47 +0000 Subject: [PATCH] [pre-commit.ci] auto fixes from pre-commit.com hooks for more information, see https://pre-commit.ci --- intel_extension_for_transformers/neural_chat/chatbot.py | 2 +- .../pipeline/plugins/retrieval/parser/context_utils.py | 2 +- .../neural_chat/pipeline/plugins/retrieval/parser/parser.py | 2 +- intel_extension_for_transformers/neural_chat/prompts/prompt.py | 2 +- 4 files changed, 4 insertions(+), 4 deletions(-) diff --git a/intel_extension_for_transformers/neural_chat/chatbot.py b/intel_extension_for_transformers/neural_chat/chatbot.py index 2ef2b70bae42..a7f0c3cc5d75 100644 --- a/intel_extension_for_transformers/neural_chat/chatbot.py +++ b/intel_extension_for_transformers/neural_chat/chatbot.py @@ -323,7 +323,7 @@ def build_chatbot(config: PipelineConfig=None): logging.error("build_chatbot: plugin init failed") return adapter.register_plugin_instance(plugin_name, plugins[plugin_name]["instance"]) - + if get_latest_error(): return else: diff --git a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/context_utils.py b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/context_utils.py index 5468739200d0..aba8f8962ecd 100644 --- a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/context_utils.py +++ b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/context_utils.py @@ -95,7 +95,7 @@ def get_relation(table_coords, caption_coords, table_page_number, caption_page_n y_distance = 0 y_close = y_distance < threshold return same_page and x_overlap and y_close, y_distance - + raw_pdf_elements = partition_pdf( filename=pdf_path, infer_table_structure=True, diff --git a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/parser.py b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/parser.py index 72529a5099a3..cbc00aa4310c 100644 --- a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/parser.py +++ b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/parser/parser.py @@ -51,7 +51,7 @@ def load(self, input, **kwargs): self.process = kwargs['process'] self.table_summary_model_name_or_path = kwargs['table_summary_model_name_or_path'] self.table_summary_mode = kwargs['table_summary_mode'] if 'table_summary_mode' in kwargs else 'none' - + if isinstance(input, str): if os.path.isfile(input): data_collection = self.parse_document(input) diff --git a/intel_extension_for_transformers/neural_chat/prompts/prompt.py b/intel_extension_for_transformers/neural_chat/prompts/prompt.py index a0a9c0958451..f1f2b6fb0f1a 100644 --- a/intel_extension_for_transformers/neural_chat/prompts/prompt.py +++ b/intel_extension_for_transformers/neural_chat/prompts/prompt.py @@ -331,4 +331,4 @@ def generate_sqlcoder_prompt(qurey, metadata_file): {table_content} --- ### Generated Summary: -""" \ No newline at end of file +"""