From 8492c621c723ae4a5249a705c8c147408fb57921 Mon Sep 17 00:00:00 2001 From: Naren Dasan Date: Tue, 3 Sep 2024 14:33:33 -0600 Subject: [PATCH] docs: Adding words to the refit and engine caching tutorials Signed-off-by: Naren Dasan --- docsrc/conf.py | 2 +- py/torch_tensorrt/dynamo/utils.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/docsrc/conf.py b/docsrc/conf.py index c4ddd6eaec..daa1a30100 100644 --- a/docsrc/conf.py +++ b/docsrc/conf.py @@ -93,7 +93,7 @@ sphinx_gallery_conf = { "examples_dirs": "../examples", "gallery_dirs": "tutorials/_rendered_examples/", - "ignore_pattern": "utils.py" + "ignore_pattern": "utils.py", } # Setup the breathe extension diff --git a/py/torch_tensorrt/dynamo/utils.py b/py/torch_tensorrt/dynamo/utils.py index fa6eefff1f..2af7922cd1 100644 --- a/py/torch_tensorrt/dynamo/utils.py +++ b/py/torch_tensorrt/dynamo/utils.py @@ -347,7 +347,7 @@ def unwrap_tensor_shape( if isinstance(tensor, int): tensor_shape.append(tensor) elif isinstance(tensor, torch.SymInt): - min_max_opt: Dict[str, int] = extract_var_range_info(tensor) + min_max_opt = extract_var_range_info(tensor) tensor_shape.append((min_max_opt["min"], min_max_opt["max"])) elif isinstance(tensor, (torch.Tensor, FakeTensor)): for dimension in tensor.shape: