diff --git a/runtimes/huggingface/poetry.lock b/runtimes/huggingface/poetry.lock index 4926c679f..5f4252bf0 100644 --- a/runtimes/huggingface/poetry.lock +++ b/runtimes/huggingface/poetry.lock @@ -2359,13 +2359,13 @@ tests = ["pytest", "pytest-cov", "pytest-pep8"] [[package]] name = "optimum" -version = "1.19.1" +version = "1.19.2" description = "Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality." optional = false python-versions = ">=3.7.0" files = [ - {file = "optimum-1.19.1-py3-none-any.whl", hash = "sha256:ca474589682fe10f7827c85260d116603a5823d5c251c453620584cbc06fa5f6"}, - {file = "optimum-1.19.1.tar.gz", hash = "sha256:fd723b723bb7fe57d98b4afbd5f9e1d923d31a9843eff32331725162994ec849"}, + {file = "optimum-1.19.2-py3-none-any.whl", hash = "sha256:66f0fafda050ee6671bab6899852b9bf95afac766d99aa54a40699c7dee598bf"}, + {file = "optimum-1.19.2.tar.gz", hash = "sha256:fc22e07f084d867bd9bce32fd0d737f7c4863514ea5d90c7acccf5dcfe5f2296"}, ] [package.dependencies] @@ -2396,7 +2396,7 @@ exporters-gpu = ["onnx", "onnxruntime-gpu", "timm"] exporters-tf = ["h5py", "numpy (<1.24.0)", "onnx", "onnxruntime", "tensorflow (>=2.4,<=2.12.1)", "tf2onnx", "timm", "transformers[sentencepiece] (>=4.26.0,<4.38.0)"] furiosa = ["optimum-furiosa"] graphcore = ["optimum-graphcore"] -habana = ["optimum-habana", "transformers (>=4.37.0,<4.38.0)"] +habana = ["optimum-habana", "transformers (>=4.38.0,<4.39.0)"] intel = ["optimum-intel (>=1.15.0)"] neural-compressor = ["optimum-intel[neural-compressor] (>=1.15.0)"] neuron = ["optimum-neuron[neuron] (>=0.0.20)", "transformers (==4.36.2)"]