From f45bb7f3fe559dc33e8aa89d24b98041f8df1eab Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 13 May 2024 12:29:22 +0000 Subject: [PATCH] build(deps): bump optimum from 1.19.1 to 1.19.2 in /runtimes/huggingface (#1754) Bumps [optimum](https://github.com/huggingface/optimum) from 1.19.1 to 1.19.2. - [Release notes](https://github.com/huggingface/optimum/releases) - [Commits](https://github.com/huggingface/optimum/compare/v1.19.1...v1.19.2) --- updated-dependencies: - dependency-name: optimum dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> --- runtimes/huggingface/poetry.lock | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/runtimes/huggingface/poetry.lock b/runtimes/huggingface/poetry.lock index 4926c679f..5f4252bf0 100644 --- a/runtimes/huggingface/poetry.lock +++ b/runtimes/huggingface/poetry.lock @@ -2359,13 +2359,13 @@ tests = ["pytest", "pytest-cov", "pytest-pep8"] [[package]] name = "optimum" -version = "1.19.1" +version = "1.19.2" description = "Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality." optional = false python-versions = ">=3.7.0" files = [ - {file = "optimum-1.19.1-py3-none-any.whl", hash = "sha256:ca474589682fe10f7827c85260d116603a5823d5c251c453620584cbc06fa5f6"}, - {file = "optimum-1.19.1.tar.gz", hash = "sha256:fd723b723bb7fe57d98b4afbd5f9e1d923d31a9843eff32331725162994ec849"}, + {file = "optimum-1.19.2-py3-none-any.whl", hash = "sha256:66f0fafda050ee6671bab6899852b9bf95afac766d99aa54a40699c7dee598bf"}, + {file = "optimum-1.19.2.tar.gz", hash = "sha256:fc22e07f084d867bd9bce32fd0d737f7c4863514ea5d90c7acccf5dcfe5f2296"}, ] [package.dependencies] @@ -2396,7 +2396,7 @@ exporters-gpu = ["onnx", "onnxruntime-gpu", "timm"] exporters-tf = ["h5py", "numpy (<1.24.0)", "onnx", "onnxruntime", "tensorflow (>=2.4,<=2.12.1)", "tf2onnx", "timm", "transformers[sentencepiece] (>=4.26.0,<4.38.0)"] furiosa = ["optimum-furiosa"] graphcore = ["optimum-graphcore"] -habana = ["optimum-habana", "transformers (>=4.37.0,<4.38.0)"] +habana = ["optimum-habana", "transformers (>=4.38.0,<4.39.0)"] intel = ["optimum-intel (>=1.15.0)"] neural-compressor = ["optimum-intel[neural-compressor] (>=1.15.0)"] neuron = ["optimum-neuron[neuron] (>=0.0.20)", "transformers (==4.36.2)"]