diff --git a/runtimes/huggingface/poetry.lock b/runtimes/huggingface/poetry.lock index ed8d91efc..142b3f91a 100644 --- a/runtimes/huggingface/poetry.lock +++ b/runtimes/huggingface/poetry.lock @@ -2080,12 +2080,13 @@ tests = ["pytest", "pytest-cov", "pytest-pep8"] [[package]] name = "optimum" -version = "1.9.1" +version = "1.11.0" description = "Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality." optional = false python-versions = ">=3.7.0" files = [ - {file = "optimum-1.9.1.tar.gz", hash = "sha256:77c7c19527652d59351c8d4eac517d694c64b316f4cb5e292a63e21c2503c304"}, + {file = "optimum-1.11.0-py3-none-any.whl", hash = "sha256:2d1d6202aff0b8c2978818991c8974a7b3672a25fc5b54498d43936cd065d021"}, + {file = "optimum-1.11.0.tar.gz", hash = "sha256:ba4821db92fa0394d90971b0c23ded612095dec5192921f35ad370f6dd41f991"}, ] [package.dependencies] @@ -2095,10 +2096,10 @@ datasets = [ {version = ">=1.2.1", optional = true, markers = "extra == \"onnxruntime\""}, ] evaluate = {version = "*", optional = true, markers = "extra == \"onnxruntime\""} -huggingface_hub = ">=0.8.0" +huggingface-hub = ">=0.8.0" numpy = "*" onnx = {version = "*", optional = true, markers = "extra == \"onnxruntime\""} -onnxruntime = {version = ">=1.9.0", optional = true, markers = "extra == \"onnxruntime\""} +onnxruntime = {version = ">=1.11.0", optional = true, markers = "extra == \"onnxruntime\""} packaging = "*" protobuf = {version = ">=3.20.1", optional = true, markers = "extra == \"onnxruntime\""} sympy = "*" @@ -2107,22 +2108,24 @@ transformers = {version = ">=4.26.0", extras = ["sentencepiece"]} [package.extras] benchmark = ["evaluate (>=0.2.0)", "optuna", "scikit-learn", "seqeval", "torchvision", "tqdm"] -dev = ["Pillow", "black (>=23.1,<24.0)", "diffusers (>=0.17.0)", "parameterized", "pytest", "pytest-xdist", "requests", "ruff (>=0.0.241,<=0.0.259)", "sacremoses", "torchaudio", "torchvision"] +dev = ["Pillow", "black (>=23.1,<24.0)", "diffusers (>=0.17.0)", "einops", "invisible-watermark", "parameterized", "pytest", "pytest-xdist", "requests", "ruff (>=0.0.241,<=0.0.259)", "sacremoses", "torchaudio", "torchvision"] +diffusers = ["diffusers"] exporters = ["onnx", "onnxruntime", "timm"] exporters-gpu = ["onnx", "onnxruntime-gpu", "timm"] -exporters-tf = ["h5py", "numpy (<1.24.0)", "onnx", "onnxruntime", "tensorflow (>=2.4,<2.11)", "tf2onnx", "timm"] +exporters-tf = ["h5py", "numpy (<1.24.0)", "onnx", "onnxruntime", "tensorflow (>=2.4)", "tf2onnx", "timm"] +furiosa = ["optimum-furiosa"] graphcore = ["optimum-graphcore"] habana = ["optimum-habana", "transformers (<4.29.0)"] -intel = ["optimum-intel"] -neural-compressor = ["optimum-intel[neural-compressor]"] +intel = ["optimum-intel (>=1.10.1)"] +neural-compressor = ["optimum-intel[neural-compressor] (>=1.9.2)"] neuron = ["optimum-neuron[neuron]"] neuronx = ["optimum-neuron[neuronx]"] -nncf = ["optimum-intel[nncf]"] -onnxruntime = ["datasets (>=1.2.1)", "evaluate", "onnx", "onnxruntime (>=1.9.0)", "protobuf (>=3.20.1)"] -onnxruntime-gpu = ["datasets (>=1.2.1)", "evaluate", "onnx", "onnxruntime-gpu (>=1.9.0)", "protobuf (>=3.20.1)"] -openvino = ["optimum-intel[openvino]"] +nncf = ["optimum-intel[nncf] (>=1.10.1)"] +onnxruntime = ["datasets (>=1.2.1)", "evaluate", "onnx", "onnxruntime (>=1.11.0)", "protobuf (>=3.20.1)"] +onnxruntime-gpu = ["datasets (>=1.2.1)", "evaluate", "onnx", "onnxruntime-gpu (>=1.11.0)", "protobuf (>=3.20.1)"] +openvino = ["optimum-intel[openvino] (>=1.10.1)"] quality = ["black (>=23.1,<24.0)", "ruff (>=0.0.241,<=0.0.259)"] -tests = ["Pillow", "diffusers (>=0.17.0)", "parameterized", "pytest", "pytest-xdist", "requests", "sacremoses", "torchaudio", "torchvision"] +tests = ["Pillow", "diffusers (>=0.17.0)", "einops", "invisible-watermark", "parameterized", "pytest", "pytest-xdist", "requests", "sacremoses", "torchaudio", "torchvision"] [[package]] name = "orjson"