From 561ac8de3e4c6525711802409731eef1281c7b48 Mon Sep 17 00:00:00 2001 From: Yuwen Hu Date: Thu, 30 May 2024 13:21:17 +0800 Subject: [PATCH 1/4] Update dependency to fit latest ipex-llm for embedding supports --- .../llama-index-embeddings-ipex-llm/pyproject.toml | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml index 3db75ee5d7fa52..99df26c1d73eb0 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml @@ -35,15 +35,16 @@ version = "0.1.2" [tool.poetry.dependencies] python = ">=3.9,<4.0" llama-index-core = "^0.10.0" -ipex-llm = {allow-prereleases = true, extras = ["llama-index"], version = ">=2.1.0b20240514"} +ipex-llm = {allow-prereleases = true, extras = ["llama-index"], version = ">=2.1.0b20240529"} torch = {optional = true, source = "ipex-xpu-src-us", version = "2.1.0a0"} torchvision = {optional = true, source = "ipex-xpu-src-us", version = "0.16.0a0"} intel_extension_for_pytorch = {optional = true, source = "ipex-xpu-src-us", version = "2.1.10+xpu"} bigdl-core-xe-21 = {optional = true, version = "*"} -bigdl-core-xe-esimd-21 = {optional = true, version = "*"} +bigdl-core-xe-batch-21 = {optional = true, version = "*"} +bigdl-core-xe-addons-21 = {optional = true, version = "*"} [tool.poetry.extras] -xpu = ["bigdl-core-xe-21", "bigdl-core-xe-esimd-21", "intel_extension_for_pytorch", "torch", "torchvision"] +xpu = ["bigdl-core-xe-21", "bigdl-core-xe-addons-21", "bigdl-core-xe-batch-21", "intel_extension_for_pytorch", "torch", "torchvision"] [tool.poetry.group.dev.dependencies] black = {extras = ["jupyter"], version = "<=23.9.1,>=23.7.0"} From df8791368274f8523db3fdfdb80e6cd06263f358 Mon Sep 17 00:00:00 2001 From: Yuwen Hu Date: Thu, 30 May 2024 13:24:05 +0800 Subject: [PATCH 2/4] bump version for llama-index-ipex-llm-embedding --- .../embeddings/llama-index-embeddings-ipex-llm/pyproject.toml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml index 99df26c1d73eb0..fa59eadeda87c7 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-ipex-llm/pyproject.toml @@ -30,7 +30,7 @@ license = "MIT" name = "llama-index-embeddings-ipex-llm" packages = [{include = "llama_index/"}] readme = "README.md" -version = "0.1.2" +version = "0.1.3" [tool.poetry.dependencies] python = ">=3.9,<4.0" From 39bf054c63d30339b910d39e3219cbbfbba147a4 Mon Sep 17 00:00:00 2001 From: Yuwen Hu Date: Thu, 30 May 2024 13:25:51 +0800 Subject: [PATCH 3/4] Update dependency to fit latest ipex-llm for llm supports --- .../llms/llama-index-llms-ipex-llm/pyproject.toml | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/llama-index-integrations/llms/llama-index-llms-ipex-llm/pyproject.toml b/llama-index-integrations/llms/llama-index-llms-ipex-llm/pyproject.toml index b02b8f6d05361f..c61f063f74e78a 100644 --- a/llama-index-integrations/llms/llama-index-llms-ipex-llm/pyproject.toml +++ b/llama-index-integrations/llms/llama-index-llms-ipex-llm/pyproject.toml @@ -30,20 +30,21 @@ license = "MIT" name = "llama-index-llms-ipex-llm" packages = [{include = "llama_index/"}] readme = "README.md" -version = "0.1.3" +version = "0.1.4" [tool.poetry.dependencies] python = ">=3.9,<4.0" llama-index-core = "^0.10.0" -ipex-llm = {allow-prereleases = true, extras = ["llama-index"], version = ">=2.1.0b20240514"} +ipex-llm = {allow-prereleases = true, extras = ["llama-index"], version = ">=2.1.0b20240529"} torch = {optional = true, source = "ipex-xpu-src-us", version = "2.1.0a0"} torchvision = {optional = true, source = "ipex-xpu-src-us", version = "0.16.0a0"} intel_extension_for_pytorch = {optional = true, source = "ipex-xpu-src-us", version = "2.1.10+xpu"} bigdl-core-xe-21 = {optional = true, version = "*"} -bigdl-core-xe-esimd-21 = {optional = true, version = "*"} +bigdl-core-xe-batch-21 = {optional = true, version = "*"} +bigdl-core-xe-addons-21 = {optional = true, version = "*"} [tool.poetry.extras] -xpu = ["bigdl-core-xe-21", "bigdl-core-xe-esimd-21", "intel_extension_for_pytorch", "torch", "torchvision"] +xpu = ["bigdl-core-xe-21", "bigdl-core-xe-addons-21", "bigdl-core-xe-batch-21", "intel_extension_for_pytorch", "torch", "torchvision"] [tool.poetry.group.dev.dependencies] black = {extras = ["jupyter"], version = "<=23.9.1,>=23.7.0"} From d2ec511b0565fbb30e3d2804eb3cee0c2c19bdc3 Mon Sep 17 00:00:00 2001 From: Yuwen Hu Date: Thu, 30 May 2024 13:36:23 +0800 Subject: [PATCH 4/4] Small doc updates for llm cpu --- docs/docs/examples/llm/ipex_llm.ipynb | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/docs/docs/examples/llm/ipex_llm.ipynb b/docs/docs/examples/llm/ipex_llm.ipynb index bc7d88829e9281..c500e80dda1ba6 100644 --- a/docs/docs/examples/llm/ipex_llm.ipynb +++ b/docs/docs/examples/llm/ipex_llm.ipynb @@ -10,7 +10,9 @@ "\n", "This example goes over how to use LlamaIndex to interact with [`ipex-llm`](https://github.com/intel-analytics/ipex-llm/) for text generation and chat on CPU. \n", "\n", - "For more examples and usage, refer to [Examples](https://github.com/run-llama/llama_index/tree/main/llama-index-integrations/llms/llama-index-llms-ipex-llm/examples)." + "> **Note**\n", + ">\n", + "> You could refer to [here](https://github.com/run-llama/llama_index/tree/main/llama-index-integrations/llms/llama-index-llms-ipex-llm/examples) for full examples of `IpexLLM`. Please note that for running on Intel CPU, please specify `-d 'cpu'` in command argument when running the examples." ] }, {