diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index a59ec960cbeee3..fd64f393c9ab5d 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -42,7 +42,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.3" +version = "0.10.5" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py index 1d1a394832ef3d..b719ae619d56f8 100644 --- a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py +++ b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py @@ -50,7 +50,7 @@ def __init__( ) -> None: """Init params.""" import torch - from llama_index.finetuning.embeddings.adapter_utils import ( + from llama_index.embeddings.adapter import ( BaseAdapter, LinearLayer, ) @@ -108,7 +108,7 @@ def from_model_path( **kwargs (Any): Additional kwargs (see __init__) """ - from llama_index.finetuning.embeddings.adapter_utils import LinearLayer + from llama_index.embeddings.adapter import LinearLayer model_cls = model_cls or LinearLayer model = model_cls.load(model_path) diff --git a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py index 495f6689c9b025..5f51874f8dae35 100644 --- a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py +++ b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py @@ -6,7 +6,7 @@ import torch import transformers from llama_index.core.utils import print_text -from llama_index.embeddings.adapter.utils import BaseAdapter +from llama_index.embeddings.adapter import BaseAdapter from sentence_transformers.util import cos_sim from torch import Tensor, nn from torch.optim import Optimizer diff --git a/llama-index-finetuning/pyproject.toml b/llama-index-finetuning/pyproject.toml index b073a6e23459ce..6c5b63661f15c6 100644 --- a/llama-index-finetuning/pyproject.toml +++ b/llama-index-finetuning/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index finetuning" license = "MIT" name = "llama-index-finetuning" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" @@ -32,7 +32,7 @@ llama-index-core = "^0.10.1" llama-index-llms-openai = "^0.1.1" llama-index-llms-gradient = "^0.1.1" llama-index-postprocessor-cohere-rerank = "^0.1.1" -llama-index-embeddings-adapter = "^0.1.1" +llama-index-embeddings-adapter = "^0.1.2" sentence-transformers = "^2.3.0" [tool.poetry.group.dev.dependencies] diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py index 2376cd580b2d71..bbdfb36e5536a3 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py +++ b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py @@ -1,3 +1,12 @@ -from llama_index.embeddings.adapter.base import AdapterEmbeddingModel +from llama_index.embeddings.adapter.base import ( + AdapterEmbeddingModel, + LinearAdapterEmbeddingModel, +) +from llama_index.embeddings.adapter.utils import BaseAdapter, LinearLayer -__all__ = ["AdapterEmbeddingModel"] +__all__ = [ + "AdapterEmbeddingModel", + "LinearAdapterEmbeddingModel", + "BaseAdapter", + "LinearLayer", +] diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml index 9a8c719a27aa8e..c3c2fc05ac5a7d 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index embeddings adapter integration" license = "MIT" name = "llama-index-embeddings-adapter" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-nomic/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-nomic/pyproject.toml index 9bf3662c665e0f..e6e3b0c19139a7 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-nomic/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-nomic/pyproject.toml @@ -8,7 +8,7 @@ check-hidden = true skip = "*.csv,*.html,*.json,*.jsonl,*.pdf,*.txt,*.ipynb" [tool.llamahub] -classes = ["NomicEmbedding"] +classes = ["NomicEmbedding", "NomicHFEmbedding"] contains_example = false import_path = "llama_index.embeddings.nomic" @@ -24,7 +24,7 @@ description = "llama-index embeddings nomic integration" license = "MIT" name = "llama-index-embeddings-nomic" readme = "README.md" -version = "0.1.1" +version = "0.1.3" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/__init__.cpython-311.pyc b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/__init__.cpython-311.pyc new file mode 100644 index 00000000000000..cc9463641386fa Binary files /dev/null and b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/__init__.cpython-311.pyc differ diff --git a/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/base.cpython-311.pyc b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/base.cpython-311.pyc new file mode 100644 index 00000000000000..6f4558c12f4f55 Binary files /dev/null and b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/base.cpython-311.pyc differ diff --git a/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/retriever.cpython-311.pyc b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/retriever.cpython-311.pyc new file mode 100644 index 00000000000000..e694d11789e609 Binary files /dev/null and b/llama-index-integrations/indices/llama-index-indices-managed-llama-cloud/llama_index/indices/managed/llama_cloud/__pycache__/retriever.cpython-311.pyc differ diff --git a/llama-index-integrations/llms/llama-index-llms-nvidia-tensorrt/pyproject.toml b/llama-index-integrations/llms/llama-index-llms-nvidia-tensorrt/pyproject.toml index a76c0adfc15764..8620cd2ef1fee4 100644 --- a/llama-index-integrations/llms/llama-index-llms-nvidia-tensorrt/pyproject.toml +++ b/llama-index-integrations/llms/llama-index-llms-nvidia-tensorrt/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index llms nvidia tensorrt integration" license = "MIT" name = "llama-index-llms-nvidia-tensorrt" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/llama-index-integrations/llms/llama-index-llms-vllm/pyproject.toml b/llama-index-integrations/llms/llama-index-llms-vllm/pyproject.toml index bd5499519609c1..3e08676380b091 100644 --- a/llama-index-integrations/llms/llama-index-llms-vllm/pyproject.toml +++ b/llama-index-integrations/llms/llama-index-llms-vllm/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index llms vllm integration" license = "MIT" name = "llama-index-llms-vllm" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/llama-index-integrations/readers/llama-index-readers-minio/README.md b/llama-index-integrations/readers/llama-index-readers-minio/README.md new file mode 100644 index 00000000000000..d0bb5d20dd2d13 --- /dev/null +++ b/llama-index-integrations/readers/llama-index-readers-minio/README.md @@ -0,0 +1,9 @@ +# Minio Reader + +## Install + +`pip install llama-index-readers-minio` + +## Import + +`from llama_index.readers.minio import MinioReader, BotoMinioReader` diff --git a/llama-index-integrations/vector_stores/llama-index-vector-stores-mongodb/pyproject.toml b/llama-index-integrations/vector_stores/llama-index-vector-stores-mongodb/pyproject.toml index b9d0fc834ab4c3..de01ba5aae5e6c 100644 --- a/llama-index-integrations/vector_stores/llama-index-vector-stores-mongodb/pyproject.toml +++ b/llama-index-integrations/vector_stores/llama-index-vector-stores-mongodb/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index vector_stores mongodb integration" license = "MIT" name = "llama-index-vector-stores-mongodb" readme = "README.md" -version = "0.1.2" +version = "0.1.3" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/pyproject.toml b/pyproject.toml index 619039efd02069..aff6f76f1c2dd2 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -41,7 +41,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.3" +version = "0.10.5" [tool.poetry.dependencies] python = ">=3.8.1,<4.0"