From d0c3704eee39c541c03a6c430dc7c8dcbee15faf Mon Sep 17 00:00:00 2001 From: Logan Markewich Date: Thu, 15 Feb 2024 20:33:12 -0600 Subject: [PATCH] final version bumps --- llama-index-core/pyproject.toml | 2 +- .../llama_index/finetuning/embeddings/adapter.py | 4 ++-- .../finetuning/embeddings/adapter_utils.py | 2 +- llama-index-finetuning/pyproject.toml | 4 ++-- .../llama_index/embeddings/adapter/__init__.py | 13 +++++++++++-- .../llama-index-embeddings-adapter/pyproject.toml | 2 +- pyproject.toml | 2 +- 7 files changed, 19 insertions(+), 10 deletions(-) diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index a59ec960cbeee..fd64f393c9ab5 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -42,7 +42,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.3" +version = "0.10.5" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py index 1d1a394832ef3..b719ae619d56f 100644 --- a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py +++ b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter.py @@ -50,7 +50,7 @@ def __init__( ) -> None: """Init params.""" import torch - from llama_index.finetuning.embeddings.adapter_utils import ( + from llama_index.embeddings.adapter import ( BaseAdapter, LinearLayer, ) @@ -108,7 +108,7 @@ def from_model_path( **kwargs (Any): Additional kwargs (see __init__) """ - from llama_index.finetuning.embeddings.adapter_utils import LinearLayer + from llama_index.embeddings.adapter import LinearLayer model_cls = model_cls or LinearLayer model = model_cls.load(model_path) diff --git a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py index 495f6689c9b02..5f51874f8dae3 100644 --- a/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py +++ b/llama-index-finetuning/llama_index/finetuning/embeddings/adapter_utils.py @@ -6,7 +6,7 @@ import torch import transformers from llama_index.core.utils import print_text -from llama_index.embeddings.adapter.utils import BaseAdapter +from llama_index.embeddings.adapter import BaseAdapter from sentence_transformers.util import cos_sim from torch import Tensor, nn from torch.optim import Optimizer diff --git a/llama-index-finetuning/pyproject.toml b/llama-index-finetuning/pyproject.toml index b073a6e23459c..6c5b63661f15c 100644 --- a/llama-index-finetuning/pyproject.toml +++ b/llama-index-finetuning/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index finetuning" license = "MIT" name = "llama-index-finetuning" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" @@ -32,7 +32,7 @@ llama-index-core = "^0.10.1" llama-index-llms-openai = "^0.1.1" llama-index-llms-gradient = "^0.1.1" llama-index-postprocessor-cohere-rerank = "^0.1.1" -llama-index-embeddings-adapter = "^0.1.1" +llama-index-embeddings-adapter = "^0.1.2" sentence-transformers = "^2.3.0" [tool.poetry.group.dev.dependencies] diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py index 2376cd580b2d7..bbdfb36e5536a 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py +++ b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/llama_index/embeddings/adapter/__init__.py @@ -1,3 +1,12 @@ -from llama_index.embeddings.adapter.base import AdapterEmbeddingModel +from llama_index.embeddings.adapter.base import ( + AdapterEmbeddingModel, + LinearAdapterEmbeddingModel, +) +from llama_index.embeddings.adapter.utils import BaseAdapter, LinearLayer -__all__ = ["AdapterEmbeddingModel"] +__all__ = [ + "AdapterEmbeddingModel", + "LinearAdapterEmbeddingModel", + "BaseAdapter", + "LinearLayer", +] diff --git a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml index 9a8c719a27aa8..c3c2fc05ac5a7 100644 --- a/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml +++ b/llama-index-integrations/embeddings/llama-index-embeddings-adapter/pyproject.toml @@ -24,7 +24,7 @@ description = "llama-index embeddings adapter integration" license = "MIT" name = "llama-index-embeddings-adapter" readme = "README.md" -version = "0.1.1" +version = "0.1.2" [tool.poetry.dependencies] python = ">=3.8.1,<3.12" diff --git a/pyproject.toml b/pyproject.toml index 619039efd0206..aff6f76f1c2dd 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -41,7 +41,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.3" +version = "0.10.5" [tool.poetry.dependencies] python = ">=3.8.1,<4.0"