From 910db520492b7cc5bb7d11d9d072f09c6ba62727 Mon Sep 17 00:00:00 2001 From: Logan Date: Mon, 3 Jun 2024 09:09:38 -0600 Subject: [PATCH] v0.10.43 (#13902) --- CHANGELOG.md | 28 ++++++++++++++++ docs/docs/CHANGELOG.md | 28 ++++++++++++++++ llama-index-core/llama_index/core/__init__.py | 2 +- llama-index-core/pyproject.toml | 2 +- poetry.lock | 32 +++++++++---------- pyproject.toml | 4 +-- 6 files changed, 76 insertions(+), 20 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 5c5d4b193d616..23b6ec6342af1 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,33 @@ # ChangeLog +## [2024-06-02] + +### `llama-index-core` [0.10.43] + +- use default UUIDs when possible for property graph index vector stores (#13886) +- avoid empty or duplicate inserts in property graph index (#13891) +- Fix cur depth for `get_rel_map` in simple property graph store (#13888) +- (bandaid) disable instrumentation from logging generators (#13901) +- Add backwards compatibility to Dispatcher.get_dispatch_event() method (#13895) +- Fix: Incorrect naming of acreate_plan in StructuredPlannerAgent (#13879) + +### `llama-index-graph-stores-neo4j` [0.2.2] + +- Handle cases where type is missing (neo4j property graph) (#13875) +- Rename `Neo4jPGStore` to `Neo4jPropertyGraphStore` (with backward compat) (#13891) + +### `llama-index-llms-openai` [0.1.22] + +- Improve the retry mechanism of OpenAI (#13878) + +### `llama-index-readers-web` [0.1.18] + +- AsyncWebPageReader: made it actually async; it was exhibiting blocking behavior (#13897) + +### `llama-index-vector-stores-opensearch` [0.1.10] + +- Fix/OpenSearch filter logic (#13804) + ## [2024-05-31] ### `llama-index-core` [0.10.42] diff --git a/docs/docs/CHANGELOG.md b/docs/docs/CHANGELOG.md index 5c5d4b193d616..23b6ec6342af1 100644 --- a/docs/docs/CHANGELOG.md +++ b/docs/docs/CHANGELOG.md @@ -1,5 +1,33 @@ # ChangeLog +## [2024-06-02] + +### `llama-index-core` [0.10.43] + +- use default UUIDs when possible for property graph index vector stores (#13886) +- avoid empty or duplicate inserts in property graph index (#13891) +- Fix cur depth for `get_rel_map` in simple property graph store (#13888) +- (bandaid) disable instrumentation from logging generators (#13901) +- Add backwards compatibility to Dispatcher.get_dispatch_event() method (#13895) +- Fix: Incorrect naming of acreate_plan in StructuredPlannerAgent (#13879) + +### `llama-index-graph-stores-neo4j` [0.2.2] + +- Handle cases where type is missing (neo4j property graph) (#13875) +- Rename `Neo4jPGStore` to `Neo4jPropertyGraphStore` (with backward compat) (#13891) + +### `llama-index-llms-openai` [0.1.22] + +- Improve the retry mechanism of OpenAI (#13878) + +### `llama-index-readers-web` [0.1.18] + +- AsyncWebPageReader: made it actually async; it was exhibiting blocking behavior (#13897) + +### `llama-index-vector-stores-opensearch` [0.1.10] + +- Fix/OpenSearch filter logic (#13804) + ## [2024-05-31] ### `llama-index-core` [0.10.42] diff --git a/llama-index-core/llama_index/core/__init__.py b/llama-index-core/llama_index/core/__init__.py index 3273f71426a20..ab66be34e8b83 100644 --- a/llama-index-core/llama_index/core/__init__.py +++ b/llama-index-core/llama_index/core/__init__.py @@ -1,6 +1,6 @@ """Init file of LlamaIndex.""" -__version__ = "0.10.42" +__version__ = "0.10.43" import logging from logging import NullHandler diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index 55dca9770c289..c30ef192d4172 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -43,7 +43,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.42" +version = "0.10.43" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/poetry.lock b/poetry.lock index ec22b209350f8..91851b472b5b6 100644 --- a/poetry.lock +++ b/poetry.lock @@ -356,13 +356,13 @@ css = ["tinycss2 (>=1.1.0,<1.3)"] [[package]] name = "certifi" -version = "2024.2.2" +version = "2024.6.2" description = "Python package for providing Mozilla's CA Bundle." optional = false python-versions = ">=3.6" files = [ - {file = "certifi-2024.2.2-py3-none-any.whl", hash = "sha256:dc383c07b76109f368f6106eee2b593b04a011ea4d55f652c6ca24a754d1cdd1"}, - {file = "certifi-2024.2.2.tar.gz", hash = "sha256:0569859f95fc761b18b45ef421b1290a0f65f147e92a1e5eb3e635f9a5e4e66f"}, + {file = "certifi-2024.6.2-py3-none-any.whl", hash = "sha256:ddc6c8ce995e6987e7faf5e3f1b02b302836a0e5d98ece18392cb1a36c72ad56"}, + {file = "certifi-2024.6.2.tar.gz", hash = "sha256:3cd43f1c6fa7dedc5899d69d3ad0398fd018ad1a17fba83ddaf78aa46c747516"}, ] [[package]] @@ -1504,13 +1504,13 @@ files = [ [[package]] name = "llama-index-agent-openai" -version = "0.2.6" +version = "0.2.7" description = "llama-index agent openai integration" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_agent_openai-0.2.6-py3-none-any.whl", hash = "sha256:fddf5a21b5cc584d4aab4517edcdd7ccdba7bf2b66ba6f1ae31d6b87800b4246"}, - {file = "llama_index_agent_openai-0.2.6.tar.gz", hash = "sha256:32d95ab8ebd8fb8174e34427e6cfcddb56003da7029be9a1d3fe002540d13b9a"}, + {file = "llama_index_agent_openai-0.2.7-py3-none-any.whl", hash = "sha256:34be65011a508dd8cab0c9a606594f28075b98b0cebe69e3c543adc8564fee0d"}, + {file = "llama_index_agent_openai-0.2.7.tar.gz", hash = "sha256:13ce535f03e32c821763c01e26af4222f3981178622414d3868013a1946e8124"}, ] [package.dependencies] @@ -1536,13 +1536,13 @@ llama-index-llms-openai = ">=0.1.1,<0.2.0" [[package]] name = "llama-index-core" -version = "0.10.42" +version = "0.10.43" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_core-0.10.42-py3-none-any.whl", hash = "sha256:31825850aace51891275cffd7ff5bb1dbb6a5adbb18a55d2fa8f66ad70aa0128"}, - {file = "llama_index_core-0.10.42.tar.gz", hash = "sha256:6bcbfefd776ae461552494ac296ccfffa1d0110b98707ec56fd58bb2387f50fa"}, + {file = "llama_index_core-0.10.43-py3-none-any.whl", hash = "sha256:355a5f7808dfa09dc27b6bd564926fa697e0279344b829732a9f91b8000157ac"}, + {file = "llama_index_core-0.10.43.tar.gz", hash = "sha256:c7dce5158e112d3cbc8d331c303e830b2eabcae55c9636e8050dc85320522c1b"}, ] [package.dependencies] @@ -1640,13 +1640,13 @@ query-tools = ["guidance (>=0.0.64,<0.0.65)", "jsonpath-ng (>=1.6.0,<2.0.0)", "l [[package]] name = "llama-index-llms-openai" -version = "0.1.21" +version = "0.1.22" description = "llama-index llms openai integration" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_llms_openai-0.1.21-py3-none-any.whl", hash = "sha256:fd2fa240238c7dca0170bf55957bb088d9b6acf63c80924eb59498f3a5ac0df3"}, - {file = "llama_index_llms_openai-0.1.21.tar.gz", hash = "sha256:e9d631135160ac87093cd7d5ae3e32c4debf8eeaafa29ee4739675867155d773"}, + {file = "llama_index_llms_openai-0.1.22-py3-none-any.whl", hash = "sha256:84a8c910671460ad724ed818192f209f7481e71bcc6528553ba7e66db2e14bcd"}, + {file = "llama_index_llms_openai-0.1.22.tar.gz", hash = "sha256:729bf2ea7043517465e1d585089512b77d8b3ce92233a67c138d5d621061ed56"}, ] [package.dependencies] @@ -4088,13 +4088,13 @@ files = [ [[package]] name = "typing-extensions" -version = "4.12.0" +version = "4.12.1" description = "Backported and Experimental Type Hints for Python 3.8+" optional = false python-versions = ">=3.8" files = [ - {file = "typing_extensions-4.12.0-py3-none-any.whl", hash = "sha256:b349c66bea9016ac22978d800cfff206d5f9816951f12a7d0ec5578b0a819594"}, - {file = "typing_extensions-4.12.0.tar.gz", hash = "sha256:8cbcdc8606ebcb0d95453ad7dc5065e6237b6aa230a31e81d0f440c30fed5fd8"}, + {file = "typing_extensions-4.12.1-py3-none-any.whl", hash = "sha256:6024b58b69089e5a89c347397254e35f1bf02a907728ec7fee9bf0fe837d203a"}, + {file = "typing_extensions-4.12.1.tar.gz", hash = "sha256:915f5e35ff76f56588223f15fdd5938f9a1cf9195c0de25130c627e4d597f6d1"}, ] [[package]] @@ -4440,4 +4440,4 @@ test = ["big-O", "jaraco.functools", "jaraco.itertools", "jaraco.test", "more-it [metadata] lock-version = "2.0" python-versions = ">=3.8.1,<4.0" -content-hash = "b4ec495bd49bb8ff53dec13714555c4edf7812200eeb2ec946879f9563306130" +content-hash = "c4e17dac95167cb9096dfb2ed7df5ca72e7f753be5c320f5a64a0cb5a6ce930d" diff --git a/pyproject.toml b/pyproject.toml index 032bdd54b7ad5..cfd35c7ce398b 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -44,7 +44,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.42" +version = "0.10.43" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" @@ -57,7 +57,7 @@ llama-index-agent-openai = ">=0.1.4,<0.3.0" llama-index-readers-file = "^0.1.4" llama-index-readers-llama-parse = "^0.1.2" llama-index-indices-managed-llama-cloud = "^0.1.2" -llama-index-core = "0.10.42" +llama-index-core = "0.10.43" llama-index-multi-modal-llms-openai = "^0.1.3" llama-index-cli = "^0.1.2"