diff --git a/CHANGELOG.md b/CHANGELOG.md index 43fc303dfb75e..bfd3c7726ea0d 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,5 +1,82 @@ # ChangeLog +## [2024-03-31] + +### `llama-index-core` [0.10.26] + +- pass proper query bundle in QueryFusionRetriever (#12387) +- Update llama_parse_json_element.py to fix error on lists (#12402) +- Add node postprocessors to retriever tool (#12415) +- Fix bug where user specified llm is not respected in fallback logic in element node parsers(#12403) +- log proper LLM response key for async callback manager events (#12421) +- Deduplicate the two built-in react system prompts; Also make it read from a Markdown file (#12307) +- fix bug in BatchEvalRunner for multi-evaluator eval_kwargs_lists (#12418) +- add the callback manager event for vector store index insert_nodes (#12443) +- fixes an issue with serializing chat messages into chat stores when they contain pydantic API objects (#12394) +- fixes an issue with slow memory.get() operation (caused by multiple calls to get_all()) (#12394) +- fixes an issue where an agent+tool message pair is cut from the memory (#12394) +- Added `FnNodeMapping` for object index (#12391) +- Make object mapping optional / hidden for object index (#12391) +- Make object index easier to create from existing vector db (#12391) +- When LLM failed to follow the react response template, tell it so #12300 + +### `llama-index-embeddings-cohere` [0.1.5] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-embeddings-itrex` [0.1.0] + +- add Intel Extension for Transformers embedding model (#12410) + +### `llama-index-graph-stores-neo4j` [0.1.4] + +- make neo4j query insensitive (#12337) + +### `llama-index-llms-cohere` [0.1.5] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-llms-ipex-llm` [0.1.0] + +- add ipex-llm integration (#12322) + +### `llama-index-llms-litellm` [0.1.4] + +- Fix litellm ChatMessage role validation error (#12449) + +### `llama-index-llms-openai` [0.1.14] + +- Use `FunctionCallingLLM` base class in OpenAI (#12227) + +### `llama-index-packs-self-rag` [0.1.4] + +- Fix llama-index-core dep (#12374) + +### `llama-index-postprocessor-cohere-rerank` [0.1.4] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-postprocessor-rankllm-rerank` [0.1.1] + +- Added RankLLM rerank (#12296) +- RankLLM fixes (#12399) + +### `llama-index-readers-papers` [0.1.4] + +- Fixed bug with path names (#12366) + +### `llama-index-vector-stores-analyticdb` [0.1.1] + +- Add AnalyticDB VectorStore (#12230) + +### `llama-index-vector-stores-kdbai` [0.1.4] + +- Fixed typo in imports/readme (#12370) + +### `llama-index-vector-stores-qdrant` [0.1.5] + +- add `in` filter operator for qdrant (#12376) + ## [2024-03-27] ### `llama-index-core` [0.10.25] diff --git a/docs/docs/CHANGELOG.md b/docs/docs/CHANGELOG.md index 43fc303dfb75e..bfd3c7726ea0d 100644 --- a/docs/docs/CHANGELOG.md +++ b/docs/docs/CHANGELOG.md @@ -1,5 +1,82 @@ # ChangeLog +## [2024-03-31] + +### `llama-index-core` [0.10.26] + +- pass proper query bundle in QueryFusionRetriever (#12387) +- Update llama_parse_json_element.py to fix error on lists (#12402) +- Add node postprocessors to retriever tool (#12415) +- Fix bug where user specified llm is not respected in fallback logic in element node parsers(#12403) +- log proper LLM response key for async callback manager events (#12421) +- Deduplicate the two built-in react system prompts; Also make it read from a Markdown file (#12307) +- fix bug in BatchEvalRunner for multi-evaluator eval_kwargs_lists (#12418) +- add the callback manager event for vector store index insert_nodes (#12443) +- fixes an issue with serializing chat messages into chat stores when they contain pydantic API objects (#12394) +- fixes an issue with slow memory.get() operation (caused by multiple calls to get_all()) (#12394) +- fixes an issue where an agent+tool message pair is cut from the memory (#12394) +- Added `FnNodeMapping` for object index (#12391) +- Make object mapping optional / hidden for object index (#12391) +- Make object index easier to create from existing vector db (#12391) +- When LLM failed to follow the react response template, tell it so #12300 + +### `llama-index-embeddings-cohere` [0.1.5] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-embeddings-itrex` [0.1.0] + +- add Intel Extension for Transformers embedding model (#12410) + +### `llama-index-graph-stores-neo4j` [0.1.4] + +- make neo4j query insensitive (#12337) + +### `llama-index-llms-cohere` [0.1.5] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-llms-ipex-llm` [0.1.0] + +- add ipex-llm integration (#12322) + +### `llama-index-llms-litellm` [0.1.4] + +- Fix litellm ChatMessage role validation error (#12449) + +### `llama-index-llms-openai` [0.1.14] + +- Use `FunctionCallingLLM` base class in OpenAI (#12227) + +### `llama-index-packs-self-rag` [0.1.4] + +- Fix llama-index-core dep (#12374) + +### `llama-index-postprocessor-cohere-rerank` [0.1.4] + +- Bump cohere version to 5.1.1 (#12279) + +### `llama-index-postprocessor-rankllm-rerank` [0.1.1] + +- Added RankLLM rerank (#12296) +- RankLLM fixes (#12399) + +### `llama-index-readers-papers` [0.1.4] + +- Fixed bug with path names (#12366) + +### `llama-index-vector-stores-analyticdb` [0.1.1] + +- Add AnalyticDB VectorStore (#12230) + +### `llama-index-vector-stores-kdbai` [0.1.4] + +- Fixed typo in imports/readme (#12370) + +### `llama-index-vector-stores-qdrant` [0.1.5] + +- add `in` filter operator for qdrant (#12376) + ## [2024-03-27] ### `llama-index-core` [0.10.25] diff --git a/docs/docs/api_reference/embeddings/huggingface_itrex.md b/docs/docs/api_reference/embeddings/huggingface_itrex.md new file mode 100644 index 0000000000000..98c987f9abda1 --- /dev/null +++ b/docs/docs/api_reference/embeddings/huggingface_itrex.md @@ -0,0 +1,4 @@ +::: llama_index.embeddings.huggingface_itrex + options: + members: + - QuantizedBgeEmbedding diff --git a/docs/docs/api_reference/llms/ipex_llm.md b/docs/docs/api_reference/llms/ipex_llm.md new file mode 100644 index 0000000000000..aec44b11d7902 --- /dev/null +++ b/docs/docs/api_reference/llms/ipex_llm.md @@ -0,0 +1,4 @@ +::: llama_index.llms.ipex_llm + options: + members: + - IpexLLM diff --git a/docs/docs/api_reference/postprocessor/rankllm_rerank.md b/docs/docs/api_reference/postprocessor/rankllm_rerank.md new file mode 100644 index 0000000000000..cff89b6afdc50 --- /dev/null +++ b/docs/docs/api_reference/postprocessor/rankllm_rerank.md @@ -0,0 +1,4 @@ +::: llama_index.postprocessor.rankllm_rerank + options: + members: + - CLASS diff --git a/docs/docs/api_reference/storage/vector_store/analyticdb.md b/docs/docs/api_reference/storage/vector_store/analyticdb.md new file mode 100644 index 0000000000000..b9652e5f42489 --- /dev/null +++ b/docs/docs/api_reference/storage/vector_store/analyticdb.md @@ -0,0 +1,4 @@ +::: llama_index.vector_stores.analyticdb + options: + members: + - AnalyticDBVectorStore diff --git a/docs/mkdocs.yml b/docs/mkdocs.yml index 6dc8415fb4062..d0b591a3936db 100644 --- a/docs/mkdocs.yml +++ b/docs/mkdocs.yml @@ -182,6 +182,7 @@ nav: - ./examples/embeddings/llamafile.ipynb - ./examples/embeddings/premai.ipynb - ./examples/embeddings/alephalpha.ipynb + - ./examples/embeddings/itrex.ipynb - Evaluation: - ./examples/evaluation/TonicValidateEvaluators.ipynb - ./examples/evaluation/semantic_similarity_eval.ipynb @@ -291,6 +292,7 @@ nav: - ./examples/llm/premai.ipynb - ./examples/llm/solar.ipynb - ./examples/llm/alephalpha.ipynb + - ./examples/llm/ipex_llm.ipynb - Low Level: - ./examples/low_level/oss_ingestion_retrieval.ipynb - ./examples/low_level/fusion_retriever.ipynb @@ -351,6 +353,7 @@ nav: - ./examples/node_postprocessor/OptimizerDemo.ipynb - ./examples/node_postprocessor/TimeWeightedPostprocessorDemo.ipynb - ./examples/node_postprocessor/JinaRerank.ipynb + - ./examples/node_postprocessor/rankLLM.ipynb - Object Stores: - ./examples/objects/object_index.ipynb - Output Parsers: @@ -676,6 +679,7 @@ nav: - ./api_reference/embeddings/google.md - ./api_reference/embeddings/gradient.md - ./api_reference/embeddings/huggingface.md + - ./api_reference/embeddings/huggingface_itrex.md - ./api_reference/embeddings/huggingface_optimum.md - ./api_reference/embeddings/huggingface_optimum_intel.md - ./api_reference/embeddings/index.md @@ -750,6 +754,7 @@ nav: - ./api_reference/llms/groq.md - ./api_reference/llms/huggingface.md - ./api_reference/llms/index.md + - ./api_reference/llms/ipex_llm.md - ./api_reference/llms/konko.md - ./api_reference/llms/langchain.md - ./api_reference/llms/litellm.md @@ -904,6 +909,7 @@ nav: - ./api_reference/postprocessor/presidio.md - ./api_reference/postprocessor/prev_next.md - ./api_reference/postprocessor/rankgpt_rerank.md + - ./api_reference/postprocessor/rankllm_rerank.md - ./api_reference/postprocessor/sbert_rerank.md - ./api_reference/postprocessor/sentence_optimizer.md - ./api_reference/postprocessor/similarity.md @@ -1185,6 +1191,7 @@ nav: - Storage: - ./api_reference/storage/storage_context.md - Vector Store: + - ./api_reference/storage/vector_store/analyticdb.md - ./api_reference/storage/vector_store/astra_db.md - ./api_reference/storage/vector_store/awadb.md - ./api_reference/storage/vector_store/azureaisearch.md @@ -1763,6 +1770,9 @@ plugins: - ../llama-index-integrations/postprocessor/llama-index-postprocessor-voyageai-rerank - ../llama-index-integrations/readers/llama-index-readers-gcs - ../llama-index-integrations/readers/llama-index-readers-readme + - ../llama-index-integrations/embeddings/llama-index-embeddings-huggingface-itrex + - ../llama-index-integrations/postprocessor/llama-index-postprocessor-rankllm-rerank + - ../llama-index-integrations/llms/llama-index-llms-ipex-llm - redirects: redirect_maps: ./api/llama_index.vector_stores.MongoDBAtlasVectorSearch.html: api_reference/storage/vector_store/mongodb.md diff --git a/llama-index-core/llama_index/core/__init__.py b/llama-index-core/llama_index/core/__init__.py index c66a4cbf3dc8b..52dc645670418 100644 --- a/llama-index-core/llama_index/core/__init__.py +++ b/llama-index-core/llama_index/core/__init__.py @@ -1,6 +1,6 @@ """Init file of LlamaIndex.""" -__version__ = "0.10.25.post3" +__version__ = "0.10.26" import logging from logging import NullHandler diff --git a/llama-index-core/pyproject.toml b/llama-index-core/pyproject.toml index aca720db73b4c..b01d921ea9ca2 100644 --- a/llama-index-core/pyproject.toml +++ b/llama-index-core/pyproject.toml @@ -43,7 +43,7 @@ name = "llama-index-core" packages = [{include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.25.post3" +version = "0.10.26" [tool.poetry.dependencies] SQLAlchemy = {extras = ["asyncio"], version = ">=1.4.49"} diff --git a/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/BUILD b/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/BUILD deleted file mode 100644 index dabf212d7e716..0000000000000 --- a/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/BUILD +++ /dev/null @@ -1 +0,0 @@ -python_tests() diff --git a/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/__init__.py b/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/__init__.py deleted file mode 100644 index e69de29bb2d1d..0000000000000 diff --git a/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/test_llms_ipex_llm.py b/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/test_llms_ipex_llm.py deleted file mode 100644 index 7858f5222caf3..0000000000000 --- a/llama-index-integrations/llms/llama-index-llms-ipex-llm/tests/test_llms_ipex_llm.py +++ /dev/null @@ -1,7 +0,0 @@ -from llama_index.core.base.llms.base import BaseLLM -from llama_index.llms.ipex_llm import IpexLLM - - -def test_embedding_class(): - names_of_base_classes = [b.__name__ for b in IpexLLM.__mro__] - assert BaseLLM.__name__ in names_of_base_classes diff --git a/llama-index-integrations/vector_stores/llama-index-vector-stores-kdbai/pyproject.toml b/llama-index-integrations/vector_stores/llama-index-vector-stores-kdbai/pyproject.toml index f3c95866170b3..5ea635e1d30f9 100644 --- a/llama-index-integrations/vector_stores/llama-index-vector-stores-kdbai/pyproject.toml +++ b/llama-index-integrations/vector_stores/llama-index-vector-stores-kdbai/pyproject.toml @@ -30,7 +30,6 @@ exclude = ["**/BUILD"] license = "MIT" name = "llama-index-vector-stores-kdbai" readme = "README.md" -version = "0.1.3" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" diff --git a/poetry.lock b/poetry.lock index e86f55b6d682a..16172bf9b17b5 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1,4 +1,4 @@ -# This file is automatically @generated by Poetry 1.6.1 and should not be changed by hand. +# This file is automatically @generated by Poetry 1.5.1 and should not be changed by hand. [[package]] name = "aiohttp" @@ -339,20 +339,6 @@ d = ["aiohttp (>=3.7.4)", "aiohttp (>=3.7.4,!=3.9.0)"] jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"] uvloop = ["uvloop (>=0.15.2)"] -[[package]] -name = "bs4" -version = "0.0.2" -description = "Dummy package for Beautiful Soup (beautifulsoup4)" -optional = false -python-versions = "*" -files = [ - {file = "bs4-0.0.2-py2.py3-none-any.whl", hash = "sha256:abf8742c0805ef7f662dce4b51cca104cffe52b835238afc169142ab9b3fbccc"}, - {file = "bs4-0.0.2.tar.gz", hash = "sha256:a48685c58f50fe127722417bae83fe6badf500d54b55f7e39ffe43b798653925"}, -] - -[package.dependencies] -beautifulsoup4 = "*" - [[package]] name = "certifi" version = "2024.2.2" @@ -1185,13 +1171,13 @@ testing = ["jaraco.test (>=5.4)", "pytest (>=6)", "pytest-checkdocs (>=2.4)", "p [[package]] name = "ipykernel" -version = "6.29.3" +version = "6.29.4" description = "IPython Kernel for Jupyter" optional = false python-versions = ">=3.8" files = [ - {file = "ipykernel-6.29.3-py3-none-any.whl", hash = "sha256:5aa086a4175b0229d4eca211e181fb473ea78ffd9869af36ba7694c947302a21"}, - {file = "ipykernel-6.29.3.tar.gz", hash = "sha256:e14c250d1f9ea3989490225cc1a542781b095a18a19447fcf2b5eaf7d0ac5bd2"}, + {file = "ipykernel-6.29.4-py3-none-any.whl", hash = "sha256:1181e653d95c6808039c509ef8e67c4126b3b3af7781496c7cbfb5ed938a27da"}, + {file = "ipykernel-6.29.4.tar.gz", hash = "sha256:3d44070060f9475ac2092b760123fadf105d2e2493c24848b6691a7c4f42af5c"}, ] [package.dependencies] @@ -1518,13 +1504,13 @@ llama-index-llms-openai = ">=0.1.1,<0.2.0" [[package]] name = "llama-index-core" -version = "0.10.25" +version = "0.10.26" description = "Interface between LLMs and your data" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_core-0.10.25-py3-none-any.whl", hash = "sha256:39a0af13f74e57bdf06b86de881cafb80020fc2ae0b0aa0f3a042e17139766ef"}, - {file = "llama_index_core-0.10.25.tar.gz", hash = "sha256:407813e4247704d3cf0957cec772889507e26692d52679a155e22f26efc6aa1b"}, + {file = "llama_index_core-0.10.26-py3-none-any.whl", hash = "sha256:e08886b9d353fb45f63d0011e78326627f1ee5c7761585fdd85a9a0dfdbf0365"}, + {file = "llama_index_core-0.10.26.tar.gz", hash = "sha256:8ea27093a053aed79b7b5d80694152066ddaa5b70116308d48db382c76367a87"}, ] [package.dependencies] @@ -1550,6 +1536,7 @@ tiktoken = ">=0.3.3" tqdm = ">=4.66.1,<5.0.0" typing-extensions = ">=4.5.0" typing-inspect = ">=0.8.0" +wrapt = "*" [package.extras] gradientai = ["gradientai (>=1.4.0)"] @@ -1629,13 +1616,13 @@ query-tools = ["guidance (>=0.0.64,<0.0.65)", "jsonpath-ng (>=1.6.0,<2.0.0)", "l [[package]] name = "llama-index-llms-openai" -version = "0.1.13" +version = "0.1.14" description = "llama-index llms openai integration" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_llms_openai-0.1.13-py3-none-any.whl", hash = "sha256:84b7f2d1699d882d6a92f7e8a8b203701e9a32e42924e01bccabddbe9955d3f7"}, - {file = "llama_index_llms_openai-0.1.13.tar.gz", hash = "sha256:c0fd932255ac9bf72b6b02c3811eebbf3431aa7603aeaab31c811547f444b1ca"}, + {file = "llama_index_llms_openai-0.1.14-py3-none-any.whl", hash = "sha256:13cec467962a6ccb9e63451c7febe8e9c2ed536bd6b1058239c2b4fd86776060"}, + {file = "llama_index_llms_openai-0.1.14.tar.gz", hash = "sha256:7eba66882ae84fa42b188941234b84267c48e449ef214f511756dcad3f9f0b62"}, ] [package.dependencies] @@ -1690,18 +1677,17 @@ llama-index-program-openai = ">=0.1.1,<0.2.0" [[package]] name = "llama-index-readers-file" -version = "0.1.12" +version = "0.1.13" description = "llama-index readers file integration" optional = false python-versions = "<4.0,>=3.8.1" files = [ - {file = "llama_index_readers_file-0.1.12-py3-none-any.whl", hash = "sha256:029eecb9371c6c621cce026fa987212a55160c6cceafef5f470366ec410cc148"}, - {file = "llama_index_readers_file-0.1.12.tar.gz", hash = "sha256:6065cbf80b0fb5d189558b8b90adbbdc92acb86147fca192d88fcc2704ad2af3"}, + {file = "llama_index_readers_file-0.1.13-py3-none-any.whl", hash = "sha256:692988b8c3ca2807d21a171351078b634ada2ff1682ebe4a484f82da384dfc55"}, + {file = "llama_index_readers_file-0.1.13.tar.gz", hash = "sha256:830f06ec7b34437fc3bb5f268d235c5c7640296adb148d8f92277dceb7f0846d"}, ] [package.dependencies] beautifulsoup4 = ">=4.12.3,<5.0.0" -bs4 = ">=0.0.2,<0.0.3" llama-index-core = ">=0.10.1,<0.11.0" pymupdf = ">=1.23.21,<2.0.0" pypdf = ">=4.0.1,<5.0.0" @@ -2377,7 +2363,7 @@ files = [ [package.dependencies] numpy = [ {version = ">=1.20.3", markers = "python_version < \"3.10\""}, - {version = ">=1.21.0", markers = "python_version >= \"3.10\" and python_version < \"3.11\""}, + {version = ">=1.21.0", markers = "python_version >= \"3.10\""}, {version = ">=1.23.2", markers = "python_version >= \"3.11\""}, ] python-dateutil = ">=2.8.2" @@ -2656,13 +2642,13 @@ tests = ["pytest"] [[package]] name = "pycparser" -version = "2.21" +version = "2.22" description = "C parser in Python" optional = false -python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*" +python-versions = ">=3.8" files = [ - {file = "pycparser-2.21-py2.py3-none-any.whl", hash = "sha256:8ee45429555515e1f6b185e78100aea234072576aa43ab53aefcae078162fca9"}, - {file = "pycparser-2.21.tar.gz", hash = "sha256:e644fdec12f7872f86c58ff790da456218b10f863970249516d60a5eaca77206"}, + {file = "pycparser-2.22-py3-none-any.whl", hash = "sha256:c3702b6d3dd8c7abc1afa565d7e63d53a1d0bd86cdc24edd75470f4de499cfcc"}, + {file = "pycparser-2.22.tar.gz", hash = "sha256:491c8be9c040f5390f5bf44a5b07752bd07f56edf992381b05c701439eec10f6"}, ] [[package]] @@ -3679,7 +3665,7 @@ files = [ ] [package.dependencies] -greenlet = {version = "!=0.4.17", optional = true, markers = "platform_machine == \"aarch64\" or platform_machine == \"ppc64le\" or platform_machine == \"x86_64\" or platform_machine == \"amd64\" or platform_machine == \"AMD64\" or platform_machine == \"win32\" or platform_machine == \"WIN32\" or extra == \"asyncio\""} +greenlet = {version = "!=0.4.17", optional = true, markers = "platform_machine == \"win32\" or platform_machine == \"WIN32\" or platform_machine == \"AMD64\" or platform_machine == \"amd64\" or platform_machine == \"x86_64\" or platform_machine == \"ppc64le\" or platform_machine == \"aarch64\" or extra == \"asyncio\""} typing-extensions = ">=4.6.0" [package.extras] @@ -3918,13 +3904,13 @@ files = [ [[package]] name = "types-docutils" -version = "0.20.0.20240317" +version = "0.20.0.20240331" description = "Typing stubs for docutils" optional = false python-versions = ">=3.8" files = [ - {file = "types-docutils-0.20.0.20240317.tar.gz", hash = "sha256:23657aab0de58634d111914b677b1855867f16cd9a9ea110254e23b48653e1a8"}, - {file = "types_docutils-0.20.0.20240317-py3-none-any.whl", hash = "sha256:4f11b3986b74f39169313ab528ffac101c45fca9c36c4cd22dbeec6143c99b7f"}, + {file = "types-docutils-0.20.0.20240331.tar.gz", hash = "sha256:ac99cdf34040c982081f54237d6017f8f5dafe0bebb818a598bf97a65f5b1715"}, + {file = "types_docutils-0.20.0.20240331-py3-none-any.whl", hash = "sha256:b9042e1cf064b4a82c87a71ed3c5f0f96e81fb6d402ca4daa6ced65a91397679"}, ] [[package]] @@ -4302,4 +4288,4 @@ testing = ["big-O", "jaraco.functools", "jaraco.itertools", "more-itertools", "p [metadata] lock-version = "2.0" python-versions = ">=3.8.1,<4.0" -content-hash = "a2fdd83822152319bfaee4601b086157de06e39d4fe56bb44d1896c1247711c0" +content-hash = "5e2631ae424d8c1b6fa9e7d93d731237e13e78be6d0da183ef8f1755f56be9c5" diff --git a/pyproject.toml b/pyproject.toml index 6337c7f84c9b6..b57bc3235a7a5 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -44,7 +44,7 @@ name = "llama-index" packages = [{from = "_llama-index", include = "llama_index"}] readme = "README.md" repository = "https://github.com/run-llama/llama_index" -version = "0.10.25" +version = "0.10.26" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" @@ -57,7 +57,7 @@ llama-index-agent-openai = ">=0.1.4,<0.3.0" llama-index-readers-file = "^0.1.4" llama-index-readers-llama-parse = "^0.1.2" llama-index-indices-managed-llama-cloud = "^0.1.2" -llama-index-core = "^0.10.25" +llama-index-core = "^0.10.26" llama-index-multi-modal-llms-openai = "^0.1.3" llama-index-cli = "^0.1.2"