From 624c721ac4e7b22e9da4a650f3e06e414ba91972 Mon Sep 17 00:00:00 2001 From: Marcus Schiesser <mail@marcusschiesser.de> Date: Mon, 15 Jul 2024 16:51:15 +0200 Subject: [PATCH] chore: update to llamaindex 0.10.55 (#165) --- .changeset/curly-lemons-confess.md | 5 +++++ helpers/python.ts | 2 +- templates/types/streaming/fastapi/pyproject.toml | 2 +- 3 files changed, 7 insertions(+), 2 deletions(-) create mode 100644 .changeset/curly-lemons-confess.md diff --git a/.changeset/curly-lemons-confess.md b/.changeset/curly-lemons-confess.md new file mode 100644 index 00000000..9733dad0 --- /dev/null +++ b/.changeset/curly-lemons-confess.md @@ -0,0 +1,5 @@ +--- +"create-llama": patch +--- + +Update to LlamaIndex 0.10.55 diff --git a/helpers/python.ts b/helpers/python.ts index fd755440..c3f8037c 100644 --- a/helpers/python.ts +++ b/helpers/python.ts @@ -121,7 +121,7 @@ const getAdditionalDependencies = ( case "llamacloud": dependencies.push({ name: "llama-index-indices-managed-llama-cloud", - version: "^0.2.1", + version: "^0.2.5", }); break; } diff --git a/templates/types/streaming/fastapi/pyproject.toml b/templates/types/streaming/fastapi/pyproject.toml index e5e4d058..c669162e 100644 --- a/templates/types/streaming/fastapi/pyproject.toml +++ b/templates/types/streaming/fastapi/pyproject.toml @@ -14,7 +14,7 @@ fastapi = "^0.109.1" uvicorn = { extras = ["standard"], version = "^0.23.2" } python-dotenv = "^1.0.0" aiostream = "^0.5.2" -llama-index = "0.10.54post1" +llama-index = "0.10.55" cachetools = "^5.3.3" [build-system] -- GitLab