From 624c721ac4e7b22e9da4a650f3e06e414ba91972 Mon Sep 17 00:00:00 2001
From: Marcus Schiesser <mail@marcusschiesser.de>
Date: Mon, 15 Jul 2024 16:51:15 +0200
Subject: [PATCH] chore: update to llamaindex 0.10.55 (#165)

---
 .changeset/curly-lemons-confess.md               | 5 +++++
 helpers/python.ts                                | 2 +-
 templates/types/streaming/fastapi/pyproject.toml | 2 +-
 3 files changed, 7 insertions(+), 2 deletions(-)
 create mode 100644 .changeset/curly-lemons-confess.md

diff --git a/.changeset/curly-lemons-confess.md b/.changeset/curly-lemons-confess.md
new file mode 100644
index 00000000..9733dad0
--- /dev/null
+++ b/.changeset/curly-lemons-confess.md
@@ -0,0 +1,5 @@
+---
+"create-llama": patch
+---
+
+Update to LlamaIndex 0.10.55
diff --git a/helpers/python.ts b/helpers/python.ts
index fd755440..c3f8037c 100644
--- a/helpers/python.ts
+++ b/helpers/python.ts
@@ -121,7 +121,7 @@ const getAdditionalDependencies = (
         case "llamacloud":
           dependencies.push({
             name: "llama-index-indices-managed-llama-cloud",
-            version: "^0.2.1",
+            version: "^0.2.5",
           });
           break;
       }
diff --git a/templates/types/streaming/fastapi/pyproject.toml b/templates/types/streaming/fastapi/pyproject.toml
index e5e4d058..c669162e 100644
--- a/templates/types/streaming/fastapi/pyproject.toml
+++ b/templates/types/streaming/fastapi/pyproject.toml
@@ -14,7 +14,7 @@ fastapi = "^0.109.1"
 uvicorn = { extras = ["standard"], version = "^0.23.2" }
 python-dotenv = "^1.0.0"
 aiostream = "^0.5.2"
-llama-index = "0.10.54post1"
+llama-index = "0.10.55"
 cachetools = "^5.3.3"
 
 [build-system]
-- 
GitLab