From ba95ca3fb686de28bc98d14e09f1017b89395ad0 Mon Sep 17 00:00:00 2001
From: Marcus Schiesser <mail@marcusschiesser.de>
Date: Fri, 2 Feb 2024 15:46:11 +0700
Subject: [PATCH] feat(cl): Use condense plus context chat engine for FastAPI
 as default

---
 .changeset/tasty-buckets-design.md                           | 5 +++++
 .../templates/components/vectordbs/python/mongo/index.py     | 2 +-
 .../templates/components/vectordbs/python/none/index.py      | 2 +-
 .../templates/components/vectordbs/python/pg/index.py        | 2 +-
 4 files changed, 8 insertions(+), 3 deletions(-)
 create mode 100644 .changeset/tasty-buckets-design.md

diff --git a/.changeset/tasty-buckets-design.md b/.changeset/tasty-buckets-design.md
new file mode 100644
index 000000000..0221c1fe5
--- /dev/null
+++ b/.changeset/tasty-buckets-design.md
@@ -0,0 +1,5 @@
+---
+"create-llama": patch
+---
+
+Use condense plus context chat engine for FastAPI as default
diff --git a/packages/create-llama/templates/components/vectordbs/python/mongo/index.py b/packages/create-llama/templates/components/vectordbs/python/mongo/index.py
index d9b14dd22..a80590b5c 100644
--- a/packages/create-llama/templates/components/vectordbs/python/mongo/index.py
+++ b/packages/create-llama/templates/components/vectordbs/python/mongo/index.py
@@ -20,4 +20,4 @@ def get_chat_engine():
     )
     index = VectorStoreIndex.from_vector_store(store, service_context)
     logger.info("Finished connecting to index from MongoDB.")
-    return index.as_chat_engine(similarity_top_k=5)
+    return index.as_chat_engine(similarity_top_k=5, chat_mode="condense_plus_context")
diff --git a/packages/create-llama/templates/components/vectordbs/python/none/index.py b/packages/create-llama/templates/components/vectordbs/python/none/index.py
index 0170d6e83..4404c66e2 100644
--- a/packages/create-llama/templates/components/vectordbs/python/none/index.py
+++ b/packages/create-llama/templates/components/vectordbs/python/none/index.py
@@ -22,4 +22,4 @@ def get_chat_engine():
     storage_context = StorageContext.from_defaults(persist_dir=STORAGE_DIR)
     index = load_index_from_storage(storage_context, service_context=service_context)
     logger.info(f"Finished loading index from {STORAGE_DIR}")
-    return index.as_chat_engine()
+    return index.as_chat_engine(similarity_top_k=5, chat_mode="condense_plus_context")
diff --git a/packages/create-llama/templates/components/vectordbs/python/pg/index.py b/packages/create-llama/templates/components/vectordbs/python/pg/index.py
index 5c902772a..510c21a53 100644
--- a/packages/create-llama/templates/components/vectordbs/python/pg/index.py
+++ b/packages/create-llama/templates/components/vectordbs/python/pg/index.py
@@ -13,4 +13,4 @@ def get_chat_engine():
     store = init_pg_vector_store_from_env()
     index = VectorStoreIndex.from_vector_store(store, service_context)
     logger.info("Finished connecting to index from PGVector.")
-    return index.as_chat_engine(similarity_top_k=5)
+    return index.as_chat_engine(similarity_top_k=5, chat_mode="condense_plus_context")
-- 
GitLab