From 3bc77f7d7fa8ce67b5bcaffa4f91348d0bcba11d Mon Sep 17 00:00:00 2001
From: Yi Ding <yi.s.ding@gmail.com>
Date: Tue, 9 Apr 2024 12:42:16 -0700
Subject: [PATCH] gpt-4-turbo GA (#698)

---
 .changeset/fast-mayflies-grab.md | 5 +++++
 examples/agent/wiki.ts           | 2 +-
 examples/multimodal/rag.ts       | 2 +-
 examples/toolsStream.ts          | 2 +-
 examples/vision.ts               | 2 +-
 packages/core/src/llm/azure.ts   | 4 ++++
 packages/core/src/llm/open_ai.ts | 1 +
 7 files changed, 14 insertions(+), 4 deletions(-)
 create mode 100644 .changeset/fast-mayflies-grab.md

diff --git a/.changeset/fast-mayflies-grab.md b/.changeset/fast-mayflies-grab.md
new file mode 100644
index 000000000..de9387dc4
--- /dev/null
+++ b/.changeset/fast-mayflies-grab.md
@@ -0,0 +1,5 @@
+---
+"llamaindex": patch
+---
+
+gpt-4-turbo GA
diff --git a/examples/agent/wiki.ts b/examples/agent/wiki.ts
index 65c00a3fe..de2e1f90d 100644
--- a/examples/agent/wiki.ts
+++ b/examples/agent/wiki.ts
@@ -1,7 +1,7 @@
 import { OpenAI, OpenAIAgent, WikipediaTool } from "llamaindex";
 
 async function main() {
-  const llm = new OpenAI({ model: "gpt-4-turbo-preview" });
+  const llm = new OpenAI({ model: "gpt-4-turbo" });
   const wikiTool = new WikipediaTool();
 
   // Create an OpenAIAgent with the Wikipedia tool
diff --git a/examples/multimodal/rag.ts b/examples/multimodal/rag.ts
index f7d945c80..0213de3cc 100644
--- a/examples/multimodal/rag.ts
+++ b/examples/multimodal/rag.ts
@@ -13,7 +13,7 @@ Settings.chunkSize = 512;
 Settings.chunkOverlap = 20;
 
 // Update llm
-Settings.llm = new OpenAI({ model: "gpt-4-vision-preview", maxTokens: 512 });
+Settings.llm = new OpenAI({ model: "gpt-4-turbo", maxTokens: 512 });
 
 // Update callbackManager
 Settings.callbackManager = new CallbackManager({
diff --git a/examples/toolsStream.ts b/examples/toolsStream.ts
index 10e8400df..f52d9049a 100644
--- a/examples/toolsStream.ts
+++ b/examples/toolsStream.ts
@@ -1,7 +1,7 @@
 import { ChatResponseChunk, OpenAI } from "llamaindex";
 
 async function main() {
-  const llm = new OpenAI({ model: "gpt-4-turbo-preview" });
+  const llm = new OpenAI({ model: "gpt-4-turbo" });
 
   const args: Parameters<typeof llm.chat>[0] = {
     additionalChatOptions: {
diff --git a/examples/vision.ts b/examples/vision.ts
index 2cf841517..6ee5d33d5 100644
--- a/examples/vision.ts
+++ b/examples/vision.ts
@@ -1,7 +1,7 @@
 import { OpenAI } from "llamaindex";
 
 (async () => {
-  const llm = new OpenAI({ model: "gpt-4-vision-preview", temperature: 0.1 });
+  const llm = new OpenAI({ model: "gpt-4-turbo", temperature: 0.1 });
 
   // complete api
   const response1 = await llm.complete({ prompt: "How are you?" });
diff --git a/packages/core/src/llm/azure.ts b/packages/core/src/llm/azure.ts
index 92d772280..81caf19a0 100644
--- a/packages/core/src/llm/azure.ts
+++ b/packages/core/src/llm/azure.ts
@@ -19,6 +19,10 @@ const ALL_AZURE_OPENAI_CHAT_MODELS = {
   },
   "gpt-4": { contextWindow: 8192, openAIModel: "gpt-4" },
   "gpt-4-32k": { contextWindow: 32768, openAIModel: "gpt-4-32k" },
+  "gpt-4-turbo": {
+    contextWindow: 128000,
+    openAIModel: "gpt-4-turbo",
+  },
   "gpt-4-vision-preview": {
     contextWindow: 128000,
     openAIModel: "gpt-4-vision-preview",
diff --git a/packages/core/src/llm/open_ai.ts b/packages/core/src/llm/open_ai.ts
index ffc5a176b..00426c221 100644
--- a/packages/core/src/llm/open_ai.ts
+++ b/packages/core/src/llm/open_ai.ts
@@ -96,6 +96,7 @@ export const GPT4_MODELS = {
   "gpt-4": { contextWindow: 8192 },
   "gpt-4-32k": { contextWindow: 32768 },
   "gpt-4-32k-0613": { contextWindow: 32768 },
+  "gpt-4-turbo": { contextWindow: 128000 },
   "gpt-4-turbo-preview": { contextWindow: 128000 },
   "gpt-4-1106-preview": { contextWindow: 128000 },
   "gpt-4-0125-preview": { contextWindow: 128000 },
-- 
GitLab