From 3bc77f7d7fa8ce67b5bcaffa4f91348d0bcba11d Mon Sep 17 00:00:00 2001 From: Yi Ding <yi.s.ding@gmail.com> Date: Tue, 9 Apr 2024 12:42:16 -0700 Subject: [PATCH] gpt-4-turbo GA (#698) --- .changeset/fast-mayflies-grab.md | 5 +++++ examples/agent/wiki.ts | 2 +- examples/multimodal/rag.ts | 2 +- examples/toolsStream.ts | 2 +- examples/vision.ts | 2 +- packages/core/src/llm/azure.ts | 4 ++++ packages/core/src/llm/open_ai.ts | 1 + 7 files changed, 14 insertions(+), 4 deletions(-) create mode 100644 .changeset/fast-mayflies-grab.md diff --git a/.changeset/fast-mayflies-grab.md b/.changeset/fast-mayflies-grab.md new file mode 100644 index 000000000..de9387dc4 --- /dev/null +++ b/.changeset/fast-mayflies-grab.md @@ -0,0 +1,5 @@ +--- +"llamaindex": patch +--- + +gpt-4-turbo GA diff --git a/examples/agent/wiki.ts b/examples/agent/wiki.ts index 65c00a3fe..de2e1f90d 100644 --- a/examples/agent/wiki.ts +++ b/examples/agent/wiki.ts @@ -1,7 +1,7 @@ import { OpenAI, OpenAIAgent, WikipediaTool } from "llamaindex"; async function main() { - const llm = new OpenAI({ model: "gpt-4-turbo-preview" }); + const llm = new OpenAI({ model: "gpt-4-turbo" }); const wikiTool = new WikipediaTool(); // Create an OpenAIAgent with the Wikipedia tool diff --git a/examples/multimodal/rag.ts b/examples/multimodal/rag.ts index f7d945c80..0213de3cc 100644 --- a/examples/multimodal/rag.ts +++ b/examples/multimodal/rag.ts @@ -13,7 +13,7 @@ Settings.chunkSize = 512; Settings.chunkOverlap = 20; // Update llm -Settings.llm = new OpenAI({ model: "gpt-4-vision-preview", maxTokens: 512 }); +Settings.llm = new OpenAI({ model: "gpt-4-turbo", maxTokens: 512 }); // Update callbackManager Settings.callbackManager = new CallbackManager({ diff --git a/examples/toolsStream.ts b/examples/toolsStream.ts index 10e8400df..f52d9049a 100644 --- a/examples/toolsStream.ts +++ b/examples/toolsStream.ts @@ -1,7 +1,7 @@ import { ChatResponseChunk, OpenAI } from "llamaindex"; async function main() { - const llm = new OpenAI({ model: "gpt-4-turbo-preview" }); + const llm = new OpenAI({ model: "gpt-4-turbo" }); const args: Parameters<typeof llm.chat>[0] = { additionalChatOptions: { diff --git a/examples/vision.ts b/examples/vision.ts index 2cf841517..6ee5d33d5 100644 --- a/examples/vision.ts +++ b/examples/vision.ts @@ -1,7 +1,7 @@ import { OpenAI } from "llamaindex"; (async () => { - const llm = new OpenAI({ model: "gpt-4-vision-preview", temperature: 0.1 }); + const llm = new OpenAI({ model: "gpt-4-turbo", temperature: 0.1 }); // complete api const response1 = await llm.complete({ prompt: "How are you?" }); diff --git a/packages/core/src/llm/azure.ts b/packages/core/src/llm/azure.ts index 92d772280..81caf19a0 100644 --- a/packages/core/src/llm/azure.ts +++ b/packages/core/src/llm/azure.ts @@ -19,6 +19,10 @@ const ALL_AZURE_OPENAI_CHAT_MODELS = { }, "gpt-4": { contextWindow: 8192, openAIModel: "gpt-4" }, "gpt-4-32k": { contextWindow: 32768, openAIModel: "gpt-4-32k" }, + "gpt-4-turbo": { + contextWindow: 128000, + openAIModel: "gpt-4-turbo", + }, "gpt-4-vision-preview": { contextWindow: 128000, openAIModel: "gpt-4-vision-preview", diff --git a/packages/core/src/llm/open_ai.ts b/packages/core/src/llm/open_ai.ts index ffc5a176b..00426c221 100644 --- a/packages/core/src/llm/open_ai.ts +++ b/packages/core/src/llm/open_ai.ts @@ -96,6 +96,7 @@ export const GPT4_MODELS = { "gpt-4": { contextWindow: 8192 }, "gpt-4-32k": { contextWindow: 32768 }, "gpt-4-32k-0613": { contextWindow: 32768 }, + "gpt-4-turbo": { contextWindow: 128000 }, "gpt-4-turbo-preview": { contextWindow: 128000 }, "gpt-4-1106-preview": { contextWindow: 128000 }, "gpt-4-0125-preview": { contextWindow: 128000 }, -- GitLab