diff --git a/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx b/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx index ffc0f91e1fb706b59a802f7e595e37db3113c0c4..b1718afe907f9f40a1ab7075fb96d7794f4ee817 100644 --- a/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx +++ b/frontend/src/components/LLMSelection/OpenAiOptions/index.jsx @@ -85,6 +85,7 @@ function OpenAIModelSelection({ apiKey, settings }) { "gpt-3.5-turbo", "gpt-3.5-turbo-1106", "gpt-4", + "gpt-4-turbo-preview", "gpt-4-1106-preview", "gpt-4-32k", ].map((model) => { diff --git a/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js b/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js index 270f8a968e06a0f8a7f2021501f7274e8fd04504..1f8cce9889f04844f9d058dd06f225ca89b3c24b 100644 --- a/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js +++ b/frontend/src/components/Modals/MangeWorkspace/Settings/ChatModelPreference/useGetProviderModels.js @@ -8,6 +8,7 @@ const PROVIDER_DEFAULT_MODELS = { "gpt-3.5-turbo", "gpt-3.5-turbo-1106", "gpt-4", + "gpt-4-turbo-preview", "gpt-4-1106-preview", "gpt-4-32k", ], diff --git a/server/utils/AiProviders/openAi/index.js b/server/utils/AiProviders/openAi/index.js index a25f69eade44b013aab76d7aa15cc36d1abe4bf3..120e7284483750e54cba8e44615ce32bc0d2e4df 100644 --- a/server/utils/AiProviders/openAi/index.js +++ b/server/utils/AiProviders/openAi/index.js @@ -52,6 +52,8 @@ class OpenAiLLM { return 8192; case "gpt-4-1106-preview": return 128000; + case "gpt-4-turbo-preview": + return 128000; case "gpt-4-32k": return 32000; default: @@ -65,6 +67,7 @@ class OpenAiLLM { "gpt-3.5-turbo", "gpt-3.5-turbo-1106", "gpt-4-1106-preview", + "gpt-4-turbo-preview", "gpt-4-32k", ]; const isPreset = validModels.some((model) => modelName === model);