diff --git a/helpers/providers/azure.ts b/helpers/providers/azure.ts
index b343d3b388b3efdc3466daa058bbf6d15cb30272..74f901423568bf9151301e4127648d27de9fb64d 100644
--- a/helpers/providers/azure.ts
+++ b/helpers/providers/azure.ts
@@ -57,7 +57,7 @@ export async function askAzureQuestions({
   askModels,
 }: ModelConfigQuestionsParams): Promise<ModelConfigParams> {
   const config: ModelConfigParams = {
-    apiKey: openAiKey,
+    apiKey: openAiKey || process.env.AZURE_OPENAI_KEY,
     model: DEFAULT_MODEL,
     embeddingModel: DEFAULT_EMBEDDING_MODEL,
     dimensions: getDimensions(DEFAULT_EMBEDDING_MODEL),
@@ -67,29 +67,6 @@ export async function askAzureQuestions({
     },
   };
 
-  if (!config.apiKey) {
-    const { key } = await prompts(
-      {
-        type: "text",
-        name: "key",
-        message: askModels
-          ? "Please provide your Azure OpenAI API key (or leave blank to use AZURE_OPENAI_KEY env variable):"
-          : "Please provide your Azure OpenAI API key (leave blank to skip):",
-        validate: (value: string) => {
-          if (askModels && !value) {
-            if (process.env.AZURE_OPENAI_KEY) {
-              return true;
-            }
-            return "AZURE_OPENAI_KEY env variable is not set - key is required";
-          }
-          return true;
-        },
-      },
-      questionHandlers,
-    );
-    config.apiKey = key || process.env.AZURE_OPENAI_KEY;
-  }
-
   // use default model values in CI or if user should not be asked
   const useDefaults = ciInfo.isCI || !askModels;
   if (!useDefaults) {