DOCSDEPLOY AI MODEL

Deploy AI Model

Loading last updated info...

Overview

EagleEye's Marketplace deployment automatically provisions an Azure OpenAI resource, but it does not deploy a chat model inside it.

This is intentional. Azure OpenAI model availability depends on the quotas available in each customer's subscription. So marketplace deployment is independent of model quotas.

This extra step is only required if you want to use:

If you plan to use Non-AI Tags only, you can skip this guide.

At the moment, EagleEye supports chat models only, so make sure you choose a compatible chat model.

EagleEye supported models include:

  • gpt-4.1-mini (recommended)
  • gpt-4o-mini
  • gpt-4.1-nano
  • gpt-5-chat, gpt-5.1-chat, gpt-5.2-chat
  • gpt-4.1, gpt-4o
  • gpt-3.5-turbo

Do not use:

  • Reasoning models such as gpt-5, gpt-5-mini, gpt-5-nano, gpt-5-pro, gpt-5-codex, gpt-5.4-mini, o1, o1-mini, o3-mini, o4-mini
  • Non-OpenAI Foundry catalog models such as Llama, Mistral, Grok, DeepSeek, Cohere, Kimi, gpt-oss-*, and Phi
  • Non-chat models such as embeddings, image models, audio models, speech models, and legacy completion models

Step 1 - Deploy an AI chat model in Azure AI Foundry

You have two options. Pick the one that best fits your workflow.

Option A - Deploy via Azure AI Foundry portal

  1. Go to your EagleEye resource group used during Step 3 installation (e.g., Northwind.EagleEye.RG).
  2. Find your openai-shared-xxx resource and open it. Image
  3. Click Go to Azure AI Foundry portal. Image
  4. In Foundry portal, click Model catalog. Image
  5. Pick a compatible chat model > Use this model Image
  6. On the deployment screen, set:
    • Deployment name to exactly eagleeye
    • Deployment type to Global Standard if available. If not, choose another compatible deployment type with available quota.
    • Tokens per Minute Rate Limit to at least 100 Image
  7. Click Deploy and wait for the deployment to become active.
Important: The deployment name must be exactly eagleeye. If you use a different name, EagleEye will fail to find the model at runtime.

Option B - Deploy via AI agent

If you have an AI coding agent connected to Azure MCP, or you have az CLI installed and signed in with the same account you used to deploy EagleEye, paste this prompt into the agent:

The agent should:

  1. Ask for the resource group you used during Step 3 installation (e.g., Northwind.EagleEye.RG).
  2. Find the openai-shared-xxx Azure OpenAI resource in that resource group.
  3. Show which compatible non-reasoning chat models have available quota.
  4. Ask which model and capacity you want to deploy.
  5. Create the deployment with the exact name eagleeye.
Tip: Review the commands your agent proposes before approving them. AI agents can still make mistakes.

Step 2 - Verify AI email analysis works

Use EagleEye's built-in test flow to verify the model deployment without sending any real emails:

1. Open your EagleEye Portal.

2. Go to TagsNew Tag, or edit an existing tag. Image

3. Set the tag type to AI or AI with Mentions. Image

4. In the AI decision prompt field, enter a simple test prompt such as:

5. Paste sample email content such as:

6. Click Test.

Expected result: the response box shows valid JSON such as {"Result": true} or {"Result": false}.

“Checked by” present = True“Checked by” not present = False

⚠️ If the model is not deployed, EagleEye will show an error banner instead of a JSON result.

References