Skip to main content

Copilot settings

iconExpress and Enterprise

Applicable to the Express and Enterprise Editions only.

Prophecy's AI capabilities, including Copilot and Agents, are powered by external LLMs.

  • The SaaS deployment uses a Prophecy-managed OpenAI subscription with GPT-4o and GPT-4o mini.
  • Dedicated SaaS deployments connect to customer-managed endpoints.

This page describes how to connect Prophecy to your customer-managed LLM.

Prerequisites

To add your LLM provider and model details to Copilot settings:

  • You must be logged in as a Prophecy cluster admin.
  • Copilot must already be enabled in your Prophecy deployment. Reach out to the Support team to enable Copilot.
info

See Prophecy AI for a list of supported LLM providers and models.

Configure Copilot settings

Prophecy lets you add your LLM credentials and model choices directly in the UI in Settings > Admin > Copilot Settings. Copilot Settings includes three subtabs:

You must provide all settings in YAML format. These values directly override those in your Kubernetes deployment.

AI Model Providers Creds

In the AI Model Providers Creds subtab, you provide the credentials to connect to your LLM provider.

You can add multiple credentials here, but Copilot will only connect to the models defined in the Available AI Models subtab.

The following YAML example shows the required fields for different LLM providers.

{
"azure_openai": { "api_key": "********", "api_endpoint": "********" },
"openai": { "api_key": "********" },
"gemini": { "api_key": "********" },
"vertex_ai": {},
}
note

When editing the credentials in Prophecy, you will not be able to see previously-entered values. Likewise, once you save your credentials, values will be masked by asterisks **** as shown in the example above.

Set up Vertex AI

Vertex AI does not require an API key. Instead, you provide a Google Cloud service account file.

Reach out to the Support team with the following formation to add Vertex AI credentials to your deployment:

  • The service account file that can authenticate your connection to Vertex AI
  • (Optional) The Vertex AI region, if it differs from the service account region
  • (Optional) The Vertex AI project, if it differs from the service account project

Available AI Models

In the Available AI Models subtab, you define two models:

  • Smart model: Prophecy uses this model for complex tasks.

    Recommended Model: gpt-4o

  • Fast model: Prophecy uses this model for easy and quick tasks.

    Recommended Model: gpt-4o-mini

You can select models from any of the providers you configured in the AI Model Providers Creds subtab.

The following YAML example shows how to format your smart and fast model configuration.

{
"smart_model": { "provider": "openai", "model_name": "gpt-4.1" },
"fast_model": { "provider": "openai", "model_name": "gpt-4.1-mini" },
}

Available AI Speech Models

In the Available AI Speech Models subtab, you define different models for speech-to-text and text-to-speech operations.

You can select models from any of the providers you configured in the AI Model Providers Creds subtab.

The following YAML example shows how to format your speech-to-text (stt) and text-to-speech (tts) model configuration.

{
"stt": { "provider": "openai", "model_name": "whisper-1" },
"tts": { "provider": "openai", "model_name": "tts-1" },
}