On This Page

Home / Cribl Copilot/Custom AI Providers (Bring Your Own Model)

Custom AI Providers (Bring Your Own Model)

Custom AI Providers enable a Bring Your Own Model (BYOM) approach, allowing you to connect your own large language model (LLM) backend to Cribl AI features. This gives you the flexibility to route supported Cribl Copilot capabilities through your preferred AI infrastructure instead of using the default Cribl-managed backend.

Why Use Custom AI Providers?

By default, Cribl-managed AI features use models selected and operated by Cribl, with traffic following Cribl-managed routes and service contracts.

While this default configuration is secure, many enterprises require deeper levels of control. Custom AI providers allow you to replace the default backend with your own approved AI vendors, ensuring that your keys, models, and traffic stay within your existing cloud governance boundaries.

Integrating your own AI backend provides the following key benefits for your deployment:

  • Vendor standardization: Use the AI providers your organization has already vetted and standardized on.
  • Data privacy and compliance: Keep AI traffic within specific regulatory or contractual boundaries. This is essential for organizations with strict data residency requirements or those operating in highly regulated industries.
  • Direct governance: Maintain full control over your API keys and the specific model versions used for inference.
  • Cost management: Gain direct visibility into how AI inference is billed. By using your own provider contracts, you can apply existing cloud credits and committed spend to your Cribl AI usage.

How Custom AI Providers Work

Custom AI providers add a configurable backend to the existing Cribl AI architecture. This setup allows the platform to route agentic workflows through a local service that communicates directly with your self-managed models (such as Anthropic or OpenAI) instead of the default Cribl AI provider.

The goal is to route supported agentic workflows through an AI service that can call either:

  • Your self-managed models (for example, Anthropic or OpenAI), or
  • The Cribl cloud-hosted AI backend, where needed.

Main Components

Custom AI Provider Component Overview
Custom AI Provider Component Overview

The following core components of the custom AI provider feature work together to route and process AI requests within your managed environment:

  • Cribl Leader Node: The Leader Node serves the UI and API. All AI requests initiated in the UI are first received by the Leader.
  • AI service: A service running on the Leader Node that manages AI agents. When a custom AI provider is enabled, it calls your configured AI provider directly using your stored credentials.
  • Cloud AI backend: Cribl-hosted backend. This continues to handle workflows that are not yet migrated to the local AI Service or customer-managed providers.

Example Workflow

When you plug in a custom AI provider, you are essentially replacing the “brain” behind Cribl’s AI features. Once configured in AI Settings, the Cribl Leader Node stops sending requests to the default Cribl cloud and instead sends them directly to your hosted model.

To see this in practice, consider the KQL assistant in Cribl Search. Without needing to know Kusto Query Language, an analyst can use your custom LLM to investigate data.

  1. Natural language input: An analyst types, “Show me the top 10 source IPs with the most failed login attempts in the last hour” into the Copilot search box.

  2. Custom inference: Cribl sends this request to your configured provider (such as Anthropic or OpenAI).

  3. KQL generation: Your LLM processes the request and returns the specific KQL syntax:

    dataset="auth_logs" | where status=="failure" | summarize count() by src_ip | top 10 by count_

  4. Execution: Cribl Search executes the query in KQL syntax against your data.

Supported Custom AI Providers and Features

Review the following sections for details on currently supported custom AI providers and the specific Copilot features they power.

Supported providers currently include:

  • Anthropic via Amazon Bedrock
  • OpenAI via Microsoft Foundry

Cribl is committed to an agnostic AI strategy. While initial support focuses on AWS and Azure ecosystems, the framework is designed to expand to additional model families and providers in future releases.

Deployment Limitations

Before configuring a custom AI provider, note the following constraints:

  • Single provider scope: You can configure only one custom AI provider per Workspace (in Cribl.Cloud) or Global setting (on-prem).
  • Available in Cribl.Cloud commercial environments only: Custom AI provider support is not currently available for Cribl.Cloud Government environments.
  • Inference only: Custom AI providers are used for inference only. Fine-tuned or custom-trained models are not supported at this time.

Supported Copilot Features

When a custom AI provider is active, the following features route through your managed LLM:

Capabilities Using Cribl-Managed Models

The following features currently rely on the global Copilot chat path and continue to use the default Cribl-managed models: