Back to Kibana

Elastic Managed LLMs

docs/reference/connectors-kibana/elastic-managed-llm.md

9.4.02.2 KB
Original Source

Elastic Managed LLMs

Elastic provides built-in LLMs through managed AI connectors.
These connectors are accessed and managed through the Elastic {{infer-cap}} Service (EIS), which is the single entry point for using Elastic Managed LLMs.

Prerequisites

  • Requires the manage_inference cluster privilege (the built-in inference_admin role grants this privilege)
  • {applies_to}ece: {applies_to}eck: {applies_to}self: ga 9.3+ For on-premises installations ({{ece}}, {{eck}}, or self managed clusters), Elastic Managed LLMs are only available through EIS with Cloud Connect. Your {{stack}} version must be 9.3 or later.

Available models

Elastic Managed LLMs are available exclusively through the Elastic {{infer-cap}} Service. You can find the list of supported models on the EIS documentation page.

Region and hosting

The Elastic Managed LLMs use third party service providers for {{infer}}. Refer to the Elastic {{infer-cap}} Service page for details.

Data protection

Customer projects or deployments hosted in any cloud service provider or region have access to Elastic Managed LLMs in the AWS US region us-east-1. All data is encrypted in transit. The LLMs are configured for zero data retention: none of the prompts or outputs are stored by the service provider.

Only request metadata is logged in AWS CloudWatch. No information related to prompts is retained. Logged metadata includes the timestamp, model used, region, and request status.

Read more at our AI Data FAQs to learn about our data practices for AI related features.

Pricing

Elastic Managed LLMs incur a cost per million tokens for input and output tokens. Refer to the Elastic pricing page that correspond to your Elastic setup for details.