Skip to content
43 changes: 32 additions & 11 deletions docs/reference/connectors-kibana/elastic-managed-llm.md
Original file line number Diff line number Diff line change
@@ -1,31 +1,54 @@
---
navigation_title: "Elastic Managed LLM"
navigation_title: "Elastic Managed LLMs"
applies_to:
stack: ga 9.0
serverless: ga
---

# Elastic Managed LLM
# Elastic Managed LLMs

Elastic provides a default LLM vetted for GenAI product features across our platform.
Details of the currently used model are available in the [model card](https://raw.githubusercontent.com/elastic/kibana/refs/heads/main/docs/reference/resources/Elastic_Managed_LLM_model_card.pdf).
Elastic provides built-in LLMs through the Elastic Managed LLM connector vetted for GenAI product features across our platform.
Currently, the following built-in LLMs are available:

* [General Purpose LLM v2](#general-purpose-llm-v2)
* [General Purpose LLM v1](#general-purpose-llm-v1)

## General Purpose LLM v2
```{applies_to}
stack: ga 9.3
serverless: ga
```

A high-performance, general-purpose language model suitable for the widest range of use cases.
Model details are available in the [model card](https://raw.githubusercontent.com/elastic/kibana/refs/heads/main/docs/reference/resources/General_Purpose_LLM_v2_model_card.pdf).

## General Purpose LLM v1
```{applies_to}
stack: ga 9.0
serverless: ga
```

General Purpose LLM v1 is the default LLM to use with this connector.

::::{note}
The default LLM may change in the future based on evaluations of performance, security, and accuracy.
* The default LLM may change in the future based on evaluations of performance, security, and accuracy.
* This LLM was called Elastic Managed LLM before 9.3.
::::

Details of the currently used model are available in the [model card](https://raw.githubusercontent.com/elastic/kibana/refs/heads/main/docs/reference/resources/Elastic_Managed_LLM_model_card.pdf).

## Prerequisites

* Requires the `manage_inference` [cluster privilege](https://www.elastic.co/docs/reference/elasticsearch/security-privileges#privileges-list-cluster) (the built-in `inference_admin` role grants this privilege)

## Region and hosting

The Elastic Managed LLM is currently proxying to AWS Bedrock in AWS US regions, beginning with `us-east-1`.
The Elastic Managed LLMs user a third party service provider for inference. Currently, this is AWS Bedrock in AWS US regions, beginning with `us-east-1`. Additional providers may be added in future.

## Data protection

Customer projects or deployments hosted in any cloud service provider or region have access to the Elastic Managed LLM in the AWS US region `us-east-1`.
All data is encrypted in transit. The LLM is configured for zero data retention: none of the prompts or outputs are stored by the model.
Customer projects or deployments hosted in any cloud service provider or region have access to the Elastic Managed LLMs in the AWS US region `us-east-1`.
All data is encrypted in transit. The LLMs are configured for zero data retention: none of the prompts or outputs are stored by the service provider.

Only request metadata is logged in AWS CloudWatch.
No information related to prompts is retained.
Expand All @@ -35,6 +58,4 @@ Read more at our [AI Data FAQs](https://www.elastic.co/trust/ai-data-faq) to lea

## Pricing

The Elastic Managed LLM incurs a cost per million tokens for input and output tokens. Refer to the Elastic Cloud [pricing pages](https://www.elastic.co/pricing) for details.


The Elastic Managed LLMs incur a cost per million tokens for input and output tokens. Refer to the Elastic Cloud [pricing pages](https://www.elastic.co/pricing) for details.
Binary file not shown.
Loading