Elastic announced it is
strengthening its relationship with Amazon Web Services (AWS) by
leveraging the latest generative artificial intelligence (AI) services
from AWS. As part of this collaboration, Elastic is offering large
language model (LLM) observability support for Amazon Bedrock in Elastic
Observability. Amazon Bedrock is a fully managed service that offers a
choice of high-performing foundation models (FMs) from leading AI
companies via a single API, along with a broad set of capabilities
organizations need to build generative AI applications with security,
privacy, and responsible AI.
The new integration offers Site Reliability Engineers (SREs) detailed
insights into their Amazon Bedrock LLMs performance and usage. SREs can
now leverage Elastic Observability to monitor invocations, errors, and
latency metrics. This allows them to more proactively prevent incidents
and identify root causes, ensuring optimal performance for their Amazon
Bedrock-powered generative AI applications. Additionally, Elastic AI
Assistant, which utilizes Amazon Bedrock, helps SREs accurately analyze
data, generate visualizations, and provides actionable recommendations
for issue resolution.
"As LLM-based applications are growing, it's essential for developers
and SREs to be able to monitor, optimize, and troubleshoot how they
perform," said Santosh Krishnan, general manager of Security and Observability Solutions at Elastic. "Today's
integration simplifies the collection of metrics and logs from Amazon
Bedrock, in turn streamlining the process of gaining valuable and
actionable insights."
Support for Amazon Bedrock LLM visibility is available today.