Couchbase, Inc. announced that its
Capella AI Model Services have integrated
NVIDIA NIM microservices, part of the
NVIDIA AI Enterprise
software platform, to streamline deployment of AI-powered applications,
providing enterprises a powerful solution for privately running
generative (GenAI) models.
Capella AI Model Services, which were recently introduced
as part of a comprehensive Capella AI Services offering for
streamlining the development of agentic applications, provide managed
endpoints for LLMs and embedding models so enterprises can meet privacy,
performance, scalability and latency requirements within their
organizational boundary. Capella AI Model Services, powered by NVIDIA AI
Enterprise, minimize latency by bringing AI closer to the data,
combining GPU-accelerated performance and enterprise-grade security to
empower organizations to seamlessly operate their AI workloads. The
collaboration enhances Capella's agentic AI and retrieval-augmented
generation (RAG) capabilities, allowing customers to efficiently power
high-throughput AI-powered applications while maintaining model
flexibility.
"Enterprises require a unified and highly performant data platform to
underpin their AI efforts and support the full application lifecycle -
from development through deployment and optimization," said Matt McDonough,
SVP of product and partners at Couchbase. "By integrating NVIDIA NIM
microservices into Capella AI Model Services, we're giving customers the
flexibility to run their preferred AI models in a secure and governed
way, while providing better performance for AI workloads and seamless
integration of AI with transactional and analytical data. Capella AI
Services allow customers to accelerate their RAG and agentic
applications with confidence, knowing they can scale and optimize their
applications as business needs evolve."
Capella Delivers Fully Integrated User Experience with NVIDIA AI Enterprise, Enabling Flexible, Scalable AI Model Deployment
Enterprises building and deploying high-throughput AI applications
can face challenges with ensuring agent reliability and compliance, as
unreliable AI responses can damage brand reputation. PII data leaks can
violate privacy regulations and managing multiple specialized databases
can create unsustainable operational overhead. Couchbase is helping
address these challenges with Capella AI Model Services, which
streamline agent application development and operations by keeping
models and data colocated in a unified platform, facilitating agentic
operations as they happen. For example, agent conversation transcripts
must be captured and compared in real time to elevate model response
accuracy. Capella also delivers built-in capabilities like semantic
caching, guardrail creation and agent monitoring with RAG workflows.
Capella AI Model Services with NVIDIA NIM provides Couchbase
customers a cost-effective solution that accelerates agent delivery by
simplifying model deployment while maximizing resource utilization and
performance. The solution leverages pre-tested LLMs and tools including NVIDIA NeMo Guardrails
to help organizations accelerate AI development while enforcing
policies and safeguards against AI hallucinations. NVIDIA's rigorously
tested, production-ready NIM microservices are optimized for reliability
and fine-tuned for specific business needs.
"Integrating NVIDIA AI software into Couchbase's Capella AI Model
Services enables developers to quickly deploy, scale and optimize
applications," said Anne Hecht, senior
director of enterprise software at NVIDIA. "Access to NVIDIA NIM
microservices further accelerates AI deployment with optimized models,
delivering low-latency performance and security for real-time
intelligent applications."