Artificial Intelligence | News | Insights | AiThority
[bsfp-cryptocurrency style=”widget-18″ align=”marquee” columns=”6″ coins=”selected” coins-count=”6″ coins-selected=”BTC,ETH,XRP,LTC,EOS,ADA,XLM,NEO,LTC,EOS,XEM,DASH,USDT,BNB,QTUM,XVG,ONT,ZEC,STEEM” currency=”USD” title=”Cryptocurrency Widget” show_title=”0″ icon=”” scheme=”light” bs-show-desktop=”1″ bs-show-tablet=”1″ bs-show-phone=”1″ custom-css-class=”” custom-id=”” css=”.vc_custom_1523079266073{margin-bottom: 0px !important;padding-top: 0px !important;padding-bottom: 0px !important;}”]

DataStax Integrates with LangChain, Enables Developers to Easily Build Production-Ready Generative AI Applications

Support for Astra DB Vector Database and Apache Cassandra Now Available Out-of-the-Box for Any LangChain User

DataStax, the company that powers generative AI applications with real-time, scalable data, announced a new integration with LangChain, the most popular orchestration framework for developing applications with large language models (LLMs). The integration makes it easy to add Astra DB – the real-time database for developers building production Gen AI applications – or Apache Cassandra®, as a new vector source in the LangChain framework.

As many companies implement retrieval augmented generation (RAG) – the process of providing context from outside data sources to deliver more accurate LLM query responses – into their generative AI applications, they require a vector store that gives them real-time updates with zero latency on critical, real-life production workloads.

Recommended AI News: Samuel, Son & Co., Limited Selects ION’s Reval for Global Treasury and Risk Management

Generative AI applications built with RAG stacks require a vector-enabled database and an orchestration framework like LangChain, to provide memory or context to LLMs for accurate and relevant answers. Developers use LangChain as the leading AI-first toolkit to connect their application to different data sources.

The new integration lets developers leverage the power of the Astra DB vector database for their LLM, AI assistant, and real-time generative AI projects through the LangChain plugin architecture for vector stores. Together, Astra DB and LangChain help developers to take advantage of framework features like vector similarity search, semantic caching, term-based search, LLM-response caching, and data injection from Astra DB (or Cassandra) into prompt templates.

Related Posts
1 of 40,870

Recommended AI News: Luzia Closes $10Million Series A, Reinforces Position as Leading AI Assistant

“In a RAG application, the model receives supplementary data or context from various sources — most often a database that can store vectors,” said Harrison Chase, CEO, LangChain. “Building a generative AI app requires a robust, powerful database, and we ensure our users have access to the best options on the market via our simple plugin architecture. With integrations like DataStax’s LangChain connector, incorporating Astra DB or Apache Cassandra as a vector store becomes a seamless and intuitive process.”

“Developers at startups and enterprises alike are using LangChain to build generative AI apps, so a deep native integration is a must-have,” said Ed Anuff, CPO, DataStax. “The ability for developers to easily use Astra DB as their vector database of choice, directly from LangChain, streamlines the process of building the personalized AI applications that companies need. In fact, we’re already seeing customers benefit from our joint technologies as healthcare AI company, Skypoint, is using Astra DB and LangChain to power its generative AI healthcare model.”

Recommended AI News: Mastercard Expands Its Consulting Services With Economics and AI Practices

October 26 at 9am PT, where LangChain founder and CEO, Harrison Chase, and SkyPoint founder and CEO, Tisson Mathew, discuss their experience building production RAG applications.

[To share your insights with us, please write to sghosh@martechseries.com]

Comments are closed.