Follow Datanami:
May 9, 2024

Databricks Enhances Enterprise AI with RAG Applications and Improved Model Serving

May 9, 2024 — Databricks recently announced the general availability of Vector Search and major updates to Model Serving that make it easy for enterprises to build high-quality RAG applications with native capabilities available directly in the Databricks Data Intelligence Platform. In this recent blog post, Akhil Gupta and Oliver Chiu discuss the practical implications of these updates and how they streamline the deployment of AI applications in the enterprise sector.


In December, we announced a new suite of tools to get Generative AI applications to production using Retrieval Augmented Generation (RAG). Since then, we have seen an explosion of RAG applications being built by thousands of customers on the Databricks Data Intelligence Platform.

Today, we are excited to make several announcements to make it easy for enterprises to build high-quality RAG applications with native capabilities available directly in the Databricks Data Intelligence Platform – including the General Availability of Vector Search and major updates to Model Serving.

The Challenge of High Quality AI Applications

As we collaborated closely with our customers to build and deploy AI applications, we’ve identified that the greatest challenge is achieving the high standard of quality required for customer facing systems. Developers spend an inordinate amount of time and effort to ensure that the output of AI applications is accurate, safe, and governed before making it available to their customers and often cite accuracy and quality as the biggest blockers to unlocking the value of these exciting new technologies.

Traditionally, the primary focus to maximize quality has been to deploy an LLM that provides the highest quality baseline reasoning and knowledge capabilities. But, recent research has shown that base model quality is only one of many determinants of the quality of your AI application. LLMs without enterprise context and guidance still hallucinate because they don’t by default have a good understanding of your data. AI applications can also expose confidential or inappropriate data if they don’t understand governance and have proper access controls.

An AI Systems Approach to Quality

Achieving production quality in GenAI applications requires a comprehensive approach involving multiple components that cover all aspects of the GenAI process: data preparation, retrieval models, language models (either SaaS or open source), ranking, post-processing pipelines, prompt engineering, and training on custom enterprise data. Together these components constitute an AI System.

Today, we are excited to announce major updates and more details to help customers build production-quality GenAI applications:

  • General availability of Vector Search, a serverless vector database purpose-built for customers to augment their LLMs with enterprise data.
  • General availability in the coming weeks of Model Serving Foundation Model API which allows you to access and query state-of-the-art LLMs from a serving endpoint
  • Major updates to Model Serving
    • A new user interface making it easier than ever before to deploy, serve, monitor, govern, and query LLMs
    • Support for additional state of the art models – Claude3, Gemini, DBRX and Llama3
    • Performance improvements to deploy and query large LLMs
    • Better governance and auditability with support for inference tables across all types of serving endpoints.

We also previously announced the following that helps deploy production-quality GenAI:

Over the course of this week, we’ll have detailed blogs on how you can use these new capabilities to build high-quality RAG apps. We’ll also share an insider’s blog on how we built DBRX, an open, general-purpose LLM created by Databricks.


Source: Akhil Gupta and Oliver Chiu, Databricks

Tags: , ,
Datanami