Databricks Model Serving offers fully managed, production ML capabilities built natively within the Databricks Lakehouse Platform
Databricks, the lakehouse company, today announced the launch of Databricks Model Serving to provide simplified production machine learning (ML) natively within the Databricks Lakehouse Platform. Model Serving removes the complexity of building and maintaining complicated infrastructure for intelligent applications. Now, organizations can leverage the Databricks Lakehouse Platform to integrate real-time machine learning systems across their business, from personalized recommendations to customer service chatbots, without the need to configure and manage the underlying infrastructure. Deep integration within the Lakehouse Platform offers data and model lineage, governance and monitoring throughout the ML lifecycle, from experimentation to training to production. Databricks Model Serving is now generally available on AWS and Azure.
With the opportunities surrounding generative artificial intelligence (AI) taking center stage, businesses feel the urgency to prioritize AI investments across the board. Leveraging AI/ML enables organizations to uncover insights from their data, make accurate, instant predictions that deliver business value, and drive new AI-led experiences for their customers. For example, AI can enable a bank to quickly identify and combat fraudulent charges on a customer’s account or give a retailer the ability to instantly suggest complementary accessories based on a customer’s clothing purchases. Most of these experiences are integrated in real-time applications. However, implementing these real-time ML systems has remained a challenge for many organizations because of the burden placed on ML experts to design and maintain infrastructure that can dynamically scale to meet demand.
“Databricks Model Serving accelerates data science teams’ path to production by simplifying deployments, reducing overhead and delivering a fully integrated experience directly within the Databricks Lakehouse,” said Patrick Wendell, Co-Founder and VP of Engineering at Databricks. “This offering will let customers deploy far more models, with lower time to production, while also lowering the total cost of ownership and the burden of managing complex infrastructure.”
Databricks Model Serving removes the complexity of building and operating these systems and offers native integrations across the lakehouse, including Databricks’ Unity Catalog, Feature Store and MLflow. It delivers a highly available, low latency service for model serving, giving businesses the ability to easily integrate ML predictions into their production workloads. Fully managed by Databricks, Model Serving quickly scales up from zero and back down as demand changes, reducing operational costs and ensuring customers pay only for the compute they use.
“As a leading global appliance company, Electrolux is committed to delivering the best experiences for our consumers at scale — we sell approximately 60 million household products in around 120 markets every year. Moving to Databricks Model Serving has supported our ambitions and enabled us to move quickly: we reduced our inference latency by 10x, helping us deliver relevant, accurate predictions even faster,” said Daniel Edsgärd, Head of Data Science at Electrolux. “By doing model serving on the same platform where our data lives and where we train models, we have been able to accelerate deployments and reduce maintenance, ultimately helping us deliver for our customers and drive more enjoyable and sustainable living around the world.”
Databricks’ unified, data-centric approach to machine learning from the lakehouse enables businesses to embed AI at scale and allows models to be served by the data and ML training platform. Lakehouse provides a consistent view of data throughout the entire ML lifecycle, which accelerates deployments and reduces errors, without having to stitch together disparate services. With Databricks, organizations can manage the entire ML process – from data preparation and experimentation to model training, deployment and monitoring – all in one place. Databricks Model Serving integrates with Lakehouse Platform capabilities, including:
- Feature Store: Provides automated online lookups to prevent online/offline skew. Define features once during model training, and Databricks will automatically retrieve and join the relevant features in the future.
- MLflow Integration: Natively connects to MLflow Model Registry, enabling fast and easy deployment of models. After providing the underlying model, Databricks will automatically prepare a production-ready container for model deployment.
- Unified Data Governance: Manage and govern all data and ML assets with Unity Catalog, including those consumed and produced by model serving.
Databricks is committed to driving innovation with its Lakehouse Platform and delivering more capabilities that make powerful, real-time machine learning accessible to any organization. This includes new quality and diagnostic features coming soon for Databricks Model Serving which will automatically capture requests and responses in a Delta table to monitor and debug models and generate training data sets. Databricks is also enabling GPU-based inference support, which is available in preview.
To learn more about how Databricks enables production ML at scale with the Lakehouse, read the Databricks blog or sign up to join the ML Virtual Event on March 14, 2023, to hear from Databricks product leaders and customers.
Visit AITechPark for cutting-edge Tech Trends around AI, ML, Cybersecurity, along with AITech News, and timely updates from industry professionals!