Main AI News:
Databricks has unveiled its latest innovation: the public preview of GPU and LLM optimization support for Databricks Model Serving. This transformative feature empowers users to effortlessly deploy a diverse array of AI models, including LLMs and Vision models, directly onto the Lakehouse Platform.
Databricks Model Serving heralds a new era of AI deployment by offering automatic optimization for LLM Serving. This means achieving top-tier performance without the burdensome task of manual configuration. What sets this product apart is its distinction as the first serverless GPU serving solution integrated into a unified data and AI platform. This all-encompassing platform facilitates the seamless creation and deployment of GenAI applications, covering the entire spectrum from data ingestion to model deployment and ongoing monitoring.