Use cases

Region availability

Lakebase Autoscaling is the latest version of Lakebase, with autoscaling compute, scale-to-zero, branching, and instant restore. For supported regions, see Region availability. If you are a Lakebase Provisioned user, see Lakebase Provisioned.

Lakebase Autoscaling supports three primary patterns: serve lakehouse data in Postgres, run an application backend, and power AI agents and ML. Each pattern uses Postgres alongside Unity Catalog to give your application a low-latency database that stays in sync with the lakehouse.

Serve lakehouse data

Lakehouse rows synced into Lakebase Postgres for low-latency app reads

Synced tables bring Unity Catalog data into your Lakebase database for low-latency transactional reads. Pick a source table, choose a sync mode, and the pipeline is fully managed. No sync scripts, no external orchestration, no jobs to monitor. Continuous mode keeps data within seconds of the source. Triggered mode balances freshness and cost with scheduled incremental updates. Your application always serves the latest analytics alongside its own operational data.

First steps Learning path

Application backend

Standard Postgres clients connect to Lakebase Postgres

Your application connects to Lakebase the same way it connects to any Postgres database. Use the drivers and frameworks you already know. When your app gets a traffic spike, autoscaling adds compute without dropping connections. When traffic stops, scale-to-zero suspends the database and reactivates in hundreds of milliseconds on the next query. You don't provision for peak and you don't pay for idle. For development, branching gives every developer an isolated copy of the production database with no data seeding, no storage duplication, and no waiting.

First steps Learning path

AI agents and ML

Agents store memory and models read features through Lakebase Postgres, synced with the lakehouse

Lakebase serves as the backend for AI agent memory and real-time feature serving. Agents built with LangGraph or the OpenAI Agents SDK store conversation state and long-term memory in Postgres. Models served with Mosaic AI access feature data through Online Feature Stores that are powered by Lakebase Autoscaling. Both benefit from automatic scaling, scale-to-zero, and Unity Catalog governance.

First steps Learning path