SYSTEM Cited by 1 source
Unity AI Gateway (Databricks)¶
Unity AI Gateway is Databricks' productised instance of the AI-gateway provider abstraction pattern, specialised to coding agents + MCP integrations rather than just application LLM calls. Its job is to be the single governance + cost + telemetry plane for all coding-tool traffic in a Databricks customer's fleet.
Three-pillar architecture (from the 2026-04-17 launch post)¶
- Centralised security and audit.
- Every agent data-access flow logged in Unity Catalog (same governance substrate as Lakehouse data + ML assets).
- All tracing in MLflow (specifically MLflow 3 GenAI tracing — named for Claude Code integration).
- MCP servers "managed in Databricks" — the gateway is the policy point for MCP traffic, not just LLM traffic.
- Single-identity plane: developers authenticate once with Databricks credentials for all tools (GitHub, Atlassian, etc.), "no separate logins per service".
- Single bill and cost limits.
- Foundation Model API provides first-party inference for OpenAI, Anthropic, Gemini, and open models like Qwen.
- Admins can also "bring external capacity in", extending governance "to all your tokens, regardless of where they flow" — patterns/unified-billing-across-providers.
- Gateway-enforced budgets are per-developer, not per-tool — admins give each developer one budget and the developer burns it on whichever tool of choice (Cursor / Codex / Gemini CLI / Claude Code / …).
- Full observability in the Lakehouse.
- Coding-tool metrics + traces land in Unity-Catalog-managed Delta tables via OpenTelemetry ingestion.
- Joinable with other Lakehouse datasets (Workday for adoption-by-org / region / seniority; PR-cycle data for velocity quantification) — patterns/telemetry-to-lakehouse.
- Surfaces rate-limit hits as a proactive capacity-planning signal.
Supported clients (at launch)¶
- Cursor
- Codex CLI
- Gemini CLI
- Claude Code — integration referenced via MLflow 3 tracing docs.
Relation to existing wiki entities¶
- Same shape as Cloudflare's internal AI engineering stack (sources/2026-04-20-cloudflare-internal-ai-engineering-stack) — single proxy, BYOK, central telemetry, single-identity plane. Databricks specialises for coding-agent clients + MCP governance where the Cloudflare instance specialised for internal application workloads.
- Parent pattern: patterns/ai-gateway-provider-abstraction.
- Audit substrate: systems/unity-catalog.
- Tracing substrate: systems/mlflow.
- Governed tool surface: systems/model-context-protocol.
What the post does not disclose¶
- Gateway internals: routing, fallback, rate-limiter algorithm, streaming handling, per-provider adapter shape.
- MCP-governance mechanics: how the gateway inspects MCP traffic, auth flow from coding-tool → gateway → MCP → data source.
- Telemetry schema landing in Delta tables.
- Latency / throughput / cost-per-token / adoption numbers.
Tier-3 Databricks post — ingested because the problem framing (coding-agent sprawl) and three-pillar architecture are substantive, not because the internals are disclosed.
Seen in¶
- sources/2026-04-17-databricks-governing-coding-agent-sprawl-with-unity-ai-gateway — launch post and currently the only ingested source.
Related¶
- systems/unity-catalog — audit + logging substrate.
- systems/mlflow — tracing substrate.
- systems/model-context-protocol — governed tool surface.
- systems/databricks-foundation-model-api — inference capacity the gateway routes to.
- concepts/coding-agent-sprawl — problem class it addresses.
- concepts/centralized-ai-governance — three-pillar framing.
- patterns/ai-gateway-provider-abstraction — parent pattern.
- patterns/central-proxy-choke-point — architectural posture.
- patterns/telemetry-to-lakehouse — observability shape.
- patterns/unified-billing-across-providers — cost posture.
- companies/databricks.