Skip to content

GRAFANA 2026-05-01 Tier 2

Read original ↗

Grafana — Faster fixes, less context sharing: how Grafana Assistant learns your infrastructure before you even ask

Summary

Grafana Labs' launch post for infrastructure memory in Grafana Assistant — a zero-configuration background capability that runs a "swarm of AI agents" across a Grafana Cloud stack's connected Prometheus, Loki, and Tempo data sources, extracts a fixed five-category knowledge schema for each discovered service group (identity + key metrics + deployment topology + dependencies + log structure), stores those extractions as searchable chunks in a vector database, and serves them back at question time via semantic search "in milliseconds." The whole pipeline refreshes weekly on its own and inherits the access-control boundary of the data sources it was generated from — users see only memories derived from data sources they can already read. Load-bearing framing: "Assistant runs this infrastructure memory in the background with zero configuration" + "This isn't a feature you configure, enable, or maintain." The architectural bet is that customer telemetry itself is a sufficient discovery substrate — the data is already there, the agent just has to systematically walk it and summarise.

Key takeaways

  1. A "swarm of AI agents" does the discovery, not the user. The four-stage pipeline is explicitly agent-driven: "A swarm of AI agents does the heavy lifting: (1) Data source discovery… (2) Metrics scans: Agents query your Prometheus data sources in parallel to find services, deployments, and infrastructure components. (3) Enrichments via logs and traces: Loki and Tempo data sources get correlated with their corresponding metrics, adding context about log formats, trace structures, and service dependencies. (4) Structured knowledge generation: For each discovered service group, agents produce documentation…" (Source: this post.) Canonicalised as patterns/swarm-of-discovery-agents-for-context-prebuild.
  2. The knowledge schema is a fixed five-category contract per service group. Every discovered service group is summarised into the same five axes: (a) Identity and purpose"what the service is, what it does, which namespace and cluster it belongs to, and what technology stack it uses"; (b) Key metrics"the metric names and labels relevant to the service, including golden signals such as latency, error rate, traffic, and saturation. Not generic guesses, but the actual metric names from your Prometheus data sources"; (c) Deployment topology"Kubernetes resources, replica counts, scaling configurations, and container details"; (d) Dependencies"Upstream and downstream service connections, database and cache relationships, message queue interactions, and external integrations"; (e) Log structure"Available log labels and their values, detected log formats (JSON, logfmt, or unstructured), common patterns, and extracted field names." Canonicalised as patterns/five-category-service-knowledge-schema — a distinct instance from Meta's per-module precompute engine (which produces freeform markdown) and Cloudflare's "Agents that remember" extraction pipeline (which produces thread summaries).
  3. Telemetry itself is the discovery substrate. No agent questionnaire, no service-catalog integration, no manual annotation. "Your existing telemetry data is the input. The assistant reads what's already in your Prometheus, Loki, and Tempo data sources and builds its understanding from there. If you have metrics, you get this infrastructure memory capability." Canonicalised as concepts/telemetry-as-discovery-substrate. The structural bet: any service worth asking the agent about is already emitting metrics / logs / traces; therefore the observability tier is the knowledge substrate. This eliminates a whole class of service-catalog staleness failures common to Backstage-style approaches where the catalog is a hand-maintained register that drifts from the actual infrastructure.
  4. Storage is a vector database, retrieval is semantic search "in milliseconds". "This knowledge is stored as searchable chunks in a vector database, so when you or the assistant need information about a specific service, it can be retrieved in milliseconds through semantic search." The memory tier is therefore not a structured catalog (like the Grafana Cloud entity graph) or a raw-log index — it's an embedding store keyed by natural-language queries over vector similarity. Canonicalised as concepts/semantic-search-over-agent-memory.
  5. Weekly refresh is the published cadence. "The whole process refreshes automatically on a weekly cadence, so your assistant's understanding of your infrastructure stays current as your environment evolves." This is the frequency-at-which-context drift is bounded — weekly is slow enough that the swarm's inference cost is amortised, fast enough that most infra evolution (new deployment, renamed metric, added service dependency) is absorbed within one iteration. Canonicalised as patterns/weekly-refresh-cadence-for-agent-context. Users can also "trigger a manual scan if you want to refresh the knowledge base ahead of the next automatic cycle" — an escape hatch for deliberate refresh after a planned rollout.
  6. Access-control inheritance: memory is scoped to the data sources it came from. "Assistant also respects your organization's access controls. Each memory is linked to the data sources used to generate it, so users only see knowledge derived from data sources they have permission to access." The ACL design decision is load-bearing: instead of copying the data source's ACL onto each generated memory chunk and re-enforcing it at query time, Grafana tags each memory with its generating data source and filters by the user's data-source access list at retrieval time. Canonicalised as concepts/acl-propagated-agent-memory. Failure mode avoided: a developer with access only to namespace=frontend Prometheus data cannot exfiltrate a summary of namespace=payments service dependencies just because the memory extraction was run by a higher-privileged service account.
  7. Service group as the unit of knowledge extraction. The post is explicit that extraction happens per service group, not per individual metric / log stream / K8s workload: "For each discovered service group, agents produce documentation covering five areas…" and "You can review what the assistant has learned by navigating to the Assistant settings and browsing the discovered service groups." The service-group abstraction — a clustered collection of workloads that share an identity (e.g. all pods behind the checkout-api Service) — is the granularity at which the schema's five categories are coherent. Below that (single pod) is too narrow; above it (entire namespace) is too broad. Canonicalised as concepts/service-group.
  8. Zero configuration as the headline UX commitment. The post uses this phrase four times in different framings: "zero configuration", "This isn't a feature you configure, enable, or maintain", "There are no setup steps, no configuration files, no scheduled jobs to manage", "If you have metrics, you get this infrastructure memory capability." The commitment is architecturally load-bearing: any requirement that users pre- declare their services, annotate their telemetry, or maintain a separate catalog would defeat the pitch. The swarm must succeed purely on what's already emitted.
  9. The claimed outcome is context-parity between experienced and new responders. "But this functionality is especially powerful for teams where not everyone has the full picture of the infrastructure. A developer investigating an issue in their service can ask about upstream dependencies and get accurate answers, even if they've never looked at those systems before." This is the target user story — the Assistant pre-loads what a senior SRE would know about the upstream services, so a junior developer on their first incident can query that knowledge in natural language. Complement to concepts/tribal-knowledge at the observability-stack altitude: Meta's precompute-engine captures tribal knowledge in code; Grafana captures it in infrastructure-shape observed through telemetry.
  10. "Not generic guesses, but the actual metric names." A minor but load-bearing quote: the golden-signals framing is generic observability vocabulary, but the Assistant's memory records the actual PromQL metric names from this customer's Prometheushttp_request_duration_seconds, process_cpu_seconds_total, whatever the customer's instrumentation emits. When the agent is asked "show me latency for checkout-api," it already knows the right PromQL expression to write. This is the mechanism by which the Assistant transitions from a generic LLM-over-docs experience into a per-customer-grounded one.

Systems / concepts / patterns introduced

Systems

  • systems/grafana-assistant — Grafana's AI assistant product. This ingest canonicalises it on the wiki for the first time (the earlier 2026-04-22 AI-Observability and Assistant-everywhere posts were skipped as pure marketing). Named surface: the infrastructure-memory subsystem, the vector-DB-backed memory store, the weekly-refresh scheduler, the per-memory ACL linkage.

Concepts

  • concepts/agent-infrastructure-memory — the generalised concept: a background-populated, long-lived, queryable data store describing the user's runtime infrastructure, built for AI-agent consumption.
  • concepts/telemetry-as-discovery-substrate — the architectural bet that live observability data (metrics + logs + traces) is a sufficient input to discover infrastructure shape without additional declarative input.
  • concepts/semantic-search-over-agent-memory — retrieval model: natural-language query → vector similarity over embedded memory chunks → relevant service knowledge in milliseconds.
  • concepts/acl-propagated-agent-memory — ACL design for agent memories: tag each memory with its data-source origin and enforce the data source's existing ACL at retrieval time, rather than copying ACLs onto the memory itself.
  • concepts/service-group — the unit of knowledge extraction; a clustered collection of workloads sharing an identity that is coherent for the five-category schema.

Patterns

  • patterns/swarm-of-discovery-agents-for-context-prebuild — use many parallel agents to walk a customer's telemetry surface area (data sources → metrics → enrichment via logs/traces → per-service-group summarisation) and produce a searchable corpus of structured service knowledge ahead of any user question.
  • patterns/five-category-service-knowledge-schema — the fixed five-axis schema (identity, key metrics, deployment topology, dependencies, log structure) as the canonical contract for per-service-group knowledge; distinct from Meta's freeform markdown per-module format.
  • patterns/weekly-refresh-cadence-for-agent-context — bounded-staleness discipline for agent memory: rebuild the whole corpus on a weekly background cadence with optional manual trigger.

Operational numbers

  • Retrieval latency: "retrieved in milliseconds through semantic search" — no p50/p99 disclosed.
  • Refresh cadence: weekly (automatic) + on-demand manual trigger.
  • Pipeline stages: 4 (data-source discovery → metrics scan → log/trace enrichment → structured knowledge generation).
  • Knowledge categories per service group: 5 (identity, metrics, topology, dependencies, logs).
  • Required data sources: Prometheus is the minimum; Loki and Tempo add enrichment. "If you have metrics, you get this infrastructure memory capability."

Caveats

  • Tier-2 launch post with substantial product-marketing framing. The architecture content is real (swarm-of-agents pipeline + five-category schema + vector-DB memory + semantic retrieval + ACL propagation + weekly refresh), but it is interleaved with a "faster fixes, less context sharing" pitch and there is no mechanism-level disclosure of: which LLM(s) power the swarm; which vector DB is used; embedding model; chunk strategy; retrieval top-K; how dependencies are inferred from traces; how service-group boundaries are detected from Prometheus labels; how the weekly refresh handles partial failures; whether memories are deleted when source data sources are removed; how inherited ACLs are re-evaluated when data-source permissions change between refreshes. These are all architecturally interesting questions that a future deep-dive post would need to answer.
  • No production numbers. No disclosure of total memories per stack, compute cost per weekly refresh, LLM token spend, vector-DB storage footprint, query QPS, or any tail-latency percentile beyond "milliseconds."
  • "Service group" boundary is undefined in the post. Presumably inferred from Prometheus service-like labels or K8s workload groupings, but the exact detection heuristic is not disclosed.
  • "Swarm of AI agents" is marketing framing to an unspecified extent. The word "swarm" could refer to a literal multi-agent system with coordination (plausible given Grafana's parallel work on agentic observability surfaces like gcx and the 2026-04-21 AI Observability launch), or to a more conventional parallel map-reduce pipeline dressed up in agent vocabulary. No architecture diagram, no agent inventory, no inter-agent protocol disclosure.
  • Claim that this is a differentiator vs. alternatives is uncorroborated. The post closes with the aspirational framing "an assistant that genuinely understands the infrastructure it's helping you observe and knows your system well enough to ask the right questions on its own" — forward-looking claim without comparative numbers against the Datadog Bits / New Relic AI / Honeycomb Query Assistant baselines.
  • Private-preview-to-GA status is unstated in this post. Implied "available today" framing but no explicit GA date or region rollout schedule.
  • Does not address: cross-stack knowledge (does memory span multiple Grafana Cloud stacks in the same account?); on-prem Grafana Enterprise / OSS compatibility (the 2026-04-22 Assistant-everywhere post announced self-managed availability — whether this capability follows is unstated); how the Assistant reconciles conflicting information across refresh cycles; whether the five-category schema is extensible by customers.

Source

Last updated · 445 distilled / 1,275 read