PATTERN Cited by 1 source
Rapid fleet-patching via managed service¶
What it is¶
When a security vulnerability ships in software the vendor also operates as a managed service, use the managed-service deployment substrate itself as the primary remediation channel:
- Develop + test the patch internally.
- Roll it to the managed-service fleet under the vendor's staged-rollout discipline — canary → region-by-region → full fleet.
- Honour customer maintenance-window contracts with pre-notification override for the urgent patch.
- Publish the CVE + distribute patches to self-hosted tiers after the managed-service fleet is safe.
The managed service is the primary defence mechanism for the plurality of users; self-hosted users are informed via CVE + community channels on a separate track.
When to use¶
- Vulnerability ships in both a managed service and a self-hosted distribution.
- Most active users are on the managed tier (common for saas databases, edge platforms, saas devtools).
- Vendor operates a fleet-patching capability with per-instance state tracking + rollback + observability.
- Vulnerability class is patchable without customer action (no config migration, no schema change, no client-side upgrade).
- Disclosure timeline can tolerate "patch first, disclose second" — i.e. there isn't a known in-the-wild exploit the public needs to defend against immediately.
Don't use when the vulnerability requires customer-side action (credential rotation, config migration, client-library upgrade), when detection itself is already public, when the managed-service fleet is small compared to self-hosted users, or when the managed-service deployment substrate cannot deploy the patch (e.g. client-side vulnerability).
Structure¶
| Phase | Owner | Activity | Typical duration |
|---|---|---|---|
| 1. Detection | Vendor security eng | Internal discovery + triage | T=0 |
| 2. Validate + develop + test fix | Vendor eng | Patch built + tested | T+hours to T+days |
| 3. Managed-fleet rollout planning | Vendor SRE + security | Rollout plan + staging + per-instance state tracking | T+days |
| 4. Managed-fleet rollout (non-maintenance-window customers) | Vendor | Canary → staged → full fleet | T+days |
| 5. Pre-notification to maintenance-window customers | Vendor | Email / console notice; ~15-h lead time per MongoDB precedent | T+~5 days |
| 6. Managed-fleet rollout (maintenance-window customers) | Vendor | Forced patch after notification | T+~6 days |
| 7. Public CVE publication | Vendor | Technical disclosure artefact | T+~7 days |
| 8. Self-hosted tier patch + notification | Vendor → customer | CVE + patched-version distribution + community channels | T+~7 – T+N days |
| 9. Operational retrospective | Vendor | Timeline + trust-layer blog post | T+N days |
MongoDB CVE-2025-14847 is the canonical instantiation — detection 2025-12-12 19:00 ET, majority of Atlas fleet 2025-12-17, full fleet 2025-12-18, CVE 2025-12-19, community forum 2025-12-23, this retrospective 2025-12-30.
Properties it gives you¶
- Plurality-of-users-first protection. Managed-tier users are safe at T=0 of the public disclosure clock; self-hosted users learn alongside attackers.
- Collapsed coordinated-disclosure window. Not 90 days — as little as 7 days, bounded by how long fleet-patching takes rather than by a fixed industry convention.
- Maintenance-window contract preserved. The customer-courtesy window still means something for routine updates; the emergency path is pre-notified rather than silent.
- Three-tier rollout respected. Atlas / EA / Community each remediate through an appropriate channel for their responsibility line.
- Trust-layer communication decoupled from CVE. CVE is the technical artefact (class / severity / versions); vendor blog post is the operational-trust artefact (timeline / decisions).
- Internal-discovery flywheel. Vendors who invest in internal security research can use this pattern routinely; vendors who rely on external reports cannot set the disclosure clock.
Properties it does NOT give you¶
- Zero-day protection. Pre-detection exploitation is not addressed; DiD layers below the patched defect must catch what the patch didn't.
- Self-hosted-tier velocity. EA + Community users remain on the slower track; the pattern optimises the managed tier.
- Arbitrary-patch safety. Velocity is not a substitute for staging — a fleet-wide patch that itself causes outages is a Cloudflare 2025-12-05 / 2025-11-18 shape of self-inflicted incident. Fleet-patching velocity ↔ staged-rollout discipline is a two-sided constraint.
- Exploit-in-the-wild scenarios. If a working exploit is already public, the coordinated-disclosure clock is already running on defender-time and patch-first-disclose-later no longer applies.
- Customer-action-required vulnerabilities. Credential rotations, config migrations, and client-library upgrades can't be patched server-side; the pattern falls back to ordinary coordinated disclosure with reporter-vendor-embargo coordination.
Operational requirements¶
- Per-instance state tracking — which customers are on which binary version, which patches landed, which deferred.
- Rollback tooling — fleet-wide fast revert if the patch misbehaves.
- Customer-courtesy contract — documented maintenance window + emergency- override policy before the first emergency, not during it.
- Communication infrastructure — status page, per-customer email, community forum, CVE-publication pipeline.
- Retrospective discipline — trust-layer publication of the timeline + decisions even when nothing failed; MongoDB's 2025-12-30 post is the canonical shape.
Seen in¶
- sources/2025-12-30-mongodb-server-security-update-december-2025 — Atlas fleet of hundreds of thousands of instances patched in ~6 days; CVE publication 1 day after fleet remediation; EA + Community handled through community forum + patched builds. Canonical wiki instance.
Related¶
- concepts/fleet-patching — the operational capability this pattern realises.
- concepts/coordinated-disclosure — the industry norm this pattern collapses.
- concepts/maintenance-window — the customer-courtesy contract this pattern negotiates.
- concepts/shared-responsibility-model — the line that determines which customers get the managed-tier vs self-hosted- tier rollout shape.
- concepts/defense-in-depth — the broader security posture this pattern contributes to at the trust-communication layer.
- patterns/pre-disclosure-patch-rollout — the disclosure- sequencing pattern this pattern enables.
- patterns/staged-rollout — the safety discipline this pattern must apply to its fleet-patch velocity.
- patterns/progressive-configuration-rollout — the config-push sibling; a variant of the same fleet-wide-rollout problem.