Skip to content

SYSTEM Cited by 2 sources

Gumby (Dropbox 7th-gen flexible GPU platform)

Definition

Gumby is Dropbox's 7th-generation flexible GPU platform — a Crush compute server with PCIe GPU slots added on top. Positioned for mixed inference and embedding workloads where the right accelerator size varies widely across tenants. Rolled out 2025 alongside systems/godzilla (the dense multi-GPU tier).

Form-factor envelope

  • TDP range: 75 W → 600 W
  • Form factors: HHHL (half-height half-length) and FHFL (full-height full-length)

The wide envelope is the whole point — Gumby's name "signals flexibility" (Dropbox's framing). Rather than one GPU SKU per server, Gumby can host anything from a low-end inference card to a high-TDP training-class GPU in a CPU-dominant chassis.

Target workloads

Relationship to Godzilla

  • Gumby — one or few GPUs, variable TDP, CPU-dominant chassis, inference-leaning.
  • Godzilla — up to 8 interconnected GPUs, GPU-dominant chassis, training-leaning.

Together they form Dropbox's 7th-gen accelerator tier; neither one alone serves the full range of workloads that Dash and related ML features demand.

Software axis — low-bit inference

Gumby's inference workloads are the substrate for Dropbox's low-bit inference strategy: multiple quantization formats (pre-MXFP AWQ/HQQ- style A16W4 or A8W8; MXFP / NVFP on Tensor Core-capable silicon) run across Gumby depending on whether each workload is memory-bound vs compute-bound. No single strategy fits all Dash workloads — conversational AI, multimodal search, document understanding, and speech processing each hit different points on the latency vs throughput frontier (Source: sources/2026-02-12-dropbox-how-low-bit-inference-enables-efficient-ai).

Seen in

Last updated · 200 distilled / 1,178 read