Skip to content

CONCEPT Cited by 1 source

Early Fusion (Multimodal)

Early fusion is the multimodal-ML strategy of concatenating raw (or minimally-encoded) modality inputs into a single feature vector before training, so one joint model learns directly over the combined input. It sits at the earliest possible fusion point in the pipeline — upstream of any modality-specific representation learning.

When it survives production

"Small, tightly controlled cohorts with consistent modality availability" (Source: sources/2026-04-22-databricks-multimodal-data-integration-production-architectures-for-healthcare-ai).

The requirements are strong because early fusion gives up two things other strategies keep:

  • No graceful degradation on missing modalities. Zero-padding a concatenated input lets the model silently learn that zeros are informative; at deployment, a genuinely missing modality is indistinguishable from a present-but-zero one.
  • Dimensionality imbalance swamps the learner. High- dimensional inputs (genomics variants) drown lower-dimensional ones (structured EHR fields) unless per-modality scaling + regularisation is engineered carefully.

Failure mode called out in ingested sources

"Scales poorly with high-dimensional genomics and large feature sets." (Source: sources/2026-04-22-databricks-multimodal-data-integration-production-architectures-for-healthcare-ai.) Genomics profiling can contribute ≥10⁶-dim feature vectors; EHR features sit in the 10²–10³ range. Direct concatenation hands the model a 1000:1 dimensionality gradient to fight through.

Contrast with other fusion strategies

  • concepts/intermediate-fusion — encode each modality separately, then merge hidden representations. Buys dimensionality-balance at the cost of more training discipline.
  • concepts/late-fusion — train per-modality models, combine predictions. Buys graceful degradation on missing modalities at the cost of ignoring cross-modal interactions until the very end.
  • concepts/attention-based-fusion — dynamic weighting across modalities and time. Buys temporal / interaction modelling at the cost of validation complexity.

See patterns/fusion-strategy-selection-by-deployment-reality for the decision-framework shape Databricks' post frames the fusion choice inside.

Seen in

  • sources/2026-04-22-databricks-multimodal-data-integration-production-architectures-for-healthcare-ai — Databricks names early fusion as the first of four fusion strategies and immediately scopes it to small / tightly- controlled cohorts with consistent modality availability; warns it "scales poorly with high-dimensional genomics and large feature sets". The post's decision framework — "match fusion to your deployment reality" — uses early fusion as the canonical example of a strategy that works in research cohorts and breaks under production sparsity.
Last updated · 517 distilled / 1,221 read