Breaking a complex problem into smaller, independently solvable subproblems.
Decomposition is a foundational strategy in AI and machine learning whereby a complex problem is partitioned into smaller, more tractable subproblems that can be solved independently or semi-independently before their solutions are recombined. Rather than confronting a monolithic challenge with a single monolithic method, decomposition allows practitioners to apply specialized techniques to each component — matching the right model, algorithm, or representation to each subproblem's particular structure. This modularity often yields solutions that are faster to develop, easier to debug, and more computationally efficient than end-to-end approaches applied to the full problem.
In machine learning, decomposition appears in many forms. Hierarchical models break prediction tasks into coarse-to-fine stages; mixture-of-experts architectures decompose the input space so that specialized sub-networks handle distinct regions; and multi-task learning decomposes a shared objective into constituent tasks that inform one another. In optimization, techniques like coordinate descent and block decomposition solve high-dimensional problems by iteratively optimizing subsets of variables, making otherwise intractable parameter spaces manageable. Probabilistic graphical models exploit conditional independence structure — itself a form of decomposition — to perform exact or approximate inference efficiently.
Decomposition is especially critical in domains where direct approaches are computationally infeasible. Large language models use attention mechanisms that decompose sequence relationships into pairwise interactions; reinforcement learning agents decompose long-horizon tasks into subgoals or options; and computer vision pipelines historically decomposed scene understanding into detection, segmentation, and recognition stages. Even neural architecture search and AutoML can be framed as decomposing the model-design problem into searchable components. The principle's power lies in its generality: whenever a problem has exploitable structure — temporal, spatial, hierarchical, or statistical — decomposition provides a principled way to leverage that structure for gains in efficiency, interpretability, and scalability.