Large-scale models that represent and reason over abstract, compositional concepts rather than raw tokens.
Large Concept Models (LCMs) are a class of foundation models designed to represent, compose, and reason over high-level conceptual primitives rather than—or in addition to—raw tokens or pixels. Where conventional large language models operate primarily on surface-level token sequences, LCMs aim to encode semantically meaningful units such as objects, relations, affordances, procedures, and abstract categories into structured latent spaces where these concepts can be explicitly manipulated, combined, and queried. This orientation toward concept-level representations draws on ideas from disentangled representation learning, symbolic AI, and multimodal alignment.
Architecturally, LCMs span a wide range of approaches: contrastive and self-supervised pretraining that clusters semantically related inputs, structured latent-variable models that expose interpretable concept dimensions, neuro-symbolic hybrids that interface neural representations with symbolic reasoning engines, and concept bottleneck models that route predictions through human-interpretable intermediate variables. A defining characteristic is that the learned concept space supports compositional generalization—the ability to correctly handle novel combinations of known concepts—which standard token-prediction objectives do not explicitly encourage. Evaluation therefore emphasizes compositional benchmarks, causal intervention tests, and concept disentanglement metrics rather than perplexity or accuracy alone.
The practical motivation for LCMs is substantial. By exposing concept-level interfaces, these models promise improved sample efficiency when transferring to new tasks, stronger causal understanding, greater interpretability for human oversight, and more reliable behavior in safety-critical settings. Downstream applications include robotics (where grounding actions to object affordances matters), scientific modeling (where abstract variables must be identified and manipulated), and human–AI collaboration (where shared conceptual vocabulary aids communication). The term gained explicit traction in the research community around 2024–2025 as interest grew in augmenting large multimodal models with structured concept representations and APIs for compositional reasoning.
LCMs sit at the intersection of several long-running research threads—cognitive-science-inspired concept learning, structured generative models, and large-scale pretraining—and represent an effort to reconcile the empirical power of modern foundation models with the systematic, interpretable reasoning that symbolic approaches have historically offered. Whether through soft concept vectors, discrete symbolic APIs, or hybrid architectures, the central bet is that making concepts first-class citizens in model design will yield more robust, transferable, and trustworthy AI systems.