AI's ability to apply implicit, everyday world knowledge to novel situations.
Commonsense reasoning refers to the capacity of AI systems to draw on implicit, background knowledge about the world — the kind of understanding humans acquire through lived experience rather than formal instruction. This includes knowing that objects fall when dropped, that people feel hurt when insulted, or that a restaurant visit typically ends with paying a bill. Unlike factual recall or logical deduction from explicit premises, commonsense reasoning requires systems to fill in unstated assumptions, resolve ambiguity, and make plausible inferences about situations they have never directly encountered.
The core challenge lies in the breadth and informality of the knowledge involved. Commonsense encompasses physical intuitions, social norms, causal relationships, temporal reasoning, and cultural conventions — all of which humans absorb implicitly but which are notoriously difficult to encode in machine-readable form. Early approaches, such as Doug Lenat's Cyc project begun in 1984, attempted to manually curate millions of everyday facts and rules into structured knowledge bases. While ambitious, these efforts revealed just how vast and context-dependent commonsense knowledge truly is, and hand-crafted ontologies struggled to scale.
Modern approaches leverage large-scale language models and neural networks trained on massive text corpora, which appear to absorb substantial commonsense knowledge implicitly through exposure to human-generated language. Benchmarks such as Winograd Schema Challenge, CommonsenseQA, and HellaSwag have been developed to probe these capabilities systematically. Transformer-based models have shown impressive performance on many such benchmarks, though researchers debate whether high scores reflect genuine reasoning or sophisticated pattern matching — a distinction with significant practical implications.
Commonsense reasoning matters because it underpins nearly every real-world AI application that involves natural language understanding, autonomous decision-making, or human-robot interaction. A system that lacks commonsense will misinterpret instructions, fail in novel situations, and behave in ways that seem bizarre or unsafe to human users. Closing the commonsense gap remains one of the central open problems in AI, sitting at the intersection of knowledge representation, natural language processing, and cognitive science.