When a single neuron encodes multiple unrelated concepts due to representational compression.
Incidental polysemanticity refers to the phenomenon in which individual neurons or internal representations within a neural network come to encode multiple distinct, often semantically unrelated concepts. Unlike deliberate design choices, this property emerges spontaneously as a byproduct of training — the network discovers that packing several meanings into a single unit is an efficient strategy for handling the enormous diversity of patterns present in real-world data. The term distinguishes this emergent behavior from intentional or architectural forms of polysemanticity, emphasizing that it arises without explicit instruction.
The mechanism behind incidental polysemanticity is closely tied to the concept of superposition, where a network with fewer neurons than concepts it needs to represent learns to overlap those representations in activation space. Because many features are sparse — they rarely co-occur — the network can afford to conflate them without suffering significant interference during typical inputs. The result is that a single neuron may fire strongly for, say, a legal term in one context and a culinary ingredient in another, with no obvious shared structure between the two.
This phenomenon poses significant challenges for mechanistic interpretability, the field concerned with reverse-engineering what neural networks have learned. When neurons are polysemantic, standard techniques like probing classifiers or activation maximization yield ambiguous or misleading results, making it harder to assign clean functional roles to individual units. It also raises safety concerns: if a neuron's behavior is context-dependent in opaque ways, predicting or auditing a model's responses in edge cases becomes substantially more difficult.
Research into incidental polysemanticity has accelerated alongside the growth of large language models, with groups at Anthropic, OpenAI, and academic institutions working to quantify its prevalence and develop tools to mitigate it. Techniques such as sparse autoencoders have been proposed to decompose polysemantic neurons into more interpretable, monosemantic features. Understanding and addressing incidental polysemanticity is now considered a central challenge in building transparent and reliably controllable AI systems.