Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Hidden Layer

Hidden Layer

An intermediate neural network layer that learns internal representations of data.

Year: 1986Generality: 796
Back to Vocab

A hidden layer is any layer in a neural network that sits between the input layer and the output layer, invisible to the raw data and the final predictions alike. Each neuron in a hidden layer receives signals from the previous layer, multiplies them by learned weights, adds a bias term, and passes the result through a nonlinear activation function such as ReLU or sigmoid. This transformation allows the network to encode increasingly abstract features of the input — edges and textures in early vision layers, semantic concepts in deeper ones — rather than relying on hand-crafted feature engineering.

The power of hidden layers lies in their collective ability to approximate arbitrarily complex functions. By stacking multiple hidden layers, a network builds a hierarchy of representations where each layer refines the abstractions produced by the one before it. The universal approximation theorem formalizes this intuition, showing that even a single sufficiently wide hidden layer can represent any continuous function — though in practice, depth tends to be far more parameter-efficient than width for learning structured data.

The number, width, and connectivity pattern of hidden layers are among the most consequential architectural decisions in model design. Shallow networks with one or two hidden layers work well for tabular data and simpler tasks, while deep networks with dozens or hundreds of layers — as in ResNets or Transformers — are necessary for high-dimensional problems like image classification, speech recognition, and language modeling. Techniques like batch normalization, residual connections, and dropout were developed largely to make training deep stacks of hidden layers stable and effective.

Hidden layers became practically significant with the popularization of backpropagation in 1986, which provided an efficient algorithm for computing gradients through multiple layers and updating weights accordingly. Before this, training networks with more than one hidden layer was computationally intractable. The subsequent decades of research into activation functions, initialization schemes, and optimization algorithms have all been aimed at unlocking the representational capacity that hidden layers provide.

Related

Related

Model Layer
Model Layer

A discrete computational stage in a neural network that transforms input representations progressively.

Generality: 794
Neural Network
Neural Network

A layered system of interconnected nodes that learns patterns from data.

Generality: 947
Hierarchy of Generalizations
Hierarchy of Generalizations

A layered framework where neural networks learn increasingly abstract data representations.

Generality: 695
Expressive Hidden States
Expressive Hidden States

Internal neural network representations that richly capture complex patterns and long-range dependencies.

Generality: 416
DNN (Deep Neural Network)
DNN (Deep Neural Network)

Neural networks with many layers that learn hierarchical representations from raw data.

Generality: 871
ANN (Artificial Neural Networks)
ANN (Artificial Neural Networks)

Layered computational models that learn from data by adjusting weighted connections.

Generality: 928