Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Observatory
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Whitening

Whitening

A data preprocessing technique that removes correlations and normalizes feature scales.

Year: 1986Generality: 550
Back to Vocab

Whitening is a data preprocessing transformation that converts input features into a representation with zero mean, unit variance, and no linear correlations between dimensions. The goal is to produce data whose covariance matrix is the identity matrix — a condition described as "white" by analogy to white noise, which has a flat power spectrum. This standardized form removes redundant statistical structure from the input, placing all features on equal footing before they enter a learning algorithm.

The most common implementations rely on eigendecomposition of the data's covariance matrix. In PCA whitening, the data is first projected onto the principal components and then scaled so each component has unit variance, simultaneously decorrelating and normalizing the features. ZCA (Zero-phase Component Analysis) whitening applies an additional rotation that keeps the transformed data as close as possible to the original space, which can be advantageous when the spatial or semantic structure of inputs matters — as with image pixels. Both approaches require estimating the covariance matrix from training data, which can be expensive or unstable in very high dimensions, motivating approximations and regularization strategies.

Whitening matters in machine learning because poorly scaled or correlated inputs can dramatically slow optimization. When features differ in magnitude or are strongly correlated, the loss landscape becomes elongated and ill-conditioned, causing gradient descent to oscillate or converge slowly. Whitening reshapes this landscape toward a more spherical geometry, allowing larger learning rates and faster convergence. It was particularly influential in early neural network research and remains relevant in unsupported feature learning, generative models, and reinforcement learning settings where input normalization is not otherwise handled.

Modern deep learning has partially displaced explicit whitening with techniques like Batch Normalization, which performs an adaptive, layer-wise normalization during training. Nevertheless, whitening retains practical importance as an offline preprocessing step for shallow models, kernel methods, and scenarios where controlling the statistical properties of inputs is essential for reproducibility and stability.

Related

Related

PCA (Principal Component Analysis)
PCA (Principal Component Analysis)

Dimensionality reduction technique that projects data onto its highest-variance directions.

Generality: 871
Batch Normalization
Batch Normalization

A technique that normalizes layer inputs to accelerate and stabilize neural network training.

Generality: 794
Layer Normalization
Layer Normalization

Normalizes activations across features within a layer to stabilize neural network training.

Generality: 731
Dimensionality Reduction
Dimensionality Reduction

Transforming high-dimensional data into fewer dimensions while preserving essential structure.

Generality: 838
LN (Layer Normalization)
LN (Layer Normalization)

A normalization technique that stabilizes neural network training by standardizing each layer's inputs.

Generality: 731
Variance Scaling
Variance Scaling

A weight initialization strategy that preserves consistent activation variance across neural network layers.

Generality: 620