Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Feature Importance

Feature Importance

Methods that rank input variables by their contribution to a model's predictions.

Year: 1995Generality: 728
Back to Vocab

Feature importance refers to a family of techniques in machine learning that quantify how much each input variable contributes to a model's predictive output. Rather than treating a trained model as a black box, feature importance methods expose the internal logic of predictions, revealing which signals the model relies on most heavily. This serves multiple practical purposes: it helps practitioners validate that a model is learning meaningful patterns rather than spurious correlations, guides feature selection to reduce dimensionality, and satisfies regulatory or business requirements for explainability.

The mechanics of feature importance vary significantly depending on the method and model type. In linear models, the magnitude of learned coefficients (after normalization) provides a direct measure of each feature's influence. Tree-based models such as Random Forests and Gradient Boosting Machines compute importance by tracking how much each feature reduces impurity (e.g., Gini impurity or variance) across all splits in the ensemble. Permutation importance, a model-agnostic approach, measures the drop in performance when a single feature's values are randomly shuffled, breaking its relationship with the target. SHAP (SHapley Additive exPlanations) values offer a theoretically grounded, game-theoretic framework that attributes each prediction to individual features consistently across any model architecture.

Each method carries its own assumptions and failure modes. Impurity-based importance in tree models tends to inflate the apparent importance of high-cardinality features. Permutation importance can be misleading when features are correlated, since shuffling one correlated feature may not fully disrupt the model's access to that information. SHAP addresses many of these issues but is computationally expensive for large datasets and complex models. Choosing the right method requires understanding both the model's structure and the data's statistical properties.

Feature importance has become indispensable as machine learning models are deployed in high-stakes domains such as healthcare, finance, and criminal justice, where understanding model behavior is not optional. It bridges the gap between predictive performance and human interpretability, enabling teams to debug models, communicate decisions to stakeholders, and build systems that are both accurate and trustworthy.

Related

Related

Random Forest
Random Forest

An ensemble of decision trees that improves accuracy and resists overfitting.

Generality: 796
Feature Design
Feature Design

Transforming raw data into informative inputs that improve machine learning model performance.

Generality: 792
Feature Extraction
Feature Extraction

Transforming raw data into compact, informative representations that improve model learning.

Generality: 838
Black Box Problem
Black Box Problem

The challenge of understanding why and how ML models reach their decisions.

Generality: 792
Interpretability
Interpretability

The degree to which humans can understand why an AI system made a decision.

Generality: 800
Attribute Sampling
Attribute Sampling

Selecting a random subset of features when training models to improve performance.

Generality: 521