Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Decision Tree

Decision Tree

A tree-structured model that makes predictions through sequential feature-based splits.

Year: 1986Generality: 838
Back to Vocab

A decision tree is a supervised learning model that represents a series of hierarchical decisions as a branching tree structure. Each internal node tests a specific input feature, each branch corresponds to the outcome of that test, and each leaf node holds a final prediction — a class label for classification tasks or a continuous value for regression. This architecture closely mirrors human reasoning processes, making decision trees among the most interpretable models in machine learning.

Training a decision tree involves recursively partitioning the input data by selecting the feature and threshold that best separates the target values at each node. For classification, common splitting criteria include Gini impurity and information gain (based on entropy); for regression, variance reduction is typically used. The algorithm greedily selects the locally optimal split at each step, continuing until a stopping condition is met — such as a maximum tree depth, a minimum number of samples per leaf, or pure node composition. Without constraints, trees tend to overfit the training data, memorizing noise rather than learning generalizable patterns. Pruning techniques, either during growth (pre-pruning) or after (post-pruning), help address this by removing branches that provide little predictive power.

Decision trees gained significant traction in machine learning through Ross Quinlan's ID3 algorithm in the 1980s, followed by C4.5 and C5.0, which added support for continuous attributes, missing values, and more robust pruning. These advances made decision trees practical for real-world datasets. A key strength is their ability to handle both numerical and categorical features without requiring normalization or extensive preprocessing, and their learned structure can be directly inspected and explained to non-technical stakeholders.

While individual decision trees are often outperformed by more complex models, they serve as the foundational building block for powerful ensemble methods such as Random Forests and Gradient Boosted Trees, which combine many trees to dramatically improve predictive accuracy and robustness. Understanding decision trees is therefore essential not only for their direct applications but as a gateway to some of the most effective algorithms in modern machine learning.

Related

Related

Root (Decision Tree)
Root (Decision Tree)

The topmost node in a decision tree, representing the first splitting decision.

Generality: 575
Random Forest
Random Forest

An ensemble of decision trees that improves accuracy and resists overfitting.

Generality: 796
Ensemble Methods
Ensemble Methods

Combining multiple trained models to produce predictions stronger than any single model.

Generality: 771
Bagging
Bagging

Ensemble method that trains multiple models on random data subsets and aggregates predictions.

Generality: 694
Ensemble Algorithm
Ensemble Algorithm

Combines multiple models to boost predictive accuracy, robustness, and generalization.

Generality: 796
Ensemble Learning
Ensemble Learning

Combining multiple models to produce predictions more accurate than any single model.

Generality: 836