Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. DQN (Deep Q-Networks)

DQN (Deep Q-Networks)

Reinforcement learning method combining Q-learning with deep neural networks for complex environments.

Year: 2015Generality: 694
Back to Vocab

Deep Q-Networks (DQN) are a class of reinforcement learning algorithms that use deep neural networks to approximate the Q-value function — a measure of how valuable it is to take a particular action from a given state. Classical Q-learning maintains a lookup table mapping every state-action pair to an expected cumulative reward, but this approach becomes computationally intractable in environments with high-dimensional inputs like raw pixel images. DQN sidesteps this limitation by training a convolutional neural network to generalize across similar states, effectively compressing the Q-table into a learned function that scales to complex, real-world-like problems.

Two key innovations made DQN training stable enough to work in practice. The first is experience replay, in which the agent stores past transitions in a memory buffer and samples random mini-batches during training, breaking the temporal correlations that would otherwise destabilize gradient updates. The second is a target network — a periodically frozen copy of the main network used to compute training targets, preventing the feedback loop that arises when both the predictions and the targets shift simultaneously. Together, these techniques transformed a theoretically appealing but practically fragile idea into a reliable learning algorithm.

DQN gained widespread attention after DeepMind demonstrated it learning to play dozens of Atari 2600 games directly from pixel inputs, achieving superhuman performance on several titles without any game-specific engineering. This result was striking because a single architecture and set of hyperparameters was applied uniformly across games with very different dynamics, suggesting that DQN had learned genuinely general strategies rather than narrow heuristics.

The impact of DQN extends well beyond Atari. It established a template for deep reinforcement learning research and spawned a family of improvements — including Double DQN, Dueling DQN, Prioritized Experience Replay, and Rainbow — each addressing specific weaknesses in the original formulation. DQN remains a foundational reference point in RL, taught in virtually every modern course on the subject and serving as a baseline against which newer algorithms are routinely benchmarked.

Related

Related

DRL (Deep Reinforcement Learning)
DRL (Deep Reinforcement Learning)

Neural networks combined with reinforcement learning to master complex sequential decision-making tasks.

Generality: 796
Q-Learning
Q-Learning

A model-free reinforcement learning algorithm that learns optimal action values through experience.

Generality: 792
Q-Value
Q-Value

Expected cumulative reward for taking an action in a given state under a policy.

Generality: 756
RL (Reinforcement Learning)
RL (Reinforcement Learning)

A learning paradigm where an agent maximizes cumulative rewards through environmental interaction.

Generality: 908
DNN (Deep Neural Network)
DNN (Deep Neural Network)

Neural networks with many layers that learn hierarchical representations from raw data.

Generality: 871
GQN (Generative Query Network)
GQN (Generative Query Network)

A neural architecture that infers and renders 3D scenes from limited viewpoint observations.

Generality: 292