Looking for indexed pages…
| Deep Q-Network (DQN) | |
| 💡No image available | |
| Overview | |
| Core idea | Approximate the action-value function with a neural network |
| Application | Reinforcement learning in games and control tasks |
| Also known as | Deep Q-learning (with deep neural networks) |
| Key techniques | Experience replay and target networks |
A Deep Q-Network (DQN) is a reinforcement learning method that combines Q-learning with deep neural networks to learn value functions directly from high-dimensional inputs such as pixels. Proposed by researchers including Volodymyr Mnih and colleagues, DQN demonstrated strong performance on Atari 2600 games using raw screen observations.
DQN is a foundational approach in modern deep reinforcement learning and is closely associated with techniques such as experience replay and a target network. These design choices help stabilize training and reduce harmful correlations in the data.
In reinforcement learning, an agent interacts with an environment to maximize cumulative reward. The goal of Q-learning is to learn the action-value function (often written as Q(s, a)), which estimates how much future reward an agent can expect after taking action a in state s. In traditional tabular Q-learning, this function is stored explicitly for each state-action pair; in high-dimensional domains, this becomes impractical, so DQN uses a neural network to approximate Q(s, a).
DQN was made especially notable by work associated with DeepMind and public benchmarks such as Atari games from Atari 2600. The approach built on earlier value-based methods, including Q-learning and the broader paradigm of reinforcement learning, typically formalized using Markov decision processes (MDPs) such as those studied in Markov decision process.
At a high level, DQN trains a neural network to predict Q-values for each possible action given the current observation. The central learning target uses a temporal-difference (TD) update derived from Q-learning, but DQN must address instability caused by bootstrapping and moving targets.
Two mechanisms are standard in DQN:
These ideas are closely related to the stabilizing strategies used in Temporal difference learning and the use of function approximation in Neural network.
Given a minibatch sampled from the replay buffer, DQN updates network parameters by minimizing a loss that measures the difference between predicted Q-values and TD targets. If the online network estimates Q(s, a), and a target network estimates the target value for next states, the update encourages the online network’s prediction for the chosen action to match the reward plus a discounted estimate of future value.
In the original DQN formulation, exploration is commonly handled using an ε-greedy policy: with probability ε the agent selects a random action, and with probability 1−ε it chooses the action with the highest estimated Q-value. This connects to broader exploration strategies studied in reinforcement learning, including approaches such as those discussed under Epsilon-greedy. During training, ε is typically annealed to gradually shift from exploration to exploitation.
Because DQN was influential, many improvements have been proposed to address overestimation bias, inefficient exploration, and limitations of the basic architecture. Well-known extensions include Double DQN, which reduces value overestimation by decoupling action selection from action evaluation. Another line of work is Dueling networks, which separate the estimation of state value and action advantages, improving learning in states where actions share similar outcomes.
Variants also include modifications to the loss function, alternative exploration methods, and distributional value learning methods. These developments reflect a broader effort across deep reinforcement learning to improve stability and performance on benchmark tasks such as those represented by Atari 2600.
DQN is often treated as a milestone in deep reinforcement learning because it showed that a single agent could learn policies from raw sensory inputs without hand-designed features. The method’s influence is evident in subsequent research on deep RL algorithms, neural network architectures, and training protocols.
DQN’s core ideas—value-based learning with neural function approximation, plus stabilizing training via experience replay and target networks—are widely reused or adapted in later systems. In this sense, DQN serves as both a specific algorithm and a template for building reinforcement learning agents, connecting to concepts in Deep reinforcement learning and the broader study of agent-environment learning.
Categories: Deep reinforcement learning, Reinforcement learning algorithms, Neural network models, Markov decision processes
This article was generated by AI using GPT Wiki. Content may contain inaccuracies. Generated on March 27, 2026. Made by Lattice Partners.
7.9s$0.00161,634 tokens