Publication record · 18.cifr/2015.mnih.dqn-atari
18.cifr/2015.mnih.dqn-atariThe theory of reinforcement learning provides a normative account, deeply rooted in psychological and neuroscientific perspectives on animal behaviour, of how agents may optimize their control of an environment. To use reinforcement learning successfully in situations approaching real-world complexity, however, agents are confronted with a difficult task: they must derive efficient representations of the environment from high-dimensional sensory inputs, and use these to generalize past experience to new situations. Here we describe a method that achieves this by combining the reinforcement learning algorithm with a deep neural network. We present a single architecture that learns control policies directly from high-dimensional sensory input using end-to-end reinforcement learning.
Computing related research...
Loading DOI…
Sign in to run agents. GPU access requires an institutional membership.
How to get GPU access: Your university, lab, or company can become a CIFR institutional member. Members get GPU-accelerated runs for all their researchers. Contact us
No invocations yet — be the first to call this agent.
DQN struggles on sparse-reward games requiring long-horizon planning (e.g., Montezuma's Revenge). The authors suggest hierarchical RL, model-based planning, and memory-augmented architectures as remedies. Extensions to continuous action spaces and real-world robotics are also flagged.