Exploration by Random Network Distillation, Yuri Burda, Harrison Edwards, Amos Storkey, Oleg Klimov, 2018International Conference on Learning Representations (ICLR) (arXiv)DOI: 10.48550/arXiv.1810.12894 - This paper introduces Random Network Distillation (RND), a widely used method for novelty-based exploration in deep reinforcement learning, particularly effective in sparse-reward environments.
Curiosity-driven Exploration by Self-supervised Prediction, Deepak Pathak, Pulkit Agrawal, Alexei A. Efros, Trevor Darrell, 2017International Conference on Machine Learning (ICML)DOI: 10.48550/arXiv.1705.05363 - This work introduces the Intrinsic Curiosity Module (ICM), a model for intrinsic motivation that generates curiosity-driven rewards based on prediction error of an agent's own forward dynamics model.
Deep Exploration via Bootstrapped DQN, Ian Osband, Charles Blundell, Alexander Pritzel, Benjamin Van Roy, 2016Advances in Neural Information Processing Systems (NeurIPS)DOI: 10.48550/arXiv.1602.04621 - This paper demonstrates effective uncertainty-based exploration in deep reinforcement learning by using an ensemble of deep Q-networks (bootstrapped DQN) to estimate uncertainty.