Accelerating Multi-step Sparse Reward Reinforcement Learning
X. Yang, Z. Ji
Chapter from the book: Spezi E. & Bray M. 2024. Proceedings of the Cardiff University Engineering Research Conference 2023.
Chapter from the book: Spezi E. & Bray M. 2024. Proceedings of the Cardiff University Engineering Research Conference 2023.
After the great successes of deep reinforcement learning (DRL) in recent years, developing methods to speed up DRL algorithms for more complex tasks closer to those in the real world has become increasingly important. In particular, there is a lack of research on long-horizon tasks that contain multiple subtasks or intermediate steps and can only provide sparse rewards at task completion point. This paper suggests to 1) use human priors to decompose a task and provide abstract demonstrations – the correct sequences of steps to guide exploration and learning, and 2) adjust the exploration parameters adaptively according to the online performances of the policy. The proposed ideas are implemented on three popular DRL algorithms, and experimental results on gridworld and manipulation tasks prove the concept and effectiveness of the proposed techniques.
Yang X. & Ji Z. 2024. Accelerating Multi-step Sparse Reward Reinforcement Learning. In: Spezi E. & Bray M (eds.), Proceedings of the Cardiff University Engineering Research Conference 2023. Cardiff: Cardiff University Press. DOI: https://doi.org/10.18573/conf1.u
This chapter distributed under the terms of the Creative Commons Attribution + Noncommercial + NoDerivatives 4.0 license. Copyright is retained by the author(s)
This book has been peer reviewed. See our Peer Review Policies for more information.
Published on May 1, 2024