• Part of
    Ubiquity Network logo
    Publish with us Cyhoeddi gyda ni

    Read Chapter
  • No readable formats available
  • Accelerating Multi-step Sparse Reward Reinforcement Learning

    X. Yang, Z. Ji

    Chapter from the book: Spezi E. & Bray M. 2024. Proceedings of the Cardiff University Engineering Research Conference 2023.

     Download

    After the great successes of deep reinforcement learning (DRL) in recent years, developing methods to speed up DRL algorithms for more complex tasks closer to those in the real world has become increasingly important. In particular, there is a lack of research on long-horizon tasks that contain multiple subtasks or intermediate steps and can only provide sparse rewards at task completion point. This paper suggests to 1) use human priors to decompose a task and provide abstract demonstrations – the correct sequences of steps to guide exploration and learning, and 2) adjust the exploration parameters adaptively according to the online performances of the policy. The proposed ideas are implemented on three popular DRL algorithms, and experimental results on gridworld and manipulation tasks prove the concept and effectiveness of the proposed techniques.

    Chapter Metrics:

    How to cite this chapter
    Yang X. & Ji Z. 2024. Accelerating Multi-step Sparse Reward Reinforcement Learning. In: Spezi E. & Bray M (eds.), Proceedings of the Cardiff University Engineering Research Conference 2023. Cardiff: Cardiff University Press. DOI: https://doi.org/10.18573/conf1.u
    License

    This chapter distributed under the terms of the Creative Commons Attribution + Noncommercial + NoDerivatives 4.0 license. Copyright is retained by the author(s)

    Peer Review Information

    This book has been peer reviewed. See our Peer Review Policies for more information.

    Additional Information

    Published on May 1, 2024

    DOI
    https://doi.org/10.18573/conf1.u