GAN-based Intrinsic Exploration for Sample Efficient Reinforcement Learning

Research output: Contribution to journalConference articlepeer-review

Abstract

In this study, we address the problem of efficient exploration in reinforcement learning. Most common exploration approaches depend on random action selection, however these approaches do not work well in environments with sparse or no rewards. We propose Generative Adversarial Network-based Intrinsic Reward Module that learns the distribution of the observed states and sends an intrinsic reward that is computed as high for states that are out of distribution, in order to lead agent to unexplored states. We evaluate our approach in Super Mario Bros for a no reward setting and in Montezuma’s Revenge for a sparse reward setting and show that our approach is indeed capable of exploring efficiently. We discuss a few weaknesses and conclude by discussing future works.

Original languageEnglish
Pages (from-to)264-272
Number of pages9
JournalInternational Conference on Agents and Artificial Intelligence
Volume2
DOIs
Publication statusPublished - 2022
Event14th International Conference on Agents and Artificial Intelligence , ICAART 2022 - Virtual, Online
Duration: 3 Feb 20225 Feb 2022

Bibliographical note

Publisher Copyright:
© 2021 by SCITEPRESS – Science and Technology Publications, Lda. All rights reserved.

Funding

This work is supported by Istanbul Technical University BAP Grant NO: MOA-2019-42321.

FundersFunder number
Istanbul Teknik ÜniversitesiMOA-2019-42321

    Keywords

    • Deep Learning
    • Efficient Exploration in Reinforcement Learning
    • Generative Adversarial Networks
    • Reinforcement Learning

    Fingerprint

    Dive into the research topics of 'GAN-based Intrinsic Exploration for Sample Efficient Reinforcement Learning'. Together they form a unique fingerprint.

    Cite this