Attention-based Curiosity-driven Exploration in Deep Reinforcement Learning

October 23, 2019 Β· Entered Twilight Β· πŸ› IEEE International Conference on Acoustics, Speech, and Signal Processing

πŸŒ… TWILIGHT: Old Age
Predates the code-sharing era β€” a pioneer of its time

"Last commit was 6.0 years ago (β‰₯5 year threshold)"

Evidence collected by the PWNC Scanner

Repo contents: .gitignore, LICENSE, README.md, figures, jupyter, log, log_experimental, src

Authors Patrik Reizinger, MÑrton Szemenyei arXiv ID 1910.10840 Category cs.LG: Machine Learning Cross-listed cs.AI, stat.ML Citations 22 Venue IEEE International Conference on Acoustics, Speech, and Signal Processing Repository https://github.com/rpatrik96/AttA2C/ ⭐ 29 Last Checked 2 months ago
Abstract
Reinforcement Learning enables to train an agent via interaction with the environment. However, in the majority of real-world scenarios, the extrinsic feedback is sparse or not sufficient, thus intrinsic reward formulations are needed to successfully train the agent. This work investigates and extends the paradigm of curiosity-driven exploration. First, a probabilistic approach is taken to exploit the advantages of the attention mechanism, which is successfully applied in other domains of Deep Learning. Combining them, we propose new methods, such as AttA2C, an extension of the Actor-Critic framework. Second, another curiosity-based approach - ICM - is extended. The proposed model utilizes attention to emphasize features for the dynamic models within ICM, moreover, we also modify the loss function, resulting in a new curiosity formulation, which we call rational curiosity. The corresponding implementation can be found at https://github.com/rpatrik96/AttA2C/.
Community shame:
Not yet rated
Community Contributions

Found the code? Know the venue? Think something is wrong? Let us know!

πŸ“œ Similar Papers

In the same crypt β€” Machine Learning