학술논문

Non-Markovian environment and experience replay / 非マルコフ環境と経験再生
Document Type
Journal Article
Source
Proceedings of the Annual Conference of JSAI. 2023, :2
Subject
Experience Replay
Reinforcement Learning
強化学習
経験再生
Language
Japanese
ISSN
2758-7347
Abstract
This paper explores solutions to the challenges encountered in applying reinforcement learning (RL) algorithms to non-Markovian environments, leveraging the hippocampal capacity for experience replay. Many trial-and-error iterations are necessary for such environments to train a discriminator capable of distinguishing states using contextual information. In contrast to artificial agents, animals can quickly reproduce successful behaviors even in non-Markovian tasks characterized by complex reward and state transition conditions. Recent research highlights the role of the rodent hippocampus in path planning and solving such tasks through repeated experience replays before initiating movement. We propose a novel RL model that effectively solves non-Markovian tasks by replaying previously successful action patterns before action selection and applying replay-based temporal biases to action values. This model ruminates past successful behaviors and significantly reduces the number of trial iterations. Our model presents a promising approach to tackling the challenges of RL in non-Markovian environments, offering opportunities to further interconnect neuroscience and AI research.

Online Access