학술논문

Exploration With Intrinsic Motivation Using Object–Action–Outcome Latent Space
Document Type
Periodical
Source
IEEE Transactions on Cognitive and Developmental Systems IEEE Trans. Cogn. Dev. Syst. Cognitive and Developmental Systems, IEEE Transactions on. 15(2):325-336 Jun, 2023
Subject
Computing and Processing
Signal Processing and Analysis
Robots
Robot sensing systems
Computational modeling
Space exploration
Predictive models
Task analysis
Robot learning
Developmental robotics
effect prediction
intrinsic motivation (IM)
open-ended learning
representation learning
Language
ISSN
2379-8920
2379-8939
Abstract
One effective approach for equipping artificial agents with sensorimotor skills is to use self-exploration. To do this efficiently is critical, as time and data collection are costly. In this study, we propose an exploration mechanism that blends action, object, and action outcome representations into a latent space, where local regions are formed to host forward model (FM) learning. The agent uses intrinsic motivation to select the FM with the highest learning progress (LP) to adopt at a given exploration step. This parallels how infants learn, as high LP indicates that the learning problem is neither too easy nor too difficult in the selected region. The proposed approach is validated with a simulated robot in a tabletop environment. The simulation scene comprises a robot and various objects, where the robot interacts with one of them each time using a set of parameterized actions and learns the outcomes of these interactions. With the proposed approach, the robot organizes its curriculum of learning as in existing intrinsic motivation approaches and outperforms them in learning speed. Moreover, the learning regime demonstrates features that partially match infant development; in particular, the proposed system learns to predict the outcomes of different skills in a staged manner.