학술논문

Distributional Cloning for Stabilized Imitation Learning via ADMM
Document Type
Conference
Source
2023 IEEE International Conference on Data Mining (ICDM) ICDM Data Mining (ICDM), 2023 IEEE International Conference on. :818-827 Dec, 2023
Subject
Communication, Networking and Broadcast Technologies
Computing and Processing
Training
Supervised learning
Cloning
Ordinary differential equations
Behavioral sciences
Data mining
Task analysis
imitation learning
neural ordinary differential equations
Language
ISSN
2374-8486
Abstract
The two leading solution paradigms for imitation learning (IL), BC and GAIL, each suffers from notable drawbacks. BC, a supervised learning approach to mimic expert actions, is vulnerable to covariate shift. GAIL applies adversarial training to minimize the discrepancy between expert and learner behaviors, which is prone to unstable training and mode collapse. In this work, we propose DC – Distributional Cloning – a novel IL approach for addressing the covariate shift and mode collapse problems simultaneously. DC directly maximizes the likelihood of observed expert and learner demonstrations, and gradually encourages the learner to evolve towards expert behaviors based on an averaging effect. The DC solution framework contains two stages in each training loop, where in stage one the mixed expert and learner state distribution is estimated via SoftFlow, and in stage two the learner policy is trained to match both the expert’s policy and state distribution via ADMM. Experimental evaluation of DC compared with several baselines in 10 different physics-based control tasks reveal superior results in learner policy performance, training stability, and mode distribution preservation.