학술논문

Learning a Generalizable Trajectory Sampling Distribution for Model Predictive Control
Document Type
Periodical
Source
IEEE Transactions on Robotics IEEE Trans. Robot. Robotics, IEEE Transactions on. 40:2111-2127 2024
Subject
Robotics and Control Systems
Computing and Processing
Components, Circuits, Devices and Systems
Trajectory
Robots
Planning
Costs
Training
Quadrotors
Training data
Deep generative models
deep learning in robotics and automation
motion and path planning
nonholonomic motion planning
Language
ISSN
1552-3098
1941-0468
Abstract
We propose a sample-based model predictive control (MPC) method for collision-free navigation that uses a normalizing flow as a sampling distribution, conditioned on the start, goal, environment, and cost parameters. This representation allows us to learn a distribution that accounts for both the dynamics of the robot and complex obstacle geometries. We propose a way to incorporate this sampling distribution into two sampling-based MPC methods, MPPI, and iCEM. However, when deploying these methods, the robot may encounter an out-of-distribution (OOD) environment. To generalize our method to OOD environments, we also present an approach that performs projection on the representation of the environment. This projection changes the environment representation to be more in-distribution while also optimizing trajectory quality in the true environment. Our simulation results on a 2-D double-integrator, a 12-DoF quadrotor and a seven-DoF kinematic manipulator suggest that using a learned sampling distribution with projection outperforms MPC baselines on both in-distribution and OOD environments over different cost functions, including OOD environments generated from real-world data.