학술논문

Deep Reinforcement Learning Under Signal Temporal Logic Constraints Using Lagrangian Relaxation
Document Type
Periodical
Source
IEEE Access Access, IEEE. 10:114814-114828 2022
Subject
Aerospace
Bioengineering
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Computing and Processing
Engineered Materials, Dielectrics and Plasmas
Engineering Profession
Fields, Waves and Electromagnetics
General Topics for Engineers
Geoscience
Nuclear Engineering
Photonics and Electrooptics
Power, Energy and Industry Applications
Robotics and Control Systems
Signal Processing and Analysis
Transportation
Task analysis
Performance analysis
Optimal control
Trajectory
Q-learning
Design methodology
Reinforcement learning
Deep learning
Constrained reinforcement learning
deep reinforcement learning
Lagrangian relaxation
signal temporal logic
Language
ISSN
2169-3536
Abstract
Deep reinforcement learning (DRL) has attracted much attention as an approach to solve optimal control problems without mathematical models of systems. On the other hand, in general, constraints may be imposed on optimal control problems. In this study, we consider the optimal control problems with constraints to complete temporal control tasks. We describe the constraints using signal temporal logic (STL), which is useful for time sensitive control tasks since it can specify continuous signals within bounded time intervals. To deal with the STL constraints, we introduce an extended constrained Markov decision process (CMDP), which is called a $\tau $ -CMDP. We formulate the STL-constrained optimal control problem as the $\tau $ -CMDP and propose a two-phase constrained DRL algorithm using the Lagrangian relaxation method. Through simulations, we also demonstrate the learning performance of the proposed algorithm.