학술논문

Fingerprint Networked Reinforcement Learning via Multiagent Modeling for Improving Decision Making in an Urban Food–Energy–Water Nexus
Document Type
Periodical
Source
IEEE Transactions on Systems, Man, and Cybernetics: Systems IEEE Trans. Syst. Man Cybern, Syst. Systems, Man, and Cybernetics: Systems, IEEE Transactions on. 53(7):4324-4338 Jul, 2023
Subject
Signal Processing and Analysis
Robotics and Control Systems
Power, Energy and Industry Applications
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Computing and Processing
General Topics for Engineers
Decision making
Games
Training
Sustainable development
Analytical models
Optimization
Numerical models
fingerprint networked reinforcement learning (FNRL)
food–energy–water (FEW) nexus
long short-term memory (LSTM) network
multiagent system (MAS)
Language
ISSN
2168-2216
2168-2232
Abstract
Food–energy–water (FEW) nexus analyses are critical to sustainable development. Nexus analyses form a unique multiagent decision-making arena that requires using a system engineering approach to simultaneously improve food and water security as well as energy efficiency. To tackle the complexity in decision making within a FEW nexus with respect to dynamic behaviors and interactive logics, we model the FEW nexus as a multiagent system (MAS) under a mixed competitive and cooperative environment from the perspective of a Markov game. Then, we propose a fingerprint networked reinforcement learning (FNRL) framework for the collective learning of a MAS by following the logic flows of human decision making. FNRL can alleviate the problems caused by stationary issues in a MAS environment by integrating a long short-term-memory-driven neural network model into the context of multiagent reinforcement learning (RL) to extract fingerprint information via historical data. Numerical simulations for an urban FEW nexus analysis in Florida (USA) demonstrate that applying the FNRL framework can drive agents in the MAS toward achieving optimality via RL in a dynamic environment.