학술논문

Toward Transparent Load Disaggregation—A Framework for Quantitative Evaluation of Explainability Using Explainable AI
Document Type
Periodical
Source
IEEE Transactions on Consumer Electronics IEEE Trans. Consumer Electron. Consumer Electronics, IEEE Transactions on. 70(1):4345-4356 Feb, 2024
Subject
Power, Energy and Industry Applications
Components, Circuits, Devices and Systems
Fields, Waves and Electromagnetics
Artificial intelligence
Visualization
Aggregates
Computational modeling
Complexity theory
Robustness
Prediction algorithms
Deep neural networks
explainable AI (XAI)
non-intrusive load monitoring
load disaggregation
Language
ISSN
0098-3063
1558-4127
Abstract
Load Disaggregation, or Non-intrusive Load Monitoring (NILM), refers to the process of estimating energy consumption of individual domestic appliances from aggregated household consumption. Recently, Deep Learning (DL) approaches have seen increased adoption in NILM community. However, DL NILM models are often treated as black-box algorithms, which introduces algorithmic transparency and explainability concerns, hindering wider adoption. Recent works have investigated explainability of DL NILM, however they are limited to computationally expensive methods or simple classification problems. In this work, we present a methodology for explainability of regression-based DL NILM with visual explanations, using explainable AI (XAI). Two explainability levels are provided. Sequence-level explanations highlight important features of predicted time-series sequence of interest, while point-level explanations enable visualising explanations at a point in time. To facilitate wider adoption of XAI, we define desirable properties of NILM explanations - faithfulness, robustness and effective complexity. Addressing the limitation of existing XAI NILM approaches that don’t assess the quality of explanations, desirable properties of explanations are used for quantitative evaluation of explainability. We show that proposed framework enables better understanding of NILM outputs and helps improve design by providing a visualization strategy and rigorous evaluation of quality of XAI methods, leading to transparency of outcomes.