학술논문

Processing-in-Memory Using Optically-Addressed Phase Change Memory
Document Type
Conference
Source
2023 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED) Low Power Electronics and Design (ISLPED), 2023 IEEE/ACM International Symposium on. :1-6 Aug, 2023
Subject
Components, Circuits, Devices and Systems
Computing and Processing
Costs
Optical design
Artificial neural networks
Programming
Throughput
Optical network units
Energy efficiency
optical computing
phase change memory
processing-in-memory
deep neural networks
Language
Abstract
Today's Deep Neural Network (DNN) inference systems contain hundreds of billions of parameters, resulting in significant latency and energy overheads during inference due to frequent data transfers between compute and memory units. Processing-in-Memory (PiM) has emerged as a viable solution to tackle this problem by avoiding the expensive data movement. PiM approaches based on electrical devices suffer from throughput and energy efficiency issues. In contrast, Optically-addressed Phase Change Memory (OPCM) operates with light and achieves much higher throughput and energy efficiency compared to its electrical counterparts. This paper introduces a system-level design that takes the OPCM programming overhead into consideration, and identifies that the programming cost dominates the DNN inference on OPCM-based PiM architectures. We explore the design space of this system and identify the most energy-efficient OPCM array size and batch size. We propose a novel thresholding and reordering technique on the weight blocks to further reduce the programming overhead. Combining these optimizations, our approach achieves up to 65.2 × higher throughput than existing photonic accelerators for practical DNN workloads.