학술논문

FlexGibbs: Reconfigurable Parallel Gibbs Sampling Accelerator for Structured Graphs
Document Type
Conference
Source
2019 IEEE 27th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM) FCCM Field-Programmable Custom Computing Machines (FCCM), 2019 IEEE 27th Annual International Symposium on. :334-334 Apr, 2019
Subject
Components, Circuits, Devices and Systems
Computing and Processing
Signal Processing and Analysis
Bayes methods
Markov processes
Source separation
Hardware
Acceleration
Deep learning
Bayesian inference
Gibbs sampling
Markov random field
hardware accelerator
probabilistic graphical model
Language
ISSN
2576-2621
Abstract
Many consider one of the key components to the success of deep learning as its compatibility with existing accelerators, mainly GPU. While GPUs are great at handling linear algebra kernels commonly found in deep learning, they are not the optimal architecture for handling unsupervised learning methods such as Bayesian models and inference. As a step towards, achieving better understanding of architectures for probabilistic models, Gibbs sampling, one of the most commonly used algorithms for Bayesian inference, is studied with a focus on parallelism that converges to the target distribution and parameterized components. We propose FlexGibbs, a reconfigurable parallel Gibbs sampling inference accelerator for structured graphs. We designed an architecture optimal for solving Markov Random Field tasks using an array of parallel Gibbs samplers, enabled by chromatic scheduling. We show that for sound source separation application, FlexGibbs configured on the FPGA fabric of Xilinx Zync CPU-FPGA SoC achieved Gibbs sampling inference speedup of 1048x and 99.85% reduction in energy over running it on ARM Cortex-A53.