FlexGibbs: Reconfigurable Parallel Gibbs Sampling Accelerator for Structured Graphs

Many consider one of the key components to the success of deep learning as its compatibility with existing accelerators, mainly GPU. While GPUs are great at handling linear algebra kernels commonly found in deep learning, they are not the optimal architecture for handling unsupervised learning methods such as Bayesian models and inference. As a step towards, achieving better understanding of architectures for probabilistic models, Gibbs sampling, one of the most commonly used algorithms for Bayesian inference, is studied with a focus on parallelism that converges to the target distribution and parameterized components. We propose FlexGibbs, a reconfigurable parallel Gibbs sampling inference accelerator for structured graphs. We designed an architecture optimal for solving Markov Random Field tasks using an array of parallel Gibbs samplers, enabled by chromatic scheduling. We show that for sound source separation application, FlexGibbs configured on the FPGA fabric of Xilinx Zync CPU-FPGA SoC achieved Gibbs sampling inference speedup of 1048x and 99.85% reduction in energy over running it on ARM Cortex-A53.

[1]  Rob A. Rutenbar,et al.  Real-Time and Low-Power Streaming Source Separation Using Markov Random Field , 2018, ACM J. Emerg. Technol. Comput. Syst..

[2]  Arthur Gretton,et al.  Parallel Gibbs Sampling: From Colored Fields to Thin Junction Trees , 2011, AISTATS.

[3]  Rob A. Rutenbar,et al.  Stereophonic spectrogram segmentation using Markov random fields , 2012, 2012 IEEE International Workshop on Machine Learning for Signal Processing.