A Deep Learning Prediction Process Accelerator Based FPGA

Recently, machine learning is widely used in applications and cloud services. And as the emerging field of machine learning, deep learning shows excellent ability in solving complex learning problems. To give users better experience, high performance implementations of deep learning applications seem very important. As a common means to accelerate algorithms, FPGA has high performance, low power consumption, small size and other characteristics. So we use FPGA to design a deep learning accelerator, the accelerator focuses on the implementation of the prediction process, data access optimization and pipeline structure. Compared with Core 2 CPU 2.3GHz, our accelerator can achieve promising result.

[1]  Berin Martini,et al.  Large-Scale FPGA-based Convolutional Networks , 2011 .

[2]  Kunle Olukotun,et al.  A Large-Scale Architecture for Restricted Boltzmann Machines , 2010, 2010 18th IEEE Annual International Symposium on Field-Programmable Custom Computing Machines.

[3]  Chao Wang,et al.  Architecture Support for Task Out-of-Order Execution in MPSoCs , 2015, IEEE Transactions on Computers.

[4]  Hong Yu,et al.  Heterogeneous Cloud Framework for Big Data Genome Sequencing , 2015, IEEE/ACM Transactions on Computational Biology and Bioinformatics.

[5]  Paul Chow,et al.  High-Performance Reconfigurable Hardware Architecture for Restricted Boltzmann Machines , 2010, IEEE Transactions on Neural Networks.

[6]  Yoshua Bengio,et al.  On the Expressive Power of Deep Architectures , 2011, ALT.

[7]  Ninghui Sun,et al.  DianNao: a small-footprint high-throughput accelerator for ubiquitous machine-learning , 2014, ASPLOS.

[8]  Geoffrey E. Hinton A Practical Guide to Training Restricted Boltzmann Machines , 2012, Neural Networks: Tricks of the Trade.

[9]  Lijuan Wang,et al.  Multi-Phase Ant Colony System for Multi-Party Data-Intensive Service Provision , 2016, IEEE Transactions on Services Computing.

[10]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[11]  Yoshua Bengio,et al.  Greedy Layer-Wise Training of Deep Networks , 2006, NIPS.