FPGA-based generic neural network architecture

In this paper, we defined a generic architecture for the extraction phase of a multi layer neural network algorithm to be implemented on a Virtex-4 FPGA. This architecture can be applied to any multi layer neural network composed of a given number of layers and a given number of neurons in each layer. In addition this architecture enhances the density of the FPGA by supporting the two concepts of time multiplexing and partial dynamic reconfiguration. Several networks with different sizes were implemented based on this generic architecture. Based on those implementations, we'll analyse the performances of a virtex-4 via a multi layer neural network by analyzing the variation of the minimum period and the number of occupied resources. This work was made in collaboration with the NodBox company (thierry.fargas@nodbox.biz) and Xilinx company (jean-louis.brelet@xilinx.com).

[1]  Michel Paindavoine,et al.  Implementation of JPEG2000 arithmetic decoder using dynamic reconfiguration of FPGA , 2004, 2004 International Conference on Image Processing, 2004. ICIP '04..

[2]  Brad Hutchings,et al.  Density enhancement of a neural network using FPGAs and run-time reconfiguration , 1994, Proceedings of IEEE Workshop on FPGA's for Custom Computing Machines.

[3]  Jürgen Teich,et al.  Platform-independent methodology for partial reconfiguration , 2004, CF '04.

[4]  Brad Hutchings,et al.  FPGA-based stochastic neural networks-implementation , 1994, Proceedings of IEEE Workshop on FPGA's for Custom Computing Machines.

[5]  Philip James-Roxby,et al.  Adapting constant multipliers in a neural network implementation , 2000, Proceedings 2000 IEEE Symposium on Field-Programmable Custom Computing Machines (Cat. No.PR00871).