A High Energy-Efficiency FPGA-Based LSTM Accelerator Architecture Design by Structured Pruning and Normalized Linear Quantization
暂无分享,去创建一个
Zhihong Huang | Yong Zheng | Tianli Li | Haigang Yang | Yiping Jia | Zhihong Huang | Haigang Yang | Yong Zheng | Yiping Jia | Tianli Li
[1] Norbert Wehn,et al. FINN-L: Library Extensions and Design Trade-Off Analysis for Variable Precision LSTM Networks on FPGAs , 2018, 2018 28th International Conference on Field Programmable Logic and Applications (FPL).
[2] Tobi Delbrück,et al. DeltaRNN: A Power-efficient Recurrent Neural Network Accelerator , 2018, FPGA.
[3] Chunhua Deng,et al. PermDNN: Efficient Compressed DNN Architecture with Permuted Diagonal Matrices , 2018, 2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO).
[4] Song Han,et al. Exploring the Regularity of Sparse Structure in Convolutional Neural Networks , 2017, ArXiv.
[5] Song Han,et al. ESE: Efficient Speech Recognition Engine with Sparse LSTM on FPGA , 2016, FPGA.
[6] Shuang Wu,et al. Training and Inference with Integers in Deep Neural Networks , 2018, ICLR.
[7] Qinru Qiu,et al. C-LSTM: Enabling Efficient LSTM using Structured Compression Techniques on FPGAs , 2018, FPGA.
[8] Tetsuya Asai,et al. Dither NN: An Accurate Neural Network with Dithering for Low Bit-Precision Hardware , 2018, 2018 International Conference on Field-Programmable Technology (FPT).
[9] Pavlo Molchanov,et al. Importance Estimation for Neural Network Pruning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[10] Bin Liu,et al. Ternary Weight Networks , 2016, ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[11] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.