Work-in-Progress: BPNet: Branch-pruned Conditional Neural Network for Systematic Time-accuracy Tradeoff in DNN Inference

Recently, there have been attempts to execute the neural network conditionally with auxiliary classifiers allowing early termination depending on the difficulty of the input, which can reduce the execution time or energy consumption without any or with negligible accuracy decrease. However, these studies do not consider how many or where the auxiliary classifiers, or branches, should be added in a systematic fashion. In this paper, we propose Branch-pruned Conditional Neural Network (BPNet) and its methodology in which the time-accuracy tradeoff for the conditional neural network can be found systematically. We applied BPNet to SqueezeNet, ResNet-20, and VGG-16 with CIFAR-10 and 100. BPNet achieves on average 2.0x of speedups without any accuracy drop on average compared to the base network.

[1]  Sungroh Yoon,et al.  Big/little deep neural network for ultra low power inference , 2015, 2015 International Conference on Hardware/Software Codesign and System Synthesis (CODES+ISSS).

[2]  Sherief Reda,et al.  Runtime configurable deep neural networks for energy-accuracy trade-off , 2016, 2016 International Conference on Hardware/Software Codesign and System Synthesis (CODES+ISSS).

[3]  Partha Pratim Pande,et al.  Trading-Off Accuracy and Energy of Deep Inference on Embedded Systems: A Co-Design Approach , 2018, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[4]  Saeid Nahavandi,et al.  Openga, a C++ genetic algorithm library , 2017, 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC).

[5]  Kalyanmoy Deb,et al.  An Evolutionary Many-Objective Optimization Algorithm Using Reference-Point-Based Nondominated Sorting Approach, Part I: Solving Problems With Box Constraints , 2014, IEEE Transactions on Evolutionary Computation.