Pruning the Convolution Neural Network (SqueezeNet) based on L2 Normalization of Activation Maps

In recent years, deep learning models have become popular in real-time embedded application, but there are many complexities for the hardware deployment because of limited resources such as memory, computational power, and energy. Network pruning is one of the promising technique to solve these problems. This paper presents an approach to prune the Convolution Neural Network (SqueezeNet) based on L2 nor-malization of activation maps on the CIFAR-10 dataset without introducing the network sparsity in the pruned model. Results show that proposed approach reduces the SqueezeNet model by 67% without a significant drop in the accuracy of the model (optimal pruning efficiency result)