Low-Rank Deep Convolutional Neural Network for Multitask Learning

In this paper, we propose a novel multitask learning method based on the deep convolutional network. The proposed deep network has four convolutional layers, three max-pooling layers, and two parallel fully connected layers. To adjust the deep network to multitask learning problem, we propose to learn a low-rank deep network so that the relation among different tasks can be explored. We proposed to minimize the number of independent parameter rows of one fully connected layer to explore the relations among different tasks, which is measured by the nuclear norm of the parameter of one fully connected layer, and seek a low-rank parameter matrix. Meanwhile, we also propose to regularize another fully connected layer by sparsity penalty so that the useful features learned by the lower layers can be selected. The learning problem is solved by an iterative algorithm based on gradient descent and back-propagation algorithms. The proposed algorithm is evaluated over benchmark datasets of multiple face attribute prediction, multitask natural language processing, and joint economics index predictions. The evaluation results show the advantage of the low-rank deep CNN model over multitask problems.

[1]  Xiaodong Liu,et al.  Representation Learning Using Multi-Task Deep Neural Networks for Semantic Classification and Information Retrieval , 2015, NAACL.

[2]  Michael G. Jacobson,et al.  The Impact of the Biomass Crop Assistance Program on the United States Forest Products Market: An Application of the Global Forest Products Model , 2019, Forests.

[3]  Siu-Yeung Cho,et al.  Human face recognition by adaptive processing of tree structures representation , 2007, Neural Computing and Applications.

[4]  Yoshua Bengio,et al.  Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.

[5]  Hanliang Fu,et al.  Visualized analysis of knowledge development in green building based on bibliographic data mining , 2018, The Journal of Supercomputing.

[6]  Amy Loutfi,et al.  A review of unsupervised feature learning and deep learning for time-series modeling , 2014, Pattern Recognit. Lett..

[7]  Feng Wang,et al.  Survey on swarm intelligence based routing protocols for wireless sensor networks: An extensive study , 2016, 2016 IEEE International Conference on Industrial Technology (ICIT).

[8]  Li Wang,et al.  Learning image convolutional representations and complete tags jointly , 2017, Neural Computing and Applications.

[9]  Xiaofei He,et al.  Multi-Target Regression via Robust Low-Rank Learning , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Bin Jiang,et al.  Multiple-relations-constrained image classification with limited training samples via Pareto optimization , 2018, Neural Computing and Applications.

[11]  Lawrence Carin,et al.  Multi-Task Learning for Classification with Dirichlet Process Priors , 2007, J. Mach. Learn. Res..

[12]  Guohui Zhang,et al.  Cross-domain attribute representation based on convolutional neural network , 2018, ICIC.

[13]  Yang Liu,et al.  Local Patches Alignment Embedding Based Localization for Wireless Sensor Networks , 2013, Wirel. Pers. Commun..

[14]  Taoufiq Gadi,et al.  A Novel Deep Learning Approach for Recognizing Stereotypical Motor Movements within and across Subjects on the Autism Spectrum Disorder , 2018, Comput. Intell. Neurosci..

[15]  Massimiliano Pontil,et al.  Regularized multi--task learning , 2004, KDD.

[16]  Wei Liu,et al.  Efficient Multi-Class Selective Sampling on Graphs , 2016, UAI.

[17]  Peng Yang,et al.  A Min-Max Optimization Framework For Online Graph Classification , 2015, CIKM.

[18]  Xiaoou Tang,et al.  Facial Landmark Detection by Deep Multi-task Learning , 2014, ECCV.

[19]  Gavin C. Cawley,et al.  Leave-One-Out Cross-Validation Based Model Selection Criteria for Weighted LS-SVMs , 2006, The 2006 IEEE International Joint Conference on Neural Network Proceedings.

[20]  Nikolaos Doulamis,et al.  FAST-MDL: Fast Adaptive Supervised Training of multi-layered deep learning models for consistent object tracking and classification , 2016, 2016 IEEE International Conference on Imaging Systems and Techniques (IST).

[21]  Yunyi Jia,et al.  Cascade heterogeneous face sketch-photo synthesis via dual-scale Markov Network , 2018, J. Exp. Theor. Artif. Intell..

[22]  Yang Zhong,et al.  Face attribute prediction using off-the-shelf CNN features , 2016, 2016 International Conference on Biometrics (ICB).

[23]  Jasha Droppo,et al.  Multi-task learning in deep neural networks for improved phoneme recognition , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[24]  Jason Weston,et al.  A unified architecture for natural language processing: deep neural networks with multitask learning , 2008, ICML '08.

[25]  Jürgen Schmidhuber,et al.  Deep learning in neural networks: An overview , 2014, Neural Networks.

[26]  Nikolaos Doulamis,et al.  Deep Learning for Computer Vision: A Brief Review , 2018, Comput. Intell. Neurosci..

[27]  Liang Dong,et al.  Recognition of Emotions Using Multichannel EEG Data and DBN-GC-Based Ensemble Deep Learning Framework , 2018, Comput. Intell. Neurosci..

[28]  Abdul Jabbar,et al.  An improved Urdu stemming algorithm for text mining based on multi-step hybrid approach , 2018, J. Exp. Theor. Artif. Intell..

[29]  Guohui Zhang,et al.  Learning Convolutional Ranking-Score Function by Query Preference Regularization , 2017, IDEAL.

[30]  Siu-Yeung Cho,et al.  A face emotion tree structure representation with probabilistic recursive neural network modeling , 2010, Neural Computing and Applications.

[31]  R. Zhang,et al.  Vaccine Adjuvant Incorporation Strategy Dictates Peptide Amphiphile Micelle Immunostimulatory Capacity , 2018, The AAPS Journal.

[32]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[33]  Robert E. Mercer,et al.  Supervised classification of spam emails with natural language stylometry , 2015, Neural Computing and Applications.

[34]  James L. Smith,et al.  Informing the Use of Vibrotactile Feedback for Information Communication , 2014 .

[35]  R. Zhang,et al.  Peptide Amphiphile Micelle Vaccine Size and Charge Influence the Host Antibody Response. , 2018, ACS biomaterials science & engineering.

[36]  Sebastian Ruder,et al.  An Overview of Multi-Task Learning in Deep Neural Networks , 2017, ArXiv.

[37]  Massimiliano Pontil,et al.  Multi-Task Feature Learning , 2006, NIPS.

[38]  Jean-Philippe Vert,et al.  Clustered Multi-Task Learning: A Convex Formulation , 2008, NIPS.

[39]  Tania Stathaki,et al.  Recent Developments in Deep Learning for Engineering Applications , 2018, Comput. Intell. Neurosci..

[40]  B. Ulery,et al.  Synthetic Vaccine Characterization and Design , 2018 .

[41]  Guohui Zhang,et al.  A Novel Image Tag Completion Method Based on Convolutional Neural Transformation , 2017, ICANN.

[42]  Juhan Nam,et al.  Multimodal Deep Learning , 2011, ICML.

[43]  Wentao Mao,et al.  Leave-one-out cross-validation-based model selection for multi-input multi-output support vector machine , 2012, Neural Computing and Applications.

[44]  Hanliang Fu,et al.  Study on Threshold Selection Methods in Calculation of Ocean Environmental Design Parameters , 2019, IEEE Access.

[45]  Laith H. Baniata,et al.  A Neural Machine Translation Model for Arabic Dialects That Utilises Multitask Learning (MTL) , 2018, Comput. Intell. Neurosci..

[46]  R. Zhang,et al.  Instructive Design of Triblock Peptide Amphiphiles for Structurally Complex Micelle Fabrication. , 2018, ACS biomaterials science & engineering.

[47]  Gang Hu,et al.  Deep Learning Methods for Underwater Target Feature Extraction and Recognition , 2018, Comput. Intell. Neurosci..

[48]  Shun Ishizaki,et al.  Analysis system for Sinhalese unit structure , 1992, J. Exp. Theor. Artif. Intell..

[49]  Yang Liu,et al.  An Integrated Approach to Sink and Sensor Role Selection in Wireless Sensor Networks: Using Dynamic Programming , 2014, Ad Hoc Sens. Wirel. Networks.

[50]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[51]  R. Zhang,et al.  Immunomodulatory vasoactive intestinal peptide amphiphile micelles. , 2018, Biomaterials science.

[52]  Caroline Lyon,et al.  Using single layer networks for discrete, sequential data: An example from Natural Language Processing , 2005, Neural Computing & Applications.

[53]  Michael S. Lew,et al.  Deep learning for visual understanding: A review , 2016, Neurocomputing.

[54]  Xiaogang Xiong,et al.  Multilayer Hybrid Deep-Learning Method for Waste Classification and Recycling , 2018, Comput. Intell. Neurosci..

[55]  Yulin Wang,et al.  Designing wearable vibrotactile notifications for information communication , 2016, Int. J. Hum. Comput. Stud..

[56]  Rich Caruana,et al.  Multitask Learning , 1997, Machine-mediated learning.

[57]  Yongzhao Zhan,et al.  An SVM–AdaBoost-based face detection system , 2014, J. Exp. Theor. Artif. Intell..