Learning Multi-Domain Adversarial Neural Networks for Text Classification

Deep neural networks have been applied to learn transferable features for adapting text classification models from a source domain to a target domain. Conventional domain adaptation used to adapt models from an individual specific domain with sufficient labeled data to another individual specific target domain without any (or with little) labeled data. However, in this paradigm, we lose sight of correlation among different domains where common knowledge could be shared to improve the performance of both the source domain and the target domain. Multi-domain learning proposes learning the sharable features from multiple source domains and the target domain. However, previous work mainly focuses on improving the performance of the target domain and lacks the effective mechanism to ensure that the shared feature space is not contaminated by domain-specific features. In this paper, we use an adversarial training strategy and orthogonality constraints to guarantee that the private and shared features do not collide with each other, which can improve the performances of both the source domains and the target domain. The experimental results, on a standard sentiment domain adaptation dataset and a consumption intention identification dataset labeled by us, show that our approach dramatically outperforms state-of-the-art baselines, and it is general enough to be applied to more scenarios.

[1]  Luc Van Gool,et al.  Disentangled Person Image Generation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[2]  Jacob Cohen,et al.  Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit. , 1968 .

[3]  Koby Crammer,et al.  Multi-domain learning by confidence-weighted parameter combination , 2010, Machine Learning.

[4]  Yuexin Wu,et al.  We know what you want to buy: a demographic-based system for product recommendation on microblogs , 2014, KDD.

[5]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[6]  Koby Crammer,et al.  Online Methods for Multi-Domain Learning and Adaptation , 2008, EMNLP.

[7]  Michael I. Jordan,et al.  Learning Transferable Features with Deep Adaptation Networks , 2015, ICML.

[8]  Yann LeCun,et al.  Disentangling factors of variation in deep representation using adversarial training , 2016, NIPS.

[9]  Yoshua Bengio,et al.  How transferable are features in deep neural networks? , 2014, NIPS.

[10]  Scott E. Reed,et al.  Weakly-supervised Disentangling with Recurrent Transformations for 3D View Synthesis , 2015, NIPS.

[11]  John Blitzer,et al.  Biographies, Bollywood, Boom-boxes and Blenders: Domain Adaptation for Sentiment Classification , 2007, ACL.

[12]  Feng Liu,et al.  Disentangling Features in 3D Face Shapes for Joint Face Reconstruction and Recognition , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[13]  Xiaoming Liu,et al.  Disentangled Representation Learning GAN for Pose-Invariant Face Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Wojciech Zaremba,et al.  An Empirical Exploration of Recurrent Network Architectures , 2015, ICML.

[15]  Xuanjing Huang,et al.  Adversarial Multi-task Learning for Text Classification , 2017, ACL.

[16]  George Trigeorgis,et al.  Domain Separation Networks , 2016, NIPS.

[17]  James H. Martin,et al.  Abstract Meaning Representation Parsing using LSTM Recurrent Neural Networks , 2017, ACL.

[18]  Xiaogang Wang,et al.  Learning Deep Feature Representations with Domain Guided Dropout for Person Re-identification , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Yue Cao,et al.  Transferable Representation Learning with Deep Adaptation Networks , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[20]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[21]  Jun Yan,et al.  Active Sentiment Domain Adaptation , 2017, ACL.

[22]  Ting Liu,et al.  Domain Adaptation via Tree Kernel Based Maximum Mean Discrepancy for User Consumption Intention Identification , 2018, IJCAI.

[23]  Victor S. Lempitsky,et al.  Unsupervised Domain Adaptation by Backpropagation , 2014, ICML.

[24]  Lejian Liao,et al.  Can Syntax Help? Improving an LSTM-based Sentence Compression Model for New Domains , 2017, ACL.

[25]  Qiang Yang,et al.  Cross-domain sentiment classification via spectral feature alignment , 2010, WWW '10.

[26]  Yongxin Yang,et al.  A Unified Perspective on Multi-Domain and Multi-Task Learning , 2014, ICLR.

[27]  Qiang Yang,et al.  Transfer Learning via Dimensionality Reduction , 2008, AAAI.

[28]  Fangzhao Wu,et al.  Sentiment Domain Adaptation with Multiple Sources , 2016, ACL.

[29]  Carolyn Penstein Rosé,et al.  Multi-Domain Learning: When Do Domains Matter? , 2012, EMNLP-CoNLL.

[30]  Philip S. Yu,et al.  Deep Learning of Transferable Representation for Scalable Domain Adaptation , 2016, IEEE Transactions on Knowledge and Data Engineering.

[31]  Yu Zhang,et al.  End-to-End Adversarial Memory Network for Cross-domain Sentiment Classification , 2017, IJCAI.

[32]  Yoshua Bengio,et al.  Domain Adaptation for Large-Scale Sentiment Classification: A Deep Learning Approach , 2011, ICML.

[33]  Yu Zhang,et al.  Hierarchical Attention Transfer Network for Cross-Domain Sentiment Classification , 2018, AAAI.

[34]  Jian-Yun Nie,et al.  Mining User Consumption Intention from Social Media Using Domain Adaptive Convolutional Neural Network , 2015, AAAI.

[35]  Byron C. Wallace,et al.  Learning Disentangled Representations of Texts with Application to Biomedical Abstracts , 2018, EMNLP.

[36]  Bohyung Han,et al.  Learning Multi-domain Convolutional Neural Networks for Visual Tracking , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[37]  Ming Zhou,et al.  Multi-Domain Adaptation for SMT Using Multi-Task Learning , 2013, EMNLP.