Transferable attention networks for adversarial domain adaptation

Abstract Domain adaptation is one of the fundamental challenges in transfer learning. How to effectively transfer knowledge from labeled source domain to unlabeled target domain is critical for domain adaptation, as it benefits to reduce the considerable performance gap due to domain shift. Existing methods of domain adaptation address this issue via matching the global features across domains. However, not all features are transferable for domain adaptation, while forcefully matching the untransferable features may lead to negative transfer. In this paper, we propose a novel method dubbed transferable attention networks (TAN) to address this issue. The proposed TAN focuses on the feature alignment by utilizing adversarial optimization. Specifically, we utilize the self-attention mechanism to weight extracted features, such that the influence of untransferable features can be effectively eliminated. Meanwhile, to exploit the complex multi-modal structures of domain adaptation, we use learned features and classifier predictions as the condition to train the adversarial networks. Furthermore, we further propose that the accurately transferable features should enable domain discrepancy to minimum. Three loss functions are introduced into the adversarial networks: classification loss, attention transfer loss, and condition transfer loss. Extensive experiments on Office-31, ImageCLEF-DA, Office-Home, and VisDA-2017 datasets testify that the proposed approach yields state-of-the-art results.

[1]  Yi Yang,et al.  Deep Adversarial Attention Alignment for Unsupervised Domain Adaptation: the Benefit of Target Expectation Maximization , 2018, ECCV.

[2]  Ivor W. Tsang,et al.  Domain Adaptation via Transfer Component Analysis , 2009, IEEE Transactions on Neural Networks.

[3]  Ming Shao,et al.  Latent Low-Rank Transfer Subspace Learning for Missing Modality Recognition , 2014, AAAI.

[4]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[5]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[6]  Michael I. Jordan,et al.  Deep Transfer Learning with Joint Adaptation Networks , 2016, ICML.

[7]  Trevor Darrell,et al.  DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition , 2013, ICML.

[8]  Trevor Darrell,et al.  Simultaneous Deep Transfer Across Domains and Tasks , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[9]  Jianmin Wang,et al.  Transferable Attention for Domain Adaptation , 2019, AAAI.

[10]  Xiaogang Wang,et al.  Multi-context Attention for Human Pose Estimation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  Trevor Darrell,et al.  Adversarial Discriminative Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Yue Cao,et al.  Transferable Representation Learning with Deep Adaptation Networks , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Jun Fu,et al.  Dual Attention Network for Scene Segmentation , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Qingming Huang,et al.  Deep Unsupervised Convolutional Domain Adaptation , 2017, ACM Multimedia.

[15]  Michael I. Jordan,et al.  Unsupervised Domain Adaptation with Residual Transfer Networks , 2016, NIPS.

[16]  Qiang Yang,et al.  A Survey on Transfer Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.

[17]  Sethuraman Panchanathan,et al.  Deep Hashing Network for Unsupervised Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[19]  Michael I. Jordan,et al.  Transferable Adversarial Training: A General Approach to Adapting Deep Classifiers , 2019, ICML.

[20]  Carlos D. Castillo,et al.  Generate to Adapt: Aligning Domains Using Generative Adversarial Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[21]  Yuan Shi,et al.  Geodesic flow kernel for unsupervised domain adaptation , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[22]  Yoshua Bengio,et al.  How transferable are features in deep neural networks? , 2014, NIPS.

[23]  Niall M. Adams,et al.  Dataset Shift in Machine Learning , 2010 .

[24]  Michael I. Jordan,et al.  Conditional Adversarial Domain Adaptation , 2017, NeurIPS.

[25]  Jianmin Wang,et al.  Transferability vs. Discriminability: Batch Spectral Penalization for Adversarial Domain Adaptation , 2019, ICML.

[26]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[27]  Gene H. Golub,et al.  Singular value decomposition and least squares solutions , 1970, Milestones in Matrix Computation.

[28]  Jianmin Wang,et al.  Multi-Adversarial Domain Adaptation , 2018, AAAI.

[29]  Taesung Park,et al.  CyCADA: Cycle-Consistent Adversarial Domain Adaptation , 2017, ICML.

[30]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[31]  Trevor Darrell,et al.  Adapting Visual Category Models to New Domains , 2010, ECCV.

[32]  Mingsheng Long,et al.  Simultaneous Learning of Pivots and Representations for Cross-Domain Sentiment Classification , 2020, AAAI.

[33]  Pedro H. O. Pinheiro,et al.  Unsupervised Domain Adaptation with Similarity Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.