Confident Anchor-Induced Multi-Source Free Domain Adaptation

Unsupervised domain adaptation has attracted appealing academic attentions by transferring knowledge from labeled source domain to unlabeled target domain. However, most existing methods assume the source data are drawn from a single domain, which cannot be successfully applied to explore complementarily transferable knowledge from multiple source domains with large distribution discrepancies. Moreover, they require access to source data during training, which are inefficient and unpractical due to privacy preservation and memory storage. To address these challenges, we develop a novel Confident-Anchor-induced multisource-free Domain Adaptation (CAiDA) model, which is a pioneer exploration of knowledge adaptation from multiple source domains to the unlabeled target domain without any source data, but with only pre-trained source models. Specifically, a source-specific transferable perception module is proposed to automatically quantify the contributions of the complementary knowledge transferred from multi-source domains to the target domain. To generate pseudo labels for the target domain without access to the source data, we develop a confident-anchorinduced pseudo label generator by constructing a confident anchor group and assigning each unconfident target sample with a semantic-nearest confident anchor. Furthermore, a class-relationship-aware consistency loss is proposed to preserve consistent inter-class relationships by aligning soft confusion matrices across domains. Theoretical analysis answers why multi-source domains are better than a single source domain, and establishes a novel learning bound to show the effectiveness of exploiting multi-source domains. Experiments on several representative datasets illustrate the superiority of our proposed CAiDA model. The code is available at https://github.com/Learning-group123/CAiDA.

[1]  David J. C. MacKay,et al.  Unsupervised Classifiers, Mutual Information and 'Phantom Targets' , 1991, NIPS.

[2]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[3]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[4]  Yishay Mansour,et al.  Multiple Source Adaptation and the Rényi Divergence , 2009, UAI.

[5]  Andreas Krause,et al.  Discriminative Clustering by Regularized Information Maximization , 2010, NIPS.

[6]  Yuan Shi,et al.  Geodesic flow kernel for unsupervised domain adaptation , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[7]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[8]  George Trigeorgis,et al.  Domain Separation Networks , 2016, NIPS.

[9]  Mengjie Zhang,et al.  Deep Reconstruction-Classification Networks for Unsupervised Domain Adaptation , 2016, ECCV.

[10]  François Laviolette,et al.  Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..

[11]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[12]  Dacheng Tao,et al.  Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Shatha Jaradat,et al.  Deep Cross-Domain Fashion Recommendation , 2017, RecSys.

[14]  Sethuraman Panchanathan,et al.  Deep Hashing Network for Unsupervised Domain Adaptation , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Mehryar Mohri,et al.  Algorithms and Theory for Multiple-Source Adaptation , 2018, NeurIPS.

[16]  Liang Lin,et al.  Deep Cocktail Network: Multi-source Unsupervised Domain Adaptation with Category Shift , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[17]  Swami Sankaranarayanan,et al.  MetaReg: Towards Domain Generalization using Meta-Regularization , 2018, NeurIPS.

[18]  Taesung Park,et al.  CyCADA: Cycle-Consistent Adversarial Domain Adaptation , 2017, ICML.

[19]  Regina Barzilay,et al.  Multi-Source Domain Adaptation with Mixture of Experts , 2018, EMNLP.

[20]  José M. F. Moura,et al.  Adversarial Multiple Source Domain Adaptation , 2018, NeurIPS.

[21]  Bo Wang,et al.  Moment Matching for Multi-Source Domain Adaptation , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[22]  Dongdong Hou,et al.  Semantic-Transferable Weakly-Supervised Endoscopic Lesions Segmentation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[23]  Namil Kim,et al.  Drop to Adapt: Learning Discriminative Features for Unsupervised Domain Adaptation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[24]  Paolo Russo,et al.  Towards Multi-source Adaptive Semantic Segmentation , 2019, ICIAP.

[25]  Xiao Wang,et al.  Eavesdrop the Composition Proportion of Training Labels in Federated Learning , 2019, ArXiv.

[26]  Kurt Keutzer,et al.  Multi-source Domain Adaptation for Semantic Segmentation , 2019, NeurIPS.

[27]  Sergey Levine,et al.  Meta-Learning with Implicit Gradients , 2019, NeurIPS.

[28]  Michael I. Jordan,et al.  Transferable Normalization: Towards Improving Transferability of Deep Neural Networks , 2019, NeurIPS.

[29]  Sungeun Hong,et al.  Progressive Domain Adaptation from a Source Pre-trained Model. , 2020 .

[30]  Tat-Seng Chua,et al.  Multi-source Domain Adaptation for Visual Sentiment Classification , 2020, AAAI.

[31]  Gan Sun,et al.  Dual Relation Semi-Supervised Multi-Label Learning , 2020, AAAI.

[32]  Maneesh Singh,et al.  Progressive Domain Adaptation for Object Detection , 2019, 2020 IEEE Winter Conference on Applications of Computer Vision (WACV).

[33]  Zi Huang,et al.  Progressive Graph Learning for Open-Set Domain Adaptation , 2020, ICML.

[34]  Jiahua Dong,et al.  CSCL: Critical Semantic-Consistent Learning for Unsupervised Domain Adaptation , 2020, ECCV.

[35]  Hau-San Wong,et al.  Model Adaptation: Unsupervised Domain Adaptation Without Source Data , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  Xiao Wang,et al.  Towards Class Imbalance in Federated Learning , 2020, ArXiv.

[37]  Jiashi Feng,et al.  Do We Really Need to Access the Source Data? Source Hypothesis Transfer for Unsupervised Domain Adaptation , 2020, ICML.

[38]  Ser-Nam Lim,et al.  Curriculum Manager for Source Selection in Multi-Source Domain Adaptation , 2020, ECCV.

[39]  Kotagiri Ramamohanarao,et al.  Learning with Bounded Instance- and Label-dependent Label Noise , 2017, ICML.

[40]  Xiaowei Xu,et al.  What Can Be Transferred: Unsupervised Domain Adaptation for Endoscopic Lesions Segmentation , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[41]  Bingbing Ni,et al.  Learning to Combine: Knowledge Aggregation for Multi-Source Domain Adaptation , 2020, ECCV.

[42]  K. Keutzer,et al.  Multi-source Distilling Domain Adaptation , 2019, AAAI.

[43]  Feng Liu,et al.  Learning Deep Kernels for Non-Parametric Two-Sample Tests , 2020, ICML.

[44]  J. Weijer,et al.  Unsupervised Domain Adaptation without Source Data by Casting a BAIT , 2020, ArXiv.

[45]  Lu Cao,et al.  Multisource Selective Transfer Framework in Multiobjective Optimization Problems , 2020, IEEE Transactions on Evolutionary Computation.

[46]  Bo Yuan,et al.  Clarinet: A One-step Approach Towards Budget-friendly Unsupervised Domain Adaptation , 2020, IJCAI.

[47]  Gang Niu,et al.  Confidence Scores Make Instance-dependent Label-noise Learning Possible , 2019, ICML.

[48]  Jie Lu,et al.  Multi-Source Contribution Learning for Domain Adaptation , 2021, IEEE Transactions on Neural Networks and Learning Systems.

[49]  Jie Lu,et al.  Learning Bounds for Open-Set Learning , 2021, ICML.

[50]  Masashi Sugiyama,et al.  Provably End-to-end Label-Noise Learning without Anchor Points , 2021, ICML.

[51]  Mingkui Tan,et al.  Source-free Domain Adaptation via Avatar Prototype Generation and Adaptation , 2021, IJCAI.

[52]  Dripta S. Raychaudhuri,et al.  Unsupervised Multi-source Domain Adaptation Without Access to Source Data , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[53]  Xiaowei Xu,et al.  Weakly-Supervised Cross-Domain Adaptation for Endoscopic Lesions Segmentation , 2020, IEEE Transactions on Circuits and Systems for Video Technology.

[54]  Long Lan,et al.  TOHAN: A One-step Approach towards Few-shot Hypothesis Adaptation , 2021, ArXiv.

[55]  Alex Olshevsky,et al.  Optimal Vaccine Allocation for Pandemic Stabilization , 2021, 2109.04612.

[56]  Yun Fu,et al.  Generic Multi-label Annotation via Adaptive Graph and Marginalized Augmentation , 2021, ACM Trans. Knowl. Discov. Data.

[57]  Feng Liu,et al.  Open Set Domain Adaptation: Theoretical Bound and Algorithm , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[58]  Alexander Olshevsky,et al.  Optimal Lockdown for Pandemic Control , 2020, 2010.12923.

[59]  Feng Liu,et al.  How does the Combined Risk Affect the Performance of Unsupervised Domain Adaptation Approaches? , 2020, AAAI.

[60]  Zijian Wang,et al.  Learning to Diversify for Single Domain Generalization , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[61]  Mingming Gong,et al.  Class2Simi: A Noise Reduction Perspective on Learning with Noisy Labels , 2020, ICML.

[62]  Errui Ding,et al.  Unsupervised Multi-Source Domain Adaptation for Person Re-Identification , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).