Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Jun:174:106230.
doi: 10.1016/j.neunet.2024.106230. Epub 2024 Mar 11.

Source-free unsupervised domain adaptation: A survey

Affiliations

Source-free unsupervised domain adaptation: A survey

Yuqi Fang et al. Neural Netw. 2024 Jun.

Abstract

Unsupervised domain adaptation (UDA) via deep learning has attracted appealing attention for tackling domain-shift problems caused by distribution discrepancy across different domains. Existing UDA approaches highly depend on the accessibility of source domain data, which is usually limited in practical scenarios due to privacy protection, data storage and transmission cost, and computation burden. To tackle this issue, many source-free unsupervised domain adaptation (SFUDA) methods have been proposed recently, which perform knowledge transfer from a pre-trained source model to the unlabeled target domain with source data inaccessible. A comprehensive review of these works on SFUDA is of great significance. In this paper, we provide a timely and systematic literature review of existing SFUDA approaches from a technical perspective. Specifically, we categorize current SFUDA studies into two groups, i.e., white-box SFUDA and black-box SFUDA, and further divide them into finer subcategories based on different learning strategies they use. We also investigate the challenges of methods in each subcategory, discuss the advantages/disadvantages of white-box and black-box SFUDA methods, conclude the commonly used benchmark datasets, and summarize the popular techniques for improved generalizability of models learned without using source data. We finally discuss several promising future directions in this field.

Keywords: Domain adaptation; Source-free; Survey; Unsupervised learning.

PubMed Disclaimer

Conflict of interest statement

Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Figures

Fig. 1.
Fig. 1.
Illustration of (a) conventional unsupervised domain adaptation (UDA), (b) white-box source-free UDA (SFUDA), and (c) black-box SFUDA. Compared with (a) conventional UDA that relies on labeled source data {XS, YS} and unlabeled target data XT, (b, c) SFUDA performs knowledge transfer by directly leveraging a pre-trained source model ΦS and unlabeled target data XT. The difference between (b) white-box SFUDA and (c) black-box SFUDA lies in whether the learnable parameters of the source model ΦS are accessible or not. API: application programming interface.
Fig. 2.
Fig. 2.
Taxonomy of existing source-free unsupervised domain adaptation (SFUDA) methods and future outlook.
Fig. 3.
Fig. 3.
Overview of the number of SFUDA studies (in terms of published years) that have been collected In this survey.
Fig. 4.
Fig. 4.
Illustration of Batch Normalization Statistics Transfer methods for source Image generation. By matching batch normalization (BN) statistics between the upper and lower branches, source-like data can be generated by preserving the target content but with source style. Unsupervised domain adaptation (UDA) is then performed between source-like data and target data.
Fig. 5.
Fig. 5.
Illustration of Surrogate Source Data Construction methods for source data generation. These methods first construct surrogate/proxy source data by selecting appropriate samples from the target domain and then perform standard unsupervised domain adaptation (UDA).
Fig. 6.
Fig. 6.
Illustration of Generative Adversarial Network (GAN) based Image Generation methods for source data generation. Typically, a pre-defined label and random noise act as the Inputs of a GAN-based generator. By utilizing the pre-trained source model, they synthesize source data for cross-domain adaptation. LCE: Cross-entropy loss function.
Fig. 7.
Fig. 7.
Illustration of Self-Supervised Knowledge Distillation methods for source-free unsupervised domain adaptation. With target data from different augmentations as Inputs, a teacher-student framework Is utilized to exploit target features, where parameters of teacher network are usually exponential moving average (EMA) of those of student network. Aug-α and Aug-β denote two data augmentation methods (e.g., flip, rotation, shift, noise addition, distortion, etc.), respectively. LKD: Knowledge distillation loss function.
Fig. 8.
Fig. 8.
Illustration of Domain Alignment via Statistics methods for source-free unsupervised domain adaptation. The corresponding methods leverage batch statistics stored in the pre-trained source model to approximate the distribution of inaccessible source data, and then perform cross-domain adaptation by reducing distribution discrepancy between source and target domains.
Fig. 9.
Fig. 9.
Illustration of Contrastive Learning methods for source-free unsupervised domain adaptation. These methods exploit discriminative representations among unlabeled target data by pulling instances of similar categories closer and pushing instances of different categories away in feature space.
Fig. 10.
Fig. 10.
Illustration of Uncertainty-Guided Adaptation methods for source-free unsupervised domain adaptation. These studies utilize uncertainty to guide target predictions, and such valuable information can be measured by Monte Carlo Dropout, entropy, etc.
Fig. 11.
Fig. 11.
Illustration of Hidden Structure Mining methods for source-free unsupervised domain adaptation. These methods take into consideration intrinsic feature structures of the target domain and iterate between target model refinement and clustering centroid update.

References

    1. Abusitta A, Wahab OA, & Fung BC (2021). VirtualGAN: Reducing mode collapse in generative adversarial networks using virtual mapping. In 2021 International Joint Conference on Neural Networks (pp. 1–6). IEEE.
    1. Agarwal P, Paudel DP, Zaech JN, & Van Gool L (2022). Unsupervised robust domain adaptation without source data. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (pp. 2009–2018).
    1. Ahmed SM, Lohit S, Peng KC, Jones MJ, & Roy-Chowdhury AK (2022). Cross-modal knowledge transfer without task-relevant source data. In European Conference on Computer Vision (pp. 111–127). Springer.
    1. Ahmed SM, Raychaudhuri DS, Paul S, Oymak S, & Roy-Chowdhury AK (2021). Unsupervised multi-source domain adaptation without access to source data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 10103–10112).
    1. Arazo E, Ortego D, Albert P, O’Connor N, & McGuinness K (2019). Unsupervised label noise modeling and loss correction. In International Conference on Machine Learning (pp. 312–321). PMLR.

Publication types

LinkOut - more resources