Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023;53(12):15909-15922.
doi: 10.1007/s10489-022-04288-4. Epub 2022 Nov 29.

Semi-supervised adversarial discriminative domain adaptation

Affiliations

Semi-supervised adversarial discriminative domain adaptation

Thai-Vu Nguyen et al. Appl Intell (Dordr). 2023.

Abstract

Domain adaptation is a potential method to train a powerful deep neural network across various datasets. More precisely, domain adaptation methods train the model on training data and test that model on a completely separate dataset. The adversarial-based adaptation method became popular among other domain adaptation methods. Relying on the idea of GAN, the adversarial-based domain adaptation tries to minimize the distribution between the training and testing dataset based on the adversarial learning process. We observe that the semi-supervised learning approach can combine with the adversarial-based method to solve the domain adaptation problem. In this paper, we propose an improved adversarial domain adaptation method called Semi-Supervised Adversarial Discriminative Domain Adaptation (SADDA), which can outperform other prior domain adaptation methods. We also show that SADDA has a wide range of applications and illustrate the promise of our method for image classification and sentiment classification problems.

Keywords: Domain adaptation; Semi-supervised adversarial discriminative domain adaptation; Semi-supervised domain adaptation.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1
Examples of images from different datasets. (a) Some digit images from MNIST [5], USPS [6], MNIST-M [7], and SVHN [8] datasets. (b) Some object images from the “bird” category in CALTECH [9], LABELME [10], PASCAL [11], and SUN [12] datasets
Fig. 2
Fig. 2
An overview of the SADDA. Firstly, training the source encoder (Ms) and the classification (Cs) using the source labeled images (Xs, Ys). Secondly, training a target encoder (Mt) through the domain adversarial process. Finally, in the testing phase, concatenate the target encoder (Mt) and the classification (Cs) to create the complete model, which will predict the label of the target dataset precisely
Fig. 3
Fig. 3
t-SNE embedding of digit classification, using (2 x 2 x 256) dimensional representation, with Source only (on the left) and SADDA (on the right) on the target dataset. Note that SADDA minimizes intra-class distance and maximizes inter-class distance
Fig. 4
Fig. 4
The discriminator loss and adversarial loss in the LABEL → CALTECH experiment
Fig. 5
Fig. 5
The overview of the SADDA method for the digit recognition task. We found that Global Average Pooling (GAP) [55] increased model stability and reduce the number of parameter
Fig. 6
Fig. 6
The overview of the SADDA method for the object recognition task on the VLCS [23] dataset. With the input image’s shape is 64 x 64 x 3
Fig. 7
Fig. 7
The overview of the SADDA method for the sentiment classification task. The input sentence has a max length equal to 50. In the design above, to prevent overfitting, the LSTM layer is always followed by the dropout layer with 0.2 rates. The numbers under the particular layer are the output shape of that layer

References

    1. Gretton A, Smola A, Huang J, Schmittfull M, Borgwardt K, Schölkopf B (2009) Covariate shift and local learning by distribution matching, pp 131–160. Cambridge, MA USA: MIT Press
    1. Long M, Cao Y, Wang J, Jordan MI (2015) Learning transferable features with deep adaptation networks. arXiv:1502.02791 - PubMed
    1. Nguyen A, Nguyen N, Tran K, Tjiputra E, Tran QD (2020) Autonomous navigation in complex environments with deep multimodal fusion network. In: 2020 IEEE/RSJ international conference on intelligent robots and systems (IROS)
    1. Donahue J, Jia Y, Vinyals O, Hoffman J, Zhang N, Tzeng E, Darrell T (2013) Decaf: A deep convolutional activation feature for generic visual recognition
    1. Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–2324. doi: 10.1109/5.726791. - DOI

LinkOut - more resources