Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2018 Dec:31:6799-6810.

Extracting Relationships by Multi-Domain Matching

Affiliations

Extracting Relationships by Multi-Domain Matching

Yitong Li et al. Adv Neural Inf Process Syst. 2018 Dec.

Abstract

In many biological and medical contexts, we construct a large labeled corpus by aggregating many sources to use in target prediction tasks. Unfortunately, many of the sources may be irrelevant to our target task, so ignoring the structure of the dataset is detrimental. This work proposes a novel approach, the Multiple Domain Matching Network (MDMN), to exploit this structure. MDMN embeds all data into a shared feature space while learning which domains share strong statistical relationships. These relationships are often insightful in their own right, and they allow domains to share strength without interference from irrelevant data. This methodology builds on existing distribution-matching approaches by assuming that source domains are varied and outcomes multi-factorial. Therefore, each domain should only match a relevant subset. Theoretical analysis shows that the proposed approach can have a tighter generalization bound than existing multiple-domain adaptation approaches. Empirically, we show that the proposed methodology handles higher numbers of source domains (up to 21 empirically), and provides state-of-the-art performance on image, text, and multi-channel time series classification, including clinical outcome data in an open label trial evaluating a novel treatment for Autism Spectrum Disorder.

PubMed Disclaimer

Figures

Figure 1:
Figure 1:
Figure 1(a) visualizes previous multiple-domain adaptation methods. Figure 1(b) visualizes the proposed method, with domain adaptation between all domains.
Figure 2:
Figure 2:
Figure 2 is a visualization of the graph induced on 22 patients by the proposed model, MDMN. Each node represents one subject and the target domain is shown in blue. Note that although the target is only strongly connected to one source domain, the links between source domains allow them to share strength and make more robust predictions. The lines are labeled by the mean of the directional weights learned in MDMN.
Figure 3:
Figure 3:
The framework of MDMN.
Figure 4:
Figure 4:
Visualization of feature spaces of different models by t-SNE. Each color represents one dataset of MNIST, MNISTM, SVHN and USPS. The testing target domain is MNISTM. The digit label is shown in the plot. The goal is to adapt generalized feature from source domains to the target domains; the digits should cluster together rather than the color clustering.
Figure 5:
Figure 5:
Relative classification accuracy by subject on two EEG datasets. The accuracy without subtracting the baseline performance is given in appendix C.2.

References

    1. Ao S, Li X, and Ling CX. Fast generalized distillation for semi-supervised domain adaptation. In AAAI, 2017.
    1. Arjovsky M, Chintala S, and Bottou L. Wasserstein gan. arXiv preprint arXiv:1701.07875, 2017.
    1. Ash JT, Schapire RE, and Engelhardt BE. Unsupervised domain adaptation using approximate label matching. arXiv preprint arXiv:1602.04889, 2016.
    1. Ben-David S, Blitzer J, Crammer K, Kulesza A, Pereira F, and Vaughan JW. A theory of learning from different domains. Machine Learning, 2010.
    1. Ben-David S and Urner R. Domain adaptation–can quantity compensate for quality? Annals of Mathematics and Artificial Intelligence, 2014.