Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2011 Oct 1;58(3):785-92.
doi: 10.1016/j.neuroimage.2011.06.029. Epub 2011 Jun 25.

Effects of hardware heterogeneity on the performance of SVM Alzheimer's disease classifier

Affiliations

Effects of hardware heterogeneity on the performance of SVM Alzheimer's disease classifier

Ahmed Abdulkadir et al. Neuroimage. .

Abstract

Fully automated machine learning methods based on structural magnetic resonance imaging (MRI) data can assist radiologists in the diagnosis of Alzheimer's disease (AD). These algorithms require large data sets to learn the separation of subjects with and without AD. Training and test data may come from heterogeneous hardware settings, which can potentially affect the performance of disease classification. A total of 518 MRI sessions from 226 healthy controls and 191 individuals with probable AD from the multicenter Alzheimer's Disease Neuroimaging Initiative (ADNI) were used to investigate whether grouping data by acquisition hardware (i.e. vendor, field strength, coil system) is beneficial for the performance of a support vector machine (SVM) classifier, compared to the case where data from different hardware is mixed. We compared the change of the SVM decision value resulting from (a) changes in hardware against the effect of disease and (b) changes resulting simply from rescanning the same subject on the same machine. Maximum accuracy of 87% was obtained with a training set of all 417 subjects. Classifiers trained with 95 subjects in each diagnostic group and acquired with heterogeneous scanner settings had an empirical detection accuracy of 84.2±2.4% when tested on an independent set of the same size. These results mirror the accuracy reported in recent studies. Encouragingly, classifiers trained on images acquired with homogenous and heterogeneous hardware settings had equivalent cross-validation performances. Two scans of the same subject acquired on the same machine had very similar decision values and were generally classified into the same group. Higher variation was introduced when two acquisitions of the same subject were performed on two scanners with different field strengths. The variation was unbiased and similar for both diagnostic groups. The findings of the study encourage the pooling of data from different sites to increase the number of training samples and thereby improving performance of disease classifiers. Although small, a change in hardware could lead to a change of the decision value and thus diagnostic grouping. The findings of this study provide estimators for diagnostic accuracy of an automated disease diagnosis method involving scans acquired with different sets of hardware. Furthermore, we show that the level of confidence in the performance estimation significantly depends on the size of the training sample, and hence should be taken into account in a clinical setting.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1
Flow-chart representation of the comparison of an example computation to the mean accuracy distribution of random sets. Comparison of PURE set accuracy with expected accuracy of mixed sets using z-test.
Fig. 2
Fig. 2
Box-plots of leave-one-out cross-validation (LOO-CV) accuracy as function of group size (x-axis) obtained by 500 random permutations.
Fig. 3
Fig. 3
Changes of the SVM decision value (y-axis) between back-to-back scans (BTB), separately for diagnostic group (first panel) and field strength (second panel), and with changes between two field strengths (thus also two systems), separately for diagnostic group (third panel) for acquisitions of 96 subjects. Change in field strength (FS) does not introduce systematic bias (one-sample t-test p>0.05). A: Training set composed of 316 images. B: Training set composed of 80 images.
Fig. 4
Fig. 4
Variance of decision values when comparing back-to-back scans or change of field strength compared to the effect of group. A: Training set size=316. B: Training set size=80. CN: healthy controls, AD: subject with probable AD, BTB: back-to-back (within subjects), FS: field strength (within subjects).

References

    1. Acosta-Cabronero J, Williams GB, Pereira JMS, Pengas G, Nestor PJ. The impact of skull-stripping and radio-frequency bias correction on grey-matter segmentation for voxel-based morphometry. Neuroimage. 2008;39:1654–1665. - PubMed
    1. Ashburner J. A fast diffeomorphic image registration algorithm. Neuroimage. 2007;38:95–113. - PubMed
    1. Ashburner J, Friston KJ. Voxel-based morphometry — the methods. Neuroimage. 2000;11:805–821. - PubMed
    1. Ashburner J, Friston KJ. Unified segmentation. Neuroimage. 2005;26:839–851. - PubMed
    1. Baron JC, Chételat G, Desgranges B, Perchey G, Landeau B, de la Sayette V, Eustache F. In vivo mapping of gray matter loss with voxel-based morphometry in mild Alzheimer’s disease. Neuroimage. 2001;14:298–309. - PubMed

Publication types

MeSH terms