Classifier design for computer-aided diagnosis: effects of finite sample size on the mean performance of classical and neural network classifiers
- PMID: 10619251
- DOI: 10.1118/1.598805
Classifier design for computer-aided diagnosis: effects of finite sample size on the mean performance of classical and neural network classifiers
Abstract
Classifier design is one of the key steps in the development of computer-aided diagnosis (CAD) algorithms. A classifier is designed with case samples drawn from the patient population. Generally, the sample size available for classifier design is limited, which introduces variance and bias into the performance of the trained classifier, relative to that obtained with an infinite sample size. For CAD applications, a commonly used performance index for a classifier is the area, Az, under the receiver operating characteristic (ROC) curve. We have conducted a computer simulation study to investigate the dependence of the mean performance, in terms of Az, on design sample size for a linear discriminant and two nonlinear classifiers, the quadratic discriminant and the backpropagation neural network (ANN). The performances of the classifiers were compared for four types of class distributions that have specific properties: multivariate normal distributions with equal covariance matrices and unequal means, unequal covariance matrices and unequal means, and unequal covariance matrices and equal means, and a feature space where the two classes were uniformly distributed in disjoint checkerboard regions. We evaluated the performances of the classifiers in feature spaces of dimensionality ranging from 3 to 15, and design sample sizes from 20 to 800 per class. The dependence of the resubstitution and hold-out performance on design (training) sample size (Nt) was investigated. For multivariate normal class distributions with equal covariance matrices, the linear discriminant is the optimal classifier. It was found that its Az-versus-1/Nt curves can be closely approximated by linear dependences over the range of sample sizes studied. In the feature spaces with unequal covariance matrices where the quadratic discriminant is optimal, the linear discriminant is inferior to the quadratic discriminant or the ANN when the design sample size is large. However, when the design sample is small, a relatively simple classifier, such as the linear discriminant or an ANN with very few hidden nodes, may be preferred because performance bias increases with the complexity of the classifier. In the regime where the classifier performance is dominated by the 1/Nt term, the performance in the limit of infinite sample size can be estimated as the intercept (1/Nt= 0) of a linear regression of Az versus 1/Nt. The understanding of the performance of the classifiers under the constraint of a finite design sample size is expected to facilitate the selection of a proper classifier for a given classification task and the design of an efficient resampling scheme.
Similar articles
-
Effect of finite sample size on feature selection and classification: a simulation study.Med Phys. 2010 Feb;37(2):907-20. doi: 10.1118/1.3284974. Med Phys. 2010. PMID: 20229900 Free PMC article.
-
Feature selection and classifier performance in computer-aided diagnosis: the effect of finite sample size.Med Phys. 2000 Jul;27(7):1509-22. doi: 10.1118/1.599017. Med Phys. 2000. PMID: 10947254 Free PMC article.
-
Classifier performance prediction for computer-aided diagnosis using a limited dataset.Med Phys. 2008 Apr;35(4):1559-70. doi: 10.1118/1.2868757. Med Phys. 2008. PMID: 18491550 Free PMC article.
-
Trainable fusion rules. I. Large sample size case.Neural Netw. 2006 Dec;19(10):1506-16. doi: 10.1016/j.neunet.2006.01.018. Epub 2006 Apr 3. Neural Netw. 2006. PMID: 16580815 Review.
-
Optimization of Network Topology in Computer-Aided Detection Schemes Using Phased Searching with NEAT in a Time-Scaled Framework.Cancer Inform. 2014 Oct 13;13(Suppl 1):17-27. doi: 10.4137/CIN.S13885. eCollection 2014. Cancer Inform. 2014. PMID: 25392680 Free PMC article. Review.
Cited by
-
Evaluation of data augmentation via synthetic images for improved breast mass detection on mammograms using deep learning.J Med Imaging (Bellingham). 2020 Jan;7(1):012703. doi: 10.1117/1.JMI.7.1.012703. Epub 2019 Nov 22. J Med Imaging (Bellingham). 2020. PMID: 31763356 Free PMC article.
-
Potential of computer-aided diagnosis of high spectral and spatial resolution (HiSS) MRI in the classification of breast lesions.J Magn Reson Imaging. 2014 Jan;39(1):59-67. doi: 10.1002/jmri.24145. Epub 2013 Sep 10. J Magn Reson Imaging. 2014. PMID: 24023011 Free PMC article.
-
Bladder Cancer Treatment Response Assessment in CT using Radiomics with Deep-Learning.Sci Rep. 2017 Aug 18;7(1):8738. doi: 10.1038/s41598-017-09315-w. Sci Rep. 2017. PMID: 28821822 Free PMC article.
-
Dynamic multiple thresholding breast boundary detection algorithm for mammograms.Med Phys. 2010 Jan;37(1):391-401. doi: 10.1118/1.3273062. Med Phys. 2010. PMID: 20175501 Free PMC article.
-
Noise injection for training artificial neural networks: a comparison with weight decay and early stopping.Med Phys. 2009 Oct;36(10):4810-8. doi: 10.1118/1.3213517. Med Phys. 2009. PMID: 19928111 Free PMC article.
Publication types
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
Other Literature Sources
Research Materials
Miscellaneous