Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2014 Aug 12;9(8):e104314.
doi: 10.1371/journal.pone.0104314. eCollection 2014.

Learning a weighted meta-sample based parameter free sparse representation classification for microarray data

Affiliations

Learning a weighted meta-sample based parameter free sparse representation classification for microarray data

Bo Liao et al. PLoS One. .

Abstract

Sparse representation classification (SRC) is one of the most promising classification methods for supervised learning. This method can effectively exploit discriminating information by introducing a [Symbol: see text]1 regularization terms to the data. With the desirable property of sparisty, SRC is robust to both noise and outliers. In this study, we propose a weighted meta-sample based non-parametric sparse representation classification method for the accurate identification of tumor subtype. The proposed method includes three steps. First, we extract the weighted meta-samples for each sub class from raw data, and the rationality of the weighting strategy is proven mathematically. Second, sparse representation coefficients can be obtained by [Symnbol: see text]1 regularization of underdetermined linear equations. Thus, data dependent sparsity can be adaptively tuned. A simple characteristic function is eventually utilized to achieve classification. Asymptotic time complexity analysis is applied to our method. Compared with some state-of-the-art classifiers, the proposed method has lower time complexity and more flexibility. Experiments on eight samples of publicly available gene expression profile data show the effectiveness of the proposed method.

PubMed Disclaimer

Conflict of interest statement

Competing Interests: The authors have declared that no competing interests exist.

Figures

Figure 1
Figure 1. Illustration of meta-sample model: each column vector of can be represented within a linear combination of meta-samples in , and the column of corresponds to the linear combination coefficients.
Figure 2
Figure 2. Optimal classification accuracy of MSRC achieved on COLON; the -axis represents the number of meta-samples (left) and the regularization parameter (right).
Classification accuracy is more sensitive to the number of meta-samples rather than to the regularization parameter.
Figure 3
Figure 3. The flowchart of PFMSRC scheme.
Figure 4
Figure 4. Comparison of prediction accuracy on four binary classification datasets by varying the number of samples from per subclass; when is larger than 10 the model based method prediction accuracy decreases as increases.
Figure 5
Figure 5. Comparison of prediction accuracy on four multiclass classification datasets by varying the number of samples from per subclass; when is larger than 10 the performance degradation of model based methods is less significant than that of binary classification.
Figure 6
Figure 6. Comparison of prediction accuracy on four binary classification datasets by varying the number of top selected genes.
Figure 7
Figure 7. Comparison of prediction accuracy on four multiclass classification datasets by varying the number of top selected genes.

Similar articles

Cited by

References

    1. Alizadeh AA, Eisen MB, Davis RE, Ma C, Lossos IS, et al. (2000) Distinct types of diffuse large b-cell lymphoma identified by gene expression profiling. Nature 403: 503–511. - PubMed
    1. West M (2003) Bayesian factor regression models in the large p, small n paradigm. Bayesian statistics 7: 723–732.
    1. Liu B, Fang B, Liu X, Chen J, Huang Z (2013) Large margin subspace learning for feature selection. Pattern Recognition.
    1. Cai D, He X, Zhou K, Han J, Bao H (2007) Locality sensitive discriminant analysis. In: IJCAI. pp. 708–713.
    1. Sugiyama M (2006) Local fisher discriminant analysis for supervised dimensionality reduction. In: Proceedings of the 23rd international conference on Machine learning ACM, pp. 905–912.

Publication types

LinkOut - more resources