Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2015:2015:821798.
doi: 10.1155/2015/821798. Epub 2015 Sep 28.

An Efficient Feature Subset Selection Algorithm for Classification of Multidimensional Dataset

Affiliations

An Efficient Feature Subset Selection Algorithm for Classification of Multidimensional Dataset

Senthilkumar Devaraj et al. ScientificWorldJournal. 2015.

Abstract

Multidimensional medical data classification has recently received increased attention by researchers working on machine learning and data mining. In multidimensional dataset (MDD) each instance is associated with multiple class values. Due to its complex nature, feature selection and classifier built from the MDD are typically more expensive or time-consuming. Therefore, we need a robust feature selection technique for selecting the optimum single subset of the features of the MDD for further analysis or to design a classifier. In this paper, an efficient feature selection algorithm is proposed for the classification of MDD. The proposed multidimensional feature subset selection (MFSS) algorithm yields a unique feature subset for further analysis or to build a classifier and there is a computational advantage on MDD compared with the existing feature selection algorithms. The proposed work is applied to benchmark multidimensional datasets. The number of features was reduced to 3% minimum and 30% maximum by using the proposed MFSS. In conclusion, the study results show that MFSS is an efficient feature selection algorithm without affecting the classification accuracy even for the reduced number of features. Also the proposed MFSS algorithm is suitable for both problem transformation and algorithm adaptation and it has great potentials in those applications generating multidimensional datasets.

PubMed Disclaimer

Figures

Figure 1
Figure 1
The relationship among different classification paradigms.
Figure 2
Figure 2
Proposed MFSS for multidimensional dataset.
Figure 3
Figure 3
Hamming score-Naive Bayes.
Figure 4
Figure 4
Hamming score-SVM.
Figure 5
Figure 5
Hamming score-IBk.
Figure 6
Figure 6
Hamming score-J48.
Figure 7
Figure 7
Exact match: Naive Bayes.
Figure 8
Figure 8
Exact match-SVM.
Figure 9
Figure 9
Exact match-IBk.
Figure 10
Figure 10
Exact match-J48.
Figure 11
Figure 11
Features selected using proposed MFSS.

Similar articles

Cited by

References

    1. Read J. A pruned problem transformation method for multi-label classification. Proceedings of the 6th New Zealand Computer Science Research Student Conference (NZCSRSC '08); April 2008; pp. 143–150.
    1. Mohamad M. S., Deris S., Yatim S. M., Othman M. R. Feature selection method using genetic algorithm for the classification of small and high dimension data. Proceedings of the 1st International Symposium on Information and Communication Technology; 2004; pp. 1–4.
    1. Zhang D., Chen S., Zhou Z.-H. Constraint score: a new filter method for feature selection with pairwise constraints. Pattern Recognition. 2008;41(5):1440–1451. doi: 10.1016/j.patcog.2007.10.009. - DOI
    1. Grande J., del Rosario Suarez M., Villar J. R. Innovations in Hybrid Intelligent Systems. Vol. 44. Berlin, Germany: Springer; 2007. A feature selection method using a fuzzy mutual information measure; pp. 56–63. (Advances in Soft Computing). - DOI
    1. Zhang M.-L., Zhou Z.-H. A review on multi-label learning algorithms. IEEE Transactions on Knowledge and Data Engineering. 2014;26(8):1819–1837. doi: 10.1109/TKDE.2013.39. - DOI

LinkOut - more resources