Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2025 Jul-Aug;22(4):1606-1614.
doi: 10.1109/TCBBIO.2025.3562809.

PScL-SDNNMAE: Protein Subcellular Localization Prediction Using Classical and Masked Autoencoder-Based Multi-View Features With Ensemble Feature Selection

PScL-SDNNMAE: Protein Subcellular Localization Prediction Using Classical and Masked Autoencoder-Based Multi-View Features With Ensemble Feature Selection

Shenjian Gu et al. IEEE Trans Comput Biol Bioinform. 2025 Jul-Aug.

Abstract

Accurate prediction of protein subcellular localization is critical for understanding cellular functions and guiding drug design. However, current computational methods have limited and insufficient performance and as such, there exist few efficient vision learners based on self-supervised learning for extracting deep and informative features. To address it, we propose a novel bioimage-based method, termed PScL-SDNNMAE, to effectively predict the subcellular localizations of proteins in human cells. PScL-SDNNMAE first extracts classical features using traditional image descriptors. Next, the masked autoencoder (MAE) is first trained using the training image data and then used to extract the MAE-based deep features. In the feature selection phase, PScL-SDNNMAE applies the Analysis of Variance (ANOVA), Mutual Information (MI) and stepwise discriminant analysis (SDA) to select the optimal features from the classical feature sets. Finally, PScL-SDNNMAE trains the deep neural network (DNN) classifier using the super feature set generated by integrating all the classical optimal and MAE-based deep features. Extensive benchmark experiments including 10-fold cross-validation on the training dataset and independent test on the independent dataset illustrate more advanced performance and generalization capability of PScL-SDNNMAE than other existing state-of-the-art predictors. Moreover, the experiments also demonstrate the effectiveness of self-supervised learning methods in learning representations of IHC images, as well as the significant potential for pre-training on massive unlabeled datasets in the future.

PubMed Disclaimer

Similar articles