Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2022 Sep 28;22(19):7384.
doi: 10.3390/s22197384.

A Review on Multiscale-Deep-Learning Applications

Affiliations
Review

A Review on Multiscale-Deep-Learning Applications

Elizar Elizar et al. Sensors (Basel). .

Abstract

In general, most of the existing convolutional neural network (CNN)-based deep-learning models suffer from spatial-information loss and inadequate feature-representation issues. This is due to their inability to capture multiscale-context information and the exclusion of semantic information throughout the pooling operations. In the early layers of a CNN, the network encodes simple semantic representations, such as edges and corners, while, in the latter part of the CNN, the network encodes more complex semantic features, such as complex geometric shapes. Theoretically, it is better for a CNN to extract features from different levels of semantic representation because tasks such as classification and segmentation work better when both simple and complex feature maps are utilized. Hence, it is also crucial to embed multiscale capability throughout the network so that the various scales of the features can be optimally captured to represent the intended task. Multiscale representation enables the network to fuse low-level and high-level features from a restricted receptive field to enhance the deep-model performance. The main novelty of this review is the comprehensive novel taxonomy of multiscale-deep-learning methods, which includes details of several architectures and their strengths that have been implemented in the existing works. Predominantly, multiscale approaches in deep-learning networks can be classed into two categories: multiscale feature learning and multiscale feature fusion. Multiscale feature learning refers to the method of deriving feature maps by examining kernels over several sizes to collect a larger range of relevant features and predict the input images' spatial mapping. Multiscale feature fusion uses features with different resolutions to find patterns over short and long distances, without a deep network. Additionally, several examples of the techniques are also discussed according to their applications in satellite imagery, medical imaging, agriculture, and industrial and manufacturing systems.

Keywords: artificial intelligence; convolutional neural network; deep learning; machine learning; multiscale features; neural network.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
The primary taxonomy of multiscale-deep-learning architecture used in classification and segmentation tasks.
Figure 2
Figure 2
Multiscale receptive fields of deep-feature maps that are used to activate the visual semantics and their contexts. Multiscale representations help in better segmenting the objects by combining low-level and high-level representations.
Figure 3
Figure 3
Multiscale CNN, defined as a network with multiple distinct CNN networks with various contextual input sizes that run concurrently, whereby the outputs are combined at the end of the network to obtain rich multiscale semantic features.
Figure 4
Figure 4
The spatial-pyramid-pooling module extracts information from different scales that varies among different subregions. Using a four-level pyramid, the pooling kernels cover the whole, half, and small portions of the image. A more powerful representation could be fused with information from the different subregions within these receptive fields.
Figure 5
Figure 5
Multilevel spatial bin, with the example of bin-size-6 resultant feature maps segmented into 6 × 6 subsets.
Figure 6
Figure 6
In ASSP, the atrous convolution uses a parameter called the dilation rate that adjusts the field of view to allow a wider receptive field for better semantic-segmentation results. By increasing the dilation rate at each block, the spatial resolution can be preserved, and a deeper network can be built by capturing features at multiple scales.
Figure 7
Figure 7
In early fusion, all local attributes (shapes and colors) are retrieved from identical regions and locally concatenated before encoding. In late fusion, image representations are derived independently for each attribute and concatenated afterward.
Figure 8
Figure 8
Feature-pyramid-network (FPN) model that combines low- and high-resolution features via a top-down pathway to enrich semantic features at all levels.

Similar articles

Cited by

References

    1. Gao K., Niu S., Ji Z., Wu M., Chen Q., Xu R., Yuan S., Fan W., Chen Y., Dong J. Double-Branched and Area-Constraint Fully Convolutional Networks for Automated Serous Retinal Detachment Segmentation in SD-OCT Images. Comput. Methods Programs Biomed. 2019;176:69–80. doi: 10.1016/j.cmpb.2019.04.027. - DOI - PubMed
    1. Teng L., Li H., Karim S. DMCNN: A Deep Multiscale Convolutional Neural Network Model for Medical Image Segmentation. J. Healthc. Eng. 2019;2019:8597606. doi: 10.1155/2019/8597606. - DOI - PMC - PubMed
    1. Sermanet P., Lecun Y. Traffic Sign Recognition with Multi-Scale Convolutional Networks; Proceedings of the International Joint Conference on Neural Networks; San Jose, CA, USA. 31 July–5 August 2011; pp. 2809–2813. - DOI
    1. Buyssens P., Elmoataz A., Lézoray O. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) Volume 7725 LNCS. Springer; Berlin/Heidelberg, Germany: 2013. Multiscale Convolutional Neural Networks for Vision–Based Classification of Cells; pp. 342–352. - DOI
    1. Zamri N.F.M., Tahir N.M., Ali M.S.A.M., Ashar N.D.K., Al-misreb A.A. Mini-Review of Street Crime Prediction and Classification Methods. J. Kejuruter. 2021;33:391.

LinkOut - more resources