Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Sep 16;23(18):7927.
doi: 10.3390/s23187927.

Intelligent ADL Recognition via IoT-Based Multimodal Deep Learning Framework

Affiliations

Intelligent ADL Recognition via IoT-Based Multimodal Deep Learning Framework

Madiha Javeed et al. Sensors (Basel). .

Abstract

Smart home monitoring systems via internet of things (IoT) are required for taking care of elders at home. They provide the flexibility of monitoring elders remotely for their families and caregivers. Activities of daily living are an efficient way to effectively monitor elderly people at home and patients at caregiving facilities. The monitoring of such actions depends largely on IoT-based devices, either wireless or installed at different places. This paper proposes an effective and robust layered architecture using multisensory devices to recognize the activities of daily living from anywhere. Multimodality refers to the sensory devices of multiple types working together to achieve the objective of remote monitoring. Therefore, the proposed multimodal-based approach includes IoT devices, such as wearable inertial sensors and videos recorded during daily routines, fused together. The data from these multi-sensors have to be processed through a pre-processing layer through different stages, such as data filtration, segmentation, landmark detection, and 2D stick model. In next layer called the features processing, we have extracted, fused, and optimized different features from multimodal sensors. The final layer, called classification, has been utilized to recognize the activities of daily living via a deep learning technique known as convolutional neural network. It is observed from the proposed IoT-based multimodal layered system's results that an acceptable mean accuracy rate of 84.14% has been achieved.

Keywords: IoT; activities of daily living recognition; deep learning; multimodal data; patient monitoring; smart homes.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
The architecture diagram for multimodal IoT-based deep learning framework via ADL recognition.
Figure 2
Figure 2
Sample signals after filters applied for motion sensor data.
Figure 3
Figure 3
Detailed view of data segmentation applied over the inertial signal has been presented using multiple colors in the figure. The red dotted box shows single segment of data.
Figure 4
Figure 4
(a) Real video frame and (b) extracted human figure after background extraction for bending activity in Berkeley-MHAD dataset.
Figure 5
Figure 5
(a) Human silhouette (b) 2D stick model, where each red dot represents the body point detected, green lines show the upper body skeleton, and orange lines give the lower body skeleton.
Figure 6
Figure 6
Extracted LPCCs result for the Jumping Jacks ADL over Berkeley-MHAD dataset.
Figure 7
Figure 7
Upward motion direction flow in Jumping in Place ADL.
Figure 8
Figure 8
Features optimization via genetic algorithm explained through a detailed view.
Figure 9
Figure 9
Proposed CNN model for multimodal IoT-based ADL recognition over Berkeley-MHAD.
Figure 10
Figure 10
Sample frame sequences from the Berkeley-MHAD [22] dataset.
Figure 11
Figure 11
Sample frame sequences from Opportunity++ [21] dataset.
Figure 12
Figure 12
Examples of problematic ADL activities over Berkeley-MHAD, where red dotted circles point out the skeleton extraction problems.

Similar articles

References

    1. Ali M., Ali A.A., Taha A.-E., Dhaou I.B., Gia T.N. Intelligent Autonomous Elderly Patient Home Monitoring System; Proceedings of the ICC 2019—2019 IEEE International Conference on Communications (ICC); Shanghai, China. 21–23 May 2019; pp. 1–6. - DOI
    1. Madiha J., Ahmad J., Kim K. Wearable Sensors based Exertion Recognition using Statistical Features and Random Forest for Physical Healthcare Monitoring; Proceedings of the 2021 International Bhurban Conference on Applied Sciences and Technologies (IBCAST); Islamabad, Pakistan. 12–16 January 2021; pp. 512–517. - DOI
    1. Zhou X., Zhang L. SA-FPN: An effective feature pyramid network for crowded human detection. Appl. Intell. 2022;52:12556–12568. doi: 10.1007/s10489-021-03121-8. - DOI
    1. Liu Y., Wang K., Liu L., Lan H., Lin L. TCGL: Temporal Contrastive Graph for Self-Supervised Video Representation Learning. IEEE Trans. Image Process. 2022;31:1978–1993. doi: 10.1109/TIP.2022.3147032. - DOI - PubMed
    1. Gaddam A., Mukhopadhyay S.C., Gupta G.S. Trial & experimentation of a smart home monitoring system for elderly; Proceedings of the 2011 IEEE International Instrumentation and Measurement Technology Conference; Hangzhou, China. 9–12 May 2011; pp. 1–6. - DOI

LinkOut - more resources