Revolutionizing dementia detection: Leveraging vision and Swin transformers for early diagnosis
- PMID: 38619385
- DOI: 10.1002/ajmg.b.32979
Revolutionizing dementia detection: Leveraging vision and Swin transformers for early diagnosis
Abstract
Dementia, an increasingly prevalent neurological disorder with a projected threefold rise globally by 2050, necessitates early detection for effective management. The risk notably increases after age 65. Dementia leads to a progressive decline in cognitive functions, affecting memory, reasoning, and problem-solving abilities. This decline can impact the individual's ability to perform daily tasks and make decisions, underscoring the crucial importance of timely identification. With the advent of technologies like computer vision and deep learning, the prospect of early detection becomes even more promising. Employing sophisticated algorithms on imaging data, such as positron emission tomography scans, facilitates the recognition of subtle structural brain changes, enabling diagnosis at an earlier stage for potentially more effective interventions. In an experimental study, the Swin transformer algorithm demonstrated superior overall accuracy compared to the vision transformer and convolutional neural network, emphasizing its efficiency. Detecting dementia early is essential for proactive management, personalized care, and implementing preventive measures, ultimately enhancing outcomes for individuals and lessening the overall burden on healthcare systems.
Keywords: CNN; PET scans; Swin transformer; deep learning; dementia; vision transformer.
© 2024 Wiley Periodicals LLC.
References
REFERENCES
-
- Ali, A. M., Benjdira, B., Koubaa, A., El‐Shafai, W., Khan, Z., & Boulila, W. (2023). Vision transformers in image restoration: A survey. Sensors, 23(5), 2385. https://doi.org/10.3390/s23052385
-
- Cai, H., Feng, X., Yin, R., Zhao, Y., Guo, L., Fan, X., & Liao, J. (2023). MIST: Multiple instance learning network based on Swin transformer for whole slide image classification of colorectal adenomas. The Journal of Pathology, 259(2), 125–135. https://doi.org/10.1002/path.6027
-
- Chen, C.‐F. R., Fan, Q., & Panda, R. (2021). Crossvit: Cross‐attention multi‐scale vision transformer for image classification. In Proceedings of the IEEE/CVF International Conference on Computer Vision. IEEE Xplore Publisher. https://doi.org/10.48550/arXiv.2103.14899
-
- Chen, S., Tian, X., Wang, Y., Song, Y., Zhang, Y., Zhao, J., & Chen, J.‐C. (2023). DAEGAN: Generative adversarial network based on dual‐domain attention‐enhanced encoder‐decoder for low‐dose PET imaging. Biomedical Signal Processing and Control, 86, 105197. https://doi.org/10.1016/j.bspc.2023.105189
-
- Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., & Houlsby, N. (2020). An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv:2010.11929. https://doi.org/10.48550/arXiv.2010.11929
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Medical