Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Jun 1;43(8):2683-2692.
doi: 10.1002/hbm.25813. Epub 2022 Feb 25.

Attention module improves both performance and interpretability of four-dimensional functional magnetic resonance imaging decoding neural network

Affiliations

Attention module improves both performance and interpretability of four-dimensional functional magnetic resonance imaging decoding neural network

Zhoufan Jiang et al. Hum Brain Mapp. .

Abstract

Decoding brain cognitive states from neuroimaging signals is an important topic in neuroscience. In recent years, deep neural networks (DNNs) have been recruited for multiple brain state decoding and achieved good performance. However, the open question of how to interpret the DNN black box remains unanswered. Capitalizing on advances in machine learning, we integrated attention modules into brain decoders to facilitate an in-depth interpretation of DNN channels. A four-dimensional (4D) convolution operation was also included to extract temporo-spatial interaction within the fMRI signal. The experiments showed that the proposed model obtains a very high accuracy (97.4%) and outperforms previous researches on the seven different task benchmarks from the Human Connectome Project (HCP) dataset. The visualization analysis further illustrated the hierarchical emergence of task-specific masks with depth. Finally, the model was retrained to regress individual traits within the HCP and to classify viewing images from the BOLD5000 dataset, respectively. Transfer learning also achieves good performance. Further visualization analysis shows that, after transfer learning, low-level attention masks remained similar to the source domain, whereas high-level attention masks changed adaptively. In conclusion, the proposed 4D model with attention module performed well and facilitated interpretation of DNNs, which is helpful for subsequent research.

Keywords: attention module; brain decoding; deep learning; functional magnetic resonance imaging; neuroimaging.

PubMed Disclaimer

Conflict of interest statement

The authors declare no potential conflict of interests.

Figures

FIGURE 1
FIGURE 1
The proposed neural network. (a) The model consists of a 4D convolution layer, four 3D attention modules, and a fully‐connected layer to provide labeled task classes. (b) The attention module, which includes the main branch and an attention branch composed of down‐sample and up‐sample paths, was connected by a shortcut skip
FIGURE 2
FIGURE 2
Performance evaluation on the HCP dataset. (a) The average confusion matrix showed a nice block diagonal architecture. (b) The 3DCNN and 4DCNN comparisons used different frames as input (frames = 7, 11, and 15). In terms of dynamic change over a long range, 4DCNN outperformed. (c) The classification performance with or without the attention module (frame = 15). Decoders with attention and a relatively longer 4D‐kernel performed better
FIGURE 3
FIGURE 3
Visualization of attention masks on the HCP dataset. (a)–(d) Examples show the average focused regions on four attention stages (from low‐level to high‐level) of different tasks (language, motor, and relational). Each of the attention masks was color‐coded with a color gradient indicating the enhancement (positive with red) or diminishment (negative with blue) of the feature maps. [Correction added on March 11, 2022, after first online publication: Figure 3 has been updated to correct the task labels in 3c.]
FIGURE 4
FIGURE 4
Prediction of individual traits. (a) An example showing that the transfer learning model yielded significant predictions of gF. (b) The attention masks from low‐level to high‐level after transfer learning. The focused regions of high‐level change adaptively
FIGURE 5
FIGURE 5
Visualization of attention masks on the BOLD5000 dataset. (a)–(d) Attention masks from low‐level to high‐level after transfer learning. The examples show the attention masks of four participants, which employed LOSO cross‐validation. The masks adaptively change to fit different subjects' brain structures

Similar articles

Cited by

References

    1. Barch, D. M. , Burgess, G. C. , Harms, M. P. , Petersen, S. E. , Schlaggar, B. L. , Corbetta, M. , … Consortium, W. U.‐M. H. (2013). Function in the human connectome: Task‐fMRI and individual differences in behavior. NeuroImage, 80, 169–189. 10.1016/j.neuroimage.2013.05.033 - DOI - PMC - PubMed
    1. Chang, N. , Pyles, J. A. , Marcus, A. , Gupta, A. , Tarr, M. J. , & Aminoff, E. M. (2019). BOLD5000, a public fMRI dataset while viewing 5000 visual images. Scientific Data, 6(1), 49. 10.1038/s41597-019-0052-3 - DOI - PMC - PubMed
    1. Chen, C.‐F. , Kreutz‐Delgado, K. , Sereno, M. I. , & Huang, R.‐S. (2019). Unraveling the spatiotemporal brain dynamics during a simulated reach‐to‐eat task. NeuroImage, 185, 58–71. 10.1016/j.neuroimage.2018.10.028 - DOI - PMC - PubMed
    1. Clark, D. , & Badea, C. (2019). Convolutional regularization methods for 4D, x‐ray CT reconstruction. Medical Imaging 2019: Physics of Medical Imaging. Proceedings of SPIE, San Diego, CA. 10948, 574–585. 10.1117/12.2512816 - DOI
    1. Esteban, O. , Birman, D. , Schaer, M. , Koyejo, O. O. , Poldrack, R. A. , & Gorgolewski, K. J. (2017). MRIQC: Advancing the automatic prediction of image quality in MRI from unseen sites. PLoS One, 12(9), e0184661. 10.1371/journal.pone.0184661 - DOI - PMC - PubMed

Publication types