Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Oct 26:16:1015752.
doi: 10.3389/fnins.2022.1015752. eCollection 2022.

Reconstruction of perceived face images from brain activities based on multi-attribute constraints

Affiliations

Reconstruction of perceived face images from brain activities based on multi-attribute constraints

Xiaoyuan Hou et al. Front Neurosci. .

Abstract

Reconstruction of perceived faces from brain signals is a hot topic in brain decoding and an important application in the field of brain-computer interfaces. Existing methods do not fully consider the multiple facial attributes represented in face images, and their different activity patterns at multiple brain regions are often ignored, which causes the reconstruction performance very poor. In the current study, we propose an algorithmic framework that efficiently combines multiple face-selective brain regions for precise multi-attribute perceived face reconstruction. Our framework consists of three modules: a multi-task deep learning network (MTDLN), which is developed to simultaneously extract the multi-dimensional face features attributed to facial expression, identity and gender from one single face image, a set of linear regressions (LR), which is built to map the relationship between the multi-dimensional face features and the brain signals from multiple brain regions, and a multi-conditional generative adversarial network (mcGAN), which is used to generate the perceived face images constrained by the predicted multi-dimensional face features. We conduct extensive fMRI experiments to evaluate the reconstruction performance of our framework both subjectively and objectively. The results show that, compared with the traditional methods, our proposed framework better characterizes the multi-attribute face features in a face image, better predicts the face features from brain signals, and achieves better reconstruction performance of both seen and unseen face images in both visual effects and quantitative assessment. Moreover, besides the state-of-the-art intra-subject reconstruction performance, our proposed framework can also realize inter-subject face reconstruction to a certain extent.

Keywords: brain decoding; facial expression; functional MRI; multi-conditional generative adversarial network; perceived face reconstruction.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

FIGURE 1
FIGURE 1
Schematic illustration of the perceived face reconstruction framework. (A) Algorithmic workflow. (B) The multi-task deep learning network (MTDLN) module. (C) The multi-conditional generative adversarial network (mcGAN) module.
FIGURE 2
FIGURE 2
Schematic diagram of eight candidate multi-task network architectures for simultaneously classifying facial expression, identity, and gender. The architectures vary at which layer the network is split into three branches, sharing no layers (A), one block (B), two blocks (C), three blocks (D), four blocks (E), five blocks (F), five blocks and one fully-connected layer (G), and five blocks and two fully connected layers (H). Each bar represents one convolutional neural network (CNN) layer. Specifically, the bars with red borders represent layers in the identity branch, the bars with green borders represent layers in the expression branch, and the bars with blue borders represent layers in the gender branch. The architecture (E) surrounded by a dashed line is the best-performing architecture.
FIGURE 3
FIGURE 3
Illustration of multi-conditional generative adversarial network (mcGAN) module structure.
FIGURE 4
FIGURE 4
Representative samples of reconstructed faces from five different feature-extraction models. From top row to bottom row: PCA, VAE (VanRullen and Reddy, 2019), pre-trained VGG-Face, re-trained VGG-Face, multi-task deep learning network (MTDLN), and Ground truth.
FIGURE 5
FIGURE 5
Intra-subject perceived face reconstruction from functional magnetic resonance imaging (fMRI) signals. (A) Reconstruction of seen faces from brain activities in each individual subject. (B) Reconstruction of unseen faces from brain activities in each individual subject. Sub1: subject 1 (top row); Sub2: subject 2 (middle row); Ground truth: original face image stimuli (bottom row).
FIGURE 6
FIGURE 6
Inter-subject face reconstruction from functional magnetic resonance imaging (fMRI) signals. (A) Reconstruction of seen faces from brain activities in both subjects. (B) Reconstruction of unseen faces from brain activities in both subjects. Model2_Sub1: Reconstruction of faces from subject 1’s brain activities using the framework trained with subject 2’s fMRI data (top row); Model1_Sub2: Reconstruction of faces from subject 2’s brain activities using the framework trained with subject 1’s fMRI data (middle row). Ground truth: original face image stimuli (bottom row).
FIGURE 7
FIGURE 7
Reconstruction accuracy of each brain region of interest to facial expression, identity and gender. The two subplots represent performance from brain signals of Subject 1 (A) and Subject 2 (B), respectively. Bars represent group-average accuracy (±SEM across 40 pairs). Black dashed line indicates the chance level reconstruction accuracy of 50%. *Indicates group-level significance p < 0.05.

References

    1. Bao P., She L., McGill M., Tsao D. Y. (2020). A map of object space in primate inferotemporal cortex. Nature 583 103–108. 10.1038/s41586-020-2350-5 - DOI - PMC - PubMed
    1. Benoit A., Caplier A., Durette B., Hérault J. (2010). Using human visual system modeling for bio-inspired low level image processing. Comput. Vis. Image Underst. 114 758–773. 10.1016/j.cviu.2010.01.011 - DOI
    1. Bi H., Li N., Guan H., Lu D., Yang L. (2019). “A multi-scale conditional generative adversarial network for face sketch synthesis,” in Proceeding of the 2019 IEEE international conference on image processing (ICIP) (Piscataway, NJ: IEEE; ), 3876–3880. 10.1109/ICIP.2019.8803629 - DOI
    1. Brouwer G. J., Heeger D. J. (2009). Decoding and reconstructing color from responses in human visual cortex. J. Neurosci. 29 13992–14003. 10.1523/JNEUROSCI.3577-09.2009 - DOI - PMC - PubMed
    1. Bruce V., Young A. (1986). Understanding face recognition. Br. J. Psychol. 77 305–327. 10.1111/j.2044-8295.1986.tb02199.x - DOI - PubMed

LinkOut - more resources