Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Jun 27;9(1):14.
doi: 10.1186/s40708-022-00162-8.

Efficient emotion recognition using hyperdimensional computing with combinatorial channel encoding and cellular automata

Affiliations

Efficient emotion recognition using hyperdimensional computing with combinatorial channel encoding and cellular automata

Alisha Menon et al. Brain Inform. .

Abstract

In this paper, a hardware-optimized approach to emotion recognition based on the efficient brain-inspired hyperdimensional computing (HDC) paradigm is proposed. Emotion recognition provides valuable information for human-computer interactions; however, the large number of input channels (> 200) and modalities (> 3 ) involved in emotion recognition are significantly expensive from a memory perspective. To address this, methods for memory reduction and optimization are proposed, including a novel approach that takes advantage of the combinatorial nature of the encoding process, and an elementary cellular automaton. HDC with early sensor fusion is implemented alongside the proposed techniques achieving two-class multi-modal classification accuracies of > 76% for valence and > 73% for arousal on the multi-modal AMIGOS and DEAP data sets, almost always better than state of the art. The required vector storage is seamlessly reduced by 98% and the frequency of vector requests by at least 1/5. The results demonstrate the potential of efficient hyperdimensional computing for low-power, multi-channeled emotion recognition tasks.

Keywords: Brain-inspired; Emotion recognition; Hardware efficient; Hyperdimensional computing; Memory optimization; Multi-modal sensor fusion; Wearable.

PubMed Disclaimer

Conflict of interest statement

The authors declare that they have no competing interests.

Figures

Fig. 1
Fig. 1
Sensor fusion datapath from electrodes to a fused hypervector for the three-modality emotion recognition system used in AMIGOS with GSR, ECG, and EEG sensor inputs. The sensor inputs are pre-processed into a set of features which are then mapped into the HD space to create a set of spatial encoder (SE) inputs. These vectors are encoded within each modality, and then finally fused together to create one vector representing information from all of the channels. The process is detailed in Sect. 2
Fig. 2
Fig. 2
HDC early fusion detailed architecture for m modalities with the four main blocks: map into HDS, spatial encoder, temporal encoder for n-gram of size n+1, and associative memory. Sensor fusion occurs early in the datapath, directly after the spatial encoder
Fig. 3
Fig. 3
HDC a late fusion and b early fusion architectures for a three-modality emotion recognition system. The late fusion architecture fuses after the temporal encoder, resulting in 3 parallel temporal encoders—one per modality. In comparison, the early fusion architecture fuses before the temporal encoder, resulting in only 1 temporal encoder
Fig. 4
Fig. 4
iM and FP vectors used to map into HDS to generate unique SE vectors per channel for a ‘unoptimized’ with distinct iM and FP vectors for all channels, b ‘iM vectors constant per modality’ with the same iM vectors between different modalities, and c ‘FP constant per feature channel’ with the same FP vectors between channels of the same modality
Fig. 5
Fig. 5
‘Combinatorial pairs’ feature channel vector set generation demonstrated for 7 stored vectors. iM loops through vector bank after exhausting available sequential pairs for FP. Hybrid method follows by burst re-generating the vector bank with rule 90 so that new combinatorial pairs can be formed for more feature channels. Generation of 18 feature channel vector sets using a bank of only 7 vectors is shown
Fig. 6
Fig. 6
Arousal and valence accuracies and required vector storage for the various memory optimization as compared to unoptimized for a AMIGOS and b DEAP data sets. Optimizations include ‘unoptimized’ with distinct iM and FP vectors for all channels, ‘iM vectors constant per modality’ with the same iM vectors between different modalities, ‘FP constant per feature channel’ with the same FP vectors between channels of the same modality, and ‘Rule 90 generation’ with generated FP and iM vectors on top of the previous optimizations
Fig. 7
Fig. 7
Number of unique feature channel vector sets {iM, PFP, NFP} that can be generated, and hence the number of channels that can be encoded, using the combinatorial pair technique as the number of stored vectors increases
Fig. 8
Fig. 8
Vector request rate (number of vectors requested per input channel) as the number of vectors stored increases while using rule 90 on a small set of continuously re-populating vectors as compared to only rule 90 for a AMIGOS and b DEAP
Fig. 9
Fig. 9
Average valence and arousal accuracy for the various memory optimizations proposed as the hypervector length decreases for a AMIGOS and b DEAP. The data labels are shown for the most optimized version: rule 90 generation

References

    1. Yin Z, Zhao M, Wang Y, Yang J, Zhang J. Recognition of emotions using multimodal physiological signals and an ensemble deep learning model. Comput Methods Programs Biomed. 2017;140:93–110. doi: 10.1016/j.cmpb.2016.12.005. - DOI - PubMed
    1. Zeng Z, Pantic M, Roisman GI, Huang TS. A survey of affect recognition methods: audio, visual, and spontaneous expressions. IEEE Trans Pattern Anal Mach Intell. 2008;31(1):39–58. doi: 10.1109/TPAMI.2008.52. - DOI - PubMed
    1. Shojaeilangari S, Yau WY, Nandakumar K, Li J, Teoh EK. Robust representation and recognition of facial emotions using extreme sparse learning. IEEE Trans Image Process. 2015;24(7):2140–2152. doi: 10.1109/TIP.2015.2416634. - DOI - PubMed
    1. Kapoor A, Burleson W, Picard RW. Automatic prediction of frustration. Int J Hum Comput Stud. 2007;65(8):724–736. doi: 10.1016/j.ijhcs.2007.02.003. - DOI
    1. Picard RW, Vyzas E, Healey J. Toward machine emotional intelligence: analysis of affective physiological state. IEEE Trans Pattern Anal Mach Intell. 2001;23(10):1175–1191. doi: 10.1109/34.954607. - DOI