Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2021 Apr 29:15:642251.
doi: 10.3389/fnins.2021.642251. eCollection 2021.

Decoding Covert Speech From EEG-A Comprehensive Review

Affiliations
Review

Decoding Covert Speech From EEG-A Comprehensive Review

Jerrin Thomas Panachakel et al. Front Neurosci. .

Abstract

Over the past decade, many researchers have come up with different implementations of systems for decoding covert or imagined speech from EEG (electroencephalogram). They differ from each other in several aspects, from data acquisition to machine learning algorithms, due to which, a comparison between different implementations is often difficult. This review article puts together all the relevant works published in the last decade on decoding imagined speech from EEG into a single framework. Every important aspect of designing such a system, such as selection of words to be imagined, number of electrodes to be recorded, temporal and spatial filtering, feature extraction and classifier are reviewed. This helps a researcher to compare the relative merits and demerits of the different approaches and choose the one that is most optimal. Speech being the most natural form of communication which human beings acquire even without formal education, imagined speech is an ideal choice of prompt for evoking brain activity patterns for a BCI (brain-computer interface) system, although the research on developing real-time (online) speech imagery based BCI systems is still in its infancy. Covert speech based BCI can help people with disabilities to improve their quality of life. It can also be used for covert communication in environments that do not support vocal communication. This paper also discusses some future directions, which will aid the deployment of speech imagery based BCI for practical applications, rather than only for laboratory experiments.

Keywords: brain-computer interfaces (BCI); covert speech; electroencephalogram (EEG); imagined speech; inner speech; neurorehabilitation; speech imagery.

PubMed Disclaimer

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Figures

Figure 1
Figure 1
Distribution of the modalities used in the literature on decoding imagined speech. “Others” include functional magnetic resonance imaging (fMRI), functional near-infrared spectroscopy (fNIRS), intracortical electroencephalography (ICE) etc.
Figure 2
Figure 2
Flowchart detailing the database searches, the number of abstracts screened, the criteria applied for screening the papers, and the full texts retrieved. The number of records in each stage is given within parenthesis.
Figure 3
Figure 3
Various steps involved in the development of a system for decoding imagined speech from EEG. This paper is organized in the same order as above.
Figure 4
Figure 4
Simplified representation of dual stream prediction model (DSPM) for imagined speech. The dorsal stream is in yellow boxes, whereas the ventral stream is in blue boxes. The red circle represents the truncation of information at primary motor cortex in the case of speech imagery. pSTG, posterior superior temporal gyrus; STS, superior temporal sulcus. The primary auditory cortex lies in the superior temporal gyrus and extends into Heschl's gyri. Though Heschl's gyri is involved in speech perception, the region is not activated during speech imagery.
Figure 5
Figure 5
Graph showing the number of electrodes used for data acquisition in various works on decoding imagined speech from EEG. X and Y-axes represent the number of electrodes and articles, respectively.
Figure 6
Figure 6
Graph showing the sampling rates used for data acquisition by the various works in the literature on decoding imagined speech from EEG. X-axis gives the sampling rates and Y-axis gives the number of articles using each specific sampling frequency.
Figure 7
Figure 7
A typical experimental setup used for recording EEG during speech imagery. The subject wears an EEG electrode cap. A monitor cues the subject on the prompt that must be imagined speaking. An optional chin rest prevents artifacts due to unintentional head movements. Figure adapted with permission from Prof. Supratim Ray, Centre for Neuroscience, Indian Institute of Science, Bangalore.
Figure 8
Figure 8
Comparison of the popularity of frequency bands used in works on decoding imagined speech from EEG. Darker shades of black represent more popular frequency bands. Common EEG frequency bands are given in different colors.
Figure 9
Figure 9
Comparison of popular machine learning algorithms used for decoding imagined speech from EEG. The x-axis gives the number of articles using each algorithm.

Similar articles

Cited by

References

    1. Abdulkader S. N., Atia A., Mostafa M.-S. M. (2015). Brain computer interfacing: applications and challenges. Egypt. Inform. J. 16, 213–230. 10.1016/j.eij.2015.06.002 - DOI
    1. Abe K., Takahashi T., Takikawa Y., Arai H., Kitazawa S. (2011). Applying independent component analysis to detect silent speech in magnetic resonance imaging signals. Eur. J. Neurosci. 34, 1189–1199. 10.1111/j.1460-9568.2011.07856.x - DOI - PubMed
    1. Abramson M., Goldinger S. D. (1997). What the reader's eye tells the mind's ear: silent reading activates inner speech. Percept. Psychophys. 59, 1059–1068. 10.3758/BF03205520 - DOI - PubMed
    1. Agarap A. F. (2018). Deep learning using rectified linear units (ReLU). arXiv preprint arXiv:1803.08375.
    1. Alderson-Day B., Weis S., McCarthy-Jones S., Moseley P., Smailes D., Fernyhough C. (2015). The brain's conversation with itself: neural substrates of dialogic inner speech. Soc. Cogn. Affect. Neurosci. 11, 110–120. 10.1093/scan/nsv094 - DOI - PMC - PubMed

LinkOut - more resources