Brain-to-speech decoding will require linguistic and pragmatic data
- PMID: 30256217
- DOI: 10.1088/1741-2552/aae466
Brain-to-speech decoding will require linguistic and pragmatic data
Abstract
Objective: Advances in electrophysiological methods such as electrocorticography (ECoG) have enabled researchers to decode phonemes, syllables, and words from brain activity. The ultimate aspiration underlying these efforts is the development of a brain-machine interface (BMI) that will enable speakers to produce real-time, naturalistic speech. In the effort to create such a device, researchers have typically followed a bottom-up approach whereby low-level units of language (e.g. phonemes, syllables, or letters) are decoded from articulation areas (e.g. premotor cortex) with the aim of assembling these low-level units into words and sentences.
Approach: In this paper, we recommend that researchers supplement the existing bottom-up approach with a novel top-down approach. According to the top-down proposal, initial decoding of top-down information may facilitate the subsequent decoding of downstream representations by constraining the hypothesis space from which low-level units are selected.
Main results: We identify types and sources of top-down information that may crucially inform BMI decoding ecosystems: communicative intentions (e.g. speech acts), situational pragmatics (e.g. recurrent communicative pressures), and formal linguistic data (e.g. syntactic rules and constructions, lexical collocations, speakers' individual speech histories).
Significance: Given the inherently interactive nature of communication, we further propose that BMIs be entrained on neural responses associated with interactive dialogue tasks, as opposed to the typical practice of entraining BMIs with non-interactive presentations of language stimuli.
Similar articles
-
Decoding speech using the timing of neural signal modulation.Annu Int Conf IEEE Eng Med Biol Soc. 2016 Aug;2016:1532-1535. doi: 10.1109/EMBC.2016.7591002. Annu Int Conf IEEE Eng Med Biol Soc. 2016. PMID: 28268618
-
Decoding spoken phonemes from sensorimotor cortex with high-density ECoG grids.Neuroimage. 2018 Oct 15;180(Pt A):301-311. doi: 10.1016/j.neuroimage.2017.10.011. Epub 2017 Oct 7. Neuroimage. 2018. PMID: 28993231 Free PMC article.
-
Decoding articulatory and phonetic components of naturalistic continuous speech from the distributed language network.J Neural Eng. 2023 Aug 14;20(4). doi: 10.1088/1741-2552/ace9fb. J Neural Eng. 2023. PMID: 37487487
-
Brain-Computer Interface: Applications to Speech Decoding and Synthesis to Augment Communication.Neurotherapeutics. 2022 Jan;19(1):263-273. doi: 10.1007/s13311-022-01190-2. Epub 2022 Jan 31. Neurotherapeutics. 2022. PMID: 35099768 Free PMC article. Review.
-
Linguistic signs in action: The neuropragmatics of speech acts.Brain Lang. 2023 Jan;236:105203. doi: 10.1016/j.bandl.2022.105203. Epub 2022 Dec 2. Brain Lang. 2023. PMID: 36470125 Free PMC article. Review.
Cited by
-
Optimising non-invasive brain-computer interface systems for free communication between naïve human participants.Sci Rep. 2019 Dec 10;9(1):18705. doi: 10.1038/s41598-019-55166-y. Sci Rep. 2019. PMID: 31822715 Free PMC article.
-
Decoding spoken English from intracortical electrode arrays in dorsal precentral gyrus.J Neural Eng. 2020 Nov 25;17(6):066007. doi: 10.1088/1741-2552/abbfef. J Neural Eng. 2020. PMID: 33236720 Free PMC article.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources
Other Literature Sources