*Result*: Decoding articulatory and phonetic components of naturalistic continuous speech from the distributed language network.

Title:
Decoding articulatory and phonetic components of naturalistic continuous speech from the distributed language network.
Authors:
Thomas TM; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America., Singh A; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America., Bullock LP; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America., Liang D; Department of Computer Science, Rice University, Houston, TX 77005, United States of America., Morse CW; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America., Scherschligt X; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America., Seymour JP; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Department of Electrical & Computer Engineering, Rice University, Houston, TX 77005, United States of America., Tandon N; Vivian L. Smith Department of Neurosurgery, McGovern Medical School, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Texas Institute for Restorative Neurotechnologies, University of Texas Health Science Center at Houston, Houston, TX 77030, United States of America.; Memorial Hermann Hospital, Texas Medical Center, Houston, TX 77030, United States of America.
Source:
Journal of neural engineering [J Neural Eng] 2023 Aug 14; Vol. 20 (4). Date of Electronic Publication: 2023 Aug 14.
Publication Type:
Journal Article; Research Support, N.I.H., Extramural
Language:
English
Journal Info:
Publisher: Institute of Physics Pub Country of Publication: England NLM ID: 101217933 Publication Model: Electronic Cited Medium: Internet ISSN: 1741-2552 (Electronic) Linking ISSN: 17412552 NLM ISO Abbreviation: J Neural Eng Subsets: MEDLINE
Imprint Name(s):
Original Publication: Bristol, U.K. : Institute of Physics Pub., 2004-
Contributed Indexing:
Keywords: brain–computer interfaces; intracranial; speech decoding; stereo-electroencephalography
Entry Date(s):
Date Created: 20230724 Date Completed: 20230815 Latest Revision: 20251011
Update Code:
20260130
DOI:
10.1088/1741-2552/ace9fb
PMID:
37487487
Database:
MEDLINE

*Further Information*

*Objective.The speech production network relies on a widely distributed brain network. However, research and development of speech brain-computer interfaces (speech-BCIs) has typically focused on decoding speech only from superficial subregions readily accessible by subdural grid arrays-typically placed over the sensorimotor cortex. Alternatively, the technique of stereo-electroencephalography (sEEG) enables access to distributed brain regions using multiple depth electrodes with lower surgical risks, especially in patients with brain injuries resulting in aphasia and other speech disorders.Approach.To investigate the decoding potential of widespread electrode coverage in multiple cortical sites, we used a naturalistic continuous speech production task. We obtained neural recordings using sEEG from eight participants while they read aloud sentences. We trained linear classifiers to decode distinct speech components (articulatory components and phonemes) solely based on broadband gamma activity and evaluated the decoding performance using nested five-fold cross-validation.Main Results.We achieved an average classification accuracy of 18.7% across 9 places of articulation (e.g. bilabials, palatals), 26.5% across 5 manner of articulation (MOA) labels (e.g. affricates, fricatives), and 4.81% across 38 phonemes. The highest classification accuracies achieved with a single large dataset were 26.3% for place of articulation, 35.7% for MOA, and 9.88% for phonemes. Electrodes that contributed high decoding power were distributed across multiple sulcal and gyral sites in both dominant and non-dominant hemispheres, including ventral sensorimotor, inferior frontal, superior temporal, and fusiform cortices. Rather than finding a distinct cortical locus for each speech component, we observed neural correlates of both articulatory and phonetic components in multiple hubs of a widespread language production network.Significance.These results reveal the distributed cortical representations whose activity can enable decoding speech components during continuous speech through the use of this minimally invasive recording method, elucidating language neurobiology and neural targets for future speech-BCIs.
(© 2023 IOP Publishing Ltd.)*