Clark School Researchers Help Decode Speech Recognition

Clark School Researchers Help Decode Speech Recognition

Clark School Researchers Help Decode Speech Recognition

Participants listened to words (acoustic waveform, top left), while neural signals were recorded from cortical surface electrode arrays (top right, red circles) implanted over superior and middle temporal gyrus (STG, MTG). Speech-induced cortical field potentials (bottom right, gray curves) recorded at multiple electrode sites were used to fit multi-input, multi-output models for offline decoding. The models take as input time-varying neural signals at multiple electrodes and output a spectrogram consisting of time-varying spectral power across a range of acoustic frequencies (180–7,000 Hz, bottom left). To assess decoding accuracy, the reconstructed spectrogram is compared to the spectrogram of the original acoustic waveform. Photo courtesy PLoS Biology.
Participants listened to words (acoustic waveform, top left), while neural signals were recorded from cortical surface electrode arrays (top right, red circles) implanted over superior and middle temporal gyrus (STG, MTG). Speech-induced cortical field potentials (bottom right, gray curves) recorded at multiple electrode sites were used to fit multi-input, multi-output models for offline decoding. The models take as input time-varying neural signals at multiple electrodes and output a spectrogram consisting of time-varying spectral power across a range of acoustic frequencies (180–7,000 Hz, bottom left). To assess decoding accuracy, the reconstructed spectrogram is compared to the spectrogram of the original acoustic waveform. Photo courtesy PLoS Biology.

Clark School Professor Shihab Shamma (electrical and computer engineering/Institute for Systems Research [ISR]), former ISR postdoctoral researcher Stephen David*, and alumnus Nima Mesgarani** (Ph.D. '08, electrical engineering) are three of the authors of a new study on how the human auditory system processes speech, published in the Jan. 31, 2012, edition of PLoS Biology.

”Reconstructing Speech from Human Auditory Cortex” details recent progress made in understanding the human brain's computational mechanisms for decoding speech. The researchers took advantage of rare neurosurgical procedures for the treatment of epilepsy, in which neural activity is measured directly from the brain’s cortical surface—a unique opportunity for characterizing how the human brain performs speech recognition. The recordings helped researchers understand what speech sounds could be reconstructed, or decoded, from higher order brain areas in the human auditory system.

The decoded speech representations allowed readout and identification of individual words directly from brain activity during single trial sound presentations. The results provide insights into higher order neural speech processing and suggest it may be possible to readout intended speech directly from brain activity. Potential applications include devices for those who have lost the ability to speak through illness or injury.

Brian N. Pasley, Helen Wills Neuroscience Institute, University of California Berkeley is the paper’s lead author. In addition to the Clark School-affiliated co-authors, additional co-authors include Robert Knight, University of California San Francisco and University of California Berkeley; Adeen Flinker, University of California Berkeley; Edward Chang, University of California San Francisco; and Nathan Crone, Johns Hopkins University.

* Stephen David is now an assistant professor at Oregon Health & Science University, where he heads the Laboratory of Brain, Hearing, and Behavior in the Oregon Hearing Research Center.

** Nima Mesgarani is currently a postdoctoral researcher in the Neurological Surgery Department of the University of California, San Francisco School of Medicine.

| Read a story about this research in USA Today |

 

Related Articles:
Radio interview with Jonathan Simon on "the cocktail party problem"
Auditory Cortex Study Reveals Cells' "Individuality"

February 6, 2012


Prev   Next

Current Headlines

Saikat Guha Elevated to IEEE Fellow

How Microscopic Metals Could Shift Catalysts Production

Christmas Tree Burning Demonstrations Shed Light on Holiday Safety Practices

Fischell Institute Advances Surgical Innovation for LMICs

Looking A“Head” to the Future of Autonomous Robots

Two Maryland Engineers Named to National Academy of Inventors

Autonomy Summit Explores Potential and Challenges of AI

Trio of Maryland Engineers Named to Forbes’ 30 Under 30 Lists

News Resources

Return to Newsroom

Search News

Archived News

Events Resources

Events Calendar