Technologies

time icon March 28, 2017

Decoding Speech Sounds From The Human Brain For A Communication Neuroprosthetic Device

Technology description

Physicians and scientists at UCSF have developed a method for decoding speech signals from the human brain surface . The researchers mapped the electrical activity patterns in the vSMC that correspond to 57 different consonant-vowel syllables commonly found in American English. This information was generated from patients who had been fitted with high density multi-electrode arrays prior to surgery. The researchers recorded the electrical activity in the vSMC while the patients read aloud a series of consonant-vowel syllables. Comparison with an audio recording revealed that specific electrical activity patterns corresponded with the movements of the lips, tongue, and jaw required to produce certain sounds. These findings have recently been published in the leading scientific journal, Nature [2]. The abstract and a link to the full article can be foundhere.

This method can be used as a platform to develop systems, software, and devices to decode speech. Since the same patterns of vocal tract movement are used in most languages, this technique could be utilized to universally decode speech from any language.

Speech requires the precise movement of the lips, tongue, and jaw in order to produce the wide variety of sounds that comprise any given language. These movements are controlled by a small region on the surface of the brain known as the ventral sensorimotor cortex (vSMC).

Certain neurodegenerative disorders such as Lou Gehrigs’s disease/ALS or multiple sclerosis, along with paralysis due to injury or stroke, can leave individuals unable to speak due to their inability to move the required muscles. In many of these cases, the patients still retain the cognitive ability to compose speech and visualize the muscle movements required to generate that speech. This inability to communicate is particularly acute for a subset of patients suffering from locked-in syndrome (LIS). These individuals are fully conscious but are only, at best, able to move their eyes. The inability of these patients to effectively communicate is compounded by the fact that stroke-related LIS has 5- and 10-year post-onset survival rates of over 80% [1]. Various speech generating devices have been developed to assist LIS patients and others with severe motor-related speech defects. However, none of these devices have been able to offer efficient communication since they are limited by the speed at which patients can select individual letters and/or words by movement of a finger or an eye. Despite these limitations, the market for speech generating devices is projected to grow to $505M by 2018.

Application area

  • Neuroprosthetic device to allow speech generation in patients mute due to paralysis or injury to the larynx
  • Devices to aid in the diagnosis of speech motor disorders (i.e. stuttering, aphasia, dysarthria

Advantages


Allows the automated production of artificial speech
Does not require selection of individual letters and/or words on an external device by the patient, leading to significantly faster communication
Allows effective communication for individuals suffering from locked-in syndrome or other severe motor-related speech defects.
Could be used for real-time speech decoding
Could be used to decode speech from any language

由于技术保密工作限制,技术信息无法完全展现,请通过邮箱或短信联系我们,获取更多技术资料。

More information

Categories
  • Neurology
  • Information technology
Keywords:

audio recording revealed

leading scientific journal

ventral sensorimotor cortex

lou gehrigs’s disease/als

offer efficient communication

下载 PDF 文档


感兴趣

Contact us

知繁业茂-yintrust logo知繁业茂-Branchly Innovation logo 知繁业茂-autmasia logo迈科技 logo