New Technology is a Stepping Stone to a Neural Speech Prosthesis, Researchers Say
A state-of-the-art brain-machine interface created by UC San Francisco neuroscientists can generate natural-sounding synthetic speech by using brain activity to control a virtual vocal tract – an anatomically detailed computer simulation including the lips, jaw, tongue and larynx. The study was conducted in research participants with intact speech, but the technology could one day restore the voices of people who have lost the ability to speak due to paralysis and other forms of neurological damage.
Stroke, traumatic brain injury, and neurodegenerative diseases such as Parkinson’s disease, multiple sclerosis and amyotrophic lateral sclerosis (ALS, or Lou Gehrig’s disease) often result in an irreversible loss of the ability to speak. Some people with severe speech disabilities learn to spell out their thoughts letter-by-letter using assistive devices that track very small eye or facial muscle movements. However, producing text or synthesized speech with such devices is laborious, error-prone, and painfully slow, typically permitting a maximum of 10 words per minute, compared to the 100 to 150 words per minute of natural speech.
The new system being developed in the laboratory of Edward Chang, MD – described April 24, 2019, in Nature – demonstrates that it is possible to create a synthesized version of a person’s voice that can be controlled by the activity of their brain’s speech centers. In the future, this approach could not only restore fluent communication to individuals with severe speech disability, the authors say, but could also reproduce some of the musicality of the human voice that conveys the speaker’s emotions and personality.
“For the first time, this study demonstrates that we can generate entire spoken sentences based on an individual’s brain activity,” said Chang, a professor of neurological surgery and member of the UCSF Weill Institute for Neuroscience. “This is an exhilarating proof of principle that with technology that is already within reach, we should be able to build a device that is clinically viable in patients with speech loss.”
Virtual Vocal Tract Improves Naturalistic Speech Synthesis
The research was led by Gopala Anumanchipalli, PhD, a speech scientist, and Josh Chartier, a bioengineering graduate student in the Chang lab. It builds on a recent study in which the pair described for the first time how the human brain’s speech centers choreograph the movements of the lips, jaw, tongue, and other vocal tract components to produce fluent speech.
From that work, Anumanchipalli and Chartier realized that previous attempts to directly decode speech from brain activity might have met with limited success because these brain regions do not directly represent the acoustic properties of speech sounds, but rather the instructions needed to coordinate the movements of the mouth and throat during speech.
“The relationship between the movements of the vocal tract and the speech sounds that are produced is a complicated one,” Anumanchipalli said. “We reasoned that if these speech centers in the brain are encoding movements rather than sounds, we should try to do the same in decoding those signals.”
In their new study, Anumancipali and Chartier asked five volunteers being treated at the UCSF Epilepsy Center – patients with intact speech who had electrodes temporarily implanted in their brains to map the source of their seizures in preparation for neurosurgery – to read several hundred sentences aloud while the researchers recorded activity from a brain region known to be involved in language production.
Based on the audio recordings of participants’ voices, the researchers used linguistic principles to reverse engineer the vocal tract movements needed to produce those sounds: pressing the lips together here, tightening vocal cords there, shifting the tip of the tongue to the roof of the mouth, then relaxing it, and so on.
This detailed mapping of sound to anatomy allowed the scientists to create a realistic virtual vocal tract for each participant that could be controlled by their brain activity. This comprised two “neural network” machine learning algorithms: a decoder that transforms brain activity patterns produced during speech into movements of the virtual vocal tract, and a synthesizer that converts these vocal tract movements into a synthetic approximation of the participant’s voice.
The synthetic speech produced by these algorithms was significantly better than synthetic speech directly decoded from participants’ brain activity without the inclusion of simulations of the speakers’ vocal tracts, the researchers found. The algorithms produced sentences that were understandable to hundreds of human listeners in crowdsourced transcription tests conducted on the Amazon Mechanical Turk platform.
As is the case with natural speech, the transcribers were more successful when they were given shorter lists of words to choose from, as would be the case with caregivers who are primed to the kinds of phrases or requests patients might utter. The transcribers accurately identified 69 percent of synthesized words from lists of 25 alternatives and transcribed 43 percent of sentences with perfect accuracy. With a more challenging 50 words to choose from, transcribers’ overall accuracy dropped to 47 percent, though they were still able to understand 21 percent of synthesized sentences perfectly.
“We still have a ways to go to perfectly mimic spoken language,” Chartier acknowledged. “We’re quite good at synthesizing slower speech sounds like ‘sh’ and ‘z’ as well as maintaining the rhythms and intonations of speech and the speaker’s gender and identity, but some of the more abrupt sounds like ‘b’s and ‘p’s get a bit fuzzy. Still, the levels of accuracy we produced here would be an amazing improvement in real-time communication compared to what’s currently available.”
Artificial Intelligence, Linguistics, and Neuroscience Fueled Advance
The researchers are currently experimenting with higher-density electrode arrays and more advanced machine learning algorithms that they hope will improve the synthesized speech even further. The next major test for the technology is to determine whether someone who can’t speak could learn to use the system without being able to train it on their own voice and to make it generalize to anything they wish to say.
Learn more: Synthetic Speech Generated from Brain Recordings
The Latest on: Neural speech prosthesis
via Google News
The Latest on: Neural speech prosthesis
- As brain plans movements, middle frontal gyrus is listeningon January 12, 2021 at 8:38 am
A brain-computer interface study reveals one brain region's surprising role in planning movements exclusively in response to sounds.
- neural networkson January 8, 2021 at 4:00 pm
In our modern connected age, our devices have become far more powerful and useful when they could draw upon resources of a global data network. The downside of a cloud-connected device is the risk ...
- Brain-Computer Interfaces: Separating Fact From Fiction On Musk’s Brain Implant Claimson January 6, 2021 at 4:01 pm
Our mind’s eye fills with everything from the Borg and neural interfaces of Star Trek ... This is also because speech isn’t just produced by a single part of the brain, but distributed ...
- Precise sense of touch is still out of reach for bionic hands, say scientistson December 22, 2020 at 1:44 pm
Bionic prosthetic limbs have seen huge technological ... “One problem with current neural electrodes is that you can’t tell during the implantation surgery which part of the nerve corresponds ...
- Study sheds new light on how the brain distinguishes speech from noiseon December 20, 2020 at 9:19 am
acetylcholine modulation enhances neural discrimination of tones from noise stimuli, which may contribute to processing important acoustic signals such as speech. Additionally, they describe novel ...
- Michael Beauchamp Labon December 18, 2020 at 3:49 am
The Beauchamp Lab studies the neural mechanisms for multisensory integration and visual ... such as during conversation when we make use of both the auditory information we hear in spoken speech and ...
- This is your brain on code: Researchers decipher neural mechanics of computer programmingon December 15, 2020 at 10:00 am
But despite our increasing reliance on technology, almost nothing is known about the neural mechanisms of computer programming. "People want to know what makes someone a good programmer," Liu said.
- Vocalizations Return to Paralyzed Man, Via Wireless Brain-Computer Interfaceon December 10, 2020 at 4:00 pm
Our results support the feasibility of neural prostheses that may have the potential to provide near-conversational synthetic speech output for individuals with severely impaired speech motor control.
- Scientists develop AI that can turn brain activity into texton August 5, 2020 at 8:02 pm
At the moment, it works on neural patterns interpreted when someone ... “We are not there yet but we think this could be the basis of a speech prosthesis,” Joseph Makin, co-author of the ...
- New app gives throat cancer patients their voice backon January 11, 2019 at 4:26 am
"I'm not very good at using the voice prosthesis," he added ... advanced statistical models such as artificial neural networks. "You use speech models with certain parameters to generate ...
via Bing News