Neuroscientists are striving to give a voice to people unable to speak in a fast-advancing quest to harness brainwaves to restore or enhance physical abilities.
Researchers at universities across California and companies, such as New York-based Precision Neuroscience, are among those making headway toward generating naturalistic speech through a combination of brain implants and artificial intelligence.
Investment and attention have long been focused on implants that enable severely disabled people to operate computer keyboards, control robotic arms, or regain some use of their own paralyzed limbs. But some labs are making strides by concentrating on technology that converts thought patterns into speech.
“We are making great progress—and making brain-to-synthetic voice as fluent as chat between two speaking people is a major goal,” said Edward Chang, a neurosurgeon at the University of California, San Francisco. “The AI algorithms we are using are getting faster, and we are learning with every new participant in our studies.”
Chang and colleagues, including from the University of California, Berkeley last month published a paper in Nature Neuroscience detailing their work with a woman with quadriplegia, or paralysis of the limbs and torso, who had not been able to speak for 18 years after suffering a stroke.
She trained a deep-learning neural network by silently attempting to say sentences composed using 1,024 different words. The audio of her voice was created by streaming her neural data to a joint speech synthesis and text-decoding model.
The technique reduced the lag between the patient’s brain signals and the resultant audio from the eight seconds the group had achieved previously to one second. This is much closer to the 100-200 millisecond time gap in normal speech. The system’s median decoding speed was 47.5 words per minute, or about a third the rate of normal conversation.