In a medical first, researchers harnessed the brainwaves of a paralyzed man unable to speak and turned what he intended to say into sentences on a computer screen.
It will take years of additional research but the study, reported Wednesday, marks an important step toward one day restoring more natural communication for people who can’t talk because of injury or illness.
“Most of us take for granted how easily we communicate through speech,” said Dr Edward Chang, a neurosurgeon at the University of California, San Francisco, who led the work. “It’s exciting to think we’re at the very beginning of a new chapter, a new field” to ease the devastation of patients who have lost that ability.
Today, people who can’t speak or write because of paralysis have very limited ways of communicating. For example, the man in the experiment, who was not identified to protect his privacy, uses a pointer attached to a baseball cap that lets him move his head to touch words or letters on a screen. Other devices can pick up patients’ eye movements. But it’s a frustratingly slow and limited substitution for speech.
In recent years, experiments with mind-controlled prosthetics have allowed paralyzed people to shake hands or take a drink using a robotic arm – they imagine moving and those brain signals are relayed through a computer to the artificial limb.
Chang’s team built on that work to develop a “speech neuroprosthetic” – a device that decodes the brainwaves that normally control the vocal tract, the tiny muscle movements of the lips, jaw, tongue and larynx that form each consonant and vowel.
The man who volunteered to test the device was in his late 30s. Fifteen years ago he suffered a brain-stem stroke that caused widespread paralysis and robbed him of speech. The researchers implanted electrodes on the surface of the man’s brain, over the area that controls speech.
A computer analyzed the patterns when he attempted to say common words such as “water” or “good”, eventually learning to differentiate between 50 words that could generate more than 1,000 sentences.
Prompted with such questions as “How are you today?” or “Are you thirsty” the device allowed the man to answer “I am very good” or “No I am not thirsty” – not voicing the words but translating them into text, the team reported in the New England Journal of Medicine.
It takes about three to four seconds for the word to appear on the screen after the man tries to say it, said lead author David Moses, an engineer in Chang’s lab. That’s not nearly as fast as speaking, but quicker than tapping out a response.
In an accompanying editorial, Harvard neurologists Leigh Hochberg and Sydney Cash called the work a “pioneering demonstration.
They suggested improvements but said if the technology pans out it could help people with injuries, strokes or illnesses like Lou Gehrig’s disease whose “brains prepare messages for delivery but those messages are trapped”.
Chang’s lab has spent years mapping the brain activity that leads to speech. First, researchers temporarily placed electrodes in the brains of volunteers undergoing surgery for epilepsy, so they could match brain activity to spoken words.
Only then was it time to try the experiment with someone unable to speak. How did they know the device interpreted the volunteer’s words correctly? They started by having him try to say specific sentences such as “Please bring my glasses” rather than answering open-ended questions until the machine-translated accurately most of the time.
The next steps include improving the device’s speed, accuracy and vocabulary size, and maybe one day allowing users to communicate with a computer-generated voice rather than text on a screen.
According to theguardian.com