Brain signals translated into speech using artificial intelligence



[ad_1]

Masahiro Fujita in a hospital bed, with a respiratory probe. In front of him, a monitor covered with Japanese panels.

People with paralyzing diseases such as motor neurons often rely on technology to help them talk.Credit: BJ Warnick / Alamy

In an effort to provide a voice for people who can not speak, neuroscientists have devised a device that can turn brain signals into words.

This technology is not yet precise enough for use outside the laboratory, although it can synthesize whole sentences for most intelligible. Its creators have described their speech decoding device in a study1 published on April 24 in Nature.

Scientists previously used artificial intelligence to translate simple words2,3Chethan Pandarinath, a neuroengineering engineer at Emory University in Atlanta, Georgia, co-authored a commentary accompanying the study. "Making the jump from one syllable to one sentence is technically quite difficult and is among the things that make the current work so impressive," he says.

Mapping of movements

Many people who have lost the ability to speak speak using technology that forces them to make small movements to control a cursor that selects letters or words on a screen. The British physicist Stephen Hawking, suffering from motor neuron disease, is a famous example. He used a speech-activated device powered by a cheek muscle, said study leader Edward Chang, a neurosurgeon at the University of California at San Francisco.

As people who use such devices have to type the words letter by letter, these devices can be very slow and produce up to ten words per minute, Chang explains. The natural spoken speech is on average 150 words per minute. "It's the effectiveness of the vocal tract that allows us to do that," he says. Chang and his team decided to model the vocal system when building their decoder.

A hand holding a set of intracranial electrodes: a plastic sheet the size of a palm, studded with metal, attached to wires.

The researchers implanted electrodes similar to these in the skull of participants to record the signals of their brain.Credit: UCSF

The researchers worked with five people who had electrodes implanted on the surface of the brain as part of the treatment for epilepsy. First, the team recorded brain activity as participants read hundreds of sentences out loud. Then, Chang and his colleagues combined these recordings with data from previous experiences that determined how the movements of the tongue, lips, jaw and larynx created sound.

The team formed an in-depth learning algorithm on this data and then integrated the program into their decoder. The device transforms cerebral signals into estimated movements of the vocal tract and transforms these movements into synthetic speech. People who listened to 101 synthesized sentences could average 70% of the words, Chang explains.

Two examples of participants reading a sentence, followed by the synthesized version of the sentence generated from their brain activity.

Download MP3

Credit: Chang lab, Department of Neurosurgery UCSF

In another experiment, the researchers asked a participant to read sentences out loud and then to mimic the same sentences by moving their mouths without producing sound. The sentences synthesized in this test were of lower quality than those created from audible speeches, explains Chang, but the results are always encouraging.

Future intelligible

Stephanie Riès, a neuroscientist at San Diego State University in California, more easily understands the word created by associating brain activity with movements of the vocal tract and translating them into sounds.

But we do not know if the new speech decoder would work with words that people only think, says Amy Orsborn, a neural engineer at the University of Washington in Seattle. "The paper does a very good job of showing that it works for an imitated speech," she says. "But how would it work if someone does not move their mouth?"

Marc Slutzky, a neurologist at Northwestern University in Chicago, Illinois, shares this view and says the decoder's performance leaves room for improvement. He notes that listeners identified the synthesized speech by selecting words from a set of choices; as the number of choices increased, people had more and more trouble understanding words.

The study "is a very important step, but there is still a long way to go before the synthesized speech is easily understood," says Slutzky.

Sign up for the everyday Nature Briefing email

Stay abreast of what matters in science and why hand-picked Nature and other publications around the world.

S & # 39; register

[ad_2]

Source link