2019
DOI: 10.1101/708206
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Machine translation of cortical activity to text with an encoder-decoder framework

Abstract: AbstractA decade after the first successful attempt to decode speech directly from human brain signals, accuracy and speed remain far below that of natural speech or typing. Here we show how to achieve high accuracy from the electrocorticogram at natural-speech rates, even with few data (on the order of half an hour of spoken speech). Taking a cue from recent advances in machine translation and automatic speech recognition, we train a recurrent neural network to map neural sign… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
103
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 60 publications
(106 citation statements)
references
References 30 publications
2
103
0
1
Order By: Relevance
“…A recent study found that AI can transform a person’s brain waves recorded during speech production into real text. Makin et al (2020) studied four people with epilepsy who underwent brain surgery and had implanted electrodes directly over the inferior frontal cortices where words and speech are produced. The four epilepsy patients read sentences aloud, and brain signals from intracranial electrodes recorded during speech production were the inputs into an encoder recurrent neural network often used for data containing temporal information such as spontaneous brain activity.…”
Section: Acknowledgementsmentioning
confidence: 99%
“…A recent study found that AI can transform a person’s brain waves recorded during speech production into real text. Makin et al (2020) studied four people with epilepsy who underwent brain surgery and had implanted electrodes directly over the inferior frontal cortices where words and speech are produced. The four epilepsy patients read sentences aloud, and brain signals from intracranial electrodes recorded during speech production were the inputs into an encoder recurrent neural network often used for data containing temporal information such as spontaneous brain activity.…”
Section: Acknowledgementsmentioning
confidence: 99%
“…To date, it has been considered in relatively few BCI studies, as the field has favored MI, P300 or SSVEP paradigms (see [ 8 ] for review). However, a DS-BCI offers the possibility of a more naturalistic form of communication [ 9 , 10 , 11 ], relying on neural recordings corresponding to units of language rather than some unrelated brain activity [ 12 , 13 ]. One recent study has demonstrated the potential for spoken sentences to be synthesized from neural activity [ 14 ] and another has shown speech reconstruction directly from the auditory cortex while subjects listened to overt speech [ 15 ].…”
Section: Introductionmentioning
confidence: 99%
“…Following a similar approach, a 'brain-to-text' system was presented in [33] to decode speech from ECoG signals, which was able to achieve word and phone error rates below 25% and 50%, respectively. More recently, in [210], seq2seq models were used to decode speech from ECoG signals, achieving WERs as low as 3%. In [47], a pilot study showed that ECoG recordings from temporal areas can be used to synthesise speech in real time.…”
Section: ) Ssis Based On Brain Activity Signalsmentioning
confidence: 99%