2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU) 2015
DOI: 10.1109/asru.2015.7404814
|View full text |Cite
|
Sign up to set email alerts
|

Using bidirectional lstm recurrent neural networks to learn high-level abstractions of sequential features for automated scoring of non-native spontaneous speech

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
32
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 69 publications
(32 citation statements)
references
References 17 publications
0
32
0
Order By: Relevance
“…In this section, we will briefly overview the preliminary work of bidirectional long short term memory (BiLSTM) network [25] and then address how we can apply it to the EEG feature extraction task.…”
Section: Preliminarymentioning
confidence: 99%
See 1 more Smart Citation
“…In this section, we will briefly overview the preliminary work of bidirectional long short term memory (BiLSTM) network [25] and then address how we can apply it to the EEG feature extraction task.…”
Section: Preliminarymentioning
confidence: 99%
“…However, one shortcoming of conventional LSTM is that it only make use of the previous context. The BiLSTM module is able to process data using two directions with separate hidden layers, respectively [25]. As a result, compared with the traditional LSTM model, BiLSTM can access the long-range context in both input directions, and hence it could be better used to model time sequences.…”
Section: Preliminarymentioning
confidence: 99%
“…This is further illustrated by the fact that for general questions, the speech-only model performed as well as the text-only model. We also note that recent work by Yu et al (2016) used neural networks to learn high-level abstractions from frame-to-frame acoustic properties of the signal and showed that these features provided a very limited gain over the features considered in this study.…”
Section: Discussionmentioning
confidence: 99%
“…For our initial work on investigating the use of automated speech rating in dialog systems, we relaxed the aforementioned constraint in favor of creating a real‐time‐able system. To this end, we implemented a hybrid recurrent neural network framework that comes with minimal manual effort and cost and high scoring accuracy and speed (Yu et al, ).…”
Section: Speech Scoringmentioning
confidence: 99%
“…For our initial work on investigating the use of automated speech rating in dialog systems, we relaxed the aforementioned constraint in favor of creating a real-time-able system. To this end, we implemented a hybrid recurrent neural network framework that comes with minimal manual effort and cost and high scoring accuracy and speed (Yu et al, 2015). In the proposed framework, we used generic time-sequence features extracted directly from the audio input instead of manually designed features, thus saving on human transcription effort and expert knowledge for training and optimizing the speech recognition engine for the rater.…”
Section: Speech Scoringmentioning
confidence: 99%