2019
DOI: 10.1007/s11265-019-01484-3
|View full text |Cite
|
Sign up to set email alerts
|

Spoken Language Understanding of Human-Machine Conversations for Language Learning Applications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(4 citation statements)
references
References 29 publications
0
4
0
Order By: Relevance
“…Based on the provisions of information theory, we argue the solution rule (2) in the context of the relative entropy functional [ 35 , 36 , 37 ] (3): where is the selective probability distribution of the studied (empirical) speech signal relative to the etalon probability distribution , . Assume that the distribution law is normal: , where is a sample matrix of autocorrelation of the speech signal of dimension .…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Based on the provisions of information theory, we argue the solution rule (2) in the context of the relative entropy functional [ 35 , 36 , 37 ] (3): where is the selective probability distribution of the studied (empirical) speech signal relative to the etalon probability distribution , . Assume that the distribution law is normal: , where is a sample matrix of autocorrelation of the speech signal of dimension .…”
Section: Methodsmentioning
confidence: 99%
“…Let , i.e., two competing hypotheses, and , are tested for a priori unknown autocorrelation matrices and . The verification will be performed using the asymptotic minimax criterion of the likelihood ratio [ 35 , 36 , 37 ] based on data from a sample , . Under such conditions, the hypothesis will be considered true if the condition is satisfied, where is the plausibility function of the signal provided that hypothesis is confirmed, and is the plausibility function of the signal .…”
Section: Methodsmentioning
confidence: 99%
“…SLU task aims to extract structured semantic representations from speech signals [3,4]. Conventional cascaded SLU systems consist of an automatic speech recognition (ASR) module as well as a downstream natural language understanding (NLU) module [5]. On the other hand, end-to-end (E2E) SLU systems directly extract users' intentions from input speech to avoid error propagation seen in the above cascaded method [1,6].…”
Section: Introductionmentioning
confidence: 99%
“…These representations are crucial to enable speech as the primary mode of human-computer interaction (HCI) [2]. Traditional SLU solutions rely on the text transcription generated by an automatic speech recognition (ASR) module, followed by a natural language understanding (NLU) system, responsible for extracting semantics from the ASR output [3]. As described in [4], in such scenarios the ASR typically operates on chunks of the incoming speech signal and outputs the transcript for each segment.…”
Section: Introductionmentioning
confidence: 99%