2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2017
DOI: 10.1109/asru.2017.8268972
|View full text |Cite
|
Sign up to set email alerts
|

Unwritten languages demand attention too! Word discovery with encoder-decoder models

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2

Relationship

3
4

Authors

Journals

citations
Cited by 15 publications
(23 citation statements)
references
References 21 publications
0
23
0
Order By: Relevance
“…Different from these related works and inspired by [9], this paper presented word segmentation from speech, in a bilingual setup and for a real language documentation scenario (Mboshi). The proposed approach first performs AUD to generate pseudophones from speech, and then uses these units in an encoderdecoder NMT for word segmentation.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Different from these related works and inspired by [9], this paper presented word segmentation from speech, in a bilingual setup and for a real language documentation scenario (Mboshi). The proposed approach first performs AUD to generate pseudophones from speech, and then uses these units in an encoderdecoder NMT for word segmentation.…”
Section: Discussionmentioning
confidence: 99%
“…However, there is no similar constraint for the source symbols, as discussed by [5]. Rather than enforcing additional constraints on the alignments, as in the latter reference, we propose to reverse the architecture and to translate from WRL words into UL symbols, following [9]. This "reverse" architecture notably prevents the attention model from ignoring some UL symbols.…”
Section: Word Segmentations From Attentionmentioning
confidence: 99%
See 1 more Smart Citation
“…This semi-supervised task lies between speech translation and keyword spotting, with cross-lingual supervision being used for word segmentation [30,31,32,33]. Bilingual setups for word segmentation were discussed by [34,35,36,37], but applied to speech transcripts (true phones). Among the most relevant to our approach are the works of [24] on speech-to-translation alignment using attentional Neural Machine Translation (NMT) and of [31,32] for language documentation.…”
Section: Related Workmentioning
confidence: 99%
“…• Neural Segmentation (bilingual): the method applied in this paper was presented in [37]. It post-processes a NMT system's soft-alignment probability matrices to generate hard segmentation.…”
Section: Unsupervised Word Discovery Experimentsmentioning
confidence: 99%