2019
DOI: 10.48550/arxiv.1911.08870
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Comparative Study on End-to-end Speech to Text Translation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…There are two main research paradigms for ST, the end-to-end model, and the cascaded system (Sperber and Paulik, 2020;nie, 2019). End-to-end ST Previous works (Bérard et al, 2016;Duong et al, 2016) have proved the potential for end-to-end ST, which has attracted intensive attentions (Vila et al, 2018;Salesky et al, 2018Salesky et al, , 2019bDi Gangi et al, 2019a;Bahar et al, 2019a;Di Gangi et al, 2019b;Inaguma et al, 2020). It's proved that pre-training (Weiss et al, 2017;Bérard et al, 2018;Bansal et al, 2018;Stoian et al, 2020) and multi-task learning (Vydana et al, 2020) can significantly improve the performance.…”
Section: Related Workmentioning
confidence: 99%
“…There are two main research paradigms for ST, the end-to-end model, and the cascaded system (Sperber and Paulik, 2020;nie, 2019). End-to-end ST Previous works (Bérard et al, 2016;Duong et al, 2016) have proved the potential for end-to-end ST, which has attracted intensive attentions (Vila et al, 2018;Salesky et al, 2018Salesky et al, , 2019bDi Gangi et al, 2019a;Bahar et al, 2019a;Di Gangi et al, 2019b;Inaguma et al, 2020). It's proved that pre-training (Weiss et al, 2017;Bérard et al, 2018;Bansal et al, 2018;Stoian et al, 2020) and multi-task learning (Vydana et al, 2020) can significantly improve the performance.…”
Section: Related Workmentioning
confidence: 99%
“…For speech translation, there are two main research paradigms, the end-to-end model and the cascaded system (Sperber and Paulik, 2020;nie, 2019). End-to-end ST Previous works (Bérard et al, 2016;Duong et al, 2016) have given the first proof of the potential for end-to-end speech-totext translation, which has attracted intensive attentions recently (Vila et al, 2018;Salesky et al, 2018Salesky et al, , 2019bDi Gangi et al, 2019a;Bahar et al, 2019a;Di Gangi et al, 2019b;Inaguma et al, 2020). Many works have proved that pre-training then transferring (Weiss et al, 2017;Bérard et al, 2018;Bansal et al, 2018;Stoian et al, 2020) and multi-task learning (Vydana et al, 2020) can significantly improve the performance of end-to-end models.…”
Section: Related Workmentioning
confidence: 99%
“…Automatic Sign Language Recognition (ASLR) is a challenging task and an active research field with the aim of reducing the dependency of sign language interpreters in the daily lives of the Deaf. Among the many similar problems attempted by deep learning researchers, sign language recognition bears a resemblance to video-based action recognition because of its shared medium of information (Varol et al, 2017), and to speech recognition and machine translation problems (Bahar et al, 2019;Bahdanau et al, 2017), due to its linguistic nature. However, there are certain aspects of ASLR that makes the task more challenging, one of which is the asynchronous multi-articulatory nature of the sign (Sutton-Spence and Woll, 1999).…”
Section: Introductionmentioning
confidence: 99%