ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054759
|View full text |Cite
|
Sign up to set email alerts
|

End-end Speech-to-Text Translation with Modality Agnostic Meta-Learning

Abstract: End-to-end Speech Translation (ST) models have several advantages such as lower latency, smaller model size, and less error compounding over conventional pipelines that combine Automatic Speech Recognition (ASR) and text Machine Translation (MT) models. However, collecting large amounts of parallel data for ST task is more difficult compared to the ASR and MT tasks. Previous studies have proposed the use of transfer learning approaches to overcome the above difficulty. These approaches benefit from weakly supe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
44
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 39 publications
(44 citation statements)
references
References 21 publications
(20 reference statements)
0
44
0
Order By: Relevance
“…This is the only submission to use an end-to-end approach for the speech track. The authors use transformer-based models combining the wait-k strategy with a modality-agnostic meta learning approach (Indurthi et al, 2020) to address data sparsity. They also use the ST task along with ASR and MT as the source task, a minor variation explored compared to the original paper.…”
Section: Submissionsmentioning
confidence: 99%
See 4 more Smart Citations
“…This is the only submission to use an end-to-end approach for the speech track. The authors use transformer-based models combining the wait-k strategy with a modality-agnostic meta learning approach (Indurthi et al, 2020) to address data sparsity. They also use the ST task along with ASR and MT as the source task, a minor variation explored compared to the original paper.…”
Section: Submissionsmentioning
confidence: 99%
“…BHANSS (Lakumarapu et al, 2020) built their end-to-end system adopting the Transformer architecture (Vaswani et al, 2017a) coupled with the meta-learning approach proposed in (Indurthi et al, 2020). Meta-learning is used mitigate the issue of over-fitting when the training data is limited, as in the ST case, and allows their system to take advantage of the available ASR and MT data.…”
Section: Submissionsmentioning
confidence: 99%
See 3 more Smart Citations