2022
DOI: 10.3390/info13020069
|View full text |Cite
|
Sign up to set email alerts
|

A Bidirectional Context Embedding Transformer for Automatic Speech Recognition

Abstract: Transformers have become popular in building end-to-end automatic speech recognition (ASR) systems. However, transformer ASR systems are usually trained to give output sequences in the left-to-right order, disregarding the right-to-left context. Currently, the existing transformer-based ASR systems that employ two decoders for bidirectional decoding are complex in terms of computation and optimization. The existing ASR transformer with a single decoder for bidirectional decoding requires extra methods (such as… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 31 publications
0
3
0
Order By: Relevance
“…The difference from the bidirectional Transformer [ 13 , 14 ] lies in the unrestricted nature of the NLT decoding process in this paper, which can decode in any order without directional constraints. Specifically, while bidirectional LSTMs use both forward and backward information simultaneously during decoding, this paper allows decoding to proceed independently in any direction.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…The difference from the bidirectional Transformer [ 13 , 14 ] lies in the unrestricted nature of the NLT decoding process in this paper, which can decode in any order without directional constraints. Specifically, while bidirectional LSTMs use both forward and backward information simultaneously during decoding, this paper allows decoding to proceed independently in any direction.…”
Section: Methodsmentioning
confidence: 99%
“…Additionally, it adopts a more lightweight model, accelerating training speed and holding significant importance for advancing speech recognition. Subsequently, Transformer decoders have undergone various improvements, such as refining positional information [ 25 ], incorporating positional information into encoder outputs [ 26 ], and enhancing bidirectional semantic information [ 13 , 14 , 27 ]. However, regardless of the type of decoder, there are certain limitations.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation