2021
DOI: 10.20944/preprints202101.0081.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What All Do Audio Transformer Models Hear? Probing Acoustic Representations for Language Delivery and Its Structure

Abstract: In recent times, BERT based transformer models have become an inseparable part of the 'tech stack' of text processing models. Similar progress is being observed in the speech domain with a multitude of models observing state-of-the-art results by using audio transformer models to encode speech. This begs the question of what are these audio transformer models learning. Moreover, although the standard methodology is to choose the last layer embedding for any downstream task, but is it the optimal choice? We try… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
references
References 21 publications
0
0
0
Order By: Relevance