2020
DOI: 10.48550/arxiv.2010.03763
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Assessing Phrasal Representation and Composition in Transformers

Abstract: Deep transformer models have pushed performance on NLP tasks to new limits, suggesting sophisticated treatment of complex linguistic inputs, such as phrases. However, we have limited understanding of how these models handle representation of phrases, and whether this reflects sophisticated composition of phrase meaning like that done by humans. In this paper, we present systematic analysis of phrasal representations in state-of-the-art pre-trained transformers. We use tests leveraging human judgments of phrase… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 45 publications
0
1
0
Order By: Relevance
“…The handling of idiomaticity is an important part of natural language processing, due to the ubiquity of idiomatic multiword expressions (MWEs) in natural language (Sag et al, 2002). As such, it is an area where the performance of state-of-the-art Transformer-based models has been investigated (Yu and Ettinger, 2020;Garcia et al, 2021b;Nandakumar et al, 2019), with the general finding being that, through pre-training alone, these models have limited abilities at handling idiomaticity. However, these models are extremely effective at transfer learning through finetuning, and thus are able to perform much better on supervised idiomatic tasks (Fakharian and Cook, 2021;Kurfalı and Östling, 2020), where significant amounts of labelled data is provided.…”
Section: Introductionmentioning
confidence: 99%
“…The handling of idiomaticity is an important part of natural language processing, due to the ubiquity of idiomatic multiword expressions (MWEs) in natural language (Sag et al, 2002). As such, it is an area where the performance of state-of-the-art Transformer-based models has been investigated (Yu and Ettinger, 2020;Garcia et al, 2021b;Nandakumar et al, 2019), with the general finding being that, through pre-training alone, these models have limited abilities at handling idiomaticity. However, these models are extremely effective at transfer learning through finetuning, and thus are able to perform much better on supervised idiomatic tasks (Fakharian and Cook, 2021;Kurfalı and Östling, 2020), where significant amounts of labelled data is provided.…”
Section: Introductionmentioning
confidence: 99%