Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.168
|View full text |Cite
|
Sign up to set email alerts
|

Retrofitting Structure-aware Transformer Language Model for End Tasks

Abstract: We consider retrofitting structure-aware Transformer language model for facilitating end tasks by proposing to exploit syntactic distance to encode both the phrasal constituency and dependency connection into the language model. A middle-layer structural learning strategy is leveraged for structure integration, accomplished with main semantic task training under multi-task learning scheme. Experimental results show that the retrofitted structure-aware Transformer language model achieves improved perplexity, me… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 36 publications
(27 citation statements)
references
References 30 publications
(46 reference statements)
0
10
0
Order By: Relevance
“…Previous works for ABSA unfortunately merely make use of the syntactic dependency edge features (i.e., the tree structure) [18,27,31,54]. Without modeling the syntactic dependency labels attached to the dependency arcs, prior studies are limited by treating all word-word relations in the graph equally [16,19,20,23]. Intuitively, the dependency edges with different labels can reveal the relationship more informatively between target aspect and the crucial clues within context, as exemplified in Fig.…”
Section: Syntax Fusion Layermentioning
confidence: 99%
“…Previous works for ABSA unfortunately merely make use of the syntactic dependency edge features (i.e., the tree structure) [18,27,31,54]. Without modeling the syntactic dependency labels attached to the dependency arcs, prior studies are limited by treating all word-word relations in the graph equally [16,19,20,23]. Intuitively, the dependency edges with different labels can reveal the relationship more informatively between target aspect and the crucial clues within context, as exemplified in Fig.…”
Section: Syntax Fusion Layermentioning
confidence: 99%
“…1 Our code is available at: https://github.com/JRC1995/Continuous-RvNN Nangia & Bowman, 2018;Choi et al, 2018;Maillard et al, 2019;Havrylov et al, 2019;Shen et al, 2019a) and some of these structure-aware methods (Shen et al, 2019a;Qian et al, 2020) also exhibit better systematicity (Fodor & Pylyshyn, 1988). Notably, even contemporary Transformer-based methods (Vaswani et al, 2017) have benefited from structural biases in multiple natural language tasks (Wang et al, 2019;Fei et al, 2020).…”
Section: Proceedings Of the 38 Th International Conference On Machinementioning
confidence: 99%
“…In recent years, Transformers (Vaswani et al, 2017) have also been extended either to better support tree structured inputs (Shiv & Quirk, 2019;Ahmed et al, 2019) or to have a better inductive bias to induce hierarchical structures by constraining self-attention (Wang et al, 2019;Nguyen et al, 2020;Shen et al, 2021) or by pushing intermediate representations to have constituent information (Fei et al, 2020). However, the fundamental capability of Transformers for composing sequences according to their latent structures in a length-generalizable manner is shown to be lacking (Tran et al, 2018;Shen et al, 2019a;Hahn, 2020).…”
Section: Related Workmentioning
confidence: 99%
“…The development of ABSA was motivated by this need for granularity, allowing for the extraction of more detailed insights from textual data. ABSA's relevance extends across various domains [10][11][12][13], from enhancing customer service to refining product features based on consumer feedback, thereby playing a pivotal role in data-driven decision-making processes.…”
Section: Introductionmentioning
confidence: 99%