Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1548
|View full text |Cite
|
Sign up to set email alerts
|

Linguistically-Informed Self-Attention for Semantic Role Labeling

Abstract: Current state-of-the-art semantic role labeling (SRL) uses a deep neural network with no explicit linguistic features. However, prior work has shown that gold syntax trees can dramatically improve SRL decoding, suggesting the possibility of increased accuracy from explicit modeling of syntax. In this work, we present linguistically-informed self-attention (LISA): a neural network model that combines multi-head self-attention with multi-task learning across dependency parsing, part-ofspeech tagging, predicate d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

9
381
0
2

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 363 publications
(422 citation statements)
references
References 46 publications
9
381
0
2
Order By: Relevance
“…He et al (2018b) propose a k-th order argument pruning algorithm based on systematic dependency trees. Strubell et al (2018) propose a self-attention based neural MTL model which incorporate dependency parsing as a auxiliary task for SRL. propose a MTL framework using hard parameter strategy to incorporate constituent parsing loss into semantic tasks, i.e.…”
Section: Related Workmentioning
confidence: 99%
“…He et al (2018b) propose a k-th order argument pruning algorithm based on systematic dependency trees. Strubell et al (2018) propose a self-attention based neural MTL model which incorporate dependency parsing as a auxiliary task for SRL. propose a MTL framework using hard parameter strategy to incorporate constituent parsing loss into semantic tasks, i.e.…”
Section: Related Workmentioning
confidence: 99%
“…By its nature, the self-attention-based model directly captures the relation between words in the sentence, which is convenient to incorporate syntactic dependency information. Strubell et al (2018) replace one attention head with pre-trained syntactic dependency information, which can be viewed as a hard way to inject syntax into the neural model. Enlightened by the machine translation model proposed by Shaw et al (2018), we introduce the Relation-Aware method to incorporate syntactic dependencies, which is a softer way to encode richer structural information.…”
Section: Introductionmentioning
confidence: 99%
“…Several studies have tried to develop models with better generalization capacities (Yang et al, 2015), (FitzGerald et al, 2015). In recent works, PropBank SRL systems have evolved and span classifier approaches have been replaced by current state of the art sequence tagging models that use recurrent neural networks (He et al, 2017) and neural atten- tion (Tan et al, 2017;Strubell et al, 2018). However, these parsers still suffer performances drops of up to 12 points in F-measure on OOD with respect to ID.…”
Section: Generalization To Propbank Parsingmentioning
confidence: 99%