2020
DOI: 10.48550/arxiv.2010.04303
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

How Can Self-Attention Networks Recognize Dyck-n Languages?

Abstract: We focus on the recognition of Dyck-n (D n ) languages with self-attention (SA) networks, which has been deemed to be a difficult task for these networks. We compare the performance of two variants of SA, one with a starting symbol (SA + ) and one without (SA − ). Our results show that SA + is able to generalize to longer sequences and deeper dependencies. For D 2 , we find that SA − completely breaks down on long sequences whereas the accuracy of SA + is 58.82%. We find attention maps learned by SA + to be am… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 17 publications
0
0
0
Order By: Relevance
“…Connections between LSTMs and counter automata have also been established empirically (Suzgun et al, 2019a) and theoretically (Merrill et al, 2020). More recently, multiple works have investigated the ability of Transformers to recognize various regular, context-free (Ebrahimi et al, 2020;Yao et al, 2021;Bhattamishra et al, 2020b), and mildly context-sensitive languages (Wang, 2021).…”
Section: H Additional Related Workmentioning
confidence: 99%
“…Connections between LSTMs and counter automata have also been established empirically (Suzgun et al, 2019a) and theoretically (Merrill et al, 2020). More recently, multiple works have investigated the ability of Transformers to recognize various regular, context-free (Ebrahimi et al, 2020;Yao et al, 2021;Bhattamishra et al, 2020b), and mildly context-sensitive languages (Wang, 2021).…”
Section: H Additional Related Workmentioning
confidence: 99%