2021 36th IEEE/ACM International Conference on Automated Software Engineering (ASE) 2021
DOI: 10.1109/ase51524.2021.9678712
|View full text |Cite
|
Sign up to set email alerts
|

Thinking Like a Developer? Comparing the Attention of Humans with Neural Models of Code

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(14 citation statements)
references
References 46 publications
0
13
1
Order By: Relevance
“…Besides our work, there have been other studies that also try to explain the mechanisms of pre-trained models for code [1,30,32,38]. Karmakar and Robbes [24] applied four probing tasks on pretrained code models to investigate whether pre-trained models can learn different aspects of source code such as syntactic, structural, surface-level, and semantic information.…”
Section: Related Work 71 Understanding Pre-trained Models For Codementioning
confidence: 99%
“…Besides our work, there have been other studies that also try to explain the mechanisms of pre-trained models for code [1,30,32,38]. Karmakar and Robbes [24] applied four probing tasks on pretrained code models to investigate whether pre-trained models can learn different aspects of source code such as syntactic, structural, surface-level, and semantic information.…”
Section: Related Work 71 Understanding Pre-trained Models For Codementioning
confidence: 99%
“…Attention studies of neural models of code. Paltenghi & Pradel (2021) have compared the attention weights of neural models of code and developers' visual attention when performing a code summarization task, and found a strong positive correlation on the copy attention mechanism for an instance of a pointer network (Vinyals et al, 2015). Wan et al (2022) and have then shown how the attention weights of pre-trained models on source code capture important properties of the abstract syntax tree of the program.…”
Section: Relation To Existing Workmentioning
confidence: 99%
“…We study four approaches: max, mean, rollout and follow-up attention. Apart from the rollout attention, which has been introduced by Abnar & Zuidema (2020), the other three are either inspired by the work of Paltenghi & Pradel (2021) or a novel contribution of this work, such as the follow-up attention.…”
Section: Extraction Functions For Interaction Matrixmentioning
confidence: 99%
See 2 more Smart Citations