Proceedings of the 44th International Conference on Software Engineering 2022
DOI: 10.1145/3510003.3510162
|View full text |Cite
|
Sign up to set email alerts
|

VarCLR

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 19 publications
(10 citation statements)
references
References 40 publications
0
7
0
Order By: Relevance
“…Previous studies [25,29,44] spend much effort building tremendous datasets to contain various source codes as much as possible. This is because the effectiveness of their model highly depends on the fed training set.…”
Section: Datasetmentioning
confidence: 99%
See 2 more Smart Citations
“…Previous studies [25,29,44] spend much effort building tremendous datasets to contain various source codes as much as possible. This is because the effectiveness of their model highly depends on the fed training set.…”
Section: Datasetmentioning
confidence: 99%
“…Metrics. Unlike previous studies [25,29,44], that focus on specific types of optimizations, DeGPT involves several optimizations, including structure simplification, appending comments, and variable renaming. Moreover, some optimizations are adopted for the first time on the output of the decompiler.…”
Section: Datasetmentioning
confidence: 99%
See 1 more Smart Citation
“…Machine learning models are widely-used in binary program analysis tasks (Pei et al, 2020;2021a;Jin et al, 2022;Chen et al, 2022b;Xu et al, 2023b;a;Wang et al, 2022). However, these models are typically designed for specific downstream tasks such as binary code similarity detection (Pei et al, 2020;Xu et al, 2023a;Wang et al, 2022), variable name prediction (Chen et al, 2022b;Xu et al, 2023b), and binary code type inference (Pei et al, 2021a). In contrast, Nova + is a pre-trained binary code model that can be generalized to various downstream tasks, and it is shown outperforming the existing state-of-the-art techniques in three downstream tasks.…”
Section: Binary Code Modelsmentioning
confidence: 99%
“…Recent studies have shown that state-of-the-art models heavily rely on variables [13,28], specific tokens [29], and even structures [30]. Chen et al [31] focus on semantic representations of program variables, and study how well models can learn similarity between variables that have similar meaning (e.g., minimum and minimal). Ding et al [32] explore the problem of learning functional similarities (and dissimilarities) between codes, towards which they rename variables to inject variable-misuse bugs in order to generate buggy programs that are structurally similar to benign ones.…”
Section: Related Workmentioning
confidence: 99%