Proceedings of the 28th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Softw 2020
DOI: 10.1145/3368089.3409671
|View full text |Cite
|
Sign up to set email alerts
|

Correlations between deep neural network model coverage criteria and model quality

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
52
2

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 55 publications
(58 citation statements)
references
References 64 publications
4
52
2
Order By: Relevance
“…d) Difference between Testing and Defect Detection: Recent paper [50] on correlations between coverage criteria and model quality suggests that coverage guided testing complements gradient-based adversarial attack. They discover that adversarial samples found by FNN coverage guided testing can be further utilised to retrain more robust models.…”
Section: H Threats To Validitymentioning
confidence: 99%
“…d) Difference between Testing and Defect Detection: Recent paper [50] on correlations between coverage criteria and model quality suggests that coverage guided testing complements gradient-based adversarial attack. They discover that adversarial samples found by FNN coverage guided testing can be further utilised to retrain more robust models.…”
Section: H Threats To Validitymentioning
confidence: 99%
“…Model testing has also been leveraged for many other domains such as image classification [79,89], automatic speech recognition [90], text classification [74], and machine translation [91,92]. Recently, Yan et al [93] have studied many coverage criteria and measured their correlations with model quality (i.e., model robustness against adversarial attacks), and empirical results show that existing criteria can not faithfully reflect model quality.…”
Section: Effects O F Configurable Parametersmentioning
confidence: 99%
“…DeepXplore [52] proposed a metric called neuron coverage for whitebox testing of DL models and leveraged gradient-based techniques to search for more effective tests. While various other metrics [39,43] have also been proposed recently, the correlation between such metrics and the robustness of models is still unclear [25,36,69]. Meanwhile, there are also a series of work targeting specific applications, such as autonomous driving, including DeepTest [63], DeepRoad [71], and DeepBillboard [78].…”
Section: Related Work DL Modelmentioning
confidence: 99%
“…Due to the popularity of DL models and the critical importance of their reliability, a growing body of research efforts have been dedicated to testing DL models, with focus on adversarial attacks [14,21,32,[46][47][48] for model robustness, the discussion on various metrics for DL model testing [36,39,43,52,69], and testing DL models for specific applications [63,71,78]. Meanwhile, both running and testing DL models inevitably involve the underlying DL libraries, which serve as central pieces of infrastructures for building, training, optimizing and deploying DL models.…”
Section: Introductionmentioning
confidence: 99%