2023
DOI: 10.1109/tse.2023.3243522
|View full text |Cite
|
Sign up to set email alerts
|

Black-Box Testing of Deep Neural Networks through Test Case Diversity

Abstract: Deep Neural Networks (DNNs) have been extensively used in many areas including image processing, medical diagnostics and autonomous driving. However, DNNs can exhibit erroneous behaviours that may lead to critical errors, especially when used in safety-critical systems. Inspired by testing techniques for traditional software systems, researchers have proposed neuron coverage criteria, as an analogy to source code coverage, to guide the testing of DNNs. Despite very active research on DNN coverage, several rece… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

1
16
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 24 publications
(17 citation statements)
references
References 86 publications
1
16
0
Order By: Relevance
“…This adaptive selection strategy for test cases has proven effective for numerical input spaces 13 . For the input space of DL systems, its applicability was also confirmed by some recent studies 20‐22 …”
Section: Lightweight Art For DL Systemssupporting
confidence: 53%
See 2 more Smart Citations
“…This adaptive selection strategy for test cases has proven effective for numerical input spaces 13 . For the input space of DL systems, its applicability was also confirmed by some recent studies 20‐22 …”
Section: Lightweight Art For DL Systemssupporting
confidence: 53%
“…13 For the input space of DL systems, its applicability was also confirmed by some recent studies. [20][21][22] Diversity, as a fundamental principle of test case selection, 14 can be expressed in various ways. Categorization is one such approach.…”
Section: The Strategy Of Test Case Selectionmentioning
confidence: 99%
See 1 more Smart Citation
“…Furthermore, the performance of most coverage measures is assessed using adversarial inputs, thus focussing on the robustness of the model instead of correctness. Irrespective of the claimed sensitivity of these measures to adversarial inputs, studies have failed to find a significant correlation between these coverage measures and their fault detection capability [5,18,48].…”
mentioning
confidence: 99%
“…Another widely used adequacy measure for traditional and AIbased systems is test suite diversity computed on test inputs or outputs [5,13,29,50]. The diversity metrics are designed based on the intuition that similar test cases exercise similar parts of the source code or training examples, thus revealing the same faults.…”
mentioning
confidence: 99%