TY - JOUR
T1 - Structural Test Coverage Criteria for Deep Neural Networks
AU - Sun, Youcheng
AU - Huang, Xiaowei
AU - Kroening, Daniel
AU - Sharp, James
AU - Hill, Matthew
AU - Ashmore, Rob
PY - 2019/10/8
Y1 - 2019/10/8
N2 - Deep neural networks (DNNs) have a wide range of applications, and software employing them must bethoroughly tested, especially in safety-critical domains. However, traditional software test coverage metricscannot be applied directly to DNNs. In this paper, inspired by the MC/DC coverage criterion, we proposea family of four novel test coverage criteria that are tailored to structural features of DNNs and their semantics. We validate the criteria by demonstrating that test inputs that are generated with guidance by ourproposed coverage criteria are able to capture undesired behaviours in a DNN. Test cases are generated using a symbolic approach and a gradient-based heuristic search. By comparing them with existing methods,we show that our criteria achieve a balance between their ability to find bugs (proxied using adversarial examples and correlation with functional coverage) and the computational cost of test input generation. Ourexperiments are conducted on state-of-the-art DNNs obtained using popular open source datasets, includingMNIST, CIFAR-10 and ImageNet.
AB - Deep neural networks (DNNs) have a wide range of applications, and software employing them must bethoroughly tested, especially in safety-critical domains. However, traditional software test coverage metricscannot be applied directly to DNNs. In this paper, inspired by the MC/DC coverage criterion, we proposea family of four novel test coverage criteria that are tailored to structural features of DNNs and their semantics. We validate the criteria by demonstrating that test inputs that are generated with guidance by ourproposed coverage criteria are able to capture undesired behaviours in a DNN. Test cases are generated using a symbolic approach and a gradient-based heuristic search. By comparing them with existing methods,we show that our criteria achieve a balance between their ability to find bugs (proxied using adversarial examples and correlation with functional coverage) and the computational cost of test input generation. Ourexperiments are conducted on state-of-the-art DNNs obtained using popular open source datasets, includingMNIST, CIFAR-10 and ImageNet.
UR - https://pure.qub.ac.uk/en/publications/structural-test-coverage-criteria-for-deep-neural-networks(dd5b518d-3b55-4556-a267-95c6a10d62e9).html
U2 - 10.1145/3358233
DO - 10.1145/3358233
M3 - Article
SN - 1539-9087
JO - ACM Transactions on Embedded Computing Systems
JF - ACM Transactions on Embedded Computing Systems
ER -