Back to Search Start Over

Quantifying safety risks of deep neural networks.

Authors :
Xu, Peipei
Ruan, Wenjie
Huang, Xiaowei
Source :
Complex & Intelligent Systems; Aug2023, Vol. 9 Issue 4, p3801-3818, 18p
Publication Year :
2023

Abstract

Safety concerns on the deep neural networks (DNNs) have been raised when they are applied to critical sectors. In this paper, we define safety risks by requesting the alignment of network's decision with human perception. To enable a general methodology for quantifying safety risks, we define a generic safety property and instantiate it to express various safety risks. For the quantification of risks, we take the maximum radius of safe norm balls, in which no safety risk exists. The computation of the maximum safe radius is reduced to the computation of their respective Lipschitz metrics—the quantities to be computed. In addition to the known adversarial example, reachability example, and invariant example, in this paper, we identify a new class of risk—uncertainty example—on which humans can tell easily, but the network is unsure. We develop an algorithm, inspired by derivative-free optimization techniques and accelerated by tensor-based parallelization on GPUs, to support an efficient computation of the metrics. We perform evaluations on several benchmark neural networks, including ACSC-Xu, MNIST, CIFAR-10, and ImageNet networks. The experiments show that our method can achieve competitive performance on safety quantification in terms of the tightness and the efficiency of computation. Importantly, as a generic approach, our method can work with a broad class of safety risks and without restrictions on the structure of neural networks. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
21994536
Volume :
9
Issue :
4
Database :
Complementary Index
Journal :
Complex & Intelligent Systems
Publication Type :
Academic Journal
Accession number :
167361365
Full Text :
https://doi.org/10.1007/s40747-022-00790-x