Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks

Download or Read eBook Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks PDF written by Kyungmi Lee (S. M.) and published by . This book was released on 2020 with total page 93 pages. Available in PDF, EPUB and Kindle.
Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks
Author :
Publisher :
Total Pages : 93
Release :
ISBN-10 : OCLC:1192484009
ISBN-13 :
Rating : 4/5 (09 Downloads)

Book Synopsis Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks by : Kyungmi Lee (S. M.)

Book excerpt: Deep neural networks are known to be vulnerable to adversarial perturbations, which are often imperceptible to humans but can alter predictions of machine learning systems. Since the exact value of adversarial robustness is difficult to obtain for complex deep neural networks, accuracy of the models against perturbed examples generated by attack methods is empirically used as a proxy to adversarial robustness. However, failure of attack methods to find adversarial perturbations cannot be equated with being robust. In this work, we identify three common cases that lead to overestimation of accuracy against perturbed examples generated by bounded first-order attack methods: 1) the value of cross-entropy loss numerically becoming zero when using standard floating point representation, resulting in non-useful gradients; 2) innately non-differentiable functions in deep neural networks, such as Rectified Linear Unit (ReLU) activation and MaxPool operation, incurring “gradient masking” [2]; and 3) certain regularization methods used during training inducing the model to be less amenable to first-order approximation. We show that these phenomena exist in a wide range of deep neural networks, and that these phenomena are not limited to specific defense methods they have been previously investigated for. For each case, we propose compensation methods that either address sources of inaccurate gradient computation, such as numerical saturation for near zero values and non-differentiability, or reduce the total number of back-propagations for iterative attacks by approximating second-order information. These compensation methods can be combined with existing attack methods for a more precise empirical evaluation metric. We illustrate the impact of these three phenomena with examples of practical interest, such as benchmarking model capacity and regularization techniques for robustness. Furthermore, we show that the gap between adversarial accuracy and the guaranteed lower bound of robustness can be partially explained by these phenomena. Overall, our work shows that overestimated adversarial accuracy that is not indicative of robustness is prevalent even for conventionally trained deep neural networks, and highlights cautions of using empirical evaluation without guaranteed bounds.


Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks Related Books

Improved Methodology for Evaluating Adversarial Robustness in Deep Neural Networks
Language: en
Pages: 93
Authors: Kyungmi Lee (S. M.)
Categories:
Type: BOOK - Published: 2020 - Publisher:

DOWNLOAD EBOOK

Deep neural networks are known to be vulnerable to adversarial perturbations, which are often imperceptible to humans but can alter predictions of machine learn
Evaluating and Understanding Adversarial Robustness in Deep Learning
Language: en
Pages: 175
Authors: Jinghui Chen
Categories:
Type: BOOK - Published: 2021 - Publisher:

DOWNLOAD EBOOK

Deep Neural Networks (DNNs) have made many breakthroughs in different areas of artificial intelligence. However, recent studies show that DNNs are vulnerable to
On the Robustness of Neural Network: Attacks and Defenses
Language: en
Pages: 158
Authors: Minhao Cheng
Categories:
Type: BOOK - Published: 2021 - Publisher:

DOWNLOAD EBOOK

Neural networks provide state-of-the-art results for most machine learning tasks. Unfortunately, neural networks are vulnerable to adversarial examples. That is
Advances in Reliably Evaluating and Improving Adversarial Robustness
Language: en
Pages:
Authors: Jonas Rauber
Categories:
Type: BOOK - Published: 2021 - Publisher:

DOWNLOAD EBOOK

Machine learning has made enormous progress in the last five to ten years. We can now make a computer, a machine, learn complex perceptual tasks from data rathe
Evaluating and Certifying the Adversarial Robustness of Neural Language Models
Language: en
Pages: 0
Authors: Muchao Ye
Categories:
Type: BOOK - Published: 2024 - Publisher:

DOWNLOAD EBOOK

Language models (LMs) built by deep neural networks (DNNs) have achieved great success in various areas of artificial intelligence, which have played an increas