Abstract
Adversarial examples have proved efficacious in fooling deep neural networks recently. Many researchers have studied this issue of adversarial examples by evaluating neural networks against their attack techniques and increasing the robustness of neural networks with their defense techniques. To the best of our knowledge, adversarial training is one of the most effective defense techniques against the adversarial examples. However, the method is not able to cope with new attacks because it requires attack techniques in the training phase. In this paper, we propose a novel defense technique, Attack-Less Adversarial Training (ALAT) method, which is independent from any attack techniques, thereby is useful in preventing future attacks. Specifically, ALAT regenerates every pixel of an image into different pixel value, which commonly eliminates the majority of the adversarial noises in the adversarial example. This pixel regeneration is useful in defense because the adversarial noises are the core problem that make the neural networks produce high misclassification rate. Our experiment results with several benchmark datasets show that our method not only relieves over-fitting issue during the training of neural networks with a large number of epochs, but also boosts the robustness of the neural network.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
We modify and extend the scripts from CleverHans. Our scripts are available at https://github.com/canboy123/alat.
References
Adeli H, Wu M (1998) Regularization neural network for construction cost estimation. J Constr Eng Manag 124(1):18–24
Agarwal A, Vatsa M, Singh R, Ratha N (2021) Cognitive data augmentation for adversarial defense via pixel masking. Patt Recogn Lett
Athalye A, Carlini N, Wagner D (2018) Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In: Dy J, Krause A (eds) Proceedings of the 35th international conference on machine learning, proceedings of machine learning research. http://proceedings.mlr.press/v80/athalye18a.html, vol 80. PMLR, Stockholmsmässan, Stockholm Sweden, pp 274–283
Athalye A, Engstrom L, Ilyas A, Kwok K (2018) Synthesizing robust adversarial examples. In: Dy J, Krause A (eds) Proceedings of the 35th international conference on machine learning, proceedings of machine learning research, vol 80. PMLR, Stockholmsmässan, Stockholm Sweden, pp 284–293
Biggio B, Roli F (2018) Wild patterns: Ten years after the rise of adversarial machine learning. Patt Recogn 84:317–331
Brown TB, Mané D, Roy A, Abadi M, Gilmer J (2017) Adversarial patch. arXiv:1712.09665
Carlini N, Wagner D (2017) Adversarial examples are not easily detected: Bypassing ten detection methods. In: Proceedings of the 10th ACM workshop on artificial intelligence and security. ACM, pp 3–14
Carlini N, Wagner D (2017) Towards evaluating the robustness of neural networks. In: Proceedings of the IEEE symposium on security and privacy (SP). IEEE, pp 39–57
Dhillon GS, Azizzadenesheli K, Bernstein JD, Kossaifi J, Khanna A, Lipton ZC, Anandkumar A (2018) Stochastic activation pruning for robust adversarial defense. In: International conference on learning representations. https://openreview.net/forum?id=H1uR4GZRZ
Dong Y, Liao F, Pang T, Su H, Zhu J, Hu X, Li J (2018) Boosting adversarial attacks with momentum. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9185–9193
Eykholt K, Evtimov I, Fernandes E, Li B, Rahmati A, Xiao C, Prakash A, Kohno T, Song D (2018) Robust physical-world attacks on deep learning visual classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1625–1634
Goodfellow IJ, Shlens J, Szegedy C (2014) Explaining and harnessing adversarial examples. arXiv:1412.6572
Guo C, Rana M, Cisse M, van der Maaten L (2018) Countering adversarial images using input transformations. In: Proceedings of the international conference on learning representations. https://openreview.net/forum?id=SyJ7ClWCb
Ho J, Lee BG, Kang DK (2020) Uni-image: universal image construction for robust neural model. Neural Netw
Jafarnia-Jahromi M, Chowdhury T, Wu HT, Mukherjee S (2019) PPD: Permutation phase defense against adversarial examples in deep learning. https://openreview.net/forum?id=HkElFj0qYQ
Krantz J (2012) Experiencing sensation and perception. Upper Saddle River, NJ
Krizhevsky A, Nair V, Hinton G (2010) Cifar-10 (canadian institute for advanced research). http://www.cs.toronto.edu/kriz/cifar.html
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Proceedings of the advances in neural information processing systems, pp 1097–1105
Krogh A, Hertz JA (1992) A simple weight decay can improve generalization. In: Proceedings of the advances in neural information processing systems, pp 950–957
Kurakin A, Goodfellow I, Bengio S (2016) Adversarial examples in the physical world. arXiv:1607.02533
Kurakin A, Goodfellow I, Bengio S (2017) Adversarial machine learning at scale. In: Proceedings of the international conference on learning representations. https://openreview.net/forum?id=BJm4T4Kgx
Land EH (1959) Experiments in color vision. Sci Am 200(5):84–99
LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Madry A, Makelov A, Schmidt L, Tsipras D, Vladu A (2018) Towards deep learning models resistant to adversarial attacks. In: Proceedings of the international conference on learning representations. https://openreview.net/forum?id=rJzIBfZAb
McLaughlin N, Del Rincon JM, Miller P (2015) Data-augmentation for reducing dataset bias in person re-identification. In: 2015 12th IEEE international conference on advanced video and signal based surveillance (AVSS). IEEE, pp 1–6
Nowlan SJ, Hinton GE (1992) Simplifying neural networks by soft weight-sharing. Neural Comput 4(4):473–493
Papernot N, McDaniel P, Goodfellow I (2016) Transferability in machine learning:, from phenomena to black-box attacks using adversarial samples. arXiv:1605.07277
Papernot N, McDaniel P, Goodfellow I, Jha S, Celik ZB, Swami A. (2017) Practical black-box attacks against machine learning. In: Proceedings of the 2017 ACM on Asia conference on computer and communications security. ACM, pp 506–519
Ren H, Huang T, Yan H (2021) Adversarial examples: attacks and defenses in the physical world. Int J Mach Learn Cybern 1–12
Rony J, Hafemann LG, Oliveira LS, Ayed IB, Sabourin R, Granger E (2018) Decoupling direction and norm for efficient gradient-based l2 adversarial attacks and defenses. arXiv:1811.09600
Shi Y, Han Y, Zhang Q, Kuang X (2020) Adaptive iterative attack towards explainable adversarial robustness. Patt Recogn 107309
Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15(1):1929–1958
Sutanto RE, Lee S (2021) Real-time adversarial attack detection with deep image prior initialized as a high-level representation based blurring network. Electronics 10(1):52
Szegedy C, Zaremba W, Sutskever I, Bruna J, Erhan D, Goodfellow I, Fergus R (2013) Intriguing properties of neural networks. arXiv:1312.6199
Theagarajan R, Bhanu B (2020) Defending black box facial recognition classifiers against adversarial attacks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pp 812–813
Tramèr F, Kurakin A, Papernot N, Goodfellow I, Boneh D, McDaniel P (2018) Ensemble adversarial training: Attacks and defenses. In: International conference on learning representations. https://openreview.net/forum?id=rkZvSe-RZ
Wang J, Perez L (2017) The effectiveness of data augmentation in image classification using deep learning. Convolut Neural Netw Vis Recognit
Wong E, Rice L, Kolter JZ (2010) Fast is better than free: Revisiting adversarial training. In: International conference on learning representations. https://openreview.net/forum?id=BJx040EFvH
Xiao H, Rasul K, Vollgraf R (2017) Fashion-mnist:, a novel image dataset for benchmarking machine learning algorithms. arXiv:1708.07747
Yang P, Chen J, Hsieh CJ, Wang JL, Jordan MI (2020) Greedy attack and gumbel attack: Generating adversarial examples for discrete data. J Mach Learn Res 21(43):1–36
Yuan X, He P, Zhu Q, Bhat RR, Li X (2017) Adversarial examples: Attacks and defenses for deep learning. arXiv:1712.07107
Zhang X, Wang J, Wang T, Jiang R, Xu J, Zhao L (2021) Robust feature learning for adversarial defense via hierarchical feature alignment. Inf Sci 560:256–270
Acknowledgements
This work was supported by Institute for Information and Communications Technology Promotion (IITP) grant funded by the Korea government (MSIT) (No.2018-0-00245, Development of prevention technology against AI dysfunction induced by deception attack).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declare that they have no conflict of interest.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Ho, J., Lee, BG. & Kang, DK. Attack-less adversarial training for a robust adversarial defense. Appl Intell 52, 4364–4381 (2022). https://doi.org/10.1007/s10489-021-02523-y
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-02523-y