skip to main content
research-article

Towards Query-Efficient Black-Box Attacks: A Universal Dual Transferability-Based Framework

Published:08 May 2023Publication History
Skip Abstract Section

Abstract

Adversarial attacks have threatened the application of deep neural networks in security-sensitive scenarios. Most existing black-box attacks fool the target model by interacting with it many times and producing global perturbations. However, all pixels are not equally crucial to the target model; thus, indiscriminately treating all pixels will increase query overhead inevitably. In addition, existing black-box attacks take clean samples as start points, which also limits query efficiency. In this article, we propose a novel black-box attack framework, constructed on a strategy of dual transferability (DT), to perturb the discriminative areas of clean examples within limited queries. The first kind of transferability is the transferability of model interpretations. Based on this property, we identify the discriminative areas of clean samples for generating local perturbations. The second is the transferability of adversarial examples, which helps us to produce local pre-perturbations for further improving query efficiency. We achieve the two kinds of transferability through an independent auxiliary model and do not incur extra query overhead. After identifying discriminative areas and generating pre-perturbations, we use the pre-perturbed samples as better start points and further perturb them locally in a black-box manner to search the corresponding adversarial examples. The DT strategy is general; thus, the proposed framework can be applied to different types of black-box attacks. We conduct extensive experiments to show that, under various system settings, our framework can significantly improve the query efficiency of existing black-box attacks and attack success rates.

REFERENCES

  1. [1] Alzantot Moustafa, Sharma Yash, Chakraborty Supriyo, Zhang Huan, Hsieh Cho-Jui, and Srivastava Mani B.. 2019. Genattack: Practical black-box attacks with gradient-free optimization. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO’19). ACM, New York, NY, 11111119. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. [2] Bhagoji Arjun Nitin, He Warren, Li Bo, and Song Dawn. 2018. Practical black-box attacks on deep neural networks using efficient query mechanisms. In Proceedings of the European Conference on Computer Vision (ECCV’18). Springer International Publishing, 158174.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. [3] Brendel Wieland, Rauber Jonas, and Bethge Matthias. 2018. Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. In International Conference on Learning Representations (ICLR’18). OpenReview.net, Vancouver, 112.Google ScholarGoogle Scholar
  4. [4] Carlini Nicholas and Wagner David. 2017. Towards evaluating the robustness of neural networks. In IEEE Symposium on Security and Privacy (SP’17). IEEE Computer Society, Seattle, 3957. Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Carlini Nicholas and Wagner David. 2018. Audio adversarial examples: Targeted attacks on speech-to-text. In IEEE Security and Privacy Workshops (SPW’17). IEEE Computer Society, San Francisco, 17. Google ScholarGoogle ScholarCross RefCross Ref
  6. [6] Chattopadhay Aditya, Sarkar Anirban, Howlader Prantik, and Balasubramanian Vineeth N.. 2018. Grad-CAM++: Generalized gradient-based visual explanations for deep convolutional networks. In 2018 IEEE Winter Conference on Applications of Computer Vision (WACV’18). IEEE Computer Society, Lake Tahoe, 839847. Google ScholarGoogle ScholarCross RefCross Ref
  7. [7] Chen Pin-Yu, Zhang Huan, Sharma Yash, Yi Jinfeng, and Hsieh Cho-Jui. 2017. ZOO: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. In Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security (AISec’17). ACM, Dallas, 1526. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. [8] Dabkowski Piotr and Gal Yarin. 2017. Real time image saliency for black box classifiers. In Advances in Neural Information Processing Systems, I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (Eds.). Vol. 30. Curran Associates, Inc., Long Beach.Google ScholarGoogle Scholar
  9. [9] Dong Xiaoyi, Han Jiangfan, Chen Dongdong, Liu Jiayang, Bian Huanyu, Ma Zehua, Li Hongsheng, Wang Xiaogang, Zhang Weiming, and Yu Nenghai. 2020. Robust superpixel-guided attentional adversarial attack. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’20). IEEE Computer Society, Los Alamitos, 1289212901. Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Dong Yinpeng, Liao Fangzhou, Pang Tianyu, Su Hang, Zhu Jun, Hu Xiaolin, and Li Jianguo. 2018. Boosting adversarial attacks with momentum. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’18). IEEE Computer Society, 91859193. Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Du Yali, Fang Meng, Yi Jinfeng, Cheng Jun, and Tao Dacheng. 2018. Towards query efficient black-box attacks: An input-free perspective. In Proceedings of the 11th ACM Workshop on Artificial Intelligence and Security (AISec’18). ACM, 1324. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. [12] Fezza Sid Ahmed, Bakhti Yassine, Hamidouche Wassim, and Déforges Olivier. 2019. Perceptual evaluation of adversarial attacks for CNN-based image classification. In International Conference on Quality of Multimedia Experience (QoMEX’19). IEEE, 16. Google ScholarGoogle ScholarCross RefCross Ref
  13. [13] Fu Ruigang, Hu Qingyong, Dong Xiaohu, Guo Yulan, Gao Yinghui, and Li Biao. 2020. Axiom-based Grad-CAM: Towards accurate visualization and explanation of CNNs. In British Machine Vision Conference (BMVC). BMVA Press, UK, 113.Google ScholarGoogle Scholar
  14. [14] Goodfellow Ian J., Shlens Jonathon, and Szegedy Christian. 2015. Explaining and harnessing adversarial examples. In International Conference on Learning Representations (ICLR’15). OpenReview.net, San Diego, 111.Google ScholarGoogle Scholar
  15. [15] Guo Chuan, Gardner Jacob, You Yurong, Wilson Andrew Gordon, and Weinberger Kilian. 2019. Simple black-box adversarial attacks. In Proceedings of the 36th International Conference on Machine Learning (ICML’19). PMLR, Long Beach, 24842493.Google ScholarGoogle Scholar
  16. [16] He Kaiming, Zhang Xiangyu, Ren Shaoqing, and Sun Jian. 2016. Deep residual learning for image recognition. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’16). IEEE Computer Society, 770778.Google ScholarGoogle ScholarCross RefCross Ref
  17. [17] Hinton Geoffrey, Vinyals Oriol, and Dean Jeff. 2015. Distilling the knowledge in a neural network. arXiv:1503.02531 (2015).Google ScholarGoogle Scholar
  18. [18] Ilyas Andrew, Engstrom Logan, Athalye Anish, and Lin Jessy. 2018. Black-box adversarial attacks with limited queries and information. In Proceedings of the 35th International Conference on Machine Learning (ICML’18), Vol. 80. PMLR, Stockholmsmässan, 21372146.Google ScholarGoogle Scholar
  19. [19] Jiang Peng-Tao, Zhang Chang-Bin, Hou Qibin, Cheng Ming-Ming, and Wei Yunchao. 2021. LayerCAM: Exploring hierarchical class activation maps for localization. IEEE Transactions on Image Processing 30 (2021), 58755888.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Kurakin Alexey, Goodfellow Ian, and Bengio Samy. 2017. Adversarial examples in the physical world. In International Conference on Learning Representations (ICLR’17), 117.Google ScholarGoogle Scholar
  21. [21] Kurakin Alexey, Goodfellow Ian, and Bengio Samy. 2017. Adversarial machine learning at scale. International Conference on Learning Representations Workshop, (ICLR Workshop). OpenReview.net, Toulon, 114.Google ScholarGoogle Scholar
  22. [22] Larson Eric Cooper and Chandler Damon Michael. 2010. Most apparent distortion: Full-reference image quality assessment and the role of strategy. Journal of Electronic Imaging 19, 1 (2010), 011006.Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Li Xurong, Ji Shouling, Han Meng, Ji Juntao, Ren Zhenyu, Liu Yushan, and Wu Chunming. 2021. Adversarial examples versus cloud-based detectors: A black-box empirical study. IEEE Transactions on Dependable and Secure Computing 18, 4 (2021), 19331949.Google ScholarGoogle Scholar
  24. [24] Liu Yanpei, Chen Xinyun, Liu Chang, and Song Dawn. 2017. Delving into transferable adversarial examples and black-box attacks. In International Conference on Learning Representations (ICLR’17). OpenReview.net, Toulon, 114.Google ScholarGoogle Scholar
  25. [25] Madry Aleksander, Makelov Aleksandar, Schmidt Ludwig, Tsipras Dimitris, and Vladu Adrian. 2018. Towards deep learning models resistant to adversarial attacks. In International Conference on Learning Representations (ICLR’18). Open-Review.net, Vancouver, 123.Google ScholarGoogle Scholar
  26. [26] Moosavi-Dezfooli Seyed Mohsen, Fawzi Alhussein, Fawzi Omar, and Frossard Pascal. 2017. Universal adversarial perturbations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’17). IEEE Computer Society, 17651773.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Moosavi-Dezfooli Seyed-Mohsen, Fawzi Alhussein, and Frossard Pascal. 2016. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16). IEEE Computer Society, 25742582.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Muhammad Mohammed Bany and Yeasin Mohammed. 2020. Eigen-cam: Class activation map using principal components. In International Joint Conference on Neural Networks (IJCNN’20). IEEE, 17.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Narodytska Nina and Kasiviswanathan Shiva. 2017. Simple black-box adversarial attacks on deep neural networks. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW’17), Vol. 2. IEEE Computer, 13101318.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Omeiza Daniel, Speakman Skyler, Cintas Celia, and Weldermariam Komminist. 2019. Smooth Grad-CAM++: An enhanced inference level visualization technique for deep convolutional neural network models. arXiv preprint arXiv:1908.01224 (2019).Google ScholarGoogle Scholar
  31. [31] Papernot Nicolas, McDaniel Patrick, Goodfellow Ian, Jha Somesh, Celik Z. Berkay, and Swami Ananthram. 2017. Practical black-box attacks against machine learning. In Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security (AisaCCS’17). ACM, 506519.Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. [32] Papernot Nicolas, McDaniel Patrick, Jha Somesh, Fredrikson Matt, Celik Z. Berkay, and Swami Ananthram. 2016. The limitations of deep learning in adversarial settings. In IEEE European symposium on security and privacy (EuroS&P’16). IEEE, 372387.Google ScholarGoogle Scholar
  33. [33] Ramaswamy Harish Guruprasad et al. 2020. Ablation-CAM: Visual explanations for deep convolutional network via gradient-free localization. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV’20). IEEE, 983991.Google ScholarGoogle Scholar
  34. [34] Ribeiro Marco Tulio, Singh Sameer, and Guestrin Carlos. 2016. “Why should I trust you?”: Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (SIGKDD’16). ACM, 11351144.Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. [35] Selvaraju Ramprasaath R., Cogswell Michael, Das Abhishek, Vedantam Ramakrishna, Parikh Devi, and Batra Dhruv. 2017. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision. IEEE Computer Society, 618626.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Simonyan Karen and Zisserman Andrew. 2015. Very deep convolutional networks for large-scale image recognition. In International Conference on Learning Representations (ICLR’15). San Diego, 114.Google ScholarGoogle Scholar
  37. [37] Su Jiawei, Vargas Danilo Vasconcellos, and Sakurai Kouichi. 2019. One pixel attack for fooling deep neural networks. IEEE Transactions on Evolutionary Computation 23, 5 (2019), 828841.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Suya Fnu, Chi Jianfeng, Evans David, and Tian Yuan. 2020. Hybrid batch attacks: Finding black-box adversarial examples with limited queries. In 29th USENIX Security Symposium (USENIX’20). USENIX Association, 13271344.Google ScholarGoogle Scholar
  39. [39] Szegedy Christian, Vanhoucke Vincent, Ioffe Sergey, Shlens Jon, and Wojna Zbigniew. 2016. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16). IEEE Computer Society, 28182826.Google ScholarGoogle ScholarCross RefCross Ref
  40. [40] Szegedy Christian, Zaremba Wojciech, Sutskever Ilya, Bruna Joan, Erhan Dumitru, Goodfellow Ian, and Fergus Rob. 2014. Intriguing properties of neural networks. In International Conference on Learning Representations (ICLR’14). Banff, AB, 110.Google ScholarGoogle Scholar
  41. [41] Tu Chun-Chen, Ting Paishun, Chen Pin-Yu, Liu Sijia, Zhang Huan, Yi Jinfeng, Hsieh Cho-Jui, and Cheng Shin-Ming. 2019. Autozoom: Autoencoder-based zeroth order optimization method for attacking black-box neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI’19), Vol. 33. AAAI Press, 742749.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. [42] Wang Xianmin, Li Jing, Kuang Xiaohui, Tan Yu-an, and Li Jin. 2019. The security of machine learning in an adversarial setting: A survey. Journal of Parallel and Distributed Computing 130 (2019), 1223.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Wierstra Daan, Schaul Tom, Glasmachers Tobias, Sun Yi, Peters Jan, and Schmidhuber Jürgen. 2014. Natural evolution strategies. Journal of Machine Learning Research 15, 1 (2014), 949980.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. [44] Xie Cihang, Wang Jianyu, Zhang Zhishuai, Zhou Yuyin, Xie Lingxi, and Yuille Alan. 2017. Adversarial examples for semantic segmentation and object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV’17). IEEE Computer Society, 13691378.Google ScholarGoogle ScholarCross RefCross Ref
  45. [45] Xie Cihang, Zhang Zhishuai, Zhou Yuyin, Bai Song, Wang Jianyu, Ren Zhou, and Yuille Alan L.. 2019. Improving transferability of adversarial examples with input diversity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR’19). IEEE, 27302739.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Zhang Jiliang and Li Chen. 2020. Adversarial examples: Opportunities and challenges. IEEE Transactions on Neural Networks and Learning Systems 31, 7 (2020), 25782593.Google ScholarGoogle Scholar
  47. [47] Zhang Wei Emma, Sheng Quan Z., Alhazmi Ahoud, and Li Chenliang. 2020. Adversarial attacks on deep-learning models in natural language processing: A survey. ACM Transactions on Intelligent Systems and Technology 11, 3 (2020), 141.Google ScholarGoogle ScholarDigital LibraryDigital Library
  48. [48] Zhou Bolei, Khosla Aditya, Lapedriza Agata, Oliva Aude, and Torralba Antonio. 2016. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16). IEEE Computer Society, 29212929.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Towards Query-Efficient Black-Box Attacks: A Universal Dual Transferability-Based Framework

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Transactions on Intelligent Systems and Technology
        ACM Transactions on Intelligent Systems and Technology  Volume 14, Issue 4
        August 2023
        481 pages
        ISSN:2157-6904
        EISSN:2157-6912
        DOI:10.1145/3596215
        • Editor:
        • Huan Liu
        Issue’s Table of Contents

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 8 May 2023
        • Online AM: 13 February 2023
        • Accepted: 23 January 2023
        • Revised: 20 October 2022
        • Received: 1 May 2022
        Published in tist Volume 14, Issue 4

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
      • Article Metrics

        • Downloads (Last 12 months)236
        • Downloads (Last 6 weeks)12

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Full Text

      View this article in Full Text.

      View Full Text