skip to main content
10.1145/3373376.3378460acmconferencesArticle/Chapter ViewAbstractPublication PagesasplosConference Proceedingsconference-collections
research-article
Open Access

DeepSniffer: A DNN Model Extraction Framework Based on Learning Architectural Hints

Authors Info & Claims
Published:13 March 2020Publication History

ABSTRACT

As deep neural networks (DNNs) continue their reach into a wide range of application domains, the neural network architecture of DNN models becomes an increasingly sensitive subject, due to either intellectual property protection or risks of adversarial attacks. Previous studies explore to leverage architecture-level events disposed in hardware platforms to extract the model architecture information. They pose the following limitations: requiring a priori knowledge of victim models, lacking in robustness and generality, or obtaining incomplete information of the victim model architecture.

Our paper proposes DeepSniffer, a learning-based model extraction framework to obtain the complete model architecture information without any prior knowledge of the victim model. It is robust to architectural and system noises introduced by the complex memory hierarchy and diverse run-time system optimizations. The basic idea of DeepSniffer is to learn the relation between extracted architectural hints (e.g., volumes of memory reads/writes obtained by side-channel or bus snooping attacks) and model internal architectures. Taking GPU platforms as a show case, DeepSniffer conducts model extraction by learning both the architecture-level execution features of kernels and the inter-layer temporal association information introduced by the common practice of DNN design. We demonstrate that DeepSniffer works experimentally in the context of an off-the-shelf Nvidia GPU platform running a variety of DNN models. The extracted models are directly helpful to the attempting of crafting adversarial inputs. Our experimental results show that DeepSniffer achieves a high accuracy of model extraction and thus improves the adversarial attack success rate from 14.6%$\sim$25.5% (without network architecture knowledge) to 75.9% (with extracted network architecture). The DeepSniffer project has been released in Github.

References

  1. 2019. HMTT: Hybrid Memory Trace Toolkit. http://asg.ict.ac.cn/hmtt/Google ScholarGoogle Scholar
  2. Naveed Akhtar and Ajmal Mian. 2018. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey. CoRR abs/1801.00553 (2018). arXiv:1801.00553Google ScholarGoogle ScholarCross RefCross Ref
  3. Scott Alfeld, Xiaojin Zhu, and Paul Barford. 2016. Data Poisoning Attacks Against Autoregressive Models (AAAI'16). AAAI Press, 1452? 1458.Google ScholarGoogle Scholar
  4. Shumeet Baluja and Ian Fischer. 2017. Adversarial transformation networks: Learning to generate adversarial examples. arXiv preprint arXiv:1703.09387 (2017).Google ScholarGoogle Scholar
  5. Erik-Oliver Blass and William Robertson. 2012. TRESOR-HUNT: Attacking CPU-bound Encryption (ACSAC '12). ACM, New York, NY, USA, 71--78.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Robert Callan, Alenka Zajic, and Milos Prvulovic. 2014. A Practical Methodology for Measuring the Side-Channel Signal Available to the Attacker for Instruction-Level Events (MICRO-47). IEEE Computer Society, Washington, DC, USA, 242?254.Google ScholarGoogle Scholar
  7. Nicholas Carlini and David Wagner. 2017. Towards evaluating the robustness of neural networks. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE, 39--57.Google ScholarGoogle ScholarCross RefCross Ref
  8. Tianqi Chen, Thierry Moreau, Ziheng Jiang, Haichen Shen, Eddie Q. Yan, Leyuan Wang, Yuwei Hu, Luis Ceze, Carlos Guestrin, and Arvind Krishnamurthy. 2018. TVM: End-to-End Optimization Stack for Deep Learning. CoRR abs/1802.04799 (2018). arXiv:1802.04799Google ScholarGoogle Scholar
  9. Moustapha Cisse, Yossi Adi, Natalia Neverova, and Joseph Keshet. 2017. Houdini: Fooling deep structured prediction models. arXiv preprint arXiv:1707.05373 (2017).Google ScholarGoogle Scholar
  10. Ronan Collobert and Jason Weston. 2008. A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning (ICML '08). ACM, New York, NY, USA, 160--167.Google ScholarGoogle Scholar
  11. Swagatam Das and Ponnuthurai Nagaratnam Suganthan. 2011. Differential evolution: a survey of the state-of-the-art. IEEE transactions on evolutionary computation 15, 1 (2011), 4--31.Google ScholarGoogle Scholar
  12. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009. IEEE, 248--255.Google ScholarGoogle ScholarCross RefCross Ref
  13. M. Dey, A. Nazari, A. Zajic, and M. Prvulovic. 2018. EMPROF: Memory Profiling Via EM-Emanation in IoT and Hand-Held Devices. In 2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO). 881--893.Google ScholarGoogle Scholar
  14. Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. 2015. Explaining and harnessing adversarial examples. Proceedings of the International Conference on Learning Representations (2015).Google ScholarGoogle Scholar
  15. Alex Graves, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. 2006. Connectionist Temporal Classification: Labelling Unsegmented Sequence Data with Recurrent Neural Networks (ICML '06). ACM, New York, NY, USA, 369--376.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Alex Graves and Navdeep Jaitly. 2014. Towards End-to-end Speech Recognition with Recurrent Neural Networks (ICML'14). II-1764- 1772.Google ScholarGoogle Scholar
  17. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition. 770--778.Google ScholarGoogle ScholarCross RefCross Ref
  18. Weizhe Hua, Zhiru Zhang, and G. Edward Suh. 2018. Reverse Engineering Convolutional Neural Networks Through Side-channel Information Leaks (DAC '18). ACM, New York, NY, USA, 4:1--4:6.Google ScholarGoogle Scholar
  19. Andrew Huang. 2003. Keeping Secrets in Hardware: The Microsoft Xbox Case Study. In Revised Papers from the 4th International Workshop on Cryptographic Hardware and Embedded Systems (CHES '02). Springer-Verlag, London, UK, UK, 213--227.Google ScholarGoogle Scholar
  20. Gao Huang, Zhuang Liu, Laurens Van Der Maaten, and Kilian QWeinberger. [n.d.]. Densely connected convolutional networks. In CVPR 2017.Google ScholarGoogle ScholarCross RefCross Ref
  21. Yongbing Huang, Licheng Chen, Zehan Cui, Yuan Ruan, Yungang Bao, Mingyu Chen, and Ninghui Sun. 2014. HMTT: A Hybrid Hardware/ Software Tracing System for Bridging the DRAM Access Trace's Semantic Gap. ACM Trans. Archit. Code Optim. 11, 1, Article 7 (Feb. 2014), 25 pages.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2016. Binarized neural networks. In Advances in neural information processing systems. 4107--4115.Google ScholarGoogle Scholar
  23. Forrest N Iandola, Song Han, Matthew W Moskewicz, Khalid Ashraf, William J Dally, and Kurt Keutzer. 2016. Squeezenet: Alexnet-level accuracy with 50x fewer parameters and < 0.5 mb model size. arXiv preprint arXiv:1602.07360 (2016).Google ScholarGoogle Scholar
  24. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ImageNet Classification with Deep Convolutional Neural Networks (NIPS'12). Curran Associates Inc., USA, 1097--1105.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Alexey Kurakin, Ian Goodfellow, and Samy Bengio. 2016. Adversarial examples in the physical world. arXiv preprint arXiv:1607.02533 (2016).Google ScholarGoogle Scholar
  26. Chang Liu, Austin Harris, Martin Maas, Michael Hicks, Mohit Tiwari, and Elaine Shi. 2015. GhostRider: A Hardware-Software System for Memory Trace Oblivious Computation (ASPLOS '15). ACM, New York, NY, USA, 87--101.Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Chang Liu, Michael Hicks, and Elaine Shi. 2013. Memory Trace Oblivious Program Execution. In Proceedings of the 2013 IEEE 26th Computer Security Foundations Symposium (CSF '13). IEEE Computer Society, Washington, DC, USA, 51--65.Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Yanpei Liu, Xinyun Chen, Chang Liu, and Dawn Song. 2017. Delving into Transferable Adversarial Examples and Black-box Attacks. ICLR abs/1611.02770 (2017). arXiv:1611.02770Google ScholarGoogle Scholar
  29. Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, Omar Fawzi, and Pascal Frossard. 2017. Universal adversarial perturbations. arXiv preprint (2017).Google ScholarGoogle Scholar
  30. Seyed Mohsen Moosavi Dezfooli, Alhussein Fawzi, and Pascal Frossard. 2016. Deepfool: a simple and accurate method to fool deep neural networks. In Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).Google ScholarGoogle ScholarCross RefCross Ref
  31. Konda Reddy Mopuri, Utsav Garg, and R Venkatesh Babu. 2017. Fast feature fool: A data independent approach to universal adversarial perturbations. arXiv preprint arXiv:1707.05572 (2017).Google ScholarGoogle Scholar
  32. Hoda Naghibijouybari, Ajaya Neupane, Zhiyun Qian, and Nael Abu- Ghazaleh. 2018. Rendered Insecure: GPU Side Channel Attacks Are Practical (CCS '18). ACM, New York, NY, USA, 2139--2153.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Papernot Nicolas, D. McDaniel Patrick, Jha Somesh, Fredrikson Matt, Celik Z. Berkay, and Swami Ananthram. 2015. The Limitations of Deep Learning in Adversarial Settings. CoRR abs/1511.07528 (2015). arXiv:1511.07528Google ScholarGoogle Scholar
  34. Nvidia. [n.d.]. CUDA toolkit documentation. http://docs.nvidia.com/ cuda/profiler-users-guide/index.htmlGoogle ScholarGoogle Scholar
  35. NVIDIA. 2016. NVIDIA Tesla K40 Active GPU Accelerator. http: //www.pny.com/nvidia-tesla-k40-active-gpu-accelerator.Google ScholarGoogle Scholar
  36. Nvidia. 2017. NVIDIA cuDNN GPU Accelerated Deep Learning. https: //developer.nvidia.com/cudnnGoogle ScholarGoogle Scholar
  37. Nicolas Papernot, Patrick McDaniel, Ian Goodfellow, Somesh Jha, Z. Berkay Celik, and Ananthram Swami. 2017. Practical Black-Box Attacks Against Machine Learning (ASIA CCS '17). ACM, New York, NY, USA, 506--519.Google ScholarGoogle Scholar
  38. Nicolas Papernot, Patrick McDaniel, Arunesh Sinha, and MichaelWellman. 2016. Towards the science of security and privacy in machine learning. arXiv preprint arXiv:1611.03814 (2016).Google ScholarGoogle Scholar
  39. Nicolas Papernot, Patrick D. McDaniel, and Ian J. Goodfellow. 2016. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples. CoRR abs/1605.07277 (2016). arXiv:1605.07277 http://arxiv.org/abs/1605.07277Google ScholarGoogle Scholar
  40. PyTorch. [n.d.]. Pytorch Tutorials. http://pytorch.org/tutorials/Google ScholarGoogle Scholar
  41. Tractica Report. 2016. Artificial Intelligience Market Forecasts.Google ScholarGoogle Scholar
  42. M. Rhu, M. O'Connor, N. Chatterjee, J. Pool, Y. Kwon, and S.W. Keckler. 2018. Compressing DMA Engine: Leveraging Activation Sparsity for Training Deep Neural Networks. In 2018 IEEE International Symposium on High Performance Computer Architecture (HPCA). 78--91.Google ScholarGoogle Scholar
  43. Sayantan Sarkar, Ankan Bansal, Upal Mahbub, and Rama Chellappa. 2017. UPSET and ANGRI: Breaking High Performance Image Classifiers. arXiv preprint arXiv:1707.01159 (2017).Google ScholarGoogle Scholar
  44. Bernt Schiele Mario Fritz Seong Joon Oh, Max Augustin. 2018. Towards Reverse-Engineering Black-Box Neural Networks. ICLR abs/1605.07277 (2018). https://arxiv.org/abs/1711.01768Google ScholarGoogle Scholar
  45. Karen Simonyan and Andrew Zisserman. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR abs/1409.1556 (2014). arXiv:1409.1556 http://arxiv.org/abs/1409.1556Google ScholarGoogle Scholar
  46. Emil Stefanov, Marten van Dijk, Elaine Shi, Christopher Fletcher, Ling Ren, Xiangyao Yu, and Srinivas Devadas. 2013. Path ORAM: An Extremely Simple Oblivious RAM Protocol (CCS '13). ACM, New York, NY, USA, 299--310.Google ScholarGoogle ScholarDigital LibraryDigital Library
  47. Jiawei Su, Danilo Vasconcellos Vargas, and Sakurai Kouichi. 2017. One pixel attack for fooling deep neural networks. arXiv preprint arXiv:1710.08864 (2017).Google ScholarGoogle Scholar
  48. Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to Sequence Learning with Neural Networks. In Advances in Neural Information Processing Systems 27, Z. Ghahramani, M.Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger (Eds.). Curran Associates, Inc., 3104--3112.Google ScholarGoogle ScholarDigital LibraryDigital Library
  49. Christian Szegedy, Sergey Ioffe, Vincent Vanhoucke, and Alexander A Alemi. 2017. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning.. In AAAI. 4278--4284.Google ScholarGoogle Scholar
  50. Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jonathon Shlens, and Zbigniew Wojna. 2015. Rethinking the Inception Architecture for Computer Vision. CoRR abs/1512.00567 (2015). arXiv:1512.00567Google ScholarGoogle Scholar
  51. Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2013. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199 (2013).Google ScholarGoogle Scholar
  52. Jin-Hua Tao, Zi-Dong Du, Qi Guo, Hui-Ying Lan, Lei Zhang, Sheng- Yuan Zhou, Cong Liu, Hai-Feng Liu, Shan Tang, and Allen Rush. 2017. BENCHIP: Benchmarking Intelligence Processors. arXiv preprint arXiv:1710.08315 (2017).Google ScholarGoogle Scholar
  53. TechCrunch. 2017. Nvidia is powering the world's first level 3 selfdriving production car.Google ScholarGoogle Scholar
  54. TensorFlow. [n.d.]. Post-training quantization. https://www. tensorflow.org/lite/performance/post_training_quantization.Google ScholarGoogle Scholar
  55. Florian Tramèr, Fan Zhang, Ari Juels, Michael K. Reiter, and Thomas Ristenpart. 2016. Stealing Machine Learning Models via Prediction APIs (SEC'16). USENIX Association, Berkeley, CA, USA, 601--618.Google ScholarGoogle Scholar
  56. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems. 6000--6010.Google ScholarGoogle Scholar
  57. Binghui Wang and Neil Zhenqiang Gong. 2018. Stealing Hyperparameters in Machine Learning. CoRR abs/1802.05351 (2018). arXiv:1802.05351Google ScholarGoogle Scholar
  58. Waymo. 2017. IntroducingWaymo's suite of custom-build, self-driving hardware. https://medium.com/waymo/introducing-waymos-suiteof- custom-built-self-driving-hardware-c47d1714563/Google ScholarGoogle Scholar
  59. Lingxiao Wei, Yannan Liu, Bo Luo, Yu Li, and Qiang Xu. 2018. I Know What You See: Power Side-Channel Attack on Convolutional Neural Network Accelerators. CoRR abs/1803.05847 (2018). arXiv:1803.05847Google ScholarGoogle ScholarDigital LibraryDigital Library
  60. Nicholas Wilt. 2013. The cuda handbook: A comprehensive guide to gpu programming. Pearson Education.Google ScholarGoogle Scholar
  61. Wayne Xiong, Jasha Droppo, Xuedong Huang, Frank Seide, Mike Seltzer, Andreas Stolcke, Dong Yu, and Geoffrey Zweig. 2017. The Microsoft 2016 conversational speech recognition system. In Acoustics, Speech and Signal Processing (ICASSP), 2017 IEEE International Conference on. IEEE, 5255--5259.Google ScholarGoogle ScholarCross RefCross Ref
  62. Mengjia Yan, ChristopherW. Fletcher, and Josep Torrellas. 2018. Cache Telepathy: Leveraging Shared Resource Attacks to Learn DNN Architectures. CoRR abs/1808.04761 (2018). arXiv:1808.04761Google ScholarGoogle Scholar
  63. Xingcheng Zhang, Zhizhong Li, Chen Change Loy, and Dahua Lin. 2016. PolyNet:APursuit of Structural Diversity in Very Deep Networks. CoRR abs/1611.05725 (2016). arXiv:1611.05725Google ScholarGoogle Scholar
  64. Barret Zoph, Vijay Vasudevan, Jonathon Shlens, and Quoc V. Le. 2017. Learning Transferable Architectures for Scalable Image Recognition. CoRR abs/1707.07012 (2017). arXiv:1707.07012Google ScholarGoogle Scholar

Index Terms

  1. DeepSniffer: A DNN Model Extraction Framework Based on Learning Architectural Hints

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          ASPLOS '20: Proceedings of the Twenty-Fifth International Conference on Architectural Support for Programming Languages and Operating Systems
          March 2020
          1412 pages
          ISBN:9781450371025
          DOI:10.1145/3373376

          Copyright © 2020 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 13 March 2020

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

          Acceptance Rates

          Overall Acceptance Rate535of2,713submissions,20%

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader