skip to main content
research-article

DeepPeep: Exploiting Design Ramifications to Decipher the Architecture of Compact DNNs

Authors Info & Claims
Published:28 October 2020Publication History
Skip Abstract Section

Abstract

The remarkable predictive performance of deep neural networks (DNNs) has led to their adoption in service domains of unprecedented scale and scope. However, the widespread adoption and growing commercialization of DNNs have underscored the importance of intellectual property (IP) protection. Devising techniques to ensure IP protection has become necessary due to the increasing trend of outsourcing the DNN computations on the untrusted accelerators in cloud-based services. The design methodologies and hyper-parameters of DNNs are crucial information, and leaking them may cause massive economic loss to the organization. Furthermore, the knowledge of DNN’s architecture can increase the success probability of an adversarial attack where an adversary perturbs the inputs and alters the prediction.

In this work, we devise a two-stage attack methodology “DeepPeep,” which exploits the distinctive characteristics of design methodologies to reverse-engineer the architecture of building blocks in compact DNNs. We show the efficacy of “DeepPeep” on P100 and P4000 GPUs. Additionally, we propose intelligent design maneuvering strategies for thwarting IP theft through the DeepPeep attack and proposed “Secure MobileNet-V1.” Interestingly, compared to vanilla MobileNet-V1, secure MobileNet-V1 provides a significant reduction in inference latency (≈60%) and improvement in predictive performance (≈2%) with very low memory and computation overheads.

References

  1. Jeremy Howard. 2019. imagenette. Retrieved from https://github.com/fastai/imagenette.Google ScholarGoogle Scholar
  2. Hal Conick. 2017. The past, present and future of AI in marketing. Marketing News 51, 1 (2017), 26--35.Google ScholarGoogle Scholar
  3. CUDA Nvidia. 2008. Cublas library. NVIDIA Corporation, Santa Clara, California 15, 27 (2008), 31.Google ScholarGoogle Scholar
  4. Dario Amodei and Danny Hernandez. 2018. AI and Compute. Retrieved from https://openai.com/blog/ai-and-compute/.Google ScholarGoogle Scholar
  5. Robert A. Bridges, Neena Imam, and Tiffany M. Mintz. 2016. Understanding GPU power: A survey of profiling, modeling, and simulation methods. ACM Comput. Surv. 49, 3 (2016), 1--27.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Victor Costan and Srinivas Devadas. 2016. Intel SGX explained. IACR Cryptol. ePrint Arch. 2016, 86 (2016), 1--118.Google ScholarGoogle Scholar
  7. Victor Costan, Ilia A. Lebedev, and Srinivas Devadas. 2016. Sanctum: Minimal hardware extensions for strong software isolation. In Proceedings of the USENIX Security Symposium.Google ScholarGoogle Scholar
  8. Alessandro Del Sole. 2018. Introducing Microsoft cognitive services. In Microsoft Computer Vision APIs Distilled. Springer, 1--4.Google ScholarGoogle Scholar
  9. Amir Gholami, Kiseok Kwon, Bichen Wu, Zizheng Tai, Xiangyu Yue, Peter H. Jin, Sicheng Zhao, and Kurt Keutzer. 2018. SqueezeNext: Hardware-aware neural network design. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops.Google ScholarGoogle ScholarCross RefCross Ref
  10. Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. 2014. Explaining and harnessing adversarial examples. Arxiv Preprint Arxiv:1412.6572 (2014).Google ScholarGoogle Scholar
  11. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16).Google ScholarGoogle ScholarCross RefCross Ref
  12. Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. MobileNets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017).Google ScholarGoogle Scholar
  13. Jie Hu, Li Shen, and Gang Sun. 2018. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 7132--7141.Google ScholarGoogle ScholarCross RefCross Ref
  14. W. Hua, Z. Zhang, and G. E. Suh. 2018. Reverse engineering convolutional neural networks through side-channel information leaks. In Proceedings of the Design Automation Conference. 1--6.Google ScholarGoogle Scholar
  15. G. Huang, Z. Liu, L. van der Maaten, and K. Q. Weinberger. 2017. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 2261--2269.Google ScholarGoogle Scholar
  16. Forrest N. Iandola, Matthew W. Moskewicz, Khalid Ashraf, Song Han, William J. Dally, and Kurt Keutzer. 2016. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and MB model size. CoRR abs/1602.07360 (2016).Google ScholarGoogle Scholar
  17. N. K. Jha, S. Mittal, and G. Mattela. 2019. The ramifications of making deep neural networks compact. In Proceedings of the 32nd International Conference on VLSI Design and 18th International Conference on Embedded Systems (VLSID’19). 215--220.Google ScholarGoogle Scholar
  18. Nandan Kumar Jha, Rajat Saini, Subhrajit Nag, and Sparsh Mittal. 2020. E2GC: Energy-efficient group convolution in deep neural networks. In Proceedings of the 33rd International Conference on VLSI Design and 19th International Conference on Embedded Systems (VLSID’20). IEEE, 155--160.Google ScholarGoogle ScholarCross RefCross Ref
  19. Yangqing Jia, Evan Shelhamer, Jeff Donahue, Sergey Karayev, Jonathan Long, Ross Girshick, Sergio Guadarrama, and Trevor Darrell. 2014. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the International Conference on Multimedia (MM’14). 675--678.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. Alex Kaplunovich and Yelena Yesha. 2017. Cloud big data decision support system for machine learning on AWS: Analytics of analytics. In Proceedings of the IEEE International Conference on Big Data (Big Data’17). IEEE, 3508--3516.Google ScholarGoogle ScholarCross RefCross Ref
  21. B. Kehoe, S. Patil, P. Abbeel, and K. Goldberg. 2015. A survey of research on cloud robotics and automation. IEEE Trans. Automat. Sci. Eng. 12, 2 (Apr. 2015), 398--409.Google ScholarGoogle ScholarCross RefCross Ref
  22. Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. 2012. CIFAR-100 (Canadian Institute for Advanced Research). (2012). Retrieved from http://www.cs.toronto.edu/ kriz/cifar.html.Google ScholarGoogle Scholar
  23. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ImageNet classification with deep convolutional neural networks. In Proceedings of the Conference on Neural Information Processing Systems. 1097--1105.Google ScholarGoogle Scholar
  24. Liangzhen Lai, Naveen Suda, and Vikas Chandra. 2018. Not all ops are created equal! arXiv preprint arXiv:1801.04326 (2018).Google ScholarGoogle Scholar
  25. Wei Li, Kai Liu, Lin Yan, Fei Cheng, YunQiu Lv, and LiZhe Zhang. 2019. FRD-CNN: Object detection based on small-scale convolutional neural networks and feature reuse. Sci. Rep. 9, 1 (2019), 1--12.Google ScholarGoogle Scholar
  26. Y. Li and A. Pedram. 2017. CATERPILLAR: Coarse grain reconfigurable architecture for accelerating the training of deep neural networks. In Proceedings of the IEEE International Conference on Application-specific Systems. 1--10.Google ScholarGoogle Scholar
  27. Qi Liu, Tao Liu, Zihao Liu, Yanzhi Wang, Yier Jin, and Wujie Wen. 2018. Security analysis and enhancement of model compressed deep learning systems under adversarial attacks. In Proceedings of the Asia and South Pacific Design Automation Conference. 721--726.Google ScholarGoogle ScholarCross RefCross Ref
  28. T. Liu, W. Wen, and Y. Jin. 2018. SIN2: Stealth infection on neural network–A low-cost agile neural Trojan attack methodology. In Proceedings of the IEEE International Symposium on Hardware Oriented Security and Trust. 227--230.Google ScholarGoogle Scholar
  29. Yanpei Liu, Xinyun Chen, Chang Liu, and Dawn Song. 2016. Delving into transferable adversarial examples and black-box attacks. Arxiv Preprint Arxiv:1611.02770 (2016).Google ScholarGoogle Scholar
  30. Yunhui Long, Vincent Bindschaedler, Lei Wang, Diyue Bu, Xiaofeng Wang, Haixu Tang, Carl A. Gunter, and Kai Chen. 2018. Understanding membership inferences on well-generalized learning models. CoRR abs/1802.04889 (2018).Google ScholarGoogle Scholar
  31. Ningning Ma, Xiangyu Zhang, Hai-Tao Zheng, and Jian Sun. 2018. ShuffleNet V2: Practical guidelines for efficient CNN architecture design. In Proceedings of the European Conference on Computer Vision.Google ScholarGoogle ScholarCross RefCross Ref
  32. Asit K. Mishra, Joseph L. Hellerstein, Walfredo Cirne, and Chita R. Das. 2010. Towards characterizing cloud backend workloads: Insights from Google compute clusters. ACM SIGMETRICS Perf. Eval. Rev. 37, 4 (2010), 34--41.Google ScholarGoogle ScholarDigital LibraryDigital Library
  33. Sparsh Mittal, S. B. Abhinaya, Manish Reddy, and Irfan Ali. 2018. A survey of techniques for improving security of GPUs. Hardw. Syst. Secur. J. 2, 3 (2018), 266--285.Google ScholarGoogle ScholarCross RefCross Ref
  34. Seong Joon Oh, Max Augustin, Mario Fritz, and Bernt Schiele. 2018. Towards reverse-engineering black-box neural networks. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  35. Bita Darvish Rouhani, Huili Chen, and Farinaz Koushanfar. 2019. DeepSigns: A generic watermarking framework for IP protection of deep learning models. In Proceedings of the International Conference on Architectural Support for Programming Languages and Operating Systems.Google ScholarGoogle Scholar
  36. Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. 2015. Imagenet large scale visual recognition challenge. Int. J. Comput. Vision 115, 3 (2015), 211--252.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L. Chen. 2018. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 4510--4520.Google ScholarGoogle Scholar
  38. R. Shokri, M. Stronati, C. Song, and V. Shmatikov. 2017. Membership inference attacks against machine learning models. In Proceedings of the IEEE Symposium on Security and Privacy (SP’17).Google ScholarGoogle Scholar
  39. Ion Stoica, Dawn Song, Raluca Ada Popa, David Patterson, Michael W. Mahoney, Randy Katz, Anthony D. Joseph, Michael Jordan, Joseph M. Hellerstein, Joseph E. Gonzalez, et al. 2017. A Berkeley view of systems challenges for AI. Arxiv Preprint Arxiv:1712.05855 (2017).Google ScholarGoogle Scholar
  40. Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. 2015. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  41. Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  42. Florian Tramer and Dan Boneh. 2019. Slalom: Fast, verifiable and private execution of neural networks in trusted hardware. In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  43. Florian Tramèr, Fan Zhang, Ari Juels, Michael K. Reiter, and Thomas Ristenpart. 2016. Stealing machine learning models via prediction APIs. In Proceedings of the 25th USENIX Security Symposium (SEC’16). 601--618.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. B. Wang and N. Z. Gong. 2018. Stealing hyperparameters in machine learning. In Proceedings of the IEEE Symposium on Security and Privacy (SP’18).Google ScholarGoogle Scholar
  45. Linnan Wang, Jinmian Ye, Yiyang Zhao, Wei Wu, Ang Li, Shuaiwen Leon Song, Zenglin Xu, and Tim Kraska. 2018. Superneurons: Dynamic GPU memory management for training deep neural networks. In Proceedings of the 23rd ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming. 41--53.Google ScholarGoogle ScholarDigital LibraryDigital Library
  46. Lilian Weng. 2018. Attention? Attention! Retrieved from https://lilianweng.github.io/lil-log/2018/06/24/attention-attention.html.Google ScholarGoogle Scholar
  47. Bichen Wu, Forrest Iandola, Peter H. Jin, and Kurt Keutzer. 2017. Squeezedet: Unified, small, low power fully convolutional neural networks for real-time object detection for autonomous driving. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. 129--137.Google ScholarGoogle ScholarCross RefCross Ref
  48. Saining Xie, Ross B. Girshick, Piotr Dollár, Zhuowen Tu, and Kaiming He. 2017. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5987--5995.Google ScholarGoogle ScholarCross RefCross Ref
  49. Saining Xie, Alexander Kirillov, Ross Girshick, and Kaiming He. 2019. Exploring randomly wired neural networks for image recognition. In Proceedings of the IEEE International Conference on Computer Vision (ICCV’19).Google ScholarGoogle ScholarCross RefCross Ref
  50. Keyulu Xu, Jingling Li, Mozhi Zhang, Simon S. Du, Ken-ichi Kawarabayashi, and Stefanie Jegelka. 2020. What can neural networks reason about? In Proceedings of the International Conference on Learning Representations.Google ScholarGoogle Scholar
  51. Mengjia Yan, Christopher Fletcher, and Josep Torrellas. 2020. Cache telepathy: Leveraging shared resource attacks to learn DNN architectures. In Proceedings of the 29th USENIX Security Symposium (USENIX Security’20).Google ScholarGoogle Scholar
  52. Sergey Zagoruyko and Nikos Komodakis. 2016. Wide residual networks. In Proceedings of the British Machine Vision Conference.Google ScholarGoogle ScholarCross RefCross Ref
  53. Xiangyu Zhang, Xinyu Zhou, Mengxiao Lin, and Jian Sun. 2018. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle ScholarCross RefCross Ref
  54. Shijun Zhao, Qianying Zhang, Guangyao Hu, Yu Qin, and Dengguo Feng. 2014. Providing root of trust for ARM trustzone using on-chip SRAM. In Proceedings of the 4th International Workshop on Trustworthy Embedded Devices (TrustED’14). 25--36.Google ScholarGoogle ScholarDigital LibraryDigital Library
  55. Barret Zoph, Vijay Vasudevan, Jonathon Shlens, and Quoc V. Le. 2018. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.Google ScholarGoogle Scholar

Index Terms

  1. DeepPeep: Exploiting Design Ramifications to Decipher the Architecture of Compact DNNs

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image ACM Journal on Emerging Technologies in Computing Systems
        ACM Journal on Emerging Technologies in Computing Systems  Volume 17, Issue 1
        January 2021
        232 pages
        ISSN:1550-4832
        EISSN:1550-4840
        DOI:10.1145/3425108
        • Editor:
        • Ramesh Karri
        Issue’s Table of Contents

        Copyright © 2020 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 28 October 2020
        • Revised: 1 July 2020
        • Accepted: 1 July 2020
        • Received: 1 August 2019
        Published in jetc Volume 17, Issue 1

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format