About | Contact Us | Register | Login
ProceedingsSeriesJournalsSearchEAI
Quality, Reliability, Security and Robustness in Heterogeneous Systems. 17th EAI International Conference, QShine 2021, Virtual Event, November 29–30, 2021, Proceedings

Research Article

Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning

Download(Requires a free EAI acccount)
5 downloads
Cite
BibTeX Plain Text
  • @INPROCEEDINGS{10.1007/978-3-030-91424-0_2,
        author={Chengyao Xu and Shangqi Lai},
        title={Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning},
        proceedings={Quality, Reliability, Security and Robustness in Heterogeneous Systems. 17th EAI International Conference, QShine 2021, Virtual Event, November 29--30, 2021, Proceedings},
        proceedings_a={QSHINE},
        year={2021},
        month={11},
        keywords={},
        doi={10.1007/978-3-030-91424-0_2}
    }
    
  • Chengyao Xu
    Shangqi Lai
    Year: 2021
    Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning
    QSHINE
    Springer
    DOI: 10.1007/978-3-030-91424-0_2
Chengyao Xu1, Shangqi Lai1,*
  • 1: Faculty of Information Technology
*Contact email: shangqi.lai@monash.edu

Abstract

In recent years, deep neural networks (DNNs) have achieved great success in many areas and have been deployed as cloud services to bring convenience to people’s daily lives. However, the widespread use of DNNs in the cloud brings critical privacy concerns. Researchers have proposed many solutions to address the privacy concerns of deploying DNN in the cloud, and one major category of solutions rely on a trusted execution environment (TEE). Nonetheless, the DNN inference requires extensive memory and computing resources to achieve accurate decision-making, which does not operate well in TEE with restricted memory space. This paper proposes a network pruning algorithm based on mean shift clustering to reduce the model size and improve the inference performance in TEE. The core idea of our design is to use a mean shift algorithm to aggregate the weight values automatically and prune the network based on the distance between the weight and center. Our experiments prune three popular networks on the CIFAR-10 dataset. The experimental results show that our algorithm successfully reduces the network size without affecting its accuracy. The inference in TEE is accelerated by 20%.

Published
2021-11-17
Appears in
SpringerLink
http://dx.doi.org/10.1007/978-3-030-91424-0_2
Copyright © 2021–2025 ICST
EBSCOProQuestDBLPDOAJPortico
EAI Logo

About EAI

  • Who We Are
  • Leadership
  • Research Areas
  • Partners
  • Media Center

Community

  • Membership
  • Conference
  • Recognition
  • Sponsor Us

Publish with EAI

  • Publishing
  • Journals
  • Proceedings
  • Books
  • EUDL