
Research Article
Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning
@INPROCEEDINGS{10.1007/978-3-030-91424-0_2, author={Chengyao Xu and Shangqi Lai}, title={Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning}, proceedings={Quality, Reliability, Security and Robustness in Heterogeneous Systems. 17th EAI International Conference, QShine 2021, Virtual Event, November 29--30, 2021, Proceedings}, proceedings_a={QSHINE}, year={2021}, month={11}, keywords={}, doi={10.1007/978-3-030-91424-0_2} }
- Chengyao Xu
Shangqi Lai
Year: 2021
Accelerating TEE-Based DNN Inference Using Mean Shift Network Pruning
QSHINE
Springer
DOI: 10.1007/978-3-030-91424-0_2
Abstract
In recent years, deep neural networks (DNNs) have achieved great success in many areas and have been deployed as cloud services to bring convenience to people’s daily lives. However, the widespread use of DNNs in the cloud brings critical privacy concerns. Researchers have proposed many solutions to address the privacy concerns of deploying DNN in the cloud, and one major category of solutions rely on a trusted execution environment (TEE). Nonetheless, the DNN inference requires extensive memory and computing resources to achieve accurate decision-making, which does not operate well in TEE with restricted memory space. This paper proposes a network pruning algorithm based on mean shift clustering to reduce the model size and improve the inference performance in TEE. The core idea of our design is to use a mean shift algorithm to aggregate the weight values automatically and prune the network based on the distance between the weight and center. Our experiments prune three popular networks on the CIFAR-10 dataset. The experimental results show that our algorithm successfully reduces the network size without affecting its accuracy. The inference in TEE is accelerated by 20%.