Privacy-preserving Offloading in Edge Intelligence Systems with Inductive Learning and Local Differential Privacy

Publication Name

IEEE Transactions on Network and Service Management


We address privacy and latency issues in edge-cloud computing environments where the neural network training is centralized. This paper considers the scenario where the edge devices are the only data sources for the deep learning model to be trained on the central server. Improper access to the massive amounts of data generated by edge devices could lead to privacy concerns. As a result, existing solutions for preserving privacy and reducing network latency in the edge environment rely on auxiliary datasets with no privacy risks or pre-trained models to build the client side feature extractor. However, finding auxiliary datasets or pre-trained models is not always guaranteed and may be challenging. To bridge this gap and eliminate the reliance on auxiliary datasets or pre-trained models of existing solutions, this paper presents DeepGuess, a privacy-preserving and latency-aware deep-learning framework. DeepGuess introduces a new learning mechanism enabled by the AutoEncoder architecture: inductive learning. With inductive learning, sensitive data stays on devices and is not explicitly sent to the central server to engage in back-propagations. To further enhance privacy, we propose a new local differential privacy algorithm that allows edge devices to apply random noise to features extracted from their sensitive data before being transferred to the non-trusted central server. The experimental evaluation of DeepGuess with various datasets and in a real-world scenario shows that our solution achieves comparable or even higher accuracy than existing solutions while reducing data transfer over the network by more than 50%.

Open Access Status

This publication is not available as open access



Link to publisher version (DOI)