Abstract

Available data may differ from true data in many cases due to sensing errors, especially for the Internet of Things (IoT). Although privacy-preserving data mining has been widely studied during the last decade, little attention has been paid to data values containing errors. Differential privacy, which is the de facto standard privacy metric, can be achieved by adding noise to a target value that must be protected. However, if the target value already contains errors, there is no reason to add extra noise. In this paper, a novel privacy model called true-value-based differential privacy (TDP) is proposed. This model applies traditional differential privacy to the “true value” unknown by the data owner or anonymizer but not to the “measured value” containing errors. Based on TDP, the amount of noise added by differential privacy techniques can be reduced by approximately 20% by our solution. As a result, the error of generated histograms can be reduced by 40.4% and 29.6% on average according to mean square error and Jensen–Shannon divergence, respectively. We validate this result on synthetic and five real data sets. Moreover, we proved that the privacy protection level does not decrease as long as the measurement error is not overestimated.

Highlights

  • Significant amounts of Internet of Things (IoT) data are generated every day by many different sensors, such as thermal cameras, home appliance sensors, automotive sensors, and smartphone-equipped sensors

  • CONTRIBUTION First, we propose true-value-based differential privacy (TDP), which is a novel concept of differential privacy considering sensing noise

  • Our research aims at protecting the true value

Read more

Summary

Introduction

Significant amounts of IoT data are generated every day by many different sensors, such as thermal cameras, home appliance sensors, automotive sensors, and smartphone-equipped sensors. These IoT data can be used for health monitoring [1], context-aware recommendation (or recommender) systems [2], navigation [3], and other applications. Differential privacy has been used in many studies, such as [7]–[9], as it is one of the most critical privacy metrics [10] It is considered an important concept for data analysis [11], [12]

Objectives
Results
Discussion
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call