Abstract
The huge amount of sensory data collected from mobile devices has offered great potentials to promote more significant services based on user data extracted from sensor readings. However, releasing user data could also seriously threaten user privacy. It is possible to directly collect sensitive information from released user data without user permissions. Furthermore, third party users can also infer sensitive information contained in released data in a latent manner by utilizing data mining techniques. In this paper, we formally define these two types of threats as inherent data privacy and latent data privacy and construct a data-sanitization strategy that can optimize the tradeoff between data utility and customized two types of privacy. The key novel idea lies that the developed strategy can combat against powerful third party users with broad knowledge about users and launching optimal inference attacks. We show that our strategy does not reduce the benefit brought by user data much, while sensitive information can still be protected. To the best of our knowledge, this is the first work that preserves both inherent data privacy and latent data privacy.
Similar content being viewed by others
References
Enck W, Gilbert P, Chun B-G, Cox LP, Jung J, McDaniel P, Sheth AN (2010) Taintdroid: an information-flow tracking system for realtime privacy monitoring on smartphones. In: Proceedings of the 9th USENIX conference on operating systems design and implementation, pp 1–6
http://abcnews.go.com/Technology/gaydar-facebook-friends/story?id=8633224
Stevens J (1986) Applied multivariate statistics for the social sciences. L. Erlbaum Associates Inc, Hillsdale
Shokri R, Theodorakopoulos G, Troncoso C, Hubaux J-P, Le Boudec J-Y (2012) Protecting location privacy: optimal strategy against localization attacks. In: Proceedings of the 2012 ACM conference on computer and communications security, pp 617–627
Salamatian S, Zhang A, Calmon F, Bhamidipati S, Fawaz N, Kveton B, Oliveira P, Taft N (2013) How to hide the elephant- or the donkey- in the room: practical privacy against statistical inference for large data. In: Global conference on signal and information processing (GlobalSIP), 2013 IEEE, pp 269–272
Kosinski M, Stillwell D, Graepel T (2013) Private traits and attributes are predictable from digital records of human behavior. Proc Natl Acad Sci 10(15):5802–5805
Barbaro M, Zeller T (2006) A face is exposed for AOL searcher. N Y Times 9:2008
Sweeney L (2002) k-anonymity: a model for protecting privacy. Int J Uncertain Fuzziness Knowl Based Syst 10(5):557–570
Zhang L, Wang X, Lu J, Li P, Cai Z (2016) An efficient privacy preserving data aggregation approach for mobile sensing. Secur Commun Netw 9(16):3844–53
Zhang L, Cai Z, Wang X (2016) Fakemask: a novel privacy preserving approach for smartphones. IEEE Trans Netw Serv Manag 13(2):335–348
Wang Y, Cai Z, Ying G, Gao Y, Tong X, Wu G (2016) An incentive mechanism with privacy protection in mobile crowdsourcing systems. Comput Netw 102:157–171
He Z, Cai Z, Yu J, Wang X, Sun Y, Li Y (2016) Cost-efficient strategies for restraining rumor spreading in mobile social networks. IEEE Trans Veh Technol PP(99):1–1
Wang Y, Cai Z, Yin G, Gao Y, Tong X, Han Q (2016) A game theory-based trust measurement model for social networks. Comput Soc Netw 3(1):2
Li N, Li T, Venkatasubramanian S (2007) t-closeness: privacy beyond k-anonymity and l-diversity. In: ICDE, pp 106–115
Machanavajjhala A, Kifer D, Gehrke J, Venkitasubramaniam M (2007) L-diversity: privacy beyond k-anonymity. ACM Trans Knowl Discov Data 1(1):1–12
Gruteser M, Grunwald D (2003) Anonymous usage of location-based services through spatial and temporal cloaking. In: Proceedings of the 1st international conference on mobile systems, applications and services, pp 31–42
Bordenabe NE, Chatzikokolakis K, Palamidessi C (2014) Optimal geo-indistinguishable mechanisms for location privacy. In: Proceedings of the 2014 ACM SIGSAC conference on computer and communications security, ser. CCS ’14. ACM, New York, pp 251–262
Li N, Qardaji W, Su D, Wu Y, Yang W (2013) Membership privacy: a unifying framework for privacy definitions. In: Proceedings of the 2013 ACM SIGSAC conference on computer and communications security, pp 889–900
Tramer F, Huang Z, Ayday E, Hubaux J-P (2015) Differential privacy with bounded priors: reconciling utility and privacy in genome-wide association studies. In: 22nd ACM conference on computer and communications security
He Z, Cai Z, Han Q, Tong W, Sun L, Li Y (2016) An energy efficient privacy-preserving content sharing scheme in mobile social networks. Pers Ubiquit Comput 20(5):833–846
Ioannidis S, Montanari A, Weinsberg U, Bhagat S, Fawaz N, Taft N (2014) Privacy tradeoffs in predictive analytics. SIGMETRICS Perform Eval Rev 42(1):57–69
Chaabane A, Acs G, Kaafar MA (2012) You are what you like! information leakage through users’ interests. In: NDSS
Fung BCM, Wang K, Chen R, Yu PS (2010) Privacy-preserving data publishing: a survey of recent developments. ACM Comput Surv 42(4):1–14
Li T, Li N, Zhang J, Molloy I (2012) Slicing: a new approach for privacy preserving data publishing. IEEE Trans Knowl Data Eng 24(3):561–574
He Z, Cai Z, Li Y (2016) Customized privacy preserving for classification based applications. In: Workshop on privacy-aware mobile computing (PAMCO), pp 37–42
Han M, Li J, Cai Z, Han Q (2016) Privacy reserved influence maximization in gps-enabled cyber-physical and online social networks. In: The 9th IEEE international conference on social computing and networking
Allard T, Hébrail G, Masseglia F, Pacitti E (2015) Chiaroscuro: transparency and privacy for massive personal time-series clustering. In: Proceedings of the 2015 ACM SIGMOD international conference on management of data, New York, NY, USA, pp 779–794
Shokri R, Shmatikov V (2015) Privacy-preserving deep learning. In: Proceedings of the 22nd ACM SIGSAC conference on computer and communications security, pp 1310–1321
Hu X, Yuan M, Yao J, Deng Y, Chen L, Yang Q, Guan H, Zeng J (2015) Differential privacy in telco big data platform. Proc VLDB Endow 8(12):1692–1703
Cai Z, He Z, Guan X, Li Y (2016) Collective data-sanitization for preventing sensitive information inference attacks in social networks. IEEE Trans Dependable Secure Comput PP(99):1–1
Shokri R, Theodorakopoulos G, Le Boudec J-Y, Hubaux J-P (2011) Quantifying location privacy. In: 2011 IEEE symposium on security and privacy (SP), pp 247–262
Dwork C, McSherry F, Nissim K, Smith A (2006) Calibrating noise to sensitivity in private data analysis. In: Proceedings of the third conference on theory of cryptography, pp 265–284
Acknowledgements
This work is partly supported by the National Science Foundation (NSF) under Grant No. CNS-1252292, NSF of China under Contract 61373083, 61370084, 61502116, 61371185, and 61373027.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
He, Z., Cai, Z., Sun, Y. et al. Customized privacy preserving for inherent data and latent data. Pers Ubiquit Comput 21, 43–54 (2017). https://doi.org/10.1007/s00779-016-0972-2
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00779-016-0972-2