| Literature DB >> 31435181 |
Yang Cao1, Masatoshi Yoshikawa2, Yonghui Xiao3, Li Xiong1.
Abstract
Differential Privacy (DP) has received increasing attention as a rigorous privacy framework. Many existing studies employ traditional DP mechanisms (e.g., the Laplace mechanism) as primitives to continuously release private data for protecting privacy at each time point (i.e., event-level privacy), which assume that the data at different time points are independent, or that adversaries do not have knowledge of correlation between data. However, continuously generated data tend to be temporally correlated, and such correlations can be acquired by adversaries. In this paper, we investigate the potential privacy loss of a traditional DP mechanism under temporal correlations. First, we analyze the privacy leakage of a DP mechanism under temporal correlation that can be modeled using Markov Chain. Our analysis reveals that, the event-level privacy loss of a DP mechanism may increase over time. We call the unexpected privacy loss temporal privacy leakage (TPL). Although TPL may increase over time, we find that its supremum may exist in some cases. Second, we design efficient algorithms for calculating TPL. Third, we propose data releasing mechanisms that convert any existing DP mechanism into one against TPL. Experiments confirm that our approach is efficient and effective.Entities:
Keywords: Differential privacy; Markov model; correlated data; streaming data; time series
Year: 2018 PMID: 31435181 PMCID: PMC6704013 DOI: 10.1109/TKDE.2018.2824328
Source DB: PubMed Journal: IEEE Trans Knowl Data Eng ISSN: 1041-4347 Impact factor: 6.977