Abstract

A user generates <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula> independent and identically distributed data random variables with a probability mass function that must be guarded from a querier. The querier must recover, with a prescribed accuracy, a given function of the data from each of <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula> independent and identically distributed query responses upon eliciting them from the user. The user chooses the data probability mass function and devises the random query responses to maximize distribution privacy as gauged by the (Kullback-Leibler) divergence between the former and the querier&#x2019;s best estimate of it based on the <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula> query responses. Considering an arbitrary function, a basic achievable lower bound for distribution privacy is provided that does not depend on <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula> and corresponds to worst-case privacy. Worst-case privacy equals the logsum cardinalities of inverse atoms under the given function, with the number of summands decreasing as the querier recovers the function with improving accuracy. Next, upper (converse) and lower (achievability) bounds for distribution privacy, dependent on <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula>, are developed. The former improves upon worst-case privacy and the latter does so under suitable assumptions; both converge to it as <inline-formula> <tex-math notation="LaTeX">$n$ </tex-math></inline-formula> grows. The converse and achievability proofs identify explicit strategies for the user and the querier.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call