Abstract

Differential privacy is nowadays considered the "gold standard" when releasing information, e.g., statistics, on sensitive data. To avoid leaking too much sensitive data noise-adding mechanisms may be used. ϵ-differential privacy measures the amount of privacy ϵ that such a mechanism ensures. Of course, adding too much noise results into useless, random information, while adding not enough may lead to privacy violations. This problem is known as the privacy-utility trade-off and raises a natural optimality question: how can we maximise utility for a given amount of privacy ϵ.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.