General inferential limits under differential and Pufferfish privacy

  • Abstract
  • Literature Map
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

General inferential limits under differential and Pufferfish privacy

Similar Papers
  • Conference Article
  • Cite Count Icon 6
  • 10.1109/bigcom.2019.00029
Real-location Reporting Based Differential Privacy Trajectory Protection for Mobile Crowdsensing
  • Aug 1, 2019
  • Xin Chen + 4 more

Mobile Crowdsensing (MCS) has become an effective technology for urban data sensing and acquisition. But this also brings the risk of trajectory privacy disclosure for participants. Most of the existing efforts attempt to add noise into the reported location information to achieve the trajectory privacy protection of the participating users. However, in many scenarios, the participants are required to report the real-location information (e.g., high-quality map generation, traffic flow monitoring, etc.). To address this problem, we propose a differential privacy based trajectory privacy protection scheme with real-location reporting in MCS. First, we present the definition of trajectory privacy protection based on real path reporting under differential privacy. Second, we give a differential trajectory privacy protection framework to achieve participants trajectory privacy protection under Bayesian inference attacks. Finally, we analyze and prove that differential trajectory privacy problem is an NP-Hard problem. Meanwhile, we also design an approximate algorithm to report participants road segment with the trajectory privacy guarantee. The experimental results on both the simulated data set and the real data set show that our proposed participant's trajectory privacy protection scheme has a good performance.

  • Book Chapter
  • Cite Count Icon 17
  • 10.1007/978-3-319-05912-9_7
What Should We Protect? Defining Differential Privacy for Social Network Analysis
  • Jan 1, 2014
  • Christine Task + 1 more

Privacy of social network data is a growing concern that threatens to limit access to this valuable data source. Analysis of the graph structure of social networks can provide valuable information for revenue generation and social science research, but unfortunately, ensuring this analysis does not violate individual privacy is difficult. Simply anonymizing graphs or even releasing only aggregate results of analysis may not provide sufficient protection. Differential privacy is an alternative privacy model, popular in data-mining over tabular data, that uses noise to obscure individuals’ contributions to aggregate results and offers a very strong mathematical guarantee that individuals’ presence in the data-set is hidden. Analyses that were previously vulnerable to identification of individuals and extraction of private data may be safely released under differential-privacy guarantees. We review two existing standards for adapting differential privacy to network data and analyze the feasibility of several common social-network analysis techniques under these standards. Additionally, we propose out-link privacy and partition privacy, novel standards for differential privacy over network data, and introduce powerful private algorithms for common network analysis techniques that were infeasible to privatize under previous differential privacy standards.

  • Conference Article
  • Cite Count Icon 23
  • 10.1145/3373718.3394796
Deciding Differential Privacy for Programs with Finite Inputs and Outputs
  • Jul 8, 2020
  • Gilles Barthe + 4 more

Differential privacy is a de facto standard for statistical computations over databases that contain private data. Its main and rather surprising strength is to guarantee individual privacy and yet allow for accurate statistical results. Thanks to its mathematical definition, differential privacy is also a natural target for formal analysis. A broad line of work develops and uses logical methods for proving privacy. A more recent and complementary line of work uses statistical methods for finding privacy violations. Although both lines of work are practically successful, they elide the fundamental question of decidability. This paper studies the decidability of differential privacy. We first establish that checking differential privacy is undecidable even if one restricts to programs having a single Boolean input and a single Boolean output. Then, we define a non-trivial class of programs and provide a decision procedure for checking the differential privacy of a program in this class. Our procedure takes as input a program P parametrized by a privacy budget ϵ and either establishes the differential privacy for all possible values of ϵ or generates a counter-example. In addition, our procedure works for both to ϵ-differential privacy and (ϵ, δ)-differential privacy. Technically, the decision procedure is based on a novel and judicious encoding of the semantics of programs in our class into a decidable fragment of the first-order theory of the reals with exponentiation. We implement our procedure and use it for (dis)proving privacy bounds for many well-known examples, including randomized response, histogram, report noisy max and sparse vector.

  • Research Article
  • Cite Count Icon 2
  • 10.3233/jcs-230036
Universal optimality and robust utility bounds for metric differential privacy1
  • Oct 13, 2023
  • Journal of Computer Security
  • Natasha Fernandes + 3 more

We study the privacy-utility trade-off in the context of metric differential privacy. Ghosh et al. introduced the idea of universal optimality to characterise the “best” mechanism for a certain query that simultaneously satisfies (a fixed) ε-differential privacy constraint whilst at the same time providing better utility compared to any other ε-differentially private mechanism for the same query. They showed that the Geometric mechanism is universally optimal for the class of counting queries. On the other hand, Brenner and Nissim showed that outside the space of counting queries, and for the Bayes risk loss function, no such universally optimal mechanisms exist. Except for the universal optimality of the Laplace mechanism, there have been no generalisations of these universally optimal results to other classes of differentially-private mechanisms. In this paper, we use metric differential privacy and quantitative information flow as the fundamental principle for studying universal optimality. Metric differential privacy is a generalisation of both standard (i.e., central) differential privacy and local differential privacy, and it is increasingly being used in various application domains, for instance in location privacy and in privacy-preserving machine learning. Similar to the approaches adopted by Ghosh et al. and Brenner and Nissim, we measure utility in terms of loss functions, and we interpret the notion of a privacy mechanism as an information-theoretic channel satisfying constraints defined by ε-differential privacy and a metric meaningful to the underlying state space. Using this framework we are able to clarify Nissim and Brenner’s negative results by (a) that in fact all privacy types contain optimal mechanisms relative to certain kinds of non-trivial loss functions, and (b) extending and generalising their negative results beyond Bayes risk specifically to a wide class of non-trivial loss functions. Our exploration suggests that universally optimal mechanisms are indeed rare within privacy types. We therefore propose weaker universal benchmarks of utility called privacy type capacities. We show that such capacities always exist and can be computed using a convex optimisation algorithm. Further, we illustrate these ideas on a selection of examples with several different underlying metrics.

  • Research Article
  • Cite Count Icon 1
  • 10.1109/jsait.2021.3056359
Low Influence, Utility, and Independence in Differential Privacy: A Curious Case of (3 2)
  • Feb 2, 2021
  • IEEE Journal on Selected Areas in Information Theory
  • Rafael G L D'Oliveira + 3 more

We study the relationship between randomized low influence functions and differentially private mechanisms. Our main aim is to formally determine whether differentially private mechanisms are low influence and whether low influence randomized functions can be differentially private. We show that differential privacy does not necessarily imply low influence in a formal sense. However, low influence implies approximate differential privacy. These results hold for both independent and non-independent randomized mechanisms, where an important instance of the former is the widely-used additive noise techniques in the differential privacy literature. Our study also reveals the interesting dynamics between utility, low influence, and independence of a differentially private mechanism. As the name of this article suggests, we show that any two such features are simultaneously possible. However, in order to have a differentially private mechanism that has both utility and low influence, even under a very mild utility condition, one has to employ non-independent mechanisms.

  • Conference Article
  • Cite Count Icon 3
  • 10.1145/3411501.3419431
SVM Learning for Default Prediction of Credit Card under Differential Privacy
  • Nov 9, 2020
  • Jianping Cai + 2 more

Currently, financial institutions utilize personal sensitive information extensively in machine learning. It results in significant privacy risks to customers. As an essential standard of privacy, differential privacy is often applied to machine learning in recent years. To establish a prediction model of credit card default under the premise of protecting personal privacy, we consider the problems of customer data contribution difference and data sample distribution imbalance, propose weighted SVM algorithm under differential privacy. Through theoretical analysis, we have ensured the security of differential privacy. The algorithm solves the problem of prediction result deviation caused by sample distribution imbalance and effectively reduces the data sensitivity and noise error. The experimental results show that the algorithm proposed in this paper can accurately predict whether a customer is default while protecting personal privacy.

  • Research Article
  • Cite Count Icon 11
  • 10.1007/s12630-023-02557-5
Interpreting frequentist hypothesis tests: insights from Bayesian inference
  • Oct 1, 2023
  • Canadian Journal of Anesthesia/Journal canadien d'anesthésie
  • David Sidebotham + 3 more

Randomized controlled trials are one of the best ways of quantifying the effectiveness of medical interventions. Therefore, when the authors of a randomized superiority trial report that differences in the primary outcome between the intervention group and the control group are “significant” (i.e., P ≤ 0.05), we might assume that the intervention has an effect on the outcome. Similarly, when differences between the groups are “not significant,” we might assume that the intervention does not have an effect on the outcome. Nevertheless, both assumptions are frequently incorrect.In this article, we explore the relationship that exists between real treatment effects and declarations of statistical significance based on P values and confidence intervals. We explain why, in some circumstances, the chance an intervention is ineffective when P ≤ 0.05 exceeds 25% and the chance an intervention is effective when P > 0.05 exceeds 50%.Over the last decade, there has been increasing interest in Bayesian methods as an alternative to frequentist hypothesis testing. We provide a robust but nontechnical introduction to Bayesian inference and explain why a Bayesian posterior distribution overcomes many of the problems associated with frequentist hypothesis testing.Notwithstanding the current interest in Bayesian methods, frequentist hypothesis testing remains the default method for statistical inference in medical research. Therefore, we propose an interim solution to the “significance problem” based on simplified Bayesian metrics (e.g., Bayes factor, false positive risk) that can be reported along with traditional P values and confidence intervals. We calculate these metrics for four well-known multicentre trials. We provide links to online calculators so readers can easily estimate these metrics for published trials. In this way, we hope decisions on incorporating the results of randomized trials into clinical practice can be enhanced, minimizing the chance that useful treatments are discarded or that ineffective treatments are adopted.

  • Research Article
  • 10.1186/s12911-025-03109-1
Enhancing privacy protection of physical examination data through synthetic algorithms based on differential privacy
  • Sep 1, 2025
  • BMC Medical Informatics and Decision Making
  • Weili Zhang + 4 more

BackgroundHealth physical examinations play a crucial role in early detection of cancer and chronic disease. However, privacy concerns limit the utilization of this kind of data for health interventions and research. Synthetic data methods based on differential privacy are increasingly used to create complete datasets that protect privacy while enabling data analysis and result interpretation. Hence, the use of synthetic algorithms based on differential privacy for privacy protection of physical examination data is a promising research direction.MethodsThree synthetic algorithms, PrivBayes, PeGS, and DP-Gibbs were used to generate complete synthetic datasets that adhere to differential privacy standards using physical examination data composed of categorical data, which compared with the existing algorithm Private-PGM.ResultsCompared with the existing algorithm, DP-Gibbs can provide privacy preserving capacity of 4.686 (ε = 0.5), while the existing algorithm only with 2.012. In addition, DP-Gibbs provides 0.620 of precision, 0.539 of F1-score, 0.342 of Kappa Coefficient, and 0.765 of AUC-score. The corresponding statistical results of existing algorithm are 0.520, 0.321, 0.188 and 0.695.ConclusionsThe main contributions of this study are the exploration of combination models incorporating different noise forms and Bayesian synthetic algorithms, alongside a comparative analysis against existing algorithms. This study explored the balance between privacy protection and data utility under different levels of privacy protection, and DP-Gibbs offers more stable technical support for de-identifying physical examination data prior to sharing and analysis, which realized the mining and application of a wider range of medical data under the requirements of privacy protection. By leveraging this effective privacy protection technique, clinical researchers can extract valuable insights on diseases and population health from the physical examination data without the risk of leaking private information.Supplementary InformationThe online version contains supplementary material available at 10.1186/s12911-025-03109-1.

  • Conference Article
  • Cite Count Icon 1
  • 10.1109/pimrc.2017.8292647
Composition properties of Bayesian differential privacy
  • Oct 1, 2017
  • Jun Zhao

Differential privacy is a rigorous privacy standard that has been applied to a range of data analysis tasks. To broaden the application scenarios of differential privacy when data records have dependencies, the notion of Bayesian differential privacy has been recently proposed. However, it is unknown whether Bayesian differential privacy preserves three nice properties of differential privacy: sequential composability, parallel composability, and post-processing. In this paper, we provide an affirmative answer to this question; i.e., Bayesian differential privacy still have these properties. The idea behind sequential composability is that if we have $m$ algorithms $Y_1, Y_2, \ldots, Y_m$, where $Y_{\ell}$ is independently $\epsilon_{\ell}$-Bayesian differential private for ${\ell}=1,2,\ldots,m$, then by feeding the result of $Y_1$ into $Y_2$, the result of $Y_2$ into $Y_3$, and so on, we will finally have an $\sum_{\ell=1}^m \epsilon_{\ell}$-Bayesian differential private algorithm. For parallel composability, we consider the situation where a database is partitioned into $m$ disjoint subsets. The $\ell$-th subset is input to a Bayesian differential private algorithm $Y_{\ell}$, for ${\ell}=1,2,\ldots,m$. Then the parallel composition of $Y_1$, $Y_2$, $\ldots$, $Y_m$ will be $\max_{\ell=1}^m \epsilon_{\ell}$-Bayesian differential private. The post-processing property means that a data analyst, without additional knowledge about the private database, cannot compute a function of the output of a Bayesian differential private algorithm and reduce its privacy guarantee.

  • Conference Article
  • Cite Count Icon 58
  • 10.1109/asonam.2012.73
A Guide to Differential Privacy Theory in Social Network Analysis
  • Aug 1, 2012
  • C Task + 1 more

Privacy of social network data is a growing concern which threatens to limit access to this valuable data source. Analysis of the graph structure of social networks can provide valuable information for revenue generation and social science research, but unfortunately, ensuring this analysis does not violate individual privacy is difficult. Simply anonymizing graphs or even releasing only aggregate results of analysis may not provide sufficient protection. Differential privacy is an alternative privacy model, popular in data-mining over tabular data, which uses noise to obscure individuals' contributions to aggregate results and offers a very strong mathematical guarantee that individuals' presence in the data-set is hidden. Analyses that were previously vulnerable to identification of individuals and extraction of private data may be safely released under differential-privacy guarantees. We review two existing standards for adapting differential privacy to network data and analyse the feasibility of several common social-network analysis techniques under these standards. Additionally, we propose out-link privacy, a novel standard for differential privacy over network data, and introduce two powerful out-link private algorithms for common network analysis techniques that were infeasible to privatize under previous differential privacy standards.

  • Conference Article
  • Cite Count Icon 29
  • 10.1145/2247596.2247604
Differentially private search log sanitization with optimal output utility
  • Mar 27, 2012
  • Yuan Hong + 3 more

Web search logs contain extremely sensitive data, as evidenced by the recent AOL incident. However, storing and analyzing search logs can be very useful for many purposes (i.e. investigating human behavior). Thus, an important research question is how to privately sanitize search logs. Several search log anonymization techniques have been proposed with concrete privacy models. However, in all of these solutions, the output utility of the techniques is only evaluated rather than being maximized in any fashion. Indeed, for effective search log anonymization, it is desirable to derive the outputs with optimal utility while meeting the privacy standard. In this paper, we propose utility-maximizing sanitization based on the rigorous privacy standard of differential privacy, in the context of search logs. Specifically, we utilize optimization models to maximize the output utility of the sanitization for different applications, while ensuring that the production process satisfies differential privacy. An added benefit is that our novel randomization strategy maintains the schema integrity in the output search logs. A comprehensive evaluation on real search logs validates the approach and demonstrates its robustness and scalability.

  • Research Article
  • 10.3991/ijim.v19i19.57619
Advanced Privacy-Utility Optimization Techniques in Federated Learning with Differential Privacy for IoMT – A Review
  • Oct 7, 2025
  • International Journal of Interactive Mobile Technologies (iJIM)
  • Shaista Ashraf Farooqi + 2 more

This paper reviews advanced optimization techniques to address the privacy-utility tradeoff in federated learning with differential privacy (FL-DP), focusing on applications in the Internet of Medical Things (IoMT). IoMT systems face significant challenges, including heterogeneous, non-IID data distributions, resource-constrained devices, and stringent privacy regulations such as HIPAA and GDPR, making it complex to ensure robust privacy while maintaining high model utility. The review explores methods such as adaptive privacy budgeting, which dynamically adjusts privacy parameters (∈) based on data sensitivity and device capabilities, and client selection strategies that enhance global model accuracy by prioritizing high-quality data contributions while effectively managing privacy budgets. Techniques like gradient clipping and noise scaling are examined for their ability to mitigate the negative impact of differential privacy (DP) noise, ensuring stability in real-time applications like remote patient monitoring and anomaly detection. This study analyzes existing techniques and identifies gaps in advancing scalable and efficient FL-DP frameworks in IoMT. Future directions include AI-driven adaptive privacy mechanisms and energy-efficient optimization algorithms to enhance the scalability, performance, and sustainability of FL-DP in IoMT environments. These advancements aim to develop secure, high-performance IoMT systems that comply with privacy standards while addressing real-world healthcare challenges.

  • Research Article
  • Cite Count Icon 7
  • 10.1145/3589207
Contextual Linear Types for Differential Privacy
  • May 17, 2023
  • ACM Transactions on Programming Languages and Systems
  • Matías Toro + 6 more

Language support for differentially private programming is both crucial and delicate. While elaborate program logics can be very expressive, type-system-based approaches using linear types tend to be more lightweight and amenable to automatic checking and inference, and in particular in the presence of higher-order programming. Since the seminal design of Fuzz , which is restricted to ϵ-differential privacy in its original design, significant progress has been made to support more advanced variants of differential privacy, like (ϵ, δ )-differential privacy. However, supporting these advanced privacy variants while also supporting higher-order programming in full has proven to be challenging. We present Jazz , a language and type system that uses linear types and latent contextual effects to support both advanced variants of differential privacy and higher-order programming. Latent contextual effects allow delaying the payment of effects for connectives such as products, sums, and functions, yielding advantages in terms of precision of the analysis and annotation burden upon elimination, as well as modularity. We formalize the core of Jazz , prove it sound for privacy via a logical relation for metric preservation, and illustrate its expressive power through a number of case studies drawn from the recent differential privacy literature.

  • Research Article
  • Cite Count Icon 2
  • 10.1093/jssam/smac012
Accuracy Gains from Privacy Amplification Through Sampling for Differential Privacy
  • Jun 3, 2022
  • Journal of Survey Statistics and Methodology
  • Jingchen Hu + 2 more

Recent research in differential privacy demonstrated that (sub)sampling can amplify the level of protection. For example, for ϵ-differential privacy and simple random sampling with sampling rate r, the actual privacy guarantee is approximately rϵ, if a value of ϵ is used to protect the output from the sample. In this paper, we study whether these amplification effects can be exploited systematically to improve the accuracy of the privatized estimate. Specifically, assuming the agency has information for the full population, we ask under which circumstances accuracy gains could be expected, if the privatized estimate would be computed on a random sample instead of the full population. We find that accuracy gains can be achieved for certain regimes. However, gains can typically only be expected, if the sensitivity of the output with respect to small changes in the database does not depend too strongly on the size of the database. We only focus on algorithms that achieve differential privacy by adding noise to the final output and illustrate the accuracy implications for two commonly used statistics: the mean and the median. We see our research as a first step toward understanding the conditions required for accuracy gains in practice and we hope that these findings will stimulate further research broadening the scope of differential privacy algorithms and outputs considered.

  • Conference Article
  • 10.1109/nana.2019.00053
Edge Weight Differential Privacy Based Spectral Query Algorithm
  • Oct 1, 2019
  • Lingjie Zhou + 5 more

The macro nature of social networks has always been a hot topic of scientific research, which is of great help in studying the characteristics of human social behavior. The spectrum is closely related to the nature of the network and determines certain information, including community division, loop, diameter, degree and so on. To ensure the security of data, the differential privacy security framework is used to propose three algorithms for privacy protection of edge weights in spectrum query. Firstly, the basic concepts of edge weight neighbor graph and edge weight difference privacy are given. For the single singular value query, the L1 global sensitivity of the function is proved, and an algorithm satisfying ε-differential privacy is designed by using Laplace mechanism. It can guarantee the privacy of edge weights when we query someone singular value. However, it cannot provide reasonable privacy protection for multiple singular value queries. So the L2 global sensitivity of the query for multiple singular values is further proved, and the Gaussian mechanism is used to design an algorithm that satisfies (ε, δ)-differential privacy. Finally, to combine with multi-singular values query strategy and spectral decomposition, a novel graph data publishing algorithm which can guarantee edge weight (ε, δ)-differential privacy is proposed. To verify the availability of these methods, experimental tests have been carried out in both model networks and actual networks, which shows that the algorithm can better guarantee the availability of data.

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.