Abstract

Information bottleneck (IB) and privacy funnel (PF) are two closely related optimization problems which have found applications in machine learning, design of privacy algorithms, capacity problems (e.g., Mrs. Gerber’s Lemma), and strong data processing inequalities, among others. In this work, we first investigate the functional properties of IB and PF through a unified theoretical framework. We then connect them to three information-theoretic coding problems, namely hypothesis testing against independence, noisy source coding, and dependence dilution. Leveraging these connections, we prove a new cardinality bound on the auxiliary variable in IB, making its computation more tractable for discrete random variables. In the second part, we introduce a general family of optimization problems, termed “bottleneck problems”, by replacing mutual information in IB and PF with other notions of mutual information, namely f-information and Arimoto’s mutual information. We then argue that, unlike IB and PF, these problems lead to easily interpretable guarantees in a variety of inference tasks with statistical constraints on accuracy and privacy. While the underlying optimization problems are non-convex, we develop a technique to evaluate bottleneck problems in closed form by equivalently expressing them in terms of lower convex or upper concave envelope of certain functions. By applying this technique to a binary case, we derive closed form expressions for several bottleneck problems.

Highlights

  • Optimization formulations that involve information-theoretic quantities have been instrumental in a variety of learning problems found in machine learning

  • We provide a comprehensive analysis of Information bottleneck (IB) and privacy funnel (PF) from an information-theoretic perspective, as well as a survey of several formulations connected to the IB and PF that have been introduced in the information theory and machine learning literature

  • It is worth noting that LIB ( β) and LPF ( β) correspond to lines of slope β supporting

Read more

Summary

Introduction

Optimization formulations that involve information-theoretic quantities (e.g., mutual information) have been instrumental in a variety of learning problems found in machine learning. For which the upper bound becomes tight This provides a significantly simpler proof for the fact that in this special case the optimal bottleneck variable T is Gaussian than the original proof given in [26]. We extend the Witsenhausen and Wyner’s approach [3] for analytically computing IB and PF This technique converts solving the optimization problems in IB and PF to determining the convex and concave envelopes of a certain function, respectively. We apply this technique to binary X and Y and derive a closed form expression for PF(r )– we call this result Mr Gerber’s Lemma. Focusing on binary X and Y, we derive closed form expressions for some of the bottleneck problems

Related Work
Notation
Information Bottleneck and Privacy Funnel
Gaussian IB and PF
Evaluation of IB and PF
Operational Meaning of IB and PF
Noisy Source Coding
Test against Independence with Communication Constraint
Dependence Dilution
Cardinality Bound
Deterministic Information Bottleneck
Family of Bottleneck Problems
Guessing Bottleneck Problems
Arimoto Bottleneck Problems
Summary and Concluding Remarks
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call