Abstract

In the Minimum Description Length (MDL) principle, learning from the data is equivalent to an optimal coding problem. We show that the codes that achieve optimal compression in MDL are critical in a very precise sense. First, when they are taken as generative models of samples, they generate samples with broad empirical distributions and with a high value of the relevance, defined as the entropy of the empirical frequencies. These results are derived for different statistical models (Dirichlet model, independent and pairwise dependent spin models, and restricted Boltzmann machines). Second, MDL codes sit precisely at a second order phase transition point where the symmetry between the sampled outcomes is spontaneously broken. The order parameter controlling the phase transition is the coding cost of the samples. The phase transition is a manifestation of the optimality of MDL codes, and it arises because codes that achieve a higher compression do not exist. These results suggest a clear interpretation of the widespread occurrence of statistical criticality as a characterization of samples which are maximally informative on the underlying generative process.

Highlights

  • It is not infrequent to find empirical data which exhibits broad frequency distributions in the most disparate domains

  • We find that Normalized Maximum Likelihood (NML) are critical in a very precise sense, because they sit at a second order phase transition that separates typical from atypical behavior

  • The aim of this paper is to elucidate the properties of efficient representations of data corresponding to universal codes that arise in Minimum Description Length (MDL)

Read more

Summary

Introduction

It is not infrequent to find empirical data which exhibits broad frequency distributions in the most disparate domains. A straight line in the rank plot corresponds to a power law frequency distribution, where the number of outcomes that are observed k times behave as mk ∼ k−μ−1. It has recently been suggested that broad distributions arise from efficient representations, i.e., when the data samples relevant variables, which are those carrying the maximal amount of information on the generative process [7,8,9].

Objectives
Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call