Abstract

Divergence functions are the non-symmetric “distance” on the manifold, Μθ, of parametric probability density functions over a measure space, (Χ,μ). Classical information geometry prescribes, on Μθ: (i) a Riemannian metric given by the Fisher information; (ii) a pair of dual connections (giving rise to the family of α-connections) that preserve the metric under parallel transport by their joint actions; and (iii) a family of divergence functions ( α-divergence) defined on Μθ x Μθ, which induce the metric and the dual connections. Here, we construct an extension of this differential geometric structure from Μθ (that of parametric probability density functions) to the manifold, Μ, of non-parametric functions on X, removing the positivity and normalization constraints. The generalized Fisher information and α-connections on M are induced by an α-parameterized family of divergence functions, reflecting the fundamental convex inequality associated with any smooth and strictly convex function. The infinite-dimensional manifold, M, has zero curvature for all these α-connections; hence, the generally non-zero curvature of M can be interpreted as arising from an embedding of Μθ into Μ. Furthermore, when a parametric model (after a monotonic scaling) forms an affine submanifold, its natural and expectation parameters form biorthogonal coordinates, and such a submanifold is dually flat for α = ± 1, generalizing the results of Amari’s α-embedding. The present analysis illuminates two different types of duality in information geometry, one concerning the referential status of a point (measurable function) expressed in the divergence function (“referential duality”) and the other concerning its representation under an arbitrary monotone scaling (“representational duality”).

Highlights

  • Information geometry is a differential geometric study of the manifold of probability measures or probability density functions [1]

  • The present analysis illuminates two different types of duality in information geometry, one concerning the referential status of a point expressed in the divergence function (“referential duality”) and the other concerning its representation under an arbitrary monotone scaling

  • The goal of the present paper is to extend these non-parametric results by showing links among three inter-connected mathematical topics that underlie information geometry, namely: (i) divergence functions measuring the non-symmetric distance of any two points on the manifold; (ii) convex analysis and the associated Legendre–Fenchel transformation linking the natural and expectation parameters of parametric models; and (iii) the resulting dual Riemannian structure involving the Fisher metric and the family of α-connections

Read more

Summary

Introduction

Information geometry is a differential geometric study of the manifold of probability measures or probability density functions [1]. Efron [18], through investigating a one-parameter family of statistical models, elucidated the meaning of curvature for asymptotic statistical inference and pointed out its flatness for the exponential model In his reaction to Efron’s work, Dawid [19] invoked the differential geometric notion of linear connections on a manifold as preserving parallelism during vector transportation and pointed out other possible constructions of linear connections on Mθ , in addition to the non-flat Levi-Civita connection associated with the Fisher metric. Have Γ(±1) zero curvatures for both exponential and mixture families, but affine coordinates were found to yield Γ(1) and Γ(−1) , themselves zero for the exponential and mixture families, respectively This classic information geometry dealing with parametric statistical models has been investigated in the non-parametric setting using the tools of infinite-dimensional analysis [21,22,23], with non-parametric. We carefully delineate two senses of duality associated with such manifolds, one related to the reference/comparison status of any pair of points (functions) and the other related to properly scaled representations of them

Parametric Information Geometry Revisited
Divergence Function and Induced Statistical Manifold
Induced Dual Riemannian Geometry
Goals and Approach
Information Geometry on Infinite-Dimensional Function Space
Differentiable Manifold in the Infinite-Dimensional Setting
Fundamental Convex Inequality and Divergence
Conjugate-Scaled Representations of Measurable Functions
Canonical Divergence
Finite-Dimensional Parametric Models
Riemannian Geometry of Parametric Models
Example
Affine Embedded Submanifold
Biorthogonality of Natural and Expectation Parameters
Dually Flat Affine Manifolds
Proofs
Discussions
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call