Abstract

Divergence is a discrepancy measure between two objects, such as functions, vectors, matrices, and so forth. In particular, divergences defined on probability distributions are widely employed in probabilistic forecasting. As the dissimilarity measure, the divergence should satisfy some conditions. In this paper, we consider two conditions: The first one is the scale-invariance property and the second is that the divergence is approximated by the sample mean of a loss function. The first requirement is an important feature for dissimilarity measures. The divergence will depend on which system of measurements we used to measure the objects. Scale-invariant divergence is transformed in a consistent way when the system of measurements is changed to the other one. The second requirement is formalized such that the divergence is expressed by using the so-called composite score. We study the relation between composite scores and scale-invariant divergences, and we propose a new class of divergences called H¨older divergence that satisfies two conditions above. We present some theoretical properties of H¨older divergence. We show that H¨older divergence unifies existing divergences from the viewpoint of scale-invariance.

Highlights

  • Nowadays, divergence measures are ubiquitous in the field of information sciences

  • We proposed Hölder divergence as defined from the composite score, and showed that the Hölder divergence has the scale-invariance property

  • We proved that the composite score satisfying the scale-invariance property leads to the Hölder divergence

Read more

Summary

Introduction

Divergence measures are ubiquitous in the field of information sciences. The divergence is a discrepancy measure between two objects, such as functions, vectors, matrices, and so forth. The first one is the scale-invariance property, and the second one is that the divergence should be represented by using the so-called composite score [5], that is an extension of scores [6]. The Kullback-Leibler divergence that is one of the most popular divergences has the scale-invariance property for the measurement of training samples [7]. Dissimilarity measures should be expressed as the form of composite scores This is a useful property, when the divergence is employed for the statistical inference of the probability densities. The score [2,5,6,8,9,10] is the class of dissimilarity measures that are calculated through the sample mean of the observed data. Ω theoretical results to any compact set in the multi-dimensional Euclidean space is straightforward

Divergences and Scores
Scale-Invariance of Divergences
Divergence for Statistical Inference
Hölder Divergences
Conjugate Relation
Bregman Divergence and Hölder Divergence
Characterization of Hölder Scores
Conclusions
Proof of Theorem 3
Proof of Lemma 2
Proofs of Theorem 4

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.