Abstract

Mutual information of two random variables can be easily obtained from their Shannon entropies. However, when nonadditive entropies are involved, the calculus of the mutual information is more complex. Here we discuss the basic matter about information from Shannon entropy. Then we analyse the case of the generalized nonadditive Tsallis entropy.

Highlights

  • In many applications of engineering and telecommunication, it is often desired to increase or decrease the dependency of two random variables

  • The mutual information can be decomposed into a sum of entropies [1], when the Shannon entropy is used

  • Due to its entropic index, which can be used as tuning parameter, this entropy is involved in several applications, in particular for image processing and image registration [6]

Read more

Summary

Introduction

In many applications of engineering and telecommunication, it is often desired to increase or decrease the dependency of two random variables. This dependency is linked to the mutual information, which is its measure. When nonadditive entropies are involved, the approach to find the mutual information is not so simple. We discuss the basic matter concerning the mutual information when Tsallis entropy is involved.

Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call