Rescaling Entropy and Divergence Rates - Archive ouverte HAL Accéder directement au contenu
Article Dans Une Revue IEEE Transactions on Information Theory Année : 2015

Rescaling Entropy and Divergence Rates

Loïck Lhote

Résumé

Based on rescaling by some suitable sequence instead of the number of time units, the usual notion of divergence rate is here extended to define and determine meaningful generalized divergence rates. Rescaling entropy rates appears as a special case. Suitable rescaling is naturally induced by the asymptotic behavior of the marginal divergences. Closed form formulas are obtained as soon as the marginal divergences behave like powers of some analytical functions. A wide class of countable Markov chains is proven to satisfy this property. Most divergence and entropy functionals defined in the literature are concerned, e.g., the classical Shannon, Kullback-Leibler, R´enyi, Tsallis. For illustration purposes, Ferreri or Basu-Harris-Hjort-Jones – among others – are also considered.
Fichier non déposé

Dates et versions

hal-01196817 , version 1 (10-09-2015)

Identifiants

Citer

Loïck Lhote, Valérie Girardin. Rescaling Entropy and Divergence Rates. IEEE Transactions on Information Theory, 2015, 61 (11), pp.5868-5882. ⟨10.1109/TIT.2015.2476486⟩. ⟨hal-01196817⟩
122 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More