Further results on generalized conditional entropies
RAIRO - Theoretical Informatics and Applications - Informatique Théorique et Applications, Tome 49 (2015) no. 1, pp. 67-92.

We further examine some properties of the conditional Rényi and Tsallis–Havrda–Charvát (THC) entropies. Such properties are interesting from the viewpoint of applications in studying protocols of quantum information science and foundations of quantum mechanics. In particular, we consider properties of the conditional Rényi and THC entropies with respect to conditioning on more. We also exemplify that the desired property can be violated with the conditional min-entropy. Applications of such results to the TCH entropy rate are considered. Connections between generalized conditional entropies and error probability are examined. Several relations between various conditional entropies are obtained. It is shown that such relations can be used for bounding the conditional Rényi and TCH entropies.

Reçu le :
Accepté le :
DOI : 10.1051/ita/2014029
Classification : 94A17, 62B10, 39B62
Mots-clés : Rényi entropy, Tsallis–Havrda–Charvát entropy, entropy rate, index of coincidence, error probability, Fano inequality
Rastegin, Alexey E. 1

1 Department of Theoretical Physics, Irkutsk State University, Gagarin Bv. 20, 664003 Irkutsk, Russia
@article{ITA_2015__49_1_67_0,
     author = {Rastegin, Alexey E.},
     title = {Further results on generalized conditional entropies},
     journal = {RAIRO - Theoretical Informatics and Applications - Informatique Th\'eorique et Applications},
     pages = {67--92},
     publisher = {EDP-Sciences},
     volume = {49},
     number = {1},
     year = {2015},
     doi = {10.1051/ita/2014029},
     mrnumber = {3342174},
     zbl = {1395.94219},
     language = {en},
     url = {http://archive.numdam.org/articles/10.1051/ita/2014029/}
}
TY  - JOUR
AU  - Rastegin, Alexey E.
TI  - Further results on generalized conditional entropies
JO  - RAIRO - Theoretical Informatics and Applications - Informatique Théorique et Applications
PY  - 2015
SP  - 67
EP  - 92
VL  - 49
IS  - 1
PB  - EDP-Sciences
UR  - http://archive.numdam.org/articles/10.1051/ita/2014029/
DO  - 10.1051/ita/2014029
LA  - en
ID  - ITA_2015__49_1_67_0
ER  - 
%0 Journal Article
%A Rastegin, Alexey E.
%T Further results on generalized conditional entropies
%J RAIRO - Theoretical Informatics and Applications - Informatique Théorique et Applications
%D 2015
%P 67-92
%V 49
%N 1
%I EDP-Sciences
%U http://archive.numdam.org/articles/10.1051/ita/2014029/
%R 10.1051/ita/2014029
%G en
%F ITA_2015__49_1_67_0
Rastegin, Alexey E. Further results on generalized conditional entropies. RAIRO - Theoretical Informatics and Applications - Informatique Théorique et Applications, Tome 49 (2015) no. 1, pp. 67-92. doi : 10.1051/ita/2014029. http://archive.numdam.org/articles/10.1051/ita/2014029/

N. Alon and J.H. Spencer, The Probabilistic Method, 3rd edition. John Wiley & Sons, Hoboken (2008). | MR | Zbl

D. Bacon and W. Van Dam, Recent progress in quantum algorithms. Commun. ACM 53 (2010) 84–93. | DOI

M. Ben-Bassat and J. Raviv, Rényi’s entropy and error probability. IEEE Trans. Inf. Theory 24 (1978) 324–331. | DOI | Zbl

I. Bengtsson and K. Życzkowski, Geometry of Quantum States: An Introduction to Quantum Entanglement. Cambridge University Press, Cambridge (2006). | MR | Zbl

S.L. Braunstein and C.M. Caves, Information-theoretic Bell inequalities. Phys. Rev. Lett. 61 (1988) 662–665. | DOI | MR

F. Buscemi, M.J.W. Hall, M. Ozawa and M.M. Wilde, Noise and disturbance in quantum measurements: An information-theoretic approach. Phys. Rev. Lett. 112 (2014) 050401. | DOI

C. Cachin, Entropy measures and unconditional security in cryptography. Ph.D. thesis, Swiss Federal Institute of Technology, Zürich (1997).

R. Chaves and T. Fritz, Entropic approach to local realism and noncontextuality. Phys. Rev. A 85 (2012) 032113. | DOI

R. Chaves and T. Fritz, Entropic inequalities and marginal problems. IEEE Trans. Inf. Theory 59 (2013) 803–817. | DOI | MR | Zbl

I. Csiszár, Information-type measures of difference of probability distributions and indirect observations. Studia Sci. Math. Hungar. 2 (1967) 299–318. | MR | Zbl

I. Csiszár, Axiomatic characterizations of information measures. Entropy 10 (2008) 261–273. | DOI | Zbl

T.M. Cover and J.A. Thomas, Elements of Information Theory. John Wiley & Sons, New York (1991). | MR | Zbl

Z. Daróczy, Generalized information functions. Inform. Control 16 (1970) 36–51. | DOI | MR | Zbl

Z. Daróczy, Inequalities for some infinite series. Acta Math. Hungar. 75 (1997) 5–8. | DOI | MR | Zbl

D. De Falco and D. Tamascelli, An introduction to quantum annealing. RAIRO: ITA 45 (2011) 99–116. | Numdam | MR | Zbl

D. Erdogmus and J.C. Principe, Lower and upper bounds for misclassification probability based on Rényi’s information. J. VLSI Signal Process. 37 (2004) 305–317. | DOI | Zbl

R.M. Fano, Transmission of Information: A Statistical Theory of Communications. MIT Press and John Wiley & Sons, New York (1961). | MR

M. Feder and N. Merhav, Relations between entropy and error probability. IEEE Trans. Inf. Theory 40 (1994) 259–266. | DOI | Zbl

S. Furuichi, Information-theoretical properties of Tsallis entropies. J. Math. Phys. 47 (2006) 023302. | DOI | MR | Zbl

M. Gell-Mann and C. Tsallis, Nonextensive Entropy – Interdisciplinary Applications. Oxford University Press, Oxford (2004). | Zbl

I.M. Georgescu, S. Ashhab and F. Nori, Quantum simulation. Rev. Mod. Phys. 86 (2014) 153–185. | DOI

L. Golshani, E. Pasha and G. Yari, Some properties of Rényi entropy and Rényi entropy rate. Inf. Sci. 179 (2009) 2426–2433. | DOI | MR | Zbl

P. Harremoës and F. Topsøe, Inequalities between entropy and index of coincidence derived from information diagrams. IEEE Trans. Inf. Theory 47 (2001) 2944–2960. | DOI | MR | Zbl

J. Havrda and F. Charvát, Quantification methods of classification processes: concept of structural α-entropy. Kybernetika 3 (1967) 30–35. | MR | Zbl

F. Hiai, M. Mosonyi, D. Petz and C. Bény, Quantum f-divergences and error correction. Rev. Math. Phys. 23 (2011) 691–747. | DOI | MR | Zbl

S.-W. Ho and S. Verdú, On the interplay between conditional entropy and error probability. IEEE Trans. Inf. Theory 56 (2011) 5930–5942. | DOI | MR | Zbl

S.-W. Ho and R.W. Yeung, On the discontinuity of the Shannon information measures. IEEE Trans. Inf. Theory 55 (2009) 5362–5374. | DOI | MR | Zbl

P. Jizba and T. Arimitsu, The world according to Rényi: thermodynamics of multifractal systems. Ann. Phys. 312 (2004) 17–59 | DOI | MR | Zbl

R. Kamimura, Minimizing α-information for generalization and interpretation. Algorithmica 22 (1998) 173–197. | DOI | MR | Zbl

S. Kullback and R. A. Leibler, On information and sufficiency. Ann. Math. Stat. 22 (1951) 79–86. | DOI | MR | Zbl

H. Maassen and J.B. M. Uffink, Generalized entropic uncertainty relations. Phys. Rev. Lett. 60 (1988) 1103–1106. | DOI | MR

A.W. Marshall, I. Olkin and B.C. Arnold, Inequalities: Theory of Majorization and Its Applications, 2nd edition. Springer-Verlag, New York (2011). | MR | Zbl

M. Matić, C.E.M. Pearce and J. Pečarić, On an inequality for the entropy of a probability distribution. Acta Math. Hungar. 85 (1999) 345–349. | DOI | MR | Zbl

J. Radhakrishnan, An entropy proof of Brégman’s theorem. J. Comb. Theory Ser. A 77 (1997) 161–164. | DOI | MR | Zbl

A.E. Rastegin, Convexity inequalities for estimating generalized conditional entropies from below. Kybernetika 48 (2012) 242–253 | MR | Zbl

A.E. Rastegin, Bounds of the Pinsker and Fannes types on the Tsallis relative entropy. Math. Phys. Anal. Geom. 16 (2013) 213–228. | DOI | MR | Zbl

A.E. Rastegin, Uncertainty relations for MUBs and SIC-POVMs in terms of generalized entropies. Eur. Phys. J. D 67 (2013) 269. | DOI

A.E. Rastegin, Tests for quantum contextuality in terms of q-entropies. Quantum Inf. Comput. 14 (2014) 0996–1013. | MR

A. Rényi, On measures of entropy and information, in Proc. of the 4th Berkeley Symposium on Mathematical Statistics and Probability, edited by J. Neyman. University of California Press, Berkeley CA (1961) 547–561. | MR | Zbl

A. Rényi, Statistics and information theory. Studia Sci. Math. Hungar. 2 (1967) 249–256. | MR | Zbl

W. Shitong and F.L. Chung, Note on the equivalence relationship between Rényi-entropy based and Tsallis-entropy based image thresholding. Pattern Recognit. Lett. 26 (2005) 2309–2312. | DOI

A. Teixeira, A. Matos and L. Antunes, Conditional Rényi entropies. IEEE Trans. Inf. Theory 58 (2012) 4273–4277. | DOI | MR | Zbl

C. Tsallis, Possible generalization of Boltzmann–Gibbs statistics. J. Stat. Phys. 52 (1988) 479–487. | DOI | MR | Zbl

I. Vajda, Bounds of the minimal error probability on checking a finite or countable number of hypotheses. Problemy Peredači Informacii 4 (1968) 9–19, (in Russian); translated as Probl. Inf. Transm. 4 (1968) 6–14. | MR

K. Życzkowski, Rényi extrapolation of Shannon entropy. Open Sys. Inf. Dyn. 10 (2003) 297–310, corrigendum in arXiv:quant-ph/0305062v2. | DOI | MR | Zbl

Cité par Sources :