Yuri M. Suhov

Learn More
{ We discuss a family of estimators for the entropy rate of a stationary ergodic process and prove their pointwise and mean consistency under a Doeblin-type mixing condition. The estimators are Cess aro averages of longest match-lengths, and their consistency follows from a generalized ergodic theorem due to Maker. We provide examples of their performance(More)
This article addresses the issue of the proof of the entropy power inequality (EPI), an important tool in the analysis of Gaussian channels of information transmission, proposed by Shannon. We analyse continuity properties of the mutual entropy of the input and output signals in an additive memoryless channel and discuss assumptions under which the(More)
In this paper, we review Fisher information matrices properties in weighted version and discuss inequalities/bounds on it by using reduced weight functions. In particular, an extended form of the Fisher information inequality previously established in [6] is given. Further, along with generalized De-Bruijn's identity, we provide new interpretation of the(More)
The Shannon Noiseless coding theorem (the data-compression principle) asserts that for an information source with an alphabet X = {0,. .. , − 1} and an asymptotic equipartition property, one can reduce the number of stored strings (x 0 ,. .. , x n−1) ∈ X n to nh with an arbitrary small error-probability. Here h is the entropy rate of the source (calculated(More)