TY - GEN
T1 - Data processing inequalities based on a certain structured class of information measures with application to estimation theory
AU - Merhav, Neri
PY - 2012
Y1 - 2012
N2 - We study data processing inequalities (DPI's) that are derived from a certain class of generalized information measures, where a series of convex functions and multiplicative likelihood ratios are nested alternately. A certain choice of the convex functions leads to an information measure that extends the notion of the Bhattacharyya distance: While the ordinary Bhattacharyya distance is based on the geometric mean of two replicas of the channel's conditional distribution, the more general one allows an arbitrary number of replicas. We apply the DPI induced by this information measure to a detailed study of lower bounds of parameter estimation under additive white Gaussian noise (AWGN) and show that in certain cases, tighter bounds can be obtained by using more than two replicas. While the resulting bound may not compete favorably with the best bounds available for the ordinary AWGN channel, the advantage of the new lower bound, becomes significant in the presence of channel uncertainty, like unknown fading. This is explained by the convexity property of the information measure.
AB - We study data processing inequalities (DPI's) that are derived from a certain class of generalized information measures, where a series of convex functions and multiplicative likelihood ratios are nested alternately. A certain choice of the convex functions leads to an information measure that extends the notion of the Bhattacharyya distance: While the ordinary Bhattacharyya distance is based on the geometric mean of two replicas of the channel's conditional distribution, the more general one allows an arbitrary number of replicas. We apply the DPI induced by this information measure to a detailed study of lower bounds of parameter estimation under additive white Gaussian noise (AWGN) and show that in certain cases, tighter bounds can be obtained by using more than two replicas. While the resulting bound may not compete favorably with the best bounds available for the ordinary AWGN channel, the advantage of the new lower bound, becomes significant in the presence of channel uncertainty, like unknown fading. This is explained by the convexity property of the information measure.
UR - http://www.scopus.com/inward/record.url?scp=84867520709&partnerID=8YFLogxK
U2 - https://doi.org/10.1109/ISIT.2012.6283061
DO - https://doi.org/10.1109/ISIT.2012.6283061
M3 - منشور من مؤتمر
SN - 9781467325790
T3 - IEEE International Symposium on Information Theory - Proceedings
SP - 1271
EP - 1275
BT - 2012 IEEE International Symposium on Information Theory Proceedings, ISIT 2012
T2 - 2012 IEEE International Symposium on Information Theory, ISIT 2012
Y2 - 1 July 2012 through 6 July 2012
ER -