By Topic

Csisza´r's cutoff rates for the general hypothesis testing problem

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
F. Alajaji ; Dept. of Math. & Stat., Queens Univ., Kingston, Ont., Canada ; Po-Ning Chen ; Z. Rached

In , Csisza´r established the concept of forward β-cutoff rate for the error exponent hypothesis testing problem based on independent and identically distributed (i.i.d.) observations. Given β<0, he defined the forward β-cutoff rate as the number R0≥0 that provides the best possible lower bound in the form β(E-R0) to the type 1 error exponent function for hypothesis testing where 00 is the rate of exponential convergence to 0 of the type 2 error probability. He then demonstrated that the forward β-cutoff rate is given by D1(1-β)/(X||ˆX), where Dα(X||ˆX) denotes the Re´nyi α-divergence [19], α>0, α≠1. Similarly, for 0<β<1, Csisza´r also established the concept of reverse β-cutoff rate for the correct exponent hypothesis testing problem. In this work, we extend Csisza´r's results by investigating the forward and reverse β-cutoff rates for the hypothesis testing between two arbitrary sources with memory. We demonstrate that the lim inf Re´nyi α-divergence rate provides the expression for the forward β-cutoff rate. We also show that if the log-likelihood large deviation spectrum admits a limit, then the reverse β-cutoff rate equals the liminf α-divergence rate, where α=(1/1-β) and 0<β<βmax, where βmax is the largest β<1 for which the lim inf (1/1-β)-divergence rate is finite. For βmax≤β<1, we show that the reverse cutoff rate is in general only upper-bounded by the lim inf Re´nyi divergence rate. Unlike in , where the alphabet for the source coding cutoff rate problem was assumed to be finite, we assume arbitrary (countable or continuous) source alphabet. We also provide several examples to illustrate our forward and reverse β-cutoff rates results and the techniques employed to establish them.

Published in:

IEEE Transactions on Information Theory  (Volume:50 ,  Issue: 4 )