TY - JOUR
T1 - Arbitrary source models and Bayesian codebooks in rate-distortion theory
AU - Kontoyiannis, Ioannis
AU - Zhang, Junshan
N1 - Funding Information:
Manuscript received February 28, 2001; revised March 14, 2002. The work of I. Kontoyiannis was supported in part by the National Science Foundation under Grants 0073378-CCR, DMS-9615444, and by USDA-IFAFS under Grant 00-52100-9615. The work of J. Zhang was supported in part by Arizona State University (ASU) Faculty Initiation Grant DM11001. I. Kontoyiannis is with the Division of Applied Mathematics and the Department of Computer Science, Brown University, Providence, RI 02912 USA (e-mail: yiannis@dam.brown.edu). J. Zhang is with the Department of Electrical Engineering, Arizona State University, Tempe, AZ 85287-7206 USA (e-mail: junshan.zhang@asu.edu). Communicated by P. A. Chou, Associate Editor for Source Coding. Publisher Item Identifier 10.1109/TIT.2002.800493.
PY - 2002/8
Y1 - 2002/8
N2 - We characterize the best achievable performance of lossy compression algorithms operating on arbitrary random sources, and with respect to general distortion measures. Direct and converse coding theorems are given for variable-rate codes operating at a fixed distortion level, emphasizing: a) nonasymptotic results, b) optimal or near-optimal redundancy bounds, and c) results with probability one. This development is based in part on the observation that there is a precise correspondence between compression algorithms and probability measures on the reproduction alphabet. This is analogous to the Kraft inequality in lossless data compression. In the case of stationary ergodic sources our results reduce to the classical coding theorems. As an application of these general results, we examine the performance of codes based on mixture codebooks for discrete memoryless sources. A mixture codebook (or Bayesian codebook) is a random codebook generated from a mixture over some class of reproduction distributions. We demonstrate the existence of universal mixture codebooks, and show that it is possible to universally encode memoryless sources with redundancy of approximately (d/2) log n bits, where d is the dimension of the simplex of probability distributions on the reproduction alphabet.
AB - We characterize the best achievable performance of lossy compression algorithms operating on arbitrary random sources, and with respect to general distortion measures. Direct and converse coding theorems are given for variable-rate codes operating at a fixed distortion level, emphasizing: a) nonasymptotic results, b) optimal or near-optimal redundancy bounds, and c) results with probability one. This development is based in part on the observation that there is a precise correspondence between compression algorithms and probability measures on the reproduction alphabet. This is analogous to the Kraft inequality in lossless data compression. In the case of stationary ergodic sources our results reduce to the classical coding theorems. As an application of these general results, we examine the performance of codes based on mixture codebooks for discrete memoryless sources. A mixture codebook (or Bayesian codebook) is a random codebook generated from a mixture over some class of reproduction distributions. We demonstrate the existence of universal mixture codebooks, and show that it is possible to universally encode memoryless sources with redundancy of approximately (d/2) log n bits, where d is the dimension of the simplex of probability distributions on the reproduction alphabet.
KW - Data compression
KW - Mixture codebooks
KW - Rate-distortion theory
KW - Redundancy rate
UR - http://www.scopus.com/inward/record.url?scp=0036671690&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0036671690&partnerID=8YFLogxK
U2 - 10.1109/TIT.2002.800493
DO - 10.1109/TIT.2002.800493
M3 - Article
AN - SCOPUS:0036671690
SN - 0018-9448
VL - 48
SP - 2276
EP - 2290
JO - IRE Professional Group on Information Theory
JF - IRE Professional Group on Information Theory
IS - 8
ER -