source-coding references

[Rissanen1981Universal] J. Rissanen and G. Jr. Langdon. Universal modeling and coding. IEEE Trans. Inform. Theory, 27(1):12-23, Jan 1981. [ bib | .pdf ]
Keywords: information-theory source-coding
[Krichevsky1981performance] R. Krichevsky and V. Trofimov. The performance of universal encoding. IEEE Trans. Inform. Theory, 27(2):199-207, Mar 1981. [ bib | .pdf ]
Universal coding theory is surveyed from the viewpoint of the interplay between delay and redundancy. The price for universality turns out to be acceptably small.

Keywords: information-theory source-coding
[Tjalkens1992universal] T.J. Tjalkens and F.M.J. Willems. A universal variable-to-fixed length source code based on Lawrence's algorithm. IEEE Trans. Inform. Theory, 38(2):247-253, Mar 1992. [ bib | .pdf ]
It is shown that the modified Lawrence algorithm is universal over the class of binary memoryless sources and that the rate converges asymptotically optimally fast to the source entropy. It is proven that no codes exist that have a better asymptotic performance. The asymptotic bounds show that universal variable-to-fixed-length codes can have a significantly lower redundancy than universal fixed-to-variable-length codes with the same number of codewords

Keywords: information-theory source-coding
[Merhav1993Universal] N. Merhav and M. Feder. Universal schemes for sequential decision from individual data sequences. IEEE Trans. Inform. Theory, 39(4):1280-1292, Jul 1993. [ bib | .pdf ]
Sequential decision algorithms are investigated in relation to a family of additive performance criteria for individual data sequences. Simple universal sequential schemes are known, under certain conditions, to approach optimality uniformly as fast as n-1 log n, where n is the sample size. For the case of finite-alphabet observations, the class of schemes that can be implemented by finite-state machines (FSMs) is studied. It is shown that Markovian machines with sufficiently long memory exist, which are asymptotically nearly as good as any given deterministic or randomized FSM for the purpose of sequential decision. For the continuous-valued observation case, a useful class of parametric schemes is discussed with special attention to the recursive least squares algorithm

Keywords: information-theory source-coding
[Merhav1993Some] N. Merhav, M. Feder, and M. Gutman. Some properties of sequential predictors for binary Markov sources. IEEE Trans. Inform. Theory, 39(3):887-892, May 1993. [ bib | .pdf ]
Universal predictions of the next outcome of a binary sequence drawn from a Markov source with unknown parameters is considered. For a given source, the predictability is defined as the least attainable expected fraction of prediction errors. A lower bound is derived on the maximum rate at which the predictability is asymptotically approached uniformly over all sources in the Markov class. This bound is achieved by a simple majority predictor. For Bernoulli sources, bounds on the large deviations performance are investigated. A lower bound is derived for the probability that the fraction of errors will exceed the predictability by a prescribed amount ?>0. This bound is achieved by the same predictor if ? is sufficiently small

Keywords: information-theory source-coding
[Feder1996Hierarchical] M. Feder and N. Merhav. Hierarchical universal coding. IEEE Trans. Inform. Theory, 42(5):1354-1364, Sep 1996. [ bib | .pdf ]
In an earlier paper, we proved a strong version of the redundancy-capacity converse theorem of universal coding, stating that for ?most? sources in a given class, the universal coding redundancy is essentially lower-bounded by the capacity of the channel induced by this class. Since this result holds for general classes of sources, it extends Rissanen's (1986) strong converse theorem for parametric families. While our earlier result has established strong optimality only for mixture codes weighted by the capacity-achieving prior, our first result herein extends this finding to a general prior. For some cases our technique also leads to a simplified proof of the above mentioned strong converse theorem. The major interest in this paper, however, is in extending the theory of universal coding to hierarchical structures of classes, where each class may have a different capacity. In this setting, one wishes to incur redundancy essentially as small as that corresponding to the active class, and not the union of classes. Our main result is that the redundancy of a code based on a two-stage mixture (first, within each class, and then over the classes), is no worse than that of any other code for ?most? sources of ?most? classes. If, in addition, the classes can be efficiently distinguished by a certain decision rule, then the best attainable redundancy is given explicitly by the capacity of the active class plus the normalized negative logarithm of the prior probability assigned to this class. These results suggest some interesting guidelines as for the choice of the prior. We also discuss some examples with a natural hierarchical partition into classes

Keywords: information-theory source-coding
[Krichevskiy1998Laplace's] R. E. Krichevskiy. Laplace's law of succession and universal encoding. IEEE Trans. Inform. Theory, 44(1):296-303, Jan 1998. [ bib | .pdf ]
Keywords: information-theory source-coding
[Meron2004Finite-memory] E. Meron and M. Feder. Finite-memory universal prediction of individual sequences. IEEE Trans. Inform. Theory, 50(7):1506-1523, Jul 2004. [ bib | .pdf ]
The problem of predicting the next outcome of an individual binary sequence under the constraint that the universal predictor has a finite memory, is explored. In this analysis, the finite-memory universal predictors are either deterministic or random time-invariant finite-state (FS) machines with K states (K-state machines). The paper provides bounds on the asymptotic achievable regret of these constrained universal predictors as a function of K, the number of their states, for long enough sequences. The specific results are as follows. When the universal predictors are deterministic machines, the comparison class consists of constant predictors, and prediction is with respect to the 0-1 loss function (Hamming distance), we get tight bounds indicating that the optimal asymptotic regret is 1/(2K). In that case of K-state deterministic universal predictors, the constant predictors comparison class, but prediction is with respect to the self-information (code length) and the square-error loss functions, we show an upper bound on the regret (coding redundancy) of O(K/sup -2/3/) and a lower bound of /spl Theta/(K/sup -4/5/). For these loss functions, if the predictor is allowed to be a random K-state machine, i.e., a machine with random state transitions, we get a lower bound of /spl Theta/(1/K) on the regret, with a matching upper bound of O(1/K) for the square-error loss, and an upper bound of O(logK/K) Throughout the paper for the self-information loss. In addition, we provide results for all these loss functions in the case where the comparison class consists of all predictors that are order-L Markov machines.

Keywords: information-theory source-coding

This file was generated by bibtex2html 1.97.