Elsevier

Information Processing & Management

Volume 30, Issue 6, November–December 1994, Pages 791-804
Information Processing & Management

An empirical evaluation of coding methods for multi-symbol alphabets

https://doi.org/10.1016/0306-4573(94)90007-8Get rights and content

Abstract

Many contemporary data compression schemes distinguish two distinct components: modelling and coding. Modelling involves estimating probabilities for the input symbols, while coding involves generating a sequence of bits in the output based on those probabilities. Several different methods for coding have been proposed. The best known are Huffman's code and the more recent technique of arithmetic coding, but other viable methods include various fixed codes, fast approximations to arithmetic coding, and splay coding. Each offers a different trade-off between compression speed, memory requirements, and the amount of compression obtained. This paper evaluates the performance of these methods in several situations and determines which is the most suitable for particular classes of application.

References (42)

  • G.V. Cormack et al.

    Algorithms for adaptive Huffman codes

    Information Processing Letters

    (1984)
  • P.G. Howard et al.

    Analysis of arithmetic coding for data compression

    Information Processing & Management

    (1992)
  • D. Knuth

    Dynamic Huffman coding

    Journal of Algorithms

    (1985)
  • T.C. Bell et al.

    Text Compression

    (January 1990)
  • T.C. Bell et al.

    Data compression in full-text retrieval systems

    Journal of the American Society for Information Science

    (October 1993)
  • J. Bentley et al.

    A locally adaptive data compression scheme

    Communications of the ACM

    (April 1986)
  • A. Bookstein et al.

    Is Huffman coding dead?

  • A. Bookstein et al.

    A systematic approach to compressing a full-text retrieval system

    Information Processing & Management

    (1992)
  • D. Chevion et al.

    High efficiency, multiplication free approximation of arithmetic coding

  • J.G. Cleary et al.

    A comparison of enumerative and adaptive codes

    IEEE Transactions on Information Theory

    (March 1984)
  • J.G. Cleary et al.

    Data compression using adaptive coding and partial string matching

    IEEE Transactions on Communications

    (April 1984)
  • J.B. Connell

    A Huffman-Shannon-Fano code

    Proc. IEEE

    (July 1973)
  • P. Elias

    Universal codeword sets and representations of the integers

    IEEE Transactions on Information Theory

    (March 1975)
  • N. Faller

    An adaptive system for data compression

  • P.M. Fenwick

    A new data structure for cumulative probability tables

    Software—Practice and Experience

    (March 1994)
  • E.R. Fiala et al.

    Data compression with finite windows

    Communications of the ACM

    (April 1989)
  • R.G. Gallager

    Variations on a theme by Huffman

    IEEE Transactions on Information Theory

    (November 1978)
  • P. Guttman

    Practical dictionary/arithmetic data compression synthesis

  • National Institute of Standards. Special Publication 500-207

  • D. Hirschberg et al.

    Efficient decoding of prefix codes

    Communications of the ACM

    (April 1990)
  • R.N. Horspool et al.

    Constructing word-based text compression algorithms

  • Cited by (0)

    A preliminary presentation of this work was made at the 1993 IEEE Data Compression Conference.

    View full text