[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
article
Free access

An adaptive dependency source model for data compression

Published: 01 January 1989 Publication History

Abstract

By dynamically recoding data on the basis of current intercharacter probabilities, the entropy of encoded messages can be significantly reduced.

References

[1]
Gallager, R.G. Variations on a theme by Huffman. IEEE Trans. Inf. Theory IT-24, 6 (Nov. 1978), 668-674.
[2]
Huffman, D.A. A method for the construction of minimumredundancy codes. Proc. Inst. Electr. Radio Eng. 40, 9 (Sept. 1952), 1098-1101.
[3]
Witten, I.H., Neal, R.M., and Cleary, J.G. Arithmetic coding for data compression. Commun. ACM 30, 6 (June 1987), 520-540.

Cited By

View all
  • (2005)A hybrid lossless compression of still images using Markov model and linear predictionImage Analysis and Processing10.1007/3-540-60298-4_259(203-208)Online publication date: 2-Jun-2005
  • (2005)DZ A text compression algorithm for natural languagesCombinatorial Pattern Matching10.1007/3-540-56024-6_16(193-204)Online publication date: 4-Jun-2005
  • (2003)Intelligent measurement multiplex systemSecond IEEE International Workshop on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, 2003. Proceedings10.1109/IDAACS.2003.1249506(13-15)Online publication date: 2003
  • Show More Cited By

Recommendations

Reviews

Timothy C. Bell

This paper describes a novel method for data compression that was inspired by a recent description of arithmetic coding [1]. The method is relatively simple to implement, although its compression and speed characteristics are not competitive with currently popular techniques such as Ziv-Lempel coding. Like Huffman coding, arithmetic coding achieves compression by representing high probability symbols with a small number of bits and vice versa. Arithmetic coding performs near optimally for a given probability distribution; the real challenge is to determine the probabilities as accurately as possible, a task known as modeling. Witten, Neal, and Cleary demonstrate arithmetic coding with a simple model that uses a character's relative frequency as an estimate of its probability [1]. Abrahamson observes that considering a character's context will result in a more accurate estimate. For example, while the relative frequency of a u in English text is normally low, after the letter q it is high. Although Abrahamson does not use the term, such a model is often called a first-order context model. Witten, Neal, and Cleary mention context models in passing and refer to a method that uses contexts to achieve a particularly good compression rate (they quote a compression of 2.2 bits/character on English text) [2]. Abrahamson's proposed scheme is not a full first-order model. It maintains the rank of the characters for each first-order context (where frequency determines rank). To code a character, the scheme determines its rank for the current context and codes this rank via an adaptive zero-order model. This approach works best when the probability distribution of ranks is independent of the context, but such a situation is unlikely to occur in practice. Abrahamson empirically compares the compression of his scheme to that of a straight zero-order model and finds that his method performs a little better, with a compression of 3.9 bits/character for a file of English text. He pays considerable attention to the periodic halving of frequency counts in order to achieve a recency effect, including two rather unconventional graphs that show the relationship between halving frequency, speed, and compression. Although the proposed compression method compares favorably with some simple methods, it offers little advantage over Ziv-Lempel methods (which are fast and give at least as good compression) or high-order context models like those of Cleary and Witten [2] (which achieve vastly superior compression rates).

Access critical reviews of Computing literature here

Become a reviewer for Computing Reviews.

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Communications of the ACM
Communications of the ACM  Volume 32, Issue 1
Jan. 1989
138 pages
ISSN:0001-0782
EISSN:1557-7317
DOI:10.1145/63238
Issue’s Table of Contents
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 January 1989
Published in CACM Volume 32, Issue 1

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)41
  • Downloads (Last 6 weeks)7
Reflects downloads up to 04 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2005)A hybrid lossless compression of still images using Markov model and linear predictionImage Analysis and Processing10.1007/3-540-60298-4_259(203-208)Online publication date: 2-Jun-2005
  • (2005)DZ A text compression algorithm for natural languagesCombinatorial Pattern Matching10.1007/3-540-56024-6_16(193-204)Online publication date: 4-Jun-2005
  • (2003)Intelligent measurement multiplex systemSecond IEEE International Workshop on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, 2003. Proceedings10.1109/IDAACS.2003.1249506(13-15)Online publication date: 2003
  • (2001)Optimal encoding of non-stationary sourcesInformation Sciences: an International Journal10.1016/S0020-0255(01)00103-7135:1-2(87-105)Online publication date: 28-Jun-2001
  • (1998)Optimal lossless compression of a class of dynamic sourcesProceedings DCC '98 Data Compression Conference (Cat. No.98TB100225)10.1109/DCC.1998.672221(501-510)Online publication date: 1998
  • (1998)An adaptive multialphabet arithmetic coding for video compressionIEEE Transactions on Circuits and Systems for Video Technology10.1109/76.6640978:2(130-137)Online publication date: 1-Apr-1998
  • (1997)A corpus for the evaluation of lossless compression algorithmsProceedings DCC '97. Data Compression Conference10.1109/DCC.1997.582019(201-210)Online publication date: 1997
  • (1994)CRUSH: a comparative lossless compression packageProceedings of IGARSS '94 - 1994 IEEE International Geoscience and Remote Sensing Symposium10.1109/IGARSS.1994.399112(310-312)Online publication date: 1994
  • (1994)Earth science data compression issues and activitiesRemote Sensing Reviews10.1080/027572594095322329:4(271-298)Online publication date: Jun-1994
  • (1993)A Knowledge Based Sensor Fusion EditorMultisensor Fusion for Computer Vision10.1007/978-3-662-02957-2_20(325-341)Online publication date: 1993
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Full Access

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media