scispace - formally typeset
Search or ask a question
Topic

Lempel–Ziv–Stac

About: Lempel–Ziv–Stac is a research topic. Over the lifetime, 254 publications have been published within this topic receiving 12732 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: The compression ratio achieved by the proposed universal code uniformly approaches the lower bounds on the compression ratios attainable by block-to-variable codes and variable- to-block codes designed to match a completely specified source.
Abstract: A universal algorithm for sequential data compression is presented. Its performance is investigated with respect to a nonprobabilistic model of constrained sources. The compression ratio achieved by the proposed universal code uniformly approaches the lower bounds on the compression ratios attainable by block-to-variable codes and variable-to-block codes designed to match a completely specified source.

5,844 citations

Journal ArticleDOI
TL;DR: A new compression algorithm is introduced that is based on principles not found in existing commercial methods in that it dynamically adapts to the redundancy characteristics of the data being compressed, and serves to illustrate system problems inherent in using any compression scheme.
Abstract: Data stored on disks and tapes or transferred over communications links in commercial computer systems generally contains significant redundancy. A mechanism or procedure which recodes the data to lessen the redundancy could possibly double or triple the effective data densitites in stored or communicated data. Moreover, if compression is automatic, it can also aid in the rise of software development costs. A transparent compression mechanism could permit the use of "sloppy" data structures, in that empty space or sparse encoding of data would not greatly expand the use of storage space or transfer time; however , that requires a good compression procedure. Several problems encountered when common compression methods are integrated into computer systems have prevented the widespread use of automatic data compression. For example (1) poor runtime execution speeds interfere in the attainment of very high data rates; (2) most compression techniques are not flexible enough to process different types of redundancy; (3) blocks of compressed data that have unpredictable lengths present storage space management problems. Each compression ' This article was written while Welch was employed at Sperry Research Center; he is now employed with Digital Equipment Corporation. 8 m, 2 /R4/OflAb l strategy poses a different set of these problems and, consequently , the use of each strategy is restricted to applications where its inherent weaknesses present no critical problems. This article introduces a new compression algorithm that is based on principles not found in existing commercial methods. This algorithm avoids many of the problems associated with older methods in that it dynamically adapts to the redundancy characteristics of the data being compressed. An investigation into possible application of this algorithm yields insight into the compressibility of various types of data and serves to illustrate system problems inherent in using any compression scheme. For readers interested in simple but subtle procedures, some details of this algorithm and its implementations are also described. The focus throughout this article will be on transparent compression in which the computer programmer is not aware of the existence of compression except in system performance. This form of compression is "noiseless," the decompressed data is an exact replica of the input data, and the compression apparatus is given no special program information, such as data type or usage statistics. Transparency is perceived to be important because putting an extra burden on the application programmer would cause

2,426 citations

Journal Article
TL;DR: This article describes a simple general-purpose data compression algorithm, called Byte Pair Encoding (BPE), which provides almost as much compression as the popular Lempel, Ziv, and Welch method.
Abstract: Data compression is becoming increasingly important as a way to stretch disk space and speed up data transfers. This article describes a simple general-purpose data compression algorithm, called Byte Pair Encoding (BPE), which provides almost as much compression as the popular Lempel, Ziv, and Welch (LZW) method [3, 2]. (I mention the LZW method in particular because it delivers good overall performance and is widely used.) BPE’s compression speed is somewhat slower than LZW’s, but BPE’s expansion is faster. The main advantage of BPE is the small, fast expansion routine, ideal for applications with limited memory. The accompanying C code provides an efficient implementation of the algorithm.

657 citations

Book
01 Jan 1991
TL;DR: In this article, the authors present a guide to data compression techniques, including Shannon-Fano and Huffman coding techniques, lossy compression, JPEG compression algorithm, and fractal compression.
Abstract: From the Publisher: Topics in this guide to data compression techniques include the Shannon-Fano and Huffman coding techniques, Lossy compression, the JPEG compression algorithm, and fractal compression. Readers also study adaptive Huffman coding, arithmetic coding, dictionary compression methods, and learn to write C programs for nearly any environment. The disk illustrates each learned technique and demonstrates how data compression works.

618 citations

Book
01 Jul 1991
TL;DR: In this paper, the authors present a guide to data compression techniques, including Shannon-Fano and Huffman coding techniques, lossy compression, JPEG compression algorithm, and fractal compression.
Abstract: From the Publisher: Topics in this guide to data compression techniques include the Shannon-Fano and Huffman coding techniques, Lossy compression, the JPEG compression algorithm, and fractal compression. Readers also study adaptive Huffman coding, arithmetic coding, dictionary compression methods, and learn to write C programs for nearly any environment. The disk illustrates each learned technique and demonstrates how data compression works.

548 citations


Network Information
Related Topics (5)
Data compression
43.6K papers, 756.5K citations
69% related
Computational complexity theory
30.8K papers, 711.2K citations
65% related
Noise (signal processing)
61K papers, 621.1K citations
63% related
Correlation clustering
19.3K papers, 602.5K citations
62% related
Wavelet transform
52.7K papers, 954.3K citations
62% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20181
20179
201614
201522
201418
201317