Image Entropy coding originated in the 1940s with the introduction of
Shannon–Fano coding, Its highly efficient DCT-based compression algorithm was largely responsible for the wide proliferation of
digital images and
digital photos.
Lempel–Ziv–Welch (LZW) is a
lossless compression algorithm developed in 1984. It is used in the
GIF format, introduced in 1987.
DEFLATE, a lossless compression algorithm specified in 1996, is used in the
Portable Network Graphics (PNG) format.
Wavelet compression, the use of
wavelets in image compression, began after the development of DCT coding. In contrast to the DCT algorithm used by the original JPEG format, JPEG 2000 instead uses
discrete wavelet transform (DWT) algorithms. JPEG 2000 technology, which includes the
Motion JPEG 2000 extension, was selected as the
video coding standard for
digital cinema in 2004.
Audio Audio data compression, not to be confused with
dynamic range compression, has the potential to reduce the transmission
bandwidth and storage requirements of audio data.
Audio compression formats compression algorithms are implemented in
software as audio
codecs. In both lossy and lossless compression,
information redundancy is reduced, using methods such as
coding,
quantization, DCT and
linear prediction to reduce the amount of information used to represent the uncompressed data. Lossy audio compression algorithms provide higher compression and are used in numerous audio applications including
Vorbis and
MP3. These algorithms almost all rely on
psychoacoustics to eliminate or reduce fidelity of less audible sounds, thereby reducing the space required to store or transmit them. The acceptable trade-off between loss of audio quality and transmission or storage size depends upon the application. For example, one 640 MB
compact disc (CD) holds approximately one hour of uncompressed
high fidelity music, less than 2 hours of music compressed losslessly, or 7 hours of music compressed in the
MP3 format at a medium
bit rate. A digital sound recorder can typically store around 200 hours of clearly intelligible speech in 640 MB.
Perceptual coding was first used for
speech coding compression, with
linear predictive coding (LPC). Initial concepts for LPC date back to the work of
Fumitada Itakura (
Nagoya University) and Shuzo Saito (
Nippon Telegraph and Telephone) in 1966. During the 1970s,
Bishnu S. Atal and
Manfred R. Schroeder at
Bell Labs developed a form of LPC called
adaptive predictive coding (APC), a perceptual coding algorithm that exploited the masking properties of the human ear, followed in the early 1980s with the
code-excited linear prediction (CELP) algorithm which achieved a significant
compression ratio for its time.
Dolby Digital, and AAC. MDCT was proposed by J. P. Princen, A. W. Johnson and A. B. Bradley in 1987, following earlier work by Princen and Bradley in 1986. The world's first commercial
broadcast automation audio compression system was developed by Oscar Bonello, an engineering professor at the
University of Buenos Aires. This broadcast automation system was launched in 1987 under the name
Audicom. A literature compendium for a large variety of audio coding systems was published in the IEEE's
Journal on Selected Areas in Communications (
JSAC), in February 1988. While there were some papers from before that time, this collection documented an entire variety of finished, working audio coders, nearly all of them using perceptual techniques and some kind of frequency analysis and back-end noiseless coding. Most video codecs are used alongside audio compression techniques to store the separate but complementary data streams as one combined package using so-called
container formats. The DCT, which is fundamental to modern video compression, It was the first
video coding format based on DCT compression. H.261 was developed by a number of companies, including
Hitachi,
PictureTel,
NTT,
BT and
Toshiba. The most popular
video coding standards used for codecs have been the
MPEG standards.
MPEG-1 was developed by the
Motion Picture Experts Group (MPEG) in 1991, and it was designed to compress
VHS-quality video. It was succeeded in 1994 by
MPEG-2/
H.262, MPEG-2 became the standard video format for
DVD and
SD digital television.
H.264/MPEG-4 AVC was developed in 2003 by a number of organizations, primarily Panasonic,
Godo Kaisha IP Bridge and
LG Electronics. AVC commercially introduced the modern
context-adaptive binary arithmetic coding (CABAC) and
context-adaptive variable-length coding (CAVLC) algorithms. AVC is the main video encoding standard for
Blu-ray Discs, and is widely used by video sharing websites and streaming internet services such as
YouTube,
Netflix,
Vimeo, and
iTunes Store, web software such as
Adobe Flash Player and
Microsoft Silverlight, and various
HDTV broadcasts over terrestrial and satellite television.
Genetics Genetics compression algorithms are the latest generation of lossless algorithms that compress data (typically sequences of nucleotides) using both conventional compression algorithms and genetic algorithms adapted to the specific datatype. In 2012, a team of scientists from Johns Hopkins University published a genetic compression algorithm that does not use a reference genome for compression. HAPZIPPER was tailored for
HapMap data and achieves over 20-fold compression (95% reduction in file size), providing 2- to 4-fold better compression and is less computationally intensive than the leading general-purpose compression utilities. For this, Chanda, Elhaik, and Bader introduced MAF-based encoding (MAFE), which reduces the heterogeneity of the dataset by sorting SNPs by their minor allele frequency, thus homogenizing the dataset. Other algorithms developed in 2009 and 2013 (DNAZip and GenomeZip) have compression ratios of up to 1200-fold—allowing 6 billion basepair diploid human genomes to be stored in 2.5 megabytes (relative to a reference genome or averaged over many genomes). For a benchmark in genetics/genomics data compressors, see == Outlook and currently unused potential ==