Discrete cosine transform

Revision as of 19:37, 19 May 2025 by imported>Kvng (review: let ref speak for itself. move punct out of math blocks.)
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Template:Short description

A discrete cosine transform (DCT) expresses a finite sequence of data points in terms of a sum of cosine functions oscillating at different frequencies. The DCT, first proposed by Nasir Ahmed in 1972, is a widely used transformation technique in signal processing and data compression. It is used in most digital media, including digital images (such as JPEG and HEIF), digital video (such as MPEG and Template:Nowrap), digital audio (such as Dolby Digital, MP3 and AAC), digital television (such as SDTV, HDTV and VOD), digital radio (such as AAC+ and DAB+), and speech coding (such as AAC-LD, Siren and Opus). DCTs are also important to numerous other applications in science and engineering, such as digital signal processing, telecommunication devices, reducing network bandwidth usage, and spectral methods for the numerical solution of partial differential equations.

A DCT is a Fourier-related transform similar to the discrete Fourier transform (DFT), but using only real numbers. The DCTs are generally related to Fourier series coefficients of a periodically and symmetrically extended sequence whereas DFTs are related to Fourier series coefficients of only periodically extended sequences. DCTs are equivalent to DFTs of roughly twice the length, operating on real data with even symmetry (since the Fourier transform of a real and even function is real and even), whereas in some variants the input or output data are shifted by half a sample.

There are eight standard DCT variants, of which four are common. The most common variant of discrete cosine transform is the type-II DCT, which is often called simply the DCT. This was the original DCT as first proposed by Ahmed. Its inverse, the type-III DCT, is correspondingly often called simply the inverse DCT or the IDCT. Two related transforms are the discrete sine transform (DST), which is equivalent to a DFT of real and odd functions, and the modified discrete cosine transform (MDCT), which is based on a DCT of overlapping data. Multidimensional DCTs (MD DCTs) are developed to extend the concept of DCT to multidimensional signals. A variety of fast algorithms have been developed to reduce the computational complexity of implementing DCT. One of these is the integer DCT (IntDCT),<ref name="Stankovic"/> an integer approximation of the standard DCT,<ref name="Britanak2010" />Template:Rp used in several ISO/IEC and ITU-T international standards.<ref name="Stankovic"/><ref name="Britanak2010"/>

DCT compression, also known as block compression, compresses data in sets of discrete DCT blocks.<ref name="Alikhani"/> DCT blocks sizes including 8x8 pixels for the standard DCT, and varied integer DCT sizes between 4x4 and 32x32 pixels.<ref name="Stankovic"/><ref name="apple"/> The DCT has a strong energy compaction property,<ref name="pubDCT"/><ref name="pubRaoYip"/> capable of achieving high quality at high data compression ratios.<ref name="Barbero"/><ref name="Lea">Template:Cite journal</ref> However, blocky compression artifacts can appear when heavy DCT compression is applied.

HistoryEdit

The DCT was first conceived by Nasir Ahmed while working at Kansas State University. The concept was proposed to the National Science Foundation in 1972. The DCT was originally intended for image compression.<ref name="Ahmed" /><ref name="Stankovic"/> Ahmed developed a practical DCT algorithm with his PhD students T. Raj Natarajan and K. R. Rao at the University of Texas at Arlington in 1973.<ref name="Ahmed"/> They presented their results in a January 1974 paper, titled Discrete Cosine Transform.<ref name="pubDCT" /><ref name="pubRaoYip" /><ref name="t81">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> It described what is now called the type-II DCT (DCT-II),<ref name="Britanak2010" />Template:Rp as well as the type-III inverse DCT (IDCT).<ref name="pubDCT"/>

Since its introduction in 1974, there has been significant research on the DCT.<ref name="t81"/> In 1977, Wen-Hsiung Chen published a paper with C. Harrison Smith and Stanley C. Fralick presenting a fast DCT algorithm.<ref name="A Fast Computational Algorithm for">Template:Cite journal</ref><ref name="t81"/> Further developments include a 1978 paper by M. J. Narasimha and A. M. Peterson, and a 1984 paper by B. G. Lee.<ref name="t81"/> These research papers, along with the original 1974 Ahmed paper and the 1977 Chen paper, were cited by the Joint Photographic Experts Group as the basis for JPEG's lossy image compression algorithm in 1992.<ref name="t81"/><ref name="chen">Template:Cite journal</ref>

The discrete sine transform (DST) was derived from the DCT, by replacing the Neumann condition at x=0 with a Dirichlet condition.<ref name="Britanak2010" />Template:Rp The DST was described in the 1974 DCT paper by Ahmed, Natarajan and Rao.<ref name="pubDCT"/> A type-I DST (DST-I) was later described by Anil K. Jain in 1976, and a type-II DST (DST-II) was then described by H.B. Kekra and J.K. Solanka in 1978.<ref>Template:Cite journal</ref>

In 1975, John A. Roese and Guner S. Robinson adapted the DCT for inter-frame motion-compensated video coding. They experimented with the DCT and the fast Fourier transform (FFT), developing inter-frame hybrid coders for both, and found that the DCT is the most efficient due to its reduced complexity, capable of compressing image data down to 0.25-bit per pixel for a videotelephone scene with image quality comparable to an intra-frame coder requiring 2-bit per pixel.<ref>Template:Cite book</ref><ref name="Roese">Template:Cite journal</ref> In 1979, Anil K. Jain and Jaswant R. Jain further developed motion-compensated DCT video compression,<ref>Template:Cite book</ref><ref name="ITU">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> also called block motion compensation.<ref name="ITU"/> This led to Chen developing a practical video compression algorithm, called motion-compensated DCT or adaptive scene coding, in 1981.<ref name="ITU"/> Motion-compensated DCT later became the standard coding technique for video compression from the late 1980s onwards.<ref name="Ghanbari"/><ref name="Li">Template:Cite book</ref>

A DCT variant, the modified discrete cosine transform (MDCT), was developed by John P. Princen, A.W. Johnson and Alan B. Bradley at the University of Surrey in 1987,<ref>Template:Cite book</ref> following earlier work by Princen and Bradley in 1986.<ref>Template:Cite journal</ref> The MDCT is used in most modern audio compression formats, such as Dolby Digital (AC-3),<ref name="Luo"/><ref name="Britanak2011"/> MP3 (which uses a hybrid DCT-FFT algorithm),<ref name="Guckert">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> Advanced Audio Coding (AAC),<ref name=brandenburg>{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> and Vorbis (Ogg).<ref name="vorbis-mdct"/>

Nasir Ahmed also developed a lossless DCT algorithm with Giridhar Mandyam and Neeraj Magotra at the University of New Mexico in 1995. This allows the DCT technique to be used for lossless compression of images. It is a modification of the original DCT algorithm, and incorporates elements of inverse DCT and delta modulation. It is a more effective lossless compression algorithm than entropy coding.<ref>Template:Cite journal</ref> Lossless DCT is also known as LDCT.<ref>Template:Cite book</ref>

ApplicationsEdit

The DCT is the most widely used transformation technique in signal processing,<ref name="Muchahary">Template:Cite book</ref> and by far the most widely used linear transform in data compression.<ref>Template:Cite book</ref> Uncompressed digital media as well as lossless compression have high memory and bandwidth requirements, which is significantly reduced by the DCT lossy compression technique,<ref name="Barbero">Template:Cite journal</ref><ref name="Lea"/> capable of achieving data compression ratios from 8:1 to 14:1 for near-studio-quality,<ref name="Barbero"/> up to 100:1 for acceptable-quality content.<ref name="Lea"/> DCT compression standards are used in digital media technologies, such as digital images, digital photos,<ref name="Atlantic">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref><ref name="epfl">Template:Cite news</ref> digital video,<ref name="Ghanbari">Template:Cite book</ref><ref name="Lee1995">Template:Cite journal</ref> streaming media,<ref name="Lee">Template:Cite book</ref> digital television, streaming television, video on demand (VOD),<ref name="Lea"/> digital cinema,<ref name="Luo"/> high-definition video (HD video), and high-definition television (HDTV).<ref name="Barbero"/><ref name="Shishikui">Template:Cite book</ref>

The DCT, and in particular the DCT-II, is often used in signal and image processing, especially for lossy compression, because it has a strong energy compaction property.<ref name="pubDCT"/><ref name="pubRaoYip"/> In typical applications, most of the signal information tends to be concentrated in a few low-frequency components of the DCT. For strongly correlated Markov processes, the DCT can approach the compaction efficiency of the Karhunen-Loève transform (which is optimal in the decorrelation sense). As explained below, this stems from the boundary conditions implicit in the cosine functions.

DCTs are widely employed in solving partial differential equations by spectral methods, where the different variants of the DCT correspond to slightly different even and odd boundary conditions at the two ends of the array.

DCTs are closely related to Chebyshev polynomials, and fast DCT algorithms (below) are used in Chebyshev approximation of arbitrary functions by series of Chebyshev polynomials, for example in Clenshaw–Curtis quadrature.

General applicationsEdit

The DCT is widely used in many applications, which include the following.

Template:Columns-list

Visual media standardsEdit

The DCT-II is an important image compression technique. It is used in image compression standards such as JPEG, and video compression standards such as Template:Nowrap, MJPEG, MPEG, DV, Theora and Daala. There, the two-dimensional DCT-II of <math>N \times N</math> blocks are computed and the results are quantized and entropy coded. In this case, <math>N</math> is typically 8 and the DCT-II formula is applied to each row and column of the block. The result is an 8 × 8 transform coefficient array in which the <math>(0,0)</math> element (top-left) is the DC (zero-frequency) component and entries with increasing vertical and horizontal index values represent higher vertical and horizontal spatial frequencies.

The integer DCT, an integer approximation of the DCT,<ref name="Britanak2010"/><ref name="Stankovic"/> is used in Advanced Video Coding (AVC),<ref name="Wang">Template:Cite journal</ref><ref name="Stankovic"/> introduced in 2003, and High Efficiency Video Coding (HEVC),<ref name="apple"/><ref name="Stankovic"/> introduced in 2013. The integer DCT is also used in the High Efficiency Image Format (HEIF), which uses a subset of the HEVC video coding format for coding still images.<ref name="apple"/> AVC uses 4 x 4 and 8 x 8 blocks. HEVC and HEIF use varied block sizes between 4 x 4 and 32 x 32 pixels.<ref name="apple"/><ref name="Stankovic"/> Template:As of, AVC is by far the most commonly used format for the recording, compression and distribution of video content, used by 91% of video developers, followed by HEVC which is used by 43% of developers.<ref name="Bitmovin">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref>

Image formatsEdit

Image compression standard Year Common applications
JPEG<ref name="Stankovic"/> 1992 The most widely used image compression standard<ref name="Hudson">Template:Cite journal</ref><ref>{{#invoke:citation/CS1|citation CitationClass=web

}}</ref> and digital image format.<ref name="Baraniuk">Template:Cite news</ref>

JPEG XR 2009 Open XML Paper Specification
WebP 2010 A graphic format that supports the lossy compression of digital images. Developed by Google.
High Efficiency Image Format (HEIF) 2013 Image file format based on HEVC compression. It improves compression over JPEG,<ref name="apple">{{#invoke:citation/CS1|citation CitationClass=web

}}</ref> and supports animation with much more efficient compression than the animated GIF format.<ref>{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref>

BPG 2014 Based on HEVC compression
JPEG XL<ref name="jxl">{{#invoke:citation/CS1|citation CitationClass=web

}}</ref> ||2020|| A royalty-free raster-graphics file format that supports both lossy and lossless compression.

Video formatsEdit

Video coding standard Year Common applications
Template:Nowrap<ref name="video-standards">{{#invoke:citation/CS1|citation CitationClass=web

}}</ref><ref>{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref> ||1988|| First of a family of video coding standards. Used primarily in older video conferencing and video telephone products.

Motion JPEG (MJPEG)<ref>Template:Cite book</ref> 1992 QuickTime, video editing, non-linear editing, digital cameras
MPEG-1 Video<ref name="Rao">Template:Cite book</ref> 1993 Digital video distribution on CD or Internet video
MPEG-2 Video (Template:Nowrap)<ref name="Rao"/> 1995 Storage and handling of digital images in broadcast applications, digital television, HDTV, cable, satellite, high-speed Internet, DVD video distribution
DV 1995 Camcorders, digital cassettes
H.263 (MPEG-4 Part 2)<ref name="video-standards"/> 1996 Video telephony over public switched telephone network (PSTN), Template:Nowrap, Integrated Services Digital Network (ISDN)<ref>Template:Cite news</ref><ref>Template:Cite book</ref>
Advanced Video Coding (AVC, Template:Nowrap, MPEG-4)<ref name="Stankovic"/><ref name="Wang"/> 2003 Popular HD video recording, compression and distribution format, Internet video, YouTube, Blu-ray Discs, HDTV broadcasts, web browsers, streaming television, mobile devices, consumer devices, Netflix,<ref name="Encodes">Template:Cite news</ref> video telephony, FaceTime<ref name="AppleInsider standards 1">{{#invoke:citation/CS1|citation CitationClass=web

}}</ref>

Theora 2004 Internet video, web browsers
VC-1 2006 Windows media, Blu-ray Discs
Apple ProRes 2007 citation CitationClass=web

}}</ref>

VP9 2010 A video codec developed by Google used in the WebM container format with HTML5.
High Efficiency Video Coding (HEVC, Template:Nowrap)<ref name="Stankovic"/><ref name="apple"/> 2013 Successor to the Template:Nowrap standard, having substantially improved compression capability
Daala 2013 Research video format by Xiph.org
AV1<ref name="AV1">{{#invoke:citation/CS1|citation CitationClass=web

}}</ref> ||2018|| An open source format based on VP10 (VP9's internal successor), Daala and Thor; used by content providers such as YouTube<ref name="YT AV1 Beta Playlist">{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref><ref name="YT AV1">{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref> and Netflix.<ref name="Netflix AV1 Android">{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref><ref name="Netflix AV1 TV">{{#invoke:citation/CS1|citation

CitationClass=web

}}</ref>

MDCT audio standardsEdit

Template:Further

General audioEdit

Audio compression standard Year Common applications
Dolby Digital (AC-3)<ref name="Luo" /><ref name="Britanak2011" /> 1991 Cinema, digital cinema, DVD, Blu-ray, streaming media, video games
Adaptive Transform Acoustic Coding (ATRAC)<ref name="Luo"/> 1992 MiniDisc
MP3<ref name="Guckert"/><ref name="Stankovic"/> 1993 Digital audio distribution, MP3 players, portable media players, streaming media
Perceptual Audio Coder (PAC)<ref name="Luo"/> 1996 Digital audio radio service (DARS)
Advanced Audio Coding (AAC / MP4 Audio)<ref name=brandenburg/><ref name="Luo"/> 1997 Digital audio distribution, portable media players, streaming media, game consoles, mobile devices, iOS, iTunes, Android, BlackBerry
High-Efficiency Advanced Audio Coding (AAC+)<ref name="Herre" /><ref name="Britanak" />Template:Rp 1997 Digital radio, digital audio broadcasting (DAB+),<ref name="Britanak" /> Digital Radio Mondiale (DRM)
Cook Codec 1998 RealAudio
Windows Media Audio (WMA)<ref name="Luo"/> 1999 Windows Media
Vorbis<ref name="vorbis-mdct" /><ref name="Luo" /> 2000 Digital audio distribution, radio stations, streaming media, video games, Spotify, Wikipedia
High-Definition Coding (HDC)<ref name="Jones"/> 2002 Digital radio, HD Radio
Dynamic Resolution Adaptation (DRA)<ref name="Luo"/> 2008 China national audio standard, China Multimedia Mobile Broadcasting, DVB-H
Opus<ref name="Valin" /> 2012 VoIP,<ref name="homepage" /> mobile telephony, WhatsApp,<ref name="Register" /><ref name="Hazra" /><ref name="Srivastava" /> PlayStation 4<ref name="PlayStation" />
Dolby AC-4<ref name="Dolby AC-4" /> 2015 ATSC 3.0, ultra-high-definition television (UHD TV)
MPEG-H 3D Audio<ref name="Bleidt" />

Speech codingEdit

Speech coding standard Year Common applications
AAC-LD (LD-MDCT)<ref name="Schnell" /> 1999 Mobile telephony, voice-over-IP (VoIP), iOS, FaceTime<ref name="AppleInsider standards 1"/>
Siren<ref name="Hersent" /> 1999 VoIP, wideband audio, G.722.1
G.722.1<ref name="Lutzky" /> 1999 VoIP, wideband audio, G.722
G.729.1<ref name="Nagireddi" /> 2006 G.729, VoIP, wideband audio,<ref name="Nagireddi" /> mobile telephony
EVRC-WB<ref name="Britanak"/>Template:Rp 2007 Wideband audio
G.718<ref name="ITU-T" /> 2008 VoIP, wideband audio, mobile telephony
G.719<ref name="Britanak"/> 2008 Teleconferencing, videoconferencing, voice mail
CELT<ref name="Terriberry" /> 2011 VoIP,<ref name="ekiga" /><ref name="FreeSWITCH" /> mobile telephony
Enhanced Voice Services (EVS)<ref name="EVS" /> 2014 Mobile telephony, VoIP, wideband audio

Multidimensional DCTEdit

Template:See also

Multidimensional DCTs (MD DCTs) have several applications, mainly 3-D DCTs such as the 3-D DCT-II, which has several new applications like Hyperspectral Imaging coding systems,<ref name="appDCT">Template:Citation</ref> variable temporal length 3-D DCT coding,<ref name="app2DCT">Template:Citation</ref> video coding algorithms,<ref name="app3DCT">Template:Citation</ref> adaptive video coding<ref name="app4DCT">Template:Citation</ref> and 3-D Compression.<ref name="app5DCT">Template:Citation</ref> Due to enhancement in the hardware, software and introduction of several fast algorithms, the necessity of using MD DCTs is rapidly increasing. DCT-IV has gained popularity for its applications in fast implementation of real-valued polyphase filtering banks,<ref>Template:Cite book</ref> lapped orthogonal transform<ref>Template:Cite journal</ref>Template:Sfn and cosine-modulated wavelet bases.<ref>Template:Cite journal</ref>

Digital signal processingEdit

DCT plays an important role in digital signal processing specifically data compression. The DCT is widely implemented in digital signal processors (DSP), as well as digital signal processing software. Many companies have developed DSPs based on DCT technology. DCTs are widely used for applications such as encoding, decoding, video, audio, multiplexing, control signals, signaling, and analog-to-digital conversion. DCTs are also commonly used for high-definition television (HDTV) encoder/decoder chips.<ref name="Stankovic"/>

Compression artifactsEdit

A common issue with DCT compression in digital media are blocky compression artifacts,<ref name="Katsaggelos">Template:Cite book</ref> caused by DCT blocks.<ref name="Alikhani">{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> In a DCT algorithm, an image (or frame in an image sequence) is divided into square blocks which are processed independently from each other, then the DCT blocks is taken within each block and the resulting DCT coefficients are quantized. This process can cause blocking artifacts, primarily at high data compression ratios.<ref name="Katsaggelos"/> This can also cause the mosquito noise effect, commonly found in digital video.<ref>{{#invoke:citation/CS1|citation |CitationClass=web }}</ref>

DCT blocks are often used in glitch art.<ref name="Alikhani"/> The artist Rosa Menkman makes use of DCT-based compression artifacts in her glitch art,<ref name="Menkman">Template:Cite book</ref> particularly the DCT blocks found in most digital media formats such as JPEG digital images and MP3 audio.<ref name="Alikhani"/> Another example is Jpegs by German photographer Thomas Ruff, which uses intentional JPEG artifacts as the basis of the picture's style.<ref>Template:Cite book</ref><ref>{{#invoke:citation/CS1|citation |CitationClass=web }}</ref>

Informal overviewEdit

Like any Fourier-related transform, DCTs express a function or a signal in terms of a sum of sinusoids with different frequencies and amplitudes. Like the DFT, a DCT operates on a function at a finite number of discrete data points. The obvious distinction between a DCT and a DFT is that the former uses only cosine functions, while the latter uses both cosines and sines (in the form of complex exponentials). However, this visible difference is merely a consequence of a deeper distinction: a DCT implies different boundary conditions from the DFT or other related transforms.

The Fourier-related transforms that operate on a function over a finite domain, such as the DFT or DCT or a Fourier series, can be thought of as implicitly defining an extension of that function outside the domain. That is, once you write a function <math>f(x)</math> as a sum of sinusoids, you can evaluate that sum at any <math>x</math>, even for <math>x</math> where the original <math>f(x)</math> was not specified. The DFT, like the Fourier series, implies a periodic extension of the original function. A DCT, like a cosine transform, implies an even extension of the original function.

File:DCT-symmetries.svg
Illustration of the implicit even/odd extensions of DCT input data, for N=11 data points (red dots), for the four most common types of DCT (types I-IV). Note the subtle differences at the interfaces between the data and the extensions: in DCT-II and DCT-IV both the end points are replicated in the extensions but not in DCT-I or DCT-III (and a zero point is inserted at the sign reversal extension in DCT-III).

However, because DCTs operate on finite, discrete sequences, two issues arise that do not apply for the continuous cosine transform. First, one has to specify whether the function is even or odd at both the left and right boundaries of the domain (i.e. the min-n and max-n boundaries in the definitions below, respectively). Second, one has to specify around what point the function is even or odd. In particular, consider a sequence abcd of four equally spaced data points, and say that we specify an even left boundary. There are two sensible possibilities: either the data are even about the sample a, in which case the even extension is dcbabcd, or the data are even about the point halfway between a and the previous point, in which case the even extension is dcbaabcd (a is repeated).

Each boundary can be either even or odd (2 choices per boundary) and can be symmetric about a data point or the point halfway between two data points (2 choices per boundary), for a total of 2 × 2 × 2 × 2 = 16 possibilities. These choices lead to all the standard variations of DCTs and also discrete sine transforms (DSTs). Half of these possibilities, those where the left boundary is even, correspond to the 8 types of DCT; the other half are the 8 types of DST.

These different boundary conditions strongly affect the applications of the transform and lead to uniquely useful properties for the various DCT types. Most directly, when using Fourier-related transforms to solve partial differential equations by spectral methods, the boundary conditions are directly specified as a part of the problem being solved. Or, for the MDCT (based on the type-IV DCT), the boundary conditions are intimately involved in the MDCT's critical property of time-domain aliasing cancellation. In a more subtle fashion, the boundary conditions are responsible for the energy compactification properties that make DCTs useful for image and audio compression, because the boundaries affect the rate of convergence of any Fourier-like series.

In particular, it is well known that any discontinuities in a function reduce the rate of convergence of the Fourier series so that more sinusoids are needed to represent the function with a given accuracy. The same principle governs the usefulness of the DFT and other transforms for signal compression; the smoother a function is, the fewer terms in its DFT or DCT are required to represent it accurately, and the more it can be compressed.Template:Efn However, the implicit periodicity of the DFT means that discontinuities usually occur at the boundaries: any random segment of a signal is unlikely to have the same value at both the left and right boundaries.Template:Efn In contrast, a DCT where both boundaries are even always yields a continuous extension at the boundaries (although the slope is generally discontinuous). This is why DCTs, and in particular DCTs of types I, II, V, and VI (the types that have two even boundaries) generally perform better for signal compression than DFTs and DSTs. In practice, a type-II DCT is usually preferred for such applications, in part for reasons of computational convenience.

Formal definitionEdit

Formally, the discrete cosine transform is a linear, invertible function <math> f : \R^{N} \to \R^{N} </math> (where <math> \R</math> denotes the set of real numbers), or equivalently an invertible Template:Mvar × Template:Mvar square matrix. There are several variants of the DCT with slightly modified definitions. The Template:Mvar real numbers <math>~ x_0,\ \ldots\ x_{N - 1} ~</math> are transformed into the Template:Mvar real numbers <math> X_0,\, \ldots,\, X_{N - 1} </math> according to one of the formulas:

DCT-IEdit

<math>X_k
= \frac{1}{2} (x_0 + (-1)^k x_{N-1}) 
+ \sum_{n=1}^{N-2} x_n \cos \left[\, \tfrac{\ \pi}{\,N-1\,} \, n \, k \,\right]
\qquad \text{ for } ~ k = 0,\ \ldots\ N-1 ~.</math>

Some authors further multiply the <math>x_0 </math> and <math> x_{N-1} </math> terms by <math> \sqrt{2\,}\,</math> and correspondingly multiply the <math>X_0</math> and <math>X_{N-1}</math> terms by <math>1/\sqrt{2\,} \,</math> which, if one further multiplies by an overall scale factor of <math display="inline">\sqrt{\tfrac{2}{N-1\,}\,}</math>, makes the DCT-I matrix orthogonal but breaks the direct correspondence with a real-even DFT.

The DCT-I is exactly equivalent (up to an overall scale factor of 2), to a DFT of <math> 2(N-1) </math> real numbers with even symmetry. For example, a DCT-I of <math>N = 5 </math> real numbers <math> a\ b\ c\ d\ e </math> is exactly equivalent to a DFT of eight real numbers Template:Not a typo (even symmetry), divided by two. (In contrast, DCT types II-IV involve a half-sample shift in the equivalent DFT.)

Note, however, that the DCT-I is not defined for <math>N</math> less than 2, while all other DCT types are defined for any positive <math>N</math>.

Thus, the DCT-I corresponds to the boundary conditions: <math>x_n</math> is even around <math>n = 0</math> and even around <math>n = N - 1</math>; similarly for <math>X_k</math>.

DCT-IIEdit

<math>X_k =
\sum_{n=0}^{N-1} x_n \cos \left[\, \tfrac{\,\pi\,}{N} \left( n + \tfrac{1}{2} \right) k \, \right]
\qquad \text{ for } ~ k = 0,\ \dots\ N-1 ~.</math>

The DCT-II is probably the most commonly used form, and is often simply referred to as the DCT.<ref name="pubDCT"/><ref name="pubRaoYip"/>

This transform is exactly equivalent (up to an overall scale factor of 2) to a DFT of <math>4N</math> real inputs of even symmetry, where the even-indexed elements are zero. That is, it is half of the DFT of the <math>4N</math> inputs <math> y_n ,</math> where <math> y_{2n} = 0</math>, <math> y_{2n+1} = x_n </math> for <math>0 \leq n < N</math>, <math>y_{2N} = 0</math>, and <math>y_{4N-n} = y_n</math> for <math>0 < n < 2N</math>. DCT-II transformation is also possible using <math>2N</math> signal followed by a multiplication by half shift. This is demonstrated by Makhoul.Template:Cn

Some authors further multiply the <math>X_0</math> term by <math>1/\sqrt{N\,} \,</math> and multiply the rest of the matrix by an overall scale factor of <math display="inline">\sqrt{{2}/{N}}</math> (see below for the corresponding change in DCT-III). This makes the DCT-II matrix orthogonal, but breaks the direct correspondence with a real-even DFT of half-shifted input. This is the normalization used by Matlab.<ref>{{#invoke:citation/CS1|citation |CitationClass=web }}</ref> In many applications, such as JPEG, the scaling is arbitrary because scale factors can be combined with a subsequent computational step (e.g. the quantization step in JPEG<ref>Template:Cite book</ref>), and a scaling can be chosen that allows the DCT to be computed with fewer multiplications.<ref>Template:Cite journal</ref><ref>Template:Cite journal</ref>

The DCT-II implies the boundary conditions: <math>x_n</math> is even around <math>n = -1/2</math> and even around <math>n = N - 1/2 \,</math>; <math> X_k </math> is even around <math>k = 0</math> and odd around <math>k = N</math>.

DCT-IIIEdit

<math> X_k =
 \tfrac{1}{2} x_0 +
 \sum_{n=1}^{N-1} x_n \cos \left[\, \tfrac{\,\pi\,}{N} \left( k + \tfrac{1}{2} \right) n \,\right]
 \qquad \text{ for } ~ k = 0,\ \ldots\ N-1 ~.</math>

Because it is the inverse of DCT-II up to a scale factor (see below), this form is sometimes simply referred to as "the inverse DCT" ("IDCT").<ref name="pubRaoYip"/>

Some authors divide the <math>x_0</math> term by <math>\sqrt{2}</math> instead of by 2 (resulting in an overall <math>x_0/\sqrt{2}</math> term) and multiply the resulting matrix by an overall scale factor of <math display="inline"> \sqrt{2/N}</math> (see above for the corresponding change in DCT-II), so that the DCT-II and DCT-III are transposes of one another. This makes the DCT-III matrix orthogonal, but breaks the direct correspondence with a real-even DFT of half-shifted output.

The DCT-III implies the boundary conditions: <math>x_n</math> is even around <math>n = 0</math> and odd around <math>n = N ;</math> <math>X_k</math> is even around <math>k = -1/2</math> and even around <math>k = N - 1/2.</math>

DCT-IVEdit

<math>X_k =
\sum_{n=0}^{N-1} x_n \cos \left[\, \tfrac{\,\pi\,}{N} \, \left(n + \tfrac{1}{2} \right)\left(k + \tfrac{1}{2} \right) \,\right]
\qquad \text{ for } k = 0,\ \ldots\ N-1 ~.</math>

The DCT-IV matrix becomes orthogonal (and thus, being clearly symmetric, its own inverse) if one further multiplies by an overall scale factor of <math display="inline"> \sqrt{2/N}.</math>

A variant of the DCT-IV, where data from different transforms are overlapped, is called the modified discrete cosine transform (MDCT).<ref>Template:Harvnb</ref>

The DCT-IV implies the boundary conditions: <math> x_n </math> is even around <math>n = -1/2</math> and odd around <math>n = N - 1/2;</math> similarly for <math>X_k.</math>

DCT V-VIIIEdit

DCTs of types I–IV treat both boundaries consistently regarding the point of symmetry: they are even/odd around either a data point for both boundaries or halfway between two data points for both boundaries. By contrast, DCTs of types V-VIII imply boundaries that are even/odd around a data point for one boundary and halfway between two data points for the other boundary.

In other words, DCT types I–IV are equivalent to real-even DFTs of even order (regardless of whether <math> N </math> is even or odd), since the corresponding DFT is of length <math> 2(N-1) </math> (for DCT-I) or <math> 4 N </math> (for DCT-II & III) or <math> 8 N </math> (for DCT-IV). The four additional types of discrete cosine transform<ref>Template:Harvnb</ref> correspond essentially to real-even DFTs of logically odd order, which have factors of <math> N \pm {1}/{2} </math> in the denominators of the cosine arguments.

However, these variants seem to be rarely used in practice. One reason, perhaps, is that FFT algorithms for odd-length DFTs are generally more complicated than FFT algorithms for even-length DFTs (e.g. the simplest radix-2 algorithms are only for even lengths), and this increased intricacy carries over to the DCTs as described below.

(The trivial real-even array, a length-one DFT (odd length) of a single number Template:Mvar , corresponds to a DCT-V of length <math> N = 1 .</math>)

Inverse transformsEdit

Using the normalization conventions above, the inverse of DCT-I is DCT-I multiplied by 2/(N − 1). The inverse of DCT-IV is DCT-IV multiplied by 2/N. The inverse of DCT-II is DCT-III multiplied by 2/N and vice versa.<ref name="pubRaoYip"/>

Like for the DFT, the normalization factor in front of these transform definitions is merely a convention and differs between treatments. For example, some authors multiply the transforms by <math display="inline">\sqrt{2/N}</math> so that the inverse does not require any additional multiplicative factor. Combined with appropriate factors of Template:Sqrt (see above), this can be used to make the transform matrix orthogonal.

Multidimensional DCTsEdit

Multidimensional variants of the various DCT types follow straightforwardly from the one-dimensional definitions: they are simply a separable product (equivalently, a composition) of DCTs along each dimension.

M-D DCT-IIEdit

For example, a two-dimensional DCT-II of an image or a matrix is simply the one-dimensional DCT-II, from above, performed along the rows and then along the columns (or vice versa). That is, the 2D DCT-II is given by the formula (omitting normalization and other scale factors, as above):

<math>

\begin{align} X_{k_1,k_2} &=

\sum_{n_1=0}^{N_1-1}

\left( \sum_{n_2=0}^{N_2-1}

x_{n_1,n_2} 

\cos \left[\frac{\pi}{N_2} \left(n_2+\frac{1}{2}\right) k_2 \right]\right) \cos \left[\frac{\pi}{N_1} \left(n_1+\frac{1}{2}\right) k_1 \right]\\ &= \sum_{n_1=0}^{N_1-1}

\sum_{n_2=0}^{N_2-1}
x_{n_1,n_2} 

\cos \left[\frac{\pi}{N_1} \left(n_1+\frac{1}{2}\right) k_1 \right] \cos \left[\frac{\pi}{N_2} \left(n_2+\frac{1}{2}\right) k_2 \right] . \end{align} </math>

The inverse of a multi-dimensional DCT is just a separable product of the inverses of the corresponding one-dimensional DCTs (see above), e.g. the one-dimensional inverses applied along one dimension at a time in a row-column algorithm.

The 3-D DCT-II is only the extension of 2-D DCT-II in three dimensional space and mathematically can be calculated by the formula

<math>

X_{k_1,k_2,k_3} =

\sum_{n_1=0}^{N_1-1}
\sum_{n_2=0}^{N_2-1}

\sum_{n_3=0}^{N_3-1}

x_{n_1,n_2,n_3} 

\cos \left[\frac{\pi}{N_1} \left(n_1+\frac{1}{2}\right) k_1 \right] \cos \left[\frac{\pi}{N_2} \left(n_2+\frac{1}{2}\right) k_2 \right] \cos \left[\frac{\pi}{N_3} \left(n_3+\frac{1}{2}\right) k_3 \right],\quad \text{for } k_i = 0,1,2,\dots,N_i-1. </math>

The inverse of 3-D DCT-II is 3-D DCT-III and can be computed from the formula given by

<math>

x_{n_1,n_2,n_3} =

\sum_{k_1=0}^{N_1-1}
\sum_{k_2=0}^{N_2-1}

\sum_{k_3=0}^{N_3-1}

X_{k_1,k_2,k_3} 

\cos \left[\frac{\pi}{N_1} \left(n_1+\frac{1}{2}\right) k_1 \right] \cos \left[\frac{\pi}{N_2} \left(n_2+\frac{1}{2}\right) k_2 \right] \cos \left[\frac{\pi}{N_3} \left(n_3+\frac{1}{2}\right) k_3 \right],\quad \text{for } n_i=0,1,2,\dots,N_i-1. </math>

Technically, computing a two-, three- (or -multi) dimensional DCT by sequences of one-dimensional DCTs along each dimension is known as a row-column algorithm. As with multidimensional FFT algorithms, however, there exist other methods to compute the same thing while performing the computations in a different order (i.e. interleaving/combining the algorithms for the different dimensions). Owing to the rapid growth in the applications based on the 3-D DCT, several fast algorithms are developed for the computation of 3-D DCT-II. Vector-Radix algorithms are applied for computing M-D DCT to reduce the computational complexity and to increase the computational speed. To compute 3-D DCT-II efficiently, a fast algorithm, Vector-Radix Decimation in Frequency (VR DIF) algorithm was developed.

3-D DCT-II VR DIFEdit

In order to apply the VR DIF algorithm the input data is to be formulated and rearranged as follows.<ref>Template:Cite journal</ref><ref name=":0">Template:Cite journal</ref> The transform size N × N × N is assumed to be 2.

File:Stages of the 3-D DCT-II VR DIF algorithm.jpg
The four basic stages of computing 3-D DCT-II using VR DIF Algorithm.
<math>

\begin{array}{lcl}\tilde{x}(n_1,n_2,n_3) =x(2n_1,2n_2,2n_3)\\ \tilde{x}(n_1,n_2,N-n_3-1)=x(2n_1,2n_2,2n_3+1)\\ \tilde{x}(n_1,N-n_2-1,n_3)=x(2n_1,2n_2+1,2n_3)\\ \tilde{x}(n_1,N-n_2-1,N-n_3-1)=x(2n_1,2n_2+1,2n_3+1)\\ \tilde{x}(N-n_1-1,n_2,n_3)=x(2n_1+1,2n_2,2n_3)\\ \tilde{x}(N-n_1-1,n_2,N-n_3-1)=x(2n_1+1,2n_2,2n_3+1)\\ \tilde{x}(N-n_1-1,N-n_2-1,n_3)=x(2n_1+1,2n_2+1,2n_3)\\ \tilde{x}(N-n_1-1,N-n_2-1,N-n_3-1)=x(2n_1+1,2n_2+1,2n_3+1)\\ \end{array} </math>

where <math>0\leq n_1,n_2,n_3 \leq \frac{N}{2} -1</math>

The figure to the adjacent shows the four stages that are involved in calculating 3-D DCT-II using VR DIF algorithm. The first stage is the 3-D reordering using the index mapping illustrated by the above equations. The second stage is the butterfly calculation. Each butterfly calculates eight points together as shown in the figure just below, where <math>c(\varphi_i)=\cos(\varphi_i)</math>.

The original 3-D DCT-II now can be written as

<math>X(k_1,k_2,k_3)=\sum_{n_1=1}^{N-1}\sum_{n_2=1}^{N-1}\sum_{n_3=1}^{N-1}\tilde{x}(n_1,n_2,n_3) \cos(\varphi k_1)\cos(\varphi k_2)\cos(\varphi k_3)

</math>

where <math>\varphi_i= \frac{\pi}{2N}(4N_i+1),\text{ and } i= 1,2,3.</math>

If the even and the odd parts of <math>k_1,k_2</math> and <math>k_3</math> and are considered, the general formula for the calculation of the 3-D DCT-II can be expressed as

File:Single butterfly of the 3-D DCT-II VR DIF algorithm.jpg
The single butterfly stage of VR DIF algorithm.
<math>X(k_1,k_2,k_3)=\sum_{n_1=1}^{\tfrac N 2 -1}\sum_{n_2=1}^{\tfrac N 2 -1}\sum_{n_1=1}^{\tfrac N 2 -1}\tilde{x}_{ijl}(n_1,n_2,n_3) \cos(\varphi (2k_1+i)\cos(\varphi (2k_2+j)

\cos(\varphi (2k_3+l))</math>

where

<math>\tilde{x}_{ijl}(n_1,n_2,n_3)=\tilde{x}(n_1,n_2,n_3)+(-1)^l\tilde{x}\left(n_1,n_2,n_3+\frac{n}{2}\right) </math>
<math>+(-1)^j\tilde{x}\left(n_1,n_2+\frac{n}{2},n_3\right)+(-1)^{j+l}\tilde{x}\left(n_1,n_2+\frac{n}{2},n_3+\frac{n}{2}\right) </math>
<math>+(-1)^i\tilde{x}\left(n_1+\frac{n}{2},n_2,n_3\right)+(-1)^{i+j}\tilde{x}\left(n_1+\frac{n}{2}+\frac{n}{2},n_2,n_3\right) </math>
<math>+(-1)^{i+l}\tilde{x}\left(n_1+\frac{n}{2},n_2,n_3+\frac{n}{3}\right)</math>
<math>+(-1)^{i+j+l}\tilde{x}\left(n_1+\frac{n}{2},n_2+\frac{n}{2},n_3+\frac{n}{2}\right) \text{ where } i,j,l= 0 \text{ or } 1.</math>
Arithmetic complexityEdit

The whole 3-D DCT calculation needs <math>~ [\log_2 N] ~</math> stages, and each stage involves <math>~ \tfrac{1}{8}\ N^3 ~</math> butterflies. The whole 3-D DCT requires <math>~ \left[ \tfrac{1}{8}\ N^3 \log_2 N \right] ~</math> butterflies to be computed. Each butterfly requires seven real multiplications (including trivial multiplications) and 24 real additions (including trivial additions). Therefore, the total number of real multiplications needed for this stage is <math>~ \left[ \tfrac{7}{8}\ N^3\ \log_2 N \right] ~,</math> and the total number of real additions i.e. including the post-additions (recursive additions) which can be calculated directly after the butterfly stage or after the bit-reverse stage are given by<ref name=":0" /> <math>~ \underbrace{\left[\frac{3}{2}N^3 \log_2N\right]}_\text{Real}+\underbrace{\left[\frac{3}{2}N^3 \log_2N-3N^3+3N^2\right]}_\text{Recursive} = \left[\frac{9}{2}N^3 \log_2N-3N^3+3N^2\right] ~.</math>

The conventional method to calculate MD-DCT-II is using a Row-Column-Frame (RCF) approach which is computationally complex and less productive on most advanced recent hardware platforms. The number of multiplications required to compute VR DIF Algorithm when compared to RCF algorithm are quite a few in number. The number of Multiplications and additions involved in RCF approach are given by <math>~\left[\frac{3}{2}N^3 \log_2 N \right]~</math> and <math>~ \left[\frac{9}{2}N^3 \log_2 N - 3N^3 + 3N^2 \right] ~,</math> respectively. From Table 1, it can be seen that the total number

TABLE 1 Comparison of VR DIF & RCF Algorithms for computing 3D-DCT-II
Transform Size 3D VR Mults RCF Mults 3D VR Adds RCF Adds
8 × 8 × 8 2.625 4.5 10.875 10.875
16 × 16 × 16 3.5 6 15.188 15.188
32 × 32 × 32 4.375 7.5 19.594 19.594
64 × 64 × 64 5.25 9 24.047 24.047

of multiplications associated with the 3-D DCT VR algorithm is less than that associated with the RCF approach by more than 40%. In addition, the RCF approach involves matrix transpose and more indexing and data swapping than the new VR algorithm. This makes the 3-D DCT VR algorithm more efficient and better suited for 3-D applications that involve the 3-D DCT-II such as video compression and other 3-D image processing applications.

The main consideration in choosing a fast algorithm is to avoid computational and structural complexities. As the technology of computers and DSPs advances, the execution time of arithmetic operations (multiplications and additions) is becoming very fast, and regular computational structure becomes the most important factor.<ref>Template:Cite journal</ref> Therefore, although the above proposed 3-D VR algorithm does not achieve the theoretical lower bound on the number of multiplications,<ref>Template:Cite journal</ref> it has a simpler computational structure as compared to other 3-D DCT algorithms. It can be implemented in place using a single butterfly and possesses the properties of the Cooley–Tukey FFT algorithm in 3-D. Hence, the 3-D VR presents a good choice for reducing arithmetic operations in the calculation of the 3-D DCT-II, while keeping the simple structure that characterize butterfly-style Cooley–Tukey FFT algorithms.

File:DCT-8x8.png
Two-dimensional DCT frequencies from the JPEG DCT

The image to the right shows a combination of horizontal and vertical frequencies for an Template:Nobr <math>(~ N_1 = N_2 = 8 ~)</math> two-dimensional DCT. Each step from left to right and top to bottom is an increase in frequency by 1/2 cycle. For example, moving right one from the top-left square yields a half-cycle increase in the horizontal frequency. Another move to the right yields two half-cycles. A move down yields two half-cycles horizontally and a half-cycle vertically. The source data Template:Nobr is transformed to a linear combination of these 64 frequency squares.

MD-DCT-IVEdit

The M-D DCT-IV is just an extension of 1-D DCT-IV on to Template:Mvar dimensional domain. The 2-D DCT-IV of a matrix or an image is given by

<math> X_{k,\ell} =
\sum_{n=0}^{N-1} \; \sum_{m=0}^{M-1} \ x_{n,m} \cos\left(\ \frac{\,( 2 m + 1 )( 2 k + 1 )\ \pi \,}{4N} \ \right) \cos\left(\ \frac{\, ( 2n + 1 )( 2 \ell + 1 )\ \pi \,}{4M} \ \right) ~,</math>
for <math>~~ k = 0,\ 1,\ 2\ \ldots\ N-1 ~~</math> and <math>~~ \ell= 0,\ 1,\ 2,\ \ldots\ M-1 ~.</math>

We can compute the MD DCT-IV using the regular row-column method or we can use the polynomial transform method<ref>Template:Cite book</ref> for the fast and efficient computation. The main idea of this algorithm is to use the Polynomial Transform to convert the multidimensional DCT into a series of 1-D DCTs directly. MD DCT-IV also has several applications in various fields.

ComputationEdit

Although the direct application of these formulas would require <math>~ \mathcal{O}(N^2) ~</math> operations, it is possible to compute the same thing with only <math>~ \mathcal{O}(N \log N ) ~</math> complexity by factorizing the computation similarly to the fast Fourier transform (FFT). One can also compute DCTs via FFTs combined with <math>~\mathcal{O}(N)~</math> pre- and post-processing steps. In general, <math>~\mathcal{O}(N \log N )~</math> methods to compute DCTs are known as fast cosine transform (FCT) algorithms.

The most efficient algorithms, in principle, are usually those that are specialized directly for the DCT, as opposed to using an ordinary FFT plus <math>~ \mathcal{O}(N) ~</math> extra operations (see below for an exception). However, even "specialized" DCT algorithms (including all of those that achieve the lowest known arithmetic counts, at least for power-of-two sizes) are typically closely related to FFT algorithms – since DCTs are essentially DFTs of real-even data, one can design a fast DCT algorithm by taking an FFT and eliminating the redundant operations due to this symmetry. This can even be done automatically Template:Harv. Algorithms based on the Cooley–Tukey FFT algorithm are most common, but any other FFT algorithm is also applicable. For example, the Winograd FFT algorithm leads to minimal-multiplication algorithms for the DFT, albeit generally at the cost of more additions, and a similar algorithm was proposed by Template:Harv for the DCT. Because the algorithms for DFTs, DCTs, and similar transforms are all so closely related, any improvement in algorithms for one transform will theoretically lead to immediate gains for the other transforms as well Template:Harv.

While DCT algorithms that employ an unmodified FFT often have some theoretical overhead compared to the best specialized DCT algorithms, the former also have a distinct advantage: Highly optimized FFT programs are widely available. Thus, in practice, it is often easier to obtain high performance for general lengths Template:Mvar with FFT-based algorithms.Template:Efn Specialized DCT algorithms, on the other hand, see widespread use for transforms of small, fixed sizes such as the Template:Nobr DCT-II used in JPEG compression, or the small DCTs (or MDCTs) typically used in audio compression. (Reduced code size may also be a reason to use a specialized DCT for embedded-device applications.)

In fact, even the DCT algorithms using an ordinary FFT are sometimes equivalent to pruning the redundant operations from a larger FFT of real-symmetric data, and they can even be optimal from the perspective of arithmetic counts. For example, a type-II DCT is equivalent to a DFT of size <math>~ 4N ~</math> with real-even symmetry whose even-indexed elements are zero. One of the most common methods for computing this via an FFT (e.g. the method used in FFTPACK and FFTW) was described by Template:Harvtxt and Template:Harvtxt, and this method in hindsight can be seen as one step of a radix-4 decimation-in-time Cooley–Tukey algorithm applied to the "logical" real-even DFT corresponding to the DCT-II.Template:Efn Because the even-indexed elements are zero, this radix-4 step is exactly the same as a split-radix step. If the subsequent size <math>~ N ~</math> real-data FFT is also performed by a real-data split-radix algorithm (as in Template:Harvtxt), then the resulting algorithm actually matches what was long the lowest published arithmetic count for the power-of-two DCT-II (<math>~ 2 N \log_2 N - N + 2 ~</math> real-arithmetic operationsTemplate:Efn).

A recent reduction in the operation count to <math>~ \tfrac{17}{9} N \log_2 N + \mathcal{O}(N)</math> also uses a real-data FFT.<ref>Template:Cite journal</ref> So, there is nothing intrinsically bad about computing the DCT via an FFT from an arithmetic perspective – it is sometimes merely a question of whether the corresponding FFT algorithm is optimal. (As a practical matter, the function-call overhead in invoking a separate FFT routine might be significant for small <math>~ N ~,</math> but this is an implementation rather than an algorithmic question since it can be solved by unrolling or inlining.)

Example of IDCTEdit

File:DCT filter comparison.png
An example showing eight different filters applied to a test image (top left) by multiplying its DCT spectrum (top right) with each filter.

Consider this 8x8 grayscale image of capital letter A.

File:Letter-a-8x8.png
Original size, scaled 10x (nearest neighbor), scaled 10x (bilinear).
File:Dct-table.png
Basis functions of the discrete cosine transformation with corresponding coefficients (specific for our image).
DCT of the image = <math> \begin{bmatrix} 6.1917 & -0.3411 & 1.2418 & 0.1492 & 0.1583 & 0.2742 & -0.0724 & 0.0561 \\ 0.2205 & 0.0214 & 0.4503 & 0.3947 & -0.7846 & -0.4391 & 0.1001 & -0.2554 \\ 1.0423 & 0.2214 & -1.0017 & -0.2720 & 0.0789 & -0.1952 & 0.2801 & 0.4713 \\ -0.2340 & -0.0392 & -0.2617 & -0.2866 & 0.6351 & 0.3501 & -0.1433 & 0.3550 \\ 0.2750 & 0.0226 & 0.1229 & 0.2183 & -0.2583 & -0.0742 & -0.2042 & -0.5906 \\ 0.0653 & 0.0428 & -0.4721 & -0.2905 & 0.4745 & 0.2875 & -0.0284 & -0.1311 \\ 0.3169 & 0.0541 & -0.1033 & -0.0225 & -0.0056 & 0.1017 & -0.1650 & -0.1500 \\ -0.2970 & -0.0627 & 0.1960 & 0.0644 & -0.1136 & -0.1031 & 0.1887 & 0.1444 \\ \end{bmatrix} </math>.

Each basis function is multiplied by its coefficient and then this product is added to the final image.

File:Idct-animation.gif
On the left is the final image. In the middle is the weighted function (multiplied by a coefficient) which is added to the final image. On the right is the current function and corresponding coefficient. Images are scaled (using bilinear interpolation) by factor 10×.

See alsoEdit

NotesEdit

Template:Reflist

ReferencesEdit

Template:Reflist

Further readingEdit

External linksEdit

Template:Sister project

Template:Compression Methods Template:Compression formats Template:DSP Template:Telecommunications