Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
White noise
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Type of signal in signal processing}} {{Other uses|White Noise (disambiguation){{!}}White Noise}} [[File:White noise.svg|thumb|right|The [[waveform]] of a [[Gaussian noise|Gaussian]] white noise signal plotted on a graph]] {{Colors of noise}} In [[signal processing]], '''white noise''' is a random [[signal]] having equal intensity at different [[frequencies]], giving it a constant [[power spectral density]].<ref>{{cite book|last=Carter, Mancini|first=Bruce, Ron|title=Op Amps for Everyone|year=2009|publisher=Texas Instruments|isbn=978-0-08-094948-2|pages=10–11}}</ref> The term is used with this or similar meanings in many scientific and technical disciplines, including [[physics]], [[acoustical engineering]], [[telecommunications]], and [[statistical forecasting]]. White noise refers to a statistical model for signals and signal sources, not to any specific signal. White noise draws its name from [[White#Optics|white light]],<ref>{{cite book |last=Stein |first=Michael L. |date=1999 |title=Interpolation of Spatial Data: Some Theory for Kriging |publisher=Springer |page=40 |isbn=978-1-4612-7166-6 |quote=white light is approximately an equal mixture of all visible frequencies of light, which was demonstrated by Isaac Newton|doi=10.1007/978-1-4612-1494-6 |series=Springer Series in Statistics }}</ref> although light that appears white generally does not have a flat power spectral density over the [[visible band]]. [[File:White-noise-mv255-240x180.png|thumb|An image of [[salt-and-pepper noise]]]] In [[discrete time]], white noise is a [[discrete signal]] whose [[sample (signal)|samples]] are regarded as a sequence of [[serial correlation|serially uncorrelated]] [[random variable]]s with zero [[mean]] and finite [[variance]]; a single realization of white noise is a '''random shock'''. In some contexts, it is also required that the samples be [[statistical independence|independent]] and have identical [[probability distribution]] (in other words [[independent and identically distributed random variables]] are the simplest representation of white noise).<ref>{{cite book |last=Stein |first=Michael L. |date=1999 |title=Interpolation of Spatial Data: Some Theory for Kriging |publisher=Springer |page=40 |isbn=978-1-4612-7166-6 |quote=The best-known generalized process is white noise, which can be thought of as a continuous time analogue to a sequence of independent and identically distributed observations.|doi=10.1007/978-1-4612-1494-6 |series=Springer Series in Statistics }}</ref> In particular, if each sample has a [[normal distribution]] with zero mean, the signal is said to be [[additive white Gaussian noise]].<ref>{{Cite book|last=Diebold|first= Frank|title=Elements of Forecasting |edition=Fourth |year=2007}}</ref> The samples of a white noise signal may be [[sequential]] in time, or arranged along one or more spatial dimensions. In [[digital image processing]], the [[pixel]]s of a white noise image are typically arranged in a rectangular grid, and are assumed to be independent random variables with [[uniform probability distribution]] over some interval. The concept can be defined also for signals spread over more complicated domains, such as a [[sphere]] or a [[torus]]. [[File:White-noise-sound-20sec-mono-44100Hz.ogg|right|thumb|The sound of white noise]] An '''{{Vanchor|infinite-bandwidth white noise signal}}''' is a purely theoretical construction. The bandwidth of white noise is limited in practice by the mechanism of noise generation, by the transmission medium and by finite observation capabilities. Thus, random signals are considered white noise if they are observed to have a flat spectrum over the range of frequencies that are relevant to the context. For an [[audio signal]], the relevant range is the band of audible sound frequencies (between 20 and 20,000 [[Hz]]). Such a signal is heard by the human ear as a hissing sound, resembling the /h/ sound in a sustained aspiration. On the other hand, the ''sh'' sound {{IPA|/ʃ/}} in ''ash'' is a colored noise because it has a [[formant]] structure. In [[music]] and [[acoustics]], the term ''white noise'' may be used for any signal that has a similar hissing sound.<!-- Other uses of the term --> In the context of [[Phylogenetic comparative methods|phylogenetically based statistical methods]], the term ''white noise'' can refer to a lack of phylogenetic pattern in comparative data.<ref>{{cite journal |title=Developmental trait evolution in trilobites |journal=Evolution |volume=66 |issue=2 |pages=314–329 |year=2011 |pmid=22276531 |doi=10.1111/j.1558-5646.2011.01447.x |last1=Fusco |first1=G |last2=Garland |first2=T. Jr |last3=Hunt |first3=G |last4=Hughes |first4=NC |s2cid=14726662 |doi-access=free }}</ref> In nontechnical contexts, it is sometimes used to mean "random talk without meaningful contents".<ref name=shipman> [[Claire Shipman]] (2005), ''[[Good Morning America]]'': "The political [[rhetoric]] on [[Social Security (United States)|Social Security]] is white noise." Said on [[American Broadcasting Company|ABC]]'s ''[[Good Morning America]]'' TV show, January 11, 2005.</ref><ref> [[Don DeLillo]] (1985), ''[[White Noise (novel)|White Noise]]'' </ref> ==Statistical properties== {{unreferenced section|date=January 2022}} [[File:Noise.jpg|thumb|240px|[[Spectrogram]] of [[pink noise]] (left) and white noise (right), shown with linear frequency axis (vertical) versus time axis (horizontal)]] Any distribution of values is possible (although it must have zero [[DC component]]). Even a binary signal which can only take on the values 1 or -1 will be white if the [[sequence]] is statistically uncorrelated. Noise having a continuous distribution, such as a [[normal distribution]], can of course be white. It is often incorrectly assumed that [[Gaussian noise]] (i.e., noise with a Gaussian amplitude distribution{{snd}}see [[normal distribution]]) necessarily refers to white noise, yet neither property implies the other. Gaussianity refers to the probability distribution with respect to the value, in this context the probability of the signal falling within any particular range of amplitudes, while the term 'white' refers to the way the signal power is distributed (i.e., independently) over time or among frequencies. One form of white noise is the generalized mean-square derivative of the [[Wiener process]] or [[Brownian motion]]. A generalization to [[random element]]s on infinite dimensional spaces, such as [[random field]]s, is the [[white noise measure]]. ==Practical applications== {{more citations needed section|date=January 2022}} ===Music=== White noise is commonly used in the production of [[electronic music]], usually either directly or as an input for a filter to create other types of noise signal. It is used extensively in [[audio synthesis]], typically to recreate percussive instruments such as [[cymbal]]s or [[snare drum]]s which have high noise content in their frequency domain.<ref>{{Cite web |last=Clark |first=Dexxter |title=Did you know all these white noise secrets? (music production tips) |url=https://www.learnhowtoproducemusic.com/blog-how-to-start-music-production/white-noise-in-music-production |access-date=2022-07-25 |website=www.learnhowtoproducemusic.com |language=en}}</ref> A simple example of white noise is a nonexistent radio station (static). ===Electronics engineering=== White noise is also used to obtain the [[impulse response]] of an electrical circuit, in particular of [[amplifier]]s and other audio equipment. It is not used for testing loudspeakers as its spectrum contains too great an amount of high-frequency content. [[Pink noise]], which differs from white noise in that it has equal energy in each octave, is used for testing transducers such as loudspeakers and microphones. ===Computing=== White noise is used as the basis of some [[hardware random number generator|random number generators]]. For example, [[Random.org]] uses a system of atmospheric antennas to generate random digit patterns from sources that can be well-modeled by white noise.<ref>{{Cite news |last=O'Connell |first=Pamela LiCalzi |date=8 April 2004 |title=Lottery Numbers and Books With a Voice |work=[[The New York Times]] |url=https://www.nytimes.com/2004/04/08/technology/online-diary.html |access-date=25 July 2022 |archive-url=https://web.archive.org/web/20090726093822/http://www.nytimes.com/2004/04/08/technology/circuits/08diar.html |archive-date=26 July 2009}}</ref> ===Tinnitus treatment=== White noise is a common synthetic noise source used for sound masking by a [[tinnitus masker]].<ref>{{ cite book|last=Jastreboff |first= P. J.|chapter= Tinnitus Habituation Therapy (THT) and Tinnitus Retraining Therapy (TRT)|title= Tinnitus Handbook|location= San Diego|publisher=Singular|year= 2000|pages=357–376}}</ref> [[White noise machine]]s and other white noise sources are sold as privacy enhancers and sleep aids (see [[music and sleep]]) and to mask [[tinnitus]].<ref>{{cite journal |title=Evidence based complementary intervention for insomnia |journal=Hawaii Med J |volume=61 |issue=9 |pages=192, 213 |date=September 2002 |pmid=12422383 |url=http://cogprints.org/5032/1/2002_H.M.J_White-noise_for_PTSD.pdf |last1=López |first1=HH |last2=Bracha |first2=AS |last3=Bracha |first3=HS}}</ref> The Marpac Sleep-Mate was the first domestic use white noise machine built in 1962 by traveling salesman Jim Buckwalter.<ref>{{Cite news|last=Green|first=Penelope|date=2018-12-27|title=The Sound of Silence|language=en-US|work=The New York Times|url=https://www.nytimes.com/2018/12/27/style/white-noise-machines.html|access-date=2021-05-20|issn=0362-4331}}</ref> Alternatively, the use of an AM radio tuned to unused frequencies ("static") is a simpler and more cost-effective source of white noise.<ref>{{Cite journal | issn = 0016-867X | volume = 58 | issue = 2 | pages = 28–34 | last = Noell | first = Courtney A |author2=William L Meyerhoff | title = Tinnitus. Diagnosis and treatment of this elusive symptom | journal = Geriatrics | date = February 2003 | pmid=12596495}}</ref> However, white noise generated from a common commercial radio receiver tuned to an unused frequency is extremely vulnerable to being contaminated with spurious signals, such as adjacent radio stations, harmonics from non-adjacent radio stations, electrical equipment in the vicinity of the receiving antenna causing interference, or even atmospheric events such as solar flares and especially lightning. ===Work environment=== The effects of white noise upon cognitive function are mixed. Recently, a small study found that white noise background stimulation improves cognitive functioning among secondary students with [[attention deficit hyperactivity disorder]] (ADHD), while decreasing performance of non-ADHD students.<ref>{{Cite journal| volume = 6| issue = 1| page = 55| last = Soderlund| first = Goran|author2=Sverker Sikstrom |author3=Jan Loftesnes |author4=Edmund Sonuga Barke | title = The effects of background white noise on memory performance in inattentive school children| journal = Behavioral and Brain Functions| year = 2010| doi=10.1186/1744-9081-6-55| pmid = 20920224| pmc = 2955636| doi-access = free}}</ref><ref>{{Cite journal | doi = 10.1111/j.1469-7610.2007.01749.x | pmid = 17683456 | issn = 0021-9630 | volume = 48 | issue = 8 | pages = 840–847 | last = Söderlund | first = Göran |author2=Sverker Sikström |author3=Andrew Smart | title = Listen to the noise: Noise is beneficial for cognitive performance in ADHD. | journal = Journal of Child Psychology and Psychiatry | year = 2007 | citeseerx = 10.1.1.452.530 }}</ref> Other work indicates it is effective in improving the mood and performance of workers by masking background office noise,<ref>{{Cite journal | doi = 10.1177/0013916592243006 | volume = 24 | issue = 3 | pages = 381–395 | last = Loewen | first = Laura J. |author2=Peter Suedfeld | title = Cognitive and Arousal Effects of Masking Office Noise | journal = Environment and Behavior | date = 1992-05-01 | bibcode = 1992EnvBe..24..381L | s2cid = 144443528 }}</ref> but decreases cognitive performance in complex card sorting tasks.<ref>{{Cite journal | issn = 0022-1309 | volume = 120 | issue = 3 | pages = 339–355 | last = Baker | first = Mary Anne |author2=Dennis H. Holding | title = The effects of noise and speech on cognitive task performance. | journal = Journal of General Psychology | date = July 1993 | doi=10.1080/00221309.1993.9711152| pmid = 8138798 }}</ref> Similarly, an experiment was carried out on sixty-six healthy participants to observe the benefits of using white noise in a learning environment. The experiment involved the participants identifying different images whilst having different sounds in the background. Overall the experiment showed that white noise does in fact have benefits in relation to learning. The experiments showed that white noise improved the participants' learning abilities and their recognition memory slightly.<ref>Rausch, V. H. (2014). White noise improves learning by modulating activity in dopaminergic midbrain regions and right superior temporal sulcus . Journal of cognitive neuroscience, 1469-1480</ref> ==Mathematical definitions== ===White noise vector=== A [[random vector]] (that is, a random variable with values in ''R<sup>n</sup>'') is said to be a white noise vector or white random vector if its components each have a [[probability distribution]] with zero mean and finite [[variance]],{{what|reason=Why aren't the variances required to be identical, like in the Gaussian case (two paragraphs below) and the discrete-time case (next section)?|date=October 2023}} and are [[statistically independent]]: that is, their [[joint probability distribution]] must be the product of the distributions of the individual components.<ref name="fessler"> Jeffrey A. Fessler (1998), ''[https://web.archive.org/web/20131218214647/http://andywilliamson.org/_/wp-content/uploads/2010/04/White-Noise.pdf On Transformations of Random Vectors.]'' Technical report 314, Dept. of Electrical Engineering and Computer Science, Univ. of Michigan. ([[PDF]])</ref> A necessary (but, [[normally distributed and uncorrelated does not imply independent|in general, not sufficient]]) condition for statistical independence of two variables is that they be [[correlation|statistically uncorrelated]]; that is, their [[covariance]] is zero. Therefore, the [[covariance matrix]] ''R'' of the components of a white noise vector ''w'' with ''n'' elements must be an ''n'' by ''n'' [[diagonal matrix]], where each diagonal element ''R<sub>ii</sub>'' is the [[variance]] of component ''w<sub>i</sub>''; and the [[Correlation and dependence#Correlation matrices|correlation]] matrix must be the ''n'' by ''n'' identity matrix. If, in addition to being independent, every variable in ''w'' also has a [[normal distribution]] with zero mean and the same variance <math>\sigma^2</math>, ''w'' is said to be a Gaussian white noise vector. In that case, the joint distribution of ''w'' is a [[multivariate normal distribution]]; the independence between the variables then implies that the distribution has [[elliptical distribution|spherical symmetry]] in ''n''-dimensional space. Therefore, any [[orthogonal transformation]] of the vector will result in a Gaussian white random vector. In particular, under most types of [[discrete Fourier transform]], such as [[FFT]] and [[discrete Hartley transform|Hartley]], the transform ''W'' of ''w'' will be a Gaussian white noise vector, too; that is, the ''n'' Fourier coefficients of ''w'' will be independent Gaussian variables with zero mean and the same variance <math>\sigma^2</math>. The [[power spectrum]] ''P'' of a random vector ''w'' can be defined as the expected value of the [[squared modulus]] of each coefficient of its Fourier transform ''W'', that is, ''P<sub>i</sub>'' = E(|''W<sub>i</sub>''|<sup>2</sup>). Under that definition, a Gaussian white noise vector will have a perfectly flat power spectrum, with ''P<sub>i</sub>'' = ''σ''<sup>2</sup> for all ''i''. If ''w'' is a white random vector, but not a Gaussian one, its Fourier coefficients ''W<sub>i</sub>'' will not be completely independent of each other; although for large ''n'' and common probability distributions the dependencies are very subtle, and their pairwise correlations can be assumed to be zero. Often the weaker condition statistically uncorrelated is used in the definition of white noise, instead of statistically independent. However, some of the commonly expected properties of white noise (such as flat power spectrum) may not hold for this weaker version. Under this assumption, the stricter version can be referred to explicitly as independent white noise vector.<ref name="ezivot">Eric Zivot and Jiahui Wang (2006), [http://faculty.washington.edu/ezivot/econ584/notes/timeSeriesConcepts.pdf Modeling Financial Time Series with S-PLUS]. Second Edition. ([[PDF]])</ref>{{rp|p.60}} Other authors use strongly white and weakly white instead.<ref name="diebold">[[Francis X. Diebold]] (2007), ''[https://www.sas.upenn.edu/~fdiebold/Teaching221/FullBook.pdf Elements of Forecasting],'' 4th edition. ([[PDF]])</ref> An example of a random vector that is Gaussian white noise in the weak but not in the strong sense is <math>x=[x_1,x_2]</math> where <math>x_1</math> is a normal random variable with zero mean, and <math>x_2</math> is equal to <math>+x_1</math> or to <math>-x_1</math>, with equal probability. These two variables are uncorrelated and individually normally distributed, but they are not jointly normally distributed and are not independent. If <math>x</math> is rotated by 45 degrees, its two components will still be uncorrelated, but their distribution will no longer be normal. In some situations, one may relax the definition by allowing each component of a white random vector <math>w</math> to have non-zero expected value <math>\mu</math>. In [[image processing]] especially, where samples are typically restricted to positive values, one often takes <math>\mu</math> to be one half of the maximum sample value. In that case, the Fourier coefficient <math>W_0</math> corresponding to the zero-frequency component (essentially, the average of the <math>w_i</math>) will also have a non-zero expected value <math>\mu\sqrt{n}</math>; and the power spectrum <math>P</math> will be flat only over the non-zero frequencies. ===Discrete-time white noise=== A discrete-time [[stochastic process]] <math>W(n)</math> is a generalization of a random vector with a finite number of components to infinitely many components. A discrete-time stochastic process <math>W(n)</math> is called white noise if its mean is equal to zero for all <math>n</math> , i.e. <math>\operatorname{E}[W(n)] = 0</math> and if the autocorrelation function <math>R_{W}(n) = \operatorname{E}[W(k+n)W(k)]</math> has a nonzero value only for <math>n = 0</math>, i.e. <math>R_{W}(n) = \sigma^2 \delta(n)</math>.{{citation needed|date=February 2024}}{{what|reason=Why aren't the W(n) required to be statistically independent, as for the finite-component white noise of the last section?|date=October 2023}} ===Continuous-time white noise=== In order to define the notion of white noise in the theory of [[continuous-time]] signals, one must replace the concept of a random vector by a continuous-time random signal; that is, a random process that generates a function <math>w</math> of a real-valued parameter <math>t</math>. Such a process is said to be white noise in the strongest sense if the value <math>w(t)</math> for any time <math>t</math> is a random variable that is statistically independent of its entire history before <math>t</math>. A weaker definition requires independence only between the values <math>w(t_1)</math> and <math>w(t_2)</math> at every pair of distinct times <math>t_1</math> and <math>t_2</math>. An even weaker definition requires only that such pairs <math>w(t_1)</math> and <math>w(t_2)</math> be uncorrelated.<ref name=econterms>[http://economics.about.com/od/economicsglossary/g/whitenoise.htm ''White noise process''] {{Webarchive|url=https://web.archive.org/web/20160911134507/http://economics.about.com/od/economicsglossary/g/whitenoise.htm |date=2016-09-11 }}. By Econterms via About.com. Accessed on 2013-02-12.</ref> As in the discrete case, some authors adopt the weaker definition for white noise, and use the qualifier independent to refer to either of the stronger definitions. Others use weakly white and strongly white to distinguish between them. However, a precise definition of these concepts is not trivial, because some quantities that are finite sums in the finite discrete case must be replaced by integrals that may not converge. Indeed, the set of all possible instances of a signal <math>w</math> is no longer a finite-dimensional space <math>\mathbb{R}^n</math>, but an infinite-dimensional [[function space]]. Moreover, by any definition a white noise signal <math>w</math> would have to be essentially discontinuous at every point; therefore even the simplest operations on <math>w</math>, like integration over a finite interval, require advanced mathematical machinery. Some authors{{citation needed|date=October 2023}}{{what|reason=Since the definition proposed in this section is not remotely workable in a mathematical sense, I doubt that any authors do this. Instead, we are looking at a heuristic only.|date=October 2023}} require each value <math>w(t)</math> to be a real-valued random variable with expectation <math>\mu</math> and some finite variance <math>\sigma^2</math>. Then the covariance <math>\mathrm{E}(w(t_1)\cdot w(t_2))</math> between the values at two times <math>t_1</math> and <math>t_2</math> is well-defined: it is zero if the times are distinct, and <math>\sigma^2</math> if they are equal. However, by this definition, the integral : <math>W_{[a,a+r]} = \int_a^{a+r} w(t)\, dt</math> over any interval with positive width <math>r</math> would be simply the width times the expectation: <math>r\mu</math>.{{what|reason=The *expectation value* of the mean is zero. And this is not a problem.|date=October 2023}} This property renders the concept inadequate as a model of white noise signals either in a physical or mathematical sense.{{what|reason=Why?|date=October 2023}} Therefore, most authors define the signal <math>w</math> indirectly by specifying random values for the integrals of <math>w(t)</math> and <math>|w(t)|^2</math> over each interval <math>[a,a+r]</math>. In this approach, however, the value of <math>w(t)</math> at an isolated time cannot be defined as a real-valued random variable{{Citation needed|reason=an authoritative work on white noise given one such example should be given|date=January 2017}}. Also the covariance <math>\mathrm{E}(w(t_1)\cdot w(t_2))</math> becomes infinite when <math>t_1=t_2</math>; and the [[autocorrelation]] function <math>\mathrm{R}(t_1,t_2)</math> must be defined as <math>N \delta(t_1-t_2)</math>, where <math>N</math> is some real constant and <math>\delta</math> is the [[Dirac delta function]].{{what|reason=Correlation can only take on values in [0,1], so N must be 1 and delta must take the value 1 for t_1 = t_2; it is not the dirac measure here. However, all these concepts are fishy.|date=October 2023}} In this approach, one usually specifies that the integral <math>W_I</math> of <math>w(t)</math> over an interval <math>I=[a,b]</math> is a real random variable with normal distribution, zero mean, and variance <math>(b-a)\sigma^2</math>; and also that the covariance <math>\mathrm{E}(W_I\cdot W_J)</math> of the integrals <math>W_I</math>, <math>W_J</math> is <math>r\sigma^2</math>, where <math>r</math> is the width of the intersection <math>I\cap J</math> of the two intervals <math>I,J</math>. This model is called a Gaussian white noise signal (or process). In the mathematical field known as [[white noise analysis]], a Gaussian white noise <math>w</math> is defined as a stochastic tempered distribution, i.e. a random variable with values in the space <math>\mathcal S'(\mathbb R)</math> of [[Distribution (mathematics)#Tempered distribution|tempered distributions]]. Analogous to the case for finite-dimensional random vectors, a probability law on the infinite-dimensional space <math>\mathcal S'(\mathbb R)</math> can be defined via its characteristic function (existence and uniqueness are guaranteed by an extension of the Bochner–Minlos theorem, which goes under the name Bochner–Minlos–Sazanov theorem); analogously to the case of the multivariate normal distribution <math>X \sim \mathcal N_n (\mu , \Sigma )</math>, which has characteristic function : <math>\forall k \in \mathbb R^n: \quad \mathrm{E}(\mathrm e^{\mathrm{i} \langle k, X \rangle }) = \mathrm e^{\mathrm i \langle k, \mu \rangle - \frac 1 2 \langle k, \Sigma k \rangle } ,</math> the white noise <math>w : \Omega \to \mathcal S'(\mathbb R)</math> must satisfy : <math>\forall \varphi \in \mathcal S (\mathbb R) : \quad \mathrm{E}(\mathrm e^{\mathrm{i} \langle w, \varphi \rangle }) = \mathrm e^{- \frac 1 2 \| \varphi \|_2^2},</math> where <math>\langle w, \varphi \rangle</math> is the natural pairing of the tempered distribution <math>w(\omega)</math> with the Schwartz function <math>\varphi</math>, taken scenariowise for <math>\omega \in \Omega</math>, and <math>\| \varphi \|_2^2 = \int_{\mathbb R} \vert \varphi (x) \vert^2\,\mathrm d x </math>. ==Mathematical applications== ===Time series analysis and regression=== In [[statistics]] and [[econometrics]] one often assumes that an observed series of data values is the sum of the values generated by a [[deterministic]] [[linear model|linear process]], depending on certain [[Dependent and independent variables|independent (explanatory) variables]], and on a series of random noise values. Then [[regression analysis]] is used to infer the parameters of the model process from the observed data, e.g. by [[ordinary least squares]], and to [[hypothesis testing|test the null hypothesis]] that each of the parameters is zero against the alternative hypothesis that it is non-zero. Hypothesis testing typically assumes that the noise values are mutually uncorrelated with zero mean and have the same Gaussian probability distribution{{snd}}in other words, that the noise is Gaussian white (not just white). If there is non-zero correlation between the noise values underlying different observations then the estimated model parameters are still [[bias of an estimator|unbiased]], but estimates of their uncertainties (such as [[confidence interval]]s) will be biased (not accurate on average). This is also true if the noise is [[heteroskedastic]]{{snd}}that is, if it has different variances for different data points. Alternatively, in the subset of regression analysis known as [[time series analysis]] there are often no explanatory variables other than the past values of the variable being modeled (the [[dependent variable]]). In this case the noise process is often modeled as a [[Moving average model|moving average]] process, in which the current value of the dependent variable depends on current and past values of a sequential white noise process. ===Random vector transformations=== These two ideas are crucial in applications such as [[channel estimation]] and [[Mixing console#Channel equalization|channel equalization]] in [[telecommunications|communications]] and [[sound reproduction|audio]]. These concepts are also used in [[data compression]]. <!-- This does not seem to be incorrect but seems to be original research, sort of. Needs to be trimmed to the bare essentials. --> In particular, by a suitable linear transformation (a [[coloring transformation]]), a white random vector can be used to produce a non-white random vector (that is, a list of random variables) whose elements have a prescribed [[covariance matrix]]. Conversely, a random vector with known covariance matrix can be transformed into a white random vector by a suitable [[whitening transformation]]. ==Generation== White noise may be generated digitally with a [[digital signal processor]], [[microprocessor]], or [[microcontroller]]. Generating white noise typically entails feeding an appropriate stream of random numbers to a [[digital-to-analog converter]]. The quality of the white noise will depend on the quality of the algorithm used.<ref>{{cite web |url=https://qualityassignmenthelp.com/wp-content/uploads/2017/03/Gaussian-noise_How-to-Generate-.pdf |title=How to Generate White Gaussian Noise |author=Matt Donadio |access-date=2012-09-19 |archive-date=2021-02-24 |archive-url=https://web.archive.org/web/20210224163707/https://qualityassignmenthelp.com/wp-content/uploads/2017/03/Gaussian-noise_How-to-Generate-.pdf }}</ref> ==Informal use== The term is sometimes used as a [[colloquialism]] to describe a backdrop of ambient sound, creating an indistinct or seamless commotion. Following are some examples: *Chatter from multiple conversations within the acoustics of a confined space. *The [[pleonastic]] [[jargon]] used by politicians to mask a point that they don't want noticed.<ref>{{citation |url=https://www.merriam-webster.com/dictionary/white%20noise |title=white noise |publisher=Merriam-Webster |access-date=2022-05-06}}</ref> *[[Music]] that is disagreeable, harsh, dissonant or [[Consonance and dissonance|discordant]] with no [[melody]]. The term can also be used metaphorically, as in the novel ''[[White Noise (novel)|White Noise]]'' (1985) by [[Don DeLillo]] which explores the symptoms of [[Modernity#Cultural and philosophical|modern culture]] that came together so as to make it difficult for an individual to actualize their ideas and personality. ==See also== {{columns-list|colwidth=30em| *[[Bochner–Minlos theorem]] *[[Brownian noise]] *[[Dirac delta function]] *[[Independent component analysis]] *[[MyNoise]] *[[Noise (electronics)]] *[[Noise (video)]] *[[Olfactory white]] *[[Pink noise]] *[[Principal component analysis]] *[[Sound masking]] }} ==References== {{reflist}} ==External links== {{Commons category|White noise}} {{Noise}} {{Stochastic processes}} {{DEFAULTSORT:White Noise}} [[Category:Noise (electronics)]] [[Category:Statistical signal processing]] [[Category:Data compression]] [[Category:Sound]] [[Category:Acoustics]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:Ambox
(
edit
)
Template:Citation
(
edit
)
Template:Citation needed
(
edit
)
Template:Cite book
(
edit
)
Template:Cite journal
(
edit
)
Template:Cite news
(
edit
)
Template:Cite web
(
edit
)
Template:Colors of noise
(
edit
)
Template:Columns-list
(
edit
)
Template:Commons category
(
edit
)
Template:IPA
(
edit
)
Template:More citations needed section
(
edit
)
Template:Noise
(
edit
)
Template:Other uses
(
edit
)
Template:Reflist
(
edit
)
Template:Rp
(
edit
)
Template:Short description
(
edit
)
Template:Sister project
(
edit
)
Template:Snd
(
edit
)
Template:Stochastic processes
(
edit
)
Template:Unreferenced
(
edit
)
Template:Unreferenced section
(
edit
)
Template:Vanchor
(
edit
)
Template:Webarchive
(
edit
)
Template:What
(
edit
)