Hubbry Logo
Additive white Gaussian noiseAdditive white Gaussian noiseMain
Open search
Additive white Gaussian noise
Community hub
Additive white Gaussian noise
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Additive white Gaussian noise
Additive white Gaussian noise
from Wikipedia

Additive white Gaussian noise (AWGN) is a basic noise model used in information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics:

  • Additive because it is added to any noise that might be intrinsic to the information system.
  • White refers to the idea that it has uniform power spectral density across the frequency band for the information system. It is an analogy to the color white which may be realized by uniform emissions at all frequencies in the visible spectrum.
  • Gaussian because it has a normal distribution in the time domain with an average time domain value of zero (Gaussian process).

Wideband noise comes from many natural noise sources, such as the thermal vibrations of atoms in conductors (referred to as thermal noise or Johnson–Nyquist noise), shot noise, black-body radiation from the earth and other warm objects, and from celestial sources such as the Sun. The central limit theorem of probability theory indicates that the summation of many random processes will tend to have distribution called Gaussian or Normal.

AWGN is often used as a channel model in which the only impairment to communication is a linear addition of wideband or white noise with a constant spectral density (expressed as watts per hertz of bandwidth) and a Gaussian distribution of amplitude. The model does not account for fading, frequency selectivity, interference, nonlinearity or dispersion. However, it produces simple and tractable mathematical models which are useful for gaining insight into the underlying behavior of a system before these other phenomena are considered.

The AWGN channel is a good model for many satellite and deep space communication links. It is not a good model for most terrestrial links because of multipath, terrain blocking, interference, etc. However, for terrestrial path modeling, AWGN is commonly used to simulate background noise of the channel under study, in addition to multipath, terrain blocking, interference, ground clutter and self interference that modern radio systems encounter in terrestrial operation.

Channel capacity

[edit]

The AWGN channel is represented by a series of outputs at discrete-time event index . is the sum of the input and noise, , where is independent and identically distributed and drawn from a zero-mean normal distribution with variance (the noise). The are further assumed to not be correlated with the .

The capacity of the channel is infinite unless the noise is nonzero, and the are sufficiently constrained. The most common constraint on the input is the so-called "power" constraint, requiring that for a codeword transmitted through the channel, we have:

where represents the maximum channel power. Therefore, the channel capacity for the power-constrained channel is given by:[clarification needed]

where is the distribution of . Expand , writing it in terms of the differential entropy:

But and are independent, therefore:

Evaluating the differential entropy of a Gaussian gives:

Because and are independent and their sum gives :

From this bound, we infer from a property of the differential entropy that

Therefore, the channel capacity is given by the highest achievable bound on the mutual information:

Where is maximized when:

Thus the channel capacity for the AWGN channel is given by:

Channel capacity and sphere packing

[edit]

Suppose that we are sending messages through the channel with index ranging from to , the number of distinct possible messages. If we encode the messages to bits, then we define the rate as:

A rate is said to be achievable if there is a sequence of codes so that the maximum probability of error tends to zero as approaches infinity. The capacity is the highest achievable rate.

Consider a codeword of length sent through the AWGN channel with noise level . When received, the codeword vector variance is now , and its mean is the codeword sent. The vector is very likely to be contained in a sphere of radius around the codeword sent. If we decode by mapping every message received onto the codeword at the center of this sphere, then an error occurs only when the received vector is outside of this sphere, which is very unlikely.

Each codeword vector has an associated sphere of received codeword vectors which are decoded to it and each such sphere must map uniquely onto a codeword. Because these spheres therefore must not intersect, we are faced with the problem of sphere packing. How many distinct codewords can we pack into our -bit codeword vector? The received vectors have a maximum energy of and therefore must occupy a sphere of radius . Each codeword sphere has radius . The volume of an n-dimensional sphere is directly proportional to , so the maximum number of uniquely decodeable spheres that can be packed into our sphere with transmission power P is:

By this argument, the rate R can be no more than .

Achievability

[edit]

In this section, we show achievability of the upper bound on the rate from the last section.

A codebook, known to both encoder and decoder, is generated by selecting codewords of length n, i.i.d. Gaussian with variance and mean zero. For large n, the empirical variance of the codebook will be very close to the variance of its distribution, thereby avoiding violation of the power constraint probabilistically.

Received messages are decoded to a message in the codebook which is uniquely jointly typical. If there is no such message or if the power constraint is violated, a decoding error is declared.

Let denote the codeword for message , while is, as before the received vector. Define the following three events:

  1. Event :the power of the received message is larger than .
  2. Event : the transmitted and received codewords are not jointly typical.
  3. Event : is in , the typical set where , which is to say that the incorrect codeword is jointly typical with the received vector.

An error therefore occurs if , or any of the occur. By the law of large numbers, goes to zero as n approaches infinity, and by the joint Asymptotic Equipartition Property the same applies to . Therefore, for a sufficiently large , both and are each less than . Since and are independent for , we have that and are also independent. Therefore, by the joint AEP, . This allows us to calculate , the probability of error as follows:

Therefore, as n approaches infinity, goes to zero and . Therefore, there is a code of rate R arbitrarily close to the capacity derived earlier.

Coding theorem converse

[edit]

Here we show that rates above the capacity are not achievable.

Suppose that the power constraint is satisfied for a codebook, and further suppose that the messages follow a uniform distribution. Let be the input messages and the output messages. Thus the information flows as:

Making use of Fano's inequality gives:

where as

Let be the encoded message of codeword index i. Then:

Let be the average power of the codeword of index i:

where the sum is over all input messages . and are independent, thus the expectation of the power of is, for noise level :

And, if is normally distributed, we have that

Therefore,

We may apply Jensen's equality to , a concave (downward) function of x, to get:

Because each codeword individually satisfies the power constraint, the average also satisfies the power constraint. Therefore,

which we may apply to simplify the inequality above and get:

Therefore, it must be that . Therefore, R must be less than a value arbitrarily close to the capacity derived earlier, as .

Effects in time domain

[edit]
Zero crossings of a noisy cosine

In serial data communications, the AWGN mathematical model is used to model the timing error caused by random jitter (RJ).

The graph to the right shows an example of timing errors associated with AWGN. The variable Δt represents the uncertainty in the zero crossing. As the amplitude of the AWGN is increased, the signal-to-noise ratio decreases. This results in increased uncertainty Δt.[1]

When affected by AWGN, the average number of either positive-going or negative-going zero crossings per second at the output of a narrow bandpass filter when the input is a sine wave is

where

ƒ0 = the center frequency of the filter,
B = the filter bandwidth,
SNR = the signal-to-noise power ratio in linear terms.

Effects in phasor domain

[edit]
AWGN contributions in the phasor domain

In modern communication systems, bandlimited AWGN cannot be ignored. When modeling bandlimited AWGN in the phasor domain, statistical analysis reveals that the amplitudes of the real and imaginary contributions are independent variables which follow the Gaussian distribution model. When combined, the resultant phasor's magnitude is a Rayleigh-distributed random variable, while the phase is uniformly distributed from 0 to 2π.

The graph to the right shows an example of how bandlimited AWGN can affect a coherent carrier signal. The instantaneous response of the noise vector cannot be precisely predicted, however, its time-averaged response can be statistically predicted. As shown in the graph, we confidently predict that the noise phasor will reside about 38% of the time inside the 1σ circle, about 86% of the time inside the 2σ circle, and about 98% of the time inside the 3σ circle.[1]

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Additive white Gaussian noise (AWGN) is a for in communication systems, representing random disturbances added to a transmitted signal, where the exhibits a Gaussian (normal) with zero and a flat power spectral density across all frequencies, making it uncorrelated in time and independent of the signal. This model approximates real-world phenomena such as thermal generated by the random motion of electrons in conductors or resistors at . The "additive" aspect signifies that the noise is linearly superimposed onto the clean signal, resulting in a received signal y(t)=x(t)+n(t)y(t) = x(t) + n(t), where x(t)x(t) is the transmitted signal and n(t)n(t) is the noise process. "White" describes the noise's power , which is constant (equal power per unit ) over the entire , implying infinite bandwidth and delta-function in the . "Gaussian" refers to the distribution of the noise samples, which follows a N(0,σ2)\mathcal{N}(0, \sigma^2), with variance σ2\sigma^2 determining the . These properties ensure that AWGN samples are statistically independent, simplifying mathematical analysis in both continuous-time and discrete-time settings. AWGN serves as the foundational channel model in for evaluating the performance limits of communication systems, particularly in deriving fundamental bounds like the Shannon capacity. For a bandlimited AWGN channel with bandwidth WW Hz and SNR=P/(N0W)\text{SNR} = P / (N_0 W) (where PP is the average signal power and N0N_0 is the ), the capacity is C=Wlog2(1+SNR)C = W \log_2 (1 + \text{SNR}) bits per second, representing the maximum reliable data rate. This theorem, established by in 1948, underscores AWGN's role in quantifying the effects of noise on reliable transmission and guiding the design of modulation, coding, and error-correcting schemes in digital communications. Beyond communications, AWGN models are applied in , systems, and to simulate environmental perturbations and assess system robustness.

Fundamentals

Definition and Characteristics

Additive white Gaussian (AWGN) is a canonical noise model employed in and to represent the random disturbances that degrade in communication systems. It combines three essential properties: additivity, whiteness, and Gaussian distribution, making it a versatile approximation for various physical noise sources. This model assumes the noise is superimposed on the signal without altering its form, possesses uniform power across frequencies, and exhibits a bell-shaped typical of many natural random processes. The additive nature of AWGN signifies that the noise is linearly added to the transmitted signal, resulting in a received signal that is the of the original signal and the noise component, with the noise being statistically independent of the signal itself. This ensures that the noise does not depend on the signal's content or , allowing for simplified in system design. In practical terms, this models scenarios where external or internal disturbances overlay the desired information without multiplicative effects. Whiteness describes the noise's power spectral density as constant over all frequencies of interest, implying equal energy contribution from each frequency band and uncorrelated samples in discrete-time representations. The Gaussian aspect means the noise values follow a normal probability distribution with zero , reflecting the cumulative effect of numerous independent microscopic fluctuations as per the . Additionally, AWGN is a , where its , variance, and correlation structure remain invariant over time, facilitating consistent statistical treatment. AWGN originates as an idealized representation of thermal in electronic circuits, particularly the Johnson-Nyquist arising from the random thermal agitation of charge carriers in conductors at equilibrium temperature. First observed experimentally and theoretically derived in the late , this provides a foundational physical basis for the AWGN model, enabling its widespread use to approximate real-world impairments like those in amplifiers and transmission lines.

Historical Development

The foundations of additive white Gaussian noise (AWGN) were laid in the late through experimental and theoretical work on in electrical conductors. In 1928, John B. Johnson published experimental findings demonstrating that arises from the random agitation of electrons in resistors, with a power proportional to and resistance. That same year, provided a rigorous theoretical derivation, confirming the noise's white spectrum and Gaussian distribution—stemming from the statistical superposition of numerous independent motions via the . This Johnson-Nyquist theorem established as inherently additive, Gaussian, and spectrally flat, providing the physical basis for AWGN models in subsequent . The integration of AWGN into information theory occurred in 1948 with Claude Shannon's foundational paper, "A Mathematical Theory of Communication," which formalized noisy channels using AWGN to quantify reliable data transmission limits. Shannon's noisy channel coding theorem specifically targeted the AWGN channel, proving that arbitrarily low error rates are achievable at rates below the channel capacity, fundamentally shaping modern communications. Post-World War II, the AWGN model saw rapid adoption in the 1950s for and systems, where it served as a standard benchmark for noise impairment analysis and system design amid growing electronic warfare and long-distance voice transmission needs. By the 1960s, advanced under this framework, with Peter Elias and Amiel Feinstein extending Shannon's ideas to practical error-correcting codes; their 1955 and 1954 works, respectively, derived bounds on error probabilities for discrete noisy channels, enabling codes that approach theoretical limits. As of November 2025, AWGN remains central to digital communications standards like 5G New Radio (NR) and (IEEE 802.11), where it underpins link-level simulations and performance evaluations in specifications. However, emerging technologies such as quantum communications are driving extensions to non-Gaussian noise models to account for photon loss, decoherence, and non-classical effects beyond traditional thermal assumptions.

Statistical and Spectral Properties

Gaussian Distribution Aspects

The Gaussian component of additive white Gaussian noise (AWGN) is characterized by its (PDF), which for a single noise sample nn follows a nN(0,σ2)n \sim \mathcal{N}(0, \sigma^2). The PDF is given by f(n)=1σ2πexp(n22σ2),f(n) = \frac{1}{\sigma \sqrt{2\pi}} \exp\left( -\frac{n^2}{2\sigma^2} \right),
Add your contribution
Related Hubs
User Avatar
No comments yet.