Hubbry Logo
Time–frequency analysisTime–frequency analysisMain
Open search
Time–frequency analysis
Community hub
Time–frequency analysis
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Time–frequency analysis
Time–frequency analysis
from Wikipedia

In signal processing, time–frequency analysis comprises those techniques that study a signal in both the time and frequency domains simultaneously, using various time–frequency representations. Rather than viewing a 1-dimensional signal (a function, real or complex-valued, whose domain is the real line) and some transform (another function whose domain is the real line, obtained from the original via some transform), time–frequency analysis studies a two-dimensional signal – a function whose domain is the two-dimensional real plane, obtained from the signal via a time–frequency transform.[1][2]

The mathematical motivation for this study is that functions and their transform representation are tightly connected, and they can be understood better by studying them jointly, as a two-dimensional object, rather than separately. A simple example is that the 4-fold periodicity of the Fourier transform – and the fact that two-fold Fourier transform reverses direction – can be interpreted by considering the Fourier transform as a 90° rotation in the associated time–frequency plane: 4 such rotations yield the identity, and 2 such rotations simply reverse direction (reflection through the origin).

The practical motivation for time–frequency analysis is that classical Fourier analysis assumes that signals are infinite in time or periodic, while many signals in practice are of short duration, and change substantially over their duration. For example, traditional musical instruments do not produce infinite duration sinusoids, but instead begin with an attack, then gradually decay. This is poorly represented by traditional methods, which motivates time–frequency analysis.

One of the most basic forms of time–frequency analysis is the short-time Fourier transform (STFT), but more sophisticated techniques have been developed, notably wavelets and least-squares spectral analysis methods for unevenly spaced data.

Motivation

[edit]

In signal processing, time–frequency analysis[3] is a body of techniques and methods used for characterizing and manipulating signals whose statistics vary in time, such as transient signals.

It is a generalization and refinement of Fourier analysis, for the case when the signal frequency characteristics are varying with time. Since many signals of interest – such as speech, music, images, and medical signals – have changing frequency characteristics, time–frequency analysis has broad scope of applications.

Whereas the technique of the Fourier transform can be extended to obtain the frequency spectrum of any slowly growing locally integrable signal, this approach requires a complete description of the signal's behavior over all time. Indeed, one can think of points in the (spectral) frequency domain as smearing together information from across the entire time domain. While mathematically elegant, such a technique is not appropriate for analyzing a signal with indeterminate future behavior. For instance, one must presuppose some degree of indeterminate future behavior in any telecommunications systems to achieve non-zero entropy (if one already knows what the other person will say one cannot learn anything).

To harness the power of a frequency representation without the need of a complete characterization in the time domain, one first obtains a time–frequency distribution of the signal, which represents the signal in both the time and frequency domains simultaneously. In such a representation the frequency domain will only reflect the behavior of a temporally localized version of the signal. This enables one to talk sensibly about signals whose component frequencies vary in time.

For instance rather than using tempered distributions to globally transform the following function into the frequency domain one could instead use these methods to describe it as a signal with a time varying frequency.

Once such a representation has been generated other techniques in time–frequency analysis may then be applied to the signal in order to extract information from the signal, to separate the signal from noise or interfering signals, etc.

Time–frequency distribution functions

[edit]

Formulations

[edit]

There are several different ways to formulate a valid time–frequency distribution function, resulting in several well-known time–frequency distributions, such as:

More information about the history and the motivation of development of time–frequency distribution can be found in the entry Time–frequency representation.

Ideal TF distribution function

[edit]

A time–frequency distribution function ideally has the following properties:[citation needed]

  1. High resolution in both time and frequency, to make it easier to be analyzed and interpreted.
  2. No cross-term to avoid confusing real components from artifacts or noise.
  3. A list of desirable mathematical properties to ensure such methods benefit real-life application.
  4. Lower computational complexity to ensure the time needed to represent and process a signal on a time–frequency plane allows real-time implementations.

Below is a brief comparison of some selected time–frequency distribution functions.[4]

Clarity Cross-term Good mathematical properties[clarification needed] Computational complexity
Gabor transform Worst No Worst Low
Wigner distribution function Best Yes Best High
Gabor–Wigner distribution function Good Almost eliminated Good High
Cone-shape distribution function Good No (eliminated, in time) Good Medium (if recursively defined)

To analyze the signals well, choosing an appropriate time–frequency distribution function is important. Which time–frequency distribution function should be used depends on the application being considered, as shown by reviewing a list of applications.[5] The high clarity of the Wigner distribution function (WDF) obtained for some signals is due to the auto-correlation function inherent in its formulation; however, the latter also causes the cross-term problem. Therefore, if we want to analyze a single-term signal, using the WDF may be the best approach; if the signal is composed of multiple components, some other methods like the Gabor transform, Gabor-Wigner distribution or Modified B-Distribution functions may be better choices.

As an illustration, magnitudes from non-localized Fourier analysis cannot distinguish the signals:


But time–frequency analysis can.

TF analysis and random processes[6]

[edit]

For a random process x(t), we cannot find the explicit value of x(t).

The value of x(t) is expressed as a probability function.

General random processes

[edit]
  • Auto-covariance function (ACF)
In usual, we suppose that for any t,
(alternative definition of the auto-covariance function)
  • Power spectral density (PSD)

Stationary random processes

[edit]

for any , Therefore, PSD, White noise:

, where is some constant.

  • When x(t) is stationary,

, (invariant with )

, (nonzero only when )

Additive white noise

[edit]
  • For additive white noise (AWN),
  • Filter Design for a signal in additive white noise



: energy of the signal

 : area of the time frequency distribution of the signal

The PSD of the white noise is


Non-stationary random processes

[edit]
  • If varies with and is nonzero when , then is a non-stationary random process.
  • If
    1. 's have zero mean for all 's
    2. 's are mutually independent for all 's and 's
then:
  • if , then

Short-time Fourier transform

[edit]

should be satisfied. Otherwise, for zero-mean random process,

  • Decompose by the AF and the FRFT. Any non-stationary random process can be expressed as a summation of the fractional Fourier transform (or chirp multiplication) of stationary random process.

Applications

[edit]

The following applications need not only the time–frequency distribution functions but also some operations to the signal. The Linear canonical transform (LCT) is really helpful. By LCTs, the shape and location on the time–frequency plane of a signal can be in the arbitrary form that we want it to be. For example, the LCTs can shift the time–frequency distribution to any location, dilate it in the horizontal and vertical direction without changing its area on the plane, shear (or twist) it, and rotate it (Fractional Fourier transform). This powerful operation, LCT, make it more flexible to analyze and apply the time–frequency distributions. The time-frequency analysis have been applied in various applications like, disease detection from biomedical signals and images, vital sign extraction from physiological signals, brain-computer interface from brain signals, machinery fault diagnosis from vibration signals, interference mitigation in spread spectrum communication systems.[7][8]

Instantaneous frequency estimation

[edit]

The definition of instantaneous frequency is the time rate of change of phase, or

where is the instantaneous phase of a signal. We can know the instantaneous frequency from the time–frequency plane directly if the image is clear enough. Because the high clarity is critical, we often use WDF to analyze it.

TF filtering and signal decomposition

[edit]

The goal of filter design is to remove the undesired component of a signal. Conventionally, we can just filter in the time domain or in the frequency domain individually as shown below.

The filtering methods mentioned above can't work well for every signal which may overlap in the time domain or in the frequency domain. By using the time–frequency distribution function, we can filter in the Euclidean time–frequency domain or in the fractional domain by employing the fractional Fourier transform. An example is shown below.

Filter design in time–frequency analysis always deals with signals composed of multiple components, so one cannot use WDF due to cross-term. The Gabor transform, Gabor–Wigner distribution function, or Cohen's class distribution function may be better choices.

The concept of signal decomposition relates to the need to separate one component from the others in a signal; this can be achieved through a filtering operation which require a filter design stage. Such filtering is traditionally done in the time domain or in the frequency domain; however, this may not be possible in the case of non-stationary signals that are multicomponent as such components could overlap in both the time domain and also in the frequency domain; as a consequence, the only possible way to achieve component separation and therefore a signal decomposition is to implement a time–frequency filter.

Sampling theory

[edit]

By the Nyquist–Shannon sampling theorem, we can conclude that the minimum number of sampling points without aliasing is equivalent to the area of the time–frequency distribution of a signal. (This is actually just an approximation, because the TF area of any signal is infinite.) Below is an example before and after we combine the sampling theory with the time–frequency distribution:

It is noticeable that the number of sampling points decreases after we apply the time–frequency distribution.

When we use the WDF, there might be the cross-term problem (also called interference). On the other hand, using Gabor transform causes an improvement in the clarity and readability of the representation, therefore improving its interpretation and application to practical problems.

Consequently, when the signal we tend to sample is composed of single component, we use the WDF; however, if the signal consists of more than one component, using the Gabor transform, Gabor-Wigner distribution function, or other reduced interference TFDs may achieve better results.

The Balian–Low theorem formalizes this, and provides a bound on the minimum number of time–frequency samples needed.

Modulation and multiplexing

[edit]

Conventionally, the operation of modulation and multiplexing concentrates in time or in frequency, separately. By taking advantage of the time–frequency distribution, we can make it more efficient to modulate and multiplex. All we have to do is to fill up the time–frequency plane. We present an example as below.

As illustrated in the upper example, using the WDF is not smart since the serious cross-term problem make it difficult to multiplex and modulate.

Electromagnetic wave propagation

[edit]

We can represent an electromagnetic wave in the form of a 2 by 1 matrix

which is similar to the time–frequency plane. When electromagnetic wave propagates through free-space, the Fresnel diffraction occurs. We can operate with the 2 by 1 matrix

by LCT with parameter matrix

where z is the propagation distance and is the wavelength. When electromagnetic wave pass through a spherical lens or be reflected by a disk, the parameter matrix should be

and

respectively, where ƒ is the focal length of the lens and R is the radius of the disk. These corresponding results can be obtained from

Optics, acoustics, and biomedicine

[edit]

Light is an electromagnetic wave, so time–frequency analysis applies to optics in the same way as for general electromagnetic wave propagation.

Similarly, it is a characteristic of acoustic signals, that their frequency components undergo abrupt variations in time and would hence be not well represented by a single frequency component analysis covering their entire durations.

As acoustic signals are used as speech in communication between the human-sender and -receiver, their undelayedly transmission in technical communication systems is crucial, which makes the use of simpler TFDs, such as the Gabor transform, suitable to analyze these signals in real-time by reducing computational complexity.

If frequency analysis speed is not a limitation, a detailed feature comparison with well defined criteria should be made before selecting a particular TFD. Another approach is to define a signal dependent TFD that is adapted to the data. In biomedicine, one can use time–frequency distribution to analyze the electromyography (EMG), electroencephalography (EEG), electrocardiogram (ECG) or otoacoustic emissions (OAEs).

History

[edit]

Early work in time–frequency analysis can be seen in the Haar wavelets (1909) of Alfréd Haar, though these were not significantly applied to signal processing. More substantial work was undertaken by Dennis Gabor, such as Gabor atoms (1947), an early form of wavelets, and the Gabor transform, a modified short-time Fourier transform. The Wigner–Ville distribution (Ville 1948, in a signal processing context) was another foundational step.

Particularly in the 1930s and 1940s, early time–frequency analysis developed in concert with quantum mechanics (Wigner developed the Wigner–Ville distribution in 1932 in quantum mechanics, and Gabor was influenced by quantum mechanics – see Gabor atom); this is reflected in the shared mathematics of the position-momentum plane and the time–frequency plane – as in the Heisenberg uncertainty principle (quantum mechanics) and the Gabor limit (time–frequency analysis), ultimately both reflecting a symplectic structure.

An early practical motivation for time–frequency analysis was the development of radar – see ambiguity function.

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
Time–frequency analysis is a branch of that develops mathematical tools to represent and analyze non-stationary signals by simultaneously capturing their evolution in both time and domains. This joint representation addresses the limitations of classical , which assumes stationarity and loses temporal localization, enabling the study of signals where content varies over time, such as in speech, seismic waves, and biomedical data. Essential for understanding dynamic processes, it provides insights into instantaneous , distribution, and phase changes. The field's theoretical foundations emerged in the early , with Eugene P. Wigner's 1932 development of the in , later adapted by Jean Ville in 1948 for signal analysis as the Wigner–Ville distribution (WVD), offering high time-frequency resolution despite interference issues. A pivotal advancement came in 1946 when introduced the (STFT), or , which applies a sliding window to the to localize frequencies in time, laying the groundwork for modern time-frequency representations. The 1980s saw the rise of wavelet transforms, pioneered by Alexandre Grossmann and Jean Morlet in 1984, which use scalable, oscillating functions for multi-resolution analysis, excelling in detecting transients and varying scales without fixed time-frequency trade-offs. Key methods in time–frequency analysis encompass linear approaches like the STFT and continuous/discrete wavelet transforms (CWT/DWT), which provide straightforward but resolution-limited views, and quadratic distributions such as the WVD and pseudo-WVD, which achieve superior concentration at the cost of cross-term artifacts. Advanced techniques, including synchrosqueezing and empirical mode decomposition, further refine these by reallocating energy for sharper representations and handling nonlinearity. These tools satisfy properties like marginals (integrating to time or signals) and invertibility, ensuring faithful signal reconstruction. Applications span diverse domains, including where wavelet and Hilbert–Huang transforms detect damage in time-varying vibrations of bridges and buildings; , employing Morlet wavelets for EEG oscillatory patterns in developmental studies; and , using high-resolution methods for seismic attribute analysis. In communications and , STFT and WVD enable efficient modulation recognition and target detection in non-stationary environments. Ongoing research as of 2025 focuses on adaptive, data-driven methods such as neural operator-based frequency transforms and time-frequency extracting transforms (TFET) to mitigate computational demands and enhance real-time processing.

Introduction

Motivation

Non-stationary signals are those whose spectral characteristics, such as content or amplitude, vary with time, in contrast to stationary signals that maintain a constant spectrum over time. Classic examples include signals, where the instantaneous increases or decreases linearly over time, and frequency-modulated waves, such as those used in systems to sweep across a range of frequencies. These signals arise in diverse applications, including seismic exploration, audio processing, and biomedical signal , where the evolving structure carries critical about the underlying . Traditional time-domain analyses, such as , capture temporal correlations but fail to reveal frequency-specific variations, treating the signal as a whole without isolating how spectral components evolve. Similarly, frequency-domain methods like the provide a global by integrating the signal over its entire duration, thereby losing all temporal localization and averaging out time-varying features. For instance, in a non-stationary signal, the assumes a stationary , masking changes in content and rendering it inadequate for detecting transient events or modulations. This limitation is particularly evident in the Heisenberg uncertainty principle, which imposes a fundamental trade-off between time and resolution in signal representations. An illustrative example is a piecewise cosine signal, defined as cos(2π10t)\cos(2\pi \cdot 10 t) for 0t<0.50 \leq t < 0.5 and cos(2π20t)\cos(2\pi \cdot 20 t) for 0.5t<10.5 \leq t < 1, representing a sudden frequency shift from 10 Hz to 20 Hz midway through. The Fourier transform of this signal yields a spectrum with peaks at both 10 Hz and 20 Hz, but it cannot indicate that the lower frequency dominates the first half while the higher one appears only in the second half, thus obscuring the temporal evolution of the frequency content. To address these shortcomings, time-frequency analysis introduces the time-frequency plane, a two-dimensional representation that maps the energy density of the signal as a function of both time and frequency, enabling visualization and quantification of how spectral components localize and vary temporally. This plane provides a joint distribution where the energy at any point (t,f)(t, f) reflects the signal's power concentration, offering a more intuitive and informative depiction of non-stationary behavior compared to separate time or frequency views.

Fundamental Concepts

Time and frequency represent dual domains in signal analysis, where a signal x(t)x(t) in the time domain is transformed into its frequency-domain counterpart X(f)X(f) via the Fourier transform, defined as X(f)=x(t)ej2πftdt.X(f) = \int_{-\infty}^{\infty} x(t) e^{-j 2\pi f t} \, dt. This integral transform decomposes the signal into its constituent frequencies, revealing the spectral content that is obscured in the time domain alone. The inverse Fourier transform recovers the original signal, underscoring the duality: operations in one domain correspond to equivalent operations in the other, such as convolution becoming multiplication. This duality forms the cornerstone of frequency-domain analysis but assumes stationarity, limiting its applicability to signals whose frequency content varies over time. A fundamental limitation arises from the Heisenberg-Gabor uncertainty principle, which quantifies the inherent trade-off in localizing a signal simultaneously in time and frequency. For a signal x(t)x(t), the product of the time duration Δt\Delta t (standard deviation of x(t)2|x(t)|^2) and the frequency bandwidth Δf\Delta f (standard deviation of X(f)2|X(f)|^2) satisfies ΔtΔf14π.\Delta t \cdot \Delta f \geq \frac{1}{4\pi}. Equality holds for Gaussian signals, illustrating the minimal achievable resolution; narrower time localization broadens the frequency spread, and vice versa. This principle, originally from quantum mechanics and adapted to signals by Gabor, imposes physical and mathematical constraints on time-frequency representations, preventing perfect joint resolution. For non-stationary signals, the concept of instantaneous frequency provides a local frequency measure, defined for the analytic signal z(t)=x(t)+jx^(t)z(t) = x(t) + j \hat{x}(t) (where x^(t)\hat{x}(t) is the of x(t)x(t)) as fi(t)=12πddtarg{z(t)}.f_i(t) = \frac{1}{2\pi} \frac{d}{dt} \arg\{z(t)\}. This derivative of the phase captures the signal's frequency evolution, with physical meaning under Bedrosian's conditions for amplitude- and frequency-modulated waveforms. It enables interpretation of or frequency-modulated signals but requires the analytic representation to avoid ambiguities in real signals. Energy conservation ensures that time-frequency representations preserve the signal's total energy, achieved through marginal distributions. Integrating the representation over frequency yields the instantaneous energy density x(t)2|x(t)|^2, while integrating over time yields the power spectrum X(f)2|X(f)|^2; thus, the double integral over the time-frequency plane equals the total signal energy x(t)2dt\int |x(t)|^2 \, dt. This property, inherent to quadratic distributions like those in Cohen's class, maintains physical interpretability by linking local and global energy measures.

Time-Frequency Representations

Linear Representations

Linear representations in time-frequency analysis provide additive decompositions of signals into basis functions localized in both time and frequency domains, offering a straightforward approach without interference artifacts from multiple components. These methods, primarily exemplified by the short-time Fourier transform (STFT), segment the signal using a sliding window and apply Fourier analysis to each segment, yielding a two-dimensional representation that balances temporal and spectral information. Unlike higher-order representations, linear methods maintain computational efficiency through direct implementation via fast Fourier transform algorithms, making them suitable for real-time processing despite inherent resolution trade-offs. The short-time Fourier transform is defined as X(t,f)=x(τ)g(τt)ej2πfτdτ,X(t,f) = \int_{-\infty}^{\infty} x(\tau) g(\tau - t) e^{-j 2\pi f \tau} \, d\tau, where x(τ)x(\tau) is the input signal and g(τ)g(\tau) is a window function centered at time tt, modulating the signal to capture frequency content ff locally. This formulation, introduced by in his foundational work on signal decomposition, enables the analysis of non-stationary signals by localizing the in time through the window. The choice of window gg significantly influences the transform's performance, with common options including rectangular, Hamming, or to control leakage and resolution. A notable variant is the Gabor transform, which employs a Gaussian window g(τ)=eπτ2g(\tau) = e^{-\pi \tau^2} to achieve optimal time-frequency localization, as the Gaussian minimizes the product of time and frequency spreads in accordance with the uncertainty principle. This choice provides the tightest concentration in the time-frequency plane among linear methods, making it particularly effective for signals requiring precise joint resolution, such as in optical and acoustic applications. The Gabor transform thus extends the STFT by prioritizing minimal uncertainty, though it retains the fixed window size limitation. In discrete settings, the STFT is computed over sampled signals, producing the discrete-time representation X(m,n)X(m,n) for time index mm and frequency bin nn, often visualized as a spectrogram X(m,n)2|X(m,n)|^2 to display energy distribution. The spectrogram's quality depends on window selection; for instance, the Hamming window g(k)=0.540.46cos(2πk/(N1))g(k) = 0.54 - 0.46 \cos(2\pi k / (N-1)) for length NN reduces spectral sidelobes compared to a rectangular window, improving frequency resolution for closely spaced components, while a Gaussian window enhances overall time-frequency compactness at the cost of broader mainlobe width in frequency. These effects arise from the window's Fourier transform properties, where narrower time windows yield coarser frequency resolution, and vice versa. Overlap between consecutive windows, typically 50-75%, further refines temporal detail without aliasing, as unified in practical synthesis frameworks. The primary advantages of linear representations like the STFT stem from their linearity, which preserves the superposition principle: the transform of a sum of signals equals the sum of their transforms, facilitating analysis of multicomponent signals without cross-term distortions. This property, inherent to the convolution-based structure, also supports efficient inversion for signal reconstruction via overlap-add methods. Computationally, the STFT leverages the fast Fourier transform for O(NlogN)O(N \log N) complexity per frame, enabling scalability in large datasets. However, a key disadvantage is the fixed resolution dictated by the window size, imposing a constant Heisenberg-like uncertainty that prevents simultaneous high precision in time and frequency, as ΔtΔf1/(4π)\Delta t \cdot \Delta f \geq 1/(4\pi) for optimal windows. This compromise limits adaptability to signals with varying frequency scales, though it ensures artifact-free representations.

Quadratic Representations

Quadratic time-frequency representations (TFRs) form a class of energy distributions that achieve optimal concentration of a signal's energy in the time-frequency plane for linear frequency-modulated components, surpassing the resolution limits of linear TFRs, though they introduce interference artifacts known as cross-terms when analyzing multi-component signals. These representations are inherently bilinear in the signal, leading to their high localization properties but also to the generation of spurious oscillations between auto-components. The instantaneous autocorrelation function, defined as Rx(t,τ)=x(t+τ/2)x(tτ/2)R_x(t, \tau) = x(t + \tau/2) x^*(t - \tau/2), serves as a foundational element for deriving these distributions. The Wigner-Ville distribution (WVD) stands as the canonical quadratic TFR, originally introduced in quantum mechanics and later adapted for signal analysis. It is formulated as Wx(t,f)=Rx(t,τ)ej2πfτdτ,W_x(t,f) = \int_{-\infty}^{\infty} R_x(t,\tau) e^{-j2\pi f \tau} \, d\tau, where the Fourier transform of the instantaneous autocorrelation yields a distribution that satisfies key desirable properties such as marginals and finite energy conservation for mono-component signals. For a linear chirp signal, the WVD exhibits perfect energy concentration along the instantaneous frequency trajectory, demonstrating its superior resolution. To address computational challenges in discrete implementations, the pseudo-Wigner-Ville distribution (PWVD) applies a time-smoothing window to the autocorrelation before Fourier transformation, reducing the integration limits and enabling efficient calculation via fast Fourier transforms. Introduced in the context of discrete-time signal processing, the PWVD takes the form PWx(t,f)=h(τ)Rx(t,τ)ej2πfτdτ,PW_x(t,f) = \int_{-\infty}^{\infty} h(\tau) R_x(t,\tau) e^{-j2\pi f \tau} \, d\tau, where h(τ)h(\tau) is a low-pass window function that trades some time resolution for reduced aliasing and computational complexity in practical applications. Cross-terms in quadratic TFRs arise from the bilinear structure, particularly when the signal comprises multiple components, as the distribution includes products of the form 2Re{xi(t+τ/2)xj(tτ/2)ej2πfτdτ}2 \operatorname{Re} \left\{ \int x_i(t + \tau/2) x_j^*(t - \tau/2) e^{-j2\pi f \tau} \, d\tau \right\} for iji \neq j, manifesting as oscillatory artifacts midway between the auto-components in the time-frequency plane. For instance, in a two-tone signal x(t)=cos(2πf1t)+cos(2πf2t)x(t) = \cos(2\pi f_1 t) + \cos(2\pi f_2 t), the WVD displays high-amplitude cross-terms oscillating at frequency (f1+f2)/2(f_1 + f_2)/2, which can obscure the true signal components despite the sharp auto-term localization. The Choi-Williams distribution mitigates these cross-terms by incorporating an exponential kernel within the Cohen class framework, suppressing interferences while preserving auto-component resolution. Defined as CWx(t,f)=Ax(τ,ν)eστ2ν2ej2π(νtfτ)dτdν,CW_x(t,f) = \int_{-\infty}^{\infty} \int_{-\infty}^{\infty} A_x(\tau, \nu) e^{-\sigma \tau^2 \nu^2} e^{j 2 \pi (\nu t - f \tau)} \, d\tau \, d\nu, where Ax(τ,ν)=x(u+τ/2)x(uτ/2)ej2πνuduA_x(\tau, \nu) = \int_{-\infty}^{\infty} x(u + \tau/2) x^*(u - \tau/2) e^{-j 2 \pi \nu u} \, du is the function and σ>0\sigma > 0 controls the trade-off between cross-term reduction and resolution, this distribution effectively attenuates bilinear cross-products by emphasizing separable contributions in the ambiguity domain. For multi-component signals like superposed chirps, the Choi-Williams approach reduces cross-term amplitudes by orders of magnitude compared to the WVD, as verified in numerical evaluations.

Properties and Formulations

Desirable Properties

Time-frequency distributions (TFDs) are expected to possess several core mathematical properties to ensure they faithfully represent the content of a signal in both time and domains. These include finite , which requires that the double integral of the of the TFD over time and be finite for finite-energy signals, guaranteeing a well-behaved representation. Real-valuedness stipulates that the TFD is real for real-valued signals, with non-negativity being a desirable but not always achievable feature, aiding interpretability as an . Time-shift invariance means that shifting the signal in time shifts the TFD accordingly: if x(t)x(t) is the signal, then the TFD of x(tt0)x(t - t_0) equals the TFD of x(t)x(t) shifted by t0t_0. Similarly, frequency-shift invariance ensures that modulating the signal in frequency shifts the TFD in frequency: the TFD of x(t)ej2πν0tx(t) e^{j 2 \pi \nu_0 t} is the TFD of x(t)x(t) shifted by ν0\nu_0 in . A fundamental set of marginal properties links the TFD to the signal's power distributions. The time marginal is obtained by integrating the TFD over : W(t,f)df=x(t)2,\int_{-\infty}^{\infty} W(t, f) \, df = |x(t)|^2, yielding the instantaneous power of the signal. The frequency marginal is the over time: W(t,f)dt=X(f)2,\int_{-\infty}^{\infty} W(t, f) \, dt = |X(f)|^2, recovering the power , where X(f)X(f) is the of x(t)x(t). These properties ensure consistency with classical one-dimensional analyses. For multi-component signals, composed of multiple distinct frequency-modulated components, a desirable ridge property requires that the TFD concentrates the energy of each component along a distinct in the time-frequency plane, corresponding to its instantaneous frequency , while minimizing interference between components. This allows clear separation and identification of individual signal constituents, essential for applications like signal . However, achieving perfect ridge separation without artifacts is challenging due to inherent trade-offs in TFD design. Invertibility is another key attribute, enabling reconstruction of the original signal from the TFD. For the z(t)z(t), it can be recovered as z(t)=0Wz(t,f)ej2πftdfz(t) = \int_{0}^{\infty} W_z(t, f) e^{j 2 \pi f t} \, df (up to normalization constants depending on convention), with the real signal obtained via the real part after applying the inverse . This property holds for specific TFDs like the Wigner-Ville distribution when properly defined for the , preserving signal information completely. No single kernel exists that yields an ideal TFD with delta-like concentration for arbitrary signals, perfect marginals, positivity, and absence of cross-terms between components. This impossibility arises from the Heisenberg uncertainty principle, which fundamentally limits simultaneous time and frequency localization, leading to unavoidable cross-terms in high-resolution quadratic distributions or smearing in positive ones. Seminal analyses demonstrate that satisfying all desirable properties simultaneously is mathematically infeasible, necessitating trade-offs via kernel selection.

Kernel-Based Formulations

Kernel-based formulations provide a unified framework for constructing quadratic time-frequency distributions, allowing flexible adjustment of desirable properties through the choice of a kernel function. These methods generalize specific distributions like the Wigner-Ville distribution (WVD) and spectrogram by incorporating a kernel that modulates the ambiguity domain representation of the signal, thereby enabling control over resolution, interference suppression, and other characteristics. Cohen's class represents the foundational kernel-based approach, encompassing a broad family of quadratic time-frequency distributions defined by the general form: Cx(t,f)=Ax(τ,θ)ϕ(τ,θ)ej2π(fτtθ)dτdθ,C_x(t,f) = \iint_{-\infty}^{\infty} A_x(\tau,\theta) \phi(\tau,\theta) e^{-j2\pi (f \tau - t \theta)} \, d\tau \, d\theta, where Ax(τ,θ)A_x(\tau,\theta) is the signal's , defined as Ax(τ,θ)=x(u+τ/2)x(uτ/2)ej2πθuduA_x(\tau,\theta) = \int x(u + \tau/2) x^*(u - \tau/2) e^{-j 2\pi \theta u} du, and ϕ(τ,θ)\phi(\tau,\theta) is the kernel that determines the specific distribution within the class. This formulation ensures under time and shifts, a key property for analyzing non-stationary signals. Specific distributions arise from particular kernel choices. For the Wigner-Ville distribution, the kernel is ϕ(τ,θ)=1\phi(\tau,\theta) = 1, yielding high time-frequency resolution but suffering from prominent cross-term interference for multicomponent signals. In contrast, the spectrogram corresponds to the kernel ϕ(τ,θ)=G(θ)2\phi(\tau,\theta) = |G(\theta)|^2, where G(θ)G(\theta) is a low-pass window function in the frequency domain; this choice applies smoothing to reduce cross-terms at the expense of resolution. The smoothed pseudo-Wigner-Ville distribution (SPWVD) extends the pseudo-WVD by incorporating separable smoothing in both time and frequency directions, with a kernel of the form ϕ(τ,θ)=h(τ)G(θ)2\phi(\tau,\theta) = h(\tau) |G(\theta)|^2, where h(τ)h(\tau) is a time-smoothing . This mitigates cross-terms while preserving much of the WVD's localization for linear frequency-modulated components. Further extensions include the affine class, which generalizes Cohen's class to achieve under affine transformations (time shifts and scalings). Kernels in the affine class satisfy conditions such as K^(f1,f2;t,f)=(f0/f)K^(f1f0/f,f2f0/f;0,f0)ej2πt(f1f2)\hat{K}(f_1, f_2; t, f) = (f_0 / f) \hat{K}(f_1 f_0 / f, f_2 f_0 / f; 0, f_0) e^{j 2\pi t (f_1 - f_2)} in the formulation; this framework is particularly suited for signals with varying local bandwidths. Kernel design in these formulations involves inherent trade-offs: kernels that preserve auto-terms (e.g., sharp ϕ1\phi \approx 1) enhance concentration and resolution but amplify cross-term artifacts, while smoothing kernels (e.g., low-pass ϕ\phi) suppress interference through averaging, often blurring true signal components and reducing overall sharpness. Optimal kernel selection depends on the signal's characteristics and application priorities, such as interference rejection in noisy environments versus precise instantaneous frequency tracking.

Analysis of Random Processes

Stationary Processes

In time-frequency analysis, stationary random processes are characterized by time-invariant statistical properties, such as constant and that depend only on the time lag. For such processes, time-frequency representations (TFRs) simplify significantly, as the energy distribution becomes independent of time and reduces to the power (PSD). This invariance allows TFRs to focus on frequency-domain characteristics without temporal variations, making them particularly suitable for analyzing signals like broadband noise or periodic stationary waveforms. A key result for stationary processes is that the expected value of a TFR, denoted E[W(t,f)]E[W(t,f)], equals the PSD Sx(f)S_x(f), which is independent of time tt. This holds for quadratic TFRs within Cohen's class, such as the Wigner-Ville distribution (WVD), where the ensemble average over realizations yields the of the function. Consequently, time averaging of the TFR across tt recovers the PSD, providing a robust estimate for stationary signals. The , a linear TFR based on the , exemplifies this behavior for stationary noise, exhibiting uniform power distribution across time while reflecting the frequency content of the PSD. For stationary white noise, a classic example, the PSD is flat (constant power across all frequencies), resulting in a TFR that appears as a uniform horizontal strip in the time-frequency plane, with no concentration in specific frequencies or times. This flat spectrum underscores the process's lack of structure, aiding in estimation and . This framework extends the Wiener-Khinchin theorem, which relates the PSD to the of the for stationary processes, to the joint time-frequency domain. In TFRs, the theorem implies that the time-marginal of the expected TFR integrates to the autocorrelation's transform, preserving while distributing power solely over . This connection facilitates spectral estimation from finite realizations of stationary processes.

Non-Stationary Processes

Non-stationary random processes exhibit statistical properties that vary with time, necessitating time-frequency representations that capture local spectral evolution, unlike the time-invariant power spectral density used for stationary processes. In time-frequency analysis, the time-dependent power spectral density for such processes is defined as Sx(t,f)=E[X(t,f)2]S_x(t,f) = E[|X(t,f)|^2], where X(t,f)X(t,f) denotes a time-frequency transform of the process x(t)x(t), and E[]E[\cdot] is the expectation operator. This formulation allows for the characterization of spectral changes over time, particularly in classes like cyclostationary processes, where statistical periodicities lead to time-varying spectra. The (STFT) provides a practical means to estimate the local of non-stationary processes through E[STFTx(t,f)2]E[|{\rm STFT}_x(t,f)|^2], which approximates the time-dependent power under suitable conditions. This expectation yields a smoothed time- distribution that reveals how the process's content evolves, with the length trading off time and resolution. For cyclostationary signals, such as amplitude-modulated noise, the STFT-based estimate highlights periodic components, enabling detection of hidden periodicities in the non-stationarity. The Wigner-Ville distribution extends to non-stationary random processes via the time-localized autocorrelation function Rx(t,τ)=E[x(t+τ/2)x(tτ/2)]R_x(t,\tau) = E[x(t + \tau/2) x^*(t - \tau/2)], from which the Wigner-Ville spectrum is obtained as Wx(t,f)=Rx(t,τ)ej2πfτdτW_x(t,f) = \int R_x(t,\tau) e^{-j 2\pi f \tau} \, d\tau. This quadratic representation offers high resolution but introduces variance due to cross-terms and the random nature of the process, often requiring kernels to mitigate errors in finite realizations. Unlike stationary cases, the time dependence in Rx(t,τ)R_x(t,\tau) accounts for evolving second-order statistics, providing a bilinear of the power spectrum. Practical examples illustrate these concepts, such as Doppler-shifted noise, where a stationary noise process undergoes a linear shift due to relative motion, resulting in a time-swept captured by the time-dependent PSD. Similarly, modulated random processes, like frequency-modulated , exhibit chirp-like spectral trajectories, where STFT or Wigner-Ville estimates reveal the instantaneous variation and associated variance in spectral localization. These cases underscore the challenges of non-uniform spectral spreading and the need for adaptive estimation to handle time-varying variance.

Applications

Signal Processing Techniques

Time-frequency analysis provides powerful tools for manipulating signals by exploiting their joint time and frequency characteristics, enabling techniques such as , filtering, and that are particularly effective for non-stationary signals. These methods operate directly in the time-frequency (TF) plane, where signal components manifest as localized energy concentrations or ridges, allowing for targeted processing that preserves temporal and spectral details better than traditional time- or frequency-only approaches. Instantaneous frequency (IF) estimation is a fundamental technique in TF analysis, often achieved by extracting ridges from time-frequency representations (TFRs), which correspond to the loci of maximum energy along the signal's frequency trajectory. The IF for the i-th component at time t is typically estimated as f^i(t)=argmaxfW(t,f)\hat{f}_i(t) = \arg\max_f W(t,f), where W(t,f)W(t,f) denotes the TFR, such as the Wigner-Ville distribution (WVD) or (STFT) spectrogram. This ridge extraction enhances resolution for multi-component signals by identifying prominent curves in the TF plane, mitigating cross-term interference in quadratic TFRs through adaptive algorithms that track energy maxima while suppressing noise-induced artifacts. For instance, in signals with intersecting components, iterative refinement of ridge paths ensures accurate IF recovery, as demonstrated in applications to oscillatory signals where traditional phase derivative methods fail due to amplitude variations. TF filtering involves applying masks or operators in the TF plane to isolate or suppress specific regions, facilitating removal and component separation. By defining a binary or soft M(t,f)M(t,f) that retains signal-dominant areas while attenuating or interferers, the filtered signal is reconstructed via inverse TF transform, such as x^(t)=M(t,f)Wx(t,f)ej2πftdf\hat{x}(t) = \int M(t,f) W_x(t,f) e^{j2\pi f t} df, preserving phase information critical for . This approach excels in non-stationary environments, where time-varying Wiener filters are designed using TF covariance structures to optimize mean-square error, outperforming fixed filters by adapting to local signal statistics. For component separation, masks target distinct ridges, enabling extraction of individual modes from overlapping TF supports without requiring prior knowledge of signal structure. Signal decomposition in the TF domain employs sparse approximation techniques like and basis pursuit to represent signals as sums of TF atoms, such as Gabor functions, selected iteratively to minimize residual energy. In , the algorithm greedily selects the dictionary element maximizing inner product correlation at each step, yielding a decomposition x(t)n=1Ncngγn(t)x(t) \approx \sum_{n=1}^N c_n g_{\gamma_n}(t), where gγng_{\gamma_n} are time-frequency atoms and γn\gamma_n parameterize their location, scale, and chirp rate. This method is particularly suited to TF dictionaries, capturing localized signal structures like transients or s with high efficiency, as shown in pattern extraction from noisy data where it achieves sparse representations with fewer terms than orthogonal bases. Basis pursuit extends this by solving an 1\ell_1-norm minimization problem for exact recovery under incoherence conditions, enhancing robustness in underdetermined TF expansions. Handling in TF representations often relies on thresholding the to reveal underlying signal s, suppressing uniform spread across frequencies. Hard thresholding discards coefficients below a -dependent level λ\lambda, while soft thresholding shrinks them toward zero, with λ\lambda typically set as σ2logN\sigma \sqrt{2 \log N}
Add your contribution
Related Hubs
User Avatar
No comments yet.