Recent from talks
Contribute something
Nothing was collected or created yet.
Multivariate random variable
View on Wikipedia| Part of a series on statistics |
| Probability theory |
|---|
In probability and statistics, a multivariate random variable or random vector is a list or vector of mathematical variables each of whose value is unknown, either because the value has not yet occurred or because there is imperfect knowledge of its value. The individual variables in a random vector are grouped together because they are all part of a single mathematical system — often they represent different properties of an individual statistical unit. For example, while a given person has a specific age, height and weight, the representation of these features of an unspecified person from within a group would be a random vector. Normally each element of a random vector is a real number.
Random vectors are often used as the underlying implementation of various types of aggregate random variables, e.g. a random matrix, random tree, random sequence, stochastic process, etc.
Formally, a multivariate random variable is a column vector (or its transpose, which is a row vector) whose components are random variables on the probability space , where is the sample space, is the sigma-algebra (the collection of all events), and is the probability measure (a function returning each event's probability).
Probability distribution
[edit]Every random vector gives rise to a probability measure on with the Borel algebra as the underlying sigma-algebra. This measure is also known as the joint probability distribution, the joint distribution, or the multivariate distribution of the random vector.
The distributions of each of the component random variables are called marginal distributions. The conditional probability distribution of given is the probability distribution of when is known to be a particular value.
The cumulative distribution function of a random vector is defined as[1]: p.15
| Eq.1 |
where .
Operations on random vectors
[edit]Random vectors can be subjected to the same kinds of algebraic operations as can non-random vectors: addition, subtraction, multiplication by a scalar, and the taking of inner products.
Affine transformations
[edit]Similarly, a new random vector can be defined by applying an affine transformation to a random vector :
- , where is an matrix and is an column vector.
If is an invertible matrix and has a probability density function , then the probability density of is
- .
Invertible mappings
[edit]More generally we can study invertible mappings of random vectors.[2]: p.284–285
Let be a one-to-one mapping from an open subset of onto a subset of , let have continuous partial derivatives in and let the Jacobian determinant of be zero at no point of . Assume that the real random vector has a probability density function and satisfies . Then the random vector is of probability density
where denotes the indicator function and set denotes support of .
Expected value
[edit]The expected value or mean of a random vector is a fixed vector whose elements are the expected values of the respective random variables.[3]: p.333
| Eq.2 |
Covariance and cross-covariance
[edit]Definitions
[edit]The covariance matrix (also called second central moment or variance-covariance matrix) of an random vector is an matrix whose (i,j)th element is the covariance between the i th and the j th random variables. The covariance matrix is the expected value, element by element, of the matrix computed as , where the superscript T refers to the transpose of the indicated vector:[2]: p. 464 [3]: p.335
| Eq.3 |
By extension, the cross-covariance matrix between two random vectors and ( having elements and having elements) is the matrix[3]: p.336
| Eq.4 |
where again the matrix expectation is taken element-by-element in the matrix. Here the (i,j)th element is the covariance between the i th element of and the j th element of .
Properties
[edit]The covariance matrix is a symmetric matrix, i.e.[2]: p. 466
- .
The covariance matrix is a positive semidefinite matrix, i.e.[2]: p. 465
- .
The cross-covariance matrix is simply the transpose of the matrix , i.e.
- .
Uncorrelatedness
[edit]Two random vectors and are called uncorrelated if
- .
They are uncorrelated if and only if their cross-covariance matrix is zero.[3]: p.337
Correlation and cross-correlation
[edit]Definitions
[edit]The correlation matrix (also called second moment) of an random vector is an matrix whose (i,j)th element is the correlation between the i th and the j th random variables. The correlation matrix is the expected value, element by element, of the matrix computed as , where the superscript T refers to the transpose of the indicated vector:[4]: p.190 [3]: p.334
| Eq.5 |
By extension, the cross-correlation matrix between two random vectors and ( having elements and having elements) is the matrix
| Eq.6 |
Properties
[edit]The correlation matrix is related to the covariance matrix by
- .
Similarly for the cross-correlation matrix and the cross-covariance matrix:
Orthogonality
[edit]Two random vectors of the same size and are called orthogonal if
- .
Independence
[edit]Two random vectors and are called independent if for all and
where and denote the cumulative distribution functions of and and denotes their joint cumulative distribution function. Independence of and is often denoted by . Written component-wise, and are called independent if for all
- .
Characteristic function
[edit]The characteristic function of a random vector with components is a function that maps every vector to a complex number. It is defined by[2]: p. 468
- .
Further properties
[edit]Expectation of a quadratic form
[edit]One can take the expectation of a quadratic form in the random vector as follows:[5]: p.170–171
where is the covariance matrix of and refers to the trace of a matrix — that is, to the sum of the elements on its main diagonal (from upper left to lower right). Since the quadratic form is a scalar, so is its expectation.
Proof: Let be an random vector with and and let be an non-stochastic matrix.
Then based on the formula for the covariance, if we denote and , we see that:
Hence
which leaves us to show that
This is true based on the fact that one can cyclically permute matrices when taking a trace without changing the end result (e.g.: ).
We see that
And since
is a scalar, then
trivially. Using the permutation we get:
and by plugging this into the original formula we get:
Expectation of the product of two different quadratic forms
[edit]One can take the expectation of the product of two different quadratic forms in a zero-mean Gaussian random vector as follows:[5]: pp. 162–176
where again is the covariance matrix of . Again, since both quadratic forms are scalars and hence their product is a scalar, the expectation of their product is also a scalar.
Applications
[edit]Portfolio theory
[edit]In portfolio theory in finance, an objective often is to choose a portfolio of risky assets such that the distribution of the random portfolio return has desirable properties. For example, one might want to choose the portfolio return having the lowest variance for a given expected value. Here the random vector is the vector of random returns on the individual assets, and the portfolio return p (a random scalar) is the inner product of the vector of random returns with a vector w of portfolio weights — the fractions of the portfolio placed in the respective assets. Since p = wT, the expected value of the portfolio return is wTE() and the variance of the portfolio return can be shown to be wTCw, where C is the covariance matrix of .
Regression theory
[edit]In linear regression theory, we have data on n observations on a dependent variable y and n observations on each of k independent variables xj. The observations on the dependent variable are stacked into a column vector y; the observations on each independent variable are also stacked into column vectors, and these latter column vectors are combined into a design matrix X (not denoting a random vector in this context) of observations on the independent variables. Then the following regression equation is postulated as a description of the process that generated the data:
where β is a postulated fixed but unknown vector of k response coefficients, and e is an unknown random vector reflecting random influences on the dependent variable. By some chosen technique such as ordinary least squares, a vector is chosen as an estimate of β, and the estimate of the vector e, denoted , is computed as
Then the statistician must analyze the properties of and , which are viewed as random vectors since a randomly different selection of n cases to observe would have resulted in different values for them.
Vector time series
[edit]The evolution of a k×1 random vector through time can be modelled as a vector autoregression (VAR) as follows:
where the i-periods-back vector observation is called the i-th lag of , c is a k × 1 vector of constants (intercepts), Ai is a time-invariant k × k matrix and is a k × 1 random vector of error terms.
References
[edit]- ^ Gallager, Robert G. (2013). Stochastic Processes Theory for Applications. Cambridge University Press. ISBN 978-1-107-03975-9.
- ^ a b c d e Taboga, Marco (2017). Lectures on Probability Theory and Mathematical Statistics. CreateSpace Independent Publishing Platform. ISBN 978-1981369195.
- ^ a b c d e Gubner, John A. (2006). Probability and Random Processes for Electrical and Computer Engineers. Cambridge University Press. ISBN 978-0-521-86470-1.
- ^ Papoulis, Athanasius (1991). Probability, Random Variables and Stochastic Processes (Third ed.). McGraw-Hill. ISBN 0-07-048477-5.
- ^ a b Kendrick, David (1981). Stochastic Control for Economic Models. McGraw-Hill. ISBN 0-07-033962-7.
Further reading
[edit]- Stark, Henry; Woods, John W. (2012). "Random Vectors". Probability, Statistics, and Random Processes for Engineers (Fourth ed.). Pearson. pp. 295–339. ISBN 978-0-13-231123-6.
Multivariate random variable
View on GrokipediaFundamentals
Definition
In probability theory, a multivariate random variable, also known as a random vector, is formally defined as a measurable function from a probability space to the -dimensional Euclidean space , where and with each being a univariate random variable.[5] This definition ensures that the joint behavior of the components can be analyzed through the induced probability measure on .[5] Unlike a univariate random variable, which captures the probabilistic behavior of a single scalar outcome, a multivariate random variable emphasizes the interdependence and joint distribution among multiple components, allowing for the study of relationships such as correlation or dependence across dimensions.[5] While the standard formulation targets real-valued spaces , the concept extends to more general settings, including complex vector spaces in applications like signal processing, where the random vector is a measurable mapping to .[6]Notation and Examples
Standard notation for a multivariate random variable employs boldface letters to denote vectors, with representing a -dimensional random vector, where each is a univariate random variable and the superscript indicates transposition to a column vector. A specific realization or observed value of this vector is denoted by lowercase , allowing distinction between the random entity and its outcomes. To illustrate, consider a bivariate case where captures the height and weight of adult humans in a population; here, might represent height in inches and weight in pounds, with joint variability reflecting biological dependencies.[4] For a continuous example, a multivariate normal random vector has mean vector and positive semi-definite covariance matrix , commonly modeling correlated measurements like stock returns or sensor data in dimensions. In the discrete setting, rolling two fair six-sided dice yields , where and are the outcomes (each uniformly distributed over ), and the joint probability mass function is for , demonstrating uniform joint support over 36 points.[7] Basic vector operations on multivariate random variables proceed component-wise: addition combines vectors, while scalar multiplication scales them, though these lack probabilistic interpretation in isolation here. The foundations of multivariate analysis trace to 19th-century statistics, notably Karl Pearson's 1896 exploration of correlations among multiple organ measurements in biological data, which laid groundwork for handling joint variability beyond univariate cases.[8]Joint Distributions
Discrete Multivariate Random Variables
A discrete multivariate random variable is a vector where each component takes values in a countable set, typically integers or subsets thereof.[9] The probability structure is fully specified by the joint probability mass function (PMF), which assigns probabilities to each possible outcome in the support.[10] The joint PMF is defined as , where is a specific realization, and it satisfies for all in the support.[11] The normalization condition requires that the probabilities sum to unity over the entire support: , where the sum is taken over all possible .[10] The support of is the finite or countably infinite set of points in (or a subset) where .[9] Consider the example of two independent fair coin flips, where with indicating the outcome of the first flip (0 for tails, 1 for heads) and for the second. The support is , and the joint PMF is uniform due to independence and fairness: for each point in the support.[12] This can be represented in a table:| 0 | 1 | |
|---|---|---|
| 0 | 0.25 | 0.25 |
| 1 | 0.25 | 0.25 |