Hubbry Logo
Random variableRandom variableMain
Open search
Random variable
Community hub
Random variable
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Random variable
Random variable
from Wikipedia

A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. The term 'random variable' in its mathematical definition refers to neither randomness nor variability[1] but instead is a mathematical function in which

  • the domain is the set of possible outcomes in a sample space (e.g. the set which are the possible upper sides of a flipped coin heads or tails as the result from tossing a coin); and
  • the range is a measurable space (e.g. corresponding to the domain above, the range might be the set if say heads mapped to -1 and mapped to 1). Typically, the range of a random variable is a subset of the real numbers.
This graph shows how a random variable is a function from all possible outcomes to real values. It also shows how a random variable is used for defining probability mass functions.

Informally, randomness typically represents some fundamental element of chance, such as in the roll of a die; it may also represent uncertainty, such as measurement error.[2] However, the interpretation of probability is philosophically complicated, and even in specific cases is not always straightforward. The purely mathematical analysis of random variables is independent of such interpretational difficulties, and can be based upon a rigorous axiomatic setup.

In the formal mathematical language of measure theory, a random variable is defined as a measurable function from a probability measure space (called the sample space) to a measurable space. This allows consideration of the pushforward measure, which is called the distribution of the random variable; the distribution is thus a probability measure on the set of all possible values of the random variable. It is possible for two random variables to have identical distributions but to differ in significant ways; for instance, they may be independent.

It is common to consider the special cases of discrete random variables and absolutely continuous random variables, corresponding to whether a random variable is valued in a countable subset or in an interval of real numbers. There are other important possibilities, especially in the theory of stochastic processes, wherein it is natural to consider random sequences or random functions. Sometimes a random variable is taken to be automatically valued in the real numbers, with more general random quantities instead being called random elements.

According to George Mackey, Pafnuty Chebyshev was the first person "to think systematically in terms of random variables".[3]

Definition

[edit]

A random variable is a measurable function from a sample space as a set of possible outcomes to a measurable space . For the measurability of to be meaningful, the sample space needs to belong to a probability triple (see the measure-theoretic definition). A random variable is often denoted by capital Roman letters such as .[4]

The probability that takes on a value in a measurable set is written as

.

Standard case

[edit]

In many cases, is real-valued, i.e. . In some contexts, the term random element (see extensions) is used to denote a random variable not of this form.

When the image (or range) of is finite or countably infinite, the random variable is called a discrete random variable[5]: 399  and its distribution is a discrete probability distribution, i.e. can be described by a probability mass function that assigns a probability to each value in the image of . If the image is uncountably infinite (usually an interval) then is called a continuous random variable.[6][7] In the special case that it is absolutely continuous, its distribution can be described by a probability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous.[8]

Any random variable can be described by its cumulative distribution function, which describes the probability that the random variable will be less than or equal to a certain value.

Extensions

[edit]

The term "random variable" in statistics is traditionally limited to the real-valued case (). In this case, the structure of the real numbers makes it possible to define quantities such as the expected value and variance of a random variable, its cumulative distribution function, and the moments of its distribution.

However, the definition above is valid for any measurable space of values. Thus one can consider random elements of other sets , such as random Boolean values, categorical values, complex numbers, vectors, matrices, sequences, trees, sets, shapes, manifolds, and functions. One may then specifically refer to a random variable of type , or an -valued random variable.

This more general concept of a random element is particularly useful in disciplines such as graph theory, machine learning, natural language processing, and other fields in discrete mathematics and computer science, where one is often interested in modeling the random variation of non-numerical data structures. In some cases, it is nonetheless convenient to represent each element of , using one or more real numbers. In this case, a random element may optionally be represented as a vector of real-valued random variables (all defined on the same underlying probability space , which allows the different random variables to covary). For example:

  • A random word may be represented as a random integer that serves as an index into the vocabulary of possible words. Alternatively, it can be represented as a random indicator vector, whose length equals the size of the vocabulary, where the only values of positive probability are , , and the position of the 1 indicates the word.
  • A random sentence of given length may be represented as a vector of random words.
  • A random graph on given vertices may be represented as a matrix of random variables, whose values specify the adjacency matrix of the random graph.
  • A random function may be represented as a collection of random variables , giving the function's values at the various points in the function's domain. The are ordinary real-valued random variables provided that the function is real-valued. For example, a stochastic process is a random function of time, a random vector is a random function of some index set such as , and a random field is a random function on any set (typically time, space, or a discrete set).

Distribution functions

[edit]

If a random variable defined on the probability space is given, we can ask questions like "How likely is it that the value of is equal to 2?". This is the same as the probability of the event which is often written as or for short.

Recording all these probabilities of outputs of a random variable yields the probability distribution of . The probability distribution "forgets" about the particular probability space used to define and only records the probabilities of various output values of . Such a probability distribution, if is real-valued, can always be captured by its cumulative distribution function

and sometimes also using a probability density function, . In measure-theoretic terms, we use the random variable to "push-forward" the measure on to a measure on . The measure is called the "(probability) distribution of " or the "law of ". [9] The density , the Radon–Nikodym derivative of with respect to some reference measure on (often, this reference measure is the Lebesgue measure in the case of continuous random variables, or the counting measure in the case of discrete random variables). The underlying probability space is a technical device used to guarantee the existence of random variables, sometimes to construct them, and to define notions such as correlation and dependence or independence based on a joint distribution of two or more random variables on the same probability space. In practice, one often disposes of the space altogether and just puts a measure on that assigns measure 1 to the whole real line, i.e., one works with probability distributions instead of random variables. See the article on quantile functions for fuller development.

Examples

[edit]

Discrete random variable

[edit]

Consider an experiment where a person is chosen at random. An example of a random variable may be the person's height. Mathematically, the random variable is interpreted as a function which maps the person to their height. Associated with the random variable is a probability distribution that allows the computation of the probability that the height is in any subset of possible values, such as the probability that the height is between 180 and 190 cm, or the probability that the height is either less than 150 or more than 200 cm.

Another random variable may be the person's number of children; this is a discrete random variable with non-negative integer values. It allows the computation of probabilities for individual integer values – the probability mass function (PMF) – or for sets of values, including infinite sets. For example, the event of interest may be "an even number of children". For both finite and infinite event sets, their probabilities can be found by adding up the PMFs of the elements; that is, the probability of an even number of children is the infinite sum .

In examples such as these, the sample space is often suppressed, since it is mathematically hard to describe, and the possible values of the random variables are then treated as a sample space. But when two random variables are measured on the same sample space of outcomes, such as the height and number of children being computed on the same random persons, it is easier to track their relationship if it is acknowledged that both height and number of children come from the same random person, for example so that questions of whether such random variables are correlated or not can be posed.

If are countable sets of real numbers, and , then is a discrete distribution function. Here for , for . Taking for instance an enumeration of all rational numbers as , one gets a discrete function that is not necessarily a step function (piecewise constant).

Coin toss

[edit]

The possible outcomes for one coin toss can be described by the sample space . We can introduce a real-valued random variable that models a $1 payoff for a successful bet on heads as follows:

If the coin is a fair coin, Y has a probability mass function given by:

Dice roll

[edit]
If the sample space is the set of possible numbers rolled on two dice, and the random variable of interest is the sum S of the numbers on the two dice, then S is a discrete random variable whose distribution is described by the probability mass function plotted as the height of picture columns here.

A random variable can also be used to describe the process of rolling dice and the possible outcomes. The most obvious representation for the two-dice case is to take the set of pairs of numbers n1 and n2 from {1, 2, 3, 4, 5, 6} (representing the numbers on the two dice) as the sample space. The total number rolled (the sum of the numbers in each pair) is then a random variable X given by the function that maps the pair to the sum: and (if the dice are fair) has a probability mass function fX given by:

Continuous random variable

[edit]

Formally, a continuous random variable is a random variable whose cumulative distribution function is continuous everywhere.[10] There are no "gaps", which would correspond to numbers which have a finite probability of occurring. Instead, continuous random variables almost never take an exact prescribed value c (formally, ) but there is a positive probability that its value will lie in particular intervals which can be arbitrarily small. Continuous random variables usually admit probability density functions (PDF), which characterize their CDF and probability measures; such distributions are also called absolutely continuous; but some continuous distributions are singular, or mixes of an absolutely continuous part and a singular part.

An example of a continuous random variable would be one based on a spinner that can choose a horizontal direction. Then the values taken by the random variable are directions. We could represent these directions by North, West, East, South, Southeast, etc. However, it is commonly more convenient to map the sample space to a random variable which takes values which are real numbers. This can be done, for example, by mapping a direction to a bearing in degrees clockwise from North. The random variable then takes values which are real numbers from the interval [0, 360), with all parts of the range being "equally likely". In this case, X = the angle spun. Any real number has probability zero of being selected, but a positive probability can be assigned to any range of values. For example, the probability of choosing a number in [0, 180] is 12. Instead of speaking of a probability mass function, we say that the probability density of X is 1/360. The probability of a subset of [0, 360) can be calculated by multiplying the measure of the set by 1/360. In general, the probability of a set for a given continuous random variable can be calculated by integrating the density over the given set.

More formally, given any interval , a random variable is called a "continuous uniform random variable" (CURV) if the probability that it takes a value in a subinterval depends only on the length of the subinterval. This implies that the probability of falling in any subinterval is proportional to the length of the subinterval, that is, if acdb, one has

where the last equality results from the unitarity axiom of probability. The probability density function of a CURV is given by the indicator function of its interval of support normalized by the interval's length: Of particular interest is the uniform distribution on the unit interval . Samples of any desired probability distribution can be generated by calculating the quantile function of on a randomly-generated number distributed uniformly on the unit interval. This exploits properties of cumulative distribution functions, which are a unifying framework for all random variables.

Mixed type

[edit]

A mixed random variable is a random variable whose cumulative distribution function is neither discrete nor everywhere-continuous.[10] It can be realized as a mixture of a discrete random variable and a continuous random variable; in which case the CDF will be the weighted average of the CDFs of the component variables.[10]

An example of a random variable of mixed type would be based on an experiment where a coin is flipped and the spinner is spun only if the result of the coin toss is heads. If the result is tails, X = −1; otherwise X = the value of the spinner as in the preceding example. There is a probability of 12 that this random variable will have the value −1. Other ranges of values would have half the probabilities of the last example.

Most generally, every probability distribution on the real line is a mixture of discrete part, singular part, and an absolutely continuous part; see Lebesgue's decomposition theorem § Refinement. The discrete part is concentrated on a countable set, but this set may be dense (like the set of all rational numbers).

Measure-theoretic definition

[edit]

The most formal, axiomatic definition of a random variable involves measure theory. Continuous random variables are defined in terms of sets of numbers, along with functions that map such sets to probabilities. Because of various difficulties (e.g. the Banach–Tarski paradox) that arise if such sets are insufficiently constrained, it is necessary to introduce what is termed a sigma-algebra to constrain the possible sets over which probabilities can be defined. Normally, a particular such sigma-algebra is used, the Borel σ-algebra, which allows for probabilities to be defined over any sets that can be derived either directly from continuous intervals of numbers or by a finite or countably infinite number of unions and/or intersections of such intervals.[11]

The measure-theoretic definition is as follows.

Let be a probability space and a measurable space. Then an -valued random variable is a measurable function , which means that, for every subset , its preimage is -measurable; , where .[12] This definition enables us to measure any subset in the target space by looking at its preimage, which by assumption is measurable.

In more intuitive terms, a member of is a possible outcome, a member of is a measurable subset of possible outcomes, the function gives the probability of each such measurable subset, represents the set of values that the random variable can take (such as the set of real numbers), and a member of is a "well-behaved" (measurable) subset of (those for which the probability may be determined). The random variable is then a function from any outcome to a quantity, such that the outcomes leading to any useful subset of quantities for the random variable have a well-defined probability.

When is a topological space, then the most common choice for the σ-algebra is the Borel σ-algebra , which is the σ-algebra generated by the collection of all open sets in . In such case the -valued random variable is called an -valued random variable. Moreover, when the space is the real line , then such a real-valued random variable is called simply a random variable.

Real-valued random variables

[edit]

In this case the observation space is the set of real numbers. Recall, is the probability space. For a real observation space, the function is a real-valued random variable if

This definition is a special case of the above because the set generates the Borel σ-algebra on the set of real numbers, and it suffices to check measurability on any generating set. Here we can prove measurability on this generating set by using the fact that .

Moments

[edit]

The probability distribution of a random variable is often characterised by a small number of parameters, which also have a practical interpretation. For example, it is often enough to know what its "average value" is. This is captured by the mathematical concept of expected value of a random variable, denoted , and also called the first moment. In general, is not equal to . Once the "average value" is known, one could then ask how far from this average value the values of typically are, a question that is answered by the variance and standard deviation of a random variable. can be viewed intuitively as an average obtained from an infinite population, the members of which are particular evaluations of .

Mathematically, this is known as the (generalised) problem of moments: for a given class of random variables , find a collection of functions such that the expectation values fully characterise the distribution of the random variable .

Moments can only be defined for real-valued functions of random variables (or complex-valued, etc.). If the random variable is itself real-valued, then moments of the variable itself can be taken, which are equivalent to moments of the identity function of the random variable. However, even for non-real-valued random variables, moments can be taken of real-valued functions of those variables. For example, for a categorical random variable X that can take on the nominal values "red", "blue" or "green", the real-valued function can be constructed; this uses the Iverson bracket, and has the value 1 if has the value "green", 0 otherwise. Then, the expected value and other moments of this function can be determined.

Functions of random variables

[edit]

A new random variable Y can be defined by applying a real Borel measurable function to the outcomes of a real-valued random variable . That is, . The cumulative distribution function of is then

If function is invertible (i.e., exists, where is 's inverse function) and is either increasing or decreasing, then the previous relation can be extended to obtain

With the same hypotheses of invertibility of , assuming also differentiability, the relation between the probability density functions can be found by differentiating both sides of the above expression with respect to , in order to obtain[10]

If there is no invertibility of but each admits at most a countable number of roots (i.e., a finite, or countably infinite, number of such that ) then the previous relation between the probability density functions can be generalized with

where , according to the inverse function theorem. The formulas for densities do not demand to be increasing.

In the measure-theoretic, axiomatic approach to probability, if a random variable on and a Borel measurable function , then is also a random variable on , since the composition of measurable functions is also measurable. (However, this is not necessarily true if is Lebesgue measurable.[citation needed]) The same procedure that allowed one to go from a probability space to can be used to obtain the distribution of .

Example 1

[edit]

Let be a real-valued, continuous random variable and let .

If , then , so

If , then

so

Example 2

[edit]

Suppose is a random variable with a cumulative distribution

where is a fixed parameter. Consider the random variable Then,

The last expression can be calculated in terms of the cumulative distribution of so

which is the cumulative distribution function (CDF) of an exponential distribution.

Example 3

[edit]

Suppose is a random variable with a standard normal distribution, whose density is

Consider the random variable We can find the density using the above formula for a change of variables:

In this case the change is not monotonic, because every value of has two corresponding values of (one positive and negative). However, because of symmetry, both halves will transform identically, i.e.,

The inverse transformation is

and its derivative is

Then,

This is a chi-squared distribution with one degree of freedom.

Example 4

[edit]

Suppose is a random variable with a normal distribution, whose density is

Consider the random variable We can find the density using the above formula for a change of variables:

In this case the change is not monotonic, because every value of has two corresponding values of (one positive and negative). Differently from the previous example, in this case however, there is no symmetry and we have to compute the two distinct terms:

The inverse transformation is

and its derivative is

Then,

This is a noncentral chi-squared distribution with one degree of freedom.

Some properties

[edit]
  • The probability distribution of the sum of two independent random variables is the convolution of each of their distributions.
  • Probability distributions are not a vector space—they are not closed under linear combinations, as these do not preserve non-negativity or total integral 1—but they are closed under convex combination, thus forming a convex subset of the space of functions (or measures).

Equivalence of random variables

[edit]

There are several different senses in which random variables can be considered to be equivalent. Two random variables can be equal, equal almost surely, or equal in distribution.

In increasing order of strength, the precise definition of these notions of equivalence is given below.

Equality in distribution

[edit]

If the sample space is a subset of the real line, random variables X and Y are equal in distribution (denoted ) if they have the same distribution functions:

To be equal in distribution, random variables need not be defined on the same probability space. Two random variables having equal moment generating functions have the same distribution. This provides, for example, a useful method of checking equality of certain functions of independent, identically distributed (IID) random variables. However, the moment generating function exists only for distributions that have a defined Laplace transform.

Almost sure equality

[edit]

Two random variables X and Y are equal almost surely (denoted ) if, and only if, the probability that they are different is zero:

For all practical purposes in probability theory, this notion of equivalence is as strong as actual equality. It is associated to the following distance:

where "ess sup" represents the essential supremum in the sense of measure theory.

Equality

[edit]

Finally, the two random variables X and Y are equal if they are equal as functions on their measurable space:

This notion is typically the least useful in probability theory because in practice and in theory, the underlying measure space of the experiment is rarely explicitly characterized or even characterizable.

Practical difference between notions of equivalence

[edit]

Since we rarely explicitly construct the probability space underlying a random variable, the difference between these notions of equivalence is somewhat subtle. Essentially, two random variables considered in isolation are "practically equivalent" if they are equal in distribution -- but once we relate them to other random variables defined on the same probability space, then they only remain "practically equivalent" if they are equal almost surely.

For example, consider the real random variables A, B, C, and D all defined on the same probability space. Suppose that A and B are equal almost surely (), but A and C are only equal in distribution (). Then , but in general (not even in distribution). Similarly, we have that the expectation values , but in general . Therefore, two random variables that are equal in distribution (but not equal almost surely) can have different covariances with a third random variable.

Convergence

[edit]

A significant theme in mathematical statistics consists of obtaining convergence results for certain sequences of random variables; for instance the law of large numbers and the central limit theorem.

There are various senses in which a sequence of random variables can converge to a random variable . These are explained in the article on convergence of random variables.

See also

[edit]

References

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
In , a random variable is a function that assigns a to each outcome in the of a random experiment, enabling the quantification of through numerical values rather than qualitative descriptions. This mapping allows probabilities to be defined over the possible values of the variable, facilitating in fields such as , , and . Random variables are broadly classified into two types: discrete and continuous. A discrete random variable takes on a countable number of distinct values, such as the number of heads in a series of coin flips, where the possible outcomes form a finite or countably infinite set. Its probability distribution is described by a probability mass function (PMF), which assigns a probability to each possible value, with the sum of these probabilities equaling 1. In contrast, a continuous random variable can assume any value within a continuous range, such as the exact time until an event occurs, representing an uncountably infinite set of outcomes. The distribution of a continuous random variable is characterized by a probability density function (PDF), where probabilities are computed as integrals over intervals, and the total area under the PDF equals 1. Key properties of random variables include the (or ), which represents the long-run average value of the variable over many repetitions of the experiment, and the variance, which measures the spread or dispersion of the variable's values around the . For a discrete random variable XX, the expected value is E(X)=xP(X=x)E(X) = \sum x \cdot P(X = x), while the variance is Var(X)=E[(XE(X))2]=E(X2)[E(X)]2Var(X) = E[(X - E(X))^2] = E(X^2) - [E(X)]^2. These properties extend to continuous cases via integrals, providing foundational tools for deriving further statistics like standard deviation and for applications in modeling real-world phenomena. Random variables also form the basis for joint distributions when multiple variables are considered together, allowing analysis of dependence and in multivariate settings.

Basic Concepts

Definition

In the early , the concept of a random variable emerged as a key element in the axiomatization of . The Italian Paolo Cantelli introduced the term variabile casuale (random variable) around 1913 in his work on probability limits, providing an early formal recognition of variables whose values depend on chance outcomes. This idea was further developed through Andrey Kolmogorov's seminal 1933 monograph Grundbegriffe der Wahrscheinlichkeitsrechnung (Foundations of the Theory of Probability), which established the modern axiomatic framework for probability and defined random variables rigorously within it. The English term "random variable" was used by J. V. Uspensky in his 1937 textbook Introduction to Mathematical Probability. Intuitively, a random variable XX assigns a numerical value to each possible outcome of a random experiment, thereby quantifying uncertain phenomena in a measurable way. For instance, in an experiment consisting of tossing a three times, the random variable XX might represent the number of heads obtained, mapping each outcome sequence (e.g., HHT) to the 2. This abstraction allows probabilities to be associated with the values taken by XX rather than directly with the underlying outcomes. Formally, a random variable XX is defined as a X:ΩR,X: \Omega \to \mathbb{R}, where (Ω,F,P)(\Omega, \mathcal{F}, P) is a . Here, Ω\Omega is the sample space, the set of all possible outcomes of the random experiment; F\mathcal{F} is a σ\sigma-algebra of subsets of Ω\Omega, known as the event space, which specifies the collection of measurable events; and PP is a probability measure on F\mathcal{F} that assigns a value between 0 and 1 to each event, satisfying Kolmogorov's axioms of probability (non-negativity, normalization, and countable additivity). The measurability of XX ensures compatibility with the probability structure, requiring that for every xRx \in \mathbb{R}, the preimage set {ωΩ:X(ω)x}\{\omega \in \Omega : X(\omega) \leq x\} belongs to F\mathcal{F}. This condition guarantees that events defined in terms of XX, such as {Xx}\{X \leq x\}, are well-defined and assignable probabilities under PP. Random variables may take discrete or continuous values, but the general definition encompasses both cases.

Probability Space

A probability space provides the mathematical foundation for defining random variables and modeling uncertainty in a rigorous manner. It is formally defined as a triple (Ω,F,P)(\Omega, \mathcal{F}, P), where Ω\Omega is the consisting of all possible outcomes of a random experiment, F\mathcal{F} is a sigma-algebra of subsets of Ω\Omega (known as events) that is closed under countable unions, intersections, and complements, and PP is a assigning to each event in F\mathcal{F} a value between 0 and 1, with the normalization condition P(Ω)=1P(\Omega) = 1. The axioms governing the probability measure PP were established by Andrey Kolmogorov in his seminal 1933 work, providing an axiomatic basis for probability theory. These axioms are: (1) non-negativity, stating that P(A)0P(A) \geq 0 for every event AFA \in \mathcal{F}; (2) normalization, P(Ω)=1P(\Omega) = 1; and (3) countable additivity, which asserts that if {Ai}i=1\{A_i\}_{i=1}^\infty is a countable collection of pairwise disjoint events in F\mathcal{F}, then P(i=1Ai)=i=1P(Ai).P\left( \bigcup_{i=1}^\infty A_i \right) = \sum_{i=1}^\infty P(A_i). These axioms ensure that probabilities behave consistently for complex events built from simpler ones. Examples of probability spaces illustrate their versatility across discrete and continuous settings. In a finite discrete case, such as a toss, the is Ω={H,T}\Omega = \{H, T\} (heads or tails), the sigma-algebra F\mathcal{F} is the power set of Ω\Omega with four elements, and PP assigns equal probability 1/21/2 to each singleton event. For a continuous case, consider a uniform distribution over the unit interval, where Ω=[0,1]\Omega = [0,1], F\mathcal{F} is the Borel sigma-algebra generated by open intervals, and PP is the restricted to [0,1][0,1], so P([a,b])=baP([a,b]) = b - a for 0ab10 \leq a \leq b \leq 1. Every random variable is defined on a (Ω,F,P)(\Omega, \mathcal{F}, P), which guarantees its measurability with respect to F\mathcal{F} and allows the assignment of probabilities to the variable's outcomes.

Types of Random Variables

Discrete Random Variables

A discrete random variable is a random variable whose range, or set of possible values, is countable, meaning it consists of either a finite number of distinct outcomes or a countably infinite number, such as the non-negative integers. Unlike more general random variables defined on a , discrete random variables assign positive probabilities only to these countable points, with the total probability summing to 1 across the entire support. The (PMF) of a discrete random variable XX, denoted pX(x)p_X(x) or simply p(x)p(x), provides the probability that XX takes a specific value xx in its range, so p(x)=P(X=x)p(x) = P(X = x). This function satisfies two key properties: p(x)0p(x) \geq 0 for all xx in the range, and the sum of p(x)p(x) over all possible xx equals 1, i.e., xp(x)=1\sum_{x} p(x) = 1. The support of XX, denoted supp(X)\operatorname{supp}(X), is the smallest set containing all xx such that p(x)>0p(x) > 0, ensuring probabilities are concentrated only on these points. Common examples of discrete random variables include the , which takes only the values 0 or 1, and the , which takes values in the non-negative integers {0,1,2,}\{0, 1, 2, \dots\}. To compute probabilities for intervals, the probability that XX falls between integers aa and bb (inclusive), where aba \leq b, is given by summing the PMF over those values: P(aXb)=x=abp(x)P(a \leq X \leq b) = \sum_{x=a}^{b} p(x). This summation leverages the countable nature of the support, allowing exact calculation via the discrete probabilities.

Continuous Random Variables

A continuous random variable is defined as a random variable whose possible values form an , such as the real line R\mathbb{R} or a continuous interval within it, with the probability of the variable equaling any specific point being zero: P(X=x)=0P(X = x) = 0 for every xx in the support. This contrasts with discrete random variables, which assign positive probabilities to countable points. Probabilities for continuous random variables are determined over intervals rather than at individual points, reflecting the infinite and uncountable nature of their range. Specifically, the probability P(aXb)P(a \leq X \leq b) for an interval [a,b][a, b] is obtained by integrating a non-negative density function over that interval, ensuring the total probability across the entire support equals 1. This integral-based approach allows for modeling phenomena with inherently continuous outcomes, such as physical measurements. Although continuous random variables lack positive probability at single points, they can approximate discrete distributions in limiting scenarios, such as when the number of discrete categories increases indefinitely. Representative examples include the uniform distribution on the interval [0,1], which assigns equal likelihood to all points within a bounded continuous range, and the , commonly used to model waiting times between events in continuous-time processes. A defining mathematical property in standard usage is that the (CDF) of an absolutely continuous random variable—which is the typical sense of "continuous" in introductory contexts—is absolutely continuous with respect to , meaning it can be expressed as the integral of a density function and possesses no jumps. Singular continuous distributions, discussed separately, represent a more advanced case without densities.

Singular and Mixed Random Variables

In , a singular continuous random variable is characterized by a (CDF) that is continuous everywhere but not absolutely continuous with respect to , implying the absence of a while lacking discrete jumps. This means the distribution is supported on a set of Lebesgue measure zero, yet it assigns positive probability to intervals without concentrating mass at points. A example is the , whose CDF is the —also known as the —which is constant on the intervals removed in the construction of the and increases continuously from 0 to 1 over [0,1], with support confined to the Cantor set of measure zero. Mixed random variables arise when the distribution combines discrete and continuous components, resulting in a CDF that exhibits jumps at discrete points alongside regions of continuous increase. For instance, consider a random variable XX with P(X=0)=0.5P(X=0) = 0.5 and, conditional on X>0X > 0, XX on (0,1](0,1] with probability 0.5; here, the distribution places a point mass at 0 while spreading the remaining probability continuously over an interval. In general, the CDF of a mixed random variable can be expressed as F(x)=yxpy+xf(t)dt+Fs(x),F(x) = \sum_{y \leq x} p_y + \int_{-\infty}^x f(t) \, dt + F_s(x), where py\sum p_y captures the discrete jumps, f(t)dt\int f(t) \, dt the absolutely continuous part, and Fs(x)F_s(x) the singular continuous component, though the latter is often absent in practical mixed cases. The Lebesgue decomposition theorem provides the foundational result for classifying all probability distributions on the real line, stating that any such distribution μ\mu can be uniquely decomposed as μ=μd+μac+μs\mu = \mu_d + \mu_{ac} + \mu_s, where μd\mu_d is the discrete (atomic) part, μac\mu_{ac} is absolutely continuous with respect to , and μs\mu_s is singular continuous. This theorem underscores that singular continuous distributions form a distinct class, separate from both discrete and absolutely continuous types. In applications, singular and fully mixed distributions (including singular parts) are rare, as most probabilistic models in and rely on purely discrete or absolutely continuous random variables for tractability; singular examples like the primarily serve theoretical purposes in measure theory and .

Distribution Functions

Cumulative Distribution Function

The (CDF) of a real-valued random variable XX, denoted FX(x)F_X(x), is defined as FX(x)=P(Xx)F_X(x) = P(X \leq x) for all xRx \in \mathbb{R}. This function provides a complete description of the of XX, applicable to discrete, continuous, or mixed cases. The CDF possesses several fundamental properties: it is non-decreasing, meaning FX(a)FX(b)F_X(a) \leq F_X(b) whenever a<ba < b; right-continuous, so FX(x)=limyx+FX(y)F_X(x) = \lim_{y \to x^+} F_X(y); and it satisfies the boundary conditions limxFX(x)=0\lim_{x \to -\infty} F_X(x) = 0 and limxFX(x)=1\lim_{x \to \infty} F_X(x) = 1. These ensure that FX(x)F_X(x) maps the real line to the interval [0,1][0, 1] in a consistent manner with probability axioms. Probabilities over intervals can be computed directly from the CDF: for any a<ba < b, P(a<Xb)=FX(b)FX(a)P(a < X \leq b) = F_X(b) - F_X(a). This property allows the CDF to specify all finite-dimensional distributions, thereby uniquely determining the law (or distribution) of XX. The form of the CDF reveals the type of random variable: discontinuities or jumps correspond to discrete components, where the jump size at a point equals the probability mass there, while continuous and differentiable portions indicate absolutely continuous parts. The quantile function, or generalized inverse of the CDF, is defined for u(0,1)u \in (0,1) as FX1(u)=inf{x:FX(x)u},F_X^{-1}(u) = \inf\{x : F_X(x) \geq u\}, providing the smallest xx such that the CDF reaches at least uu. This function is non-decreasing and left-continuous, facilitating the generation of random variables from uniform distributions via the inverse transform sampling method.

Probability Mass and Density Functions

For discrete random variables, the probability mass function (PMF), denoted p(x)p(x), assigns to each possible value xx in the support the probability p(x)=P(X=x)0p(x) = P(X = x) \geq 0. This function fully characterizes the distribution, as the probability of XX taking any finite or countable set of values AA is P(XA)=xAp(x)P(X \in A) = \sum_{x \in A} p(x). The PMF relates to the cumulative distribution function (CDF) F(x)=P(Xx)F(x) = P(X \leq x) through the jumps in the CDF, specifically p(x)=F(x)F(x)p(x) = F(x) - F(x^-), where F(x)=limyxF(y)F(x^-) = \lim_{y \uparrow x} F(y) denotes the left-hand limit at xx. A fundamental property is normalization: xp(x)=1\sum_{x} p(x) = 1, with the sum taken over the countable support of XX. The PMF enables computation of expectations for functions of the random variable. For a measurable function gg, the expectation is E[g(X)]=xg(x)p(x)E[g(X)] = \sum_{x} g(x) p(x), provided the sum converges absolutely. This includes key quantities like the mean E[X]=xxp(x)E[X] = \sum_{x} x p(x), assuming finite support or appropriate convergence. For absolutely continuous random variables, the probability density function (PDF), denoted f(x)f(x), provides a density with respect to Lebesgue measure such that probabilities are given by integrals: P(a<Xb)=abf(x)dxP(a < X \leq b) = \int_{a}^{b} f(x) \, dx. The PDF is obtained from the CDF as its derivative where differentiable: f(x)=ddxF(x)f(x) = \frac{d}{dx} F(x). Conversely, the CDF recovers via F(x)=xf(t)dtF(x) = \int_{-\infty}^{x} f(t) \, dt. The PDF satisfies f(x)0f(x) \geq 0 for all xx and the normalization condition f(x)dx=1\int_{-\infty}^{\infty} f(x) \, dx = 1. Expectations using the PDF follow E[g(X)]=g(x)f(x)dxE[g(X)] = \int_{-\infty}^{\infty} g(x) f(x) \, dx, again assuming absolute integrability. For instance, the mean is E[X]=xf(x)dxE[X] = \int_{-\infty}^{\infty} x f(x) \, dx. While the PDF uniquely determines the distribution for absolutely continuous cases (up to sets of Lebesgue measure zero), representations involving generalized functions like Dirac deltas are not unique, as one can add such components without altering probabilities under integration. Singular distributions, which have a continuous CDF but are not absolutely continuous with respect to , admit no ordinary PDF.

Examples

Discrete Examples

A Bernoulli random variable is the simplest discrete random variable, taking only two possible values: 1 (representing success) with probability pp and 0 (representing failure) with probability 1p1 - p, where 0<p<10 < p < 1. The probability mass function (PMF) is given by P(X=x)=px(1p)1x,x{0,1}.P(X = x) = p^x (1 - p)^{1 - x}, \quad x \in \{0, 1\}. For example, if p=0.6p = 0.6, then P(X=1)=0.6P(X = 1) = 0.6 and P(X=0)=0.4P(X = 0) = 0.4. The binomial random variable generalizes the Bernoulli by representing the number of successes in nn independent Bernoulli trials, each with success probability pp. Its support is the integers k=0,1,,nk = 0, 1, \dots, n, and the PMF is P(X=k)=(nk)pk(1p)nk,P(X = k) = \binom{n}{k} p^k (1 - p)^{n - k}, where (nk)=n!k!(nk)!\binom{n}{k} = \frac{n!}{k!(n - k)!} is the binomial coefficient counting the number of ways to choose kk successes out of nn trials. Thus, a binomial random variable is the sum of nn independent Bernoulli random variables with the same pp. A classic binomial example is the number of heads in nn tosses of a fair coin, where success is heads with p=0.5p = 0.5. For n=3n = 3, the probability of exactly 2 heads is P(X=2)=(32)(0.5)2(0.5)1=3×0.125=0.375=38.P(X = 2) = \binom{3}{2} (0.5)^2 (0.5)^{1} = 3 \times 0.125 = 0.375 = \frac{3}{8}. The expected value (mean) is np=3×0.5=1.5np = 3 \times 0.5 = 1.5. Another common discrete example is the outcome of a fair six-sided dice roll, which follows a discrete uniform distribution on the set {1,2,3,4,5,6}\{1, 2, 3, 4, 5, 6\}, with each outcome equally likely. The PMF is P(X=x)=16,x=1,2,,6.P(X = x) = \frac{1}{6}, \quad x = 1, 2, \dots, 6. This distribution assigns equal probability 1/N1/N to each of NN possible outcomes.

Continuous Examples

The uniform distribution on the interval [a,b][a, b], where a<ba < b, serves as a foundational example of a continuous random variable, representing equal likelihood across a finite range. Its probability density function (PDF) is defined as f(x)=1ba,axb,f(x) = \frac{1}{b - a}, \quad a \leq x \leq b, and f(x)=0f(x) = 0 otherwise. The corresponding cumulative distribution function (CDF) is F(x)=xaba,axb,F(x) = \frac{x - a}{b - a}, \quad a \leq x \leq b, with F(x)=0F(x) = 0 for x<ax < a and F(x)=1F(x) = 1 for x>bx > b. For the standard uniform distribution on [0,1][0, 1], the probability P(0.2<X<0.5)P(0.2 < X < 0.5) is computed by integrating the PDF over the interval, yielding 0.30.3. The exponential distribution, parameterized by rate λ>0\lambda > 0, exemplifies continuous random variables in modeling waiting times until the first event in a Poisson process. Its PDF is f(x)=λeλx,x0,f(x) = \lambda e^{-\lambda x}, \quad x \geq 0, and f(x)=0f(x) = 0 otherwise. The probability that the waiting time exceeds t0t \geq 0 is P(X>t)=eλtP(X > t) = e^{-\lambda t}. For λ=1\lambda = 1, the , or , is 1/λ=11/\lambda = 1. The normal distribution, with mean μ\mu and variance σ2>0\sigma^2 > 0, is a cornerstone continuous distribution characterized by its symmetric bell-shaped curve. Its PDF is f(x)=12πσ2exp((xμ)22σ2),<x<.f(x) = \frac{1}{\sqrt{2\pi \sigma^2}} \exp\left( -\frac{(x - \mu)^2}{2\sigma^2} \right), \quad -\infty < x < \infty.
Add your contribution
Related Hubs
User Avatar
No comments yet.