Hubbry Logo
Conditional dependenceConditional dependenceMain
Open search
Conditional dependence
Community hub
Conditional dependence
logo
8 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Conditional dependence
Conditional dependence
from Wikipedia
A Bayesian network illustrating conditional dependence

In probability theory, conditional dependence is a relationship between two or more events that are dependent when a third event occurs.[1] For example, if and are two events that individually increase the probability of a third event and do not directly affect each other, then initially (when it has not been observed whether or not the event occurs)[2][3] ( are independent).

But suppose that now is observed to occur. If event occurs then the probability of occurrence of the event will decrease because its positive relation to is less necessary as an explanation for the occurrence of (similarly, event occurring will decrease the probability of occurrence of ). Hence, now the two events and are conditionally negatively dependent on each other because the probability of occurrence of each is negatively dependent on whether the other occurs. We have[4]

Conditional dependence of A and B given C is the logical negation of conditional independence .[5] In conditional independence two events (which may be dependent or not) become independent given the occurrence of a third event.[6]

Example

[edit]

In essence probability is influenced by a person's information about the possible occurrence of an event. For example, let the event be 'I have a new phone'; event be 'I have a new watch'; and event be 'I am happy'; and suppose that having either a new phone or a new watch increases the probability of my being happy. Let us assume that the event has occurred – meaning 'I am happy'. Now if another person sees my new watch, he/she will reason that my likelihood of being happy was increased by my new watch, so there is less need to attribute my happiness to a new phone.

To make the example more numerically specific, suppose that there are four possible states given in the middle four columns of the following table, in which the occurrence of event is signified by a in row and its non-occurrence is signified by a and likewise for and That is, and The probability of is for every

Event Probability of event
0 1 0 1
0 0 1 1
0 1 1 1

and so

Event Probability of event
0 0 0 1
0 1 0 1
0 0 1 1
0 0 0 1

In this example, occurs if and only if at least one of occurs. Unconditionally (that is, without reference to ), and are independent of each other because —the sum of the probabilities associated with a in row —is while But conditional on having occurred (the last three columns in the table), we have while Since in the presence of the probability of is affected by the presence or absence of and are mutually dependent conditional on

See also

[edit]

References

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
In , conditional dependence describes a relationship between two or more random variables or where their statistical dependence persists or emerges even after accounting for the influence of one or more conditioning variables. Specifically, for random variables X and Y given Z, conditional dependence holds if the P(X | Y, Z) differs from P(X | Z) for some values where P(Y, Z) > 0, meaning knowledge of Y provides additional about X beyond what Z alone offers. This contrasts with unconditional dependence, where P(X, Y) ≠ P(X)P(Y), and can manifest in scenarios where variables appear independent marginally but become dependent upon conditioning, or vice versa, as seen in examples like indicators (e.g., and bacterial infection) that are independent overall but dependent given the presence of fever. Conditional dependence plays a central role in probabilistic modeling, particularly in graphical models such as Bayesian networks, where it helps encode complex joint distributions through conditional relationships and d-separation criteria to identify independencies. In and , measuring conditional dependence is essential for tasks like , where irrelevant variables are screened out given others, and causal discovery, which distinguishes direct effects from spurious correlations. Various metrics have been developed to quantify it, including kernel-based approaches using reproducing kernel Hilbert spaces for non-linear dependencies and simple coefficients based on for practical computation in high dimensions. These concepts underpin advancements in , enabling efficient in large-scale systems by exploiting conditional structures to reduce .

Core Concepts

Definition

Conditional dependence refers to a relationship between random variables or events where the of one is influenced by the other, even after incorporating information from a conditioning variable or set. Intuitively, it arises when knowing the outcome of one variable alters the expected behavior of another, despite accounting for the conditioning factor, reflecting a residual association not explained by the conditioner alone. Formally, two random variables XX and YY are conditionally dependent given a third variable ZZ (with P(Z=z)>0P(Z = z) > 0) if there exist values x,y,zx, y, z in their supports such that P(X=x,Y=yZ=z)P(X=xZ=z)P(Y=yZ=z).P(X = x, Y = y \mid Z = z) \neq P(X = x \mid Z = z) \, P(Y = y \mid Z = z). This inequality indicates that the joint conditional distribution does not factorize into the product of the marginal conditionals, signifying dependence. Unlike unconditional (marginal) dependence, which assesses association without conditioning, conditional dependence can emerge or disappear based on the conditioner; notably, XX and YY may be unconditionally independent yet conditionally dependent given ZZ, as in collider bias where ZZ is a common effect of XX and YY, inducing spurious association upon conditioning. Conversely, unconditional dependence may vanish under certain conditioning, highlighting the context-specific nature of probabilistic relationships. The concept was first formalized within modern in the early 20th century, building on Andrei Kolmogorov's axiomatic foundations established in 1933, which provided the rigorous framework for conditional probabilities underlying dependence relations.

Relation to Unconditional Dependence

Unconditional dependence between two random variables XX and YY occurs when their does not factorize into the product of their marginal distributions, that is, when P(X,Y)P(X)P(Y)P(X, Y) \neq P(X) P(Y). This contrasts with conditional dependence, which, as defined earlier, evaluates the joint distribution relative to a conditioning variable ZZ. In essence, unconditional dependence captures marginal associations without additional context, while conditional dependence reveals how these associations may alter given knowledge of ZZ. Conditioning on ZZ can induce conditional independence from unconditional dependence, particularly in scenarios involving a . For instance, if ZZ directly influences both XX and YY (as in a where arrows point from ZZ to XX and from ZZ to YY), XX and YY exhibit unconditional dependence due to their shared origin, but become conditionally independent given ZZ, as the influence of the is accounted for. This structure, known as a or , illustrates how conditioning removes spurious associations propagated through ZZ. Conversely, conditioning can induce conditional dependence where unconditional previously held, a phenomenon exemplified by the V-structure in directed acyclic graphs. In a V-structure, arrows converge on ZZ from both XX and YY (i.e., XZYX \to Z \leftarrow Y), rendering XX and YY unconditionally independent since they lack a direct path of influence. However, conditioning on ZZ—the common effect—creates a dependence between XX and YY, as observing ZZ provides evidence that selects paths linking the two causes through the at ZZ. This is the basis for "explaining away," where evidence for one cause (say, XX) reduces the likelihood of the alternative cause (YY) given the observed effect ZZ, thereby inducing negative conditional dependence between the causes. Overall, conditioning on ZZ can thus create new dependencies, remove existing ones, or even invert the direction of association between XX and YY, fundamentally altering the dependence structure depending on the underlying causal relationships. These dynamics underscore the importance of graphical models like directed acyclic graphs in visualizing how marginal and conditional dependencies interact.

Formal Framework

Probabilistic Formulation

In , conditional dependence between two events AA and BB given a third event CC with P(C)>0P(C) > 0 is defined as the failure of the equality P(ABC)P(AC)P(BC)P(A \cap B \mid C) \neq P(A \mid C) P(B \mid C), where the is given by P(AC)=P(AC)/P(C)P(A \mid C) = P(A \cap C)/P(C). This inequality indicates that the occurrence of AA affects the probability of BB (or vice versa) even after accounting for CC. For random variables, consider random variables XX, YY, and ZZ defined on a probability space. The joint conditional probability mass or density function encapsulates the probabilistic structure. Specifically, the joint conditional distribution satisfies P(X,YZ)=P(XY,Z)P(YZ)P(X, Y \mid Z) = P(X \mid Y, Z) P(Y \mid Z), derived from the chain rule for conditional probabilities: starting from the joint distribution P(X,Y,Z)=P(XY,Z)P(Y,Z)=P(XY,Z)P(YZ)P(Z)P(X, Y, Z) = P(X \mid Y, Z) P(Y, Z) = P(X \mid Y, Z) P(Y \mid Z) P(Z), dividing by P(Z)P(Z) yields the conditional form, assuming P(Z)>0P(Z) > 0. Conditional dependence holds when this factorization does not imply P(XY,Z)=P(XZ)P(X \mid Y, Z) = P(X \mid Z), i.e., when P(X,YZ)P(XZ)P(YZ)P(X, Y \mid Z) \neq P(X \mid Z) P(Y \mid Z). Unconditional dependence arises as the special case where ZZ is a constant event with probability 1. In the discrete case, for random variables taking values in countable sets, the conditional joint probability mass function is pX,YZ(x,yz)=pX,Y,Z(x,y,z)/pZ(z)p_{X,Y \mid Z}(x,y \mid z) = p_{X,Y,Z}(x,y,z) / p_Z(z) for pZ(z)>0p_Z(z) > 0, and the marginal conditionals are pXZ(xz)=ypX,YZ(x,yz)p_{X \mid Z}(x \mid z) = \sum_y p_{X,Y \mid Z}(x,y \mid z) and similarly for YY. Dependence occurs if pX,YZ(x,yz)pXZ(xz)pYZ(yz)p_{X,Y \mid Z}(x,y \mid z) \neq p_{X \mid Z}(x \mid z) p_{Y \mid Z}(y \mid z) for some x,y,zx, y, z with pZ(z)>0p_Z(z) > 0. For continuous random variables with joint density fX,Y,Zf_{X,Y,Z}, the conditional joint density is fX,YZ(x,yz)=fX,Y,Z(x,y,z)/fZ(z)f_{X,Y \mid Z}(x,y \mid z) = f_{X,Y,Z}(x,y,z) / f_Z(z) for fZ(z)>0f_Z(z) > 0, with marginal conditionals fXZ(xz)=fX,YZ(x,yz)dyf_{X \mid Z}(x \mid z) = \int f_{X,Y \mid Z}(x,y \mid z) \, dy and analogously for YY. Conditional dependence is present when fX,YZ(x,yz)fXZ(xz)fYZ(yz)f_{X,Y \mid Z}(x,y \mid z) \neq f_{X \mid Z}(x \mid z) f_{Y \mid Z}(y \mid z) for some x,y,zx, y, z with fZ(z)>0f_Z(z) > 0. From an axiomatic perspective in measure-theoretic probability, conditional dependence is framed using sigma-algebras. Let (Ω,F,P)(\Omega, \mathcal{F}, P) be a , and let σ(X)\sigma(X), σ(Y)\sigma(Y), σ(Z)\sigma(Z) be the sigma-algebras generated by measurable functions XX, YY, Z:ΩRZ: \Omega \to \mathbb{R}, respectively. The random variables XX and YY are conditionally dependent given ZZ if σ(X)\sigma(X) and σ(Y)\sigma(Y) are not conditionally independent given σ(Z)\sigma(Z), meaning there exist events Aσ(X)A \in \sigma(X), Bσ(Y)B \in \sigma(Y) such that P(ABσ(Z))P(Aσ(Z))P(Bσ(Z))P(A \cap B \mid \sigma(Z)) \neq P(A \mid \sigma(Z)) P(B \mid \sigma(Z)) on a set of positive probability, where conditional probability given a sigma-algebra is defined via the Radon-Nikodym derivative of the restricted measures. Equivalently, for bounded measurable functions ff on the range of XX and gg on the range of YY, E[f(X)g(Y)σ(Z)]E[f(X)σ(Z)]E[g(Y)σ(Z)]E[f(X) g(Y) \mid \sigma(Z)] \neq E[f(X) \mid \sigma(Z)] E[g(Y) \mid \sigma(Z)] . This setup ensures the formulation aligns with Kolmogorov's axioms extended to conditional expectations.

Measure of Conditional Dependence

One prominent measure of conditional dependence is the , denoted I(X;YZ)I(X; Y \mid Z), which quantifies the amount of information shared between random variables XX and YY after conditioning on ZZ. Defined in terms of entropies as I(X;YZ)=H(XZ)+H(YZ)H(X,YZ)I(X; Y \mid Z) = H(X \mid Z) + H(Y \mid Z) - H(X, Y \mid Z), where H(XZ)H(X \mid Z) is the of XX given ZZ measuring the remaining in XX after observing ZZ, and similarly for the other terms, this metric captures the expected reduction in uncertainty about one variable from knowing the other, conditional on ZZ. It equals zero if and only if XX and YY are conditionally independent given ZZ, providing a symmetric, non-negative measure applicable to both discrete and continuous variables without assuming . For jointly Gaussian random variables, partial correlation offers a computationally efficient alternative, measuring the between XX and YY after removing the linear effects of ZZ. The coefficient is given by ρXYZ=ρXYρXZρYZ(1ρXZ2)(1ρYZ2),\rho_{XY \cdot Z} = \frac{\rho_{XY} - \rho_{XZ} \rho_{YZ}}{\sqrt{(1 - \rho_{XZ}^2)(1 - \rho_{YZ}^2)}},
Add your contribution
Related Hubs
User Avatar
No comments yet.