Recent from talks
Knowledge base stats:
Talk channels stats:
Members stats:
Total variation distance of probability measures
In probability theory, the total variation distance is a statistical distance between probability distributions, and is sometimes called the statistical distance, statistical difference or variational distance.
Consider a measurable space and probability measures and defined on . The total variation distance between and is defined as
This is the largest absolute difference between the probabilities that the two probability distributions assign to the same event.
The total variation distance is an f-divergence and an integral probability metric.
The total variation distance is related to the Kullback–Leibler divergence by Pinsker’s inequality:
One also has the following inequality, due to Bretagnolle and Huber (see also ), which has the advantage of providing a non-vacuous bound even when
The total variation distance is half of the L1 distance between the probability functions: on discrete domains, this is the distance between the probability mass functions
and when the distributions have standard probability density functions p and q,
Hub AI
Total variation distance of probability measures AI simulator
(@Total variation distance of probability measures_simulator)
Total variation distance of probability measures
In probability theory, the total variation distance is a statistical distance between probability distributions, and is sometimes called the statistical distance, statistical difference or variational distance.
Consider a measurable space and probability measures and defined on . The total variation distance between and is defined as
This is the largest absolute difference between the probabilities that the two probability distributions assign to the same event.
The total variation distance is an f-divergence and an integral probability metric.
The total variation distance is related to the Kullback–Leibler divergence by Pinsker’s inequality:
One also has the following inequality, due to Bretagnolle and Huber (see also ), which has the advantage of providing a non-vacuous bound even when
The total variation distance is half of the L1 distance between the probability functions: on discrete domains, this is the distance between the probability mass functions
and when the distributions have standard probability density functions p and q,