Hubbry Logo
Arithmetic meanArithmetic meanMain
Open search
Arithmetic mean
Community hub
Arithmetic mean
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Arithmetic mean
Arithmetic mean
from Wikipedia

In mathematics and statistics, the arithmetic mean ( /ˌærɪθˈmɛtɪk/ arr-ith-MET-ik), arithmetic average, or just the mean or average is the sum of a collection of numbers divided by the count of numbers in the collection.[1] The collection is often a set of results from an experiment, an observational study, or a survey. The term "arithmetic mean" is preferred in some contexts in mathematics and statistics because it helps to distinguish it from other types of means, such as geometric and harmonic.

Arithmetic means are also frequently used in economics, anthropology, history, and almost every other academic field to some extent. For example, per capita income is the arithmetic average of the income of a nation's population.

While the arithmetic mean is often used to report central tendencies, it is not a robust statistic: it is greatly influenced by outliers (values much larger or smaller than most others). For skewed distributions, such as the distribution of income for which a few people's incomes are substantially higher than most people's, the arithmetic mean may not coincide with one's notion of "middle". In that case, robust statistics, such as the median, may provide a better description of central tendency.

Definition

[edit]

The arithmetic mean of a set of observed data is equal to the sum of the numerical values of each observation, divided by the total number of observations. Symbolically, for a data set consisting of the values , the arithmetic mean is defined by the formula:[2]

In simpler terms, the formula for the arithmetic mean is:

For example, if the monthly salaries of employees are , then the arithmetic mean is:

Example
Person Salary
A 2500
B 2700
C 2300
D 2650
E 2450
Average 2520

If the data set is a statistical population (i.e. consists of every possible observation and not just a subset of them), then the mean of that population is called the population mean and denoted by the Greek letter . If the data set is a statistical sample (a subset of the population), it is called the sample mean (which for a data set is denoted as ).

The arithmetic mean can be similarly defined for vectors in multiple dimensions, not only scalar values; this is often referred to as a centroid. More generally, because the arithmetic mean is a convex combination (meaning its coefficients sum to ), it can be defined on a convex space, not only a vector space.

History

[edit]

Statistician Churchill Eisenhart, senior researcher fellow at the U. S. National Bureau of Standards, traced the history of the arithmetic mean in detail. In the modern age, it started to be used as a way of combining various observations that should be identical, but were not such as estimates of the direction of magnetic north. In 1635, mathematician Henry Gellibrand described as "meane" the midpoint of a lowest and highest number, not quite the arithmetic mean. In 1668, a person known as "D. B." was quoted in the Transactions of the Royal Society describing "taking the mean" of five values:[3]

In this Table, he [Capt. Sturmy] notes the greatest difference to be 14 minutes; and so taking the mean for the true Variation, he concludes it then and there to be just 1. deg. 27. min.

— D.B., p. 726

Motivating properties

[edit]

The arithmetic mean has several properties that make it interesting, especially as a measure of central tendency. These include:

  • If numbers have a mean , then . Since is the distance from a given number to the mean, one way to interpret this property is by saying that the numbers to the left of the mean are balanced by the numbers to the right. The mean is the only number for which the residuals (deviations from the estimate) sum to zero. This can also be interpreted as saying that the mean is translationally invariant in the sense that for any real number , .
  • If it is required to use a single number as a "typical" value for a set of known numbers , then the arithmetic mean of the numbers does this best since it minimizes the sum of squared deviations from the typical value: the sum of . The sample mean is also the best single predictor because it has the lowest root mean squared error.[4] If the arithmetic mean of a population of numbers is desired, then the estimate of it that is unbiased is the arithmetic mean of a sample drawn from the population.
  • The arithmetic mean is independent of scale of the units of measurement, in the sense that So, for example, calculating a mean of liters and then converting to gallons is the same as converting to gallons first and then calculating the mean. This is also called first order homogeneity.

Additional properties

[edit]
  • The arithmetic mean of a sample is always between the largest and smallest values in that sample.
  • The arithmetic mean of any amount of equal-sized number groups together is the arithmetic mean of the arithmetic means of each group.

Contrast with median

[edit]

The arithmetic mean differs from the median, which is the value that separates the higher half and lower half of a data set. When the values in a data set form an arithmetic progression, the median and arithmetic mean are equal. For example, in the data set , both the mean and median are .

In other cases, the mean and median can differ significantly. For instance, in the data set , the arithmetic mean is , while the median is . This occurs because the mean is sensitive to extreme values and may not accurately reflect the central tendency of most data points.

This distinction has practical implications across different fields. For example, since the 1980s, the median income in the United States has increased at a slower rate than the arithmetic mean income.[5]

Similarly, in climate studies, daily mean temperature distributions tend to approximate a normal distribution, whereas annual or monthly rainfall totals often display a skewed distribution, with some periods having unusually high totals while most have relatively low amounts. In such cases, the median can provide a more representative measure of central tendency.[6]

Generalizations

[edit]

Weighted average

[edit]

A weighted average, or weighted mean, is an average in which some data points count more heavily than others in that they are given more weight in the calculation.[7] For example, the arithmetic mean of and is , or equivalently . In contrast, a weighted mean in which the first number receives, for example, twice as much weight as the second (perhaps because it is assumed to appear twice as often in the general population from which these numbers were sampled) would be calculated as . Here the weights, which necessarily sum to one, are and , the former being twice the latter. The arithmetic mean (sometimes called the "unweighted average" or "equally weighted average") can be interpreted as a special case of a weighted average in which all weights are equal to the same number ( in the above example and in a situation with numbers being averaged).

Functions

[edit]
In calculus, and especially multivariable calculus, the mean of a function is loosely defined as the ”average" value of the function over its domain.

Continuous probability distributions

[edit]
Comparison of two log-normal distributions with equal median, but different skewness, resulting in various means and modes

If a numerical property, and any sample of data from it, can take on any value from a continuous range instead of, for example, just integers, then the probability of a number falling into some range of possible values can be described by integrating a continuous probability distribution across this range, even when the naive probability for a sample number taking one certain value from infinitely many is zero. In this context, the analog of a weighted average, in which there are infinitely many possibilities for the precise value of the variable in each range, is called the mean of the probability distribution. The most widely encountered probability distribution is called the normal distribution; it has the property that all measures of its central tendency, including not just the mean but also the median mentioned above and the mode (the three Ms[8]), are equal. This equality does not hold for other probability distributions, as illustrated for the log-normal distribution here.

Angles

[edit]

Particular care is needed when using cyclic data, such as phases or angles. Taking the arithmetic mean of 1° and 359° yields a result of 180°. This is incorrect for two reasons:

  1. Angle measurements are only defined up to an additive constant of 360° ( or , if measuring in radians). Thus, these could easily be called 1° and -1°, or 361° and 719°, since each one of them produces a different average.
  2. In this situation, 0° (or 360°) is geometrically a better average value: there is lower dispersion about it (the points are both 1° from it and 179° from 180°, the putative average).

In general application, such an oversight will lead to the average value artificially moving towards the middle of the numerical range. A solution to this problem is to use the optimization formulation (that is, define the mean as the central point: the point about which one has the lowest dispersion) and redefine the difference as a modular distance (i.e. the distance on the circle: so the modular distance between 1° and 359° is 2°, not 358°).

Proof without words of the AM–GM inequality:
PR is the diameter of a circle centered on O; its radius AO is the arithmetic mean of a and b. Triangle PGR is a right triangle from Thales's theorem, enabling use of the geometric mean theorem to show that its altitude GQ is the geometric mean. For any ratio a:b, AO ≥ GQ.

Symbols and encoding

[edit]

The arithmetic mean is often denoted by a bar (vinculum or macron), as in .[4]

See also

[edit]

Notes

[edit]

References

[edit]

Further reading

[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
The arithmetic mean, also known as the , is a fundamental measure of in and statistics, calculated as the sum of a set of numerical values divided by the number of values in the set. For a finite of nn numbers x1,x2,,xnx_1, x_2, \dots, x_n, it is expressed by the xˉ=1ni=1nxi\bar{x} = \frac{1}{n} \sum_{i=1}^n x_i, where the result represents a typical or central value within the . This simple yet powerful provides a balanced summary of data, assuming equal importance for each value, and is distinct from other means like the geometric or , which handle multiplicative or rate-based data differently. In , the arithmetic mean serves as an for the population known as the , making it essential for descriptive and inferential analyses across disciplines such as , physics, and social sciences. It possesses several key mathematical properties that enhance its utility: the mean always lies between the minimum and maximum values of the (inclusive of equality in trivial cases); the sum of the deviations of each value from the mean equals zero; and it utilizes all points, providing a complete representation of the set. However, its sensitivity to extreme values (outliers) can skew results in non-symmetric distributions, prompting the use of alternatives like the in such scenarios. These properties stem from its algebraic foundation, allowing for straightforward computation and integration into more complex models, such as weighted means where values have varying importance. The concept of the arithmetic mean traces its roots to ancient mathematical practices, with systematic exploration emerging in Greek antiquity through studies of proportions and ratios, though its formal adoption as a statistical tool gained prominence in the amid debates on and averaging techniques. Early astronomers and surveyors, including figures like and Thomas Simpson, refined its application for reducing observational errors, establishing it as a of modern despite initial skepticism regarding its representativeness in uneven datasets. Today, it remains ubiquitous in computational algorithms, , and everyday , underscoring its enduring relevance in quantifying averages and trends.

Fundamentals

Definition

The arithmetic mean, commonly referred to as the mean or average, is a fundamental measure of central tendency in statistics and mathematics, defined as the sum of a finite set of numerical values divided by the number of values in the set. It provides a single value that summarizes the "center" of the data and is applicable to any finite collection of real numbers, assuming no additional weighting is applied. For a set of nn numbers x1,x2,,xnx_1, x_2, \dots, x_n, the unweighted arithmetic mean xˉ\bar{x} is calculated using the formula xˉ=1ni=1nxi,\bar{x} = \frac{1}{n} \sum_{i=1}^n x_i, where i=1nxi\sum_{i=1}^n x_i represents the summation of the values (the total obtained by adding them together). This formula assumes a basic understanding of summation as the process of adding multiple terms. For instance, consider the numbers 2, 4, 4, 4, 8, 10: their sum is 32, and with n=6n = 6, the arithmetic mean is 326=1635.33\frac{32}{6} = \frac{16}{3} \approx 5.33. In statistical contexts, a distinction is made between the population mean μ\mu, which is the arithmetic mean of all elements in an entire population, and the sample mean xˉ\bar{x}, which is the arithmetic mean computed from a subset (sample) of the population used to estimate μ\mu. This differentiation is crucial for inferential , where the sample mean serves as an estimator for the unknown population parameter.

Calculation

The arithmetic mean, denoted as xˉ\bar{x}, of a finite set of numbers x1,x2,,xnx_1, x_2, \dots, x_n where n>0n > 0 is computed by first calculating the sum S=i=1nxiS = \sum_{i=1}^n x_i and then dividing by the number of observations nn: xˉ=1ni=1nxi=Sn.\bar{x} = \frac{1}{n} \sum_{i=1}^n x_i = \frac{S}{n}. This process involves iterating through the dataset once to accumulate the sum, followed by a single division operation. For a simple example with three values, consider the numbers 2, 4, and 6. The sum is S=2+4+6=12S = 2 + 4 + 6 = 12, and dividing by n=3n = 3 gives xˉ=12/3=4\bar{x} = 12 / 3 = 4. For larger datasets, the same procedure applies but may benefit from organized presentation. Consider the following table of 10 temperature readings in degrees :
IndexValue
122.5
224.1
321.8
423.0
525.2
622.9
723.7
824.5
921.3
1022.8
The sum is S=231.8S = 231.8, and with n=10n = 10, the is xˉ=231.8/10=23.18\bar{x} = 231.8 / 10 = 23.18. Another example demonstrates the relationship between the mean, the number of observations, and the total sum. In a class of 40 students where the average number of books read by each student is 7, the total number of books read by all students is 40×7=28040 \times 7 = 280. This shows that the sum of the values equals the arithmetic mean multiplied by the number of observations (sum = mean × n). Another example illustrates efficient recalculation when the dataset changes. For instance, consider a scenario where the average weight of 49 students is 39 kg. Seven students with an average weight of 40 kg leave, and seven new students with an average weight of 54 kg join. The new average can be calculated efficiently: the net weight gain is 7×(5440)=987 \times (54 - 40) = 98 kg, so the increase in average is 98/49=298 / 49 = 2 kg, resulting in a new average of 39+2=4139 + 2 = 41 kg. In computational practice, the direct method has a of O(n)O(n), as it performs a linear pass over the for additions and a constant-time division. For large datasets, iterative accumulation can help manage memory and intermediate results, but care is needed to avoid overflow in fixed-precision arithmetic. One numerically approach uses recursive updating starting from the first value: initialize xˉ1=x1\bar{x}_1 = x_1, then for each subsequent k=2k = 2 to nn, update xˉk=xˉk1+xkxˉk1k\bar{x}_k = \bar{x}_{k-1} + \frac{x_k - \bar{x}_{k-1}}{k}. This method centers updates around the current estimate, reducing the magnitude of additions and mitigating rounding errors when values are clustered. Additionally, in , rounding errors can accumulate during . Pairwise summation mitigates this by recursively summing pairs of numbers (e.g., sum adjacent pairs, then sum those results pairwise, and so on), bounding the error growth to O(logn)O(\log n) times the unit roundoff rather than O(n)O(n). This method is particularly useful for high-precision requirements. Edge cases require special handling. For a single value (n=1n=1), the mean is the value itself: xˉ=x1\bar{x} = x_1. If all values are zero, the mean is zero. However, the mean of an (n=0n=0) is undefined, as it involves .

Properties

Motivating properties

The arithmetic mean possesses several intuitive properties that make it a natural choice for summarizing the of a , particularly when equal importance is assigned to each observation. One key motivating property is its additivity, which states that the mean of the sum of two or more equals the sum of their individual s, scaled appropriately by the number of observations. For instance, if a is partitioned into subsets, the overall can be computed as a weighted of the subset s, facilitating efficient calculations for large or divided . This property is particularly useful in aggregating from multiple sources without recomputing from scratch. Another foundational attribute is the arithmetic mean's , which ensures that the mean of a of random variables is the corresponding of their means: aX+bY=aX+bY\overline{aX + bY} = a\overline{X} + b\overline{Y}, where aa and bb are constants. This underpins its compatibility with , such as , where predictions and parameter estimates rely on averaging transformed while preserving structural relationships. It motivates the mean's role in modeling additive processes, like forecasting totals from component averages in or . A compelling reason for preferring the arithmetic mean arises from its optimality in minimizing the sum of squared deviations from the data points. Consider a constant model μ^\hat{\mu} estimating a fixed value for all observations x1,x2,,xnx_1, x_2, \dots, x_n; the value of μ^\hat{\mu} that minimizes i=1n(xiμ^)2\sum_{i=1}^n (x_i - \hat{\mu})^2 is precisely the arithmetic mean xˉ=1ni=1nxi\bar{x} = \frac{1}{n} \sum_{i=1}^n x_i. To see this, expand the sum: (xiμ^)2=(xixˉ+xˉμ^)2=(xixˉ)2+2(xˉμ^)(xixˉ)+n(xˉμ^)2\sum (x_i - \hat{\mu})^2 = \sum (x_i - \bar{x} + \bar{x} - \hat{\mu})^2 = \sum (x_i - \bar{x})^2 + 2(\bar{x} - \hat{\mu})\sum (x_i - \bar{x}) + n(\bar{x} - \hat{\mu})^2. The cross-term vanishes because (xixˉ)=0\sum (x_i - \bar{x}) = 0, leaving (xixˉ)2+n(xˉμ^)2\sum (x_i - \bar{x})^2 + n(\bar{x} - \hat{\mu})^2, which is minimized at μ^=xˉ\hat{\mu} = \bar{x} since the second term is nonnegative and zero only when μ^=xˉ\hat{\mu} = \bar{x}. This least-squares property positions the mean as the best constant predictor under squared error loss, a criterion central to many statistical applications. In symmetric distributions, the arithmetic mean further justifies its use through its alignment with the concept of a balance point, analogous to the center of mass in physics. For a set of masses at positions xix_i, the center of mass xˉ=miximi\bar{x} = \frac{\sum m_i x_i}{\sum m_i} reduces to the arithmetic mean when masses are equal (mi=1m_i = 1), representing the point where the dataset is equilibrated. This ensures the is an unbiased of the , as deviations above and below cancel out on average, providing a stable measure of without directional bias. Such properties make it ideal for symmetric data in fields like physics and . Practically, these motivate the arithmetic mean's widespread application in averaging errors or predictions under assumptions of equal weighting. In error analysis, it computes the average deviation to assess model performance, as squared errors emphasize larger discrepancies while the mean provides an interpretable summary. Similarly, in predictive modeling like ensemble methods, averaging forecasts from multiple models reduces variance and improves accuracy, leveraging the mean's additivity and least-squares efficiency for reliable point estimates.

Additional properties

The arithmetic mean exhibits as an aggregation function, meaning that applying the operation twice to a yields the same result as applying it once: if Xˉ\bar{X} denotes the arithmetic mean of the values in XX, then the arithmetic mean of {Xˉ,Xˉ,,Xˉ}\{\bar{X}, \bar{X}, \dots, \bar{X}\} (with nn copies) is again Xˉ\bar{X}. For x1,x2,,xn>0x_1, x_2, \dots, x_n > 0, the arithmetic mean satisfies the AM-GM-HM inequality : the arithmetic mean (AM) is at least the (GM), which is at least the (HM), i.e., AMGMHM\mathrm{AM} \geq \mathrm{GM} \geq \mathrm{HM}, with equality if and only if all xix_i are equal. This relationship follows from the convexity of the logarithmic function in the proof of AM ≥ GM (via ) and a similar argument for GM ≥ HM using the reciprocal function. As a convex combination of the input values with equal weights 1/n1/n, the arithmetic mean preserves the bounds of the dataset: for real numbers x1x2xnx_1 \leq x_2 \leq \dots \leq x_n, it holds that minixiXˉmaxixi\min_i x_i \leq \bar{X} \leq \max_i x_i. The arithmetic mean is particularly sensitive to outliers, as a single extreme value disproportionately influences the overall average due to its linear weighting of all observations. This contrasts with more robust measures like the median. Quantitatively, the variance of the sample mean Xˉ\bar{X} from an independent random sample of size nn drawn from a population with variance σ2\sigma^2 is Var(Xˉ)=σ2/n\mathrm{Var}(\bar{X}) = \sigma^2 / n, which decreases with larger nn and underscores the mean's stability under repeated sampling but vulnerability to skewed data. A key mathematical property is that the arithmetic mean minimizes the sum of squared deviations from the data points. To derive this, consider the objective function S(μ)=i=1n(xiμ)2.S(\mu) = \sum_{i=1}^n (x_i - \mu)^2. Differentiating with respect to μ\mu gives dSdμ=2i=1n(xiμ)=0,\frac{dS}{d\mu} = -2 \sum_{i=1}^n (x_i - \mu) = 0, which simplifies to i=1nxi=nμ\sum_{i=1}^n x_i = n\mu, so μ=Xˉ\mu = \bar{X}. The second d2Sdμ2=2n>0\frac{d^2S}{d\mu^2} = 2n > 0 confirms a minimum. Alternatively, expanding S(y^)S(\hat{y}) for any estimate y^\hat{y} yields S(y^)=(xiXˉ)2+n(y^Xˉ)2(xiXˉ)2=S(Xˉ)S(\hat{y}) = \sum (x_i - \bar{X})^2 + n(\hat{y} - \bar{X})^2 \geq \sum (x_i - \bar{X})^2 = S(\bar{X}), with equality only if y^=Xˉ\hat{y} = \bar{X}.

Historical Context

Early origins

The concept of the arithmetic mean emerged in ancient civilizations through practical applications in astronomy, , and theoretical , often without formal mathematical notation. In around 2000 BCE, astronomers calculated mean positions of celestial bodies to predict movements, employing computed mean values, such as the mean lunar month of 29;30,30 days (approximately 29.53 days), based on long-term observations of variations between 29 and 30 days. These computations, recorded on clay tablets, employed and arithmetic progressions to approximate planetary and lunar positions, enabling long-term calendars and predictions. In , circa 1650 BCE, the demonstrates implicit use of averaging in problems, such as dividing loaves of bread or measures of among workers using unit fractions and proportional shares for fair allocation in labor or contexts. This approach supported administrative tasks in and , where equitable division of supplies was essential. Greek thinkers further conceptualized the arithmetic mean in both musical theory and . The Pythagoreans, around the 6th century BCE, applied numerical averages to harmonics, identifying the arithmetic mean as one of three classical means (alongside geometric and ) to explain intervals in music; for example, they related string lengths to frequency ratios like 2:1 for octaves, using averages to harmonize scales. In , (4th century BCE) distinguished the "mean according to arithmetic proportion" as a fixed midpoint—such as 6 between 10 and 2—in his from the Nicomachean Ethics, advocating virtue as an intermediate state between excess and deficiency, though relative to individual circumstances rather than strict arithmetic equality. During the medieval Islamic period, scholars like (9th century CE) integrated averages into practical computations for and astronomy. In his treatise Kitab al-Jabr wa'l-Muqabala, al-Khwarizmi addressed problems by dividing estates proportionally among heirs using algebraic methods to resolve Qur'anic rules for complex family distributions. His astronomical work, Zij al-Sindhind, included tables of mean motions for planetary positions to refine calendars and almanacs. Roman agricultural practices also relied on averaging for yield estimation, as detailed by Lucius Junius Moderatus Columella in De Re Rustica (1st century CE). Columella recommended assessing average crop outputs over multiple seasons to guide farm management; for , he cited typical yields of 10-15 modii per iugerum (about 6-9 bushels per acre) on good , derived from observational averages to optimize planting and labor. These estimates emerged from empirical and measurement needs, where merchants and farmers averaged quantities of goods like or wine to standardize exchanges without precise notation.

Formal development

The formal development of the arithmetic mean as a rigorous mathematical and statistical concept began in the 17th and 18th centuries with foundational work in probability and analysis. In 1713, Jacob Bernoulli's posthumously published Ars Conjectandi introduced the weak law of large numbers, demonstrating that the arithmetic mean of a large number of independent Bernoulli trials converges in probability to the expected value, thereby establishing averaging as a principled method for estimating probabilities in repeated experiments. In the early 18th century, Roger Cotes discussed the arithmetic mean in the context of error analysis in his posthumous Opera Miscellanea (1722). Later, Thomas Simpson's 1755 treatise explicitly advocated taking the arithmetic mean of multiple observations to minimize errors in astronomical measurements, influencing its use in probability and statistics. Later in the 18th century, Leonhard Euler advanced the notation and theoretical framework for summation in his 1755 treatise Institutiones calculi differentialis, where he introduced the sigma symbol (Σ) to denote sums, facilitating the precise expression of the arithmetic mean as the total sum divided by the number of terms in analytical contexts. The 19th century saw the arithmetic mean integrated into statistical estimation and probabilistic theory. In 1809, Carl Friedrich Gauss's Theoria Motus Corporum Coelestium formalized the method of , proving that under the assumption of normally distributed errors, the arithmetic mean serves as the maximum likelihood estimator for the true value, marking a pivotal shift toward its use as an optimal statistical estimator in astronomy and beyond. Shortly thereafter, in 1810, Pierre-Simon Laplace's memoir on probability extended this by proving an early version of the , showing that the distribution of the sum of independent random variables approximates a , which implies that the arithmetic mean of sufficiently large samples tends to follow a centered on the population mean. By the , the arithmetic mean achieved standardization in statistical practice and education. William Sealy Gosset's 1908 paper "The Probable Error of a Mean," published under the pseudonym "," introduced the t-test for inferring population means from small samples, embedding the arithmetic mean centrally in hypothesis testing procedures for comparing group averages. Ronald Fisher's influential textbook Statistical Methods for Research Workers further codified its role, presenting the arithmetic mean alongside variance and other measures in accessible tables and methods for experimental design, promoting its widespread adoption in biological and social sciences. This progression culminated in a transition from manual, calculations to computational tools, enabling efficient computation of arithmetic means in large datasets. During the 1920s and 1930s, mechanical tabulating machines from facilitated of sums and averages in statistical bureaus, while post-World War II electronic computers and software like SAS (introduced in 1966) automated mean calculations, integrating them into modern workflows.

Comparisons with Other Measures

Contrast with median

The median, in contrast to the arithmetic mean, is defined as the middle value in a dataset when the observations are ordered from smallest to largest; for an even number of observations, it is the average of the two central values. This measure represents the point that divides the data into two equal halves, providing a robust indicator of without relying on all values equally. A key distinction lies in their sensitivity to outliers: the arithmetic mean can be heavily influenced by extreme values, as it incorporates every observation proportionally, whereas the median remains unaffected by values beyond the central position. For instance, in the dataset {1, 2, 3, 100}, the arithmetic mean is 26.5, pulled upward by the outlier, while the median is 2.5, better reflecting the cluster of smaller values. This sensitivity often leads to the mean exceeding the median in datasets with positive outliers, such as income distributions where wealth inequality results in a few high earners distorting the average. The choice between the two depends on data symmetry and distribution shape. In symmetric distributions, such as human heights approximating a , the mean and median coincide, making the mean preferable for its additional properties like additivity. However, for skewed distributions like house prices, where a few luxury properties inflate the mean, the median provides a more representative "typical" value. In a , which models such positive skew (e.g., certain biological or financial data), the mean exceeds the median due to the right tail. From a statistical perspective, the arithmetic mean is the maximum likelihood and asymptotically efficient under parametric assumptions like normality, minimizing variance among unbiased . In contrast, the serves as the maximum likelihood for the and is preferred in non-parametric settings or robust analyses, where it resists outliers and requires fewer distributional assumptions. This makes the particularly valuable when data may violate normality, ensuring more reliable inferences in skewed or contaminated samples.

Contrast with mode

The mode of a dataset is defined as the value that appears most frequently, serving as a measure of central tendency that identifies the peak or peaks in the distribution of . In contrast, the arithmetic mean treats all values equally by summing them and dividing by the number of observations, providing a balanced summary that incorporates every data point without emphasis on frequency. This fundamental difference means the mean reflects the overall "center of mass" of the data, while the mode highlights concentrations or modal values, particularly useful in multimodal distributions where multiple peaks exist. The arithmetic mean is most applicable to quantitative on interval or scales, such as calculating the test score in a class (e.g., scores of 85, 90, 92, and 85 yield a of 88), where numerical averaging provides meaningful insight. Conversely, the mode excels with categorical or nominal , identifying the most common category, like the most frequent in a (e.g., appearing 15 times out of 50 observations). In scenarios involving discrete counts or preferences, the mode captures typical occurrences that the might obscure, as averaging categories lacks interpretive value. A key limitation of the mode is that it may not exist if all values are unique or appear equally often, or it may not be unique in bimodal or multimodal datasets, leading to in representation. For instance, in the {1, 1, 2, 3}, the mode is 1 due to its highest frequency, while the arithmetic mean is $1.75,calculatedas, calculated as \frac{1+1+2+3}{4}$. In a uniform distribution, such as rolling a fair die where each outcome from 1 to 6 is equally likely, no mode exists because frequencies are identical, yet the of 3.5 clearly indicates the central value. The , while always definable for numerical data, can sometimes mislead in skewed or discrete datasets by pulling toward extremes, though it remains robust in its comprehensive inclusion of all points. In descriptive statistics, the arithmetic mean and mode are often used together alongside the median to provide a complete profile of central tendency, revealing different aspects of the data's structure—such as average performance via the mean and prevalent categories via the mode—for more informed analysis.

Generalizations

Weighted arithmetic mean

The weighted arithmetic mean assigns non-negative weights wiw_i to each data value xix_i (for i=1i = 1 to nn) to reflect their relative importance, extending the standard arithmetic mean for cases where data points contribute unequally. The value is computed as xˉ=i=1nwixii=1nwi,\bar{x} = \frac{\sum_{i=1}^n w_i x_i}{\sum_{i=1}^n w_i}, where the denominator normalizes the weights to ensure they sum to unity if they do not already. If the weights are predefined to sum to 1, the denominator is simply 1, simplifying the expression while maintaining the proportional influence of each wiw_i. This measure is widely applied in for calculating grade point averages (GPAs), where course grades are weighted by the number of credit hours to account for varying course loads. In , it determines a portfolio's as the weighted sum of individual asset returns, with weights corresponding to the proportion of capital invested in each asset. The weighted arithmetic mean inherits the linearity of the unweighted version, meaning it can be expressed as a of the xix_i, which facilitates its use in optimization and regression contexts; however, the choice of weights influences the mean's sensitivity to outliers, amplifying the impact of heavily weighted points. It reduces to the unweighted arithmetic mean when all wi=1/nw_i = 1/n, unifying the two concepts under equal weighting. For illustration, consider test scores of 90, 80, and 70 with respective weights of 0.5, 0.3, and 0.2 (e.g., reflecting differing assessment importances); the weighted is 0.5×90+0.3×80+0.2×70=830.5 \times 90 + 0.3 \times 80 + 0.2 \times 70 = 83. A notable special case is the exponential , a time-weighted variant used in time series analysis, where weights decline exponentially for older observations to emphasize recent data while still incorporating historical values as an infinite weighted sum.

Arithmetic mean in probability distributions

In , the arithmetic mean of a random variable represents its , which serves as the population mean under the probability distribution governing the variable. For a discrete XX with probability mass function p(x)p(x), the expected value is given by E[X]=xxp(x)E[X] = \sum_x x \, p(x). For a continuous XX with f(x)f(x), it is E[X]=xf(x)dxE[X] = \int_{-\infty}^{\infty} x \, f(x) \, dx. This expected value quantifies the long-run average value of the random variable over many independent realizations. When estimating the from a sample, the sample Xˉ=1ni=1nXi\bar{X} = \frac{1}{n} \sum_{i=1}^n X_i is used, where X1,,XnX_1, \dots, X_n are independent and identically distributed (i.i.d.) observations from the distribution. This sample is an unbiased of the population , meaning E[Xˉ]=E[X]E[\bar{X}] = E[X], ensuring that on average, it equals the true across repeated samples. A key result facilitating inference about the population is the (CLT), which states that for i.i.d. random variables with finite μ\mu and variance σ2>0\sigma^2 > 0, the distribution of the standardized sample n(Xˉμ)/σ\sqrt{n} (\bar{X} - \mu)/\sigma
Add your contribution
Related Hubs
User Avatar
No comments yet.