Recent from talks
Nothing was collected or created yet.
Linear discriminant analysis
View on Wikipedia
Linear discriminant analysis (LDA), normal discriminant analysis (NDA), canonical variates analysis (CVA), or discriminant function analysis is a generalization of Fisher's linear discriminant, a method used in statistics and other fields, to find a linear combination of features that characterizes or separates two or more classes of objects or events. The resulting combination may be used as a linear classifier, or, more commonly, for dimensionality reduction before later classification.
LDA is closely related to analysis of variance (ANOVA) and regression analysis, which also attempt to express one dependent variable as a linear combination of other features or measurements.[2][3] However, ANOVA uses categorical independent variables and a continuous dependent variable, whereas discriminant analysis has continuous independent variables and a categorical dependent variable (i.e. the class label).[4] Logistic regression and probit regression are more similar to LDA than ANOVA is, as they also explain a categorical variable by the values of continuous independent variables. These other methods are preferable in applications where it is not reasonable to assume that the independent variables are normally distributed, which is a fundamental assumption of the LDA method.
LDA is also closely related to principal component analysis (PCA) and factor analysis in that they both look for linear combinations of variables which best explain the data.[5] LDA explicitly attempts to model the difference between the classes of data. PCA, in contrast, does not take into account any difference in class, and factor analysis builds the feature combinations based on differences rather than similarities. Discriminant analysis is also different from factor analysis in that it is not an interdependence technique: a distinction between independent variables and dependent variables (also called criterion variables) must be made.
LDA works when the measurements made on independent variables for each observation are continuous quantities. When dealing with categorical independent variables, the equivalent technique is discriminant correspondence analysis.[6][7]
Discriminant analysis is used when groups are known a priori (unlike in cluster analysis). Each case must have a score on one or more quantitative predictor measures, and a score on a group measure.[8] In simple terms, discriminant function analysis is classification - the act of distributing things into groups, classes or categories of the same type.
History
[edit]The original dichotomous discriminant analysis was developed by Sir Ronald Fisher in 1936.[9] It is different from an ANOVA or MANOVA, which is used to predict one (ANOVA) or multiple (MANOVA) continuous dependent variables by one or more independent categorical variables. Discriminant function analysis is useful in determining whether a set of variables is effective in predicting category membership.[10]
LDA for two classes
[edit]Consider a set of observations (also called features, attributes, variables or measurements) for each sample of an object or event with known class . This set of samples is called the training set in a supervised learning context. The classification problem is then to find a good predictor for the class of any sample of the same distribution (not necessarily from the training set) given only an observation .[11]: 338
LDA approaches the problem by assuming that the conditional probability density functions and are both the normal distribution with mean and covariance parameters and , respectively. Under this assumption, the Bayes-optimal solution is to predict points as being from the second class if the log of the likelihood ratios is bigger than some threshold T, so that:
Without any further assumptions, the resulting classifier is referred to as quadratic discriminant analysis (QDA).
LDA instead makes the additional simplifying homoscedasticity assumption (i.e. that the class covariances are identical, so ) and that the covariances have full rank. In this case, several terms cancel:
- because both sides are scalar and transpose to each other ( is Hermitian)
and the above decision criterion becomes a threshold on the dot product
for some threshold constant c, where
This means that the criterion of an input being in a class is purely a function of this linear combination of the known observations.
It is often useful to see this conclusion in geometrical terms: the criterion of an input being in a class is purely a function of projection of multidimensional-space point onto vector (thus, we only consider its direction). In other words, the observation belongs to if corresponding is located on a certain side of a hyperplane perpendicular to . The location of the plane is defined by the threshold .
Assumptions
[edit]The assumptions of discriminant analysis are the same as those for MANOVA. The analysis is quite sensitive to outliers and the size of the smallest group must be larger than the number of predictor variables.[8]
- Multivariate normality: Independent variables are normal for each level of the grouping variable.[10][8]
- Homogeneity of variance/covariance (homoscedasticity): Variances among group variables are the same across levels of predictors. Can be tested with Box's M statistic.[10] It has been suggested, however, that linear discriminant analysis be used when covariances are equal, and that quadratic discriminant analysis may be used when covariances are not equal.[8]
- Independence: Participants are assumed to be randomly sampled, and a participant's score on one variable is assumed to be independent of scores on that variable for all other participants.[10][8]
It has been suggested that discriminant analysis is relatively robust to slight violations of these assumptions,[12] and it has also been shown that discriminant analysis may still be reliable when using dichotomous variables (where multivariate normality is often violated).[13]
Discriminant functions
[edit]Discriminant analysis works by creating one or more linear combinations of predictors, creating a new latent variable for each function. These functions are called discriminant functions. The number of functions possible is either where = number of groups, or (the number of predictors), whichever is smaller. The first function created maximizes the differences between groups on that function. The second function maximizes differences on that function, but also must not be correlated with the previous function. This continues with subsequent functions with the requirement that the new function not be correlated with any of the previous functions.
Given group , with sets of sample space, there is a discriminant rule such that if , then . Discriminant analysis then, finds “good” regions of to minimize classification error, therefore leading to a high percent correct classified in the classification table.[14]
Each function is given a discriminant score[clarification needed] to determine how well it predicts group placement.
- Structure Correlation Coefficients: The correlation between each predictor and the discriminant score of each function. This is a zero-order correlation (i.e., not corrected for the other predictors).[15]
- Standardized Coefficients: Each predictor's weight in the linear combination that is the discriminant function. Like in a regression equation, these coefficients are partial (i.e., corrected for the other predictors). Indicates the unique contribution of each predictor in predicting group assignment.
- Functions at Group Centroids: Mean discriminant scores for each grouping variable are given for each function. The farther apart the means are, the less error there will be in classification.
Discrimination rules
[edit]- Maximum likelihood: Assigns to the group that maximizes population (group) density.[16]
- Bayes Discriminant Rule: Assigns to the group that maximizes , where πi represents the prior probability of that classification, and represents the population density.[16]
- Fisher's linear discriminant rule: Maximizes the ratio between SSbetween and SSwithin, and finds a linear combination of the predictors to predict group.[16]
Eigenvalues
[edit]An eigenvalue in discriminant analysis is the characteristic root of each function.[clarification needed] It is an indication of how well that function differentiates the groups, where the larger the eigenvalue, the better the function differentiates.[8] This however, should be interpreted with caution, as eigenvalues have no upper limit.[10][8] The eigenvalue can be viewed as a ratio of SSbetween and SSwithin as in ANOVA when the dependent variable is the discriminant function, and the groups are the levels of the IV[clarification needed].[10] This means that the largest eigenvalue is associated with the first function, the second largest with the second, etc..
Effect size
[edit]Some suggest the use of eigenvalues as effect size measures, however, this is generally not supported.[10] Instead, the canonical correlation is the preferred measure of effect size. It is similar to the eigenvalue, but is the square root of the ratio of SSbetween and SStotal. It is the correlation between groups and the function.[10] Another popular measure of effect size is the percent of variance[clarification needed] for each function. This is calculated by: (λx/Σλi) X 100 where λx is the eigenvalue for the function and Σλi is the sum of all eigenvalues. This tells us how strong the prediction is for that particular function compared to the others.[10] Percent correctly classified can also be analyzed as an effect size. The kappa value can describe this while correcting for chance agreement.[10]Kappa normalizes across all categorizes rather than biased by a significantly good or poorly performing classes.[clarification needed][17]
Canonical discriminant analysis for k classes
[edit]Canonical discriminant analysis (CDA) finds axes (k − 1 canonical coordinates, k being the number of classes) that best separate the categories. These linear functions are uncorrelated and define, in effect, an optimal k − 1 space through the n-dimensional cloud of data that best separates (the projections in that space of) the k groups. See “Multiclass LDA” for details below.
Because LDA uses canonical variates, it was initially often referred as the "method of canonical variates"[18] or canonical variates analysis (CVA).[19]
Fisher's linear discriminant
[edit]The terms Fisher's linear discriminant and LDA are often used interchangeably, although Fisher's original article[2] actually describes a slightly different discriminant, which does not make some of the assumptions of LDA such as normally distributed classes or equal class covariances.
Suppose two classes of observations have means and covariances . Then the linear combination of features will have means and variances for . Fisher defined the separation between these two distributions to be the ratio of the variance between the classes to the variance within the classes:
This measure is, in some sense, a measure of the signal-to-noise ratio for the class labelling. It can be shown that the maximum separation occurs when
When the assumptions of LDA are satisfied, the above equation is equivalent to LDA.

Be sure to note that the vector is the normal to the discriminant hyperplane. As an example, in a two dimensional problem, the line that best divides the two groups is perpendicular to .
Generally, the data points to be discriminated are projected onto ; then the threshold that best separates the data is chosen from analysis of the one-dimensional distribution. There is no general rule for the threshold. However, if projections of points from both classes exhibit approximately the same distributions, a good choice would be the hyperplane between projections of the two means, and . In this case the parameter c in threshold condition can be found explicitly:
- .
Otsu's method is related to Fisher's linear discriminant, and was created to binarize the histogram of pixels in a grayscale image by optimally picking the black/white threshold that minimizes intra-class variance and maximizes inter-class variance within/between grayscales assigned to black and white pixel classes.
Multiclass LDA
[edit]

In the case where there are more than two classes, the analysis used in the derivation of the Fisher discriminant can be extended to find a subspace which appears to contain all of the class variability.[20] This generalization is due to C. R. Rao.[21] Suppose that each of C classes has a mean and the same covariance . Then the scatter between class variability may be defined by the sample covariance of the class means
where is the mean of the class means. The class separation in a direction in this case will be given by
This means that when is an eigenvector of the separation will be equal to the corresponding eigenvalue.
If is diagonalizable, the variability between features will be contained in the subspace spanned by the eigenvectors corresponding to the C − 1 largest eigenvalues (since is of rank C − 1 at most). These eigenvectors are primarily used in feature reduction, as in PCA. The eigenvectors corresponding to the smaller eigenvalues will tend to be very sensitive to the exact choice of training data, and it is often necessary to use regularisation as described in the next section.
If classification is required, instead of dimension reduction, there are a number of alternative techniques available. For instance, the classes may be partitioned, and a standard Fisher discriminant or LDA used to classify each partition. A common example of this is "one against the rest" where the points from one class are put in one group, and everything else in the other, and then LDA applied. This will result in C classifiers, whose results are combined. Another common method is pairwise classification, where a new classifier is created for each pair of classes (giving C(C − 1)/2 classifiers in total), with the individual classifiers combined to produce a final classification.
Incremental LDA
[edit]The typical implementation of the LDA technique requires that all the samples are available in advance. However, there are situations where the entire data set is not available and the input data are observed as a stream. In this case, it is desirable for the LDA feature extraction to have the ability to update the computed LDA features by observing the new samples without running the algorithm on the whole data set. For example, in many real-time applications such as mobile robotics or on-line face recognition, it is important to update the extracted LDA features as soon as new observations are available. An LDA feature extraction technique that can update the LDA features by simply observing new samples is an incremental LDA algorithm, and this idea has been extensively studied over the last two decades.[22] Chatterjee and Roychowdhury proposed an incremental self-organized LDA algorithm for updating the LDA features.[23] In other work, Demir and Ozmehmet proposed online local learning algorithms for updating LDA features incrementally using error-correcting and the Hebbian learning rules.[24] Later, Aliyari et al. derived fast incremental algorithms to update the LDA features by observing the new samples.[22]
Practical use
[edit]In practice, the class means and covariances are not known. They can, however, be estimated from the training set. Either the maximum likelihood estimate or the maximum a posteriori estimate may be used in place of the exact value in the above equations. Although the estimates of the covariance may be considered optimal in some sense, this does not mean that the resulting discriminant obtained by substituting these values is optimal in any sense, even if the assumption of normally distributed classes is correct.
Another complication in applying LDA and Fisher's discriminant to real data occurs when the number of measurements of each sample (i.e., the dimensionality of each data vector) exceeds the number of samples in each class.[5] In this case, the covariance estimates do not have full rank, and so cannot be inverted. There are a number of ways to deal with this. One is to use a pseudo inverse instead of the usual matrix inverse in the above formulae. However, better numeric stability may be achieved by first projecting the problem onto the subspace spanned by .[25] Another strategy to deal with small sample size is to use a shrinkage estimator of the covariance matrix, which can be expressed mathematically as
where is the identity matrix, and is the shrinkage intensity or regularisation parameter. This leads to the framework of regularized discriminant analysis[26] or shrinkage discriminant analysis.[27]
Also, in many practical cases linear discriminants are not suitable. LDA and Fisher's discriminant can be extended for use in non-linear classification via the kernel trick. Here, the original observations are effectively mapped into a higher dimensional non-linear space. Linear classification in this non-linear space is then equivalent to non-linear classification in the original space. The most commonly used example of this is the kernel Fisher discriminant.
LDA can be generalized to multiple discriminant analysis, where c becomes a categorical variable with N possible states, instead of only two. Analogously, if the class-conditional densities are normal with shared covariances, the sufficient statistic for are the values of N projections, which are the subspace spanned by the N means, affine projected by the inverse covariance matrix. These projections can be found by solving a generalized eigenvalue problem, where the numerator is the covariance matrix formed by treating the means as the samples, and the denominator is the shared covariance matrix. See “Multiclass LDA” above for details.
Applications
[edit]In addition to the examples given below, LDA is applied in positioning and product management.
Bankruptcy prediction
[edit]In bankruptcy prediction based on accounting ratios and other financial variables, linear discriminant analysis was the first statistical method applied to systematically explain which firms entered bankruptcy vs. survived. Despite limitations including known nonconformance of accounting ratios to the normal distribution assumptions of LDA, Edward Altman's 1968 model[28] is still a leading model in practical applications.[29][30][31]
Face recognition
[edit]In computerised face recognition, each face is represented by a large number of pixel values. Linear discriminant analysis is primarily used here to reduce the number of features to a more manageable number before classification. Each of the new dimensions is a linear combination of pixel values, which form a template. The linear combinations obtained using Fisher's linear discriminant are called Fisher faces, while those obtained using the related principal component analysis are called eigenfaces.
Marketing
[edit]In marketing, discriminant analysis was once often used to determine the factors which distinguish different types of customers and/or products on the basis of surveys or other forms of collected data. Logistic regression or other methods are now more commonly used. The use of discriminant analysis in marketing can be described by the following steps:
- Formulate the problem and gather data—Identify the salient attributes consumers use to evaluate products in this category—Use quantitative marketing research techniques (such as surveys) to collect data from a sample of potential customers concerning their ratings of all the product attributes. The data collection stage is usually done by marketing research professionals. Survey questions ask the respondent to rate a product from one to five (or 1 to 7, or 1 to 10) on a range of attributes chosen by the researcher. Anywhere from five to twenty attributes are chosen. They could include things like: ease of use, weight, accuracy, durability, colourfulness, price, or size. The attributes chosen will vary depending on the product being studied. The same question is asked about all the products in the study. The data for multiple products is codified and input into a statistical program such as R, SPSS or SAS. (This step is the same as in Factor analysis).
- Estimate the Discriminant Function Coefficients and determine the statistical significance and validity—Choose the appropriate discriminant analysis method. The direct method involves estimating the discriminant function so that all the predictors are assessed simultaneously. The stepwise method enters the predictors sequentially. The two-group method should be used when the dependent variable has two categories or states. The multiple discriminant method is used when the dependent variable has three or more categorical states. Use Wilks's Lambda to test for significance in SPSS or F stat in SAS. The most common method used to test validity is to split the sample into an estimation or analysis sample, and a validation or holdout sample. The estimation sample is used in constructing the discriminant function. The validation sample is used to construct a classification matrix which contains the number of correctly classified and incorrectly classified cases. The percentage of correctly classified cases is called the hit ratio.
- Plot the results on a two dimensional map, define the dimensions, and interpret the results. The statistical program (or a related module) will map the results. The map will plot each product (usually in two-dimensional space). The distance of products to each other indicate either how different they are. The dimensions must be labelled by the researcher. This requires subjective judgement and is often very challenging. See perceptual mapping.
Biomedical studies
[edit]The main application of discriminant analysis in medicine is the assessment of severity state of a patient and prognosis of disease outcome. For example, during retrospective analysis, patients are divided into groups according to severity of disease – mild, moderate, and severe form. Then results of clinical and laboratory analyses are studied to reveal statistically different variables in these groups. Using these variables, discriminant functions are built to classify disease severity in future patients. Additionally, Linear Discriminant Analysis (LDA) can help select more discriminative samples for data augmentation, improving classification performance.[32]
In biology, similar principles are used in order to classify and define groups of different biological objects, for example, to define phage types of Salmonella enteritidis based on Fourier transform infrared spectra,[33] to detect animal source of Escherichia coli studying its virulence factors[34] etc.
Earth science
[edit]This method can be used to separate the alteration zones[clarification needed]. For example, when different data from various zones are available, discriminant analysis can find the pattern within the data and classify it effectively.[35]
Comparison to logistic regression
[edit]Discriminant function analysis is very similar to logistic regression, and both can be used to answer the same research questions.[10] Logistic regression does not have as many assumptions and restrictions as discriminant analysis. However, when discriminant analysis’ assumptions are met, it is more powerful than logistic regression.[36] Unlike logistic regression, discriminant analysis can be used with small sample sizes. It has been shown that when sample sizes are equal, and homogeneity of variance/covariance holds, discriminant analysis is more accurate.[8] Despite all these advantages, logistic regression has none-the-less become the common choice, since the assumptions of discriminant analysis are rarely met.[9][8]
Linear discriminant in high dimensions
[edit]Geometric anomalies in higher dimensions lead to the well-known curse of dimensionality. Nevertheless, proper utilization of concentration of measure phenomena can make computation easier.[37] An important case of these blessing of dimensionality phenomena was highlighted by Donoho and Tanner: if a sample is essentially high-dimensional then each point can be separated from the rest of the sample by linear inequality, with high probability, even for exponentially large samples.[38] These linear inequalities can be selected in the standard (Fisher's) form of the linear discriminant for a rich family of probability distribution.[39] In particular, such theorems are proven for log-concave distributions including multidimensional normal distribution (the proof is based on the concentration inequalities for log-concave measures[40]) and for product measures on a multidimensional cube (this is proven using Talagrand's concentration inequality for product probability spaces). Data separability by classical linear discriminants simplifies the problem of error correction for artificial intelligence systems in high dimension.[41]
See also
[edit]References
[edit]- ^ Holtel, Frederik (2023-02-20). "Linear Discriminant Analysis (LDA) Can Be So Easy". Medium. Retrieved 2024-05-18.
- ^ a b Fisher, R. A. (1936). "The Use of Multiple Measurements in Taxonomic Problems" (PDF). Annals of Eugenics. 7 (2): 179–188. doi:10.1111/j.1469-1809.1936.tb02137.x. hdl:2440/15227.
- ^ McLachlan, G. J. (2004). Discriminant Analysis and Statistical Pattern Recognition. Wiley Interscience. ISBN 978-0-471-69115-0. MR 1190469.
- ^ Analyzing Quantitative Data: An Introduction for Social Researchers, Debra Wetcher-Hendricks, p.288
- ^ a b Martinez, A. M.; Kak, A. C. (2001). "PCA versus LDA" (PDF). IEEE Transactions on Pattern Analysis and Machine Intelligence. 23 (2): 228–233. doi:10.1109/34.908974. Archived from the original (PDF) on 2008-10-11. Retrieved 2010-06-30.
- ^ Abdi, H. (2007) "Discriminant correspondence analysis." In: N.J. Salkind (Ed.): Encyclopedia of Measurement and Statistic. Thousand Oaks (CA): Sage. pp. 270–275.
- ^ Perriere, G.; Thioulouse, J. (2003). "Use of Correspondence Discriminant Analysis to predict the subcellular location of bacterial proteins". Computer Methods and Programs in Biomedicine. 70 (2): 99–105. doi:10.1016/s0169-2607(02)00011-1. PMID 12507786.
- ^ a b c d e f g h i Büyüköztürk, Ş. & Çokluk-Bökeoğlu, Ö. (2008). Discriminant function analysis: Concept and application. Egitim Arastirmalari - Eurasian Journal of Educational Research, 33, 73-92.
- ^ a b Cohen et al. Applied Multiple Regression/Correlation Analysis for the Behavioural Sciences 3rd ed. (2003). Taylor & Francis Group.
- ^ a b c d e f g h i j k Hansen, John (2005). "Using SPSS for Windows and Macintosh: Analyzing and Understanding Data". The American Statistician. 59: 113. doi:10.1198/tas.2005.s139.
- ^ Venables, W. N.; Ripley, B. D. (2002). Modern Applied Statistics with S (4th ed.). Springer Verlag. ISBN 978-0-387-95457-8.
- ^ Peter A. Lachenbruch. (1975). Discriminant analysis. NY: Hafner
- ^ Klecka, William R. (1980). Discriminant analysis. Quantitative Applications in the Social Sciences Series, No. 19. Thousand Oaks, CA: Sage Publications.
- ^ Hardle, W., Simar, L. (2007). Applied Multivariate Statistical Analysis. Springer Berlin Heidelberg. pp. 289–303.
- ^ Garson, G. D. (2008). Discriminant function analysis. https://web.archive.org/web/20080312065328/http://www2.chass.ncsu.edu/garson/pA765/discrim.htm.
- ^ a b c Hardle, W., Simar, L. (2007). Applied Multivariate Statistical Analysis. Springer Berlin Heidelberg. pp. 289-303.
- ^ Israel, Steven A. (June 2006). "Performance Metrics: How and When". Geocarto International. 21 (2): 23–32. Bibcode:2006GeoIn..21...23I. doi:10.1080/10106040608542380. ISSN 1010-6049. S2CID 122376081.
- ^ Nabney, Ian (2002). Netlab: Algorithms for Pattern Recognition. p. 274. ISBN 1-85233-440-1.
- ^ Magwene, Paul (2023). "Chapter 14: Canonical Variates Analysis". Statistical Computing for Biologists.
- ^ Garson, G. D. (2008). Discriminant function analysis. "PA 765: Discriminant Function Analysis". Archived from the original on 2008-03-12. Retrieved 2008-03-04. .
- ^ Rao, R. C. (1948). "The utilization of multiple measurements in problems of biological classification". Journal of the Royal Statistical Society, Series B. 10 (2): 159–203. doi:10.1111/j.2517-6161.1948.tb00008.x. JSTOR 2983775.
- ^ a b Aliyari Ghassabeh, Youness; Rudzicz, Frank; Moghaddam, Hamid Abrishami (2015-06-01). "Fast incremental LDA feature extraction". Pattern Recognition. 48 (6): 1999–2012. Bibcode:2015PatRe..48.1999A. doi:10.1016/j.patcog.2014.12.012.
- ^ Chatterjee, C.; Roychowdhury, V.P. (1997-05-01). "On self-organizing algorithms and networks for class-separability features". IEEE Transactions on Neural Networks. 8 (3): 663–678. doi:10.1109/72.572105. ISSN 1045-9227. PMID 18255669.
- ^ Demir, G. K.; Ozmehmet, K. (2005-03-01). "Online Local Learning Algorithms for Linear Discriminant Analysis". Pattern Recognit. Lett. 26 (4): 421–431. Bibcode:2005PaReL..26..421D. doi:10.1016/j.patrec.2004.08.005. ISSN 0167-8655.
- ^ Yu, H.; Yang, J. (2001). "A direct LDA algorithm for high-dimensional data — with application to face recognition". Pattern Recognition. 34 (10): 2067–2069. Bibcode:2001PatRe..34.2067Y. CiteSeerX 10.1.1.70.3507. doi:10.1016/s0031-3203(00)00162-x.
- ^ Friedman, J. H. (1989). "Regularized Discriminant Analysis" (PDF). Journal of the American Statistical Association. 84 (405): 165–175. CiteSeerX 10.1.1.382.2682. doi:10.2307/2289860. JSTOR 2289860. MR 0999675.
- ^ Ahdesmäki, M.; Strimmer, K. (2010). "Feature selection in omics prediction problems using cat scores and false nondiscovery rate control". Annals of Applied Statistics. 4 (1): 503–519. arXiv:0903.2003. doi:10.1214/09-aoas277. S2CID 2508935.
- ^ Altman, Edward I. (1968). "Financial Ratios, Discriminant Analysis and the Prediction of Corporate Bankruptcy". The Journal of Finance. 23 (4): 589–609. doi:10.2307/2978933. JSTOR 2978933.
- ^ Agarwal, Vineet; Taffler, Richard (2005). "Twenty-five years of z-scores in the UK: do they really work?" (PDF).
- ^ Agarwal, Vineet; Taffler, Richard (2007). "Twenty-Five Years of the Taffler Z-Score Model: Does It Really Have Predictive Ability?". Accounting and Business Research. 37 (4): 285–300. doi:10.1080/00014788.2007.9663313.
- ^ Bimpong, Patrick; et al. (2020). "Assessing Predictive Power and Earnings Manipulations. Applied Study on Listed Consumer Goods and Service Companies in Ghana Using 3 Z-Score Models". Expert Journal of Finance. 8 (1): 1–26.
- ^ Moradi, M; Demirel, H (2024). "Alzheimer's disease classification using 3D conditional progressive GAN-and LDA-based data selection". Signal, Image and Video Processing. 18 (2): 1847–1861. doi:10.1007/s11760-023-02878-4.
- ^ Preisner, O; Guiomar, R; Machado, J; Menezes, JC; Lopes, JA (2010). "Application of Fourier transform infrared spectroscopy and chemometrics for differentiation of Salmonella enterica serovar Enteritidis phage types". Appl Environ Microbiol. 76 (11): 3538–3544. Bibcode:2010ApEnM..76.3538P. doi:10.1128/aem.01589-09. PMC 2876429. PMID 20363777.
- ^ David, DE; Lynne, AM; Han, J; Foley, SL (2010). "Evaluation of virulence factor profiling in the characterization of veterinary Escherichia coli isolates". Appl Environ Microbiol. 76 (22): 7509–7513. Bibcode:2010ApEnM..76.7509D. doi:10.1128/aem.00726-10. PMC 2976202. PMID 20889790.
- ^ Tahmasebi, P.; Hezarkhani, A.; Mortazavi, M. (2010). "Application of discriminant analysis for alteration separation; sungun copper deposit, East Azerbaijan, Iran. Australian" (PDF). Journal of Basic and Applied Sciences. 6 (4): 564–576.
- ^ Trevor Hastie; Robert Tibshirani; Jerome Friedman. The Elements of Statistical Learning. Data Mining, Inference, and Prediction (second ed.). Springer. p. 128.
- ^ Kainen P.C. (1997) Utilizing geometric anomalies of high dimension: When complexity makes computation easier. In: Kárný M., Warwick K. (eds) Computer Intensive Methods in Control and Signal Processing: The Curse of Dimensionality, Springer, 1997, pp. 282–294.
- ^ Donoho, D., Tanner, J. (2009) Observed universality of phase transitions in high-dimensional geometry, with implications for modern data analysis and signal processing, Phil. Trans. R. Soc. A 367, 4273–4293.
- ^ Gorban, Alexander N.; Golubkov, Alexander; Grechuck, Bogdan; Mirkes, Evgeny M.; Tyukin, Ivan Y. (2018). "Correction of AI systems by linear discriminants: Probabilistic foundations". Information Sciences. 466: 303–322. arXiv:1811.05321. doi:10.1016/j.ins.2018.07.040. S2CID 52876539.
- ^ Guédon, O., Milman, E. (2011) Interpolating thin-shell and sharp large-deviation estimates for isotropic log-concave measures, Geom. Funct. Anal. 21 (5), 1043–1068.
- ^ Gorban, Alexander N.; Makarov, Valeri A.; Tyukin, Ivan Y. (July 2019). "The unreasonable effectiveness of small neural ensembles in high-dimensional brain". Physics of Life Reviews. 29: 55–88. arXiv:1809.07656. Bibcode:2019PhLRv..29...55G. doi:10.1016/j.plrev.2018.09.005. PMID 30366739.
Further reading
[edit]- Duda, R. O.; Hart, P. E.; Stork, D. H. (2000). Pattern Classification (2nd ed.). Wiley Interscience. ISBN 978-0-471-05669-0. MR 1802993.
- Hilbe, J. M. (2009). Logistic Regression Models. Chapman & Hall/CRC Press. ISBN 978-1-4200-7575-5.
- Mika, S.; et al. (1999). "Fisher discriminant analysis with kernels". Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (Cat. No.98TH8468). pp. 41–48. CiteSeerX 10.1.1.35.9904. doi:10.1109/NNSP.1999.788121. ISBN 978-0-7803-5673-3. S2CID 8473401.
- McFarland, H. Richard; Donald, St. P. Richards (2001). "Exact Misclassification Probabilities for Plug-In Normal Quadratic Discriminant Functions. I. The Equal-Means Case". Journal of Multivariate Analysis. 77 (1): 21–53. doi:10.1006/jmva.2000.1924.
- McFarland, H. Richard; Donald, St. P. Richards (2002). "Exact Misclassification Probabilities for Plug-In Normal Quadratic Discriminant Functions. II. The Heterogeneous Case". Journal of Multivariate Analysis. 82 (2): 299–330. doi:10.1006/jmva.2001.2034.
- Haghighat, M.; Abdel-Mottaleb, M.; Alhalabi, W. (2016). "Discriminant Correlation Analysis: Real-Time Feature Level Fusion for Multimodal Biometric Recognition". IEEE Transactions on Information Forensics and Security. 11 (9): 1984–1996. doi:10.1109/TIFS.2016.2569061. S2CID 15624506.
External links
[edit]- Discriminant Correlation Analysis (DCA) of the Haghighat article (see above)
- ALGLIB contains open-source LDA implementation in C# / C++ / Pascal / VBA.
- LDA in Python- LDA implementation in Python
- LDA tutorial using MS Excel
- Biomedical statistics. Discriminant analysis
- StatQuest: Linear Discriminant Analysis (LDA) clearly explained on YouTube
- Course notes, Discriminant function analysis by G. David Garson, NC State University
- Discriminant analysis tutorial in Microsoft Excel by Kardi Teknomo
- Course notes, Discriminant function analysis by David W. Stockburger, Missouri State University Archived 2016-03-03 at the Wayback Machine
- Discriminant function analysis (DA) by John Poulsen and Aaron French, San Francisco State University
Linear discriminant analysis
View on GrokipediaHistorical Development
Origins in Statistics
The origins of linear discriminant analysis lie in early efforts to separate groups using linear combinations of multiple variables, rooted in biometric and anthropological applications. Karl Pearson laid foundational concepts in multivariate analysis through his 1901 development of principal components analysis, published in the Philosophical Magazine, which involved constructing linear combinations of variables to best represent systems of points in multivariate space.[6] This work provided tools for dimensionality reduction that later influenced discriminant methods. In the 1920s, Pearson extended these ideas with the coefficient of racial likeness, a statistical measure designed to quantify differences between populations using linear functions of correlated variables, particularly for classifying human groups from physical traits such as cranial indices.[7] Concurrently, Prasanta Chandra Mahalanobis contributed to discriminant concepts through his development of distance measures in anthropometric studies, starting around 1920 with analyses of race mixture in Bengal, which accounted for variable correlations to better separate ethnic groups.[8] These pre-Fisher innovations found practical use in anthropology and biometrics for species classification, where linear separators based on multivariate measurements—such as skull dimensions or body proportions—were employed to distinguish human races or animal taxa without probabilistic classification rules.[9] Such applications highlighted the utility of linear methods for group discrimination in empirical sciences. This statistical groundwork set the stage for Ronald Fisher's 1936 formalization of the technique.Key Contributions and Evolution
Ronald Fisher introduced linear discriminant analysis in his seminal 1936 paper, where he proposed a method to find a linear combination of multiple measurements that maximizes the separation between taxonomic groups, specifically applied to classifying three species of iris flowers using four morphological features: sepal length, sepal width, petal length, and petal width.[10] This approach, known as Fisher's linear discriminant, derived coefficients for a discriminant function that achieved perfect separation in the binary classification between Iris setosa and Iris versicolor, with no overlap in the projected values across the 50 samples per species, demonstrating the method's efficacy for distinguishing populations with multivariate normal distributions.[10] Following World War II, C. Radhakrishna Rao advanced the theoretical foundations in 1948 by generalizing Fisher's discriminant criterion to multiple populations and linking it to canonical correlations, providing a unified framework for biological classification problems through the maximization of between-group variance relative to within-group variance. Rao's criterion, which involves solving a generalized eigenvalue problem, extended the method's applicability beyond binary cases and established connections to multivariate analysis techniques, influencing subsequent developments in statistical discrimination. In the 1970s and 1980s, computational advancements made eigenvalue-based solutions for linear discriminant analysis more tractable, building on Harold Hotelling's earlier contributions to multivariate analysis, including his 1936 introduction of canonical correlation analysis that provided the mathematical basis for extracting discriminant directions via eigenvalue decomposition.[11] These methods gained practical utility with improved computing resources, enabling efficient implementation of the generalized eigenvalue problem central to LDA for high-dimensional data.[11] Modern milestones in the 1990s integrated linear discriminant analysis into machine learning, notably through Belhumeur et al.'s 1997 work applying it to face recognition, where "Fisherfaces" outperformed principal component analysis by projecting data onto class-specific directions that enhance separability under varying illumination and pose. In the 2000s, online variants emerged for streaming data, such as Pang et al.'s 2005 incremental linear discriminant analysis, which updates the discriminant subspace efficiently as new data arrives without full recomputation, addressing concept drift in dynamic environments like sensor networks. Since the 2010s, LDA has been extended in kernel and deep learning frameworks, incorporating nonlinear mappings and neural architectures for improved performance on complex datasets as of 2025.[12]Fundamental Principles
Core Assumptions
Linear discriminant analysis (LDA) relies on several key statistical assumptions to ensure the validity of its discriminant functions and classification boundaries. Central to the method is the assumption that the observations within each class are independently and identically distributed (i.i.d.), which underpins the probabilistic framework for separating classes based on linear combinations of features.[13] This independence allows the log-posterior ratio between classes to exhibit linearity, facilitating optimal separation along linear decision boundaries when the other distributional assumptions hold.[14] A foundational assumption is multivariate normality for each class: the feature vectors for class are drawn from a multivariate Gaussian distribution , where is the class-specific mean vector. Additionally, LDA assumes homoscedasticity, meaning the covariance matrix is identical across all classes (), which simplifies the decision rule to a linear form and avoids the need for class-specific quadratic terms.[2] These normality and equal covariance assumptions enable the derivation of maximum likelihood estimates for the parameters and ensure that the method achieves the Bayes optimal classifier under the model. The model also incorporates prior probabilities for each class , representing the relative frequency of occurrence in the population; these are often assumed equal () unless empirical evidence suggests otherwise, such as through sample proportions. Violations of these assumptions can compromise performance: for instance, heteroscedasticity (unequal covariances) introduces bias in boundary estimation, prompting the use of quadratic discriminant analysis (QDA) as an alternative that relaxes the equal covariance constraint.[13] In high-dimensional settings where the number of features exceeds the sample size, the normality assumption may lead to overfitting or biased covariance estimates, reducing the method's reliability unless regularized variants are employed.Binary Classification Framework
Linear discriminant analysis in the binary classification framework addresses the problem of distinguishing between two classes, typically labeled as class 0 and class 1, where the data from each class is assumed to follow a multivariate normal distribution with respective means and , and a shared covariance matrix . The objective is to derive a linear projection that maximizes the separation between the projected class means while minimizing the within-class variability, thereby facilitating effective classification in a lower-dimensional space.[15] The core mechanism relies on Fisher's criterion, which seeks to maximize the ratio of between-class scatter to within-class scatter for a projection vector . This is formalized as the objective function where represents the between-class scatter matrix, capturing the variance due to differences in class means, and denotes the within-class scatter matrix, reflecting the common variability within each class.[15] Maximizing yields the optimal projection vector , which points in the direction that best discriminates the classes by solving the generalized eigenvalue problem inherent in the criterion. This projection maps the original high-dimensional data onto a one-dimensional line, where the projected distributions of the two classes exhibit maximal separation relative to their spreads.[15] For classifying a new observation , the projected value is compared against a threshold: assign to class 1 if , and to class 0 otherwise, where the threshold is typically set to to account for class priors and , assuming equal misclassification costs.[15][13] As an illustrative example, consider two-dimensional data consisting of two Gaussian blobs centered at distinct means with identical covariance structures; the optimal LDA projection aligns with the vector connecting the means, transforming the data into a one-dimensional space where the classes are well-separated along this axis for straightforward thresholding.Mathematical Derivation
Discriminant Functions
In linear discriminant analysis (LDA), the discriminant functions arise from the application of Bayes' theorem under the assumption of multivariate Gaussian class-conditional densities with equal covariance matrices across classes. Specifically, for a feature vector , the posterior probability of class is given by , where is the prior probability of class and is the class-conditional density with mean and common covariance . Taking the logarithm of the posterior, the classification rule assigns to the class that maximizes the discriminant score , which includes a quadratic term in . However, since is the same for all classes, the term is common across all and can be ignored for maximization, yielding the linear form This linear function represents the log-posterior up to a constant, enabling efficient computation of class posteriors.[16] When comparing two classes and , the log-odds ratio simplifies further to , where is a constant. For equal priors (), this reduces to a purely linear boundary in . Geometrically, the decision boundary where forms a hyperplane perpendicular to the vector , which points in the direction that maximizes the separation between the projected class means while accounting for the covariance structure.[16]Eigenvalue and Effect Size Analysis
In linear discriminant analysis, the optimal discriminant directions are determined by solving the generalized eigenvalue problem , where denotes the between-class scatter matrix, the within-class scatter matrix, the eigenvectors representing projection directions, and the corresponding eigenvalues that measure the separation achieved along each direction. The eigenvalues serve as indicators of discriminatory power, with higher values signifying greater class separation relative to within-class variability; these are typically ordered in descending magnitude to prioritize the most effective directions. The eigenvector associated with the largest eigenvalue corresponds to Fisher's linear discriminant, which maximizes the ratio of between-class to within-class variance. In the binary classification setting, the analysis reduces to a single non-zero eigenvalue, expressed as , which equals the squared Mahalanobis distance between the two class means and . This eigenvalue quantifies the overall separability of the two classes under the assumptions of equal covariance matrices. Furthermore, in binary LDA, the trace of the eigenvalue matrix, , directly corresponds to this squared Mahalanobis distance, providing a scalar summary of the discrimination strength.[5] Effect sizes in LDA are evaluated using multivariate criteria to assess the overall discriminatory capability. Wilks' lambda, defined as , ranges from 0 to 1, with values approaching 0 indicating strong group separation as the between-class variance dominates the total variance.[17] For multivariate extensions beyond the binary case, Pillai's trace offers a complementary measure, computed as the sum of the squared canonical correlations between the discriminant functions and the dependent variables; higher values reflect superior discrimination by emphasizing the proportion of variance explained by the between-class component.[18] These metrics enable statistical testing of whether the derived discriminants significantly differentiate the classes, with Wilks' lambda often converted to an F-statistic for hypothesis evaluation.[17]Extensions to Multiple Classes
Canonical Discriminant Analysis
Canonical discriminant analysis extends linear discriminant analysis to scenarios involving more than two classes, providing a framework for dimensionality reduction through the identification of linear combinations of features, known as canonical variates, that maximize the separation between multiple class means while minimizing within-class variability.[19] In this multivariate generalization, originally formalized for multiple groups by C.R. Rao in 1948, the method derives directions in feature space that capture the essential differences among classes, with the number of meaningful canonical variates limited to , where is the dimensionality of the input data. This approach is particularly useful for supervised dimension reduction, projecting high-dimensional data onto a lower-dimensional space where class distinctions are accentuated. The foundational setup involves defining the between-class scatter matrix and the within-class scatter matrix . For classes with prior probabilities (typically estimated as the proportion of samples in class ), class-conditional means , and overall mean , the between-class scatter matrix is given by which quantifies the dispersion of the class means around the grand mean.[20] The within-class scatter matrix is where is the covariance matrix of class , assuming multivariate normality within each class; in practice, is often pooled across classes as the average within-class covariance.[21] These matrices form the basis for the generalized eigenvalue problem , solved to obtain the eigenvectors corresponding to the largest eigenvalues , which indicate the discriminatory power of each direction. The canonical variates are the projections of the centered data onto these eigenvectors: the -th canonical variable is , with normalized such that .[19] The eigenvalues relate to the canonical correlations , which measure the strength of association between the original variables and the class structure, providing an interpretation akin to the roots in multivariate analysis of variance (MANOVA).[22] The first few canonical variates, ordered by decreasing , are selected for projection, as they successively maximize the ratio of between-class to within-class variance. Unlike principal component analysis (PCA), which seeks directions maximizing total variance without regard to class labels, canonical discriminant analysis explicitly optimizes class separability by maximizing the trace of or equivalent criteria, making it a supervised alternative for tasks requiring clear group distinctions.[21] This focus on between-group variance ensures that the reduced representation preserves discriminatory information, though it requires at least samples per class for identifiability.Multiclass and Incremental Variants
In multiclass linear discriminant analysis (LDA), classification proceeds by assigning an observation to the class that maximizes the discriminant function , where is the mean vector of class , is the shared covariance matrix, and is the prior probability of class . This formulation yields up to non-trivial discriminant directions for classes, as the between-class scatter matrix has rank at most . The inclusion of priors naturally accommodates unbalanced classes by weighting the contributions of each class according to their estimated prevalence in the data.[23] Incremental variants of LDA address scenarios where data arrives sequentially or in streams, enabling updates to the model without recomputing the full scatter matrices from scratch. One early approach is the incremental LDA (ILDA) algorithm, which supports both sequential (one sample at a time) and chunk-based updates to the between-class scatter matrix and within-class scatter matrix using rank-1 modifications for new data points or batches. This method is particularly suited for streaming data classification, maintaining discriminative performance while avoiding storage of the entire dataset. Another variant is the incremental orthogonal centroid algorithm (IOCA), which extends the orthogonal centroid method to compute LDA projections incrementally for binary and multiclass settings without retaining all historical data, by iteratively orthogonalizing class centroids in the feature space.[12] A key challenge in these incremental methods is ensuring numerical stability during updates, as repeated rank-1 modifications to scatter matrices can accumulate errors or lead to ill-conditioning, often mitigated through techniques like QR decomposition or regularization of the matrices. Such approaches are valuable in large-scale machine learning applications, reducing the computational time from for full LDA recomputation (with samples and features) to per update, facilitating real-time adaptation in dynamic environments.[24]Practical Implementation
Decision Rules and Classification
In binary linear discriminant analysis, classification decisions are made by evaluating the discriminant function for a new observation , assigning it to class 1 if and to class 0 otherwise, where the threshold under the assumption of equal prior probabilities for the two classes. This rule arises from the Bayes optimal decision boundary that minimizes misclassification error when class-conditional densities are Gaussian with equal covariance .[3] For multiclass problems with classes, linear discriminant analysis extends the binary framework by computing class-specific discriminant scores for each class , and assigning to the class maximizing , which approximates the maximum a posteriori probability under Gaussian assumptions with shared covariance. Equivalently, this corresponds to classifying to the nearest class centroid in the low-dimensional discriminant subspace spanned by the leading eigenvectors of the between-class scatter matrix, using the Mahalanobis distance metric defined by .[25] To evaluate classification performance, the resubstitution error rate—computed by applying the decision rule to the training data—provides a lower bound but systematically underestimates the true generalization error due to overfitting.[25] Cross-validation, such as -fold partitioning of the data, yields a more unbiased estimate by training on subsets and testing on held-out portions, averaging the resulting error rates across folds. Leave-one-out cross-validation is especially efficient for linear discriminant analysis given its parametric nature, as refitting the model after removing a single observation involves minimal recomputation of means and covariance, enabling exact assessment of prediction accuracy for small-to-moderate datasets.[25] In multiclass settings, the confusion matrix tabulates predicted versus actual class labels across all categories, quantifying per-class error rates and overall accuracy to highlight imbalances in discrimination performance. When multiple classes yield identical maximum discriminant scores for an observation—resulting in a tie—resolution typically involves random assignment among the tied classes to maintain probabilistic consistency, or preferentially selecting the class with the highest prior probability if priors differ.[26]Computational Considerations
Implementing linear discriminant analysis (LDA) involves several computational challenges, particularly related to numerical stability and scalability in high-dimensional settings. The core computations require estimating the within-class scatter matrix and solving for its inverse , which is used in deriving the discriminant functions. Direct matrix inversion can be numerically unstable due to potential ill-conditioning of , especially when the data exhibits near-collinear features. To mitigate this, Cholesky decomposition is commonly employed to compute factor by factor, avoiding explicit inversion and improving stability by ensuring positive definiteness assumptions hold through the lower triangular factorization , where is the Cholesky factor. A frequent issue arises when the number of features exceeds the sample size (i.e., ), rendering singular and preventing its inversion. Shrinkage estimators address this by blending the sample covariance with a target matrix, such as the identity, to ensure invertibility; for instance, Friedman's regularized discriminant analysis (RDA) introduces parameters that shrink the covariance estimates toward a common or diagonal form, balancing bias and variance effectively in small-sample scenarios.[27] For scalability to large , regularized variants incorporate ridge penalties by adding a multiple of the identity matrix to , yielding for some , which stabilizes estimation without drastically altering the discriminant directions. Approximate methods further enhance efficiency, such as randomized singular value decomposition (SVD) to low-rank approximate or the generalized eigenvalue problem, reducing the effective dimensionality before full eigendecomposition.[28][29] The time complexity of standard LDA is dominated by the eigendecomposition of the matrix in the generalized eigenvalue problem, incurring operations, alongside for computing scatter matrices from samples, making it feasible for moderate but challenging for very high dimensions. Incremental variants can update these computations online for streaming data, though they retain similar per-update costs.[30] Practical implementations are available in major statistical and machine learning libraries. In R, thelda() function from the MASS package performs LDA with options for priors and cross-validation. Python's scikit-learn provides LinearDiscriminantAnalysis in the discriminant_analysis module, supporting shrinkage via the shrinkage parameter for regularized estimation. MATLAB's classify function, part of the Statistics and Machine Learning Toolbox, handles LDA classification with built-in support for linear and quadratic variants.[31]
