Recent from talks
Knowledge base stats:
Talk channels stats:
Members stats:
Confirmatory factor analysis
In statistics, confirmatory factor analysis (CFA) is a special form of factor analysis, most commonly used in social science research. It is used to test whether measures of a construct are consistent with a researcher's understanding of the nature of that construct (or factor). As such, the objective of confirmatory factor analysis is to test whether the data fit a hypothesized measurement model. This hypothesized model is based on theory and/or previous analytic research. CFA was first developed by Jöreskog (1969) and has built upon and replaced older methods of analyzing construct validity such as the MTMM Matrix as described in Campbell & Fiske (1959).
In confirmatory factor analysis, the researcher first develops a hypothesis about what factors they believe are underlying the measures used (e.g., "Depression" being the factor underlying the Beck Depression Inventory and the Hamilton Rating Scale for Depression) and may impose constraints on the model based on these a priori hypotheses. By imposing these constraints, the researcher is forcing the model to be consistent with their theory. For example, if it is posited that there are two factors accounting for the covariance in the measures, and that these factors are unrelated to each other, the researcher can create a model where the correlation between factor A and factor B is constrained to zero. Model fit measures could then be obtained to assess how well the proposed model captured the covariance between all the items or measures in the model. If the constraints the researcher has imposed on the model are inconsistent with the sample data, then the results of statistical tests of model fit will indicate a poor fit, and the model will be rejected. If the fit is poor, it may be due to some items measuring multiple factors. It might also be that some items within a factor are more related to each other than others.
For some applications, the requirement of "zero loadings" (for indicators not supposed to load on a certain factor) has been regarded as too strict. A newly developed analysis method, "exploratory structural equation modeling", specifies hypotheses about the relation between observed indicators and their supposed primary latent factors while allowing for estimation of loadings with other latent factors as well.
In confirmatory factor analysis, researchers are typically interested in studying the degree to which responses on a p x 1 vector of observable random variables can be used to assign a value to one or more unobserved variable(s) . The investigation is largely accomplished by estimating and evaluating the loading of each item used to tap aspects of the unobserved latent variable. That is, y[i] is the vector of observed responses predicted by the unobserved latent variable , which is defined as:
,
where is the p x 1 vector of observed random variables, are the unobserved latent variables and is a p x k matrix with k equal to the number of latent variables. Since, are imperfect measures of , the model also consists of error, . Estimates in the maximum likelihood (ML) case generated by iteratively minimizing the fit function,
where is the variance-covariance matrix implied by the proposed factor analysis model and is the observed variance-covariance matrix. That is, values are found for free model parameters that minimize the difference between the model-implied variance-covariance matrix and observed variance-covariance matrix.
Hub AI
Confirmatory factor analysis AI simulator
(@Confirmatory factor analysis_simulator)
Confirmatory factor analysis
In statistics, confirmatory factor analysis (CFA) is a special form of factor analysis, most commonly used in social science research. It is used to test whether measures of a construct are consistent with a researcher's understanding of the nature of that construct (or factor). As such, the objective of confirmatory factor analysis is to test whether the data fit a hypothesized measurement model. This hypothesized model is based on theory and/or previous analytic research. CFA was first developed by Jöreskog (1969) and has built upon and replaced older methods of analyzing construct validity such as the MTMM Matrix as described in Campbell & Fiske (1959).
In confirmatory factor analysis, the researcher first develops a hypothesis about what factors they believe are underlying the measures used (e.g., "Depression" being the factor underlying the Beck Depression Inventory and the Hamilton Rating Scale for Depression) and may impose constraints on the model based on these a priori hypotheses. By imposing these constraints, the researcher is forcing the model to be consistent with their theory. For example, if it is posited that there are two factors accounting for the covariance in the measures, and that these factors are unrelated to each other, the researcher can create a model where the correlation between factor A and factor B is constrained to zero. Model fit measures could then be obtained to assess how well the proposed model captured the covariance between all the items or measures in the model. If the constraints the researcher has imposed on the model are inconsistent with the sample data, then the results of statistical tests of model fit will indicate a poor fit, and the model will be rejected. If the fit is poor, it may be due to some items measuring multiple factors. It might also be that some items within a factor are more related to each other than others.
For some applications, the requirement of "zero loadings" (for indicators not supposed to load on a certain factor) has been regarded as too strict. A newly developed analysis method, "exploratory structural equation modeling", specifies hypotheses about the relation between observed indicators and their supposed primary latent factors while allowing for estimation of loadings with other latent factors as well.
In confirmatory factor analysis, researchers are typically interested in studying the degree to which responses on a p x 1 vector of observable random variables can be used to assign a value to one or more unobserved variable(s) . The investigation is largely accomplished by estimating and evaluating the loading of each item used to tap aspects of the unobserved latent variable. That is, y[i] is the vector of observed responses predicted by the unobserved latent variable , which is defined as:
,
where is the p x 1 vector of observed random variables, are the unobserved latent variables and is a p x k matrix with k equal to the number of latent variables. Since, are imperfect measures of , the model also consists of error, . Estimates in the maximum likelihood (ML) case generated by iteratively minimizing the fit function,
where is the variance-covariance matrix implied by the proposed factor analysis model and is the observed variance-covariance matrix. That is, values are found for free model parameters that minimize the difference between the model-implied variance-covariance matrix and observed variance-covariance matrix.