Discriminant validity


In psychology, discriminant validity tests whether concepts or measurements that are not supposed to be related are actually unrelated.
Campbell and Fiske introduced the concept of discriminant validity within their discussion on evaluating test validity. They stressed the importance of using both discriminant and convergent validation techniques when assessing new tests. A successful evaluation of discriminant validity shows that a test of a concept is not highly correlated with other tests designed to measure theoretically different concepts.
In showing that two scales do not correlate, it is necessary to correct for attenuation in the correlation due to measurement error. It is possible to calculate the extent to which the two scales overlap by using the following formula where is correlation between x and y, is the reliability of x, and is the reliability of y:
Although there is no standard value for discriminant validity, a result less than 0.85 suggests that discriminant validity likely exists between the two scales. A result greater than 0.85, however, suggests that the two constructs overlap greatly and they are likely measuring the same thing, and therefore, discriminant validity between them cannot be claimed.
Consider researchers developing a new scale designed to measure narcissism. They may want to show discriminant validity with a scale measuring self-esteem. Narcissism and self-esteem are theoretically different concepts, and therefore it is important that the researchers show that their new scale measures narcissism and not simply self-esteem.
First, the average inter-item correlations within and between the two scales can be calculated:
The correction for attenuation formula can then be applied:
Since 0.607 is less than 0.85, it can be concluded that discriminant validity exists between the scale measuring narcissism and the scale measuring self-esteem. The two scales measure theoretically different constructs.
Recommended approaches to test for discriminant validity on the construct level are AVE-SE comparisons and the assessment of the HTMT ration. Simulation tests reveal that the former performs poorly for variance-based structural equation models, e.g. PLS, but well for covariance-based SEM, e.g. Amos, and the latter performs well for both types of SEM. Voorhees et al. recommend combining both methods for covariance-based SEM with a HTMT cutoff of 0.85. A recommended approach to test for discriminant validity on the item level is exploratory factor analysis.