The finicky effects of measurement error: When does the bias go down? When does the bias go up?

So… the other day there was a most interesting discussion over on Twitter/X about the effects of measurement error in linear regression. If you’d like to look at how the discussion went, you can get started here:

Or here, which is where more discussion around the effects of measurement error took place:

I wanted to add my 2 cents to this as well (I mean, I am a psychometrician after all). But the more I thought about it, the more I realized that it would take a lot more than 2 or 3 posts to explain everything I wanted to explain. Which is why I’m writing this in blog format now.

BIG DISCLAIMER: Nothing that I’m about to write down here is new. In fact, this is a combination of my lecture on the effects of measurement error from my Intro to SEM class, as well as (mostly) showcasing examples in R code from the Bollen’s oh-so-classic-of-classics Structural Equations with Latent Variables . More specifically, all of this is taken from Chapter 5: The Consequences of Measurement Error, pp.154 — 167

The model

Bollen does an excellent job at describing exactly what most of us in the social sciences would like to fit as opposed to what we are actually fitting.

The previous path diagram implies the following system of structural equations:

\eta = \gamma\xi + \zeta

x = \lambda_{1}\xi + \delta

y = \lambda_{2}\eta + \epsilon

Covariance and correlation between x and y

This is the simplest case that I think we’re all familiar with. I’ll reproduce the same results as in Bollen’s Chapter 5, but he makes the assumption that \lambda_{1}=\lambda_{2}=1 . Remember that the standard assumptions of errors from Classical Test Theory also apply here (population mean of 0, uncorrelated errors, etc.)

Cov(\xi, \eta) = Cov(\xi, \gamma\xi + \zeta) = \gamma Cov(\xi, \xi) = \gamma Var(\xi)

Cov(x,y) =Cov(\xi + \delta, \eta + \epsilon) = Cov(\xi, \eta) + 0 + 0 + 0=Cov(\xi, \eta)

So the covariance of the latent variables is the same as the covariance between the observed values measured with error.

The variance is where things get interesting. Indeed, based on the model above, it follows that:

Var(x) = Var(\xi + \delta) = Var(\xi) + Var(\delta)

Var(y) = Var(\eta + \epsilon) = Var(\eta) + Var(\epsilon)

And since variances are strictly positive, then the following inequality follows trivially:

\frac{Cov(\xi, \eta)}{\sqrt{[Var(\xi)+Var(\delta)][Var(\eta)+Var(\epsilon)]}} \leq \frac{Cov(\xi, \eta)}{\sqrt{Var(\xi)Var(\eta)}}

Corr(x, y) \leq Corr(\xi, \eta)

Simple linear regression between x and y

Although I’m sure some of you are already anticipating what’s going to be in this section, I believe a few of you may be surprised.

For the simple linear regression model y = \gamma^{*}x + \zeta^{*} , we can use some of the previous results to figure out an expression for the regression coefficient \gamma^{*} , i.e., the coefficient we would obtain if we ignore measurement error. The slope for a simple linear regression can be expressed as:

\gamma^{*}=\frac{Cov(x,y)}{Var(x)}

But above we derived that Cov(x,y)=Cov(\xi, \eta)=\gamma Var(\xi)

Therefore, we can put it together with the expression for \gamma^{*} to obtain:

\gamma^{*}=\gamma \left [ \frac{Var(\xi)}{Var(x)} \right ]

So \gamma^{*} is biased downwards, depending on the reliability of x, and only equals \gamma when there is no measurement error. Now, reliability ranges from 0 to 1, so we know that ratio has to be positive and less than 1, shrinking the estimate of \gamma^{*} to 0.

So far so good… except there’s a little bit of a “sleight of hand” going on here. Which is the role that \lambda_{1}, \lambda_{2} are playing. For (\lambda_{1}, \lambda_{2}) \leq 1 AND \frac{\lambda_{2}} {\lambda_{1}} \leq 1 the shrinking towards 0 will happen. HOWEVER, if both of the previous conditions on the loadings for x and y are violated, then anything goes. The bias for the regression coefficient \gamma^{*} can go up or it can go down. Bollen himself claims in p.156: The attenuation of \gamma^{*} need not follow since \frac{\lambda_{2}} {\lambda_{1}} might exceed 1.

For a very easy example:

Now, Bollen himself admits that this would be an “unusual situation” (whatever that’s supposed to mean). But, mathematically, it is not impossible. Which means that, in the presence of measurement error, things can increase, decrease or stay the same depending on the measurement model that we are assuming relating the observed and latent variables.

Multiple regression between matrix of predictors X and dependent variable y : ANYTHING GOES.

The much more common case of multiple regression is much more finicky, because we now need to account for the interrelationship of the measurement error across the various predictors. This follows from the identity in Bollen equation (5.25) where he related the matrix of regression coefficients from the latent variables (measured without error) to the matrix of coefficients of te observed variables (measured with error):

\mathbf{\Gamma^{*'}} =  \mathbf{\Sigma}^{-1}_{xx'} \mathbf{\Sigma}_{x\xi'} \mathbf{\Gamma'}

So the coefficients of the observed variables, measured with error, are partly a function of the covariance matrix of the predictors themselves (with error) TIMES the covariance between the observed and latent predictors TIMES the matrix of regression coefficients for the latent variables. As I’m sure you can imagine, we’re in an anything goes type scenario, because depending on the \mathbf{\Sigma}^{-1}_{xx'} matrix and the \mathbf{\Sigma}_{x\xi'} matrix, the coefficients in \mathbf{\Gamma^{*'}} could be anything when compared to the coefficients in \mathbf{\Gamma'}

Bollen presents a very interesting example. Say you’re in a situation with 3 predictors and one dependent variable. But out of the 3 predictors, only 1 of them is measured with error. What would happen to the regression coefficients of the rest of the equation? Now the model (in equation form) looks like this:

x_{1} = \xi_{1} + \delta

x_{2} = \xi_{2}

x_{3} = \xi_{3}

\eta = \gamma_{1}^{*}x_{1} + \gamma_{2}^{*} \xi_{2} + \gamma_{3}^{*} \xi_{3} +\nu^{*}

Now look at what happens if we plug in some numbers:

Yes, we see the expected attenuation of the regression coefficient in the presence of measurement error. But look at the effect that that the measurement error in one predictor had in the other predictors, even when measurement error is UNcorrelated ….the coefficients measured without error become inflated. Which is described in equation (5.32) in Bollen:

plim(\hat{\gamma}^{*}_{2}) = \gamma_{2} + b_{\xi_{1} \xi_{2} \cdot x_{1} \xi_{3} \cdots \xi_{q} } \gamma_{1}

So the coefficient \hat{\gamma}^{*}_{2} depends partly on the coefficient for the predictors measured WITHOUT error (that would be \gamma_{2} for \xi_{2} and \gamma_{1} for \xi_{1} ) as well as the relationship between \xi_{1} and \xi_{2} , captured in b . Bollen (1989) concludes (p.163):

The direction of asymptotic bias is difficult to predict, since it depends on the magnitude and sign of the second term on the right hand side.

If you would like to consider a case where all the observed variables have measurement errors with different reliabilities. Look at the following:

So now the predictor with the highest reliability, x_{1} exhibits an inflation of its multiple regression coefficient, and the predictors with the lowest reliability exhibit downward biases.

A lot of our intuitions (and I’m including myself here) regarding the effect that measurement error has on the associations within variables comes mostly from what measurement error does to the variances in a covariance-correlation setting. HOWEVER, without understanding the measurement model that relates the (faulty) observables to the (true) latent variables, we cannot ascertain that the bias for the coefficients will always go down. True, I’m willing to concede that in the majority of cases that we encounter in empirical research, the bias is (i) probably towards 0 more often than not; and (ii) usually, at least one other thing would be biased towards 0. But we need the measurement model to understand where the bias may go.

I’m of the idea that the world we study is infinitely more complicated than what we can imagine. And if we can “imagine” situations where measurement error can bias coefficients upwards as opposed to downwards, I’d say it stands to reason to think that there will be cases in real life where measurement error acts contrary to what our (not-entirely-correct) intuition says.

Anyhoo… my 2 cents ¯\_(ツ)_/¯