Auxiliary variables and congeniality in multiple imputation

Meng’s concept of congeniality in multiple imputation (MI) is I think a tricky one (for me anyway!). Loosely speaking congeniality is about whether the imputation and analysis models make different assumptions about the data. Meng gave a definition in his 1994 paper, but I prefer the one given in a more recent paper by Xie and Meng, which is what I and Rachael Hughes used in our paper this year on different methods of combining bootstrapping with MI. In words (see the papers for the same in equations) it is that there exists a Bayesian model for the data such that:

  • given complete/full data, the posterior mean of the parameter of interest matches the point estimate given by fitting our analysis model of interest to that data, and the posterior variance matches the variance estimator calculated by our analysis model fit.
  • the conditional distribution of the missing data given the observed in this Bayesian model matches that used by our imputation model.

If they are congenial and the models are correctly specified, Rubin’s variance estimator is (asymptotically) unbiased for the true repeated sampling variance of the MI point estimator(s).

One of the potentially useful features of MI are that we can include variables in the imputation stage which we then don’t use in the analysis model. Including such auxiliary variables in the imputation model can increase the likelihood that the MAR assumption holds when the auxiliary variable is associated with the probability of missingness, and can increase efficiency according to how strongly it is correlated with the variable(s) being imputed. A nice paper (among many) on the potential of including auxiliary variables in MI is Hardt et al 2012. In this post, I’ll consider whether including auxiliary variables in the imputation model leads to uncongeniality. The post was prompted following a discussion earlier in the year with my colleague Paul von Hippel.

Read more

Convergence plots for smcfcs in R

The smcfcs package in R imputes missing values of covariates compatibly (congenially) with the user’s specified outcome or substantive model. Just like the regular chained equations (fully conditional specification) multiple imputation method, smcfcs is an iterative procedure, and users should check that they have used enough iterations for the process to have (hopefully) converged to its stationary distribution. The smcfcs package returns a matrix with the parameter estimates of the outcome/substantive model from each imputed dataset and iteration within these. But it requires the user to figure out how to appropriately plot these.

Read more

What does correlation in a Bland-Altman plot mean?

The Bland-Altman plot is a very popular approach for analysing data from a method agreement study, which I was teaching students about today. We have measurements of a sample of subjects using one measurement technique or method, and a second measurement on each, taken using a new technique or method. The objective is to see how closely the measurements from the two methods agree. If they are very similar, we could use the new method, which may be cheaper, easier or less invasive to use, rather than the old method. The Bland-Altman plot plots the pairwise differences between the measurements against their average. Sometimes one sees a correlation between the pair-wise differences and averages. What is the interpretation of such a correlation?

Read more