I assume that you have multivariate dependent variable y_ij, where i=1,...,n and j=1,...p. I think it is easiest first to think how CV would work as WAIC is approximation of CV. There's two options.
The first options is the pure M-open approach (see
http://dx.doi.org/10.1214/12-SS102 for discussion of M-open etc.) where you compute the llpd only for the observed outputs. For example, if p=2 the collection of lppd_i's contain mix of terms p_post(-i)(y_i1), p_post(-i)(y_i2), p_post(-i)(y_i1,y_i2) depending which observations y_ij are missing. As same observations are missing for all models, obtained lppd_CV is ok for model comparison.
The second option is to use mix of M-open and M-completed/closed approaches. For example, if p=2 the collection of lppd_i's contain the terms p_post(-i)(y_i1,y_i2), where in case of missing data we integrate over the missing data distribution using the imputation model (which may be problematic if the imputation algorithm is not consistent as Mike warned).
In the case of measurement error, I assume that you have all observations and it's just that your model is slightly more complicated.
WAIC can be used in these cases, too. The first option is simpler to implement.
Cody> "Secondly, WAIC, like other metrics, seems to be derived under assumptions of independent errors. Some people have found that measures like AIC and BIC perform reasonably in model comparison in spite of violating these assumptions (using simulated data). Would it be insane to include a WAIC comparison in a paper comparing spatial models?"
If WAIC is derived from cross-validation it does not need the assumption of independent errors. The prediction task consists of a collection of independent marginal predictions (1a). It depends on your decision problem, whether this the correct prediction task or should you care more about the joint predictions (1b). Although WAIC and CV are asymptotically equal, they assume slightly different prediction tasks: WAIC) interested in the predictions only at the observed covariate values (2a), CV) interested in the predictions also at the not yet observed covariate values (2b). If your answer is 1a+2a, then WAIC is sane choice although it seems that it can break sometimes when using very flexible models (like Gaussian process or Markov random field spatial models). See the slides
http://www.lce.hut.fi/~ave/slides_trondheim.pdf (we are writing a paper about this and further developments).
Mike> "Actually AIC is an approximation to WAIC, which is an approximation to the KL divergence between the predictive posterior predictive distribution and the true data distribution (assuming one exists). It's really not as assumption of independent errors for WAIC, but rather an assumption that each datum is really two independent data (as in cross validation)."
cross-validation assumes exhangeability, not independence
To summarise: if your choice is between WAIC or some other *IC, use WAIC. If you can use CV, it is a more robust choice.