Hi Jorge,
the GOF test for colext in our vignette is wrong: it always indicates a fitting model. We will correct this.
It is my understanding that no GOF test can be conducted directly for a binary response. The deviance and other measures of fit are uninformative about fit in the case of a binary response (this can be read somewhere in the GLM bible by McCullagh & Nelder for instance). One has to aggregate the response in some way before the usual GOF techniques can be applied.
What I did in the context of a dynamic occupancy model fitted in BUGS to binary data with three secondary reps was to aggregate the data to detection frequencies and conducted a Bayesian posterior predictive check of GOF to this binomial variant of the response, based on a chisquare discrepancy. (Note that I still fitted the model to the binary response though.) This indicated my model did not fit; hence, that appeared like an improvement ;)
I hope that something similar can be done for a likelihood analysis in unmarked. We have to check this out.
Kind regards ----- Marc
--
You received this message because you are subscribed to the Google Groups "unmarked" group.
To unsubscribe from this group and stop receiving emails from it, send an email to unmarked+u...@googlegroups.com.
For more options, visit https://groups.google.com/groups/opt_out.
Brooks, S. P., E. A., Catchpole, and B. J. T. Morgan. 2000. Bayesian animal survival estimation. Statistical Science 15:357–376.
McCullagh, P., and J. A. Nelder. 1983. Generalized linear models. Chapman and Hall.
Sillett, T. S., R. B. Chandler, J. A. Royal, M. Kery, and S. A. Morrison. 2012. Hierarchical distance sampling models to estimate population size and habitat-specific abundance of an island endemic. Ecological Applications 22:1997–2006.
Dear Eric,
thank you for remining me that in „my own“ vignette, there is a counter example to my claim that the GOF test we give there does not work ... To be honest, I had forgotten this. In a couple of other examples with simulated data for a binary response, the test failed to indicate a non-fitting model. I don’t understand why it does show lack of fit in our example, because from my reading of the literature (for instance some parts in McCullagh and Nelder), we can’t do any GOF test directly for a binary response, but we always have to aggregate the response first. I will check this in our case and report back sometime.
Couple of more remarks:
(1) GOF for binary response: the key thing to me seems to be whether your response is strictly binary (i.e., only 0 or 1) or whether it is some aggregation of binary numbers. If you have counts with values other than strictly zeroes and ones, then Chisquare and similar GOF tests should be OK. In the Island jay paper (Sillett et al. 2012), the response was a vector containing the counts of birds observed per distance class, and so this should be OK with our test. If you have a logistic regression (i.e., binomial GLM) with non-binary data (i.e., where the sample size or binomial index, N, is greater than 1 for some or most responses), then the test should also be OK, because then you have naturally „aggregated“ data (remember the binomial is the sum of N Bernoullis). However, it is my understanding that with binary responses, no discrepancy measure can be directly used to inform about lack of fit.
(2) Residuals: residuals have an important place in model criticism also in GLMs, and I like to think about the hierarchical models that we fit using unmarked (or BUGS, JAGS etc) as 2 or more nested GLMs. So of course, there should be a way how we can compute residuals and inspect them to see whether they truly look like a starry sky or whether there is some remaining structure perhaps related to some covariates. I don’t have much experience in this in the context of GLMs, but I have long thought that I should explore residual checking in these HMs much more.
(3) Quantifying the magnitude of lack of fit: Let’s assume you fit a hierarchial GLM with a reponse other than binary and do a parametric boostrap GOF test using some discrepancy measure such as Chisquare. Then, rather than using the parametric boostrap GOF to simply give you a “yes-no“ answer to the question of whether the model fits, it may be more interesting to give you an indication of how far your model is from a model that fits. I think that one could perhaps compute some „lack of fit statistic“ by dividing the GOF measure for your model by that of a perfectly fitting model (which would be the mean of the bootstrap distribution). ---- In the capture-recapture literature (e.g., CJS model), people have long quantified the lack of fit of a model using a c-hat statistics and often inflate their SEs by the square root of that to account for „overdispersion“ (assuming that all lack of fit is some sort of unstructured noise rather due to a structural failure of the model). It appears to me that if this is a sensible thing to do for CJS models, then perhaps something like that could be done in the kinds of HMs we fit in unmarked as well ?
Kind regards ---- Marc