Danielle (sorry in advance, for the long-winded answer),
In my previous post, I mentioned one way of proceeding to decide on
which c-hat to use when you check c-hat of each sub-global model, when
the global model is too complex to be fit à la Burnham and Anderson.
When you mentioned that you had several c-hat, I assumed that you
obtained them from such sub-global models (sorry if I misread). Yes, it
is difficult (computer-intensive) to implement if you have several
"sub-global" models. Another approach is to use the c-hat of the top-
ranking model.
If you used a "dredging approach", did you get these multiple c-hats
from each model considered or did you dredge several times (i.e.,
different sets of candidate models)? Fitting a limited number of models
supported by biological hypotheses is harder to do than using an
automated approach such as dredging, but in my opinion the former is
superior (but others may disagree).
Sensitivity of model selection to c-hat:
Regarding the sensitivity of model selection to c-hat, this is a
general exercise that you can do as soon as you have overdispersion. If
the ranking of the models change substantially when you increase c-hat
from 1 to the observed value, this suggests that there is a lot of
model selection uncertainty.
AIC vs AICc:
If you treat the number of sites as the number of observations (>
1536), then in your case, both will yield very similar values.
Adjusting model selection and inference for overdispersion:
Yes, in each model, the adjustment is sqrt(c-hat) * SE. You can do it
manually by using the equations in Burnham and Anderson (2002) or use
different functions in AICcmodavg with the c.hat argument (i.e.,
?aictab, ?modavgPred). I suggest you check up on Burnham and Anderson
(2002) or Anderson (2008: Model-based inference in the life sciences:
a primer on evidence. Springer, New York, NY, USA.) which address these
topics.
Marc
--
____________________________________
Marc J. Mazerolle
Département des sciences du bois et de la forêt
2405 rue de la Terrasse
Université Laval
Québec, Québec G1V 0A6, Canada
-------- Message initial --------
De: Danielle Rappaport <
daniell...@gmail.com>
Reply-to:
unma...@googlegroups.com
À:
unma...@googlegroups.com
Objet: Re: [unmarked] occu and goodness of fit (mb.gof.test)
Date: Fri, 26 Oct 2018 08:18:07 -0400
Marc,
I'm grateful for your assistance (and patience in providing multiple responses).
That all makes sense. However, to be sure that I understand your suggestion regarding the influence of the c-hat: you say that for every instance where I can't evaluate the goodness of fit performance of the global model, I estimate the c-hat of the "sub-global" models and see how the corresponding AIC, and thus model ranking, changes with the value of the c-hat. 1) Must I calculate the c-hat for ALL of the sub-global models? This would be computationally unviable for me considering that I have arrived at my model set through a dredging procedure that yields thousands of possible models. Is it justifiable to reassess the model ranking/AIC in light of c-hat values just for the best-ranking models (delta AIC<2)? Also, to confirm, I am to inflate the standard error by sqrt(c-hat), correct? 2) Is there a built-in function within unmarked or AICcmodavg to programmatically perform the process described above--inflate the variance of the top-ranking sub-global models by using their c-hat scores and then re-evaluate model ranking? Apologies if this is in the literature; I just couldn't find it. 3) Lastly, it just came on my radar that I may want to opt for AICc instead of AIC; however, I don't reckon my n (>1536) to be small and therefore don't think I need a sample size bias correction. That said, I read that regardless of sample size, AICc is more general and is thus often used in place of AIC. Do you have any recommendations? I already calculated AIC and would prefer to stick with that criterion if there is no added benefit of a sample size bias correction in my case.
Thanks a million, and I will try my best not to pester you with further questions,
Danielle
On Thu, Oct 25, 2018 at 8:21 PM Marc J. Mazerolle <
marc.ma...@sbf.ulaval.ca> wrote:
> Danielle,
>
> In their book, Burnham and Anderson (2002: Model selection and multimodel inference: a practical information-theoretic approach. Springer., section 6.5.1, p. 306) recommend that if you cannot run a single global model, to estimate the c-hat of different "sub-global" models and use the smallest estimate. Is this what you have been doing? Thus, in your case, I would use the smallest c-hat among your models and conduct multimodel inference. Regarding your question on the influence of the c-hat, you can check it easily by changing the c-hat and seeing how the model ranking changes with increasing values of c-hat.
>
> Best,
>
> Marc
>
> ____________________
> Marc J. Mazerolle
> Département des sciences du bois et de la forêt
> 2405 rue de la Terrasse
> Université Laval
> Québec, Québec G1V OA6, Canada
> Tel:
(418) 656-2131 ext. 7120
> Email:
marc.ma...@sbf.ulaval.ca
> De :
unma...@googlegroups.com <
unma...@googlegroups.com> de la part de Danielle Rappaport <
daniell...@gmail.com>
> Envoyé : 25 octobre 2018 15:48
> À : unmarked
> Objet : Re: [unmarked] occu and goodness of fit (mb.gof.test)
>
> Marc,
> Thanks for your speedy reply and help.
> There are a small subset of top-ranked models (delta AIC < 2 & c-hat < 3). In terms of selecting the one upon which I should advance my inferences, do you advise I select the model with the lowest AIC or the lowest c-hat score (they don't always correspond)? I'm not sure whether the difference in c-hat is negligible or not (e.g. 2.4 vs 2.5 vs 2.6).
> Best,
> Danielle
>
>
> On Wednesday, October 24, 2018 at 8:55:03 PM UTC-4, Marc Mazerolle wrote:
> > Hi Danielle,
> >
> > if your goodness-of-fit test indicates lack of fit (P < alpha), it means that the data and model "disagree". The difficult part is determining the source of the lack of fit. If your c-hat estimate is not too large (say < 3, but hopefully closer to 1), the disagreement is probably due to data varying slightly more than expected for the distribution at hand. In this case, then you could use your c-hat estimate to inflate the variances and conduct (conservative) inferences. However, the disagreement could also be due to having a lousy model to begin with (lack of fit due to wrong model structure), and this lousy model structure would yield a large c-hat. In your case, I would pursue with the c-hat = 2.4. To know how large a c-hat is too large would require a small simulation study.
> >
> > Best,
> >
> > Marc
> >
> > ____________________
> > Marc J. Mazerolle
> > Département des sciences du bois et de la forêt
> > 2405 rue de la Terrasse
> > Université Laval
> > Québec, Québec G1V OA6, Canada
> > Tel:
(418) 656-2131 ext. 7120
> > Email:
marc.ma...@sbf.ulaval.ca