I am developing a species distribution model for a rare, data-scarce species. To explore the effect of sample prevalence, I fitted several models with different presence–absence ratios by randomly sampling pseudo-absences while keeping presences fixed. After choosing priors for parameters and hyperparameters and fitting the models with INLA, I evaluated their performance. To check whether predicted probabilities are ecologically realistic, I plotted a reliability (calibration) diagram.
The plot suggests miscalibration. I have not found studies that use INLA and then explicitly recalibrate predicted probabilities, as is commonly done for some machine-learning models. In your experience, is post-hoc calibration (e.g., logistic recalibration/Platt scaling, isotonic or beta calibration) advisable in this context, or would it add bias? Alternatively, is adjusting only the intercept to account for the discrepancy between sample and true prevalence sufficient?
Thank you in advance,
Lola R