I'm looking for guidance on legitimate interpretations of the elpd. Briefly: can the elpd difference between two models be interpreted as the log odds ratio of the posterior model probabilities?Example:Model 1 elpd_loo = -50Model 2 elpd_loo = -48abs difference: 2From these results, could one state that Model 2 "describes the data" (in the sense of expected prediction error) about 7 times better than model 1 (i.e. exp(2)), on average? Similarly, if one converted the elpd difference to base 2 (2 / log(2) = 2.89), could one say that the average expected information gain of Model 2 relative to Model 1 is 2.89 bits?
I am seeking ways to describe the results of loo such that they could be more easily understood or communicated to non-mathematical audiences. Apologies if my interpretations are way off base (I am a member of that non-mathematical audience).
--
You received this message because you are subscribed to the Google Groups "Stan users mailing list" group.
To unsubscribe from this group and stop receiving emails from it, send an email to stan-users+...@googlegroups.com.
To post to this group, send email to stan-...@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
--
You received this message because you are subscribed to a topic in the Google Groups "Stan users mailing list" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/stan-users/oUIKmpDGzFw/unsubscribe.
To unsubscribe from this group and all its topics, send an email to stan-users+...@googlegroups.com.
"Comparing the models on PSIS-LOO reveals an estimated difference in elpd of 10.2 (with a standard error of 5.1) in favor of Model A. "
So in my application I basically state that one model is preferred according to the LOO elpd estimate. Given that I'm comparing one model with and one model without an interaction term, and there's a pretty obvious interaction just from plotting the data, I think the model comparison results for my application are uncontroversial.
More generally though, I think it would be of great practical value if Aki and co-authors could include an example of converting elpd to a more intuitive scale for some given application (as stated in Aki's second reply, above).
Hi Daniel,Sorry for the late reply.I'm afraid I didn't state much more than what is written at the bottom of p. 20 of Aki et al's paper (v5):"Comparing the models on PSIS-LOO reveals an estimated difference in elpd of 10.2 (with a standard error of 5.1) in favor of Model A. "
So in my application I basically state that one model is preferred according to the LOO elpd estimate. Given that I'm comparing one model with and one model without an interaction term, and there's a pretty obvious interaction just from plotting the data, I think the model comparison results for my application are uncontroversial.
More generally though, I think it would be of great practical value if Aki and co-authors could include an example of converting elpd to a more intuitive scale for some given application (as stated in Aki's second reply, above).
Remember it is the difference that is important, the absolute magnitude is more a function of the number of data points than anything else.
Avi
References: