On Aug 1, 2014, at 12:03 AM, Ben Goodrich <
goodri...@gmail.com> wrote:
> On Thursday, July 31, 2014 11:08:54 PM UTC-4, Bob Carpenter wrote:
> Not for me! I still want to know why the Rhats for lp__ are
> usually so much higher and whether it could be problematic for
> the parameter estimates if lp__ estimates don't seem to have
> converged according to Rhat.
>
> Perhaps, but we already have examples where the Rhat is close to 1 for everything, including lp__, despite the chains not having converged.
I know that Rhat near 1 is necessary for convergence, but
not sufficient.
The question is really whether Rhat not having converged has
dire implications for the other parameters, which do have Rhat
near 1.
> To me, I think the bigger question is whether it makes sense to make a decision about a chain-level phenomenon (convergence) from the margins?
We can reject chain-level convergence of a vector
if one of the elements hasn't converged, right?
> A related question is that since the Markov Chain operates in the unconstrained space, does it make sense to evaluate convergence in the constrained space?
I don't see why not. It's just a transform of the parameters.
Just like X^2 and I know that makes sense to monitor.
> The lp__ is at least a function of all the parameters, and it is the log-posterior with respect to the parameters in the unconstrained space.
Right. Does that mean we should always run long enough that
Rhat for lp__ is close to 1?
> The Rhat makes more sense when you really care about one thing (which is presumably not lp__) and want to know how much is variance is inflated by having only a given number of draws.
I understand that, too. We could invent all kinds of crazy functions
of parameters that wouldn't converge. I'm just asking if lp__ is
one of them or whether its nonconvergence signals problems for the chain
as a whole.
- Bob