Hi Christopher,
Thank you kindly for the quick and clear response. I hope your patience extends far enough to answer some additional questions :).
My first has to do with the particulars of the four trial state changes. I would like to implement the exact algorithm in OxCal, if possible, so I would like to implement those, too. With what probability is each step trial state chosen relative to the others? If there are 99 t_i's is an update on g attempted 1/100 times on average (excluding the shifts and expansions/contractions). Sticking to the four trial state changes, to guarantee detailed balance it is necessary that the proposal distribution be symmetric (well, that's not quite right: using symmetric proposal distributions is sufficient for achieving detailed balance with typical sampling approaches), which is easy to maintain for the first three types of trial state changes, but requires some care for the fourth.
Let s be a rescaling factor and mu the current mean for the vector t. The proposal t' = mu + (t-mu)*s seems likely what one would use to do the expansion/contraction (yes? did I screw something up, or do you use a different equation?). How do you draw s to ensure that detailed balance is maintained?
My last question has to do with the trial state change for t_i. The article reads, "the acceptance of this trial move depends on the product of the KDE distribution probability given in Equation 9 for all the other events, and the likelihood associated with any direct measurements on t_i." I know that the KDE approach mixes frequentist and Bayesian reasoning, so perhaps my question is, as physicists sometimes say, not even wrong. Nevertheless, here goes: is it more accurate to say that the second component is the posterior probability of t_i given all available information aside from the KDE? Or do you envision the KDE component of the conditional density to be something like a prior for the event?
To remove the possibility of ambiguity on the preceding point, Equation 9 gives a formula for p(t_i|g,t_{-i}) (up to a constant of proportionality), where t_{-i} represents all the elements of the vector t aside from the i-th element. The only measurement that bears on t_i is a radiocarbon determination for which the fraction modern value is phi_{m,i} (m for measurement) with associated uncertainty on the fraction modern of sigma_{m,i} (i.e., there is no stratigraphic or other information to help constrain t_i further). The probability of t_i given the radiocarbon measurement is (assuming I haven't messed up any of the math -- and I don't know how to get a text symbol proportional to)
p(t_i|phi_{m,i},sigma_{m,i}) is proportional to p(phi_{m,i}|t_i,sigma_{m,i}) p(t_i)
where phi_{m,i}|t_i,sigma_{m,i} is normally distributed and depends on the calibration curve in the usual way and p(t_i) is the prior on t_i (often assumed to be an improper uniform distribution on the interval -Inf to Inf, but not necessarily so). As I interpret it, the quantity that enters into the Metropolis step for t_i is:
z_i = p(t_i|g,t_{-i}) * p(phi_{m,i}|t_i,sigma_{m,i})
Do I understand correctly? Why not the following instead?
z_i = p(t_i|g,t_{-i}) * p(phi_{m,i}|t_i,sigma_{m,i}) p(t_i)
All the best,
Mike Price