Dear Sara,
I am not sure how much discussion has been had on this forum regarding
the estimation of marginal likelihoods for the purpose of computing
Bayes factors. However my understanding is that the method of
estimating Bayes factors implemented in Tracer (using the harmonic
mean of posterior likelihoods to estimate the marginal likelihood) is
so problematic as to be essentially useless. Of course I have used it
myself in previous publications, but as far as I can tell most serious
Bayesian statisticians now laugh at its use. My understanding is that
the estimator has infinite variance and so in practice the uncertainty
in the estimate of the marginal likelihood is so great that any result
you get is practically useless. The error estimates of the marginal
likelihoods produced by bootstrapping are almost certainly
unreliable.
I have been very tempted to suggest that this estimator is removed
from Tracer for these reasons. Luckily, some of the BEAST development
team are working on alternative methods of estimating Bayes factors
such as thermodynamic integration, which have much better statistical
properties. In the mean time I would personally avoid the use of the
harmonic mean estimator.
With regards to your particular questions here are a few thoughts:
(1) Comparing strict versus relaxed clocks:
The comparison between lognormal relaxed and strict is *relatively*
easy. Use a lognormal relaxed clock first. If there is no appreciable
probably mass near zero in the marginal posterior distribution of
ucld.stdev then you can't use a strict clock. However if the marginal
distribution of ucld.stdev extends down to (abuts) zero, then the data
can't reject a strict clock.
(2) Constant size versus Exponential growth:
Use exponential growth first. If the marginal posterior distribution
of the growthRate includes 0, then your data is compatible with
constant size. You must ensure that the operator on the
exponential.growthRate is a randomWalkOperator and not a scaleOperator
for this "test" to be valid.
FINALLY: the important thing about model choice is the sensitivity of
the estimated *parameter of interest* to changes in the model and
prior. So in many respects its more important to identify which
aspects of the modeling have an impact on the *answer you care about*
than to find the "right" model.
Cheers
Alexei