Hi Kyle, Guido -
absolute values of DIC are not all that informative - they are highly dependent on the number of subjects, trials, and the task - conditions, difficulty, number of parameters, etc. So it is difficult to compare values between experiments. Also, while relative DIC values (for the same dataset) are meant to be used for model comparison, it is known that the standard form of DIC (reported by PyMC) is a little biased toward models with complexity, though there are alternative forms that are less biased - e.g. see Plummer 2008, where he uses "popt" to more appropriately penalize complexity.
DIC values between alternative models can vary a lot more than e.g. AIC or BIC, because the DIC is computed over the entire posterior distribution rather than just at the ML value. The deviance is a function of probability density, so if the posterior is very peaked, the density can be > 1 and hence the DIC is negative (as in many of Guido's cases). It may be surprising that some of the values for alternative models of the same dataset are massively different, but this can occur if the posterior is much more spread out across a range of parameter values.
On the other hand, pD itself (the term in DIC used to penalize complexity) can also be negative. This is because standard DIC assumes that the mean of the posterior is a good estimator, i.e. an optimistic measure of the degree to which the model fits the data, where the difference between the (mean deviance across the posterior) and the (deviance at the mean value) is meant to indicate the degree to which we may have overfit the data by using the mean as the estimator. But the mean is not always a good estimator. (The easiest way to see this is if the posterior is bimodal, the mean between two modes may have zero likelihood, and hence the deviance at the mean value would be very large, leading to negative pD (and potentially negative DIC). Bimodality is not the only case that this can occur, though.
So in general, for more complex models than simple linear models, the DIC should be used with caution. One could still use it if they check some of the above issues, but I think it should be supplemented with posterior predictive checks, or using the Bayes Factor. (There are a few tricks to get the bayes factor from MCMC, though this is not currently supported directly in HDDM). The BF is not immune to problems either, as it can be sensitive to the priors, but informative priors can help a lot (e.g. from other data sets or from a training sample within the current data set).
Michael