Note that in the two examples you gave R is the same size, *but the sign was flipped* (in the second case R is negative). The p value is the odds that the R value is different from zero. From Clauset et al., pg. 19:
"The basic idea behind the likelihood ratio test is to compute the likelihood of
the data under two competing distributions. The one with the higher likelihood is
then the better fit. Alternatively one can calculate the ratio of the two likelihoods, or
equivalently the logarithm R of the ratio, which is positive or negative depending on
which distribution is better, or zero in the event of a tie.
The sign of the log likelihood ratio alone, however, will not definitively indicate
which model is the better fit because, like other quantities, it is subject to statistical
fluctuation. If its true value, meaning its expected value over many independent data
sets drawn from the same distribution, is close to zero, then the fluctuations could
change the sign of the ratio and hence the results of the test cannot be trusted. In
order to make a firm choice between distributions we need a log likelihood ratio that
is sufficiently positive or negative that it could not plausibly be the result of a chance
fluctuation from a true result that is close to zero.
To make a quantitative judgment about whether the observed value of R is sufficiently far from zero, we need to know the size of the expected fluctuations, i.e., we
need to know the standard deviation σ on R. This we can estimate from our data
using a method proposed by Vuong [63]. This method gives a p-value that tells us
whether the observed sign of R is statistically significant. If this p-value is small (say
p < 0.1) then it is unlikely that the observed sign is a chance result of fluctuations
and the sign is a reliable indicator of which model is the better fit to the data. If p
is large on the other hand, the sign is not reliable and the test does not favor either
model over the other. It is one of the advantages of this approach that it can tell us
not only which of two hypotheses is favored, but also when the data are insufficient to
favor either of them"