# permutations where statistic is greater than or equal to observed
--------------------------------------------------------------------------------
# permutations
rather than
1 + #permutations where ...
---------------------------------
1 + # permutations
the latter is the correct thing, in that the true labelling is one of
the possible permutations. For more details, see
"Nonparametric Permutation Tests for Functional Neuroimaging: A Primer
with Examples."
http://www.fil.ion.ucl.ac.uk/spm/doc/papers/NicholsHolmes.pdf
f
> --
> You received this message because you are subscribed to the Google Groups
> "Princeton MVPA Toolbox for Matlab" group.
> To post to this group, send email to mvpa-t...@googlegroups.com.
> To unsubscribe from this group, send email to
> mvpa-toolbox...@googlegroups.com.
> For more options, visit this group at
> http://groups.google.com/group/mvpa-toolbox?hl=en.
>
a) do permute_test and wavestrapper_results perform the exact same
analysis? if not, then maybe it's not a surprise/problem that they differ.
b) 1000 iterations isn't that many. perhaps the results would stabilize
after more iterations?
g
> --
> You received this message because you are subscribed to the Google
> Groups "Princeton MVPA Toolbox for Matlab" group.
> To post to this group, send email to mvpa-t...@googlegroups.com.
> To unsubscribe from this group, send email to
> mvpa-toolbox...@googlegroups.com.
> For more options, visit this group at
> http://groups.google.com/group/mvpa-toolbox?hl=en.
--
---
Greg Detre
cell: 617 642 3902
email: gr...@gregdetre.co.uk
web: http://www.gregdetre.co.uk
0
-- = 0
1000
1
-- = 0.0009
1001
f
That was my thought. I had wondered about this till I read the paper
and asked a couple of other statisticians. But I'd be interested in
any sources that indicate otherwise.
f
if your real result scores better than all 1000 scrambled null results,
then p = 0 is the right answer for it to give, isn't it? can you explain
that would be indicative of a bug?
g
"Given exchangeability under the null hypothesis, the observed data is
equally likely to have arisen from any possible labeling. Hence, the
statistics associated with each of the possible labeling are also
equally likely. Thus, we have the permutation (or randomization)
distribution of our statistic: the permutation distribution is the
sampling distribution of the statistic under the null hypothesis,
given the data observed. Under the null hypothesis, the observed
statistic is randomly chosen from the set of statistics corresponding
to all possible relabelings. This gives us a way to formalize our
"surprise" at an outcome: the probability of an outcome as or more
extreme than the one observed, the P-value, is the proportion of
statistic values in the permutation distribution greater or equal to
that observed.
***
The actual labeling used in the experiment is one of the possible
labelings, so if the observed statistic is the largest of the
permutation distribution, the P-value is 1/N, where N is the number of
possible labelings of the initial randomization scheme.
***
"
f
Correct, my (1000+1) was in case you already had the value for 1000
permutations.
f
> Not a bug, just the incorrect permutation test procedure. I'm going to
> cite the Nichols&Holmes paper to spare everyone having to look for the
> passage :) (page 4)
Let me insert my .1 of non-scientific rumble.
I think that I am placing some different meaning into word 'likely' in
following:
> "Given exchangeability under the null hypothesis, the OBSERVED DATA is
> equally LIKELY to have arisen from any possible labeling. Hence, the
> STATISTICS associated with each of the possible labeling are also
> equally LIKELY.
although indeed, under H0, observed data is as likely as any other, its
statistic is not (therefore we have some non-uniform distribution of
statistics usually). Hence incorporating it into the random sample test
(+1 in enumerator) and accounting for it in the 'sample size' (+1 in
denominator) might provide tiny but unnecessary conservative bias
(just consider the case when you don't know that # of permutations is
20, like in the authors' paper, then you assess on 20 permutations and
get your p=(1+1)/(20+1)=0.095 instead of true 1/20=0.05).
So far I see +1 just as a safety bias to avoid 0s in
p-statistics instead of a simple mathematician-unfriendly max:
p = max(p_estimate, MC simulation "resolution")
= max((number_of samples>alpha)/number_of_permutations, 1/number_of_permutations)
= max((number_of samples>alpha), 1)/number_of_permutations
instead of suggested conservative
((number_of samples>alpha) + 1) /(number of permutations +1)
or am I very wrong?
--
.-.
=------------------------------ /v\ ----------------------------=
Keep in touch // \\ (yoh@|www.)onerussian.com
Yaroslav Halchenko /( )\ ICQ#: 60653192
Linux User ^^-^^ [175555]
I should have been clearer: the only thing that matters is the 1 in
the numerator (which is what the paper passage alludes to). I add one
in the denominator because I don't want to have to remember whether I
did 9999 or 10000 permutations :)
f
> I add one
> in the denominator because I don't want to have to remember whether I
> did 9999 or 10000 permutations :)
;-)
in any case, indeed, +1 seems to be just a guarding term which doesn't
matter much whenever reasonable number of permutations is used.
Actually, that's a really good way of looking at it, and I think I now
understand what you were getting at. If you could try all the
permutations, the true order of examples would certainly be one of the
possibilities. I'm not sure always adding it in is necessarily ad-hoc,
but it could mean the results are a little bit conservative. I have
something under review comparing p-values obtained with permutation
and analytical tests, as the number of permutations increases, and the
results seem to bear that.
f
> something under review comparing p-values obtained with permutation
> and analytical tests, as the number of permutations increases, and the
> results seem to bear that.
cool -- is preprint available for general (or not so general) public ;)?